/* SPDX-License-Identifier: GPL-2.0 */ #ifndef _ASM_POWERPC_BOOK3S_32_MMU_HASH_H_ #define _ASM_POWERPC_BOOK3S_32_MMU_HASH_H_ /* * 32-bit hash table MMU support */ /* * BATs */ /* Block size masks */ #define BL_128K 0x000 #define BL_256K 0x001 #define BL_512K 0x003 #define BL_1M 0x007 #define BL_2M 0x00F #define BL_4M 0x01F #define BL_8M 0x03F #define BL_16M 0x07F #define BL_32M 0x0FF #define BL_64M 0x1FF #define BL_128M 0x3FF #define BL_256M 0x7FF /* BAT Access Protection */ #define BPP_XX 0x00 /* No access */ #define BPP_RX 0x01 /* Read only */ #define BPP_RW 0x02 /* Read/write */ #ifndef __ASSEMBLY__ /* Contort a phys_addr_t into the right format/bits for a BAT */ #ifdef CONFIG_PHYS_64BIT #define BAT_PHYS_ADDR(x) ((u32)((x & 0x00000000fffe0000ULL) | \ ((x & 0x0000000e00000000ULL) >> 24) | \ ((x & 0x0000000100000000ULL) >> 30))) #define PHYS_BAT_ADDR(x) (((u64)(x) & 0x00000000fffe0000ULL) | \ (((u64)(x) << 24) & 0x0000000e00000000ULL) | \ (((u64)(x) << 30) & 0x0000000100000000ULL)) #else #define BAT_PHYS_ADDR(x) (x) #define PHYS_BAT_ADDR(x) ((x) & 0xfffe0000) #endif struct ppc_bat { u32 batu; u32 batl; }; #endif /* !__ASSEMBLY__ */ /* * Hash table */ /* Values for PP (assumes Ks=0, Kp=1) */ #define PP_RWXX 0 /* Supervisor read/write, User none */ #define PP_RWRX 1 /* Supervisor read/write, User read */ #define PP_RWRW 2 /* Supervisor read/write, User read/write */ #define PP_RXRX 3 /* Supervisor read, User read */ /* Values for Segment Registers */ #define SR_NX 0x10000000 /* No Execute */ #define SR_KP 0x20000000 /* User key */ #define SR_KS 0x40000000 /* Supervisor key */ #ifdef __ASSEMBLY__ #include .macro uus_addi sr reg1 reg2 imm .if NUM_USER_SEGMENTS > \sr addi \reg1,\reg2,\imm .endif .endm .macro uus_mtsr sr reg1 .if NUM_USER_SEGMENTS > \sr mtsr \sr, \reg1 .endif .endm /* * This isync() shouldn't be necessary as the kernel is not excepted to run * any instruction in userspace soon after the update of segments and 'rfi' * instruction is used to return to userspace, but hash based cores * (at least G3) seem to exhibit a random behaviour when the 'isync' is not * there. 603 cores don't have this behaviour so don't do the 'isync' as it * saves several CPU cycles. */ .macro uus_isync #ifdef CONFIG_PPC_BOOK3S_604 BEGIN_MMU_FTR_SECTION isync END_MMU_FTR_SECTION_IFSET(MMU_FTR_HPTE_TABLE) #endif .endm .macro update_user_segments_by_4 tmp1 tmp2 tmp3 tmp4 uus_addi 1, \tmp2, \tmp1, 0x111 uus_addi 2, \tmp3, \tmp1, 0x222 uus_addi 3, \tmp4, \tmp1, 0x333 uus_mtsr 0, \tmp1 uus_mtsr 1, \tmp2 uus_mtsr 2, \tmp3 uus_mtsr 3, \tmp4 uus_addi 4, \tmp1, \tmp1, 0x444 uus_addi 5, \tmp2, \tmp2, 0x444 uus_addi 6, \tmp3, \tmp3, 0x444 uus_addi 7, \tmp4, \tmp4, 0x444 uus_mtsr 4, \tmp1 uus_mtsr 5, \tmp2 uus_mtsr 6, \tmp3 uus_mtsr 7, \tmp4 uus_addi 8, \tmp1, \tmp1, 0x444 uus_addi 9, \tmp2, \tmp2, 0x444 uus_addi 10, \tmp3, \tmp3, 0x444 uus_addi 11, \tmp4, \tmp4, 0x444 uus_mtsr 8, \tmp1 uus_mtsr 9, \tmp2 uus_mtsr 10, \tmp3 uus_mtsr 11, \tmp4 uus_addi 12, \tmp1, \tmp1, 0x444 uus_addi 13, \tmp2, \tmp2, 0x444 uus_addi 14, \tmp3, \tmp3, 0x444 uus_addi 15, \tmp4, \tmp4, 0x444 uus_mtsr 12, \tmp1 uus_mtsr 13, \tmp2 uus_mtsr 14, \tmp3 uus_mtsr 15, \tmp4 uus_isync .endm #else /* * This macro defines the mapping from contexts to VSIDs (virtual * segment IDs). We use a skew on both the context and the high 4 bits * of the 32-bit virtual address (the "effective segment ID") in order * to spread out the entries in the MMU hash table. Note, if this * function is changed then hash functions will have to be * changed to correspond. */ #define CTX_TO_VSID(c, id) ((((c) * (897 * 16)) + (id * 0x111)) & 0xffffff) /* * Hardware Page Table Entry * Note that the xpn and x bitfields are used only by processors that * support extended addressing; otherwise, those bits are reserved. */ struct hash_pte { unsigned long v:1; /* Entry is valid */ unsigned long vsid:24; /* Virtual segment identifier */ unsigned long h:1; /* Hash algorithm indicator */ unsigned long api:6; /* Abbreviated page index */ unsigned long rpn:20; /* Real (physical) page number */ unsigned long xpn:3; /* Real page number bits 0-2, optional */ unsigned long r:1; /* Referenced */ unsigned long c:1; /* Changed */ unsigned long w:1; /* Write-thru cache mode */ unsigned long i:1; /* Cache inhibited */ unsigned long m:1; /* Memory coherence */ unsigned long g:1; /* Guarded */ unsigned long x:1; /* Real page number bit 3, optional */ unsigned long pp:2; /* Page protection */ }; typedef struct { unsigned long id; void __user *vdso; } mm_context_t; void update_bats(void); static inline void cleanup_cpu_mmu_context(void) { } /* patch sites */ extern s32 patch__hash_page_A0, patch__hash_page_A1, patch__hash_page_A2; extern s32 patch__hash_page_B, patch__hash_page_C; extern s32 patch__flush_hash_A0, patch__flush_hash_A1, patch__flush_hash_A2; extern s32 patch__flush_hash_B; #include #include static __always_inline void update_user_segment(u32 n, u32 val) { if (n << 28 < TASK_SIZE) mtsr(val + n * 0x111, n << 28); } static __always_inline void update_user_segments(u32 val) { val &= 0xf0ffffff; update_user_segment(0, val); update_user_segment(1, val); update_user_segment(2, val); update_user_segment(3, val); update_user_segment(4, val); update_user_segment(5, val); update_user_segment(6, val); update_user_segment(7, val); update_user_segment(8, val); update_user_segment(9, val); update_user_segment(10, val); update_user_segment(11, val); update_user_segment(12, val); update_user_segment(13, val); update_user_segment(14, val); update_user_segment(15, val); } int __init find_free_bat(void); unsigned int bat_block_size(unsigned long base, unsigned long top); #endif /* !__ASSEMBLY__ */ /* We happily ignore the smaller BATs on 601, we don't actually use * those definitions on hash32 at the moment anyway */ #define mmu_virtual_psize MMU_PAGE_4K #define mmu_linear_psize MMU_PAGE_256M #endif /* _ASM_POWERPC_BOOK3S_32_MMU_HASH_H_ */