root/arch/arm64/include/asm/pgtable-hwdef.h

/* [<][>][^][v][top][bottom][index][help] */

INCLUDED FROM


   1 /* SPDX-License-Identifier: GPL-2.0-only */
   2 /*
   3  * Copyright (C) 2012 ARM Ltd.
   4  */
   5 #ifndef __ASM_PGTABLE_HWDEF_H
   6 #define __ASM_PGTABLE_HWDEF_H
   7 
   8 #include <asm/memory.h>
   9 
  10 /*
  11  * Number of page-table levels required to address 'va_bits' wide
  12  * address, without section mapping. We resolve the top (va_bits - PAGE_SHIFT)
  13  * bits with (PAGE_SHIFT - 3) bits at each page table level. Hence:
  14  *
  15  *  levels = DIV_ROUND_UP((va_bits - PAGE_SHIFT), (PAGE_SHIFT - 3))
  16  *
  17  * where DIV_ROUND_UP(n, d) => (((n) + (d) - 1) / (d))
  18  *
  19  * We cannot include linux/kernel.h which defines DIV_ROUND_UP here
  20  * due to build issues. So we open code DIV_ROUND_UP here:
  21  *
  22  *      ((((va_bits) - PAGE_SHIFT) + (PAGE_SHIFT - 3) - 1) / (PAGE_SHIFT - 3))
  23  *
  24  * which gets simplified as :
  25  */
  26 #define ARM64_HW_PGTABLE_LEVELS(va_bits) (((va_bits) - 4) / (PAGE_SHIFT - 3))
  27 
  28 /*
  29  * Size mapped by an entry at level n ( 0 <= n <= 3)
  30  * We map (PAGE_SHIFT - 3) at all translation levels and PAGE_SHIFT bits
  31  * in the final page. The maximum number of translation levels supported by
  32  * the architecture is 4. Hence, starting at at level n, we have further
  33  * ((4 - n) - 1) levels of translation excluding the offset within the page.
  34  * So, the total number of bits mapped by an entry at level n is :
  35  *
  36  *  ((4 - n) - 1) * (PAGE_SHIFT - 3) + PAGE_SHIFT
  37  *
  38  * Rearranging it a bit we get :
  39  *   (4 - n) * (PAGE_SHIFT - 3) + 3
  40  */
  41 #define ARM64_HW_PGTABLE_LEVEL_SHIFT(n) ((PAGE_SHIFT - 3) * (4 - (n)) + 3)
  42 
  43 #define PTRS_PER_PTE            (1 << (PAGE_SHIFT - 3))
  44 
  45 /*
  46  * PMD_SHIFT determines the size a level 2 page table entry can map.
  47  */
  48 #if CONFIG_PGTABLE_LEVELS > 2
  49 #define PMD_SHIFT               ARM64_HW_PGTABLE_LEVEL_SHIFT(2)
  50 #define PMD_SIZE                (_AC(1, UL) << PMD_SHIFT)
  51 #define PMD_MASK                (~(PMD_SIZE-1))
  52 #define PTRS_PER_PMD            PTRS_PER_PTE
  53 #endif
  54 
  55 /*
  56  * PUD_SHIFT determines the size a level 1 page table entry can map.
  57  */
  58 #if CONFIG_PGTABLE_LEVELS > 3
  59 #define PUD_SHIFT               ARM64_HW_PGTABLE_LEVEL_SHIFT(1)
  60 #define PUD_SIZE                (_AC(1, UL) << PUD_SHIFT)
  61 #define PUD_MASK                (~(PUD_SIZE-1))
  62 #define PTRS_PER_PUD            PTRS_PER_PTE
  63 #endif
  64 
  65 /*
  66  * PGDIR_SHIFT determines the size a top-level page table entry can map
  67  * (depending on the configuration, this level can be 0, 1 or 2).
  68  */
  69 #define PGDIR_SHIFT             ARM64_HW_PGTABLE_LEVEL_SHIFT(4 - CONFIG_PGTABLE_LEVELS)
  70 #define PGDIR_SIZE              (_AC(1, UL) << PGDIR_SHIFT)
  71 #define PGDIR_MASK              (~(PGDIR_SIZE-1))
  72 #define PTRS_PER_PGD            (1 << (MAX_USER_VA_BITS - PGDIR_SHIFT))
  73 
  74 /*
  75  * Section address mask and size definitions.
  76  */
  77 #define SECTION_SHIFT           PMD_SHIFT
  78 #define SECTION_SIZE            (_AC(1, UL) << SECTION_SHIFT)
  79 #define SECTION_MASK            (~(SECTION_SIZE-1))
  80 
  81 /*
  82  * Contiguous page definitions.
  83  */
  84 #ifdef CONFIG_ARM64_64K_PAGES
  85 #define CONT_PTE_SHIFT          5
  86 #define CONT_PMD_SHIFT          5
  87 #elif defined(CONFIG_ARM64_16K_PAGES)
  88 #define CONT_PTE_SHIFT          7
  89 #define CONT_PMD_SHIFT          5
  90 #else
  91 #define CONT_PTE_SHIFT          4
  92 #define CONT_PMD_SHIFT          4
  93 #endif
  94 
  95 #define CONT_PTES               (1 << CONT_PTE_SHIFT)
  96 #define CONT_PTE_SIZE           (CONT_PTES * PAGE_SIZE)
  97 #define CONT_PTE_MASK           (~(CONT_PTE_SIZE - 1))
  98 #define CONT_PMDS               (1 << CONT_PMD_SHIFT)
  99 #define CONT_PMD_SIZE           (CONT_PMDS * PMD_SIZE)
 100 #define CONT_PMD_MASK           (~(CONT_PMD_SIZE - 1))
 101 /* the the numerical offset of the PTE within a range of CONT_PTES */
 102 #define CONT_RANGE_OFFSET(addr) (((addr)>>PAGE_SHIFT)&(CONT_PTES-1))
 103 
 104 /*
 105  * Hardware page table definitions.
 106  *
 107  * Level 1 descriptor (PUD).
 108  */
 109 #define PUD_TYPE_TABLE          (_AT(pudval_t, 3) << 0)
 110 #define PUD_TABLE_BIT           (_AT(pudval_t, 1) << 1)
 111 #define PUD_TYPE_MASK           (_AT(pudval_t, 3) << 0)
 112 #define PUD_TYPE_SECT           (_AT(pudval_t, 1) << 0)
 113 
 114 /*
 115  * Level 2 descriptor (PMD).
 116  */
 117 #define PMD_TYPE_MASK           (_AT(pmdval_t, 3) << 0)
 118 #define PMD_TYPE_TABLE          (_AT(pmdval_t, 3) << 0)
 119 #define PMD_TYPE_SECT           (_AT(pmdval_t, 1) << 0)
 120 #define PMD_TABLE_BIT           (_AT(pmdval_t, 1) << 1)
 121 
 122 /*
 123  * Section
 124  */
 125 #define PMD_SECT_VALID          (_AT(pmdval_t, 1) << 0)
 126 #define PMD_SECT_USER           (_AT(pmdval_t, 1) << 6)         /* AP[1] */
 127 #define PMD_SECT_RDONLY         (_AT(pmdval_t, 1) << 7)         /* AP[2] */
 128 #define PMD_SECT_S              (_AT(pmdval_t, 3) << 8)
 129 #define PMD_SECT_AF             (_AT(pmdval_t, 1) << 10)
 130 #define PMD_SECT_NG             (_AT(pmdval_t, 1) << 11)
 131 #define PMD_SECT_CONT           (_AT(pmdval_t, 1) << 52)
 132 #define PMD_SECT_PXN            (_AT(pmdval_t, 1) << 53)
 133 #define PMD_SECT_UXN            (_AT(pmdval_t, 1) << 54)
 134 
 135 /*
 136  * AttrIndx[2:0] encoding (mapping attributes defined in the MAIR* registers).
 137  */
 138 #define PMD_ATTRINDX(t)         (_AT(pmdval_t, (t)) << 2)
 139 #define PMD_ATTRINDX_MASK       (_AT(pmdval_t, 7) << 2)
 140 
 141 /*
 142  * Level 3 descriptor (PTE).
 143  */
 144 #define PTE_VALID               (_AT(pteval_t, 1) << 0)
 145 #define PTE_TYPE_MASK           (_AT(pteval_t, 3) << 0)
 146 #define PTE_TYPE_PAGE           (_AT(pteval_t, 3) << 0)
 147 #define PTE_TABLE_BIT           (_AT(pteval_t, 1) << 1)
 148 #define PTE_USER                (_AT(pteval_t, 1) << 6)         /* AP[1] */
 149 #define PTE_RDONLY              (_AT(pteval_t, 1) << 7)         /* AP[2] */
 150 #define PTE_SHARED              (_AT(pteval_t, 3) << 8)         /* SH[1:0], inner shareable */
 151 #define PTE_AF                  (_AT(pteval_t, 1) << 10)        /* Access Flag */
 152 #define PTE_NG                  (_AT(pteval_t, 1) << 11)        /* nG */
 153 #define PTE_DBM                 (_AT(pteval_t, 1) << 51)        /* Dirty Bit Management */
 154 #define PTE_CONT                (_AT(pteval_t, 1) << 52)        /* Contiguous range */
 155 #define PTE_PXN                 (_AT(pteval_t, 1) << 53)        /* Privileged XN */
 156 #define PTE_UXN                 (_AT(pteval_t, 1) << 54)        /* User XN */
 157 #define PTE_HYP_XN              (_AT(pteval_t, 1) << 54)        /* HYP XN */
 158 
 159 #define PTE_ADDR_LOW            (((_AT(pteval_t, 1) << (48 - PAGE_SHIFT)) - 1) << PAGE_SHIFT)
 160 #ifdef CONFIG_ARM64_PA_BITS_52
 161 #define PTE_ADDR_HIGH           (_AT(pteval_t, 0xf) << 12)
 162 #define PTE_ADDR_MASK           (PTE_ADDR_LOW | PTE_ADDR_HIGH)
 163 #else
 164 #define PTE_ADDR_MASK           PTE_ADDR_LOW
 165 #endif
 166 
 167 /*
 168  * AttrIndx[2:0] encoding (mapping attributes defined in the MAIR* registers).
 169  */
 170 #define PTE_ATTRINDX(t)         (_AT(pteval_t, (t)) << 2)
 171 #define PTE_ATTRINDX_MASK       (_AT(pteval_t, 7) << 2)
 172 
 173 /*
 174  * 2nd stage PTE definitions
 175  */
 176 #define PTE_S2_RDONLY           (_AT(pteval_t, 1) << 6)   /* HAP[2:1] */
 177 #define PTE_S2_RDWR             (_AT(pteval_t, 3) << 6)   /* HAP[2:1] */
 178 #define PTE_S2_XN               (_AT(pteval_t, 2) << 53)  /* XN[1:0] */
 179 
 180 #define PMD_S2_RDONLY           (_AT(pmdval_t, 1) << 6)   /* HAP[2:1] */
 181 #define PMD_S2_RDWR             (_AT(pmdval_t, 3) << 6)   /* HAP[2:1] */
 182 #define PMD_S2_XN               (_AT(pmdval_t, 2) << 53)  /* XN[1:0] */
 183 
 184 #define PUD_S2_RDONLY           (_AT(pudval_t, 1) << 6)   /* HAP[2:1] */
 185 #define PUD_S2_RDWR             (_AT(pudval_t, 3) << 6)   /* HAP[2:1] */
 186 #define PUD_S2_XN               (_AT(pudval_t, 2) << 53)  /* XN[1:0] */
 187 
 188 /*
 189  * Memory Attribute override for Stage-2 (MemAttr[3:0])
 190  */
 191 #define PTE_S2_MEMATTR(t)       (_AT(pteval_t, (t)) << 2)
 192 #define PTE_S2_MEMATTR_MASK     (_AT(pteval_t, 0xf) << 2)
 193 
 194 /*
 195  * EL2/HYP PTE/PMD definitions
 196  */
 197 #define PMD_HYP                 PMD_SECT_USER
 198 #define PTE_HYP                 PTE_USER
 199 
 200 /*
 201  * Highest possible physical address supported.
 202  */
 203 #define PHYS_MASK_SHIFT         (CONFIG_ARM64_PA_BITS)
 204 #define PHYS_MASK               ((UL(1) << PHYS_MASK_SHIFT) - 1)
 205 
 206 #define TTBR_CNP_BIT            (UL(1) << 0)
 207 
 208 /*
 209  * TCR flags.
 210  */
 211 #define TCR_T0SZ_OFFSET         0
 212 #define TCR_T1SZ_OFFSET         16
 213 #define TCR_T0SZ(x)             ((UL(64) - (x)) << TCR_T0SZ_OFFSET)
 214 #define TCR_T1SZ(x)             ((UL(64) - (x)) << TCR_T1SZ_OFFSET)
 215 #define TCR_TxSZ(x)             (TCR_T0SZ(x) | TCR_T1SZ(x))
 216 #define TCR_TxSZ_WIDTH          6
 217 #define TCR_T0SZ_MASK           (((UL(1) << TCR_TxSZ_WIDTH) - 1) << TCR_T0SZ_OFFSET)
 218 
 219 #define TCR_EPD0_SHIFT          7
 220 #define TCR_EPD0_MASK           (UL(1) << TCR_EPD0_SHIFT)
 221 #define TCR_IRGN0_SHIFT         8
 222 #define TCR_IRGN0_MASK          (UL(3) << TCR_IRGN0_SHIFT)
 223 #define TCR_IRGN0_NC            (UL(0) << TCR_IRGN0_SHIFT)
 224 #define TCR_IRGN0_WBWA          (UL(1) << TCR_IRGN0_SHIFT)
 225 #define TCR_IRGN0_WT            (UL(2) << TCR_IRGN0_SHIFT)
 226 #define TCR_IRGN0_WBnWA         (UL(3) << TCR_IRGN0_SHIFT)
 227 
 228 #define TCR_EPD1_SHIFT          23
 229 #define TCR_EPD1_MASK           (UL(1) << TCR_EPD1_SHIFT)
 230 #define TCR_IRGN1_SHIFT         24
 231 #define TCR_IRGN1_MASK          (UL(3) << TCR_IRGN1_SHIFT)
 232 #define TCR_IRGN1_NC            (UL(0) << TCR_IRGN1_SHIFT)
 233 #define TCR_IRGN1_WBWA          (UL(1) << TCR_IRGN1_SHIFT)
 234 #define TCR_IRGN1_WT            (UL(2) << TCR_IRGN1_SHIFT)
 235 #define TCR_IRGN1_WBnWA         (UL(3) << TCR_IRGN1_SHIFT)
 236 
 237 #define TCR_IRGN_NC             (TCR_IRGN0_NC | TCR_IRGN1_NC)
 238 #define TCR_IRGN_WBWA           (TCR_IRGN0_WBWA | TCR_IRGN1_WBWA)
 239 #define TCR_IRGN_WT             (TCR_IRGN0_WT | TCR_IRGN1_WT)
 240 #define TCR_IRGN_WBnWA          (TCR_IRGN0_WBnWA | TCR_IRGN1_WBnWA)
 241 #define TCR_IRGN_MASK           (TCR_IRGN0_MASK | TCR_IRGN1_MASK)
 242 
 243 
 244 #define TCR_ORGN0_SHIFT         10
 245 #define TCR_ORGN0_MASK          (UL(3) << TCR_ORGN0_SHIFT)
 246 #define TCR_ORGN0_NC            (UL(0) << TCR_ORGN0_SHIFT)
 247 #define TCR_ORGN0_WBWA          (UL(1) << TCR_ORGN0_SHIFT)
 248 #define TCR_ORGN0_WT            (UL(2) << TCR_ORGN0_SHIFT)
 249 #define TCR_ORGN0_WBnWA         (UL(3) << TCR_ORGN0_SHIFT)
 250 
 251 #define TCR_ORGN1_SHIFT         26
 252 #define TCR_ORGN1_MASK          (UL(3) << TCR_ORGN1_SHIFT)
 253 #define TCR_ORGN1_NC            (UL(0) << TCR_ORGN1_SHIFT)
 254 #define TCR_ORGN1_WBWA          (UL(1) << TCR_ORGN1_SHIFT)
 255 #define TCR_ORGN1_WT            (UL(2) << TCR_ORGN1_SHIFT)
 256 #define TCR_ORGN1_WBnWA         (UL(3) << TCR_ORGN1_SHIFT)
 257 
 258 #define TCR_ORGN_NC             (TCR_ORGN0_NC | TCR_ORGN1_NC)
 259 #define TCR_ORGN_WBWA           (TCR_ORGN0_WBWA | TCR_ORGN1_WBWA)
 260 #define TCR_ORGN_WT             (TCR_ORGN0_WT | TCR_ORGN1_WT)
 261 #define TCR_ORGN_WBnWA          (TCR_ORGN0_WBnWA | TCR_ORGN1_WBnWA)
 262 #define TCR_ORGN_MASK           (TCR_ORGN0_MASK | TCR_ORGN1_MASK)
 263 
 264 #define TCR_SH0_SHIFT           12
 265 #define TCR_SH0_MASK            (UL(3) << TCR_SH0_SHIFT)
 266 #define TCR_SH0_INNER           (UL(3) << TCR_SH0_SHIFT)
 267 
 268 #define TCR_SH1_SHIFT           28
 269 #define TCR_SH1_MASK            (UL(3) << TCR_SH1_SHIFT)
 270 #define TCR_SH1_INNER           (UL(3) << TCR_SH1_SHIFT)
 271 #define TCR_SHARED              (TCR_SH0_INNER | TCR_SH1_INNER)
 272 
 273 #define TCR_TG0_SHIFT           14
 274 #define TCR_TG0_MASK            (UL(3) << TCR_TG0_SHIFT)
 275 #define TCR_TG0_4K              (UL(0) << TCR_TG0_SHIFT)
 276 #define TCR_TG0_64K             (UL(1) << TCR_TG0_SHIFT)
 277 #define TCR_TG0_16K             (UL(2) << TCR_TG0_SHIFT)
 278 
 279 #define TCR_TG1_SHIFT           30
 280 #define TCR_TG1_MASK            (UL(3) << TCR_TG1_SHIFT)
 281 #define TCR_TG1_16K             (UL(1) << TCR_TG1_SHIFT)
 282 #define TCR_TG1_4K              (UL(2) << TCR_TG1_SHIFT)
 283 #define TCR_TG1_64K             (UL(3) << TCR_TG1_SHIFT)
 284 
 285 #define TCR_IPS_SHIFT           32
 286 #define TCR_IPS_MASK            (UL(7) << TCR_IPS_SHIFT)
 287 #define TCR_A1                  (UL(1) << 22)
 288 #define TCR_ASID16              (UL(1) << 36)
 289 #define TCR_TBI0                (UL(1) << 37)
 290 #define TCR_TBI1                (UL(1) << 38)
 291 #define TCR_HA                  (UL(1) << 39)
 292 #define TCR_HD                  (UL(1) << 40)
 293 #define TCR_NFD0                (UL(1) << 53)
 294 #define TCR_NFD1                (UL(1) << 54)
 295 
 296 /*
 297  * TTBR.
 298  */
 299 #ifdef CONFIG_ARM64_PA_BITS_52
 300 /*
 301  * This should be GENMASK_ULL(47, 2).
 302  * TTBR_ELx[1] is RES0 in this configuration.
 303  */
 304 #define TTBR_BADDR_MASK_52      (((UL(1) << 46) - 1) << 2)
 305 #endif
 306 
 307 #ifdef CONFIG_ARM64_VA_BITS_52
 308 /* Must be at least 64-byte aligned to prevent corruption of the TTBR */
 309 #define TTBR1_BADDR_4852_OFFSET (((UL(1) << (52 - PGDIR_SHIFT)) - \
 310                                  (UL(1) << (48 - PGDIR_SHIFT))) * 8)
 311 #endif
 312 
 313 #endif

/* [<][>][^][v][top][bottom][index][help] */