~ [ source navigation ] ~ [ diff markup ] ~ [ identifier search ] ~

TOMOYO Linux Cross Reference
Linux/arch/riscv/include/asm/pgtable-64.h

Version: ~ [ linux-6.11.5 ] ~ [ linux-6.10.14 ] ~ [ linux-6.9.12 ] ~ [ linux-6.8.12 ] ~ [ linux-6.7.12 ] ~ [ linux-6.6.58 ] ~ [ linux-6.5.13 ] ~ [ linux-6.4.16 ] ~ [ linux-6.3.13 ] ~ [ linux-6.2.16 ] ~ [ linux-6.1.114 ] ~ [ linux-6.0.19 ] ~ [ linux-5.19.17 ] ~ [ linux-5.18.19 ] ~ [ linux-5.17.15 ] ~ [ linux-5.16.20 ] ~ [ linux-5.15.169 ] ~ [ linux-5.14.21 ] ~ [ linux-5.13.19 ] ~ [ linux-5.12.19 ] ~ [ linux-5.11.22 ] ~ [ linux-5.10.228 ] ~ [ linux-5.9.16 ] ~ [ linux-5.8.18 ] ~ [ linux-5.7.19 ] ~ [ linux-5.6.19 ] ~ [ linux-5.5.19 ] ~ [ linux-5.4.284 ] ~ [ linux-5.3.18 ] ~ [ linux-5.2.21 ] ~ [ linux-5.1.21 ] ~ [ linux-5.0.21 ] ~ [ linux-4.20.17 ] ~ [ linux-4.19.322 ] ~ [ linux-4.18.20 ] ~ [ linux-4.17.19 ] ~ [ linux-4.16.18 ] ~ [ linux-4.15.18 ] ~ [ linux-4.14.336 ] ~ [ linux-4.13.16 ] ~ [ linux-4.12.14 ] ~ [ linux-4.11.12 ] ~ [ linux-4.10.17 ] ~ [ linux-4.9.337 ] ~ [ linux-4.4.302 ] ~ [ linux-3.10.108 ] ~ [ linux-2.6.32.71 ] ~ [ linux-2.6.0 ] ~ [ linux-2.4.37.11 ] ~ [ unix-v6-master ] ~ [ ccs-tools-1.8.9 ] ~ [ policy-sample ] ~
Architecture: ~ [ i386 ] ~ [ alpha ] ~ [ m68k ] ~ [ mips ] ~ [ ppc ] ~ [ sparc ] ~ [ sparc64 ] ~

  1 /* SPDX-License-Identifier: GPL-2.0-only */
  2 /*
  3  * Copyright (C) 2012 Regents of the University of California
  4  */
  5 
  6 #ifndef _ASM_RISCV_PGTABLE_64_H
  7 #define _ASM_RISCV_PGTABLE_64_H
  8 
  9 #include <linux/bits.h>
 10 #include <linux/const.h>
 11 #include <asm/errata_list.h>
 12 
 13 extern bool pgtable_l4_enabled;
 14 extern bool pgtable_l5_enabled;
 15 
 16 #define PGDIR_SHIFT_L3  30
 17 #define PGDIR_SHIFT_L4  39
 18 #define PGDIR_SHIFT_L5  48
 19 #define PGDIR_SHIFT     (pgtable_l5_enabled ? PGDIR_SHIFT_L5 : \
 20                 (pgtable_l4_enabled ? PGDIR_SHIFT_L4 : PGDIR_SHIFT_L3))
 21 /* Size of region mapped by a page global directory */
 22 #define PGDIR_SIZE      (_AC(1, UL) << PGDIR_SHIFT)
 23 #define PGDIR_MASK      (~(PGDIR_SIZE - 1))
 24 
 25 /* p4d is folded into pgd in case of 4-level page table */
 26 #define P4D_SHIFT_L3   30
 27 #define P4D_SHIFT_L4   39
 28 #define P4D_SHIFT_L5   39
 29 #define P4D_SHIFT      (pgtable_l5_enabled ? P4D_SHIFT_L5 : \
 30                 (pgtable_l4_enabled ? P4D_SHIFT_L4 : P4D_SHIFT_L3))
 31 #define P4D_SIZE       (_AC(1, UL) << P4D_SHIFT)
 32 #define P4D_MASK       (~(P4D_SIZE - 1))
 33 
 34 /* pud is folded into pgd in case of 3-level page table */
 35 #define PUD_SHIFT      30
 36 #define PUD_SIZE       (_AC(1, UL) << PUD_SHIFT)
 37 #define PUD_MASK       (~(PUD_SIZE - 1))
 38 
 39 #define PMD_SHIFT       21
 40 /* Size of region mapped by a page middle directory */
 41 #define PMD_SIZE        (_AC(1, UL) << PMD_SHIFT)
 42 #define PMD_MASK        (~(PMD_SIZE - 1))
 43 
 44 /* Page 4th Directory entry */
 45 typedef struct {
 46         unsigned long p4d;
 47 } p4d_t;
 48 
 49 #define p4d_val(x)      ((x).p4d)
 50 #define __p4d(x)        ((p4d_t) { (x) })
 51 #define PTRS_PER_P4D    (PAGE_SIZE / sizeof(p4d_t))
 52 
 53 /* Page Upper Directory entry */
 54 typedef struct {
 55         unsigned long pud;
 56 } pud_t;
 57 
 58 #define pud_val(x)      ((x).pud)
 59 #define __pud(x)        ((pud_t) { (x) })
 60 #define PTRS_PER_PUD    (PAGE_SIZE / sizeof(pud_t))
 61 
 62 /* Page Middle Directory entry */
 63 typedef struct {
 64         unsigned long pmd;
 65 } pmd_t;
 66 
 67 #define pmd_val(x)      ((x).pmd)
 68 #define __pmd(x)        ((pmd_t) { (x) })
 69 
 70 #define PTRS_PER_PMD    (PAGE_SIZE / sizeof(pmd_t))
 71 
 72 /*
 73  * rv64 PTE format:
 74  * | 63 | 62 61 | 60 54 | 53  10 | 9             8 | 7 | 6 | 5 | 4 | 3 | 2 | 1 | 0
 75  *   N      MT     RSV    PFN      reserved for SW   D   A   G   U   X   W   R   V
 76  */
 77 #define _PAGE_PFN_MASK  GENMASK(53, 10)
 78 
 79 /*
 80  * [63] Svnapot definitions:
 81  * 0 Svnapot disabled
 82  * 1 Svnapot enabled
 83  */
 84 #define _PAGE_NAPOT_SHIFT       63
 85 #define _PAGE_NAPOT             BIT(_PAGE_NAPOT_SHIFT)
 86 /*
 87  * Only 64KB (order 4) napot ptes supported.
 88  */
 89 #define NAPOT_CONT_ORDER_BASE 4
 90 enum napot_cont_order {
 91         NAPOT_CONT64KB_ORDER = NAPOT_CONT_ORDER_BASE,
 92         NAPOT_ORDER_MAX,
 93 };
 94 
 95 #define for_each_napot_order(order)                                             \
 96         for (order = NAPOT_CONT_ORDER_BASE; order < NAPOT_ORDER_MAX; order++)
 97 #define for_each_napot_order_rev(order)                                         \
 98         for (order = NAPOT_ORDER_MAX - 1;                                       \
 99              order >= NAPOT_CONT_ORDER_BASE; order--)
100 #define napot_cont_order(val)   (__builtin_ctzl((val.pte >> _PAGE_PFN_SHIFT) << 1))
101 
102 #define napot_cont_shift(order) ((order) + PAGE_SHIFT)
103 #define napot_cont_size(order)  BIT(napot_cont_shift(order))
104 #define napot_cont_mask(order)  (~(napot_cont_size(order) - 1UL))
105 #define napot_pte_num(order)    BIT(order)
106 
107 #ifdef CONFIG_RISCV_ISA_SVNAPOT
108 #define HUGE_MAX_HSTATE         (2 + (NAPOT_ORDER_MAX - NAPOT_CONT_ORDER_BASE))
109 #else
110 #define HUGE_MAX_HSTATE         2
111 #endif
112 
113 /*
114  * [62:61] Svpbmt Memory Type definitions:
115  *
116  *  00 - PMA    Normal Cacheable, No change to implied PMA memory type
117  *  01 - NC     Non-cacheable, idempotent, weakly-ordered Main Memory
118  *  10 - IO     Non-cacheable, non-idempotent, strongly-ordered I/O memory
119  *  11 - Rsvd   Reserved for future standard use
120  */
121 #define _PAGE_NOCACHE_SVPBMT    (1UL << 61)
122 #define _PAGE_IO_SVPBMT         (1UL << 62)
123 #define _PAGE_MTMASK_SVPBMT     (_PAGE_NOCACHE_SVPBMT | _PAGE_IO_SVPBMT)
124 
125 /*
126  * [63:59] T-Head Memory Type definitions:
127  * bit[63] SO - Strong Order
128  * bit[62] C - Cacheable
129  * bit[61] B - Bufferable
130  * bit[60] SH - Shareable
131  * bit[59] Sec - Trustable
132  * 00110 - NC   Weakly-ordered, Non-cacheable, Bufferable, Shareable, Non-trustable
133  * 01110 - PMA  Weakly-ordered, Cacheable, Bufferable, Shareable, Non-trustable
134  * 10010 - IO   Strongly-ordered, Non-cacheable, Non-bufferable, Shareable, Non-trustable
135  */
136 #define _PAGE_PMA_THEAD         ((1UL << 62) | (1UL << 61) | (1UL << 60))
137 #define _PAGE_NOCACHE_THEAD     ((1UL << 61) | (1UL << 60))
138 #define _PAGE_IO_THEAD          ((1UL << 63) | (1UL << 60))
139 #define _PAGE_MTMASK_THEAD      (_PAGE_PMA_THEAD | _PAGE_IO_THEAD | (1UL << 59))
140 
141 static inline u64 riscv_page_mtmask(void)
142 {
143         u64 val;
144 
145         ALT_SVPBMT(val, _PAGE_MTMASK);
146         return val;
147 }
148 
149 static inline u64 riscv_page_nocache(void)
150 {
151         u64 val;
152 
153         ALT_SVPBMT(val, _PAGE_NOCACHE);
154         return val;
155 }
156 
157 static inline u64 riscv_page_io(void)
158 {
159         u64 val;
160 
161         ALT_SVPBMT(val, _PAGE_IO);
162         return val;
163 }
164 
165 #define _PAGE_NOCACHE           riscv_page_nocache()
166 #define _PAGE_IO                riscv_page_io()
167 #define _PAGE_MTMASK            riscv_page_mtmask()
168 
169 /* Set of bits to preserve across pte_modify() */
170 #define _PAGE_CHG_MASK  (~(unsigned long)(_PAGE_PRESENT | _PAGE_READ |  \
171                                           _PAGE_WRITE | _PAGE_EXEC |    \
172                                           _PAGE_USER | _PAGE_GLOBAL |   \
173                                           _PAGE_MTMASK))
174 
175 static inline int pud_present(pud_t pud)
176 {
177         return (pud_val(pud) & _PAGE_PRESENT);
178 }
179 
180 static inline int pud_none(pud_t pud)
181 {
182         return (pud_val(pud) == 0);
183 }
184 
185 static inline int pud_bad(pud_t pud)
186 {
187         return !pud_present(pud);
188 }
189 
190 #define pud_leaf        pud_leaf
191 static inline bool pud_leaf(pud_t pud)
192 {
193         return pud_present(pud) && (pud_val(pud) & _PAGE_LEAF);
194 }
195 
196 static inline int pud_user(pud_t pud)
197 {
198         return pud_val(pud) & _PAGE_USER;
199 }
200 
201 static inline void set_pud(pud_t *pudp, pud_t pud)
202 {
203         WRITE_ONCE(*pudp, pud);
204 }
205 
206 static inline void pud_clear(pud_t *pudp)
207 {
208         set_pud(pudp, __pud(0));
209 }
210 
211 static inline pud_t pfn_pud(unsigned long pfn, pgprot_t prot)
212 {
213         return __pud((pfn << _PAGE_PFN_SHIFT) | pgprot_val(prot));
214 }
215 
216 static inline unsigned long _pud_pfn(pud_t pud)
217 {
218         return __page_val_to_pfn(pud_val(pud));
219 }
220 
221 static inline pmd_t *pud_pgtable(pud_t pud)
222 {
223         return (pmd_t *)pfn_to_virt(__page_val_to_pfn(pud_val(pud)));
224 }
225 
226 static inline struct page *pud_page(pud_t pud)
227 {
228         return pfn_to_page(__page_val_to_pfn(pud_val(pud)));
229 }
230 
231 #define mm_p4d_folded  mm_p4d_folded
232 static inline bool mm_p4d_folded(struct mm_struct *mm)
233 {
234         if (pgtable_l5_enabled)
235                 return false;
236 
237         return true;
238 }
239 
240 #define mm_pud_folded  mm_pud_folded
241 static inline bool mm_pud_folded(struct mm_struct *mm)
242 {
243         if (pgtable_l4_enabled)
244                 return false;
245 
246         return true;
247 }
248 
249 #define pmd_index(addr) (((addr) >> PMD_SHIFT) & (PTRS_PER_PMD - 1))
250 
251 static inline pmd_t pfn_pmd(unsigned long pfn, pgprot_t prot)
252 {
253         unsigned long prot_val = pgprot_val(prot);
254 
255         ALT_THEAD_PMA(prot_val);
256 
257         return __pmd((pfn << _PAGE_PFN_SHIFT) | prot_val);
258 }
259 
260 static inline unsigned long _pmd_pfn(pmd_t pmd)
261 {
262         return __page_val_to_pfn(pmd_val(pmd));
263 }
264 
265 #define mk_pmd(page, prot)    pfn_pmd(page_to_pfn(page), prot)
266 
267 #define pmd_ERROR(e) \
268         pr_err("%s:%d: bad pmd %016lx.\n", __FILE__, __LINE__, pmd_val(e))
269 
270 #define pud_ERROR(e)   \
271         pr_err("%s:%d: bad pud %016lx.\n", __FILE__, __LINE__, pud_val(e))
272 
273 #define p4d_ERROR(e)   \
274         pr_err("%s:%d: bad p4d %016lx.\n", __FILE__, __LINE__, p4d_val(e))
275 
276 static inline void set_p4d(p4d_t *p4dp, p4d_t p4d)
277 {
278         if (pgtable_l4_enabled)
279                 WRITE_ONCE(*p4dp, p4d);
280         else
281                 set_pud((pud_t *)p4dp, (pud_t){ p4d_val(p4d) });
282 }
283 
284 static inline int p4d_none(p4d_t p4d)
285 {
286         if (pgtable_l4_enabled)
287                 return (p4d_val(p4d) == 0);
288 
289         return 0;
290 }
291 
292 static inline int p4d_present(p4d_t p4d)
293 {
294         if (pgtable_l4_enabled)
295                 return (p4d_val(p4d) & _PAGE_PRESENT);
296 
297         return 1;
298 }
299 
300 static inline int p4d_bad(p4d_t p4d)
301 {
302         if (pgtable_l4_enabled)
303                 return !p4d_present(p4d);
304 
305         return 0;
306 }
307 
308 static inline void p4d_clear(p4d_t *p4d)
309 {
310         if (pgtable_l4_enabled)
311                 set_p4d(p4d, __p4d(0));
312 }
313 
314 static inline p4d_t pfn_p4d(unsigned long pfn, pgprot_t prot)
315 {
316         return __p4d((pfn << _PAGE_PFN_SHIFT) | pgprot_val(prot));
317 }
318 
319 static inline unsigned long _p4d_pfn(p4d_t p4d)
320 {
321         return __page_val_to_pfn(p4d_val(p4d));
322 }
323 
324 static inline pud_t *p4d_pgtable(p4d_t p4d)
325 {
326         if (pgtable_l4_enabled)
327                 return (pud_t *)pfn_to_virt(__page_val_to_pfn(p4d_val(p4d)));
328 
329         return (pud_t *)pud_pgtable((pud_t) { p4d_val(p4d) });
330 }
331 #define p4d_page_vaddr(p4d)     ((unsigned long)p4d_pgtable(p4d))
332 
333 static inline struct page *p4d_page(p4d_t p4d)
334 {
335         return pfn_to_page(__page_val_to_pfn(p4d_val(p4d)));
336 }
337 
338 #define pud_index(addr) (((addr) >> PUD_SHIFT) & (PTRS_PER_PUD - 1))
339 
340 #define pud_offset pud_offset
341 pud_t *pud_offset(p4d_t *p4d, unsigned long address);
342 
343 static inline void set_pgd(pgd_t *pgdp, pgd_t pgd)
344 {
345         if (pgtable_l5_enabled)
346                 WRITE_ONCE(*pgdp, pgd);
347         else
348                 set_p4d((p4d_t *)pgdp, (p4d_t){ pgd_val(pgd) });
349 }
350 
351 static inline int pgd_none(pgd_t pgd)
352 {
353         if (pgtable_l5_enabled)
354                 return (pgd_val(pgd) == 0);
355 
356         return 0;
357 }
358 
359 static inline int pgd_present(pgd_t pgd)
360 {
361         if (pgtable_l5_enabled)
362                 return (pgd_val(pgd) & _PAGE_PRESENT);
363 
364         return 1;
365 }
366 
367 static inline int pgd_bad(pgd_t pgd)
368 {
369         if (pgtable_l5_enabled)
370                 return !pgd_present(pgd);
371 
372         return 0;
373 }
374 
375 static inline void pgd_clear(pgd_t *pgd)
376 {
377         if (pgtable_l5_enabled)
378                 set_pgd(pgd, __pgd(0));
379 }
380 
381 static inline p4d_t *pgd_pgtable(pgd_t pgd)
382 {
383         if (pgtable_l5_enabled)
384                 return (p4d_t *)pfn_to_virt(__page_val_to_pfn(pgd_val(pgd)));
385 
386         return (p4d_t *)p4d_pgtable((p4d_t) { pgd_val(pgd) });
387 }
388 #define pgd_page_vaddr(pgd)     ((unsigned long)pgd_pgtable(pgd))
389 
390 static inline struct page *pgd_page(pgd_t pgd)
391 {
392         return pfn_to_page(__page_val_to_pfn(pgd_val(pgd)));
393 }
394 #define pgd_page(pgd)   pgd_page(pgd)
395 
396 #define p4d_index(addr) (((addr) >> P4D_SHIFT) & (PTRS_PER_P4D - 1))
397 
398 #define p4d_offset p4d_offset
399 p4d_t *p4d_offset(pgd_t *pgd, unsigned long address);
400 
401 #ifdef CONFIG_TRANSPARENT_HUGEPAGE
402 static inline int pte_devmap(pte_t pte);
403 static inline pte_t pmd_pte(pmd_t pmd);
404 
405 static inline int pmd_devmap(pmd_t pmd)
406 {
407         return pte_devmap(pmd_pte(pmd));
408 }
409 
410 static inline int pud_devmap(pud_t pud)
411 {
412         return 0;
413 }
414 
415 static inline int pgd_devmap(pgd_t pgd)
416 {
417         return 0;
418 }
419 #endif
420 
421 #endif /* _ASM_RISCV_PGTABLE_64_H */
422 

~ [ source navigation ] ~ [ diff markup ] ~ [ identifier search ] ~

kernel.org | git.kernel.org | LWN.net | Project Home | SVN repository | Mail admin

Linux® is a registered trademark of Linus Torvalds in the United States and other countries.
TOMOYO® is a registered trademark of NTT DATA CORPORATION.

sflogo.php