~ [ source navigation ] ~ [ diff markup ] ~ [ identifier search ] ~

TOMOYO Linux Cross Reference
Linux/include/linux/kmsan.h

Version: ~ [ linux-6.11.5 ] ~ [ linux-6.10.14 ] ~ [ linux-6.9.12 ] ~ [ linux-6.8.12 ] ~ [ linux-6.7.12 ] ~ [ linux-6.6.58 ] ~ [ linux-6.5.13 ] ~ [ linux-6.4.16 ] ~ [ linux-6.3.13 ] ~ [ linux-6.2.16 ] ~ [ linux-6.1.114 ] ~ [ linux-6.0.19 ] ~ [ linux-5.19.17 ] ~ [ linux-5.18.19 ] ~ [ linux-5.17.15 ] ~ [ linux-5.16.20 ] ~ [ linux-5.15.169 ] ~ [ linux-5.14.21 ] ~ [ linux-5.13.19 ] ~ [ linux-5.12.19 ] ~ [ linux-5.11.22 ] ~ [ linux-5.10.228 ] ~ [ linux-5.9.16 ] ~ [ linux-5.8.18 ] ~ [ linux-5.7.19 ] ~ [ linux-5.6.19 ] ~ [ linux-5.5.19 ] ~ [ linux-5.4.284 ] ~ [ linux-5.3.18 ] ~ [ linux-5.2.21 ] ~ [ linux-5.1.21 ] ~ [ linux-5.0.21 ] ~ [ linux-4.20.17 ] ~ [ linux-4.19.322 ] ~ [ linux-4.18.20 ] ~ [ linux-4.17.19 ] ~ [ linux-4.16.18 ] ~ [ linux-4.15.18 ] ~ [ linux-4.14.336 ] ~ [ linux-4.13.16 ] ~ [ linux-4.12.14 ] ~ [ linux-4.11.12 ] ~ [ linux-4.10.17 ] ~ [ linux-4.9.337 ] ~ [ linux-4.4.302 ] ~ [ linux-3.10.108 ] ~ [ linux-2.6.32.71 ] ~ [ linux-2.6.0 ] ~ [ linux-2.4.37.11 ] ~ [ unix-v6-master ] ~ [ ccs-tools-1.8.9 ] ~ [ policy-sample ] ~
Architecture: ~ [ i386 ] ~ [ alpha ] ~ [ m68k ] ~ [ mips ] ~ [ ppc ] ~ [ sparc ] ~ [ sparc64 ] ~

  1 /* SPDX-License-Identifier: GPL-2.0 */
  2 /*
  3  * KMSAN API for subsystems.
  4  *
  5  * Copyright (C) 2017-2022 Google LLC
  6  * Author: Alexander Potapenko <glider@google.com>
  7  *
  8  */
  9 #ifndef _LINUX_KMSAN_H
 10 #define _LINUX_KMSAN_H
 11 
 12 #include <linux/dma-direction.h>
 13 #include <linux/gfp.h>
 14 #include <linux/kmsan-checks.h>
 15 #include <linux/types.h>
 16 
 17 struct page;
 18 struct kmem_cache;
 19 struct task_struct;
 20 struct scatterlist;
 21 struct urb;
 22 
 23 #ifdef CONFIG_KMSAN
 24 
 25 /**
 26  * kmsan_task_create() - Initialize KMSAN state for the task.
 27  * @task: task to initialize.
 28  */
 29 void kmsan_task_create(struct task_struct *task);
 30 
 31 /**
 32  * kmsan_task_exit() - Notify KMSAN that a task has exited.
 33  * @task: task about to finish.
 34  */
 35 void kmsan_task_exit(struct task_struct *task);
 36 
 37 /**
 38  * kmsan_init_shadow() - Initialize KMSAN shadow at boot time.
 39  *
 40  * Allocate and initialize KMSAN metadata for early allocations.
 41  */
 42 void __init kmsan_init_shadow(void);
 43 
 44 /**
 45  * kmsan_init_runtime() - Initialize KMSAN state and enable KMSAN.
 46  */
 47 void __init kmsan_init_runtime(void);
 48 
 49 /**
 50  * kmsan_memblock_free_pages() - handle freeing of memblock pages.
 51  * @page:       struct page to free.
 52  * @order:      order of @page.
 53  *
 54  * Freed pages are either returned to buddy allocator or held back to be used
 55  * as metadata pages.
 56  */
 57 bool __init __must_check kmsan_memblock_free_pages(struct page *page,
 58                                                    unsigned int order);
 59 
 60 /**
 61  * kmsan_alloc_page() - Notify KMSAN about an alloc_pages() call.
 62  * @page:  struct page pointer returned by alloc_pages().
 63  * @order: order of allocated struct page.
 64  * @flags: GFP flags used by alloc_pages()
 65  *
 66  * KMSAN marks 1<<@order pages starting at @page as uninitialized, unless
 67  * @flags contain __GFP_ZERO.
 68  */
 69 void kmsan_alloc_page(struct page *page, unsigned int order, gfp_t flags);
 70 
 71 /**
 72  * kmsan_free_page() - Notify KMSAN about a free_pages() call.
 73  * @page:  struct page pointer passed to free_pages().
 74  * @order: order of deallocated struct page.
 75  *
 76  * KMSAN marks freed memory as uninitialized.
 77  */
 78 void kmsan_free_page(struct page *page, unsigned int order);
 79 
 80 /**
 81  * kmsan_copy_page_meta() - Copy KMSAN metadata between two pages.
 82  * @dst: destination page.
 83  * @src: source page.
 84  *
 85  * KMSAN copies the contents of metadata pages for @src into the metadata pages
 86  * for @dst. If @dst has no associated metadata pages, nothing happens.
 87  * If @src has no associated metadata pages, @dst metadata pages are unpoisoned.
 88  */
 89 void kmsan_copy_page_meta(struct page *dst, struct page *src);
 90 
 91 /**
 92  * kmsan_slab_alloc() - Notify KMSAN about a slab allocation.
 93  * @s:      slab cache the object belongs to.
 94  * @object: object pointer.
 95  * @flags:  GFP flags passed to the allocator.
 96  *
 97  * Depending on cache flags and GFP flags, KMSAN sets up the metadata of the
 98  * newly created object, marking it as initialized or uninitialized.
 99  */
100 void kmsan_slab_alloc(struct kmem_cache *s, void *object, gfp_t flags);
101 
102 /**
103  * kmsan_slab_free() - Notify KMSAN about a slab deallocation.
104  * @s:      slab cache the object belongs to.
105  * @object: object pointer.
106  *
107  * KMSAN marks the freed object as uninitialized.
108  */
109 void kmsan_slab_free(struct kmem_cache *s, void *object);
110 
111 /**
112  * kmsan_kmalloc_large() - Notify KMSAN about a large slab allocation.
113  * @ptr:   object pointer.
114  * @size:  object size.
115  * @flags: GFP flags passed to the allocator.
116  *
117  * Similar to kmsan_slab_alloc(), but for large allocations.
118  */
119 void kmsan_kmalloc_large(const void *ptr, size_t size, gfp_t flags);
120 
121 /**
122  * kmsan_kfree_large() - Notify KMSAN about a large slab deallocation.
123  * @ptr: object pointer.
124  *
125  * Similar to kmsan_slab_free(), but for large allocations.
126  */
127 void kmsan_kfree_large(const void *ptr);
128 
129 /**
130  * kmsan_map_kernel_range_noflush() - Notify KMSAN about a vmap.
131  * @start:      start of vmapped range.
132  * @end:        end of vmapped range.
133  * @prot:       page protection flags used for vmap.
134  * @pages:      array of pages.
135  * @page_shift: page_shift passed to vmap_range_noflush().
136  *
137  * KMSAN maps shadow and origin pages of @pages into contiguous ranges in
138  * vmalloc metadata address range. Returns 0 on success, callers must check
139  * for non-zero return value.
140  */
141 int __must_check kmsan_vmap_pages_range_noflush(unsigned long start,
142                                                 unsigned long end,
143                                                 pgprot_t prot,
144                                                 struct page **pages,
145                                                 unsigned int page_shift);
146 
147 /**
148  * kmsan_vunmap_kernel_range_noflush() - Notify KMSAN about a vunmap.
149  * @start: start of vunmapped range.
150  * @end:   end of vunmapped range.
151  *
152  * KMSAN unmaps the contiguous metadata ranges created by
153  * kmsan_map_kernel_range_noflush().
154  */
155 void kmsan_vunmap_range_noflush(unsigned long start, unsigned long end);
156 
157 /**
158  * kmsan_ioremap_page_range() - Notify KMSAN about a ioremap_page_range() call.
159  * @addr:       range start.
160  * @end:        range end.
161  * @phys_addr:  physical range start.
162  * @prot:       page protection flags used for ioremap_page_range().
163  * @page_shift: page_shift argument passed to vmap_range_noflush().
164  *
165  * KMSAN creates new metadata pages for the physical pages mapped into the
166  * virtual memory. Returns 0 on success, callers must check for non-zero return
167  * value.
168  */
169 int __must_check kmsan_ioremap_page_range(unsigned long addr, unsigned long end,
170                                           phys_addr_t phys_addr, pgprot_t prot,
171                                           unsigned int page_shift);
172 
173 /**
174  * kmsan_iounmap_page_range() - Notify KMSAN about a iounmap_page_range() call.
175  * @start: range start.
176  * @end:   range end.
177  *
178  * KMSAN unmaps the metadata pages for the given range and, unlike for
179  * vunmap_page_range(), also deallocates them.
180  */
181 void kmsan_iounmap_page_range(unsigned long start, unsigned long end);
182 
183 /**
184  * kmsan_handle_dma() - Handle a DMA data transfer.
185  * @page:   first page of the buffer.
186  * @offset: offset of the buffer within the first page.
187  * @size:   buffer size.
188  * @dir:    one of possible dma_data_direction values.
189  *
190  * Depending on @direction, KMSAN:
191  * * checks the buffer, if it is copied to device;
192  * * initializes the buffer, if it is copied from device;
193  * * does both, if this is a DMA_BIDIRECTIONAL transfer.
194  */
195 void kmsan_handle_dma(struct page *page, size_t offset, size_t size,
196                       enum dma_data_direction dir);
197 
198 /**
199  * kmsan_handle_dma_sg() - Handle a DMA transfer using scatterlist.
200  * @sg:    scatterlist holding DMA buffers.
201  * @nents: number of scatterlist entries.
202  * @dir:   one of possible dma_data_direction values.
203  *
204  * Depending on @direction, KMSAN:
205  * * checks the buffers in the scatterlist, if they are copied to device;
206  * * initializes the buffers, if they are copied from device;
207  * * does both, if this is a DMA_BIDIRECTIONAL transfer.
208  */
209 void kmsan_handle_dma_sg(struct scatterlist *sg, int nents,
210                          enum dma_data_direction dir);
211 
212 /**
213  * kmsan_handle_urb() - Handle a USB data transfer.
214  * @urb:    struct urb pointer.
215  * @is_out: data transfer direction (true means output to hardware).
216  *
217  * If @is_out is true, KMSAN checks the transfer buffer of @urb. Otherwise,
218  * KMSAN initializes the transfer buffer.
219  */
220 void kmsan_handle_urb(const struct urb *urb, bool is_out);
221 
222 /**
223  * kmsan_unpoison_entry_regs() - Handle pt_regs in low-level entry code.
224  * @regs:       struct pt_regs pointer received from assembly code.
225  *
226  * KMSAN unpoisons the contents of the passed pt_regs, preventing potential
227  * false positive reports. Unlike kmsan_unpoison_memory(),
228  * kmsan_unpoison_entry_regs() can be called from the regions where
229  * kmsan_in_runtime() returns true, which is the case in early entry code.
230  */
231 void kmsan_unpoison_entry_regs(const struct pt_regs *regs);
232 
233 /**
234  * kmsan_get_metadata() - Return a pointer to KMSAN shadow or origins.
235  * @addr:      kernel address.
236  * @is_origin: whether to return origins or shadow.
237  *
238  * Return NULL if metadata cannot be found.
239  */
240 void *kmsan_get_metadata(void *addr, bool is_origin);
241 
242 /**
243  * kmsan_enable_current(): Enable KMSAN for the current task.
244  *
245  * Each kmsan_enable_current() current call must be preceded by a
246  * kmsan_disable_current() call. These call pairs may be nested.
247  */
248 void kmsan_enable_current(void);
249 
250 /**
251  * kmsan_disable_current(): Disable KMSAN for the current task.
252  *
253  * Each kmsan_disable_current() current call must be followed by a
254  * kmsan_enable_current() call. These call pairs may be nested.
255  */
256 void kmsan_disable_current(void);
257 
258 /**
259  * memset_no_sanitize_memory(): Fill memory without KMSAN instrumentation.
260  * @s: address of kernel memory to fill.
261  * @c: constant byte to fill the memory with.
262  * @n: number of bytes to fill.
263  *
264  * This is like memset(), but without KMSAN instrumentation.
265  */
266 static inline void *memset_no_sanitize_memory(void *s, int c, size_t n)
267 {
268         return __memset(s, c, n);
269 }
270 
271 extern bool kmsan_enabled;
272 extern int panic_on_kmsan;
273 
274 /*
275  * KMSAN performs a lot of consistency checks that are currently enabled by
276  * default. BUG_ON is normally discouraged in the kernel, unless used for
277  * debugging, but KMSAN itself is a debugging tool, so it makes little sense to
278  * recover if something goes wrong.
279  */
280 #define KMSAN_WARN_ON(cond)                                           \
281         ({                                                            \
282                 const bool __cond = WARN_ON(cond);                    \
283                 if (unlikely(__cond)) {                               \
284                         WRITE_ONCE(kmsan_enabled, false);             \
285                         if (panic_on_kmsan) {                         \
286                                 /* Can't call panic() here because */ \
287                                 /* of uaccess checks. */              \
288                                 BUG();                                \
289                         }                                             \
290                 }                                                     \
291                 __cond;                                               \
292         })
293 
294 #else
295 
296 static inline void kmsan_init_shadow(void)
297 {
298 }
299 
300 static inline void kmsan_init_runtime(void)
301 {
302 }
303 
304 static inline bool __must_check kmsan_memblock_free_pages(struct page *page,
305                                                           unsigned int order)
306 {
307         return true;
308 }
309 
310 static inline void kmsan_task_create(struct task_struct *task)
311 {
312 }
313 
314 static inline void kmsan_task_exit(struct task_struct *task)
315 {
316 }
317 
318 static inline void kmsan_alloc_page(struct page *page, unsigned int order,
319                                     gfp_t flags)
320 {
321 }
322 
323 static inline void kmsan_free_page(struct page *page, unsigned int order)
324 {
325 }
326 
327 static inline void kmsan_copy_page_meta(struct page *dst, struct page *src)
328 {
329 }
330 
331 static inline void kmsan_slab_alloc(struct kmem_cache *s, void *object,
332                                     gfp_t flags)
333 {
334 }
335 
336 static inline void kmsan_slab_free(struct kmem_cache *s, void *object)
337 {
338 }
339 
340 static inline void kmsan_kmalloc_large(const void *ptr, size_t size,
341                                        gfp_t flags)
342 {
343 }
344 
345 static inline void kmsan_kfree_large(const void *ptr)
346 {
347 }
348 
349 static inline int __must_check kmsan_vmap_pages_range_noflush(
350         unsigned long start, unsigned long end, pgprot_t prot,
351         struct page **pages, unsigned int page_shift)
352 {
353         return 0;
354 }
355 
356 static inline void kmsan_vunmap_range_noflush(unsigned long start,
357                                               unsigned long end)
358 {
359 }
360 
361 static inline int __must_check kmsan_ioremap_page_range(unsigned long start,
362                                                         unsigned long end,
363                                                         phys_addr_t phys_addr,
364                                                         pgprot_t prot,
365                                                         unsigned int page_shift)
366 {
367         return 0;
368 }
369 
370 static inline void kmsan_iounmap_page_range(unsigned long start,
371                                             unsigned long end)
372 {
373 }
374 
375 static inline void kmsan_handle_dma(struct page *page, size_t offset,
376                                     size_t size, enum dma_data_direction dir)
377 {
378 }
379 
380 static inline void kmsan_handle_dma_sg(struct scatterlist *sg, int nents,
381                                        enum dma_data_direction dir)
382 {
383 }
384 
385 static inline void kmsan_handle_urb(const struct urb *urb, bool is_out)
386 {
387 }
388 
389 static inline void kmsan_unpoison_entry_regs(const struct pt_regs *regs)
390 {
391 }
392 
393 static inline void kmsan_enable_current(void)
394 {
395 }
396 
397 static inline void kmsan_disable_current(void)
398 {
399 }
400 
401 static inline void *memset_no_sanitize_memory(void *s, int c, size_t n)
402 {
403         return memset(s, c, n);
404 }
405 
406 #define KMSAN_WARN_ON WARN_ON
407 
408 #endif
409 
410 #endif /* _LINUX_KMSAN_H */
411 

~ [ source navigation ] ~ [ diff markup ] ~ [ identifier search ] ~

kernel.org | git.kernel.org | LWN.net | Project Home | SVN repository | Mail admin

Linux® is a registered trademark of Linus Torvalds in the United States and other countries.
TOMOYO® is a registered trademark of NTT DATA CORPORATION.

sflogo.php