~ [ source navigation ] ~ [ diff markup ] ~ [ identifier search ] ~

TOMOYO Linux Cross Reference
Linux/arch/x86/include/asm/preempt.h

Version: ~ [ linux-6.12-rc7 ] ~ [ linux-6.11.7 ] ~ [ linux-6.10.14 ] ~ [ linux-6.9.12 ] ~ [ linux-6.8.12 ] ~ [ linux-6.7.12 ] ~ [ linux-6.6.60 ] ~ [ linux-6.5.13 ] ~ [ linux-6.4.16 ] ~ [ linux-6.3.13 ] ~ [ linux-6.2.16 ] ~ [ linux-6.1.116 ] ~ [ linux-6.0.19 ] ~ [ linux-5.19.17 ] ~ [ linux-5.18.19 ] ~ [ linux-5.17.15 ] ~ [ linux-5.16.20 ] ~ [ linux-5.15.171 ] ~ [ linux-5.14.21 ] ~ [ linux-5.13.19 ] ~ [ linux-5.12.19 ] ~ [ linux-5.11.22 ] ~ [ linux-5.10.229 ] ~ [ linux-5.9.16 ] ~ [ linux-5.8.18 ] ~ [ linux-5.7.19 ] ~ [ linux-5.6.19 ] ~ [ linux-5.5.19 ] ~ [ linux-5.4.285 ] ~ [ linux-5.3.18 ] ~ [ linux-5.2.21 ] ~ [ linux-5.1.21 ] ~ [ linux-5.0.21 ] ~ [ linux-4.20.17 ] ~ [ linux-4.19.323 ] ~ [ linux-4.18.20 ] ~ [ linux-4.17.19 ] ~ [ linux-4.16.18 ] ~ [ linux-4.15.18 ] ~ [ linux-4.14.336 ] ~ [ linux-4.13.16 ] ~ [ linux-4.12.14 ] ~ [ linux-4.11.12 ] ~ [ linux-4.10.17 ] ~ [ linux-4.9.337 ] ~ [ linux-4.4.302 ] ~ [ linux-3.10.108 ] ~ [ linux-2.6.32.71 ] ~ [ linux-2.6.0 ] ~ [ linux-2.4.37.11 ] ~ [ unix-v6-master ] ~ [ ccs-tools-1.8.12 ] ~ [ policy-sample ] ~
Architecture: ~ [ i386 ] ~ [ alpha ] ~ [ m68k ] ~ [ mips ] ~ [ ppc ] ~ [ sparc ] ~ [ sparc64 ] ~

Diff markup

Differences between /arch/x86/include/asm/preempt.h (Version linux-6.12-rc7) and /arch/i386/include/asm-i386/preempt.h (Version linux-6.9.12)


  1 /* SPDX-License-Identifier: GPL-2.0 */              1 
  2 #ifndef __ASM_PREEMPT_H                           
  3 #define __ASM_PREEMPT_H                           
  4                                                   
  5 #include <asm/rmwcc.h>                            
  6 #include <asm/percpu.h>                           
  7 #include <asm/current.h>                          
  8                                                   
  9 #include <linux/static_call_types.h>              
 10                                                   
 11 /* We use the MSB mostly because its available    
 12 #define PREEMPT_NEED_RESCHED    0x80000000        
 13                                                   
 14 /*                                                
 15  * We use the PREEMPT_NEED_RESCHED bit as an i    
 16  * that a decrement hitting 0 means we can and    
 17  */                                               
 18 #define PREEMPT_ENABLED (0 + PREEMPT_NEED_RESC    
 19                                                   
 20 /*                                                
 21  * We mask the PREEMPT_NEED_RESCHED bit so as     
 22  * that think a non-zero value indicates we ca    
 23  */                                               
 24 static __always_inline int preempt_count(void)    
 25 {                                                 
 26         return raw_cpu_read_4(pcpu_hot.preempt    
 27 }                                                 
 28                                                   
 29 static __always_inline void preempt_count_set(    
 30 {                                                 
 31         int old, new;                             
 32                                                   
 33         old = raw_cpu_read_4(pcpu_hot.preempt_    
 34         do {                                      
 35                 new = (old & PREEMPT_NEED_RESC    
 36                         (pc & ~PREEMPT_NEED_RE    
 37         } while (!raw_cpu_try_cmpxchg_4(pcpu_h    
 38 }                                                 
 39                                                   
 40 /*                                                
 41  * must be macros to avoid header recursion he    
 42  */                                               
 43 #define init_task_preempt_count(p) do { } whil    
 44                                                   
 45 #define init_idle_preempt_count(p, cpu) do { \    
 46         per_cpu(pcpu_hot.preempt_count, (cpu))    
 47 } while (0)                                       
 48                                                   
 49 /*                                                
 50  * We fold the NEED_RESCHED bit into the preem    
 51  * preempt_enable() can decrement and test for    
 52  * single instruction.                            
 53  *                                                
 54  * We invert the actual bit, so that when the     
 55  * need to resched (the bit is cleared) and ca    
 56  */                                               
 57                                                   
 58 static __always_inline void set_preempt_need_r    
 59 {                                                 
 60         raw_cpu_and_4(pcpu_hot.preempt_count,     
 61 }                                                 
 62                                                   
 63 static __always_inline void clear_preempt_need    
 64 {                                                 
 65         raw_cpu_or_4(pcpu_hot.preempt_count, P    
 66 }                                                 
 67                                                   
 68 static __always_inline bool test_preempt_need_    
 69 {                                                 
 70         return !(raw_cpu_read_4(pcpu_hot.preem    
 71 }                                                 
 72                                                   
 73 /*                                                
 74  * The various preempt_count add/sub methods      
 75  */                                               
 76                                                   
 77 static __always_inline void __preempt_count_ad    
 78 {                                                 
 79         raw_cpu_add_4(pcpu_hot.preempt_count,     
 80 }                                                 
 81                                                   
 82 static __always_inline void __preempt_count_su    
 83 {                                                 
 84         raw_cpu_add_4(pcpu_hot.preempt_count,     
 85 }                                                 
 86                                                   
 87 /*                                                
 88  * Because we keep PREEMPT_NEED_RESCHED set wh    
 89  * a decrement which hits zero means we have n    
 90  * reschedule.                                    
 91  */                                               
 92 static __always_inline bool __preempt_count_de    
 93 {                                                 
 94         return GEN_UNARY_RMWcc("decl", __my_cp    
 95                                __percpu_arg([v    
 96 }                                                 
 97                                                   
 98 /*                                                
 99  * Returns true when we need to resched and ca    
100  */                                               
101 static __always_inline bool should_resched(int    
102 {                                                 
103         return unlikely(raw_cpu_read_4(pcpu_ho    
104 }                                                 
105                                                   
106 #ifdef CONFIG_PREEMPTION                          
107                                                   
108 extern asmlinkage void preempt_schedule(void);    
109 extern asmlinkage void preempt_schedule_thunk(    
110                                                   
111 #define preempt_schedule_dynamic_enabled          
112 #define preempt_schedule_dynamic_disabled         
113                                                   
114 extern asmlinkage void preempt_schedule_notrac    
115 extern asmlinkage void preempt_schedule_notrac    
116                                                   
117 #define preempt_schedule_notrace_dynamic_enabl    
118 #define preempt_schedule_notrace_dynamic_disab    
119                                                   
120 #ifdef CONFIG_PREEMPT_DYNAMIC                     
121                                                   
122 DECLARE_STATIC_CALL(preempt_schedule, preempt_    
123                                                   
124 #define __preempt_schedule() \                    
125 do { \                                            
126         __STATIC_CALL_MOD_ADDRESSABLE(preempt_    
127         asm volatile ("call " STATIC_CALL_TRAM    
128 } while (0)                                       
129                                                   
130 DECLARE_STATIC_CALL(preempt_schedule_notrace,     
131                                                   
132 #define __preempt_schedule_notrace() \            
133 do { \                                            
134         __STATIC_CALL_MOD_ADDRESSABLE(preempt_    
135         asm volatile ("call " STATIC_CALL_TRAM    
136 } while (0)                                       
137                                                   
138 #else /* PREEMPT_DYNAMIC */                       
139                                                   
140 #define __preempt_schedule() \                    
141         asm volatile ("call preempt_schedule_t    
142                                                   
143 #define __preempt_schedule_notrace() \            
144         asm volatile ("call preempt_schedule_n    
145                                                   
146 #endif /* PREEMPT_DYNAMIC */                      
147                                                   
148 #endif /* PREEMPTION */                           
149                                                   
150 #endif /* __ASM_PREEMPT_H */                      
151                                                   

~ [ source navigation ] ~ [ diff markup ] ~ [ identifier search ] ~

kernel.org | git.kernel.org | LWN.net | Project Home | SVN repository | Mail admin

Linux® is a registered trademark of Linus Torvalds in the United States and other countries.
TOMOYO® is a registered trademark of NTT DATA CORPORATION.

sflogo.php