~ [ source navigation ] ~ [ diff markup ] ~ [ identifier search ] ~

TOMOYO Linux Cross Reference
Linux/arch/parisc/lib/lusercopy.S

Version: ~ [ linux-6.12-rc7 ] ~ [ linux-6.11.7 ] ~ [ linux-6.10.14 ] ~ [ linux-6.9.12 ] ~ [ linux-6.8.12 ] ~ [ linux-6.7.12 ] ~ [ linux-6.6.60 ] ~ [ linux-6.5.13 ] ~ [ linux-6.4.16 ] ~ [ linux-6.3.13 ] ~ [ linux-6.2.16 ] ~ [ linux-6.1.116 ] ~ [ linux-6.0.19 ] ~ [ linux-5.19.17 ] ~ [ linux-5.18.19 ] ~ [ linux-5.17.15 ] ~ [ linux-5.16.20 ] ~ [ linux-5.15.171 ] ~ [ linux-5.14.21 ] ~ [ linux-5.13.19 ] ~ [ linux-5.12.19 ] ~ [ linux-5.11.22 ] ~ [ linux-5.10.229 ] ~ [ linux-5.9.16 ] ~ [ linux-5.8.18 ] ~ [ linux-5.7.19 ] ~ [ linux-5.6.19 ] ~ [ linux-5.5.19 ] ~ [ linux-5.4.285 ] ~ [ linux-5.3.18 ] ~ [ linux-5.2.21 ] ~ [ linux-5.1.21 ] ~ [ linux-5.0.21 ] ~ [ linux-4.20.17 ] ~ [ linux-4.19.323 ] ~ [ linux-4.18.20 ] ~ [ linux-4.17.19 ] ~ [ linux-4.16.18 ] ~ [ linux-4.15.18 ] ~ [ linux-4.14.336 ] ~ [ linux-4.13.16 ] ~ [ linux-4.12.14 ] ~ [ linux-4.11.12 ] ~ [ linux-4.10.17 ] ~ [ linux-4.9.337 ] ~ [ linux-4.4.302 ] ~ [ linux-3.10.108 ] ~ [ linux-2.6.32.71 ] ~ [ linux-2.6.0 ] ~ [ linux-2.4.37.11 ] ~ [ unix-v6-master ] ~ [ ccs-tools-1.8.12 ] ~ [ policy-sample ] ~
Architecture: ~ [ i386 ] ~ [ alpha ] ~ [ m68k ] ~ [ mips ] ~ [ ppc ] ~ [ sparc ] ~ [ sparc64 ] ~

Diff markup

Differences between /arch/parisc/lib/lusercopy.S (Version linux-6.12-rc7) and /arch/i386/lib/lusercopy.S (Version linux-5.7.19)


  1 /* SPDX-License-Identifier: GPL-2.0-or-later *    
  2 /*                                                
  3  *    User Space Access Routines                  
  4  *                                                
  5  *    Copyright (C) 2000-2002 Hewlett-Packard     
  6  *    Copyright (C) 2000 Richard Hirst <rhirst    
  7  *    Copyright (C) 2001 Matthieu Delahaye <de    
  8  *    Copyright (C) 2003 Randolph Chung <tausq    
  9  *    Copyright (C) 2017 Helge Deller <deller@g    
 10  *    Copyright (C) 2017 John David Anglin <dav    
 11  */                                               
 12                                                   
 13 /*                                                
 14  * These routines still have plenty of room fo    
 15  * (word & doubleword load/store, dual issue,     
 16  */                                               
 17                                                   
 18 /*                                                
 19  * The following routines assume that space re    
 20  * the space id associated with the current us    
 21  */                                               
 22                                                   
 23                                                   
 24         .text                                     
 25                                                   
 26 #include <asm/assembly.h>                         
 27 #include <asm/errno.h>                            
 28 #include <linux/linkage.h>                        
 29                                                   
 30         /*                                        
 31          * unsigned long lclear_user(void *to,    
 32          *                                        
 33          * Returns 0 for success.                 
 34          * otherwise, returns number of bytes     
 35          */                                       
 36                                                   
 37 ENTRY_CFI(lclear_user)                            
 38         comib,=,n   0,%r25,$lclu_done             
 39 $lclu_loop:                                       
 40         addib,<>    -1,%r25,$lclu_loop            
 41 1:      stbs,ma     %r0,1(%sr3,%r26)              
 42                                                   
 43 $lclu_done:                                       
 44         bv          %r0(%r2)                      
 45         copy        %r25,%r28                     
 46                                                   
 47 2:      b           $lclu_done                    
 48         ldo         1(%r25),%r25                  
 49                                                   
 50         ASM_EXCEPTIONTABLE_ENTRY(1b,2b)           
 51 ENDPROC_CFI(lclear_user)                          
 52                                                   
 53                                                   
 54 /*                                                
 55  * unsigned long pa_memcpy(void *dstp, const v    
 56  *                                                
 57  * Inputs:                                        
 58  * - sr1 already contains space of source regi    
 59  * - sr2 already contains space of destination    
 60  *                                                
 61  * Returns:                                       
 62  * - number of bytes that could not be copied.    
 63  *   On success, this will be zero.               
 64  *                                                
 65  * This code is based on a C-implementation of    
 66  * Randolph Chung, which in turn was derived f    
 67  *                                                
 68  * Several strategies are tried to try to get     
 69  * conditions. In the optimal case, we copy by    
 70  * at a time using general registers.  Unalign    
 71  * aligning the destination and then using shi    
 72  * cases by falling back to a byte-at-a-time c    
 73  *                                                
 74  * Testing with various alignments and buffer     
 75  * often >10x faster than a simple byte-at-a-t    
 76  * aligned operands. It is interesting to note    
 77  * (written in C) is actually quite fast alrea    
 78  * it by 30-40% for aligned copies because of     
 79  * cases the glibc version is still slightly f    
 80  * credibility that gcc can generate very good    
 81  *                                                
 82  * Possible optimizations:                        
 83  * - add cache prefetching                        
 84  * - try not to use the post-increment address    
 85  *   additional interlocks. Assumption is that    
 86  *   machines (pre PA8000 processors)             
 87  */                                               
 88                                                   
 89         dst = arg0                                
 90         src = arg1                                
 91         len = arg2                                
 92         end = arg3                                
 93         t1  = r19                                 
 94         t2  = r20                                 
 95         t3  = r21                                 
 96         t4  = r22                                 
 97         srcspc = sr1                              
 98         dstspc = sr2                              
 99                                                   
100         t0 = r1                                   
101         a1 = t1                                   
102         a2 = t2                                   
103         a3 = t3                                   
104         a0 = t4                                   
105                                                   
106         save_src = ret0                           
107         save_dst = ret1                           
108         save_len = r31                            
109                                                   
110 ENTRY_CFI(pa_memcpy)                              
111         /* Last destination address */            
112         add     dst,len,end                       
113                                                   
114         /* short copy with less than 16 bytes?    
115         cmpib,COND(>>=),n 15,len,.Lbyte_loop      
116                                                   
117         /* same alignment? */                     
118         xor     src,dst,t0                        
119         extru   t0,31,2,t1                        
120         cmpib,<>,n  0,t1,.Lunaligned_copy         
121                                                   
122 #ifdef CONFIG_64BIT                               
123         /* only do 64-bit copies if we can get    
124         extru   t0,31,3,t1                        
125         cmpib,<>,n  0,t1,.Lalign_loop32           
126                                                   
127         /* loop until we are 64-bit aligned */    
128 .Lalign_loop64:                                   
129         extru   dst,31,3,t1                       
130         cmpib,=,n       0,t1,.Lcopy_loop_16_st    
131 20:     ldb,ma  1(srcspc,src),t1                  
132 21:     stb,ma  t1,1(dstspc,dst)                  
133         b       .Lalign_loop64                    
134         ldo     -1(len),len                       
135                                                   
136         ASM_EXCEPTIONTABLE_ENTRY(20b,.Lcopy_do    
137         ASM_EXCEPTIONTABLE_ENTRY(21b,.Lcopy_do    
138                                                   
139 .Lcopy_loop_16_start:                             
140         ldi     31,t0                             
141 .Lcopy_loop_16:                                   
142         cmpb,COND(>>=),n t0,len,.Lword_loop       
143                                                   
144 10:     ldd     0(srcspc,src),t1                  
145 11:     ldd     8(srcspc,src),t2                  
146         ldo     16(src),src                       
147 12:     std,ma  t1,8(dstspc,dst)                  
148 13:     std,ma  t2,8(dstspc,dst)                  
149 14:     ldd     0(srcspc,src),t1                  
150 15:     ldd     8(srcspc,src),t2                  
151         ldo     16(src),src                       
152 16:     std,ma  t1,8(dstspc,dst)                  
153 17:     std,ma  t2,8(dstspc,dst)                  
154                                                   
155         ASM_EXCEPTIONTABLE_ENTRY(10b,.Lcopy_do    
156         ASM_EXCEPTIONTABLE_ENTRY(11b,.Lcopy16_    
157         ASM_EXCEPTIONTABLE_ENTRY(12b,.Lcopy_do    
158         ASM_EXCEPTIONTABLE_ENTRY(13b,.Lcopy_do    
159         ASM_EXCEPTIONTABLE_ENTRY(14b,.Lcopy_do    
160         ASM_EXCEPTIONTABLE_ENTRY(15b,.Lcopy16_    
161         ASM_EXCEPTIONTABLE_ENTRY(16b,.Lcopy_do    
162         ASM_EXCEPTIONTABLE_ENTRY(17b,.Lcopy_do    
163                                                   
164         b       .Lcopy_loop_16                    
165         ldo     -32(len),len                      
166                                                   
167 .Lword_loop:                                      
168         cmpib,COND(>>=),n 3,len,.Lbyte_loop       
169 20:     ldw,ma  4(srcspc,src),t1                  
170 21:     stw,ma  t1,4(dstspc,dst)                  
171         b       .Lword_loop                       
172         ldo     -4(len),len                       
173                                                   
174         ASM_EXCEPTIONTABLE_ENTRY(20b,.Lcopy_do    
175         ASM_EXCEPTIONTABLE_ENTRY(21b,.Lcopy_do    
176                                                   
177 #endif /* CONFIG_64BIT */                         
178                                                   
179         /* loop until we are 32-bit aligned */    
180 .Lalign_loop32:                                   
181         extru   dst,31,2,t1                       
182         cmpib,=,n       0,t1,.Lcopy_loop_8        
183 20:     ldb,ma  1(srcspc,src),t1                  
184 21:     stb,ma  t1,1(dstspc,dst)                  
185         b       .Lalign_loop32                    
186         ldo     -1(len),len                       
187                                                   
188         ASM_EXCEPTIONTABLE_ENTRY(20b,.Lcopy_do    
189         ASM_EXCEPTIONTABLE_ENTRY(21b,.Lcopy_do    
190                                                   
191                                                   
192 .Lcopy_loop_8:                                    
193         cmpib,COND(>>=),n 15,len,.Lbyte_loop      
194                                                   
195 10:     ldw     0(srcspc,src),t1                  
196 11:     ldw     4(srcspc,src),t2                  
197 12:     stw,ma  t1,4(dstspc,dst)                  
198 13:     stw,ma  t2,4(dstspc,dst)                  
199 14:     ldw     8(srcspc,src),t1                  
200 15:     ldw     12(srcspc,src),t2                 
201         ldo     16(src),src                       
202 16:     stw,ma  t1,4(dstspc,dst)                  
203 17:     stw,ma  t2,4(dstspc,dst)                  
204                                                   
205         ASM_EXCEPTIONTABLE_ENTRY(10b,.Lcopy_do    
206         ASM_EXCEPTIONTABLE_ENTRY(11b,.Lcopy8_f    
207         ASM_EXCEPTIONTABLE_ENTRY(12b,.Lcopy_do    
208         ASM_EXCEPTIONTABLE_ENTRY(13b,.Lcopy_do    
209         ASM_EXCEPTIONTABLE_ENTRY(14b,.Lcopy_do    
210         ASM_EXCEPTIONTABLE_ENTRY(15b,.Lcopy8_f    
211         ASM_EXCEPTIONTABLE_ENTRY(16b,.Lcopy_do    
212         ASM_EXCEPTIONTABLE_ENTRY(17b,.Lcopy_do    
213                                                   
214         b       .Lcopy_loop_8                     
215         ldo     -16(len),len                      
216                                                   
217 .Lbyte_loop:                                      
218         cmpclr,COND(<>) len,%r0,%r0               
219         b,n     .Lcopy_done                       
220 20:     ldb     0(srcspc,src),t1                  
221         ldo     1(src),src                        
222 21:     stb,ma  t1,1(dstspc,dst)                  
223         b       .Lbyte_loop                       
224         ldo     -1(len),len                       
225                                                   
226         ASM_EXCEPTIONTABLE_ENTRY(20b,.Lcopy_do    
227         ASM_EXCEPTIONTABLE_ENTRY(21b,.Lcopy_do    
228                                                   
229 .Lcopy_done:                                      
230         bv      %r0(%r2)                          
231         sub     end,dst,ret0                      
232                                                   
233                                                   
234         /* src and dst are not aligned the sam    
235         /* need to go the hard way */             
236 .Lunaligned_copy:                                 
237         /* align until dst is 32bit-word-align    
238         extru   dst,31,2,t1                       
239         cmpib,=,n       0,t1,.Lcopy_dstaligned    
240 20:     ldb     0(srcspc,src),t1                  
241         ldo     1(src),src                        
242 21:     stb,ma  t1,1(dstspc,dst)                  
243         b       .Lunaligned_copy                  
244         ldo     -1(len),len                       
245                                                   
246         ASM_EXCEPTIONTABLE_ENTRY(20b,.Lcopy_do    
247         ASM_EXCEPTIONTABLE_ENTRY(21b,.Lcopy_do    
248                                                   
249 .Lcopy_dstaligned:                                
250                                                   
251         /* store src, dst and len in safe plac    
252         copy    src,save_src                      
253         copy    dst,save_dst                      
254         copy    len,save_len                      
255                                                   
256         /* len now needs give number of words     
257         SHRREG  len,2,len                         
258                                                   
259         /*                                        
260          * Copy from a not-aligned src to an a    
261          * Handles 4 words per loop.              
262          */                                       
263                                                   
264         depw,z src,28,2,t0                        
265         subi 32,t0,t0                             
266         mtsar t0                                  
267         extru len,31,2,t0                         
268         cmpib,= 2,t0,.Lcase2                      
269         /* Make src aligned by rounding it dow    
270         depi 0,31,2,src                           
271                                                   
272         cmpiclr,<> 3,t0,%r0                       
273         b,n .Lcase3                               
274         cmpiclr,<> 1,t0,%r0                       
275         b,n .Lcase1                               
276 .Lcase0:                                          
277         cmpb,COND(=) %r0,len,.Lcda_finish         
278         nop                                       
279                                                   
280 1:      ldw,ma 4(srcspc,src), a3                  
281         ASM_EXCEPTIONTABLE_ENTRY(1b,.Lcda_rdfa    
282 1:      ldw,ma 4(srcspc,src), a0                  
283         ASM_EXCEPTIONTABLE_ENTRY(1b,.Lcda_rdfa    
284         b,n .Ldo3                                 
285 .Lcase1:                                          
286 1:      ldw,ma 4(srcspc,src), a2                  
287         ASM_EXCEPTIONTABLE_ENTRY(1b,.Lcda_rdfa    
288 1:      ldw,ma 4(srcspc,src), a3                  
289         ASM_EXCEPTIONTABLE_ENTRY(1b,.Lcda_rdfa    
290         ldo -1(len),len                           
291         cmpb,COND(=),n %r0,len,.Ldo0              
292 .Ldo4:                                            
293 1:      ldw,ma 4(srcspc,src), a0                  
294         ASM_EXCEPTIONTABLE_ENTRY(1b,.Lcda_rdfa    
295         shrpw a2, a3, %sar, t0                    
296 1:      stw,ma t0, 4(dstspc,dst)                  
297         ASM_EXCEPTIONTABLE_ENTRY(1b,.Lcopy_don    
298 .Ldo3:                                            
299 1:      ldw,ma 4(srcspc,src), a1                  
300         ASM_EXCEPTIONTABLE_ENTRY(1b,.Lcda_rdfa    
301         shrpw a3, a0, %sar, t0                    
302 1:      stw,ma t0, 4(dstspc,dst)                  
303         ASM_EXCEPTIONTABLE_ENTRY(1b,.Lcopy_don    
304 .Ldo2:                                            
305 1:      ldw,ma 4(srcspc,src), a2                  
306         ASM_EXCEPTIONTABLE_ENTRY(1b,.Lcda_rdfa    
307         shrpw a0, a1, %sar, t0                    
308 1:      stw,ma t0, 4(dstspc,dst)                  
309         ASM_EXCEPTIONTABLE_ENTRY(1b,.Lcopy_don    
310 .Ldo1:                                            
311 1:      ldw,ma 4(srcspc,src), a3                  
312         ASM_EXCEPTIONTABLE_ENTRY(1b,.Lcda_rdfa    
313         shrpw a1, a2, %sar, t0                    
314 1:      stw,ma t0, 4(dstspc,dst)                  
315         ASM_EXCEPTIONTABLE_ENTRY(1b,.Lcopy_don    
316         ldo -4(len),len                           
317         cmpb,COND(<>) %r0,len,.Ldo4               
318         nop                                       
319 .Ldo0:                                            
320         shrpw a2, a3, %sar, t0                    
321 1:      stw,ma t0, 4(dstspc,dst)                  
322         ASM_EXCEPTIONTABLE_ENTRY(1b,.Lcopy_don    
323                                                   
324 .Lcda_rdfault:                                    
325 .Lcda_finish:                                     
326         /* calculate new src, dst and len and     
327         sub     dst,save_dst,t0                   
328         add     save_src,t0,src                   
329         b       .Lbyte_loop                       
330         sub     save_len,t0,len                   
331                                                   
332 .Lcase3:                                          
333 1:      ldw,ma 4(srcspc,src), a0                  
334         ASM_EXCEPTIONTABLE_ENTRY(1b,.Lcda_rdfa    
335 1:      ldw,ma 4(srcspc,src), a1                  
336         ASM_EXCEPTIONTABLE_ENTRY(1b,.Lcda_rdfa    
337         b .Ldo2                                   
338         ldo 1(len),len                            
339 .Lcase2:                                          
340 1:      ldw,ma 4(srcspc,src), a1                  
341         ASM_EXCEPTIONTABLE_ENTRY(1b,.Lcda_rdfa    
342 1:      ldw,ma 4(srcspc,src), a2                  
343         ASM_EXCEPTIONTABLE_ENTRY(1b,.Lcda_rdfa    
344         b .Ldo1                                   
345         ldo 2(len),len                            
346                                                   
347                                                   
348         /* fault exception fixup handlers: */     
349 #ifdef CONFIG_64BIT                               
350 .Lcopy16_fault:                                   
351         b       .Lcopy_done                       
352 10:     std,ma  t1,8(dstspc,dst)                  
353         ASM_EXCEPTIONTABLE_ENTRY(10b,.Lcopy_do    
354 #endif                                            
355                                                   
356 .Lcopy8_fault:                                    
357         b       .Lcopy_done                       
358 10:     stw,ma  t1,4(dstspc,dst)                  
359         ASM_EXCEPTIONTABLE_ENTRY(10b,.Lcopy_do    
360 ENDPROC_CFI(pa_memcpy)                            
361                                                   
362         .end                                      
                                                      

~ [ source navigation ] ~ [ diff markup ] ~ [ identifier search ] ~

kernel.org | git.kernel.org | LWN.net | Project Home | SVN repository | Mail admin

Linux® is a registered trademark of Linus Torvalds in the United States and other countries.
TOMOYO® is a registered trademark of NTT DATA CORPORATION.

sflogo.php