~ [ source navigation ] ~ [ diff markup ] ~ [ identifier search ] ~

TOMOYO Linux Cross Reference
Linux/arch/mips/lib/csum_partial.S

Version: ~ [ linux-6.12-rc7 ] ~ [ linux-6.11.7 ] ~ [ linux-6.10.14 ] ~ [ linux-6.9.12 ] ~ [ linux-6.8.12 ] ~ [ linux-6.7.12 ] ~ [ linux-6.6.60 ] ~ [ linux-6.5.13 ] ~ [ linux-6.4.16 ] ~ [ linux-6.3.13 ] ~ [ linux-6.2.16 ] ~ [ linux-6.1.116 ] ~ [ linux-6.0.19 ] ~ [ linux-5.19.17 ] ~ [ linux-5.18.19 ] ~ [ linux-5.17.15 ] ~ [ linux-5.16.20 ] ~ [ linux-5.15.171 ] ~ [ linux-5.14.21 ] ~ [ linux-5.13.19 ] ~ [ linux-5.12.19 ] ~ [ linux-5.11.22 ] ~ [ linux-5.10.229 ] ~ [ linux-5.9.16 ] ~ [ linux-5.8.18 ] ~ [ linux-5.7.19 ] ~ [ linux-5.6.19 ] ~ [ linux-5.5.19 ] ~ [ linux-5.4.285 ] ~ [ linux-5.3.18 ] ~ [ linux-5.2.21 ] ~ [ linux-5.1.21 ] ~ [ linux-5.0.21 ] ~ [ linux-4.20.17 ] ~ [ linux-4.19.323 ] ~ [ linux-4.18.20 ] ~ [ linux-4.17.19 ] ~ [ linux-4.16.18 ] ~ [ linux-4.15.18 ] ~ [ linux-4.14.336 ] ~ [ linux-4.13.16 ] ~ [ linux-4.12.14 ] ~ [ linux-4.11.12 ] ~ [ linux-4.10.17 ] ~ [ linux-4.9.337 ] ~ [ linux-4.4.302 ] ~ [ linux-3.10.108 ] ~ [ linux-2.6.32.71 ] ~ [ linux-2.6.0 ] ~ [ linux-2.4.37.11 ] ~ [ unix-v6-master ] ~ [ ccs-tools-1.8.12 ] ~ [ policy-sample ] ~
Architecture: ~ [ i386 ] ~ [ alpha ] ~ [ m68k ] ~ [ mips ] ~ [ ppc ] ~ [ sparc ] ~ [ sparc64 ] ~

Diff markup

Differences between /arch/mips/lib/csum_partial.S (Architecture sparc64) and /arch/i386/lib/csum_partial.S (Architecture i386)


  1 /*                                                
  2  * This file is subject to the terms and condi    
  3  * License.  See the file "COPYING" in the mai    
  4  * for more details.                              
  5  *                                                
  6  * Quick'n'dirty IP checksum ...                  
  7  *                                                
  8  * Copyright (C) 1998, 1999 Ralf Baechle          
  9  * Copyright (C) 1999 Silicon Graphics, Inc.      
 10  * Copyright (C) 2007  Maciej W. Rozycki          
 11  * Copyright (C) 2014 Imagination Technologies    
 12  */                                               
 13 #include <linux/errno.h>                          
 14 #include <linux/export.h>                         
 15 #include <asm/asm.h>                              
 16 #include <asm/asm-offsets.h>                      
 17 #include <asm/regdef.h>                           
 18                                                   
 19 #ifdef CONFIG_64BIT                               
 20 /*                                                
 21  * As we are sharing code base with the mips32    
 22  * register definitions). We need to redefine     
 23  * the n64 ABI register naming to the o32 ABI     
 24  */                                               
 25 #undef t0                                         
 26 #undef t1                                         
 27 #undef t2                                         
 28 #undef t3                                         
 29 #define t0      $8                                
 30 #define t1      $9                                
 31 #define t2      $10                               
 32 #define t3      $11                               
 33 #define t4      $12                               
 34 #define t5      $13                               
 35 #define t6      $14                               
 36 #define t7      $15                               
 37                                                   
 38 #define USE_DOUBLE                                
 39 #endif                                            
 40                                                   
 41 #ifdef USE_DOUBLE                                 
 42                                                   
 43 #define LOAD   ld                                 
 44 #define LOAD32 lwu                                
 45 #define ADD    daddu                              
 46 #define NBYTES 8                                  
 47                                                   
 48 #else                                             
 49                                                   
 50 #define LOAD   lw                                 
 51 #define LOAD32 lw                                 
 52 #define ADD    addu                               
 53 #define NBYTES 4                                  
 54                                                   
 55 #endif /* USE_DOUBLE */                           
 56                                                   
 57 #define UNIT(unit)  ((unit)*NBYTES)               
 58                                                   
 59 #define ADDC(sum,reg)                             
 60         .set    push;                             
 61         .set    noat;                             
 62         ADD     sum, reg;                         
 63         sltu    v1, sum, reg;                     
 64         ADD     sum, v1;                          
 65         .set    pop                               
 66                                                   
 67 #define ADDC32(sum,reg)                           
 68         .set    push;                             
 69         .set    noat;                             
 70         addu    sum, reg;                         
 71         sltu    v1, sum, reg;                     
 72         addu    sum, v1;                          
 73         .set    pop                               
 74                                                   
 75 #define CSUM_BIGCHUNK1(src, offset, sum, _t0,     
 76         LOAD    _t0, (offset + UNIT(0))(src);     
 77         LOAD    _t1, (offset + UNIT(1))(src);     
 78         LOAD    _t2, (offset + UNIT(2))(src);     
 79         LOAD    _t3, (offset + UNIT(3))(src);     
 80         ADDC(_t0, _t1);                           
 81         ADDC(_t2, _t3);                           
 82         ADDC(sum, _t0);                           
 83         ADDC(sum, _t2)                            
 84                                                   
 85 #ifdef USE_DOUBLE                                 
 86 #define CSUM_BIGCHUNK(src, offset, sum, _t0, _    
 87         CSUM_BIGCHUNK1(src, offset, sum, _t0,     
 88 #else                                             
 89 #define CSUM_BIGCHUNK(src, offset, sum, _t0, _    
 90         CSUM_BIGCHUNK1(src, offset, sum, _t0,     
 91         CSUM_BIGCHUNK1(src, offset + 0x10, sum    
 92 #endif                                            
 93                                                   
 94 /*                                                
 95  * a0: source address                             
 96  * a1: length of the area to checksum             
 97  * a2: partial checksum                           
 98  */                                               
 99                                                   
100 #define src a0                                    
101 #define sum v0                                    
102                                                   
103         .text                                     
104         .set    noreorder                         
105         .align  5                                 
106 LEAF(csum_partial)                                
107 EXPORT_SYMBOL(csum_partial)                       
108         move    sum, zero                         
109         move    t7, zero                          
110                                                   
111         sltiu   t8, a1, 0x8                       
112         bnez    t8, .Lsmall_csumcpy               
113          move   t2, a1                            
114                                                   
115         andi    t7, src, 0x1                      
116                                                   
117 .Lhword_align:                                    
118         beqz    t7, .Lword_align                  
119          andi   t8, src, 0x2                      
120                                                   
121         lbu     t0, (src)                         
122         LONG_SUBU       a1, a1, 0x1               
123 #ifdef __MIPSEL__                                 
124         sll     t0, t0, 8                         
125 #endif                                            
126         ADDC(sum, t0)                             
127         PTR_ADDU        src, src, 0x1             
128         andi    t8, src, 0x2                      
129                                                   
130 .Lword_align:                                     
131         beqz    t8, .Ldword_align                 
132          sltiu  t8, a1, 56                        
133                                                   
134         lhu     t0, (src)                         
135         LONG_SUBU       a1, a1, 0x2               
136         ADDC(sum, t0)                             
137         sltiu   t8, a1, 56                        
138         PTR_ADDU        src, src, 0x2             
139                                                   
140 .Ldword_align:                                    
141         bnez    t8, .Ldo_end_words                
142          move   t8, a1                            
143                                                   
144         andi    t8, src, 0x4                      
145         beqz    t8, .Lqword_align                 
146          andi   t8, src, 0x8                      
147                                                   
148         LOAD32  t0, 0x00(src)                     
149         LONG_SUBU       a1, a1, 0x4               
150         ADDC(sum, t0)                             
151         PTR_ADDU        src, src, 0x4             
152         andi    t8, src, 0x8                      
153                                                   
154 .Lqword_align:                                    
155         beqz    t8, .Loword_align                 
156          andi   t8, src, 0x10                     
157                                                   
158 #ifdef USE_DOUBLE                                 
159         ld      t0, 0x00(src)                     
160         LONG_SUBU       a1, a1, 0x8               
161         ADDC(sum, t0)                             
162 #else                                             
163         lw      t0, 0x00(src)                     
164         lw      t1, 0x04(src)                     
165         LONG_SUBU       a1, a1, 0x8               
166         ADDC(sum, t0)                             
167         ADDC(sum, t1)                             
168 #endif                                            
169         PTR_ADDU        src, src, 0x8             
170         andi    t8, src, 0x10                     
171                                                   
172 .Loword_align:                                    
173         beqz    t8, .Lbegin_movement              
174          LONG_SRL       t8, a1, 0x7               
175                                                   
176 #ifdef USE_DOUBLE                                 
177         ld      t0, 0x00(src)                     
178         ld      t1, 0x08(src)                     
179         ADDC(sum, t0)                             
180         ADDC(sum, t1)                             
181 #else                                             
182         CSUM_BIGCHUNK1(src, 0x00, sum, t0, t1,    
183 #endif                                            
184         LONG_SUBU       a1, a1, 0x10              
185         PTR_ADDU        src, src, 0x10            
186         LONG_SRL        t8, a1, 0x7               
187                                                   
188 .Lbegin_movement:                                 
189         beqz    t8, 1f                            
190          andi   t2, a1, 0x40                      
191                                                   
192 .Lmove_128bytes:                                  
193         CSUM_BIGCHUNK(src, 0x00, sum, t0, t1,     
194         CSUM_BIGCHUNK(src, 0x20, sum, t0, t1,     
195         CSUM_BIGCHUNK(src, 0x40, sum, t0, t1,     
196         CSUM_BIGCHUNK(src, 0x60, sum, t0, t1,     
197         LONG_SUBU       t8, t8, 0x01              
198         .set    reorder                           
199         PTR_ADDU        src, src, 0x80            
200         bnez    t8, .Lmove_128bytes               
201         .set    noreorder                         
202                                                   
203 1:                                                
204         beqz    t2, 1f                            
205          andi   t2, a1, 0x20                      
206                                                   
207 .Lmove_64bytes:                                   
208         CSUM_BIGCHUNK(src, 0x00, sum, t0, t1,     
209         CSUM_BIGCHUNK(src, 0x20, sum, t0, t1,     
210         PTR_ADDU        src, src, 0x40            
211                                                   
212 1:                                                
213         beqz    t2, .Ldo_end_words                
214          andi   t8, a1, 0x1c                      
215                                                   
216 .Lmove_32bytes:                                   
217         CSUM_BIGCHUNK(src, 0x00, sum, t0, t1,     
218         andi    t8, a1, 0x1c                      
219         PTR_ADDU        src, src, 0x20            
220                                                   
221 .Ldo_end_words:                                   
222         beqz    t8, .Lsmall_csumcpy               
223          andi   t2, a1, 0x3                       
224         LONG_SRL        t8, t8, 0x2               
225                                                   
226 .Lend_words:                                      
227         LOAD32  t0, (src)                         
228         LONG_SUBU       t8, t8, 0x1               
229         ADDC(sum, t0)                             
230         .set    reorder                           
231         PTR_ADDU        src, src, 0x4             
232         bnez    t8, .Lend_words                   
233         .set    noreorder                         
234                                                   
235 /* unknown src alignment and < 8 bytes to go      
236 .Lsmall_csumcpy:                                  
237         move    a1, t2                            
238                                                   
239         andi    t0, a1, 4                         
240         beqz    t0, 1f                            
241          andi   t0, a1, 2                         
242                                                   
243         /* Still a full word to go  */            
244         ulw     t1, (src)                         
245         PTR_ADDIU       src, 4                    
246 #ifdef USE_DOUBLE                                 
247         dsll    t1, t1, 32                        
248 #endif                                            
249         ADDC(sum, t1)                             
250                                                   
251 1:      move    t1, zero                          
252         beqz    t0, 1f                            
253          andi   t0, a1, 1                         
254                                                   
255         /* Still a halfword to go  */             
256         ulhu    t1, (src)                         
257         PTR_ADDIU       src, 2                    
258                                                   
259 1:      beqz    t0, 1f                            
260          sll    t1, t1, 16                        
261                                                   
262         lbu     t2, (src)                         
263          nop                                      
264                                                   
265 #ifdef __MIPSEB__                                 
266         sll     t2, t2, 8                         
267 #endif                                            
268         or      t1, t2                            
269                                                   
270 1:      ADDC(sum, t1)                             
271                                                   
272         /* fold checksum */                       
273 #ifdef USE_DOUBLE                                 
274         dsll32  v1, sum, 0                        
275         daddu   sum, v1                           
276         sltu    v1, sum, v1                       
277         dsra32  sum, sum, 0                       
278         addu    sum, v1                           
279 #endif                                            
280                                                   
281         /* odd buffer alignment? */               
282 #if defined(CONFIG_CPU_MIPSR2) || defined(CONF    
283     defined(CONFIG_CPU_LOONGSON64)                
284         .set    push                              
285         .set    arch=mips32r2                     
286         wsbh    v1, sum                           
287         movn    sum, v1, t7                       
288         .set    pop                               
289 #else                                             
290         beqz    t7, 1f                  /* odd    
291          lui    v1, 0x00ff                        
292         addu    v1, 0x00ff                        
293         and     t0, sum, v1                       
294         sll     t0, t0, 8                         
295         srl     sum, sum, 8                       
296         and     sum, sum, v1                      
297         or      sum, sum, t0                      
298 1:                                                
299 #endif                                            
300         .set    reorder                           
301         /* Add the passed partial csum.  */       
302         ADDC32(sum, a2)                           
303         jr      ra                                
304         .set    noreorder                         
305         END(csum_partial)                         
306                                                   
307                                                   
308 /*                                                
309  * checksum and copy routines based on memcpy.    
310  *                                                
311  *      csum_partial_copy_nocheck(src, dst, le    
312  *      __csum_partial_copy_kernel(src, dst, l    
313  *                                                
314  * See "Spec" in memcpy.S for details.  Unlike    
315  * function in this file use the standard call    
316  */                                               
317                                                   
318 #define src a0                                    
319 #define dst a1                                    
320 #define len a2                                    
321 #define sum v0                                    
322 #define odd t8                                    
323                                                   
324 /*                                                
325  * All exception handlers simply return 0.        
326  */                                               
327                                                   
328 /* Instruction type */                            
329 #define LD_INSN 1                                 
330 #define ST_INSN 2                                 
331 #define LEGACY_MODE 1                             
332 #define EVA_MODE    2                             
333 #define USEROP   1                                
334 #define KERNELOP 2                                
335                                                   
336 /*                                                
337  * Wrapper to add an entry in the exception ta    
338  * in case the insn causes a memory exception.    
339  * Arguments:                                     
340  * insn    : Load/store instruction               
341  * type    : Instruction type                     
342  * reg     : Register                             
343  * addr    : Address                              
344  * handler : Exception handler                    
345  */                                               
346 #define EXC(insn, type, reg, addr)                
347         .if \mode == LEGACY_MODE;                 
348 9:              insn reg, addr;                   
349                 .section __ex_table,"a";          
350                 PTR_WD  9b, .L_exc;               
351                 .previous;                        
352         /* This is enabled in EVA mode */         
353         .else;                                    
354                 /* If loading from user or sto    
355                 .if ((\from == USEROP) && (typ    
356                     ((\to == USEROP) && (type     
357 9:                      __BUILD_EVA_INSN(insn#    
358                         .section __ex_table,"a    
359                         PTR_WD  9b, .L_exc;       
360                         .previous;                
361                 .else;                            
362                         /* EVA without excepti    
363                         insn reg, addr;           
364                 .endif;                           
365         .endif                                    
366                                                   
367 #undef LOAD                                       
368                                                   
369 #ifdef USE_DOUBLE                                 
370                                                   
371 #define LOADK   ld /* No exception */             
372 #define LOAD(reg, addr)         EXC(ld, LD_INS    
373 #define LOADBU(reg, addr)       EXC(lbu, LD_IN    
374 #define LOADL(reg, addr)        EXC(ldl, LD_IN    
375 #define LOADR(reg, addr)        EXC(ldr, LD_IN    
376 #define STOREB(reg, addr)       EXC(sb, ST_INS    
377 #define STOREL(reg, addr)       EXC(sdl, ST_IN    
378 #define STORER(reg, addr)       EXC(sdr, ST_IN    
379 #define STORE(reg, addr)        EXC(sd, ST_INS    
380 #define ADD    daddu                              
381 #define SUB    dsubu                              
382 #define SRL    dsrl                               
383 #define SLL    dsll                               
384 #define SLLV   dsllv                              
385 #define SRLV   dsrlv                              
386 #define NBYTES 8                                  
387 #define LOG_NBYTES 3                              
388                                                   
389 #else                                             
390                                                   
391 #define LOADK   lw /* No exception */             
392 #define LOAD(reg, addr)         EXC(lw, LD_INS    
393 #define LOADBU(reg, addr)       EXC(lbu, LD_IN    
394 #define LOADL(reg, addr)        EXC(lwl, LD_IN    
395 #define LOADR(reg, addr)        EXC(lwr, LD_IN    
396 #define STOREB(reg, addr)       EXC(sb, ST_INS    
397 #define STOREL(reg, addr)       EXC(swl, ST_IN    
398 #define STORER(reg, addr)       EXC(swr, ST_IN    
399 #define STORE(reg, addr)        EXC(sw, ST_INS    
400 #define ADD    addu                               
401 #define SUB    subu                               
402 #define SRL    srl                                
403 #define SLL    sll                                
404 #define SLLV   sllv                               
405 #define SRLV   srlv                               
406 #define NBYTES 4                                  
407 #define LOG_NBYTES 2                              
408                                                   
409 #endif /* USE_DOUBLE */                           
410                                                   
411 #ifdef CONFIG_CPU_LITTLE_ENDIAN                   
412 #define LDFIRST LOADR                             
413 #define LDREST  LOADL                             
414 #define STFIRST STORER                            
415 #define STREST  STOREL                            
416 #define SHIFT_DISCARD SLLV                        
417 #define SHIFT_DISCARD_REVERT SRLV                 
418 #else                                             
419 #define LDFIRST LOADL                             
420 #define LDREST  LOADR                             
421 #define STFIRST STOREL                            
422 #define STREST  STORER                            
423 #define SHIFT_DISCARD SRLV                        
424 #define SHIFT_DISCARD_REVERT SLLV                 
425 #endif                                            
426                                                   
427 #define FIRST(unit) ((unit)*NBYTES)               
428 #define REST(unit)  (FIRST(unit)+NBYTES-1)        
429                                                   
430 #define ADDRMASK (NBYTES-1)                       
431                                                   
432 #ifndef CONFIG_CPU_DADDI_WORKAROUNDS              
433         .set    noat                              
434 #else                                             
435         .set    at=v1                             
436 #endif                                            
437                                                   
438         .macro __BUILD_CSUM_PARTIAL_COPY_USER     
439                                                   
440         li      sum, -1                           
441         move    odd, zero                         
442         /*                                        
443          * Note: dst & src may be unaligned, l    
444          * Temps                                  
445          */                                       
446         /*                                        
447          * The "issue break"s below are very a    
448          * Issue delays for dcache fills will     
449          * load queue full replay traps, etc.     
450          *                                        
451          * If len < NBYTES use byte operations    
452          */                                       
453         sltu    t2, len, NBYTES                   
454         and     t1, dst, ADDRMASK                 
455         bnez    t2, .Lcopy_bytes_checklen\@       
456          and    t0, src, ADDRMASK                 
457         andi    odd, dst, 0x1                     
458         bnez    t1, .Ldst_unaligned\@             
459          nop                                      
460         bnez    t0, .Lsrc_unaligned_dst_aligne    
461         /*                                        
462          * use delay slot for fall-through        
463          * src and dst are aligned; need to co    
464          */                                       
465 .Lboth_aligned\@:                                 
466          SRL    t0, len, LOG_NBYTES+3    # +3     
467         beqz    t0, .Lcleanup_both_aligned\@ #    
468          nop                                      
469         SUB     len, 8*NBYTES           # subt    
470         .align  4                                 
471 1:                                                
472         LOAD(t0, UNIT(0)(src))                    
473         LOAD(t1, UNIT(1)(src))                    
474         LOAD(t2, UNIT(2)(src))                    
475         LOAD(t3, UNIT(3)(src))                    
476         LOAD(t4, UNIT(4)(src))                    
477         LOAD(t5, UNIT(5)(src))                    
478         LOAD(t6, UNIT(6)(src))                    
479         LOAD(t7, UNIT(7)(src))                    
480         SUB     len, len, 8*NBYTES                
481         ADD     src, src, 8*NBYTES                
482         STORE(t0, UNIT(0)(dst))                   
483         ADDC(t0, t1)                              
484         STORE(t1, UNIT(1)(dst))                   
485         ADDC(sum, t0)                             
486         STORE(t2, UNIT(2)(dst))                   
487         ADDC(t2, t3)                              
488         STORE(t3, UNIT(3)(dst))                   
489         ADDC(sum, t2)                             
490         STORE(t4, UNIT(4)(dst))                   
491         ADDC(t4, t5)                              
492         STORE(t5, UNIT(5)(dst))                   
493         ADDC(sum, t4)                             
494         STORE(t6, UNIT(6)(dst))                   
495         ADDC(t6, t7)                              
496         STORE(t7, UNIT(7)(dst))                   
497         ADDC(sum, t6)                             
498         .set    reorder                           
499         ADD     dst, dst, 8*NBYTES                
500         bgez    len, 1b                           
501         .set    noreorder                         
502         ADD     len, 8*NBYTES           # reve    
503                                                   
504         /*                                        
505          * len == the number of bytes left to     
506          */                                       
507 .Lcleanup_both_aligned\@:                         
508 #define rem t7                                    
509         beqz    len, .Ldone\@                     
510          sltu   t0, len, 4*NBYTES                 
511         bnez    t0, .Lless_than_4units\@          
512          and    rem, len, (NBYTES-1)    # rem     
513         /*                                        
514          * len >= 4*NBYTES                        
515          */                                       
516         LOAD(t0, UNIT(0)(src))                    
517         LOAD(t1, UNIT(1)(src))                    
518         LOAD(t2, UNIT(2)(src))                    
519         LOAD(t3, UNIT(3)(src))                    
520         SUB     len, len, 4*NBYTES                
521         ADD     src, src, 4*NBYTES                
522         STORE(t0, UNIT(0)(dst))                   
523         ADDC(t0, t1)                              
524         STORE(t1, UNIT(1)(dst))                   
525         ADDC(sum, t0)                             
526         STORE(t2, UNIT(2)(dst))                   
527         ADDC(t2, t3)                              
528         STORE(t3, UNIT(3)(dst))                   
529         ADDC(sum, t2)                             
530         .set    reorder                           
531         ADD     dst, dst, 4*NBYTES                
532         beqz    len, .Ldone\@                     
533         .set    noreorder                         
534 .Lless_than_4units\@:                             
535         /*                                        
536          * rem = len % NBYTES                     
537          */                                       
538         beq     rem, len, .Lcopy_bytes\@          
539          nop                                      
540 1:                                                
541         LOAD(t0, 0(src))                          
542         ADD     src, src, NBYTES                  
543         SUB     len, len, NBYTES                  
544         STORE(t0, 0(dst))                         
545         ADDC(sum, t0)                             
546         .set    reorder                           
547         ADD     dst, dst, NBYTES                  
548         bne     rem, len, 1b                      
549         .set    noreorder                         
550                                                   
551         /*                                        
552          * src and dst are aligned, need to co    
553          * A loop would do only a byte at a ti    
554          * mispredicts.  Can't do an explicit     
555          * because can't assume read-access to    
556          * STREST dst, which doesn't require r    
557          *                                        
558          * This code should perform better tha    
559          * wide-issue mips processors because     
560          * more instruction-level parallelism.    
561          */                                       
562 #define bits t2                                   
563         beqz    len, .Ldone\@                     
564          ADD    t1, dst, len    # t1 is just p    
565         li      bits, 8*NBYTES                    
566         SLL     rem, len, 3     # rem = number    
567         LOAD(t0, 0(src))                          
568         SUB     bits, bits, rem # bits = numbe    
569         SHIFT_DISCARD t0, t0, bits                
570         STREST(t0, -1(t1))                        
571         SHIFT_DISCARD_REVERT t0, t0, bits         
572         .set reorder                              
573         ADDC(sum, t0)                             
574         b       .Ldone\@                          
575         .set noreorder                            
576 .Ldst_unaligned\@:                                
577         /*                                        
578          * dst is unaligned                       
579          * t0 = src & ADDRMASK                    
580          * t1 = dst & ADDRMASK; T1 > 0            
581          * len >= NBYTES                          
582          *                                        
583          * Copy enough bytes to align dst         
584          * Set match = (src and dst have same     
585          */                                       
586 #define match rem                                 
587         LDFIRST(t3, FIRST(0)(src))                
588         ADD     t2, zero, NBYTES                  
589         LDREST(t3, REST(0)(src))                  
590         SUB     t2, t2, t1      # t2 = number     
591         xor     match, t0, t1                     
592         STFIRST(t3, FIRST(0)(dst))                
593         SLL     t4, t1, 3               # t4 =    
594         SHIFT_DISCARD t3, t3, t4                  
595         /* no SHIFT_DISCARD_REVERT to handle o    
596         ADDC(sum, t3)                             
597         beq     len, t2, .Ldone\@                 
598          SUB    len, len, t2                      
599         ADD     dst, dst, t2                      
600         beqz    match, .Lboth_aligned\@           
601          ADD    src, src, t2                      
602                                                   
603 .Lsrc_unaligned_dst_aligned\@:                    
604         SRL     t0, len, LOG_NBYTES+2    # +2     
605         beqz    t0, .Lcleanup_src_unaligned\@     
606          and    rem, len, (4*NBYTES-1)   # rem    
607 1:                                                
608 /*                                                
609  * Avoid consecutive LD*'s to the same registe    
610  * implementations can't issue them in the sam    
611  * It's OK to load FIRST(N+1) before REST(N) b    
612  * are to the same unit (unless src is aligned    
613  */                                               
614         LDFIRST(t0, FIRST(0)(src))                
615         LDFIRST(t1, FIRST(1)(src))                
616         SUB     len, len, 4*NBYTES                
617         LDREST(t0, REST(0)(src))                  
618         LDREST(t1, REST(1)(src))                  
619         LDFIRST(t2, FIRST(2)(src))                
620         LDFIRST(t3, FIRST(3)(src))                
621         LDREST(t2, REST(2)(src))                  
622         LDREST(t3, REST(3)(src))                  
623         ADD     src, src, 4*NBYTES                
624 #ifdef CONFIG_CPU_SB1                             
625         nop                             # impr    
626 #endif                                            
627         STORE(t0, UNIT(0)(dst))                   
628         ADDC(t0, t1)                              
629         STORE(t1, UNIT(1)(dst))                   
630         ADDC(sum, t0)                             
631         STORE(t2, UNIT(2)(dst))                   
632         ADDC(t2, t3)                              
633         STORE(t3, UNIT(3)(dst))                   
634         ADDC(sum, t2)                             
635         .set    reorder                           
636         ADD     dst, dst, 4*NBYTES                
637         bne     len, rem, 1b                      
638         .set    noreorder                         
639                                                   
640 .Lcleanup_src_unaligned\@:                        
641         beqz    len, .Ldone\@                     
642          and    rem, len, NBYTES-1  # rem = le    
643         beq     rem, len, .Lcopy_bytes\@          
644          nop                                      
645 1:                                                
646         LDFIRST(t0, FIRST(0)(src))                
647         LDREST(t0, REST(0)(src))                  
648         ADD     src, src, NBYTES                  
649         SUB     len, len, NBYTES                  
650         STORE(t0, 0(dst))                         
651         ADDC(sum, t0)                             
652         .set    reorder                           
653         ADD     dst, dst, NBYTES                  
654         bne     len, rem, 1b                      
655         .set    noreorder                         
656                                                   
657 .Lcopy_bytes_checklen\@:                          
658         beqz    len, .Ldone\@                     
659          nop                                      
660 .Lcopy_bytes\@:                                   
661         /* 0 < len < NBYTES  */                   
662 #ifdef CONFIG_CPU_LITTLE_ENDIAN                   
663 #define SHIFT_START 0                             
664 #define SHIFT_INC 8                               
665 #else                                             
666 #define SHIFT_START 8*(NBYTES-1)                  
667 #define SHIFT_INC -8                              
668 #endif                                            
669         move    t2, zero        # partial word    
670         li      t3, SHIFT_START # shift           
671 #define COPY_BYTE(N)                    \         
672         LOADBU(t0, N(src));             \         
673         SUB     len, len, 1;            \         
674         STOREB(t0, N(dst));             \         
675         SLLV    t0, t0, t3;             \         
676         addu    t3, SHIFT_INC;          \         
677         beqz    len, .Lcopy_bytes_done\@; \       
678          or     t2, t0                            
679                                                   
680         COPY_BYTE(0)                              
681         COPY_BYTE(1)                              
682 #ifdef USE_DOUBLE                                 
683         COPY_BYTE(2)                              
684         COPY_BYTE(3)                              
685         COPY_BYTE(4)                              
686         COPY_BYTE(5)                              
687 #endif                                            
688         LOADBU(t0, NBYTES-2(src))                 
689         SUB     len, len, 1                       
690         STOREB(t0, NBYTES-2(dst))                 
691         SLLV    t0, t0, t3                        
692         or      t2, t0                            
693 .Lcopy_bytes_done\@:                              
694         ADDC(sum, t2)                             
695 .Ldone\@:                                         
696         /* fold checksum */                       
697         .set    push                              
698         .set    noat                              
699 #ifdef USE_DOUBLE                                 
700         dsll32  v1, sum, 0                        
701         daddu   sum, v1                           
702         sltu    v1, sum, v1                       
703         dsra32  sum, sum, 0                       
704         addu    sum, v1                           
705 #endif                                            
706                                                   
707 #if defined(CONFIG_CPU_MIPSR2) || defined(CONF    
708     defined(CONFIG_CPU_LOONGSON64)                
709         .set    push                              
710         .set    arch=mips32r2                     
711         wsbh    v1, sum                           
712         movn    sum, v1, odd                      
713         .set    pop                               
714 #else                                             
715         beqz    odd, 1f                 /* odd    
716          lui    v1, 0x00ff                        
717         addu    v1, 0x00ff                        
718         and     t0, sum, v1                       
719         sll     t0, t0, 8                         
720         srl     sum, sum, 8                       
721         and     sum, sum, v1                      
722         or      sum, sum, t0                      
723 1:                                                
724 #endif                                            
725         .set    pop                               
726         .set reorder                              
727         jr      ra                                
728         .set noreorder                            
729         .endm                                     
730                                                   
731         .set noreorder                            
732 .L_exc:                                           
733         jr      ra                                
734          li     v0, 0                             
735                                                   
736 FEXPORT(__csum_partial_copy_nocheck)              
737 EXPORT_SYMBOL(__csum_partial_copy_nocheck)        
738 #ifndef CONFIG_EVA                                
739 FEXPORT(__csum_partial_copy_to_user)              
740 EXPORT_SYMBOL(__csum_partial_copy_to_user)        
741 FEXPORT(__csum_partial_copy_from_user)            
742 EXPORT_SYMBOL(__csum_partial_copy_from_user)      
743 #endif                                            
744 __BUILD_CSUM_PARTIAL_COPY_USER LEGACY_MODE USE    
745                                                   
746 #ifdef CONFIG_EVA                                 
747 LEAF(__csum_partial_copy_to_user)                 
748 __BUILD_CSUM_PARTIAL_COPY_USER EVA_MODE KERNEL    
749 END(__csum_partial_copy_to_user)                  
750                                                   
751 LEAF(__csum_partial_copy_from_user)               
752 __BUILD_CSUM_PARTIAL_COPY_USER EVA_MODE USEROP    
753 END(__csum_partial_copy_from_user)                
754 #endif                                            
                                                      

~ [ source navigation ] ~ [ diff markup ] ~ [ identifier search ] ~

kernel.org | git.kernel.org | LWN.net | Project Home | SVN repository | Mail admin

Linux® is a registered trademark of Linus Torvalds in the United States and other countries.
TOMOYO® is a registered trademark of NTT DATA CORPORATION.

sflogo.php