~ [ source navigation ] ~ [ diff markup ] ~ [ identifier search ] ~

TOMOYO Linux Cross Reference
Linux/arch/riscv/lib/memset.S

Version: ~ [ linux-6.12-rc7 ] ~ [ linux-6.11.7 ] ~ [ linux-6.10.14 ] ~ [ linux-6.9.12 ] ~ [ linux-6.8.12 ] ~ [ linux-6.7.12 ] ~ [ linux-6.6.60 ] ~ [ linux-6.5.13 ] ~ [ linux-6.4.16 ] ~ [ linux-6.3.13 ] ~ [ linux-6.2.16 ] ~ [ linux-6.1.116 ] ~ [ linux-6.0.19 ] ~ [ linux-5.19.17 ] ~ [ linux-5.18.19 ] ~ [ linux-5.17.15 ] ~ [ linux-5.16.20 ] ~ [ linux-5.15.171 ] ~ [ linux-5.14.21 ] ~ [ linux-5.13.19 ] ~ [ linux-5.12.19 ] ~ [ linux-5.11.22 ] ~ [ linux-5.10.229 ] ~ [ linux-5.9.16 ] ~ [ linux-5.8.18 ] ~ [ linux-5.7.19 ] ~ [ linux-5.6.19 ] ~ [ linux-5.5.19 ] ~ [ linux-5.4.285 ] ~ [ linux-5.3.18 ] ~ [ linux-5.2.21 ] ~ [ linux-5.1.21 ] ~ [ linux-5.0.21 ] ~ [ linux-4.20.17 ] ~ [ linux-4.19.323 ] ~ [ linux-4.18.20 ] ~ [ linux-4.17.19 ] ~ [ linux-4.16.18 ] ~ [ linux-4.15.18 ] ~ [ linux-4.14.336 ] ~ [ linux-4.13.16 ] ~ [ linux-4.12.14 ] ~ [ linux-4.11.12 ] ~ [ linux-4.10.17 ] ~ [ linux-4.9.337 ] ~ [ linux-4.4.302 ] ~ [ linux-3.10.108 ] ~ [ linux-2.6.32.71 ] ~ [ linux-2.6.0 ] ~ [ linux-2.4.37.11 ] ~ [ unix-v6-master ] ~ [ ccs-tools-1.8.12 ] ~ [ policy-sample ] ~
Architecture: ~ [ i386 ] ~ [ alpha ] ~ [ m68k ] ~ [ mips ] ~ [ ppc ] ~ [ sparc ] ~ [ sparc64 ] ~

Diff markup

Differences between /arch/riscv/lib/memset.S (Version linux-6.12-rc7) and /arch/mips/lib/memset.S (Version linux-5.15.171)


  1 /* SPDX-License-Identifier: GPL-2.0-only */    << 
  2 /*                                                  1 /*
  3  * Copyright (C) 2013 Regents of the Universit !!   2  * This file is subject to the terms and conditions of the GNU General Public
                                                   >>   3  * License.  See the file "COPYING" in the main directory of this archive
                                                   >>   4  * for more details.
                                                   >>   5  *
                                                   >>   6  * Copyright (C) 1998, 1999, 2000 by Ralf Baechle
                                                   >>   7  * Copyright (C) 1999, 2000 Silicon Graphics, Inc.
                                                   >>   8  * Copyright (C) 2007 by Maciej W. Rozycki
                                                   >>   9  * Copyright (C) 2011, 2012 MIPS Technologies, Inc.
  4  */                                                10  */
                                                   >>  11 #include <asm/asm.h>
                                                   >>  12 #include <asm/asm-offsets.h>
                                                   >>  13 #include <asm/export.h>
                                                   >>  14 #include <asm/regdef.h>
                                                   >>  15 
                                                   >>  16 #if LONGSIZE == 4
                                                   >>  17 #define LONG_S_L swl
                                                   >>  18 #define LONG_S_R swr
                                                   >>  19 #else
                                                   >>  20 #define LONG_S_L sdl
                                                   >>  21 #define LONG_S_R sdr
                                                   >>  22 #endif
                                                   >>  23 
                                                   >>  24 #ifdef CONFIG_CPU_MICROMIPS
                                                   >>  25 #define STORSIZE (LONGSIZE * 2)
                                                   >>  26 #define STORMASK (STORSIZE - 1)
                                                   >>  27 #define FILL64RG t8
                                                   >>  28 #define FILLPTRG t7
                                                   >>  29 #undef  LONG_S
                                                   >>  30 #define LONG_S LONG_SP
                                                   >>  31 #else
                                                   >>  32 #define STORSIZE LONGSIZE
                                                   >>  33 #define STORMASK LONGMASK
                                                   >>  34 #define FILL64RG a1
                                                   >>  35 #define FILLPTRG t0
                                                   >>  36 #endif
  5                                                    37 
                                                   >>  38 #define LEGACY_MODE 1
                                                   >>  39 #define EVA_MODE    2
  6                                                    40 
  7 #include <linux/linkage.h>                     !!  41 /*
  8 #include <asm/asm.h>                           !!  42  * No need to protect it with EVA #ifdefery. The generated block of code
                                                   >>  43  * will never be assembled if EVA is not enabled.
                                                   >>  44  */
                                                   >>  45 #define __EVAFY(insn, reg, addr) __BUILD_EVA_INSN(insn##e, reg, addr)
                                                   >>  46 #define ___BUILD_EVA_INSN(insn, reg, addr) __EVAFY(insn, reg, addr)
                                                   >>  47 
                                                   >>  48 #define EX(insn,reg,addr,handler)                       \
                                                   >>  49         .if \mode == LEGACY_MODE;                       \
                                                   >>  50 9:              insn    reg, addr;                      \
                                                   >>  51         .else;                                          \
                                                   >>  52 9:              ___BUILD_EVA_INSN(insn, reg, addr);     \
                                                   >>  53         .endif;                                         \
                                                   >>  54         .section __ex_table,"a";                        \
                                                   >>  55         PTR_WD  9b, handler;                            \
                                                   >>  56         .previous
                                                   >>  57 
                                                   >>  58         .macro  f_fill64 dst, offset, val, fixup, mode
                                                   >>  59         EX(LONG_S, \val, (\offset +  0 * STORSIZE)(\dst), \fixup)
                                                   >>  60         EX(LONG_S, \val, (\offset +  1 * STORSIZE)(\dst), \fixup)
                                                   >>  61         EX(LONG_S, \val, (\offset +  2 * STORSIZE)(\dst), \fixup)
                                                   >>  62         EX(LONG_S, \val, (\offset +  3 * STORSIZE)(\dst), \fixup)
                                                   >>  63 #if ((defined(CONFIG_CPU_MICROMIPS) && (LONGSIZE == 4)) || !defined(CONFIG_CPU_MICROMIPS))
                                                   >>  64         EX(LONG_S, \val, (\offset +  4 * STORSIZE)(\dst), \fixup)
                                                   >>  65         EX(LONG_S, \val, (\offset +  5 * STORSIZE)(\dst), \fixup)
                                                   >>  66         EX(LONG_S, \val, (\offset +  6 * STORSIZE)(\dst), \fixup)
                                                   >>  67         EX(LONG_S, \val, (\offset +  7 * STORSIZE)(\dst), \fixup)
                                                   >>  68 #endif
                                                   >>  69 #if (!defined(CONFIG_CPU_MICROMIPS) && (LONGSIZE == 4))
                                                   >>  70         EX(LONG_S, \val, (\offset +  8 * STORSIZE)(\dst), \fixup)
                                                   >>  71         EX(LONG_S, \val, (\offset +  9 * STORSIZE)(\dst), \fixup)
                                                   >>  72         EX(LONG_S, \val, (\offset + 10 * STORSIZE)(\dst), \fixup)
                                                   >>  73         EX(LONG_S, \val, (\offset + 11 * STORSIZE)(\dst), \fixup)
                                                   >>  74         EX(LONG_S, \val, (\offset + 12 * STORSIZE)(\dst), \fixup)
                                                   >>  75         EX(LONG_S, \val, (\offset + 13 * STORSIZE)(\dst), \fixup)
                                                   >>  76         EX(LONG_S, \val, (\offset + 14 * STORSIZE)(\dst), \fixup)
                                                   >>  77         EX(LONG_S, \val, (\offset + 15 * STORSIZE)(\dst), \fixup)
                                                   >>  78 #endif
                                                   >>  79         .endm
                                                   >>  80 
                                                   >>  81         .align  5
                                                   >>  82 
                                                   >>  83         /*
                                                   >>  84          * Macro to generate the __bzero{,_user} symbol
                                                   >>  85          * Arguments:
                                                   >>  86          * mode: LEGACY_MODE or EVA_MODE
                                                   >>  87          */
                                                   >>  88         .macro __BUILD_BZERO mode
                                                   >>  89         /* Initialize __memset if this is the first time we call this macro */
                                                   >>  90         .ifnotdef __memset
                                                   >>  91         .set __memset, 1
                                                   >>  92         .hidden __memset /* Make sure it does not leak */
                                                   >>  93         .endif
                                                   >>  94 
                                                   >>  95         sltiu           t0, a2, STORSIZE        /* very small region? */
                                                   >>  96         .set            noreorder
                                                   >>  97         bnez            t0, .Lsmall_memset\@
                                                   >>  98          andi           t0, a0, STORMASK        /* aligned? */
                                                   >>  99         .set            reorder
                                                   >> 100 
                                                   >> 101 #ifdef CONFIG_CPU_MICROMIPS
                                                   >> 102         move            t8, a1                  /* used by 'swp' instruction */
                                                   >> 103         move            t9, a1
                                                   >> 104 #endif
                                                   >> 105         .set            noreorder
                                                   >> 106 #ifndef CONFIG_CPU_DADDI_WORKAROUNDS
                                                   >> 107         beqz            t0, 1f
                                                   >> 108          PTR_SUBU       t0, STORSIZE            /* alignment in bytes */
                                                   >> 109 #else
                                                   >> 110         .set            noat
                                                   >> 111         li              AT, STORSIZE
                                                   >> 112         beqz            t0, 1f
                                                   >> 113          PTR_SUBU       t0, AT                  /* alignment in bytes */
                                                   >> 114         .set            at
                                                   >> 115 #endif
                                                   >> 116         .set            reorder
                                                   >> 117 
                                                   >> 118 #ifndef CONFIG_CPU_NO_LOAD_STORE_LR
                                                   >> 119         R10KCBARRIER(0(ra))
                                                   >> 120 #ifdef __MIPSEB__
                                                   >> 121         EX(LONG_S_L, a1, (a0), .Lfirst_fixup\@) /* make word/dword aligned */
                                                   >> 122 #else
                                                   >> 123         EX(LONG_S_R, a1, (a0), .Lfirst_fixup\@) /* make word/dword aligned */
                                                   >> 124 #endif
                                                   >> 125         PTR_SUBU        a0, t0                  /* long align ptr */
                                                   >> 126         PTR_ADDU        a2, t0                  /* correct size */
                                                   >> 127 
                                                   >> 128 #else /* CONFIG_CPU_NO_LOAD_STORE_LR */
                                                   >> 129 #define STORE_BYTE(N)                           \
                                                   >> 130         EX(sb, a1, N(a0), .Lbyte_fixup\@);      \
                                                   >> 131         .set            noreorder;              \
                                                   >> 132         beqz            t0, 0f;                 \
                                                   >> 133          PTR_ADDU       t0, 1;                  \
                                                   >> 134         .set            reorder;
                                                   >> 135 
                                                   >> 136         PTR_ADDU        a2, t0                  /* correct size */
                                                   >> 137         PTR_ADDU        t0, 1
                                                   >> 138         STORE_BYTE(0)
                                                   >> 139         STORE_BYTE(1)
                                                   >> 140 #if LONGSIZE == 4
                                                   >> 141         EX(sb, a1, 2(a0), .Lbyte_fixup\@)
                                                   >> 142 #else
                                                   >> 143         STORE_BYTE(2)
                                                   >> 144         STORE_BYTE(3)
                                                   >> 145         STORE_BYTE(4)
                                                   >> 146         STORE_BYTE(5)
                                                   >> 147         EX(sb, a1, 6(a0), .Lbyte_fixup\@)
                                                   >> 148 #endif
                                                   >> 149 0:
                                                   >> 150         ori             a0, STORMASK
                                                   >> 151         xori            a0, STORMASK
                                                   >> 152         PTR_ADDIU       a0, STORSIZE
                                                   >> 153 #endif /* CONFIG_CPU_NO_LOAD_STORE_LR */
                                                   >> 154 1:      ori             t1, a2, 0x3f            /* # of full blocks */
                                                   >> 155         xori            t1, 0x3f
                                                   >> 156         andi            t0, a2, 0x40-STORSIZE
                                                   >> 157         beqz            t1, .Lmemset_partial\@  /* no block to fill */
                                                   >> 158 
                                                   >> 159         PTR_ADDU        t1, a0                  /* end address */
                                                   >> 160 1:      PTR_ADDIU       a0, 64
                                                   >> 161         R10KCBARRIER(0(ra))
                                                   >> 162         f_fill64 a0, -64, FILL64RG, .Lfwd_fixup\@, \mode
                                                   >> 163         bne             t1, a0, 1b
                                                   >> 164 
                                                   >> 165 .Lmemset_partial\@:
                                                   >> 166         R10KCBARRIER(0(ra))
                                                   >> 167         PTR_LA          t1, 2f                  /* where to start */
                                                   >> 168 #ifdef CONFIG_CPU_MICROMIPS
                                                   >> 169         LONG_SRL        t7, t0, 1
                                                   >> 170 #endif
                                                   >> 171 #if LONGSIZE == 4
                                                   >> 172         PTR_SUBU        t1, FILLPTRG
                                                   >> 173 #else
                                                   >> 174         .set            noat
                                                   >> 175         LONG_SRL        AT, FILLPTRG, 1
                                                   >> 176         PTR_SUBU        t1, AT
                                                   >> 177         .set            at
                                                   >> 178 #endif
                                                   >> 179         PTR_ADDU        a0, t0                  /* dest ptr */
                                                   >> 180         jr              t1
                                                   >> 181 
                                                   >> 182         /* ... but first do longs ... */
                                                   >> 183         f_fill64 a0, -64, FILL64RG, .Lpartial_fixup\@, \mode
                                                   >> 184 2:      andi            a2, STORMASK            /* At most one long to go */
                                                   >> 185 
                                                   >> 186         .set            noreorder
                                                   >> 187         beqz            a2, 1f
                                                   >> 188 #ifndef CONFIG_CPU_NO_LOAD_STORE_LR
                                                   >> 189          PTR_ADDU       a0, a2                  /* What's left */
                                                   >> 190         .set            reorder
                                                   >> 191         R10KCBARRIER(0(ra))
                                                   >> 192 #ifdef __MIPSEB__
                                                   >> 193         EX(LONG_S_R, a1, -1(a0), .Llast_fixup\@)
                                                   >> 194 #else
                                                   >> 195         EX(LONG_S_L, a1, -1(a0), .Llast_fixup\@)
                                                   >> 196 #endif
                                                   >> 197 #else /* CONFIG_CPU_NO_LOAD_STORE_LR */
                                                   >> 198          PTR_SUBU       t0, $0, a2
                                                   >> 199         .set            reorder
                                                   >> 200         move            a2, zero                /* No remaining longs */
                                                   >> 201         PTR_ADDIU       t0, 1
                                                   >> 202         STORE_BYTE(0)
                                                   >> 203         STORE_BYTE(1)
                                                   >> 204 #if LONGSIZE == 4
                                                   >> 205         EX(sb, a1, 2(a0), .Lbyte_fixup\@)
                                                   >> 206 #else
                                                   >> 207         STORE_BYTE(2)
                                                   >> 208         STORE_BYTE(3)
                                                   >> 209         STORE_BYTE(4)
                                                   >> 210         STORE_BYTE(5)
                                                   >> 211         EX(sb, a1, 6(a0), .Lbyte_fixup\@)
                                                   >> 212 #endif
                                                   >> 213 0:
                                                   >> 214 #endif /* CONFIG_CPU_NO_LOAD_STORE_LR */
                                                   >> 215 1:      move            a2, zero
                                                   >> 216         jr              ra
                                                   >> 217 
                                                   >> 218 .Lsmall_memset\@:
                                                   >> 219         PTR_ADDU        t1, a0, a2
                                                   >> 220         beqz            a2, 2f
                                                   >> 221 
                                                   >> 222 1:      PTR_ADDIU       a0, 1                   /* fill bytewise */
                                                   >> 223         R10KCBARRIER(0(ra))
                                                   >> 224         .set            noreorder
                                                   >> 225         bne             t1, a0, 1b
                                                   >> 226          EX(sb, a1, -1(a0), .Lsmall_fixup\@)
                                                   >> 227         .set            reorder
                                                   >> 228 
                                                   >> 229 2:      move            a2, zero
                                                   >> 230         jr              ra                      /* done */
                                                   >> 231         .if __memset == 1
                                                   >> 232         END(memset)
                                                   >> 233         .set __memset, 0
                                                   >> 234         .hidden __memset
                                                   >> 235         .endif
  9                                                   236 
 10 /* void *memset(void *, int, size_t) */        !! 237 #ifdef CONFIG_CPU_NO_LOAD_STORE_LR
 11 SYM_FUNC_START(__memset)                       !! 238 .Lbyte_fixup\@:
 12         move t0, a0  /* Preserve return value  !! 239         /*
 13                                                !! 240          * unset_bytes = (#bytes - (#unaligned bytes)) - (-#unaligned bytes remaining + 1) + 1
 14         /* Defer to byte-oriented fill for sma !! 241          *      a2     =             a2                -              t0                   + 1
 15         sltiu a3, a2, 16                       !! 242          */
 16         bnez a3, 4f                            !! 243         PTR_SUBU        a2, t0
                                                   >> 244         PTR_ADDIU       a2, 1
                                                   >> 245         jr              ra
                                                   >> 246 #endif /* CONFIG_CPU_NO_LOAD_STORE_LR */
                                                   >> 247 
                                                   >> 248 .Lfirst_fixup\@:
                                                   >> 249         /* unset_bytes already in a2 */
                                                   >> 250         jr      ra
 17                                                   251 
                                                   >> 252 .Lfwd_fixup\@:
 18         /*                                        253         /*
 19          * Round to nearest XLEN-aligned addre !! 254          * unset_bytes = partial_start_addr +  #bytes   -     fault_addr
 20          * greater than or equal to start addr !! 255          *      a2     =         t1         + (a2 & 3f) - $28->task->BUADDR
 21          */                                       256          */
 22         addi a3, t0, SZREG-1                   !! 257         PTR_L           t0, TI_TASK($28)
 23         andi a3, a3, ~(SZREG-1)                !! 258         andi            a2, 0x3f
 24         beq a3, t0, 2f  /* Skip if already ali !! 259         LONG_L          t0, THREAD_BUADDR(t0)
 25         /* Handle initial misalignment */      !! 260         LONG_ADDU       a2, t1
 26         sub a4, a3, t0                         !! 261         LONG_SUBU       a2, t0
                                                   >> 262         jr              ra
                                                   >> 263 
                                                   >> 264 .Lpartial_fixup\@:
                                                   >> 265         /*
                                                   >> 266          * unset_bytes = partial_end_addr +      #bytes     -     fault_addr
                                                   >> 267          *      a2     =       a0         + (a2 & STORMASK) - $28->task->BUADDR
                                                   >> 268          */
                                                   >> 269         PTR_L           t0, TI_TASK($28)
                                                   >> 270         andi            a2, STORMASK
                                                   >> 271         LONG_L          t0, THREAD_BUADDR(t0)
                                                   >> 272         LONG_ADDU       a2, a0
                                                   >> 273         LONG_SUBU       a2, t0
                                                   >> 274         jr              ra
                                                   >> 275 
                                                   >> 276 .Llast_fixup\@:
                                                   >> 277         /* unset_bytes already in a2 */
                                                   >> 278         jr              ra
                                                   >> 279 
                                                   >> 280 .Lsmall_fixup\@:
                                                   >> 281         /*
                                                   >> 282          * unset_bytes = end_addr - current_addr + 1
                                                   >> 283          *      a2     =    t1    -      a0      + 1
                                                   >> 284          */
                                                   >> 285         PTR_SUBU        a2, t1, a0
                                                   >> 286         PTR_ADDIU       a2, 1
                                                   >> 287         jr              ra
                                                   >> 288 
                                                   >> 289         .endm
                                                   >> 290 
                                                   >> 291 /*
                                                   >> 292  * memset(void *s, int c, size_t n)
                                                   >> 293  *
                                                   >> 294  * a0: start of area to clear
                                                   >> 295  * a1: char to fill with
                                                   >> 296  * a2: size of area to clear
                                                   >> 297  */
                                                   >> 298 
                                                   >> 299 LEAF(memset)
                                                   >> 300 EXPORT_SYMBOL(memset)
                                                   >> 301         move            v0, a0                  /* result */
                                                   >> 302         beqz            a1, 1f
                                                   >> 303 
                                                   >> 304         andi            a1, 0xff                /* spread fillword */
                                                   >> 305         LONG_SLL                t1, a1, 8
                                                   >> 306         or              a1, t1
                                                   >> 307         LONG_SLL                t1, a1, 16
                                                   >> 308 #if LONGSIZE == 8
                                                   >> 309         or              a1, t1
                                                   >> 310         LONG_SLL                t1, a1, 32
                                                   >> 311 #endif
                                                   >> 312         or              a1, t1
 27 1:                                                313 1:
 28         sb a1, 0(t0)                           !! 314 #ifndef CONFIG_EVA
 29         addi t0, t0, 1                         !! 315 FEXPORT(__bzero)
 30         bltu t0, a3, 1b                        !! 316 EXPORT_SYMBOL(__bzero)
 31         sub a2, a2, a4  /* Update count */     !! 317 #endif
 32                                                !! 318         __BUILD_BZERO LEGACY_MODE
 33 2: /* Duff's device with 32 XLEN stores per it !! 319 
 34         /* Broadcast value into all bytes */   !! 320 #ifdef CONFIG_EVA
 35         andi a1, a1, 0xff                      !! 321 LEAF(__bzero)
 36         slli a3, a1, 8                         !! 322 EXPORT_SYMBOL(__bzero)
 37         or a1, a3, a1                          !! 323         __BUILD_BZERO EVA_MODE
 38         slli a3, a1, 16                        !! 324 END(__bzero)
 39         or a1, a3, a1                          !! 325 #endif
 40 #ifdef CONFIG_64BIT                            << 
 41         slli a3, a1, 32                        << 
 42         or a1, a3, a1                          << 
 43 #endif                                         << 
 44                                                << 
 45         /* Calculate end address */            << 
 46         andi a4, a2, ~(SZREG-1)                << 
 47         add a3, t0, a4                         << 
 48                                                << 
 49         andi a4, a4, 31*SZREG  /* Calculate re << 
 50         beqz a4, 3f            /* Shortcut if  << 
 51         neg a4, a4                             << 
 52         addi a4, a4, 32*SZREG  /* Calculate in << 
 53                                                << 
 54         /* Adjust start address with offset */ << 
 55         sub t0, t0, a4                         << 
 56                                                << 
 57         /* Jump into loop body */              << 
 58         /* Assumes 32-bit instruction lengths  << 
 59         la a5, 3f                              << 
 60 #ifdef CONFIG_64BIT                            << 
 61         srli a4, a4, 1                         << 
 62 #endif                                         << 
 63         add a5, a5, a4                         << 
 64         jr a5                                  << 
 65 3:                                             << 
 66         REG_S a1,        0(t0)                 << 
 67         REG_S a1,    SZREG(t0)                 << 
 68         REG_S a1,  2*SZREG(t0)                 << 
 69         REG_S a1,  3*SZREG(t0)                 << 
 70         REG_S a1,  4*SZREG(t0)                 << 
 71         REG_S a1,  5*SZREG(t0)                 << 
 72         REG_S a1,  6*SZREG(t0)                 << 
 73         REG_S a1,  7*SZREG(t0)                 << 
 74         REG_S a1,  8*SZREG(t0)                 << 
 75         REG_S a1,  9*SZREG(t0)                 << 
 76         REG_S a1, 10*SZREG(t0)                 << 
 77         REG_S a1, 11*SZREG(t0)                 << 
 78         REG_S a1, 12*SZREG(t0)                 << 
 79         REG_S a1, 13*SZREG(t0)                 << 
 80         REG_S a1, 14*SZREG(t0)                 << 
 81         REG_S a1, 15*SZREG(t0)                 << 
 82         REG_S a1, 16*SZREG(t0)                 << 
 83         REG_S a1, 17*SZREG(t0)                 << 
 84         REG_S a1, 18*SZREG(t0)                 << 
 85         REG_S a1, 19*SZREG(t0)                 << 
 86         REG_S a1, 20*SZREG(t0)                 << 
 87         REG_S a1, 21*SZREG(t0)                 << 
 88         REG_S a1, 22*SZREG(t0)                 << 
 89         REG_S a1, 23*SZREG(t0)                 << 
 90         REG_S a1, 24*SZREG(t0)                 << 
 91         REG_S a1, 25*SZREG(t0)                 << 
 92         REG_S a1, 26*SZREG(t0)                 << 
 93         REG_S a1, 27*SZREG(t0)                 << 
 94         REG_S a1, 28*SZREG(t0)                 << 
 95         REG_S a1, 29*SZREG(t0)                 << 
 96         REG_S a1, 30*SZREG(t0)                 << 
 97         REG_S a1, 31*SZREG(t0)                 << 
 98         addi t0, t0, 32*SZREG                  << 
 99         bltu t0, a3, 3b                        << 
100         andi a2, a2, SZREG-1  /* Update count  << 
101                                                << 
102 4:                                             << 
103         /* Handle trailing misalignment */     << 
104         beqz a2, 6f                            << 
105         add a3, t0, a2                         << 
106 5:                                             << 
107         sb a1, 0(t0)                           << 
108         addi t0, t0, 1                         << 
109         bltu t0, a3, 5b                        << 
110 6:                                             << 
111         ret                                    << 
112 SYM_FUNC_END(__memset)                         << 
113 SYM_FUNC_ALIAS_WEAK(memset, __memset)          << 
114 SYM_FUNC_ALIAS(__pi_memset, __memset)          << 
115 SYM_FUNC_ALIAS(__pi___memset, __memset)        << 
                                                      

~ [ source navigation ] ~ [ diff markup ] ~ [ identifier search ] ~

kernel.org | git.kernel.org | LWN.net | Project Home | SVN repository | Mail admin

Linux® is a registered trademark of Linus Torvalds in the United States and other countries.
TOMOYO® is a registered trademark of NTT DATA CORPORATION.

sflogo.php