~ [ source navigation ] ~ [ diff markup ] ~ [ identifier search ] ~

TOMOYO Linux Cross Reference
Linux/arch/arm64/lib/memset.S

Version: ~ [ linux-6.12-rc7 ] ~ [ linux-6.11.7 ] ~ [ linux-6.10.14 ] ~ [ linux-6.9.12 ] ~ [ linux-6.8.12 ] ~ [ linux-6.7.12 ] ~ [ linux-6.6.60 ] ~ [ linux-6.5.13 ] ~ [ linux-6.4.16 ] ~ [ linux-6.3.13 ] ~ [ linux-6.2.16 ] ~ [ linux-6.1.116 ] ~ [ linux-6.0.19 ] ~ [ linux-5.19.17 ] ~ [ linux-5.18.19 ] ~ [ linux-5.17.15 ] ~ [ linux-5.16.20 ] ~ [ linux-5.15.171 ] ~ [ linux-5.14.21 ] ~ [ linux-5.13.19 ] ~ [ linux-5.12.19 ] ~ [ linux-5.11.22 ] ~ [ linux-5.10.229 ] ~ [ linux-5.9.16 ] ~ [ linux-5.8.18 ] ~ [ linux-5.7.19 ] ~ [ linux-5.6.19 ] ~ [ linux-5.5.19 ] ~ [ linux-5.4.285 ] ~ [ linux-5.3.18 ] ~ [ linux-5.2.21 ] ~ [ linux-5.1.21 ] ~ [ linux-5.0.21 ] ~ [ linux-4.20.17 ] ~ [ linux-4.19.323 ] ~ [ linux-4.18.20 ] ~ [ linux-4.17.19 ] ~ [ linux-4.16.18 ] ~ [ linux-4.15.18 ] ~ [ linux-4.14.336 ] ~ [ linux-4.13.16 ] ~ [ linux-4.12.14 ] ~ [ linux-4.11.12 ] ~ [ linux-4.10.17 ] ~ [ linux-4.9.337 ] ~ [ linux-4.4.302 ] ~ [ linux-3.10.108 ] ~ [ linux-2.6.32.71 ] ~ [ linux-2.6.0 ] ~ [ linux-2.4.37.11 ] ~ [ unix-v6-master ] ~ [ ccs-tools-1.8.12 ] ~ [ policy-sample ] ~
Architecture: ~ [ i386 ] ~ [ alpha ] ~ [ m68k ] ~ [ mips ] ~ [ ppc ] ~ [ sparc ] ~ [ sparc64 ] ~

Diff markup

Differences between /arch/arm64/lib/memset.S (Version linux-6.12-rc7) and /arch/sparc/lib/memset.S (Version linux-2.6.32.71)


  1 /* SPDX-License-Identifier: GPL-2.0-only */    !!   1 /* linux/arch/sparc/lib/memset.S: Sparc optimized memset, bzero and clear_user code
  2 /*                                             !!   2  * Copyright (C) 1991,1996 Free Software Foundation
  3  * Copyright (C) 2013 ARM Ltd.                 !!   3  * Copyright (C) 1996,1997 Jakub Jelinek (jj@sunsite.mff.cuni.cz)
  4  * Copyright (C) 2013 Linaro.                  !!   4  * Copyright (C) 1996 David S. Miller (davem@caip.rutgers.edu)
  5  *                                                  5  *
  6  * This code is based on glibc cortex strings  !!   6  * Returns 0, if ok, and number of bytes not yet set if exception
  7  * be found @                                  !!   7  * occurs and we were called as clear_user.
  8  *                                             << 
  9  * http://bazaar.launchpad.net/~linaro-toolcha << 
 10  * files/head:/src/aarch64/                    << 
 11  */                                                 8  */
 12                                                     9 
 13 #include <linux/linkage.h>                     !!  10 #include <asm/ptrace.h>
 14 #include <asm/assembler.h>                     << 
 15 #include <asm/cache.h>                         << 
 16                                                    11 
 17 /*                                             !!  12 /* Work around cpp -rob */
 18  * Fill in the buffer with character c (alignm !!  13 #define ALLOC #alloc
 19  *                                             !!  14 #define EXECINSTR #execinstr
 20  * Parameters:                                 !!  15 #define EX(x,y,a,b)                             \
 21  *      x0 - buf                               !!  16 98:     x,y;                                    \
 22  *      x1 - c                                 !!  17         .section .fixup,ALLOC,EXECINSTR;        \
 23  *      x2 - n                                 !!  18         .align  4;                              \
 24  * Returns:                                    !!  19 99:     ba 30f;                                 \
 25  *      x0 - buf                               !!  20          a, b, %o0;                             \
 26  */                                            !!  21         .section __ex_table,ALLOC;              \
                                                   >>  22         .align  4;                              \
                                                   >>  23         .word   98b, 99b;                       \
                                                   >>  24         .text;                                  \
                                                   >>  25         .align  4
                                                   >>  26 
                                                   >>  27 #define EXT(start,end,handler)                  \
                                                   >>  28         .section __ex_table,ALLOC;              \
                                                   >>  29         .align  4;                              \
                                                   >>  30         .word   start, 0, end, handler;         \
                                                   >>  31         .text;                                  \
                                                   >>  32         .align  4
                                                   >>  33 
                                                   >>  34 /* Please don't change these macros, unless you change the logic
                                                   >>  35  * in the .fixup section below as well.
                                                   >>  36  * Store 64 bytes at (BASE + OFFSET) using value SOURCE. */
                                                   >>  37 #define ZERO_BIG_BLOCK(base, offset, source)    \
                                                   >>  38         std     source, [base + offset + 0x00]; \
                                                   >>  39         std     source, [base + offset + 0x08]; \
                                                   >>  40         std     source, [base + offset + 0x10]; \
                                                   >>  41         std     source, [base + offset + 0x18]; \
                                                   >>  42         std     source, [base + offset + 0x20]; \
                                                   >>  43         std     source, [base + offset + 0x28]; \
                                                   >>  44         std     source, [base + offset + 0x30]; \
                                                   >>  45         std     source, [base + offset + 0x38];
                                                   >>  46 
                                                   >>  47 #define ZERO_LAST_BLOCKS(base, offset, source)  \
                                                   >>  48         std     source, [base - offset - 0x38]; \
                                                   >>  49         std     source, [base - offset - 0x30]; \
                                                   >>  50         std     source, [base - offset - 0x28]; \
                                                   >>  51         std     source, [base - offset - 0x20]; \
                                                   >>  52         std     source, [base - offset - 0x18]; \
                                                   >>  53         std     source, [base - offset - 0x10]; \
                                                   >>  54         std     source, [base - offset - 0x08]; \
                                                   >>  55         std     source, [base - offset - 0x00];
                                                   >>  56 
                                                   >>  57         .text
                                                   >>  58         .align 4
                                                   >>  59 
                                                   >>  60         .globl  __bzero_begin
                                                   >>  61 __bzero_begin:
                                                   >>  62 
                                                   >>  63         .globl  __bzero, __memset, 
                                                   >>  64         .globl  memset
                                                   >>  65         .globl  __memset_start, __memset_end
                                                   >>  66 __memset_start:
                                                   >>  67 __memset:
                                                   >>  68 memset:
                                                   >>  69         and     %o1, 0xff, %g3
                                                   >>  70         sll     %g3, 8, %g2
                                                   >>  71         or      %g3, %g2, %g3
                                                   >>  72         sll     %g3, 16, %g2
                                                   >>  73         or      %g3, %g2, %g3
                                                   >>  74         b       1f
                                                   >>  75          mov    %o2, %o1
                                                   >>  76 3:
                                                   >>  77         cmp     %o2, 3
                                                   >>  78         be      2f
                                                   >>  79          EX(stb %g3, [%o0], sub %o1, 0)
                                                   >>  80 
                                                   >>  81         cmp     %o2, 2
                                                   >>  82         be      2f
                                                   >>  83          EX(stb %g3, [%o0 + 0x01], sub %o1, 1)
 27                                                    84 
 28 dstin           .req    x0                     !!  85         EX(stb  %g3, [%o0 + 0x02], sub %o1, 2)
 29 val             .req    w1                     << 
 30 count           .req    x2                     << 
 31 tmp1            .req    x3                     << 
 32 tmp1w           .req    w3                     << 
 33 tmp2            .req    x4                     << 
 34 tmp2w           .req    w4                     << 
 35 zva_len_x       .req    x5                     << 
 36 zva_len         .req    w5                     << 
 37 zva_bits_x      .req    x6                     << 
 38                                                << 
 39 A_l             .req    x7                     << 
 40 A_lw            .req    w7                     << 
 41 dst             .req    x8                     << 
 42 tmp3w           .req    w9                     << 
 43 tmp3            .req    x9                     << 
 44                                                << 
 45 SYM_FUNC_START(__pi_memset)                    << 
 46         mov     dst, dstin      /* Preserve re << 
 47         and     A_lw, val, #255                << 
 48         orr     A_lw, A_lw, A_lw, lsl #8       << 
 49         orr     A_lw, A_lw, A_lw, lsl #16      << 
 50         orr     A_l, A_l, A_l, lsl #32         << 
 51                                                << 
 52         cmp     count, #15                     << 
 53         b.hi    .Lover16_proc                  << 
 54         /*All store maybe are non-aligned..*/  << 
 55         tbz     count, #3, 1f                  << 
 56         str     A_l, [dst], #8                 << 
 57 1:                                             << 
 58         tbz     count, #2, 2f                  << 
 59         str     A_lw, [dst], #4                << 
 60 2:                                                 86 2:
 61         tbz     count, #1, 3f                  !!  87         sub     %o2, 4, %o2
 62         strh    A_lw, [dst], #2                !!  88         add     %o1, %o2, %o1
 63 3:                                             !!  89         b       4f
 64         tbz     count, #0, 4f                  !!  90          sub    %o0, %o2, %o0
 65         strb    A_lw, [dst]                    << 
 66 4:                                             << 
 67         ret                                    << 
 68                                                    91 
 69 .Lover16_proc:                                 !!  92 __bzero:
 70         /*Whether  the start address is aligne !!  93         mov     %g0, %g3
 71         neg     tmp2, dst                      << 
 72         ands    tmp2, tmp2, #15                << 
 73         b.eq    .Laligned                      << 
 74 /*                                             << 
 75 * The count is not less than 16, we can use st << 
 76 * then adjust the dst aligned with 16.This pro << 
 77 * memory address at alignment boundary.        << 
 78 */                                             << 
 79         stp     A_l, A_l, [dst] /*non-aligned  << 
 80         /*make the dst aligned..*/             << 
 81         sub     count, count, tmp2             << 
 82         add     dst, dst, tmp2                 << 
 83                                                << 
 84 .Laligned:                                     << 
 85         cbz     A_l, .Lzero_mem                << 
 86                                                << 
 87 .Ltail_maybe_long:                             << 
 88         cmp     count, #64                     << 
 89         b.ge    .Lnot_short                    << 
 90 .Ltail63:                                      << 
 91         ands    tmp1, count, #0x30             << 
 92         b.eq    3f                             << 
 93         cmp     tmp1w, #0x20                   << 
 94         b.eq    1f                             << 
 95         b.lt    2f                             << 
 96         stp     A_l, A_l, [dst], #16           << 
 97 1:                                                 94 1:
 98         stp     A_l, A_l, [dst], #16           !!  95         cmp     %o1, 7
 99 2:                                             !!  96         bleu    7f
100         stp     A_l, A_l, [dst], #16           !!  97          andcc  %o0, 3, %o2
101 /*                                             !!  98 
102 * The last store length is less than 16,use st !!  99         bne     3b
103 * It will lead some bytes written twice and th << 
104 */                                             << 
105 3:                                             << 
106         ands    count, count, #15              << 
107         cbz     count, 4f                      << 
108         add     dst, dst, count                << 
109         stp     A_l, A_l, [dst, #-16]   /* Rep << 
110 4:                                                100 4:
111         ret                                    !! 101          andcc  %o0, 4, %g0
                                                   >> 102 
                                                   >> 103         be      2f
                                                   >> 104          mov    %g3, %g2
                                                   >> 105 
                                                   >> 106         EX(st   %g3, [%o0], sub %o1, 0)
                                                   >> 107         sub     %o1, 4, %o1
                                                   >> 108         add     %o0, 4, %o0
                                                   >> 109 2:
                                                   >> 110         andcc   %o1, 0xffffff80, %o3    ! Now everything is 8 aligned and o1 is len to run
                                                   >> 111         be      9f
                                                   >> 112          andcc  %o1, 0x78, %o2
                                                   >> 113 10:
                                                   >> 114         ZERO_BIG_BLOCK(%o0, 0x00, %g2)
                                                   >> 115         subcc   %o3, 128, %o3
                                                   >> 116         ZERO_BIG_BLOCK(%o0, 0x40, %g2)
                                                   >> 117 11:
                                                   >> 118         EXT(10b, 11b, 20f)
                                                   >> 119         bne     10b
                                                   >> 120          add    %o0, 128, %o0
                                                   >> 121 
                                                   >> 122         orcc    %o2, %g0, %g0
                                                   >> 123 9:
                                                   >> 124         be      13f
                                                   >> 125          andcc  %o1, 7, %o1
                                                   >> 126 
                                                   >> 127         srl     %o2, 1, %o3
                                                   >> 128         set     13f, %o4
                                                   >> 129         sub     %o4, %o3, %o4
                                                   >> 130         jmp     %o4
                                                   >> 131          add    %o0, %o2, %o0
                                                   >> 132 
                                                   >> 133 12:
                                                   >> 134         ZERO_LAST_BLOCKS(%o0, 0x48, %g2)
                                                   >> 135         ZERO_LAST_BLOCKS(%o0, 0x08, %g2)
                                                   >> 136 13:
                                                   >> 137         be      8f
                                                   >> 138          andcc  %o1, 4, %g0
                                                   >> 139 
                                                   >> 140         be      1f
                                                   >> 141          andcc  %o1, 2, %g0
112                                                   142 
113         /*                                     !! 143         EX(st   %g3, [%o0], and %o1, 7)
114         * Critical loop. Start at a new cache  !! 144         add     %o0, 4, %o0
115         * 64 bytes per line, this ensures the  << 
116         */                                     << 
117         .p2align        L1_CACHE_SHIFT         << 
118 .Lnot_short:                                   << 
119         sub     dst, dst, #16/* Pre-bias.  */  << 
120         sub     count, count, #64              << 
121 1:                                                145 1:
122         stp     A_l, A_l, [dst, #16]           !! 146         be      1f
123         stp     A_l, A_l, [dst, #32]           !! 147          andcc  %o1, 1, %g0
124         stp     A_l, A_l, [dst, #48]           << 
125         stp     A_l, A_l, [dst, #64]!          << 
126         subs    count, count, #64              << 
127         b.ge    1b                             << 
128         tst     count, #0x3f                   << 
129         add     dst, dst, #16                  << 
130         b.ne    .Ltail63                       << 
131 .Lexitfunc:                                    << 
132         ret                                    << 
133                                                   148 
134         /*                                     !! 149         EX(sth  %g3, [%o0], and %o1, 3)
135         * For zeroing memory, check to see if  !! 150         add     %o0, 2, %o0
136         * zero entire 'cache' lines.           << 
137         */                                     << 
138 .Lzero_mem:                                    << 
139         cmp     count, #63                     << 
140         b.le    .Ltail63                       << 
141         /*                                     << 
142         * For zeroing small amounts of memory, << 
143         * the line-clear code.                 << 
144         */                                     << 
145         cmp     count, #128                    << 
146         b.lt    .Lnot_short /*count is at leas << 
147                                                << 
148         mrs     tmp1, dczid_el0                << 
149         tbnz    tmp1, #4, .Lnot_short          << 
150         mov     tmp3w, #4                      << 
151         and     zva_len, tmp1w, #15     /* Saf << 
152         lsl     zva_len, tmp3w, zva_len        << 
153                                                << 
154         ands    tmp3w, zva_len, #63            << 
155         /*                                     << 
156         * ensure the zva_len is not less than  << 
157         * It is not meaningful to use ZVA if t << 
158         */                                     << 
159         b.ne    .Lnot_short                    << 
160 .Lzero_by_line:                                << 
161         /*                                     << 
162         * Compute how far we need to go to bec << 
163         * already at quad-word alignment.      << 
164         */                                     << 
165         cmp     count, zva_len_x               << 
166         b.lt    .Lnot_short             /* Not << 
167         sub     zva_bits_x, zva_len_x, #1      << 
168         neg     tmp2, dst                      << 
169         ands    tmp2, tmp2, zva_bits_x         << 
170         b.eq    2f                      /* Alr << 
171         /* Not aligned, check that there's eno << 
172         sub     tmp1, count, tmp2              << 
173         /*                                     << 
174         * grantee the remain length to be ZVA  << 
175         * avoid to make the 2f's process over  << 
176         cmp     tmp1, #64                      << 
177         ccmp    tmp1, zva_len_x, #8, ge /* NZC << 
178         b.lt    .Lnot_short                    << 
179         /*                                     << 
180         * We know that there's at least 64 byt << 
181         * to overrun by 64 bytes.              << 
182         */                                     << 
183         mov     count, tmp1                    << 
184 1:                                                151 1:
185         stp     A_l, A_l, [dst]                !! 152         bne,a   8f
186         stp     A_l, A_l, [dst, #16]           !! 153          EX(stb %g3, [%o0], and %o1, 1)
187         stp     A_l, A_l, [dst, #32]           !! 154 8:
188         subs    tmp2, tmp2, #64                !! 155         retl
189         stp     A_l, A_l, [dst, #48]           !! 156          clr    %o0
190         add     dst, dst, #64                  !! 157 7:
191         b.ge    1b                             !! 158         be      13b
192         /* We've overrun a bit, so adjust dst  !! 159          orcc   %o1, 0, %g0
193         add     dst, dst, tmp2                 !! 160 
194 2:                                             !! 161         be      0f
195         sub     count, count, zva_len_x        !! 162 8:
196 3:                                             !! 163          add    %o0, 1, %o0
197         dc      zva, dst                       !! 164         subcc   %o1, 1, %o1
198         add     dst, dst, zva_len_x            !! 165         bne     8b
199         subs    count, count, zva_len_x        !! 166          EX(stb %g3, [%o0 - 1], add %o1, 1)
200         b.ge    3b                             !! 167 0:
201         ands    count, count, zva_bits_x       !! 168         retl
202         b.ne    .Ltail_maybe_long              !! 169          clr    %o0
                                                   >> 170 __memset_end:
                                                   >> 171 
                                                   >> 172         .section .fixup,#alloc,#execinstr
                                                   >> 173         .align  4
                                                   >> 174 20:
                                                   >> 175         cmp     %g2, 8
                                                   >> 176         bleu    1f
                                                   >> 177          and    %o1, 0x7f, %o1
                                                   >> 178         sub     %g2, 9, %g2
                                                   >> 179         add     %o3, 64, %o3
                                                   >> 180 1:
                                                   >> 181         sll     %g2, 3, %g2
                                                   >> 182         add     %o3, %o1, %o0
                                                   >> 183         b 30f
                                                   >> 184          sub    %o0, %g2, %o0
                                                   >> 185 21:
                                                   >> 186         mov     8, %o0
                                                   >> 187         and     %o1, 7, %o1
                                                   >> 188         sub     %o0, %g2, %o0
                                                   >> 189         sll     %o0, 3, %o0
                                                   >> 190         b 30f
                                                   >> 191          add    %o0, %o1, %o0
                                                   >> 192 30:
                                                   >> 193 /* %o4 is faulting address, %o5 is %pc where fault occurred */
                                                   >> 194         save    %sp, -104, %sp
                                                   >> 195         mov     %i5, %o0
                                                   >> 196         mov     %i7, %o1
                                                   >> 197         call    lookup_fault
                                                   >> 198          mov    %i4, %o2
203         ret                                       199         ret
204 SYM_FUNC_END(__pi_memset)                      !! 200          restore
205                                                << 
206 SYM_FUNC_ALIAS(__memset, __pi_memset)          << 
207 EXPORT_SYMBOL(__memset)                        << 
208                                                   201 
209 SYM_FUNC_ALIAS_WEAK(memset, __pi_memset)       !! 202         .globl __bzero_end
210 EXPORT_SYMBOL(memset)                          !! 203 __bzero_end:
                                                      

~ [ source navigation ] ~ [ diff markup ] ~ [ identifier search ] ~

kernel.org | git.kernel.org | LWN.net | Project Home | SVN repository | Mail admin

Linux® is a registered trademark of Linus Torvalds in the United States and other countries.
TOMOYO® is a registered trademark of NTT DATA CORPORATION.

sflogo.php