1 /* SPDX-License-Identifier: GPL-2.0 */ 1 /* SPDX-License-Identifier: GPL-2.0 */ 2 /* strlen.S: Sparc optimized strlen code !! 2 /* 3 * Hand optimized from GNU libc's strlen !! 3 * strlen.S (c) 1995 David Mosberger (davidm@cs.arizona.edu) 4 * Copyright (C) 1991,1996 Free Software Found !! 4 * 5 * Copyright (C) 1996,2008 David S. Miller (da !! 5 * Finds length of a 0-terminated string. Optimized for the 6 * Copyright (C) 1996, 1997 Jakub Jelinek (jj@ !! 6 * Alpha architecture: >> 7 * >> 8 * - memory accessed as aligned quadwords only >> 9 * - uses bcmpge to compare 8 bytes in parallel >> 10 * - does binary search to find 0 byte in last >> 11 * quadword (HAKMEM needed 12 instructions to >> 12 * do this instead of the 9 instructions that >> 13 * binary search needs). 7 */ 14 */ >> 15 #include <asm/export.h> >> 16 .set noreorder >> 17 .set noat 8 18 9 #include <linux/export.h> !! 19 .align 3 10 #include <linux/linkage.h> !! 20 11 #include <asm/asm.h> !! 21 .globl strlen 12 !! 22 .ent strlen 13 #define LO_MAGIC 0x01010101 !! 23 14 #define HI_MAGIC 0x80808080 !! 24 strlen: 15 !! 25 ldq_u $1, 0($16) # load first quadword ($16 may be misaligned) 16 .text !! 26 lda $2, -1($31) 17 ENTRY(strlen) !! 27 insqh $2, $16, $2 18 mov %o0, %o1 !! 28 andnot $16, 7, $0 19 andcc %o0, 3, %g0 !! 29 or $2, $1, $1 20 BRANCH32(be, pt, 9f) !! 30 cmpbge $31, $1, $2 # $2 <- bitmask: bit i == 1 <==> i-th byte == 0 21 sethi %hi(HI_MAGIC), %o4 !! 31 bne $2, found 22 ldub [%o0], %o5 !! 32 23 BRANCH_REG_ZERO(pn, %o5, 11f) !! 33 loop: ldq $1, 8($0) 24 add %o0, 1, %o0 !! 34 addq $0, 8, $0 # addr += 8 25 andcc %o0, 3, %g0 !! 35 nop # helps dual issue last two insns 26 BRANCH32(be, pn, 4f) !! 36 cmpbge $31, $1, $2 27 or %o4, %lo(HI_MAGIC), %o3 !! 37 beq $2, loop 28 ldub [%o0], %o5 !! 38 29 BRANCH_REG_ZERO(pn, %o5, 12f) !! 39 found: blbs $2, done # make aligned case fast 30 add %o0, 1, %o0 !! 40 negq $2, $3 31 andcc %o0, 3, %g0 !! 41 and $2, $3, $2 32 BRANCH32(be, pt, 5f) !! 42 33 sethi %hi(LO_MAGIC), %o4 !! 43 and $2, 0x0f, $1 34 ldub [%o0], %o5 !! 44 addq $0, 4, $3 35 BRANCH_REG_ZERO(pn, %o5, 13f) !! 45 cmoveq $1, $3, $0 36 add %o0, 1, %o0 !! 46 37 BRANCH32(ba, pt, 8f) !! 47 and $2, 0x33, $1 38 or %o4, %lo(LO_MAGIC), %o2 !! 48 addq $0, 2, $3 39 9: !! 49 cmoveq $1, $3, $0 40 or %o4, %lo(HI_MAGIC), %o3 !! 50 41 4: !! 51 and $2, 0x55, $1 42 sethi %hi(LO_MAGIC), %o4 !! 52 addq $0, 1, $3 43 5: !! 53 cmoveq $1, $3, $0 44 or %o4, %lo(LO_MAGIC), %o2 !! 54 45 8: !! 55 done: subq $0, $16, $0 46 ld [%o0], %o5 !! 56 ret $31, ($26) 47 2: !! 57 48 sub %o5, %o2, %o4 !! 58 .end strlen 49 andcc %o4, %o3, %g0 !! 59 EXPORT_SYMBOL(strlen) 50 BRANCH32(be, pt, 8b) << 51 add %o0, 4, %o0 << 52 << 53 /* Check every byte. */ << 54 srl %o5, 24, %g7 << 55 andcc %g7, 0xff, %g0 << 56 BRANCH32(be, pn, 1f) << 57 add %o0, -4, %o4 << 58 srl %o5, 16, %g7 << 59 andcc %g7, 0xff, %g0 << 60 BRANCH32(be, pn, 1f) << 61 add %o4, 1, %o4 << 62 srl %o5, 8, %g7 << 63 andcc %g7, 0xff, %g0 << 64 BRANCH32(be, pn, 1f) << 65 add %o4, 1, %o4 << 66 andcc %o5, 0xff, %g0 << 67 BRANCH32_ANNUL(bne, pt, 2b) << 68 ld [%o0], %o5 << 69 add %o4, 1, %o4 << 70 1: << 71 retl << 72 sub %o4, %o1, %o0 << 73 11: << 74 retl << 75 mov 0, %o0 << 76 12: << 77 retl << 78 mov 1, %o0 << 79 13: << 80 retl << 81 mov 2, %o0 << 82 ENDPROC(strlen) << 83 EXPORT_SYMBOL(strlen) <<
Linux® is a registered trademark of Linus Torvalds in the United States and other countries.
TOMOYO® is a registered trademark of NTT DATA CORPORATION.