1*4882a593Smuzhiyun/* SPDX-License-Identifier: GPL-2.0-only */ 2*4882a593Smuzhiyun/* 3*4882a593Smuzhiyun * Copyright (C) 2013 ARM Ltd. 4*4882a593Smuzhiyun * Copyright (C) 2013 Linaro. 5*4882a593Smuzhiyun * 6*4882a593Smuzhiyun * This code is based on glibc cortex strings work originally authored by Linaro 7*4882a593Smuzhiyun * be found @ 8*4882a593Smuzhiyun * 9*4882a593Smuzhiyun * http://bazaar.launchpad.net/~linaro-toolchain-dev/cortex-strings/trunk/ 10*4882a593Smuzhiyun * files/head:/src/aarch64/ 11*4882a593Smuzhiyun */ 12*4882a593Smuzhiyun 13*4882a593Smuzhiyun#include <linux/linkage.h> 14*4882a593Smuzhiyun#include <asm/assembler.h> 15*4882a593Smuzhiyun 16*4882a593Smuzhiyun/* 17*4882a593Smuzhiyun* compare memory areas(when two memory areas' offset are different, 18*4882a593Smuzhiyun* alignment handled by the hardware) 19*4882a593Smuzhiyun* 20*4882a593Smuzhiyun* Parameters: 21*4882a593Smuzhiyun* x0 - const memory area 1 pointer 22*4882a593Smuzhiyun* x1 - const memory area 2 pointer 23*4882a593Smuzhiyun* x2 - the maximal compare byte length 24*4882a593Smuzhiyun* Returns: 25*4882a593Smuzhiyun* x0 - a compare result, maybe less than, equal to, or greater than ZERO 26*4882a593Smuzhiyun*/ 27*4882a593Smuzhiyun 28*4882a593Smuzhiyun/* Parameters and result. */ 29*4882a593Smuzhiyunsrc1 .req x0 30*4882a593Smuzhiyunsrc2 .req x1 31*4882a593Smuzhiyunlimit .req x2 32*4882a593Smuzhiyunresult .req x0 33*4882a593Smuzhiyun 34*4882a593Smuzhiyun/* Internal variables. */ 35*4882a593Smuzhiyundata1 .req x3 36*4882a593Smuzhiyundata1w .req w3 37*4882a593Smuzhiyundata2 .req x4 38*4882a593Smuzhiyundata2w .req w4 39*4882a593Smuzhiyunhas_nul .req x5 40*4882a593Smuzhiyundiff .req x6 41*4882a593Smuzhiyunendloop .req x7 42*4882a593Smuzhiyuntmp1 .req x8 43*4882a593Smuzhiyuntmp2 .req x9 44*4882a593Smuzhiyuntmp3 .req x10 45*4882a593Smuzhiyunpos .req x11 46*4882a593Smuzhiyunlimit_wd .req x12 47*4882a593Smuzhiyunmask .req x13 48*4882a593Smuzhiyun 49*4882a593SmuzhiyunSYM_FUNC_START_WEAK_PI(memcmp) 50*4882a593Smuzhiyun cbz limit, .Lret0 51*4882a593Smuzhiyun eor tmp1, src1, src2 52*4882a593Smuzhiyun tst tmp1, #7 53*4882a593Smuzhiyun b.ne .Lmisaligned8 54*4882a593Smuzhiyun ands tmp1, src1, #7 55*4882a593Smuzhiyun b.ne .Lmutual_align 56*4882a593Smuzhiyun sub limit_wd, limit, #1 /* limit != 0, so no underflow. */ 57*4882a593Smuzhiyun lsr limit_wd, limit_wd, #3 /* Convert to Dwords. */ 58*4882a593Smuzhiyun /* 59*4882a593Smuzhiyun * The input source addresses are at alignment boundary. 60*4882a593Smuzhiyun * Directly compare eight bytes each time. 61*4882a593Smuzhiyun */ 62*4882a593Smuzhiyun.Lloop_aligned: 63*4882a593Smuzhiyun ldr data1, [src1], #8 64*4882a593Smuzhiyun ldr data2, [src2], #8 65*4882a593Smuzhiyun.Lstart_realigned: 66*4882a593Smuzhiyun subs limit_wd, limit_wd, #1 67*4882a593Smuzhiyun eor diff, data1, data2 /* Non-zero if differences found. */ 68*4882a593Smuzhiyun csinv endloop, diff, xzr, cs /* Last Dword or differences. */ 69*4882a593Smuzhiyun cbz endloop, .Lloop_aligned 70*4882a593Smuzhiyun 71*4882a593Smuzhiyun /* Not reached the limit, must have found a diff. */ 72*4882a593Smuzhiyun tbz limit_wd, #63, .Lnot_limit 73*4882a593Smuzhiyun 74*4882a593Smuzhiyun /* Limit % 8 == 0 => the diff is in the last 8 bytes. */ 75*4882a593Smuzhiyun ands limit, limit, #7 76*4882a593Smuzhiyun b.eq .Lnot_limit 77*4882a593Smuzhiyun /* 78*4882a593Smuzhiyun * The remained bytes less than 8. It is needed to extract valid data 79*4882a593Smuzhiyun * from last eight bytes of the intended memory range. 80*4882a593Smuzhiyun */ 81*4882a593Smuzhiyun lsl limit, limit, #3 /* bytes-> bits. */ 82*4882a593Smuzhiyun mov mask, #~0 83*4882a593SmuzhiyunCPU_BE( lsr mask, mask, limit ) 84*4882a593SmuzhiyunCPU_LE( lsl mask, mask, limit ) 85*4882a593Smuzhiyun bic data1, data1, mask 86*4882a593Smuzhiyun bic data2, data2, mask 87*4882a593Smuzhiyun 88*4882a593Smuzhiyun orr diff, diff, mask 89*4882a593Smuzhiyun b .Lnot_limit 90*4882a593Smuzhiyun 91*4882a593Smuzhiyun.Lmutual_align: 92*4882a593Smuzhiyun /* 93*4882a593Smuzhiyun * Sources are mutually aligned, but are not currently at an 94*4882a593Smuzhiyun * alignment boundary. Round down the addresses and then mask off 95*4882a593Smuzhiyun * the bytes that precede the start point. 96*4882a593Smuzhiyun */ 97*4882a593Smuzhiyun bic src1, src1, #7 98*4882a593Smuzhiyun bic src2, src2, #7 99*4882a593Smuzhiyun ldr data1, [src1], #8 100*4882a593Smuzhiyun ldr data2, [src2], #8 101*4882a593Smuzhiyun /* 102*4882a593Smuzhiyun * We can not add limit with alignment offset(tmp1) here. Since the 103*4882a593Smuzhiyun * addition probably make the limit overflown. 104*4882a593Smuzhiyun */ 105*4882a593Smuzhiyun sub limit_wd, limit, #1/*limit != 0, so no underflow.*/ 106*4882a593Smuzhiyun and tmp3, limit_wd, #7 107*4882a593Smuzhiyun lsr limit_wd, limit_wd, #3 108*4882a593Smuzhiyun add tmp3, tmp3, tmp1 109*4882a593Smuzhiyun add limit_wd, limit_wd, tmp3, lsr #3 110*4882a593Smuzhiyun add limit, limit, tmp1/* Adjust the limit for the extra. */ 111*4882a593Smuzhiyun 112*4882a593Smuzhiyun lsl tmp1, tmp1, #3/* Bytes beyond alignment -> bits.*/ 113*4882a593Smuzhiyun neg tmp1, tmp1/* Bits to alignment -64. */ 114*4882a593Smuzhiyun mov tmp2, #~0 115*4882a593Smuzhiyun /*mask off the non-intended bytes before the start address.*/ 116*4882a593SmuzhiyunCPU_BE( lsl tmp2, tmp2, tmp1 )/*Big-endian.Early bytes are at MSB*/ 117*4882a593Smuzhiyun /* Little-endian. Early bytes are at LSB. */ 118*4882a593SmuzhiyunCPU_LE( lsr tmp2, tmp2, tmp1 ) 119*4882a593Smuzhiyun 120*4882a593Smuzhiyun orr data1, data1, tmp2 121*4882a593Smuzhiyun orr data2, data2, tmp2 122*4882a593Smuzhiyun b .Lstart_realigned 123*4882a593Smuzhiyun 124*4882a593Smuzhiyun /*src1 and src2 have different alignment offset.*/ 125*4882a593Smuzhiyun.Lmisaligned8: 126*4882a593Smuzhiyun cmp limit, #8 127*4882a593Smuzhiyun b.lo .Ltiny8proc /*limit < 8: compare byte by byte*/ 128*4882a593Smuzhiyun 129*4882a593Smuzhiyun and tmp1, src1, #7 130*4882a593Smuzhiyun neg tmp1, tmp1 131*4882a593Smuzhiyun add tmp1, tmp1, #8/*valid length in the first 8 bytes of src1*/ 132*4882a593Smuzhiyun and tmp2, src2, #7 133*4882a593Smuzhiyun neg tmp2, tmp2 134*4882a593Smuzhiyun add tmp2, tmp2, #8/*valid length in the first 8 bytes of src2*/ 135*4882a593Smuzhiyun subs tmp3, tmp1, tmp2 136*4882a593Smuzhiyun csel pos, tmp1, tmp2, hi /*Choose the maximum.*/ 137*4882a593Smuzhiyun 138*4882a593Smuzhiyun sub limit, limit, pos 139*4882a593Smuzhiyun /*compare the proceeding bytes in the first 8 byte segment.*/ 140*4882a593Smuzhiyun.Ltinycmp: 141*4882a593Smuzhiyun ldrb data1w, [src1], #1 142*4882a593Smuzhiyun ldrb data2w, [src2], #1 143*4882a593Smuzhiyun subs pos, pos, #1 144*4882a593Smuzhiyun ccmp data1w, data2w, #0, ne /* NZCV = 0b0000. */ 145*4882a593Smuzhiyun b.eq .Ltinycmp 146*4882a593Smuzhiyun cbnz pos, 1f /*diff occurred before the last byte.*/ 147*4882a593Smuzhiyun cmp data1w, data2w 148*4882a593Smuzhiyun b.eq .Lstart_align 149*4882a593Smuzhiyun1: 150*4882a593Smuzhiyun sub result, data1, data2 151*4882a593Smuzhiyun ret 152*4882a593Smuzhiyun 153*4882a593Smuzhiyun.Lstart_align: 154*4882a593Smuzhiyun lsr limit_wd, limit, #3 155*4882a593Smuzhiyun cbz limit_wd, .Lremain8 156*4882a593Smuzhiyun 157*4882a593Smuzhiyun ands xzr, src1, #7 158*4882a593Smuzhiyun b.eq .Lrecal_offset 159*4882a593Smuzhiyun /*process more leading bytes to make src1 aligned...*/ 160*4882a593Smuzhiyun add src1, src1, tmp3 /*backwards src1 to alignment boundary*/ 161*4882a593Smuzhiyun add src2, src2, tmp3 162*4882a593Smuzhiyun sub limit, limit, tmp3 163*4882a593Smuzhiyun lsr limit_wd, limit, #3 164*4882a593Smuzhiyun cbz limit_wd, .Lremain8 165*4882a593Smuzhiyun /*load 8 bytes from aligned SRC1..*/ 166*4882a593Smuzhiyun ldr data1, [src1], #8 167*4882a593Smuzhiyun ldr data2, [src2], #8 168*4882a593Smuzhiyun 169*4882a593Smuzhiyun subs limit_wd, limit_wd, #1 170*4882a593Smuzhiyun eor diff, data1, data2 /*Non-zero if differences found.*/ 171*4882a593Smuzhiyun csinv endloop, diff, xzr, ne 172*4882a593Smuzhiyun cbnz endloop, .Lunequal_proc 173*4882a593Smuzhiyun /*How far is the current SRC2 from the alignment boundary...*/ 174*4882a593Smuzhiyun and tmp3, tmp3, #7 175*4882a593Smuzhiyun 176*4882a593Smuzhiyun.Lrecal_offset:/*src1 is aligned now..*/ 177*4882a593Smuzhiyun neg pos, tmp3 178*4882a593Smuzhiyun.Lloopcmp_proc: 179*4882a593Smuzhiyun /* 180*4882a593Smuzhiyun * Divide the eight bytes into two parts. First,backwards the src2 181*4882a593Smuzhiyun * to an alignment boundary,load eight bytes and compare from 182*4882a593Smuzhiyun * the SRC2 alignment boundary. If all 8 bytes are equal,then start 183*4882a593Smuzhiyun * the second part's comparison. Otherwise finish the comparison. 184*4882a593Smuzhiyun * This special handle can garantee all the accesses are in the 185*4882a593Smuzhiyun * thread/task space in avoid to overrange access. 186*4882a593Smuzhiyun */ 187*4882a593Smuzhiyun ldr data1, [src1,pos] 188*4882a593Smuzhiyun ldr data2, [src2,pos] 189*4882a593Smuzhiyun eor diff, data1, data2 /* Non-zero if differences found. */ 190*4882a593Smuzhiyun cbnz diff, .Lnot_limit 191*4882a593Smuzhiyun 192*4882a593Smuzhiyun /*The second part process*/ 193*4882a593Smuzhiyun ldr data1, [src1], #8 194*4882a593Smuzhiyun ldr data2, [src2], #8 195*4882a593Smuzhiyun eor diff, data1, data2 /* Non-zero if differences found. */ 196*4882a593Smuzhiyun subs limit_wd, limit_wd, #1 197*4882a593Smuzhiyun csinv endloop, diff, xzr, ne/*if limit_wd is 0,will finish the cmp*/ 198*4882a593Smuzhiyun cbz endloop, .Lloopcmp_proc 199*4882a593Smuzhiyun.Lunequal_proc: 200*4882a593Smuzhiyun cbz diff, .Lremain8 201*4882a593Smuzhiyun 202*4882a593Smuzhiyun/* There is difference occurred in the latest comparison. */ 203*4882a593Smuzhiyun.Lnot_limit: 204*4882a593Smuzhiyun/* 205*4882a593Smuzhiyun* For little endian,reverse the low significant equal bits into MSB,then 206*4882a593Smuzhiyun* following CLZ can find how many equal bits exist. 207*4882a593Smuzhiyun*/ 208*4882a593SmuzhiyunCPU_LE( rev diff, diff ) 209*4882a593SmuzhiyunCPU_LE( rev data1, data1 ) 210*4882a593SmuzhiyunCPU_LE( rev data2, data2 ) 211*4882a593Smuzhiyun 212*4882a593Smuzhiyun /* 213*4882a593Smuzhiyun * The MS-non-zero bit of DIFF marks either the first bit 214*4882a593Smuzhiyun * that is different, or the end of the significant data. 215*4882a593Smuzhiyun * Shifting left now will bring the critical information into the 216*4882a593Smuzhiyun * top bits. 217*4882a593Smuzhiyun */ 218*4882a593Smuzhiyun clz pos, diff 219*4882a593Smuzhiyun lsl data1, data1, pos 220*4882a593Smuzhiyun lsl data2, data2, pos 221*4882a593Smuzhiyun /* 222*4882a593Smuzhiyun * We need to zero-extend (char is unsigned) the value and then 223*4882a593Smuzhiyun * perform a signed subtraction. 224*4882a593Smuzhiyun */ 225*4882a593Smuzhiyun lsr data1, data1, #56 226*4882a593Smuzhiyun sub result, data1, data2, lsr #56 227*4882a593Smuzhiyun ret 228*4882a593Smuzhiyun 229*4882a593Smuzhiyun.Lremain8: 230*4882a593Smuzhiyun /* Limit % 8 == 0 =>. all data are equal.*/ 231*4882a593Smuzhiyun ands limit, limit, #7 232*4882a593Smuzhiyun b.eq .Lret0 233*4882a593Smuzhiyun 234*4882a593Smuzhiyun.Ltiny8proc: 235*4882a593Smuzhiyun ldrb data1w, [src1], #1 236*4882a593Smuzhiyun ldrb data2w, [src2], #1 237*4882a593Smuzhiyun subs limit, limit, #1 238*4882a593Smuzhiyun 239*4882a593Smuzhiyun ccmp data1w, data2w, #0, ne /* NZCV = 0b0000. */ 240*4882a593Smuzhiyun b.eq .Ltiny8proc 241*4882a593Smuzhiyun sub result, data1, data2 242*4882a593Smuzhiyun ret 243*4882a593Smuzhiyun.Lret0: 244*4882a593Smuzhiyun mov result, #0 245*4882a593Smuzhiyun ret 246*4882a593SmuzhiyunSYM_FUNC_END_PI(memcmp) 247*4882a593SmuzhiyunEXPORT_SYMBOL_NOKASAN(memcmp) 248