| /OK3568_Linux_fs/kernel/arch/x86/crypto/ |
| H A D | glue_helper-asm-avx.S | 8 #define load_8way(src, x0, x1, x2, x3, x4, x5, x6, x7) \ argument 15 vmovdqu (6*16)(src), x6; \ 18 #define store_8way(dst, x0, x1, x2, x3, x4, x5, x6, x7) \ argument 25 vmovdqu x6, (6*16)(dst); \ 28 #define store_cbc_8way(src, dst, x0, x1, x2, x3, x4, x5, x6, x7) \ argument 34 vpxor (5*16)(src), x6, x6; \ 36 store_8way(dst, x0, x1, x2, x3, x4, x5, x6, x7); 44 #define load_ctr_8way(iv, bswap, x0, x1, x2, x3, x4, x5, x6, x7, t0, t1, t2) \ argument 65 vpshufb t1, x7, x6; \ 72 #define store_ctr_8way(src, dst, x0, x1, x2, x3, x4, x5, x6, x7) \ argument [all …]
|
| H A D | glue_helper-asm-avx2.S | 8 #define load_16way(src, x0, x1, x2, x3, x4, x5, x6, x7) \ argument 15 vmovdqu (6*32)(src), x6; \ 18 #define store_16way(dst, x0, x1, x2, x3, x4, x5, x6, x7) \ argument 25 vmovdqu x6, (6*32)(dst); \ 28 #define store_cbc_16way(src, dst, x0, x1, x2, x3, x4, x5, x6, x7, t0) \ argument 37 vpxor (5*32+16)(src), x6, x6; \ 39 store_16way(dst, x0, x1, x2, x3, x4, x5, x6, x7); 55 #define load_ctr_16way(iv, bswap, x0, x1, x2, x3, x4, x5, x6, x7, t0, t0x, t1, \ argument 81 vpshufb t1, t2, x6; \ 88 #define store_ctr_16way(src, dst, x0, x1, x2, x3, x4, x5, x6, x7) \ argument [all …]
|
| H A D | camellia-aesni-avx-asm_64.S | 51 #define roundsm16(x0, x1, x2, x3, x4, x5, x6, x7, t0, t1, t2, t3, t4, t5, t6, \ argument 69 vpshufb t4, x6, x6; \ 84 filter_8bit(x6, t2, t3, t7, t6); \ 96 vaesenclast t4, x6, x6; \ 104 filter_8bit(x6, t0, t1, t7, t6); \ 136 vpxor x6, x1, x1; \ 142 vpxor x0, x6, x6; \ 148 vpxor x6, x3, x3; \ 152 vpxor x1, x6, x6; \ 169 vpxor t1, x6, x6; \ [all …]
|
| H A D | camellia-aesni-avx2-asm_64.S | 63 #define roundsm32(x0, x1, x2, x3, x4, x5, x6, x7, t0, t1, t2, t3, t4, t5, t6, \ argument 79 vpshufb t4, x6, x6; \ 92 filter_8bit(x6, t2, t3, t7, t4); \ 94 vextracti128 $1, x6, t2##_x; \ 114 vaesenclast t4##_x, x6##_x, x6##_x; \ 116 vinserti128 $1, t2##_x, x6, x6; \ 140 filter_8bit(x6, t0, t1, t7, t6); \ 162 vpxor x6, x1, x1; \ 174 vpxor x0, x6, x6; \ 184 vpxor x6, x3, x3; \ [all …]
|
| /OK3568_Linux_fs/kernel/arch/arm64/crypto/ |
| H A D | aes-neonbs-core.S | 113 .macro mul_gf16_2, x0, x1, x2, x3, x4, x5, x6, x7, \ 125 eor \t0, \x4, \x6 127 mul_gf4_n_gf4 \t0, \t1, \y0, \y1, \t3, \x6, \x7, \y2, \y3, \t2 132 eor \x6, \x6, \t0 137 .macro inv_gf256, x0, x1, x2, x3, x4, x5, x6, x7, \ 139 eor \t3, \x4, \x6 142 eor \s1, \x7, \x6 166 and \s1, \x6, \x2 185 mul_gf16_2 \x0, \x1, \x2, \x3, \x4, \x5, \x6, \x7, \ 227 .macro add_round_key, x0, x1, x2, x3, x4, x5, x6, x7 [all …]
|
| H A D | poly1305-core.S_shipped | 70 ldp x6,x17,[x0,#16] // [along with is_base2_26] 78 lsr x16,x6,#32 100 csel x6,x6,x14,eq 113 adc x6,x6,x3 130 mul x10,x6,x9 // h2*5*r1 132 mul x11,x6,x7 // h2*r0 138 and x6,x14,#3 142 adc x6,x6,xzr 147 stp x6,xzr,[x0,#16] // [and clear is_base2_26] 158 ldp x6,x7,[x0,#16] // [along with is_base2_26] [all …]
|
| H A D | aes-modes.S | 133 enc_prepare w8, x6, x7 134 encrypt_block v4, w8, x6, x7, w9 135 enc_switch_key w3, x2, x6 140 enc_prepare w3, x2, x6 147 encrypt_block v0, w3, x2, x6, w7 149 encrypt_block v1, w3, x2, x6, w7 151 encrypt_block v2, w3, x2, x6, w7 153 encrypt_block v3, w3, x2, x6, w7 163 encrypt_block v4, w3, x2, x6, w7 180 enc_prepare w8, x6, x7 [all …]
|
| H A D | aes-ce-ccm-core.S | 35 add x6, x4, #16 42 ld1 {v5.4s}, [x6], #16 /* load 2nd round key */ 45 4: ld1 {v3.4s}, [x6], #16 /* load next round key */ 48 5: ld1 {v4.4s}, [x6], #16 /* load next round key */ 52 ld1 {v5.4s}, [x6], #16 /* load next round key */ 127 ldr x8, [x6, #8] /* load lower ctr */ 131 ld1 {v1.8b}, [x6] /* load upper ctr */ 181 str x8, [x6, #8] /* store lsb end of ctr (BE) */
|
| H A D | sha512-core.S_shipped | 101 eor x6,x24,x24,ror#23 107 eor x16,x16,x6,ror#18 // Sigma1(e) 108 ror x6,x20,#28 115 eor x17,x6,x17,ror#34 // Sigma0(a) 122 ldp x5,x6,[x1],#2*8 169 rev x6,x6 // 3 178 add x24,x24,x6 // h+=X[i] 369 str x6,[sp,#24] 372 eor x6,x21,x21,ror#23 378 eor x16,x16,x6,ror#18 // Sigma1(e) [all …]
|
| /OK3568_Linux_fs/kernel/arch/arm64/boot/dts/microchip/ |
| H A D | sparx5.dtsi | 87 reg = <0x6 0x1110000c 0x24>; 114 reg = <0x6 0x00300000 0x10000>, /* GIC Dist */ 115 <0x6 0x00340000 0xc0000>, /* GICR */ 116 <0x6 0x00200000 0x2000>, /* GICC */ 117 <0x6 0x00210000 0x2000>, /* GICV */ 118 <0x6 0x00220000 0x2000>; /* GICH */ 125 reg = <0x6 0x00000000 0xd0>; 142 reg = <0x6 0x00100000 0x20>; 155 reg = <0x6 0x00102000 0x20>; 168 reg = <0x6 0x00104000 0x40>; [all …]
|
| /OK3568_Linux_fs/kernel/arch/arm64/lib/ |
| H A D | crc32.S | 26 ldp x5, x6, [x8] 30 CPU_BE( rev x6, x6 ) 51 crc32\c\()x w8, w8, x6 57 ldp x5, x6, [x1, #-16] 61 CPU_BE( rev x6, x6 ) 65 crc32\c\()x w0, w0, x6
|
| H A D | copy_page.S | 30 ldp x6, x7, [x1, #32] 50 stnp x6, x7, [x0, #32 - 256] 51 ldp x6, x7, [x1, #32] 70 stnp x6, x7, [x0, #32 - 256]
|
| /OK3568_Linux_fs/kernel/arch/arm64/kernel/ |
| H A D | head.S | 312 mrs_s x6, SYS_ID_AA64MMFR2_EL1 313 and x6, x6, #(0xf << ID_AA64MMFR2_LVA_SHIFT) 315 cbnz x6, 1f 319 adr_l x6, vabits_actual 320 str x5, [x6] 322 dc ivac, x6 // Invalidate potentially stale cache line 341 adr_l x6, idmap_t0sz 342 str x5, [x6] 344 dc ivac, x6 // Invalidate potentially stale cache line 362 create_table_entry x0, x3, EXTRA_SHIFT, x4, x5, x6 [all …]
|
| /OK3568_Linux_fs/kernel/arch/arm/crypto/ |
| H A D | aes-neonbs-core.S | 168 .macro mul_gf16_2, x0, x1, x2, x3, x4, x5, x6, x7, \ 180 veor \t0, \x4, \x6 182 mul_gf4_n_gf4 \t0, \t1, \y0, \y1, \t3, \x6, \x7, \y2, \y3, \t2 187 veor \x6, \x6, \t0 192 .macro inv_gf256, x0, x1, x2, x3, x4, x5, x6, x7, \ 194 veor \t3, \x4, \x6 197 veor \s1, \x7, \x6 221 vand \s1, \x6, \x2 240 mul_gf16_2 \x0, \x1, \x2, \x3, \x4, \x5, \x6, \x7, \ 260 .macro shift_rows, x0, x1, x2, x3, x4, x5, x6, x7, \ [all …]
|
| /OK3568_Linux_fs/kernel/arch/arm64/kvm/hyp/nvhe/ |
| H A D | host.S | 27 stp x6, x7, [x0, #CPU_XREG_OFFSET(6)] 46 ldp x6, x7, [x29, #CPU_XREG_OFFSET(6)] 83 hyp_kimg_va lr, x6 90 hyp_kimg_va x0, x6 93 mov x6, x3 129 hyp_pa x5, x6 211 ldp x6, x7, [x18, #CPU_XREG_OFFSET(6)] 223 stp x6, x7, [x18, #CPU_XREG_OFFSET(6)]
|
| /OK3568_Linux_fs/u-boot/arch/arm/cpu/armv8/ |
| H A D | transition.S | 15 switch_el x6, 1f, 0f, 0f 32 1: armv8_switch_to_el2_m x4, x5, x6 38 switch_el x6, 0f, 1f, 0f 44 1: armv8_switch_to_el1_m x4, x5, x6
|
| H A D | cache.S | 30 mrs x6, ccsidr_el1 /* read the new cssidr_el1 */ 31 and x2, x6, #7 /* x2 <- log2(cache line size)-4 */ 34 and x3, x3, x6, lsr #3 /* x3 <- max number of #ways */ 37 and x4, x4, x6, lsr #13 /* x4 <- max number of #sets */ 45 mov x6, x3 /* x6 <- working copy of #ways */ 47 lsl x7, x6, x5 55 2: subs x6, x6, #1 /* decrement the way */
|
| H A D | exceptions.S | 85 stp x5, x6, [sp, #-16]! 96 sub x6, sp, #(8*30) 106 sub x6, sp, #(8*30) 117 sub x6, sp, #(8*30) 125 stp x7, x6, [sp, #-16]! 143 ldp x5, x6, [sp],#16
|
| H A D | sleep.S | 95 stp x6, x7, [sp, #32] 110 mrs x6, mair_el2 118 stp x6, x7, [x0, #32] 124 ldp x6, x7, [sp, #32] 166 ldp x6, x7, [x0, #32] 176 msr mair_el2, x6
|
| /OK3568_Linux_fs/external/security/rk_tee_user/v2/export-ta_arm32/include/ |
| H A D | sincosf.h | 68 double x3, x4, x5, x6, s, c, c1, c2, s1; in sincosf_poly() local 82 x6 = x4 * x2; in sincosf_poly() 88 *cosp = c + x6 * c2; in sincosf_poly() 96 double x3, x4, x6, x7, s, c, c1, c2, s1; in sinf_poly() local 114 x6 = x4 * x2; in sinf_poly() 117 return c + x6 * c2; in sinf_poly()
|
| /OK3568_Linux_fs/external/security/rk_tee_user/v2/export-ta_arm64/include/ |
| H A D | sincosf.h | 68 double x3, x4, x5, x6, s, c, c1, c2, s1; in sincosf_poly() local 82 x6 = x4 * x2; in sincosf_poly() 88 *cosp = c + x6 * c2; in sincosf_poly() 96 double x3, x4, x6, x7, s, c, c1, c2, s1; in sinf_poly() local 114 x6 = x4 * x2; in sinf_poly() 117 return c + x6 * c2; in sinf_poly()
|
| /OK3568_Linux_fs/u-boot/arch/arm/dts/ |
| H A D | rv1108-sdram-ddr3-400.dtsi | 34 0x6 37 0x6 41 0x6 44 0x6
|
| H A D | rk3288-fennec.dts | 21 0x6 0x0 0x8 0x4 0x17 0x24 0xd 0x6 26 0x0 0xc3 0x6 0x2>;
|
| H A D | .rk3288-fennec.dtb.pre.tmp | |
| /OK3568_Linux_fs/kernel/tools/testing/selftests/arm64/fp/ |
| H A D | sve-test.S | 249 mov x6, x1 251 _adrz x0, x6, 2 256 mov x0, x6 270 mov x6, x1 272 _adrp x0, x6, 2 277 mov x0, x6 383 mov x1, x6 392 mov x2, x6 408 mov x1, x6 417 mov x2, x6
|