Lines Matching refs:tiv
90 #define load_xts_8way(iv, src, dst, x0, x1, x2, x3, x4, x5, x6, x7, tiv, t0, \ argument
95 vmovdqu (iv), tiv; \
96 vpxor (0*16)(src), tiv, x0; \
97 vmovdqu tiv, (0*16)(dst); \
100 gf128mul_x_ble(tiv, t0, t1); \
101 vpxor (1*16)(src), tiv, x1; \
102 vmovdqu tiv, (1*16)(dst); \
104 gf128mul_x_ble(tiv, t0, t1); \
105 vpxor (2*16)(src), tiv, x2; \
106 vmovdqu tiv, (2*16)(dst); \
108 gf128mul_x_ble(tiv, t0, t1); \
109 vpxor (3*16)(src), tiv, x3; \
110 vmovdqu tiv, (3*16)(dst); \
112 gf128mul_x_ble(tiv, t0, t1); \
113 vpxor (4*16)(src), tiv, x4; \
114 vmovdqu tiv, (4*16)(dst); \
116 gf128mul_x_ble(tiv, t0, t1); \
117 vpxor (5*16)(src), tiv, x5; \
118 vmovdqu tiv, (5*16)(dst); \
120 gf128mul_x_ble(tiv, t0, t1); \
121 vpxor (6*16)(src), tiv, x6; \
122 vmovdqu tiv, (6*16)(dst); \
124 gf128mul_x_ble(tiv, t0, t1); \
125 vpxor (7*16)(src), tiv, x7; \
126 vmovdqu tiv, (7*16)(dst); \
128 gf128mul_x_ble(tiv, t0, t1); \
129 vmovdqu tiv, (iv);