xref: /OK3568_Linux_fs/kernel/arch/s390/crypto/crc32-vx.c (revision 4882a59341e53eb6f0b4789bf948001014eff981)
1*4882a593Smuzhiyun // SPDX-License-Identifier: GPL-2.0
2*4882a593Smuzhiyun /*
3*4882a593Smuzhiyun  * Crypto-API module for CRC-32 algorithms implemented with the
4*4882a593Smuzhiyun  * z/Architecture Vector Extension Facility.
5*4882a593Smuzhiyun  *
6*4882a593Smuzhiyun  * Copyright IBM Corp. 2015
7*4882a593Smuzhiyun  * Author(s): Hendrik Brueckner <brueckner@linux.vnet.ibm.com>
8*4882a593Smuzhiyun  */
9*4882a593Smuzhiyun #define KMSG_COMPONENT	"crc32-vx"
10*4882a593Smuzhiyun #define pr_fmt(fmt)	KMSG_COMPONENT ": " fmt
11*4882a593Smuzhiyun 
12*4882a593Smuzhiyun #include <linux/module.h>
13*4882a593Smuzhiyun #include <linux/cpufeature.h>
14*4882a593Smuzhiyun #include <linux/crc32.h>
15*4882a593Smuzhiyun #include <crypto/internal/hash.h>
16*4882a593Smuzhiyun #include <asm/fpu/api.h>
17*4882a593Smuzhiyun 
18*4882a593Smuzhiyun 
19*4882a593Smuzhiyun #define CRC32_BLOCK_SIZE	1
20*4882a593Smuzhiyun #define CRC32_DIGEST_SIZE	4
21*4882a593Smuzhiyun 
22*4882a593Smuzhiyun #define VX_MIN_LEN		64
23*4882a593Smuzhiyun #define VX_ALIGNMENT		16L
24*4882a593Smuzhiyun #define VX_ALIGN_MASK		(VX_ALIGNMENT - 1)
25*4882a593Smuzhiyun 
26*4882a593Smuzhiyun struct crc_ctx {
27*4882a593Smuzhiyun 	u32 key;
28*4882a593Smuzhiyun };
29*4882a593Smuzhiyun 
30*4882a593Smuzhiyun struct crc_desc_ctx {
31*4882a593Smuzhiyun 	u32 crc;
32*4882a593Smuzhiyun };
33*4882a593Smuzhiyun 
34*4882a593Smuzhiyun /* Prototypes for functions in assembly files */
35*4882a593Smuzhiyun u32 crc32_le_vgfm_16(u32 crc, unsigned char const *buf, size_t size);
36*4882a593Smuzhiyun u32 crc32_be_vgfm_16(u32 crc, unsigned char const *buf, size_t size);
37*4882a593Smuzhiyun u32 crc32c_le_vgfm_16(u32 crc, unsigned char const *buf, size_t size);
38*4882a593Smuzhiyun 
39*4882a593Smuzhiyun /*
40*4882a593Smuzhiyun  * DEFINE_CRC32_VX() - Define a CRC-32 function using the vector extension
41*4882a593Smuzhiyun  *
42*4882a593Smuzhiyun  * Creates a function to perform a particular CRC-32 computation. Depending
43*4882a593Smuzhiyun  * on the message buffer, the hardware-accelerated or software implementation
44*4882a593Smuzhiyun  * is used.   Note that the message buffer is aligned to improve fetch
45*4882a593Smuzhiyun  * operations of VECTOR LOAD MULTIPLE instructions.
46*4882a593Smuzhiyun  *
47*4882a593Smuzhiyun  */
48*4882a593Smuzhiyun #define DEFINE_CRC32_VX(___fname, ___crc32_vx, ___crc32_sw)		    \
49*4882a593Smuzhiyun 	static u32 __pure ___fname(u32 crc,				    \
50*4882a593Smuzhiyun 				unsigned char const *data, size_t datalen)  \
51*4882a593Smuzhiyun 	{								    \
52*4882a593Smuzhiyun 		struct kernel_fpu vxstate;				    \
53*4882a593Smuzhiyun 		unsigned long prealign, aligned, remaining;		    \
54*4882a593Smuzhiyun 									    \
55*4882a593Smuzhiyun 		if (datalen < VX_MIN_LEN + VX_ALIGN_MASK)		    \
56*4882a593Smuzhiyun 			return ___crc32_sw(crc, data, datalen);		    \
57*4882a593Smuzhiyun 									    \
58*4882a593Smuzhiyun 		if ((unsigned long)data & VX_ALIGN_MASK) {		    \
59*4882a593Smuzhiyun 			prealign = VX_ALIGNMENT -			    \
60*4882a593Smuzhiyun 				  ((unsigned long)data & VX_ALIGN_MASK);    \
61*4882a593Smuzhiyun 			datalen -= prealign;				    \
62*4882a593Smuzhiyun 			crc = ___crc32_sw(crc, data, prealign);		    \
63*4882a593Smuzhiyun 			data = (void *)((unsigned long)data + prealign);    \
64*4882a593Smuzhiyun 		}							    \
65*4882a593Smuzhiyun 									    \
66*4882a593Smuzhiyun 		aligned = datalen & ~VX_ALIGN_MASK;			    \
67*4882a593Smuzhiyun 		remaining = datalen & VX_ALIGN_MASK;			    \
68*4882a593Smuzhiyun 									    \
69*4882a593Smuzhiyun 		kernel_fpu_begin(&vxstate, KERNEL_VXR_LOW);		    \
70*4882a593Smuzhiyun 		crc = ___crc32_vx(crc, data, aligned);			    \
71*4882a593Smuzhiyun 		kernel_fpu_end(&vxstate, KERNEL_VXR_LOW);		    \
72*4882a593Smuzhiyun 									    \
73*4882a593Smuzhiyun 		if (remaining)						    \
74*4882a593Smuzhiyun 			crc = ___crc32_sw(crc, data + aligned, remaining);  \
75*4882a593Smuzhiyun 									    \
76*4882a593Smuzhiyun 		return crc;						    \
77*4882a593Smuzhiyun 	}
78*4882a593Smuzhiyun 
DEFINE_CRC32_VX(crc32_le_vx,crc32_le_vgfm_16,crc32_le)79*4882a593Smuzhiyun DEFINE_CRC32_VX(crc32_le_vx, crc32_le_vgfm_16, crc32_le)
80*4882a593Smuzhiyun DEFINE_CRC32_VX(crc32_be_vx, crc32_be_vgfm_16, crc32_be)
81*4882a593Smuzhiyun DEFINE_CRC32_VX(crc32c_le_vx, crc32c_le_vgfm_16, __crc32c_le)
82*4882a593Smuzhiyun 
83*4882a593Smuzhiyun 
84*4882a593Smuzhiyun static int crc32_vx_cra_init_zero(struct crypto_tfm *tfm)
85*4882a593Smuzhiyun {
86*4882a593Smuzhiyun 	struct crc_ctx *mctx = crypto_tfm_ctx(tfm);
87*4882a593Smuzhiyun 
88*4882a593Smuzhiyun 	mctx->key = 0;
89*4882a593Smuzhiyun 	return 0;
90*4882a593Smuzhiyun }
91*4882a593Smuzhiyun 
crc32_vx_cra_init_invert(struct crypto_tfm * tfm)92*4882a593Smuzhiyun static int crc32_vx_cra_init_invert(struct crypto_tfm *tfm)
93*4882a593Smuzhiyun {
94*4882a593Smuzhiyun 	struct crc_ctx *mctx = crypto_tfm_ctx(tfm);
95*4882a593Smuzhiyun 
96*4882a593Smuzhiyun 	mctx->key = ~0;
97*4882a593Smuzhiyun 	return 0;
98*4882a593Smuzhiyun }
99*4882a593Smuzhiyun 
crc32_vx_init(struct shash_desc * desc)100*4882a593Smuzhiyun static int crc32_vx_init(struct shash_desc *desc)
101*4882a593Smuzhiyun {
102*4882a593Smuzhiyun 	struct crc_ctx *mctx = crypto_shash_ctx(desc->tfm);
103*4882a593Smuzhiyun 	struct crc_desc_ctx *ctx = shash_desc_ctx(desc);
104*4882a593Smuzhiyun 
105*4882a593Smuzhiyun 	ctx->crc = mctx->key;
106*4882a593Smuzhiyun 	return 0;
107*4882a593Smuzhiyun }
108*4882a593Smuzhiyun 
crc32_vx_setkey(struct crypto_shash * tfm,const u8 * newkey,unsigned int newkeylen)109*4882a593Smuzhiyun static int crc32_vx_setkey(struct crypto_shash *tfm, const u8 *newkey,
110*4882a593Smuzhiyun 			   unsigned int newkeylen)
111*4882a593Smuzhiyun {
112*4882a593Smuzhiyun 	struct crc_ctx *mctx = crypto_shash_ctx(tfm);
113*4882a593Smuzhiyun 
114*4882a593Smuzhiyun 	if (newkeylen != sizeof(mctx->key))
115*4882a593Smuzhiyun 		return -EINVAL;
116*4882a593Smuzhiyun 	mctx->key = le32_to_cpu(*(__le32 *)newkey);
117*4882a593Smuzhiyun 	return 0;
118*4882a593Smuzhiyun }
119*4882a593Smuzhiyun 
crc32be_vx_setkey(struct crypto_shash * tfm,const u8 * newkey,unsigned int newkeylen)120*4882a593Smuzhiyun static int crc32be_vx_setkey(struct crypto_shash *tfm, const u8 *newkey,
121*4882a593Smuzhiyun 			     unsigned int newkeylen)
122*4882a593Smuzhiyun {
123*4882a593Smuzhiyun 	struct crc_ctx *mctx = crypto_shash_ctx(tfm);
124*4882a593Smuzhiyun 
125*4882a593Smuzhiyun 	if (newkeylen != sizeof(mctx->key))
126*4882a593Smuzhiyun 		return -EINVAL;
127*4882a593Smuzhiyun 	mctx->key = be32_to_cpu(*(__be32 *)newkey);
128*4882a593Smuzhiyun 	return 0;
129*4882a593Smuzhiyun }
130*4882a593Smuzhiyun 
crc32le_vx_final(struct shash_desc * desc,u8 * out)131*4882a593Smuzhiyun static int crc32le_vx_final(struct shash_desc *desc, u8 *out)
132*4882a593Smuzhiyun {
133*4882a593Smuzhiyun 	struct crc_desc_ctx *ctx = shash_desc_ctx(desc);
134*4882a593Smuzhiyun 
135*4882a593Smuzhiyun 	*(__le32 *)out = cpu_to_le32p(&ctx->crc);
136*4882a593Smuzhiyun 	return 0;
137*4882a593Smuzhiyun }
138*4882a593Smuzhiyun 
crc32be_vx_final(struct shash_desc * desc,u8 * out)139*4882a593Smuzhiyun static int crc32be_vx_final(struct shash_desc *desc, u8 *out)
140*4882a593Smuzhiyun {
141*4882a593Smuzhiyun 	struct crc_desc_ctx *ctx = shash_desc_ctx(desc);
142*4882a593Smuzhiyun 
143*4882a593Smuzhiyun 	*(__be32 *)out = cpu_to_be32p(&ctx->crc);
144*4882a593Smuzhiyun 	return 0;
145*4882a593Smuzhiyun }
146*4882a593Smuzhiyun 
crc32c_vx_final(struct shash_desc * desc,u8 * out)147*4882a593Smuzhiyun static int crc32c_vx_final(struct shash_desc *desc, u8 *out)
148*4882a593Smuzhiyun {
149*4882a593Smuzhiyun 	struct crc_desc_ctx *ctx = shash_desc_ctx(desc);
150*4882a593Smuzhiyun 
151*4882a593Smuzhiyun 	/*
152*4882a593Smuzhiyun 	 * Perform a final XOR with 0xFFFFFFFF to be in sync
153*4882a593Smuzhiyun 	 * with the generic crc32c shash implementation.
154*4882a593Smuzhiyun 	 */
155*4882a593Smuzhiyun 	*(__le32 *)out = ~cpu_to_le32p(&ctx->crc);
156*4882a593Smuzhiyun 	return 0;
157*4882a593Smuzhiyun }
158*4882a593Smuzhiyun 
__crc32le_vx_finup(u32 * crc,const u8 * data,unsigned int len,u8 * out)159*4882a593Smuzhiyun static int __crc32le_vx_finup(u32 *crc, const u8 *data, unsigned int len,
160*4882a593Smuzhiyun 			      u8 *out)
161*4882a593Smuzhiyun {
162*4882a593Smuzhiyun 	*(__le32 *)out = cpu_to_le32(crc32_le_vx(*crc, data, len));
163*4882a593Smuzhiyun 	return 0;
164*4882a593Smuzhiyun }
165*4882a593Smuzhiyun 
__crc32be_vx_finup(u32 * crc,const u8 * data,unsigned int len,u8 * out)166*4882a593Smuzhiyun static int __crc32be_vx_finup(u32 *crc, const u8 *data, unsigned int len,
167*4882a593Smuzhiyun 			      u8 *out)
168*4882a593Smuzhiyun {
169*4882a593Smuzhiyun 	*(__be32 *)out = cpu_to_be32(crc32_be_vx(*crc, data, len));
170*4882a593Smuzhiyun 	return 0;
171*4882a593Smuzhiyun }
172*4882a593Smuzhiyun 
__crc32c_vx_finup(u32 * crc,const u8 * data,unsigned int len,u8 * out)173*4882a593Smuzhiyun static int __crc32c_vx_finup(u32 *crc, const u8 *data, unsigned int len,
174*4882a593Smuzhiyun 			     u8 *out)
175*4882a593Smuzhiyun {
176*4882a593Smuzhiyun 	/*
177*4882a593Smuzhiyun 	 * Perform a final XOR with 0xFFFFFFFF to be in sync
178*4882a593Smuzhiyun 	 * with the generic crc32c shash implementation.
179*4882a593Smuzhiyun 	 */
180*4882a593Smuzhiyun 	*(__le32 *)out = ~cpu_to_le32(crc32c_le_vx(*crc, data, len));
181*4882a593Smuzhiyun 	return 0;
182*4882a593Smuzhiyun }
183*4882a593Smuzhiyun 
184*4882a593Smuzhiyun 
185*4882a593Smuzhiyun #define CRC32_VX_FINUP(alg, func)					      \
186*4882a593Smuzhiyun 	static int alg ## _vx_finup(struct shash_desc *desc, const u8 *data,  \
187*4882a593Smuzhiyun 				   unsigned int datalen, u8 *out)	      \
188*4882a593Smuzhiyun 	{								      \
189*4882a593Smuzhiyun 		return __ ## alg ## _vx_finup(shash_desc_ctx(desc),	      \
190*4882a593Smuzhiyun 					      data, datalen, out);	      \
191*4882a593Smuzhiyun 	}
192*4882a593Smuzhiyun 
193*4882a593Smuzhiyun CRC32_VX_FINUP(crc32le, crc32_le_vx)
194*4882a593Smuzhiyun CRC32_VX_FINUP(crc32be, crc32_be_vx)
195*4882a593Smuzhiyun CRC32_VX_FINUP(crc32c, crc32c_le_vx)
196*4882a593Smuzhiyun 
197*4882a593Smuzhiyun #define CRC32_VX_DIGEST(alg, func)					      \
198*4882a593Smuzhiyun 	static int alg ## _vx_digest(struct shash_desc *desc, const u8 *data, \
199*4882a593Smuzhiyun 				     unsigned int len, u8 *out)		      \
200*4882a593Smuzhiyun 	{								      \
201*4882a593Smuzhiyun 		return __ ## alg ## _vx_finup(crypto_shash_ctx(desc->tfm),    \
202*4882a593Smuzhiyun 					      data, len, out);		      \
203*4882a593Smuzhiyun 	}
204*4882a593Smuzhiyun 
205*4882a593Smuzhiyun CRC32_VX_DIGEST(crc32le, crc32_le_vx)
206*4882a593Smuzhiyun CRC32_VX_DIGEST(crc32be, crc32_be_vx)
207*4882a593Smuzhiyun CRC32_VX_DIGEST(crc32c, crc32c_le_vx)
208*4882a593Smuzhiyun 
209*4882a593Smuzhiyun #define CRC32_VX_UPDATE(alg, func)					      \
210*4882a593Smuzhiyun 	static int alg ## _vx_update(struct shash_desc *desc, const u8 *data, \
211*4882a593Smuzhiyun 				     unsigned int datalen)		      \
212*4882a593Smuzhiyun 	{								      \
213*4882a593Smuzhiyun 		struct crc_desc_ctx *ctx = shash_desc_ctx(desc);	      \
214*4882a593Smuzhiyun 		ctx->crc = func(ctx->crc, data, datalen);		      \
215*4882a593Smuzhiyun 		return 0;						      \
216*4882a593Smuzhiyun 	}
217*4882a593Smuzhiyun 
218*4882a593Smuzhiyun CRC32_VX_UPDATE(crc32le, crc32_le_vx)
219*4882a593Smuzhiyun CRC32_VX_UPDATE(crc32be, crc32_be_vx)
220*4882a593Smuzhiyun CRC32_VX_UPDATE(crc32c, crc32c_le_vx)
221*4882a593Smuzhiyun 
222*4882a593Smuzhiyun 
223*4882a593Smuzhiyun static struct shash_alg crc32_vx_algs[] = {
224*4882a593Smuzhiyun 	/* CRC-32 LE */
225*4882a593Smuzhiyun 	{
226*4882a593Smuzhiyun 		.init		=	crc32_vx_init,
227*4882a593Smuzhiyun 		.setkey		=	crc32_vx_setkey,
228*4882a593Smuzhiyun 		.update		=	crc32le_vx_update,
229*4882a593Smuzhiyun 		.final		=	crc32le_vx_final,
230*4882a593Smuzhiyun 		.finup		=	crc32le_vx_finup,
231*4882a593Smuzhiyun 		.digest		=	crc32le_vx_digest,
232*4882a593Smuzhiyun 		.descsize	=	sizeof(struct crc_desc_ctx),
233*4882a593Smuzhiyun 		.digestsize	=	CRC32_DIGEST_SIZE,
234*4882a593Smuzhiyun 		.base		=	{
235*4882a593Smuzhiyun 			.cra_name	 = "crc32",
236*4882a593Smuzhiyun 			.cra_driver_name = "crc32-vx",
237*4882a593Smuzhiyun 			.cra_priority	 = 200,
238*4882a593Smuzhiyun 			.cra_flags	 = CRYPTO_ALG_OPTIONAL_KEY,
239*4882a593Smuzhiyun 			.cra_blocksize	 = CRC32_BLOCK_SIZE,
240*4882a593Smuzhiyun 			.cra_ctxsize	 = sizeof(struct crc_ctx),
241*4882a593Smuzhiyun 			.cra_module	 = THIS_MODULE,
242*4882a593Smuzhiyun 			.cra_init	 = crc32_vx_cra_init_zero,
243*4882a593Smuzhiyun 		},
244*4882a593Smuzhiyun 	},
245*4882a593Smuzhiyun 	/* CRC-32 BE */
246*4882a593Smuzhiyun 	{
247*4882a593Smuzhiyun 		.init		=	crc32_vx_init,
248*4882a593Smuzhiyun 		.setkey		=	crc32be_vx_setkey,
249*4882a593Smuzhiyun 		.update		=	crc32be_vx_update,
250*4882a593Smuzhiyun 		.final		=	crc32be_vx_final,
251*4882a593Smuzhiyun 		.finup		=	crc32be_vx_finup,
252*4882a593Smuzhiyun 		.digest		=	crc32be_vx_digest,
253*4882a593Smuzhiyun 		.descsize	=	sizeof(struct crc_desc_ctx),
254*4882a593Smuzhiyun 		.digestsize	=	CRC32_DIGEST_SIZE,
255*4882a593Smuzhiyun 		.base		=	{
256*4882a593Smuzhiyun 			.cra_name	 = "crc32be",
257*4882a593Smuzhiyun 			.cra_driver_name = "crc32be-vx",
258*4882a593Smuzhiyun 			.cra_priority	 = 200,
259*4882a593Smuzhiyun 			.cra_flags	 = CRYPTO_ALG_OPTIONAL_KEY,
260*4882a593Smuzhiyun 			.cra_blocksize	 = CRC32_BLOCK_SIZE,
261*4882a593Smuzhiyun 			.cra_ctxsize	 = sizeof(struct crc_ctx),
262*4882a593Smuzhiyun 			.cra_module	 = THIS_MODULE,
263*4882a593Smuzhiyun 			.cra_init	 = crc32_vx_cra_init_zero,
264*4882a593Smuzhiyun 		},
265*4882a593Smuzhiyun 	},
266*4882a593Smuzhiyun 	/* CRC-32C LE */
267*4882a593Smuzhiyun 	{
268*4882a593Smuzhiyun 		.init		=	crc32_vx_init,
269*4882a593Smuzhiyun 		.setkey		=	crc32_vx_setkey,
270*4882a593Smuzhiyun 		.update		=	crc32c_vx_update,
271*4882a593Smuzhiyun 		.final		=	crc32c_vx_final,
272*4882a593Smuzhiyun 		.finup		=	crc32c_vx_finup,
273*4882a593Smuzhiyun 		.digest		=	crc32c_vx_digest,
274*4882a593Smuzhiyun 		.descsize	=	sizeof(struct crc_desc_ctx),
275*4882a593Smuzhiyun 		.digestsize	=	CRC32_DIGEST_SIZE,
276*4882a593Smuzhiyun 		.base		=	{
277*4882a593Smuzhiyun 			.cra_name	 = "crc32c",
278*4882a593Smuzhiyun 			.cra_driver_name = "crc32c-vx",
279*4882a593Smuzhiyun 			.cra_priority	 = 200,
280*4882a593Smuzhiyun 			.cra_flags	 = CRYPTO_ALG_OPTIONAL_KEY,
281*4882a593Smuzhiyun 			.cra_blocksize	 = CRC32_BLOCK_SIZE,
282*4882a593Smuzhiyun 			.cra_ctxsize	 = sizeof(struct crc_ctx),
283*4882a593Smuzhiyun 			.cra_module	 = THIS_MODULE,
284*4882a593Smuzhiyun 			.cra_init	 = crc32_vx_cra_init_invert,
285*4882a593Smuzhiyun 		},
286*4882a593Smuzhiyun 	},
287*4882a593Smuzhiyun };
288*4882a593Smuzhiyun 
289*4882a593Smuzhiyun 
crc_vx_mod_init(void)290*4882a593Smuzhiyun static int __init crc_vx_mod_init(void)
291*4882a593Smuzhiyun {
292*4882a593Smuzhiyun 	return crypto_register_shashes(crc32_vx_algs,
293*4882a593Smuzhiyun 				       ARRAY_SIZE(crc32_vx_algs));
294*4882a593Smuzhiyun }
295*4882a593Smuzhiyun 
crc_vx_mod_exit(void)296*4882a593Smuzhiyun static void __exit crc_vx_mod_exit(void)
297*4882a593Smuzhiyun {
298*4882a593Smuzhiyun 	crypto_unregister_shashes(crc32_vx_algs, ARRAY_SIZE(crc32_vx_algs));
299*4882a593Smuzhiyun }
300*4882a593Smuzhiyun 
301*4882a593Smuzhiyun module_cpu_feature_match(VXRS, crc_vx_mod_init);
302*4882a593Smuzhiyun module_exit(crc_vx_mod_exit);
303*4882a593Smuzhiyun 
304*4882a593Smuzhiyun MODULE_AUTHOR("Hendrik Brueckner <brueckner@linux.vnet.ibm.com>");
305*4882a593Smuzhiyun MODULE_LICENSE("GPL");
306*4882a593Smuzhiyun 
307*4882a593Smuzhiyun MODULE_ALIAS_CRYPTO("crc32");
308*4882a593Smuzhiyun MODULE_ALIAS_CRYPTO("crc32-vx");
309*4882a593Smuzhiyun MODULE_ALIAS_CRYPTO("crc32c");
310*4882a593Smuzhiyun MODULE_ALIAS_CRYPTO("crc32c-vx");
311