1/* SPDX-License-Identifier: GPL-2.0-or-later */
2/*
3 * Core of the accelerated CRC algorithm.
4 * In your file, define the constants and CRC_FUNCTION_NAME
5 * Then include this file.
6 *
7 * Calculate the checksum of data that is 16 byte aligned and a multiple of
8 * 16 bytes.
9 *
10 * The first step is to reduce it to 1024 bits. We do this in 8 parallel
11 * chunks in order to mask the latency of the vpmsum instructions. If we
12 * have more than 32 kB of data to checksum we repeat this step multiple
13 * times, passing in the previous 1024 bits.
14 *
15 * The next step is to reduce the 1024 bits to 64 bits. This step adds
16 * 32 bits of 0s to the end - this matches what a CRC does. We just
17 * calculate constants that land the data in this 32 bits.
18 *
19 * We then use fixed point Barrett reduction to compute a mod n over GF(2)
20 * for n = CRC using POWER8 instructions. We use x = 32.
21 *
22 * https://en.wikipedia.org/wiki/Barrett_reduction
23 *
24 * Copyright (C) 2015 Anton Blanchard <anton@au.ibm.com>, IBM
25*/
26
27#include <asm/ppc_asm.h>
28#include <asm/ppc-opcode.h>
29
30#define MAX_SIZE	32768
31
32	.text
33
34#if defined(__BIG_ENDIAN__) && defined(REFLECT)
35#define BYTESWAP_DATA
36#elif defined(__LITTLE_ENDIAN__) && !defined(REFLECT)
37#define BYTESWAP_DATA
38#else
39#undef BYTESWAP_DATA
40#endif
41
42#define off16		r25
43#define off32		r26
44#define off48		r27
45#define off64		r28
46#define off80		r29
47#define off96		r30
48#define off112		r31
49
50#define const1		v24
51#define const2		v25
52
53#define byteswap	v26
54#define	mask_32bit	v27
55#define	mask_64bit	v28
56#define zeroes		v29
57
58#ifdef BYTESWAP_DATA
59#define VPERM(A, B, C, D) vperm	A, B, C, D
60#else
61#define VPERM(A, B, C, D)
62#endif
63
64/* unsigned int CRC_FUNCTION_NAME(unsigned int crc, void *p, unsigned long len) */
65FUNC_START(CRC_FUNCTION_NAME)
66	std	r31,-8(r1)
67	std	r30,-16(r1)
68	std	r29,-24(r1)
69	std	r28,-32(r1)
70	std	r27,-40(r1)
71	std	r26,-48(r1)
72	std	r25,-56(r1)
73
74	li	off16,16
75	li	off32,32
76	li	off48,48
77	li	off64,64
78	li	off80,80
79	li	off96,96
80	li	off112,112
81	li	r0,0
82
83	/* Enough room for saving 10 non volatile VMX registers */
84	subi	r6,r1,56+10*16
85	subi	r7,r1,56+2*16
86
87	stvx	v20,0,r6
88	stvx	v21,off16,r6
89	stvx	v22,off32,r6
90	stvx	v23,off48,r6
91	stvx	v24,off64,r6
92	stvx	v25,off80,r6
93	stvx	v26,off96,r6
94	stvx	v27,off112,r6
95	stvx	v28,0,r7
96	stvx	v29,off16,r7
97
98	mr	r10,r3
99
100	vxor	zeroes,zeroes,zeroes
101	vspltisw v0,-1
102
103	vsldoi	mask_32bit,zeroes,v0,4
104	vsldoi	mask_64bit,zeroes,v0,8
105
106	/* Get the initial value into v8 */
107	vxor	v8,v8,v8
108	MTVRD(v8, R3)
109#ifdef REFLECT
110	vsldoi	v8,zeroes,v8,8	/* shift into bottom 32 bits */
111#else
112	vsldoi	v8,v8,zeroes,4	/* shift into top 32 bits */
113#endif
114
115#ifdef BYTESWAP_DATA
116	addis	r3,r2,.byteswap_constant@toc@ha
117	addi	r3,r3,.byteswap_constant@toc@l
118
119	lvx	byteswap,0,r3
120	addi	r3,r3,16
121#endif
122
123	cmpdi	r5,256
124	blt	.Lshort
125
126	rldicr	r6,r5,0,56
127
128	/* Checksum in blocks of MAX_SIZE */
1291:	lis	r7,MAX_SIZE@h
130	ori	r7,r7,MAX_SIZE@l
131	mr	r9,r7
132	cmpd	r6,r7
133	bgt	2f
134	mr	r7,r6
1352:	subf	r6,r7,r6
136
137	/* our main loop does 128 bytes at a time */
138	srdi	r7,r7,7
139
140	/*
141	 * Work out the offset into the constants table to start at. Each
142	 * constant is 16 bytes, and it is used against 128 bytes of input
143	 * data - 128 / 16 = 8
144	 */
145	sldi	r8,r7,4
146	srdi	r9,r9,3
147	subf	r8,r8,r9
148
149	/* We reduce our final 128 bytes in a separate step */
150	addi	r7,r7,-1
151	mtctr	r7
152
153	addis	r3,r2,.constants@toc@ha
154	addi	r3,r3,.constants@toc@l
155
156	/* Find the start of our constants */
157	add	r3,r3,r8
158
159	/* zero v0-v7 which will contain our checksums */
160	vxor	v0,v0,v0
161	vxor	v1,v1,v1
162	vxor	v2,v2,v2
163	vxor	v3,v3,v3
164	vxor	v4,v4,v4
165	vxor	v5,v5,v5
166	vxor	v6,v6,v6
167	vxor	v7,v7,v7
168
169	lvx	const1,0,r3
170
171	/*
172	 * If we are looping back to consume more data we use the values
173	 * already in v16-v23.
174	 */
175	cmpdi	r0,1
176	beq	2f
177
178	/* First warm up pass */
179	lvx	v16,0,r4
180	lvx	v17,off16,r4
181	VPERM(v16,v16,v16,byteswap)
182	VPERM(v17,v17,v17,byteswap)
183	lvx	v18,off32,r4
184	lvx	v19,off48,r4
185	VPERM(v18,v18,v18,byteswap)
186	VPERM(v19,v19,v19,byteswap)
187	lvx	v20,off64,r4
188	lvx	v21,off80,r4
189	VPERM(v20,v20,v20,byteswap)
190	VPERM(v21,v21,v21,byteswap)
191	lvx	v22,off96,r4
192	lvx	v23,off112,r4
193	VPERM(v22,v22,v22,byteswap)
194	VPERM(v23,v23,v23,byteswap)
195	addi	r4,r4,8*16
196
197	/* xor in initial value */
198	vxor	v16,v16,v8
199
2002:	bdz	.Lfirst_warm_up_done
201
202	addi	r3,r3,16
203	lvx	const2,0,r3
204
205	/* Second warm up pass */
206	VPMSUMD(v8,v16,const1)
207	lvx	v16,0,r4
208	VPERM(v16,v16,v16,byteswap)
209	ori	r2,r2,0
210
211	VPMSUMD(v9,v17,const1)
212	lvx	v17,off16,r4
213	VPERM(v17,v17,v17,byteswap)
214	ori	r2,r2,0
215
216	VPMSUMD(v10,v18,const1)
217	lvx	v18,off32,r4
218	VPERM(v18,v18,v18,byteswap)
219	ori	r2,r2,0
220
221	VPMSUMD(v11,v19,const1)
222	lvx	v19,off48,r4
223	VPERM(v19,v19,v19,byteswap)
224	ori	r2,r2,0
225
226	VPMSUMD(v12,v20,const1)
227	lvx	v20,off64,r4
228	VPERM(v20,v20,v20,byteswap)
229	ori	r2,r2,0
230
231	VPMSUMD(v13,v21,const1)
232	lvx	v21,off80,r4
233	VPERM(v21,v21,v21,byteswap)
234	ori	r2,r2,0
235
236	VPMSUMD(v14,v22,const1)
237	lvx	v22,off96,r4
238	VPERM(v22,v22,v22,byteswap)
239	ori	r2,r2,0
240
241	VPMSUMD(v15,v23,const1)
242	lvx	v23,off112,r4
243	VPERM(v23,v23,v23,byteswap)
244
245	addi	r4,r4,8*16
246
247	bdz	.Lfirst_cool_down
248
249	/*
250	 * main loop. We modulo schedule it such that it takes three iterations
251	 * to complete - first iteration load, second iteration vpmsum, third
252	 * iteration xor.
253	 */
254	.balign	16
2554:	lvx	const1,0,r3
256	addi	r3,r3,16
257	ori	r2,r2,0
258
259	vxor	v0,v0,v8
260	VPMSUMD(v8,v16,const2)
261	lvx	v16,0,r4
262	VPERM(v16,v16,v16,byteswap)
263	ori	r2,r2,0
264
265	vxor	v1,v1,v9
266	VPMSUMD(v9,v17,const2)
267	lvx	v17,off16,r4
268	VPERM(v17,v17,v17,byteswap)
269	ori	r2,r2,0
270
271	vxor	v2,v2,v10
272	VPMSUMD(v10,v18,const2)
273	lvx	v18,off32,r4
274	VPERM(v18,v18,v18,byteswap)
275	ori	r2,r2,0
276
277	vxor	v3,v3,v11
278	VPMSUMD(v11,v19,const2)
279	lvx	v19,off48,r4
280	VPERM(v19,v19,v19,byteswap)
281	lvx	const2,0,r3
282	ori	r2,r2,0
283
284	vxor	v4,v4,v12
285	VPMSUMD(v12,v20,const1)
286	lvx	v20,off64,r4
287	VPERM(v20,v20,v20,byteswap)
288	ori	r2,r2,0
289
290	vxor	v5,v5,v13
291	VPMSUMD(v13,v21,const1)
292	lvx	v21,off80,r4
293	VPERM(v21,v21,v21,byteswap)
294	ori	r2,r2,0
295
296	vxor	v6,v6,v14
297	VPMSUMD(v14,v22,const1)
298	lvx	v22,off96,r4
299	VPERM(v22,v22,v22,byteswap)
300	ori	r2,r2,0
301
302	vxor	v7,v7,v15
303	VPMSUMD(v15,v23,const1)
304	lvx	v23,off112,r4
305	VPERM(v23,v23,v23,byteswap)
306
307	addi	r4,r4,8*16
308
309	bdnz	4b
310
311.Lfirst_cool_down:
312	/* First cool down pass */
313	lvx	const1,0,r3
314	addi	r3,r3,16
315
316	vxor	v0,v0,v8
317	VPMSUMD(v8,v16,const1)
318	ori	r2,r2,0
319
320	vxor	v1,v1,v9
321	VPMSUMD(v9,v17,const1)
322	ori	r2,r2,0
323
324	vxor	v2,v2,v10
325	VPMSUMD(v10,v18,const1)
326	ori	r2,r2,0
327
328	vxor	v3,v3,v11
329	VPMSUMD(v11,v19,const1)
330	ori	r2,r2,0
331
332	vxor	v4,v4,v12
333	VPMSUMD(v12,v20,const1)
334	ori	r2,r2,0
335
336	vxor	v5,v5,v13
337	VPMSUMD(v13,v21,const1)
338	ori	r2,r2,0
339
340	vxor	v6,v6,v14
341	VPMSUMD(v14,v22,const1)
342	ori	r2,r2,0
343
344	vxor	v7,v7,v15
345	VPMSUMD(v15,v23,const1)
346	ori	r2,r2,0
347
348.Lsecond_cool_down:
349	/* Second cool down pass */
350	vxor	v0,v0,v8
351	vxor	v1,v1,v9
352	vxor	v2,v2,v10
353	vxor	v3,v3,v11
354	vxor	v4,v4,v12
355	vxor	v5,v5,v13
356	vxor	v6,v6,v14
357	vxor	v7,v7,v15
358
359#ifdef REFLECT
360	/*
361	 * vpmsumd produces a 96 bit result in the least significant bits
362	 * of the register. Since we are bit reflected we have to shift it
363	 * left 32 bits so it occupies the least significant bits in the
364	 * bit reflected domain.
365	 */
366	vsldoi	v0,v0,zeroes,4
367	vsldoi	v1,v1,zeroes,4
368	vsldoi	v2,v2,zeroes,4
369	vsldoi	v3,v3,zeroes,4
370	vsldoi	v4,v4,zeroes,4
371	vsldoi	v5,v5,zeroes,4
372	vsldoi	v6,v6,zeroes,4
373	vsldoi	v7,v7,zeroes,4
374#endif
375
376	/* xor with last 1024 bits */
377	lvx	v8,0,r4
378	lvx	v9,off16,r4
379	VPERM(v8,v8,v8,byteswap)
380	VPERM(v9,v9,v9,byteswap)
381	lvx	v10,off32,r4
382	lvx	v11,off48,r4
383	VPERM(v10,v10,v10,byteswap)
384	VPERM(v11,v11,v11,byteswap)
385	lvx	v12,off64,r4
386	lvx	v13,off80,r4
387	VPERM(v12,v12,v12,byteswap)
388	VPERM(v13,v13,v13,byteswap)
389	lvx	v14,off96,r4
390	lvx	v15,off112,r4
391	VPERM(v14,v14,v14,byteswap)
392	VPERM(v15,v15,v15,byteswap)
393
394	addi	r4,r4,8*16
395
396	vxor	v16,v0,v8
397	vxor	v17,v1,v9
398	vxor	v18,v2,v10
399	vxor	v19,v3,v11
400	vxor	v20,v4,v12
401	vxor	v21,v5,v13
402	vxor	v22,v6,v14
403	vxor	v23,v7,v15
404
405	li	r0,1
406	cmpdi	r6,0
407	addi	r6,r6,128
408	bne	1b
409
410	/* Work out how many bytes we have left */
411	andi.	r5,r5,127
412
413	/* Calculate where in the constant table we need to start */
414	subfic	r6,r5,128
415	add	r3,r3,r6
416
417	/* How many 16 byte chunks are in the tail */
418	srdi	r7,r5,4
419	mtctr	r7
420
421	/*
422	 * Reduce the previously calculated 1024 bits to 64 bits, shifting
423	 * 32 bits to include the trailing 32 bits of zeros
424	 */
425	lvx	v0,0,r3
426	lvx	v1,off16,r3
427	lvx	v2,off32,r3
428	lvx	v3,off48,r3
429	lvx	v4,off64,r3
430	lvx	v5,off80,r3
431	lvx	v6,off96,r3
432	lvx	v7,off112,r3
433	addi	r3,r3,8*16
434
435	VPMSUMW(v0,v16,v0)
436	VPMSUMW(v1,v17,v1)
437	VPMSUMW(v2,v18,v2)
438	VPMSUMW(v3,v19,v3)
439	VPMSUMW(v4,v20,v4)
440	VPMSUMW(v5,v21,v5)
441	VPMSUMW(v6,v22,v6)
442	VPMSUMW(v7,v23,v7)
443
444	/* Now reduce the tail (0 - 112 bytes) */
445	cmpdi	r7,0
446	beq	1f
447
448	lvx	v16,0,r4
449	lvx	v17,0,r3
450	VPERM(v16,v16,v16,byteswap)
451	VPMSUMW(v16,v16,v17)
452	vxor	v0,v0,v16
453	bdz	1f
454
455	lvx	v16,off16,r4
456	lvx	v17,off16,r3
457	VPERM(v16,v16,v16,byteswap)
458	VPMSUMW(v16,v16,v17)
459	vxor	v0,v0,v16
460	bdz	1f
461
462	lvx	v16,off32,r4
463	lvx	v17,off32,r3
464	VPERM(v16,v16,v16,byteswap)
465	VPMSUMW(v16,v16,v17)
466	vxor	v0,v0,v16
467	bdz	1f
468
469	lvx	v16,off48,r4
470	lvx	v17,off48,r3
471	VPERM(v16,v16,v16,byteswap)
472	VPMSUMW(v16,v16,v17)
473	vxor	v0,v0,v16
474	bdz	1f
475
476	lvx	v16,off64,r4
477	lvx	v17,off64,r3
478	VPERM(v16,v16,v16,byteswap)
479	VPMSUMW(v16,v16,v17)
480	vxor	v0,v0,v16
481	bdz	1f
482
483	lvx	v16,off80,r4
484	lvx	v17,off80,r3
485	VPERM(v16,v16,v16,byteswap)
486	VPMSUMW(v16,v16,v17)
487	vxor	v0,v0,v16
488	bdz	1f
489
490	lvx	v16,off96,r4
491	lvx	v17,off96,r3
492	VPERM(v16,v16,v16,byteswap)
493	VPMSUMW(v16,v16,v17)
494	vxor	v0,v0,v16
495
496	/* Now xor all the parallel chunks together */
4971:	vxor	v0,v0,v1
498	vxor	v2,v2,v3
499	vxor	v4,v4,v5
500	vxor	v6,v6,v7
501
502	vxor	v0,v0,v2
503	vxor	v4,v4,v6
504
505	vxor	v0,v0,v4
506
507.Lbarrett_reduction:
508	/* Barrett constants */
509	addis	r3,r2,.barrett_constants@toc@ha
510	addi	r3,r3,.barrett_constants@toc@l
511
512	lvx	const1,0,r3
513	lvx	const2,off16,r3
514
515	vsldoi	v1,v0,v0,8
516	vxor	v0,v0,v1		/* xor two 64 bit results together */
517
518#ifdef REFLECT
519	/* shift left one bit */
520	vspltisb v1,1
521	vsl	v0,v0,v1
522#endif
523
524	vand	v0,v0,mask_64bit
525#ifndef REFLECT
526	/*
527	 * Now for the Barrett reduction algorithm. The idea is to calculate q,
528	 * the multiple of our polynomial that we need to subtract. By
529	 * doing the computation 2x bits higher (ie 64 bits) and shifting the
530	 * result back down 2x bits, we round down to the nearest multiple.
531	 */
532	VPMSUMD(v1,v0,const1)	/* ma */
533	vsldoi	v1,zeroes,v1,8	/* q = floor(ma/(2^64)) */
534	VPMSUMD(v1,v1,const2)	/* qn */
535	vxor	v0,v0,v1	/* a - qn, subtraction is xor in GF(2) */
536
537	/*
538	 * Get the result into r3. We need to shift it left 8 bytes:
539	 * V0 [ 0 1 2 X ]
540	 * V0 [ 0 X 2 3 ]
541	 */
542	vsldoi	v0,v0,zeroes,8	/* shift result into top 64 bits */
543#else
544	/*
545	 * The reflected version of Barrett reduction. Instead of bit
546	 * reflecting our data (which is expensive to do), we bit reflect our
547	 * constants and our algorithm, which means the intermediate data in
548	 * our vector registers goes from 0-63 instead of 63-0. We can reflect
549	 * the algorithm because we don't carry in mod 2 arithmetic.
550	 */
551	vand	v1,v0,mask_32bit	/* bottom 32 bits of a */
552	VPMSUMD(v1,v1,const1)		/* ma */
553	vand	v1,v1,mask_32bit	/* bottom 32bits of ma */
554	VPMSUMD(v1,v1,const2)		/* qn */
555	vxor	v0,v0,v1		/* a - qn, subtraction is xor in GF(2) */
556
557	/*
558	 * Since we are bit reflected, the result (ie the low 32 bits) is in
559	 * the high 32 bits. We just need to shift it left 4 bytes
560	 * V0 [ 0 1 X 3 ]
561	 * V0 [ 0 X 2 3 ]
562	 */
563	vsldoi	v0,v0,zeroes,4		/* shift result into top 64 bits of */
564#endif
565
566	/* Get it into r3 */
567	MFVRD(R3, v0)
568
569.Lout:
570	subi	r6,r1,56+10*16
571	subi	r7,r1,56+2*16
572
573	lvx	v20,0,r6
574	lvx	v21,off16,r6
575	lvx	v22,off32,r6
576	lvx	v23,off48,r6
577	lvx	v24,off64,r6
578	lvx	v25,off80,r6
579	lvx	v26,off96,r6
580	lvx	v27,off112,r6
581	lvx	v28,0,r7
582	lvx	v29,off16,r7
583
584	ld	r31,-8(r1)
585	ld	r30,-16(r1)
586	ld	r29,-24(r1)
587	ld	r28,-32(r1)
588	ld	r27,-40(r1)
589	ld	r26,-48(r1)
590	ld	r25,-56(r1)
591
592	blr
593
594.Lfirst_warm_up_done:
595	lvx	const1,0,r3
596	addi	r3,r3,16
597
598	VPMSUMD(v8,v16,const1)
599	VPMSUMD(v9,v17,const1)
600	VPMSUMD(v10,v18,const1)
601	VPMSUMD(v11,v19,const1)
602	VPMSUMD(v12,v20,const1)
603	VPMSUMD(v13,v21,const1)
604	VPMSUMD(v14,v22,const1)
605	VPMSUMD(v15,v23,const1)
606
607	b	.Lsecond_cool_down
608
609.Lshort:
610	cmpdi	r5,0
611	beq	.Lzero
612
613	addis	r3,r2,.short_constants@toc@ha
614	addi	r3,r3,.short_constants@toc@l
615
616	/* Calculate where in the constant table we need to start */
617	subfic	r6,r5,256
618	add	r3,r3,r6
619
620	/* How many 16 byte chunks? */
621	srdi	r7,r5,4
622	mtctr	r7
623
624	vxor	v19,v19,v19
625	vxor	v20,v20,v20
626
627	lvx	v0,0,r4
628	lvx	v16,0,r3
629	VPERM(v0,v0,v16,byteswap)
630	vxor	v0,v0,v8	/* xor in initial value */
631	VPMSUMW(v0,v0,v16)
632	bdz	.Lv0
633
634	lvx	v1,off16,r4
635	lvx	v17,off16,r3
636	VPERM(v1,v1,v17,byteswap)
637	VPMSUMW(v1,v1,v17)
638	bdz	.Lv1
639
640	lvx	v2,off32,r4
641	lvx	v16,off32,r3
642	VPERM(v2,v2,v16,byteswap)
643	VPMSUMW(v2,v2,v16)
644	bdz	.Lv2
645
646	lvx	v3,off48,r4
647	lvx	v17,off48,r3
648	VPERM(v3,v3,v17,byteswap)
649	VPMSUMW(v3,v3,v17)
650	bdz	.Lv3
651
652	lvx	v4,off64,r4
653	lvx	v16,off64,r3
654	VPERM(v4,v4,v16,byteswap)
655	VPMSUMW(v4,v4,v16)
656	bdz	.Lv4
657
658	lvx	v5,off80,r4
659	lvx	v17,off80,r3
660	VPERM(v5,v5,v17,byteswap)
661	VPMSUMW(v5,v5,v17)
662	bdz	.Lv5
663
664	lvx	v6,off96,r4
665	lvx	v16,off96,r3
666	VPERM(v6,v6,v16,byteswap)
667	VPMSUMW(v6,v6,v16)
668	bdz	.Lv6
669
670	lvx	v7,off112,r4
671	lvx	v17,off112,r3
672	VPERM(v7,v7,v17,byteswap)
673	VPMSUMW(v7,v7,v17)
674	bdz	.Lv7
675
676	addi	r3,r3,128
677	addi	r4,r4,128
678
679	lvx	v8,0,r4
680	lvx	v16,0,r3
681	VPERM(v8,v8,v16,byteswap)
682	VPMSUMW(v8,v8,v16)
683	bdz	.Lv8
684
685	lvx	v9,off16,r4
686	lvx	v17,off16,r3
687	VPERM(v9,v9,v17,byteswap)
688	VPMSUMW(v9,v9,v17)
689	bdz	.Lv9
690
691	lvx	v10,off32,r4
692	lvx	v16,off32,r3
693	VPERM(v10,v10,v16,byteswap)
694	VPMSUMW(v10,v10,v16)
695	bdz	.Lv10
696
697	lvx	v11,off48,r4
698	lvx	v17,off48,r3
699	VPERM(v11,v11,v17,byteswap)
700	VPMSUMW(v11,v11,v17)
701	bdz	.Lv11
702
703	lvx	v12,off64,r4
704	lvx	v16,off64,r3
705	VPERM(v12,v12,v16,byteswap)
706	VPMSUMW(v12,v12,v16)
707	bdz	.Lv12
708
709	lvx	v13,off80,r4
710	lvx	v17,off80,r3
711	VPERM(v13,v13,v17,byteswap)
712	VPMSUMW(v13,v13,v17)
713	bdz	.Lv13
714
715	lvx	v14,off96,r4
716	lvx	v16,off96,r3
717	VPERM(v14,v14,v16,byteswap)
718	VPMSUMW(v14,v14,v16)
719	bdz	.Lv14
720
721	lvx	v15,off112,r4
722	lvx	v17,off112,r3
723	VPERM(v15,v15,v17,byteswap)
724	VPMSUMW(v15,v15,v17)
725
726.Lv15:	vxor	v19,v19,v15
727.Lv14:	vxor	v20,v20,v14
728.Lv13:	vxor	v19,v19,v13
729.Lv12:	vxor	v20,v20,v12
730.Lv11:	vxor	v19,v19,v11
731.Lv10:	vxor	v20,v20,v10
732.Lv9:	vxor	v19,v19,v9
733.Lv8:	vxor	v20,v20,v8
734.Lv7:	vxor	v19,v19,v7
735.Lv6:	vxor	v20,v20,v6
736.Lv5:	vxor	v19,v19,v5
737.Lv4:	vxor	v20,v20,v4
738.Lv3:	vxor	v19,v19,v3
739.Lv2:	vxor	v20,v20,v2
740.Lv1:	vxor	v19,v19,v1
741.Lv0:	vxor	v20,v20,v0
742
743	vxor	v0,v19,v20
744
745	b	.Lbarrett_reduction
746
747.Lzero:
748	mr	r3,r10
749	b	.Lout
750
751FUNC_END(CRC_FUNCTION_NAME)
752