-
Notifications
You must be signed in to change notification settings - Fork 0
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
lib/raid6: Add AVX512 optimized gen_syndrome functions
Optimize RAID6 gen_syndrom functions to take advantage of the 512-bit ZMM integer instructions introduced in AVX512. AVX512 optimized gen_syndrom functions, which is simply based on avx2.c written by Yuanhan Liu and sse2.c written by hpa. The patch was tested and benchmarked before submission on a hardware that has AVX512 flags to support such instructions Cc: H. Peter Anvin <[email protected]> Cc: Jim Kukunas <[email protected]> Cc: Fenghua Yu <[email protected]> Signed-off-by: Megha Dey <[email protected]> Signed-off-by: Gayatri Kammela <[email protected]> Reviewed-by: Fenghua Yu <[email protected]> Signed-off-by: Shaohua Li <[email protected]>
- Loading branch information
Showing
6 changed files
with
320 additions
and
3 deletions.
There are no files selected for viewing
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,294 @@ | ||
/* -*- linux-c -*- -------------------------------------------------------- | ||
* | ||
* Copyright (C) 2016 Intel Corporation | ||
* | ||
* Author: Gayatri Kammela <[email protected]> | ||
* Author: Megha Dey <[email protected]> | ||
* | ||
* Based on avx2.c: Copyright 2012 Yuanhan Liu All Rights Reserved | ||
* Based on sse2.c: Copyright 2002 H. Peter Anvin - All Rights Reserved | ||
* | ||
* This program is free software; you can redistribute it and/or modify | ||
* it under the terms of the GNU General Public License as published by | ||
* the Free Software Foundation, Inc., 53 Temple Place Ste 330, | ||
* Boston MA 02111-1307, USA; either version 2 of the License, or | ||
* (at your option) any later version; incorporated herein by reference. | ||
* | ||
* ----------------------------------------------------------------------- | ||
*/ | ||
|
||
/* | ||
* AVX512 implementation of RAID-6 syndrome functions | ||
* | ||
*/ | ||
|
||
#ifdef CONFIG_AS_AVX512 | ||
|
||
#include <linux/raid/pq.h> | ||
#include "x86.h" | ||
|
||
static const struct raid6_avx512_constants { | ||
u64 x1d[8]; | ||
} raid6_avx512_constants __aligned(512) = { | ||
{ 0x1d1d1d1d1d1d1d1dULL, 0x1d1d1d1d1d1d1d1dULL, | ||
0x1d1d1d1d1d1d1d1dULL, 0x1d1d1d1d1d1d1d1dULL, | ||
0x1d1d1d1d1d1d1d1dULL, 0x1d1d1d1d1d1d1d1dULL, | ||
0x1d1d1d1d1d1d1d1dULL, 0x1d1d1d1d1d1d1d1dULL,}, | ||
}; | ||
|
||
static int raid6_have_avx512(void) | ||
{ | ||
return boot_cpu_has(X86_FEATURE_AVX2) && | ||
boot_cpu_has(X86_FEATURE_AVX) && | ||
boot_cpu_has(X86_FEATURE_AVX512F) && | ||
boot_cpu_has(X86_FEATURE_AVX512BW) && | ||
boot_cpu_has(X86_FEATURE_AVX512VL) && | ||
boot_cpu_has(X86_FEATURE_AVX512DQ); | ||
} | ||
|
||
static void raid6_avx5121_gen_syndrome(int disks, size_t bytes, void **ptrs) | ||
{ | ||
u8 **dptr = (u8 **)ptrs; | ||
u8 *p, *q; | ||
int d, z, z0; | ||
|
||
z0 = disks - 3; /* Highest data disk */ | ||
p = dptr[z0+1]; /* XOR parity */ | ||
q = dptr[z0+2]; /* RS syndrome */ | ||
|
||
kernel_fpu_begin(); | ||
|
||
asm volatile("vmovdqa64 %0,%%zmm0\n\t" | ||
"vpxorq %%zmm1,%%zmm1,%%zmm1" /* Zero temp */ | ||
: | ||
: "m" (raid6_avx512_constants.x1d[0])); | ||
|
||
for (d = 0; d < bytes; d += 64) { | ||
asm volatile("prefetchnta %0\n\t" | ||
"vmovdqa64 %0,%%zmm2\n\t" /* P[0] */ | ||
"prefetchnta %1\n\t" | ||
"vmovdqa64 %%zmm2,%%zmm4\n\t" /* Q[0] */ | ||
"vmovdqa64 %1,%%zmm6" | ||
: | ||
: "m" (dptr[z0][d]), "m" (dptr[z0-1][d])); | ||
for (z = z0-2; z >= 0; z--) { | ||
asm volatile("prefetchnta %0\n\t" | ||
"vpcmpgtb %%zmm4,%%zmm1,%%k1\n\t" | ||
"vpmovm2b %%k1,%%zmm5\n\t" | ||
"vpaddb %%zmm4,%%zmm4,%%zmm4\n\t" | ||
"vpandq %%zmm0,%%zmm5,%%zmm5\n\t" | ||
"vpxorq %%zmm5,%%zmm4,%%zmm4\n\t" | ||
"vpxorq %%zmm6,%%zmm2,%%zmm2\n\t" | ||
"vpxorq %%zmm6,%%zmm4,%%zmm4\n\t" | ||
"vmovdqa64 %0,%%zmm6" | ||
: | ||
: "m" (dptr[z][d])); | ||
} | ||
asm volatile("vpcmpgtb %%zmm4,%%zmm1,%%k1\n\t" | ||
"vpmovm2b %%k1,%%zmm5\n\t" | ||
"vpaddb %%zmm4,%%zmm4,%%zmm4\n\t" | ||
"vpandq %%zmm0,%%zmm5,%%zmm5\n\t" | ||
"vpxorq %%zmm5,%%zmm4,%%zmm4\n\t" | ||
"vpxorq %%zmm6,%%zmm2,%%zmm2\n\t" | ||
"vpxorq %%zmm6,%%zmm4,%%zmm4\n\t" | ||
"vmovntdq %%zmm2,%0\n\t" | ||
"vpxorq %%zmm2,%%zmm2,%%zmm2\n\t" | ||
"vmovntdq %%zmm4,%1\n\t" | ||
"vpxorq %%zmm4,%%zmm4,%%zmm4" | ||
: | ||
: "m" (p[d]), "m" (q[d])); | ||
} | ||
|
||
asm volatile("sfence" : : : "memory"); | ||
kernel_fpu_end(); | ||
} | ||
|
||
const struct raid6_calls raid6_avx512x1 = { | ||
raid6_avx5121_gen_syndrome, | ||
NULL, /* XOR not yet implemented */ | ||
raid6_have_avx512, | ||
"avx512x1", | ||
1 /* Has cache hints */ | ||
}; | ||
|
||
/* | ||
* Unrolled-by-2 AVX512 implementation | ||
*/ | ||
static void raid6_avx5122_gen_syndrome(int disks, size_t bytes, void **ptrs) | ||
{ | ||
u8 **dptr = (u8 **)ptrs; | ||
u8 *p, *q; | ||
int d, z, z0; | ||
|
||
z0 = disks - 3; /* Highest data disk */ | ||
p = dptr[z0+1]; /* XOR parity */ | ||
q = dptr[z0+2]; /* RS syndrome */ | ||
|
||
kernel_fpu_begin(); | ||
|
||
asm volatile("vmovdqa64 %0,%%zmm0\n\t" | ||
"vpxorq %%zmm1,%%zmm1,%%zmm1" /* Zero temp */ | ||
: | ||
: "m" (raid6_avx512_constants.x1d[0])); | ||
|
||
/* We uniformly assume a single prefetch covers at least 64 bytes */ | ||
for (d = 0; d < bytes; d += 128) { | ||
asm volatile("prefetchnta %0\n\t" | ||
"prefetchnta %1\n\t" | ||
"vmovdqa64 %0,%%zmm2\n\t" /* P[0] */ | ||
"vmovdqa64 %1,%%zmm3\n\t" /* P[1] */ | ||
"vmovdqa64 %%zmm2,%%zmm4\n\t" /* Q[0] */ | ||
"vmovdqa64 %%zmm3,%%zmm6" /* Q[1] */ | ||
: | ||
: "m" (dptr[z0][d]), "m" (dptr[z0][d+64])); | ||
for (z = z0-1; z >= 0; z--) { | ||
asm volatile("prefetchnta %0\n\t" | ||
"prefetchnta %1\n\t" | ||
"vpcmpgtb %%zmm4,%%zmm1,%%k1\n\t" | ||
"vpcmpgtb %%zmm6,%%zmm1,%%k2\n\t" | ||
"vpmovm2b %%k1,%%zmm5\n\t" | ||
"vpmovm2b %%k2,%%zmm7\n\t" | ||
"vpaddb %%zmm4,%%zmm4,%%zmm4\n\t" | ||
"vpaddb %%zmm6,%%zmm6,%%zmm6\n\t" | ||
"vpandq %%zmm0,%%zmm5,%%zmm5\n\t" | ||
"vpandq %%zmm0,%%zmm7,%%zmm7\n\t" | ||
"vpxorq %%zmm5,%%zmm4,%%zmm4\n\t" | ||
"vpxorq %%zmm7,%%zmm6,%%zmm6\n\t" | ||
"vmovdqa64 %0,%%zmm5\n\t" | ||
"vmovdqa64 %1,%%zmm7\n\t" | ||
"vpxorq %%zmm5,%%zmm2,%%zmm2\n\t" | ||
"vpxorq %%zmm7,%%zmm3,%%zmm3\n\t" | ||
"vpxorq %%zmm5,%%zmm4,%%zmm4\n\t" | ||
"vpxorq %%zmm7,%%zmm6,%%zmm6" | ||
: | ||
: "m" (dptr[z][d]), "m" (dptr[z][d+64])); | ||
} | ||
asm volatile("vmovntdq %%zmm2,%0\n\t" | ||
"vmovntdq %%zmm3,%1\n\t" | ||
"vmovntdq %%zmm4,%2\n\t" | ||
"vmovntdq %%zmm6,%3" | ||
: | ||
: "m" (p[d]), "m" (p[d+64]), "m" (q[d]), | ||
"m" (q[d+64])); | ||
} | ||
|
||
asm volatile("sfence" : : : "memory"); | ||
kernel_fpu_end(); | ||
} | ||
|
||
const struct raid6_calls raid6_avx512x2 = { | ||
raid6_avx5122_gen_syndrome, | ||
NULL, /* XOR not yet implemented */ | ||
raid6_have_avx512, | ||
"avx512x2", | ||
1 /* Has cache hints */ | ||
}; | ||
|
||
#ifdef CONFIG_X86_64 | ||
|
||
/* | ||
* Unrolled-by-4 AVX2 implementation | ||
*/ | ||
static void raid6_avx5124_gen_syndrome(int disks, size_t bytes, void **ptrs) | ||
{ | ||
u8 **dptr = (u8 **)ptrs; | ||
u8 *p, *q; | ||
int d, z, z0; | ||
|
||
z0 = disks - 3; /* Highest data disk */ | ||
p = dptr[z0+1]; /* XOR parity */ | ||
q = dptr[z0+2]; /* RS syndrome */ | ||
|
||
kernel_fpu_begin(); | ||
|
||
asm volatile("vmovdqa64 %0,%%zmm0\n\t" | ||
"vpxorq %%zmm1,%%zmm1,%%zmm1\n\t" /* Zero temp */ | ||
"vpxorq %%zmm2,%%zmm2,%%zmm2\n\t" /* P[0] */ | ||
"vpxorq %%zmm3,%%zmm3,%%zmm3\n\t" /* P[1] */ | ||
"vpxorq %%zmm4,%%zmm4,%%zmm4\n\t" /* Q[0] */ | ||
"vpxorq %%zmm6,%%zmm6,%%zmm6\n\t" /* Q[1] */ | ||
"vpxorq %%zmm10,%%zmm10,%%zmm10\n\t" /* P[2] */ | ||
"vpxorq %%zmm11,%%zmm11,%%zmm11\n\t" /* P[3] */ | ||
"vpxorq %%zmm12,%%zmm12,%%zmm12\n\t" /* Q[2] */ | ||
"vpxorq %%zmm14,%%zmm14,%%zmm14" /* Q[3] */ | ||
: | ||
: "m" (raid6_avx512_constants.x1d[0])); | ||
|
||
for (d = 0; d < bytes; d += 256) { | ||
for (z = z0; z >= 0; z--) { | ||
asm volatile("prefetchnta %0\n\t" | ||
"prefetchnta %1\n\t" | ||
"prefetchnta %2\n\t" | ||
"prefetchnta %3\n\t" | ||
"vpcmpgtb %%zmm4,%%zmm1,%%k1\n\t" | ||
"vpcmpgtb %%zmm6,%%zmm1,%%k2\n\t" | ||
"vpcmpgtb %%zmm12,%%zmm1,%%k3\n\t" | ||
"vpcmpgtb %%zmm14,%%zmm1,%%k4\n\t" | ||
"vpmovm2b %%k1,%%zmm5\n\t" | ||
"vpmovm2b %%k2,%%zmm7\n\t" | ||
"vpmovm2b %%k3,%%zmm13\n\t" | ||
"vpmovm2b %%k4,%%zmm15\n\t" | ||
"vpaddb %%zmm4,%%zmm4,%%zmm4\n\t" | ||
"vpaddb %%zmm6,%%zmm6,%%zmm6\n\t" | ||
"vpaddb %%zmm12,%%zmm12,%%zmm12\n\t" | ||
"vpaddb %%zmm14,%%zmm14,%%zmm14\n\t" | ||
"vpandq %%zmm0,%%zmm5,%%zmm5\n\t" | ||
"vpandq %%zmm0,%%zmm7,%%zmm7\n\t" | ||
"vpandq %%zmm0,%%zmm13,%%zmm13\n\t" | ||
"vpandq %%zmm0,%%zmm15,%%zmm15\n\t" | ||
"vpxorq %%zmm5,%%zmm4,%%zmm4\n\t" | ||
"vpxorq %%zmm7,%%zmm6,%%zmm6\n\t" | ||
"vpxorq %%zmm13,%%zmm12,%%zmm12\n\t" | ||
"vpxorq %%zmm15,%%zmm14,%%zmm14\n\t" | ||
"vmovdqa64 %0,%%zmm5\n\t" | ||
"vmovdqa64 %1,%%zmm7\n\t" | ||
"vmovdqa64 %2,%%zmm13\n\t" | ||
"vmovdqa64 %3,%%zmm15\n\t" | ||
"vpxorq %%zmm5,%%zmm2,%%zmm2\n\t" | ||
"vpxorq %%zmm7,%%zmm3,%%zmm3\n\t" | ||
"vpxorq %%zmm13,%%zmm10,%%zmm10\n\t" | ||
"vpxorq %%zmm15,%%zmm11,%%zmm11\n" | ||
"vpxorq %%zmm5,%%zmm4,%%zmm4\n\t" | ||
"vpxorq %%zmm7,%%zmm6,%%zmm6\n\t" | ||
"vpxorq %%zmm13,%%zmm12,%%zmm12\n\t" | ||
"vpxorq %%zmm15,%%zmm14,%%zmm14" | ||
: | ||
: "m" (dptr[z][d]), "m" (dptr[z][d+64]), | ||
"m" (dptr[z][d+128]), "m" (dptr[z][d+192])); | ||
} | ||
asm volatile("vmovntdq %%zmm2,%0\n\t" | ||
"vpxorq %%zmm2,%%zmm2,%%zmm2\n\t" | ||
"vmovntdq %%zmm3,%1\n\t" | ||
"vpxorq %%zmm3,%%zmm3,%%zmm3\n\t" | ||
"vmovntdq %%zmm10,%2\n\t" | ||
"vpxorq %%zmm10,%%zmm10,%%zmm10\n\t" | ||
"vmovntdq %%zmm11,%3\n\t" | ||
"vpxorq %%zmm11,%%zmm11,%%zmm11\n\t" | ||
"vmovntdq %%zmm4,%4\n\t" | ||
"vpxorq %%zmm4,%%zmm4,%%zmm4\n\t" | ||
"vmovntdq %%zmm6,%5\n\t" | ||
"vpxorq %%zmm6,%%zmm6,%%zmm6\n\t" | ||
"vmovntdq %%zmm12,%6\n\t" | ||
"vpxorq %%zmm12,%%zmm12,%%zmm12\n\t" | ||
"vmovntdq %%zmm14,%7\n\t" | ||
"vpxorq %%zmm14,%%zmm14,%%zmm14" | ||
: | ||
: "m" (p[d]), "m" (p[d+64]), "m" (p[d+128]), | ||
"m" (p[d+192]), "m" (q[d]), "m" (q[d+64]), | ||
"m" (q[d+128]), "m" (q[d+192])); | ||
} | ||
|
||
asm volatile("sfence" : : : "memory"); | ||
kernel_fpu_end(); | ||
} | ||
|
||
const struct raid6_calls raid6_avx512x4 = { | ||
raid6_avx5124_gen_syndrome, | ||
NULL, /* XOR not yet implemented */ | ||
raid6_have_avx512, | ||
"avx512x4", | ||
1 /* Has cache hints */ | ||
}; | ||
#endif | ||
|
||
#endif /* CONFIG_AS_AVX512 */ |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters