123456789101112131415161718192021222324252627 |
- #ifndef _ASM_X86_XOR_64_H
- #define _ASM_X86_XOR_64_H
- static struct xor_block_template xor_block_sse = {
- .name = "generic_sse",
- .do_2 = xor_sse_2,
- .do_3 = xor_sse_3,
- .do_4 = xor_sse_4,
- .do_5 = xor_sse_5,
- };
- /* Also try the AVX routines */
- #include <asm/xor_avx.h>
- /* We force the use of the SSE xor block because it can write around L2.
- We may also be able to load into the L1 only depending on how the cpu
- deals with a load to a line that is being prefetched. */
- #undef XOR_TRY_TEMPLATES
- #define XOR_TRY_TEMPLATES \
- do { \
- AVX_XOR_SPEED; \
- xor_speed(&xor_block_sse_pf64); \
- xor_speed(&xor_block_sse); \
- } while (0)
- #endif /* _ASM_X86_XOR_64_H */
|