mirror of
https://github.com/torvalds/linux.git
synced 2024-12-21 02:21:36 +00:00
7056741fd9
Optimize RAID6 recovery functions to take advantage of the 256-bit YMM integer instructions introduced in AVX2. The patch was tested and benchmarked before submission. However hardware is not yet released so benchmark numbers cannot be reported. Acked-by: "H. Peter Anvin" <hpa@zytor.com> Signed-off-by: Jim Kukunas <james.t.kukunas@linux.intel.com> Signed-off-by: NeilBrown <neilb@suse.de>
71 lines
2.0 KiB
C
71 lines
2.0 KiB
C
/* ----------------------------------------------------------------------- *
|
|
*
|
|
* Copyright 2002-2004 H. Peter Anvin - All Rights Reserved
|
|
*
|
|
* This program is free software; you can redistribute it and/or modify
|
|
* it under the terms of the GNU General Public License as published by
|
|
* the Free Software Foundation, Inc., 53 Temple Place Ste 330,
|
|
* Boston MA 02111-1307, USA; either version 2 of the License, or
|
|
* (at your option) any later version; incorporated herein by reference.
|
|
*
|
|
* ----------------------------------------------------------------------- */
|
|
|
|
/*
|
|
* raid6/x86.h
|
|
*
|
|
* Definitions common to x86 and x86-64 RAID-6 code only
|
|
*/
|
|
|
|
#ifndef LINUX_RAID_RAID6X86_H
|
|
#define LINUX_RAID_RAID6X86_H
|
|
|
|
#if (defined(__i386__) || defined(__x86_64__)) && !defined(__arch_um__)
|
|
|
|
#ifdef __KERNEL__ /* Real code */
|
|
|
|
#include <asm/i387.h>
|
|
|
|
#else /* Dummy code for user space testing */
|
|
|
|
static inline void kernel_fpu_begin(void)
|
|
{
|
|
}
|
|
|
|
static inline void kernel_fpu_end(void)
|
|
{
|
|
}
|
|
|
|
#define __aligned(x) __attribute__((aligned(x)))
|
|
|
|
#define X86_FEATURE_MMX (0*32+23) /* Multimedia Extensions */
|
|
#define X86_FEATURE_FXSR (0*32+24) /* FXSAVE and FXRSTOR instructions
|
|
* (fast save and restore) */
|
|
#define X86_FEATURE_XMM (0*32+25) /* Streaming SIMD Extensions */
|
|
#define X86_FEATURE_XMM2 (0*32+26) /* Streaming SIMD Extensions-2 */
|
|
#define X86_FEATURE_XMM3 (4*32+ 0) /* "pni" SSE-3 */
|
|
#define X86_FEATURE_SSSE3 (4*32+ 9) /* Supplemental SSE-3 */
|
|
#define X86_FEATURE_AVX (4*32+28) /* Advanced Vector Extensions */
|
|
#define X86_FEATURE_AVX2 (9*32+ 5) /* AVX2 instructions */
|
|
#define X86_FEATURE_MMXEXT (1*32+22) /* AMD MMX extensions */
|
|
|
|
/* Should work well enough on modern CPUs for testing */
|
|
static inline int boot_cpu_has(int flag)
|
|
{
|
|
u32 eax, ebx, ecx, edx;
|
|
|
|
eax = (flag & 0x100) ? 7 :
|
|
(flag & 0x20) ? 0x80000001 : 1;
|
|
ecx = 0;
|
|
|
|
asm volatile("cpuid"
|
|
: "+a" (eax), "=b" (ebx), "=d" (edx), "+c" (ecx));
|
|
|
|
return ((flag & 0x100 ? ebx :
|
|
(flag & 0x80) ? ecx : edx) >> (flag & 31)) & 1;
|
|
}
|
|
|
|
#endif /* ndef __KERNEL__ */
|
|
|
|
#endif
|
|
#endif
|