forked from Minki/linux
1da177e4c3
Initial git repository build. I'm not bothering with the full history, even though we have it. We can create a separate "historical" git archive of that later if we want to, and in the meantime it's about 3.2GB when imported into git - space that would just make the early git days unnecessarily complicated, when we don't have a lot of good infrastructure for it. Let it rip!
122 lines
3.2 KiB
C
122 lines
3.2 KiB
C
#include <linux/init.h>
|
|
#include <linux/mm.h>
|
|
#include <asm/mtrr.h>
|
|
#include <asm/msr.h>
|
|
|
|
#include "mtrr.h"
|
|
|
|
static void
|
|
amd_get_mtrr(unsigned int reg, unsigned long *base,
|
|
unsigned int *size, mtrr_type * type)
|
|
{
|
|
unsigned long low, high;
|
|
|
|
rdmsr(MSR_K6_UWCCR, low, high);
|
|
/* Upper dword is region 1, lower is region 0 */
|
|
if (reg == 1)
|
|
low = high;
|
|
/* The base masks off on the right alignment */
|
|
*base = (low & 0xFFFE0000) >> PAGE_SHIFT;
|
|
*type = 0;
|
|
if (low & 1)
|
|
*type = MTRR_TYPE_UNCACHABLE;
|
|
if (low & 2)
|
|
*type = MTRR_TYPE_WRCOMB;
|
|
if (!(low & 3)) {
|
|
*size = 0;
|
|
return;
|
|
}
|
|
/*
|
|
* This needs a little explaining. The size is stored as an
|
|
* inverted mask of bits of 128K granularity 15 bits long offset
|
|
* 2 bits
|
|
*
|
|
* So to get a size we do invert the mask and add 1 to the lowest
|
|
* mask bit (4 as its 2 bits in). This gives us a size we then shift
|
|
* to turn into 128K blocks
|
|
*
|
|
* eg 111 1111 1111 1100 is 512K
|
|
*
|
|
* invert 000 0000 0000 0011
|
|
* +1 000 0000 0000 0100
|
|
* *128K ...
|
|
*/
|
|
low = (~low) & 0x1FFFC;
|
|
*size = (low + 4) << (15 - PAGE_SHIFT);
|
|
return;
|
|
}
|
|
|
|
static void amd_set_mtrr(unsigned int reg, unsigned long base,
|
|
unsigned long size, mtrr_type type)
|
|
/* [SUMMARY] Set variable MTRR register on the local CPU.
|
|
<reg> The register to set.
|
|
<base> The base address of the region.
|
|
<size> The size of the region. If this is 0 the region is disabled.
|
|
<type> The type of the region.
|
|
<do_safe> If TRUE, do the change safely. If FALSE, safety measures should
|
|
be done externally.
|
|
[RETURNS] Nothing.
|
|
*/
|
|
{
|
|
u32 regs[2];
|
|
|
|
/*
|
|
* Low is MTRR0 , High MTRR 1
|
|
*/
|
|
rdmsr(MSR_K6_UWCCR, regs[0], regs[1]);
|
|
/*
|
|
* Blank to disable
|
|
*/
|
|
if (size == 0)
|
|
regs[reg] = 0;
|
|
else
|
|
/* Set the register to the base, the type (off by one) and an
|
|
inverted bitmask of the size The size is the only odd
|
|
bit. We are fed say 512K We invert this and we get 111 1111
|
|
1111 1011 but if you subtract one and invert you get the
|
|
desired 111 1111 1111 1100 mask
|
|
|
|
But ~(x - 1) == ~x + 1 == -x. Two's complement rocks! */
|
|
regs[reg] = (-size >> (15 - PAGE_SHIFT) & 0x0001FFFC)
|
|
| (base << PAGE_SHIFT) | (type + 1);
|
|
|
|
/*
|
|
* The writeback rule is quite specific. See the manual. Its
|
|
* disable local interrupts, write back the cache, set the mtrr
|
|
*/
|
|
wbinvd();
|
|
wrmsr(MSR_K6_UWCCR, regs[0], regs[1]);
|
|
}
|
|
|
|
static int amd_validate_add_page(unsigned long base, unsigned long size, unsigned int type)
|
|
{
|
|
/* Apply the K6 block alignment and size rules
|
|
In order
|
|
o Uncached or gathering only
|
|
o 128K or bigger block
|
|
o Power of 2 block
|
|
o base suitably aligned to the power
|
|
*/
|
|
if (type > MTRR_TYPE_WRCOMB || size < (1 << (17 - PAGE_SHIFT))
|
|
|| (size & ~(size - 1)) - size || (base & (size - 1)))
|
|
return -EINVAL;
|
|
return 0;
|
|
}
|
|
|
|
static struct mtrr_ops amd_mtrr_ops = {
|
|
.vendor = X86_VENDOR_AMD,
|
|
.set = amd_set_mtrr,
|
|
.get = amd_get_mtrr,
|
|
.get_free_region = generic_get_free_region,
|
|
.validate_add_page = amd_validate_add_page,
|
|
.have_wrcomb = positive_have_wrcomb,
|
|
};
|
|
|
|
int __init amd_init_mtrr(void)
|
|
{
|
|
set_mtrr_ops(&amd_mtrr_ops);
|
|
return 0;
|
|
}
|
|
|
|
//arch_initcall(amd_mtrr_init);
|