forked from Minki/linux
x86: cpa: convert ioremap to new API
Signed-off-by: Ingo Molnar <mingo@elte.hu> Signed-off-by: Thomas Gleixner <tglx@linutronix.de>
This commit is contained in:
parent
5f8681529c
commit
d806e5ee20
@ -19,6 +19,11 @@
|
||||
#include <asm/pgtable.h>
|
||||
#include <asm/tlbflush.h>
|
||||
|
||||
enum ioremap_mode {
|
||||
IOR_MODE_UNCACHED,
|
||||
IOR_MODE_CACHED,
|
||||
};
|
||||
|
||||
#ifdef CONFIG_X86_64
|
||||
|
||||
unsigned long __phys_addr(unsigned long x)
|
||||
@ -64,19 +69,17 @@ int page_is_ram(unsigned long pagenr)
|
||||
* Fix up the linear direct mapping of the kernel to avoid cache attribute
|
||||
* conflicts.
|
||||
*/
|
||||
static int ioremap_change_attr(unsigned long phys_addr, unsigned long size,
|
||||
pgprot_t prot)
|
||||
static int ioremap_change_attr(unsigned long paddr, unsigned long size,
|
||||
enum ioremap_mode mode)
|
||||
{
|
||||
unsigned long npages, vaddr, last_addr = phys_addr + size - 1;
|
||||
unsigned long vaddr = (unsigned long)__va(paddr);
|
||||
unsigned long nrpages = size >> PAGE_SHIFT;
|
||||
int err, level;
|
||||
|
||||
/* No change for pages after the last mapping */
|
||||
if (last_addr >= (max_pfn_mapped << PAGE_SHIFT))
|
||||
if ((paddr + size - 1) >= (max_pfn_mapped << PAGE_SHIFT))
|
||||
return 0;
|
||||
|
||||
npages = (size + PAGE_SIZE - 1) >> PAGE_SHIFT;
|
||||
vaddr = (unsigned long) __va(phys_addr);
|
||||
|
||||
/*
|
||||
* If there is no identity map for this address,
|
||||
* change_page_attr_addr is unnecessary
|
||||
@ -84,13 +87,15 @@ static int ioremap_change_attr(unsigned long phys_addr, unsigned long size,
|
||||
if (!lookup_address(vaddr, &level))
|
||||
return 0;
|
||||
|
||||
/*
|
||||
* Must use an address here and not struct page because the
|
||||
* phys addr can be a in hole between nodes and not have a
|
||||
* memmap entry.
|
||||
*/
|
||||
err = change_page_attr_addr(vaddr, npages, prot);
|
||||
|
||||
switch (mode) {
|
||||
case IOR_MODE_UNCACHED:
|
||||
default:
|
||||
err = set_memory_uc(vaddr, nrpages);
|
||||
break;
|
||||
case IOR_MODE_CACHED:
|
||||
err = set_memory_wb(vaddr, nrpages);
|
||||
break;
|
||||
}
|
||||
if (!err)
|
||||
global_flush_tlb();
|
||||
|
||||
@ -107,12 +112,12 @@ static int ioremap_change_attr(unsigned long phys_addr, unsigned long size,
|
||||
* caller shouldn't need to know that small detail.
|
||||
*/
|
||||
static void __iomem *__ioremap(unsigned long phys_addr, unsigned long size,
|
||||
unsigned long flags)
|
||||
enum ioremap_mode mode)
|
||||
{
|
||||
void __iomem *addr;
|
||||
struct vm_struct *area;
|
||||
unsigned long offset, last_addr;
|
||||
pgprot_t pgprot;
|
||||
pgprot_t prot;
|
||||
|
||||
/* Don't allow wraparound or zero size */
|
||||
last_addr = phys_addr + size - 1;
|
||||
@ -134,7 +139,15 @@ static void __iomem *__ioremap(unsigned long phys_addr, unsigned long size,
|
||||
return NULL;
|
||||
}
|
||||
|
||||
pgprot = MAKE_GLOBAL(__PAGE_KERNEL | flags);
|
||||
switch (mode) {
|
||||
case IOR_MODE_UNCACHED:
|
||||
default:
|
||||
prot = PAGE_KERNEL_NOCACHE;
|
||||
break;
|
||||
case IOR_MODE_CACHED:
|
||||
prot = PAGE_KERNEL;
|
||||
break;
|
||||
}
|
||||
|
||||
/*
|
||||
* Mappings have to be page-aligned
|
||||
@ -152,12 +165,12 @@ static void __iomem *__ioremap(unsigned long phys_addr, unsigned long size,
|
||||
area->phys_addr = phys_addr;
|
||||
addr = (void __iomem *) area->addr;
|
||||
if (ioremap_page_range((unsigned long)addr, (unsigned long)addr + size,
|
||||
phys_addr, pgprot)) {
|
||||
phys_addr, prot)) {
|
||||
remove_vm_area((void *)(PAGE_MASK & (unsigned long) addr));
|
||||
return NULL;
|
||||
}
|
||||
|
||||
if (ioremap_change_attr(phys_addr, size, pgprot) < 0) {
|
||||
if (ioremap_change_attr(phys_addr, size, mode) < 0) {
|
||||
vunmap(addr);
|
||||
return NULL;
|
||||
}
|
||||
@ -188,13 +201,13 @@ static void __iomem *__ioremap(unsigned long phys_addr, unsigned long size,
|
||||
*/
|
||||
void __iomem *ioremap_nocache(unsigned long phys_addr, unsigned long size)
|
||||
{
|
||||
return __ioremap(phys_addr, size, _PAGE_PCD | _PAGE_PWT);
|
||||
return __ioremap(phys_addr, size, IOR_MODE_UNCACHED);
|
||||
}
|
||||
EXPORT_SYMBOL(ioremap_nocache);
|
||||
|
||||
void __iomem *ioremap_cache(unsigned long phys_addr, unsigned long size)
|
||||
{
|
||||
return __ioremap(phys_addr, size, 0);
|
||||
return __ioremap(phys_addr, size, IOR_MODE_CACHED);
|
||||
}
|
||||
EXPORT_SYMBOL(ioremap_cache);
|
||||
|
||||
@ -242,7 +255,7 @@ void iounmap(volatile void __iomem *addr)
|
||||
}
|
||||
|
||||
/* Reset the direct mapping. Can block */
|
||||
ioremap_change_attr(p->phys_addr, p->size, PAGE_KERNEL);
|
||||
ioremap_change_attr(p->phys_addr, p->size, IOR_MODE_CACHED);
|
||||
|
||||
/* Finally remove it */
|
||||
o = remove_vm_area((void *)addr);
|
||||
|
Loading…
Reference in New Issue
Block a user