forked from Minki/linux
198 lines
4.3 KiB
C
198 lines
4.3 KiB
C
|
/*
|
||
|
* Copyright (C) 2004-2006 Atmel Corporation
|
||
|
*
|
||
|
* This program is free software; you can redistribute it and/or modify
|
||
|
* it under the terms of the GNU General Public License version 2 as
|
||
|
* published by the Free Software Foundation.
|
||
|
*/
|
||
|
#include <linux/vmalloc.h>
|
||
|
#include <linux/module.h>
|
||
|
|
||
|
#include <asm/io.h>
|
||
|
#include <asm/pgtable.h>
|
||
|
#include <asm/cacheflush.h>
|
||
|
#include <asm/tlbflush.h>
|
||
|
#include <asm/addrspace.h>
|
||
|
|
||
|
static inline int remap_area_pte(pte_t *pte, unsigned long address,
|
||
|
unsigned long end, unsigned long phys_addr,
|
||
|
pgprot_t prot)
|
||
|
{
|
||
|
unsigned long pfn;
|
||
|
|
||
|
pfn = phys_addr >> PAGE_SHIFT;
|
||
|
do {
|
||
|
WARN_ON(!pte_none(*pte));
|
||
|
|
||
|
set_pte(pte, pfn_pte(pfn, prot));
|
||
|
address += PAGE_SIZE;
|
||
|
pfn++;
|
||
|
pte++;
|
||
|
} while (address && (address < end));
|
||
|
|
||
|
return 0;
|
||
|
}
|
||
|
|
||
|
static inline int remap_area_pmd(pmd_t *pmd, unsigned long address,
|
||
|
unsigned long end, unsigned long phys_addr,
|
||
|
pgprot_t prot)
|
||
|
{
|
||
|
unsigned long next;
|
||
|
|
||
|
phys_addr -= address;
|
||
|
|
||
|
do {
|
||
|
pte_t *pte = pte_alloc_kernel(pmd, address);
|
||
|
if (!pte)
|
||
|
return -ENOMEM;
|
||
|
|
||
|
next = (address + PMD_SIZE) & PMD_MASK;
|
||
|
if (remap_area_pte(pte, address, next,
|
||
|
address + phys_addr, prot))
|
||
|
return -ENOMEM;
|
||
|
|
||
|
address = next;
|
||
|
pmd++;
|
||
|
} while (address && (address < end));
|
||
|
return 0;
|
||
|
}
|
||
|
|
||
|
static int remap_area_pud(pud_t *pud, unsigned long address,
|
||
|
unsigned long end, unsigned long phys_addr,
|
||
|
pgprot_t prot)
|
||
|
{
|
||
|
unsigned long next;
|
||
|
|
||
|
phys_addr -= address;
|
||
|
|
||
|
do {
|
||
|
pmd_t *pmd = pmd_alloc(&init_mm, pud, address);
|
||
|
if (!pmd)
|
||
|
return -ENOMEM;
|
||
|
next = (address + PUD_SIZE) & PUD_MASK;
|
||
|
if (remap_area_pmd(pmd, address, next,
|
||
|
phys_addr + address, prot))
|
||
|
return -ENOMEM;
|
||
|
|
||
|
address = next;
|
||
|
pud++;
|
||
|
} while (address && address < end);
|
||
|
|
||
|
return 0;
|
||
|
}
|
||
|
|
||
|
static int remap_area_pages(unsigned long address, unsigned long phys_addr,
|
||
|
size_t size, pgprot_t prot)
|
||
|
{
|
||
|
unsigned long end = address + size;
|
||
|
unsigned long next;
|
||
|
pgd_t *pgd;
|
||
|
int err = 0;
|
||
|
|
||
|
phys_addr -= address;
|
||
|
|
||
|
pgd = pgd_offset_k(address);
|
||
|
flush_cache_all();
|
||
|
BUG_ON(address >= end);
|
||
|
|
||
|
spin_lock(&init_mm.page_table_lock);
|
||
|
do {
|
||
|
pud_t *pud = pud_alloc(&init_mm, pgd, address);
|
||
|
|
||
|
err = -ENOMEM;
|
||
|
if (!pud)
|
||
|
break;
|
||
|
|
||
|
next = (address + PGDIR_SIZE) & PGDIR_MASK;
|
||
|
if (next < address || next > end)
|
||
|
next = end;
|
||
|
err = remap_area_pud(pud, address, next,
|
||
|
phys_addr + address, prot);
|
||
|
if (err)
|
||
|
break;
|
||
|
|
||
|
address = next;
|
||
|
pgd++;
|
||
|
} while (address && (address < end));
|
||
|
|
||
|
spin_unlock(&init_mm.page_table_lock);
|
||
|
flush_tlb_all();
|
||
|
return err;
|
||
|
}
|
||
|
|
||
|
/*
|
||
|
* Re-map an arbitrary physical address space into the kernel virtual
|
||
|
* address space. Needed when the kernel wants to access physical
|
||
|
* memory directly.
|
||
|
*/
|
||
|
void __iomem *__ioremap(unsigned long phys_addr, size_t size,
|
||
|
unsigned long flags)
|
||
|
{
|
||
|
void *addr;
|
||
|
struct vm_struct *area;
|
||
|
unsigned long offset, last_addr;
|
||
|
pgprot_t prot;
|
||
|
|
||
|
/*
|
||
|
* Check if we can simply use the P4 segment. This area is
|
||
|
* uncacheable, so if caching/buffering is requested, we can't
|
||
|
* use it.
|
||
|
*/
|
||
|
if ((phys_addr >= P4SEG) && (flags == 0))
|
||
|
return (void __iomem *)phys_addr;
|
||
|
|
||
|
/* Don't allow wraparound or zero size */
|
||
|
last_addr = phys_addr + size - 1;
|
||
|
if (!size || last_addr < phys_addr)
|
||
|
return NULL;
|
||
|
|
||
|
/*
|
||
|
* XXX: When mapping regular RAM, we'd better make damn sure
|
||
|
* it's never used for anything else. But this is really the
|
||
|
* caller's responsibility...
|
||
|
*/
|
||
|
if (PHYSADDR(P2SEGADDR(phys_addr)) == phys_addr)
|
||
|
return (void __iomem *)P2SEGADDR(phys_addr);
|
||
|
|
||
|
/* Mappings have to be page-aligned */
|
||
|
offset = phys_addr & ~PAGE_MASK;
|
||
|
phys_addr &= PAGE_MASK;
|
||
|
size = PAGE_ALIGN(last_addr + 1) - phys_addr;
|
||
|
|
||
|
prot = __pgprot(_PAGE_PRESENT | _PAGE_RW | _PAGE_DIRTY
|
||
|
| _PAGE_ACCESSED | _PAGE_TYPE_SMALL | flags);
|
||
|
|
||
|
/*
|
||
|
* Ok, go for it..
|
||
|
*/
|
||
|
area = get_vm_area(size, VM_IOREMAP);
|
||
|
if (!area)
|
||
|
return NULL;
|
||
|
area->phys_addr = phys_addr;
|
||
|
addr = area->addr;
|
||
|
if (remap_area_pages((unsigned long)addr, phys_addr, size, prot)) {
|
||
|
vunmap(addr);
|
||
|
return NULL;
|
||
|
}
|
||
|
|
||
|
return (void __iomem *)(offset + (char *)addr);
|
||
|
}
|
||
|
EXPORT_SYMBOL(__ioremap);
|
||
|
|
||
|
void __iounmap(void __iomem *addr)
|
||
|
{
|
||
|
struct vm_struct *p;
|
||
|
|
||
|
if ((unsigned long)addr >= P4SEG)
|
||
|
return;
|
||
|
|
||
|
p = remove_vm_area((void *)(PAGE_MASK & (unsigned long __force)addr));
|
||
|
if (unlikely(!p)) {
|
||
|
printk (KERN_ERR "iounmap: bad address %p\n", addr);
|
||
|
return;
|
||
|
}
|
||
|
|
||
|
kfree (p);
|
||
|
}
|
||
|
EXPORT_SYMBOL(__iounmap);
|