mirror of
https://github.com/torvalds/linux.git
synced 2024-12-05 10:32:35 +00:00
a95f9ac271
I've been trying to avoid this for a long time ... but per-cpu space has slowly been growing. Tejun has some patches in linux-next that pre-reserve some space (PERCPU_DYNAMIC_EARLY_SIZE) for use before slab comes online ... and this pushes ia64 above the 64K current limit on static percpu space. I could probably squeeze it back under (we are only over by 512 bytes). But I don't think that I'll be able to squeeze it down enough to build a comfortable breathing space - and I don't want to keep nibbling off a dozen bytes here and there every time some generic code bumps us back over the limit. Next available supported page size is 256K ... so we have to quadruple the available space - a bigger jump than I'd like. But perhaps it will be enough to last a few more years before it needs to be increased again. Acked-by: Tejun Heo <tj@kernel.org> Signed-off-by: Tony Luck <tony.luck@intel.com>
224 lines
6.0 KiB
C
224 lines
6.0 KiB
C
#ifndef _ASM_IA64_PAGE_H
|
|
#define _ASM_IA64_PAGE_H
|
|
/*
|
|
* Pagetable related stuff.
|
|
*
|
|
* Copyright (C) 1998, 1999, 2002 Hewlett-Packard Co
|
|
* David Mosberger-Tang <davidm@hpl.hp.com>
|
|
*/
|
|
|
|
#include <asm/intrinsics.h>
|
|
#include <asm/types.h>
|
|
|
|
/*
|
|
* The top three bits of an IA64 address are its Region Number.
|
|
* Different regions are assigned to different purposes.
|
|
*/
|
|
#define RGN_SHIFT (61)
|
|
#define RGN_BASE(r) (__IA64_UL_CONST(r)<<RGN_SHIFT)
|
|
#define RGN_BITS (RGN_BASE(-1))
|
|
|
|
#define RGN_KERNEL 7 /* Identity mapped region */
|
|
#define RGN_UNCACHED 6 /* Identity mapped I/O region */
|
|
#define RGN_GATE 5 /* Gate page, Kernel text, etc */
|
|
#define RGN_HPAGE 4 /* For Huge TLB pages */
|
|
|
|
/*
|
|
* PAGE_SHIFT determines the actual kernel page size.
|
|
*/
|
|
#if defined(CONFIG_IA64_PAGE_SIZE_4KB)
|
|
# define PAGE_SHIFT 12
|
|
#elif defined(CONFIG_IA64_PAGE_SIZE_8KB)
|
|
# define PAGE_SHIFT 13
|
|
#elif defined(CONFIG_IA64_PAGE_SIZE_16KB)
|
|
# define PAGE_SHIFT 14
|
|
#elif defined(CONFIG_IA64_PAGE_SIZE_64KB)
|
|
# define PAGE_SHIFT 16
|
|
#else
|
|
# error Unsupported page size!
|
|
#endif
|
|
|
|
#define PAGE_SIZE (__IA64_UL_CONST(1) << PAGE_SHIFT)
|
|
#define PAGE_MASK (~(PAGE_SIZE - 1))
|
|
|
|
#define PERCPU_PAGE_SHIFT 18 /* log2() of max. size of per-CPU area */
|
|
#define PERCPU_PAGE_SIZE (__IA64_UL_CONST(1) << PERCPU_PAGE_SHIFT)
|
|
|
|
|
|
#ifdef CONFIG_HUGETLB_PAGE
|
|
# define HPAGE_REGION_BASE RGN_BASE(RGN_HPAGE)
|
|
# define HPAGE_SHIFT hpage_shift
|
|
# define HPAGE_SHIFT_DEFAULT 28 /* check ia64 SDM for architecture supported size */
|
|
# define HPAGE_SIZE (__IA64_UL_CONST(1) << HPAGE_SHIFT)
|
|
# define HPAGE_MASK (~(HPAGE_SIZE - 1))
|
|
|
|
# define HAVE_ARCH_HUGETLB_UNMAPPED_AREA
|
|
#endif /* CONFIG_HUGETLB_PAGE */
|
|
|
|
#ifdef __ASSEMBLY__
|
|
# define __pa(x) ((x) - PAGE_OFFSET)
|
|
# define __va(x) ((x) + PAGE_OFFSET)
|
|
#else /* !__ASSEMBLY */
|
|
# define STRICT_MM_TYPECHECKS
|
|
|
|
extern void clear_page (void *page);
|
|
extern void copy_page (void *to, void *from);
|
|
|
|
/*
|
|
* clear_user_page() and copy_user_page() can't be inline functions because
|
|
* flush_dcache_page() can't be defined until later...
|
|
*/
|
|
#define clear_user_page(addr, vaddr, page) \
|
|
do { \
|
|
clear_page(addr); \
|
|
flush_dcache_page(page); \
|
|
} while (0)
|
|
|
|
#define copy_user_page(to, from, vaddr, page) \
|
|
do { \
|
|
copy_page((to), (from)); \
|
|
flush_dcache_page(page); \
|
|
} while (0)
|
|
|
|
|
|
#define __alloc_zeroed_user_highpage(movableflags, vma, vaddr) \
|
|
({ \
|
|
struct page *page = alloc_page_vma( \
|
|
GFP_HIGHUSER | __GFP_ZERO | movableflags, vma, vaddr); \
|
|
if (page) \
|
|
flush_dcache_page(page); \
|
|
page; \
|
|
})
|
|
|
|
#define __HAVE_ARCH_ALLOC_ZEROED_USER_HIGHPAGE
|
|
|
|
#define virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT)
|
|
|
|
#ifdef CONFIG_VIRTUAL_MEM_MAP
|
|
extern int ia64_pfn_valid (unsigned long pfn);
|
|
#else
|
|
# define ia64_pfn_valid(pfn) 1
|
|
#endif
|
|
|
|
#ifdef CONFIG_VIRTUAL_MEM_MAP
|
|
extern struct page *vmem_map;
|
|
#ifdef CONFIG_DISCONTIGMEM
|
|
# define page_to_pfn(page) ((unsigned long) (page - vmem_map))
|
|
# define pfn_to_page(pfn) (vmem_map + (pfn))
|
|
#else
|
|
# include <asm-generic/memory_model.h>
|
|
#endif
|
|
#else
|
|
# include <asm-generic/memory_model.h>
|
|
#endif
|
|
|
|
#ifdef CONFIG_FLATMEM
|
|
# define pfn_valid(pfn) (((pfn) < max_mapnr) && ia64_pfn_valid(pfn))
|
|
#elif defined(CONFIG_DISCONTIGMEM)
|
|
extern unsigned long min_low_pfn;
|
|
extern unsigned long max_low_pfn;
|
|
# define pfn_valid(pfn) (((pfn) >= min_low_pfn) && ((pfn) < max_low_pfn) && ia64_pfn_valid(pfn))
|
|
#endif
|
|
|
|
#define page_to_phys(page) (page_to_pfn(page) << PAGE_SHIFT)
|
|
#define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT)
|
|
#define pfn_to_kaddr(pfn) __va((pfn) << PAGE_SHIFT)
|
|
|
|
typedef union ia64_va {
|
|
struct {
|
|
unsigned long off : 61; /* intra-region offset */
|
|
unsigned long reg : 3; /* region number */
|
|
} f;
|
|
unsigned long l;
|
|
void *p;
|
|
} ia64_va;
|
|
|
|
/*
|
|
* Note: These macros depend on the fact that PAGE_OFFSET has all
|
|
* region bits set to 1 and all other bits set to zero. They are
|
|
* expressed in this way to ensure they result in a single "dep"
|
|
* instruction.
|
|
*/
|
|
#define __pa(x) ({ia64_va _v; _v.l = (long) (x); _v.f.reg = 0; _v.l;})
|
|
#define __va(x) ({ia64_va _v; _v.l = (long) (x); _v.f.reg = -1; _v.p;})
|
|
|
|
#define REGION_NUMBER(x) ({ia64_va _v; _v.l = (long) (x); _v.f.reg;})
|
|
#define REGION_OFFSET(x) ({ia64_va _v; _v.l = (long) (x); _v.f.off;})
|
|
|
|
#ifdef CONFIG_HUGETLB_PAGE
|
|
# define htlbpage_to_page(x) (((unsigned long) REGION_NUMBER(x) << 61) \
|
|
| (REGION_OFFSET(x) >> (HPAGE_SHIFT-PAGE_SHIFT)))
|
|
# define HUGETLB_PAGE_ORDER (HPAGE_SHIFT - PAGE_SHIFT)
|
|
extern unsigned int hpage_shift;
|
|
#endif
|
|
|
|
static __inline__ int
|
|
get_order (unsigned long size)
|
|
{
|
|
long double d = size - 1;
|
|
long order;
|
|
|
|
order = ia64_getf_exp(d);
|
|
order = order - PAGE_SHIFT - 0xffff + 1;
|
|
if (order < 0)
|
|
order = 0;
|
|
return order;
|
|
}
|
|
|
|
#endif /* !__ASSEMBLY__ */
|
|
|
|
#ifdef STRICT_MM_TYPECHECKS
|
|
/*
|
|
* These are used to make use of C type-checking..
|
|
*/
|
|
typedef struct { unsigned long pte; } pte_t;
|
|
typedef struct { unsigned long pmd; } pmd_t;
|
|
#ifdef CONFIG_PGTABLE_4
|
|
typedef struct { unsigned long pud; } pud_t;
|
|
#endif
|
|
typedef struct { unsigned long pgd; } pgd_t;
|
|
typedef struct { unsigned long pgprot; } pgprot_t;
|
|
typedef struct page *pgtable_t;
|
|
|
|
# define pte_val(x) ((x).pte)
|
|
# define pmd_val(x) ((x).pmd)
|
|
#ifdef CONFIG_PGTABLE_4
|
|
# define pud_val(x) ((x).pud)
|
|
#endif
|
|
# define pgd_val(x) ((x).pgd)
|
|
# define pgprot_val(x) ((x).pgprot)
|
|
|
|
# define __pte(x) ((pte_t) { (x) } )
|
|
# define __pgprot(x) ((pgprot_t) { (x) } )
|
|
|
|
#else /* !STRICT_MM_TYPECHECKS */
|
|
/*
|
|
* .. while these make it easier on the compiler
|
|
*/
|
|
# ifndef __ASSEMBLY__
|
|
typedef unsigned long pte_t;
|
|
typedef unsigned long pmd_t;
|
|
typedef unsigned long pgd_t;
|
|
typedef unsigned long pgprot_t;
|
|
typedef struct page *pgtable_t;
|
|
# endif
|
|
|
|
# define pte_val(x) (x)
|
|
# define pmd_val(x) (x)
|
|
# define pgd_val(x) (x)
|
|
# define pgprot_val(x) (x)
|
|
|
|
# define __pte(x) (x)
|
|
# define __pgd(x) (x)
|
|
# define __pgprot(x) (x)
|
|
#endif /* !STRICT_MM_TYPECHECKS */
|
|
|
|
#define PAGE_OFFSET RGN_BASE(RGN_KERNEL)
|
|
|
|
#define VM_DATA_DEFAULT_FLAGS (VM_READ | VM_WRITE | \
|
|
VM_MAYREAD | VM_MAYWRITE | VM_MAYEXEC | \
|
|
(((current->personality & READ_IMPLIES_EXEC) != 0) \
|
|
? VM_EXEC : 0))
|
|
|
|
#endif /* _ASM_IA64_PAGE_H */
|