mirror of
https://github.com/torvalds/linux.git
synced 2024-11-22 12:11:40 +00:00
m68k: implement the new page table range API
Add PFN_PTE_SHIFT, update_mmu_cache_range(), flush_icache_pages() and flush_dcache_folio(). Link: https://lkml.kernel.org/r/20230802151406.3735276-16-willy@infradead.org Signed-off-by: Matthew Wilcox (Oracle) <willy@infradead.org> Tested-by: Geert Uytterhoeven <geert@linux-m68k.org> Acked-by: Mike Rapoport (IBM) <rppt@kernel.org> Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
This commit is contained in:
parent
a6d01af08b
commit
5553b15a4b
@ -220,24 +220,29 @@ static inline void flush_cache_page(struct vm_area_struct *vma, unsigned long vm
|
||||
|
||||
/* Push the page at kernel virtual address and clear the icache */
|
||||
/* RZ: use cpush %bc instead of cpush %dc, cinv %ic */
|
||||
static inline void __flush_page_to_ram(void *vaddr)
|
||||
static inline void __flush_pages_to_ram(void *vaddr, unsigned int nr)
|
||||
{
|
||||
if (CPU_IS_COLDFIRE) {
|
||||
unsigned long addr, start, end;
|
||||
addr = ((unsigned long) vaddr) & ~(PAGE_SIZE - 1);
|
||||
start = addr & ICACHE_SET_MASK;
|
||||
end = (addr + PAGE_SIZE - 1) & ICACHE_SET_MASK;
|
||||
end = (addr + nr * PAGE_SIZE - 1) & ICACHE_SET_MASK;
|
||||
if (start > end) {
|
||||
flush_cf_bcache(0, end);
|
||||
end = ICACHE_MAX_ADDR;
|
||||
}
|
||||
flush_cf_bcache(start, end);
|
||||
} else if (CPU_IS_040_OR_060) {
|
||||
__asm__ __volatile__("nop\n\t"
|
||||
".chip 68040\n\t"
|
||||
"cpushp %%bc,(%0)\n\t"
|
||||
".chip 68k"
|
||||
: : "a" (__pa(vaddr)));
|
||||
unsigned long paddr = __pa(vaddr);
|
||||
|
||||
do {
|
||||
__asm__ __volatile__("nop\n\t"
|
||||
".chip 68040\n\t"
|
||||
"cpushp %%bc,(%0)\n\t"
|
||||
".chip 68k"
|
||||
: : "a" (paddr));
|
||||
paddr += PAGE_SIZE;
|
||||
} while (--nr);
|
||||
} else {
|
||||
unsigned long _tmp;
|
||||
__asm__ __volatile__("movec %%cacr,%0\n\t"
|
||||
@ -249,10 +254,14 @@ static inline void __flush_page_to_ram(void *vaddr)
|
||||
}
|
||||
|
||||
#define ARCH_IMPLEMENTS_FLUSH_DCACHE_PAGE 1
|
||||
#define flush_dcache_page(page) __flush_page_to_ram(page_address(page))
|
||||
#define flush_dcache_page(page) __flush_pages_to_ram(page_address(page), 1)
|
||||
#define flush_dcache_folio(folio) \
|
||||
__flush_pages_to_ram(folio_address(folio), folio_nr_pages(folio))
|
||||
#define flush_dcache_mmap_lock(mapping) do { } while (0)
|
||||
#define flush_dcache_mmap_unlock(mapping) do { } while (0)
|
||||
#define flush_icache_page(vma, page) __flush_page_to_ram(page_address(page))
|
||||
#define flush_icache_pages(vma, page, nr) \
|
||||
__flush_pages_to_ram(page_address(page), nr)
|
||||
#define flush_icache_page(vma, page) flush_icache_pages(vma, page, 1)
|
||||
|
||||
extern void flush_icache_user_page(struct vm_area_struct *vma, struct page *page,
|
||||
unsigned long addr, int len);
|
||||
|
@ -291,6 +291,7 @@ static inline pte_t pte_swp_clear_exclusive(pte_t pte)
|
||||
return pte;
|
||||
}
|
||||
|
||||
#define PFN_PTE_SHIFT PAGE_SHIFT
|
||||
#define pmd_pfn(pmd) (pmd_val(pmd) >> PAGE_SHIFT)
|
||||
#define pmd_page(pmd) (pfn_to_page(pmd_val(pmd) >> PAGE_SHIFT))
|
||||
|
||||
|
@ -112,6 +112,7 @@ static inline void pud_set(pud_t *pudp, pmd_t *pmdp)
|
||||
#define pte_present(pte) (pte_val(pte) & (_PAGE_PRESENT | _PAGE_PROTNONE))
|
||||
#define pte_clear(mm,addr,ptep) ({ pte_val(*(ptep)) = 0; })
|
||||
|
||||
#define PFN_PTE_SHIFT PAGE_SHIFT
|
||||
#define pte_page(pte) virt_to_page(__va(pte_val(pte)))
|
||||
#define pte_pfn(pte) (pte_val(pte) >> PAGE_SHIFT)
|
||||
#define pfn_pte(pfn, prot) __pte(((pfn) << PAGE_SHIFT) | pgprot_val(prot))
|
||||
|
@ -31,8 +31,6 @@
|
||||
do{ \
|
||||
*(pteptr) = (pteval); \
|
||||
} while(0)
|
||||
#define set_pte_at(mm,addr,ptep,pteval) set_pte(ptep,pteval)
|
||||
|
||||
|
||||
/* PMD_SHIFT determines the size of the area a second-level page table can map */
|
||||
#if CONFIG_PGTABLE_LEVELS == 3
|
||||
@ -138,11 +136,15 @@ extern void kernel_set_cachemode(void *addr, unsigned long size, int cmode);
|
||||
* tables contain all the necessary information. The Sun3 does, but
|
||||
* they are updated on demand.
|
||||
*/
|
||||
static inline void update_mmu_cache(struct vm_area_struct *vma,
|
||||
unsigned long address, pte_t *ptep)
|
||||
static inline void update_mmu_cache_range(struct vm_fault *vmf,
|
||||
struct vm_area_struct *vma, unsigned long address,
|
||||
pte_t *ptep, unsigned int nr)
|
||||
{
|
||||
}
|
||||
|
||||
#define update_mmu_cache(vma, addr, ptep) \
|
||||
update_mmu_cache_range(NULL, vma, addr, ptep, 1)
|
||||
|
||||
#endif /* !__ASSEMBLY__ */
|
||||
|
||||
/* MMU-specific headers */
|
||||
|
@ -105,6 +105,7 @@ static inline void pte_clear (struct mm_struct *mm, unsigned long addr, pte_t *p
|
||||
pte_val (*ptep) = 0;
|
||||
}
|
||||
|
||||
#define PFN_PTE_SHIFT 0
|
||||
#define pte_pfn(pte) (pte_val(pte) & SUN3_PAGE_PGNUM_MASK)
|
||||
#define pfn_pte(pfn, pgprot) \
|
||||
({ pte_t __pte; pte_val(__pte) = pfn | pgprot_val(pgprot); __pte; })
|
||||
|
@ -81,7 +81,7 @@ static inline void cache_page(void *vaddr)
|
||||
|
||||
void mmu_page_ctor(void *page)
|
||||
{
|
||||
__flush_page_to_ram(page);
|
||||
__flush_pages_to_ram(page, 1);
|
||||
flush_tlb_kernel_page(page);
|
||||
nocache_page(page);
|
||||
}
|
||||
|
Loading…
Reference in New Issue
Block a user