mirror of
https://github.com/torvalds/linux.git
synced 2024-12-04 01:51:34 +00:00
mm: remove unused hugepage for vma_alloc_folio()
The hugepage parameter was deprecated since commit ddc1a5cbc0
("mempolicy: alloc_pages_mpol() for NUMA policy without vma"), for
PMD-sized THP, it still tries only preferred node if possible in
vma_alloc_folio() by checking the order of the folio allocation.
Link: https://lkml.kernel.org/r/20241010061556.1846751-1-wangkefeng.wang@huawei.com
Signed-off-by: Kefeng Wang <wangkefeng.wang@huawei.com>
Acked-by: David Hildenbrand <david@redhat.com>
Reviewed-by: Zi Yan <ziy@nvidia.com>
Reviewed-by: Barry Song <baohua@kernel.org>
Cc: Hugh Dickins <hughd@google.com>
Cc: Matthew Wilcox <willy@infradead.org>
Cc: Ryan Roberts <ryan.roberts@arm.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
This commit is contained in:
parent
f8780515fe
commit
6359c39c9d
@ -18,7 +18,7 @@ extern void clear_page(void *page);
|
||||
#define clear_user_page(page, vaddr, pg) clear_page(page)
|
||||
|
||||
#define vma_alloc_zeroed_movable_folio(vma, vaddr) \
|
||||
vma_alloc_folio(GFP_HIGHUSER_MOVABLE | __GFP_ZERO, 0, vma, vaddr, false)
|
||||
vma_alloc_folio(GFP_HIGHUSER_MOVABLE | __GFP_ZERO, 0, vma, vaddr)
|
||||
|
||||
extern void copy_page(void * _to, void * _from);
|
||||
#define copy_user_page(to, from, vaddr, pg) copy_page(to, from)
|
||||
|
@ -983,7 +983,7 @@ struct folio *vma_alloc_zeroed_movable_folio(struct vm_area_struct *vma,
|
||||
if (vma->vm_flags & VM_MTE)
|
||||
flags |= __GFP_ZEROTAGS;
|
||||
|
||||
return vma_alloc_folio(flags, 0, vma, vaddr, false);
|
||||
return vma_alloc_folio(flags, 0, vma, vaddr);
|
||||
}
|
||||
|
||||
void tag_clear_highpage(struct page *page)
|
||||
|
@ -14,7 +14,7 @@ extern unsigned long memory_end;
|
||||
#define copy_user_page(to, from, vaddr, pg) copy_page(to, from)
|
||||
|
||||
#define vma_alloc_zeroed_movable_folio(vma, vaddr) \
|
||||
vma_alloc_folio(GFP_HIGHUSER_MOVABLE | __GFP_ZERO, 0, vma, vaddr, false)
|
||||
vma_alloc_folio(GFP_HIGHUSER_MOVABLE | __GFP_ZERO, 0, vma, vaddr)
|
||||
|
||||
#define __pa(vaddr) ((unsigned long)(vaddr))
|
||||
#define __va(paddr) ((void *)((unsigned long)(paddr)))
|
||||
|
@ -74,7 +74,7 @@ static inline void copy_page(void *to, void *from)
|
||||
#define copy_user_page(to, from, vaddr, pg) copy_page(to, from)
|
||||
|
||||
#define vma_alloc_zeroed_movable_folio(vma, vaddr) \
|
||||
vma_alloc_folio(GFP_HIGHUSER_MOVABLE | __GFP_ZERO, 0, vma, vaddr, false)
|
||||
vma_alloc_folio(GFP_HIGHUSER_MOVABLE | __GFP_ZERO, 0, vma, vaddr)
|
||||
|
||||
/*
|
||||
* These are used to make use of C type-checking..
|
||||
|
@ -35,7 +35,7 @@ static inline void copy_user_page(void *to, void *from, unsigned long vaddr,
|
||||
}
|
||||
|
||||
#define vma_alloc_zeroed_movable_folio(vma, vaddr) \
|
||||
vma_alloc_folio(GFP_HIGHUSER_MOVABLE | __GFP_ZERO, 0, vma, vaddr, false)
|
||||
vma_alloc_folio(GFP_HIGHUSER_MOVABLE | __GFP_ZERO, 0, vma, vaddr)
|
||||
|
||||
#ifndef __pa
|
||||
#define __pa(x) __phys_addr((unsigned long)(x))
|
||||
|
@ -306,7 +306,7 @@ struct folio *folio_alloc_noprof(gfp_t gfp, unsigned int order);
|
||||
struct folio *folio_alloc_mpol_noprof(gfp_t gfp, unsigned int order,
|
||||
struct mempolicy *mpol, pgoff_t ilx, int nid);
|
||||
struct folio *vma_alloc_folio_noprof(gfp_t gfp, int order, struct vm_area_struct *vma,
|
||||
unsigned long addr, bool hugepage);
|
||||
unsigned long addr);
|
||||
#else
|
||||
static inline struct page *alloc_pages_noprof(gfp_t gfp_mask, unsigned int order)
|
||||
{
|
||||
@ -326,7 +326,7 @@ static inline struct folio *folio_alloc_mpol_noprof(gfp_t gfp, unsigned int orde
|
||||
{
|
||||
return folio_alloc_noprof(gfp, order);
|
||||
}
|
||||
#define vma_alloc_folio_noprof(gfp, order, vma, addr, hugepage) \
|
||||
#define vma_alloc_folio_noprof(gfp, order, vma, addr) \
|
||||
folio_alloc_noprof(gfp, order)
|
||||
#endif
|
||||
|
||||
@ -341,7 +341,7 @@ static inline struct folio *folio_alloc_mpol_noprof(gfp_t gfp, unsigned int orde
|
||||
static inline struct page *alloc_page_vma_noprof(gfp_t gfp,
|
||||
struct vm_area_struct *vma, unsigned long addr)
|
||||
{
|
||||
struct folio *folio = vma_alloc_folio_noprof(gfp, 0, vma, addr, false);
|
||||
struct folio *folio = vma_alloc_folio_noprof(gfp, 0, vma, addr);
|
||||
|
||||
return &folio->page;
|
||||
}
|
||||
|
@ -226,7 +226,7 @@ struct folio *vma_alloc_zeroed_movable_folio(struct vm_area_struct *vma,
|
||||
{
|
||||
struct folio *folio;
|
||||
|
||||
folio = vma_alloc_folio(GFP_HIGHUSER_MOVABLE, 0, vma, vaddr, false);
|
||||
folio = vma_alloc_folio(GFP_HIGHUSER_MOVABLE, 0, vma, vaddr);
|
||||
if (folio)
|
||||
clear_user_highpage(&folio->page, vaddr);
|
||||
|
||||
|
@ -1143,7 +1143,7 @@ static struct folio *vma_alloc_anon_folio_pmd(struct vm_area_struct *vma,
|
||||
const int order = HPAGE_PMD_ORDER;
|
||||
struct folio *folio;
|
||||
|
||||
folio = vma_alloc_folio(gfp, order, vma, addr & HPAGE_PMD_MASK, true);
|
||||
folio = vma_alloc_folio(gfp, order, vma, addr & HPAGE_PMD_MASK);
|
||||
|
||||
if (unlikely(!folio)) {
|
||||
count_vm_event(THP_FAULT_FALLBACK);
|
||||
|
2
mm/ksm.c
2
mm/ksm.c
@ -2971,7 +2971,7 @@ struct folio *ksm_might_need_to_copy(struct folio *folio,
|
||||
if (!folio_test_uptodate(folio))
|
||||
return folio; /* let do_swap_page report the error */
|
||||
|
||||
new_folio = vma_alloc_folio(GFP_HIGHUSER_MOVABLE, 0, vma, addr, false);
|
||||
new_folio = vma_alloc_folio(GFP_HIGHUSER_MOVABLE, 0, vma, addr);
|
||||
if (new_folio &&
|
||||
mem_cgroup_charge(new_folio, vma->vm_mm, GFP_KERNEL)) {
|
||||
folio_put(new_folio);
|
||||
|
10
mm/memory.c
10
mm/memory.c
@ -1059,8 +1059,7 @@ static inline struct folio *folio_prealloc(struct mm_struct *src_mm,
|
||||
if (need_zero)
|
||||
new_folio = vma_alloc_zeroed_movable_folio(vma, addr);
|
||||
else
|
||||
new_folio = vma_alloc_folio(GFP_HIGHUSER_MOVABLE, 0, vma,
|
||||
addr, false);
|
||||
new_folio = vma_alloc_folio(GFP_HIGHUSER_MOVABLE, 0, vma, addr);
|
||||
|
||||
if (!new_folio)
|
||||
return NULL;
|
||||
@ -4017,8 +4016,7 @@ static struct folio *__alloc_swap_folio(struct vm_fault *vmf)
|
||||
struct folio *folio;
|
||||
swp_entry_t entry;
|
||||
|
||||
folio = vma_alloc_folio(GFP_HIGHUSER_MOVABLE, 0, vma,
|
||||
vmf->address, false);
|
||||
folio = vma_alloc_folio(GFP_HIGHUSER_MOVABLE, 0, vma, vmf->address);
|
||||
if (!folio)
|
||||
return NULL;
|
||||
|
||||
@ -4174,7 +4172,7 @@ static struct folio *alloc_swap_folio(struct vm_fault *vmf)
|
||||
gfp = vma_thp_gfp_mask(vma);
|
||||
while (orders) {
|
||||
addr = ALIGN_DOWN(vmf->address, PAGE_SIZE << order);
|
||||
folio = vma_alloc_folio(gfp, order, vma, addr, true);
|
||||
folio = vma_alloc_folio(gfp, order, vma, addr);
|
||||
if (folio) {
|
||||
if (!mem_cgroup_swapin_charge_folio(folio, vma->vm_mm,
|
||||
gfp, entry))
|
||||
@ -4713,7 +4711,7 @@ static struct folio *alloc_anon_folio(struct vm_fault *vmf)
|
||||
gfp = vma_thp_gfp_mask(vma);
|
||||
while (orders) {
|
||||
addr = ALIGN_DOWN(vmf->address, PAGE_SIZE << order);
|
||||
folio = vma_alloc_folio(gfp, order, vma, addr, true);
|
||||
folio = vma_alloc_folio(gfp, order, vma, addr);
|
||||
if (folio) {
|
||||
if (mem_cgroup_charge(folio, vma->vm_mm, gfp)) {
|
||||
count_mthp_stat(order, MTHP_STAT_ANON_FAULT_FALLBACK_CHARGE);
|
||||
|
@ -2290,7 +2290,6 @@ struct folio *folio_alloc_mpol_noprof(gfp_t gfp, unsigned int order,
|
||||
* @order: Order of the folio.
|
||||
* @vma: Pointer to VMA.
|
||||
* @addr: Virtual address of the allocation. Must be inside @vma.
|
||||
* @hugepage: Unused (was: For hugepages try only preferred node if possible).
|
||||
*
|
||||
* Allocate a folio for a specific address in @vma, using the appropriate
|
||||
* NUMA policy. The caller must hold the mmap_lock of the mm_struct of the
|
||||
@ -2301,7 +2300,7 @@ struct folio *folio_alloc_mpol_noprof(gfp_t gfp, unsigned int order,
|
||||
* Return: The folio on success or NULL if allocation fails.
|
||||
*/
|
||||
struct folio *vma_alloc_folio_noprof(gfp_t gfp, int order, struct vm_area_struct *vma,
|
||||
unsigned long addr, bool hugepage)
|
||||
unsigned long addr)
|
||||
{
|
||||
struct mempolicy *pol;
|
||||
pgoff_t ilx;
|
||||
|
@ -251,7 +251,7 @@ static int mfill_atomic_pte_copy(pmd_t *dst_pmd,
|
||||
if (!*foliop) {
|
||||
ret = -ENOMEM;
|
||||
folio = vma_alloc_folio(GFP_HIGHUSER_MOVABLE, 0, dst_vma,
|
||||
dst_addr, false);
|
||||
dst_addr);
|
||||
if (!folio)
|
||||
goto out;
|
||||
|
||||
|
Loading…
Reference in New Issue
Block a user