mirror of
https://github.com/torvalds/linux.git
synced 2024-11-26 14:12:06 +00:00
mm: munlock: fix a bug where THP tail page is encountered
Since commitff6a6da60b
("mm: accelerate munlock() treatment of THP pages") munlock skips tail pages of a munlocked THP page. However, when the head page already has PageMlocked unset, it will not skip the tail pages. Commit7225522bb4
("mm: munlock: batch non-THP page isolation and munlock+putback using pagevec") has added a PageTransHuge() check which contains VM_BUG_ON(PageTail(page)). Sasha Levin found this triggered using trinity, on the first tail page of a THP page without PageMlocked flag. This patch fixes the issue by skipping tail pages also in the case when PageMlocked flag is unset. There is still a possibility of race with THP page split between clearing PageMlocked and determining how many pages to skip. The race might result in former tail pages not being skipped, which is however no longer a bug, as during the skip the PageTail flags are cleared. However this race also affects correctness of NR_MLOCK accounting, which is to be fixed in a separate patch. Signed-off-by: Vlastimil Babka <vbabka@suse.cz> Reported-by: Sasha Levin <sasha.levin@oracle.com> Cc: Michel Lespinasse <walken@google.com> Cc: Andrea Arcangeli <aarcange@redhat.com> Cc: Rik van Riel <riel@redhat.com> Cc: Mel Gorman <mgorman@suse.de> Cc: Hugh Dickins <hughd@google.com> Cc: Bob Liu <bob.liu@oracle.com> Cc: <stable@vger.kernel.org> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
This commit is contained in:
parent
9a0bb2966e
commit
c424be1cbb
29
mm/mlock.c
29
mm/mlock.c
@ -133,7 +133,10 @@ static void __munlock_isolation_failed(struct page *page)
|
||||
|
||||
/**
|
||||
* munlock_vma_page - munlock a vma page
|
||||
* @page - page to be unlocked
|
||||
* @page - page to be unlocked, either a normal page or THP page head
|
||||
*
|
||||
* returns the size of the page as a page mask (0 for normal page,
|
||||
* HPAGE_PMD_NR - 1 for THP head page)
|
||||
*
|
||||
* called from munlock()/munmap() path with page supposedly on the LRU.
|
||||
* When we munlock a page, because the vma where we found the page is being
|
||||
@ -148,21 +151,30 @@ static void __munlock_isolation_failed(struct page *page)
|
||||
*/
|
||||
unsigned int munlock_vma_page(struct page *page)
|
||||
{
|
||||
unsigned int page_mask = 0;
|
||||
unsigned int nr_pages;
|
||||
|
||||
BUG_ON(!PageLocked(page));
|
||||
|
||||
if (TestClearPageMlocked(page)) {
|
||||
unsigned int nr_pages = hpage_nr_pages(page);
|
||||
nr_pages = hpage_nr_pages(page);
|
||||
mod_zone_page_state(page_zone(page), NR_MLOCK, -nr_pages);
|
||||
page_mask = nr_pages - 1;
|
||||
if (!isolate_lru_page(page))
|
||||
__munlock_isolated_page(page);
|
||||
else
|
||||
__munlock_isolation_failed(page);
|
||||
} else {
|
||||
nr_pages = hpage_nr_pages(page);
|
||||
}
|
||||
|
||||
return page_mask;
|
||||
/*
|
||||
* Regardless of the original PageMlocked flag, we determine nr_pages
|
||||
* after touching the flag. This leaves a possible race with a THP page
|
||||
* split, such that a whole THP page was munlocked, but nr_pages == 1.
|
||||
* Returning a smaller mask due to that is OK, the worst that can
|
||||
* happen is subsequent useless scanning of the former tail pages.
|
||||
* The NR_MLOCK accounting can however become broken.
|
||||
*/
|
||||
return nr_pages - 1;
|
||||
}
|
||||
|
||||
/**
|
||||
@ -440,7 +452,8 @@ void munlock_vma_pages_range(struct vm_area_struct *vma,
|
||||
|
||||
while (start < end) {
|
||||
struct page *page = NULL;
|
||||
unsigned int page_mask, page_increm;
|
||||
unsigned int page_mask;
|
||||
unsigned long page_increm;
|
||||
struct pagevec pvec;
|
||||
struct zone *zone;
|
||||
int zoneid;
|
||||
@ -490,7 +503,9 @@ void munlock_vma_pages_range(struct vm_area_struct *vma,
|
||||
goto next;
|
||||
}
|
||||
}
|
||||
page_increm = 1 + (~(start >> PAGE_SHIFT) & page_mask);
|
||||
/* It's a bug to munlock in the middle of a THP page */
|
||||
VM_BUG_ON((start >> PAGE_SHIFT) & page_mask);
|
||||
page_increm = 1 + page_mask;
|
||||
start += page_increm * PAGE_SIZE;
|
||||
next:
|
||||
cond_resched();
|
||||
|
Loading…
Reference in New Issue
Block a user