truncate: Add truncate_cleanup_folio()
Convert both callers of truncate_cleanup_page() to use truncate_cleanup_folio() instead. Signed-off-by: Matthew Wilcox (Oracle) <willy@infradead.org> Reviewed-by: Christoph Hellwig <hch@lst.de> Reviewed-by: William Kucharski <william.kucharski@oracle.com>
This commit is contained in:
parent
82c50f8b44
commit
efe99bba28
@ -177,21 +177,21 @@ void do_invalidatepage(struct page *page, unsigned int offset,
|
|||||||
* its lock, b) when a concurrent invalidate_mapping_pages got there first and
|
* its lock, b) when a concurrent invalidate_mapping_pages got there first and
|
||||||
* c) when tmpfs swizzles a page between a tmpfs inode and swapper_space.
|
* c) when tmpfs swizzles a page between a tmpfs inode and swapper_space.
|
||||||
*/
|
*/
|
||||||
static void truncate_cleanup_page(struct page *page)
|
static void truncate_cleanup_folio(struct folio *folio)
|
||||||
{
|
{
|
||||||
if (page_mapped(page))
|
if (folio_mapped(folio))
|
||||||
unmap_mapping_page(page);
|
unmap_mapping_page(&folio->page);
|
||||||
|
|
||||||
if (page_has_private(page))
|
if (folio_has_private(folio))
|
||||||
do_invalidatepage(page, 0, thp_size(page));
|
do_invalidatepage(&folio->page, 0, folio_size(folio));
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Some filesystems seem to re-dirty the page even after
|
* Some filesystems seem to re-dirty the page even after
|
||||||
* the VM has canceled the dirty bit (eg ext3 journaling).
|
* the VM has canceled the dirty bit (eg ext3 journaling).
|
||||||
* Hence dirty accounting check is placed after invalidation.
|
* Hence dirty accounting check is placed after invalidation.
|
||||||
*/
|
*/
|
||||||
cancel_dirty_page(page);
|
folio_cancel_dirty(folio);
|
||||||
ClearPageMappedToDisk(page);
|
folio_clear_mappedtodisk(folio);
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
@ -220,13 +220,14 @@ invalidate_complete_page(struct address_space *mapping, struct page *page)
|
|||||||
|
|
||||||
int truncate_inode_page(struct address_space *mapping, struct page *page)
|
int truncate_inode_page(struct address_space *mapping, struct page *page)
|
||||||
{
|
{
|
||||||
|
struct folio *folio = page_folio(page);
|
||||||
VM_BUG_ON_PAGE(PageTail(page), page);
|
VM_BUG_ON_PAGE(PageTail(page), page);
|
||||||
|
|
||||||
if (page->mapping != mapping)
|
if (page->mapping != mapping)
|
||||||
return -EIO;
|
return -EIO;
|
||||||
|
|
||||||
truncate_cleanup_page(page);
|
truncate_cleanup_folio(folio);
|
||||||
delete_from_page_cache(page);
|
filemap_remove_folio(folio);
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -332,7 +333,7 @@ void truncate_inode_pages_range(struct address_space *mapping,
|
|||||||
index = indices[pagevec_count(&pvec) - 1] + 1;
|
index = indices[pagevec_count(&pvec) - 1] + 1;
|
||||||
truncate_exceptional_pvec_entries(mapping, &pvec, indices);
|
truncate_exceptional_pvec_entries(mapping, &pvec, indices);
|
||||||
for (i = 0; i < pagevec_count(&pvec); i++)
|
for (i = 0; i < pagevec_count(&pvec); i++)
|
||||||
truncate_cleanup_page(pvec.pages[i]);
|
truncate_cleanup_folio(page_folio(pvec.pages[i]));
|
||||||
delete_from_page_cache_batch(mapping, &pvec);
|
delete_from_page_cache_batch(mapping, &pvec);
|
||||||
for (i = 0; i < pagevec_count(&pvec); i++)
|
for (i = 0; i < pagevec_count(&pvec); i++)
|
||||||
unlock_page(pvec.pages[i]);
|
unlock_page(pvec.pages[i]);
|
||||||
|
Loading…
Reference in New Issue
Block a user