mirror of
git://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
synced 2025-09-18 22:14:16 +00:00
mm: convert swap_readpage() to swap_read_folio()
All callers have a folio, so pass it in, saving two calls to compound_head(). Link: https://lkml.kernel.org/r/20231213215842.671461-11-willy@infradead.org Signed-off-by: Matthew Wilcox (Oracle) <willy@infradead.org> Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
This commit is contained in:
parent
3a61e6f668
commit
c9bdf768dd
5 changed files with 21 additions and 20 deletions
|
|
@ -3882,9 +3882,9 @@ vm_fault_t do_swap_page(struct vm_fault *vmf)
|
||||||
|
|
||||||
folio_add_lru(folio);
|
folio_add_lru(folio);
|
||||||
|
|
||||||
/* To provide entry to swap_readpage() */
|
/* To provide entry to swap_read_folio() */
|
||||||
folio->swap = entry;
|
folio->swap = entry;
|
||||||
swap_readpage(page, true, NULL);
|
swap_read_folio(folio, true, NULL);
|
||||||
folio->private = NULL;
|
folio->private = NULL;
|
||||||
}
|
}
|
||||||
} else {
|
} else {
|
||||||
|
|
|
||||||
18
mm/page_io.c
18
mm/page_io.c
|
|
@ -420,7 +420,7 @@ static void sio_read_complete(struct kiocb *iocb, long ret)
|
||||||
mempool_free(sio, sio_pool);
|
mempool_free(sio, sio_pool);
|
||||||
}
|
}
|
||||||
|
|
||||||
static void swap_readpage_fs(struct folio *folio, struct swap_iocb **plug)
|
static void swap_read_folio_fs(struct folio *folio, struct swap_iocb **plug)
|
||||||
{
|
{
|
||||||
struct swap_info_struct *sis = swp_swap_info(folio->swap);
|
struct swap_info_struct *sis = swp_swap_info(folio->swap);
|
||||||
struct swap_iocb *sio = NULL;
|
struct swap_iocb *sio = NULL;
|
||||||
|
|
@ -454,7 +454,7 @@ static void swap_readpage_fs(struct folio *folio, struct swap_iocb **plug)
|
||||||
*plug = sio;
|
*plug = sio;
|
||||||
}
|
}
|
||||||
|
|
||||||
static void swap_readpage_bdev_sync(struct folio *folio,
|
static void swap_read_folio_bdev_sync(struct folio *folio,
|
||||||
struct swap_info_struct *sis)
|
struct swap_info_struct *sis)
|
||||||
{
|
{
|
||||||
struct bio_vec bv;
|
struct bio_vec bv;
|
||||||
|
|
@ -474,7 +474,7 @@ static void swap_readpage_bdev_sync(struct folio *folio,
|
||||||
put_task_struct(current);
|
put_task_struct(current);
|
||||||
}
|
}
|
||||||
|
|
||||||
static void swap_readpage_bdev_async(struct folio *folio,
|
static void swap_read_folio_bdev_async(struct folio *folio,
|
||||||
struct swap_info_struct *sis)
|
struct swap_info_struct *sis)
|
||||||
{
|
{
|
||||||
struct bio *bio;
|
struct bio *bio;
|
||||||
|
|
@ -487,10 +487,10 @@ static void swap_readpage_bdev_async(struct folio *folio,
|
||||||
submit_bio(bio);
|
submit_bio(bio);
|
||||||
}
|
}
|
||||||
|
|
||||||
void swap_readpage(struct page *page, bool synchronous, struct swap_iocb **plug)
|
void swap_read_folio(struct folio *folio, bool synchronous,
|
||||||
|
struct swap_iocb **plug)
|
||||||
{
|
{
|
||||||
struct folio *folio = page_folio(page);
|
struct swap_info_struct *sis = swp_swap_info(folio->swap);
|
||||||
struct swap_info_struct *sis = page_swap_info(page);
|
|
||||||
bool workingset = folio_test_workingset(folio);
|
bool workingset = folio_test_workingset(folio);
|
||||||
unsigned long pflags;
|
unsigned long pflags;
|
||||||
bool in_thrashing;
|
bool in_thrashing;
|
||||||
|
|
@ -514,11 +514,11 @@ void swap_readpage(struct page *page, bool synchronous, struct swap_iocb **plug)
|
||||||
folio_mark_uptodate(folio);
|
folio_mark_uptodate(folio);
|
||||||
folio_unlock(folio);
|
folio_unlock(folio);
|
||||||
} else if (data_race(sis->flags & SWP_FS_OPS)) {
|
} else if (data_race(sis->flags & SWP_FS_OPS)) {
|
||||||
swap_readpage_fs(folio, plug);
|
swap_read_folio_fs(folio, plug);
|
||||||
} else if (synchronous || (sis->flags & SWP_SYNCHRONOUS_IO)) {
|
} else if (synchronous || (sis->flags & SWP_SYNCHRONOUS_IO)) {
|
||||||
swap_readpage_bdev_sync(folio, sis);
|
swap_read_folio_bdev_sync(folio, sis);
|
||||||
} else {
|
} else {
|
||||||
swap_readpage_bdev_async(folio, sis);
|
swap_read_folio_bdev_async(folio, sis);
|
||||||
}
|
}
|
||||||
|
|
||||||
if (workingset) {
|
if (workingset) {
|
||||||
|
|
|
||||||
|
|
@ -10,7 +10,8 @@ struct mempolicy;
|
||||||
/* linux/mm/page_io.c */
|
/* linux/mm/page_io.c */
|
||||||
int sio_pool_init(void);
|
int sio_pool_init(void);
|
||||||
struct swap_iocb;
|
struct swap_iocb;
|
||||||
void swap_readpage(struct page *page, bool do_poll, struct swap_iocb **plug);
|
void swap_read_folio(struct folio *folio, bool do_poll,
|
||||||
|
struct swap_iocb **plug);
|
||||||
void __swap_read_unplug(struct swap_iocb *plug);
|
void __swap_read_unplug(struct swap_iocb *plug);
|
||||||
static inline void swap_read_unplug(struct swap_iocb *plug)
|
static inline void swap_read_unplug(struct swap_iocb *plug)
|
||||||
{
|
{
|
||||||
|
|
@ -63,7 +64,7 @@ static inline unsigned int folio_swap_flags(struct folio *folio)
|
||||||
}
|
}
|
||||||
#else /* CONFIG_SWAP */
|
#else /* CONFIG_SWAP */
|
||||||
struct swap_iocb;
|
struct swap_iocb;
|
||||||
static inline void swap_readpage(struct page *page, bool do_poll,
|
static inline void swap_read_folio(struct folio *folio, bool do_poll,
|
||||||
struct swap_iocb **plug)
|
struct swap_iocb **plug)
|
||||||
{
|
{
|
||||||
}
|
}
|
||||||
|
|
|
||||||
|
|
@ -530,7 +530,7 @@ fail_put_swap:
|
||||||
* the swap entry is no longer in use.
|
* the swap entry is no longer in use.
|
||||||
*
|
*
|
||||||
* get/put_swap_device() aren't needed to call this function, because
|
* get/put_swap_device() aren't needed to call this function, because
|
||||||
* __read_swap_cache_async() call them and swap_readpage() holds the
|
* __read_swap_cache_async() call them and swap_read_folio() holds the
|
||||||
* swap cache folio lock.
|
* swap cache folio lock.
|
||||||
*/
|
*/
|
||||||
struct page *read_swap_cache_async(swp_entry_t entry, gfp_t gfp_mask,
|
struct page *read_swap_cache_async(swp_entry_t entry, gfp_t gfp_mask,
|
||||||
|
|
@ -548,7 +548,7 @@ struct page *read_swap_cache_async(swp_entry_t entry, gfp_t gfp_mask,
|
||||||
mpol_cond_put(mpol);
|
mpol_cond_put(mpol);
|
||||||
|
|
||||||
if (page_allocated)
|
if (page_allocated)
|
||||||
swap_readpage(&folio->page, false, plug);
|
swap_read_folio(folio, false, plug);
|
||||||
return folio_file_page(folio, swp_offset(entry));
|
return folio_file_page(folio, swp_offset(entry));
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
@ -665,7 +665,7 @@ struct page *swap_cluster_readahead(swp_entry_t entry, gfp_t gfp_mask,
|
||||||
if (!folio)
|
if (!folio)
|
||||||
continue;
|
continue;
|
||||||
if (page_allocated) {
|
if (page_allocated) {
|
||||||
swap_readpage(&folio->page, false, &splug);
|
swap_read_folio(folio, false, &splug);
|
||||||
if (offset != entry_offset) {
|
if (offset != entry_offset) {
|
||||||
folio_set_readahead(folio);
|
folio_set_readahead(folio);
|
||||||
count_vm_event(SWAP_RA);
|
count_vm_event(SWAP_RA);
|
||||||
|
|
@ -681,7 +681,7 @@ skip:
|
||||||
folio = __read_swap_cache_async(entry, gfp_mask, mpol, ilx,
|
folio = __read_swap_cache_async(entry, gfp_mask, mpol, ilx,
|
||||||
&page_allocated, false);
|
&page_allocated, false);
|
||||||
if (unlikely(page_allocated))
|
if (unlikely(page_allocated))
|
||||||
swap_readpage(&folio->page, false, NULL);
|
swap_read_folio(folio, false, NULL);
|
||||||
zswap_folio_swapin(folio);
|
zswap_folio_swapin(folio);
|
||||||
return folio_file_page(folio, swp_offset(entry));
|
return folio_file_page(folio, swp_offset(entry));
|
||||||
}
|
}
|
||||||
|
|
@ -839,7 +839,7 @@ static struct page *swap_vma_readahead(swp_entry_t targ_entry, gfp_t gfp_mask,
|
||||||
if (!folio)
|
if (!folio)
|
||||||
continue;
|
continue;
|
||||||
if (page_allocated) {
|
if (page_allocated) {
|
||||||
swap_readpage(&folio->page, false, &splug);
|
swap_read_folio(folio, false, &splug);
|
||||||
if (i != ra_info.offset) {
|
if (i != ra_info.offset) {
|
||||||
folio_set_readahead(folio);
|
folio_set_readahead(folio);
|
||||||
count_vm_event(SWAP_RA);
|
count_vm_event(SWAP_RA);
|
||||||
|
|
@ -857,7 +857,7 @@ skip:
|
||||||
folio = __read_swap_cache_async(targ_entry, gfp_mask, mpol, targ_ilx,
|
folio = __read_swap_cache_async(targ_entry, gfp_mask, mpol, targ_ilx,
|
||||||
&page_allocated, false);
|
&page_allocated, false);
|
||||||
if (unlikely(page_allocated))
|
if (unlikely(page_allocated))
|
||||||
swap_readpage(&folio->page, false, NULL);
|
swap_read_folio(folio, false, NULL);
|
||||||
zswap_folio_swapin(folio);
|
zswap_folio_swapin(folio);
|
||||||
return folio_file_page(folio, swp_offset(entry));
|
return folio_file_page(folio, swp_offset(entry));
|
||||||
}
|
}
|
||||||
|
|
|
||||||
|
|
@ -2225,7 +2225,7 @@ EXPORT_SYMBOL_GPL(add_swap_extent);
|
||||||
/*
|
/*
|
||||||
* A `swap extent' is a simple thing which maps a contiguous range of pages
|
* A `swap extent' is a simple thing which maps a contiguous range of pages
|
||||||
* onto a contiguous range of disk blocks. A rbtree of swap extents is
|
* onto a contiguous range of disk blocks. A rbtree of swap extents is
|
||||||
* built at swapon time and is then used at swap_writepage/swap_readpage
|
* built at swapon time and is then used at swap_writepage/swap_read_folio
|
||||||
* time for locating where on disk a page belongs.
|
* time for locating where on disk a page belongs.
|
||||||
*
|
*
|
||||||
* If the swapfile is an S_ISBLK block device, a single extent is installed.
|
* If the swapfile is an S_ISBLK block device, a single extent is installed.
|
||||||
|
|
|
||||||
Loading…
Add table
Reference in a new issue