mm: convert swap_readpage() to swap_read_folio()
authorMatthew Wilcox (Oracle) <willy@infradead.org>
Wed, 13 Dec 2023 21:58:39 +0000 (21:58 +0000)
committerAndrew Morton <akpm@linux-foundation.org>
Fri, 29 Dec 2023 19:58:31 +0000 (11:58 -0800)
All callers have a folio, so pass it in, saving two calls to
compound_head().

Link: https://lkml.kernel.org/r/20231213215842.671461-11-willy@infradead.org
Signed-off-by: Matthew Wilcox (Oracle) <willy@infradead.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
mm/memory.c
mm/page_io.c
mm/swap.h
mm/swap_state.c
mm/swapfile.c

index 5c023bef2adbf92e50906b0a983ec1e9b6491b95..cfcaf4c0198c21c2d2e0d305b2c9df56e5a6f19c 100644 (file)
@@ -3882,9 +3882,9 @@ vm_fault_t do_swap_page(struct vm_fault *vmf)
 
                                folio_add_lru(folio);
 
-                               /* To provide entry to swap_readpage() */
+                               /* To provide entry to swap_read_folio() */
                                folio->swap = entry;
-                               swap_readpage(page, true, NULL);
+                               swap_read_folio(folio, true, NULL);
                                folio->private = NULL;
                        }
                } else {
index 6736c56526bfd60b0ee2b5db83f8a32c807010d4..09c6a4f316f3e205c310c83433d8f9ee2bc36fc7 100644 (file)
@@ -420,7 +420,7 @@ static void sio_read_complete(struct kiocb *iocb, long ret)
        mempool_free(sio, sio_pool);
 }
 
-static void swap_readpage_fs(struct folio *folio, struct swap_iocb **plug)
+static void swap_read_folio_fs(struct folio *folio, struct swap_iocb **plug)
 {
        struct swap_info_struct *sis = swp_swap_info(folio->swap);
        struct swap_iocb *sio = NULL;
@@ -454,7 +454,7 @@ static void swap_readpage_fs(struct folio *folio, struct swap_iocb **plug)
                *plug = sio;
 }
 
-static void swap_readpage_bdev_sync(struct folio *folio,
+static void swap_read_folio_bdev_sync(struct folio *folio,
                struct swap_info_struct *sis)
 {
        struct bio_vec bv;
@@ -474,7 +474,7 @@ static void swap_readpage_bdev_sync(struct folio *folio,
        put_task_struct(current);
 }
 
-static void swap_readpage_bdev_async(struct folio *folio,
+static void swap_read_folio_bdev_async(struct folio *folio,
                struct swap_info_struct *sis)
 {
        struct bio *bio;
@@ -487,10 +487,10 @@ static void swap_readpage_bdev_async(struct folio *folio,
        submit_bio(bio);
 }
 
-void swap_readpage(struct page *page, bool synchronous, struct swap_iocb **plug)
+void swap_read_folio(struct folio *folio, bool synchronous,
+               struct swap_iocb **plug)
 {
-       struct folio *folio = page_folio(page);
-       struct swap_info_struct *sis = page_swap_info(page);
+       struct swap_info_struct *sis = swp_swap_info(folio->swap);
        bool workingset = folio_test_workingset(folio);
        unsigned long pflags;
        bool in_thrashing;
@@ -514,11 +514,11 @@ void swap_readpage(struct page *page, bool synchronous, struct swap_iocb **plug)
                folio_mark_uptodate(folio);
                folio_unlock(folio);
        } else if (data_race(sis->flags & SWP_FS_OPS)) {
-               swap_readpage_fs(folio, plug);
+               swap_read_folio_fs(folio, plug);
        } else if (synchronous || (sis->flags & SWP_SYNCHRONOUS_IO)) {
-               swap_readpage_bdev_sync(folio, sis);
+               swap_read_folio_bdev_sync(folio, sis);
        } else {
-               swap_readpage_bdev_async(folio, sis);
+               swap_read_folio_bdev_async(folio, sis);
        }
 
        if (workingset) {
index b81587740cf137d2780aae18974d9357c76d2b48..859ae8f0fd2d4a296d1ee62107f17e5c3ab0a627 100644 (file)
--- a/mm/swap.h
+++ b/mm/swap.h
@@ -10,7 +10,8 @@ struct mempolicy;
 /* linux/mm/page_io.c */
 int sio_pool_init(void);
 struct swap_iocb;
-void swap_readpage(struct page *page, bool do_poll, struct swap_iocb **plug);
+void swap_read_folio(struct folio *folio, bool do_poll,
+               struct swap_iocb **plug);
 void __swap_read_unplug(struct swap_iocb *plug);
 static inline void swap_read_unplug(struct swap_iocb *plug)
 {
@@ -63,7 +64,7 @@ static inline unsigned int folio_swap_flags(struct folio *folio)
 }
 #else /* CONFIG_SWAP */
 struct swap_iocb;
-static inline void swap_readpage(struct page *page, bool do_poll,
+static inline void swap_read_folio(struct folio *folio, bool do_poll,
                struct swap_iocb **plug)
 {
 }
index 874b40a1f502ea181fee978308a7fe960147a118..d2fe70e307d94b312f616bbb5dd041456fa7ce51 100644 (file)
@@ -530,7 +530,7 @@ fail_put_swap:
  * the swap entry is no longer in use.
  *
  * get/put_swap_device() aren't needed to call this function, because
- * __read_swap_cache_async() call them and swap_readpage() holds the
+ * __read_swap_cache_async() call them and swap_read_folio() holds the
  * swap cache folio lock.
  */
 struct page *read_swap_cache_async(swp_entry_t entry, gfp_t gfp_mask,
@@ -548,7 +548,7 @@ struct page *read_swap_cache_async(swp_entry_t entry, gfp_t gfp_mask,
        mpol_cond_put(mpol);
 
        if (page_allocated)
-               swap_readpage(&folio->page, false, plug);
+               swap_read_folio(folio, false, plug);
        return folio_file_page(folio, swp_offset(entry));
 }
 
@@ -665,7 +665,7 @@ struct page *swap_cluster_readahead(swp_entry_t entry, gfp_t gfp_mask,
                if (!folio)
                        continue;
                if (page_allocated) {
-                       swap_readpage(&folio->page, false, &splug);
+                       swap_read_folio(folio, false, &splug);
                        if (offset != entry_offset) {
                                folio_set_readahead(folio);
                                count_vm_event(SWAP_RA);
@@ -681,7 +681,7 @@ skip:
        folio = __read_swap_cache_async(entry, gfp_mask, mpol, ilx,
                                        &page_allocated, false);
        if (unlikely(page_allocated))
-               swap_readpage(&folio->page, false, NULL);
+               swap_read_folio(folio, false, NULL);
        zswap_folio_swapin(folio);
        return folio_file_page(folio, swp_offset(entry));
 }
@@ -839,7 +839,7 @@ static struct page *swap_vma_readahead(swp_entry_t targ_entry, gfp_t gfp_mask,
                if (!folio)
                        continue;
                if (page_allocated) {
-                       swap_readpage(&folio->page, false, &splug);
+                       swap_read_folio(folio, false, &splug);
                        if (i != ra_info.offset) {
                                folio_set_readahead(folio);
                                count_vm_event(SWAP_RA);
@@ -857,7 +857,7 @@ skip:
        folio = __read_swap_cache_async(targ_entry, gfp_mask, mpol, targ_ilx,
                                        &page_allocated, false);
        if (unlikely(page_allocated))
-               swap_readpage(&folio->page, false, NULL);
+               swap_read_folio(folio, false, NULL);
        zswap_folio_swapin(folio);
        return folio_file_page(folio, swp_offset(entry));
 }
index b22c47b11d6523c87d02e398caca5d8e577559a6..f3e23a3d26aec782f5ff4bacdb8d8276c5a74b66 100644 (file)
@@ -2225,7 +2225,7 @@ EXPORT_SYMBOL_GPL(add_swap_extent);
 /*
  * A `swap extent' is a simple thing which maps a contiguous range of pages
  * onto a contiguous range of disk blocks.  A rbtree of swap extents is
- * built at swapon time and is then used at swap_writepage/swap_readpage
+ * built at swapon time and is then used at swap_writepage/swap_read_folio
  * time for locating where on disk a page belongs.
  *
  * If the swapfile is an S_ISBLK block device, a single extent is installed.