btrfs: use a folio array throughout the defrag process
authorMatthew Wilcox (Oracle) <willy@infradead.org>
Thu, 14 Dec 2023 16:13:31 +0000 (16:13 +0000)
committerDavid Sterba <dsterba@suse.com>
Mon, 4 Mar 2024 15:24:45 +0000 (16:24 +0100)
Remove more hidden calls to compound_head() by using an array of folios
instead of pages.  Also neaten the error path in defrag_one_range() by
adjusting the length of the array instead of checking for NULL.

Reviewed-by: Johannes Thumshirn <johannes.thumshirn@wdc.com>
Signed-off-by: Matthew Wilcox (Oracle) <willy@infradead.org>
Reviewed-by: David Sterba <dsterba@suse.com>
Signed-off-by: David Sterba <dsterba@suse.com>
fs/btrfs/defrag.c

index fc5950776156eb8c06fd33fcf4196faeca9e8718..d58cfd999505cecab3575e9c27862a16268d0f38 100644 (file)
@@ -861,7 +861,7 @@ out:
  * NOTE: Caller should also wait for page writeback after the cluster is
  * prepared, here we don't do writeback wait for each page.
  */
-static struct page *defrag_prepare_one_page(struct btrfs_inode *inode, pgoff_t index)
+static struct folio *defrag_prepare_one_folio(struct btrfs_inode *inode, pgoff_t index)
 {
        struct address_space *mapping = inode->vfs_inode.i_mapping;
        gfp_t mask = btrfs_alloc_write_mask(mapping);
@@ -875,7 +875,7 @@ again:
        folio = __filemap_get_folio(mapping, index,
                                    FGP_LOCK | FGP_ACCESSED | FGP_CREAT, mask);
        if (IS_ERR(folio))
-               return &folio->page;
+               return folio;
 
        /*
         * Since we can defragment files opened read-only, we can encounter
@@ -942,7 +942,7 @@ again:
                        return ERR_PTR(-EIO);
                }
        }
-       return &folio->page;
+       return folio;
 }
 
 struct defrag_target_range {
@@ -1163,7 +1163,7 @@ static_assert(PAGE_ALIGNED(CLUSTER_SIZE));
  */
 static int defrag_one_locked_target(struct btrfs_inode *inode,
                                    struct defrag_target_range *target,
-                                   struct page **pages, int nr_pages,
+                                   struct folio **folios, int nr_pages,
                                    struct extent_state **cached_state)
 {
        struct btrfs_fs_info *fs_info = inode->root->fs_info;
@@ -1172,7 +1172,7 @@ static int defrag_one_locked_target(struct btrfs_inode *inode,
        const u64 len = target->len;
        unsigned long last_index = (start + len - 1) >> PAGE_SHIFT;
        unsigned long start_index = start >> PAGE_SHIFT;
-       unsigned long first_index = page_index(pages[0]);
+       unsigned long first_index = folios[0]->index;
        int ret = 0;
        int i;
 
@@ -1189,8 +1189,8 @@ static int defrag_one_locked_target(struct btrfs_inode *inode,
 
        /* Update the page status */
        for (i = start_index - first_index; i <= last_index - first_index; i++) {
-               ClearPageChecked(pages[i]);
-               btrfs_folio_clamp_set_dirty(fs_info, page_folio(pages[i]), start, len);
+               folio_clear_checked(folios[i]);
+               btrfs_folio_clamp_set_dirty(fs_info, folios[i], start, len);
        }
        btrfs_delalloc_release_extents(inode, len);
        extent_changeset_free(data_reserved);
@@ -1206,7 +1206,7 @@ static int defrag_one_range(struct btrfs_inode *inode, u64 start, u32 len,
        struct defrag_target_range *entry;
        struct defrag_target_range *tmp;
        LIST_HEAD(target_list);
-       struct page **pages;
+       struct folio **folios;
        const u32 sectorsize = inode->root->fs_info->sectorsize;
        u64 last_index = (start + len - 1) >> PAGE_SHIFT;
        u64 start_index = start >> PAGE_SHIFT;
@@ -1217,21 +1217,21 @@ static int defrag_one_range(struct btrfs_inode *inode, u64 start, u32 len,
        ASSERT(nr_pages <= CLUSTER_SIZE / PAGE_SIZE);
        ASSERT(IS_ALIGNED(start, sectorsize) && IS_ALIGNED(len, sectorsize));
 
-       pages = kcalloc(nr_pages, sizeof(struct page *), GFP_NOFS);
-       if (!pages)
+       folios = kcalloc(nr_pages, sizeof(struct folio *), GFP_NOFS);
+       if (!folios)
                return -ENOMEM;
 
        /* Prepare all pages */
        for (i = 0; i < nr_pages; i++) {
-               pages[i] = defrag_prepare_one_page(inode, start_index + i);
-               if (IS_ERR(pages[i])) {
-                       ret = PTR_ERR(pages[i]);
-                       pages[i] = NULL;
-                       goto free_pages;
+               folios[i] = defrag_prepare_one_folio(inode, start_index + i);
+               if (IS_ERR(folios[i])) {
+                       ret = PTR_ERR(folios[i]);
+                       nr_pages = i;
+                       goto free_folios;
                }
        }
        for (i = 0; i < nr_pages; i++)
-               wait_on_page_writeback(pages[i]);
+               folio_wait_writeback(folios[i]);
 
        /* Lock the pages range */
        lock_extent(&inode->io_tree, start_index << PAGE_SHIFT,
@@ -1251,7 +1251,7 @@ static int defrag_one_range(struct btrfs_inode *inode, u64 start, u32 len,
                goto unlock_extent;
 
        list_for_each_entry(entry, &target_list, list) {
-               ret = defrag_one_locked_target(inode, entry, pages, nr_pages,
+               ret = defrag_one_locked_target(inode, entry, folios, nr_pages,
                                               &cached_state);
                if (ret < 0)
                        break;
@@ -1265,14 +1265,12 @@ unlock_extent:
        unlock_extent(&inode->io_tree, start_index << PAGE_SHIFT,
                      (last_index << PAGE_SHIFT) + PAGE_SIZE - 1,
                      &cached_state);
-free_pages:
+free_folios:
        for (i = 0; i < nr_pages; i++) {
-               if (pages[i]) {
-                       unlock_page(pages[i]);
-                       put_page(pages[i]);
-               }
+               folio_unlock(folios[i]);
+               folio_put(folios[i]);
        }
-       kfree(pages);
+       kfree(folios);
        return ret;
 }