Add a helper dax_ioamp_zero() to zero a range. This patch basically
merges __dax_zero_page_range() and iomap_dax_zero().

Suggested-by: Christoph Hellwig <[email protected]>
Signed-off-by: Vivek Goyal <[email protected]>
---
 fs/dax.c               | 12 ++++++------
 fs/iomap/buffered-io.c |  9 +--------
 include/linux/dax.h    | 11 +++++------
 3 files changed, 12 insertions(+), 20 deletions(-)

diff --git a/fs/dax.c b/fs/dax.c
index 1b9ba6b59cdb..63303e274221 100644
--- a/fs/dax.c
+++ b/fs/dax.c
@@ -1059,23 +1059,23 @@ int generic_dax_zero_page_range(struct dax_device 
*dax_dev, pgoff_t pgoff,
 }
 EXPORT_SYMBOL_GPL(generic_dax_zero_page_range);
 
-int __dax_zero_page_range(struct block_device *bdev,
-               struct dax_device *dax_dev, sector_t sector,
-               unsigned int offset, unsigned int size)
+int dax_iomap_zero(loff_t pos, unsigned offset, unsigned size,
+                             struct iomap *iomap)
 {
        pgoff_t pgoff;
        long rc, id;
+       sector_t sector = iomap_sector(iomap, pos & PAGE_MASK);
 
-       rc = bdev_dax_pgoff(bdev, sector, PAGE_SIZE, &pgoff);
+       rc = bdev_dax_pgoff(iomap->bdev, sector, PAGE_SIZE, &pgoff);
        if (rc)
                return rc;
 
        id = dax_read_lock();
-       rc = dax_zero_page_range(dax_dev, pgoff, offset, size);
+       rc = dax_zero_page_range(iomap->dax_dev, pgoff, offset, size);
        dax_read_unlock(id);
        return rc;
 }
-EXPORT_SYMBOL_GPL(__dax_zero_page_range);
+EXPORT_SYMBOL_GPL(dax_iomap_zero);
 
 static loff_t
 dax_iomap_actor(struct inode *inode, loff_t pos, loff_t length, void *data,
diff --git a/fs/iomap/buffered-io.c b/fs/iomap/buffered-io.c
index 828444e14d09..5a5d784a110e 100644
--- a/fs/iomap/buffered-io.c
+++ b/fs/iomap/buffered-io.c
@@ -974,13 +974,6 @@ static int iomap_zero(struct inode *inode, loff_t pos, 
unsigned offset,
        return iomap_write_end(inode, pos, bytes, bytes, page, iomap, srcmap);
 }
 
-static int iomap_dax_zero(loff_t pos, unsigned offset, unsigned bytes,
-               struct iomap *iomap)
-{
-       return __dax_zero_page_range(iomap->bdev, iomap->dax_dev,
-                       iomap_sector(iomap, pos & PAGE_MASK), offset, bytes);
-}
-
 static loff_t
 iomap_zero_range_actor(struct inode *inode, loff_t pos, loff_t count,
                void *data, struct iomap *iomap, struct iomap *srcmap)
@@ -1000,7 +993,7 @@ iomap_zero_range_actor(struct inode *inode, loff_t pos, 
loff_t count,
                bytes = min_t(loff_t, PAGE_SIZE - offset, count);
 
                if (IS_DAX(inode))
-                       status = iomap_dax_zero(pos, offset, bytes, iomap);
+                       status = dax_iomap_zero(pos, offset, bytes, iomap);
                else
                        status = iomap_zero(inode, pos, offset, bytes, iomap,
                                        srcmap);
diff --git a/include/linux/dax.h b/include/linux/dax.h
index 3356b874c55d..ffaaa12f8ca9 100644
--- a/include/linux/dax.h
+++ b/include/linux/dax.h
@@ -13,6 +13,7 @@
 typedef unsigned long dax_entry_t;
 
 struct iomap_ops;
+struct iomap;
 struct dax_device;
 struct dax_operations {
        /*
@@ -228,13 +229,11 @@ int dax_invalidate_mapping_entry_sync(struct 
address_space *mapping,
                                      pgoff_t index);
 
 #ifdef CONFIG_FS_DAX
-int __dax_zero_page_range(struct block_device *bdev,
-               struct dax_device *dax_dev, sector_t sector,
-               unsigned int offset, unsigned int length);
+int dax_iomap_zero(loff_t pos, unsigned offset, unsigned size,
+                             struct iomap *iomap);
 #else
-static inline int __dax_zero_page_range(struct block_device *bdev,
-               struct dax_device *dax_dev, sector_t sector,
-               unsigned int offset, unsigned int length)
+static inline int dax_iomap_zero(loff_t pos, unsigned offset, unsigned size,
+                                struct iomap *iomap)
 {
        return -ENXIO;
 }
-- 
2.18.1

--
dm-devel mailing list
[email protected]
https://www.redhat.com/mailman/listinfo/dm-devel

Reply via email to