It is worth noting that using dededuplication does consume a lot of
memory.

Deduplication tables consume memory and eventually spill over and
consume disk space - this causes extra read and write operations for
every block of data on which deduplication is attempted cause more I/O
waits.

A system with large pools with small memory areas does not perform
deduplication well.

So, if you run the following command on your pool (pool-ssd in this
example)

sudo zdb -s pool-ssd

..you see something like the following:

  pool: pool-ssd
 state: ONLINE
  scan: scrub repaired 0B in 0 days 00:31:14 with 0 errors on Wed Sep  2 
14:35:52 2020
config:

        NAME                                            STATE     READ WRITE 
CKSUM
        pool-ssd                                        ONLINE       0     0    
 0
          mirror-0                                      ONLINE       0     0    
 0
            ata-INTEL_SSDSC2BW480H6_CVTR608104NW480EGN  ONLINE       0     0    
 0
            ata-INTEL_SSDSC2BW480H6_CVTR552500S0480EGN  ONLINE       0     0    
 0

errors: No known data errors

 dedup: DDT entries 4258028, size 448B on disk, 306B in core

bucket              allocated                       referenced          
______   ______________________________   ______________________________
refcnt   blocks   LSIZE   PSIZE   DSIZE   blocks   LSIZE   PSIZE   DSIZE
------   ------   -----   -----   -----   ------   -----   -----   -----
     1    3.63M    464G    226G    226G    3.63M    464G    226G    226G
     2     376K   46.1G   24.3G   24.3G     822K    100G   53.2G   53.2G
     4    51.3K   5.60G   2.51G   2.51G     240K   26.3G   11.8G   11.8G
     8    7.49K    925M    544M    544M    77.9K   9.41G   5.47G   5.47G
    16    3.32K    415M    261M    261M    68.8K   8.41G   5.35G   5.35G
    32      657   78.7M   62.1M   62.1M    29.3K   3.51G   2.79G   2.79G
    64      389   46.2M   42.7M   42.7M    34.0K   4.03G   3.78G   3.78G
   128      248   30.1M   29.5M   29.5M    44.8K   5.44G   5.32G   5.32G
   256      339   42.1M   40.8M   40.8M     123K   15.4G   14.9G   14.9G
   512      374   46.8M   46.1M   46.1M     271K   33.9G   33.5G   33.5G
    1K      254   31.8M   31.3M   31.3M     355K   44.4G   43.8G   43.8G
    2K        5    640K    513K    513K    12.9K   1.61G   1.20G   1.20G
    4K        4    512K    512K    512K    22.7K   2.84G   2.84G   2.84G
    8K        8      1M    897K    897K    89.3K   11.2G   10.1G   10.1G
   64K        2    256K      6K      6K     226K   28.3G    679M    679M
 Total    4.06M    517G    254G    254G    5.99M    759G    421G    421G

---
Looking at the "dedup: DDT entries 4258028, size 448B on disk, 306B in core" 
line, we have:

4258028 de-dup entries, each entry used 306 bytes, so that's ~1.2GB of
memory for just the de-dup/ table.  But it is capped at 1/4 the size of
the ARC, so you may find on a small memory system such as a raspberry pi
that a lot of the dedup table can't be stored in memory and needs to be
on-disk, and this increases the I/O waits.

It may be an idea to disable de-dup and see if it helps to resolve your
issue on a comparatively slow and memory constrained system such as a
raspberry pi when you have large pools

** Changed in: zfs-linux (Ubuntu)
       Status: New => Incomplete

** Changed in: zfs-linux (Ubuntu)
   Importance: Undecided => Medium

** Changed in: zfs-linux (Ubuntu)
     Assignee: (unassigned) => Colin Ian King (colin-king)

-- 
You received this bug notification because you are a member of Ubuntu
Bugs, which is subscribed to Ubuntu.
https://bugs.launchpad.net/bugs/1899249

Title:
  OpenZFS writing stalls, under load

To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/zfs-linux/+bug/1899249/+subscriptions

-- 
ubuntu-bugs mailing list
[email protected]
https://lists.ubuntu.com/mailman/listinfo/ubuntu-bugs

Reply via email to