This is a Proof of Concept to introduce a scalable, multi-threaded
decompression framework into fsck.erofs to reduce extraction time.

Baseline Profiling:
Using the Linux 6.7 kernel source packed with LZ4HC (4K pclusters),
perf showed a strictly synchronous execution path. The main thread
spent ~52% of its time in LZ4_decompress_safe, heavily blocked by
synchronous I/O (~32% in el0_svc/vfs_read).

First Iteration (Naive Workqueue):
A standard producer-consumer workqueue overlapping compute with pwrite()
suffered massive scheduling overhead. For 4KB LZ4 clusters, workers
spent ~44% of CPU time spinning on __arm64_sys_futex and try_to_wake_up.

Current PoC (Dynamic Pcluster Batching):
To eliminate lock contention, this patch introduces a batching context.
Instead of queuing 1 pcluster per task, the main thread collects an
array of sequential pclusters (Z_EROFS_PCLUSTER_BATCH_SIZE = 32) before
submitting a single erofs_work unit.

Results:
- Scheduling overhead (futex) dropped significantly.
- Workers stay cache-hot, decompressing 32 blocks per wakeup.
- LZ4_decompress_safe is successfully offloaded to background cores
  (~18.8% self-execution time), completely decoupled from main thread I/O.

Signed-off-by: Nithurshen <[email protected]>
---
 fsck/main.c              | 144 ++++++++++++-------------------
 include/erofs/internal.h |  15 +++-
 lib/data.c               | 182 +++++++++++++++++++++++++++++++--------
 3 files changed, 217 insertions(+), 124 deletions(-)

diff --git a/fsck/main.c b/fsck/main.c
index ab697be..1b6db42 100644
--- a/fsck/main.c
+++ b/fsck/main.c
@@ -15,6 +15,9 @@
 #include "erofs/xattr.h"
 #include "../lib/compressor.h"
 #include "../lib/liberofs_compress.h"
+#include "erofs/workqueue.h"
+
+extern struct erofs_workqueue erofs_wq;
 
 static int erofsfsck_check_inode(erofs_nid_t pnid, erofs_nid_t nid);
 
@@ -493,135 +496,96 @@ out:
 
 static int erofs_verify_inode_data(struct erofs_inode *inode, int outfd)
 {
-       struct erofs_map_blocks map = {
-               .buf = __EROFS_BUF_INITIALIZER,
-       };
+       struct erofs_map_blocks map = { .buf = __EROFS_BUF_INITIALIZER };
        bool needdecode = fsckcfg.check_decomp && !erofs_is_packed_inode(inode);
        int ret = 0;
-       bool compressed;
+       bool compressed = erofs_inode_is_data_compressed(inode->datalayout);
        erofs_off_t pos = 0;
        u64 pchunk_len = 0;
-       unsigned int raw_size = 0, buffer_size = 0;
-       char *raw = NULL, *buffer = NULL;
 
-       erofs_dbg("verify data chunk of nid(%llu): type(%d)",
-                 inode->nid | 0ULL, inode->datalayout);
+       struct z_erofs_read_ctx ctx = {
+               .pending_tasks = 0,
+               .final_err = 0,
+               .outfd = outfd,
+               .current_task = NULL
+       };
+       pthread_mutex_init(&ctx.lock, NULL);
+       pthread_cond_init(&ctx.cond, NULL);
 
-       compressed = erofs_inode_is_data_compressed(inode->datalayout);
-       while (pos < inode->i_size) {
-               unsigned int alloc_rawsize;
+       erofs_dbg("verify data chunk of nid(%llu): type(%d)", inode->nid | 
0ULL, inode->datalayout);
 
+       while (pos < inode->i_size) {
                map.m_la = pos;
                ret = erofs_map_blocks(inode, &map, EROFS_GET_BLOCKS_FIEMAP);
-               if (ret)
-                       goto out;
+               if (ret) goto out;
 
-               if (!compressed && map.m_llen != map.m_plen) {
-                       erofs_err("broken chunk length m_la %" PRIu64 " m_llen 
%" PRIu64 " m_plen %" PRIu64,
-                                 map.m_la, map.m_llen, map.m_plen);
-                       ret = -EFSCORRUPTED;
-                       goto out;
-               }
-
-               /* the last lcluster can be divided into 3 parts */
                if (map.m_la + map.m_llen > inode->i_size)
                        map.m_llen = inode->i_size - map.m_la;
 
                pchunk_len += map.m_plen;
                pos += map.m_llen;
 
-               /* should skip decomp? */
                if (map.m_la >= inode->i_size || !needdecode)
                        continue;
 
                if (outfd >= 0 && !(map.m_flags & EROFS_MAP_MAPPED)) {
-                       ret = lseek(outfd, map.m_llen, SEEK_CUR);
-                       if (ret < 0) {
-                               ret = -errno;
-                               goto out;
-                       }
                        continue;
                }
 
-               if (map.m_plen > Z_EROFS_PCLUSTER_MAX_SIZE) {
-                       if (compressed && !(map.m_flags & 
__EROFS_MAP_FRAGMENT)) {
-                               erofs_err("invalid pcluster size %" PRIu64 " @ 
offset %" PRIu64 " of nid %" PRIu64,
-                                         map.m_plen, map.m_la,
-                                         inode->nid | 0ULL);
-                               ret = -EFSCORRUPTED;
-                               goto out;
-                       }
-                       alloc_rawsize = Z_EROFS_PCLUSTER_MAX_SIZE;
-               } else {
-                       alloc_rawsize = map.m_plen;
-               }
-
-               if (alloc_rawsize > raw_size) {
-                       char *newraw = realloc(raw, alloc_rawsize);
-
-                       if (!newraw) {
+               if (compressed) {
+                       char *raw = malloc(map.m_plen);
+                       size_t buffer_size = map.m_llen > 
erofs_blksiz(inode->sbi) ? map.m_llen : erofs_blksiz(inode->sbi);
+                       char *buffer = malloc(buffer_size);
+                       if (!raw || !buffer) {
+                               free(raw); free(buffer);
                                ret = -ENOMEM;
                                goto out;
                        }
-                       raw = newraw;
-                       raw_size = alloc_rawsize;
-               }
 
-               if (compressed) {
-                       if (map.m_llen > buffer_size) {
-                               char *newbuffer;
-
-                               buffer_size = map.m_llen;
-                               newbuffer = realloc(buffer, buffer_size);
-                               if (!newbuffer) {
-                                       ret = -ENOMEM;
-                                       goto out;
-                               }
-                               buffer = newbuffer;
+                       pthread_mutex_lock(&ctx.lock);
+                       if (erofs_wq.job_count > 128) {
+                               z_erofs_read_ctx_enqueue(&ctx);
+                               while (ctx.pending_tasks > 0)
+                                       pthread_cond_wait(&ctx.cond, &ctx.lock);
                        }
-                       ret = z_erofs_read_one_data(inode, &map, raw, buffer,
-                                                   0, map.m_llen, false);
-                       if (ret)
+                       pthread_mutex_unlock(&ctx.lock);
+
+                       ret = z_erofs_read_one_data(inode, &map, raw, buffer, 
0, map.m_llen, false, map.m_la, &ctx);
+                       if (ret) {
+                               free(raw); free(buffer);
                                goto out;
+                       }
 
-                       if (outfd >= 0 && write(outfd, buffer, map.m_llen) < 0)
-                               goto fail_eio;
                } else {
-                       u64 p = 0;
-
-                       do {
-                               u64 count = min_t(u64, alloc_rawsize,
-                                                 map.m_llen);
-
-                               ret = erofs_read_one_data(inode, &map, raw, p, 
count);
-                               if (ret)
-                                       goto out;
-
-                               if (outfd >= 0 && write(outfd, raw, count) < 0)
-                                       goto fail_eio;
-                               map.m_llen -= count;
-                               p += count;
-                       } while (map.m_llen);
+                       char *raw = malloc(map.m_llen);
+                       ret = erofs_read_one_data(inode, &map, raw, 0, 
map.m_llen);
+                       if (ret == 0 && outfd >= 0)
+                               pwrite(outfd, raw, map.m_llen, map.m_la);
+                       free(raw);
+                       if (ret) goto out;
                }
        }
+       z_erofs_read_ctx_enqueue(&ctx);
+
+out:
+       pthread_mutex_lock(&ctx.lock);
+       while (ctx.pending_tasks > 0)
+               pthread_cond_wait(&ctx.cond, &ctx.lock);
+       if (ctx.final_err < 0 && ret == 0)
+               ret = ctx.final_err;
+       pthread_mutex_unlock(&ctx.lock);
 
        if (fsckcfg.print_comp_ratio) {
                if (!erofs_is_packed_inode(inode))
                        fsckcfg.logical_blocks += BLK_ROUND_UP(inode->sbi, 
inode->i_size);
                fsckcfg.physical_blocks += BLK_ROUND_UP(inode->sbi, pchunk_len);
        }
-out:
-       if (raw)
-               free(raw);
-       if (buffer)
-               free(buffer);
-       return ret < 0 ? ret : 0;
 
-fail_eio:
-       erofs_err("I/O error occurred when verifying data chunk @ nid %llu",
-                 inode->nid | 0ULL);
-       ret = -EIO;
-       goto out;
+       if (outfd >= 0 && ret == 0)
+               ftruncate(outfd, inode->i_size);
+       pthread_mutex_destroy(&ctx.lock);
+       pthread_cond_destroy(&ctx.cond);
+       return ret < 0 ? ret : 0;
 }
 
 static inline int erofs_extract_dir(struct erofs_inode *inode)
@@ -1019,6 +983,8 @@ int main(int argc, char *argv[])
 
        erofs_init_configure();
 
+       erofs_alloc_workqueue(&erofs_wq, 4, 256, NULL, NULL);
+
        fsckcfg.physical_blocks = 0;
        fsckcfg.logical_blocks = 0;
        fsckcfg.extract_path = NULL;
diff --git a/include/erofs/internal.h b/include/erofs/internal.h
index e741f1c..de9ac49 100644
--- a/include/erofs/internal.h
+++ b/include/erofs/internal.h
@@ -63,6 +63,8 @@ struct erofs_buf {
 #define BLK_ROUND_UP(sbi, addr)        \
        (roundup(addr, erofs_blksiz(sbi)) >> (sbi)->blkszbits)
 
+#define Z_EROFS_PCLUSTER_BATCH_SIZE 32
+
 struct erofs_buffer_head;
 struct erofs_bufmgr;
 
@@ -475,9 +477,20 @@ int erofs_map_blocks(struct erofs_inode *inode,
 int erofs_map_dev(struct erofs_sb_info *sbi, struct erofs_map_dev *map);
 int erofs_read_one_data(struct erofs_inode *inode, struct erofs_map_blocks 
*map,
                        char *buffer, u64 offset, size_t len);
+struct z_erofs_decompress_task;
+struct z_erofs_read_ctx {
+       pthread_mutex_t lock;
+       pthread_cond_t cond;
+       int pending_tasks;
+       int final_err;
+       int outfd;
+       struct z_erofs_decompress_task *current_task;
+};
+void z_erofs_read_ctx_enqueue(struct z_erofs_read_ctx *ctx);
 int z_erofs_read_one_data(struct erofs_inode *inode,
                        struct erofs_map_blocks *map, char *raw, char *buffer,
-                       erofs_off_t skip, erofs_off_t length, bool trimmed);
+                       erofs_off_t skip, erofs_off_t length, bool trimmed,
+                       erofs_off_t out_offset, struct z_erofs_read_ctx *ctx);
 void *erofs_read_metadata(struct erofs_sb_info *sbi, erofs_nid_t nid,
                          erofs_off_t *offset, int *lengthp);
 int z_erofs_parse_cfgs(struct erofs_sb_info *sbi, struct erofs_super_block 
*dsb);
diff --git a/lib/data.c b/lib/data.c
index 6fd1389..4d8fcef 100644
--- a/lib/data.c
+++ b/lib/data.c
@@ -9,6 +9,35 @@
 #include "erofs/trace.h"
 #include "erofs/decompress.h"
 #include "liberofs_fragments.h"
+#include "erofs/workqueue.h"
+#include <pthread.h>
+
+struct erofs_workqueue erofs_wq;
+
+/* struct z_erofs_read_ctx {
+       pthread_mutex_t lock;
+       pthread_cond_t cond;
+       int pending_tasks;
+       int final_err;
+       int outfd;
+}; */
+
+struct z_erofs_decompress_task {
+       struct erofs_work work;
+       struct z_erofs_read_ctx *ctx;
+       struct z_erofs_decompress_req reqs[Z_EROFS_PCLUSTER_BATCH_SIZE];
+       char *raw_bufs[Z_EROFS_PCLUSTER_BATCH_SIZE];
+       erofs_off_t out_offsets[Z_EROFS_PCLUSTER_BATCH_SIZE];
+       unsigned int nr_reqs;
+};
+
+void z_erofs_read_ctx_enqueue(struct z_erofs_read_ctx *ctx)
+{
+       if (ctx && ctx->current_task) {
+               erofs_queue_work(&erofs_wq, &ctx->current_task->work);
+               ctx->current_task = NULL;
+       }
+}
 
 void *erofs_bread(struct erofs_buf *buf, erofs_off_t offset, bool need_kmap)
 {
@@ -275,9 +304,45 @@ static int erofs_read_raw_data(struct erofs_inode *inode, 
char *buffer,
        return 0;
 }
 
+static void z_erofs_decompress_worker(struct erofs_work *work, void *tlsp)
+{
+    struct z_erofs_decompress_task *task = (struct z_erofs_decompress_task 
*)work;
+    struct z_erofs_read_ctx *ctx = task->ctx;
+    int i, ret = 0, first_err = 0;
+
+    for (i = 0; i < task->nr_reqs; ++i) {
+        ret = z_erofs_decompress(&task->reqs[i]);
+
+        if (ret == 0 && ctx && ctx->outfd >= 0) {
+            if (pwrite(ctx->outfd, task->reqs[i].out,
+                       task->reqs[i].decodedlength, task->out_offsets[i]) < 0)
+                ret = -errno;
+        }
+
+        if (ret < 0 && first_err == 0)
+            first_err = ret;
+
+        free(task->raw_bufs[i]);
+        if (ctx && ctx->outfd >= 0)
+            free(task->reqs[i].out);
+    }
+
+    if (ctx) {
+        pthread_mutex_lock(&ctx->lock);
+        if (first_err < 0 && ctx->final_err == 0)
+            ctx->final_err = first_err;
+        ctx->pending_tasks--;
+        if (ctx->pending_tasks == 0)
+            pthread_cond_signal(&ctx->cond);
+        pthread_mutex_unlock(&ctx->lock);
+    }
+    free(task);
+}
+
 int z_erofs_read_one_data(struct erofs_inode *inode,
                        struct erofs_map_blocks *map, char *raw, char *buffer,
-                       erofs_off_t skip, erofs_off_t length, bool trimmed)
+                       erofs_off_t skip, erofs_off_t length, bool trimmed,
+                       erofs_off_t out_offset, struct z_erofs_read_ctx *ctx)
 {
        struct erofs_sb_info *sbi = inode->sbi;
        struct erofs_map_dev mdev;
@@ -307,24 +372,40 @@ int z_erofs_read_one_data(struct erofs_inode *inode,
        if (ret < 0)
                return ret;
 
-       ret = z_erofs_decompress(&(struct z_erofs_decompress_req) {
-                       .sbi = sbi,
-                       .in = raw,
-                       .out = buffer,
-                       .decodedskip = skip,
-                       .interlaced_offset =
-                               map->m_algorithmformat == 
Z_EROFS_COMPRESSION_INTERLACED ?
-                                       erofs_blkoff(sbi, map->m_la) : 0,
-                       .inputsize = map->m_plen,
-                       .decodedlength = length,
-                       .alg = map->m_algorithmformat,
-                       .partial_decoding = trimmed ? true :
-                               !(map->m_flags & EROFS_MAP_FULL_MAPPED) ||
-                                       (map->m_flags & EROFS_MAP_PARTIAL_REF),
-                        });
-       if (ret < 0)
-               return ret;
-       return 0;
+       struct z_erofs_decompress_task *task = ctx->current_task;
+    if (!task) {
+        task = calloc(1, sizeof(*task));
+        task->ctx = ctx;
+        task->work.fn = z_erofs_decompress_worker;
+        ctx->current_task = task;
+
+        pthread_mutex_lock(&ctx->lock);
+        ctx->pending_tasks++;
+        pthread_mutex_unlock(&ctx->lock);
+    }
+
+    int idx = task->nr_reqs++;
+    task->reqs[idx] = (struct z_erofs_decompress_req) {
+        .sbi = sbi,
+        .in = raw,
+        .out = buffer,
+        .decodedskip = skip,
+        .interlaced_offset = map->m_algorithmformat == 
Z_EROFS_COMPRESSION_INTERLACED ?
+                    erofs_blkoff(sbi, map->m_la) : 0,
+        .inputsize = map->m_plen,
+        .decodedlength = length,
+        .alg = map->m_algorithmformat,
+        .partial_decoding = trimmed ? true :
+            !(map->m_flags & EROFS_MAP_FULL_MAPPED) ||
+                (map->m_flags & EROFS_MAP_PARTIAL_REF),
+    };
+    task->raw_bufs[idx] = raw;
+    task->out_offsets[idx] = out_offset;
+
+    if (task->nr_reqs == Z_EROFS_PCLUSTER_BATCH_SIZE) {
+        z_erofs_read_ctx_enqueue(ctx);
+    }
+    return 0;
 }
 
 static int z_erofs_read_data(struct erofs_inode *inode, char *buffer,
@@ -335,10 +416,17 @@ static int z_erofs_read_data(struct erofs_inode *inode, 
char *buffer,
                .buf = __EROFS_BUF_INITIALIZER,
        };
        bool trimmed;
-       unsigned int bufsize = 0;
-       char *raw = NULL;
        int ret = 0;
 
+       struct z_erofs_read_ctx ctx = {
+               .pending_tasks = 0,
+               .final_err = 0,
+               .outfd = -1,
+               .current_task = NULL
+       };
+       pthread_mutex_init(&ctx.lock, NULL);
+       pthread_cond_init(&ctx.cond, NULL);
+
        end = offset + size;
        while (end > offset) {
                map.m_la = end - 1;
@@ -374,25 +462,51 @@ static int z_erofs_read_data(struct erofs_inode *inode, 
char *buffer,
                        continue;
                }
 
-               if (map.m_plen > bufsize) {
-                       char *newraw;
+               /*
+                * If the global workqueue is getting too deep,
+                * dynamically throttle the producer by forcing the main thread
+                * to wait early. Prevents memory bloat from fast I/O out-pacing
+                * the decompression threads.
+                */
+               pthread_mutex_lock(&ctx.lock);
+               if (erofs_wq.job_count > 128) {
+                       z_erofs_read_ctx_enqueue(&ctx);
+                       while (ctx.pending_tasks > 0)
+                               pthread_cond_wait(&ctx.cond, &ctx.lock);
+               }
+               pthread_mutex_unlock(&ctx.lock);
 
-                       bufsize = map.m_plen;
-                       newraw = realloc(raw, bufsize);
-                       if (!newraw) {
-                               ret = -ENOMEM;
-                               break;
-                       }
-                       raw = newraw;
+               /* Allocate fresh raw buffer for each pcluster. */
+               char *raw = malloc(map.m_plen);
+               if (!raw) {
+                       ret = -ENOMEM;
+                       break;
                }
 
                ret = z_erofs_read_one_data(inode, &map, raw,
-                               buffer + end - offset, skip, length, trimmed);
-               if (ret < 0)
+                               buffer + end - offset, skip, length, trimmed, 
0, &ctx);
+               if (ret < 0) {
+                       free(raw);
                        break;
+               }
+       }
+       z_erofs_read_ctx_enqueue(&ctx);
+
+       /*
+        * Wait for all queued pclusters for this read request to finish
+        * before allowing the VFS layer or fsck core to consume the buffer.
+        */
+       pthread_mutex_lock(&ctx.lock);
+       while (ctx.pending_tasks > 0) {
+               pthread_cond_wait(&ctx.cond, &ctx.lock);
        }
-       if (raw)
-               free(raw);
+
+       /* Bubble up any decompression errors caught by the worker threads */
+       if (ctx.final_err < 0 && ret == 0)
+               ret = ctx.final_err;
+
+       pthread_mutex_destroy(&ctx.lock);
+       pthread_cond_destroy(&ctx.cond);
        return ret < 0 ? ret : 0;
 }
 
-- 
2.51.0


Reply via email to