Make cfq hierarhical.

Signed-off-by: Nauman Rafique <nau...@google.com>
Signed-off-by: Fabio Checconi <fa...@gandalf.sssup.it>
Signed-off-by: Paolo Valente <paolo.vale...@unimore.it>
Signed-off-by: Aristeu Rozanski <a...@redhat.com>
Signed-off-by: Vivek Goyal <vgo...@redhat.com>
---
 block/Kconfig.iosched |    8 ++++++
 block/cfq-iosched.c   |   68 ++++++++++++++++++++++++++++++++++++++++++++++--
 init/Kconfig          |    2 +-
 3 files changed, 74 insertions(+), 4 deletions(-)

diff --git a/block/Kconfig.iosched b/block/Kconfig.iosched
index dd5224d..a91a807 100644
--- a/block/Kconfig.iosched
+++ b/block/Kconfig.iosched
@@ -54,6 +54,14 @@ config IOSCHED_CFQ
          working environment, suitable for desktop systems.
          This is the default I/O scheduler.
 
+config IOSCHED_CFQ_HIER
+       bool "CFQ Hierarchical Scheduling support"
+       depends on IOSCHED_CFQ && CGROUPS
+       select GROUP_IOSCHED
+       default n
+       ---help---
+         Enable hierarhical scheduling in cfq.
+
 choice
        prompt "Default I/O scheduler"
        default DEFAULT_CFQ
diff --git a/block/cfq-iosched.c b/block/cfq-iosched.c
index 1b67303..b64c8fd 100644
--- a/block/cfq-iosched.c
+++ b/block/cfq-iosched.c
@@ -1222,6 +1222,60 @@ static void cfq_ioc_set_ioprio(struct io_context *ioc)
        ioc->ioprio_changed = 0;
 }
 
+#ifdef CONFIG_IOSCHED_CFQ_HIER
+static void changed_cgroup(struct io_context *ioc, struct cfq_io_context *cic)
+{
+       struct cfq_queue *async_cfqq = cic_to_cfqq(cic, 0);
+       struct cfq_queue *sync_cfqq = cic_to_cfqq(cic, 1);
+       struct cfq_data *cfqd = cic->key;
+       struct io_group *iog, *__iog;
+       unsigned long flags;
+       struct request_queue *q;
+
+       if (unlikely(!cfqd))
+               return;
+
+       q = cfqd->queue;
+
+       spin_lock_irqsave(q->queue_lock, flags);
+
+       iog = io_get_io_group(q, 0);
+
+       if (async_cfqq != NULL) {
+               __iog = cfqq_to_io_group(async_cfqq);
+               if (iog != __iog) {
+                       /* cgroup changed, drop the reference to async queue */
+                       cic_set_cfqq(cic, NULL, 0);
+                       cfq_put_queue(async_cfqq);
+               }
+       }
+
+       if (sync_cfqq != NULL) {
+               __iog = cfqq_to_io_group(sync_cfqq);
+
+               /*
+                * Drop reference to sync queue. A new sync queue will
+                * be assigned in new group upon arrival of a fresh request.
+                * If old queue has got requests, those reuests will be
+                * dispatched over a period of time and queue will be freed
+                * automatically.
+                */
+               if (iog != __iog) {
+                       cic_set_cfqq(cic, NULL, 1);
+                       cfq_put_queue(sync_cfqq);
+               }
+       }
+
+       spin_unlock_irqrestore(q->queue_lock, flags);
+}
+
+static void cfq_ioc_set_cgroup(struct io_context *ioc)
+{
+       call_for_each_cic(ioc, changed_cgroup);
+       ioc->cgroup_changed = 0;
+}
+#endif  /* CONFIG_IOSCHED_CFQ_HIER */
+
 static struct cfq_queue *
 cfq_find_alloc_queue(struct cfq_data *cfqd, int is_sync,
                                struct io_context *ioc, gfp_t gfp_mask)
@@ -1230,7 +1284,10 @@ cfq_find_alloc_queue(struct cfq_data *cfqd, int is_sync,
        struct cfq_io_context *cic;
        struct request_queue *q = cfqd->queue;
        struct io_queue *ioq = NULL, *new_ioq = NULL;
+       struct io_group *iog = NULL;
 retry:
+       iog = io_get_io_group(q, 1);
+
        cic = cfq_cic_lookup(cfqd, ioc);
        /* cic always exists here */
        cfqq = cic_to_cfqq(cic, is_sync);
@@ -1297,8 +1354,9 @@ alloc_ioq:
 
                cfqq->ioq = ioq;
                cfq_init_prio_data(cfqq, ioc);
-               elv_init_ioq(q->elevator, ioq, cfqq, cfqq->org_ioprio_class,
-                               cfqq->org_ioprio, is_sync);
+               elv_init_ioq(q->elevator, ioq, iog, cfqq,
+                               cfqq->org_ioprio_class, cfqq->org_ioprio,
+                               is_sync);
 
                if (is_sync) {
                        if (!cfq_class_idle(cfqq))
@@ -1330,7 +1388,7 @@ cfq_get_queue(struct cfq_data *cfqd, int is_sync, struct 
io_context *ioc,
        const int ioprio_class = task_ioprio_class(ioc);
        struct cfq_queue *async_cfqq = NULL;
        struct cfq_queue *cfqq = NULL;
-       struct io_group *iog = io_lookup_io_group_current(cfqd->queue);
+       struct io_group *iog = io_get_io_group(cfqd->queue, 1);
 
        if (!is_sync) {
                async_cfqq = io_group_async_queue_prio(iog, ioprio_class,
@@ -1489,6 +1547,10 @@ out:
        smp_read_barrier_depends();
        if (unlikely(ioc->ioprio_changed))
                cfq_ioc_set_ioprio(ioc);
+#ifdef CONFIG_IOSCHED_CFQ_HIER
+       if (unlikely(ioc->cgroup_changed))
+               cfq_ioc_set_cgroup(ioc);
+#endif
        return cic;
 err_free:
        cfq_cic_free(cic);
diff --git a/init/Kconfig b/init/Kconfig
index ab76477..1a4686d 100644
--- a/init/Kconfig
+++ b/init/Kconfig
@@ -607,7 +607,7 @@ config CGROUP_MEM_RES_CTLR_SWAP
          size is 4096bytes, 512k per 1Gbytes of swap.
 
 config GROUP_IOSCHED
-       bool "Group IO Scheduler"
+       bool
        depends on CGROUPS && ELV_FAIR_QUEUING
        default n
        ---help---
-- 
1.6.0.6

_______________________________________________
Containers mailing list
contain...@lists.linux-foundation.org
https://lists.linux-foundation.org/mailman/listinfo/containers

_______________________________________________
Devel mailing list
Devel@openvz.org
https://openvz.org/mailman/listinfo/devel

Reply via email to