Deal with task-groups whose bandwidth hasnt been explicitly set by the administrator. Unallocated CPU bandwidth is equally distributed among such "don't care" groups.
Signed-off-by : Srivatsa Vaddagiri <[EMAIL PROTECTED]> linux-2.6.18-root/include/linux/sched.h | 2 --- 1 file changed, 91 insertions(+), 5 deletions(-) --- diff -puN kernel/sched.c~dont_care kernel/sched.c --- linux-2.6.20/kernel/sched.c~dont_care 2007-04-12 09:09:48.000000000 +0530 +++ linux-2.6.20-vatsa/kernel/sched.c 2007-04-12 11:07:15.000000000 +0530 @@ -227,6 +227,12 @@ static DEFINE_PER_CPU(struct task_grp_rq /* task-group object - maintains information about each task-group */ struct task_grp { unsigned short ticks, long_ticks; /* bandwidth given to task-group */ + int left_over_pct; + int total_dont_care_grps; + int dont_care; /* Does this group care for its bandwidth ? */ + struct task_grp *parent; + struct list_head dont_care_list; + struct list_head list; struct task_grp_rq *rq[NR_CPUS]; /* runqueue pointer for every cpu */ }; @@ -7210,6 +7216,12 @@ void __init sched_init(void) init_task_grp.ticks = CPU_CONTROL_SHORT_WINDOW; /* 100% bandwidth */ init_task_grp.long_ticks = NUM_LONG_TICKS; + init_task_grp.left_over_pct = 100; /* 100% unallocated bandwidth */ + init_task_grp.parent = NULL; + init_task_grp.total_dont_care_grps = 1; /* init_task_grp itself */ + init_task_grp.dont_care = 1; + INIT_LIST_HEAD(&init_task_grp.dont_care_list); + list_add_tail(&init_task_grp.list, &init_task_grp.dont_care_list); for_each_possible_cpu(i) { struct rq *rq; @@ -7382,19 +7394,50 @@ void set_curr_task(int cpu, struct task_ #ifdef CONFIG_CPUMETER +/* Distribute left over bandwidth equally to all "dont care" task groups */ +static void recalc_dontcare(struct task_grp *tg_root) +{ + int ticks; + struct list_head *entry; + + if (!tg_root->total_dont_care_grps) + return; + + ticks = ((tg_root->left_over_pct / tg_root->total_dont_care_grps) * + CPU_CONTROL_SHORT_WINDOW) / 100; + + list_for_each(entry, &tg_root->dont_care_list) { + struct task_grp *tg; + int i; + + tg = list_entry(entry, struct task_grp, list); + tg->ticks = ticks; + for_each_possible_cpu(i) + tg->rq[i]->ticks = tg->ticks; + } +} + /* Allocate runqueue structures for the new task-group */ -static int sched_create_group(void) +static int sched_create_group(struct task_grp *tg_parent) { struct task_grp *tg; struct task_grp_rq *tgrq; int i; + if (tg_parent->parent) + /* We don't support hierarchical CPU res mgmt (yet) */ + return -EINVAL; + tg = kzalloc(sizeof(*tg), GFP_KERNEL); if (!tg) return -ENOMEM; tg->ticks = CPU_CONTROL_SHORT_WINDOW; tg->long_ticks = NUM_LONG_TICKS; + tg->parent = tg_parent; + tg->dont_care = 1; + tg->left_over_pct = 100; + INIT_LIST_HEAD(&tg->dont_care_list); for_each_possible_cpu(i) { tgrq = kzalloc(sizeof(*tgrq), GFP_KERNEL); @@ -7404,6 +7447,15 @@ static int sched_create_group(void) task_grp_rq_init(tgrq, tg); } + if (tg->parent) { + tg->parent->total_dont_care_grps++; + list_add_tail(&tg->list, &tg->parent->dont_care_list); + recalc_dontcare(tg->parent); + } else { + tg->total_dont_care_grps = 1; + list_add_tail(&tg->list, &tg->dont_care_list); + } + /* A later patch will make 'tg' accessible beyond this function */ return 0; oom: @@ -7418,6 +7470,16 @@ oom: static void sched_destroy_group(struct task_grp *tg) { int i; + struct task_grp *tg_root = tg->parent; + + if (!tg_root) + tg_root = tg; + + if (tg->dont_care) { + tg_root->total_dont_care_grps--; + list_del(&tg->list); + recalc_dontcare(tg_root); + } for_each_possible_cpu(i) kfree(tg->rq[i]); @@ -7428,12 +7490,33 @@ static void sched_destroy_group(struct t /* Assign quota to this group */ static int sched_set_quota(struct task_grp *tg, int quota) { - int i; + int i, old_quota = 0; + struct task_grp *tg_root = tg->parent; + + if (!tg_root) + tg_root = tg; + + if (!tg->dont_care) + old_quota = (tg->ticks * 100) / CPU_CONTROL_SHORT_WINDOW; + + if ((quota - old_quota) > tg_root->left_over_pct) + return -EINVAL; + + if (tg->dont_care) { + tg->dont_care = 0; + tg_root->total_dont_care_grps--; + list_del(&tg->list); + } tg->ticks = (quota * CPU_CONTROL_SHORT_WINDOW) / 100; + for_each_possible_cpu(i) { + tg->rq[i]->ticks = tg->ticks; + tg->rq[i]->long_ticks = tg->long_ticks; + } - for_each_possible_cpu(i) - tg->rq[i]->ticks = tg->ticks; + /* xxx: needs some locking */ + tg_root->left_over_pct -= (quota - old_quota); + recalc_dontcare(tg_root); return 0; } @@ -7446,7 +7529,10 @@ static inline int cpu_quota(struct task_ /* Return assigned quota for this group */ static int sched_get_quota(struct task_grp *tg) { - return cpu_quota(tg); + if (tg->dont_care) + return 0; + else + return cpu_quota(tg); } /* _ -- Regards, vatsa ------------------------------------------------------------------------- Take Surveys. Earn Cash. Influence the Future of IT Join SourceForge.net's Techsay panel and you'll get the chance to share your opinions on IT & business topics through brief surveys-and earn cash http://www.techsay.com/default.php?page=join.php&p=sourceforge&CID=DEVDEV _______________________________________________ ckrm-tech mailing list https://lists.sourceforge.net/lists/listinfo/ckrm-tech