Re: Test for CFS Bandwidth Control V6

From: Paul Turner
Date: Tue Jun 07 2011 - 22:55:20 EST


[ Sorry for the delayed response, I was out on vacation for the second
half of May until last week -- I've now caught up on email and am
preparing the next posting ]

On Mon, May 23, 2011 at 5:53 PM, Hidetoshi Seto
<seto.hidetoshi@xxxxxxxxxxxxxx> wrote:
> Hi Paul and Xiao,
>
> Please check/test a fix at the foot of this mail.
>
> (2011/05/20 11:12), Xiao Guangrong wrote:
>> Hi Paul,
>>
>> I'm so sorry for sending this mail in the new thread, since i didn't
>> receive your V6 patchset from LKML.
>>
>> It seams the patchset can not be applied, since it's conflict between
>> patch 3 and patch 5:
>>
>> ========Quote========
> (snip)
>> ========End quote========
>
> Maybe I've fixed it by hand, or git-am is so wonderful.
>
> I believe Paul will do it right for next time.
>
>>
>> I downloaded the patchset from Internet, i missed the newer version?
>>
>> I have done some test after fixed the conflict by handle, below test can cause
>> box crash:
>>
>> ========Quoteãcpu_hotlpug.sh ========
> (snip)
>> ======== End quoteãcpu_hotlpug.sh ========
>>
>> Sorry to disturb you if the bug is know.
>>
>> Thanks!
>
> Thank you for reporting it, Xiao!
>
> I confirmed that running your test cause hung-up on my box.
> And after some investigation, I found that this is an infinite loop
> in migrate_task() due to miscalculation of rq->nr_running; when a
> task is queued to throttled entity the nr_running is incremented at
> the queuing and also the unthrottling.
>
> I made a fix for this bug and it seems works well for me.
> Could you try this patch and give us your feedback, Xiao?
>
> Thanks,
> H.Seto
>
> ---
> Âkernel/sched_fair.c | Â 28 +++++++++++++---------------
> Â1 files changed, 13 insertions(+), 15 deletions(-)
>
> diff --git a/kernel/sched_fair.c b/kernel/sched_fair.c
> index 3936393..544072f 100644
> --- a/kernel/sched_fair.c
> +++ b/kernel/sched_fair.c
> @@ -1537,7 +1537,7 @@ static void unthrottle_cfs_rq(struct cfs_rq *cfs_rq)
> Â Â Â Âwalk_tg_tree_from(cfs_rq->tg, tg_unthrottle_down, tg_nop,
> Â Â Â Â Â Â Â Â Â Â Â Â Â(void *)&udd);
>
> - Â Â Â if (!cfs_rq->load.weight)
> + Â Â Â if (!cfs_rq->h_nr_running)
> Â Â Â Â Â Â Â Âreturn;
>

Why change here?

> Â Â Â Âtask_delta = cfs_rq->h_nr_running;
> @@ -1843,10 +1843,9 @@ enqueue_task_fair(struct rq *rq, struct task_struct *p, int flags)
> Â Â Â Â Â Â Â Âcfs_rq->h_nr_running++;
>
> Â Â Â Â Â Â Â Â/* end evaluation on throttled cfs_rq */
> - Â Â Â Â Â Â Â if (cfs_rq_throttled(cfs_rq)) {
> - Â Â Â Â Â Â Â Â Â Â Â se = NULL;

Hmm.. yeah this is a casualty of moving the h_nr_running computations
in-line as a part of the previous refactoring within the last
releases. This optimization (setting se = NULL to skip the second
half) obviously won't work properly with detecting whether we made it
to the end of the tree.

> - Â Â Â Â Â Â Â Â Â Â Â break;
> - Â Â Â Â Â Â Â }
> + Â Â Â Â Â Â Â if (cfs_rq_throttled(cfs_rq))
> + Â Â Â Â Â Â Â Â Â Â Â goto done;
> +
> Â Â Â Â Â Â Â Âflags = ENQUEUE_WAKEUP;
> Â Â Â Â}
>
> @@ -1855,14 +1854,14 @@ enqueue_task_fair(struct rq *rq, struct task_struct *p, int flags)
> Â Â Â Â Â Â Â Âcfs_rq->h_nr_running++;
>
> Â Â Â Â Â Â Â Âif (cfs_rq_throttled(cfs_rq))
> - Â Â Â Â Â Â Â Â Â Â Â break;
> + Â Â Â Â Â Â Â Â Â Â Â goto done;
>
> Â Â Â Â Â Â Â Âupdate_cfs_load(cfs_rq, 0);
> Â Â Â Â Â Â Â Âupdate_cfs_shares(cfs_rq);
> Â Â Â Â}
>
> - Â Â Â if (!se)
> - Â Â Â Â Â Â Â inc_nr_running(rq);
> + Â Â Â inc_nr_running(rq);
> +done:
> Â Â Â Âhrtick_update(rq);
> Â}
>
> @@ -1885,10 +1884,9 @@ static void dequeue_task_fair(struct rq *rq, struct task_struct *p, int flags)
> Â Â Â Â Â Â Â Âcfs_rq->h_nr_running--;
>
> Â Â Â Â Â Â Â Â/* end evaluation on throttled cfs_rq */
> - Â Â Â Â Â Â Â if (cfs_rq_throttled(cfs_rq)) {
> - Â Â Â Â Â Â Â Â Â Â Â se = NULL;
> - Â Â Â Â Â Â Â Â Â Â Â break;
> - Â Â Â Â Â Â Â }
> + Â Â Â Â Â Â Â if (cfs_rq_throttled(cfs_rq))
> + Â Â Â Â Â Â Â Â Â Â Â goto done;
> +
> Â Â Â Â Â Â Â Â/* Don't dequeue parent if it has other entities besides us */
> Â Â Â Â Â Â Â Âif (cfs_rq->load.weight) {
> Â Â Â Â Â Â Â Â Â Â Â Â /* Avoid double update below. */
> @@ -1910,14 +1908,14 @@ static void dequeue_task_fair(struct rq *rq, struct task_struct *p, int flags)
> Â Â Â Â Â Â Â Âcfs_rq->h_nr_running--;
>
> Â Â Â Â Â Â Â Âif (cfs_rq_throttled(cfs_rq))
> - Â Â Â Â Â Â Â Â Â Â Â break;
> + Â Â Â Â Â Â Â Â Â Â Â goto done;
>
> Â Â Â Â Â Â Â Âupdate_cfs_load(cfs_rq, 0);
> Â Â Â Â Â Â Â Âupdate_cfs_shares(cfs_rq);
> Â Â Â Â}
>
> - Â Â Â if (!se)
> - Â Â Â Â Â Â Â dec_nr_running(rq);
> + Â Â Â dec_nr_running(rq);
> +done:
> Â Â Â Âhrtick_update(rq);
> Â}
>
> --
> 1.7.4.4
>
>
>

How about instead something like the following. We can actually take
advantage of the second loop always executing by deferring the
accounting update on a throttle entity. This keeps the control flow
within dequeue_task_fair linear.

What do you think of (untested):

--- a/kernel/sched_fair.c
+++ b/kernel/sched_fair.c
@@ -1744,13 +1744,12 @@ enqueue_task_fair(struct rq *rq, struct
task_struct *p, int flags)
break;
cfs_rq = cfs_rq_of(se);
enqueue_entity(cfs_rq, se, flags);
- cfs_rq->h_nr_running++;

- /* end evaluation on throttled cfs_rq */
- if (cfs_rq_throttled(cfs_rq)) {
- se = NULL;
+ /* note: ordering with throttle check to perform
h_nr_running accounting on throttled entity below */
+ if (cfs_rq_throttled(cfs_rq))
break;
- }
+
+ cfs_rq->h_nr_running++;
flags = ENQUEUE_WAKEUP;
}

@@ -1786,13 +1785,12 @@ static void dequeue_task_fair(struct rq *rq,
struct task_struct *p, int flags)
for_each_sched_entity(se) {
cfs_rq = cfs_rq_of(se);
dequeue_entity(cfs_rq, se, flags);
- cfs_rq->h_nr_running--;

- /* end evaluation on throttled cfs_rq */
- if (cfs_rq_throttled(cfs_rq)) {
- se = NULL;
+ /* note: ordering with throttle check to perform
h_nr_running accounting on throttled entity below */
+ if (cfs_rq_throttled(cfs_rq))
break;
- }
+
+ cfs_rq->h_nr_running--;
/* Don't dequeue parent if it has other entities besides
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/