Re: [RFC v2 PATCH 3/8] sched: Bandwidth initialization for fairtask groups

From: Bharata B Rao
Date: Tue Oct 13 2009 - 23:51:21 EST


On Tue, Oct 13, 2009 at 04:27:01PM +0200, Peter Zijlstra wrote:
> On Wed, 2009-09-30 at 18:22 +0530, Bharata B Rao wrote:
>
> > diff --git a/kernel/sched.c b/kernel/sched.c
> > index c283d0f..0147f6f 100644
> > --- a/kernel/sched.c
> > +++ b/kernel/sched.c
> > @@ -262,6 +262,15 @@ static DEFINE_MUTEX(sched_domains_mutex);
> >
> > #include <linux/cgroup.h>
> >
> > +#if defined(CONFIG_FAIR_GROUP_SCHED) && defined(CONFIG_CFS_HARD_LIMITS)
> > +struct cfs_bandwidth {
> > + spinlock_t cfs_runtime_lock;
> > + ktime_t cfs_period;
> > + u64 cfs_runtime;
> > + struct hrtimer cfs_period_timer;
> > +};
> > +#endif
>
> too much cfs here..

Right, this will eventually be merged with rt_bandwidth. Dhaval already has
patches for bandwidth code unification b/n cfs and rt. As I said, the initial
focus is to show how the hard limit code looks like.

>
> > struct cfs_rq;
> >
> > static LIST_HEAD(task_groups);
> > @@ -282,6 +291,11 @@ struct task_group {
> > /* runqueue "owned" by this group on each cpu */
> > struct cfs_rq **cfs_rq;
> > unsigned long shares;
> > +#ifdef CONFIG_CFS_HARD_LIMITS
> > + struct cfs_bandwidth cfs_bandwidth;
> > + /* If set, throttle when the group exceeds its bandwidth */
> > + int hard_limit_enabled;
> > +#endif
>
> What's wrong with doing something like cfs_bandwidth.cfs_runtime ==
> RUNTIME_INF ?

Can be done.

>
> > #endif
> >
> > #ifdef CONFIG_RT_GROUP_SCHED
> > @@ -477,6 +491,16 @@ struct cfs_rq {
> > unsigned long rq_weight;
> > #endif
> > #endif
> > +#ifdef CONFIG_CFS_HARD_LIMITS
> > + /* set when the group is throttled on this cpu */
> > + int cfs_throttled;
> > +
> > + /* runtime currently consumed by the group on this rq */
> > + u64 cfs_time;
> > +
> > + /* runtime available to the group on this rq */
> > + u64 cfs_runtime;
> > +#endif
>
> too much cfs_ again.

But this is needed. It is present in rt also.

>
> > /*
> > * Number of tasks at this heirarchy.
> > */
> > @@ -665,6 +689,11 @@ struct rq {
> > /* BKL stats */
> > unsigned int bkl_count;
> > #endif
> > + /*
> > + * Protects the cfs runtime related fields of all cfs_rqs under
> > + * this rq
> > + */
> > + spinlock_t runtime_lock;
> > };
> >
> > static DEFINE_PER_CPU_SHARED_ALIGNED(struct rq, runqueues);
>
>
> > +static inline void rq_runtime_lock(struct rq *rq)
> > +{
> > + spin_lock(&rq->runtime_lock);
> > +}
> > +
> > +static inline void rq_runtime_unlock(struct rq *rq)
> > +{
> > + spin_unlock(&rq->runtime_lock);
> > +}
>
> needless obfuscation.

This is needed to keep the code clean for !CFS_HARD_LIMITS case.

>
> > CONFIG_RT_GROUP_SCHED
> > @@ -10317,6 +10617,23 @@ static struct cftype cpu_files[] = {
> > .read_u64 = cpu_shares_read_u64,
> > .write_u64 = cpu_shares_write_u64,
> > },
> > +#ifdef CONFIG_CFS_HARD_LIMITS
> > + {
> > + .name = "cfs_runtime_us",
> > + .read_s64 = cpu_cfs_runtime_read_s64,
> > + .write_s64 = cpu_cfs_runtime_write_s64,
> > + },
> > + {
> > + .name = "cfs_period_us",
> > + .read_u64 = cpu_cfs_period_read_u64,
> > + .write_u64 = cpu_cfs_period_write_u64,
> > + },
> > + {
> > + .name = "cfs_hard_limit",
> > + .read_u64 = cpu_cfs_hard_limit_read_u64,
> > + .write_u64 = cpu_cfs_hard_limit_write_u64,
> > + },
> > +#endif /* CONFIG_CFS_HARD_LIMITS */
> > #endif
> > #ifdef CONFIG_RT_GROUP_SCHED
> > {
>
> I guess that cfs_hard_limit thing is superfluous as well.

Ok, will try to remove this control and will treat the case when
runtime != RUNTIME_INF as hard limits enabled case.

Regards,
Bharata.
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/