Re: [PATCH 2/3] read_lock migrate_disable pushdown to rt_read_lock

From: Nicholas Mc Guire
Date: Thu Dec 05 2013 - 21:34:00 EST


On Thu, 05 Dec 2013, Steven Rostedt wrote:

> On Fri, 6 Dec 2013 00:44:49 +0100
> Nicholas Mc Guire <der.herr@xxxxxxx> wrote:
>
> >
> >
> > pushdown of migrate_disable/enable from read_*lock* to the rt_read_*lock*
> > api level
> >
> > general mapping to mutexes:
> >
> > read_*lock*
> > `-> rt_read_*lock*
> > `-> __spin_lock (the sleeping spin locks)
> > `-> rt_mutex
> >
> > The real read_lock* mapping:
> >
> > read_lock_irqsave -.
> > read_lock_irq `-> rt_read_lock_irqsave()
> > `->read_lock ---------. \
> > read_lock_bh ------+ \
> > `--> rt_read_lock()
> > if (rt_mutex_owner(lock) != current){
> > `-> __rt_spin_lock()
> > rt_spin_lock_fastlock()
> > `->rt_mutex_cmpxchg()
> > migrate_disable()
> > }
> > rwlock->read_depth++;
> > read_trylock mapping:
> >
> > read_trylock
> > `-> rt_read_trylock
> > if (rt_mutex_owner(lock) != current){
> > `-> rt_mutex_trylock()
> > rt_mutex_fasttrylock()
> > rt_mutex_cmpxchg()
> > migrate_disable()
> > }
> > rwlock->read_depth++;
> >
> > read_unlock* mapping:
> >
> > read_unlock_bh --------+
> > read_unlock_irq -------+
> > read_unlock_irqrestore +
> > read_unlock -----------+
> > `-> rt_read_unlock()
> > if(--rwlock->read_depth==0){
> > `-> __rt_spin_unlock()
> > rt_spin_lock_fastunlock()
> > `-> rt_mutex_cmpxchg()
> > migrate_disable()
> > }
> >
> > So calls to migrate_disable/enable() are better placed at the rt_read_*
> > level of lock/trylock/unlock as all of the read_*lock* API has this as a
> > common path. In the rt_read* API of lock/trylock/unlock the nesting level
> > is already being recorded in rwlock->read_depth, so we can push down the
> > migrate disable/enable to that level and condition it on the read_depth
> > going from 0 to 1 -> migrate_disable and 1 to 0 -> migrate_enable. This
> > eliminates the recursive calls that were needed when migrate_disable/enable
> > was done at the read_*lock* level.
> >
> > The approach to read_*_bh also eliminates the concerns raised with the
> > regards to api inbalances (read_lock_bh -> read_unlock+local_bh_enable)
> >
> > this is on top of 3.12.1-rt4 with the first batch of migration_cleanup
> > patches applied.
> >
> > No change of functional behavior
> >
> > Signed-off-by: Nicholas Mc Guire <der.herr@xxxxxxx>
> > ---
> > include/linux/rwlock_rt.h | 6 ------
> > kernel/rt.c | 9 +++++----
> > 2 files changed, 5 insertions(+), 10 deletions(-)
> >
> > diff --git a/include/linux/rwlock_rt.h b/include/linux/rwlock_rt.h
> > index 853ee36..6d86c33 100644
> > --- a/include/linux/rwlock_rt.h
> > +++ b/include/linux/rwlock_rt.h
> > @@ -33,7 +33,6 @@ extern void __rt_rwlock_init(rwlock_t *rwlock, char *name, struct lock_class_key
> > #define read_lock_irqsave(lock, flags) \
> > do { \
> > typecheck(unsigned long, flags); \
> > - migrate_disable(); \
> > flags = rt_read_lock_irqsave(lock); \
> > } while (0)
> >
> > @@ -46,14 +45,12 @@ extern void __rt_rwlock_init(rwlock_t *rwlock, char *name, struct lock_class_key
> >
> > #define read_lock(lock) \
> > do { \
> > - migrate_disable(); \
> > rt_read_lock(lock); \
> > } while (0)
> >
> > #define read_lock_bh(lock) \
> > do { \
> > local_bh_disable(); \
> > - migrate_disable(); \
> > rt_read_lock(lock); \
> > } while (0)
> >
> > @@ -77,13 +74,11 @@ extern void __rt_rwlock_init(rwlock_t *rwlock, char *name, struct lock_class_key
> > #define read_unlock(lock) \
> > do { \
> > rt_read_unlock(lock); \
> > - migrate_enable(); \
> > } while (0)
> >
> > #define read_unlock_bh(lock) \
> > do { \
> > rt_read_unlock(lock); \
> > - migrate_enable(); \
> > local_bh_enable(); \
> > } while (0)
> >
> > @@ -109,7 +104,6 @@ extern void __rt_rwlock_init(rwlock_t *rwlock, char *name, struct lock_class_key
> > typecheck(unsigned long, flags); \
> > (void) flags; \
> > rt_read_unlock(lock); \
> > - migrate_enable(); \
> > } while (0)
> >
> > #define write_unlock_irqrestore(lock, flags) \
> > diff --git a/kernel/rt.c b/kernel/rt.c
> > index 29771e2..3403000 100644
> > --- a/kernel/rt.c
> > +++ b/kernel/rt.c
> > @@ -213,19 +213,18 @@ int __lockfunc rt_read_trylock(rwlock_t *rwlock)
> > * but not when read_depth == 0 which means that the lock is
> > * write locked.
> > */
> > - migrate_disable();
> > if (rt_mutex_owner(lock) != current) {
> > ret = rt_mutex_trylock(lock);
> > - if (ret)
> > + if (ret) {
> > + migrate_disable();
> > rwlock_acquire(&rwlock->dep_map, 0, 1, _RET_IP_);
> > + }
>
> I like this patch in general, but I'm nervous about this part.
>
> What was the original reason to disable migration before checking the
> mutex owner? Seems like we should have only disabled it on success. I'm
> assuming there was some subtle reason not to.
>

I think that it was simply being treated semantically as a lock - but it
is actually not.
hold a
migrate_disable/lock per_cpu unlock/migrate_enable
lock/migrate_disable per_cpu migrate_enable/unlcok
lock/migrate_disable per_cpu unlock/migrate_enable
migrate_disable/lock per_cpu migrate_enable/unlcok
reference

are all equivalent - the only thing you need to ensure that the per cpu
object will not be accessed before both lock and migration_disable have
been sucessful. So from my understanding this is safe.

if we get migrated after a succesful trylock what would go wrong ?
the protected object will not be accessed until after the spin_trylock
returned so migration is disabled

> If there was some strange reason, I'm not convinced that your change
> makes that reason go away.
>
IF there is a reason then this patch is broken - my conclusion up to now
is that there is no such reason.

thx!
hofrat
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/