Re: [patch] cpusets: avoid looping when storing to mems_allowed ifone node remains set

From: Miao Xie
Date: Tue Sep 13 2011 - 00:10:13 EST


On fri, 9 Sep 2011 03:15:17 -0700 (pdt), David Rientjes wrote:
> {get,put}_mems_allowed() exist so that general kernel code may locklessly
> access a task's set of allowable nodes without having the chance that a
> concurrent write will cause the nodemask to be empty on configurations
> where MAX_NUMNODES > BITS_PER_LONG.
>
> This could incur a significant delay, however, especially in low memory
> conditions because the page allocator is blocking and reclaim requires
> get_mems_allowed() itself. It is not atypical to see writes to cpuset.mems
> take over 2 seconds to complete, for example. In low memory conditions,
> this is problematic because it's one of the most imporant times to change
> cpuset.mems in the first place!
>
> The only way a task's set of allowable nodes may change is through
> cpusets by writing to cpuset.mems and when attaching a task to a
> different cpuset. This is done by setting all new nodes, ensuring
> generic code is not reading the nodemask with get_mems_allowed() at the
> same time, and then clearing all the old nodes. This prevents the
> possibility that a reader will see an empty nodemask at the same time the
> writer is storing a new nodemask.
>
> If at least one node remains unchanged, though, it's possible to simply
> set all new nodes and then clear all the old nodes. Changing a task's
> nodemask is protected by cgroup_mutex so it's guaranteed that two threads
> are not changing the same task's nodemask at the same time, so the
> nodemask is guaranteed to be stored before another thread changes it and
> determines whether a node remains set or not.

This patch is dangerous if the task has a bind memory policy that was set
to be neither MPOL_F_STATIC_NODES nor MPOL_F_RELATIVE_NODES, because the
memory policy use node_remap() to rebind the allowed nodes, but node_remap()
may make the old mask and the new mask nonoverlapping. So at this condition,
the task may also see an empty node mask.

Thanks
Miao

>
> Signed-off-by: David Rientjes <rientjes@xxxxxxxxxx>
> ---
> kernel/cpuset.c | 9 ++++++---
> 1 files changed, 6 insertions(+), 3 deletions(-)
>
> diff --git a/kernel/cpuset.c b/kernel/cpuset.c
> --- a/kernel/cpuset.c
> +++ b/kernel/cpuset.c
> @@ -949,6 +949,8 @@ static void cpuset_migrate_mm(struct mm_struct *mm, const nodemask_t *from,
> static void cpuset_change_task_nodemask(struct task_struct *tsk,
> nodemask_t *newmems)
> {
> + bool masks_disjoint = !nodes_intersects(*newmems, tsk->mems_allowed);
> +
> repeat:
> /*
> * Allow tasks that have access to memory reserves because they have
> @@ -963,7 +965,6 @@ repeat:
> nodes_or(tsk->mems_allowed, tsk->mems_allowed, *newmems);
> mpol_rebind_task(tsk, newmems, MPOL_REBIND_STEP1);
>
> -
> /*
> * ensure checking ->mems_allowed_change_disable after setting all new
> * allowed nodes.
> @@ -980,9 +981,11 @@ repeat:
>
> /*
> * Allocation of memory is very fast, we needn't sleep when waiting
> - * for the read-side.
> + * for the read-side. No wait is necessary, however, if at least one
> + * node remains unchanged.
> */
> - while (ACCESS_ONCE(tsk->mems_allowed_change_disable)) {
> + while (masks_disjoint &&
> + ACCESS_ONCE(tsk->mems_allowed_change_disable)) {
> task_unlock(tsk);
> if (!task_curr(tsk))
> yield();
> --
> To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
> the body of a message to majordomo@xxxxxxxxxxxxxxx
> More majordomo info at http://vger.kernel.org/majordomo-info.html
> Please read the FAQ at http://www.tux.org/lkml/
>

--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/