Re: [PATCH 1/3] wait-simple: Introduce the simple waitqueueimplementation
From: Peter Zijlstra
Date:  Thu Dec 12 2013 - 06:18:35 EST
On Wed, Dec 11, 2013 at 08:06:37PM -0500, Paul Gortmaker wrote:
> +/*
> + * Event API
> + */
> +#define __swait_event(wq, condition)					\
> +do {									\
> +	DEFINE_SWAITER(__wait);						\
> +									\
> +	for (;;) {							\
> +		swait_prepare(&wq, &__wait, TASK_UNINTERRUPTIBLE);	\
> +		if (condition)						\
> +			break;						\
> +		schedule();						\
> +	}								\
> +	swait_finish(&wq, &__wait);					\
> +} while (0)
> +
> +#define __swait_event_interruptible(wq, condition, ret)			\
> +do {									\
> +	DEFINE_SWAITER(__wait);						\
> +									\
> +	for (;;) {							\
> +		swait_prepare(&wq, &__wait, TASK_INTERRUPTIBLE);	\
> +		if (condition)						\
> +			break;						\
> +		if (signal_pending(current)) {				\
> +			ret = -ERESTARTSYS;				\
> +			break;						\
> +		}							\
> +		schedule();						\
> +	}								\
> +	swait_finish(&wq, &__wait);					\
> +} while (0)
> +
> +#define __swait_event_interruptible_timeout(wq, condition, ret)		\
> +do {									\
> +	DEFINE_SWAITER(__wait);						\
> +									\
> +	for (;;) {							\
> +		swait_prepare(&wq, &__wait, TASK_INTERRUPTIBLE);	\
> +		if (condition)						\
> +			break;						\
> +		if (signal_pending(current)) {				\
> +			ret = -ERESTARTSYS;				\
> +			break;						\
> +		}							\
> +		ret = schedule_timeout(ret);				\
> +		if (!ret)						\
> +			break;						\
> +	}								\
> +	swait_finish(&wq, &__wait);					\
> +} while (0)
Urgh, please have a look at ___wait_event() we just killed all the
pointless replication for the normal waitqueues, please don't add more
of it.
> +unsigned int
> +__swake_up_locked(struct swait_queue_head *head, unsigned int state,
> +		  unsigned int num)
> +{
> +	struct swaiter *curr, *next;
> +	int woken = 0;
> +
> +	list_for_each_entry_safe(curr, next, &head->task_list, node) {
> +		if (wake_up_state(curr->task, state)) {
> +			__swait_dequeue(curr);
> +			/*
> +			 * The waiting task can free the waiter as
> +			 * soon as curr->task = NULL is written,
> +			 * without taking any locks. A memory barrier
> +			 * is required here to prevent the following
> +			 * store to curr->task from getting ahead of
> +			 * the dequeue operation.
> +			 */
> +			smp_wmb();
> +			curr->task = NULL;
> +			if (++woken == num)
> +				break;
> +		}
> +	}
> +	return woken;
> +}
> +
> +unsigned int
> +__swake_up(struct swait_queue_head *head, unsigned int state, unsigned int num)
> +{
> +	unsigned long flags;
> +	int woken;
> +
> +	if (!swaitqueue_active(head))
> +		return 0;
> +
> +	raw_spin_lock_irqsave(&head->lock, flags);
> +	woken = __swake_up_locked(head, state, num);
> +	raw_spin_unlock_irqrestore(&head->lock, flags);
> +	return woken;
> +}
> +EXPORT_SYMBOL(__swake_up);
Urgh, fail. Do not put unbounded loops in raw_spin_lock.
I think I posted a patch a while back to cure this.
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/