diff options
| author | Tejun Heo <tj@kernel.org> | 2017-07-18 18:41:52 -0400 |
|---|---|---|
| committer | Moyster <oysterized@gmail.com> | 2017-11-06 15:24:02 +0100 |
| commit | df88439f36936cc2c5a1eb80d13898b6bfb3cef7 (patch) | |
| tree | 4df075976043676c92058bfe37db36fbc96cdfab /drivers/target | |
| parent | d72966e1a992092d1907c2f8867b404d5afb486c (diff) | |
workqueue: restore WQ_UNBOUND/max_active==1 to be ordered
commit 5c0338c68706be53b3dc472e4308961c36e4ece1 upstream.
The combination of WQ_UNBOUND and max_active == 1 used to imply
ordered execution. After NUMA affinity 4c16bd327c74 ("workqueue:
implement NUMA affinity for unbound workqueues"), this is no longer
true due to per-node worker pools.
While the right way to create an ordered workqueue is
alloc_ordered_workqueue(), the documentation has been misleading for a
long time and people do use WQ_UNBOUND and max_active == 1 for ordered
workqueues which can lead to subtle bugs which are very difficult to
trigger.
It's unlikely that we'd see noticeable performance impact by enforcing
ordering on WQ_UNBOUND / max_active == 1 workqueues. Let's
automatically set __WQ_ORDERED for those workqueues.
Signed-off-by: Tejun Heo <tj@kernel.org>
Reported-by: Christoph Hellwig <hch@infradead.org>
Reported-by: Alexei Potashnik <alexei@purestorage.com>
Fixes: 4c16bd327c74 ("workqueue: implement NUMA affinity for unbound workqueues")
Cc: stable@vger.kernel.org # v3.10+
Signed-off-by: Willy Tarreau <w@1wt.eu>
Conflicts:
kernel/workqueue.c
Diffstat (limited to 'drivers/target')
0 files changed, 0 insertions, 0 deletions
