diff options
author | Tejun Heo <tj@kernel.org> | 2016-09-16 22:49:34 +0300 |
---|---|---|
committer | Tejun Heo <tj@kernel.org> | 2016-09-17 20:18:21 +0300 |
commit | eac0337af12b6a55f08c69429400d6530d602dff (patch) | |
tree | a154b887be6b2a848d2ca5a83c1efda5a8918c8b /mm/slab.c | |
parent | a81f80f3eb759de72d74d18e478873fc0575abc5 (diff) | |
download | linux-eac0337af12b6a55f08c69429400d6530d602dff.tar.xz |
slab, workqueue: remove keventd_up() usage
Now that workqueue can handle work item queueing from very early
during boot, there is no need to gate schedule_delayed_work_on() while
!keventd_up(). Remove it.
Signed-off-by: Tejun Heo <tj@kernel.org>
Cc: Christoph Lameter <cl@linux.com>
Cc: Pekka Enberg <penberg@kernel.org>
Cc: David Rientjes <rientjes@google.com>
Cc: Joonsoo Kim <iamjoonsoo.kim@lge.com>
Cc: Andrew Morton <akpm@linux-foundation.org>
Cc: linux-mm@kvack.org
Diffstat (limited to 'mm/slab.c')
-rw-r--r-- | mm/slab.c | 7 |
1 files changed, 1 insertions, 6 deletions
diff --git a/mm/slab.c b/mm/slab.c index b67271024135..dc69b6b625b1 100644 --- a/mm/slab.c +++ b/mm/slab.c @@ -550,12 +550,7 @@ static void start_cpu_timer(int cpu) { struct delayed_work *reap_work = &per_cpu(slab_reap_work, cpu); - /* - * When this gets called from do_initcalls via cpucache_init(), - * init_workqueues() has already run, so keventd will be setup - * at that time. - */ - if (keventd_up() && reap_work->work.func == NULL) { + if (reap_work->work.func == NULL) { init_reap_node(cpu); INIT_DEFERRABLE_WORK(reap_work, cache_reap); schedule_delayed_work_on(cpu, reap_work, |