Re: [PATCH] mm/page_isolation: fix a deadlock with printk()

From: Qian Cai
Date: Fri Oct 04 2019 - 14:53:11 EST


On Fri, 2019-10-04 at 19:47 +0200, David Hildenbrand wrote:
> On 04.10.19 18:42, Qian Cai wrote:
> > It is unsafe to call printk() while zone->lock was held, i.e.,
> >
> > zone->lock --> console_sem
> >
> > because the console could always allocate some memory in different code
> > paths and form locking chains in an opposite order,
> >
> > console_sem --> * --> zone->lock
> >
> > As the result, it triggers lockdep splats like below and in [1]. It is
> > fine to take zone->lock after has_unmovable_pages() (which has
> > dump_stack()) in set_migratetype_isolate(). While at it, remove a
> > problematic printk() in __offline_isolated_pages() only for debugging as
> > well which will always disable lockdep on debug kernels.
> >
> > The problem is probably there forever, but neither many developers will
> > run memory offline with the lockdep enabled nor admins in the field are
> > lucky enough yet to hit a perfect timing which required to trigger a
> > real deadlock. In addition, there aren't many places that call printk()
> > while zone->lock was held.
> >
> > WARNING: possible circular locking dependency detected
> > ------------------------------------------------------
> > test.sh/1724 is trying to acquire lock:
> > 0000000052059ec0 (console_owner){-...}, at: console_unlock+0x
> > 01: 328/0xa30
> >
> > but task is already holding lock:
> > 000000006ffd89c8 (&(&zone->lock)->rlock){-.-.}, at: start_iso
> > 01: late_page_range+0x216/0x538
> >
> > which lock already depends on the new lock.
> >
> > the existing dependency chain (in reverse order) is:
> >
> > -> #2 (&(&zone->lock)->rlock){-.-.}:
> > lock_acquire+0x21a/0x468
> > _raw_spin_lock+0x54/0x68
> > get_page_from_freelist+0x8b6/0x2d28
> > __alloc_pages_nodemask+0x246/0x658
> > __get_free_pages+0x34/0x78
> > sclp_init+0x106/0x690
> > sclp_register+0x2e/0x248
> > sclp_rw_init+0x4a/0x70
> > sclp_console_init+0x4a/0x1b8
> > console_init+0x2c8/0x410
> > start_kernel+0x530/0x6a0
> > startup_continue+0x70/0xd0
> >
> > -> #1 (sclp_lock){-.-.}:
> > lock_acquire+0x21a/0x468
> > _raw_spin_lock_irqsave+0xcc/0xe8
> > sclp_add_request+0x34/0x308
> > sclp_conbuf_emit+0x100/0x138
> > sclp_console_write+0x96/0x3b8
> > console_unlock+0x6dc/0xa30
> > vprintk_emit+0x184/0x3c8
> > vprintk_default+0x44/0x50
> > printk+0xa8/0xc0
> > iommu_debugfs_setup+0xf2/0x108
> > iommu_init+0x6c/0x78
> > do_one_initcall+0x162/0x680
> > kernel_init_freeable+0x4e8/0x5a8
> > kernel_init+0x2a/0x188
> > ret_from_fork+0x30/0x34
> > kernel_thread_starter+0x0/0xc
> >
> > -> #0 (console_owner){-...}:
> > check_noncircular+0x338/0x3e0
> > __lock_acquire+0x1e66/0x2d88
> > lock_acquire+0x21a/0x468
> > console_unlock+0x3a6/0xa30
> > vprintk_emit+0x184/0x3c8
> > vprintk_default+0x44/0x50
> > printk+0xa8/0xc0
> > __dump_page+0x1dc/0x710
> > dump_page+0x2e/0x58
> > has_unmovable_pages+0x2e8/0x470
> > start_isolate_page_range+0x404/0x538
> > __offline_pages+0x22c/0x1338
> > memory_subsys_offline+0xa6/0xe8
> > device_offline+0xe6/0x118
> > state_store+0xf0/0x110
> > kernfs_fop_write+0x1bc/0x270
> > vfs_write+0xce/0x220
> > ksys_write+0xea/0x190
> > system_call+0xd8/0x2b4
> >
> > other info that might help us debug this:
> >
> > Chain exists of:
> > console_owner --> sclp_lock --> &(&zone->lock)->rlock
> >
> > Possible unsafe locking scenario:
> >
> > CPU0 CPU1
> > ---- ----
> > lock(&(&zone->lock)->rlock);
> > lock(sclp_lock);
> > lock(&(&zone->lock)->rlock);
> > lock(console_owner);
> >
> > *** DEADLOCK ***
> >
> > 9 locks held by test.sh/1724:
> > #0: 000000000e925408 (sb_writers#4){.+.+}, at: vfs_write+0x201:
> > #1: 0000000050aa4280 (&of->mutex){+.+.}, at: kernfs_fop_write:
> > #2: 0000000062e5c628 (kn->count#198){.+.+}, at: kernfs_fop_write
> > #3: 00000000523236a0 (device_hotplug_lock){+.+.}, at:
> > lock_device_hotplug_sysfs+0x30/0x80
> > #4: 0000000062e70990 (&dev->mutex){....}, at: device_offline
> > #5: 0000000051fd36b0 (cpu_hotplug_lock.rw_sem){++++}, at:
> > __offline_pages+0xec/0x1338
> > #6: 00000000521ca470 (mem_hotplug_lock.rw_sem){++++}, at:
> > percpu_down_write+0x38/0x210
> > #7: 000000006ffd89c8 (&(&zone->lock)->rlock){-.-.}, at:
> > start_isolate_page_range+0x216/0x538
> > #8: 000000005205a100 (console_lock){+.+.}, at: vprintk_emit
> >
> > stack backtrace:
> > Hardware name: IBM 2964 N96 400 (z/VM 6.4.0)
> > Call Trace:
> > ([<00000000512ae218>] show_stack+0x110/0x1b0)
> > [<0000000051b6d506>] dump_stack+0x126/0x178
> > [<00000000513a4b08>] check_noncircular+0x338/0x3e0
> > [<00000000513aaaf6>] __lock_acquire+0x1e66/0x2d88
> > [<00000000513a7e12>] lock_acquire+0x21a/0x468
> > [<00000000513bb2fe>] console_unlock+0x3a6/0xa30
> > [<00000000513bde2c>] vprintk_emit+0x184/0x3c8
> > [<00000000513be0b4>] vprintk_default+0x44/0x50
> > [<00000000513beb60>] printk+0xa8/0xc0
> > [<000000005158c364>] __dump_page+0x1dc/0x710
> > [<000000005158c8c6>] dump_page+0x2e/0x58
> > [<00000000515d87c8>] has_unmovable_pages+0x2e8/0x470
> > [<000000005167072c>] start_isolate_page_range+0x404/0x538
> > [<0000000051b96de4>] __offline_pages+0x22c/0x1338
> > [<0000000051908586>] memory_subsys_offline+0xa6/0xe8
> > [<00000000518e561e>] device_offline+0xe6/0x118
> > [<0000000051908170>] state_store+0xf0/0x110
> > [<0000000051796384>] kernfs_fop_write+0x1bc/0x270
> > [<000000005168972e>] vfs_write+0xce/0x220
> > [<0000000051689b9a>] ksys_write+0xea/0x190
> > [<0000000051ba9990>] system_call+0xd8/0x2b4
> > INFO: lockdep is turned off.
> >
> > [1] https://lore.kernel.org/lkml/1568817579.5576.172.camel@xxxxxx/
> >
> > Signed-off-by: Qian Cai <cai@xxxxxx>
> > ---
> > mm/page_alloc.c | 4 ----
> > mm/page_isolation.c | 10 +++++-----
> > 2 files changed, 5 insertions(+), 9 deletions(-)
> >
> > diff --git a/mm/page_alloc.c b/mm/page_alloc.c
> > index 15c2050c629b..232bbb1dc521 100644
> > --- a/mm/page_alloc.c
> > +++ b/mm/page_alloc.c
> > @@ -8588,10 +8588,6 @@ void zone_pcp_reset(struct zone *zone)
> > BUG_ON(!PageBuddy(page));
> > order = page_order(page);
> > offlined_pages += 1 << order;
> > -#ifdef CONFIG_DEBUG_VM
> > - pr_info("remove from free list %lx %d %lx\n",
> > - pfn, 1 << order, end_pfn);
> > -#endif
> > del_page_from_free_area(page, &zone->free_area[order]);
> > for (i = 0; i < (1 << order); i++)
> > SetPageReserved((page+i));
> > diff --git a/mm/page_isolation.c b/mm/page_isolation.c
> > index 89c19c0feadb..8682ccb5fbd1 100644
> > --- a/mm/page_isolation.c
> > +++ b/mm/page_isolation.c
> > @@ -25,8 +25,6 @@ static int set_migratetype_isolate(struct page *page, int migratetype, int isol_
> >
> > zone = page_zone(page);
> >
> > - spin_lock_irqsave(&zone->lock, flags);
> > -
> > /*
> > * We assume the caller intended to SET migrate type to isolate.
> > * If it is already set, then someone else must have raced and
> > @@ -74,16 +72,18 @@ static int set_migratetype_isolate(struct page *page, int migratetype, int isol_
> > int mt = get_pageblock_migratetype(page);
> >
> > set_pageblock_migratetype(page, MIGRATE_ISOLATE);
> > +
> > + spin_lock_irqsave(&zone->lock, flags);
>
> The migratetype has to be tested and set under lock, otherwise two
> clients could race. I don't like such severe locking changes just to
> make some printing we only need for debugging work.

Ah, it could use a different lock for that then rather than reuse zone->lock
which is kind of confusion to begin with.

>
> Can't we somehow return some information (page / cause) from
> has_unmovable_pages() and print from a save place instead?

Possible too, but using a different lock to protect migratetype looks simpler.

>
> To fix the BUG, I would much rather want to see all printing getting
> ripped out instead. That's easy to backort.

I am not even sure if this worth backporting at all as it is rather difficult to
hit in the field and probably exist there for many years without anyone noticed
it.

>
> We can then come back and think about how to log such stuff in order to
> debug it.
>