Re: Regression with FALLOC_FL_PUNCH_HOLE in 3.5-rc kernel

From: Zdenek Kabelac
Date: Sun Jul 01 2012 - 18:03:16 EST


Dne 1.7.2012 22:10, Zdenek Kabelac napsal(a):
Dne 1.7.2012 20:45, Hugh Dickins napsal(a):
On Sun, 1 Jul 2012, Zdenek Kabelac wrote:
Dne 1.7.2012 01:10, Hugh Dickins napsal(a):
On Sat, 30 Jun 2012, Zdenek Kabelac wrote:
Dne 30.6.2012 21:55, Hugh Dickins napsal(a):
On Sat, 30 Jun 2012, Zdenek Kabelac wrote:

When I've used 3.5-rc kernels - I've noticed kernel deadlocks.
Ooops log included. After some experimenting - reliable way to hit
this
oops
is to run lvm test suite for 10 minutes. Since 3.5 merge window does
not
included anything related to this oops I've went for bisect.

Thanks a lot for reporting, and going to such effort to find
a reproducible testcase that you could bisect on.


Game result is commit: 3f31d07571eeea18a7d34db9af21d2285b807a17

mm/fs: route MADV_REMOVE to FALLOC_FL_PUNCH_HOLE

But this leaves me very puzzled.

Is the "lvm test suite" what I find at
git.fedorahosted.org/git/lvm2.git
under tests/ ?

Yes - that's it -

make
as root:
cd test
make check_local

(inside test subdirectory should be enough, if not - just report any
problem)

If you have something else running at the same time, which happens to
use
madvise(,,MADV_REMOVE) on a filesystem which the commit above now
enables
it on (I guess ext4 from the =y in your config), then I suppose we
should
start searching for improper memory freeing or scribbling in its
holepunch
support: something that might be corrupting the dm_region in your oops.

What the test is doing - it creates file in LVM_TEST_DIR (default is
/tmp)

I ran "LVM_TEST_DIR=/tmp make check_local":
without that it appeared to be using a subdirectory made under test/.

And being a year or two out of date in my userspace, and unfamiliar with
the syntax and whereabouts of lvm.conf, it was easiest for me to hack
lib/config/defaults.h to #define DEFAULT_ISSUE_DISCARDS 1
after I spotted a warning message about issue_discards.

and using loop device to simulate device (small size - it should fit
bellow
200MB)

Within this file second layer through virtual DM devices is created and
simulates various numbers of PV devices to play with.

This sounds much easier to set up than I was expecting:
thanks for the info, I'll try it later on today.

Sorry, I never reached it yesterday, but arrived there this morning.



So since everything now support TRIM - such operations should be passed
down to the backend file - which probably triggers the path.

What filesystem do you have for /tmp?

From your later remarks, I inferred tmpfs.


If tmpfs, then it will make much more sense if we assume your bisection
endpoint was off by one. Your bisection log was not quite complete;
and even if it did appear to converge on the commit you cite, you might
have got (un)lucky when testing the commit before it, and concluded
"good" when more attempts would have said "bad".

The commit before, 83e4fa9c16e4af7122e31be3eca5d57881d236fe
"tmpfs: support fallocate FALLOC_FL_PUNCH_HOLE", would be a
much more likely first bad commit if your /tmp is on tmpfs:
that does indeed wire up loop to pass TRIM down to tmpfs by
fallocate - that indeed played a part in my own testing.

Whereas if your /tmp is on ext4, loop has been passing TRIM down
with fallocate since v3.0. And whichever, madvise(,,MADV_REMOVE)
should be completely irrelevant.

While I've been aware of the fact that tmpfs was enhanced with trim support -
I've not tried to run on real ext4 filesystem since for my tests I'm using
tmpfs for quite some time to safe rewrites of SSD :)

So now I've checked with real ext4 - and the bug is there as well
so I've went back - it crashes on 3.4, 3.3 and 3.2 as well.

3.1 is the first kernel which does survive (checked 5 repeated runs)

Very useful research, thank you.


And you are correct, the first commit which causes crash really is
83e4fa9c16e4af when I use tmpfs as backend storage - the problem why I've
missed to properly identify this commit in my bisect is that crash usually
happens on the second pass of the lvm test suite 'make check_local' execution
- and I've been running test just once. To be sure I've run 5 run passes on
3.4.0-08568-gec9516f - which is OK, but 3.4.0-08569-g83e4fa9 is crashing
usually on second run, with commit 3f31d07571e the crash always happens in
the first pass.

I've also checked some rawhide kernel vmlinuz-3.5.0-0.rc2.git0.1.fc18.x86_64
and it's crashing as well - so it's probably not uniqueness of my config.

So is there any primary suspect in 3.2 which is worth to check - or I need
another day to play another bisect game ?

No need for a further bisect: 3.2 is when the disard/trim/fallocate
support went into drivers/block/loop.c, so these tests would be unable
to show if DM was right or wrong before then.

Well I've played meanwhile the game with minimized kernel config
and the outcome is:

last working kernel is: 3.1.0-rc1-00008-g548ef6c
first broken: 3.1.0-rc1-00009-gdfaa2ef

dfaa2ef68e80c378e610e3c8c536f1c239e8d3ef
loop: add discard support for loop devices

I don't have Fedora Rawhide on, but after hacking ISSUE_DISCARDS
I did quickly crash around where you expected; though in my case
it was in dm_rh_dec() called from mirror_end_io().

Change of Issue Discards option in lvm.conf is not needed.
I'm able to get these oopses with this setting turned off.

But I've not taken it any further than that. You've shown that it's
as much a problem with ext4 as with tmpfs, and has been a problem
ever since these tests' use of discard reached DM.

I think it's fair to assume that there's something wrong with DM's
handling of REQ_DISCARD. (Perhaps it was all correct when written,
but I think there has been a series of modifications to REQ_DISCARD
handling in the block layer, it's been rather troublesome.)


So does anyone has some idea what should be checked next ?


I've enabled more kernel debugging tools - and this came out when running test suite:

Unsure if this could be source of troubles?

kernel BUG at arch/x86/mm/physaddr.c:18!
invalid opcode: 0000 [#1] SMP DEBUG_PAGEALLOC
CPU 1
Modules linked in: dm_mirror dm_region_hash dm_log aes_generic tun bridge stp llc psmouse serio_raw cryptomgr aead arc4 crypto_algapi sr_mod cdrom iwl3945 iwl_legacy mac80211 cfg80211 crypto e1000e evdev loop dm_mod autofs4

Pid: 8, comm: kworker/1:0 Not tainted 3.1.0-rc1-00009-gdfaa2ef #30 LENOVO 6464CTO/6464CTO
RIP: 0010:[<ffffffff81021f00>] [<ffffffff81021f00>] __phys_addr+0x60/0x70
RSP: 0018:ffff880136e41aa0 EFLAGS: 00010213
RAX: ffff87ffffffffff RBX: ffff880136e41ba0 RCX: ffff880136e41b3c
RDX: ffff880136e41b48 RSI: ffff880136e41b40 RDI: 0000000000000000
RBP: ffff880136e41aa0 R08: 0000000000000000 R09: 0000000000000000
R10: 0000000000000000 R11: 0000000000000001 R12: ffff880136e41b40
R13: ffff880136e41b48 R14: ffff880136e41b3c R15: ffff88011e9bec40
FS: 0000000000000000(0000) GS:ffff880137400000(0000) knlGS:0000000000000000
CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b
CR2: 00007f34760df0c0 CR3: 0000000001604000 CR4: 00000000000006a0
DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7: 0000000000000400
Process kworker/1:0 (pid: 8, threadinfo ffff880136e40000, task ffff880136e3e090)
Stack:
ffff880136e41ad0 ffffffffa0013f3d 0000000000000800 ffff880136e41ba0
ffff88011eae6800 ffff880136e41c10 ffff880136e41b80 ffffffffa001437e
00000000239a96d0 0000000000000000 0000000000000000 0000000000000000
Call Trace:
[<ffffffffa0013f3d>] km_get_page+0x2d/0xa0 [dm_mod]
[<ffffffffa001437e>] dispatch_io+0x1ae/0x250 [dm_mod]
[<ffffffffa0013e90>] ? vm_next_page+0x20/0x20 [dm_mod]
[<ffffffffa0013f10>] ? vm_get_page+0x60/0x60 [dm_mod]
[<ffffffffa013dab0>] ? mirror_flush+0x120/0x120 [dm_mirror]
[<ffffffffa0014627>] dm_io+0xe7/0x1f0 [dm_mod]
[<ffffffffa0013f10>] ? vm_get_page+0x60/0x60 [dm_mod]
[<ffffffffa0013e90>] ? vm_next_page+0x20/0x20 [dm_mod]
[<ffffffffa013e204>] do_writes+0x324/0x650 [dm_mirror]
[<ffffffffa013e085>] ? do_writes+0x1a5/0x650 [dm_mirror]
[<ffffffffa013dab0>] ? mirror_flush+0x120/0x120 [dm_mirror]
[<ffffffffa013e9f4>] do_mirror+0x104/0x230 [dm_mirror]
[<ffffffff81050177>] process_one_work+0x197/0x460
[<ffffffff81050116>] ? process_one_work+0x136/0x460
[<ffffffffa013e8f0>] ? do_reads+0x180/0x180 [dm_mirror]
[<ffffffff81050986>] worker_thread+0x126/0x2d0
[<ffffffff81050860>] ? manage_workers.isra.28+0x1f0/0x1f0
[<ffffffff81056066>] kthread+0x96/0xa0
[<ffffffff8133b614>] kernel_thread_helper+0x4/0x10
[<ffffffff8133405d>] ? retint_restore_args+0xe/0xe
[<ffffffff81055fd0>] ? __init_kthread_worker+0x70/0x70
[<ffffffff8133b610>] ? gs_change+0xb/0xb
Code: ff ff 87 ff ff 48 39 c7 76 21 0f b6 0d a3 61 62 00 48 b8 00 00 00 00 00 78 00 00 48 01 f8 48 89 c2 48 d3 ea 48 85 d2 74 d0 0f 0b <0f> 0b 0f 0b 66 2e 0f 1f 84 00 00 00 00 00 66 90 83 f9 01 49 ba
RIP [<ffffffff81021f00>] __phys_addr+0x60/0x70
RSP <ffff880136e41aa0>
---[ end trace 040dfd9d68fccabc ]---



Zdenek
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/