[Bug 1001019] Re: Mdadm crash on raid5 reshape
James Lee
1001019 at bugs.launchpad.net
Sun Mar 17 16:32:52 UTC 2013
I've seen this bug too - I was also doing a reshape of a RAID5 array
(but without changing chunk size - just adding a new drive).
I'm also seeing this intermittently only. One thing that might be
relevant is that I'm reshaping several arrays in succession through a
script (doing them in series, but with no waiting between reshapes).
I've attached my dmesg output - similar error and stack to the above.
Let me know if there are any other diags which would help here, I can
hopefully repro this without too much trouble.
** Attachment added: "Sample dmesg output showing a repro of the bug"
https://bugs.launchpad.net/ubuntu/+source/mdadm/+bug/1001019/+attachment/3579754/+files/dmesg.txt
--
You received this bug notification because you are a member of Ubuntu
Foundations Bugs, which is subscribed to mdadm in Ubuntu.
https://bugs.launchpad.net/bugs/1001019
Title:
Mdadm crash on raid5 reshape
Status in “mdadm” package in Ubuntu:
Confirmed
Bug description:
Ubuntu 12.04 LTS
mdadm 3.2.3-2ubuntu1 0
I try to grow mdadm raid5 array with changing device number and chunk size on same time:
mdadm --grow /dev/md2 --raid-devices=4 --chunk=512 --backup-file=/root/md2_backup.img
After this /proc/mdstat don't respond and I can't gen any info about
raid status
May 17 23:27:17 protoss kernel: [40835.240145] RAID conf printout:
May 17 23:27:17 protoss kernel: [40835.240153] --- level:5 rd:4 wd:4
May 17 23:27:17 protoss kernel: [40835.240160] disk 0, o:1, dev:sdc1
May 17 23:27:17 protoss kernel: [40835.240166] disk 1, o:1, dev:sdd1
May 17 23:27:17 protoss kernel: [40835.240171] disk 2, o:1, dev:sdf1
May 17 23:27:17 protoss kernel: [40835.240176] disk 3, o:1, dev:sdg1
May 17 23:27:17 protoss kernel: [40835.240332] md: reshape of RAID array md2
May 17 23:27:17 protoss kernel: [40835.240342] md: minimum _guaranteed_ speed: 1000 KB/sec/disk.
May 17 23:27:17 protoss kernel: [40835.240348] md: using maximum available idle IO bandwidth (but not more than 200000 KB/sec) for reshape.
May 17 23:27:17 protoss kernel: [40835.240365] md: using 128k window, over a total of 1953513472k.
May 17 23:27:17 protoss kernel: [40835.507761] md: md_do_sync() got signal ... exiting
May 17 23:30:42 protoss kernel: [41040.652048] INFO: task md2_raid5:259 blocked for more than 120 seconds.
May 17 23:30:42 protoss kernel: [41040.652055] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
May 17 23:30:42 protoss kernel: [41040.652062] md2_raid5 D ffffffff81806240 0 259 2 0x00000000
May 17 23:30:42 protoss kernel: [41040.652074] ffff8801336a1c40 0000000000000046 ffffea0001d7f180 ffffffff7fffffff
May 17 23:30:42 protoss kernel: [41040.652086] ffff8801336a1fd8 ffff8801336a1fd8 ffff8801336a1fd8 0000000000013780
May 17 23:30:42 protoss kernel: [41040.652097] ffff880138b796f0 ffff880133ba2de0 ffff8801336a1c30 ffff880133ba2de0
May 17 23:30:42 protoss kernel: [41040.652108] Call Trace:
May 17 23:30:42 protoss kernel: [41040.652123] [<ffffffff8165a55f>] schedule+0x3f/0x60
May 17 23:30:42 protoss kernel: [41040.652157] [<ffffffffa00a3d2e>] resize_stripes+0x51e/0x590 [raid456]
May 17 23:30:42 protoss kernel: [41040.652167] [<ffffffff81056c9c>] ? update_shares+0xcc/0x100
May 17 23:30:42 protoss kernel: [41040.652176] [<ffffffff8105f990>] ? try_to_wake_up+0x200/0x200
May 17 23:30:42 protoss kernel: [41040.652192] [<ffffffffa00a3e1f>] check_reshape+0x7f/0xd0 [raid456]
May 17 23:38:42 protoss kernel: [41520.652310] [<ffffffff8108a3a0>] ? flush_kthread_worker+0xa0/0xa0
May 17 23:38:42 protoss kernel: [41520.652319] [<ffffffff81666bf0>] ? gs_change+0x13/0x13
May 17 23:38:42 protoss kernel: [41520.652357] INFO: task lvm:25938 blocked for more than 120 seconds.
May 17 23:38:42 protoss kernel: [41520.652362] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
May 17 23:38:42 protoss kernel: [41520.652367] lvm D ffffffff81806240 0 25938 25935 0x00000000
May 17 23:38:42 protoss kernel: [41520.652377] ffff88006c6b1ae8 0000000000000086 0000000000000000 ffff88006c6b1aa8
May 17 23:38:42 protoss kernel: [41520.652388] ffff88006c6b1fd8 ffff88006c6b1fd8 ffff88006c6b1fd8 0000000000013780
May 17 23:38:42 protoss kernel: [41520.652398] ffff880138b616f0 ffff88010c02dbc0 ffff88006c6b1ab8 ffff88013fc94040
May 17 23:38:42 protoss kernel: [41520.652409] Call Trace:
May 17 23:38:42 protoss kernel: [41520.652417] [<ffffffff8165a55f>] schedule+0x3f/0x60
May 17 23:38:42 protoss kernel: [41520.652424] [<ffffffff8165a60f>] io_schedule+0x8f/0xd0
May 17 23:38:42 protoss kernel: [41520.652434] [<ffffffff811b0e74>] dio_await_completion+0x54/0xd0
May 17 23:38:42 protoss kernel: [41520.652442] [<ffffffff811b3484>] __blockdev_direct_IO+0x954/0xd90
May 17 23:38:42 protoss kernel: [41520.652451] [<ffffffff811af6f0>] ? blkdev_get_block+0x80/0x80
May 17 23:38:42 protoss kernel: [41520.652461] [<ffffffff811af1b7>] blkdev_direct_IO+0x57/0x60
May 17 23:38:42 protoss kernel: [41520.652468] [<ffffffff811af6f0>] ? blkdev_get_block+0x80/0x80
May 17 23:38:42 protoss kernel: [41520.652478] [<ffffffff811196bb>] generic_file_aio_read+0x24b/0x280
May 17 23:38:42 protoss kernel: [41520.652488] [<ffffffff811874ac>] ? path_openat+0xfc/0x3f0
May 17 23:38:42 protoss kernel: [41520.652496] [<ffffffff81177452>] do_sync_read+0xd2/0x110
May 17 23:38:42 protoss kernel: [41520.652506] [<ffffffff8129cd03>] ? security_file_permission+0x93/0xb0
May 17 23:38:42 protoss kernel: [41520.652514] [<ffffffff811778d1>] ? rw_verify_area+0x61/0xf0
May 17 23:38:42 protoss kernel: [41520.652521] [<ffffffff81177db0>] vfs_read+0xb0/0x180
May 17 23:38:42 protoss kernel: [41520.652529] [<ffffffff81177eca>] sys_read+0x4a/0x90
May 17 23:38:42 protoss kernel: [41520.652537] [<ffffffff81664a82>] system_call_fastpath+0x16/0x1b
To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/mdadm/+bug/1001019/+subscriptions
More information about the foundations-bugs
mailing list