View Issue Details

IDProjectCategoryView StatusLast Update
0015013CentOS-7kernelpublic2018-07-08 23:49
Reporterluf 
PriorityhighSeveritycrashReproducibilityalways
Status assignedResolutionopen 
Platformamd64 (x86_64)OSCentOS 7OS Version7.5
Product Version7.5.1804 
Target VersionFixed in Version 
Summary0015013: Kernel crash during boot
DescriptionI have an old AMD Athlon.(tm) 64 X2 Dual Core Processor 5200+ ,Speed: 2700
DDR800, Single Channel

The boot of kernel 3.10.0-862.6.3.el7.x86_64 ends up with:
[ 0.000000] tsc: Fast TSC calibration failed
[ 2.390016] ata2: softreset failed (device not ready)
[ 2.390083] ata1: softreset failed (device not ready)
[ 2.390145] ata3: softreset failed (device not ready)
[ 3.744083] kfd kfd: kgd2kfd_probe failed
[ 6.624895] BUG: unable to handle kernel NULL pointer dereference at 00000000000005ec
[ 6.625018] IP: [<ffffffffc0927ffb>] read_balance+0x1db/0x4e0 [raid1]
[ 6.625161] PGD 0
[ 6.625161] Oops: 0000 [#1] SMP
[ 6.625161] Modules linked in: powernow_k8 edac_mce_amd kvm_amd raid1 kvm dm_raid raid456 async_raid6_recov async_memcpy async_pq ppdev raid6_pq async_xor snd_hda_codec_hdmi xor async_tx snd_hda_codec_realtek irqbypass snd_hda_codec_generic snd_hda_intel k8temp pcspkr snd_hda_codec snd_seq snd_hda_core snd_seq_device snd_hwdep sp5100_tco snd_pcm i2c_piix4 sg parport_pc parport floppy(+) snd_timer snd soundcore shpchp nfsd nfs_acl lockd auth_rpcgss grace sunrpc ip_tables xfs libcrc32c sr_mod cdrom ata_generic pata_acpi sd_mod crc_t10dif crct10dif_generic crct10dif_common amdkfd amd_iommu_v2 radeon i2c_algo_bit drm_kms_helper syscopyarea sysfillrect uas sysimgblt fb_sys_fops ttm drm pata_atiixp firewire_ohci serio_raw firewire_core ahci crc_itu_t libahci libata r8169 usb_storage mii i2c_core dm_mirror dm_region_hash dm_log dm_mod
[ 6.625161] CPU: 0 PID: 584 Comm: systemd-udevd Not tainted 3.10.0-862.6.3.el7.x86_64 #1
[ 6.625161] Hardware name: MSI MS-7280/MS-7280, BIOS V1.8 11/02/2007
[ 6.625161] task: ffff8bfef594bf40 ti: ffff8bff736dc000 task.ti: ffff8bff736dc000
[ 6.625161] RIP: 0010:[<ffffffffc0927ffb>] [<ffffffffc0927ffb>] read_balance+0x1db/0x4e0 [raid1]
[ 6.627458] RSP: 0018:ffff8bff736df660 EFLAGS: 00010246
[ 6.627458] RAX: 0000000000000000 RBX: 0000000000000004 RCX: 0000000000000002
[ 6.627458] RDX: 0000000000000000 RSI: ffff8dff732daf58 RDI: 0000000000000001
[ 6.627458] RBP: ffff8bff736df6e8 R08: 0000000000000004 R09: 0000000000000008
[ 6.627458] R10: 0000000013ffff80 R11: 0000000000000004 R12: 0000000000000000
[ 6.627458] R13: 0000000000000060 R14: ffff8bff7b2e5780 R15: ffff8bff7ae0c900
[ 6.627458] FS: 00007f4947ff88c0(0000) GS: ffff8bff7fa00000(0000) knlGS:0000000000000000
[ 6.627458] CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033
[ 6.627458] CR2: 00000000000005ec CR3: 00000000b36ae000 CR4: 00000000000007f0
[ 6.627458] Call Trace:
[ 6.627458] [<ffffffffc0929024>] ? r1bio_pool_alloc+0x24/0x30 [raid1]
[ 6.627458] [<ffffffffb0f97165>] ? mempool_alloc_slab+0x15/0x20
[ 6.627458] [<ffffffffc0929c33>] raid1_make_request|0x193/0x500 [raid1]
[ 6.627458] [<ffffffffb0fb5cd8>] ? zone_statisticts+0x88/0xa0
[ 6.627458] [<ffffffffb13486b0>] md_handle_request+0xd0/0x150
[ 6.627458] [<ffffffffc0780ba8>] raid_map+0x28/0x40 [dm_raid]
[ 6.627458] [<ffffffffc0036136>] __map_bio+0x96/0x190 [dm_mod]
[ 6.627458] [<ffffffffc0034680>] ? queue_io+0x80/0x80 [dm_mod]
[ 6.627458] [<ffffffffc00364437>] __clone_and_map_data_bio+0x177/0x280 [dm_mod]
[ 6.627458] [<ffffffffc0036811>] __split_and_process_bio+0x2d1/0x520 [dm_mod]
[ 6.627458] [<ffffffffc0036d7d>] dm_make_request+0x11d/0x1a0 [dm_mod]
[ 6.627458] [<ffffffffb111b0fb>] generic_make_request+0x10b/0x320
[ 6.627458] [<ffffffffb111b380>] submit_bio+0x70/0x150
[ 6.627458] [<ffffffffb0f94de4>] ? __add_to_page_cache_locked+0xe4/0x190
[ 6.627458] [<ffffffffb105ed0a>] mpage_bio_submit+0x2a/0x40
[ 6.627458] [<ffffffffb105f73f>] mpage_readpages+0x13f/0x170
[ 6.627458] [<ffffffffb1059250>] ? I_BDEV+0x10/0x10
[ 6.627458] [<ffffffffb1059250>] ? I_BDEV+0x10/0x10

while booting kernel up to 3.10.0-862.3.3.el7.x86_64 works without any issue.
The first 4 lines are common (tsc - kfd) but no oops.

Sometimes the output ends with "Hardware name" line.
Steps To ReproduceInstall new kernel 3.10.0-862.6.3.el7.x86_64 and try to boot it.
Tagskerneloops, regression, upstream
abrt_hash
URL

Activities

toracat

toracat

2018-07-07 07:01

manager   ~0032188

Last edited: 2018-07-07 07:25

View 2 revisions

This is a known bug and is being tracked upstream:

https://access.redhat.com/solutions/3520511
"Kernel panic on systems with MD Raid devices after updating to 3.10.0-862.6.3.el7 with RIP read_balance"

luf

luf

2018-07-07 07:11

reporter   ~0032189

Cool. Thanks! I wasn't able to find any info about it.
BTW any idea what is the solution? I'm not RedHat customer so I have no access to the solution.
toracat

toracat

2018-07-07 07:24

manager   ~0032190

Last edited: 2018-07-07 15:42

View 2 revisions

No solution at the moment. You'd need to stay with kernel-3.10.0-862.3.3.el7.

Or you can test-install a newer kernel such as ELRepo's kernel-lt (currently at 4.4.139) or kernel-ml (currently at 4.17.4) and see if the issue has been resolved there.

[EDIT] No bother. From what I've gathered, the issue is RH specific. Most likely upstream kernels (kernel.org) do not have this problem.

tru

tru

2018-07-07 10:44

administrator   ~0032192

thanks @toracat
imho, major upstream regression issue linking to a non public https://bugzilla.redhat.com/show_bug.cgi?id=1581845.

let's wait for a new kernel release
toracat

toracat

2018-07-07 17:10

manager   ~0032195

A couple of potential patches to fix the issue have been submitted by @pgreco. Centosplus kernels that include the patch (either v1 or v2) are available here:

Patch version1: https://people.centos.org/toracat/kernel/7/plus/bug15013v1/

Patch version2: https://people.centos.org/toracat/kernel/7/plus/bug15013v2/

Please test if you can.
tru

tru

2018-07-08 20:42

administrator   ~0032202

FYI: I have tried to reproduced the issue with a plain kickstart from 7.5.1804 GA (general availabity, ie without updates) and 7.5.1804+updates on a virtuabox hosted on my laptop also running CentOS-7... so far without success.

 https://github.com/truatpasteurdotfr/ks/blob/master/c7-GA-raid1.cfg and https://github.com/truatpasteurdotfr/ks/blob/master/c7-raid1.cfg. (single / as /dev/md0 for /dev/sd[ab]1, no swap, no separated /boot)

GA kernel: kernel-3.10.0-862.el7.x86_64 installed then `yum -y update && reboot` -> no issue
GA+updates : kernel-3.10.0-862.6.3.el7.x86_64 installed and booted without issue
tru

tru

2018-07-08 20:46

administrator   ~0032203

@luf: could you share your disk partitionning ? madm raid1 for /boot and / on plain partitions?; or do you have lvm over mdadm raid1 ?
tru

tru

2018-07-08 21:24

administrator   ~0032204

as reported on the forum, /boot and / on plain partition do NOT reproduce the issue on my test box either.
tru

tru

2018-07-08 22:04

administrator   ~0032205

Last edited: 2018-07-08 22:57

View 3 revisions

I could NOT yet reproduce the issue:
1) /boot on /dev/md0 (sd[ab]1) + / on /dev/md1 (sd[ab]2)
2) /boot on /dev/md0 (sd[ab]1) + / on /dev/md1vg/root with the VG md1vg on /dev/md1 (sd[ab]2)

maybe one need a mirrored LV to hit the issue? https://www.centos.org/forums/viewtopic.php?f=48&t=67589&start=10#p283912
' lvcreate -m 1 -i 1 --size=100M --name=v0 vg'

luf

luf

2018-07-08 22:25

reporter   ~0032206

No centosplus patched kernel (neither v1 nor v2) helped.

It's weird I have no md device. /boot xfs plain partition, system on lvm (no mirror/raid used inside lvm), some data on another lvm (seems using raid1).
I'm sorry I setup this several years ago :) and installed tens of other machines since that ...

I'm trying to track the problem down (not mounting data, ...). Hold on.
Definitely the boot process freeze (oom) after initramfs (mounted swap, var, tmp, ...).
toracat

toracat

2018-07-08 23:13

manager   ~0032207

@luf

Thanks for testing the patches. It was unfortunate that neither worked.

According to what I have gathered, RH now has a patch that fixes the issue and it will be in RH 7.6. But they may apply it to the 7.5 kernel as well. I don't know how long it takes, though.
tru

tru

2018-07-08 23:29

administrator   ~0032209

reproducer found on 3.10.0-862.6.3.el7.x86_64 just do the lvcreate
```
[root@c7-raid1 ~]# lvcreate -m 1 -i 1 --size=100M --name=v0 vg
[ 73.000849] async_tx: api initialized (async)
[ 73.028813] xor: automatically using best checksumming function:
[ 73.105329] avx : 210756.000 MB/sec
[ 73.165622] raid6: sse2x1 gen() 11984 MB/s
[ 73.225962] raid6: sse2x2 gen() 16699 MB/s
[ 73.276151] raid6: sse2x4 gen() 16898 MB/s
[ 73.285823] raid6: using algorithm sse2x4 gen() (16898 MB/s)
[ 73.286786] raid6: using ssse3x2 recovery algorithm
[ 73.310946] device-mapper: raid: Loading target version 1.13.2
[ 73.455678] device-mapper: raid: Superblocks created for new raid set
[ 73.599486] md/raid1:mdX: not clean -- starting background reconstruction
[ 73.609407] md/raid1:mdX: active with 2 out of 2 mirrors
[ 73.651862] mdX: bitmap file is out of date, doing full recovery
[ 73.695113] md: resync of RAID array mdX
[ 73.763171] BUG: unable to handle kernel NULL pointer dereference at 00000000000005ec
[ 73.818871] IP: [<ffffffffc0306ffb>] read_balance+0x1db/0x4e0 [raid1]
[ 73.859260] PGD 0
[ 73.898958] Oops: 0000 [#1] SMP
[ 73.937908] Modules linked in: dm_raid raid456 async_raid6_recov async_memcpy async_pq raid6_pq async_xor xor async_tx ip6t_rpfilter ipt_REJECT nf_reject_ipv4 ip6t_REJECT nf_reject_ipv6 xt_conntrack ip_set nfnetlink ebtable_nat ebtable_broute bridge stp llc ip6table_nat nf_conntrack_ipv6 nf_defrag_ipv6 nf_nat_ipv6 ip6table_mangle ip6table_security ip6table_raw iptable_nat nf_conntrack_ipv4 nf_defrag_ipv4 nf_nat_ipv4 nf_nat nf_conntrack iptable_mangle iptable_security iptable_raw ebtable_filter ebtables ip6table_filter ip6_tables iptable_filter ext4 mbcache jbd2 intel_powerclamp iosf_mbi crc32_pclmul ghash_clmulni_intel ppdev aesni_intel snd_intel8x0 lrw gf128mul glue_helper snd_ac97_codec ac97_bus ablk_helper cryptd snd_seq snd_seq_device snd_pcm pcspkr snd_timer sg i2c_piix4 snd i2c_core soundcore parport_pc parport video ip_tables xfs libcrc32c raid1 sd_mod crc_t10dif crct10dif_generic ahci libahci virtio_net libata crct10dif_pclmul virtio_pci crct10dif_common crc32c_intel serio_raw virtio_ring virtio dm_mirror dm_region_hash dm_log dm_mod
[ 74.093312] CPU: 0 PID: 1496 Comm: lvcreate Not tainted 3.10.0-862.6.3.el7.x86_64 #1
[ 74.094631] Hardware name: innotek GmbH VirtualBox/VirtualBox, BIOS VirtualBox 12/01/2006
[ 74.096187] task: ffffa008fb890000 ti: ffffa008fa79c000 task.ti: ffffa008fa79c000
[ 74.097445] RIP: 0010:[<ffffffffc0306ffb>] [<ffffffffc0306ffb>] read_balance+0x1db/0x4e0 [raid1]
[ 74.098957] RSP: 0018:ffffa008fa79f660 EFLAGS: 00010246
[ 74.133055] RAX: 0000000000000000 RBX: 0000000000000000 RCX: 0000000000000002
[ 74.136868] RDX: 0000000000000000 RSI: ffffa008b698f1a0 RDI: 0000000000000000
[ 74.138706] RBP: ffffa008fa79f6e8 R08: 0000000000000001 R09: 0000000000000008
[ 74.140531] R10: 0000000000031f80 R11: 0000000000000000 R12: ffffa008f9e1b680
[ 74.142243] R13: ffffa008fbb03000 R14: ffffa008b5753080 R15: ffffa008faa0cb00
[ 74.143959] FS: 00007f9d43506880(0000) GS:ffffa008ffc00000(0000) knlGS:0000000000000000
[ 74.145902] CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033
[ 74.147309] CR2: 00000000000005ec CR3: 0000000079f50000 CR4: 00000000000606f0
[ 74.148459] DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
[ 74.149609] DR3: 0000000000000000 DR6: 00000000fffe0ff0 DR7: 0000000000000400
[ 74.180865] Call Trace:
[ 74.181311] [<ffffffffc0308024>] ? r1bio_pool_alloc+0x24/0x30 [raid1]
[ 74.191339] [<ffffffffc0308c33>] raid1_make_request+0x193/0x500 [raid1]
[ 74.221139] [<ffffffff8b15ee1a>] ? sg_init_table+0x1a/0x40
[ 74.222028] [<ffffffff8b168e9d>] ? list_del+0xd/0x30
[ 74.222786] [<ffffffff8b3486b0>] md_handle_request+0xd0/0x150
[ 74.223641] [<ffffffffc076dba8>] raid_map+0x28/0x40 [dm_raid]
[ 74.233089] [<ffffffffc020a136>] __map_bio+0x96/0x190 [dm_mod]
[ 74.270908] [<ffffffffc0208680>] ? queue_io+0x80/0x80 [dm_mod]
[ 74.281231] [<ffffffffc020a437>] __clone_and_map_data_bio+0x177/0x280 [dm_mod]
[ 74.310904] [<ffffffffc020a811>] __split_and_process_bio+0x2d1/0x520 [dm_mod]
[ 74.401291] [<ffffffffc020ad7d>] dm_make_request+0x11d/0x1a0 [dm_mod]
[ 74.449083] [<ffffffff8b11b0fb>] generic_make_request+0x10b/0x320
[ 74.602480] [<ffffffff8b11b380>] submit_bio+0x70/0x150
[ 74.603105] [<ffffffff8af94dee>] ? __add_to_page_cache_locked+0xee/0x190
[ 74.604089] [<ffffffff8b05ed0a>] mpage_bio_submit+0x2a/0x40
[ 74.612825] [<ffffffff8b05f73f>] mpage_readpages+0x13f/0x170
[ 74.632590] [<ffffffff8b059250>] ? I_BDEV+0x10/0x10
[ 74.709406] [<ffffffff8b059250>] ? I_BDEV+0x10/0x10
[ 74.710156] [<ffffffff8b059b0d>] blkdev_readpages+0x1d/0x20
[ 74.710999] [<ffffffff8afa19ef>] __do_page_cache_readahead+0x1cf/0x260
[ 74.711986] [<ffffffff8afa1f89>] force_page_cache_readahead+0x99/0xe0
[ 74.712969] [<ffffffff8afa2067>] page_cache_sync_readahead+0x97/0xb0
[ 74.713930] [<ffffffff8af95c92>] generic_file_aio_read+0x2c2/0x790
[ 74.714864] [<ffffffff8b059f4c>] blkdev_aio_read+0x4c/0x70
[ 74.715701] [<ffffffff8b01a8d3>] do_sync_read+0x93/0xe0
[ 74.716496] [<ffffffff8b01b2ff>] vfs_read+0x9f/0x170
[ 74.717251] [<ffffffff8b01c1cf>] SyS_read+0x7f/0xf0
[ 74.717991] [<ffffffff8b5206e1>] ? system_call_after_swapgs+0xae/0x146
[ 74.718973] [<ffffffff8b520795>] system_call_fastpath+0x1c/0x21
[ 74.719923] [<ffffffff8b5206e1>] ? system_call_after_swapgs+0xae/0x146
[ 74.720908] Code: c1 4c 01 d0 48 89 42 10 49 8b 46 28 49 8b 17 48 8b 40 20 48 8b 92 50 03 00 00 a8 80 0f 85 e9 02 00 00 f6 c4 02 0f 85 d5 02 00 00 <8b> 82 ec 05 00 00 44 39 c8 4c 89 d7 44 89 5d bc 41 0f 4f c1 48
[ 74.724860] RIP [<ffffffffc0306ffb>] read_balance+0x1db/0x4e0 [raid1]
[ 74.743796] RSP <ffffa008fa79f660>
[ 74.744328] CR2: 00000000000005ec
[ 74.744865] ---[ end trace 9e10787b4b6f0f43 ]---
[ 74.745567] Kernel panic - not syncing: Fatal exception
[ 74.746366] Kernel Offset: 0x9e00000 from 0xffffffff81000000 (relocation range: 0xffffffff80000000-0xffffffffbfffffff)
```
luf

luf

2018-07-08 23:33

reporter   ~0032210

@tru: I can confirm that the lvm with raid1 is the root cause.
I'm now up and running with "3.10.0-862.6.3.el7.x86_64 #1 SMP".
I had to change /etc/lvm/lvm.conf (filter + global_filter out the devices for second lvm - the data one). I recreated initramfs (just to be sure) and I'm up and running now (without data LVM).

@toracat: Thanks, I'll wait on older kernel as I need the data LVM :)
tru

tru

2018-07-08 23:33

administrator   ~0032211

then you can no longer boot your 3.10.0-862.6.3.el7.x86_64 kernel which panic at boot time:
[ OK ] Started LVM2 PV scan on device 9:1.
[ 21.498335] md/raid1:mdX: not clean -- starting background reconstruction
[ 21.501623] md/raid1:mdX: active with 2 out of 2 mirrors
[ OK ] Started Create Volatile Files and Directories.
         Starting Security Auditing Service...
[ OK ] Started Device-mapper event daemon.
         Starting Device-mapper event daemon...
[ 21.537101] md: resync of RAID array mdX
[ 21.633097] BUG: unable to handle kernel NULL pointer dereference at 00000000000005ec
[ 21.641911] IP: [<ffffffffc0437ffb>] read_balance+0x1db/0x4e0 [raid1]
[ 21.643007] PGD 0
[ 21.643367] Oops: 0000 [#1] SMP
[ 21.643905] Modules linked in: dm_raid raid456 async_raid6_recov async_memcpy async_pq raid6_pq async_xor xor async_tx ext4 mbcache jbd2 intel_powerclamp iosf_mbi crc32_pclmul ghash_clmulni_intel aesni_intel lrw gf128mul glue_helper ablk_helper ppdev snd_intel8x0(+) cryptd snd_ac97_codec ac97_bus pcspkr snd_seq snd_seq_device snd_pcm sg snd_timer i2c_piix4 i2c_core floppy(+) snd soundcore parport_pc video parport ip_tables xfs libcrc32c raid1 sd_mod crc_t10dif crct10dif_generic ahci libahci libata virtio_net virtio_pci crct10dif_pclmul crct10dif_common crc32c_intel virtio_ring virtio serio_raw dm_mirror dm_region_hash dm_log dm_mod
[ 21.831667] CPU: 0 PID: 706 Comm: systemd-udevd Not tainted 3.10.0-862.6.3.el7.x86_64 #1
[ 21.922915] Hardware name: innotek GmbH VirtualBox/VirtualBox, BIOS VirtualBox 12/01/2006
[ 21.983349] task: ffff8f4536261fa0 ti: ffff8f457a83c000 task.ti: ffff8f457a83c000
[ 21.984467] RIP: 0010:[<ffffffffc0437ffb>] [<ffffffffc0437ffb>] read_balance+0x1db/0x4e0 [raid1]
[ 21.985787] RSP: 0018:ffff8f457a83f660 EFLAGS: 00010246
[ 21.986559] RAX: 0000000000000000 RBX: 0000000000000000 RCX: 0000000000000002
[ 21.993088] RDX: 0000000000000000 RSI: ffff8f45352a8978 RDI: 0000000000000000
[ 21.994128] RBP: ffff8f457a83f6e8 R08: 0000000000000001 R09: 0000000000000008
[ 21.995164] R10: 0000000000031f80 R11: 0000000000000000 R12: ffff8f457ad79680
[ 21.996206] R13: ffff8f457a398600 R14: ffff8f457a7cd800 R15: ffff8f4536839d00
[ 21.997251] FS: 00007f33dbab48c0(0000) GS:ffff8f457fc00000(0000) knlGS:0000000000000000
[ 21.998437] CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033
[ 21.999277] CR2: 00000000000005ec CR3: 000000007a900000 CR4: 00000000000606f0
[ 22.000315] DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
[ 22.001351] DR3: 0000000000000000 DR6: 00000000fffe0ff0 DR7: 0000000000000400
[ 22.002413] Call Trace:
[ 22.002792] [<ffffffffc0439024>] ? r1bio_pool_alloc+0x24/0x30 [raid1]
[ 22.003761] [<ffffffffc0439c33>] raid1_make_request+0x193/0x500 [raid1]
[ 22.004743] [<ffffffff9eccc0ba>] ? check_preempt_curr+0x8a/0xa0
[ 22.052568] [<ffffffff9eccf3ec>] ? try_to_wake_up+0x18c/0x350
[ 22.053940] [<ffffffff9f1486b0>] md_handle_request+0xd0/0x150
[ 22.054869] [<ffffffffc0793ba8>] raid_map+0x28/0x40 [dm_raid]
[ 22.055816] [<ffffffffc0334136>] __map_bio+0x96/0x190 [dm_mod]
[ 22.056763] [<ffffffffc0332680>] ? queue_io+0x80/0x80 [dm_mod]
[ 22.057711] [<ffffffffc0334437>] __clone_and_map_data_bio+0x177/0x280 [dm_mod]
[ 22.058877] [<ffffffffc0334811>] __split_and_process_bio+0x2d1/0x520 [dm_mod]
[ 22.060083] [<ffffffffc0334d7d>] dm_make_request+0x11d/0x1a0 [dm_mod]
[ 22.061155] [<ffffffff9ef1b0fb>] generic_make_request+0x10b/0x320
[ 22.062149] [<ffffffff9ef1b380>] submit_bio+0x70/0x150
[ 22.073189] [<ffffffff9ed94dee>] ? __add_to_page_cache_locked+0xee/0x190
[ 22.074373] [<ffffffff9ee5ed0a>] mpage_bio_submit+0x2a/0x40
[ 22.075354] [<ffffffff9ee5f73f>] mpage_readpages+0x13f/0x170
[ 22.076401] [<ffffffff9ee59250>] ? I_BDEV+0x10/0x10
[ 22.077397] [<ffffffff9ee59250>] ? I_BDEV+0x10/0x10
[ 22.078265] [<ffffffff9ee59b0d>] blkdev_readpages+0x1d/0x20
[ 22.079190] [<ffffffff9eda19ef>] __do_page_cache_readahead+0x1cf/0x260
[ 22.092999] [<ffffffff9eda1f89>] force_page_cache_readahead+0x99/0xe0
[ 22.094066] [<ffffffff9eda2067>] page_cache_sync_readahead+0x97/0xb0
[ 22.095099] [<ffffffff9ed95c92>] generic_file_aio_read+0x2c2/0x790
[ 22.096107] [<ffffffff9ee59f4c>] blkdev_aio_read+0x4c/0x70
[ 22.096998] [<ffffffff9ee1a8d3>] do_sync_read+0x93/0xe0
[ 22.097856] [<ffffffff9ee1b2ff>] vfs_read+0x9f/0x170
[ 22.112888] [<ffffffff9ee1c1cf>] SyS_read+0x7f/0xf0
[ 22.113806] [<ffffffff9f3206e1>] ? system_call_after_swapgs+0xae/0x146
[ 22.123560] [<ffffffff9f320795>] system_call_fastpath+0x1c/0x21
[ 22.124542] [<ffffffff9f3206e1>] ? system_call_after_swapgs+0xae/0x146
[ 22.125637] Code: c1 4c 01 d0 48 89 42 10 49 8b 46 28 49 8b 17 48 8b 40 20 48 8b 92 50 03 00 00 a8 80 0f 85 e9 02 00 00 f6 c4 02 0f 85 d5 02 00 00 <8b> 82 ec 05 00 00 44 39 c8 4c 89 d7 44 89 5d bc 41 0f 4f c1 48
[ 22.129939] RIP [<ffffffffc0437ffb>] read_balance+0x1db/0x4e0 [raid1]
[ 22.130989] RSP <ffff8f457a83f660>
[ 22.131556] CR2: 00000000000005ec
[ 22.163144] ---[ end trace 742a3760c2c90099 ]---
[ 22.173338] Kernel panic - not syncing: Fatal exception
[ 22.174217] Kernel Offset: 0x1dc00000 from 0xffffffff81000000 (relocation range: 0xffffffff80000000-0xffffffffbfffffff)
tru

tru

2018-07-08 23:38

administrator   ~0032212

kickstart reproducers
https://github.com/truatpasteurdotfr/ks/blob/master/c7-GA-raid1%2Bboot%2Blvm-15013.cfg
https://github.com/truatpasteurdotfr/ks/blob/master/c7-raid1%2Bboot%2Blvm-15013.cfg

SIGSLEEP :P
luf

luf

2018-07-08 23:49

reporter   ~0032213

@tru: the oops is during "vgchange -ay"
Once booted into 3.10.0-862.6.3 - filtered data VG (with raid1):
vgscan = see only system VG
I removed the changes in /etc/lvm/lvm.conf and run:
vgscan = ok - see both VGs (system and data)
vgchange -ay = system freeze/kernel oops

Issue History

Date Modified Username Field Change
2018-07-06 04:54 luf New Issue
2018-07-06 04:54 luf Tag Attached: kerneloops
2018-07-07 07:01 toracat Note Added: 0032188
2018-07-07 07:11 luf Note Added: 0032189
2018-07-07 07:24 toracat Note Added: 0032190
2018-07-07 07:25 toracat Note Edited: 0032188 View Revisions
2018-07-07 07:26 toracat Status new => acknowledged
2018-07-07 10:44 tru Note Added: 0032192
2018-07-07 10:45 tru Tag Attached: upstream
2018-07-07 10:45 tru Tag Attached: regression
2018-07-07 15:42 toracat Note Edited: 0032190 View Revisions
2018-07-07 17:10 toracat Note Added: 0032195
2018-07-08 20:42 tru Note Added: 0032202
2018-07-08 20:46 tru Note Added: 0032203
2018-07-08 21:24 tru Note Added: 0032204
2018-07-08 22:04 tru Note Added: 0032205
2018-07-08 22:25 luf Note Added: 0032206
2018-07-08 22:50 tru Note Edited: 0032205 View Revisions
2018-07-08 22:57 tru Note Edited: 0032205 View Revisions
2018-07-08 23:13 toracat Note Added: 0032207
2018-07-08 23:14 toracat Status acknowledged => assigned
2018-07-08 23:29 tru Note Added: 0032209
2018-07-08 23:33 luf Note Added: 0032210
2018-07-08 23:33 tru Note Added: 0032211
2018-07-08 23:38 tru Note Added: 0032212
2018-07-08 23:49 luf Note Added: 0032213