Match messages in logs (every line would be required to be present in log output Copy from "Messages before crash" column below): | |
Match messages in full crash (every line would be required to be present in crash log output Copy from "Full Crash" column below): | |
Limit to a test: (Copy from below "Failing text"): | |
Delete these reports as invalid (real bug in review or some such) | |
Bug or comment: | |
Extra info: |
Failing Test | Full Crash | Messages before crash | Comment |
---|---|---|---|
racer test 1: racer on clients: centos-105.localnet DURATION=2700 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 PGD 2c6daa067 PUD 2d47ec067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) lustre(OE) osp(OE) ofd(OE) lod(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) jbd2 mbcache lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) crc32_generic lnet(OE) dm_flakey dm_mod libcfs(OE) crc_t10dif crct10dif_generic crct10dif_common i2c_piix4 virtio_balloon virtio_console pcspkr ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm drm_panel_orientation_quirks ata_piix serio_raw virtio_blk i2c_core floppy libata CPU: 7 PID: 3735 Comm: ll_sa_3493 Kdump: loaded Tainted: P OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.16.0-3.module_el8.7.0+1218+f626c2ff 04/01/2014 task: ffff8802d7b35c40 ti: ffff8802a48c0000 task.ti: ffff8802a48c0000 RIP: 0010:[<ffffffff813f502c>] [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 RSP: 0018:ffff8802a48c3d78 EFLAGS: 00010296 RAX: ffff880278f43090 RBX: 0000000000000008 RCX: 000000010026001a RDX: 0000000000000026 RSI: ffff880278f433c0 RDI: 0000000000000008 RBP: ffff8802a48c3d88 R08: ffff8802ac042128 R09: 0000000000000000 R10: ffff8802ac0413e8 R11: ffff8802ac042128 R12: 0000000000000000 R13: ffff8800a140a4f0 R14: ffff880278f433c0 R15: ffff8802ac042128 FS: 0000000000000000(0000) GS:ffff880331bc0000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 00000000a210a000 CR4: 00000000000007e0 Call Trace: [<ffffffffa1120521>] ll_statahead_thread+0xdd1/0x2020 [lustre] [<ffffffff810df6e2>] ? pick_next_task_fair+0x702/0x8a0 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e329e>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffff810c834d>] ? finish_task_switch+0x5d/0x1b0 [<ffffffffa111f750>] ? ll_statahead_handle.constprop.30+0x1c0/0x1c0 [lustre] [<ffffffff810ba114>] kthread+0xe4/0xf0 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 | Lustre: 18198:0:(mdd_dir.c:4838:mdd_migrate_object()) lustre-MDD0002: [0x280000403:0x1:0x0]/7 is open, migrate only dentry Lustre: 18198:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 507 < left 922, rollback = 2 Lustre: 18198:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 3/12/4, destroy: 1/4/0 Lustre: 18198:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 4/4/0, xattr_set: 13/922/0 Lustre: 18198:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 5/55/0, punch: 0/0/0, quota 1/3/0 Lustre: 18198:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 15/263/4, delete: 3/6/0 Lustre: 18198:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 8/8/0, ref_del: 3/3/0 Lustre: 17163:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff8802e9081940 x1837361153323264/t4294967650(0) o101->b09ae593-5930-4106-9fab-3f290d23c7d5@0@lo:198/0 lens 376/816 e 0 to 0 dl 1752244213 ref 1 fl Interpret:H/602/0 rc 0/0 job:'dd.0' uid:0 gid:0 projid:0 chown (18156) used greatest stack depth: 10064 bytes left Lustre: 11103:0:(mdd_dir.c:4838:mdd_migrate_object()) lustre-MDD0001: [0x240000403:0x1:0x0]/5 is open, migrate only dentry Lustre: 11102:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 510 < left 877, rollback = 2 Lustre: 11102:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 8 previous similar messages Lustre: 11102:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 3/12/2, destroy: 0/0/0 Lustre: 11102:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 8 previous similar messages Lustre: 11102:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 2/2/0, xattr_set: 11/877/0 Lustre: 11102:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 8 previous similar messages Lustre: 11102:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 5/55/0, punch: 0/0/0, quota 1/3/0 Lustre: 11102:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 8 previous similar messages Lustre: 11102:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 10/203/3, delete: 0/0/0 Lustre: 11102:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 8 previous similar messages Lustre: 11102:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 4/4/0, ref_del: 0/0/0 Lustre: 11102:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 8 previous similar messages Lustre: 19775:0:(mdd_dir.c:4838:mdd_migrate_object()) lustre-MDD0000: [0x200000403:0x1:0x0]/8 is open, migrate only dentry Lustre: 19797:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 510 < left 788, rollback = 2 Lustre: 19797:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 17 previous similar messages Lustre: 11090:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 1/4/2, destroy: 0/0/0 Lustre: 11090:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 27 previous similar messages Lustre: 11090:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 2/2/0, xattr_set: 12/951/0 Lustre: 11090:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 27 previous similar messages Lustre: 11090:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 5/21/0, punch: 0/0/0, quota 7/129/0 Lustre: 11090:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 27 previous similar messages Lustre: 11090:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 5/141/1, delete: 2/2/0 Lustre: 11090:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 27 previous similar messages Lustre: 11090:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 2/2/0, ref_del: 1/1/0 Lustre: 11090:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 27 previous similar messages Lustre: 11082:0:(lod_lov.c:1402:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000404:0x2e:0x0] with magic=0xbd60bd0 Lustre: 17823:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff8802dd984140 x1837361155213952/t4294968565(0) o101->8cc7706d-621d-4586-b561-4aca88e0b647@0@lo:203/0 lens 376/816 e 0 to 0 dl 1752244218 ref 1 fl Interpret:H/602/0 rc 0/0 job:'dd.0' uid:0 gid:0 projid:0 Lustre: 11103:0:(mdd_dir.c:4838:mdd_migrate_object()) lustre-MDD0002: [0x280000403:0x1:0x0]/16 is open, migrate only dentry 14[20446]: segfault at 8 ip 00007fda02f837e8 sp 00007fff960f0eb0 error 4 in ld-2.17.so[7fda02f78000+22000] Lustre: 12839:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-OST0002: opcode 7: before 516 < left 618, rollback = 7 Lustre: 12839:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 12839:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 21 previous similar messages Lustre: 12839:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 12839:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 21 previous similar messages Lustre: 12839:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 12839:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 21 previous similar messages Lustre: 12839:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 12839:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 21 previous similar messages Lustre: 12839:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 12839:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 21 previous similar messages Lustre: 11105:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 505 < left 1990, rollback = 2 Lustre: 11105:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 31 previous similar messages Lustre: 19374:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-OST0002: opcode 7: before 516 < left 618, rollback = 7 Lustre: 19374:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 3 previous similar messages LustreError: 11103:0:(mdt_reint.c:2564:mdt_reint_migrate()) lustre-MDT0001: migrate [0x240000404:0x8e:0x0]/13 failed: rc = -116 Lustre: 17134:0:(lod_lov.c:1402:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000403:0x38:0x0] with magic=0xbd60bd0 Lustre: 17134:0:(lod_lov.c:1402:lod_parse_striping()) Skipped 1 previous similar message Lustre: 18843:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 3/12/9, destroy: 0/0/0 Lustre: 18843:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 77 previous similar messages Lustre: 18843:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 2/2/0, xattr_set: 11/877/0 Lustre: 18843:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 77 previous similar messages Lustre: 18843:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 5/55/0, punch: 0/0/0, quota 4/54/0 Lustre: 18843:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 77 previous similar messages Lustre: 18843:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 12/275/4, delete: 0/0/0 Lustre: 18843:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 77 previous similar messages Lustre: 18843:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 6/6/0, ref_del: 0/0/0 Lustre: 18843:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 77 previous similar messages Lustre: 19809:0:(mdd_dir.c:4838:mdd_migrate_object()) lustre-MDD0002: [0x280000403:0x1:0x0]/12 is open, migrate only dentry Lustre: 19809:0:(mdd_dir.c:4838:mdd_migrate_object()) Skipped 3 previous similar messages LustreError: 11105:0:(mdt_reint.c:2564:mdt_reint_migrate()) lustre-MDT0001: migrate [0x240000404:0x87:0x0]/20 failed: rc = -2 Lustre: 23052:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 504 < left 1634, rollback = 2 Lustre: 23052:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 101 previous similar messages 11[23638]: segfault at 8 ip 00007fc99e5b27e8 sp 00007ffc6f618a30 error 4 in ld-2.17.so[7fc99e5a7000+22000] Lustre: 18198:0:(mdd_dir.c:4838:mdd_migrate_object()) lustre-MDD0002: [0x280000403:0x1:0x0]/10 is open, migrate only dentry Lustre: 18198:0:(mdd_dir.c:4838:mdd_migrate_object()) Skipped 3 previous similar messages LustreError: 20036:0:(mdt_reint.c:2564:mdt_reint_migrate()) lustre-MDT0000: migrate [0x200000403:0x1:0x0]/14 failed: rc = -2 8[24664]: segfault at 8 ip 00007fac06ce87e8 sp 00007ffe189ddb30 error 4 in ld-2.17.so[7fac06cdd000+22000] Lustre: 18198:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 3/12/12, destroy: 0/0/0 Lustre: 18198:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 71 previous similar messages Lustre: 18198:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 2/2/0, xattr_set: 10/788/0 Lustre: 18198:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 71 previous similar messages Lustre: 18198:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 5/55/0, punch: 0/0/0, quota 4/54/0 Lustre: 18198:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 71 previous similar messages Lustre: 18198:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 11/239/4, delete: 0/0/0 Lustre: 18198:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 71 previous similar messages Lustre: 18198:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 5/5/0, ref_del: 0/0/0 Lustre: 18198:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 71 previous similar messages 14[24565]: segfault at 8 ip 00007f47d2e267e8 sp 00007ffe2f89e9a0 error 4 in ld-2.17.so[7f47d2e1b000+22000] LustreError: 19775:0:(mdt_reint.c:2564:mdt_reint_migrate()) lustre-MDT0002: migrate [0x240000404:0xe5:0x0]/18 failed: rc = -2 Lustre: dir [0x200000403:0x143:0x0] stripe 2 readdir failed: -2, directory is partially accessed! LustreError: 18843:0:(mdd_dir.c:4759:mdd_migrate_cmd_check()) lustre-MDD0002: '13' migration was interrupted, run 'lfs migrate -m 1 -c 3 -H crush 13' to finish migration: rc = -1 LustreError: 21008:0:(mdt_xattr.c:406:mdt_dir_layout_update()) lustre-MDT0001: [0x240000404:0xe5:0x0] migrate mdt count mismatch 2 != 1 Lustre: 19797:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 508 < left 983, rollback = 2 Lustre: 19797:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 61 previous similar messages Lustre: 20862:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff8800a3fa3240 x1837361162029184/t4294972255(0) o101->b09ae593-5930-4106-9fab-3f290d23c7d5@0@lo:230/0 lens 376/864 e 0 to 0 dl 1752244245 ref 1 fl Interpret:H/602/0 rc 0/0 job:'dd.0' uid:0 gid:0 projid:0 8[27156]: segfault at 8 ip 00007f715b2197e8 sp 00007ffed9fbf300 error 4 in ld-2.17.so[7f715b20e000+22000] LustreError: 19797:0:(mdt_reint.c:2564:mdt_reint_migrate()) lustre-MDT0002: migrate [0x240000404:0xa3:0x0]/7 failed: rc = -2 LustreError: 19797:0:(mdt_reint.c:2564:mdt_reint_migrate()) Skipped 1 previous similar message 0[27396]: segfault at 8 ip 00007ff0b7f3e7e8 sp 00007fff6e88f9f0 error 4 in ld-2.17.so[7ff0b7f33000+22000] Lustre: 11103:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 3/12/4, destroy: 0/0/0 Lustre: 11103:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 145 previous similar messages Lustre: 11103:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 2/2/0, xattr_set: 10/788/0 Lustre: 11103:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 145 previous similar messages Lustre: 11103:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 5/55/0, punch: 0/0/0, quota 4/54/0 Lustre: 11103:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 145 previous similar messages Lustre: 11103:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 12/275/3, delete: 0/0/0 Lustre: 11103:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 145 previous similar messages Lustre: 11103:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 6/6/0, ref_del: 0/0/0 Lustre: 11103:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 145 previous similar messages Lustre: 19797:0:(mdt_reint.c:2484:mdt_reint_migrate()) lustre-MDT0002: [0x280000403:0x1:0x0]/2 is open, migrate only dentry LustreError: 27857:0:(file.c:248:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a66892a8: inode [0x240000403:0x10b:0x0] mdc close failed: rc = -13 Lustre: 12839:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-OST0002: opcode 7: before 516 < left 618, rollback = 7 Lustre: 12839:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 1 previous similar message Lustre: 11103:0:(mdd_dir.c:4838:mdd_migrate_object()) lustre-MDD0000: [0x200000403:0x1:0x0]/8 is open, migrate only dentry Lustre: 11103:0:(mdd_dir.c:4838:mdd_migrate_object()) Skipped 7 previous similar messages Lustre: 21279:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-OST0002: opcode 7: before 516 < left 618, rollback = 7 Lustre: 21279:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 1 previous similar message LustreError: 23052:0:(lustre_lmv.h:500:lmv_is_sane()) unknown layout LMV: magic=0xcd40cd0 count=2 index=1 hash=crush:0x82000003 version=1 migrate_offset=1 migrate_hash=fnv_1a_64:2 pool= LustreError: 23052:0:(mdt_reint.c:2564:mdt_reint_migrate()) lustre-MDT0000: migrate [0x200000404:0x28a:0x0]/4 failed: rc = -9 14[29625]: segfault at 8 ip 00007fe2f717d7e8 sp 00007fff3effa1d0 error 4 in ld-2.17.so[7fe2f7172000+22000] LustreError: 30045:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x280000403:0x1aa:0x0]: rc = -5 LustreError: 30045:0:(llite_lib.c:3770:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 4[29850]: segfault at 8 ip 00007f36b931d7e8 sp 00007fff07107740 error 4 in ld-2.17.so[7f36b9312000+22000] LustreError: 20036:0:(mdd_dir.c:4759:mdd_migrate_cmd_check()) lustre-MDD0000: '3' migration was interrupted, run 'lfs migrate -m 2 -c 1 -H crush 3' to finish migration: rc = -1 Lustre: dir [0x280000403:0x16d:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 2 previous similar messages LustreError: 29985:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000404:0x2c7:0x0]: rc = -5 LustreError: 29985:0:(lcommon_cl.c:188:cl_file_inode_init()) Skipped 1 previous similar message LustreError: 29985:0:(llite_lib.c:3770:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 29985:0:(llite_lib.c:3770:ll_prep_inode()) Skipped 1 previous similar message Lustre: 30176:0:(lod_lov.c:1402:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000403:0x1c8:0x0] with magic=0xbd60bd0 Lustre: 30176:0:(lod_lov.c:1402:lod_parse_striping()) Skipped 1 previous similar message Lustre: 19797:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 506 < left 727, rollback = 2 Lustre: 19797:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 238 previous similar messages Lustre: dir [0x200000403:0x277:0x0] stripe 3 readdir failed: -2, directory is partially accessed! LustreError: 29422:0:(lov_object.c:1350:lov_layout_change()) lustre-clilov-ffff8800a668ae98: cannot apply new layout on [0x200000404:0x240:0x0] : rc = -5 LustreError: 29422:0:(vvp_io.c:1909:vvp_io_init()) lustre: refresh file layout [0x200000404:0x240:0x0] error -5. Lustre: 14212:0:(lod_lov.c:1402:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000403:0x1ca:0x0] with magic=0xbd60bd0 Lustre: 14212:0:(lod_lov.c:1402:lod_parse_striping()) Skipped 3 previous similar messages 14[31596]: segfault at 8 ip 00007f7b464ed7e8 sp 00007ffe4085b1c0 error 4 in ld-2.17.so[7f7b464e2000+22000] LustreError: 19670:0:(lustre_lmv.h:500:lmv_is_sane()) unknown layout LMV: magic=0xcd40cd0 count=3 index=2 hash=crush:0x82000003 version=1 migrate_offset=2 migrate_hash=fnv_1a_64:2 pool= LustreError: 31709:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000404:0x2c7:0x0]: rc = -5 LustreError: 31709:0:(lcommon_cl.c:188:cl_file_inode_init()) Skipped 2 previous similar messages LustreError: 31709:0:(llite_lib.c:3770:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 31709:0:(llite_lib.c:3770:ll_prep_inode()) Skipped 2 previous similar messages LustreError: 18986:0:(mdt_xattr.c:406:mdt_dir_layout_update()) lustre-MDT0001: [0x240000403:0x209:0x0] migrate mdt count mismatch 2 != 1 LustreError: 31751:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000404:0x2c7:0x0]: rc = -5 LustreError: 31751:0:(llite_lib.c:3770:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 18206:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0001: failed to get lu_attr of [0x240000404:0x87:0x0]: rc = -2 LustreError: 31137:0:(file.c:248:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a66892a8: inode [0x240000404:0x87:0x0] mdc close failed: rc = -2 LustreError: 16885:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0001: failed to get lu_attr of [0x240000404:0x87:0x0]: rc = -2 LustreError: 32203:0:(file.c:248:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a668ae98: inode [0x240000404:0x87:0x0] mdc close failed: rc = -2 LustreError: 18427:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0002: failed to get lu_attr of [0x280000403:0x1b:0x0]: rc = -2 LustreError: 19797:0:(mdd_dir.c:4759:mdd_migrate_cmd_check()) lustre-MDD0002: '10' migration was interrupted, run 'lfs migrate -m 0 -c 2 -H crush 10' to finish migration: rc = -1 LustreError: 19797:0:(mdd_dir.c:4759:mdd_migrate_cmd_check()) Skipped 1 previous similar message LustreError: 19797:0:(mdt_reint.c:2564:mdt_reint_migrate()) lustre-MDT0002: migrate [0x200000403:0x1:0x0]/10 failed: rc = -1 LustreError: 19797:0:(mdt_reint.c:2564:mdt_reint_migrate()) Skipped 5 previous similar messages Lustre: 19374:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 618, rollback = 7 Lustre: 19374:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 5 previous similar messages 6[1025]: segfault at 8 ip 00007fe1631d97e8 sp 00007ffd81fe72a0 error 4 in ld-2.17.so[7fe1631ce000+22000] Lustre: 30174:0:(lod_lov.c:1402:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000404:0x204:0x0] with magic=0xbd60bd0 Lustre: 30174:0:(lod_lov.c:1402:lod_parse_striping()) Skipped 3 previous similar messages Lustre: 21310:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 3/12/6, destroy: 0/0/0 Lustre: 21310:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 367 previous similar messages Lustre: 21310:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 2/2/0, xattr_set: 9/699/0 Lustre: 21310:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 367 previous similar messages Lustre: 21310:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 5/55/0, punch: 0/0/0, quota 4/54/0 Lustre: 21310:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 367 previous similar messages Lustre: 21310:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 11/239/2, delete: 0/0/0 Lustre: 21310:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 367 previous similar messages Lustre: 21310:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 5/5/0, ref_del: 0/0/0 Lustre: 21310:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 367 previous similar messages Lustre: 20036:0:(mdd_dir.c:4838:mdd_migrate_object()) lustre-MDD0002: [0x280000403:0x1:0x0]/4 is open, migrate only dentry Lustre: 20036:0:(mdd_dir.c:4838:mdd_migrate_object()) Skipped 11 previous similar messages LustreError: 30179:0:(mdt_xattr.c:406:mdt_dir_layout_update()) lustre-MDT0001: [0x240000404:0x169:0x0] migrate mdt count mismatch 1 != 2 Lustre: 11103:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 510 < left 610, rollback = 2 Lustre: 11103:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 264 previous similar messages LustreError: 6425:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000404:0x2c7:0x0]: rc = -5 LustreError: 6425:0:(lcommon_cl.c:188:cl_file_inode_init()) Skipped 1 previous similar message LustreError: 6425:0:(llite_lib.c:3770:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 6425:0:(llite_lib.c:3770:ll_prep_inode()) Skipped 1 previous similar message Lustre: 21287:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 618, rollback = 7 Lustre: 21287:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 1 previous similar message Lustre: 18984:0:(lod_lov.c:1402:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000404:0x3e7:0x0] with magic=0xbd60bd0 Lustre: 18984:0:(lod_lov.c:1402:lod_parse_striping()) Skipped 1 previous similar message LustreError: 4079:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0002: failed to get lu_attr of [0x280000403:0x32a:0x0]: rc = -2 LustreError: 5163:0:(file.c:248:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a668ae98: inode [0x280000403:0x32a:0x0] mdc close failed: rc = -2 LustreError: 5163:0:(file.c:248:ll_close_inode_openhandle()) Skipped 1 previous similar message LustreError: 19809:0:(mdd_dir.c:4759:mdd_migrate_cmd_check()) lustre-MDD0000: '17' migration was interrupted, run 'lfs migrate -m 1 -c 1 -H crush 17' to finish migration: rc = -1 18[8104]: segfault at 1c ip 00000000004017b0 sp 00007ffeb1ffff30 error 6 in 18[400000+6000] LustreError: 18843:0:(mdt_reint.c:2564:mdt_reint_migrate()) lustre-MDT0000: migrate [0x280000404:0x36e:0x0]/20 failed: rc = -2 LustreError: 18843:0:(mdt_reint.c:2564:mdt_reint_migrate()) Skipped 7 previous similar messages LustreError: 9134:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000404:0x3de:0x0]: rc = -5 LustreError: 9134:0:(lcommon_cl.c:188:cl_file_inode_init()) Skipped 4 previous similar messages LustreError: 9134:0:(llite_lib.c:3770:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 9134:0:(llite_lib.c:3770:ll_prep_inode()) Skipped 4 previous similar messages 18[9216]: segfault at 406000 ip 0000000000406000 sp 00007ffd9c838018 error 14 in 18[606000+1000] LustreError: 8432:0:(llite_lib.c:1872:ll_update_lsm_md()) lustre: [0x240000403:0x403:0x0] dir layout mismatch: LustreError: 8432:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=3 count=2 index=1 hash=crush:0x2000003 max_inherit=0 max_inherit_rr=0 version=2 migrate_offset=0 migrate_hash=invalid:0 pool= LustreError: 8432:0:(lustre_lmv.h:167:lmv_stripe_object_dump()) stripe[0] [0x240000400:0x1c:0x0] LustreError: 8432:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=1 count=3 index=1 hash=crush:0x82000003 max_inherit=0 max_inherit_rr=0 version=1 migrate_offset=2 migrate_hash=fnv_1a_64:2 pool= Lustre: 12837:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 516 < left 618, rollback = 7 Lustre: 12837:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 13 previous similar messages traps: 11[10105] trap invalid opcode ip:4059fe sp:7ffdc3f61398 error:0 in 11[400000+6000] LustreError: 18982:0:(mdt_xattr.c:406:mdt_dir_layout_update()) lustre-MDT0001: [0x240000403:0x231:0x0] migrate mdt count mismatch 1 != 2 LustreError: 9106:0:(mdd_dir.c:4759:mdd_migrate_cmd_check()) lustre-MDD0001: '9' migration was interrupted, run 'lfs migrate -m 0 -c 3 -H crush 9' to finish migration: rc = -1 LustreError: 11086:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0001: failed to get lu_attr of [0x240000404:0xa3:0x0]: rc = -2 LustreError: 9959:0:(file.c:248:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a668ae98: inode [0x240000404:0xa3:0x0] mdc close failed: rc = -2 LustreError: 20036:0:(mdd_dir.c:4759:mdd_migrate_cmd_check()) lustre-MDD0001: '16' migration was interrupted, run 'lfs migrate -m 0 -c 3 -H crush 16' to finish migration: rc = -1 LustreError: 18186:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000403:0x562:0x0]: rc = -5 LustreError: 18186:0:(lcommon_cl.c:188:cl_file_inode_init()) Skipped 1 previous similar message LustreError: 18186:0:(llite_lib.c:3770:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 18186:0:(llite_lib.c:3770:ll_prep_inode()) Skipped 1 previous similar message Lustre: 11103:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 3/12/4, destroy: 1/4/0 Lustre: 11103:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 743 previous similar messages Lustre: 11103:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 5/5/0, xattr_set: 17/1203/0 Lustre: 11103:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 743 previous similar messages Lustre: 11103:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 5/55/0, punch: 0/0/0, quota 8/130/0 Lustre: 11103:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 743 previous similar messages Lustre: 11103:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 16/299/3, delete: 3/6/0 Lustre: 11103:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 743 previous similar messages Lustre: 11103:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 9/9/0, ref_del: 3/3/0 Lustre: 11103:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 743 previous similar messages LustreError: 12408:0:(lov_object.c:1350:lov_layout_change()) lustre-clilov-ffff8800a668ae98: cannot apply new layout on [0x240000403:0x562:0x0] : rc = -5 LustreError: 12408:0:(vvp_io.c:1909:vvp_io_init()) lustre: refresh file layout [0x240000403:0x562:0x0] error -5. LustreError: 19768:0:(lov_object.c:1350:lov_layout_change()) lustre-clilov-ffff8800a668ae98: cannot apply new layout on [0x240000403:0x562:0x0] : rc = -5 Lustre: 17008:0:(lod_lov.c:1402:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000403:0x53c:0x0] with magic=0xbd60bd0 Lustre: 17008:0:(lod_lov.c:1402:lod_parse_striping()) Skipped 13 previous similar messages traps: 5[15592] trap invalid opcode ip:405955 sp:969b1469 error:0 in 5[400000+6000] 1[22527]: segfault at 8 ip 00007fdb697a97e8 sp 00007fff02e00a00 error 4 in ld-2.17.so[7fdb6979e000+22000] Lustre: 18300:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 618, rollback = 7 Lustre: 18300:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 21 previous similar messages 9[29447]: segfault at 8 ip 00007fed1c1d17e8 sp 00007ffec51a26d0 error 4 in ld-2.17.so[7fed1c1c6000+22000] LustreError: 15592:0:(file.c:248:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a668ae98: inode [0x200000403:0x63f:0x0] mdc close failed: rc = -13 Lustre: 11103:0:(mdd_dir.c:4838:mdd_migrate_object()) lustre-MDD0002: [0x280000403:0x1:0x0]/2 is open, migrate only dentry Lustre: 11103:0:(mdd_dir.c:4838:mdd_migrate_object()) Skipped 12 previous similar messages Lustre: 11103:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 501 < left 699, rollback = 2 Lustre: 11103:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 710 previous similar messages 9[2309]: segfault at 8 ip 00007f993ab8b7e8 sp 00007ffd7db74690 error 4 in ld-2.17.so[7f993ab80000+22000] Lustre: mdt_io00_011: service thread pid 21321 was inactive for 40.130 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 21321, comm: mdt_io00_011 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock+0x88/0x1c0 [mdt] [<0>] mdt_rename_source_lock+0x57/0xf0 [mdt] [<0>] mdt_reint_rename+0xde7/0x3ae0 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x73c/0xbb0 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x23d/0xd80 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 11093, comm: mdt_out00_001 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_pdo_lock+0x369/0x750 [mdt] [<0>] mdt_parent_lock+0x76/0x2a0 [mdt] [<0>] mdt_getattr_name_lock+0x1854/0x2d20 [mdt] [<0>] mdt_intent_getattr+0x2cc/0x4e0 [mdt] [<0>] mdt_intent_opc.constprop.74+0x211/0xc60 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x397/0x980 [ptlrpc] [<0>] ldlm_handle_enqueue+0x547/0x1900 [ptlrpc] [<0>] mdt_batch_getattr+0x6c/0x190 [mdt] [<0>] mdt_batch+0x465/0x195b [mdt] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x23d/0xd80 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 30205, comm: mdt00_031 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock_try+0xa0/0x250 [mdt] [<0>] mdt_getattr_name_lock+0x1739/0x2d20 [mdt] [<0>] mdt_intent_getattr+0x2cc/0x4e0 [mdt] [<0>] mdt_intent_opc.constprop.74+0x211/0xc60 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x397/0x980 [ptlrpc] [<0>] ldlm_handle_enqueue+0x547/0x1900 [ptlrpc] [<0>] tgt_enqueue+0x68/0x240 [ptlrpc] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x23d/0xd80 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: mdt_out00_000: service thread pid 11090 was inactive for 40.078 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: mdt00_016: service thread pid 17927 was inactive for 40.116 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 1 previous similar message 10[6325]: segfault at 8 ip 00007fd9a2c567e8 sp 00007fff95ac7870 error 4 in ld-2.17.so[7fd9a2c4b000+22000] Lustre: mdt00_012: service thread pid 17804 was inactive for 72.066 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 6 previous similar messages LustreError: 10797:0:(ldlm_lockd.c:257:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0001_UUID lock: ffff88009d809a80/0xe0176a8f89dcf357 lrc: 3/0,0 mode: PR/PR res: [0x240000404:0x4dd:0x0].0x0 bits 0x13/0x0 rrc: 4 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0xe0176a8f89dcf166 expref: 401 pid: 11093 timeout: 355 lvb_type: 0 Lustre: mdt00_029: service thread pid 30176 completed after 84.831s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_010: service thread pid 17633 completed after 85.141s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_003: service thread pid 12863 completed after 85.131s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_out00_004: service thread pid 29429 completed after 99.450s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_032: service thread pid 6029 completed after 85.185s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_026: service thread pid 25831 completed after 85.189s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_011: service thread pid 21321 completed after 100.339s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_out00_000: service thread pid 11090 completed after 99.903s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 30205:0:(ldlm_lockd.c:1447:ldlm_handle_enqueue()) ### lock on destroyed export ffff8802e55cb7e8 ns: mdt-lustre-MDT0001_UUID lock: ffff8800b3c96580/0xe0176a8f89dcf778 lrc: 3/0,0 mode: PR/PR res: [0x240000403:0x5a0:0x0].0x0 bits 0x1b/0x0 rrc: 4 type: IBT gid 0 flags: 0x50200000000000 nid: 0@lo remote: 0xe0176a8f89dcf75c expref: 21 pid: 30205 timeout: 0 lvb_type: 0 LustreError: lustre-MDT0001-mdc-ffff8800a66892a8: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: mdt00_005: service thread pid 17008 completed after 85.023s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: lustre-MDT0001-mdc-ffff8800a66892a8: Connection to lustre-MDT0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: lustre-MDT0001-mdc-ffff8800a66892a8: This client was evicted by lustre-MDT0001; in progress operations using this service will fail. Lustre: mdt00_012: service thread pid 17804 completed after 85.180s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_031: service thread pid 30205 completed after 100.308s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_007: service thread pid 17134 completed after 85.033s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_025: service thread pid 22519 completed after 85.032s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_006: service thread pid 17122 completed after 85.055s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_out00_001: service thread pid 11093 completed after 100.346s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_024: service thread pid 21008 completed after 85.057s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 20275:0:(file.c:6198:ll_inode_revalidate_fini()) lustre: revalidate FID [0x240000403:0x1:0x0] error: rc = -5 LustreError: 20275:0:(file.c:6198:ll_inode_revalidate_fini()) Skipped 1 previous similar message Lustre: mdt00_009: service thread pid 17163 completed after 85.005s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_002: service thread pid 11083 completed after 85.194s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_033: service thread pid 7043 completed after 85.019s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_014: service thread pid 17830 completed after 85.244s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 2290:0:(llite_lib.c:2023:ll_md_setattr()) md_setattr fails: rc = -108 Lustre: mdt00_016: service thread pid 17927 completed after 99.335s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 28598:0:(mdc_request.c:1470:mdc_read_page()) lustre-MDT0001-mdc-ffff8800a66892a8: [0x240000402:0x15:0x0] lock enqueue fails: rc = -108 Lustre: dir [0x200000403:0xa60:0x0] stripe 1 readdir failed: -108, directory is partially accessed! LustreError: 28978:0:(file.c:248:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a66892a8: inode [0x240000403:0x3f1:0x0] mdc close failed: rc = -108 LustreError: 28978:0:(file.c:248:ll_close_inode_openhandle()) Skipped 1 previous similar message Lustre: lustre-MDT0001-mdc-ffff8800a66892a8: Connection restored to 192.168.123.107@tcp (at 0@lo) LustreError: 17615:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0001: failed to get lu_attr of [0x240000403:0x5d4:0x0]: rc = -2 LustreError: 19797:0:(mdt_reint.c:2564:mdt_reint_migrate()) lustre-MDT0000: migrate [0x240000404:0x392:0x0]/8 failed: rc = -2 LustreError: 19797:0:(mdt_reint.c:2564:mdt_reint_migrate()) Skipped 8 previous similar messages Lustre: 12839:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-OST0000: opcode 7: before 515 < left 618, rollback = 7 Lustre: 12839:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 19 previous similar messages Lustre: 12839:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 12839:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 181 previous similar messages Lustre: 12839:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/1, xattr_set: 2/15/0 Lustre: 12839:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 181 previous similar messages Lustre: 12839:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 12839:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 181 previous similar messages Lustre: 12839:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 12839:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 181 previous similar messages Lustre: 12839:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 12839:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 181 previous similar messages 19[26304]: segfault at 8 ip 00007fb91781f7e8 sp 00007fff89feba40 error 4 in ld-2.17.so[7fb917814000+22000] LustreError: 4079:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000403:0xaf0:0x0]: rc = -2 LustreError: 26364:0:(file.c:248:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a66892a8: inode [0x200000403:0xa60:0x0] mdc close failed: rc = -2 LustreError: 26364:0:(file.c:248:ll_close_inode_openhandle()) Skipped 12 previous similar messages LustreError: 23052:0:(mdd_dir.c:4759:mdd_migrate_cmd_check()) lustre-MDD0000: '17' migration was interrupted, run 'lfs migrate -m 2 -c 3 -H crush 17' to finish migration: rc = -1 LustreError: 30914:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x280000404:0x199f:0x0]: rc = -5 LustreError: 30914:0:(lcommon_cl.c:188:cl_file_inode_init()) Skipped 10 previous similar messages LustreError: 30914:0:(llite_lib.c:3770:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 30914:0:(llite_lib.c:3770:ll_prep_inode()) Skipped 10 previous similar messages 4[30335]: segfault at 8 ip 00007f52701857e8 sp 00007ffc62d75ec0 error 4 in ld-2.17.so[7f527017a000+22000] Lustre: 19670:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 504 < left 1606, rollback = 2 Lustre: 19670:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 388 previous similar messages Lustre: 21310:0:(mdd_dir.c:4838:mdd_migrate_object()) lustre-MDD0000: [0x200000403:0x1:0x0]/14 is open, migrate only dentry Lustre: 21310:0:(mdd_dir.c:4838:mdd_migrate_object()) Skipped 17 previous similar messages Lustre: dir [0x240000405:0x11c:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 3 previous similar messages 14[4096]: segfault at 8 ip 00007f27c2ce77e8 sp 00007fff8c18c770 error 4 in ld-2.17.so[7f27c2cdc000+22000] 8[8046]: segfault at 8 ip 00007fedcb2ba7e8 sp 00007fffe5a9d340 error 4 in ld-2.17.so[7fedcb2af000+22000] LustreError: 14014:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x280000403:0x986:0x0]: rc = -5 LustreError: 14014:0:(lcommon_cl.c:188:cl_file_inode_init()) Skipped 11 previous similar messages LustreError: 17270:0:(llite_lib.c:3770:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 17270:0:(llite_lib.c:3770:ll_prep_inode()) Skipped 12 previous similar messages Lustre: 6029:0:(lod_lov.c:1402:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000403:0xae7:0x0] with magic=0xbd60bd0 Lustre: 6029:0:(lod_lov.c:1402:lod_parse_striping()) Skipped 11 previous similar messages LustreError: 18659:0:(file.c:248:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a668ae98: inode [0x240000403:0x842:0x0] mdc close failed: rc = -13 3[24458]: segfault at 8 ip 00007f22526e17e8 sp 00007ffe539a28a0 error 4 in ld-2.17.so[7f22526d6000+22000] Lustre: 12837:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 516 < left 618, rollback = 7 Lustre: 12837:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 109 previous similar messages 19[27517]: segfault at 8 ip 00007ff1e6ce17e8 sp 00007fffd0160e70 error 4 in ld-2.17.so[7ff1e6cd6000+22000] 18[27819]: segfault at 8 ip 00007f12f9aa47e8 sp 00007fffa676f300 error 4 in ld-2.17.so[7f12f9a99000+22000] 0[27974]: segfault at 0 ip 0000000000401d60 sp 00007ffc4b5cbd38 error 6 in 0[400000+6000] 13[28479]: segfault at 8 ip 00007fc111ca07e8 sp 00007ffd9f18da60 error 4 in ld-2.17.so[7fc111c95000+22000] 14[29010]: segfault at 8 ip 00007f3bd1c207e8 sp 00007ffcc4a49540 error 4 in ld-2.17.so[7f3bd1c15000+22000] LustreError: 19797:0:(mdt_reint.c:2564:mdt_reint_migrate()) lustre-MDT0000: migrate [0x200000403:0x1:0x0]/18 failed: rc = -2 LustreError: 19797:0:(mdt_reint.c:2564:mdt_reint_migrate()) Skipped 11 previous similar messages LustreError: 21967:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0002: failed to get lu_attr of [0x280000404:0x1d53:0x0]: rc = -2 LustreError: 21967:0:(mdd_object.c:3901:mdd_close()) Skipped 1 previous similar message LustreError: 19797:0:(mdd_dir.c:4759:mdd_migrate_cmd_check()) lustre-MDD0001: '3' migration was interrupted, run 'lfs migrate -m 0 -c 2 -H crush 3' to finish migration: rc = -1 LustreError: 19797:0:(mdd_dir.c:4759:mdd_migrate_cmd_check()) Skipped 2 previous similar messages Lustre: dir [0x200000404:0x10a6:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 1 previous similar message 5[5011]: segfault at 8 ip 00007fe6b653d7e8 sp 00007ffc8abbd880 error 4 in ld-2.17.so[7fe6b6532000+22000] LustreError: 10797:0:(ldlm_lockd.c:257:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0000_UUID lock: ffff8802e1ff7c00/0xe0176a8f8a087c1b lrc: 3/0,0 mode: PR/PR res: [0x200000404:0x1209:0x0].0x0 bits 0x1b/0x0 rrc: 4 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0xe0176a8f8a087b73 expref: 952 pid: 7760 timeout: 634 lvb_type: 0 LustreError: lustre-MDT0000-mdc-ffff8800a668ae98: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 2 previous similar messages Lustre: lustre-MDT0000-mdc-ffff8800a668ae98: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: lustre-MDT0000-mdc-ffff8800a668ae98: This client was evicted by lustre-MDT0000; in progress operations using this service will fail. LustreError: 4559:0:(file.c:248:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a668ae98: inode [0x200000404:0x1079:0x0] mdc close failed: rc = -108 LustreError: 4559:0:(file.c:248:ll_close_inode_openhandle()) Skipped 3 previous similar messages LustreError: 6303:0:(file.c:6198:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000404:0x1209:0x0] error: rc = -5 LustreError: 6303:0:(file.c:6198:ll_inode_revalidate_fini()) Skipped 120 previous similar messages Lustre: lustre-MDT0000-mdc-ffff8800a668ae98: Connection restored to 192.168.123.107@tcp (at 0@lo) Lustre: 17162:0:(lod_lov.c:1402:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000403:0xad2:0x0] with magic=0xbd60bd0 Lustre: 17162:0:(lod_lov.c:1402:lod_parse_striping()) Skipped 1 previous similar message Lustre: 18300:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 18300:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 825 previous similar messages Lustre: 18300:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 18300:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 825 previous similar messages Lustre: 18300:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 18300:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 825 previous similar messages Lustre: 18300:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 18300:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 825 previous similar messages Lustre: 18300:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 18300:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 825 previous similar messages LustreError: 6600:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000403:0xafa:0x0]: rc = -5 LustreError: 6600:0:(lcommon_cl.c:188:cl_file_inode_init()) Skipped 25 previous similar messages LustreError: 6600:0:(llite_lib.c:3770:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 6600:0:(llite_lib.c:3770:ll_prep_inode()) Skipped 24 previous similar messages 4[20439]: segfault at 0 ip 0000000000403e5f sp 00007ffef5b1d040 error 6 in 4[400000+6000] Lustre: 19797:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 512 < left 922, rollback = 2 Lustre: 19797:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 590 previous similar messages LustreError: 924:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 14 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 LustreError: 12264:0:(llite_nfs.c:430:ll_dir_get_parent_fid()) lustre: failure inode [0x240000403:0xb3d:0x0] get parent: rc = -116 Lustre: 19797:0:(mdd_dir.c:4838:mdd_migrate_object()) lustre-MDD0001: [0x240000403:0x1:0x0]/15 is open, migrate only dentry Lustre: 19797:0:(mdd_dir.c:4838:mdd_migrate_object()) Skipped 26 previous similar messages 18[4766]: segfault at 8 ip 00007f8c4d84c7e8 sp 00007ffc5b964420 error 4 in ld-2.17.so[7f8c4d841000+22000] 0[5231]: segfault at 8 ip 00007f9cd2cca7e8 sp 00007ffe7a2d95e0 error 4 in ld-2.17.so[7f9cd2cbf000+22000] 12[9544]: segfault at 8 ip 00007f47addca7e8 sp 00007ffcfae802c0 error 4 in ld-2.17.so[7f47addbf000+22000] Lustre: 12839:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 516 < left 582, rollback = 7 Lustre: 12839:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 87 previous similar messages 13[23431]: segfault at 0 ip (null) sp 00007ffff971e9e8 error 14 in 13[400000+6000] 12[23501]: segfault at 0 ip (null) sp 00007ffc67b51358 error 14 in 12[400000+6000] 18[27059]: segfault at 8 ip 00007fa3fa0967e8 sp 00007fff8a805570 error 4 in ld-2.17.so[7fa3fa08b000+22000] Lustre: 18066:0:(lod_lov.c:1402:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000404:0x2459:0x0] with magic=0xbd60bd0 Lustre: 18066:0:(lod_lov.c:1402:lod_parse_striping()) Skipped 7 previous similar messages 11[7200]: segfault at 8 ip 00007f74da78b7e8 sp 00007ffc7bd53880 error 4 in ld-2.17.so[7f74da780000+22000] 18[7446]: segfault at 8 ip 00007f7a9637f7e8 sp 00007ffcda8e69b0 error 4 in ld-2.17.so[7f7a96374000+22000] 4[8099]: segfault at 8 ip 00007f6d95e9c7e8 sp 00007ffd82f6c070 error 4 in ld-2.17.so[7f6d95e91000+22000] cp (7976) used greatest stack depth: 9920 bytes left 4[8227]: segfault at 8 ip 00007f16971237e8 sp 00007ffcf58288d0 error 4 in ld-2.17.so[7f1697118000+22000] 14[9220]: segfault at 8 ip 00007f59a389c7e8 sp 00007ffe8c9408a0 error 4 in ld-2.17.so[7f59a3891000+22000] LustreError: 14888:0:(mdt_reint.c:2564:mdt_reint_migrate()) lustre-MDT0000: migrate [0x200000403:0x2:0x0]/4 failed: rc = -2 LustreError: 14888:0:(mdt_reint.c:2564:mdt_reint_migrate()) Skipped 11 previous similar messages 10[26203]: segfault at 1c ip 00000000004017b0 sp 00007ffc2ac0b1d0 error 6 in 10[400000+6000] 10[26424]: segfault at 8 ip 00007f1fa034f7e8 sp 00007ffc6d84b650 error 4 in ld-2.17.so[7f1fa0344000+22000] LustreError: 18206:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0001: failed to get lu_attr of [0x240000403:0x1241:0x0]: rc = -2 LustreError: 18206:0:(mdd_object.c:3901:mdd_close()) Skipped 2 previous similar messages LustreError: 13199:0:(file.c:248:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a66892a8: inode [0x240000403:0x1241:0x0] mdc close failed: rc = -2 LustreError: 13199:0:(file.c:248:ll_close_inode_openhandle()) Skipped 36 previous similar messages traps: 12[31706] general protection ip:404df6 sp:7ffd7154cd88 error:0 in 12[400000+6000] 5[31971]: segfault at 8 ip 00007f1001cfe7e8 sp 00007ffc72bc1b50 error 4 in ld-2.17.so[7f1001cf3000+22000] Lustre: 11102:0:(mdt_reint.c:2484:mdt_reint_migrate()) lustre-MDT0000: [0x200000403:0x1:0x0]/6 is open, migrate only dentry 16[1650]: segfault at 0 ip (null) sp 00007fff2a42b5d8 error 14 in 16[400000+6000] LustreError: 21310:0:(mdd_dir.c:4759:mdd_migrate_cmd_check()) lustre-MDD0002: '18' migration was interrupted, run 'lfs migrate -m 0 -c 2 -H crush 18' to finish migration: rc = -1 5[3555]: segfault at 8 ip 00007f2d6d0307e8 sp 00007fffcefb6cb0 error 4 in ld-2.17.so[7f2d6d025000+22000] 19[5195]: segfault at 8 ip 00007ff4e905b7e8 sp 00007ffd3e555ad0 error 4 in ld-2.17.so[7ff4e9050000+22000] 8[5272]: segfault at 404592 ip 0000000000404914 sp 00007fffc4e680b8 error 7 in 8[400000+6000] 17[6170]: segfault at 8 ip 00007f27d58d67e8 sp 00007fff152cdd70 error 4 in ld-2.17.so[7f27d58cb000+22000] Lustre: dir [0x240000405:0xec6:0x0] stripe 2 readdir failed: -2, directory is partially accessed! Lustre: Skipped 18 previous similar messages LustreError: 8755:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000405:0x115a:0x0]: rc = -5 LustreError: 8755:0:(lcommon_cl.c:188:cl_file_inode_init()) Skipped 13 previous similar messages LustreError: 8755:0:(llite_lib.c:3770:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 8755:0:(llite_lib.c:3770:ll_prep_inode()) Skipped 13 previous similar messages 17[8050]: segfault at 8 ip 00007fc0679eb7e8 sp 00007ffdf4894390 error 4 in ld-2.17.so[7fc0679e0000+22000] 10[12325]: segfault at 8 ip 00007f756c3ed7e8 sp 00007ffdd9729110 error 4 in ld-2.17.so[7f756c3e2000+22000] LustreError: 22918:0:(lustre_lmv.h:500:lmv_is_sane()) unknown layout LMV: magic=0xcd40cd0 count=4 index=3 hash=crush:0x82000003 version=1 migrate_offset=3 migrate_hash=fnv_1a_64:2 pool= LustreError: 17163:0:(mdt_xattr.c:406:mdt_dir_layout_update()) lustre-MDT0002: [0x280000404:0x2a3f:0x0] migrate mdt count mismatch 3 != 2 0[18280]: segfault at 8 ip 00007f9c47d197e8 sp 00007ffea3067700 error 4 in ld-2.17.so[7f9c47d0e000+22000] 5[19318]: segfault at 8 ip 00007ff177b737e8 sp 00007ffea0df38c0 error 4 in ld-2.17.so[7ff177b68000+22000] 3[20005]: segfault at 8 ip 00007f17c0fce7e8 sp 00007ffe1915b7f0 error 4 in ld-2.17.so[7f17c0fc3000+22000] Lustre: dir [0x240000405:0x103a:0x0] stripe 3 readdir failed: -2, directory is partially accessed! Lustre: Skipped 15 previous similar messages LustreError: 4079:0:(mdd_orphans.c:261:mdd_orphan_delete()) lustre-MDD0002: could not delete orphan object [0x280000404:0x2adf:0x0]: rc = -2 LustreError: 4079:0:(mdd_object.c:3955:mdd_close()) lustre-MDD0002: unable to delete [0x280000404:0x2adf:0x0] from orphan list: rc = -2 16[23420]: segfault at 0 ip 0000000000403cf0 sp 00007ffcb17359b8 error 6 in 16[400000+6000] LustreError: 10797:0:(ldlm_lockd.c:257:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0002_UUID lock: ffff8802ac65cb40/0xe0176a8f8a5329b5 lrc: 3/0,0 mode: PR/PR res: [0x280000403:0x1a80:0x0].0x0 bits 0x1b/0x0 rrc: 3 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0xe0176a8f8a5328ea expref: 900 pid: 17008 timeout: 1084 lvb_type: 0 LustreError: 18984:0:(ldlm_lockd.c:1447:ldlm_handle_enqueue()) ### lock on destroyed export ffff8802e55cc138 ns: mdt-lustre-MDT0002_UUID lock: ffff880299994000/0xe0176a8f8a536f61 lrc: 3/0,0 mode: PR/PR res: [0x280000403:0x1:0x0].0x0 bits 0x12/0x0 rrc: 12 type: IBT gid 0 flags: 0x50200000000000 nid: 0@lo remote: 0xe0176a8f8a536f53 expref: 356 pid: 18984 timeout: 0 lvb_type: 0 LustreError: 18984:0:(ldlm_lockd.c:1447:ldlm_handle_enqueue()) Skipped 7 previous similar messages LustreError: lustre-MDT0002-mdc-ffff8800a66892a8: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-MDT0002-mdc-ffff8800a66892a8: Connection to lustre-MDT0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: lustre-MDT0002-mdc-ffff8800a66892a8: This client was evicted by lustre-MDT0002; in progress operations using this service will fail. LustreError: Skipped 4 previous similar messages LustreError: 26324:0:(llite_lib.c:2023:ll_md_setattr()) md_setattr fails: rc = -5 LustreError: 26324:0:(llite_lib.c:2023:ll_md_setattr()) Skipped 1 previous similar message LustreError: 26734:0:(file.c:6198:ll_inode_revalidate_fini()) lustre: revalidate FID [0x280000403:0x1:0x0] error: rc = -5 LustreError: 26734:0:(file.c:6198:ll_inode_revalidate_fini()) Skipped 4 previous similar messages LustreError: 23528:0:(mdc_request.c:1470:mdc_read_page()) lustre-MDT0002-mdc-ffff8800a66892a8: [0x280000400:0x58:0x0] lock enqueue fails: rc = -108 Lustre: dir [0x280000404:0x2a3f:0x0] stripe 0 readdir failed: -108, directory is partially accessed! Lustre: Skipped 1 previous similar message Lustre: lustre-MDT0002-mdc-ffff8800a66892a8: Connection restored to 192.168.123.107@tcp (at 0@lo) 6[4114]: segfault at 1c ip 00000000004017b0 sp 00007ffda04b1890 error 6 in 6[400000+6000] LustreError: 20036:0:(lustre_lmv.h:500:lmv_is_sane()) unknown layout LMV: magic=0xcd40cd0 count=4 index=3 hash=crush:0x82000003 version=1 migrate_offset=3 migrate_hash=fnv_1a_64:2 pool= 10[5703]: segfault at 8 ip 00007f66578e17e8 sp 00007ffd48df1580 error 4 in ld-2.17.so[7f66578d6000+22000] LustreError: 14888:0:(mdd_dir.c:4759:mdd_migrate_cmd_check()) lustre-MDD0000: '7' migration was interrupted, run 'lfs migrate -m 2 -c 2 -H crush 7' to finish migration: rc = -1 LustreError: 14888:0:(mdd_dir.c:4759:mdd_migrate_cmd_check()) Skipped 9 previous similar messages LustreError: 18986:0:(mdt_xattr.c:406:mdt_dir_layout_update()) lustre-MDT0002: [0x280000405:0x79:0x0] migrate mdt count mismatch 3 != 2 LustreError: 18206:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0001: failed to get lu_attr of [0x240000405:0xfff:0x0]: rc = -2 LustreError: 18206:0:(mdd_object.c:3901:mdd_close()) Skipped 4 previous similar messages Lustre: 7043:0:(lod_lov.c:1402:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000405:0x1561:0x0] with magic=0xbd60bd0 Lustre: 7043:0:(lod_lov.c:1402:lod_parse_striping()) Skipped 37 previous similar messages 5[8372]: segfault at 8 ip 00007faa79a327e8 sp 00007fffaab9d7f0 error 4 in ld-2.17.so[7faa79a27000+22000] 9[8437]: segfault at 8 ip 00007fc4a39c67e8 sp 00007ffdf7d96110 error 4 in ld-2.17.so[7fc4a39bb000+22000] LustreError: 13443:0:(lov_object.c:1350:lov_layout_change()) lustre-clilov-ffff8800a66892a8: cannot apply new layout on [0x280000403:0x1cf0:0x0] : rc = -5 LustreError: 13443:0:(vvp_io.c:1909:vvp_io_init()) lustre: refresh file layout [0x280000403:0x1cf0:0x0] error -5. LustreError: 11878:0:(lov_object.c:1350:lov_layout_change()) lustre-clilov-ffff8800a66892a8: cannot apply new layout on [0x280000405:0x207:0x0] : rc = -5 LustreError: 11878:0:(vvp_io.c:1909:vvp_io_init()) lustre: refresh file layout [0x280000405:0x207:0x0] error -5. Lustre: 21284:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 21284:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 2967 previous similar messages Lustre: 21284:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 21284:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 2967 previous similar messages Lustre: 21284:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 21284:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 2967 previous similar messages Lustre: 21284:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 21284:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 2967 previous similar messages Lustre: 21284:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 21284:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 2967 previous similar messages LustreError: 21318:0:(lov_object.c:1350:lov_layout_change()) lustre-clilov-ffff8800a66892a8: cannot apply new layout on [0x280000405:0x207:0x0] : rc = -5 LustreError: 22689:0:(lov_object.c:1350:lov_layout_change()) lustre-clilov-ffff8800a66892a8: cannot apply new layout on [0x200000405:0x16e2:0x0] : rc = -5 LustreError: 22689:0:(lov_object.c:1350:lov_layout_change()) Skipped 2 previous similar messages LustreError: 22689:0:(vvp_io.c:1909:vvp_io_init()) lustre: refresh file layout [0x200000405:0x16e2:0x0] error -5. LustreError: 22720:0:(lov_object.c:1350:lov_layout_change()) lustre-clilov-ffff8800a66892a8: cannot apply new layout on [0x240000403:0x1aff:0x0] : rc = -5 LustreError: 22720:0:(vvp_io.c:1909:vvp_io_init()) lustre: refresh file layout [0x240000403:0x1aff:0x0] error -5. LustreError: 18262:0:(lustre_lmv.h:500:lmv_is_sane()) unknown layout LMV: magic=0xcd40cd0 count=3 index=2 hash=crush:0x82000003 version=1 migrate_offset=2 migrate_hash=fnv_1a_64:2 pool= LustreError: 17830:0:(mdt_xattr.c:406:mdt_dir_layout_update()) lustre-MDT0002: [0x280000403:0x1fbe:0x0] migrate mdt count mismatch 2 != 1 4[31088]: segfault at 8 ip 00007f996f22f7e8 sp 00007fff54df40c0 error 4 in ld-2.17.so[7f996f224000+22000] Lustre: 18262:0:(mdt_reint.c:2484:mdt_reint_migrate()) lustre-MDT0000: [0x200000403:0x1:0x0]/10 is open, migrate only dentry ptlrpc_watchdog_fire: 18 callbacks suppressed Lustre: mdt00_033: service thread pid 7043 was inactive for 40.103 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Lustre: Skipped 2 previous similar messages Pid: 7043, comm: mdt00_033 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock+0x88/0x1c0 [mdt] [<0>] mdt_getattr_name_lock+0xc5a/0x2d20 [mdt] [<0>] mdt_intent_getattr+0x2cc/0x4e0 [mdt] [<0>] mdt_intent_opc.constprop.74+0x211/0xc60 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x397/0x980 [ptlrpc] Lustre: mdt00_000: service thread pid 11081 was inactive for 40.063 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 8 previous similar messages [<0>] ldlm_handle_enqueue+0x547/0x1900 [ptlrpc] [<0>] tgt_enqueue+0x68/0x240 [ptlrpc] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x23d/0xd80 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 17830, comm: mdt00_014 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock+0x88/0x1c0 [mdt] [<0>] mdt_getattr_name_lock+0xc5a/0x2d20 [mdt] [<0>] mdt_intent_getattr+0x2cc/0x4e0 [mdt] [<0>] mdt_intent_opc.constprop.74+0x211/0xc60 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x397/0x980 [ptlrpc] [<0>] ldlm_handle_enqueue+0x547/0x1900 [ptlrpc] [<0>] tgt_enqueue+0x68/0x240 [ptlrpc] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x23d/0xd80 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 17932, comm: mdt00_017 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock+0x88/0x1c0 [mdt] [<0>] mdt_getattr_name_lock+0xc5a/0x2d20 [mdt] [<0>] mdt_intent_getattr+0x2cc/0x4e0 [mdt] [<0>] mdt_intent_opc.constprop.74+0x211/0xc60 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x397/0x980 [ptlrpc] [<0>] ldlm_handle_enqueue+0x547/0x1900 [ptlrpc] [<0>] tgt_enqueue+0x68/0x240 [ptlrpc] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x23d/0xd80 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe LustreError: 10797:0:(ldlm_lockd.c:257:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0000_UUID lock: ffff8802a4b49a80/0xe0176a8f8a6fcc05 lrc: 3/0,0 mode: PR/PR res: [0x200000404:0x2f1f:0x0].0x0 bits 0x19/0x0 rrc: 3 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0xe0176a8f8a6fcbdb expref: 1248 pid: 14212 timeout: 1293 lvb_type: 0 Lustre: mdt00_011: service thread pid 17758 completed after 99.174s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_021: service thread pid 18984 completed after 99.220s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_017: service thread pid 17932 completed after 99.244s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_000: service thread pid 11081 completed after 99.219s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_027: service thread pid 30158 completed after 99.247s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_014: service thread pid 17830 completed after 99.222s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: lustre-MDT0000-mdc-ffff8800a66892a8: operation mds_close to node 0@lo failed: rc = -107 Lustre: lustre-MDT0000-mdc-ffff8800a66892a8: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: mdt00_035: service thread pid 27923 completed after 99.252s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 11822:0:(ldlm_lockd.c:2550:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1752245394 with bad export cookie 16147492153683927162 Lustre: mdt00_033: service thread pid 7043 completed after 99.289s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: lustre-MDT0000-mdc-ffff8800a66892a8: This client was evicted by lustre-MDT0000; in progress operations using this service will fail. LustreError: 31603:0:(vvp_io.c:1909:vvp_io_init()) lustre: refresh file layout [0x200000404:0x2f1f:0x0] error -108. LustreError: 31798:0:(file.c:6198:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000007:0x1:0x0] error: rc = -108 LustreError: 31798:0:(file.c:6198:ll_inode_revalidate_fini()) Skipped 21 previous similar messages LustreError: 31877:0:(mdc_request.c:1470:mdc_read_page()) lustre-MDT0000-mdc-ffff8800a66892a8: [0x200000403:0x1:0x0] lock enqueue fails: rc = -108 LustreError: 31877:0:(mdc_request.c:1470:mdc_read_page()) Skipped 6 previous similar messages Lustre: lustre-MDT0000-mdc-ffff8800a66892a8: Connection restored to 192.168.123.107@tcp (at 0@lo) Lustre: 19797:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 503 < left 2168, rollback = 2 Lustre: 19797:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 3074 previous similar messages Lustre: 18262:0:(mdd_dir.c:4838:mdd_migrate_object()) lustre-MDD0000: [0x200000403:0x2:0x0]/16 is open, migrate only dentry Lustre: 18262:0:(mdd_dir.c:4838:mdd_migrate_object()) Skipped 91 previous similar messages 3[1795]: segfault at 8 ip 00007f8ab02c97e8 sp 00007ffd6d194910 error 4 in ld-2.17.so[7f8ab02be000+22000] Lustre: 32728:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-OST0000: opcode 7: before 516 < left 618, rollback = 7 Lustre: 32728:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 291 previous similar messages 10[3273]: segfault at 8 ip 00007fa8876717e8 sp 00007ffc49690f00 error 4 in ld-2.17.so[7fa887666000+22000] LustreError: 12863:0:(mdt_xattr.c:406:mdt_dir_layout_update()) lustre-MDT0002: [0x280000405:0x615:0x0] migrate mdt count mismatch 1 != 2 2[3259]: segfault at 8 ip 00007efc0ea717e8 sp 00007ffd8d392910 error 4 in ld-2.17.so[7efc0ea66000+22000] LustreError: 17758:0:(mdt_xattr.c:406:mdt_dir_layout_update()) lustre-MDT0002: [0x280000405:0xf5:0x0] migrate mdt count mismatch 2 != 1 LustreError: 11082:0:(mdt_xattr.c:406:mdt_dir_layout_update()) lustre-MDT0001: [0x240000403:0x1be3:0x0] migrate mdt count mismatch 2 != 3 LustreError: 12088:0:(lov_object.c:1350:lov_layout_change()) lustre-clilov-ffff8800a66892a8: cannot apply new layout on [0x200000405:0x1dcb:0x0] : rc = -5 LustreError: 12088:0:(vvp_io.c:1909:vvp_io_init()) lustre: refresh file layout [0x200000405:0x1dcb:0x0] error -5. LustreError: 19809:0:(mdd_dir.c:4759:mdd_migrate_cmd_check()) lustre-MDD0002: '1' migration was interrupted, run 'lfs migrate -m 1 -c 2 -H crush 1' to finish migration: rc = -1 LustreError: 19809:0:(mdd_dir.c:4759:mdd_migrate_cmd_check()) Skipped 20 previous similar messages 19[19099]: segfault at 8 ip 00007ffb753ae7e8 sp 00007ffd449ca640 error 4 in ld-2.17.so[7ffb753a3000+22000] 19[19356]: segfault at 8 ip 00007f642e0b37e8 sp 00007fff9f5486e0 error 4 in ld-2.17.so[7f642e0a8000+22000] 9[20994]: segfault at 8 ip 00007f0d3a0437e8 sp 00007ffefbf1afe0 error 4 in ld-2.17.so[7f0d3a038000+22000] LustreError: 25261:0:(lov_object.c:1350:lov_layout_change()) lustre-clilov-ffff8800a668ae98: cannot apply new layout on [0x280000405:0x943:0x0] : rc = -5 LustreError: 25261:0:(lov_object.c:1350:lov_layout_change()) Skipped 3 previous similar messages LustreError: 25261:0:(vvp_io.c:1909:vvp_io_init()) lustre: refresh file layout [0x280000405:0x943:0x0] error -5. LustreError: 25261:0:(vvp_io.c:1909:vvp_io_init()) Skipped 1 previous similar message LustreError: 11086:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0001: failed to get lu_attr of [0x240000403:0x20be:0x0]: rc = -2 LustreError: 11086:0:(mdd_object.c:3901:mdd_close()) Skipped 13 previous similar messages LustreError: 19809:0:(mdt_reint.c:2564:mdt_reint_migrate()) lustre-MDT0001: migrate [0x280000403:0x253c:0x0]/19 failed: rc = -2 LustreError: 19809:0:(mdt_reint.c:2564:mdt_reint_migrate()) Skipped 92 previous similar messages 16[30710]: segfault at 0 ip (null) sp 00007ffc811cd228 error 14 in 16[400000+6000] 4[30681]: segfault at 8 ip 00007f488d9837e8 sp 00007ffed5de50e0 error 4 in ld-2.17.so[7f488d978000+22000] LustreError: 30560:0:(file.c:248:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a668ae98: inode [0x280000405:0x9dd:0x0] mdc close failed: rc = -2 LustreError: 30560:0:(file.c:248:ll_close_inode_openhandle()) Skipped 77 previous similar messages LustreError: 18066:0:(mdt_xattr.c:406:mdt_dir_layout_update()) lustre-MDT0000: [0x200000406:0x62c:0x0] migrate mdt count mismatch 2 != 3 LustreError: 31480:0:(lov_object.c:1350:lov_layout_change()) lustre-clilov-ffff8800a66892a8: cannot apply new layout on [0x240000403:0x2278:0x0] : rc = -5 LustreError: 31480:0:(lov_object.c:1350:lov_layout_change()) Skipped 3 previous similar messages LustreError: 31480:0:(vvp_io.c:1909:vvp_io_init()) lustre: refresh file layout [0x240000403:0x2278:0x0] error -5. LustreError: 31480:0:(vvp_io.c:1909:vvp_io_init()) Skipped 1 previous similar message Lustre: dir [0x200000406:0x744:0x0] stripe 3 readdir failed: -2, directory is partially accessed! Lustre: Skipped 15 previous similar messages LustreError: 7555:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x280000405:0x95a:0x0]: rc = -5 LustreError: 7555:0:(lcommon_cl.c:188:cl_file_inode_init()) Skipped 190 previous similar messages LustreError: 7555:0:(llite_lib.c:3770:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 7555:0:(llite_lib.c:3770:ll_prep_inode()) Skipped 190 previous similar messages 10[7620]: segfault at 8 ip 00007f01e3f347e8 sp 00007fff0adf6a10 error 4 in ld-2.17.so[7f01e3f29000+22000] LustreError: 16933:0:(lov_object.c:1350:lov_layout_change()) lustre-clilov-ffff8800a668ae98: cannot apply new layout on [0x280000405:0x943:0x0] : rc = -5 LustreError: 16933:0:(lov_object.c:1350:lov_layout_change()) Skipped 2 previous similar messages 16[17395]: segfault at 8 ip 00007fd8c007e7e8 sp 00007fff320994c0 error 4 in ld-2.17.so[7fd8c0073000+22000] 10[17477]: segfault at 0 ip (null) sp 00007fff9f3fc2c8 error 14 in 10[400000+6000] LustreError: 12268:0:(vvp_io.c:1909:vvp_io_init()) lustre: refresh file layout [0x240000405:0x1f28:0x0] error -5. LustreError: 158:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 18 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 LustreError: 158:0:(statahead.c:825:ll_statahead_interpret_work()) Skipped 1 previous similar message LustreError: 174:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 18 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 LustreError: 10797:0:(ldlm_lockd.c:257:expired_lock_main()) ### lock callback timer expired after 101s: evicting client at 0@lo ns: mdt-lustre-MDT0002_UUID lock: ffff8802abc39e40/0xe0176a8f8aa4d8d7 lrc: 3/0,0 mode: CR/CR res: [0x280000403:0x2a7d:0x0].0x0 bits 0xa/0x0 rrc: 4 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0xe0176a8f8aa4d87c expref: 727 pid: 11082 timeout: 1626 lvb_type: 0 LustreError: lustre-MDT0002-mdc-ffff8800a668ae98: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-MDT0002-mdc-ffff8800a668ae98: Connection to lustre-MDT0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: lustre-MDT0002-mdc-ffff8800a668ae98: This client was evicted by lustre-MDT0002; in progress operations using this service will fail. LustreError: 26853:0:(llite_lib.c:2023:ll_md_setattr()) md_setattr fails: rc = -108 LustreError: 26814:0:(file.c:6198:ll_inode_revalidate_fini()) lustre: revalidate FID [0x280000403:0x29cd:0x0] error: rc = -108 LustreError: 26814:0:(file.c:6198:ll_inode_revalidate_fini()) Skipped 15 previous similar messages Lustre: lustre-MDT0002-mdc-ffff8800a668ae98: Connection restored to 192.168.123.107@tcp (at 0@lo) Lustre: 17252:0:(lod_lov.c:1402:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000406:0xd74:0x0] with magic=0xbd60bd0 Lustre: 17252:0:(lod_lov.c:1402:lod_parse_striping()) Skipped 165 previous similar messages LustreError: 30399:0:(lov_object.c:1350:lov_layout_change()) lustre-clilov-ffff8800a668ae98: cannot apply new layout on [0x280000405:0x943:0x0] : rc = -5 LustreError: 30399:0:(lov_object.c:1350:lov_layout_change()) Skipped 16 previous similar messages LustreError: 406:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 5 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 13[31518]: segfault at 8 ip 00007fbae7bed7e8 sp 00007ffe95976680 error 4 in ld-2.17.so[7fbae7be2000+22000] 13[9123]: segfault at 8 ip 00007fc743a067e8 sp 00007fffd1642810 error 4 in ld-2.17.so[7fc7439fb000+22000] 19[25134]: segfault at 8 ip 00007f5c96e387e8 sp 00007ffe51e1a100 error 4 in ld-2.17.so[7f5c96e2d000+22000] 11[24772]: segfault at 0 ip (null) sp 00007ffdc3da5be8 error 14 in 11[400000+6000] 9[31033]: segfault at 8 ip 00007f5e49a087e8 sp 00007fff07f71d90 error 4 in ld-2.17.so[7f5e499fd000+22000] 11[31862]: segfault at 8 ip 00007efca45137e8 sp 00007ffca5b4a2c0 error 4 in ld-2.17.so[7efca4508000+22000] | Link to test |
racer test 1: racer on clients: centos-45.localnet DURATION=2700 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 PGD 2bb379067 PUD 2984c4067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) lustre(OE) osp(OE) ofd(OE) lod(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) jbd2 mbcache lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) crc32_generic lnet(OE) dm_flakey dm_mod libcfs(OE) crc_t10dif crct10dif_generic crct10dif_common virtio_console virtio_balloon i2c_piix4 pcspkr ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm drm_panel_orientation_quirks ata_piix serio_raw virtio_blk floppy i2c_core libata CPU: 5 PID: 8983 Comm: ll_sa_8797 Kdump: loaded Tainted: P OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.16.0-3.module_el8.7.0+1218+f626c2ff 04/01/2014 task: ffff88031b8e3760 ti: ffff8802b480c000 task.ti: ffff8802b480c000 RIP: 0010:[<ffffffff813f502c>] [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 RSP: 0018:ffff8802b480fd78 EFLAGS: 00010296 RAX: ffff88008f614d10 RBX: 0000000000000008 RCX: 0000000100260024 RDX: 0000000000000026 RSI: ffff88008f615040 RDI: 0000000000000008 RBP: ffff8802b480fd88 R08: ffff8802ed703ef8 R09: 0000000000000000 R10: ffff8802ed7009f8 R11: ffff8802ed703ef8 R12: 0000000000000000 R13: ffff88031b8e0010 R14: ffff88008f615040 R15: ffff8802ed703ef8 FS: 0000000000000000(0000) GS:ffff880331b40000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 00000002d8a0e000 CR4: 00000000000007e0 Call Trace: [<ffffffffa1127791>] ll_statahead_thread+0xdd1/0x2020 [lustre] [<ffffffff817e320e>] ? _raw_spin_lock+0x1e/0x20 [<ffffffff810dbb51>] ? put_prev_entity+0x31/0x400 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e329e>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffff810c834d>] ? finish_task_switch+0x5d/0x1b0 [<ffffffffa11269c0>] ? ll_statahead_handle.constprop.30+0x1c0/0x1c0 [lustre] [<ffffffff810ba114>] kthread+0xe4/0xf0 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 | Lustre: 15457:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff8800a8576940 x1837073806699648/t4294968442(0) o101->6abfd204-2290-4b04-b77a-72a76796ca30@0@lo:219/0 lens 376/864 e 0 to 0 dl 1751970169 ref 1 fl Interpret:H/602/0 rc 0/0 job:'dd.0' uid:0 gid:0 projid:0 Lustre: 10481:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-OST0002: opcode 7: before 516 < left 618, rollback = 7 Lustre: 10481:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 10481:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 10481:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 10481:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 10481:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 10481:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 516 < left 558, rollback = 7 Lustre: 10481:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 1 previous similar message Lustre: 10481:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 10481:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 10481:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 10481:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 10481:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/558/0, punch: 0/0/0, quota 1/3/0 Lustre: 10481:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 10481:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 10481:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 10481:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 10481:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 15457:0:(lod_lov.c:1402:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000401:0x28b:0x0] with magic=0xbd60bd0 11[24401]: segfault at 0 ip (null) sp 00007ffe3bda1058 error 14 in 11[400000+6000] 5[25432]: segfault at 8 ip 00007f3b4b3ad7e8 sp 00007fffcb536e90 error 4 in ld-2.17.so[7f3b4b3a2000+22000] Lustre: 15118:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 618, rollback = 7 Lustre: 15118:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 1 previous similar message Lustre: 15118:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 15118:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 15118:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 15118:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 15118:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 15118:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 15118:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 15118:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 15118:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 15118:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 15118:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 516 < left 618, rollback = 7 Lustre: 15118:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 3 previous similar messages Lustre: 15118:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 15118:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 15118:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 15118:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 15118:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 15118:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 15118:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 15118:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 15118:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 15118:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 10478:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 516 < left 618, rollback = 7 Lustre: 10478:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 5 previous similar messages Lustre: 10478:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 10478:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 10478:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 10478:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 10478:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 10478:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 10478:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 10478:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 10478:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 10478:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 5 previous similar messages 2[29021]: segfault at 8 ip 00007f9ab52867e8 sp 00007fff1c619ff0 error 4 in ld-2.17.so[7f9ab527b000+22000] 0[31981]: segfault at 8 ip 00007f5dc0b677e8 sp 00007ffce809fd30 error 4 in ld-2.17.so[7f5dc0b5c000+22000] Lustre: 10478:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 515 < left 618, rollback = 7 Lustre: 10478:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 7 previous similar messages Lustre: 10478:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 10478:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 7 previous similar messages Lustre: 10478:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/1, xattr_set: 2/15/0 Lustre: 10478:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 7 previous similar messages Lustre: 10478:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 10478:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 7 previous similar messages Lustre: 10478:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 10478:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 7 previous similar messages Lustre: 10478:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 10478:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 7 previous similar messages 11[2329]: segfault at 8 ip 00007f06c9c947e8 sp 00007ffca119a6c0 error 4 in ld-2.17.so[7f06c9c89000+22000] Lustre: 16885:0:(lod_lov.c:1402:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000401:0x6bf:0x0] with magic=0xbd60bd0 Lustre: 16885:0:(lod_lov.c:1402:lod_parse_striping()) Skipped 1 previous similar message Lustre: 10481:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-OST0000: opcode 7: before 516 < left 618, rollback = 7 Lustre: 10481:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 5 previous similar messages Lustre: 10481:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 10481:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 10481:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 10481:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 10481:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 10481:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 10481:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 10481:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 10481:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 10481:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 9645:0:(lod_lov.c:1402:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000402:0x710:0x0] with magic=0xbd60bd0 Lustre: 9645:0:(lod_lov.c:1402:lod_parse_striping()) Skipped 1 previous similar message 0[6265]: segfault at 8 ip 00007fc4e8aa57e8 sp 00007fff239ec220 error 4 in ld-2.17.so[7fc4e8a9a000+22000] 7[6855]: segfault at 8 ip 00007faaac3ae7e8 sp 00007ffddcd3c7d0 error 4 in ld-2.17.so[7faaac3a3000+22000] 16[9886]: segfault at 8 ip 00007f87aa6b57e8 sp 00007ffd73be4470 error 4 in ld-2.17.so[7f87aa6aa000+22000] Lustre: 10481:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-OST0000: opcode 7: before 516 < left 618, rollback = 7 Lustre: 10481:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 21 previous similar messages Lustre: 10481:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 10481:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 21 previous similar messages Lustre: 16270:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 16270:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 23 previous similar messages Lustre: 16270:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 16270:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 23 previous similar messages Lustre: 16270:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 16270:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 23 previous similar messages Lustre: 16270:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 16270:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 23 previous similar messages Lustre: 15625:0:(lod_lov.c:1402:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000401:0xbbc:0x0] with magic=0xbd60bd0 Lustre: 15625:0:(lod_lov.c:1402:lod_parse_striping()) Skipped 1 previous similar message 9[23884]: segfault at 8 ip 00007f9ac61107e8 sp 00007ffc2b0c6ef0 error 4 in ld-2.17.so[7f9ac6105000+22000] Lustre: 15625:0:(lod_lov.c:1402:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000402:0xe17:0x0] with magic=0xbd60bd0 Lustre: 15625:0:(lod_lov.c:1402:lod_parse_striping()) Skipped 1 previous similar message Lustre: 16270:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-OST0002: opcode 7: before 516 < left 618, rollback = 7 Lustre: 16270:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 89 previous similar messages Lustre: 16270:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 16270:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 89 previous similar messages Lustre: 10478:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 10478:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 91 previous similar messages Lustre: 10478:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 10478:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 91 previous similar messages Lustre: 10478:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 10478:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 91 previous similar messages Lustre: 10478:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 10478:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 91 previous similar messages Lustre: 15457:0:(lod_lov.c:1402:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000402:0xf78:0x0] with magic=0xbd60bd0 Lustre: 15457:0:(lod_lov.c:1402:lod_parse_striping()) Skipped 1 previous similar message 7[31699]: segfault at 8 ip 00007fc7172cf7e8 sp 00007fffd23eb7d0 error 4 in ld-2.17.so[7fc7172c4000+22000] 18[987]: segfault at 1c ip 00000000004017b0 sp 00007fff666d0770 error 6 in 18[400000+6000] Lustre: 9643:0:(lod_lov.c:1402:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000402:0x11af:0x0] with magic=0xbd60bd0 Lustre: 9643:0:(lod_lov.c:1402:lod_parse_striping()) Skipped 1 previous similar message 17[18571]: segfault at 0 ip 0000000000403e5f sp 00007ffd4a856980 error 6 in 17[400000+6000] Lustre: 16055:0:(lod_lov.c:1402:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000401:0x16ee:0x0] with magic=0xbd60bd0 Lustre: 16055:0:(lod_lov.c:1402:lod_parse_striping()) Skipped 5 previous similar messages LustreError: 20564:0:(file.c:248:ll_close_inode_openhandle()) lustre-clilmv-ffff8802ed4c37e8: inode [0x200000402:0x18cb:0x0] mdc close failed: rc = -13 5[28269]: segfault at 0 ip 0000000000403e5f sp 00007fff68391eb0 error 6 in 2[400000+6000] 1[3790]: segfault at 0 ip 0000000000403e5f sp 00007ffc70ac5680 error 6 in 1[400000+6000] Lustre: mdt00_005: service thread pid 14983 was inactive for 40.078 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 14983, comm: mdt00_005 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock_try+0xa0/0x250 [mdt] [<0>] mdt_getattr_name_lock+0x1739/0x2d20 [mdt] [<0>] mdt_intent_getattr+0x2cc/0x4e0 [mdt] [<0>] mdt_intent_opc.constprop.74+0x211/0xc60 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x397/0x980 [ptlrpc] [<0>] ldlm_handle_enqueue+0x547/0x1900 [ptlrpc] [<0>] tgt_enqueue+0x68/0x240 [ptlrpc] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x23d/0xd80 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 15332, comm: mdt00_006 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock+0x88/0x1c0 [mdt] [<0>] mdt_reint_link+0x8fb/0xff0 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x73c/0xbb0 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x23d/0xd80 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 32422, comm: mdt00_014 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock_try+0xa0/0x250 [mdt] [<0>] mdt_getattr_name_lock+0x1739/0x2d20 [mdt] [<0>] mdt_intent_getattr+0x2cc/0x4e0 [mdt] [<0>] mdt_intent_opc.constprop.74+0x211/0xc60 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x397/0x980 [ptlrpc] [<0>] ldlm_handle_enqueue+0x547/0x1900 [ptlrpc] [<0>] tgt_enqueue+0x68/0x240 [ptlrpc] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x23d/0xd80 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: mdt00_000: service thread pid 9643 was inactive for 40.058 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. LustreError: 9314:0:(ldlm_lockd.c:257:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0000_UUID lock: ffff8802c5e3b880/0xd06913fc8f3fc9c3 lrc: 3/0,0 mode: PR/PR res: [0x200000402:0x2265:0x0].0x0 bits 0x1b/0x0 rrc: 6 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0xd06913fc8f3fc999 expref: 1044 pid: 31285 timeout: 483 lvb_type: 0 Lustre: mdt00_006: service thread pid 15332 completed after 100.399s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_014: service thread pid 32422 completed after 100.364s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_005: service thread pid 14983 completed after 100.387s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 9643:0:(ldlm_lockd.c:1447:ldlm_handle_enqueue()) ### lock on destroyed export ffff8800a9e053d8 ns: mdt-lustre-MDT0000_UUID lock: ffff8802bbd796c0/0xd06913fc8f3fced9 lrc: 3/0,0 mode: PR/PR res: [0x200000402:0x2153:0x0].0x0 bits 0x13/0x0 rrc: 4 type: IBT gid 0 flags: 0x50200000000000 nid: 0@lo remote: 0xd06913fc8f3fce69 expref: 136 pid: 9643 timeout: 0 lvb_type: 0 Lustre: mdt00_000: service thread pid 9643 completed after 98.447s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: lustre-MDT0000-mdc-ffff8802ee839bf8: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-MDT0000-mdc-ffff8802ee839bf8: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: lustre-MDT0000-mdc-ffff8802ee839bf8: This client was evicted by lustre-MDT0000; in progress operations using this service will fail. LustreError: 13478:0:(file.c:6198:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000402:0x2153:0x0] error: rc = -5 LustreError: 13500:0:(file.c:248:ll_close_inode_openhandle()) lustre-clilmv-ffff8802ee839bf8: inode [0x200000401:0x1f62:0x0] mdc close failed: rc = -108 LustreError: 13500:0:(ldlm_resource.c:981:ldlm_resource_complain()) lustre-MDT0000-mdc-ffff8802ee839bf8: namespace resource [0x200000007:0x1:0x0].0x0 (ffff8802ddfd5940) refcount nonzero (2) after lock cleanup; forcing cleanup. Lustre: lustre-MDT0000-mdc-ffff8802ee839bf8: Connection restored to 192.168.123.47@tcp (at 0@lo) 14[13483]: segfault at 0 ip (null) sp 00007ffed8dc5ab8 error 14 in 14[400000+6000] Lustre: 10481:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-OST0002: opcode 7: before 516 < left 618, rollback = 7 Lustre: 10481:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 191 previous similar messages Lustre: 10481:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 10481:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 191 previous similar messages Lustre: 10481:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 10481:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 187 previous similar messages Lustre: 10481:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 10481:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 187 previous similar messages Lustre: 10481:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 10481:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 187 previous similar messages Lustre: 10481:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 10481:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 187 previous similar messages Lustre: 16885:0:(lod_lov.c:1402:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0x140:0x0] with magic=0xbd60bd0 Lustre: 16885:0:(lod_lov.c:1402:lod_parse_striping()) Skipped 1 previous similar message 12[21571]: segfault at 8 ip 00007ff3cbb587e8 sp 00007ffd0a0332e0 error 4 in ld-2.17.so[7ff3cbb4d000+22000] 3[21786]: segfault at 8 ip 00007f57755097e8 sp 00007fff53398610 error 4 in ld-2.17.so[7f57754fe000+22000] 12[26795]: segfault at 8 ip 00007f9ea39897e8 sp 00007ffded5f94e0 error 4 in ld-2.17.so[7f9ea397e000+22000] 8[27922]: segfault at 8 ip 00007f37adf467e8 sp 00007ffc92d6d980 error 4 in ld-2.17.so[7f37adf3b000+22000] 7[2214]: segfault at 0 ip (null) sp 00007fff57c64b28 error 14 in 7[400000+6000] traps: 2[6959] trap invalid opcode ip:405152 sp:7ffe38671968 error:0 in 2[400000+6000] 8[17642]: segfault at 8 ip 00007f3f764b17e8 sp 00007ffcc56a2930 error 4 in ld-2.17.so[7f3f764a6000+22000] 13[17489]: segfault at 8 ip 00007f8ed25567e8 sp 00007ffeced6ded0 error 4 in ld-2.17.so[7f8ed254b000+22000] 6[24778]: segfault at 8 ip 00007f9c43e0f7e8 sp 00007ffc5b8893e0 error 4 in ld-2.17.so[7f9c43e04000+22000] LustreError: 25875:0:(file.c:248:ll_close_inode_openhandle()) lustre-clilmv-ffff8802ed4c37e8: inode [0x200000403:0x105c:0x0] mdc close failed: rc = -13 LustreError: 25875:0:(file.c:248:ll_close_inode_openhandle()) Skipped 8 previous similar messages LustreError: 30869:0:(file.c:248:ll_close_inode_openhandle()) lustre-clilmv-ffff8802ed4c37e8: inode [0x200000402:0x3370:0x0] mdc close failed: rc = -13 4[10323]: segfault at 8 ip 00007ff9840e17e8 sp 00007ffcf95cb740 error 4 in ld-2.17.so[7ff9840d6000+22000] LustreError: 10434:0:(file.c:248:ll_close_inode_openhandle()) lustre-clilmv-ffff8802ed4c37e8: inode [0x200000403:0x16ef:0x0] mdc close failed: rc = -13 LustreError: 15598:0:(file.c:248:ll_close_inode_openhandle()) lustre-clilmv-ffff8802ed4c37e8: inode [0x200000403:0x18c4:0x0] mdc close failed: rc = -13 4[29340]: segfault at 0 ip 0000000000403e5f sp 00007ffe77cab120 error 6 in 4[400000+6000] 11[32490]: segfault at 8 ip 00007f243be437e8 sp 00007ffe02e739f0 error 4 in ld-2.17.so[7f243be38000+22000] LustreError: 32766:0:(file.c:248:ll_close_inode_openhandle()) lustre-clilmv-ffff8802ed4c37e8: inode [0x200000402:0x4071:0x0] mdc close failed: rc = -13 Lustre: mdt_io00_004: service thread pid 5027 was inactive for 40.033 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. LustreError: 9314:0:(ldlm_lockd.c:257:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0003_UUID lock: ffff8802d85e8b80/0xd06913fc8f73e69d lrc: 3/0,0 mode: PW/PW res: [0x300000400:0x44a:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000000020020 nid: 0@lo remote: 0xd06913fc8f73e688 expref: 15 pid: 10467 timeout: 746 lvb_type: 0 LustreError: lustre-OST0003-osc-ffff8802ee839bf8: operation ost_sync to node 0@lo failed: rc = -107 Lustre: lustre-OST0003-osc-ffff8802ee839bf8: Connection to lustre-OST0003 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: lustre-OST0003-osc-ffff8802ee839bf8: This client was evicted by lustre-OST0003; in progress operations using this service will fail. Lustre: 3740:0:(llite_lib.c:4215:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.47@tcp:/lustre/fid: [0x200000402:0x3c65:0x0]/ may get corrupted (rc -108) Lustre: lustre-OST0003-osc-ffff8802ee839bf8: Connection restored to 192.168.123.47@tcp (at 0@lo) Lustre: lustre-OST0003-osc-ffff8802ee839bf8: disconnect after 24s idle LustreError: 9314:0:(ldlm_lockd.c:257:expired_lock_main()) ### lock callback timer expired after 101s: evicting client at 0@lo ns: mdt-lustre-MDT0000_UUID lock: ffff8802d6d4d2c0/0xd06913fc8f7ce6d8 lrc: 3/0,0 mode: PR/PR res: [0x200000403:0x1eec:0x0].0x0 bits 0x1b/0x0 rrc: 8 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0xd06913fc8f7ce6ae expref: 788 pid: 13028 timeout: 775 lvb_type: 0 LustreError: 15457:0:(ldlm_lockd.c:1447:ldlm_handle_enqueue()) ### lock on destroyed export ffff8802d9c392a8 ns: mdt-lustre-MDT0000_UUID lock: ffff8802d622d680/0xd06913fc8f7cf16d lrc: 1/0,0 mode: --/PR res: [0x200000403:0x1eec:0x0].0x0 bits 0x1b/0x0 rrc: 6 type: IBT gid 0 flags: 0x54a01000000000 nid: 0@lo remote: 0xd06913fc8f7cf151 expref: 598 pid: 15457 timeout: 0 lvb_type: 0 LustreError: lustre-MDT0000-mdc-ffff8802ee839bf8: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-MDT0000-mdc-ffff8802ee839bf8: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: mdt_io00_004: service thread pid 5027 completed after 100.836s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: lustre-MDT0000-mdc-ffff8802ee839bf8: This client was evicted by lustre-MDT0000; in progress operations using this service will fail. LustreError: 15457:0:(ldlm_lockd.c:1447:ldlm_handle_enqueue()) Skipped 4 previous similar messages LustreError: 618:0:(file.c:6198:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000401:0x1:0x0] error: rc = -5 LustreError: 618:0:(file.c:6198:ll_inode_revalidate_fini()) Skipped 12 previous similar messages LustreError: 685:0:(file.c:248:ll_close_inode_openhandle()) lustre-clilmv-ffff8802ee839bf8: inode [0x200000401:0x1:0x0] mdc close failed: rc = -108 LustreError: 331:0:(vvp_io.c:1909:vvp_io_init()) lustre: refresh file layout [0x200000403:0x1eec:0x0] error -108. Lustre: lustre-MDT0000-mdc-ffff8802ee839bf8: Connection restored to 192.168.123.47@tcp (at 0@lo) Lustre: 16270:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-OST0002: opcode 7: before 516 < left 570, rollback = 7 Lustre: 16270:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 331 previous similar messages Lustre: 16270:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 16270:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 331 previous similar messages Lustre: 16270:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 16270:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 331 previous similar messages Lustre: 16270:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/570/0, punch: 0/0/0, quota 1/3/0 Lustre: 16270:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 331 previous similar messages Lustre: 16270:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 16270:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 331 previous similar messages Lustre: 16270:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 16270:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 331 previous similar messages 14[15050]: segfault at 8 ip 00007f3dcfcc87e8 sp 00007ffda0a016d0 error 4 in ld-2.17.so[7f3dcfcbd000+22000] Lustre: 32422:0:(lod_lov.c:1402:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000402:0x45d7:0x0] with magic=0xbd60bd0 Lustre: 32422:0:(lod_lov.c:1402:lod_parse_striping()) Skipped 13 previous similar messages 11[19461]: segfault at 8 ip 00007f43974067e8 sp 00007ffe4a2e2910 error 4 in ld-2.17.so[7f43973fb000+22000] 16[22157]: segfault at 8 ip 00007f88d89f387a sp 00007ffd555ac640 error 4 in ld-2.17.so[7f88d89e8000+22000] 5[8478]: segfault at 0 ip 0000000000403e5f sp 00007ffc2bbb8680 error 6 in 5[400000+6000] LustreError: 9827:0:(file.c:248:ll_close_inode_openhandle()) lustre-clilmv-ffff8802ed4c37e8: inode [0x200000402:0x4f06:0x0] mdc close failed: rc = -13 LustreError: 9827:0:(file.c:248:ll_close_inode_openhandle()) Skipped 9 previous similar messages Lustre: lustre-OST0000-osc-ffff8802ed4c37e8: disconnect after 22s idle Lustre: lustre-OST0003-osc-ffff8802ed4c37e8: disconnect after 22s idle LustreError: 9314:0:(ldlm_lockd.c:257:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0000_UUID lock: ffff8802c64752c0/0xd06913fc8f9e0ab5 lrc: 3/0,0 mode: CR/CR res: [0x200000404:0x1018:0x0].0x0 bits 0xa/0x0 rrc: 5 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0xd06913fc8f9e0a92 expref: 1421 pid: 32422 timeout: 974 lvb_type: 0 LustreError: lustre-MDT0000-mdc-ffff8802ed4c37e8: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 2 previous similar messages Lustre: lustre-MDT0000-mdc-ffff8802ed4c37e8: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 10498:0:(ldlm_lockd.c:2550:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1751971043 with bad export cookie 15017556407895070725 LustreError: 21291:0:(client.c:1375:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802ddcbf340 x1837074016313216/t0(0) o104->lustre-MDT0000@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' uid:4294967295 gid:4294967295 projid:4294967295 LustreError: lustre-MDT0000-mdc-ffff8802ed4c37e8: This client was evicted by lustre-MDT0000; in progress operations using this service will fail. LustreError: 14634:0:(vvp_io.c:1909:vvp_io_init()) lustre: refresh file layout [0x200000404:0x1018:0x0] error -108. LustreError: 14526:0:(file.c:6198:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000404:0xc38:0x0] error: rc = -108 LustreError: 14526:0:(file.c:6198:ll_inode_revalidate_fini()) Skipped 37 previous similar messages LustreError: 14604:0:(mdc_request.c:1470:mdc_read_page()) lustre-MDT0000-mdc-ffff8802ed4c37e8: [0x200000401:0x1:0x0] lock enqueue fails: rc = -108 Lustre: lustre-MDT0000-mdc-ffff8802ed4c37e8: Connection restored to 192.168.123.47@tcp (at 0@lo) 3[25612]: segfault at 8 ip 00007f63da07d7e8 sp 00007ffdaa0dec90 error 4 in ld-2.17.so[7f63da072000+22000] 1[27758]: segfault at 1c ip 00000000004017b0 sp 00007fff105732f0 error 6 in 1[400000+6000] 18[29132]: segfault at 8 ip 00007f1c88faf7e8 sp 00007ffc16cc4200 error 4 in ld-2.17.so[7f1c88fa4000+22000] 0[29386]: segfault at 8 ip 00007f115ffed7e8 sp 00007ffe783b4b00 error 4 in ld-2.17.so[7f115ffe2000+22000] 0[30659]: segfault at 8 ip 00007f16d1bf07e8 sp 00007ffc4d16c000 error 4 in ld-2.17.so[7f16d1be5000+22000] 9[2340]: segfault at 0 ip 0000000000403e5f sp 00007ffd4cd3f4b0 error 6 in 9[400000+6000] 18[4272]: segfault at 8 ip 00007f4abe85a7e8 sp 00007ffd6a389bd0 error 4 in ld-2.17.so[7f4abe84f000+22000] 9[5311]: segfault at 6fa86d7d ip 0000000000404d03 sp 00007ffd9f081330 error 6 in 9[400000+6000] 10[8055]: segfault at 0 ip 0000000000403e5f sp 00007ffe1c5975c0 error 6 in 10[400000+6000] LustreError: 13713:0:(file.c:248:ll_close_inode_openhandle()) lustre-clilmv-ffff8802ee839bf8: inode [0x200000404:0x1bc6:0x0] mdc close failed: rc = -13 LustreError: 13713:0:(file.c:248:ll_close_inode_openhandle()) Skipped 9 previous similar messages Lustre: 16055:0:(lod_lov.c:1402:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000404:0x1c9b:0x0] with magic=0xbd60bd0 Lustre: 16055:0:(lod_lov.c:1402:lod_parse_striping()) Skipped 3 previous similar messages 11[17409]: segfault at 8 ip 00007f9f316757e8 sp 00007fffe1718c50 error 4 in ld-2.17.so[7f9f3166a000+22000] 18[22340]: segfault at 0 ip 0000000000403e5f sp 00007fffe6f8aec0 error 6 in 18[400000+6000] 9[25101]: segfault at 8 ip 00007f9f495987e8 sp 00007ffc68aac270 error 4 in ld-2.17.so[7f9f4958d000+22000] 12[26555]: segfault at 0 ip 0000000000403e5f sp 00007ffe12e941b0 error 6 in 12[400000+6000] 19[27758]: segfault at 8 ip 00007f3949ae47e8 sp 00007ffc4cd2ecc0 error 4 in ld-2.17.so[7f3949ad9000+22000] 11[30938]: segfault at 0 ip 0000000000403e5f sp 00007ffe0db8c3a0 error 6 in 11[400000+6000] 7[32233]: segfault at 8 ip 00007f172eb937e8 sp 00007ffe75cfe3f0 error 4 in ld-2.17.so[7f172eb88000+22000] 16[4848]: segfault at 8 ip 00007f7a897c77e8 sp 00007ffc9b0f6410 error 4 in ld-2.17.so[7f7a897bc000+22000] 18[6565]: segfault at 8 ip 00007f0086afa7e8 sp 00007fff83a98430 error 4 in ld-2.17.so[7f0086aef000+22000] 5[10150]: segfault at 8 ip 00007f804f8d87e8 sp 00007ffc77b59300 error 4 in ld-2.17.so[7f804f8cd000+22000] 1[10963]: segfault at 0 ip 0000000000403e5f sp 00007fff25dfb600 error 6 in 1[400000+6000] 8[15542]: segfault at 8 ip 00007faf9a5e57e8 sp 00007fff99c73a40 error 4 in ld-2.17.so[7faf9a5da000+22000] 1[16046]: segfault at 8 ip 00007f33beef07e8 sp 00007fff8d7e8f10 error 4 in ld-2.17.so[7f33beee5000+22000] 14[23423]: segfault at 8 ip 00007fcf971877e8 sp 00007ffc0cdf9ae0 error 4 in ld-2.17.so[7fcf9717c000+22000] 6[24245]: segfault at 8 ip 00007f0bc6d197e8 sp 00007ffc7f552990 error 4 in ld-2.17.so[7f0bc6d0e000+22000] 3[24621]: segfault at 406000 ip 0000000000406000 sp 00007fff370b2238 error 14 in 3[606000+1000] 12[27219]: segfault at 8 ip 00007fa7bc8b37e8 sp 00007fffdd1cbc50 error 4 in ld-2.17.so[7fa7bc8a8000+22000] 6[31943]: segfault at 8 ip 00007f5feed107e8 sp 00007ffc7e895460 error 4 in ld-2.17.so[7f5feed05000+22000] 4[1493]: segfault at 8 ip 00007fed6a43a7e8 sp 00007ffe6f2332e0 error 4 in ld-2.17.so[7fed6a42f000+22000] 7[8047]: segfault at 0 ip 0000000000403e5f sp 00007fff5907e8c0 error 6 in 7[400000+6000] 8[10344]: segfault at 8 ip 00007f082fb917e8 sp 00007ffdeb15e580 error 4 in ld-2.17.so[7f082fb86000+22000] Lustre: 20833:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 516 < left 618, rollback = 7 Lustre: 20833:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 579 previous similar messages Lustre: 20833:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 20833:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 579 previous similar messages Lustre: 20833:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 20833:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 579 previous similar messages Lustre: 20833:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 20833:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 579 previous similar messages Lustre: 20833:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 20833:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 579 previous similar messages Lustre: 20833:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 20833:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 579 previous similar messages 16[15390]: segfault at 8 ip 00007f89f8f4f7e8 sp 00007ffde89c41f0 error 4 in ld-2.17.so[7f89f8f44000+22000] 2[16108]: segfault at 4045c9 ip 0000000000405b29 sp 00007ffc237b77e0 error 7 in 2[400000+6000] 13[16572]: segfault at 8 ip 00007f4e80c327e8 sp 00007fff31edf3d0 error 4 in ld-2.17.so[7f4e80c27000+22000] 9[16906]: segfault at 8 ip 00007fea146167e8 sp 00007fff4fd58c60 error 4 in ld-2.17.so[7fea1460b000+22000] 8[28701]: segfault at 8 ip 00007f18180e77e8 sp 00007ffeb4bd2470 error 4 in ld-2.17.so[7f18180dc000+22000] 1[29927]: segfault at 8 ip 00007fdd779a27e8 sp 00007ffeabedd550 error 4 in ld-2.17.so[7fdd77997000+22000] 3[703]: segfault at 8 ip 00007feff65587e8 sp 00007fff5c2ba7c0 error 4 in ld-2.17.so[7feff654d000+22000] 3[4283]: segfault at 0 ip (null) sp 00007ffdc03046f8 error 14 in 3[400000+6000] 8[5558]: segfault at 8 ip 00007f05e0f347e8 sp 00007ffd0d718e60 error 4 in ld-2.17.so[7f05e0f29000+22000] 7[10434]: segfault at 8 ip 00007fad57f947e8 sp 00007ffe97b8afc0 error 4 in ld-2.17.so[7fad57f89000+22000] 7[11652]: segfault at 8 ip 00007f732250c7e8 sp 00007ffc53abd050 error 4 in ld-2.17.so[7f7322501000+22000] 1[12945]: segfault at 8 ip 00007f8aa96f57e8 sp 00007ffd18726110 error 4 in ld-2.17.so[7f8aa96ea000+22000] 9[13459]: segfault at 8 ip 00007f0a736f97e8 sp 00007ffcc0b68770 error 4 in ld-2.17.so[7f0a736ee000+22000] LustreError: 13680:0:(file.c:248:ll_close_inode_openhandle()) lustre-clilmv-ffff8802ed4c37e8: inode [0x200000404:0x3c8b:0x0] mdc close failed: rc = -13 LustreError: 13680:0:(file.c:248:ll_close_inode_openhandle()) Skipped 3 previous similar messages 19[19186]: segfault at 4017b0 ip 00000000004017b0 sp 00007fff6f5118d0 error 14 18[20488]: segfault at 8 ip 00007fe5c70027e8 sp 00007ffcee213090 error 4 in ld-2.17.so[7fe5c6ff7000+22000] 12[21324]: segfault at 8 ip 00007f3b5c5a37e8 sp 00007ffee49f71b0 error 4 in ld-2.17.so[7f3b5c598000+22000] 0[21998]: segfault at 8 ip 00007f61f00777e8 sp 00007ffe895f7350 error 4 in ld-2.17.so[7f61f006c000+22000] 5[27525]: segfault at 0 ip (null) sp 00007ffd5a7be078 error 14 in 5[400000+6000] 7[816]: segfault at 8 ip 00007f3033c6d7e8 sp 00007ffe142ce210 error 4 in ld-2.17.so[7f3033c62000+22000] 0[9849]: segfault at 8 ip 00007f8f76e957e8 sp 00007ffd4f912dc0 error 4 in ld-2.17.so[7f8f76e8a000+22000] Lustre: 15332:0:(lod_lov.c:1402:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000404:0x474f:0x0] with magic=0xbd60bd0 Lustre: 15332:0:(lod_lov.c:1402:lod_parse_striping()) Skipped 21 previous similar messages 16[19349]: segfault at 8 ip 00007f9b259517e8 sp 00007ffffd5e6f80 error 4 in ld-2.17.so[7f9b25946000+22000] 13[31229]: segfault at 8 ip 00007fbde0c557e8 sp 00007ffcad986c90 error 4 in ld-2.17.so[7fbde0c4a000+22000] 18[6834]: segfault at 0 ip 0000000000401898 sp 00007fff774e4fa8 error 6 in 18[400000+6000] ptlrpc_watchdog_fire: 2 callbacks suppressed Lustre: mdt00_004: service thread pid 13028 was inactive for 40.131 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 9644, comm: mdt00_001 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock_try+0xa0/0x250 [mdt] [<0>] mdt_getattr_name_lock+0x1739/0x2d20 [mdt] [<0>] mdt_intent_getattr+0x2cc/0x4e0 [mdt] [<0>] mdt_intent_opc.constprop.74+0x211/0xc60 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x397/0x980 [ptlrpc] [<0>] ldlm_handle_enqueue+0x547/0x1900 [ptlrpc] [<0>] tgt_enqueue+0x68/0x240 [ptlrpc] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x23d/0xd80 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 11851, comm: mdt00_003 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock_try+0xa0/0x250 [mdt] [<0>] mdt_getattr_name_lock+0x1739/0x2d20 [mdt] [<0>] mdt_intent_getattr+0x2cc/0x4e0 [mdt] [<0>] mdt_intent_opc.constprop.74+0x211/0xc60 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x397/0x980 [ptlrpc] [<0>] ldlm_handle_enqueue+0x547/0x1900 [ptlrpc] [<0>] tgt_enqueue+0x68/0x240 [ptlrpc] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x23d/0xd80 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: Skipped 4 previous similar messages Pid: 13028, comm: mdt00_004 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock+0x88/0x1c0 [mdt] [<0>] mdt_reint_link+0x8fb/0xff0 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x73c/0xbb0 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x23d/0xd80 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: lustre-OST0003-osc-ffff8802ed4c37e8: disconnect after 21s idle LustreError: 9314:0:(ldlm_lockd.c:257:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0000_UUID lock: ffff8802d2764b40/0xd06913fc90292376 lrc: 3/0,0 mode: PR/PR res: [0x200000405:0x4062:0x0].0x0 bits 0x1b/0x0 rrc: 7 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0xd06913fc9029233e expref: 1313 pid: 16297 timeout: 1808 lvb_type: 0 Lustre: mdt00_004: service thread pid 13028 completed after 100.309s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_003: service thread pid 11851 completed after 100.296s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: lustre-MDT0000-mdc-ffff8802ed4c37e8: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-MDT0000-mdc-ffff8802ed4c37e8: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: lustre-MDT0000-mdc-ffff8802ed4c37e8: This client was evicted by lustre-MDT0000; in progress operations using this service will fail. Lustre: mdt00_001: service thread pid 9644 completed after 100.313s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 8644:0:(file.c:6198:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000405:0x4062:0x0] error: rc = -5 LustreError: 8644:0:(file.c:6198:ll_inode_revalidate_fini()) Skipped 38 previous similar messages LustreError: 8357:0:(mdc_request.c:1470:mdc_read_page()) lustre-MDT0000-mdc-ffff8802ed4c37e8: [0x200000401:0x1:0x0] lock enqueue fails: rc = -108 LustreError: 8357:0:(mdc_request.c:1470:mdc_read_page()) Skipped 18 previous similar messages Lustre: lustre-MDT0000-mdc-ffff8802ed4c37e8: Connection restored to 192.168.123.47@tcp (at 0@lo) 14[26923]: segfault at 8 ip 00007f34a36be7e8 sp 00007fffd8870de0 error 4 in ld-2.17.so[7f34a36b3000+22000] Lustre: 15830:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 516 < left 588, rollback = 7 Lustre: 15830:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 579 previous similar messages Lustre: 15830:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 15830:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 579 previous similar messages Lustre: 15830:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 15830:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 579 previous similar messages Lustre: 15830:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/588/0, punch: 0/0/0, quota 1/3/0 Lustre: 15830:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 579 previous similar messages Lustre: 15830:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 15830:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 579 previous similar messages Lustre: 15830:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 15830:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 579 previous similar messages 19[13293]: segfault at 8 ip 00007fe9af9267e8 sp 00007ffed6402060 error 4 in ld-2.17.so[7fe9af91b000+22000] 16[14997]: segfault at 8 ip 00007f820dbff7e8 sp 00007fff62500360 error 4 in ld-2.17.so[7f820dbf4000+22000] LustreError: 22404:0:(file.c:248:ll_close_inode_openhandle()) lustre-clilmv-ffff8802ed4c37e8: inode [0x200000406:0xf5e:0x0] mdc close failed: rc = -13 LustreError: 22404:0:(file.c:248:ll_close_inode_openhandle()) Skipped 13 previous similar messages 12[23151]: segfault at 8 ip 00007fd6853d17e8 sp 00007ffe44448640 error 4 in ld-2.17.so[7fd6853c6000+22000] 4[16647]: segfault at 8 ip 00007f1c5cb417e8 sp 00007ffe6477d230 error 4 in ld-2.17.so[7f1c5cb36000+22000] 10[16965]: segfault at 8 ip 00007f3e5a90d7e8 sp 00007ffd065e7a80 error 4 in ld-2.17.so[7f3e5a902000+22000] 3[24452]: segfault at 0 ip 0000000000403e5f sp 00007ffd572f2c50 error 6 in 3[400000+6000] 13[31291]: segfault at 8 ip 00007f7792f047e8 sp 00007ffc77937ef0 error 4 in ld-2.17.so[7f7792ef9000+22000] 18[32388]: segfault at 8 ip 00007f7eec2397e8 sp 00007ffd99809ed0 error 4 in ld-2.17.so[7f7eec22e000+22000] 0[32405]: segfault at 0 ip 0000000000403e5f sp 00007ffee30996b0 error 6 in 0[400000+6000] 8[2002]: segfault at 0 ip (null) sp 00007ffd3612f468 error 14 in 8[400000+6000] 8[2170]: segfault at 8 ip 00007fa633e1f7e8 sp 00007ffdbfbc2480 error 4 in ld-2.17.so[7fa633e14000+22000] 16[12463]: segfault at 8 ip 00007fa8b42a27e8 sp 00007fff627654a0 error 4 in ld-2.17.so[7fa8b4297000+22000] Lustre: 31285:0:(lod_lov.c:1402:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000404:0x7546:0x0] with magic=0xbd60bd0 Lustre: 31285:0:(lod_lov.c:1402:lod_parse_striping()) Skipped 29 previous similar messages 8[28849]: segfault at 8 ip 00007f0bae2c47e8 sp 00007ffe52262c50 error 4 in ld-2.17.so[7f0bae2b9000+22000] 5[30286]: segfault at 8 ip 00007f0e321957e8 sp 00007ffef8fa5290 error 4 in ld-2.17.so[7f0e3218a000+22000] 17[32635]: segfault at 8 ip 00007fea0e2817e8 sp 00007ffc8376fab0 error 4 in ld-2.17.so[7fea0e276000+22000] 18[5098]: segfault at 8 ip 00007f93e65a97e8 sp 00007ffca0b1eb20 error 4 in ld-2.17.so[7f93e659e000+22000] 19[14197]: segfault at 8 ip 00007fbd4b9247e8 sp 00007fff85c3abe0 error 4 in ld-2.17.so[7fbd4b919000+22000] 13[15315]: segfault at 0 ip 0000000000403e5f sp 00007ffdb5620260 error 6 in 13[400000+6000] 5[16088]: segfault at 8 ip 00007f3a57b0e7e8 sp 00007ffd56b1ca50 error 4 in ld-2.17.so[7f3a57b03000+22000] 16[16676]: segfault at 8 ip 00007f24137327e8 sp 00007ffd3e6b16d0 error 4 in ld-2.17.so[7f2413727000+22000] 13[29451]: segfault at 8 ip 00007f72c5efe7e8 sp 00007ffddc8f6bb0 error 4 in ld-2.17.so[7f72c5ef3000+22000] 18[2554]: segfault at 8 ip 00007fe24cd307e8 sp 00007ffcbee86630 error 4 in ld-2.17.so[7fe24cd25000+22000] 19[14605]: segfault at 8 ip 00007f99d472e7e8 sp 00007ffeba939e60 error 4 in ld-2.17.so[7f99d4723000+22000] 6[29622]: segfault at 8 ip 00007f4afd8377e8 sp 00007ffdfb168080 error 4 in ld-2.17.so[7f4afd82c000+22000] LustreError: 4049:0:(statahead.c:2457:start_statahead_thread()) lustre: unsupported statahead pattern 0X0. 13[3982]: segfault at 0 ip 0000000000403e5f sp 00007ffcdc1f8360 error 6 in 13[400000+6000] 10[6986]: segfault at 406000 ip 0000000000406000 sp 00007ffff0f37ec8 error 14 in 10[606000+1000] Lustre: 16270:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 618, rollback = 7 Lustre: 16270:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 689 previous similar messages Lustre: 16270:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 16270:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 689 previous similar messages Lustre: 16270:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 16270:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 689 previous similar messages Lustre: 16270:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 16270:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 689 previous similar messages Lustre: 16270:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 16270:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 689 previous similar messages Lustre: 16270:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 16270:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 689 previous similar messages | Link to test |
racer test 1: racer on clients: centos-55.localnet DURATION=2700 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 PGD 2a1592067 PUD 2bb7c2067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) osp(OE) ofd(OE) lod(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) crc32_generic lnet(OE) libcfs(OE) crc_t10dif crct10dif_generic crct10dif_common pcspkr virtio_console virtio_balloon i2c_piix4 ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm ata_piix floppy drm_panel_orientation_quirks serio_raw virtio_blk i2c_core libata CPU: 2 PID: 7567 Comm: ll_sa_7555 Kdump: loaded Tainted: P OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.16.0-3.module_el8.7.0+1218+f626c2ff 04/01/2014 task: ffff8802d4e69280 ti: ffff8802c6748000 task.ti: ffff8802c6748000 RIP: 0010:[<ffffffff813f502c>] [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 RSP: 0018:ffff8802c674bd78 EFLAGS: 00010296 RAX: ffff880081d7e990 RBX: 0000000000000008 RCX: 0000000100260016 RDX: 0000000000000026 RSI: ffff880081d7ecc0 RDI: 0000000000000008 RBP: ffff8802c674bd88 R08: ffff88009e1f48e8 R09: 0000000000000000 R10: ffff88009e1f6a08 R11: ffff88009e1f48e8 R12: 0000000000000000 R13: ffff88009a92dc40 R14: ffff880081d7ecc0 R15: ffff88009e1f48e8 FS: 0000000000000000(0000) GS:ffff880331a80000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 00000002acb2c000 CR4: 00000000000007e0 Call Trace: [<ffffffffa15b5791>] ll_statahead_thread+0xdd1/0x2020 [lustre] [<ffffffff817e320e>] ? _raw_spin_lock+0x1e/0x20 [<ffffffff810dbb51>] ? put_prev_entity+0x31/0x400 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e329e>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffff810c834d>] ? finish_task_switch+0x5d/0x1b0 [<ffffffffa15b49c0>] ? ll_statahead_handle.constprop.30+0x1c0/0x1c0 [lustre] [<ffffffff810ba114>] kthread+0xe4/0xf0 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 | Lustre: 12321:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff8802d2387840 x1837034453014656/t4294969148(0) o101->63098b92-42e1-4fb3-91f4-4408e3e09d3b@0@lo:455/0 lens 384/840 e 0 to 0 dl 1751932655 ref 1 fl Interpret:H/602/0 rc 0/0 job:'dd.0' uid:0 gid:0 projid:0 Lustre: mdt_io00_003: service thread pid 23031 was inactive for 40.084 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 16667, comm: mdt00_003 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_pdo_lock+0x369/0x750 [mdt] [<0>] mdt_parent_lock+0x76/0x2a0 [mdt] [<0>] mdt_getattr_name_lock+0x1854/0x2d20 [mdt] [<0>] mdt_intent_getattr+0x2cc/0x4e0 [mdt] [<0>] mdt_intent_opc.constprop.74+0x211/0xc60 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x397/0x980 [ptlrpc] [<0>] ldlm_handle_enqueue+0x547/0x1900 [ptlrpc] [<0>] tgt_enqueue+0x68/0x240 [ptlrpc] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x23d/0xd80 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: Skipped 1 previous similar message Pid: 23031, comm: mdt_io00_003 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock+0x88/0x1c0 [mdt] [<0>] mdt_rename_source_lock+0x57/0xf0 [mdt] [<0>] mdt_reint_rename+0xde7/0x3ae0 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x73c/0xbb0 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x23d/0xd80 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: mdt00_012: service thread pid 23866 was inactive for 40.138 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Lustre: mdt00_002: service thread pid 12322 was inactive for 40.137 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: mdt00_006: service thread pid 22214 was inactive for 40.122 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Pid: 23866, comm: mdt00_012 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock+0x88/0x1c0 [mdt] [<0>] mdt_getattr_name_lock+0xc5a/0x2d20 [mdt] [<0>] mdt_intent_getattr+0x2cc/0x4e0 [mdt] [<0>] mdt_intent_opc.constprop.74+0x211/0xc60 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x397/0x980 [ptlrpc] [<0>] ldlm_handle_enqueue+0x547/0x1900 [ptlrpc] [<0>] tgt_enqueue+0x68/0x240 [ptlrpc] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x23d/0xd80 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe LustreError: 11895:0:(ldlm_lockd.c:257:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0000_UUID lock: ffff8802d3281e40/0xd0ff2d47ef8f1c69 lrc: 3/0,0 mode: PR/PR res: [0x200000401:0xcb:0x0].0x0 bits 0x13/0x0 rrc: 4 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0xd0ff2d47ef8f1c4d expref: 87 pid: 23842 timeout: 207 lvb_type: 0 LustreError: 18271:0:(ldlm_lockd.c:1447:ldlm_handle_enqueue()) ### lock on destroyed export ffff8800a3c46fc8 ns: mdt-lustre-MDT0000_UUID lock: ffff8802d2b170c0/0xd0ff2d47ef8f1ef4 lrc: 3/0,0 mode: PR/PR res: [0x200000401:0x1:0x0].0x0 bits 0x13/0x0 rrc: 13 type: IBT gid 0 flags: 0x50200000000000 nid: 0@lo remote: 0xd0ff2d47ef8f1ee6 expref: 19 pid: 18271 timeout: 0 lvb_type: 0 Lustre: mdt_io00_003: service thread pid 23031 completed after 100.515s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_004: service thread pid 18271 completed after 99.271s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_001: service thread pid 12321 completed after 99.281s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: lustre-MDT0000-mdc-ffff8802db315d28: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-MDT0000-mdc-ffff8802db315d28: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: mdt00_008: service thread pid 22478 completed after 99.260s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_002: service thread pid 12322 completed after 99.294s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_009: service thread pid 23831 completed after 99.283s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: lustre-MDT0000-mdc-ffff8802db315d28: This client was evicted by lustre-MDT0000; in progress operations using this service will fail. Lustre: mdt00_012: service thread pid 23866 completed after 99.298s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_011: service thread pid 23842 completed after 99.265s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_003: service thread pid 16667 completed after 100.457s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_006: service thread pid 22214 completed after 99.284s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 27960:0:(file.c:6198:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000401:0x1:0x0] error: rc = -5 LustreError: 27552:0:(llite_lib.c:2023:ll_md_setattr()) md_setattr fails: rc = -108 LustreError: 28494:0:(file.c:248:ll_close_inode_openhandle()) lustre-clilmv-ffff8802db315d28: inode [0x200000402:0xa8:0x0] mdc close failed: rc = -108 LustreError: 27388:0:(vvp_io.c:1909:vvp_io_init()) lustre: refresh file layout [0x200000401:0xcb:0x0] error -108. Lustre: lustre-MDT0000-mdc-ffff8802db315d28: Connection restored to 192.168.123.57@tcp (at 0@lo) Lustre: 28692:0:(lod_lov.c:1402:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000402:0x18e:0x0] with magic=0xbd60bd0 LustreError: 18271:0:(mdt_handler.c:746:mdt_pack_acl2body()) lustre-MDT0000: unable to read [0x200000403:0x1d1:0x0] ACL: rc = -2 LustreError: 9619:0:(file.c:248:ll_close_inode_openhandle()) lustre-clilmv-ffff8802db315d28: inode [0x200000403:0x1ca:0x0] mdc close failed: rc = -13 LustreError: 9619:0:(file.c:248:ll_close_inode_openhandle()) Skipped 2 previous similar messages LustreError: 14479:0:(file.c:248:ll_close_inode_openhandle()) lustre-clilmv-ffff8802db315d28: inode [0x200000403:0x343:0x0] mdc close failed: rc = -13 10[24697]: segfault at 8 ip 00007f52b145e7e8 sp 00007fffd9712ac0 error 4 in ld-2.17.so[7f52b1453000+22000] 9[28407]: segfault at 8 ip 00007f7940fc97e8 sp 00007ffde1150230 error 4 in ld-2.17.so[7f7940fbe000+22000] LustreError: 32304:0:(file.c:248:ll_close_inode_openhandle()) lustre-clilmv-ffff8802db315d28: inode [0x200000402:0x66a:0x0] mdc close failed: rc = -13 LustreError: 22214:0:(mdt_handler.c:746:mdt_pack_acl2body()) lustre-MDT0000: unable to read [0x200000402:0x6ab:0x0] ACL: rc = -2 LustreError: 23872:0:(mdt_handler.c:746:mdt_pack_acl2body()) lustre-MDT0000: unable to read [0x200000402:0x6a8:0x0] ACL: rc = -2 0[7589]: segfault at 0 ip (null) sp 00007ffc6c0ce268 error 14 in 0[400000+6000] LustreError: 22041:0:(file.c:248:ll_close_inode_openhandle()) lustre-clilmv-ffff8802da8fb7e8: inode [0x200000402:0x97e:0x0] mdc close failed: rc = -13 Lustre: 16667:0:(lod_lov.c:1402:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0xc90:0x0] with magic=0xbd60bd0 Lustre: 16667:0:(lod_lov.c:1402:lod_parse_striping()) Skipped 1 previous similar message LustreError: 26063:0:(file.c:248:ll_close_inode_openhandle()) lustre-clilmv-ffff8802da8fb7e8: inode [0x200000402:0xed7:0x0] mdc close failed: rc = -13 15[32086]: segfault at 8 ip 00007f1a932db7e8 sp 00007ffeff1fb0f0 error 4 in ld-2.17.so[7f1a932d0000+22000] 14[10258]: segfault at 4017b0 ip 00000000004017b0 sp 00007fff397bd010 error 14 LustreError: 10258:0:(file.c:248:ll_close_inode_openhandle()) lustre-clilmv-ffff8802db315d28: inode [0x200000402:0x1103:0x0] mdc close failed: rc = -13 Lustre: 28708:0:(lod_lov.c:1402:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0x1065:0x0] with magic=0xbd60bd0 Lustre: 28708:0:(lod_lov.c:1402:lod_parse_striping()) Skipped 1 previous similar message LustreError: 5034:0:(file.c:248:ll_close_inode_openhandle()) lustre-clilmv-ffff8802da8fb7e8: inode [0x200000403:0x13dd:0x0] mdc close failed: rc = -13 LustreError: 5034:0:(file.c:248:ll_close_inode_openhandle()) Skipped 1 previous similar message Lustre: 23842:0:(lod_lov.c:1402:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000402:0x14ff:0x0] with magic=0xbd60bd0 Lustre: 23842:0:(lod_lov.c:1402:lod_parse_striping()) Skipped 1 previous similar message 12[25236]: segfault at 8 ip 00007f79993697e8 sp 00007fffef8a1830 error 4 in ld-2.17.so[7f799935e000+22000] 19[31338]: segfault at 8 ip 00007f9005f237e8 sp 00007ffce19c8ce0 error 4 in ld-2.17.so[7f9005f18000+22000] LustreError: 18743:0:(file.c:248:ll_close_inode_openhandle()) lustre-clilmv-ffff8802da8fb7e8: inode [0x200000403:0x19a8:0x0] mdc close failed: rc = -13 19[27279]: segfault at 0 ip 0000000000403e5f sp 00007fff3b8d47e0 error 6 in 19[400000+6000] 11[25128]: segfault at 0 ip (null) sp 00007fff375080a8 error 14 in 11[400000+6000] Lustre: 12319:0:(lod_lov.c:1402:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0x2417:0x0] with magic=0xbd60bd0 Lustre: 12319:0:(lod_lov.c:1402:lod_parse_striping()) Skipped 1 previous similar message Lustre: 28708:0:(lod_lov.c:1402:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000402:0x25e9:0x0] with magic=0xbd60bd0 Lustre: 28708:0:(lod_lov.c:1402:lod_parse_striping()) Skipped 1 previous similar message LustreError: 10053:0:(statahead.c:2457:start_statahead_thread()) lustre: unsupported statahead pattern 0X0. Lustre: 27131:0:(lod_lov.c:1402:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0x2709:0x0] with magic=0xbd60bd0 Lustre: 27131:0:(lod_lov.c:1402:lod_parse_striping()) Skipped 1 previous similar message 5[23960]: segfault at 8 ip 00007f1a10c697e8 sp 00007ffebb7b1510 error 4 in ld-2.17.so[7f1a10c5e000+22000] 6[712]: segfault at 8 ip 00007f858351e7e8 sp 00007ffc2ceaced0 error 4 in ld-2.17.so[7f8583513000+22000] LustreError: 18271:0:(mdt_handler.c:746:mdt_pack_acl2body()) lustre-MDT0000: unable to read [0x200000402:0x29a6:0x0] ACL: rc = -2 LustreError: 23872:0:(mdt_handler.c:746:mdt_pack_acl2body()) lustre-MDT0000: unable to read [0x200000402:0x2a8d:0x0] ACL: rc = -2 ptlrpc_watchdog_fire: 6 callbacks suppressed Lustre: mdt00_013: service thread pid 23872 was inactive for 40.002 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 5 previous similar messages Lustre: mdt00_009: service thread pid 23831 was inactive for 40.000 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 23831, comm: mdt00_009 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock_try+0xa0/0x250 [mdt] [<0>] mdt_getattr_name_lock+0x1739/0x2d20 [mdt] [<0>] mdt_intent_getattr+0x2cc/0x4e0 [mdt] [<0>] mdt_intent_opc.constprop.74+0x211/0xc60 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x397/0x980 [ptlrpc] [<0>] ldlm_handle_enqueue+0x547/0x1900 [ptlrpc] [<0>] tgt_enqueue+0x68/0x240 [ptlrpc] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x23d/0xd80 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 23842, comm: mdt00_011 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock_try+0xa0/0x250 [mdt] [<0>] mdt_getattr_name_lock+0x1739/0x2d20 [mdt] [<0>] mdt_intent_getattr+0x2cc/0x4e0 [mdt] [<0>] mdt_intent_opc.constprop.74+0x211/0xc60 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x397/0x980 [ptlrpc] [<0>] ldlm_handle_enqueue+0x547/0x1900 [ptlrpc] [<0>] tgt_enqueue+0x68/0x240 [ptlrpc] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x23d/0xd80 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 28708, comm: mdt00_019 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock_try+0xa0/0x250 [mdt] [<0>] mdt_getattr_name_lock+0x1739/0x2d20 [mdt] [<0>] mdt_intent_getattr+0x2cc/0x4e0 [mdt] [<0>] mdt_intent_opc.constprop.74+0x211/0xc60 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x397/0x980 [ptlrpc] [<0>] ldlm_handle_enqueue+0x547/0x1900 [ptlrpc] [<0>] tgt_enqueue+0x68/0x240 [ptlrpc] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x23d/0xd80 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: mdt00_005: service thread pid 22127 was inactive for 40.096 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 6 previous similar messages LustreError: 11895:0:(ldlm_lockd.c:257:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0000_UUID lock: ffff8802c6f77840/0xd0ff2d47efe96a36 lrc: 3/0,0 mode: PR/PR res: [0x200000402:0x2cab:0x0].0x0 bits 0x1b/0x0 rrc: 11 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0xd0ff2d47efe96a1a expref: 1102 pid: 23872 timeout: 818 lvb_type: 0 Lustre: mdt00_013: service thread pid 23872 completed after 100.185s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_014: service thread pid 28669 completed after 99.935s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_019: service thread pid 28708 completed after 99.932s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_011: service thread pid 23842 completed after 100.003s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: lustre-MDT0000-mdc-ffff8802da8fb7e8: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 6 previous similar messages Lustre: lustre-MDT0000-mdc-ffff8802da8fb7e8: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: lustre-MDT0000-mdc-ffff8802da8fb7e8: This client was evicted by lustre-MDT0000; in progress operations using this service will fail. Lustre: mdt00_009: service thread pid 23831 completed after 100.197s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 23002:0:(file.c:248:ll_close_inode_openhandle()) lustre-clilmv-ffff8802da8fb7e8: inode [0x200000401:0x1:0x0] mdc close failed: rc = -108 LustreError: 23002:0:(file.c:248:ll_close_inode_openhandle()) Skipped 1 previous similar message Lustre: mdt00_002: service thread pid 12322 completed after 99.869s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 21329:0:(file.c:6198:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000007:0x1:0x0] error: rc = -108 LustreError: 21329:0:(file.c:6198:ll_inode_revalidate_fini()) Skipped 10 previous similar messages Lustre: mdt00_012: service thread pid 23866 completed after 99.858s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_002: service thread pid 12354 completed after 99.558s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_020: service thread pid 27131 completed after 99.049s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_005: service thread pid 22127 completed after 99.061s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_008: service thread pid 22478 completed after 99.064s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: lustre-MDT0000-mdc-ffff8802da8fb7e8: Connection restored to 192.168.123.57@tcp (at 0@lo) Lustre: 22478:0:(lod_lov.c:1402:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000404:0x9d:0x0] with magic=0xbd60bd0 Lustre: 22478:0:(lod_lov.c:1402:lod_parse_striping()) Skipped 3 previous similar messages 15[12639]: segfault at 8 ip 00007f99d24d47e8 sp 00007ffdad6f3540 error 4 in ld-2.17.so[7f99d24c9000+22000] 7[18034]: segfault at 0 ip 0000000000403e5f sp 00007ffeb38b4d70 error 6 in 7[400000+6000] LustreError: 28708:0:(mdt_handler.c:746:mdt_pack_acl2body()) lustre-MDT0000: unable to read [0x200000404:0x4fa:0x0] ACL: rc = -2 16[1130]: segfault at 8 ip 00007f2858ff87e8 sp 00007ffe41974990 error 4 in ld-2.17.so[7f2858fed000+22000] Lustre: 22465:0:(lod_lov.c:1402:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0x315b:0x0] with magic=0xbd60bd0 Lustre: 22465:0:(lod_lov.c:1402:lod_parse_striping()) Skipped 1 previous similar message 3[2557]: segfault at 8 ip 00007fd32f8b47e8 sp 00007ffd8ff02240 error 4 in ld-2.17.so[7fd32f8a9000+22000] LustreError: 11895:0:(ldlm_lockd.c:257:expired_lock_main()) ### lock callback timer expired after 101s: evicting client at 0@lo ns: mdt-lustre-MDT0000_UUID lock: ffff8802b31ae580/0xd0ff2d47effccba0 lrc: 3/0,0 mode: PR/PR res: [0x200000403:0x3481:0x0].0x0 bits 0x13/0x0 rrc: 17 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0xd0ff2d47effccb92 expref: 1007 pid: 23866 timeout: 1041 lvb_type: 0 LustreError: 28706:0:(ldlm_lockd.c:1447:ldlm_handle_enqueue()) ### lock on destroyed export ffff8802c3e88958 ns: mdt-lustre-MDT0000_UUID lock: ffff8802b77243c0/0xd0ff2d47effcf0f3 lrc: 3/0,0 mode: PR/PR res: [0x200000403:0x3481:0x0].0x0 bits 0x1b/0x0 rrc: 12 type: IBT gid 0 flags: 0x50200000000000 nid: 0@lo remote: 0xd0ff2d47effcf0bb expref: 192 pid: 28706 timeout: 0 lvb_type: 0 LustreError: 11888:0:(ldlm_lockd.c:2550:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1751933583 with bad export cookie 15059805465935290478 Lustre: lustre-MDT0000-mdc-ffff8802db315d28: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 28706:0:(ldlm_lockd.c:1447:ldlm_handle_enqueue()) Skipped 7 previous similar messages LustreError: lustre-MDT0000-mdc-ffff8802db315d28: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: lustre-MDT0000-mdc-ffff8802db315d28: This client was evicted by lustre-MDT0000; in progress operations using this service will fail. LustreError: 22694:0:(file.c:6198:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000403:0x3481:0x0] error: rc = -5 LustreError: 22694:0:(file.c:6198:ll_inode_revalidate_fini()) Skipped 9 previous similar messages LustreError: 22278:0:(llite_lib.c:2023:ll_md_setattr()) md_setattr fails: rc = -5 LustreError: Skipped 3 previous similar messages LustreError: 22337:0:(vvp_io.c:1909:vvp_io_init()) lustre: refresh file layout [0x200000403:0x3481:0x0] error -108. Lustre: lustre-MDT0000-mdc-ffff8802db315d28: Connection restored to 192.168.123.57@tcp (at 0@lo) Lustre: lustre-OST0000-osc-ffff8802db315d28: disconnect after 24s idle LustreError: 11895:0:(ldlm_lockd.c:257:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0000_UUID lock: ffff88009dce7840/0xd0ff2d47effd45e8 lrc: 3/0,0 mode: PR/PR res: [0x200000405:0xa:0x0].0x0 bits 0x1b/0x0 rrc: 8 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0xd0ff2d47effd457f expref: 47 pid: 12335 timeout: 1143 lvb_type: 0 LustreError: 27131:0:(ldlm_lockd.c:1447:ldlm_handle_enqueue()) ### lock on destroyed export ffff8802b60ba548 ns: mdt-lustre-MDT0000_UUID lock: ffff8802b6258b80/0xd0ff2d47effd51b8 lrc: 3/0,0 mode: PR/PR res: [0x200000401:0x1:0x0].0x0 bits 0x13/0x0 rrc: 20 type: IBT gid 0 flags: 0x50200000000000 nid: 0@lo remote: 0xd0ff2d47effd51aa expref: 17 pid: 27131 timeout: 0 lvb_type: 0 LustreError: 27131:0:(ldlm_lockd.c:1447:ldlm_handle_enqueue()) Skipped 7 previous similar messages LustreError: lustre-MDT0000-mdc-ffff8802db315d28: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-MDT0000-mdc-ffff8802db315d28: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: lustre-MDT0000-mdc-ffff8802db315d28: This client was evicted by lustre-MDT0000; in progress operations using this service will fail. LustreError: Skipped 2 previous similar messages LustreError: 24785:0:(file.c:6198:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000401:0x1:0x0] error: rc = -5 LustreError: 24785:0:(file.c:6198:ll_inode_revalidate_fini()) Skipped 41 previous similar messages LustreError: 25492:0:(file.c:248:ll_close_inode_openhandle()) lustre-clilmv-ffff8802db315d28: inode [0x200000404:0x876:0x0] mdc close failed: rc = -108 LustreError: 25492:0:(file.c:248:ll_close_inode_openhandle()) Skipped 18 previous similar messages LustreError: 25492:0:(ldlm_resource.c:981:ldlm_resource_complain()) lustre-MDT0000-mdc-ffff8802db315d28: namespace resource [0x200000401:0x1:0x0].0x0 (ffff8800a14c5440) refcount nonzero (1) after lock cleanup; forcing cleanup. Lustre: lustre-MDT0000-mdc-ffff8802db315d28: Connection restored to 192.168.123.57@tcp (at 0@lo) 10[10993]: segfault at 8 ip 00007f78fee337e8 sp 00007ffd92925d70 error 4 in ld-2.17.so[7f78fee28000+22000] Lustre: 12322:0:(lod_lov.c:1402:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000404:0xefa:0x0] with magic=0xbd60bd0 Lustre: 12322:0:(lod_lov.c:1402:lod_parse_striping()) Skipped 1 previous similar message LustreError: 5600:0:(statahead.c:2409:start_statahead_thread()) lustre: invalid pattern 0X0. | Link to test |
racer test 1: racer on clients: centos-35.localnet DURATION=2700 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 PGD 31e784067 PUD 31b2b4067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) osp(OE) ofd(OE) lod(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) dm_flakey dm_mod libcfs(OE) loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) jbd2 mbcache crc32_generic crc_t10dif crct10dif_generic crct10dif_common virtio_console virtio_balloon pcspkr i2c_piix4 ip_tables rpcsec_gss_krb5 ata_generic drm_kms_helper pata_acpi ttm drm drm_panel_orientation_quirks ata_piix floppy serio_raw virtio_blk i2c_core libata [last unloaded: libcfs] CPU: 5 PID: 13445 Comm: ll_sa_13196 Kdump: loaded Tainted: P OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.16.0-3.module_el8.7.0+1218+f626c2ff 04/01/2014 task: ffff8802ef523760 ti: ffff8802cd8dc000 task.ti: ffff8802cd8dc000 RIP: 0010:[<ffffffff813f502c>] [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 RSP: 0018:ffff8802cd8dfd78 EFLAGS: 00010296 RAX: ffff88029612e990 RBX: 0000000000000008 RCX: 0000000100260022 RDX: 0000000000000026 RSI: ffff88029612ecc0 RDI: 0000000000000008 RBP: ffff8802cd8dfd88 R08: ffff880325051dd8 R09: 0000000000000000 R10: ffff880325054c38 R11: ffff880325051dd8 R12: 0000000000000000 R13: ffff8802ef525c40 R14: ffff88029612ecc0 R15: ffff880325051dd8 FS: 0000000000000000(0000) GS:ffff880331b40000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 00000002cf4cc000 CR4: 00000000000007e0 Call Trace: [<ffffffffa1786791>] ll_statahead_thread+0xdd1/0x2020 [lustre] [<ffffffff817e320e>] ? _raw_spin_lock+0x1e/0x20 [<ffffffff810dbb51>] ? put_prev_entity+0x31/0x400 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e329e>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffff810c834d>] ? finish_task_switch+0x5d/0x1b0 [<ffffffffa17859c0>] ? ll_statahead_handle.constprop.30+0x1c0/0x1c0 [lustre] [<ffffffff810ba114>] kthread+0xe4/0xf0 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 | LustreError: 30137:0:(file.c:248:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e8254a88: inode [0x200000401:0x11:0x0] mdc close failed: rc = -13 Lustre: 29743:0:(lod_lov.c:1402:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000402:0x2e:0x0] with magic=0xbd60bd0 12[376]: segfault at 8 ip 00007f4ea0cff7e8 sp 00007ffc259a4c70 error 4 in ld-2.17.so[7f4ea0cf4000+22000] LustreError: 2022:0:(file.c:248:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e8254a88: inode [0x200000401:0x142:0x0] mdc close failed: rc = -13 Lustre: 1007:0:(lod_lov.c:1402:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000402:0x1df:0x0] with magic=0xbd60bd0 Lustre: 1007:0:(lod_lov.c:1402:lod_parse_striping()) Skipped 1 previous similar message Lustre: 27849:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 618, rollback = 7 Lustre: 27849:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 27849:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 27849:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 27849:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 27849:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 9[4294]: segfault at 8 ip 00007f64a3b827e8 sp 00007ffeed9f1e50 error 4 in ld-2.17.so[7f64a3b77000+22000] Lustre: 30044:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 618, rollback = 7 Lustre: 30044:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 1 previous similar message Lustre: 30044:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 30044:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 30044:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 30044:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 30044:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 30044:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 30044:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 30044:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 30044:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 30044:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 27849:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 618, rollback = 7 Lustre: 27849:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 3 previous similar messages Lustre: 27849:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 27849:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 27849:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 27849:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 27849:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 27849:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 27849:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 27849:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 27849:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 27849:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 3 previous similar messages 19[6889]: segfault at 8 ip 00007fae0e2cd7e8 sp 00007ffced8d4370 error 4 in ld-2.17.so[7fae0e2c2000+22000] Lustre: 29946:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff8802efe63740 x1836925584035456/t4294978939(0) o101->65758477-e7bf-4e52-8671-5c8a9053fefc@0@lo:77/0 lens 376/816 e 0 to 0 dl 1751828842 ref 1 fl Interpret:H/602/0 rc 0/0 job:'dd.0' uid:0 gid:0 projid:0 Lustre: 27849:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 618, rollback = 7 Lustre: 27849:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 3 previous similar messages Lustre: 27849:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 27849:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 27849:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 27849:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 27849:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/78/0 Lustre: 27849:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 27849:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 27849:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 27849:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 27849:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 27848:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 618, rollback = 7 Lustre: 27848:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 1 previous similar message Lustre: 27848:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 27848:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 27848:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 27848:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 27848:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 27848:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 27848:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 27848:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 27848:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 27848:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 1 previous similar message LustreError: 9088:0:(file.c:248:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e8254a88: inode [0x200000401:0x3a9:0x0] mdc close failed: rc = -13 Lustre: 30044:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 618, rollback = 7 Lustre: 30044:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 19 previous similar messages Lustre: 30044:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 30044:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 19 previous similar messages Lustre: 30044:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 30044:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 19 previous similar messages Lustre: 30044:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 30044:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 19 previous similar messages Lustre: 30044:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 30044:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 19 previous similar messages Lustre: 30044:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 30044:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 19 previous similar messages LustreError: 12334:0:(file.c:248:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e8254a88: inode [0x200000401:0x4cb:0x0] mdc close failed: rc = -13 5[12386]: segfault at 8 ip 00007fec166967e8 sp 00007ffd43408b60 error 4 in ld-2.17.so[7fec1668b000+22000] Lustre: 30044:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 618, rollback = 7 Lustre: 30044:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 9 previous similar messages Lustre: 30044:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 30044:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 9 previous similar messages Lustre: 30044:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 30044:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 9 previous similar messages Lustre: 30044:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 30044:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 9 previous similar messages Lustre: 30044:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 30044:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 9 previous similar messages Lustre: 30044:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 30044:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 9 previous similar messages Lustre: 1268:0:(lod_lov.c:1402:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000401:0x551:0x0] with magic=0xbd60bd0 Lustre: 1268:0:(lod_lov.c:1402:lod_parse_striping()) Skipped 1 previous similar message Lustre: 27577:0:(lod_lov.c:1402:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000401:0x72f:0x0] with magic=0xbd60bd0 Lustre: 27577:0:(lod_lov.c:1402:lod_parse_striping()) Skipped 1 previous similar message Lustre: 761:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-OST0000: opcode 7: before 516 < left 618, rollback = 7 Lustre: 761:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 39 previous similar messages Lustre: 761:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 761:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 39 previous similar messages Lustre: 761:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 761:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 39 previous similar messages Lustre: 761:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 761:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 39 previous similar messages Lustre: 761:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 761:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 39 previous similar messages Lustre: 761:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 761:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 39 previous similar messages 12[24591]: segfault at 8 ip 00007fa112d927e8 sp 00007ffe74a0f570 error 4 in ld-2.17.so[7fa112d87000+22000] 7[24931]: segfault at 8 ip 00007f50fc9947e8 sp 00007ffeafe25af0 error 4 in ld-2.17.so[7f50fc989000+22000] 1[24992]: segfault at 8 ip 00007f2b081b37e8 sp 00007fff70ed5b90 error 4 in ld-2.17.so[7f2b081a8000+22000] 8[25129]: segfault at 0 ip (null) sp 00007ffc7f0ae6a8 error 14 in 8[400000+6000] 1[26889]: segfault at 8 ip 00007fc2b0c137e8 sp 00007ffecb67e600 error 4 in ld-2.17.so[7fc2b0c08000+22000] LustreError: 26889:0:(file.c:248:ll_close_inode_openhandle()) lustre-clilmv-ffff8800ad552548: inode [0x200000402:0x98a:0x0] mdc close failed: rc = -13 LustreError: 26889:0:(file.c:248:ll_close_inode_openhandle()) Skipped 1 previous similar message 9[27082]: segfault at 8 ip 00007f30f01067e8 sp 00007ffe8fb20110 error 4 in ld-2.17.so[7f30f00fb000+22000] Lustre: 1268:0:(lod_lov.c:1402:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000401:0xb7b:0x0] with magic=0xbd60bd0 Lustre: 1268:0:(lod_lov.c:1402:lod_parse_striping()) Skipped 1 previous similar message 3[1873]: segfault at 8 ip 00007f7c9d4557e8 sp 00007fffb2243150 error 4 in ld-2.17.so[7f7c9d44a000+22000] 5[3091]: segfault at 8 ip 00007fec052e37e8 sp 00007ffc9bd37a30 error 4 in ld-2.17.so[7fec052d8000+22000] 12[5505]: segfault at 8 ip 00007f4b0b1ad7e8 sp 00007ffc4f4d6050 error 4 in ld-2.17.so[7f4b0b1a2000+22000] Lustre: 28372:0:(lod_lov.c:1402:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000401:0xe61:0x0] with magic=0xbd60bd0 Lustre: 28372:0:(lod_lov.c:1402:lod_parse_striping()) Skipped 1 previous similar message Lustre: 29811:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 618, rollback = 7 Lustre: 29811:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 83 previous similar messages Lustre: 29811:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 29811:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 83 previous similar messages Lustre: 29811:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 29811:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 83 previous similar messages Lustre: 29811:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 29811:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 83 previous similar messages Lustre: 29811:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 29811:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 83 previous similar messages Lustre: 29811:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 29811:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 83 previous similar messages | Link to test |
racer test 1: racer on clients: centos-80.localnet DURATION=2700 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 PGD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) osp(OE) ofd(OE) lod(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) crc32_generic lnet(OE) libcfs(OE) crc_t10dif crct10dif_generic crct10dif_common pcspkr virtio_balloon virtio_console i2c_piix4 ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm ata_piix drm_panel_orientation_quirks serio_raw virtio_blk floppy i2c_core libata CPU: 15 PID: 28710 Comm: ll_sa_28686 Kdump: loaded Tainted: P OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.16.0-3.module_el8.7.0+1218+f626c2ff 04/01/2014 task: ffff8802d480b760 ti: ffff88029aabc000 task.ti: ffff88029aabc000 RIP: 0010:[<ffffffff813f502c>] [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 RSP: 0018:ffff88029aabfd78 EFLAGS: 00010296 RAX: ffff88027b4be990 RBX: 0000000000000008 RCX: 0000000100260013 RDX: 0000000000000026 RSI: ffff88027b4becc0 RDI: 0000000000000008 RBP: ffff88029aabfd88 R08: ffff8802d1876018 R09: 0000000000000000 R10: ffff8802d1873858 R11: ffff8802d1876018 R12: 0000000000000000 R13: ffff8802d8f449d0 R14: ffff88027b4becc0 R15: ffff8802d1876018 FS: 0000000000000000(0000) GS:ffff880331dc0000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 0000000001c10000 CR4: 00000000000007e0 Call Trace: [<ffffffffa15da831>] ll_statahead_thread+0xdb1/0x2000 [lustre] [<ffffffff817e320e>] ? _raw_spin_lock+0x1e/0x20 [<ffffffff810dbb51>] ? put_prev_entity+0x31/0x400 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e329e>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffff810c834d>] ? finish_task_switch+0x5d/0x1b0 [<ffffffffa15d9a80>] ? ll_statahead_handle.constprop.28+0x1c0/0x1c0 [lustre] [<ffffffff810ba114>] kthread+0xe4/0xf0 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 | Lustre: 22726:0:(lod_lov.c:1402:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000402:0x4c:0x0] with magic=0xbd60bd0 Lustre: 22726:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff8802d7952d40 x1835661204151552/t4294971925(0) o101->686b1cd5-4361-4331-90b1-70102095b23e@0@lo:21/0 lens 376/864 e 0 to 0 dl 1750623051 ref 1 fl Interpret:H/602/0 rc 0/0 job:'dd.0' uid:0 gid:0 projid:0 8[3373]: segfault at 8 ip 00007fae79d057e8 sp 00007ffff33399a0 error 4 in ld-2.17.so[7fae79cfa000+22000] 9[5285]: segfault at 8 ip 00007fe4749287e8 sp 00007fff3ed68ba0 error 4 in ld-2.17.so[7fe47491d000+22000] 19[9727]: segfault at 8 ip 00007fcbca9827e8 sp 00007ffdd20ee380 error 4 in ld-2.17.so[7fcbca977000+22000] LustreError: 26437:0:(mdt_handler.c:746:mdt_pack_acl2body()) lustre-MDT0000: unable to read [0x200000401:0x44b:0x0] ACL: rc = -2 LustreError: 11576:0:(mdt_handler.c:746:mdt_pack_acl2body()) lustre-MDT0000: unable to read [0x200000401:0x59f:0x0] ACL: rc = -2 13[22412]: segfault at 8 ip 00007f5b699467e8 sp 00007ffeb42e5d90 error 4 in ld-2.17.so[7f5b6993b000+22000] 19[26364]: segfault at 8 ip 00007feb564ca7e8 sp 00007ffd64439290 error 4 in ld-2.17.so[7feb564bf000+22000] LustreError: 30401:0:(file.c:248:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a28712a8: inode [0x200000402:0x8cf:0x0] mdc close failed: rc = -13 LustreError: 22693:0:(mdt_handler.c:746:mdt_pack_acl2body()) lustre-MDT0000: unable to read [0x200000402:0xb63:0x0] ACL: rc = -2 LustreError: 25905:0:(file.c:248:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a28712a8: inode [0x200000402:0xcf5:0x0] mdc close failed: rc = -13 1[7468]: segfault at 8 ip 00007f33805e27e8 sp 00007ffdb6bdbab0 error 4 in ld-2.17.so[7f33805d7000+22000] 6[29367]: segfault at 8 ip 00007f352582d7e8 sp 00007ffe75d73df0 error 4 in ld-2.17.so[7f3525822000+22000] LustreError: 30916:0:(file.c:248:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a28712a8: inode [0x200000402:0x12d2:0x0] mdc close failed: rc = -13 1[6116]: segfault at 8 ip 00007f9199c837e8 sp 00007ffcffb37500 error 4 in ld-2.17.so[7f9199c78000+22000] LustreError: 26477:0:(mdt_handler.c:746:mdt_pack_acl2body()) lustre-MDT0000: unable to read [0x200000401:0x1630:0x0] ACL: rc = -2 | Link to test |
racer test 1: racer on clients: centos-60.localnet DURATION=2700 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 PGD a3275067 PUD a0503067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) lustre(OE) osp(OE) ofd(OE) lod(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) jbd2 mbcache lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) crc32_generic lnet(OE) dm_flakey dm_mod libcfs(OE) crc_t10dif crct10dif_generic crct10dif_common pcspkr virtio_console i2c_piix4 virtio_balloon ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm ata_piix drm drm_panel_orientation_quirks libata virtio_blk serio_raw i2c_core floppy CPU: 11 PID: 20234 Comm: ll_sa_20144 Kdump: loaded Tainted: P OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.16.0-3.module_el8.7.0+1218+f626c2ff 04/01/2014 task: ffff88029ed9c9d0 ti: ffff88029dc08000 task.ti: ffff88029dc08000 RIP: 0010:[<ffffffff813f502c>] [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 RSP: 0018:ffff88029dc0bd78 EFLAGS: 00010296 RAX: ffff88008d6af310 RBX: 0000000000000008 RCX: 0000000100260018 RDX: 0000000000000026 RSI: ffff88008d6af640 RDI: 0000000000000008 RBP: ffff88029dc0bd88 R08: ffff8802de5431b8 R09: 0000000000000000 R10: ffff8802de546368 R11: ffff8802de5431b8 R12: 0000000000000000 R13: ffff8802b27c8010 R14: ffff88008d6af640 R15: ffff8802de5431b8 FS: 0000000000000000(0000) GS:ffff880331cc0000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 00000002d36a2000 CR4: 00000000000007e0 Call Trace: [<ffffffffa1127831>] ll_statahead_thread+0xdb1/0x2000 [lustre] [<ffffffff817e320e>] ? _raw_spin_lock+0x1e/0x20 [<ffffffff810dbb51>] ? put_prev_entity+0x31/0x400 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e329e>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffff810c834d>] ? finish_task_switch+0x5d/0x1b0 [<ffffffffa1126a80>] ? ll_statahead_handle.constprop.28+0x1c0/0x1c0 [lustre] [<ffffffff810ba114>] kthread+0xe4/0xf0 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 | Lustre: 10011:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 510 < left 922, rollback = 2 Lustre: 10011:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 3/12/3, destroy: 1/4/0 Lustre: 10011:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 4/4/0, xattr_set: 13/922/0 Lustre: 10011:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 5/55/0, punch: 0/0/0, quota 1/3/0 Lustre: 10011:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 15/263/2, delete: 3/6/0 Lustre: 10011:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 8/8/0, ref_del: 3/3/0 Lustre: 10009:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 507 < left 610, rollback = 2 Lustre: 10009:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 8 previous similar messages Lustre: 10009:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 3/12/4, destroy: 0/0/0 Lustre: 10009:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 8 previous similar messages Lustre: 10009:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 2/2/0, xattr_set: 8/610/0 Lustre: 10009:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 8 previous similar messages Lustre: 10009:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 5/55/0, punch: 0/0/0, quota 1/3/0 Lustre: 10009:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 8 previous similar messages Lustre: 10009:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 12/275/4, delete: 0/0/0 Lustre: 10009:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 8 previous similar messages Lustre: 10009:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 6/6/0, ref_del: 0/0/0 Lustre: 10009:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 8 previous similar messages Lustre: 10009:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 512 < left 1100, rollback = 2 Lustre: 10009:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 11 previous similar messages Lustre: 10009:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 3/12/1, destroy: 1/4/0 Lustre: 10009:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 11 previous similar messages Lustre: 10009:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 4/4/0, xattr_set: 15/1100/0 Lustre: 10009:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 11 previous similar messages Lustre: 10009:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 5/55/0, punch: 0/0/0, quota 7/129/0 Lustre: 10009:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 11 previous similar messages Lustre: 10009:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 15/263/2, delete: 3/6/0 Lustre: 10009:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 11 previous similar messages Lustre: 10009:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 8/8/0, ref_del: 3/3/0 Lustre: 10009:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 11 previous similar messages 1[20543]: segfault at 8 ip 00007fb688e057e8 sp 00007ffde6128690 error 4 in ld-2.17.so[7fb688dfa000+22000] LustreError: 10009:0:(mdd_dir.c:4747:mdd_migrate_cmd_check()) lustre-MDD0001: '19' migration was interrupted, run 'lfs migrate -m 0 -c 2 -H crush 19' to finish migration: rc = -1 LustreError: 10009:0:(mdt_reint.c:2564:mdt_reint_migrate()) lustre-MDT0001: migrate [0x200000403:0x2:0x0]/19 failed: rc = -1 Lustre: 18724:0:(mdd_dir.c:4826:mdd_migrate_object()) lustre-MDD0000: [0x200000403:0x1:0x0]/17 is open, migrate only dentry Lustre: 10009:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 508 < left 788, rollback = 2 Lustre: 10009:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 23 previous similar messages Lustre: 10009:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 3/12/4, destroy: 0/0/0 Lustre: 10009:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 23 previous similar messages Lustre: 10009:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 2/2/0, xattr_set: 10/788/0 Lustre: 10009:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 23 previous similar messages Lustre: 10009:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 5/55/0, punch: 0/0/0, quota 4/54/0 Lustre: 10009:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 23 previous similar messages Lustre: 10009:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 11/239/3, delete: 0/0/0 Lustre: 10009:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 23 previous similar messages Lustre: 10009:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 5/5/0, ref_del: 0/0/0 Lustre: 10009:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 23 previous similar messages Lustre: 11846:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 516 < left 618, rollback = 7 LustreError: 9992:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000404:0xc6:0x0]: rc = -2 LustreError: 21300:0:(file.c:248:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e77b92a8: inode [0x200000404:0xc6:0x0] mdc close failed: rc = -2 Lustre: 17922:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff8802c9dea340 x1835060326492800/t4294968612(0) o101->c2b9d73d-6a44-4224-b35a-6b1a10263169@0@lo:31/0 lens 376/840 e 0 to 0 dl 1750050016 ref 1 fl Interpret:H/602/0 rc 0/0 job:'dd.0' uid:0 gid:0 projid:0 LustreError: 10009:0:(mdd_dir.c:4747:mdd_migrate_cmd_check()) lustre-MDD0002: '12' migration was interrupted, run 'lfs migrate -m 1 -c 2 -H crush 12' to finish migration: rc = -1 LustreError: 10009:0:(mdt_reint.c:2564:mdt_reint_migrate()) lustre-MDT0002: migrate [0x200000403:0x1:0x0]/12 failed: rc = -1 LustreError: 9992:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000404:0xc6:0x0]: rc = -2 LustreError: 17085:0:(file.c:248:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e858dd28: inode [0x200000404:0xc6:0x0] mdc close failed: rc = -2 Lustre: 22591:0:(mdd_dir.c:4826:mdd_migrate_object()) lustre-MDD0000: [0x200000403:0x2:0x0]/1 is open, migrate only dentry Lustre: 11846:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 516 < left 618, rollback = 7 Lustre: 11846:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 1 previous similar message Lustre: 11846:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 11846:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 41 previous similar messages Lustre: 11846:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 11846:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 41 previous similar messages Lustre: 11846:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 11846:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 41 previous similar messages Lustre: 11846:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 11846:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 41 previous similar messages Lustre: 11846:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 11846:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 41 previous similar messages Lustre: 22130:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 512 < left 652, rollback = 2 Lustre: 22130:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 39 previous similar messages Lustre: 10011:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 3/12/4, destroy: 0/0/0 Lustre: 10011:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 13 previous similar messages Lustre: 10011:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 2/2/0, xattr_set: 9/699/0 Lustre: 10011:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 13 previous similar messages Lustre: 10011:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 5/55/0, punch: 0/0/0, quota 1/3/0 Lustre: 10011:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 13 previous similar messages Lustre: 10011:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 10/203/4, delete: 0/0/0 Lustre: 10011:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 13 previous similar messages Lustre: 10011:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 4/4/0, ref_del: 0/0/0 Lustre: 10011:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 13 previous similar messages Lustre: dir [0x240000404:0x66:0x0] stripe 2 readdir failed: -2, directory is partially accessed! Lustre: 18724:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 509 < left 936, rollback = 2 Lustre: 18724:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 19 previous similar messages LustreError: 27547:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000404:0xbc:0x0]: rc = -5 LustreError: 27547:0:(llite_lib.c:3769:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 Lustre: 22136:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff8802efb56940 x1835060331133568/t4294971905(0) o101->2c2a4d20-f6e5-4652-9661-2dfed01992b8@0@lo:75/0 lens 376/864 e 0 to 0 dl 1750050060 ref 1 fl Interpret:H/602/0 rc 0/0 job:'dd.0' uid:0 gid:0 projid:0 Lustre: 19060:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 516 < left 618, rollback = 7 Lustre: 19060:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 1 previous similar message Lustre: 19060:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 19060:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 36 previous similar messages Lustre: 19060:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 19060:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 36 previous similar messages Lustre: 19060:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 19060:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 36 previous similar messages Lustre: 19060:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 19060:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 36 previous similar messages Lustre: 19060:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 19060:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 36 previous similar messages Lustre: 13347:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff8802cd31e440 x1835060331642752/t4294970525(0) o101->c2b9d73d-6a44-4224-b35a-6b1a10263169@0@lo:78/0 lens 376/816 e 0 to 0 dl 1750050063 ref 1 fl Interpret:H/602/0 rc 0/0 job:'dd.0' uid:0 gid:0 projid:0 Lustre: 18724:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 509 < left 922, rollback = 2 Lustre: 18724:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 28 previous similar messages LustreError: 31792:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x280000403:0x141:0x0]: rc = -5 LustreError: 31792:0:(llite_lib.c:3769:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 32388:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x280000403:0x141:0x0]: rc = -5 LustreError: 32388:0:(llite_lib.c:3769:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 26521:0:(file.c:248:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e858dd28: inode [0x200000403:0x163:0x0] mdc close failed: rc = -13 18[1954]: segfault at 8 ip 00007f46eb2d37e8 sp 00007ffd17c963a0 error 4 in ld-2.17.so[7f46eb2c8000+22000] Lustre: 18724:0:(mdd_dir.c:4826:mdd_migrate_object()) lustre-MDD0001: [0x240000403:0x1:0x0]/3 is open, migrate only dentry Lustre: 18724:0:(mdd_dir.c:4826:mdd_migrate_object()) Skipped 1 previous similar message Lustre: 22428:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 510 < left 699, rollback = 2 Lustre: 22428:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 14 previous similar messages Lustre: 22428:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 3/12/2, destroy: 0/0/0 Lustre: 22428:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 16 previous similar messages Lustre: 22428:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 2/2/0, xattr_set: 9/699/0 Lustre: 22428:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 16 previous similar messages Lustre: 22428:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 5/55/0, punch: 0/0/0, quota 1/3/0 Lustre: 22428:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 16 previous similar messages Lustre: 22428:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 10/203/3, delete: 0/0/0 Lustre: 22428:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 16 previous similar messages Lustre: 22428:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 4/4/0, ref_del: 0/0/0 Lustre: 22428:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 16 previous similar messages Lustre: 11846:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 516 < left 618, rollback = 7 Lustre: 11846:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 1 previous similar message 0[17185]: segfault at 8 ip 00007fe71f09e7e8 sp 00007fffaa0ad730 error 4 in ld-2.17.so[7fe71f093000+22000] 16[17317]: segfault at 8 ip 00007fa1fc4467e8 sp 00007fffcbdac170 error 4 in ld-2.17.so[7fa1fc43b000+22000] Lustre: 15904:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 515 < left 618, rollback = 7 Lustre: 15904:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 5 previous similar messages 1[21011]: segfault at 8 ip 00007fa56cb1f7e8 sp 00007ffdf84ba450 error 4 in ld-2.17.so[7fa56cb14000+22000] Lustre: 10011:0:(mdd_dir.c:4826:mdd_migrate_object()) lustre-MDD0000: [0x200000403:0x1:0x0]/3 is open, migrate only dentry Lustre: 10011:0:(mdd_dir.c:4826:mdd_migrate_object()) Skipped 1 previous similar message 11[24014]: segfault at 8 ip 00007f550442c7e8 sp 00007ffc2876aca0 error 4 in ld-2.17.so[7f5504421000+22000] 6[25703]: segfault at 1c ip 00000000004017b0 sp 00007fffc8c0be00 error 6 in 6[400000+6000] 4[25898]: segfault at 8 ip 00007f9f31d6e7e8 sp 00007fff07659a70 error 4 in ld-2.17.so[7f9f31d63000+22000] 8[25868]: segfault at 8 ip 00007f95638be7e8 sp 00007ffdee2313f0 error 4 in ld-2.17.so[7f95638b3000+22000] Lustre: 11845:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-OST0000: opcode 7: before 516 < left 618, rollback = 7 Lustre: 11845:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 3 previous similar messages Lustre: 10011:0:(mdd_dir.c:4826:mdd_migrate_object()) lustre-MDD0000: [0x200000403:0x1:0x0]/10 is open, migrate only dentry Lustre: 17176:0:(lod_lov.c:1402:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0x73d:0x0] with magic=0xbd60bd0 1[30863]: segfault at 0 ip 0000000000403f00 sp 00007ffd0a601978 error 6 in 1[400000+6000] Lustre: 15904:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 515 < left 618, rollback = 7 Lustre: 15904:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 3 previous similar messages Lustre: 15904:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 15904:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 21 previous similar messages Lustre: 15904:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/1, xattr_set: 2/15/0 Lustre: 15904:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 21 previous similar messages Lustre: 15904:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 15904:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 21 previous similar messages Lustre: 15904:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 15904:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 21 previous similar messages Lustre: 15904:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 15904:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 21 previous similar messages Lustre: 10011:0:(mdd_dir.c:4826:mdd_migrate_object()) lustre-MDD0001: [0x240000403:0x1:0x0]/15 is open, migrate only dentry INFO: task mkdir:30510 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mkdir D ffff8800a4390880 11616 30510 16669 0x00000080 Call Trace: [<ffffffff817e19d9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df7ea>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa1d>] mutex_lock+0x2d/0x40 [<ffffffff81256f65>] filename_create+0x85/0x180 [<ffffffff81257c94>] ? getname_flags+0xc4/0x1a0 [<ffffffff81257f41>] user_path_create+0x41/0x60 [<ffffffff81259286>] SyS_mkdirat+0x46/0xe0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff81259339>] SyS_mkdir+0x19/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a INFO: task mkdir:31372 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mkdir D ffff8800a6db8880 13056 31372 16737 0x00000080 Call Trace: [<ffffffff817e19d9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df7ea>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa1d>] mutex_lock+0x2d/0x40 [<ffffffff81256f65>] filename_create+0x85/0x180 [<ffffffff81257c94>] ? getname_flags+0xc4/0x1a0 [<ffffffff81257f41>] user_path_create+0x41/0x60 [<ffffffff81259286>] SyS_mkdirat+0x46/0xe0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff81259339>] SyS_mkdir+0x19/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a INFO: task mkdir:301 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mkdir D ffff8802e1534240 13056 301 16749 0x00000080 Call Trace: [<ffffffff817e19d9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df7ea>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa1d>] mutex_lock+0x2d/0x40 [<ffffffff81256f65>] filename_create+0x85/0x180 [<ffffffff81257c94>] ? getname_flags+0xc4/0x1a0 [<ffffffff81257f41>] user_path_create+0x41/0x60 [<ffffffff81259286>] SyS_mkdirat+0x46/0xe0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff81259339>] SyS_mkdir+0x19/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a INFO: task mkdir:355 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mkdir D ffff8802db431878 13056 355 16528 0x00000080 Call Trace: [<ffffffff817e19d9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df7ea>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa1d>] mutex_lock+0x2d/0x40 [<ffffffff81256f65>] filename_create+0x85/0x180 [<ffffffff81257c94>] ? getname_flags+0xc4/0x1a0 [<ffffffff81257f41>] user_path_create+0x41/0x60 [<ffffffff81259286>] SyS_mkdirat+0x46/0xe0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff81259339>] SyS_mkdir+0x19/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a Lustre: 10011:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 509 < left 610, rollback = 2 Lustre: 10011:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 7 previous similar messages Lustre: 15904:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 516 < left 618, rollback = 7 Lustre: 15904:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 5 previous similar messages 6[5253]: segfault at 0 ip 00000000004043f0 sp 00007fffee734bb8 error 6 in 6[400000+6000] LustreError: 22428:0:(mdt_reint.c:2564:mdt_reint_migrate()) lustre-MDT0002: migrate [0x280000403:0x1:0x0]/12 failed: rc = -2 LustreError: 22428:0:(mdt_reint.c:2564:mdt_reint_migrate()) Skipped 1 previous similar message 8[8716]: segfault at 8 ip 00007f8c472217e8 sp 00007ffdd8ba2f00 error 4 in ld-2.17.so[7f8c47216000+22000] 12[11820]: segfault at 8 ip 00007f37b46dd7e8 sp 00007fffa318b6f0 error 4 in ld-2.17.so[7f37b46d2000+22000] 10[18722]: segfault at 8 ip 00007f4266d887e8 sp 00007ffe0e252d00 error 4 in ld-2.17.so[7f4266d7d000+22000] 9[21960]: segfault at 8 ip 00007f49b75077e8 sp 00007fff3dfd5990 error 4 in ld-2.17.so[7f49b74fc000+22000] 9[22383]: segfault at 8 ip 00007fc6588567e8 sp 00007ffdee09bc90 error 4 in ld-2.17.so[7fc65884b000+22000] 5[22804]: segfault at 406000 ip 0000000000406000 sp 00007fff56a6f538 error 14 in 5[606000+1000] INFO: task mkdir:30117 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mkdir D ffff8800a339aae8 13056 30117 16564 0x00000080 Call Trace: [<ffffffff817e19d9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df7ea>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa1d>] mutex_lock+0x2d/0x40 [<ffffffff81256f65>] filename_create+0x85/0x180 [<ffffffff81257c94>] ? getname_flags+0xc4/0x1a0 [<ffffffff81257f41>] user_path_create+0x41/0x60 [<ffffffff81259286>] SyS_mkdirat+0x46/0xe0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff81259339>] SyS_mkdir+0x19/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a INFO: task mkdir:30128 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mkdir D ffff880096834fc8 13056 30128 16716 0x00000080 Call Trace: [<ffffffff817e19d9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df7ea>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa1d>] mutex_lock+0x2d/0x40 [<ffffffff81256f65>] filename_create+0x85/0x180 [<ffffffff81257c94>] ? getname_flags+0xc4/0x1a0 [<ffffffff81257f41>] user_path_create+0x41/0x60 [<ffffffff81259286>] SyS_mkdirat+0x46/0xe0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff81259339>] SyS_mkdir+0x19/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a INFO: task mkdir:30197 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mkdir D ffff8802decc2980 13056 30197 16575 0x00000080 Call Trace: [<ffffffff817e19d9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df7ea>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa1d>] mutex_lock+0x2d/0x40 [<ffffffff81256f65>] filename_create+0x85/0x180 [<ffffffff81257c94>] ? getname_flags+0xc4/0x1a0 [<ffffffff81257f41>] user_path_create+0x41/0x60 [<ffffffff81259286>] SyS_mkdirat+0x46/0xe0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff81259339>] SyS_mkdir+0x19/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a INFO: task mkdir:30242 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mkdir D ffff8802dfd32140 13024 30242 16568 0x00000080 Call Trace: [<ffffffff817e19d9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df7ea>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa1d>] mutex_lock+0x2d/0x40 [<ffffffff81256f65>] filename_create+0x85/0x180 [<ffffffff81257c94>] ? getname_flags+0xc4/0x1a0 [<ffffffff81257f41>] user_path_create+0x41/0x60 [<ffffffff81259286>] SyS_mkdirat+0x46/0xe0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff81259339>] SyS_mkdir+0x19/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a INFO: task mkdir:32465 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mkdir D ffff8802d4ebeb80 13056 32465 16917 0x00000080 Call Trace: [<ffffffff817e19d9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df7ea>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa1d>] mutex_lock+0x2d/0x40 [<ffffffff81256f65>] filename_create+0x85/0x180 [<ffffffff81257c94>] ? getname_flags+0xc4/0x1a0 [<ffffffff81257f41>] user_path_create+0x41/0x60 [<ffffffff81259286>] SyS_mkdirat+0x46/0xe0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff81259339>] SyS_mkdir+0x19/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a INFO: task mkdir:32667 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mkdir D ffff8802e1530880 13056 32667 16655 0x00000080 Call Trace: [<ffffffff817e19d9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df7ea>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa1d>] mutex_lock+0x2d/0x40 [<ffffffff81256f65>] filename_create+0x85/0x180 [<ffffffff81257c94>] ? getname_flags+0xc4/0x1a0 [<ffffffff81257f41>] user_path_create+0x41/0x60 [<ffffffff81259286>] SyS_mkdirat+0x46/0xe0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff81259339>] SyS_mkdir+0x19/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a Lustre: 21519:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 516 < left 618, rollback = 7 Lustre: 21519:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 21 previous similar messages Lustre: 21519:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 21519:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 39 previous similar messages Lustre: 21519:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 21519:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 39 previous similar messages Lustre: 21519:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 21519:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 39 previous similar messages Lustre: 21519:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 21519:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 39 previous similar messages Lustre: 21519:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 21519:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 39 previous similar messages 4[1645]: segfault at 406000 ip 0000000000406000 sp 00007ffcbc5c7738 error 14 in 4 (deleted)[606000+1000] 9[3046]: segfault at 8 ip 00007fd7c0d477e8 sp 00007ffe775c0290 error 4 in ld-2.17.so[7fd7c0d3c000+22000] 10[3259]: segfault at 8 ip 00007ff9126c57e8 sp 00007ffcae1b33b0 error 4 in ld-2.17.so[7ff9126ba000+22000] 11[3830]: segfault at 0 ip (null) sp 00007fff3c1b1108 error 14 in 11[400000+6000] 2[3821]: segfault at 1c ip 00000000004017b0 sp 00007fff42ee72f0 error 6 in 2[400000+6000] 18[6437]: segfault at 8 ip 00007fbb886da7e8 sp 00007ffeb7f66b30 error 4 in ld-2.17.so[7fbb886cf000+22000] Lustre: 10011:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 509 < left 610, rollback = 2 Lustre: 10011:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 11 previous similar messages LustreError: 8091:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x280000404:0x770:0x0]: rc = -5 LustreError: 8091:0:(llite_lib.c:3769:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 10[9679]: segfault at 404540 ip 0000000000404d7a sp 00007ffcbc2568b8 error 7 in 10[400000+6000] 9[12292]: segfault at 8 ip 00007f42895527e8 sp 00007ffc29dda610 error 4 in ld-2.17.so[7f4289547000+22000] LustreError: 22130:0:(mdt_reint.c:2564:mdt_reint_migrate()) lustre-MDT0001: migrate [0x200000403:0xfdc:0x0]/5 failed: rc = -2 Lustre: 19060:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-OST0002: opcode 7: before 516 < left 618, rollback = 7 Lustre: 19060:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 99 previous similar messages LustreError: 25275:0:(file.c:248:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e858dd28: inode [0x280000403:0x7c6:0x0] mdc close failed: rc = -13 LustreError: 25275:0:(file.c:248:ll_close_inode_openhandle()) Skipped 1 previous similar message 18[3532]: segfault at 8 ip 00007f77757fb7e8 sp 00007ffc219a44e0 error 4 in ld-2.17.so[7f77757f0000+22000] 17[3783]: segfault at 0 ip 0000000000403e5f sp 00007fff2127f700 error 6 in 17[400000+6000] 19[7439]: segfault at 8 ip 00007f883dd727e8 sp 00007ffcaddc5130 error 4 in ld-2.17.so[7f883dd67000+22000] 0[7429]: segfault at 8 ip 00007fac1ea737e8 sp 00007ffdf0dc44b0 error 4 in ld-2.17.so[7fac1ea68000+22000] 15[7635]: segfault at 8 ip 00007fa2789127e8 sp 00007ffc53ed6f60 error 4 in ld-2.17.so[7fa278907000+22000] Lustre: 22428:0:(mdd_dir.c:4826:mdd_migrate_object()) lustre-MDD0001: [0x240000403:0x1:0x0]/0 is open, migrate only dentry 6[28054]: segfault at 0 ip (null) sp 00007ffcb3b25cc8 error 14 in 6[400000+6000] 5[30084]: segfault at 8 ip 00007f0d7502c7e8 sp 00007ffd0c502df0 error 4 in ld-2.17.so[7f0d75021000+22000] 11[32074]: segfault at c0cfcd ip 0000000000404768 sp 00007ffed9c70958 error 6 11[31987]: segfault at c0cfcd ip 0000000000404768 sp 00007ffe279dd6e8 error 6 in 11[400000+6000] in 11[400000+6000] 14[32091]: segfault at 5468 ip 00000000004059a7 sp 00007fffdb45b7b8 error 6 in 14[400000+6000] 1[31854]: segfault at 0 ip 0000000000403e5f sp 00007ffce4fcacd0 error 6 in 1[400000+6000] Lustre: 18724:0:(mdd_dir.c:4826:mdd_migrate_object()) lustre-MDD0000: [0x200000403:0x1:0x0]/17 is open, migrate only dentry 0[2516]: segfault at 8 ip 00007fc7b739a7e8 sp 00007ffe6f8649f0 error 4 in ld-2.17.so[7fc7b738f000+22000] Lustre: 18283:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 18283:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 183 previous similar messages Lustre: 18283:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/1, xattr_set: 2/15/0 Lustre: 18283:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 183 previous similar messages Lustre: 18283:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 18283:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 183 previous similar messages Lustre: 18283:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 18283:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 183 previous similar messages Lustre: 18283:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 18283:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 183 previous similar messages 2[5743]: segfault at 8 ip 00007f48166b97e8 sp 00007ffc77f16650 error 4 in ld-2.17.so[7f48166ae000+22000] 0[10537]: segfault at 0 ip (null) sp 00007ffc914994d8 error 14 in 0[400000+6000] 16[24365]: segfault at 0 ip (null) sp 00007ffd2c2590f8 error 14 in 16[400000+6000] 10[25997]: segfault at 8 ip 00007f3a928817e8 sp 00007ffcba3c62c0 error 4 in ld-2.17.so[7f3a92876000+22000] 19[31176]: segfault at 4017b0 ip 00000000004017b0 sp 00007ffc89ab0540 error 14 15[1985]: segfault at 0 ip (null) sp 00007ffe1fcb88b8 error 14 in 15[400000+6000] 17[7796]: segfault at 8 ip 00007fb08e68d7e8 sp 00007ffc0adecab0 error 4 in ld-2.17.so[7fb08e682000+22000] Lustre: 20432:0:(lod_lov.c:1402:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000404:0x11b2:0x0] with magic=0xbd60bd0 Lustre: 20432:0:(lod_lov.c:1402:lod_parse_striping()) Skipped 1 previous similar message Lustre: 22428:0:(mdd_dir.c:4826:mdd_migrate_object()) lustre-MDD0002: [0x280000403:0x1:0x0]/10 is open, migrate only dentry Lustre: 22428:0:(mdd_dir.c:4826:mdd_migrate_object()) Skipped 1 previous similar message 9[12472]: segfault at 0 ip 0000000000403e5f sp 00007ffe22c9d380 error 6 in 7[400000+6000] Lustre: 18283:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 516 < left 618, rollback = 7 Lustre: 18283:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 165 previous similar messages 0[23995]: segfault at 8 ip 00007f67531127e8 sp 00007ffca079e470 error 4 in ld-2.17.so[7f6753107000+22000] Lustre: 20432:0:(lod_lov.c:1402:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000404:0x133b:0x0] with magic=0xbd60bd0 Lustre: 20432:0:(lod_lov.c:1402:lod_parse_striping()) Skipped 1 previous similar message LustreError: 25331:0:(file.c:248:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e858dd28: inode [0x280000404:0x1306:0x0] mdc close failed: rc = -13 Lustre: lustre-OST0001-osc-ffff8802e77b92a8: disconnect after 21s idle Lustre: mdt00_019: service thread pid 21140 was inactive for 40.105 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 21140, comm: mdt00_019 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_pdo_lock+0x369/0x750 [mdt] [<0>] mdt_parent_lock+0x76/0x2a0 [mdt] [<0>] mdt_reint_open+0x1274/0x2ed0 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x73c/0xbb0 [mdt] [<0>] mdt_intent_open+0x93/0x480 [mdt] [<0>] mdt_intent_opc.constprop.74+0x211/0xc60 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x397/0x980 [ptlrpc] [<0>] ldlm_handle_enqueue+0x547/0x1900 [ptlrpc] [<0>] tgt_enqueue+0x68/0x240 [ptlrpc] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x23d/0xd80 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 10011, comm: mdt_io00_002 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_fini+0xabc/0xfd0 [ptlrpc] [<0>] ldlm_cli_enqueue+0x461/0xb00 [ptlrpc] [<0>] osp_md_object_lock+0x151/0x2f0 [osp] [<0>] lod_object_lock+0xdb/0x7c0 [lod] [<0>] mdd_object_lock+0x2d/0xd0 [mdd] [<0>] mdt_remote_object_lock_try+0x14c/0x189 [mdt] [<0>] mdt_object_lock_internal+0x3c4/0x470 [mdt] [<0>] mdt_rename_lock+0xd9/0x360 [mdt] [<0>] mdt_reint_migrate+0x87e/0x23e0 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x73c/0xbb0 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x23d/0xd80 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 22389, comm: mdt_io00_007 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock+0x88/0x1c0 [mdt] [<0>] mdt_rename_source_lock+0x57/0xf0 [mdt] [<0>] mdt_reint_migrate+0x1763/0x23e0 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x73c/0xbb0 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x23d/0xd80 [ptlrpc] Lustre: mdt00_007: service thread pid 17491 was inactive for 40.156 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: mdt_io00_006: service thread pid 22374 was inactive for 40.047 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 3 previous similar messages Lustre: mdt00_005: service thread pid 13846 was inactive for 40.025 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 4 previous similar messages Lustre: mdt00_010: service thread pid 18677 was inactive for 40.117 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 8 previous similar messages Lustre: mdt_io00_003: service thread pid 18724 was inactive for 72.033 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 4 previous similar messages LustreError: 9825:0:(ldlm_lockd.c:257:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0000_UUID lock: ffff8802bfe13880/0x6c82a01bb7e846e5 lrc: 3/0,0 mode: PW/PW res: [0x200000403:0x2304:0x0].0x0 bits 0x4/0x0 rrc: 7 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0x6c82a01bb7e846d0 expref: 1744 pid: 17610 timeout: 1001 lvb_type: 0 Lustre: mdt00_022: service thread pid 22069 completed after 96.469s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: 22389:0:(mdd_dir.c:4826:mdd_migrate_object()) lustre-MDD0000: [0x200000403:0x1:0x0]/10 is open, migrate only dentry Lustre: 22389:0:(mdd_dir.c:4826:mdd_migrate_object()) Skipped 1 previous similar message LustreError: 18677:0:(ldlm_lockd.c:1447:ldlm_handle_enqueue()) ### lock on destroyed export ffff8802e6d892a8 ns: mdt-lustre-MDT0000_UUID lock: ffff8802e2504000/0x6c82a01bb7e8ba5e lrc: 3/0,0 mode: PR/PR res: [0x200000403:0x1:0x0].0x0 bits 0x13/0x0 rrc: 21 type: IBT gid 0 flags: 0x50200000000000 nid: 0@lo remote: 0x6c82a01bb7e8ba49 expref: 779 pid: 18677 timeout: 0 lvb_type: 0 Lustre: mdt00_013: service thread pid 19461 completed after 96.050s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_023: service thread pid 22136 completed after 95.801s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_009: service thread pid 17922 completed after 96.001s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_006: service thread pid 17176 completed after 97.127s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_017: service thread pid 20953 completed after 95.574s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_028: service thread pid 22682 completed after 97.803s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_018: service thread pid 21108 completed after 95.467s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: lustre-MDT0000-mdc-ffff8802e77b92a8: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-MDT0000-mdc-ffff8802e77b92a8: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: mdt00_011: service thread pid 18687 completed after 96.002s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_000: service thread pid 9988 completed after 95.838s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_019: service thread pid 21140 completed after 100.206s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_007: service thread pid 17491 completed after 100.218s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_007: service thread pid 22389 completed after 100.215s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: lustre-MDT0000-mdc-ffff8802e77b92a8: This client was evicted by lustre-MDT0000; in progress operations using this service will fail. Lustre: mdt00_003: service thread pid 13347 completed after 95.467s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_029: service thread pid 22692 completed after 98.684s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_010: service thread pid 18677 completed after 95.735s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 16211:0:(file.c:6195:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000403:0x1:0x0] error: rc = -5 Lustre: mdt_io00_002: service thread pid 10011 completed after 100.161s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_005: service thread pid 13846 completed after 97.711s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_004: service thread pid 13844 completed after 100.026s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 19245:0:(file.c:248:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e77b92a8: inode [0x200000403:0x22f1:0x0] mdc close failed: rc = -108 Lustre: mdt_io00_011: service thread pid 22657 completed after 100.023s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 25485:0:(llite_lib.c:2023:ll_md_setattr()) md_setattr fails: rc = -108 LustreError: 25485:0:(llite_lib.c:2023:ll_md_setattr()) Skipped 1 previous similar message LustreError: 25444:0:(vvp_io.c:1905:vvp_io_init()) lustre: refresh file layout [0x200000403:0x22f1:0x0] error -108. Lustre: mdt00_002: service thread pid 9990 completed after 95.548s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 15118:0:(mdc_request.c:1470:mdc_read_page()) lustre-MDT0000-mdc-ffff8802e77b92a8: [0x200000403:0x2:0x0] lock enqueue fails: rc = -108 Lustre: mdt_io00_008: service thread pid 22428 completed after 100.071s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 19184:0:(mdt_reint.c:2564:mdt_reint_migrate()) lustre-MDT0000: migrate [0x200000403:0x2:0x0]/11 failed: rc = -2 Lustre: mdt_io00_004: service thread pid 19184 completed after 99.870s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_005: service thread pid 22130 completed after 99.991s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: lustre-MDT0000-mdc-ffff8802e77b92a8: Connection restored to 192.168.123.62@tcp (at 0@lo) Lustre: mdt_io00_006: service thread pid 22374 completed after 99.971s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_009: service thread pid 22503 completed after 99.849s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: 18724:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 504 < left 699, rollback = 2 Lustre: 18724:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 84 previous similar messages Lustre: mdt_io00_003: service thread pid 18724 completed after 93.404s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_001: service thread pid 10010 completed after 93.080s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). hrtimer: interrupt took 13825468 ns 14[31022]: segfault at 8 ip 00007f43a7e877e8 sp 00007ffd6db12420 error 4 in ld-2.17.so[7f43a7e7c000+22000] 6[30890]: segfault at 1c ip 00000000004017b0 sp 00007ffc01d6f580 error 6 in 6[400000+6000] 4[32526]: segfault at 8 ip 00007f5334ff77e8 sp 00007ffdea96bff0 error 4 in ld-2.17.so[7f5334fec000+22000] 14[1426]: segfault at 8 ip 00007f65a43d67e8 sp 00007ffe96983970 error 4 in ld-2.17.so[7f65a43cb000+22000] Lustre: 22657:0:(mdt_reint.c:2484:mdt_reint_migrate()) lustre-MDT0000: [0x200000403:0x2:0x0]/13 is open, migrate only dentry Lustre: 21157:0:(lod_lov.c:1402:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000404:0x15d3:0x0] with magic=0xbd60bd0 Lustre: 21157:0:(lod_lov.c:1402:lod_parse_striping()) Skipped 1 previous similar message 0[18175]: segfault at 0 ip 0000000000403e5f sp 00007ffed781d580 error 6 in 0[400000+6000] 3[23345]: segfault at 8 ip 00007fcb7f3047e8 sp 00007ffd7e2708c0 error 4 in ld-2.17.so[7fcb7f2f9000+22000] 3[23482]: segfault at 8 ip 00007fe6dc50c7e8 sp 00007ffddb963060 error 4 in ld-2.17.so[7fe6dc501000+22000] 16[28379]: segfault at 8 ip 00007fd8484bf7e8 sp 00007ffd336430a0 error 4 in ld-2.17.so[7fd8484b4000+22000] 18[29090]: segfault at 8 ip 00007f6b2baef7e8 sp 00007fff57867800 error 4 in ld-2.17.so[7f6b2bae4000+22000] 7[30274]: segfault at 8 ip 00007f46d48277e8 sp 00007ffd0cb55100 error 4 in ld-2.17.so[7f46d481c000+22000] 18[31121]: segfault at 8 ip 00007fc48459b7e8 sp 00007ffd8d4184a0 error 4 in ld-2.17.so[7fc484590000+22000] 11[12697]: segfault at 8 ip 00007ff969c327e8 sp 00007ffe62a06f70 error 4 in ld-2.17.so[7ff969c27000+22000] 16[19540]: segfault at 8 ip 00007fee37a417e8 sp 00007ffd09a4aaf0 error 4 in ld-2.17.so[7fee37a36000+22000] LustreError: 20460:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000404:0x18c8:0x0]: rc = -5 LustreError: 20460:0:(llite_lib.c:3769:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 17[20553]: segfault at 0 ip 0000000000403e5f sp 00007ffc41e03780 error 6 in 17[400000+6000] LustreError: 9993:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000405:0x97e:0x0]: rc = -2 LustreError: 15914:0:(file.c:248:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e77b92a8: inode [0x200000405:0x97e:0x0] mdc close failed: rc = -2 LustreError: 15914:0:(file.c:248:ll_close_inode_openhandle()) Skipped 6 previous similar messages LustreError: 22497:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000404:0x18c8:0x0]: rc = -5 LustreError: 22497:0:(llite_lib.c:3769:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 3[24661]: segfault at 8 ip 00007f8bb3abb7e8 sp 00007ffc60619930 error 4 in ld-2.17.so[7f8bb3ab0000+22000] 17[26257]: segfault at 8 ip 00007f3b268d57e8 sp 00007ffdb718e0b0 error 4 in ld-2.17.so[7f3b268ca000+22000] LustreError: 26319:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000404:0x18c8:0x0]: rc = -5 LustreError: 26319:0:(llite_lib.c:3769:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 Lustre: 21519:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 21519:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 357 previous similar messages Lustre: 21519:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 21519:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 357 previous similar messages Lustre: 21519:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 21519:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 357 previous similar messages Lustre: 21519:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 21519:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 357 previous similar messages Lustre: 21519:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 21519:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 357 previous similar messages Lustre: 22657:0:(mdd_dir.c:4826:mdd_migrate_object()) lustre-MDD0001: [0x240000403:0x1:0x0]/12 is open, migrate only dentry Lustre: 22657:0:(mdd_dir.c:4826:mdd_migrate_object()) Skipped 5 previous similar messages 14[2866]: segfault at 8 ip 00007f9b1041c7e8 sp 00007ffd85501060 error 4 in ld-2.17.so[7f9b10411000+22000] 1[3666]: segfault at 8 ip 00007fd76affd7e8 sp 00007ffcab0911d0 error 4 in ld-2.17.so[7fd76aff2000+22000] 8[4074]: segfault at 8 ip 00007fb55ee0c7e8 sp 00007fff952fc930 error 4 in ld-2.17.so[7fb55ee01000+22000] 10[4938]: segfault at 8 ip 00007fcbe957e7e8 sp 00007ffd993145c0 error 4 in ld-2.17.so[7fcbe9573000+22000] 5[5408]: segfault at 8 ip 00007f13162597e8 sp 00007fffde4bcef0 error 4 in ld-2.17.so[7f131624e000+22000] 1[5216]: segfault at 8 ip 00007f0d5720c7e8 sp 00007ffd2f71df50 error 4 in ld-2.17.so[7f0d57201000+22000] LustreError: 22657:0:(mdt_reint.c:2564:mdt_reint_migrate()) lustre-MDT0000: migrate [0x240000404:0x1b29:0x0]/12 failed: rc = -2 LustreError: 7656:0:(llite_lib.c:1872:ll_update_lsm_md()) lustre: [0x200000403:0x3154:0x0] dir layout mismatch: LustreError: 7656:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=2 count=3 index=0 hash=crush:0x2000003 max_inherit=0 max_inherit_rr=0 version=2 migrate_offset=0 migrate_hash=invalid:0 pool= LustreError: 7656:0:(lustre_lmv.h:167:lmv_stripe_object_dump()) stripe[0] [0x200000400:0x37:0x0] LustreError: 7656:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=1 count=4 index=0 hash=crush:0x82000003 max_inherit=0 max_inherit_rr=0 version=1 migrate_offset=3 migrate_hash=fnv_1a_64:2 pool= LustreError: 9003:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x280000404:0x1a47:0x0]: rc = -5 LustreError: 7659:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=3 count=3 index=0 hash=crush:0x2000003 max_inherit=0 max_inherit_rr=0 version=2 migrate_offset=0 migrate_hash=invalid:0 pool= LustreError: 7659:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=1 count=4 index=0 hash=crush:0x82000003 max_inherit=0 max_inherit_rr=0 version=1 migrate_offset=3 migrate_hash=fnv_1a_64:2 pool= LustreError: 9003:0:(llite_lib.c:3769:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 Lustre: 32001:0:(out_handler.c:879:out_tx_end()) lustre-MDT0001-osd: error during execution of #0 from /home/green/git/lustre-release/lustre/ptlrpc/../../lustre/target/out_handler.c:141: rc = -2 3[11159]: segfault at 8 ip 00007fe0c46957e8 sp 00007ffca3ff8ea0 error 4 in ld-2.17.so[7fe0c468a000+22000] LustreError: 22657:0:(mdt_reint.c:2564:mdt_reint_migrate()) lustre-MDT0002: migrate [0x280000403:0x1:0x0]/12 failed: rc = -2 Lustre: dir [0x240000404:0x1be0:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: dir [0x280000404:0x19e8:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 4 previous similar messages LustreError: 14685:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000403:0x1ac9:0x0]: rc = -5 LustreError: 14685:0:(lcommon_cl.c:188:cl_file_inode_init()) Skipped 1 previous similar message LustreError: 14685:0:(llite_lib.c:3769:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 14685:0:(llite_lib.c:3769:ll_prep_inode()) Skipped 1 previous similar message Lustre: dir [0x240000403:0x1a8a:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 2 previous similar messages LustreError: 22389:0:(mdd_dir.c:4747:mdd_migrate_cmd_check()) lustre-MDD0002: '4' migration was interrupted, run 'lfs migrate -m 0 -c 2 -H crush 4' to finish migration: rc = -1 Lustre: dir [0x280000403:0x1919:0x0] stripe 2 readdir failed: -2, directory is partially accessed! LustreError: 22389:0:(mdt_reint.c:2564:mdt_reint_migrate()) lustre-MDT0001: migrate [0x200000405:0xe08:0x0]/16 failed: rc = -2 LustreError: 22389:0:(mdt_reint.c:2564:mdt_reint_migrate()) Skipped 4 previous similar messages Lustre: 21140:0:(lod_lov.c:1402:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000405:0x1050:0x0] with magic=0xbd60bd0 Lustre: 21140:0:(lod_lov.c:1402:lod_parse_striping()) Skipped 1 previous similar message LustreError: 22500:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x280000404:0x1be0:0x0]: rc = -5 LustreError: 22500:0:(llite_lib.c:3769:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 18838:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0002: failed to get lu_attr of [0x280000403:0x1919:0x0]: rc = -2 LustreError: 21810:0:(file.c:248:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e858dd28: inode [0x280000403:0x1919:0x0] mdc close failed: rc = -2 Lustre: 22139:0:(lod_lov.c:1402:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000403:0x1cb0:0x0] with magic=0xbd60bd0 Lustre: 22139:0:(lod_lov.c:1402:lod_parse_striping()) Skipped 11 previous similar messages LustreError: 28448:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000403:0x1c69:0x0]: rc = -5 LustreError: 28448:0:(lcommon_cl.c:188:cl_file_inode_init()) Skipped 5 previous similar messages LustreError: 28448:0:(llite_lib.c:3769:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 28448:0:(llite_lib.c:3769:ll_prep_inode()) Skipped 5 previous similar messages Lustre: 11845:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 515 < left 618, rollback = 7 Lustre: 11845:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 191 previous similar messages 19[29149]: segfault at 8 ip 00007fbee09f57e8 sp 00007ffc1be06490 error 4 in ld-2.17.so[7fbee09ea000+22000] LustreError: 17503:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0002: failed to get lu_attr of [0x280000403:0x1919:0x0]: rc = -2 Lustre: 9988:0:(lod_lov.c:1402:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000404:0x1daa:0x0] with magic=0xbd60bd0 Lustre: 9988:0:(lod_lov.c:1402:lod_parse_striping()) Skipped 1 previous similar message LustreError: 22428:0:(mdt_reint.c:2564:mdt_reint_migrate()) lustre-MDT0002: migrate [0x280000403:0x1:0x0]/17 failed: rc = -16 1[9520]: segfault at 8 ip 00007f92cdbe37e8 sp 00007fff87d6cb10 error 4 in ld-2.17.so[7f92cdbd8000+22000] 14[14226]: segfault at 8 ip 00007fb1810a67e8 sp 00007ffecfa04d50 error 4 in ld-2.17.so[7fb18109b000+22000] 12[14471]: segfault at 8 ip 00007f9f51a8c7e8 sp 00007ffe2d039e90 error 4 in ld-2.17.so[7f9f51a81000+22000] 12[14543]: segfault at 8 ip 00007fb082e9b7e8 sp 00007ffd7cf0ef00 error 4 in ld-2.17.so[7fb082e90000+22000] 10[14685]: segfault at 8 ip 00007f12e2aca7e8 sp 00007ffd8c8bdaa0 error 4 in ld-2.17.so[7f12e2abf000+22000] 6[15323]: segfault at 8 ip 00007f6b72e107e8 sp 00007ffe6cb87a70 error 4 in ld-2.17.so[7f6b72e05000+22000] 17[20562]: segfault at 401ac0 ip 0000000000401ac0 sp 00007ffecc36c698 error 7 in 17[400000+6000] 17[20913]: segfault at 8 ip 00007fef5c6067e8 sp 00007ffe5323f130 error 4 in ld-2.17.so[7fef5c5fb000+22000] 11[21949]: segfault at 406000 ip 0000000000406000 sp 00007fffad969318 error 14 in 11[606000+1000] 15[24905]: segfault at 8 ip 00007f13e0ec47e8 sp 00007ffcd871cb10 error 4 in ld-2.17.so[7f13e0eb9000+22000] 10[25481]: segfault at 0 ip 0000000000403e5f sp 00007ffc2c27ee20 error 6 in 10[400000+6000] LustreError: 21053:0:(file.c:248:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e858dd28: inode [0x200000403:0x3a46:0x0] mdc close failed: rc = -13 LustreError: 21053:0:(file.c:248:ll_close_inode_openhandle()) Skipped 1 previous similar message Lustre: 15975:0:(lod_lov.c:1402:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000404:0x2067:0x0] with magic=0xbd60bd0 Lustre: 15975:0:(lod_lov.c:1402:lod_parse_striping()) Skipped 3 previous similar messages 18[3683]: segfault at 8 ip 00007f11423657e8 sp 00007ffec8bf0c00 error 4 in ld-2.17.so[7f114235a000+22000] Lustre: 11214:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 498 < left 1055, rollback = 2 Lustre: 11214:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 729 previous similar messages 8[9910]: segfault at 406000 ip 0000000000405fff sp 00007ffd6b83dd70 error 14 in 8[400000+6000] 8[9309]: segfault at 4045bc ip 00000000004045bc sp 00007ffcb5de50a8 error 7 in 8[400000+6000] 8[12751]: segfault at 8 ip 00007fb4a5ee07e8 sp 00007fff029dae70 error 4 in ld-2.17.so[7fb4a5ed5000+22000] 18[12945]: segfault at 1c ip 00000000004017b0 sp 00007ffd8c0bac90 error 6 in 18 (deleted)[400000+6000] 6[16121]: segfault at 8 ip 00007f7cc08197e8 sp 00007ffdc9f95590 error 4 in ld-2.17.so[7f7cc080e000+22000] | Link to test |
racer test 1: racer on clients: centos-30.localnet DURATION=2700 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 PGD 2e156d067 PUD 2c7c82067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) lustre(OE) osp(OE) ofd(OE) lod(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) jbd2 mbcache lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) crc32_generic lnet(OE) dm_flakey dm_mod libcfs(OE) crc_t10dif crct10dif_generic crct10dif_common virtio_balloon pcspkr virtio_console i2c_piix4 ip_tables rpcsec_gss_krb5 ata_generic drm_kms_helper pata_acpi ttm drm drm_panel_orientation_quirks ata_piix floppy serio_raw virtio_blk i2c_core libata CPU: 2 PID: 28975 Comm: ll_sa_28963 Kdump: loaded Tainted: P OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.16.0-3.module_el8.7.0+1218+f626c2ff 04/01/2014 task: ffff8802da325c40 ti: ffff8802ce3dc000 task.ti: ffff8802ce3dc000 RIP: 0010:[<ffffffff813f502c>] [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 RSP: 0018:ffff8802ce3dfd78 EFLAGS: 00010296 RAX: ffff8802c50bf310 RBX: 0000000000000008 RCX: 000000010026001c RDX: 0000000000000026 RSI: ffff8802c50bf640 RDI: 0000000000000008 RBP: ffff8802ce3dfd88 R08: ffff8802c8ef6368 R09: 0000000000000000 R10: ffff8802c8ef0d48 R11: ffff8802c8ef6368 R12: 0000000000000000 R13: ffff8802cbb91280 R14: ffff8802c50bf640 R15: ffff8802c8ef6368 FS: 0000000000000000(0000) GS:ffff880331a80000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 00000000a05fa000 CR4: 00000000000007e0 Call Trace: [<ffffffffa1110601>] ll_statahead_thread+0xdb1/0x2000 [lustre] [<ffffffff817e320e>] ? _raw_spin_lock+0x1e/0x20 [<ffffffff810dbb51>] ? put_prev_entity+0x31/0x400 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e329e>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffff810c834d>] ? finish_task_switch+0x5d/0x1b0 [<ffffffffa110f850>] ? ll_statahead_handle.constprop.28+0x1c0/0x1c0 [lustre] [<ffffffff810ba114>] kthread+0xe4/0xf0 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 | dd (15831) used greatest stack depth: 9600 bytes left LustreError: 15933:0:(file.c:248:ll_close_inode_openhandle()) lustre-clilmv-ffff8802f0054a88: inode [0x200000401:0x31:0x0] mdc close failed: rc = -13 Lustre: 10395:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-OST0000: opcode 7: before 516 < left 618, rollback = 7 Lustre: 10395:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 10395:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 10395:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 10395:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 10395:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 LustreError: 16231:0:(file.c:248:ll_close_inode_openhandle()) lustre-clilmv-ffff8802f0054a88: inode [0x200000401:0x81:0x0] mdc close failed: rc = -13 Lustre: 10431:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff8800a97e3c40 x1834598606298368/t4294970122(0) o101->8960cd2f-085c-4aa8-9f8c-16c929fc9fe2@0@lo:589/0 lens 376/816 e 0 to 0 dl 1749609654 ref 1 fl Interpret:H/602/0 rc 0/0 job:'dd.0' uid:0 gid:0 projid:0 Lustre: 16202:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-OST0000: opcode 7: before 516 < left 618, rollback = 7 Lustre: 16202:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 1 previous similar message Lustre: 16202:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 16202:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 16202:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 16202:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 16202:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 16202:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 16202:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 16202:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 16202:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 16202:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 10395:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 618, rollback = 7 Lustre: 10395:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 5 previous similar messages Lustre: 10395:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 10395:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 10395:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 10395:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 10395:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 10395:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 10395:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 10395:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 10395:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 10395:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 10395:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 618, rollback = 7 Lustre: 10395:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 1 previous similar message Lustre: 10395:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 10395:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 10395:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 10395:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 10395:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 10395:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 10395:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 10395:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 10395:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 10395:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 10395:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 515 < left 618, rollback = 7 Lustre: 10395:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 1 previous similar message Lustre: 10395:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 10395:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 10395:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/1, xattr_set: 2/15/0 Lustre: 10395:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 10395:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 10395:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 10395:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 10395:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 10395:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 10395:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 1 previous similar message 4[21550]: segfault at 406000 ip 0000000000406000 sp 00007ffd497dbdd8 error 14 in 1[606000+1000] 1[21558]: segfault at 406000 ip 0000000000406000 sp 00007fff7061b068 error 14 in 1[606000+1000] Lustre: 19370:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 618, rollback = 7 Lustre: 19370:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 1 previous similar message Lustre: 19370:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 19370:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 19370:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 19370:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 19370:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 19370:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 19370:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 19370:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 19370:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 19370:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 10397:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-OST0000: opcode 7: before 516 < left 618, rollback = 7 Lustre: 10397:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 15 previous similar messages Lustre: 10397:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 10397:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 15 previous similar messages Lustre: 10397:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 10397:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 15 previous similar messages Lustre: 10397:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 10397:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 15 previous similar messages Lustre: 10397:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 10397:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 15 previous similar messages Lustre: 10397:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 10397:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 15 previous similar messages Lustre: 14656:0:(lod_lov.c:1402:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000402:0x401:0x0] with magic=0xbd60bd0 12[29067]: segfault at 8 ip 00007f3e52d417e8 sp 00007ffcdea7b730 error 4 in ld-2.17.so[7f3e52d36000+22000] LustreError: 29597:0:(file.c:248:ll_close_inode_openhandle()) lustre-clilmv-ffff8802ee98b7e8: inode [0x200000402:0x3ee:0x0] mdc close failed: rc = -13 Lustre: 15105:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 516 < left 618, rollback = 7 Lustre: 15105:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 13 previous similar messages Lustre: 15105:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 15105:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 13 previous similar messages Lustre: 15105:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 15105:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 13 previous similar messages Lustre: 15105:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 15105:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 13 previous similar messages Lustre: 15105:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 15105:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 13 previous similar messages Lustre: 15105:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 15105:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 13 previous similar messages 15[1413]: segfault at 8 ip 00007fbeed3807e8 sp 00007fff899201d0 error 4 in ld-2.17.so[7fbeed375000+22000] LustreError: 1413:0:(file.c:248:ll_close_inode_openhandle()) lustre-clilmv-ffff8802ee98b7e8: inode [0x200000401:0x577:0x0] mdc close failed: rc = -13 LustreError: 4101:0:(file.c:248:ll_close_inode_openhandle()) lustre-clilmv-ffff8802ee98b7e8: inode [0x200000402:0x69f:0x0] mdc close failed: rc = -13 Lustre: 14437:0:(lod_lov.c:1402:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000402:0x638:0x0] with magic=0xbd60bd0 Lustre: 14437:0:(lod_lov.c:1402:lod_parse_striping()) Skipped 1 previous similar message LustreError: 6647:0:(statahead.c:2399:start_statahead_thread()) lustre: invalid pattern 0X0. 10[8692]: segfault at 8 ip 00007fbca47287e8 sp 00007fffbd6b89c0 error 4 in ld-2.17.so[7fbca471d000+22000] 16[9865]: segfault at 0 ip 0000000000403e5f sp 00007fff46b7b7f0 error 6 in 16[400000+6000] Lustre: 15105:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 516 < left 618, rollback = 7 Lustre: 15105:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 17 previous similar messages Lustre: 15105:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 15105:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 17 previous similar messages Lustre: 15105:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 15105:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 17 previous similar messages Lustre: 15105:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 15105:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 17 previous similar messages Lustre: 15105:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 15105:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 17 previous similar messages Lustre: 15105:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 15105:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 17 previous similar messages 17[10597]: segfault at 8 ip 00007fe0ccdf07e8 sp 00007ffc6c287b80 error 4 in ld-2.17.so[7fe0ccde5000+22000] LustreError: 10597:0:(file.c:248:ll_close_inode_openhandle()) lustre-clilmv-ffff8802ee98b7e8: inode [0x200000401:0x862:0x0] mdc close failed: rc = -13 1[16991]: segfault at 8 ip 00007f68cd6a27e8 sp 00007ffdc09eafa0 error 4 in ld-2.17.so[7f68cd697000+22000] 6[22532]: segfault at 8 ip 00007f8e86df67e8 sp 00007ffdfb5cc060 error 4 in ld-2.17.so[7f8e86deb000+22000] LustreError: 22532:0:(file.c:248:ll_close_inode_openhandle()) lustre-clilmv-ffff8802f0054a88: inode [0x200000402:0xb9e:0x0] mdc close failed: rc = -13 | Link to test |
racer test 2: racer rename: centos-5.localnet DURATION=2700 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 Lustre: lustre-OST0001-osc-MDT0000: update sequence from 0x300000402 to 0x300000405 PGD 2dd918067 PUD 2da720067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) llz4hc(OE) llz4(OE) lustre(OE) osp(OE) ofd(OE) lod(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) jbd2 mbcache lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) crc32_generic lnet(OE) dm_flakey dm_mod libcfs(OE) crc_t10dif crct10dif_generic crct10dif_common virtio_console virtio_balloon pcspkr i2c_piix4 ip_tables rpcsec_gss_krb5 ata_generic drm_kms_helper pata_acpi ttm drm drm_panel_orientation_quirks ata_piix virtio_blk i2c_core serio_raw libata floppy CPU: 11 PID: 22641 Comm: ll_sa_22487 Kdump: loaded Tainted: P OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.16.0-3.module_el8.7.0+1218+f626c2ff 04/01/2014 task: ffff88023f0c3760 ti: ffff880268608000 task.ti: ffff880268608000 RIP: 0010:[<ffffffff813f502c>] [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 RSP: 0018:ffff88026860bd78 EFLAGS: 00010296 RAX: ffff88022d111410 RBX: 0000000000000008 RCX: 0000000100260021 RDX: 0000000000000026 RSI: ffff88022d111740 RDI: 0000000000000008 RBP: ffff88026860bd88 R08: ffff880249eac248 R09: 0000000000000000 R10: ffff880249eab1b8 R11: ffff880249eac248 R12: 0000000000000000 R13: ffff88023f0c49d0 R14: ffff88022d111740 R15: ffff880249eac248 FS: 0000000000000000(0000) GS:ffff880331cc0000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 00000002e0aa2000 CR4: 00000000000007e0 Call Trace: [<ffffffffa110dbc1>] ll_statahead_thread+0xdb1/0x2000 [lustre] [<ffffffff810df6e2>] ? pick_next_task_fair+0x702/0x8a0 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e329e>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffff810c834d>] ? finish_task_switch+0x5d/0x1b0 [<ffffffffa110ce10>] ? ll_statahead_handle.constprop.28+0x1c0/0x1c0 [lustre] [<ffffffff810ba114>] kthread+0xe4/0xf0 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 | Lustre: 27278:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff8802c8e10540 x1833852604476672/t4295172078(0) o101->1ac643b5-50aa-4fc6-833e-44e40151ac16@0@lo:736/0 lens 376/18608 e 0 to 0 dl 1748900856 ref 1 fl Interpret:H/602/0 rc 0/0 job:'dd.0' uid:0 gid:0 projid:0 Lustre: 27278:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-MDT0002: opcode 0: before 515 < left 2897, rollback = 0 Lustre: 27278:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 1611 previous similar messages Lustre: 27278:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 6/24/0, destroy: 1/4/0 Lustre: 27278:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 2210 previous similar messages Lustre: 27278:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 2897/2897/0, xattr_set: 4345/40664/0 Lustre: 27278:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 2210 previous similar messages Lustre: 27278:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 28/157/0, punch: 0/0/0, quota 1/3/0 Lustre: 27278:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 2210 previous similar messages Lustre: 27278:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 7/118/0, delete: 2/5/0 Lustre: 27278:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 2210 previous similar messages Lustre: 27278:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 1/1/0, ref_del: 2/2/0 Lustre: 27278:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 2210 previous similar messages Lustre: 31870:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff88027d2bee40 x1833852604921088/t4295252671(0) o101->1ac643b5-50aa-4fc6-833e-44e40151ac16@0@lo:737/0 lens 376/33392 e 0 to 0 dl 1748900857 ref 1 fl Interpret:H/602/0 rc 0/0 job:'dd.0' uid:0 gid:0 projid:0 Lustre: 27278:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff8802aa9e7840 x1833852605998080/t4295172393(0) o101->1ac643b5-50aa-4fc6-833e-44e40151ac16@0@lo:741/0 lens 376/19432 e 0 to 0 dl 1748900861 ref 1 fl Interpret:H/602/0 rc 0/0 job:'dd.0' uid:0 gid:0 projid:0 Lustre: 10891:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff8802aba0b240 x1833852607630336/t4295253003(0) o101->d96ae9f5-416f-4987-9648-bf5c9c55fb45@0@lo:744/0 lens 376/33392 e 0 to 0 dl 1748900864 ref 1 fl Interpret:H/602/0 rc 0/0 job:'dd.0' uid:0 gid:0 projid:0 Lustre: 10886:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff8802e14d1440 x1833852610529920/t4295172715(0) o101->d96ae9f5-416f-4987-9648-bf5c9c55fb45@0@lo:754/0 lens 376/45904 e 0 to 0 dl 1748900874 ref 1 fl Interpret:H/602/0 rc 0/0 job:'dd.0' uid:0 gid:0 projid:0 Lustre: 10886:0:(mdt_recovery.c:102:mdt_req_from_lrd()) Skipped 1 previous similar message LustreError: 5018:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000409:0x2a36:0x0]: rc = -5 LustreError: 5018:0:(lcommon_cl.c:188:cl_file_inode_init()) Skipped 62 previous similar messages LustreError: 5018:0:(llite_lib.c:3769:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 5018:0:(llite_lib.c:3769:ll_prep_inode()) Skipped 62 previous similar messages Lustre: 23495:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff88025f9dee40 x1833852612979072/t4295173015(0) o101->d96ae9f5-416f-4987-9648-bf5c9c55fb45@0@lo:9/0 lens 376/45904 e 0 to 0 dl 1748900884 ref 1 fl Interpret:H/602/0 rc 0/0 job:'dd.0' uid:0 gid:0 projid:0 Lustre: 23495:0:(mdt_recovery.c:102:mdt_req_from_lrd()) Skipped 1 previous similar message Lustre: 17272:0:(lod_lov.c:1403:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000406:0x2bb1:0x0] with magic=0xbd60bd0 Lustre: 17272:0:(lod_lov.c:1403:lod_parse_striping()) Skipped 137 previous similar messages LustreError: 394:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 10 [0x240000406:0x2bb7:0x0] inode@0000000000000000: rc = -5 LustreError: 394:0:(statahead.c:825:ll_statahead_interpret_work()) Skipped 6 previous similar messages Lustre: 20454:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff8802de479940 x1833852622441472/t4295173703(0) o101->d96ae9f5-416f-4987-9648-bf5c9c55fb45@0@lo:29/0 lens 376/45904 e 0 to 0 dl 1748900904 ref 1 fl Interpret:H/602/0 rc 0/0 job:'dd.0' uid:0 gid:0 projid:0 Lustre: 20454:0:(mdt_recovery.c:102:mdt_req_from_lrd()) Skipped 3 previous similar messages Lustre: 27291:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff8802d07c3740 x1833852643091072/t4295256115(0) o101->1ac643b5-50aa-4fc6-833e-44e40151ac16@0@lo:86/0 lens 376/43216 e 0 to 0 dl 1748900961 ref 1 fl Interpret:H/602/0 rc 0/0 job:'dd.0' uid:0 gid:0 projid:0 Lustre: 27291:0:(mdt_recovery.c:102:mdt_req_from_lrd()) Skipped 2 previous similar messages LustreError: 10675:0:(ldlm_lockd.c:257:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0002_UUID lock: ffff8802d1d9f0c0/0x235b843c035bb25 lrc: 3/0,0 mode: PW/PW res: [0x340000402:0xbb3:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->786431) gid 0 flags: 0x60000400000020 nid: 0@lo remote: 0x235b843c035b3f4 expref: 3025 pid: 4236 timeout: 3403 lvb_type: 0 LustreError: 4225:0:(ldlm_lockd.c:1447:ldlm_handle_enqueue()) ### lock on destroyed export ffff8800a775ca88 ns: filter-lustre-OST0002_UUID lock: ffff880293cee940/0x235b843c04d0c11 lrc: 3/0,0 mode: --/PW res: [0x340000402:0xca4:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x50000000020000 nid: 0@lo remote: 0x235b843c04d0c03 expref: 2994 pid: 4225 timeout: 0 lvb_type: 0 LustreError: 4225:0:(ldlm_lockd.c:1447:ldlm_handle_enqueue()) Skipped 8 previous similar messages LustreError: lustre-OST0002-osc-ffff8802e56392a8: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-OST0002-osc-ffff8802e56392a8: Connection to lustre-OST0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: lustre-OST0002-osc-ffff8802e56392a8: This client was evicted by lustre-OST0002; in progress operations using this service will fail. LustreError: 12679:0:(client.c:1373:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88024eaf8040 x1833852650850816/t0(0) o104->lustre-OST0002@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' uid:4294967295 gid:4294967295 projid:4294967295 Lustre: 3799:0:(llite_lib.c:4214:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.7@tcp:/lustre/fid: [0x280000406:0x1c3d:0x0]/ may get corrupted (rc -108) Lustre: 3797:0:(llite_lib.c:4214:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.7@tcp:/lustre/fid: [0x200000409:0x2c5d:0x0]// may get corrupted (rc -108) LustreError: 13104:0:(client.c:1373:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802850bbc40 x1833852651299328/t0(0) o104->lustre-OST0002@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' uid:4294967295 gid:4294967295 projid:4294967295 LustreError: 13104:0:(client.c:1373:ptlrpc_import_delay_req()) Skipped 1 previous similar message LustreError: 13337:0:(ldlm_resource.c:981:ldlm_resource_complain()) lustre-OST0002-osc-ffff8802e56392a8: namespace resource [0x340000402:0x160f:0x0].0x0 (ffff8802682efc40) refcount nonzero (1) after lock cleanup; forcing cleanup. Lustre: lustre-OST0002-osc-ffff8802e56392a8: Connection restored to (at 0@lo) Lustre: Skipped 1 previous similar message LustreError: 10672:0:(client.c:1373:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88007aeda840 x1833852655982208/t0(0) o105->lustre-OST0003@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' uid:4294967295 gid:4294967295 projid:4294967295 LustreError: lustre-OST0003-osc-ffff8802e45db7e8: This client was evicted by lustre-OST0003; in progress operations using this service will fail. LustreError: lustre-OST0000-osc-ffff8802e45db7e8: This client was evicted by lustre-OST0000; in progress operations using this service will fail. Lustre: 3794:0:(llite_lib.c:4214:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.7@tcp:/lustre/fid: [0x280000403:0x50b2:0x0]/ may get corrupted (rc -108) LustreError: 14289:0:(ldlm_resource.c:981:ldlm_resource_complain()) lustre-OST0000-osc-ffff8802e45db7e8: namespace resource [0x2c0000401:0x20ef:0x0].0x0 (ffff88025450b9c0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 9157:0:(client.c:1373:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802caaacb40 x1833852657175680/t0(0) o104->lustre-OST0000@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' uid:4294967295 gid:4294967295 projid:4294967295 LustreError: 9157:0:(client.c:1373:ptlrpc_import_delay_req()) Skipped 36 previous similar messages Lustre: 19203:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff880293f28040 x1833852676968064/t4295177425(0) o101->d96ae9f5-416f-4987-9648-bf5c9c55fb45@0@lo:213/0 lens 376/46696 e 0 to 0 dl 1748901088 ref 1 fl Interpret:H/602/0 rc 0/0 job:'dd.0' uid:0 gid:0 projid:0 Lustre: 19203:0:(mdt_recovery.c:102:mdt_req_from_lrd()) Skipped 4 previous similar messages ptlrpc_watchdog_fire: 2 callbacks suppressed Lustre: mdt_io00_009: service thread pid 19718 was inactive for 40.096 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 19718, comm: mdt_io00_009 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_pdo_lock+0x369/0x750 [mdt] [<0>] mdt_parent_lock+0x76/0x2a0 [mdt] [<0>] mdt_lock_two_dirs+0x30/0x190 [mdt] [<0>] mdt_reint_rename+0x1192/0x3950 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x73c/0xbb0 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x23d/0xd10 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe LustreError: 10675:0:(ldlm_lockd.c:257:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0003_UUID lock: ffff880241184780/0x235b843c05ae4aa lrc: 3/0,0 mode: PW/PW res: [0x380000402:0x23dd:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000400030020 nid: 0@lo remote: 0x235b843c05adce6 expref: 3977 pid: 9796 timeout: 3558 lvb_type: 0 LustreError: 10675:0:(ldlm_lockd.c:257:expired_lock_main()) Skipped 2 previous similar messages LustreError: 10016:0:(ldlm_lockd.c:1447:ldlm_handle_enqueue()) ### lock on destroyed export ffff8802d2392548 ns: filter-lustre-OST0003_UUID lock: ffff8802378661c0/0x235b843c071f4cf lrc: 3/0,0 mode: --/PW res: [0x380000401:0x3bf4:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x50000000020000 nid: 0@lo remote: 0x235b843c071f4c8 expref: 3730 pid: 10016 timeout: 0 lvb_type: 0 LustreError: 9230:0:(client.c:1373:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802aa9e2d40 x1833852718928768/t0(0) o104->lustre-OST0003@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' uid:4294967295 gid:4294967295 projid:4294967295 LustreError: 9230:0:(client.c:1373:ptlrpc_import_delay_req()) Skipped 16 previous similar messages LustreError: lustre-OST0003-osc-ffff8802e56392a8: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 196 previous similar messages Lustre: lustre-OST0003-osc-ffff8802e56392a8: Connection to lustre-OST0003 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 2 previous similar messages LustreError: lustre-OST0003-osc-ffff8802e56392a8: This client was evicted by lustre-OST0003; in progress operations using this service will fail. LustreError: 22959:0:(ldlm_resource.c:981:ldlm_resource_complain()) lustre-OST0003-osc-ffff8802e56392a8: namespace resource [0x380000402:0x33ce:0x0].0x0 (ffff880258d6efc0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 22959:0:(ldlm_resource.c:981:ldlm_resource_complain()) Skipped 1516 previous similar messages Lustre: lustre-OST0003-osc-ffff8802e56392a8: Connection restored to (at 0@lo) Lustre: Skipped 2 previous similar messages Lustre: 19696:0:(client.c:2451:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1748901077/real 1748901077] req@ffff88025574c140 x1833852695645056/t0(0) o101->lustre-MDT0001-mdc-ffff8802e56392a8@0@lo:12/10 lens 376/14824 e 0 to 1 dl 1748901134 ref 2 fl Rpc:XQr/602/ffffffff rc 0/-1 job:'dd.0' uid:0 gid:0 projid:0 Lustre: lustre-MDT0001: Client d96ae9f5-416f-4987-9648-bf5c9c55fb45 (at 0@lo) reconnecting Lustre: mdt_io00_009: service thread pid 19718 completed after 53.120s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 2994:0:(mdt_open.c:1302:mdt_cross_open()) lustre-MDT0001: [0x240000406:0x3027:0x0] doesn't exist!: rc = -14 Lustre: 20401:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff880257b71440 x1833852745298944/t4295185973(0) o101->d96ae9f5-416f-4987-9648-bf5c9c55fb45@0@lo:422/0 lens 376/48256 e 0 to 0 dl 1748901297 ref 1 fl Interpret:H/602/0 rc 0/0 job:'dd.0' uid:0 gid:0 projid:0 Lustre: 20401:0:(mdt_recovery.c:102:mdt_req_from_lrd()) Skipped 12 previous similar messages LustreError: 10675:0:(ldlm_lockd.c:257:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0000_UUID lock: ffff8802354e4f00/0x235b843c073c6aa lrc: 3/0,0 mode: PW/PW res: [0x2c0000402:0x3622:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000080020020 nid: 0@lo remote: 0x235b843c073c6a3 expref: 2888 pid: 14244 timeout: 3666 lvb_type: 0 LustreError: 9219:0:(ldlm_lockd.c:1447:ldlm_handle_enqueue()) ### lock on destroyed export ffff8802e45d8008 ns: filter-lustre-OST0000_UUID lock: ffff880252dcb880/0x235b843c0893488 lrc: 3/0,0 mode: --/PW res: [0x2c0000402:0x358a:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x50000080020000 nid: 0@lo remote: 0x235b843c0893290 expref: 2888 pid: 9219 timeout: 0 lvb_type: 0 LustreError: lustre-OST0000-osc-ffff8802e56392a8: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 2 previous similar messages LustreError: 17495:0:(client.c:1373:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802b324d040 x1833852765000960/t0(0) o105->lustre-OST0000@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' uid:4294967295 gid:4294967295 projid:4294967295 LustreError: 17495:0:(client.c:1373:ptlrpc_import_delay_req()) Skipped 14 previous similar messages LustreError: lustre-OST0000-osc-ffff8802e56392a8: This client was evicted by lustre-OST0000; in progress operations using this service will fail. Lustre: 3803:0:(llite_lib.c:4214:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.7@tcp:/lustre/fid: [0x240000406:0x31a7:0x0]/ may get corrupted (rc -108) Lustre: lustre-OST0000-osc-ffff8802e56392a8: Connection restored to (at 0@lo) Lustre: Skipped 1 previous similar message Lustre: lustre-OST0003-osc-ffff8802e45db7e8: Connection to lustre-OST0003 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 2 previous similar messages LustreError: 9977:0:(client.c:1373:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88026f51b240 x1833852774429568/t0(0) o104->lustre-OST0003@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' uid:4294967295 gid:4294967295 projid:4294967295 LustreError: 9977:0:(client.c:1373:ptlrpc_import_delay_req()) Skipped 2 previous similar messages LustreError: lustre-OST0003-osc-ffff8802e45db7e8: This client was evicted by lustre-OST0003; in progress operations using this service will fail. Lustre: 3806:0:(llite_lib.c:4214:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.7@tcp:/lustre/fid: [0x280000403:0x5395:0x0]/ may get corrupted (rc -108) Lustre: 3806:0:(llite_lib.c:4214:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.7@tcp:/lustre/fid: [0x280000403:0x538e:0x0]/ may get corrupted (rc -108) Lustre: 3808:0:(llite_lib.c:4214:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.7@tcp:/lustre/fid: [0x280000406:0x205e:0x0]/ may get corrupted (rc -108) LustreError: 31204:0:(ldlm_resource.c:981:ldlm_resource_complain()) lustre-OST0003-osc-ffff8802e45db7e8: namespace resource [0x380000402:0x375e:0x0].0x0 (ffff880258ae5440) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 31204:0:(ldlm_resource.c:981:ldlm_resource_complain()) Skipped 271 previous similar messages Lustre: mdt_out00_004: service thread pid 20833 was inactive for 40.056 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 20833, comm: mdt_out00_004 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_pdo_lock+0x369/0x750 [mdt] [<0>] mdt_parent_lock+0x76/0x2a0 [mdt] [<0>] mdt_getattr_name_lock+0x1844/0x2cc0 [mdt] [<0>] mdt_intent_getattr+0x2cc/0x4e0 [mdt] [<0>] mdt_intent_opc.constprop.74+0x211/0xc60 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x397/0x980 [ptlrpc] [<0>] ldlm_handle_enqueue+0x547/0x1900 [ptlrpc] [<0>] mdt_batch_getattr+0x6c/0x190 [mdt] [<0>] mdt_batch+0x465/0x195b [mdt] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x23d/0xd10 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe LustreError: 20833:0:(ldlm_lockd.c:1447:ldlm_handle_enqueue()) ### lock on destroyed export ffff8800a6d8b7e8 ns: mdt-lustre-MDT0000_UUID lock: ffff880259e3c3c0/0x235b843c094e3c6 lrc: 3/0,0 mode: PR/PR res: [0x200000407:0x5749:0x0].0x0 bits 0x13/0x0 rrc: 3 type: IBT gid 0 flags: 0x50200000000000 nid: 0@lo remote: 0x235b843c08629d6 expref: 52 pid: 20833 timeout: 0 lvb_type: 0 Lustre: mdt_out00_004: service thread pid 20833 completed after 70.335s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: lustre-MDT0000-mdc-ffff8802e56392a8: This client was evicted by lustre-MDT0000; in progress operations using this service will fail. LustreError: 32488:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e56392a8: inode [0x200000409:0x2d59:0x0] mdc close failed: rc = -108 LustreError: 32488:0:(file.c:247:ll_close_inode_openhandle()) Skipped 2 previous similar messages LustreError: 30335:0:(file.c:6187:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000007:0x1:0x0] error: rc = -108 LustreError: 30335:0:(file.c:6187:ll_inode_revalidate_fini()) Skipped 36 previous similar messages Lustre: lustre-OST0003-osc-MDT0001: update sequence from 0x380000401 to 0x380000403 Lustre: ll_ost00_019: service thread pid 11540 was inactive for 40.073 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 11540, comm: ll_ost00_019 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] ofd_destroy_by_fid+0x19c/0x610 [ofd] [<0>] ofd_destroy_hdl+0x20c/0xaf0 [ofd] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x23d/0xd10 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: ll_ost00_206: service thread pid 14217 was inactive for 40.078 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 1 previous similar message Lustre: ll_ost00_058: service thread pid 6293 was inactive for 39.990 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 5 previous similar messages LustreError: 21482:0:(client.c:1373:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff880320065540 x1833852798345984/t0(0) o104->lustre-OST0000@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' uid:4294967295 gid:4294967295 projid:4294967295 LustreError: 21482:0:(client.c:1373:ptlrpc_import_delay_req()) Skipped 4 previous similar messages LustreError: 8722:0:(ldlm_lockd.c:2550:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1748901347 with bad export cookie 159235962974058847 LustreError: lustre-OST0000-osc-ffff8802e56392a8: This client was evicted by lustre-OST0000; in progress operations using this service will fail. LustreError: 8722:0:(ldlm_lockd.c:2550:ldlm_cancel_handler()) Skipped 3 previous similar messages Lustre: 3804:0:(llite_lib.c:4214:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.7@tcp:/lustre/fid: [0x280000406:0x213b:0x0]/ may get corrupted (rc -108) Lustre: ll_ost00_019: service thread pid 11540 completed after 46.165s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: ll_ost00_052: service thread pid 5577 completed after 46.141s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: ll_ost00_058: service thread pid 6293 completed after 45.457s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: ll_ost00_109: service thread pid 9160 completed after 46.258s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: ll_ost00_206: service thread pid 14217 completed after 46.259s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: ll_ost00_191: service thread pid 13100 completed after 46.234s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: ll_ost00_056: service thread pid 5658 completed after 46.239s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: ll_ost00_001: service thread pid 13293 completed after 46.340s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 10675:0:(ldlm_lockd.c:257:expired_lock_main()) ### lock callback timer expired after 101s: evicting client at 0@lo ns: filter-lustre-OST0002_UUID lock: ffff88023e10f840/0x235b843c0914764 lrc: 3/0,0 mode: PW/PW res: [0x340000401:0x6bd5:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000480000020 nid: 0@lo remote: 0x235b843c091475d expref: 4374 pid: 14217 timeout: 3811 lvb_type: 0 LustreError: 10675:0:(ldlm_lockd.c:257:expired_lock_main()) Skipped 3 previous similar messages LustreError: lustre-OST0002-osc-ffff8802e45db7e8: operation ost_sync to node 0@lo failed: rc = -107 LustreError: Skipped 7 previous similar messages LustreError: 7946:0:(ldlm_lockd.c:2550:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1748901379 with bad export cookie 159235962968368904 LustreError: lustre-OST0002-osc-ffff8802e45db7e8: This client was evicted by lustre-OST0002; in progress operations using this service will fail. LustreError: 14231:0:(ldlm_lockd.c:1447:ldlm_handle_enqueue()) ### lock on destroyed export ffff8802ca8defc8 ns: filter-lustre-OST0002_UUID lock: ffff88024b1f9a80/0x235b843c0a0d227 lrc: 3/0,0 mode: --/PW res: [0x340000400:0x513f:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x50000000020000 nid: 0@lo remote: 0x235b843c0a0d220 expref: 3789 pid: 14231 timeout: 0 lvb_type: 0 LustreError: 4094:0:(ldlm_resource.c:981:ldlm_resource_complain()) lustre-OST0002-osc-ffff8802e45db7e8: namespace resource [0x340000400:0x4ea0:0x0].0x0 (ffff88008d3387c0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 4094:0:(ldlm_resource.c:981:ldlm_resource_complain()) Skipped 43 previous similar messages Lustre: lustre-OST0002: haven't heard from client 1ac643b5-50aa-4fc6-833e-44e40151ac16 (at 0@lo) in 31 seconds. I think it's dead, and I am evicting it. exp ffff880294ffefc8, cur 1748901410 deadline 1748901409 last 1748901379 Lustre: lustre-OST0001-osc-MDT0001: update sequence from 0x300000401 to 0x300000403 Lustre: lustre-OST0002-osc-MDT0001: update sequence from 0x340000401 to 0x340000403 Lustre: lustre-OST0002-osc-ffff8802e45db7e8: Connection restored to (at 0@lo) Lustre: Skipped 3 previous similar messages LustreError: lustre-OST0002-osc-ffff8802e45db7e8: This client was evicted by lustre-OST0002; in progress operations using this service will fail. LustreError: 6785:0:(ldlm_resource.c:981:ldlm_resource_complain()) lustre-OST0002-osc-ffff8802e45db7e8: namespace resource [0x340000400:0x5fd8:0x0].0x0 (ffff880078470040) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 6785:0:(ldlm_resource.c:981:ldlm_resource_complain()) Skipped 197 previous similar messages Lustre: lustre-OST0001-osc-MDT0002: update sequence from 0x300000400 to 0x300000404 Lustre: 18329:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-MDT0000: opcode 0: before 513 < left 1897, rollback = 0 Lustre: 18329:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 1473 previous similar messages Lustre: 18329:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 6/24/0, destroy: 1/4/0 Lustre: 18329:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 1473 previous similar messages Lustre: 18329:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1897/1897/0, xattr_set: 2845/26664/0 Lustre: 18329:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 1473 previous similar messages Lustre: 18329:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 28/157/0, punch: 0/0/0, quota 1/3/0 Lustre: 18329:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 1473 previous similar messages Lustre: 18329:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 7/118/0, delete: 2/5/1 Lustre: 18329:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 1473 previous similar messages Lustre: 18329:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 1/1/0, ref_del: 2/2/1 Lustre: 18329:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 1467 previous similar messages Lustre: 20320:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff8802c8e12840 x1833852831823488/t4295192286(0) o101->d96ae9f5-416f-4987-9648-bf5c9c55fb45@0@lo:682/0 lens 376/48256 e 0 to 0 dl 1748901557 ref 1 fl Interpret:H/602/0 rc 0/0 job:'dd.0' uid:0 gid:0 projid:0 Lustre: 20320:0:(mdt_recovery.c:102:mdt_req_from_lrd()) Skipped 14 previous similar messages Lustre: lustre-OST0000-osc-MDT0002: update sequence from 0x2c0000400 to 0x2c0000403 Lustre: lustre-OST0000-osc-MDT0001: update sequence from 0x2c0000401 to 0x2c0000404 Lustre: lustre-OST0003-osc-MDT0002: update sequence from 0x380000400 to 0x380000404 Lustre: lustre-OST0003-osc-ffff8802e56392a8: Connection to lustre-OST0003 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 4 previous similar messages LustreError: lustre-OST0003-osc-ffff8802e56392a8: This client was evicted by lustre-OST0003; in progress operations using this service will fail. LustreError: 11540:0:(client.c:1373:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802e9d98f40 x1833852865741696/t0(0) o104->lustre-OST0003@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' uid:4294967295 gid:4294967295 projid:4294967295 LustreError: 11540:0:(client.c:1373:ptlrpc_import_delay_req()) Skipped 39 previous similar messages LustreError: 9215:0:(ldlm_lockd.c:1447:ldlm_handle_enqueue()) ### lock on destroyed export ffff8802e40a6fc8 ns: filter-lustre-OST0003_UUID lock: ffff8800750d8f40/0x235b843c0c21488 lrc: 3/0,0 mode: PW/PW res: [0x380000401:0x7c44:0x0].0x0 rrc: 2 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x50000000020000 nid: 0@lo remote: 0x235b843c0c21449 expref: 5213 pid: 9215 timeout: 0 lvb_type: 0 Lustre: 3806:0:(llite_lib.c:4214:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.7@tcp:/lustre/fid: [0x240000408:0x1f10:0x0]/ may get corrupted (rc -108) Lustre: 3802:0:(llite_lib.c:4214:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.7@tcp:/lustre/fid: [0x20000040a:0x1e3:0x0]/ may get corrupted (rc -108) Lustre: 3804:0:(llite_lib.c:4214:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.7@tcp:/lustre/fid: [0x240000408:0x1efd:0x0]// may get corrupted (rc -108) Lustre: 3804:0:(llite_lib.c:4214:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.7@tcp:/lustre/fid: [0x240000408:0x1ef0:0x0]// may get corrupted (rc -108) LustreError: 14147:0:(ldlm_resource.c:981:ldlm_resource_complain()) lustre-OST0003-osc-ffff8802e56392a8: namespace resource [0x380000400:0x6b30:0x0].0x0 (ffff8802d1dd9bc0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 14147:0:(ldlm_resource.c:981:ldlm_resource_complain()) Skipped 1 previous similar message hrtimer: interrupt took 22644563 ns Lustre: lustre-OST0003: haven't heard from client d96ae9f5-416f-4987-9648-bf5c9c55fb45 (at 0@lo) in 31 seconds. I think it's dead, and I am evicting it. exp ffff88009c29e678, cur 1748901588 deadline 1748901587 last 1748901557 Lustre: lustre-OST0002-osc-MDT0002: update sequence from 0x340000400 to 0x340000404 LustreError: 14147:0:(ldlm_resource.c:981:ldlm_resource_complain()) lustre-OST0003-osc-ffff8802e56392a8: namespace resource [0x380000403:0x270d:0x0].0x0 (ffff8802a7d7c2c0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 14147:0:(ldlm_resource.c:981:ldlm_resource_complain()) Skipped 53 previous similar messages LustreError: lustre-OST0003-osc-ffff8802e56392a8: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 22 previous similar messages LustreError: lustre-OST0003-osc-ffff8802e56392a8: This client was evicted by lustre-OST0003; in progress operations using this service will fail. | Link to test |
racer test 2: racer rename: centos-10.localnet DURATION=2700 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 PGD 2db2a0067 PUD 2d08fd067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) llz4hc(OE) llz4(OE) lustre(OE) osp(OE) ofd(OE) lod(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) jbd2 mbcache lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) crc32_generic lnet(OE) dm_flakey dm_mod libcfs(OE) crc_t10dif crct10dif_generic crct10dif_common pcspkr virtio_balloon virtio_console i2c_piix4 ip_tables rpcsec_gss_krb5 ata_generic drm_kms_helper pata_acpi ttm drm drm_panel_orientation_quirks i2c_core virtio_blk ata_piix serio_raw floppy libata CPU: 2 PID: 2039 Comm: ll_sa_1915 Kdump: loaded Tainted: P OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.16.0-3.module_el8.7.0+1218+f626c2ff 04/01/2014 task: ffff8800705ea4f0 ti: ffff8800655e0000 task.ti: ffff8800655e0000 RIP: 0010:[<ffffffff813f502c>] [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 RSP: 0018:ffff8800655e3d78 EFLAGS: 00010296 RAX: ffff880218da1d90 RBX: 0000000000000008 RCX: dead000000000200 RDX: 0000000000000002 RSI: ffff880218da20c0 RDI: 0000000000000008 RBP: ffff8800655e3d88 R08: ffff880331403660 R09: 0000000000000000 R10: ffff880331403640 R11: ffff880254aa93e8 R12: 0000000000000000 R13: ffff88025bce0010 R14: ffff880218da20c0 R15: ffff880254aa93e8 FS: 0000000000000000(0000) GS:ffff880331a80000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 000000027d818000 CR4: 00000000000007e0 Call Trace: [<ffffffffa112abc1>] ll_statahead_thread+0xdb1/0x2000 [lustre] [<ffffffff810df6e2>] ? pick_next_task_fair+0x702/0x8a0 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e329e>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffff810c834d>] ? finish_task_switch+0x5d/0x1b0 [<ffffffffa1129e10>] ? ll_statahead_handle.constprop.28+0x1c0/0x1c0 [lustre] [<ffffffff810ba114>] kthread+0xe4/0xf0 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 | Lustre: 19422:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff88031fef5a40 x1833413651556992/t4295160221(0) o101->ebe878d2-f553-43ee-90ff-4d334c9ddb8a@0@lo:25/0 lens 376/16544 e 0 to 0 dl 1748481875 ref 1 fl Interpret:H/602/0 rc 0/0 job:'dd.0' uid:0 gid:0 projid:0 Lustre: 17697:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff88028163f840 x1833413651916032/t4295317188(0) o101->49c1553c-0fc9-4b6e-a0af-6e671e257aff@0@lo:26/0 lens 376/14192 e 0 to 0 dl 1748481876 ref 1 fl Interpret:H/602/0 rc 0/0 job:'dd.0' uid:0 gid:0 projid:0 LustreError: 23753:0:(mdt_open.c:1302:mdt_cross_open()) lustre-MDT0001: [0x240000404:0x4bea:0x0] doesn't exist!: rc = -14 Lustre: 28960:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff880290c76940 x1833413652345344/t4295161986(0) o101->ebe878d2-f553-43ee-90ff-4d334c9ddb8a@0@lo:27/0 lens 376/21640 e 0 to 0 dl 1748481877 ref 1 fl Interpret:H/602/0 rc 0/0 job:'dd.0' uid:0 gid:0 projid:0 Lustre: 28960:0:(mdt_recovery.c:102:mdt_req_from_lrd()) Skipped 1 previous similar message Lustre: 11000:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff88008bc74b40 x1833413652832256/t4295162104(0) o101->49c1553c-0fc9-4b6e-a0af-6e671e257aff@0@lo:30/0 lens 376/34184 e 0 to 0 dl 1748481880 ref 1 fl Interpret:H/602/0 rc 0/0 job:'dd.0' uid:0 gid:0 projid:0 Lustre: 11281:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff880290c75540 x1833413658893824/t4295161139(0) o101->ebe878d2-f553-43ee-90ff-4d334c9ddb8a@0@lo:53/0 lens 376/43360 e 0 to 0 dl 1748481903 ref 1 fl Interpret:H/602/0 rc 0/0 job:'dd.0' uid:0 gid:0 projid:0 Lustre: 10990:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff88008d6d0a40 x1833413662171776/t4295161277(0) o101->ebe878d2-f553-43ee-90ff-4d334c9ddb8a@0@lo:61/0 lens 376/43360 e 0 to 0 dl 1748481911 ref 1 fl Interpret:H/602/0 rc 0/0 job:'dd.0' uid:0 gid:0 projid:0 Lustre: 11281:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff88008bc75f40 x1833413681911936/t4295320744(0) o101->49c1553c-0fc9-4b6e-a0af-6e671e257aff@0@lo:100/0 lens 376/48184 e 0 to 0 dl 1748481950 ref 1 fl Interpret:H/602/0 rc 0/0 job:'dd.0' uid:0 gid:0 projid:0 Lustre: 11281:0:(mdt_recovery.c:102:mdt_req_from_lrd()) Skipped 2 previous similar messages LustreError: 29575:0:(statahead.c:2447:start_statahead_thread()) lustre: unsupported statahead pattern 0X0. Lustre: 11964:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff880261cd0540 x1833413705501056/t4295167535(0) o101->ebe878d2-f553-43ee-90ff-4d334c9ddb8a@0@lo:141/0 lens 376/44848 e 0 to 0 dl 1748481991 ref 1 fl Interpret:H/602/0 rc 0/0 job:'dd.0' uid:0 gid:0 projid:0 Lustre: 11964:0:(mdt_recovery.c:102:mdt_req_from_lrd()) Skipped 4 previous similar messages LustreError: lustre-MDT0001-mdc-ffff8800a72fb7e8: This client was evicted by lustre-MDT0001; in progress operations using this service will fail. LustreError: 20681:0:(client.c:1373:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802591e0a40 x1833413743217536/t0(0) o105->lustre-OST0000@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' uid:4294967295 gid:4294967295 projid:4294967295 LustreError: 20681:0:(client.c:1373:ptlrpc_import_delay_req()) Skipped 2 previous similar messages LustreError: 13207:0:(ldlm_lockd.c:1447:ldlm_handle_enqueue()) ### lock on destroyed export ffff8802e5042548 ns: filter-lustre-OST0003_UUID lock: ffff8802689eb100/0xf43bd31648b5876a lrc: 3/0,0 mode: --/PW res: [0x340000402:0x2e9e:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x50000000020000 nid: 0@lo remote: 0xf43bd31648b584ed expref: 4056 pid: 13207 timeout: 0 lvb_type: 0 LustreError: lustre-OST0003-osc-ffff8800a72fb7e8: This client was evicted by lustre-OST0003; in progress operations using this service will fail. LustreError: 25491:0:(client.c:1373:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff880234b69940 x1833413743301632/t0(0) o104->lustre-OST0003@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' uid:4294967295 gid:4294967295 projid:4294967295 LustreError: 25491:0:(client.c:1373:ptlrpc_import_delay_req()) Skipped 7 previous similar messages LustreError: lustre-OST0000-osc-ffff8800a72fb7e8: This client was evicted by lustre-OST0000; in progress operations using this service will fail. Lustre: 3961:0:(llite_lib.c:4214:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.12@tcp:/lustre/fid: [0x24000040a:0xc4:0x0]// may get corrupted (rc -108) Lustre: 3959:0:(llite_lib.c:4214:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.12@tcp:/lustre/fid: [0x24000040a:0xce:0x0]// may get corrupted (rc -108) Lustre: 3965:0:(llite_lib.c:4214:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.12@tcp:/lustre/fid: [0x240000404:0x4dd1:0x0]/ may get corrupted (rc -108) Lustre: 3963:0:(llite_lib.c:4214:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.12@tcp:/lustre/fid: [0x280000406:0x1a4e:0x0]// may get corrupted (rc -108) Lustre: 3963:0:(llite_lib.c:4214:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.12@tcp:/lustre/fid: [0x280000406:0x1a2a:0x0]/ may get corrupted (rc -108) LustreError: 3454:0:(ldlm_resource.c:1097:ldlm_resource_complain()) lustre-OST0003-osc-ffff8800a72fb7e8: namespace resource [0x340000402:0x348e:0x0].0x0 (ffff88006be82d40) refcount nonzero (0) after lock cleanup; forcing cleanup. LustreError: 3454:0:(ldlm_resource.c:1097:ldlm_resource_complain()) Skipped 1 previous similar message LustreError: 3472:0:(ldlm_resource.c:1097:ldlm_resource_complain()) lustre-OST0000-osc-ffff8800a72fb7e8: namespace resource [0x2c0000401:0x310f:0x0].0x0 (ffff880248fd3c40) refcount nonzero (0) after lock cleanup; forcing cleanup. LustreError: 3472:0:(ldlm_resource.c:1097:ldlm_resource_complain()) Skipped 467 previous similar messages Lustre: 18246:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff88006aa2df40 x1833413745646080/t4295324283(0) o101->49c1553c-0fc9-4b6e-a0af-6e671e257aff@0@lo:233/0 lens 376/48488 e 0 to 0 dl 1748482083 ref 1 fl Interpret:H/602/0 rc 0/0 job:'dd.0' uid:0 gid:0 projid:0 Lustre: 18246:0:(mdt_recovery.c:102:mdt_req_from_lrd()) Skipped 8 previous similar messages LustreError: 8221:0:(statahead.c:2447:start_statahead_thread()) lustre: unsupported statahead pattern 0X0. Lustre: 28933:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 6/24/0, destroy: 1/4/0 Lustre: 28933:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 2640 previous similar messages Lustre: 28933:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1977/1977/0, xattr_set: 2965/27784/0 Lustre: 28933:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 2640 previous similar messages Lustre: 28933:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 28/157/0, punch: 0/0/0, quota 1/3/0 Lustre: 28933:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 2640 previous similar messages Lustre: 28933:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 7/118/0, delete: 2/5/1 Lustre: 28933:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 2640 previous similar messages Lustre: 28933:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 1/1/0, ref_del: 2/2/1 Lustre: 28933:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 2640 previous similar messages Lustre: 11282:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff88023b941940 x1833413808336512/t4295170346(0) o101->ebe878d2-f553-43ee-90ff-4d334c9ddb8a@0@lo:388/0 lens 376/43360 e 0 to 0 dl 1748482238 ref 1 fl Interpret:H/602/0 rc 0/0 job:'dd.0' uid:0 gid:0 projid:0 Lustre: 11282:0:(mdt_recovery.c:102:mdt_req_from_lrd()) Skipped 6 previous similar messages Lustre: 16951:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-MDT0001: opcode 0: before 514 < left 3515, rollback = 0 Lustre: 16951:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 1442 previous similar messages LustreError: lustre-MDT0000-mdc-ffff8802e5748008: This client was evicted by lustre-MDT0000; in progress operations using this service will fail. LustreError: 12471:0:(llite_lib.c:2023:ll_md_setattr()) md_setattr fails: rc = -5 LustreError: 8201:0:(file.c:6187:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000403:0x2:0x0] error: rc = -108 LustreError: lustre-MDT0002-mdc-ffff8800a72fb7e8: This client was evicted by lustre-MDT0002; in progress operations using this service will fail. LustreError: 23753:0:(ldlm_lockd.c:1447:ldlm_handle_enqueue()) ### lock on destroyed export ffff88028f50ae98 ns: mdt-lustre-MDT0002_UUID lock: ffff88026870d680/0xf43bd31648c825ec lrc: 3/0,0 mode: PR/PR res: [0x200000409:0xfa8:0x0].0x0 bits 0x1/0x0 rrc: 2 type: IBT gid 0 flags: 0x50200000000000 nid: 0@lo remote: 0xf43bd31648c825d0 expref: 2 pid: 23753 timeout: 0 lvb_type: 0 Lustre: lustre-OST0002-osc-MDT0002: update sequence from 0x380000400 to 0x380000403 Lustre: ll_ost00_198: service thread pid 26857 was inactive for 72.058 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 26857, comm: ll_ost00_198 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] ofd_destroy_by_fid+0x19c/0x610 [ofd] [<0>] ofd_destroy_hdl+0x20c/0xaf0 [ofd] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x23d/0xd10 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 Lustre: ll_ost00_020: service thread pid 2100 was inactive for 72.060 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 7 previous similar messages [<0>] 0xfffffffffffffffe Pid: 29651, comm: ll_ost00_219 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] ofd_destroy_by_fid+0x19c/0x610 [ofd] [<0>] ofd_destroy_hdl+0x20c/0xaf0 [ofd] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x23d/0xd10 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 22413, comm: ll_ost00_070 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] ofd_destroy_by_fid+0x19c/0x610 [ofd] [<0>] ofd_destroy_hdl+0x20c/0xaf0 [ofd] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x23d/0xd10 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: ll_ost00_034: service thread pid 21175 was inactive for 72.112 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 6 previous similar messages Lustre: lustre-OST0003-osc-MDT0002: update sequence from 0x340000400 to 0x340000403 Lustre: 23683:0:(out_handler.c:879:out_tx_end()) lustre-MDT0001-osd: error during execution of #2 from /home/green/git/lustre-release/lustre/ptlrpc/../../lustre/target/out_handler.c:562: rc = -2 LustreError: 23683:0:(out_lib.c:1168:out_tx_index_delete_undo()) lustre-MDT0001-osd: Oops, can not rollback index_delete yet: rc = -524 LustreError: 23683:0:(out_lib.c:1168:out_tx_index_delete_undo()) Skipped 3 previous similar messages Lustre: ll_ost00_020: service thread pid 2100 completed after 89.269s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: ll_ost00_127: service thread pid 24013 completed after 88.441s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: ll_ost00_113: service thread pid 23940 completed after 88.438s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: ll_ost00_198: service thread pid 26857 completed after 89.298s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: ll_ost00_126: service thread pid 24007 completed after 88.455s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: ll_ost00_070: service thread pid 22413 completed after 89.423s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: ll_ost00_012: service thread pid 24209 completed after 88.492s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: ll_ost00_155: service thread pid 24694 completed after 87.812s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: ll_ost00_174: service thread pid 24764 completed after 88.492s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: ll_ost00_091: service thread pid 23116 completed after 88.945s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: ll_ost00_094: service thread pid 23133 completed after 87.834s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: ll_ost00_066: service thread pid 22352 completed after 87.799s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: ll_ost00_044: service thread pid 21271 completed after 87.806s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: ll_ost00_193: service thread pid 26473 completed after 87.802s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: ll_ost00_144: service thread pid 24676 completed after 87.842s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: ll_ost00_122: service thread pid 23985 completed after 87.831s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: ll_ost00_035: service thread pid 21176 completed after 87.812s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: ll_ost00_166: service thread pid 24739 completed after 87.810s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: ll_ost00_219: service thread pid 29651 completed after 89.331s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: ll_ost00_206: service thread pid 28478 completed after 89.338s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: ll_ost00_014: service thread pid 24876 completed after 89.334s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: ll_ost00_202: service thread pid 28468 completed after 87.843s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: ll_ost00_129: service thread pid 24043 completed after 87.859s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: ll_ost00_036: service thread pid 21183 completed after 89.400s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: ll_ost00_199: service thread pid 27101 completed after 88.569s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: ll_ost00_171: service thread pid 24761 completed after 87.876s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: ll_ost00_135: service thread pid 24655 completed after 89.456s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: ll_ost00_027: service thread pid 20746 completed after 89.571s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: ll_ost00_142: service thread pid 24666 completed after 87.962s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: ll_ost00_053: service thread pid 22091 completed after 87.960s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: ll_ost00_034: service thread pid 21175 completed after 88.776s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: ll_ost00_002: service thread pid 13209 completed after 88.650s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 10932:0:(mdt_open.c:1302:mdt_cross_open()) lustre-MDT0002: [0x280000407:0x178e:0x0] doesn't exist!: rc = -14 LustreError: 11281:0:(mdt_open.c:1302:mdt_cross_open()) lustre-MDT0002: [0x280000407:0x178e:0x0] doesn't exist!: rc = -14 Lustre: lustre-OST0000-osc-MDT0002: update sequence from 0x2c0000400 to 0x2c0000403 Lustre: lustre-OST0001-osc-MDT0002: update sequence from 0x300000400 to 0x300000403 LustreError: 11961:0:(mdt_open.c:1302:mdt_cross_open()) lustre-MDT0002: [0x280000407:0x178e:0x0] doesn't exist!: rc = -14 LustreError: 11961:0:(mdt_open.c:1302:mdt_cross_open()) Skipped 2 previous similar messages Lustre: lustre-OST0000-osc-MDT0001: update sequence from 0x2c0000401 to 0x2c0000404 LustreError: 26938:0:(mdt_open.c:1302:mdt_cross_open()) lustre-MDT0002: [0x280000407:0x178e:0x0] doesn't exist!: rc = -14 Lustre: lustre-OST0002-osc-MDT0001: update sequence from 0x380000401 to 0x380000404 Lustre: lustre-OST0003-osc-MDT0001: update sequence from 0x340000401 to 0x340000404 Lustre: lustre-OST0001-osc-MDT0001: update sequence from 0x300000401 to 0x300000404 Lustre: ll_ost_out00_00: service thread pid 13221 was inactive for 40.154 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 21 previous similar messages Lustre: lustre-OST0000-osc-MDT0000: update sequence from 0x2c0000402 to 0x2c0000405 Lustre: lustre-OST0002-osc-MDT0000: update sequence from 0x380000402 to 0x380000405 LustreError: 10885:0:(ldlm_lockd.c:257:expired_lock_main()) ### lock callback timer expired after 103s: evicting client at 0@lo ns: filter-lustre-OST0003_UUID lock: ffff8802687c6940/0xf43bd3164908ec83 lrc: 3/0,0 mode: PW/PW res: [0x340000402:0x506f:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000480020020 nid: 0@lo remote: 0xf43bd3164908ec75 expref: 5148 pid: 24209 timeout: 3458 lvb_type: 0 LustreError: 10885:0:(ldlm_lockd.c:257:expired_lock_main()) Skipped 6 previous similar messages LustreError: 17427:0:(ldlm_lockd.c:2550:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1748482371 with bad export cookie 17598892061542943214 LustreError: 24687:0:(client.c:1373:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88021fcaad40 x1833413935935360/t0(0) o104->lustre-OST0003@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' uid:4294967295 gid:4294967295 projid:4294967295 LustreError: 24687:0:(client.c:1373:ptlrpc_import_delay_req()) Skipped 9 previous similar messages Lustre: lustre-OST0003-osc-ffff8800a72fb7e8: Connection to lustre-OST0003 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: lustre-OST0003-osc-ffff8800a72fb7e8: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 471 previous similar messages Lustre: Skipped 6 previous similar messages LustreError: lustre-OST0003-osc-ffff8800a72fb7e8: This client was evicted by lustre-OST0003; in progress operations using this service will fail. Lustre: 3965:0:(llite_lib.c:4214:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.12@tcp:/lustre/fid: [0x200000409:0x1594:0x0]/ may get corrupted (rc -108) Lustre: 3965:0:(llite_lib.c:4214:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.12@tcp:/lustre/fid: [0x200000409:0x15b0:0x0]/ may get corrupted (rc -108) Lustre: 3965:0:(llite_lib.c:4214:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.12@tcp:/lustre/fid: [0x280000407:0x19ee:0x0]/ may get corrupted (rc -108) Lustre: 3952:0:(llite_lib.c:4214:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.12@tcp:/lustre/fid: [0x280000408:0x4d6:0x0]/ may get corrupted (rc -108) Lustre: 3959:0:(llite_lib.c:4214:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.12@tcp:/lustre/fid: [0x200000409:0x15c6:0x0]// may get corrupted (rc -108) Lustre: 3963:0:(llite_lib.c:4214:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.12@tcp:/lustre/fid: [0x280000408:0x459:0x0]/ may get corrupted (rc -108) Lustre: 3963:0:(llite_lib.c:4214:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.12@tcp:/lustre/fid: [0x20000040a:0x46a:0x0]/ may get corrupted (rc -108) Lustre: 3963:0:(llite_lib.c:4214:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.12@tcp:/lustre/fid: [0x280000408:0x40b:0x0]/ may get corrupted (rc -108) LustreError: 2100:0:(client.c:1373:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff880230f1e440 x1833413936458368/t0(0) o104->lustre-OST0003@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' uid:4294967295 gid:4294967295 projid:4294967295 LustreError: 2100:0:(client.c:1373:ptlrpc_import_delay_req()) Skipped 33 previous similar messages Lustre: lustre-OST0003-osc-ffff8800a72fb7e8: Connection restored to 192.168.123.12@tcp (at 0@lo) Lustre: Skipped 6 previous similar messages Lustre: 23753:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff88008c9e3240 x1833413937464832/t4295338584(0) o101->49c1553c-0fc9-4b6e-a0af-6e671e257aff@0@lo:658/0 lens 376/48488 e 0 to 0 dl 1748482508 ref 1 fl Interpret:H/602/0 rc 0/0 job:'dd.0' uid:0 gid:0 projid:0 Lustre: 23753:0:(mdt_recovery.c:102:mdt_req_from_lrd()) Skipped 24 previous similar messages Lustre: lustre-OST0001-osc-MDT0000: update sequence from 0x300000402 to 0x300000405 Lustre: lustre-OST0002-osc-MDT0002: update sequence from 0x380000403 to 0x380000406 LustreError: 11997:0:(client.c:1373:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8800650f6440 x1833413952442240/t0(0) o104->lustre-MDT0001@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' uid:4294967295 gid:4294967295 projid:4294967295 LustreError: 11997:0:(client.c:1373:ptlrpc_import_delay_req()) Skipped 6 previous similar messages Lustre: ll_ost_out00_00: service thread pid 13221 completed after 120.340s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: lustre-MDT0001-mdc-ffff8802e5748008: This client was evicted by lustre-MDT0001; in progress operations using this service will fail. LustreError: 6975:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e5748008: inode [0x240000404:0x55da:0x0] mdc close failed: rc = -5 LustreError: 6975:0:(file.c:247:ll_close_inode_openhandle()) Skipped 44 previous similar messages LustreError: 2643:0:(vvp_io.c:1909:vvp_io_init()) lustre: refresh file layout [0x24000040a:0x890:0x0] error -108. LustreError: 2643:0:(vvp_io.c:1909:vvp_io_init()) Skipped 2 previous similar messages LustreError: 10993:0:(mdt_open.c:1302:mdt_cross_open()) lustre-MDT0002: [0x280000407:0x178e:0x0] doesn't exist!: rc = -14 LustreError: 10993:0:(mdt_open.c:1302:mdt_cross_open()) Skipped 3 previous similar messages Lustre: lustre-OST0003-osc-MDT0000: update sequence from 0x340000402 to 0x340000405 LustreError: 17713:0:(mdt_open.c:1302:mdt_cross_open()) lustre-MDT0002: [0x280000407:0x178e:0x0] doesn't exist!: rc = -14 LustreError: 17713:0:(mdt_open.c:1302:mdt_cross_open()) Skipped 2 previous similar messages LustreError: 10965:0:(mdt_open.c:1302:mdt_cross_open()) lustre-MDT0002: [0x280000407:0x178e:0x0] doesn't exist!: rc = -14 LustreError: 10965:0:(mdt_open.c:1302:mdt_cross_open()) Skipped 6 previous similar messages LustreError: 24764:0:(client.c:1373:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802315d6940 x1833413991332352/t0(0) o104->lustre-OST0002@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' uid:4294967295 gid:4294967295 projid:4294967295 LustreError: lustre-OST0002-osc-ffff8800a72fb7e8: This client was evicted by lustre-OST0002; in progress operations using this service will fail. Lustre: 3962:0:(llite_lib.c:4214:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.12@tcp:/lustre/fid: [0x280000408:0x716:0x0]/ may get corrupted (rc -108) Lustre: 3962:0:(llite_lib.c:4214:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.12@tcp:/lustre/fid: [0x24000040b:0x1a1:0x0]/ may get corrupted (rc -108) Lustre: 3966:0:(llite_lib.c:4214:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.12@tcp:/lustre/fid: [0x280000408:0x77b:0x0]/ may get corrupted (rc -108) Lustre: 3966:0:(llite_lib.c:4214:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.12@tcp:/lustre/fid: [0x280000408:0x765:0x0]/ may get corrupted (rc -108) Lustre: 3951:0:(llite_lib.c:4214:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.12@tcp:/lustre/fid: [0x280000408:0x782:0x0]// may get corrupted (rc -108) Lustre: 3951:0:(llite_lib.c:4214:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.12@tcp:/lustre/fid: [0x24000040a:0xb00:0x0]/ may get corrupted (rc -108) Lustre: 3958:0:(llite_lib.c:4214:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.12@tcp:/lustre/fid: [0x280000408:0x6e9:0x0]/ may get corrupted (rc -108) Lustre: 3958:0:(llite_lib.c:4214:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.12@tcp:/lustre/fid: [0x24000040b:0xf2:0x0]/ may get corrupted (rc -108) Lustre: 3964:0:(llite_lib.c:4214:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.12@tcp:/lustre/fid: [0x280000407:0x1c45:0x0]/ may get corrupted (rc -108) Lustre: 3964:0:(llite_lib.c:4214:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.12@tcp:/lustre/fid: [0x24000040b:0x198:0x0]/ may get corrupted (rc -108) Lustre: 3960:0:(llite_lib.c:4214:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.12@tcp:/lustre/fid: [0x280000408:0x70f:0x0]// may get corrupted (rc -108) Lustre: 3960:0:(llite_lib.c:4214:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.12@tcp:/lustre/fid: [0x24000040b:0x10e:0x0]/ may get corrupted (rc -108) Lustre: 3963:0:(llite_lib.c:4214:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.12@tcp:/lustre/fid: [0x280000408:0x735:0x0]// may get corrupted (rc -108) LustreError: 19868:0:(ldlm_lockd.c:2550:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1748482509 with bad export cookie 17598892061532268963 LustreError: 14747:0:(ldlm_lockd.c:1447:ldlm_handle_enqueue()) ### lock on destroyed export ffff88028cdc1bf8 ns: filter-lustre-OST0000_UUID lock: ffff880272c9d2c0/0xf43bd316494e2eb8 lrc: 3/0,0 mode: PW/PW res: [0x2c0000404:0xceb:0x0].0x0 rrc: 2 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x50000080020000 nid: 0@lo remote: 0xf43bd316494e2eb1 expref: 3758 pid: 14747 timeout: 0 lvb_type: 0 LustreError: lustre-OST0000-osc-ffff8802e5748008: This client was evicted by lustre-OST0000; in progress operations using this service will fail. Lustre: 3960:0:(llite_lib.c:4214:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.12@tcp:/lustre/fid: [0x280000407:0x1b91:0x0]/ may get corrupted (rc -108) Lustre: 3958:0:(llite_lib.c:4214:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.12@tcp:/lustre/fid: [0x20000040a:0x60f:0x0]// may get corrupted (rc -108) Lustre: 3951:0:(llite_lib.c:4214:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.12@tcp:/lustre/fid: [0x24000040b:0x1c2:0x0]// may get corrupted (rc -108) Lustre: 3964:0:(llite_lib.c:4214:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.12@tcp:/lustre/fid: [0x24000040a:0xaef:0x0]/ may get corrupted (rc -108) Lustre: 3961:0:(llite_lib.c:4214:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.12@tcp:/lustre/fid: [0x24000040a:0xa9f:0x0]/ may get corrupted (rc -108) Lustre: 3961:0:(llite_lib.c:4214:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.12@tcp:/lustre/fid: [0x24000040a:0xa4e:0x0]/ may get corrupted (rc -108) Lustre: 3960:0:(llite_lib.c:4214:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.12@tcp:/lustre/fid: [0x20000040a:0x62b:0x0]/ may get corrupted (rc -108) Lustre: 3964:0:(llite_lib.c:4214:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.12@tcp:/lustre/fid: [0x24000040a:0xa93:0x0]/ may get corrupted (rc -108) Lustre: 3954:0:(llite_lib.c:4214:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.12@tcp:/lustre/fid: [0x280000407:0x1c5a:0x0]/ may get corrupted (rc -108) Lustre: 3961:0:(llite_lib.c:4214:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.12@tcp:/lustre/fid: [0x20000040a:0x631:0x0]/ may get corrupted (rc -108) LustreError: 16736:0:(ldlm_resource.c:1097:ldlm_resource_complain()) lustre-OST0002-osc-ffff8800a72fb7e8: namespace resource [0x380000405:0x411:0x0].0x0 (ffff8800957ba340) refcount nonzero (0) after lock cleanup; forcing cleanup. LustreError: 16736:0:(ldlm_resource.c:1097:ldlm_resource_complain()) Skipped 573 previous similar messages Lustre: lustre-OST0000: haven't heard from client ebe878d2-f553-43ee-90ff-4d334c9ddb8a (at 0@lo) in 31 seconds. I think it's dead, and I am evicting it. exp ffff8802375dca88, cur 1748482541 deadline 1748482540 last 1748482510 Lustre: lustre-OST0000-osc-ffff8802e5748008: disconnect after 49s idle LustreError: 18056:0:(mdt_open.c:1302:mdt_cross_open()) lustre-MDT0002: [0x280000407:0x178e:0x0] doesn't exist!: rc = -14 LustreError: 18056:0:(mdt_open.c:1302:mdt_cross_open()) Skipped 5 previous similar messages Lustre: 18248:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 6/24/0, destroy: 1/4/0 Lustre: 18248:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 1673 previous similar messages Lustre: 18248:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 3241/3241/0, xattr_set: 4861/45480/0 Lustre: 18248:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 1673 previous similar messages Lustre: 18248:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 28/157/0, punch: 0/0/0, quota 1/3/0 Lustre: 18248:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 1673 previous similar messages Lustre: 18248:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 7/118/0, delete: 2/5/1 Lustre: 18248:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 1673 previous similar messages Lustre: 18248:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 1/1/0, ref_del: 2/2/1 Lustre: 18248:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 1673 previous similar messages ptlrpc_watchdog_fire: 30 callbacks suppressed Lustre: mdt_out00_004: service thread pid 23038 was inactive for 40.158 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Lustre: Skipped 2 previous similar messages Pid: 23038, comm: mdt_out00_004 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_pdo_lock+0x369/0x750 [mdt] [<0>] mdt_parent_lock+0x76/0x2a0 [mdt] [<0>] mdt_getattr_name_lock+0x1844/0x2cc0 [mdt] [<0>] mdt_intent_getattr+0x2cc/0x4e0 [mdt] [<0>] mdt_intent_opc.constprop.74+0x211/0xc60 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x397/0x980 [ptlrpc] [<0>] ldlm_handle_enqueue+0x547/0x1900 [ptlrpc] [<0>] mdt_batch_getattr+0x6c/0x190 [mdt] [<0>] mdt_batch+0x465/0x195b [mdt] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x23d/0xd10 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: 11281:0:(osd_internal.h:1335:osd_trans_exec_op()) lustre-MDT0000: opcode 0: before 514 < left 3567, rollback = 0 Lustre: 11281:0:(osd_internal.h:1335:osd_trans_exec_op()) Skipped 1692 previous similar messages LustreError: 23038:0:(ldlm_lockd.c:1447:ldlm_handle_enqueue()) ### lock on destroyed export ffff880261270008 ns: mdt-lustre-MDT0001_UUID lock: ffff880236ae8f40/0xf43bd316497b429a lrc: 3/0,0 mode: PR/PR res: [0x24000040a:0xc0d:0x0].0x0 bits 0x13/0x0 rrc: 2 type: IBT gid 0 flags: 0x50200000000000 nid: 0@lo remote: 0xf43bd31649673d66 expref: 40 pid: 23038 timeout: 0 lvb_type: 0 Lustre: mdt_out00_004: service thread pid 23038 completed after 99.811s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: lustre-MDT0001-mdc-ffff8802e5748008: This client was evicted by lustre-MDT0001; in progress operations using this service will fail. LustreError: 30884:0:(mdt_open.c:1302:mdt_cross_open()) lustre-MDT0002: [0x280000407:0x178e:0x0] doesn't exist!: rc = -14 LustreError: 30884:0:(mdt_open.c:1302:mdt_cross_open()) Skipped 5 previous similar messages | Link to test |
racer test 1: racer on clients: centos-65.localnet DURATION=2700 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 PGD 2cdfd2067 PUD 28a9a6067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: llz4hc(OE) llz4(OE) lustre(OE) osp(OE) ofd(OE) lod(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) crc32_generic lnet(OE) libcfs(OE) crc_t10dif crct10dif_generic crct10dif_common pcspkr virtio_console virtio_balloon i2c_piix4 ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm ata_piix drm_panel_orientation_quirks virtio_blk serio_raw libata i2c_core floppy CPU: 7 PID: 8529 Comm: ll_sa_6481 Kdump: loaded Tainted: P OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.16.0-3.module_el8.7.0+1218+f626c2ff 04/01/2014 task: ffff8802c0c8b760 ti: ffff8802d9340000 task.ti: ffff8802d9340000 RIP: 0010:[<ffffffff813f502c>] [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 RSP: 0018:ffff8802d9343d78 EFLAGS: 00010296 RAX: ffff8801edfa0a90 RBX: 0000000000000008 RCX: 0000000100260025 RDX: 0000000000000026 RSI: ffff8801edfa0dc0 RDI: 0000000000000008 RBP: ffff8802d9343d88 R08: ffff88028b42dcc8 R09: 0000000000000000 R10: ffff88028b42e6b8 R11: ffff88028b42dcc8 R12: 0000000000000000 R13: ffff88025a88dc40 R14: ffff8801edfa0dc0 R15: ffff88028b42dcc8 FS: 0000000000000000(0000) GS:ffff880331bc0000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 0000000245318000 CR4: 00000000000007e0 Call Trace: [<ffffffffa15d0bc1>] ll_statahead_thread+0xdb1/0x2000 [lustre] [<ffffffff817e320e>] ? _raw_spin_lock+0x1e/0x20 [<ffffffff810dbb51>] ? put_prev_entity+0x31/0x400 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e329e>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffff810c834d>] ? finish_task_switch+0x5d/0x1b0 [<ffffffffa15cfe10>] ? ll_statahead_handle.constprop.28+0x1c0/0x1c0 [lustre] [<ffffffff810ba114>] kthread+0xe4/0xf0 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 | cp (28009) used greatest stack depth: 10064 bytes left LustreError: 27893:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802cb819bf8: inode [0x200000403:0x7:0x0] mdc close failed: rc = -13 Lustre: 29343:0:(mdd_dir.c:4822:mdd_migrate_object()) lustre-MDD0001: [0x240000403:0x1:0x0]/11 is open, migrate only dentry 5[30758]: segfault at 0 ip 00007fe4cff2a59a sp 00007fff60633530 error 4 in ld-2.17.so[7fe4cff28000+22000] Lustre: 15025:0:(mdd_dir.c:4822:mdd_migrate_object()) lustre-MDD0002: [0x280000403:0x1:0x0]/7 is open, migrate only dentry Lustre: 15025:0:(mdd_dir.c:4822:mdd_migrate_object()) Skipped 2 previous similar messages Lustre: 30495:0:(mdd_dir.c:4822:mdd_migrate_object()) lustre-MDD0001: [0x200000403:0x2:0x0]/12 is open, migrate only dentry LustreError: 30630:0:(mdt_reint.c:2540:mdt_reint_migrate()) lustre-MDT0000: migrate [0x200000403:0x1:0x0]/5 failed: rc = -2 LustreError: 30630:0:(mdt_reint.c:2540:mdt_reint_migrate()) lustre-MDT0000: migrate [0x280000403:0x3b:0x0]/15 failed: rc = -2 LustreError: 5879:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x280000404:0xa4:0x0]: rc = -5 LustreError: 5879:0:(llite_lib.c:3769:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 15025:0:(mdt_reint.c:2540:mdt_reint_migrate()) lustre-MDT0000: migrate [0x200000404:0x9b:0x0]/10 failed: rc = -2 LustreError: 6312:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x280000404:0xa4:0x0]: rc = -5 LustreError: 6312:0:(llite_lib.c:3769:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 Lustre: 4810:0:(lod_lov.c:1403:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0x20:0x0] with magic=0xbd60bd0 Lustre: 30630:0:(mdd_dir.c:4822:mdd_migrate_object()) lustre-MDD0001: [0x240000403:0x1:0x0]/1 is open, migrate only dentry LustreError: 15025:0:(mdt_reint.c:2540:mdt_reint_migrate()) lustre-MDT0001: migrate [0x240000404:0x28:0x0]/4 failed: rc = -2 Lustre: 7549:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff8802be786440 x1833361004283904/t4294970033(0) o101->7c54f921-d369-4f1e-8f05-a66b3b2a9418@0@lo:424/0 lens 376/816 e 0 to 0 dl 1748429424 ref 1 fl Interpret:H/602/0 rc 0/0 job:'dd.0' uid:0 gid:0 projid:0 LustreError: 30495:0:(mdt_reint.c:2540:mdt_reint_migrate()) lustre-MDT0001: migrate [0x200000404:0xf4:0x0]/4 failed: rc = -2 LustreError: 30495:0:(mdt_reint.c:2540:mdt_reint_migrate()) Skipped 1 previous similar message Lustre: 15027:0:(mdd_dir.c:4822:mdd_migrate_object()) lustre-MDD0000: [0x200000403:0x1:0x0]/2 is open, migrate only dentry Lustre: 15027:0:(mdd_dir.c:4822:mdd_migrate_object()) Skipped 1 previous similar message Lustre: 28077:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff880299aa4140 x1833361005493376/t4294970511(0) o101->413fca67-bccb-4767-b306-cead9605e937@0@lo:430/0 lens 376/840 e 0 to 0 dl 1748429430 ref 1 fl Interpret:H/602/0 rc 0/0 job:'dd.0' uid:0 gid:0 projid:0 LustreError: 10962:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x280000404:0xa4:0x0]: rc = -5 LustreError: 10962:0:(lcommon_cl.c:188:cl_file_inode_init()) Skipped 1 previous similar message LustreError: 10962:0:(llite_lib.c:3769:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 10962:0:(llite_lib.c:3769:ll_prep_inode()) Skipped 1 previous similar message Lustre: 4750:0:(lod_lov.c:1403:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0x20:0x0] with magic=0xbd60bd0 Lustre: 4750:0:(lod_lov.c:1403:lod_parse_striping()) Skipped 1 previous similar message LustreError: 29901:0:(mdt_reint.c:2540:mdt_reint_migrate()) lustre-MDT0000: migrate [0x200000404:0x174:0x0]/7 failed: rc = -2 LustreError: 29901:0:(mdt_reint.c:2540:mdt_reint_migrate()) Skipped 1 previous similar message LustreError: 14293:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x280000404:0xa4:0x0]: rc = -5 LustreError: 14293:0:(llite_lib.c:3769:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 Lustre: 7430:0:(mdd_dir.c:4822:mdd_migrate_object()) lustre-MDD0000: [0x280000403:0x1:0x0]/10 is open, migrate only dentry Lustre: 7430:0:(mdd_dir.c:4822:mdd_migrate_object()) Skipped 1 previous similar message LustreError: 30495:0:(mdd_dir.c:4743:mdd_migrate_cmd_check()) lustre-MDD0000: '17' migration was interrupted, run 'lfs migrate -m 2 -c 2 -H crush 17' to finish migration: rc = -1 LustreError: 15253:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000403:0x171:0x0]: rc = -5 LustreError: 15253:0:(llite_lib.c:3769:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 12389:0:(llite_lib.c:1872:ll_update_lsm_md()) lustre: [0x240000403:0x10c:0x0] dir layout mismatch: LustreError: 12389:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=3 count=3 index=1 hash=crush:0x2000003 max_inherit=0 max_inherit_rr=0 version=2 migrate_offset=0 migrate_hash=invalid:0 pool= LustreError: 12389:0:(lustre_lmv.h:167:lmv_stripe_object_dump()) stripe[0] [0x240000400:0xd:0x0] LustreError: 12389:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=1 count=4 index=1 hash=crush:0x82000003 max_inherit=0 max_inherit_rr=0 version=1 migrate_offset=3 migrate_hash=fnv_1a_64:2 pool= LustreError: 15027:0:(mdt_reint.c:2540:mdt_reint_migrate()) lustre-MDT0001: migrate [0x280000403:0x188:0x0]/11 failed: rc = -2 LustreError: 15027:0:(mdt_reint.c:2540:mdt_reint_migrate()) Skipped 3 previous similar messages LustreError: 13578:0:(mdd_dir.c:4743:mdd_migrate_cmd_check()) lustre-MDD0002: '0' migration was interrupted, run 'lfs migrate -m 1 -c 3 -H crush 0' to finish migration: rc = -1 Lustre: 4662:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff8802bbcfda40 x1833361010317184/t4294974386(0) o101->7c54f921-d369-4f1e-8f05-a66b3b2a9418@0@lo:463/0 lens 376/816 e 0 to 0 dl 1748429463 ref 1 fl Interpret:H/602/0 rc 0/0 job:'dd.0' uid:0 gid:0 projid:0 LustreError: 20486:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x280000404:0xa4:0x0]: rc = -5 LustreError: 20486:0:(lcommon_cl.c:188:cl_file_inode_init()) Skipped 1 previous similar message LustreError: 20486:0:(llite_lib.c:3769:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 20486:0:(llite_lib.c:3769:ll_prep_inode()) Skipped 1 previous similar message Lustre: dir [0x280000403:0x188:0x0] stripe 2 readdir failed: -2, directory is partially accessed! 2[19194]: segfault at 8 ip 00007fc66f6eb7e8 sp 00007fff9b35faa0 error 4 in ld-2.17.so[7fc66f6e0000+22000] Lustre: 29535:0:(mdd_dir.c:4822:mdd_migrate_object()) lustre-MDD0002: [0x280000403:0x1:0x0]/18 is open, migrate only dentry Lustre: 29535:0:(mdd_dir.c:4822:mdd_migrate_object()) Skipped 4 previous similar messages LustreError: 29123:0:(mdt_reint.c:876:mdt_add_dirty_flag()) file attribute read error for [0x240000403:0x1f8:0x0]: -2. LustreError: 29123:0:(osd_handler.c:211:osd_trans_start()) lustre-MDT0001: can't assign tx: rc = -2 LustreError: 22020:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802cb819bf8: inode [0x240000403:0x1f8:0x0] mdc close failed: rc = -2 LustreError: 19431:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802caabdd28: inode [0x200000404:0x1e2:0x0] mdc close failed: rc = -2 15[24357]: segfault at 0 ip 0000000000403e5f sp 00007ffcb9d829a0 error 6 in 19[400000+6000] LustreError: 28414:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000403:0x224:0x0]: rc = -5 LustreError: 28414:0:(lcommon_cl.c:188:cl_file_inode_init()) Skipped 3 previous similar messages LustreError: 28414:0:(llite_lib.c:3769:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 28414:0:(llite_lib.c:3769:ll_prep_inode()) Skipped 3 previous similar messages LustreError: 1042:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802caabdd28: inode [0x240000403:0x23b:0x0] mdc close failed: rc = -2 LustreError: 4727:0:(mdd_dir.c:4743:mdd_migrate_cmd_check()) lustre-MDD0002: '14' migration was interrupted, run 'lfs migrate -m 1 -c 2 -H crush 14' to finish migration: rc = -1 LustreError: 4727:0:(mdt_reint.c:2540:mdt_reint_migrate()) lustre-MDT0002: migrate [0x240000403:0x1:0x0]/14 failed: rc = -1 LustreError: 4727:0:(mdt_reint.c:2540:mdt_reint_migrate()) Skipped 4 previous similar messages Lustre: 4765:0:(lod_lov.c:1403:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000404:0x13e:0x0] with magic=0xbd60bd0 Lustre: 4765:0:(lod_lov.c:1403:lod_parse_striping()) Skipped 1 previous similar message Lustre: 30062:0:(mdd_dir.c:4822:mdd_migrate_object()) lustre-MDD0001: [0x240000403:0x1:0x0]/8 is open, migrate only dentry Lustre: 30062:0:(mdd_dir.c:4822:mdd_migrate_object()) Skipped 8 previous similar messages LustreError: 28424:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0001: failed to get lu_attr of [0x240000403:0xde:0x0]: rc = -2 LustreError: 11198:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802caabdd28: inode [0x240000403:0xde:0x0] mdc close failed: rc = -2 LustreError: 30074:0:(mdd_dir.c:4743:mdd_migrate_cmd_check()) lustre-MDD0002: '13' migration was interrupted, run 'lfs migrate -m 1 -c 3 -H crush 13' to finish migration: rc = -1 LustreError: 15197:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000403:0x15a:0x0]: rc = -5 LustreError: 15197:0:(lcommon_cl.c:188:cl_file_inode_init()) Skipped 19 previous similar messages LustreError: 15197:0:(llite_lib.c:3769:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 15197:0:(llite_lib.c:3769:ll_prep_inode()) Skipped 19 previous similar messages LustreError: 15014:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0001: failed to get lu_attr of [0x240000403:0xde:0x0]: rc = -2 LustreError: 28734:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0000: [0x200000404:0x489:0x0] migrate mdt count mismatch 3 != 2 LustreError: 10870:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802cb819bf8: inode [0x240000404:0x2b9:0x0] mdc close failed: rc = -2 LustreError: 10870:0:(file.c:247:ll_close_inode_openhandle()) Skipped 1 previous similar message LustreError: 30470:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0002: [0x280000404:0xa2:0x0] migrate mdt count mismatch 2 != 3 Lustre: lustre-MDT0000: trigger partial OI scrub for RPC inconsistency, checking FID [0x200000403:0x409:0x0]/0xa): rc = 0 LustreError: 20222:0:(osd_index.c:204:__osd_xattr_load_by_oid()) lustre-MDT0000: can't get bonus, rc = -2 LustreError: 215:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 5 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 LustreError: 16800:0:(lov_object.c:1350:lov_layout_change()) lustre-clilov-ffff8802cb819bf8: cannot apply new layout on [0x240000404:0x34d:0x0] : rc = -5 LustreError: 16800:0:(vvp_io.c:1909:vvp_io_init()) lustre: refresh file layout [0x240000404:0x34d:0x0] error -5. LustreError: 4810:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0000: [0x200000403:0x56d:0x0] migrate mdt count mismatch 2 != 3 LustreError: 4727:0:(mdd_dir.c:4743:mdd_migrate_cmd_check()) lustre-MDD0000: '17' migration was interrupted, run 'lfs migrate -m 2 -c 1 -H crush 17' to finish migration: rc = -1 LustreError: 4727:0:(mdd_dir.c:4743:mdd_migrate_cmd_check()) Skipped 1 previous similar message 1[29956]: segfault at 0 ip 0000000000403e5f sp 00007ffec27b3f00 error 6 in 1[400000+6000] LustreError: 29351:0:(lov_object.c:1350:lov_layout_change()) lustre-clilov-ffff8802cb819bf8: cannot apply new layout on [0x240000404:0x34d:0x0] : rc = -5 LustreError: 29351:0:(lov_object.c:1350:lov_layout_change()) Skipped 2 previous similar messages Lustre: dir [0x280000404:0x307:0x0] stripe 2 readdir failed: -2, directory is partially accessed! LustreError: 29535:0:(mdd_dir.c:4743:mdd_migrate_cmd_check()) lustre-MDD0002: '11' migration was interrupted, run 'lfs migrate -m 2 -c 1 -H crush 11' to finish migration: rc = -1 Lustre: dir [0x240000404:0x53f:0x0] stripe 2 readdir failed: -2, directory is partially accessed! LustreError: 383:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 10 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 Lustre: 29535:0:(lod_lov.c:1403:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000404:0x569:0x0] with magic=0xbd60bd0 Lustre: 29535:0:(lod_lov.c:1403:lod_parse_striping()) Skipped 3 previous similar messages Lustre: lustre-MDT0001: trigger partial OI scrub for RPC inconsistency, checking FID [0x240000404:0x54c:0x0]/0xa): rc = 0 LustreError: 7890:0:(osd_index.c:204:__osd_xattr_load_by_oid()) lustre-MDT0001: can't get bonus, rc = -2 LustreError: 26632:0:(mdt_reint.c:2540:mdt_reint_migrate()) lustre-MDT0002: migrate [0x280000403:0x3ef:0x0]/15 failed: rc = -116 LustreError: 26632:0:(mdt_reint.c:2540:mdt_reint_migrate()) Skipped 9 previous similar messages LustreError: 4770:0:(lustre_lmv.h:500:lmv_is_sane()) unknown layout LMV: magic=0xcd40cd0 count=3 index=2 hash=crush:0x82000003 version=1 migrate_offset=2 migrate_hash=fnv_1a_64:2 pool= Lustre: 29535:0:(lod_lov.c:1403:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000404:0x54d:0x0] with magic=0xbd60bd0 Lustre: 29535:0:(lod_lov.c:1403:lod_parse_striping()) Skipped 1 previous similar message 19[16475]: segfault at 8 ip 00007fe8775467e8 sp 00007ffca8ed4450 error 4 in ld-2.17.so[7fe87753b000+22000] LustreError: 27882:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0002: [0x280000404:0xa2:0x0] migrate mdt count mismatch 2 != 3 LustreError: 12995:0:(llite_lib.c:1872:ll_update_lsm_md()) lustre: [0x240000404:0x61a:0x0] dir layout mismatch: LustreError: 12995:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=2 count=2 index=1 hash=crush:0x2000003 max_inherit=0 max_inherit_rr=0 version=2 migrate_offset=0 migrate_hash=invalid:0 pool= LustreError: 12995:0:(lustre_lmv.h:167:lmv_stripe_object_dump()) stripe[0] [0x240000400:0x2a:0x0] LustreError: 12995:0:(lustre_lmv.h:167:lmv_stripe_object_dump()) Skipped 6 previous similar messages LustreError: 12995:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=1 count=5 index=1 hash=crush:0x82000003 max_inherit=0 max_inherit_rr=0 version=1 migrate_offset=2 migrate_hash=crush:2000003 pool= LustreError: 30495:0:(mdd_dir.c:4743:mdd_migrate_cmd_check()) lustre-MDD0002: '6' migration was interrupted, run 'lfs migrate -m 0 -c 2 -H crush 6' to finish migration: rc = -1 Lustre: 7430:0:(mdd_dir.c:4822:mdd_migrate_object()) lustre-MDD0002: [0x280000403:0x62d:0x0]/14 is open, migrate only dentry Lustre: 7430:0:(mdd_dir.c:4822:mdd_migrate_object()) Skipped 14 previous similar messages Lustre: 13044:0:(lod_lov.c:1403:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000404:0x469:0x0] with magic=0xbd60bd0 Lustre: 13044:0:(lod_lov.c:1403:lod_parse_striping()) Skipped 7 previous similar messages LustreError: 20863:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000404:0x6e7:0x0]: rc = -5 LustreError: 20863:0:(lcommon_cl.c:188:cl_file_inode_init()) Skipped 32 previous similar messages LustreError: 20863:0:(llite_lib.c:3769:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 20863:0:(llite_lib.c:3769:ll_prep_inode()) Skipped 32 previous similar messages Lustre: dir [0x200000404:0x643:0x0] stripe 3 readdir failed: -2, directory is partially accessed! LustreError: 863:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 18 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 LustreError: 863:0:(statahead.c:825:ll_statahead_interpret_work()) Skipped 1 previous similar message LustreError: 14634:0:(statahead.c:2447:start_statahead_thread()) lustre: unsupported statahead pattern 0X0. Lustre: dir [0x240000403:0x570:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 3 previous similar messages Lustre: 3309:0:(lod_lov.c:1403:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000404:0x833:0x0] with magic=0xbd60bd0 Lustre: 3309:0:(lod_lov.c:1403:lod_parse_striping()) Skipped 9 previous similar messages LustreError: 4625:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0000: [0x200000403:0x74e:0x0] migrate mdt count mismatch 2 != 1 LustreError: 31693:0:(lov_object.c:1350:lov_layout_change()) lustre-clilov-ffff8802caabdd28: cannot apply new layout on [0x240000403:0x45a:0x0] : rc = -5 LustreError: 28521:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802caabdd28: inode [0x200000404:0x688:0x0] mdc close failed: rc = -2 Lustre: dir [0x240000404:0x724:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 9 previous similar messages LustreError: 29544:0:(mdd_dir.c:4743:mdd_migrate_cmd_check()) lustre-MDD0002: '8' migration was interrupted, run 'lfs migrate -m 0 -c 1 -H crush 8' to finish migration: rc = -1 LustreError: 29544:0:(mdd_dir.c:4743:mdd_migrate_cmd_check()) Skipped 3 previous similar messages LustreError: 408:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 19 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 LustreError: 408:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 15 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 Lustre: 3309:0:(lod_lov.c:1403:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0x663:0x0] with magic=0xbd60bd0 Lustre: 3309:0:(lod_lov.c:1403:lod_parse_striping()) Skipped 47 previous similar messages LustreError: 38:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 13 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 LustreError: 17888:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802caabdd28: inode [0x240000404:0x8ef:0x0] mdc close failed: rc = -2 LustreError: 17888:0:(file.c:247:ll_close_inode_openhandle()) Skipped 1 previous similar message LustreError: 25784:0:(lov_object.c:1350:lov_layout_change()) lustre-clilov-ffff8802caabdd28: cannot apply new layout on [0x200000403:0xae8:0x0] : rc = -5 LustreError: 25784:0:(vvp_io.c:1909:vvp_io_init()) lustre: refresh file layout [0x200000403:0xae8:0x0] error -5. LustreError: 25784:0:(vvp_io.c:1909:vvp_io_init()) Skipped 2 previous similar messages LustreError: 7494:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0002: [0x280000403:0x878:0x0] migrate mdt count mismatch 3 != 1 LustreError: 2169:0:(lov_object.c:1350:lov_layout_change()) lustre-clilov-ffff8802cb819bf8: cannot apply new layout on [0x240000404:0xb27:0x0] : rc = -5 LustreError: 2169:0:(vvp_io.c:1909:vvp_io_init()) lustre: refresh file layout [0x240000404:0xb27:0x0] error -5. LustreError: 12873:0:(lov_object.c:1350:lov_layout_change()) lustre-clilov-ffff8802caabdd28: cannot apply new layout on [0x200000403:0xae8:0x0] : rc = -5 LustreError: 256:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 19 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 LustreError: 256:0:(statahead.c:825:ll_statahead_interpret_work()) Skipped 1 previous similar message LustreError: 28731:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0000: [0x200000403:0xd13:0x0] migrate mdt count mismatch 2 != 1 Lustre: dir [0x240000404:0x956:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 6 previous similar messages 8[22291]: segfault at 8 ip 00007f596de777e8 sp 00007ffd849d8690 error 4 in ld-2.17.so[7f596de6c000+22000] LustreError: 408:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x280000403:0x83d:0x0]: rc = -5 LustreError: 408:0:(lcommon_cl.c:188:cl_file_inode_init()) Skipped 86 previous similar messages LustreError: 408:0:(llite_lib.c:3769:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 408:0:(llite_lib.c:3769:ll_prep_inode()) Skipped 86 previous similar messages LustreError: 408:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 19 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 Lustre: mdt00_031: service thread pid 30275 was inactive for 40.106 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 30275, comm: mdt00_031 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock+0x88/0x1c0 [mdt] [<0>] mdt_reint_setattr+0x1324/0x15f0 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x73c/0xbb0 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x23d/0xd10 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 30218, comm: mdt00_029 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock_try+0xa0/0x250 [mdt] [<0>] mdt_object_open_lock+0x6b9/0xc10 [mdt] [<0>] mdt_open_by_fid_lock+0x4ec/0xc10 [mdt] [<0>] mdt_reint_open+0x9d8/0x2d70 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x73c/0xbb0 [mdt] [<0>] mdt_intent_open+0x93/0x480 [mdt] [<0>] mdt_intent_opc.constprop.74+0x211/0xc60 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x397/0x980 [ptlrpc] [<0>] ldlm_handle_enqueue+0x547/0x1900 [ptlrpc] [<0>] tgt_enqueue+0x68/0x240 [ptlrpc] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x23d/0xd10 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: mdt00_045: service thread pid 30519 was inactive for 40.098 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Lustre: Skipped 1 previous similar message Pid: 30519, comm: mdt00_045 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_fini+0xabc/0xfd0 [ptlrpc] [<0>] ldlm_cli_enqueue+0x461/0xb00 [ptlrpc] [<0>] osp_md_object_lock+0x151/0x2f0 [osp] [<0>] lod_object_lock+0xdb/0x7c0 [lod] [<0>] mdd_object_lock+0x2d/0xd0 [mdd] [<0>] mdt_remote_object_lock_try+0x14c/0x189 [mdt] [<0>] mdt_object_pdo_lock+0x729/0x750 [mdt] [<0>] mdt_parent_lock+0x76/0x2a0 [mdt] [<0>] mdt_reint_unlink+0x187/0x1610 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x73c/0xbb0 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x23d/0xd10 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: mdt_io00_021: service thread pid 13044 was inactive for 72.228 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: mdt_io00_013: service thread pid 13673 was inactive for 72.105 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 2 previous similar messages Lustre: mdt_io00_019: service thread pid 4727 was inactive for 72.114 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 6 previous similar messages Lustre: mdt_io00_003: service thread pid 29343 was inactive for 72.102 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 6 previous similar messages LustreError: 14836:0:(ldlm_lockd.c:257:expired_lock_main()) ### lock callback timer expired after 101s: evicting client at 0@lo ns: mdt-lustre-MDT0002_UUID lock: ffff880245568040/0x2d318c70316b9438 lrc: 3/0,0 mode: PW/PW res: [0x280000404:0xc62:0x0].0x0 bits 0x4/0x0 rrc: 7 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0x2d318c70316b942a expref: 447 pid: 7708 timeout: 517 lvb_type: 0 Lustre: mdt00_031: service thread pid 30275 completed after 100.401s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: 13044:0:(mdd_dir.c:4822:mdd_migrate_object()) lustre-MDD0002: [0x280000403:0x1:0x0]/4 is open, migrate only dentry Lustre: 13044:0:(mdd_dir.c:4822:mdd_migrate_object()) Skipped 35 previous similar messages Lustre: mdt00_029: service thread pid 30218 completed after 100.164s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: lustre-MDT0002-mdc-ffff8802cb819bf8: operation mds_close to node 0@lo failed: rc = -107 Lustre: lustre-MDT0002-mdc-ffff8802cb819bf8: Connection to lustre-MDT0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: lustre-MDT0002-mdc-ffff8802cb819bf8: This client was evicted by lustre-MDT0002; in progress operations using this service will fail. Lustre: mdt_io00_021: service thread pid 13044 completed after 100.407s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 19864:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802cb819bf8: inode [0x280000403:0x7b9:0x0] mdc close failed: rc = -108 LustreError: 23163:0:(llite_lib.c:2023:ll_md_setattr()) md_setattr fails: rc = -108 Lustre: lustre-MDT0002-mdc-ffff8802cb819bf8: Connection restored to 192.168.123.67@tcp (at 0@lo) Lustre: mdt_io00_017: service thread pid 30062 completed after 99.860s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_013: service thread pid 13673 completed after 99.895s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_006: service thread pid 30074 completed after 99.803s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_012: service thread pid 13578 completed after 99.765s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_010: service thread pid 30630 completed after 99.971s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_016: service thread pid 29978 completed after 99.795s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_020: service thread pid 12941 completed after 99.255s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_009: service thread pid 30594 completed after 99.233s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_000: service thread pid 15025 completed after 98.872s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_019: service thread pid 4727 completed after 98.780s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_011: service thread pid 7430 completed after 98.314s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_007: service thread pid 30495 completed after 97.979s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_014: service thread pid 26632 completed after 97.962s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 15027:0:(mdd_dir.c:4743:mdd_migrate_cmd_check()) lustre-MDD0000: '2' migration was interrupted, run 'lfs migrate -m 2 -c 3 -H crush 2' to finish migration: rc = -1 LustreError: 15027:0:(mdd_dir.c:4743:mdd_migrate_cmd_check()) Skipped 9 previous similar messages LustreError: 15027:0:(mdt_reint.c:2540:mdt_reint_migrate()) lustre-MDT0000: migrate [0x240000403:0x1:0x0]/2 failed: rc = -1 LustreError: 15027:0:(mdt_reint.c:2540:mdt_reint_migrate()) Skipped 32 previous similar messages Lustre: mdt_io00_002: service thread pid 15027 completed after 97.939s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_045: service thread pid 30519 completed after 98.313s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: 3309:0:(lod_lov.c:1403:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000403:0x7fa:0x0] with magic=0xbd60bd0 Lustre: 3309:0:(lod_lov.c:1403:lod_parse_striping()) Skipped 33 previous similar messages Lustre: mdt_io00_004: service thread pid 29535 completed after 98.616s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_008: service thread pid 30549 completed after 98.560s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_003: service thread pid 29343 completed after 97.766s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_015: service thread pid 29544 completed after 97.450s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_022: service thread pid 13111 completed after 96.360s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 7460:0:(mdt_open.c:1703:mdt_reint_open()) lustre-MDT0000: name '13' present, but FID [0x200000404:0xc9c:0x0] is invalid Lustre: dir [0x240000403:0xcb2:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 2 previous similar messages 19[17331]: segfault at 8 ip 00007fc2fa17c7e8 sp 00007fffb750bac0 error 4 in ld-2.17.so[7fc2fa171000+22000] LustreError: 4754:0:(mdt_handler.c:746:mdt_pack_acl2body()) lustre-MDT0001: unable to read [0x240000403:0xf59:0x0] ACL: rc = -2 8[32128]: segfault at 8 ip 00007fdf54be27e8 sp 00007fffb2e0a9d0 error 4 in ld-2.17.so[7fdf54bd7000+22000] LustreError: 30074:0:(lustre_lmv.h:500:lmv_is_sane()) unknown layout LMV: magic=0xcd40cd0 count=3 index=2 hash=crush:0x82000003 version=1 migrate_offset=2 migrate_hash=fnv_1a_64:2 pool= LustreError: 28174:0:(mdt_handler.c:746:mdt_pack_acl2body()) lustre-MDT0001: unable to read [0x240000403:0xd65:0x0] ACL: rc = -2 9[21268]: segfault at 0 ip (null) sp 00007ffdd72932c8 error 14 in 9[400000+6000] 5[6379]: segfault at 8 ip 00007f1f910c07e8 sp 00007ffc970f8bf0 error 4 in ld-2.17.so[7f1f910b5000+22000] 19[9347]: segfault at 8 ip 00007fbd576707e8 sp 00007ffc50088e90 error 4 in ld-2.17.so[7fbd57665000+22000] Lustre: dir [0x280000404:0x1152:0x0] stripe 2 readdir failed: -2, directory is partially accessed! Lustre: Skipped 3 previous similar messages LustreError: 7712:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0000: [0x200000404:0x1643:0x0] migrate mdt count mismatch 3 != 2 LustreError: 11157:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802cb819bf8: inode [0x200000403:0x14ee:0x0] mdc close failed: rc = -2 LustreError: 11157:0:(file.c:247:ll_close_inode_openhandle()) Skipped 20 previous similar messages LustreError: 127:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 18 [0x240000404:0x1478:0x0] inode@0000000000000000: rc = -5 LustreError: 127:0:(statahead.c:825:ll_statahead_interpret_work()) Skipped 1 previous similar message LustreError: 18678:0:(lov_object.c:1350:lov_layout_change()) lustre-clilov-ffff8802caabdd28: cannot apply new layout on [0x280000405:0x43d:0x0] : rc = -5 LustreError: 18678:0:(lov_object.c:1350:lov_layout_change()) Skipped 1 previous similar message LustreError: 18678:0:(vvp_io.c:1909:vvp_io_init()) lustre: refresh file layout [0x280000405:0x43d:0x0] error -5. 13[27033]: segfault at 8 ip 00007ffa8824e7e8 sp 00007ffcd1c27110 error 4 in ld-2.17.so[7ffa88243000+22000] Lustre: 7453:0:(lod_lov.c:1403:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000403:0x1220:0x0] with magic=0xbd60bd0 Lustre: 7453:0:(lod_lov.c:1403:lod_parse_striping()) Skipped 35 previous similar messages LustreError: 4727:0:(mdd_dir.c:4743:mdd_migrate_cmd_check()) lustre-MDD0001: '5' migration was interrupted, run 'lfs migrate -m 0 -c 3 -H crush 5' to finish migration: rc = -1 LustreError: 4727:0:(mdd_dir.c:4743:mdd_migrate_cmd_check()) Skipped 13 previous similar messages LustreError: 14836:0:(ldlm_lockd.c:257:expired_lock_main()) ### lock callback timer expired after 101s: evicting client at 0@lo ns: mdt-lustre-MDT0000_UUID lock: ffff88030edf2200/0x2d318c7031792aba lrc: 3/0,0 mode: PR/PR res: [0x200000404:0x133f:0x0].0x0 bits 0x1b/0x0 rrc: 5 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0x2d318c7031792a97 expref: 865 pid: 30214 timeout: 666 lvb_type: 0 LustreError: lustre-MDT0000-mdc-ffff8802caabdd28: operation mds_hsm_state_set to node 0@lo failed: rc = -107 Lustre: lustre-MDT0000-mdc-ffff8802caabdd28: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: lustre-MDT0000-mdc-ffff8802caabdd28: This client was evicted by lustre-MDT0000; in progress operations using this service will fail. LustreError: Skipped 1 previous similar message LustreError: 32561:0:(vvp_io.c:1909:vvp_io_init()) lustre: refresh file layout [0x200000404:0x133f:0x0] error -5. LustreError: 7069:0:(llite_lib.c:2023:ll_md_setattr()) md_setattr fails: rc = -108 LustreError: 6988:0:(file.c:6187:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000404:0x1862:0x0] error: rc = -108 LustreError: 7202:0:(llite_nfs.c:430:ll_dir_get_parent_fid()) lustre: failure inode [0x200000403:0x154b:0x0] get parent: rc = -108 Lustre: lustre-MDT0000-mdc-ffff8802caabdd28: Connection restored to 192.168.123.67@tcp (at 0@lo) LustreError: 30447:0:(ldlm_lockd.c:1447:ldlm_handle_enqueue()) ### lock on destroyed export ffff8802ca931bf8 ns: mdt-lustre-MDT0000_UUID lock: ffff880284be07c0/0x2d318c7031905844 lrc: 3/0,0 mode: PR/PR res: [0x200000404:0x1643:0x0].0x0 bits 0x12/0x0 rrc: 3 type: IBT gid 0 flags: 0x50200000000000 nid: 0@lo remote: 0x2d318c7031905836 expref: 9 pid: 30447 timeout: 0 lvb_type: 0 2[11004]: segfault at 8 ip 00007f8206ca17e8 sp 00007ffd1ac48cc0 error 4 in ld-2.17.so[7f8206c96000+22000] INFO: task mrename:858 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mrename D ffff88029e1249d0 11696 858 27459 0x00000080 Call Trace: [<ffffffff817e19d9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df7ea>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa1d>] mutex_lock+0x2d/0x40 [<ffffffff81252061>] lock_rename+0x31/0xe0 [<ffffffff812586af>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed462>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff812597de>] SyS_renameat2+0xe/0x10 [<ffffffff8125981e>] SyS_rename+0x1e/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a INFO: task mrename:2775 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mrename D ffff880099e9ca80 11696 2775 27528 0x00000080 Call Trace: [<ffffffff817e19d9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df7ea>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa1d>] mutex_lock+0x2d/0x40 [<ffffffff81252061>] lock_rename+0x31/0xe0 [<ffffffff812586af>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed462>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff812597de>] SyS_renameat2+0xe/0x10 [<ffffffff8125981e>] SyS_rename+0x1e/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a INFO: task mrename:4152 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mrename D ffff8802b104aae8 10736 4152 27298 0x00000080 Call Trace: [<ffffffff817e19d9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df7ea>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa1d>] mutex_lock+0x2d/0x40 [<ffffffff81252061>] lock_rename+0x31/0xe0 [<ffffffff812586af>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed462>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff812597de>] SyS_renameat2+0xe/0x10 [<ffffffff8125981e>] SyS_rename+0x1e/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a INFO: task mrename:4227 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mrename D ffff8802c158aae8 11696 4227 27166 0x00000080 Call Trace: [<ffffffff817e19d9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df7ea>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa1d>] mutex_lock+0x2d/0x40 [<ffffffff81252061>] lock_rename+0x31/0xe0 [<ffffffff812586af>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed462>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff812597de>] SyS_renameat2+0xe/0x10 [<ffffffff8125981e>] SyS_rename+0x1e/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a INFO: task mrename:5278 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mrename D ffff8802bcbf52c0 11696 5278 27423 0x00000080 Call Trace: [<ffffffff817e19d9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df7ea>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa1d>] mutex_lock+0x2d/0x40 [<ffffffff81252061>] lock_rename+0x31/0xe0 [<ffffffff812586af>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed462>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff812597de>] SyS_renameat2+0xe/0x10 [<ffffffff8125981e>] SyS_rename+0x1e/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a INFO: task mrename:8496 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mrename D ffff88009016bd58 11696 8496 27446 0x00000080 Call Trace: [<ffffffff817e19d9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df7ea>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa1d>] mutex_lock+0x2d/0x40 [<ffffffff81252061>] lock_rename+0x31/0xe0 [<ffffffff812586af>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed462>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff812597de>] SyS_renameat2+0xe/0x10 [<ffffffff8125981e>] SyS_rename+0x1e/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a INFO: task mrename:8580 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mrename D ffff8802be654a80 10928 8580 27337 0x00000080 Call Trace: [<ffffffff817e19d9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df7ea>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa1d>] mutex_lock+0x2d/0x40 [<ffffffff81252061>] lock_rename+0x31/0xe0 [<ffffffff812586af>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed462>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff812597de>] SyS_renameat2+0xe/0x10 [<ffffffff8125981e>] SyS_rename+0x1e/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a INFO: task mrename:9145 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mrename D ffff8802c25c9280 11696 9145 27393 0x00000080 Call Trace: [<ffffffff817e19d9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df7ea>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa1d>] mutex_lock+0x2d/0x40 [<ffffffff81252061>] lock_rename+0x31/0xe0 [<ffffffff812586af>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed462>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff812597de>] SyS_renameat2+0xe/0x10 [<ffffffff8125981e>] SyS_rename+0x1e/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a INFO: task mrename:10243 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mrename D ffff88009ab53d58 11808 10243 27280 0x00000080 Call Trace: [<ffffffff817e19d9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df7ea>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa1d>] mutex_lock+0x2d/0x40 [<ffffffff81252061>] lock_rename+0x31/0xe0 [<ffffffff812586af>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed462>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff812597de>] SyS_renameat2+0xe/0x10 [<ffffffff8125981e>] SyS_rename+0x1e/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a INFO: task mrename:11854 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mrename D ffff8802c7583a00 11696 11854 27579 0x00000080 Call Trace: [<ffffffff817e19d9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df7ea>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa1d>] mutex_lock+0x2d/0x40 [<ffffffff81252061>] lock_rename+0x31/0xe0 [<ffffffff812586af>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed462>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff812597de>] SyS_renameat2+0xe/0x10 [<ffffffff8125981e>] SyS_rename+0x1e/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a LustreError: 14836:0:(ldlm_lockd.c:257:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0002_UUID lock: ffff8802cda05e00/0x2d318c70318eca32 lrc: 3/0,0 mode: PR/PR res: [0x280000405:0x803:0x0].0x0 bits 0x1b/0x0 rrc: 3 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0x2d318c70318eca16 expref: 708 pid: 30339 timeout: 761 lvb_type: 0 LustreError: lustre-MDT0002-mdc-ffff8802caabdd28: operation mds_reint to node 0@lo failed: rc = -107 Lustre: lustre-MDT0002-mdc-ffff8802caabdd28: Connection to lustre-MDT0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 20425:0:(ldlm_lockd.c:2550:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1748430022 with bad export cookie 3256538419055133309 LustreError: lustre-MDT0002-mdc-ffff8802caabdd28: This client was evicted by lustre-MDT0002; in progress operations using this service will fail. LustreError: 14077:0:(mdc_request.c:1469:mdc_read_page()) lustre-MDT0002-mdc-ffff8802caabdd28: [0x280000401:0x2a:0x0] lock enqueue fails: rc = -108 Lustre: lustre-MDT0002-mdc-ffff8802caabdd28: Connection restored to 192.168.123.67@tcp (at 0@lo) LustreError: 532:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x280000404:0xd92:0x0]: rc = -5 LustreError: 532:0:(lcommon_cl.c:188:cl_file_inode_init()) Skipped 78 previous similar messages LustreError: 532:0:(llite_lib.c:3769:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 532:0:(llite_lib.c:3769:ll_prep_inode()) Skipped 78 previous similar messages LustreError: 30495:0:(lustre_lmv.h:500:lmv_is_sane()) unknown layout LMV: magic=0xcd40cd0 count=4 index=3 hash=crush:0x82000003 version=1 migrate_offset=3 migrate_hash=fnv_1a_64:2 pool= LustreError: 4625:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0001: [0x240000404:0x1690:0x0] migrate mdt count mismatch 1 != 3 8[24613]: segfault at 4017b0 ip 00000000004017b0 sp 00007ffca0edd110 error 14 LustreError: 30495:0:(mdt_reint.c:2540:mdt_reint_migrate()) lustre-MDT0000: migrate [0x200000403:0x2:0x0]/4 failed: rc = -1 LustreError: 30495:0:(mdt_reint.c:2540:mdt_reint_migrate()) Skipped 49 previous similar messages Lustre: 29901:0:(mdd_dir.c:4822:mdd_migrate_object()) lustre-MDD0000: [0x200000403:0x1:0x0]/6 is open, migrate only dentry Lustre: 29901:0:(mdd_dir.c:4822:mdd_migrate_object()) Skipped 55 previous similar messages LustreError: 45:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 14 [0x200000403:0x15a2:0x0] inode@0000000000000000: rc = -5 Lustre: lustre-MDT0002: trigger partial OI scrub for RPC inconsistency, checking FID [0x280000405:0x94b:0x0]/0xa): rc = 0 LustreError: 12689:0:(lov_object.c:1350:lov_layout_change()) lustre-clilov-ffff8802caabdd28: cannot apply new layout on [0x280000405:0x8a5:0x0] : rc = -5 LustreError: 12689:0:(vvp_io.c:1909:vvp_io_init()) lustre: refresh file layout [0x280000405:0x8a5:0x0] error -5. Lustre: dir [0x280000406:0x136:0x0] stripe 3 readdir failed: -2, directory is partially accessed! Lustre: Skipped 26 previous similar messages LustreError: 7682:0:(vvp_io.c:1909:vvp_io_init()) lustre: refresh file layout [0x240000404:0x1659:0x0] error -5. LustreError: 7682:0:(vvp_io.c:1909:vvp_io_init()) Skipped 2 previous similar messages LustreError: 4759:0:(mdt_handler.c:746:mdt_pack_acl2body()) lustre-MDT0001: unable to read [0x240000400:0x5b:0x0] ACL: rc = -2 3[22091]: segfault at 8 ip 00007fcafe3507e8 sp 00007ffeb99edb30 error 4 in ld-2.17.so[7fcafe345000+22000] LustreError: 28333:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0002: failed to get lu_attr of [0x280000405:0x11e:0x0]: rc = -2 LustreError: 15211:0:(lov_object.c:1350:lov_layout_change()) lustre-clilov-ffff8802caabdd28: cannot apply new layout on [0x200000403:0x15a2:0x0] : rc = -5 LustreError: 15211:0:(lov_object.c:1350:lov_layout_change()) Skipped 7 previous similar messages LustreError: 14836:0:(ldlm_lockd.c:257:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0002_UUID lock: ffff880099db96c0/0x2d318c7031acc725 lrc: 3/0,0 mode: PR/PR res: [0x280000406:0x374:0x0].0x0 bits 0x1b/0x0 rrc: 5 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0x2d318c7031acc6fb expref: 529 pid: 4707 timeout: 974 lvb_type: 0 LustreError: lustre-MDT0002-mdc-ffff8802cb819bf8: operation mds_getattr_lock to node 0@lo failed: rc = -107 Lustre: lustre-MDT0002-mdc-ffff8802cb819bf8: Connection to lustre-MDT0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: Skipped 4 previous similar messages LustreError: lustre-MDT0002-mdc-ffff8802cb819bf8: This client was evicted by lustre-MDT0002; in progress operations using this service will fail. LustreError: 24927:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802cb819bf8: inode [0x280000405:0xdda:0x0] mdc close failed: rc = -108 LustreError: 28208:0:(file.c:6187:ll_inode_revalidate_fini()) lustre: revalidate FID [0x240000403:0x19fa:0x0] error: rc = -5 LustreError: 28208:0:(file.c:6187:ll_inode_revalidate_fini()) Skipped 158 previous similar messages LustreError: 24927:0:(file.c:247:ll_close_inode_openhandle()) Skipped 74 previous similar messages LustreError: 25969:0:(mdc_request.c:1469:mdc_read_page()) lustre-MDT0002-mdc-ffff8802cb819bf8: [0x280000402:0x5c:0x0] lock enqueue fails: rc = -108 LustreError: 25969:0:(mdc_request.c:1469:mdc_read_page()) Skipped 18 previous similar messages Lustre: lustre-MDT0002-mdc-ffff8802cb819bf8: Connection restored to 192.168.123.67@tcp (at 0@lo) Lustre: 6473:0:(lod_lov.c:1403:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0x1d19:0x0] with magic=0xbd60bd0 Lustre: 6473:0:(lod_lov.c:1403:lod_parse_striping()) Skipped 33 previous similar messages LustreError: 13458:0:(mdd_dir.c:4743:mdd_migrate_cmd_check()) lustre-MDD0002: '5' migration was interrupted, run 'lfs migrate -m 1 -c 3 -H crush 5' to finish migration: rc = -1 LustreError: 13458:0:(mdd_dir.c:4743:mdd_migrate_cmd_check()) Skipped 21 previous similar messages 17[13281]: segfault at 8 ip 00007fc68a35a7e8 sp 00007ffd75ea85a0 error 4 in ld-2.17.so[7fc68a34f000+22000] 3[19986]: segfault at 0 ip (null) sp 00007ffc5ce24dc8 error 14 in 3[400000+6000] LustreError: 7460:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0000: [0x200000403:0x1a59:0x0] migrate mdt count mismatch 1 != 3 LustreError: 7460:0:(mdt_xattr.c:402:mdt_dir_layout_update()) Skipped 3 previous similar messages 8[26626]: segfault at 8 ip 00007f5082a6b7e8 sp 00007ffcc62d9eb0 error 4 in ld-2.17.so[7f5082a60000+22000] LustreError: 14836:0:(ldlm_lockd.c:257:expired_lock_main()) ### lock callback timer expired after 101s: evicting client at 0@lo ns: mdt-lustre-MDT0002_UUID lock: ffff8802e41ba200/0x2d318c7031b9541e lrc: 3/0,0 mode: PR/PR res: [0x280000406:0x953:0x0].0x0 bits 0x1b/0x0 rrc: 3 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0x2d318c7031b95402 expref: 139 pid: 28032 timeout: 1111 lvb_type: 0 LustreError: lustre-MDT0002-mdc-ffff8802cb819bf8: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-MDT0002-mdc-ffff8802cb819bf8: Connection to lustre-MDT0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: lustre-MDT0002-mdc-ffff8802cb819bf8: This client was evicted by lustre-MDT0002; in progress operations using this service will fail. LustreError: Skipped 1 previous similar message LustreError: 5119:0:(file.c:6187:ll_inode_revalidate_fini()) lustre: revalidate FID [0x280000406:0x953:0x0] error: rc = -5 LustreError: 7255:0:(llite_lib.c:2023:ll_md_setattr()) md_setattr fails: rc = -108 LustreError: 7255:0:(llite_lib.c:2023:ll_md_setattr()) Skipped 3 previous similar messages LustreError: 5119:0:(file.c:6187:ll_inode_revalidate_fini()) Skipped 8 previous similar messages Lustre: lustre-MDT0002-mdc-ffff8802cb819bf8: Connection restored to 192.168.123.67@tcp (at 0@lo) LustreError: 12941:0:(lustre_lmv.h:500:lmv_is_sane()) unknown layout LMV: magic=0xcd40cd0 count=2 index=1 hash=crush:0x82000003 version=1 migrate_offset=1 migrate_hash=fnv_1a_64:2 pool= 15[8635]: segfault at 4017b0 ip 00000000004017b0 sp 00007ffc88e6fac0 error 14 Lustre: dir [0x200000403:0x1ee8:0x0] stripe 3 readdir failed: -2, directory is partially accessed! Lustre: Skipped 11 previous similar messages LustreError: 15015:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0001: failed to get lu_attr of [0x240000403:0x197e:0x0]: rc = -2 LustreError: 15015:0:(mdd_object.c:3901:mdd_close()) Skipped 1 previous similar message LustreError: 7502:0:(mdt_open.c:1703:mdt_reint_open()) lustre-MDT0000: name '13' present, but FID [0x200000404:0xc9c:0x0] is invalid LustreError: 28201:0:(mdt_open.c:1703:mdt_reint_open()) lustre-MDT0000: name '13' present, but FID [0x200000404:0xc9c:0x0] is invalid LustreError: 7453:0:(mdt_open.c:1703:mdt_reint_open()) lustre-MDT0000: name '13' present, but FID [0x200000404:0xc9c:0x0] is invalid LustreError: 31:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 9 [0x280000408:0x1f:0x0] inode@0000000000000000: rc = -5 LustreError: 31:0:(statahead.c:825:ll_statahead_interpret_work()) Skipped 7 previous similar messages 15[6838]: segfault at 8 ip 00007ff4392c27e8 sp 00007ffcf4c5f0f0 error 4 in ld-2.17.so[7ff4392b7000+22000] LustreError: 31172:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0002: failed to get lu_attr of [0x280000406:0x7ff:0x0]: rc = -2 LustreError: 31172:0:(mdd_object.c:3901:mdd_close()) Skipped 1 previous similar message LustreError: 26425:0:(lov_object.c:1350:lov_layout_change()) lustre-clilov-ffff8802caabdd28: cannot apply new layout on [0x200000403:0x236b:0x0] : rc = -5 LustreError: 26425:0:(vvp_io.c:1909:vvp_io_init()) lustre: refresh file layout [0x200000403:0x236b:0x0] error -5. LustreError: 31157:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0002: failed to get lu_attr of [0x280000408:0xb1:0x0]: rc = -2 LustreError: 31157:0:(mdd_object.c:3901:mdd_close()) Skipped 1 previous similar message 15[1051]: segfault at 8 ip 00007f23c283f7e8 sp 00007ffd0af09730 error 4 in ld-2.17.so[7f23c2834000+22000] LustreError: 16153:0:(llite_lib.c:1872:ll_update_lsm_md()) lustre: [0x280000408:0x26b:0x0] dir layout mismatch: LustreError: 16153:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=2 count=3 index=2 hash=crush:0x2000003 max_inherit=0 max_inherit_rr=0 version=2 migrate_offset=0 migrate_hash=invalid:0 pool= LustreError: 16153:0:(lustre_lmv.h:167:lmv_stripe_object_dump()) stripe[0] [0x280000400:0x6d:0x0] LustreError: 16153:0:(lustre_lmv.h:167:lmv_stripe_object_dump()) Skipped 6 previous similar messages LustreError: 16153:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=1 count=6 index=2 hash=crush:0x82000003 max_inherit=0 max_inherit_rr=0 version=1 migrate_offset=3 migrate_hash=crush:3 pool= LustreError: 4822:0:(mdt_open.c:1703:mdt_reint_open()) lustre-MDT0000: name '13' present, but FID [0x200000404:0xc9c:0x0] is invalid LustreError: 4822:0:(mdt_open.c:1703:mdt_reint_open()) Skipped 1 previous similar message 11[5862]: segfault at 8 ip 00007f962c09c7e8 sp 00007fffe83c3090 error 4 in ld-2.17.so[7f962c091000+22000] LustreError: 14836:0:(ldlm_lockd.c:257:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0001_UUID lock: ffff88028a11a980/0x2d318c7031e41316 lrc: 3/0,0 mode: PR/PR res: [0x240000403:0x2484:0x0].0x0 bits 0x1b/0x0 rrc: 4 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0x2d318c7031e412ec expref: 523 pid: 7549 timeout: 1320 lvb_type: 0 Lustre: 30062:0:(mdd_dir.c:4822:mdd_migrate_object()) lustre-MDD0001: [0x240000403:0x246c:0x0]/17 is open, migrate only dentry Lustre: 30062:0:(mdd_dir.c:4822:mdd_migrate_object()) Skipped 112 previous similar messages LustreError: 471:0:(ldlm_lockd.c:2550:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1748430582 with bad export cookie 3256538419055133064 Lustre: lustre-MDT0001-mdc-ffff8802cb819bf8: Connection to lustre-MDT0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: lustre-MDT0001-mdc-ffff8802cb819bf8: operation mds_getattr_lock to node 0@lo failed: rc = -107 LustreError: lustre-MDT0001-mdc-ffff8802cb819bf8: This client was evicted by lustre-MDT0001; in progress operations using this service will fail. Lustre: lustre-MDT0001-mdc-ffff8802cb819bf8: Connection restored to 192.168.123.67@tcp (at 0@lo) LustreError: 13673:0:(mdt_reint.c:2540:mdt_reint_migrate()) lustre-MDT0000: migrate [0x240000403:0x2311:0x0]/15 failed: rc = -2 LustreError: 13673:0:(mdt_reint.c:2540:mdt_reint_migrate()) Skipped 74 previous similar messages 14[16110]: segfault at 0 ip 0000000000403e5f sp 00007fffa6b30c70 error 6 in 14[400000+6000] 8[18187]: segfault at 8 ip 00007fedcd4777e8 sp 00007ffca0eefb40 error 4 in ld-2.17.so[7fedcd46c000+22000] LustreError: 15217:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000405:0xaf:0x0]: rc = -5 LustreError: 15217:0:(lcommon_cl.c:188:cl_file_inode_init()) Skipped 508 previous similar messages LustreError: 15217:0:(llite_lib.c:3769:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 15217:0:(llite_lib.c:3769:ll_prep_inode()) Skipped 508 previous similar messages LustreError: 13416:0:(llite_lib.c:1872:ll_update_lsm_md()) lustre: [0x200000403:0x26b3:0x0] dir layout mismatch: LustreError: 13416:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=2 count=2 index=0 hash=crush:0x2000003 max_inherit=0 max_inherit_rr=0 version=2 migrate_offset=0 migrate_hash=invalid:0 pool= LustreError: 13416:0:(lustre_lmv.h:167:lmv_stripe_object_dump()) stripe[0] [0x200000400:0x9d:0x0] LustreError: 13416:0:(lustre_lmv.h:167:lmv_stripe_object_dump()) Skipped 8 previous similar messages LustreError: 13416:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=1 count=4 index=0 hash=crush:0x82000003 max_inherit=0 max_inherit_rr=0 version=1 migrate_offset=2 migrate_hash=crush:2000003 pool= LustreError: 2517:0:(vvp_io.c:1909:vvp_io_init()) lustre: refresh file layout [0x280000408:0x3d2:0x0] error -5. LustreError: 13419:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=2 count=2 index=0 hash=crush:0x2000003 max_inherit=0 max_inherit_rr=0 version=2 migrate_offset=0 migrate_hash=invalid:0 pool= LustreError: 13419:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=1 count=4 index=0 hash=crush:0x82000003 max_inherit=0 max_inherit_rr=0 version=1 migrate_offset=2 migrate_hash=crush:2000003 pool= LustreError: 13425:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=2 count=2 index=0 hash=crush:0x2000003 max_inherit=0 max_inherit_rr=0 version=2 migrate_offset=0 migrate_hash=invalid:0 pool= LustreError: 13425:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=1 count=4 index=0 hash=crush:0x82000003 max_inherit=0 max_inherit_rr=0 version=1 migrate_offset=2 migrate_hash=crush:2000003 pool= LustreError: 13422:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=2 count=2 index=0 hash=crush:0x2000003 max_inherit=0 max_inherit_rr=0 version=2 migrate_offset=0 migrate_hash=invalid:0 pool= LustreError: 13422:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=1 count=4 index=0 hash=crush:0x82000003 max_inherit=0 max_inherit_rr=0 version=1 migrate_offset=2 migrate_hash=crush:2000003 pool= LustreError: 13209:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=2 count=2 index=0 hash=crush:0x2000003 max_inherit=0 max_inherit_rr=0 version=2 migrate_offset=0 migrate_hash=invalid:0 pool= LustreError: 13209:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=1 count=4 index=0 hash=crush:0x82000003 max_inherit=0 max_inherit_rr=0 version=1 migrate_offset=2 migrate_hash=crush:2000003 pool= LustreError: 13206:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=3 count=2 index=0 hash=crush:0x2000003 max_inherit=0 max_inherit_rr=0 version=2 migrate_offset=0 migrate_hash=invalid:0 pool= LustreError: 13206:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=1 count=4 index=0 hash=crush:0x82000003 max_inherit=0 max_inherit_rr=0 version=1 migrate_offset=2 migrate_hash=crush:2000003 pool= LustreError: 13204:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=3 count=2 index=0 hash=crush:0x2000003 max_inherit=0 max_inherit_rr=0 version=2 migrate_offset=0 migrate_hash=invalid:0 pool= LustreError: 13204:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=1 count=4 index=0 hash=crush:0x82000003 max_inherit=0 max_inherit_rr=0 version=1 migrate_offset=2 migrate_hash=crush:2000003 pool= 16[17306]: segfault at 0 ip (null) sp 00007fffc4c50bf8 error 14 in 16[400000+6000] LustreError: 32019:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000403:0x24b2:0x0]: rc = -2 LustreError: 32019:0:(mdd_object.c:3901:mdd_close()) Skipped 3 previous similar messages LustreError: 28357:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0002: [0x280000406:0x118c:0x0] migrate mdt count mismatch 1 != 3 LustreError: 28357:0:(mdt_xattr.c:402:mdt_dir_layout_update()) Skipped 7 previous similar messages LustreError: 14836:0:(ldlm_lockd.c:257:expired_lock_main()) ### lock callback timer expired after 101s: evicting client at 0@lo ns: mdt-lustre-MDT0000_UUID lock: ffff880277ebda40/0x2d318c7031ec3144 lrc: 3/0,0 mode: PR/PR res: [0x200000403:0x2732:0x0].0x0 bits 0x1b/0x0 rrc: 6 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0x2d318c7031ec3128 expref: 684 pid: 3309 timeout: 1454 lvb_type: 0 LustreError: lustre-MDT0000-mdc-ffff8802cb819bf8: operation mds_close to node 0@lo failed: rc = -107 Lustre: lustre-MDT0000-mdc-ffff8802cb819bf8: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: lustre-MDT0000-mdc-ffff8802cb819bf8: This client was evicted by lustre-MDT0000; in progress operations using this service will fail. LustreError: 16411:0:(file.c:6187:ll_inode_revalidate_fini()) lustre: revalidate FID [0x280000408:0xae3:0x0] error: rc = -5 LustreError: 16411:0:(file.c:6187:ll_inode_revalidate_fini()) Skipped 1 previous similar message LustreError: 17180:0:(llite_lib.c:2023:ll_md_setattr()) md_setattr fails: rc = -5 LustreError: 8955:0:(mdc_request.c:1469:mdc_read_page()) lustre-MDT0000-mdc-ffff8802cb819bf8: [0x200000402:0x6c:0x0] lock enqueue fails: rc = -5 LustreError: 8955:0:(mdc_request.c:1469:mdc_read_page()) Skipped 3 previous similar messages LustreError: 18389:0:(ldlm_resource.c:1097:ldlm_resource_complain()) lustre-MDT0000-mdc-ffff8802cb819bf8: namespace resource [0x200000403:0x2b1e:0x0].0x0 (ffff880296b434c0) refcount nonzero (0) after lock cleanup; forcing cleanup. Lustre: lustre-MDT0000-mdc-ffff8802cb819bf8: Connection restored to 192.168.123.67@tcp (at 0@lo) LustreError: 6013:0:(ldlm_lockd.c:1447:ldlm_handle_enqueue()) ### lock on destroyed export ffff8802cb81d3d8 ns: mdt-lustre-MDT0000_UUID lock: ffff88030ef570c0/0x2d318c7031fd6627 lrc: 3/0,0 mode: PR/PR res: [0x200000402:0x6d:0x0].0x0 bits 0x12/0x0 rrc: 2 type: IBT gid 0 flags: 0x50200000000000 nid: 0@lo remote: 0x2d318c7031fd660b expref: 20 pid: 6013 timeout: 0 lvb_type: 0 Lustre: 12550:0:(mdt_reint.c:2460:mdt_reint_migrate()) lustre-MDT0002: [0x280000403:0x1:0x0]/12 is open, migrate only dentry LustreError: 7502:0:(mdt_open.c:1703:mdt_reint_open()) lustre-MDT0000: name '13' present, but FID [0x200000404:0xc9c:0x0] is invalid LustreError: 12941:0:(mdd_dir.c:4743:mdd_migrate_cmd_check()) lustre-MDD0000: '5' migration was interrupted, run 'lfs migrate -m 2 -c 1 -H crush 5' to finish migration: rc = -1 LustreError: 12941:0:(mdd_dir.c:4743:mdd_migrate_cmd_check()) Skipped 22 previous similar messages Lustre: 30495:0:(lod_lov.c:1403:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0x2464:0x0] with magic=0xbd60bd0 Lustre: 30495:0:(lod_lov.c:1403:lod_parse_striping()) Skipped 157 previous similar messages LustreError: 1091:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802caabdd28: inode [0x200000406:0xe0:0x0] mdc close failed: rc = -2 LustreError: 1091:0:(file.c:247:ll_close_inode_openhandle()) Skipped 125 previous similar messages 15[7919]: segfault at 8 ip 00007febb860c7e8 sp 00007ffd20a79720 error 4 in ld-2.17.so[7febb8601000+22000] 3[5170]: segfault at 0 ip (null) sp 00007ffdeb222fa8 error 14 in 3[400000+6000] 3[5130]: segfault at 0 ip (null) sp 00007ffe3556ac78 error 14 in 3[400000+6000] 13[2218]: segfault at 8 ip 00007f3ba246d7e8 sp 00007ffc48ed6640 error 4 in ld-2.17.so[7f3ba2462000+22000] LustreError: 31793:0:(statahead.c:2399:start_statahead_thread()) lustre: invalid pattern 0X0. LustreError: 14836:0:(ldlm_lockd.c:257:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0000_UUID lock: ffff8800adfcc780/0x2d318c7032088608 lrc: 3/0,0 mode: PR/PR res: [0x200000406:0x299:0x0].0x0 bits 0x1b/0x0 rrc: 5 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0x2d318c703208858a expref: 232 pid: 30218 timeout: 1618 lvb_type: 0 LustreError: lustre-MDT0000-mdc-ffff8802cb819bf8: operation mds_reint to node 0@lo failed: rc = -107 Lustre: lustre-MDT0000-mdc-ffff8802cb819bf8: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: lustre-MDT0000-mdc-ffff8802cb819bf8: This client was evicted by lustre-MDT0000; in progress operations using this service will fail. LustreError: 17328:0:(llite_lib.c:2023:ll_md_setattr()) md_setattr fails: rc = -108 Lustre: lustre-MDT0000-mdc-ffff8802cb819bf8: Connection restored to 192.168.123.67@tcp (at 0@lo) 8[23679]: segfault at 8 ip 00007f42819b77e8 sp 00007ffc9473fd70 error 4 in ld-2.17.so[7f42819ac000+22000] Lustre: dir [0x280000406:0x1577:0x0] stripe 3 readdir failed: -2, directory is partially accessed! Lustre: Skipped 74 previous similar messages LustreError: 26719:0:(lov_object.c:1350:lov_layout_change()) lustre-clilov-ffff8802cb819bf8: cannot apply new layout on [0x280000406:0x14ab:0x0] : rc = -5 LustreError: 26719:0:(lov_object.c:1350:lov_layout_change()) Skipped 6 previous similar messages LustreError: 26719:0:(vvp_io.c:1909:vvp_io_init()) lustre: refresh file layout [0x280000406:0x14ab:0x0] error -5. LustreError: 26719:0:(vvp_io.c:1909:vvp_io_init()) Skipped 2 previous similar messages LustreError: 406:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 20 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 LustreError: 406:0:(statahead.c:825:ll_statahead_interpret_work()) Skipped 3 previous similar messages 16[12717]: segfault at 0 ip 0000000000403e5f sp 00007fffccad8470 error 6 in 16[400000+6000] 8[11079]: segfault at 8 ip 00007f558c48c7e8 sp 00007ffe764c3d90 error 4 in ld-2.17.so[7f558c481000+22000] 14[21945]: segfault at 8 ip 00007f2872f167e8 sp 00007ffcc75b5240 error 4 in ld-2.17.so[7f2872f0b000+22000] 1[7151]: segfault at 8 ip 00007f03ae2967e8 sp 00007ffdc47bcf30 error 4 in ld-2.17.so[7f03ae28b000+22000] LustreError: 10198:0:(llite_lib.c:1872:ll_update_lsm_md()) lustre: [0x200000407:0xc38:0x0] dir layout mismatch: LustreError: 10198:0:(llite_lib.c:1872:ll_update_lsm_md()) Skipped 6 previous similar messages LustreError: 10198:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=3 count=3 index=0 hash=crush:0x2000003 max_inherit=0 max_inherit_rr=0 version=2 migrate_offset=0 migrate_hash=invalid:0 pool= LustreError: 10198:0:(lustre_lmv.h:167:lmv_stripe_object_dump()) stripe[0] [0x200000400:0xeb:0x0] LustreError: 10198:0:(lustre_lmv.h:167:lmv_stripe_object_dump()) Skipped 41 previous similar messages LustreError: 10198:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=1 count=4 index=0 hash=crush:0x82000003 max_inherit=0 max_inherit_rr=0 version=1 migrate_offset=3 migrate_hash=fnv_1a_64:2 pool= LustreError: 18378:0:(vvp_io.c:1909:vvp_io_init()) lustre: refresh file layout [0x200000407:0xc26:0x0] error -5. LustreError: 18378:0:(vvp_io.c:1909:vvp_io_init()) Skipped 5 previous similar messages LustreError: 14836:0:(ldlm_lockd.c:257:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0001_UUID lock: ffff880280b73880/0x2d318c70322b3a27 lrc: 3/0,0 mode: PR/PR res: [0x240000404:0x3c42:0x0].0x0 bits 0x12/0x0 rrc: 17 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0x2d318c70322b39fd expref: 698 pid: 7549 timeout: 1921 lvb_type: 0 LustreError: lustre-MDT0001-mdc-ffff8802cb819bf8: operation mds_close to node 0@lo failed: rc = -107 Lustre: lustre-MDT0001-mdc-ffff8802cb819bf8: Connection to lustre-MDT0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 6013:0:(ldlm_lockd.c:1447:ldlm_handle_enqueue()) ### lock on destroyed export ffff880243874a88 ns: mdt-lustre-MDT0001_UUID lock: ffff8802d320fc00/0x2d318c70323adfe0 lrc: 3/0,0 mode: PR/PR res: [0x240000405:0x1c87:0x0].0x0 bits 0x1b/0x0 rrc: 5 type: IBT gid 0 flags: 0x50200000000000 nid: 0@lo remote: 0x2d318c70323adfc4 expref: 490 pid: 6013 timeout: 0 lvb_type: 0 LustreError: 6013:0:(ldlm_lockd.c:1447:ldlm_handle_enqueue()) Skipped 3 previous similar messages LustreError: lustre-MDT0001-mdc-ffff8802cb819bf8: This client was evicted by lustre-MDT0001; in progress operations using this service will fail. LustreError: 16572:0:(llite_lib.c:2023:ll_md_setattr()) md_setattr fails: rc = -108 LustreError: 15122:0:(file.c:6187:ll_inode_revalidate_fini()) lustre: revalidate FID [0x240000404:0x3c42:0x0] error: rc = -108 LustreError: 15122:0:(file.c:6187:ll_inode_revalidate_fini()) Skipped 173 previous similar messages LustreError: 19977:0:(mdc_request.c:1469:mdc_read_page()) lustre-MDT0001-mdc-ffff8802cb819bf8: [0x240000400:0xbc:0x0] lock enqueue fails: rc = -108 LustreError: 19977:0:(mdc_request.c:1469:mdc_read_page()) Skipped 8 previous similar messages Lustre: lustre-MDT0001-mdc-ffff8802cb819bf8: Connection restored to 192.168.123.67@tcp (at 0@lo) 4[32316]: segfault at 4017b0 ip 00000000004017b0 sp 00007ffdddab08f0 error 14 LustreError: 28193:0:(ldlm_lockd.c:1447:ldlm_handle_enqueue()) ### lock on destroyed export ffff880243874a88 ns: mdt-lustre-MDT0001_UUID lock: ffff88008fd034c0/0x2d318c70323ade58 lrc: 3/0,0 mode: PR/PR res: [0x240000400:0xda:0x0].0x0 bits 0x12/0x0 rrc: 2 type: IBT gid 0 flags: 0x50200000000000 nid: 0@lo remote: 0x2d318c70323ade3c expref: 3 pid: 28193 timeout: 0 lvb_type: 0 Lustre: 12941:0:(mdd_dir.c:4822:mdd_migrate_object()) lustre-MDD0002: [0x280000403:0x1:0x0]/3 is open, migrate only dentry Lustre: 12941:0:(mdd_dir.c:4822:mdd_migrate_object()) Skipped 145 previous similar messages LustreError: 30549:0:(mdt_reint.c:2540:mdt_reint_migrate()) lustre-MDT0001: migrate [0x240000405:0x9cb:0x0]/17 failed: rc = -2 LustreError: 30549:0:(mdt_reint.c:2540:mdt_reint_migrate()) Skipped 112 previous similar messages LustreError: 7494:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0001: [0x240000405:0x9cb:0x0] migrate mdt count mismatch 1 != 2 LustreError: 7494:0:(mdt_xattr.c:402:mdt_dir_layout_update()) Skipped 7 previous similar messages | Link to test |
racer test 1: racer on clients: centos-35.localnet DURATION=2700 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 PGD 2a16d4067 PUD 2e633f067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) osp(OE) ofd(OE) lod(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) dm_flakey dm_mod libcfs(OE) loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) jbd2 mbcache crc32_generic crc_t10dif crct10dif_generic crct10dif_common virtio_console pcspkr virtio_balloon i2c_piix4 ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm drm_panel_orientation_quirks ata_piix i2c_core serio_raw floppy virtio_blk libata [last unloaded: libcfs] CPU: 10 PID: 27545 Comm: ll_sa_27303 Kdump: loaded Tainted: P OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.16.0-3.module_el8.7.0+1218+f626c2ff 04/01/2014 task: ffff8800a3bfc9d0 ti: ffff880324198000 task.ti: ffff880324198000 RIP: 0010:[<ffffffff813f502c>] [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 RSP: 0018:ffff88032419bd78 EFLAGS: 00010296 RAX: ffff8802bcb03a10 RBX: 0000000000000008 RCX: 0000000100260022 RDX: 0000000000000026 RSI: ffff8802bcb03d40 RDI: 0000000000000008 RBP: ffff88032419bd88 R08: ffff8802f2d7c8e8 R09: 0000000000000000 R10: ffff8802f2d7d978 R11: ffff8802f2d7c8e8 R12: 0000000000000000 R13: ffff8802f2a6dc40 R14: ffff8802bcb03d40 R15: ffff8802f2d7c8e8 FS: 0000000000000000(0000) GS:ffff880331c80000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 0000000326ce2000 CR4: 00000000000007e0 Call Trace: [<ffffffffa176cbb1>] ll_statahead_thread+0xdb1/0x2000 [lustre] [<ffffffff817e320e>] ? _raw_spin_lock+0x1e/0x20 [<ffffffff810dbb51>] ? put_prev_entity+0x31/0x400 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e329e>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffff810c834d>] ? finish_task_switch+0x5d/0x1b0 [<ffffffffa176be00>] ? ll_statahead_handle.constprop.28+0x1c0/0x1c0 [lustre] [<ffffffff810ba114>] kthread+0xe4/0xf0 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 | Lustre: 32095:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff8800a9dcf340 x1832974231584640/t4294967609(0) o101->8c6aa17c-7289-4045-8354-85e0c248f99d@0@lo:713/0 lens 376/816 e 0 to 0 dl 1748060518 ref 1 fl Interpret:H/602/0 rc 0/0 job:'dd.0' uid:0 gid:0 projid:0 7[3836]: segfault at 8 ip 00007fd2002fe7e8 sp 00007ffe86db6e10 error 4 in ld-2.17.so[7fd2002f3000+22000] Lustre: 32364:0:(osd_internal.h:1332:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 618, rollback = 7 Lustre: 32364:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 32364:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 32364:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 32364:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 32364:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 32366:0:(osd_internal.h:1332:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 618, rollback = 7 Lustre: 32366:0:(osd_internal.h:1332:osd_trans_exec_op()) Skipped 1 previous similar message Lustre: 32366:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 32366:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 32366:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 32366:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 32366:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 32366:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 32366:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 32366:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 32366:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 32366:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 32366:0:(osd_internal.h:1332:osd_trans_exec_op()) lustre-OST0002: opcode 7: before 516 < left 618, rollback = 7 Lustre: 32366:0:(osd_internal.h:1332:osd_trans_exec_op()) Skipped 5 previous similar messages Lustre: 32366:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 32366:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 32366:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 32366:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 32366:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 32366:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 32366:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 32366:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 32366:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 32366:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 5 previous similar messages 10[7386]: segfault at 8 ip 00007fbb553067e8 sp 00007ffeca8eab60 error 4 in ld-2.17.so[7fbb552fb000+22000] Lustre: 32366:0:(osd_internal.h:1332:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 618, rollback = 7 Lustre: 32366:0:(osd_internal.h:1332:osd_trans_exec_op()) Skipped 9 previous similar messages Lustre: 32366:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 32366:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 9 previous similar messages Lustre: 32366:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 32366:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 9 previous similar messages Lustre: 32366:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 32366:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 9 previous similar messages Lustre: 32366:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 32366:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 9 previous similar messages Lustre: 32366:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 32366:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 9 previous similar messages LustreError: 8439:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802f0f2efc8: inode [0x200000401:0x240:0x0] mdc close failed: rc = -13 12[11084]: segfault at 0 ip (null) sp 00007ffe11861598 error 14 in 12[400000+6000] Lustre: 1887:0:(osd_internal.h:1332:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 588, rollback = 7 Lustre: 1887:0:(osd_internal.h:1332:osd_trans_exec_op()) Skipped 5 previous similar messages Lustre: 1887:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 1887:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 1887:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 1887:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 1887:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/588/0, punch: 0/0/0, quota 1/3/0 Lustre: 1887:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 1887:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 1887:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 1887:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 1887:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 5 previous similar messages 1[13319]: segfault at 8 ip 00007f7b9e2737e8 sp 00007ffe1e9fcc90 error 4 in ld-2.17.so[7f7b9e268000+22000] Lustre: 3550:0:(osd_internal.h:1332:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 618, rollback = 7 Lustre: 3550:0:(osd_internal.h:1332:osd_trans_exec_op()) Skipped 1 previous similar message Lustre: 3550:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 3550:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 3550:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 3550:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 3550:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 3550:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 3550:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 3550:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 3550:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 3550:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 1 previous similar message 2[14415]: segfault at 406000 ip 0000000000406000 sp 00007ffff03a9598 error 14 in 2[606000+1000] 2[14635]: segfault at 8 ip 00007f73c13667e8 sp 00007ffe2df18b80 error 4 in ld-2.17.so[7f73c135b000+22000] 13[14627]: segfault at 0 ip 0000000000403e5f sp 00007ffde761b1b0 error 6 in 13[400000+6000] LustreError: 16651:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800bb91dd28: inode [0x200000402:0x49d:0x0] mdc close failed: rc = -13 17[17259]: segfault at 406000 ip 0000000000406000 sp 00007ffdf6f860f8 error 14 in 17[606000+1000] 9[18141]: segfault at 8 ip 00007fad99f957e8 sp 00007ffe0f10eb60 error 4 in ld-2.17.so[7fad99f8a000+22000] Lustre: 32366:0:(osd_internal.h:1332:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 618, rollback = 7 Lustre: 32366:0:(osd_internal.h:1332:osd_trans_exec_op()) Skipped 9 previous similar messages Lustre: 32366:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 32366:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 9 previous similar messages Lustre: 32366:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 32366:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 9 previous similar messages Lustre: 32366:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 32366:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 9 previous similar messages Lustre: 32366:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 32366:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 9 previous similar messages Lustre: 32366:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 32366:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 9 previous similar messages 2[23488]: segfault at 1c ip 00000000004017b0 sp 00007fffb47d9fb0 error 6 in 2[400000+6000] Lustre: 3550:0:(osd_internal.h:1332:osd_trans_exec_op()) lustre-OST0000: opcode 7: before 516 < left 618, rollback = 7 Lustre: 3550:0:(osd_internal.h:1332:osd_trans_exec_op()) Skipped 25 previous similar messages Lustre: 3550:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 3550:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 25 previous similar messages Lustre: 3550:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 3550:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 25 previous similar messages Lustre: 3550:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 3550:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 25 previous similar messages Lustre: 3550:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 3550:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 25 previous similar messages Lustre: 3550:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 3550:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 25 previous similar messages 5[24948]: segfault at 0 ip (null) sp 00007fff68a109a8 error 14 in 5[400000+6000] 6[26264]: segfault at 8 ip 00007f9d089637e8 sp 00007ffd3e1b5050 error 4 in ld-2.17.so[7f9d08958000+22000] LustreError: 28889:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802f0f2efc8: inode [0x200000402:0x8c8:0x0] mdc close failed: rc = -13 5[32568]: segfault at 0 ip 00000000004043f0 sp 00007ffcf20f0278 error 6 in 16[400000+6000] 10[2127]: segfault at 8 ip 00007feab74887e8 sp 00007ffdc42d12f0 error 4 in ld-2.17.so[7feab747d000+22000] Lustre: 18334:0:(osd_internal.h:1332:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 516 < left 618, rollback = 7 Lustre: 18334:0:(osd_internal.h:1332:osd_trans_exec_op()) Skipped 31 previous similar messages Lustre: 18334:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 18334:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 31 previous similar messages Lustre: 18334:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 18334:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 31 previous similar messages Lustre: 18334:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 18334:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 31 previous similar messages Lustre: 18334:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 18334:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 31 previous similar messages Lustre: 18334:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 18334:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 31 previous similar messages LustreError: 5544:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800bb91dd28: inode [0x200000402:0xbb8:0x0] mdc close failed: rc = -13 17[6018]: segfault at 8 ip 00007f573c77a7e8 sp 00007ffd130e6970 error 4 in ld-2.17.so[7f573c76f000+22000] 11[10298]: segfault at 8 ip 00007f160570f7e8 sp 00007ffcc48ba690 error 4 in ld-2.17.so[7f1605704000+22000] 1[13443]: segfault at 8 ip 00007f21cf2767e8 sp 00007ffd436b23f0 error 4 in ld-2.17.so[7f21cf26b000+22000] 15[14423]: segfault at 8 ip 00007f0e1c2bc7e8 sp 00007ffc9f7934c0 error 4 in ld-2.17.so[7f0e1c2b1000+22000] 15[16398]: segfault at 8 ip 00007f6a0ac707e8 sp 00007ffde5401680 error 4 in ld-2.17.so[7f6a0ac65000+22000] traps: 18[20200] general protection ip:7f74d42a3bbb sp:7ffe4d3bda00 error:0 in ld-2.17.so[7f74d429c000+22000] 13[27786]: segfault at 8 ip 00007ff4536e17e8 sp 00007ffd74d17320 error 4 in ld-2.17.so[7ff4536d6000+22000] Lustre: 3972:0:(lod_lov.c:1414:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000401:0x1358:0x0] with magic=0xbd60bd0 Lustre: 1887:0:(osd_internal.h:1332:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 516 < left 618, rollback = 7 Lustre: 1887:0:(osd_internal.h:1332:osd_trans_exec_op()) Skipped 71 previous similar messages Lustre: 1887:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 1887:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 71 previous similar messages Lustre: 1887:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 1887:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 71 previous similar messages Lustre: 1887:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 1887:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 71 previous similar messages Lustre: 1887:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 1887:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 71 previous similar messages Lustre: 1887:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 1887:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 71 previous similar messages LustreError: 2357:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800bb91dd28: inode [0x200000402:0x1599:0x0] mdc close failed: rc = -13 1[6251]: segfault at 8 ip 00007f2124f9e7e8 sp 00007ffdc08dd720 error 4 in ld-2.17.so[7f2124f93000+22000] 4[8564]: segfault at 0 ip 00000000004016cd sp 00007ffce7aff870 error 6 in 4[400000+6000] 19[9524]: segfault at 8 ip 00007f85b7bf07e8 sp 00007ffecc212c60 error 4 in ld-2.17.so[7f85b7be5000+22000] 13[10259]: segfault at 8 ip 00007f2da41d67e8 sp 00007ffe799240d0 error 4 in ld-2.17.so[7f2da41cb000+22000] 13[10522]: segfault at 0 ip (null) sp 00007fffa3d46de8 error 14 in 13[400000+6000] 0[16728]: segfault at 8 ip 00007f78290467e8 sp 00007ffcbe2fd090 error 4 in ld-2.17.so[7f782903b000+22000] LustreError: 16728:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802f0f2efc8: inode [0x200000402:0x1a72:0x0] mdc close failed: rc = -13 Lustre: 14701:0:(lod_lov.c:1414:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000401:0x1c7a:0x0] with magic=0xbd60bd0 Lustre: 14701:0:(lod_lov.c:1414:lod_parse_striping()) Skipped 1 previous similar message 19[24058]: segfault at 8 ip 00007efdad4fa7e8 sp 00007ffca5bcd840 error 4 in ld-2.17.so[7efdad4ef000+22000] 10[27416]: segfault at 0 ip 0000000000403e5f sp 00007ffd2c343e30 error 6 in 10[400000+6000] 4[28046]: segfault at 8 ip 00007fb1cd90b7e8 sp 00007ffdd2735160 error 4 in ld-2.17.so[7fb1cd900000+22000] 3[30127]: segfault at 8 ip 00007f482b53e7e8 sp 00007ffc56e96b40 error 4 in ld-2.17.so[7f482b533000+22000] LustreError: 31432:0:(statahead.c:2399:start_statahead_thread()) lustre: invalid pattern 0X0. Lustre: 1871:0:(lod_lov.c:1414:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000402:0x2086:0x0] with magic=0xbd60bd0 Lustre: 1871:0:(lod_lov.c:1414:lod_parse_striping()) Skipped 1 previous similar message 0[5865]: segfault at 8 ip 00007fbc14f7b7e8 sp 00007ffc9d4e7970 error 4 in ld-2.17.so[7fbc14f70000+22000] 17[6758]: segfault at 8 ip 00007fc01ff797e8 sp 00007ffc723c0860 error 4 in ld-2.17.so[7fc01ff6e000+22000] Lustre: 2264:0:(lod_lov.c:1414:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000402:0x227d:0x0] with magic=0xbd60bd0 Lustre: 2264:0:(lod_lov.c:1414:lod_parse_striping()) Skipped 1 previous similar message Lustre: mdt00_011: service thread pid 4127 was inactive for 40.095 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 32474, comm: mdt00_003 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock_try+0xa0/0x250 [mdt] [<0>] mdt_getattr_name_lock+0x1729/0x2cc0 [mdt] [<0>] mdt_intent_getattr+0x2cc/0x4e0 [mdt] [<0>] mdt_intent_opc.constprop.74+0x211/0xc60 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x397/0x980 [ptlrpc] [<0>] ldlm_handle_enqueue+0x547/0x1900 [ptlrpc] [<0>] tgt_enqueue+0x68/0x240 [ptlrpc] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x257/0xcd0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 1871, comm: mdt00_006 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock+0x88/0x1c0 [mdt] [<0>] mdt_object_find_lock+0x54/0x170 [mdt] [<0>] mdt_reint_setxattr+0x133/0x10e0 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x73c/0xbb0 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x257/0xcd0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: Skipped 2 previous similar messages Pid: 4127, comm: mdt00_011 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock+0x88/0x1c0 [mdt] [<0>] mdt_intent_getxattr+0x80/0x330 [mdt] [<0>] mdt_intent_opc.constprop.74+0x211/0xc60 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x397/0x980 [ptlrpc] [<0>] ldlm_handle_enqueue+0x547/0x1900 [ptlrpc] [<0>] tgt_enqueue+0x68/0x240 [ptlrpc] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x257/0xcd0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: mdt00_012: service thread pid 5630 was inactive for 40.074 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. LustreError: 32087:0:(ldlm_lockd.c:257:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0000_UUID lock: ffff8802f4238400/0x27e01505ef8b3cdc lrc: 3/0,0 mode: PR/PR res: [0x200000402:0x2283:0x0].0x0 bits 0x1b/0x0 rrc: 14 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0x27e01505ef8b3cb2 expref: 1119 pid: 32095 timeout: 3509 lvb_type: 0 Lustre: mdt00_006: service thread pid 1871 completed after 100.034s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_011: service thread pid 4127 completed after 100.018s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_003: service thread pid 32474 completed after 99.945s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: lustre-MDT0000-mdc-ffff8800bb91dd28: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-MDT0000-mdc-ffff8800bb91dd28: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: Skipped 1 previous similar message Lustre: mdt00_010: service thread pid 3972 completed after 99.924s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: lustre-MDT0000-mdc-ffff8800bb91dd28: This client was evicted by lustre-MDT0000; in progress operations using this service will fail. Lustre: mdt00_012: service thread pid 5630 completed after 99.895s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_007: service thread pid 2264 completed after 99.559s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_016: service thread pid 14170 completed after 99.580s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_015: service thread pid 14168 completed after 99.476s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_013: service thread pid 10875 completed after 99.834s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_002: service thread pid 32096 completed after 99.552s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 8231:0:(file.c:6187:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000007:0x1:0x0] error: rc = -108 LustreError: 8208:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800bb91dd28: inode [0x200000401:0x21ee:0x0] mdc close failed: rc = -108 LustreError: 7855:0:(mdc_request.c:1469:mdc_read_page()) lustre-MDT0000-mdc-ffff8800bb91dd28: [0x200000401:0x1:0x0] lock enqueue fails: rc = -108 Lustre: lustre-MDT0000-mdc-ffff8800bb91dd28: Connection restored to (at 0@lo) 4[8419]: segfault at 0 ip 0000000000403e5f sp 00007fff646d03f0 error 6 in 4[400000+6000] Lustre: 6215:0:(osd_internal.h:1332:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 618, rollback = 7 Lustre: 6215:0:(osd_internal.h:1332:osd_trans_exec_op()) Skipped 201 previous similar messages Lustre: 6215:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 6215:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 201 previous similar messages Lustre: 6215:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 6215:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 201 previous similar messages Lustre: 6215:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 6215:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 201 previous similar messages Lustre: 6215:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 6215:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 201 previous similar messages Lustre: 6215:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 6215:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 201 previous similar messages Lustre: 3043:0:(lod_lov.c:1414:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0x160:0x0] with magic=0xbd60bd0 Lustre: 3043:0:(lod_lov.c:1414:lod_parse_striping()) Skipped 1 previous similar message 8[14700]: segfault at 8 ip 00007f98f33487e8 sp 00007ffd3e171be0 error 4 in ld-2.17.so[7f98f333d000+22000] Lustre: 32474:0:(lod_lov.c:1414:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000402:0x2525:0x0] with magic=0xbd60bd0 Lustre: 32474:0:(lod_lov.c:1414:lod_parse_striping()) Skipped 1 previous similar message LustreError: 17052:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800bb91dd28: inode [0x200000402:0x2551:0x0] mdc close failed: rc = -13 LustreError: 17052:0:(file.c:247:ll_close_inode_openhandle()) Skipped 3 previous similar messages traps: 4[23610] trap invalid opcode ip:4059da sp:7ffea1b93778 error:0 in 4[400000+6000] 10[31241]: segfault at 1c ip 00000000004017b0 sp 00007fff42e64b90 error 6 in 9 (deleted)[400000+6000] 0[32651]: segfault at 0 ip (null) sp 00007ffc358fe6c8 error 14 in 0[400000+6000] 13[2639]: segfault at 0 ip 0000000000403e5f sp 00007ffd98f61380 error 6 in 13[400000+6000] 13[2973]: segfault at 8 ip 00007fc1b74f97e8 sp 00007fff8cf53050 error 4 in ld-2.17.so[7fc1b74ee000+22000] 8[5509]: segfault at 1c ip 00000000004017b0 sp 00007ffd97cab560 error 6 in 8[400000+6000] LustreError: 6420:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802f0f2efc8: inode [0x200000403:0x9ed:0x0] mdc close failed: rc = -13 5[10969]: segfault at 8 ip 00007f5b716727e8 sp 00007fff51770ed0 error 4 in ld-2.17.so[7f5b71667000+22000] 1[12107]: segfault at 0 ip 0000000000403e5f sp 00007ffcaf260fd0 error 6 in 1[400000+6000] 18[12684]: segfault at 8 ip 00007fa37bb4a7e8 sp 00007ffe5fa06740 error 4 in ld-2.17.so[7fa37bb3f000+22000] 8[12804]: segfault at 0 ip (null) sp 00007ffc2b691f58 error 14 in 8[400000+6000] 17[12931]: segfault at 0 ip (null) sp 00007ffc1ef27cf8 error 14 in 17[400000+6000] dd (14909) used greatest stack depth: 9600 bytes left Lustre: 14170:0:(lod_lov.c:1414:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000402:0x3176:0x0] with magic=0xbd60bd0 Lustre: 14170:0:(lod_lov.c:1414:lod_parse_striping()) Skipped 1 previous similar message 1[22189]: segfault at 8 ip 00007f3cbb9c97e8 sp 00007ffc45ffc980 error 4 in ld-2.17.so[7f3cbb9be000+22000] 15[23037]: segfault at 8 ip 00007fc5ce1db7e8 sp 00007fffef674440 error 4 in ld-2.17.so[7fc5ce1d0000+22000] 12[25319]: segfault at 8 ip 00007f5ec678f7e8 sp 00007ffe2c97c9c0 error 4 in ld-2.17.so[7f5ec6784000+22000] | Link to test |
racer test 1: racer on clients: centos-70.localnet DURATION=2700 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 PGD a122f067 PUD 840d1067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) osp(OE) ofd(OE) lod(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) crc32_generic lnet(OE) libcfs(OE) crc_t10dif crct10dif_generic crct10dif_common i2c_piix4 virtio_console pcspkr virtio_balloon ip_tables rpcsec_gss_krb5 drm_kms_helper ttm ata_generic pata_acpi drm drm_panel_orientation_quirks floppy ata_piix virtio_blk i2c_core serio_raw libata CPU: 2 PID: 21293 Comm: ll_sa_13371 Kdump: loaded Tainted: P OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.16.0-3.module_el8.7.0+1218+f626c2ff 04/01/2014 task: ffff8802ac828010 ti: ffff8802cce94000 task.ti: ffff8802cce94000 RIP: 0010:[<ffffffff813f502c>] [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 RSP: 0018:ffff8802cce97d78 EFLAGS: 00010296 RAX: ffff8801ee9b4390 RBX: 0000000000000008 RCX: 0000000100260021 RDX: 0000000000000026 RSI: ffff8801ee9b46c0 RDI: 0000000000000008 RBP: ffff8802cce97d88 R08: ffff8802d91d5628 R09: 0000000000000000 R10: ffff8802d91d0358 R11: ffff8802d91d5628 R12: 0000000000000000 R13: ffff8802d061dc40 R14: ffff8801ee9b46c0 R15: ffff8802d91d5628 FS: 0000000000000000(0000) GS:ffff880331a80000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 00000000906ea000 CR4: 00000000000007e0 Call Trace: [<ffffffffa15abbb1>] ll_statahead_thread+0xdb1/0x2000 [lustre] [<ffffffff810df6e2>] ? pick_next_task_fair+0x702/0x8a0 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e329e>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffff810c834d>] ? finish_task_switch+0x5d/0x1b0 [<ffffffffa15aae00>] ? ll_statahead_handle.constprop.28+0x1c0/0x1c0 [lustre] [<ffffffff810ba114>] kthread+0xe4/0xf0 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 | Lustre: 15191:0:(mdd_dir.c:4822:mdd_migrate_object()) lustre-MDD0000: [0x200000403:0x2:0x0]/5 is open, migrate only dentry Lustre: 15191:0:(mdd_dir.c:4822:mdd_migrate_object()) lustre-MDD0001: [0x200000401:0x4:0x0]/13 is open, migrate only dentry Lustre: 31114:0:(mdd_dir.c:4822:mdd_migrate_object()) lustre-MDD0002: [0x280000403:0x1:0x0]/5 is open, migrate only dentry Lustre: 31114:0:(mdd_dir.c:4822:mdd_migrate_object()) Skipped 1 previous similar message Lustre: 30390:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff8802d8781940 x1832939896070656/t4294968153(0) o101->0826c2d5-6bc4-49c2-9c49-6fae1d6b9470@0@lo:463/0 lens 376/864 e 0 to 0 dl 1748027803 ref 1 fl Interpret:H/602/0 rc 0/0 job:'dd.0' uid:0 gid:0 projid:0 ls (28621) used greatest stack depth: 10144 bytes left Lustre: 15189:0:(mdd_dir.c:4822:mdd_migrate_object()) lustre-MDD0000: [0x200000403:0x1:0x0]/7 is open, migrate only dentry LustreError: 31026:0:(mdt_reint.c:2540:mdt_reint_migrate()) lustre-MDT0002: migrate [0x280000404:0x2:0x0]/11 failed: rc = -2 LustreError: 30969:0:(mdd_dir.c:4743:mdd_migrate_cmd_check()) lustre-MDD0002: '5' migration was interrupted, run 'lfs migrate -m 0 -c 1 -H crush 5' to finish migration: rc = -1 LustreError: 30969:0:(mdt_reint.c:2540:mdt_reint_migrate()) lustre-MDT0002: migrate [0x240000403:0x1:0x0]/5 failed: rc = -1 LustreError: 4201:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000404:0x67:0x0]: rc = -5 LustreError: 4201:0:(llite_lib.c:3769:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 2557:0:(mdt_reint.c:2540:mdt_reint_migrate()) lustre-MDT0001: migrate [0x200000404:0x44:0x0]/14 failed: rc = -2 Lustre: 30356:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff880291474b40 x1832939899377280/t4294969557(0) o101->0826c2d5-6bc4-49c2-9c49-6fae1d6b9470@0@lo:484/0 lens 376/816 e 0 to 0 dl 1748027824 ref 1 fl Interpret:H/602/0 rc 0/0 job:'dd.0' uid:0 gid:0 projid:0 Lustre: 30663:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff8802c07c2340 x1832939899853696/t4294969829(0) o101->0826c2d5-6bc4-49c2-9c49-6fae1d6b9470@0@lo:488/0 lens 376/840 e 0 to 0 dl 1748027828 ref 1 fl Interpret:H/602/0 rc 0/0 job:'dd.0' uid:0 gid:0 projid:0 LustreError: 28911:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0001: failed to get lu_attr of [0x240000403:0x4a:0x0]: rc = -2 LustreError: 1483:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802caa8c138: inode [0x240000403:0x4a:0x0] mdc close failed: rc = -2 LustreError: 1483:0:(file.c:247:ll_close_inode_openhandle()) Skipped 1 previous similar message Lustre: 31190:0:(mdd_dir.c:4822:mdd_migrate_object()) lustre-MDD0000: [0x200000403:0x1:0x0]/9 is open, migrate only dentry Lustre: 31190:0:(mdd_dir.c:4822:mdd_migrate_object()) Skipped 1 previous similar message LustreError: 6318:0:(mdd_dir.c:4743:mdd_migrate_cmd_check()) lustre-MDD0001: '17' migration was interrupted, run 'lfs migrate -m 0 -c 3 -H crush 17' to finish migration: rc = -1 LustreError: 6318:0:(mdt_reint.c:2540:mdt_reint_migrate()) lustre-MDT0001: migrate [0x200000403:0x2:0x0]/17 failed: rc = -1 LustreError: 30969:0:(mdt_reint.c:2540:mdt_reint_migrate()) lustre-MDT0000: migrate [0x280000403:0x8e:0x0]/18 failed: rc = -2 LustreError: 30969:0:(mdt_reint.c:2540:mdt_reint_migrate()) Skipped 3 previous similar messages Lustre: dir [0x280000403:0x8e:0x0] stripe 3 readdir failed: -2, directory is partially accessed! LustreError: 9781:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802cf908008: inode [0x280000404:0x2c:0x0] mdc close failed: rc = -2 LustreError: 30126:0:(mdd_dir.c:4743:mdd_migrate_cmd_check()) lustre-MDD0001: '14' migration was interrupted, run 'lfs migrate -m 2 -c 3 -H crush 14' to finish migration: rc = -1 LustreError: 30126:0:(mdt_reint.c:2540:mdt_reint_migrate()) lustre-MDT0001: migrate [0x200000403:0x1:0x0]/14 failed: rc = -1 LustreError: 30126:0:(mdt_reint.c:2540:mdt_reint_migrate()) Skipped 1 previous similar message LustreError: 31026:0:(mdd_dir.c:4743:mdd_migrate_cmd_check()) lustre-MDD0000: '17' migration was interrupted, run 'lfs migrate -m 0 -c 3 -H crush 17' to finish migration: rc = -1 5[13741]: segfault at 8 ip 00007f3b58c9e7e8 sp 00007ffde1930ee0 error 4 in ld-2.17.so[7f3b58c93000+22000] Lustre: 30126:0:(mdd_dir.c:4822:mdd_migrate_object()) lustre-MDD0001: [0x240000401:0x6:0x0]/sleep is open, migrate only dentry LustreError: 18515:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802cf908008: inode [0x200000403:0x14:0x0] mdc close failed: rc = -2 LustreError: 22630:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000404:0x185:0x0]: rc = -5 LustreError: 22630:0:(lcommon_cl.c:188:cl_file_inode_init()) Skipped 2 previous similar messages LustreError: 22630:0:(llite_lib.c:3769:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 22630:0:(llite_lib.c:3769:ll_prep_inode()) Skipped 2 previous similar messages LustreError: 30948:0:(mdt_reint.c:2540:mdt_reint_migrate()) lustre-MDT0002: migrate [0x200000404:0x192:0x0]/11 failed: rc = -2 LustreError: 30948:0:(mdt_reint.c:2540:mdt_reint_migrate()) Skipped 1 previous similar message Lustre: 30369:0:(lod_lov.c:1414:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000404:0x1b2:0x0] with magic=0xbd60bd0 LustreError: 20250:0:(lov_object.c:1350:lov_layout_change()) lustre-clilov-ffff8802cf908008: cannot apply new layout on [0x240000404:0x185:0x0] : rc = -5 LustreError: 20250:0:(vvp_io.c:1905:vvp_io_init()) lustre: refresh file layout [0x240000404:0x185:0x0] error -5. LustreError: 15191:0:(mdd_dir.c:4743:mdd_migrate_cmd_check()) lustre-MDD0002: '17' migration was interrupted, run 'lfs migrate -m 0 -c 3 -H crush 17' to finish migration: rc = -1 LustreError: 23930:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802caa8c138: inode [0x280000403:0x1a:0x0] mdc close failed: rc = -2 Lustre: 15191:0:(mdd_dir.c:4822:mdd_migrate_object()) lustre-MDD0000: [0x200000403:0x1:0x0]/18 is open, migrate only dentry Lustre: 15191:0:(mdd_dir.c:4822:mdd_migrate_object()) Skipped 3 previous similar messages Lustre: dir [0x200000404:0x192:0x0] stripe 3 readdir failed: -2, directory is partially accessed! LustreError: 24624:0:(lov_object.c:1350:lov_layout_change()) lustre-clilov-ffff8802caa8c138: cannot apply new layout on [0x200000403:0x3f9:0x0] : rc = -5 LustreError: 24624:0:(vvp_io.c:1905:vvp_io_init()) lustre: refresh file layout [0x200000403:0x3f9:0x0] error -5. LustreError: 30600:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0000: [0x200000403:0x262:0x0] migrate mdt count mismatch 3 != 1 LustreError: 29496:0:(lov_object.c:1350:lov_layout_change()) lustre-clilov-ffff8802cf908008: cannot apply new layout on [0x240000404:0x185:0x0] : rc = -5 LustreError: 29496:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000404:0x185:0x0]: rc = -5 LustreError: 29496:0:(llite_lib.c:3769:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 27232:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x280000403:0x2ec:0x0]: rc = -5 LustreError: 27232:0:(llite_lib.c:3769:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 7[32356]: segfault at 8 ip 00007f44619f27e8 sp 00007ffcee2ce9a0 error 4 in ld-2.17.so[7f44619e7000+22000] LustreError: 315:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000403:0x2df:0x0]: rc = -5 LustreError: 315:0:(llite_lib.c:3769:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 131:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 16 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 LustreError: 315:0:(llite_lib.c:3769:ll_prep_inode()) Skipped 1 previous similar message Lustre: 30704:0:(lod_lov.c:1414:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000404:0x203:0x0] with magic=0xbd60bd0 Lustre: 30704:0:(lod_lov.c:1414:lod_parse_striping()) Skipped 1 previous similar message LustreError: 30966:0:(mdt_reint.c:2540:mdt_reint_migrate()) lustre-MDT0002: migrate [0x240000404:0x18a:0x0]/12 failed: rc = -2 LustreError: 30966:0:(mdt_reint.c:2540:mdt_reint_migrate()) Skipped 3 previous similar messages Lustre: 30532:0:(lod_lov.c:1414:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000404:0x200:0x0] with magic=0xbd60bd0 Lustre: 30532:0:(lod_lov.c:1414:lod_parse_striping()) Skipped 3 previous similar messages LustreError: 31190:0:(mdd_dir.c:4743:mdd_migrate_cmd_check()) lustre-MDD0002: '13' migration was interrupted, run 'lfs migrate -m 1 -c 2 -H crush 13' to finish migration: rc = -1 LustreError: 5789:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x280000404:0x34b:0x0]: rc = -5 LustreError: 5789:0:(lcommon_cl.c:188:cl_file_inode_init()) Skipped 4 previous similar messages LustreError: 5789:0:(llite_lib.c:3769:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 5789:0:(llite_lib.c:3769:ll_prep_inode()) Skipped 3 previous similar messages Lustre: 30966:0:(mdd_dir.c:4822:mdd_migrate_object()) lustre-MDD0000: [0x200000400:0xf:0x0]/9 is open, migrate only dentry Lustre: 30966:0:(mdd_dir.c:4822:mdd_migrate_object()) Skipped 3 previous similar messages Lustre: 30659:0:(lod_lov.c:1414:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000404:0x227:0x0] with magic=0xbd60bd0 Lustre: 30659:0:(lod_lov.c:1414:lod_parse_striping()) Skipped 1 previous similar message Lustre: lustre-MDT0002: trigger partial OI scrub for RPC inconsistency, checking FID [0x280000403:0x248:0x0]/0xa): rc = 0 LustreError: 13337:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x280000404:0x29f:0x0]: rc = -5 LustreError: 13337:0:(lcommon_cl.c:188:cl_file_inode_init()) Skipped 8 previous similar messages LustreError: 13337:0:(llite_lib.c:3769:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 13337:0:(llite_lib.c:3769:ll_prep_inode()) Skipped 8 previous similar messages Lustre: 28405:0:(lod_lov.c:1414:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000404:0x35a:0x0] with magic=0xbd60bd0 Lustre: 28405:0:(lod_lov.c:1414:lod_parse_striping()) Skipped 5 previous similar messages LustreError: 15489:0:(lov_object.c:1350:lov_layout_change()) lustre-clilov-ffff8802cf908008: cannot apply new layout on [0x240000404:0x185:0x0] : rc = -5 17[12910]: segfault at 0 ip (null) sp 00007ffd3f23d008 error 14 in 17[400000+6000] Lustre: lustre-MDT0001: trigger partial OI scrub for RPC inconsistency, checking FID [0x240000403:0x306:0x0]/0xa): rc = 0 LustreError: 6207:0:(lustre_lmv.h:500:lmv_is_sane()) unknown layout LMV: magic=0xcd40cd0 count=4 index=3 hash=crush:0x82000003 version=1 migrate_offset=3 migrate_hash=fnv_1a_64:2 pool= Lustre: lustre-MDT0000: trigger partial OI scrub for RPC inconsistency, checking FID [0x200000404:0x31c:0x0]/0xa): rc = 0 LustreError: 30692:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0000: [0x200000403:0x691:0x0] migrate mdt count mismatch 3 != 2 LustreError: 30465:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0002: failed to get lu_attr of [0x280000403:0x28e:0x0]: rc = -2 LustreError: 30465:0:(mdd_object.c:3901:mdd_close()) Skipped 1 previous similar message LustreError: 16754:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802caa8c138: inode [0x280000403:0x28e:0x0] mdc close failed: rc = -2 LustreError: 218:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 19 [0x280000403:0x3ac:0x0] inode@0000000000000000: rc = -5 LustreError: 218:0:(statahead.c:825:ll_statahead_interpret_work()) Skipped 1 previous similar message LustreError: 31026:0:(mdd_dir.c:4743:mdd_migrate_cmd_check()) lustre-MDD0002: '4' migration was interrupted, run 'lfs migrate -m 0 -c 2 -H crush 4' to finish migration: rc = -1 LustreError: 28399:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000404:0x3f6:0x0]: rc = -5 LustreError: 28399:0:(lcommon_cl.c:188:cl_file_inode_init()) Skipped 21 previous similar messages LustreError: 28399:0:(llite_lib.c:3769:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 28399:0:(llite_lib.c:3769:ll_prep_inode()) Skipped 21 previous similar messages LustreError: 31114:0:(lustre_lmv.h:500:lmv_is_sane()) unknown layout LMV: magic=0xcd40cd0 count=3 index=2 hash=crush:0x82000003 version=1 migrate_offset=2 migrate_hash=fnv_1a_64:2 pool= LustreError: 30126:0:(mdt_reint.c:2540:mdt_reint_migrate()) lustre-MDT0000: migrate [0x200000403:0x5f5:0x0]/9 failed: rc = -2 LustreError: 30126:0:(mdt_reint.c:2540:mdt_reint_migrate()) Skipped 11 previous similar messages Lustre: dir [0x240000404:0x240:0x0] stripe 3 readdir failed: -2, directory is partially accessed! Lustre: Skipped 2 previous similar messages Lustre: 30969:0:(mdd_dir.c:4822:mdd_migrate_object()) lustre-MDD0000: [0x200000403:0x2:0x0]/13 is open, migrate only dentry Lustre: 30969:0:(mdd_dir.c:4822:mdd_migrate_object()) Skipped 7 previous similar messages LustreError: 30642:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0000: [0x200000403:0x5f5:0x0] migrate mdt count mismatch 2 != 1 LustreError: 6702:0:(mdd_dir.c:4743:mdd_migrate_cmd_check()) lustre-MDD0002: '12' migration was interrupted, run 'lfs migrate -m 0 -c 3 -H crush 12' to finish migration: rc = -1 LustreError: 6702:0:(mdd_dir.c:4743:mdd_migrate_cmd_check()) Skipped 2 previous similar messages LustreError: 306:0:(lov_object.c:1350:lov_layout_change()) lustre-clilov-ffff8802cf908008: cannot apply new layout on [0x240000404:0x185:0x0] : rc = -5 LustreError: 306:0:(vvp_io.c:1905:vvp_io_init()) lustre: refresh file layout [0x240000404:0x185:0x0] error -5. Lustre: 30393:0:(lod_lov.c:1414:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000404:0x552:0x0] with magic=0xbd60bd0 Lustre: 30393:0:(lod_lov.c:1414:lod_parse_striping()) Skipped 11 previous similar messages Lustre: 28988:0:(lod_lov.c:1414:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000403:0x483:0x0] with magic=0xbd60bd0 Lustre: 28988:0:(lod_lov.c:1414:lod_parse_striping()) Skipped 3 previous similar messages dd (17318) used greatest stack depth: 10016 bytes left LustreError: 16921:0:(lov_object.c:1350:lov_layout_change()) lustre-clilov-ffff8802cf908008: cannot apply new layout on [0x280000404:0x43a:0x0] : rc = -5 LustreError: 16921:0:(lov_object.c:1350:lov_layout_change()) Skipped 2 previous similar messages LustreError: 10939:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802caa8c138: inode [0x200000404:0x31c:0x0] mdc close failed: rc = -2 LustreError: 10939:0:(file.c:247:ll_close_inode_openhandle()) Skipped 1 previous similar message LustreError: 19468:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000403:0x50e:0x0]: rc = -5 LustreError: 19468:0:(lcommon_cl.c:188:cl_file_inode_init()) Skipped 21 previous similar messages LustreError: 19468:0:(llite_lib.c:3769:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 19468:0:(llite_lib.c:3769:ll_prep_inode()) Skipped 21 previous similar messages LustreError: 22101:0:(lov_object.c:1350:lov_layout_change()) lustre-clilov-ffff8802cf908008: cannot apply new layout on [0x240000404:0x185:0x0] : rc = -5 Lustre: dir [0x240000403:0x612:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 4 previous similar messages LustreError: 16586:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802caa8c138: inode [0x200000404:0x3d7:0x0] mdc close failed: rc = -2 LustreError: 315:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 3 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 LustreError: 2567:0:(mdd_dir.c:4743:mdd_migrate_cmd_check()) lustre-MDD0001: '5' migration was interrupted, run 'lfs migrate -m 0 -c 2 -H crush 5' to finish migration: rc = -1 LustreError: 2567:0:(mdd_dir.c:4743:mdd_migrate_cmd_check()) Skipped 7 previous similar messages 17[30918]: segfault at 8 ip 00007f8120f6d7e8 sp 00007ffcbb208e10 error 4 in ld-2.17.so[7f8120f62000+22000] 7[32544]: segfault at 8 ip 00007f75848047e8 sp 00007ffd08afbe30 error 4 in ld-2.17.so[7f75847f9000+22000] 14[1547]: segfault at 0 ip 0000000000403e5f sp 00007ffc8a7587a0 error 6 in 14[400000+6000] Lustre: 30729:0:(lod_lov.c:1414:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000403:0x7f1:0x0] with magic=0xbd60bd0 Lustre: 30729:0:(lod_lov.c:1414:lod_parse_striping()) Skipped 31 previous similar messages LustreError: 10476:0:(llite_lib.c:1872:ll_update_lsm_md()) lustre: [0x240000403:0x87f:0x0] dir layout mismatch: LustreError: 10476:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=2 count=2 index=1 hash=crush:0x2000003 max_inherit=0 max_inherit_rr=0 version=2 migrate_offset=0 migrate_hash=invalid:0 pool= LustreError: 10476:0:(lustre_lmv.h:167:lmv_stripe_object_dump()) stripe[0] [0x240000400:0x36:0x0] LustreError: 10476:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=1 count=3 index=1 hash=crush:0x82000003 max_inherit=0 max_inherit_rr=0 version=1 migrate_offset=2 migrate_hash=fnv_1a_64:2 pool= LustreError: 10472:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=2 count=2 index=1 hash=crush:0x2000003 max_inherit=0 max_inherit_rr=0 version=2 migrate_offset=0 migrate_hash=invalid:0 pool= LustreError: 10472:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=1 count=3 index=1 hash=crush:0x82000003 max_inherit=0 max_inherit_rr=0 version=1 migrate_offset=2 migrate_hash=fnv_1a_64:2 pool= LustreError: 7264:0:(llite_nfs.c:430:ll_dir_get_parent_fid()) lustre: failure inode [0x200000403:0x691:0x0] get parent: rc = -2 LustreError: 7262:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802caa8c138: inode [0x200000403:0x691:0x0] mdc close failed: rc = -2 LustreError: 7262:0:(file.c:247:ll_close_inode_openhandle()) Skipped 1 previous similar message LustreError: 15190:0:(mdt_reint.c:2540:mdt_reint_migrate()) lustre-MDT0001: migrate [0x200000403:0x2:0x0]/11 failed: rc = -1 LustreError: 15190:0:(mdt_reint.c:2540:mdt_reint_migrate()) Skipped 30 previous similar messages LustreError: 6970:0:(lov_object.c:1350:lov_layout_change()) lustre-clilov-ffff8802caa8c138: cannot apply new layout on [0x240000403:0x50e:0x0] : rc = -5 LustreError: 6970:0:(vvp_io.c:1905:vvp_io_init()) lustre: refresh file layout [0x240000403:0x50e:0x0] error -5. LustreError: 6970:0:(vvp_io.c:1905:vvp_io_init()) Skipped 1 previous similar message Lustre: 15191:0:(mdd_dir.c:4822:mdd_migrate_object()) lustre-MDD0000: [0x200000403:0x1:0x0]/19 is open, migrate only dentry Lustre: 15191:0:(mdd_dir.c:4822:mdd_migrate_object()) Skipped 40 previous similar messages LustreError: 31:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 3 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 Lustre: dir [0x200000404:0x9fa:0x0] stripe 2 readdir failed: -2, directory is partially accessed! Lustre: Skipped 4 previous similar messages LustreError: 30639:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0001: [0x240000403:0x2e6:0x0] migrate mdt count mismatch 3 != 2 LustreError: 28163:0:(llite_lib.c:1872:ll_update_lsm_md()) lustre: [0x280000403:0xb91:0x0] dir layout mismatch: LustreError: 28163:0:(llite_lib.c:1872:ll_update_lsm_md()) Skipped 1 previous similar message LustreError: 28163:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=3 count=3 index=2 hash=crush:0x2000003 max_inherit=0 max_inherit_rr=0 version=2 migrate_offset=0 migrate_hash=invalid:0 pool= LustreError: 28163:0:(lustre_lmv.h:167:lmv_stripe_object_dump()) stripe[0] [0x280000400:0x31:0x0] LustreError: 28163:0:(lustre_lmv.h:167:lmv_stripe_object_dump()) Skipped 9 previous similar messages LustreError: 28163:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=1 count=4 index=2 hash=crush:0x82000003 max_inherit=0 max_inherit_rr=0 version=1 migrate_offset=3 migrate_hash=fnv_1a_64:2 pool= LustreError: 28166:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=3 count=3 index=2 hash=crush:0x2000003 max_inherit=0 max_inherit_rr=0 version=2 migrate_offset=0 migrate_hash=invalid:0 pool= LustreError: 28166:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=1 count=4 index=2 hash=crush:0x82000003 max_inherit=0 max_inherit_rr=0 version=1 migrate_offset=3 migrate_hash=fnv_1a_64:2 pool= LustreError: 22882:0:(vvp_io.c:1905:vvp_io_init()) lustre: refresh file layout [0x240000403:0x913:0x0] error -5. 15[29786]: segfault at 8 ip 00007f5b6e0db7e8 sp 00007ffe68c1e0f0 error 4 in ld-2.17.so[7f5b6e0d0000+22000] Lustre: dir [0x240000404:0x934:0x0] stripe 2 readdir failed: -2, directory is partially accessed! Lustre: Skipped 19 previous similar messages LustreError: 30128:0:(vvp_io.c:1905:vvp_io_init()) lustre: refresh file layout [0x240000404:0x583:0x0] error -5. LustreError: 7368:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000403:0xc2b:0x0]: rc = -5 LustreError: 7368:0:(lcommon_cl.c:188:cl_file_inode_init()) Skipped 29 previous similar messages LustreError: 7368:0:(llite_lib.c:3769:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 7368:0:(llite_lib.c:3769:ll_prep_inode()) Skipped 29 previous similar messages 3[7801]: segfault at 8 ip 00007faf59bd47e8 sp 00007ffd8b263840 error 4 in ld-2.17.so[7faf59bc9000+22000] LustreError: 4407:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802cf908008: inode [0x280000403:0x7a3:0x0] mdc close failed: rc = -2 LustreError: 4407:0:(file.c:247:ll_close_inode_openhandle()) Skipped 4 previous similar messages LustreError: 15191:0:(mdd_dir.c:4743:mdd_migrate_cmd_check()) lustre-MDD0002: '5' migration was interrupted, run 'lfs migrate -m 1 -c 1 -H crush 5' to finish migration: rc = -1 LustreError: 15191:0:(mdd_dir.c:4743:mdd_migrate_cmd_check()) Skipped 10 previous similar messages 14[14255]: segfault at 8 ip 00007f2beba9a7e8 sp 00007ffd57d7be60 error 4 in ld-2.17.so[7f2beba8f000+22000] 14[15618]: segfault at 8 ip 00007f690b6997e8 sp 00007fffc8e1cc20 error 4 in ld-2.17.so[7f690b68e000+22000] 14[14855]: segfault at 8 ip 00007f4acc3c27e8 sp 00007ffc72d7f8e0 error 4 in ld-2.17.so[7f4acc3b7000+22000] 6[15704]: segfault at 8 ip 00007fcb83f817e8 sp 00007fff3ff40440 error 4 in ld-2.17.so[7fcb83f76000+22000] 5[15324]: segfault at 8 ip 00007f05f9b377e8 sp 00007ffe1c1458d0 error 4 in ld-2.17.so[7f05f9b2c000+22000] LustreError: 6073:0:(lov_object.c:1350:lov_layout_change()) lustre-clilov-ffff8802caa8c138: cannot apply new layout on [0x280000403:0xc21:0x0] : rc = -5 LustreError: 6073:0:(lov_object.c:1350:lov_layout_change()) Skipped 6 previous similar messages LustreError: 6073:0:(vvp_io.c:1905:vvp_io_init()) lustre: refresh file layout [0x280000403:0xc21:0x0] error -5. LustreError: 28312:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0002: [0x280000403:0xbd6:0x0] migrate mdt count mismatch 2 != 1 11[29403]: segfault at 8 ip 00007ff8d37a57e8 sp 00007ffcf58e0690 error 4 in ld-2.17.so[7ff8d379a000+22000] 6[3126]: segfault at 8 ip 00007f91071c47e8 sp 00007ffdb62870e0 error 4 in ld-2.17.so[7f91071b9000+22000] Lustre: 30376:0:(lod_lov.c:1414:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000403:0xc27:0x0] with magic=0xbd60bd0 Lustre: 30376:0:(lod_lov.c:1414:lod_parse_striping()) Skipped 15 previous similar messages LustreError: 316:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 18 [0x280000403:0xcb2:0x0] inode@0000000000000000: rc = -5 LustreError: 31702:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0002: [0x280000403:0x6de:0x0] migrate mdt count mismatch 2 != 1 LustreError: 31702:0:(mdt_xattr.c:402:mdt_dir_layout_update()) Skipped 1 previous similar message 16[14976]: segfault at 8 ip 00007fad7fffd7e8 sp 00007ffefa6ac780 error 4 in ld-2.17.so[7fad7fff2000+22000] LustreError: 31:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 18 [0x280000403:0xcb2:0x0] inode@0000000000000000: rc = -5 Lustre: dir [0x280000404:0xc89:0x0] stripe 2 readdir failed: -2, directory is partially accessed! Lustre: Skipped 1 previous similar message LustreError: 14238:0:(vvp_io.c:1905:vvp_io_init()) lustre: refresh file layout [0x280000404:0xdb2:0x0] error -5. LustreError: 28022:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000404:0x97b:0x0]: rc = -2 LustreError: 28022:0:(mdd_object.c:3901:mdd_close()) Skipped 1 previous similar message LustreError: 8584:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802cf908008: inode [0x200000404:0x97b:0x0] mdc close failed: rc = -2 LustreError: 8584:0:(file.c:247:ll_close_inode_openhandle()) Skipped 4 previous similar messages 13[24265]: segfault at 8 ip 00007f30dc22c7e8 sp 00007ffdbb0b40d0 error 4 in ld-2.17.so[7f30dc221000+22000] LustreError: 30318:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0002: [0x280000404:0xf6b:0x0] migrate mdt count mismatch 2 != 3 LustreError: 30318:0:(mdt_xattr.c:402:mdt_dir_layout_update()) Skipped 1 previous similar message LustreError: 127:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 18 [0x280000403:0xcb2:0x0] inode@0000000000000000: rc = -5 Lustre: dir [0x240000404:0xca3:0x0] stripe 3 readdir failed: -2, directory is partially accessed! LustreError: 615:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0001: failed to get lu_attr of [0x240000403:0xc1b:0x0]: rc = -2 LustreError: 615:0:(mdd_object.c:3901:mdd_close()) Skipped 1 previous similar message LustreError: 2327:0:(lov_object.c:1350:lov_layout_change()) lustre-clilov-ffff8802caa8c138: cannot apply new layout on [0x280000403:0xffc:0x0] : rc = -5 LustreError: 2327:0:(lov_object.c:1350:lov_layout_change()) Skipped 10 previous similar messages LustreError: 30393:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0002: [0x280000403:0x1083:0x0] migrate mdt count mismatch 2 != 3 LustreError: 316:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 11 [0x240000403:0xbb7:0x0] inode@0000000000000000: rc = -5 5[10360]: segfault at 8 ip 00007fee381447e8 sp 00007ffd20327a20 error 4 in ld-2.17.so[7fee38139000+22000] LustreError: 30126:0:(mdt_reint.c:2540:mdt_reint_migrate()) lustre-MDT0001: migrate [0x200000403:0x1488:0x0]/3 failed: rc = -2 LustreError: 30126:0:(mdt_reint.c:2540:mdt_reint_migrate()) Skipped 53 previous similar messages Lustre: 6207:0:(mdd_dir.c:4822:mdd_migrate_object()) lustre-MDD0000: [0x240000403:0x1:0x0]/10 is open, migrate only dentry Lustre: 6207:0:(mdd_dir.c:4822:mdd_migrate_object()) Skipped 47 previous similar messages LustreError: 15386:0:(vvp_io.c:1905:vvp_io_init()) lustre: refresh file layout [0x200000403:0x1499:0x0] error -5. LustreError: 15386:0:(vvp_io.c:1905:vvp_io_init()) Skipped 3 previous similar messages Lustre: 23567:0:(lod_lov.c:1414:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000403:0x1181:0x0] with magic=0xbd60bd0 Lustre: 23567:0:(lod_lov.c:1414:lod_parse_striping()) Skipped 57 previous similar messages LustreError: 23080:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 16 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 LustreError: 23080:0:(statahead.c:825:ll_statahead_interpret_work()) Skipped 9 previous similar messages Lustre: mdt00_000: service thread pid 15170 was inactive for 72.119 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 15170, comm: mdt00_000 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_fini+0xabc/0xfd0 [ptlrpc] [<0>] ldlm_cli_enqueue+0x461/0xb00 [ptlrpc] [<0>] osp_md_object_lock+0x151/0x2f0 [osp] [<0>] lod_object_lock+0xdb/0x7c0 [lod] [<0>] mdd_object_lock+0x2d/0xd0 [mdd] [<0>] mdt_remote_object_lock_try+0x14c/0x189 [mdt] [<0>] mdt_object_lock_internal+0x3c4/0x470 [mdt] [<0>] mdt_object_lock+0x88/0x1c0 [mdt] [<0>] mdt_reint_link+0x8d4/0xf40 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x73c/0xbb0 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x257/0xcd0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: mdt00_016: service thread pid 28488 was inactive for 72.267 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 28488, comm: mdt00_016 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock+0x88/0x1c0 [mdt] [<0>] mdt_getattr_name_lock+0xf6/0x2cc0 [mdt] [<0>] mdt_intent_getattr+0x2cc/0x4e0 [mdt] [<0>] mdt_intent_opc.constprop.74+0x211/0xc60 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x397/0x980 [ptlrpc] [<0>] ldlm_handle_enqueue+0x547/0x1900 [ptlrpc] [<0>] tgt_enqueue+0x68/0x240 [ptlrpc] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x257/0xcd0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: mdt00_033: service thread pid 30351 was inactive for 72.195 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 30351, comm: mdt00_033 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock+0x88/0x1c0 [mdt] [<0>] mdt_getattr_name_lock+0xc4a/0x2cc0 [mdt] [<0>] mdt_intent_getattr+0x2cc/0x4e0 [mdt] [<0>] mdt_intent_opc.constprop.74+0x211/0xc60 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x397/0x980 [ptlrpc] [<0>] ldlm_handle_enqueue+0x547/0x1900 [ptlrpc] [<0>] tgt_enqueue+0x68/0x240 [ptlrpc] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x257/0xcd0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: dir [0x280000403:0x122c:0x0] stripe 3 readdir failed: -2, directory is partially accessed! Lustre: Skipped 21 previous similar messages LustreError: 52:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000403:0x13be:0x0]: rc = -5 LustreError: 52:0:(lcommon_cl.c:188:cl_file_inode_init()) Skipped 119 previous similar messages LustreError: 52:0:(llite_lib.c:3769:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 52:0:(llite_lib.c:3769:ll_prep_inode()) Skipped 119 previous similar messages 14[4917]: segfault at 8 ip 00007f05ee3267e8 sp 00007fff32dc3750 error 4 in ld-2.17.so[7f05ee31b000+22000] 15[3218]: segfault at 8 ip 00007f1cf8b0c7e8 sp 00007ffc495ce050 error 4 in ld-2.17.so[7f1cf8b01000+22000] LustreError: 6330:0:(mdd_dir.c:4743:mdd_migrate_cmd_check()) lustre-MDD0000: '0' migration was interrupted, run 'lfs migrate -m 2 -c 1 -H crush 0' to finish migration: rc = -1 LustreError: 6330:0:(mdd_dir.c:4743:mdd_migrate_cmd_check()) Skipped 15 previous similar messages LustreError: 14993:0:(ldlm_lockd.c:257:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0000_UUID lock: ffff8802a02d4b40/0xa9fda64c8cb13e53 lrc: 3/0,0 mode: PR/PR res: [0x200000403:0x143e:0x0].0x0 bits 0x1b/0x0 rrc: 3 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0xa9fda64c8cb13e30 expref: 777 pid: 30532 timeout: 784 lvb_type: 0 LustreError: lustre-MDT0000-mdc-ffff8802caa8c138: operation mds_getxattr to node 0@lo failed: rc = -107 Lustre: lustre-MDT0000-mdc-ffff8802caa8c138: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: lustre-MDT0000-mdc-ffff8802caa8c138: This client was evicted by lustre-MDT0000; in progress operations using this service will fail. LustreError: 8111:0:(llite_lib.c:2023:ll_md_setattr()) md_setattr fails: rc = -5 LustreError: 29524:0:(mdc_request.c:1469:mdc_read_page()) lustre-MDT0000-mdc-ffff8802caa8c138: [0x200000402:0x4a:0x0] lock enqueue fails: rc = -5 LustreError: 29524:0:(file.c:6187:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000007:0x1:0x0] error: rc = -108 Lustre: mdt00_000: service thread pid 15170 completed after 100.196s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_016: service thread pid 28488 completed after 99.565s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 8812:0:(ldlm_resource.c:981:ldlm_resource_complain()) lustre-MDT0000-mdc-ffff8802caa8c138: namespace resource [0x200000403:0x143e:0x0].0x0 (ffff8802d53e16c0) refcount nonzero (1) after lock cleanup; forcing cleanup. Lustre: mdt00_033: service thread pid 30351 completed after 98.244s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: lustre-MDT0000-mdc-ffff8802caa8c138: Connection restored to 192.168.123.72@tcp (at 0@lo) LustreError: 30643:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0000: [0x200000405:0x114:0x0] migrate mdt count mismatch 1 != 2 LustreError: 30643:0:(mdt_xattr.c:402:mdt_dir_layout_update()) Skipped 2 previous similar messages 2[9222]: segfault at 0 ip (null) sp 00007ffc5f69f2e8 error 14 in 2[400000+6000] LustreError: 8202:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802cf908008: inode [0x200000405:0x23a:0x0] mdc close failed: rc = -2 LustreError: 8202:0:(file.c:247:ll_close_inode_openhandle()) Skipped 28 previous similar messages Lustre: dir [0x280000404:0x1553:0x0] stripe 2 readdir failed: -2, directory is partially accessed! Lustre: Skipped 5 previous similar messages LustreError: 30636:0:(mdt_handler.c:746:mdt_pack_acl2body()) lustre-MDT0000: unable to read [0x200000403:0x1c7f:0x0] ACL: rc = -2 LustreError: 131:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 12 [0x200000405:0x3eb:0x0] inode@0000000000000000: rc = -5 LustreError: 131:0:(statahead.c:825:ll_statahead_interpret_work()) Skipped 5 previous similar messages LustreError: 6167:0:(lov_object.c:1350:lov_layout_change()) lustre-clilov-ffff8802caa8c138: cannot apply new layout on [0x200000403:0x1499:0x0] : rc = -5 LustreError: 6167:0:(lov_object.c:1350:lov_layout_change()) Skipped 8 previous similar messages LustreError: 28978:0:(mdt_open.c:1302:mdt_cross_open()) lustre-MDT0000: [0x200000403:0x1b61:0x0] doesn't exist!: rc = -14 LustreError: 30659:0:(mdt_handler.c:746:mdt_pack_acl2body()) lustre-MDT0000: unable to read [0x200000403:0x1c40:0x0] ACL: rc = -2 LustreError: 2567:0:(lustre_lmv.h:500:lmv_is_sane()) unknown layout LMV: magic=0xcd40cd0 count=3 index=2 hash=crush:0x82000003 version=1 migrate_offset=2 migrate_hash=fnv_1a_64:2 pool= Lustre: 28978:0:(lod_lov.c:1414:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000404:0x1782:0x0] with magic=0xbd60bd0 Lustre: 28978:0:(lod_lov.c:1414:lod_parse_striping()) Skipped 123 previous similar messages LustreError: 30659:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0002: [0x280000403:0x1a4c:0x0] migrate mdt count mismatch 2 != 1 LustreError: 30659:0:(mdt_xattr.c:402:mdt_dir_layout_update()) Skipped 2 previous similar messages 12[32078]: segfault at 0 ip 00007fd41952959a sp 00007fff9b519ce0 error 4 in ld-2.17.so[7fd419527000+22000] LustreError: 28235:0:(vvp_io.c:1905:vvp_io_init()) lustre: refresh file layout [0x280000404:0x1555:0x0] error -5. LustreError: 28235:0:(vvp_io.c:1905:vvp_io_init()) Skipped 1 previous similar message 10[9789]: segfault at 8 ip 00007f04e94aa7e8 sp 00007ffd8f2a4d50 error 4 in ld-2.17.so[7f04e949f000+22000] Lustre: mdt_io00_000: service thread pid 15189 was inactive for 40.005 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 15189, comm: mdt_io00_000 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock+0x88/0x1c0 [mdt] [<0>] mdt_rename_source_lock+0x57/0xf0 [mdt] [<0>] mdt_reint_migrate+0x1832/0x24b0 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x73c/0xbb0 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x257/0xcd0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 30966, comm: mdt_io00_005 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_fini+0xabc/0xfd0 [ptlrpc] [<0>] ldlm_cli_enqueue+0x461/0xb00 [ptlrpc] [<0>] osp_md_object_lock+0x151/0x2f0 [osp] [<0>] lod_object_lock+0xdb/0x7c0 [lod] [<0>] mdd_object_lock+0x2d/0xd0 [mdd] [<0>] mdt_remote_object_lock_try+0x14c/0x189 [mdt] [<0>] mdt_object_lock_internal+0x3c4/0x470 [mdt] [<0>] mdt_rename_lock+0xd9/0x360 [mdt] [<0>] mdt_reint_migrate+0x87e/0x24b0 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x73c/0xbb0 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x257/0xcd0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 15191, comm: mdt_io00_002 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_rename_lock+0xd9/0x360 [mdt] [<0>] mdt_reint_migrate+0x87e/0x24b0 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x73c/0xbb0 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x257/0xcd0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: mdt_io00_018: service thread pid 6702 was inactive for 40.087 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: mdt_io00_010: service thread pid 31186 was inactive for 40.083 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 4 previous similar messages Lustre: mdt_io00_021: service thread pid 959 was inactive for 40.014 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 1 previous similar message Lustre: mdt_io00_003: service thread pid 30126 was inactive for 40.046 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 3 previous similar messages LustreError: 14993:0:(ldlm_lockd.c:257:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0001_UUID lock: ffff88009e659e40/0xa9fda64c8ce81432 lrc: 3/0,0 mode: PR/PR res: [0x240000403:0x1aa6:0x0].0x0 bits 0x1b/0x0 rrc: 5 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0xa9fda64c8ce813ec expref: 473 pid: 15178 timeout: 1214 lvb_type: 0 Lustre: mdt_io00_000: service thread pid 15189 completed after 102.001s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 14984:0:(ldlm_lockd.c:2550:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1748028870 with bad export cookie 12249129409220666952 Lustre: lustre-MDT0001-mdc-ffff8802cf908008: Connection to lustre-MDT0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: lustre-MDT0001-mdc-ffff8802cf908008: This client was evicted by lustre-MDT0001; in progress operations using this service will fail. LustreError: 13944:0:(mdc_request.c:1469:mdc_read_page()) lustre-MDT0001-mdc-ffff8802cf908008: [0x240000400:0x76:0x0] lock enqueue fails: rc = -108 LustreError: 13944:0:(mdc_request.c:1469:mdc_read_page()) Skipped 3 previous similar messages Lustre: dir [0x240000404:0x17bc:0x0] stripe 0 readdir failed: -108, directory is partially accessed! Lustre: Skipped 3 previous similar messages LustreError: 29316:0:(llite_lib.c:2023:ll_md_setattr()) md_setattr fails: rc = -108 Lustre: lustre-MDT0001-mdc-ffff8802cf908008: Connection restored to 192.168.123.72@tcp (at 0@lo) Lustre: mdt_io00_005: service thread pid 30966 completed after 101.752s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_002: service thread pid 15191 completed after 101.753s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_018: service thread pid 6702 completed after 101.244s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_020: service thread pid 13213 completed after 101.476s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_007: service thread pid 31026 completed after 101.391s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 2567:0:(mdt_reint.c:2540:mdt_reint_migrate()) lustre-MDT0002: migrate [0x200000403:0x211b:0x0]/14 failed: rc = -2 LustreError: 2567:0:(mdt_reint.c:2540:mdt_reint_migrate()) Skipped 79 previous similar messages Lustre: mdt_io00_013: service thread pid 2567 completed after 101.281s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_012: service thread pid 2557 completed after 102.313s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_010: service thread pid 31186 completed after 101.175s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_017: service thread pid 6330 completed after 100.484s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_021: service thread pid 959 completed after 99.000s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: 15190:0:(mdd_dir.c:4822:mdd_migrate_object()) lustre-MDD0000: [0x200000403:0x2:0x0]/19 is open, migrate only dentry Lustre: 15190:0:(mdd_dir.c:4822:mdd_migrate_object()) Skipped 69 previous similar messages Lustre: mdt_io00_001: service thread pid 15190 completed after 98.618s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_019: service thread pid 7119 completed after 98.885s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_016: service thread pid 6318 completed after 98.870s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_003: service thread pid 30126 completed after 95.779s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_022: service thread pid 1415 completed after 95.029s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_009: service thread pid 31143 completed after 94.569s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_008: service thread pid 31114 completed after 94.019s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). 7[4913]: segfault at 8 ip 00007f5483ec87e8 sp 00007fff429e9660 error 4 in ld-2.17.so[7f5483ebd000+22000] 7[5159]: segfault at 8 ip 00007fa0a383a7e8 sp 00007ffe105c72b0 error 4 in ld-2.17.so[7fa0a382f000+22000] LustreError: 317:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 7 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 LustreError: 317:0:(statahead.c:825:ll_statahead_interpret_work()) Skipped 12 previous similar messages 0[11318]: segfault at 8 ip 00007f25ee0f67e8 sp 00007ffea9079a30 error 4 in ld-2.17.so[7f25ee0eb000+22000] LustreError: 17674:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000404:0x1997:0x0]: rc = -5 LustreError: 17674:0:(lcommon_cl.c:188:cl_file_inode_init()) Skipped 179 previous similar messages LustreError: 17674:0:(llite_lib.c:3769:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 17674:0:(llite_lib.c:3769:ll_prep_inode()) Skipped 179 previous similar messages LustreError: 959:0:(mdd_dir.c:4743:mdd_migrate_cmd_check()) lustre-MDD0001: '1' migration was interrupted, run 'lfs migrate -m 2 -c 2 -H crush 1' to finish migration: rc = -1 LustreError: 959:0:(mdd_dir.c:4743:mdd_migrate_cmd_check()) Skipped 27 previous similar messages LustreError: 15189:0:(lustre_lmv.h:500:lmv_is_sane()) unknown layout LMV: magic=0xcd40cd0 count=4 index=3 hash=crush:0x82000003 version=1 migrate_offset=3 migrate_hash=fnv_1a_64:2 pool= LustreError: 4204:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0001: [0x240000404:0x1c48:0x0] migrate mdt count mismatch 3 != 1 LustreError: 4204:0:(mdt_xattr.c:402:mdt_dir_layout_update()) Skipped 1 previous similar message LustreError: 10038:0:(vvp_io.c:1905:vvp_io_init()) lustre: refresh file layout [0x240000403:0x199f:0x0] error -5. LustreError: 10038:0:(vvp_io.c:1905:vvp_io_init()) Skipped 3 previous similar messages LustreError: 31186:0:(lustre_lmv.h:500:lmv_is_sane()) unknown layout LMV: magic=0xcd40cd0 count=3 index=2 hash=crush:0x82000003 version=1 migrate_offset=2 migrate_hash=fnv_1a_64:2 pool= 6[7571]: segfault at 8 ip 00007ff94f17a7e8 sp 00007ffd51871f20 error 4 in ld-2.17.so[7ff94f16f000+22000] 13[12002]: segfault at 8 ip 00007f1ac9eab7e8 sp 00007ffdf34ac010 error 4 in ld-2.17.so[7f1ac9ea0000+22000] LustreError: 30659:0:(mdt_handler.c:746:mdt_pack_acl2body()) lustre-MDT0001: unable to read [0x240000405:0x339:0x0] ACL: rc = -2 LustreError: 15174:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0001: failed to get lu_attr of [0x240000403:0x1aa4:0x0]: rc = -2 LustreError: 15174:0:(mdd_object.c:3901:mdd_close()) Skipped 1 previous similar message 9[24075]: segfault at 8 ip 00007f59e80ce7e8 sp 00007ffc61d734d0 error 4 in ld-2.17.so[7f59e80c3000+22000] LustreError: 25889:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802caa8c138: inode [0x280000403:0x206a:0x0] mdc close failed: rc = -2 LustreError: 25889:0:(file.c:247:ll_close_inode_openhandle()) Skipped 43 previous similar messages LustreError: 5520:0:(llite_nfs.c:430:ll_dir_get_parent_fid()) lustre: failure inode [0x200000403:0x2355:0x0] get parent: rc = -2 LustreError: 5520:0:(llite_nfs.c:430:ll_dir_get_parent_fid()) Skipped 4 previous similar messages LustreError: 16719:0:(lov_object.c:1350:lov_layout_change()) lustre-clilov-ffff8802cf908008: cannot apply new layout on [0x280000404:0x2196:0x0] : rc = -5 LustreError: 16719:0:(lov_object.c:1350:lov_layout_change()) Skipped 12 previous similar messages Lustre: 15170:0:(lod_lov.c:1414:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0x2852:0x0] with magic=0xbd60bd0 Lustre: 15170:0:(lod_lov.c:1414:lod_parse_striping()) Skipped 163 previous similar messages Lustre: 15189:0:(mdt_reint.c:2460:mdt_reint_migrate()) lustre-MDT0002: [0x280000403:0x1:0x0]/18 is open, migrate only dentry LustreError: 7528:0:(statahead.c:2399:start_statahead_thread()) lustre: invalid pattern 0X1000. LustreError: 28930:0:(mdt_open.c:1302:mdt_cross_open()) lustre-MDT0002: [0x280000404:0x251b:0x0] doesn't exist!: rc = -14 17[14120]: segfault at 8 ip 00007fa2ec62c7e8 sp 00007ffd2d515650 error 4 in ld-2.17.so[7fa2ec621000+22000] 11[20299]: segfault at b6 ip 00000000000000b6 sp 00007ffe2bd555a8 error 14 in 11[400000+6000] LustreError: 6330:0:(lustre_lmv.h:500:lmv_is_sane()) unknown layout LMV: magic=0xcd40cd0 count=4 index=3 hash=crush:0x82000003 version=1 migrate_offset=3 migrate_hash=fnv_1a_64:2 pool= LustreError: 1927:0:(llite_nfs.c:430:ll_dir_get_parent_fid()) lustre: failure inode [0x280000404:0x2386:0x0] get parent: rc = -2 LustreError: 1927:0:(llite_nfs.c:430:ll_dir_get_parent_fid()) Skipped 1 previous similar message 3[26749]: segfault at 8 ip 00007fed37d577e8 sp 00007ffed0421560 error 4 in ld-2.17.so[7fed37d4c000+22000] 0[10855]: segfault at 8 ip 00007f33017087e8 sp 00007ffd1fa4e8b0 error 4 in ld-2.17.so[7f33016fd000+22000] 11[21694]: segfault at 8 ip 00007f8e7fa967e8 sp 00007ffe36b52240 error 4 in ld-2.17.so[7f8e7fa8b000+22000] LustreError: 14993:0:(ldlm_lockd.c:257:expired_lock_main()) ### lock callback timer expired after 101s: evicting client at 0@lo ns: mdt-lustre-MDT0000_UUID lock: ffff8802d34643c0/0xa9fda64c8d2b4f9c lrc: 3/0,0 mode: PR/PR res: [0x200000403:0x3204:0x0].0x0 bits 0x1b/0x0 rrc: 4 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0xa9fda64c8d2b4f80 expref: 604 pid: 30318 timeout: 1814 lvb_type: 0 LustreError: lustre-MDT0000-mdc-ffff8802cf908008: operation mds_reint to node 0@lo failed: rc = -107 Lustre: lustre-MDT0000-mdc-ffff8802cf908008: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: lustre-MDT0000-mdc-ffff8802cf908008: This client was evicted by lustre-MDT0000; in progress operations using this service will fail. LustreError: 24001:0:(file.c:6187:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000007:0x1:0x0] error: rc = -108 LustreError: 24001:0:(file.c:6187:ll_inode_revalidate_fini()) Skipped 14 previous similar messages LustreError: 24210:0:(mdc_request.c:1469:mdc_read_page()) lustre-MDT0000-mdc-ffff8802cf908008: [0x200000400:0x9e:0x0] lock enqueue fails: rc = -108 LustreError: 24210:0:(mdc_request.c:1469:mdc_read_page()) Skipped 6 previous similar messages Lustre: dir [0x200000403:0x3135:0x0] stripe 0 readdir failed: -108, directory is partially accessed! Lustre: Skipped 51 previous similar messages Lustre: lustre-MDT0000-mdc-ffff8802cf908008: Connection restored to 192.168.123.72@tcp (at 0@lo) LustreError: 30457:0:(mdd_orphans.c:261:mdd_orphan_delete()) lustre-MDD0001: could not delete orphan object [0x240000404:0x2834:0x0]: rc = -2 LustreError: 30457:0:(mdd_object.c:3955:mdd_close()) lustre-MDD0001: unable to delete [0x240000404:0x2834:0x0] from orphan list: rc = -2 LustreError: 959:0:(mdt_reint.c:2540:mdt_reint_migrate()) lustre-MDT0001: migrate [0x240000404:0x29ad:0x0]/1 failed: rc = -116 LustreError: 959:0:(mdt_reint.c:2540:mdt_reint_migrate()) Skipped 75 previous similar messages Lustre: 2567:0:(mdd_dir.c:4822:mdd_migrate_object()) lustre-MDD0000: [0x200000403:0x2:0x0]/16 is open, migrate only dentry Lustre: 2567:0:(mdd_dir.c:4822:mdd_migrate_object()) Skipped 87 previous similar messages LustreError: 5826:0:(mdt_handler.c:746:mdt_pack_acl2body()) lustre-MDT0002: unable to read [0x280000404:0x2aa2:0x0] ACL: rc = -2 14[9212]: segfault at 8 ip 00007fbfce52e7e8 sp 00007fff1107e530 error 4 in ld-2.17.so[7fbfce523000+22000] LustreError: 31751:0:(osd_io.c:901:osd_grow_blocksize()) object [0x340000400:0x4b7:0x0]: change block size4096 -> 1048576 error rc = -95 0[17599]: segfault at 8 ip 00007fd0691b77e8 sp 00007ffe9ef55ba0 error 4 in ld-2.17.so[7fd0691ac000+22000] LustreError: 5730:0:(mdt_handler.c:746:mdt_pack_acl2body()) lustre-MDT0002: unable to read [0x280000404:0x29b3:0x0] ACL: rc = -2 LustreError: 5854:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0002: [0x280000404:0x1c66:0x0] migrate mdt count mismatch 3 != 1 LustreError: 5854:0:(mdt_xattr.c:402:mdt_dir_layout_update()) Skipped 13 previous similar messages LustreError: 31988:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000405:0x1a27:0x0]: rc = -5 LustreError: 31988:0:(lcommon_cl.c:188:cl_file_inode_init()) Skipped 239 previous similar messages LustreError: 31988:0:(llite_lib.c:3769:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 31988:0:(llite_lib.c:3769:ll_prep_inode()) Skipped 239 previous similar messages LustreError: 31186:0:(mdd_dir.c:4743:mdd_migrate_cmd_check()) lustre-MDD0001: '0' migration was interrupted, run 'lfs migrate -m 0 -c 3 -H crush 0' to finish migration: rc = -1 LustreError: 31186:0:(mdd_dir.c:4743:mdd_migrate_cmd_check()) Skipped 27 previous similar messages LustreError: 661:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 10 [0x200000406:0x336:0x0] inode@0000000000000000: rc = -5 LustreError: 661:0:(statahead.c:825:ll_statahead_interpret_work()) Skipped 9 previous similar messages 18[26585]: segfault at 8 ip 00007f63f5fe47e8 sp 00007ffe2a554a40 error 4 in ld-2.17.so[7f63f5fd9000+22000] LustreError: 30770:0:(vvp_io.c:1905:vvp_io_init()) lustre: refresh file layout [0x280000403:0x31b3:0x0] error -5. LustreError: 30770:0:(vvp_io.c:1905:vvp_io_init()) Skipped 8 previous similar messages 0[23752]: segfault at 8 ip 00007f0cd85ae7e8 sp 00007ffd514fd550 error 4 in ld-2.17.so[7f0cd85a3000+22000] LustreError: 15439:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802caa8c138: inode [0x200000406:0x60f:0x0] mdc close failed: rc = -2 LustreError: 15439:0:(file.c:247:ll_close_inode_openhandle()) Skipped 54 previous similar messages 13[1470]: segfault at 8 ip 00007f46a14317e8 sp 00007ffd513bf1a0 error 4 in ld-2.17.so[7f46a1426000+22000] 19[7684]: segfault at 8 ip 00007f36d0d637e8 sp 00007ffce71bcda0 error 4 in ld-2.17.so[7f36d0d58000+22000] Lustre: 30642:0:(lod_lov.c:1414:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000406:0x515:0x0] with magic=0xbd60bd0 Lustre: 30642:0:(lod_lov.c:1414:lod_parse_striping()) Skipped 97 previous similar messages LustreError: 5733:0:(lov_object.c:1350:lov_layout_change()) lustre-clilov-ffff8802cf908008: cannot apply new layout on [0x280000404:0x31d9:0x0] : rc = -5 LustreError: 5733:0:(lov_object.c:1350:lov_layout_change()) Skipped 15 previous similar messages 19[28414]: segfault at 0 ip 0000000000403e5f sp 00007fff0e472080 error 6 in 19[400000+6000] 5[29069]: segfault at 8 ip 00007fb27aeb97e8 sp 00007ffc6cc39860 error 4 in ld-2.17.so[7fb27aeae000+22000] LustreError: 31702:0:(mdt_open.c:1703:mdt_reint_open()) lustre-MDT0000: name '17' present, but FID [0x200000406:0x266:0x0] is invalid LustreError: 28484:0:(mdt_open.c:1703:mdt_reint_open()) lustre-MDT0000: name '17' present, but FID [0x200000406:0x266:0x0] is invalid LustreError: 19216:0:(llite_lib.c:1872:ll_update_lsm_md()) lustre: [0x200000406:0x9a5:0x0] dir layout mismatch: LustreError: 19216:0:(llite_lib.c:1872:ll_update_lsm_md()) Skipped 1 previous similar message LustreError: 19216:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=2 count=3 index=0 hash=crush:0x2000003 max_inherit=0 max_inherit_rr=0 version=2 migrate_offset=0 migrate_hash=invalid:0 pool= LustreError: 19216:0:(lustre_lmv.h:167:lmv_stripe_object_dump()) stripe[0] [0x200000400:0xd5:0x0] LustreError: 19216:0:(lustre_lmv.h:167:lmv_stripe_object_dump()) Skipped 13 previous similar messages LustreError: 19216:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=1 count=5 index=0 hash=crush:0x82000003 max_inherit=0 max_inherit_rr=0 version=1 migrate_offset=3 migrate_hash=crush:2000003 pool= LustreError: 10785:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=2 count=3 index=0 hash=crush:0x2000003 max_inherit=0 max_inherit_rr=0 version=2 migrate_offset=0 migrate_hash=invalid:0 pool= LustreError: 10785:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=1 count=5 index=0 hash=crush:0x82000003 max_inherit=0 max_inherit_rr=0 version=1 migrate_offset=3 migrate_hash=crush:2000003 pool= LustreError: 14993:0:(ldlm_lockd.c:257:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0002_UUID lock: ffff8802d51da5c0/0xa9fda64c8d6383cf lrc: 3/0,0 mode: PR/PR res: [0x280000404:0x2ff1:0x0].0x0 bits 0x1b/0x0 rrc: 4 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0xa9fda64c8d6383b3 expref: 543 pid: 31433 timeout: 2303 lvb_type: 0 LustreError: 25265:0:(ldlm_lockd.c:2550:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1748029959 with bad export cookie 12249129409220667197 LustreError: lustre-MDT0002-mdc-ffff8802caa8c138: operation mds_reint to node 0@lo failed: rc = -107 Lustre: lustre-MDT0002-mdc-ffff8802caa8c138: Connection to lustre-MDT0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: lustre-MDT0002-mdc-ffff8802caa8c138: This client was evicted by lustre-MDT0002; in progress operations using this service will fail. LustreError: 29393:0:(llite_lib.c:2023:ll_md_setattr()) md_setattr fails: rc = -108 Lustre: lustre-MDT0002-mdc-ffff8802caa8c138: Connection restored to 192.168.123.72@tcp (at 0@lo) 12[2501]: segfault at 8 ip 00007fee354eb7e8 sp 00007ffc6c728760 error 4 in ld-2.17.so[7fee354e0000+22000] LustreError: 14471:0:(mdt_open.c:1703:mdt_reint_open()) lustre-MDT0000: name '17' present, but FID [0x200000406:0x266:0x0] is invalid LustreError: 14993:0:(ldlm_lockd.c:257:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0001_UUID lock: ffff8802a02d0400/0xa9fda64c8d693159 lrc: 3/0,0 mode: PR/PR res: [0x240000404:0x3730:0x0].0x0 bits 0x1b/0x0 rrc: 4 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0xa9fda64c8d69312f expref: 598 pid: 30633 timeout: 2428 lvb_type: 0 LustreError: lustre-MDT0001-mdc-ffff8802caa8c138: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: 11785:0:(ldlm_lockd.c:2550:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1748030084 with bad export cookie 12249129409220667190 Lustre: lustre-MDT0001-mdc-ffff8802caa8c138: Connection to lustre-MDT0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: lustre-MDT0001-mdc-ffff8802caa8c138: This client was evicted by lustre-MDT0001; in progress operations using this service will fail. LustreError: 9614:0:(file.c:6187:ll_inode_revalidate_fini()) lustre: revalidate FID [0x240000404:0x3730:0x0] error: rc = -5 LustreError: 9614:0:(file.c:6187:ll_inode_revalidate_fini()) Skipped 21 previous similar messages Lustre: lustre-MDT0001-mdc-ffff8802caa8c138: Connection restored to 192.168.123.72@tcp (at 0@lo) Lustre: 2557:0:(mdd_dir.c:4822:mdd_migrate_object()) lustre-MDD0001: [0x240000403:0x1:0x0]/13 is open, migrate only dentry Lustre: 2557:0:(mdd_dir.c:4822:mdd_migrate_object()) Skipped 78 previous similar messages LustreError: 31143:0:(mdt_reint.c:2540:mdt_reint_migrate()) lustre-MDT0000: migrate [0x280000404:0x3336:0x0]/16 failed: rc = -2 LustreError: 31143:0:(mdt_reint.c:2540:mdt_reint_migrate()) Skipped 71 previous similar messages | Link to test |
racer test 1: racer on clients: centos-30.localnet DURATION=2700 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 PGD 2c3d0a067 PUD 2eb37a067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) lustre(OE) osp(OE) ofd(OE) lod(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) jbd2 mbcache lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) crc32_generic lnet(OE) dm_flakey dm_mod libcfs(OE) crc_t10dif crct10dif_generic crct10dif_common pcspkr virtio_balloon virtio_console i2c_piix4 ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm drm_panel_orientation_quirks ata_piix floppy serio_raw virtio_blk i2c_core libata CPU: 12 PID: 28326 Comm: ll_sa_28292 Kdump: loaded Tainted: P OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.16.0-3.module_el8.7.0+1218+f626c2ff 04/01/2014 task: ffff8802eb2d5c40 ti: ffff8802ab22c000 task.ti: ffff8802ab22c000 RIP: 0010:[<ffffffff813f502c>] [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 RSP: 0018:ffff8802ab22fd78 EFLAGS: 00010296 RAX: ffff88029f394390 RBX: 0000000000000008 RCX: 000000010026001e RDX: 0000000000000026 RSI: ffff88029f3946c0 RDI: 0000000000000008 RBP: ffff8802ab22fd88 R08: ffff8800a74f0008 R09: 0000000000000000 R10: ffff8800a74f0358 R11: ffff8800a74f0008 R12: 0000000000000000 R13: ffff8802ea9a5c40 R14: ffff88029f3946c0 R15: ffff8800a74f0008 FS: 0000000000000000(0000) GS:ffff880331d00000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 00000002c3f26000 CR4: 00000000000007e0 Call Trace: [<ffffffffa1110111>] ll_statahead_thread+0xdb1/0x2000 [lustre] [<ffffffff810df6e2>] ? pick_next_task_fair+0x702/0x8a0 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e329e>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffff810c834d>] ? finish_task_switch+0x5d/0x1b0 [<ffffffffa110f360>] ? ll_statahead_handle.constprop.28+0x1c0/0x1c0 [lustre] [<ffffffff810ba114>] kthread+0xe4/0xf0 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 | Lustre: 14169:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff8802eaa31e40 x1832383867562880/t4294967671(0) o101->b2e44554-0b11-4228-809b-933680fca1f4@0@lo:167/0 lens 376/840 e 0 to 0 dl 1747497497 ref 1 fl Interpret:H/602/0 rc 0/0 job:'dd.0' uid:0 gid:0 projid:0 Lustre: 9269:0:(osd_internal.h:1332:osd_trans_exec_op()) lustre-OST0002: opcode 7: before 516 < left 618, rollback = 7 Lustre: 9269:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 9269:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 9269:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 9269:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 9269:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 14468:0:(osd_internal.h:1332:osd_trans_exec_op()) lustre-OST0002: opcode 7: before 516 < left 618, rollback = 7 Lustre: 14468:0:(osd_internal.h:1332:osd_trans_exec_op()) Skipped 1 previous similar message Lustre: 14468:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 14468:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 14468:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 14468:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 14468:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 14468:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 14468:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 14468:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 14468:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 14468:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 14468:0:(osd_internal.h:1332:osd_trans_exec_op()) lustre-OST0002: opcode 7: before 516 < left 618, rollback = 7 Lustre: 14468:0:(osd_internal.h:1332:osd_trans_exec_op()) Skipped 1 previous similar message Lustre: 14468:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 14468:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 14468:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 14468:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 14468:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 14468:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 14468:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 14468:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 14468:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 14468:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 9268:0:(osd_internal.h:1332:osd_trans_exec_op()) lustre-OST0002: opcode 7: before 516 < left 618, rollback = 7 Lustre: 9268:0:(osd_internal.h:1332:osd_trans_exec_op()) Skipped 1 previous similar message Lustre: 9268:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 9268:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 9268:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 9268:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 9268:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 9268:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 9268:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 9268:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 9268:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 9268:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 1 previous similar message LustreError: 20547:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a9ea2e98: inode [0x200000402:0x1d6:0x0] mdc close failed: rc = -13 Lustre: 9270:0:(osd_internal.h:1332:osd_trans_exec_op()) lustre-OST0002: opcode 7: before 516 < left 618, rollback = 7 Lustre: 9270:0:(osd_internal.h:1332:osd_trans_exec_op()) Skipped 5 previous similar messages Lustre: 9270:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 9270:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 9270:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 9270:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 9270:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 9270:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 9270:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 9270:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 9270:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 9270:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 5 previous similar messages LustreError: 21005:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a9ea2e98: inode [0x200000401:0x21e:0x0] mdc close failed: rc = -13 Lustre: 9268:0:(osd_internal.h:1332:osd_trans_exec_op()) lustre-OST0000: opcode 7: before 516 < left 618, rollback = 7 Lustre: 9268:0:(osd_internal.h:1332:osd_trans_exec_op()) Skipped 5 previous similar messages Lustre: 9268:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 9268:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 9268:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 9268:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 9268:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 9268:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 9268:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 9268:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 9268:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 9268:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 5 previous similar messages 7[25390]: segfault at 8 ip 00007f72ff5b87e8 sp 00007ffd6fa814d0 error 4 in ld-2.17.so[7f72ff5ad000+22000] LustreError: 25781:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a9ea2e98: inode [0x200000402:0x365:0x0] mdc close failed: rc = -13 Lustre: 9269:0:(osd_internal.h:1332:osd_trans_exec_op()) lustre-OST0000: opcode 7: before 516 < left 618, rollback = 7 Lustre: 9269:0:(osd_internal.h:1332:osd_trans_exec_op()) Skipped 17 previous similar messages Lustre: 9269:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 9269:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 17 previous similar messages Lustre: 9269:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 9269:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 17 previous similar messages Lustre: 9269:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 9269:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 17 previous similar messages Lustre: 9269:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 9269:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 17 previous similar messages Lustre: 9269:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 9269:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 17 previous similar messages traps: 8[29990] trap invalid opcode ip:40543a sp:7ffc06ecc9c8 error:0 in 8[400000+6000] 17[30437]: segfault at 8 ip 00007f8a160d77e8 sp 00007ffdb9ffcad0 error 4 in ld-2.17.so[7f8a160cc000+22000] Lustre: 20714:0:(osd_internal.h:1332:osd_trans_exec_op()) lustre-OST0002: opcode 7: before 516 < left 540, rollback = 7 Lustre: 20714:0:(osd_internal.h:1332:osd_trans_exec_op()) Skipped 33 previous similar messages Lustre: 20714:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 20714:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 33 previous similar messages Lustre: 20714:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 20714:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 33 previous similar messages Lustre: 20714:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/540/0, punch: 0/0/0, quota 4/150/0 Lustre: 20714:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 33 previous similar messages Lustre: 20714:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 20714:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 33 previous similar messages Lustre: 20714:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 20714:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 33 previous similar messages 8[1305]: segfault at 0 ip 0000000000404470 sp 00007ffee8732be8 error 6 in 10[400000+6000] 8[1926]: segfault at 8 ip 00007f7aaf7007e8 sp 00007ffda0763d20 error 4 in ld-2.17.so[7f7aaf6f5000+22000] Lustre: 16371:0:(lod_lov.c:1414:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000402:0x606:0x0] with magic=0xbd60bd0 13[4255]: segfault at 8 ip 00007fc14da047e8 sp 00007ffd39fc5fb0 error 4 in ld-2.17.so[7fc14d9f9000+22000] 5[4346]: segfault at 0 ip 0000000000403e5f sp 00007ffd32cd2270 error 6 in 5[400000+6000] LustreError: 4346:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a9deb7e8: inode [0x200000401:0x698:0x0] mdc close failed: rc = -13 13[7661]: segfault at 8 ip 00007f9ef2b0f7e8 sp 00007ffe1cbe7510 error 4 in ld-2.17.so[7f9ef2b04000+22000] Lustre: 8538:0:(lod_lov.c:1414:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000402:0x85e:0x0] with magic=0xbd60bd0 Lustre: 8538:0:(lod_lov.c:1414:lod_parse_striping()) Skipped 1 previous similar message Lustre: 20714:0:(osd_internal.h:1332:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 618, rollback = 7 Lustre: 20714:0:(osd_internal.h:1332:osd_trans_exec_op()) Skipped 43 previous similar messages Lustre: 20714:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 20714:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 43 previous similar messages Lustre: 20714:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 20714:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 43 previous similar messages Lustre: 20714:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 20714:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 43 previous similar messages Lustre: 20714:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 20714:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 43 previous similar messages Lustre: 20714:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 20714:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 43 previous similar messages 2[14925]: segfault at 8 ip 00007f76f1f277e8 sp 00007ffed7589f40 error 4 in ld-2.17.so[7f76f1f1c000+22000] 3[17326]: segfault at 406000 ip 0000000000406000 sp 00007ffccb9d6fb8 error 14 in 3[606000+1000] 14[18926]: segfault at 8 ip 00007f090c1ab7e8 sp 00007ffeb3c62f70 error 4 in ld-2.17.so[7f090c1a0000+22000] traps: 9[20910] trap invalid opcode ip:4058fa sp:7fffccbd0a58 error:0 in 12[400000+6000] 3[26731]: segfault at 8 ip 00007fd3c63b87e8 sp 00007ffdabd7a9f0 error 4 in ld-2.17.so[7fd3c63ad000+22000] Lustre: 14848:0:(lod_lov.c:1414:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000402:0xe20:0x0] with magic=0xbd60bd0 Lustre: 14848:0:(lod_lov.c:1414:lod_parse_striping()) Skipped 3 previous similar messages LustreError: 32624:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a9ea2e98: inode [0x200000401:0xfab:0x0] mdc close failed: rc = -13 Lustre: 14383:0:(osd_internal.h:1332:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 516 < left 618, rollback = 7 Lustre: 14383:0:(osd_internal.h:1332:osd_trans_exec_op()) Skipped 151 previous similar messages Lustre: 14383:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 14383:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 151 previous similar messages Lustre: 14383:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 14383:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 151 previous similar messages Lustre: 14383:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 14383:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 151 previous similar messages Lustre: 14383:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 14383:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 151 previous similar messages Lustre: 14383:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 14383:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 151 previous similar messages Lustre: 8538:0:(lod_lov.c:1414:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000402:0x10e4:0x0] with magic=0xbd60bd0 Lustre: 8538:0:(lod_lov.c:1414:lod_parse_striping()) Skipped 1 previous similar message LustreError: 9100:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a9deb7e8: inode [0x200000401:0x1262:0x0] mdc close failed: rc = -13 LustreError: 9100:0:(file.c:247:ll_close_inode_openhandle()) Skipped 3 previous similar messages 7[12511]: segfault at 8 ip 00007f3c7b8a17e8 sp 00007ffc422ebc60 error 4 in ld-2.17.so[7f3c7b896000+22000] Lustre: 11927:0:(lod_lov.c:1414:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000402:0x13dd:0x0] with magic=0xbd60bd0 Lustre: 11927:0:(lod_lov.c:1414:lod_parse_striping()) Skipped 1 previous similar message 5[17923]: segfault at 8 ip 00007f4b301837e8 sp 00007fff0913f810 error 4 in ld-2.17.so[7f4b30178000+22000] 5[19198]: segfault at 8 ip 00007f9ee77bd7e8 sp 00007ffe334ce5b0 error 4 in ld-2.17.so[7f9ee77b2000+22000] 2[24304]: segfault at 8 ip 00007f0fae5347e8 sp 00007fff9b643450 error 4 in ld-2.17.so[7f0fae529000+22000] 1[28530]: segfault at 8 ip 00007fe4360817e8 sp 00007ffc60374c30 error 4 in ld-2.17.so[7fe436076000+22000] 17[28719]: segfault at 8 ip 00007ff6c246d7e8 sp 00007ffdc78b2f30 error 4 in ld-2.17.so[7ff6c2462000+22000] 17[28779]: segfault at 8 ip 00007f0e7a1fe7e8 sp 00007fff47d8c9c0 error 4 in ld-2.17.so[7f0e7a1f3000+22000] 5[31151]: segfault at 8 ip 00007f495b1bf2fc sp 00007ffe063b0ac0 error 4 in ld-2.17.so[7f495b1b2000+22000] LustreError: 1320:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a9ea2e98: inode [0x200000402:0x1a4e:0x0] mdc close failed: rc = -13 LustreError: 1320:0:(file.c:247:ll_close_inode_openhandle()) Skipped 2 previous similar messages 10[4017]: segfault at 0 ip 00000000004043f0 sp 00007ffed44f2798 error 6 in 10[400000+6000] Lustre: 5126:0:(lod_lov.c:1414:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000401:0x1c1b:0x0] with magic=0xbd60bd0 Lustre: 5126:0:(lod_lov.c:1414:lod_parse_striping()) Skipped 1 previous similar message 14[9476]: segfault at 8 ip 00007fa09a6107e8 sp 00007fff59dddc20 error 4 in ld-2.17.so[7fa09a605000+22000] LustreError: 9713:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a9deb7e8: inode [0x200000402:0x1cdf:0x0] mdc close failed: rc = -13 Lustre: 16367:0:(lod_lov.c:1414:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000401:0x1db0:0x0] with magic=0xbd60bd0 Lustre: 16367:0:(lod_lov.c:1414:lod_parse_striping()) Skipped 1 previous similar message 17[11621]: segfault at 406000 ip 0000000000406000 sp 00007ffebb840db8 error 14 in 17[606000+1000] Lustre: 20714:0:(osd_internal.h:1332:osd_trans_exec_op()) lustre-OST0000: opcode 7: before 516 < left 618, rollback = 7 Lustre: 20714:0:(osd_internal.h:1332:osd_trans_exec_op()) Skipped 279 previous similar messages Lustre: 20714:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 20714:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 279 previous similar messages Lustre: 20714:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 20714:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 279 previous similar messages Lustre: 20714:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 20714:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 279 previous similar messages Lustre: 20714:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 20714:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 279 previous similar messages Lustre: 20714:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 20714:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 279 previous similar messages 3[29568]: segfault at 8 ip 00007f3d8dda57e8 sp 00007ffe4d6ca4d0 error 4 in ld-2.17.so[7f3d8dd9a000+22000] 5[3341]: segfault at 8 ip 00007f6892f537e8 sp 00007ffe18fe7920 error 4 in ld-2.17.so[7f6892f48000+22000] 16[5083]: segfault at 0 ip 0000000000403e5f sp 00007ffecd4a71d0 error 6 in 16 (deleted)[400000+6000] LustreError: 6776:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a9ea2e98: inode [0x200000402:0x2656:0x0] mdc close failed: rc = -13 10[8294]: segfault at 8 ip 00007f598e5257e8 sp 00007ffd0e7e4240 error 4 in ld-2.17.so[7f598e51a000+22000] 16[9919]: segfault at 8 ip 00007fa5325197e8 sp 00007ffd94ff7b00 error 4 in ld-2.17.so[7fa53250e000+22000] Lustre: 14169:0:(lod_lov.c:1414:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000401:0x2a3c:0x0] with magic=0xbd60bd0 Lustre: 14169:0:(lod_lov.c:1414:lod_parse_striping()) Skipped 1 previous similar message 5[17093]: segfault at 8 ip 00007fcfd278b7e8 sp 00007ffeee894cd0 error 4 in ld-2.17.so[7fcfd2780000+22000] 11[23234]: segfault at 8 ip 00007f06489b97e8 sp 00007ffe846573c0 error 4 in ld-2.17.so[7f06489ae000+22000] 5[24208]: segfault at 8 ip 00007fd3e6d307e8 sp 00007fffde944cd0 error 4 in ld-2.17.so[7fd3e6d25000+22000] 5[31558]: segfault at 0 ip 00000000004043f0 sp 00007ffdadc03568 error 6 in 5[400000+6000] 9[32584]: segfault at 8 ip 00007f49990577e8 sp 00007ffe0bb14170 error 4 in ld-2.17.so[7f499904c000+22000] 6[6445]: segfault at 406000 ip 0000000000406000 sp 00007fffc19ce648 error 14 in 6[606000+1000] 2[6842]: segfault at 0 ip 0000000000403e5f sp 00007fff29a16480 error 6 in 2[400000+6000] 0[7932]: segfault at 8 ip 00007ff6a3acc7e8 sp 00007ffc1c33c2e0 error 4 in ld-2.17.so[7ff6a3ac1000+22000] 2[8579]: segfault at 0 ip 0000000000403e5f sp 00007ffef3334040 error 6 in 12[400000+6000] LustreError: 8982:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a9deb7e8: inode [0x200000401:0x3288:0x0] mdc close failed: rc = -13 LustreError: 8982:0:(file.c:247:ll_close_inode_openhandle()) Skipped 1 previous similar message 18[15974]: segfault at 8 ip 00007fa908a6d7e8 sp 00007ffc2eca0460 error 4 in ld-2.17.so[7fa908a62000+22000] Lustre: mdt_io00_002: service thread pid 8556 was inactive for 40.088 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 8556, comm: mdt_io00_002 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] Lustre: mdt_out00_000: service thread pid 8543 was inactive for 40.079 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_check_lock+0xec/0x3c0 [mdt] [<0>] mdt_reint_rename+0x148f/0x3950 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x73c/0xbb0 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x257/0xcd0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 8545, comm: mdt_out00_001 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_pdo_lock+0x369/0x750 [mdt] [<0>] mdt_parent_lock+0x76/0x2a0 [mdt] [<0>] mdt_getattr_name_lock+0x1844/0x2cc0 [mdt] [<0>] mdt_intent_getattr+0x2cc/0x4e0 [mdt] [<0>] mdt_intent_opc.constprop.74+0x211/0xc60 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x397/0x980 [ptlrpc] [<0>] ldlm_handle_enqueue+0x547/0x1900 [ptlrpc] [<0>] mdt_batch_getattr+0x6c/0x190 [mdt] [<0>] mdt_batch+0x465/0x195b [mdt] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x257/0xcd0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 13821, comm: mdt_out00_002 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_pdo_lock+0x369/0x750 [mdt] [<0>] mdt_parent_lock+0x76/0x2a0 [mdt] [<0>] mdt_getattr_name_lock+0x1844/0x2cc0 [mdt] [<0>] mdt_intent_getattr+0x2cc/0x4e0 [mdt] [<0>] mdt_intent_opc.constprop.74+0x211/0xc60 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x397/0x980 [ptlrpc] [<0>] ldlm_handle_enqueue+0x547/0x1900 [ptlrpc] [<0>] mdt_batch_getattr+0x6c/0x190 [mdt] [<0>] mdt_batch+0x465/0x195b [mdt] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x257/0xcd0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: lustre-OST0002-osc-ffff8800a9deb7e8: disconnect after 21s idle Lustre: lustre-OST0003-osc-ffff8800a9deb7e8: disconnect after 21s idle Lustre: Skipped 1 previous similar message LustreError: 8318:0:(ldlm_lockd.c:257:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0000_UUID lock: ffff8802e0ca07c0/0xfa2dee2eafd111a1 lrc: 3/0,0 mode: PR/PR res: [0x200000401:0x35d4:0x0].0x0 bits 0x13/0x0 rrc: 6 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0xfa2dee2eafd11035 expref: 1198 pid: 8545 timeout: 1034 lvb_type: 0 LustreError: 8556:0:(client.c:1373:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802c4e91940 x1832384008059264/t0(0) o105->lustre-MDT0000@0@lo:15/16 lens 336/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' uid:4294967295 gid:4294967295 projid:4294967295 Lustre: mdt_io00_002: service thread pid 8556 completed after 100.129s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_017: service thread pid 11372 completed after 100.070s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 24846:0:(ldlm_lockd.c:1447:ldlm_handle_enqueue()) ### lock on destroyed export ffff8800a9dbb7e8 ns: mdt-lustre-MDT0000_UUID lock: ffff8800a56c8f40/0xfa2dee2eafd1151a lrc: 3/0,0 mode: PR/PR res: [0x200000401:0x35d4:0x0].0x0 bits 0x13/0x0 rrc: 3 type: IBT gid 0 flags: 0x50200000000000 nid: 0@lo remote: 0xfa2dee2eafd1150c expref: 446 pid: 24846 timeout: 0 lvb_type: 0 Lustre: mdt00_012: service thread pid 24846 completed after 100.036s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: lustre-MDT0000-mdc-ffff8800a9deb7e8: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-MDT0000-mdc-ffff8800a9deb7e8: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: mdt_out00_001: service thread pid 8545 completed after 100.108s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_out00_002: service thread pid 13821 completed after 100.144s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_out00_000: service thread pid 8543 completed after 100.130s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: lustre-MDT0000-mdc-ffff8800a9deb7e8: This client was evicted by lustre-MDT0000; in progress operations using this service will fail. LustreError: 17918:0:(file.c:6143:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000401:0x35d4:0x0] error: rc = -5 LustreError: 18153:0:(mdc_request.c:1469:mdc_read_page()) lustre-MDT0000-mdc-ffff8800a9deb7e8: [0x200000401:0x1:0x0] lock enqueue fails: rc = -5 LustreError: 18348:0:(ldlm_resource.c:981:ldlm_resource_complain()) lustre-MDT0000-mdc-ffff8800a9deb7e8: namespace resource [0x200000401:0x1:0x0].0x0 (ffff8802d2c7f240) refcount nonzero (2) after lock cleanup; forcing cleanup. Lustre: lustre-MDT0000-mdc-ffff8800a9deb7e8: Connection restored to 192.168.123.32@tcp (at 0@lo) 1[26481]: segfault at 8 ip 00007f547c18a7e8 sp 00007ffefaa51320 error 4 in ld-2.17.so[7f547c17f000+22000] 17[27247]: segfault at 8 ip 00007fde654187e8 sp 00007ffdc451b020 error 4 in ld-2.17.so[7fde6540d000+22000] 0[29976]: segfault at 8 ip 00007f0d294497e8 sp 00007ffe4bcaf920 error 4 in ld-2.17.so[7f0d2943e000+22000] 13[8321]: segfault at 8 ip 00007f719a4c77e8 sp 00007ffce90343d0 error 4 in ld-2.17.so[7f719a4bc000+22000] LustreError: 16655:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a9ea2e98: inode [0x200000402:0x3ebe:0x0] mdc close failed: rc = -13 LustreError: 16655:0:(file.c:247:ll_close_inode_openhandle()) Skipped 5 previous similar messages Lustre: 14468:0:(osd_internal.h:1332:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 516 < left 570, rollback = 7 Lustre: 14468:0:(osd_internal.h:1332:osd_trans_exec_op()) Skipped 313 previous similar messages Lustre: 14468:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 14468:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 313 previous similar messages Lustre: 14468:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 14468:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 313 previous similar messages Lustre: 14468:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/570/0, punch: 0/0/0, quota 1/3/0 Lustre: 14468:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 313 previous similar messages Lustre: 14468:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 14468:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 313 previous similar messages Lustre: 14468:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 14468:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 313 previous similar messages Lustre: 16371:0:(lod_lov.c:1414:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0xb7c:0x0] with magic=0xbd60bd0 Lustre: 16371:0:(lod_lov.c:1414:lod_parse_striping()) Skipped 3 previous similar messages 4[25238]: segfault at 8 ip 00007f154e60f7e8 sp 00007ffee3179180 error 4 in ld-2.17.so[7f154e604000+22000] 16[30852]: segfault at 8 ip 00007f40583567e8 sp 00007ffcc7f3abb0 error 4 in ld-2.17.so[7f405834b000+22000] 13[30854]: segfault at 8 ip 00007ff7e338c7e8 sp 00007ffd63aa6bb0 error 4 in ld-2.17.so[7ff7e3381000+22000] 13[30918]: segfault at 8 ip 00007f1a84bcb7e8 sp 00007ffdbbe8e0e0 error 4 in ld-2.17.so[7f1a84bc0000+22000] mkdir (31663) used greatest stack depth: 9968 bytes left 8[705]: segfault at 0 ip 0000000000401d60 sp 00007ffc8b116c88 error 6 in 8[400000+6000] 5[1262]: segfault at 8 ip 00007fe861bd57e8 sp 00007ffc9f81d130 error 4 in ld-2.17.so[7fe861bca000+22000] 17[11181]: segfault at 8 ip 00007fdbef8f67e8 sp 00007fffcbc5fa80 error 4 in ld-2.17.so[7fdbef8eb000+22000] 19[12714]: segfault at 8 ip 00007f71619717e8 sp 00007ffda6008860 error 4 in ld-2.17.so[7f7161966000+22000] 18[15781]: segfault at 8 ip 00007faeefea27e8 sp 00007ffcacb831f0 error 4 in ld-2.17.so[7faeefe97000+22000] 14[23800]: segfault at 8 ip 00007fca6629f7e8 sp 00007ffd55d5d3a0 error 4 in ld-2.17.so[7fca66294000+22000] 13[27096]: segfault at 8 ip 00007fce6b2ff7e8 sp 00007ffe44747040 error 4 in ld-2.17.so[7fce6b2f4000+22000] 7[664]: segfault at 8 ip 00007fa346aee7e8 sp 00007ffc5855f460 error 4 in ld-2.17.so[7fa346ae3000+22000] 12[5405]: segfault at 8 ip 00007fdfad6667e8 sp 00007ffd23207500 error 4 in ld-2.17.so[7fdfad65b000+22000] 12[6134]: segfault at 0 ip 0000000000403e5f sp 00007ffd27352250 error 6 in 12[400000+6000] Lustre: 14848:0:(lod_lov.c:1414:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0x1dfa:0x0] with magic=0xbd60bd0 Lustre: 14848:0:(lod_lov.c:1414:lod_parse_striping()) Skipped 5 previous similar messages 18[14016]: segfault at 8 ip 00007ff7341657e8 sp 00007ffcdeca4a50 error 4 in ld-2.17.so[7ff73415a000+22000] 13[15553]: segfault at 0 ip 00000000004043f0 sp 00007ffe805cc098 error 6 in 14[400000+6000] 2[17682]: segfault at 1c ip 00000000004017b0 sp 00007ffdef699940 error 6 in 13 (deleted)[400000+6000] 1[22670]: segfault at 0 ip 0000000000403e5f sp 00007fff41873bc0 error 6 in 14[400000+6000] 18[22937]: segfault at 8 ip 00007ff3b231d7e8 sp 00007fffda09ec50 error 4 in ld-2.17.so[7ff3b2312000+22000] 17[26950]: segfault at 8 ip 00007f83982c07e8 sp 00007fff83665790 error 4 in ld-2.17.so[7f83982b5000+22000] 5[31416]: segfault at 8 ip 00007f1a6434f7e8 sp 00007ffdcdc70b80 error 4 in ld-2.17.so[7f1a64344000+22000] 9[17148]: segfault at 8 ip 00007effea5487e8 sp 00007ffc057cbbd0 error 4 in ld-2.17.so[7effea53d000+22000] 2[22655]: segfault at 8 ip 00007fe52904f7e8 sp 00007ffe05f6bcd0 error 4 in ld-2.17.so[7fe529044000+22000] 3[24081]: segfault at 0 ip (null) sp 00007ffda2fc2e38 error 14 in 3[400000+6000] 16[25402]: segfault at 8 ip 00007fdbda4ed7e8 sp 00007fffe0e24460 error 4 in ld-2.17.so[7fdbda4e2000+22000] 8[26472]: segfault at 8 ip 00007f953462d7e8 sp 00007ffd8d69ff60 error 4 in ld-2.17.so[7f9534622000+22000] 19[30097]: segfault at 8 ip 00007f84d512a7e8 sp 00007ffc9bc69fc0 error 4 in ld-2.17.so[7f84d511f000+22000] 10[5827]: segfault at 8 ip 00007f0a0db4c7e8 sp 00007ffe0387b0f0 error 4 in ld-2.17.so[7f0a0db41000+22000] 2[9548]: segfault at 8 ip 00007f183ab0a7e8 sp 00007ffeabb46630 error 4 in ld-2.17.so[7f183aaff000+22000] 4[10773]: segfault at 8 ip 00007fe015aa97e8 sp 00007ffd93d04290 error 4 in ld-2.17.so[7fe015a9e000+22000] 2[11745]: segfault at 8 ip 00007ff440ff27e8 sp 00007ffd66fa2d40 error 4 in ld-2.17.so[7ff440fe7000+22000] 12[13484]: segfault at 8 ip 00007f8106bc57e8 sp 00007fffb7c0a260 error 4 in ld-2.17.so[7f8106bba000+22000] Lustre: 9269:0:(osd_internal.h:1332:osd_trans_exec_op()) lustre-OST0000: opcode 7: before 516 < left 618, rollback = 7 Lustre: 9269:0:(osd_internal.h:1332:osd_trans_exec_op()) Skipped 475 previous similar messages Lustre: 9269:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 9269:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 475 previous similar messages Lustre: 9269:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 9269:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 475 previous similar messages Lustre: 9269:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 9269:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 475 previous similar messages Lustre: 9269:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 9269:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 475 previous similar messages Lustre: 9269:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 9269:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 475 previous similar messages 11[17414]: segfault at 8 ip 00007fbd529a97e8 sp 00007ffe29831460 error 4 in ld-2.17.so[7fbd5299e000+22000] Lustre: 16371:0:(lod_lov.c:1414:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0x3578:0x0] with magic=0xbd60bd0 Lustre: 16371:0:(lod_lov.c:1414:lod_parse_striping()) Skipped 7 previous similar messages 7[31935]: segfault at 8 ip 00007fbafd77e7e8 sp 00007fff1cb39e80 error 4 in ld-2.17.so[7fbafd773000+22000] 4[31946]: segfault at 8 ip 00007f5c29efa7e8 sp 00007fff2d8dad00 error 4 in ld-2.17.so[7f5c29eef000+22000] LustreError: 31946:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a9deb7e8: inode [0x200000402:0x6dfd:0x0] mdc close failed: rc = -13 LustreError: 31946:0:(file.c:247:ll_close_inode_openhandle()) Skipped 5 previous similar messages 2[5810]: segfault at 8 ip 00007f980de477e8 sp 00007ffd0c0bd1e0 error 4 in ld-2.17.so[7f980de3c000+22000] 0[11670]: segfault at 8 ip 00007f6fabdf87e8 sp 00007ffc1567de00 error 4 in ld-2.17.so[7f6fabded000+22000] 14[18293]: segfault at 8 ip 00007f9b45f357e8 sp 00007fff8d838370 error 4 in ld-2.17.so[7f9b45f2a000+22000] 16[20004]: segfault at 8 ip 00007fdc776a67e8 sp 00007ffc071432c0 error 4 in ld-2.17.so[7fdc7769b000+22000] 14[22847]: segfault at 406000 ip 0000000000406000 sp 00007ffc8a8b6838 error 14 in 14[606000+1000] 6[22948]: segfault at 8 ip 00007f0a02f777e8 sp 00007ffdeedd8900 error 4 in ld-2.17.so[7f0a02f6c000+22000] | Link to test |
racer test 1: racer on clients: centos-75.localnet DURATION=2700 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 PGD 318f00067 PUD 318f01067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) lustre(OE) osp(OE) ofd(OE) lod(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) jbd2 mbcache lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) crc32_generic lnet(OE) dm_flakey dm_mod libcfs(OE) crc_t10dif crct10dif_generic crct10dif_common virtio_console i2c_piix4 virtio_balloon pcspkr ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm drm_panel_orientation_quirks ata_piix i2c_core virtio_blk serio_raw libata floppy CPU: 6 PID: 18161 Comm: ll_sa_18142 Kdump: loaded Tainted: P OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.16.0-3.module_el8.7.0+1218+f626c2ff 04/01/2014 task: ffff880329049280 ti: ffff8802d5968000 task.ti: ffff8802d5968000 RIP: 0010:[<ffffffff813f502c>] [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 RSP: 0018:ffff8802d596bd78 EFLAGS: 00010296 RAX: ffff8802b8f9a710 RBX: 0000000000000008 RCX: 0000000100260023 RDX: 0000000000000026 RSI: ffff8802b8f9aa40 RDI: 0000000000000008 RBP: ffff8802d596bd88 R08: ffff8802dcd37a98 R09: 0000000000000000 R10: ffff8802dcd34c38 R11: ffff8802dcd37a98 R12: 0000000000000000 R13: ffff8802e97c24f0 R14: ffff8802b8f9aa40 R15: ffff8802dcd37a98 FS: 0000000000000000(0000) GS:ffff880331b80000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 00000002ee606000 CR4: 00000000000007e0 Call Trace: [<ffffffffa1109111>] ll_statahead_thread+0xdb1/0x2000 [lustre] [<ffffffff817e320e>] ? _raw_spin_lock+0x1e/0x20 [<ffffffff810dbb51>] ? put_prev_entity+0x31/0x400 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e329e>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffff810c834d>] ? finish_task_switch+0x5d/0x1b0 [<ffffffffa1108360>] ? ll_statahead_handle.constprop.28+0x1c0/0x1c0 [lustre] [<ffffffff810ba114>] kthread+0xe4/0xf0 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 | 13[18036]: segfault at 8 ip 00007f994c22a7e8 sp 00007ffe752fa4e0 error 4 in ld-2.17.so[7f994c21f000+22000] cat (18394) used greatest stack depth: 9600 bytes left Lustre: 15893:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff8802e8b6d540 x1832238453607808/t4294972370(0) o101->e301d82d-b48d-45d9-8a98-dc3e2eb37110@0@lo:404/0 lens 376/816 e 0 to 0 dl 1747358814 ref 1 fl Interpret:H/602/0 rc 0/0 job:'dd.0' uid:0 gid:0 projid:0 Lustre: 10872:0:(osd_internal.h:1332:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 618, rollback = 7 Lustre: 10872:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 10872:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 10872:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 10872:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 10872:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 10872:0:(osd_internal.h:1332:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 618, rollback = 7 Lustre: 10872:0:(osd_internal.h:1332:osd_trans_exec_op()) Skipped 1 previous similar message Lustre: 10872:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 10872:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 10872:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 10872:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 10872:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 10872:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 10872:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 10872:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 10872:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 10872:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 16365:0:(osd_internal.h:1332:osd_trans_exec_op()) lustre-OST0002: opcode 7: before 516 < left 618, rollback = 7 Lustre: 16365:0:(osd_internal.h:1332:osd_trans_exec_op()) Skipped 1 previous similar message Lustre: 16365:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 16365:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 16365:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 16365:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 16365:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 16365:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 16365:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 16365:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 16365:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 16365:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 10868:0:(osd_internal.h:1332:osd_trans_exec_op()) lustre-OST0000: opcode 7: before 516 < left 618, rollback = 7 Lustre: 10868:0:(osd_internal.h:1332:osd_trans_exec_op()) Skipped 1 previous similar message Lustre: 10868:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 10868:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 10868:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 10868:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 10868:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 10868:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 10868:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 10868:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 10868:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 10868:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 10868:0:(osd_internal.h:1332:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 516 < left 618, rollback = 7 Lustre: 10868:0:(osd_internal.h:1332:osd_trans_exec_op()) Skipped 19 previous similar messages Lustre: 10868:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 10868:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 19 previous similar messages Lustre: 10868:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 10868:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 19 previous similar messages Lustre: 10868:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 10868:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 19 previous similar messages Lustre: 10868:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 10868:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 19 previous similar messages Lustre: 10868:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 10868:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 19 previous similar messages LustreError: 26119:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802ef489bf8: inode [0x200000401:0x34f:0x0] mdc close failed: rc = -13 Lustre: 10872:0:(osd_internal.h:1332:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 618, rollback = 7 Lustre: 10872:0:(osd_internal.h:1332:osd_trans_exec_op()) Skipped 15 previous similar messages Lustre: 10872:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 10872:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 15 previous similar messages Lustre: 10872:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 10872:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 15 previous similar messages Lustre: 10872:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 10872:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 15 previous similar messages Lustre: 10872:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 10872:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 15 previous similar messages Lustre: 10872:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 10872:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 15 previous similar messages Lustre: 10868:0:(osd_internal.h:1332:osd_trans_exec_op()) lustre-OST0002: opcode 7: before 516 < left 618, rollback = 7 Lustre: 10868:0:(osd_internal.h:1332:osd_trans_exec_op()) Skipped 25 previous similar messages Lustre: 10868:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 10868:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 25 previous similar messages Lustre: 10868:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 10868:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 25 previous similar messages Lustre: 10868:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 10868:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 25 previous similar messages Lustre: 10868:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 10868:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 25 previous similar messages Lustre: 10868:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 10868:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 25 previous similar messages 3[5343]: segfault at 0 ip 0000000000403e5f sp 00007ffc8ce4b5b0 error 6 in 3[400000+6000] 0[6748]: segfault at 8 ip 00007f41c71b17e8 sp 00007ffccf19b450 error 4 in ld-2.17.so[7f41c71a6000+22000] Lustre: 15991:0:(osd_internal.h:1332:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 618, rollback = 7 Lustre: 15991:0:(osd_internal.h:1332:osd_trans_exec_op()) Skipped 79 previous similar messages Lustre: 15991:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 15991:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 79 previous similar messages Lustre: 15991:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 15991:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 79 previous similar messages Lustre: 15991:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 15991:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 79 previous similar messages Lustre: 15991:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 15991:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 79 previous similar messages Lustre: 15991:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 15991:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 79 previous similar messages LustreError: 16571:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802ecc6b7e8: inode [0x200000402:0xba5:0x0] mdc close failed: rc = -13 19[19292]: segfault at 8 ip 00007fd3372227e8 sp 00007ffdec530d40 error 4 in ld-2.17.so[7fd337217000+22000] 6[26630]: segfault at 1c ip 00000000004017b0 sp 00007ffee2354c90 error 6 in 19 (deleted)[400000+6000] 17[26770]: segfault at 6f ip 0000000000404fb3 sp 00007ffe4bd0e4b8 error 6 in 17[400000+6000] Lustre: 3001:0:(lod_lov.c:1414:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000401:0xe7e:0x0] with magic=0xbd60bd0 Lustre: 9735:0:(lod_lov.c:1414:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000401:0xee0:0x0] with magic=0xbd60bd0 Lustre: 9735:0:(lod_lov.c:1414:lod_parse_striping()) Skipped 1 previous similar message 6[29754]: segfault at 8 ip 00007f784d1147e8 sp 00007ffdb4096be0 error 4 in ld-2.17.so[7f784d109000+22000] 2[30971]: segfault at 8 ip 00007f9cacc167e8 sp 00007ffe2ae49ac0 error 4 in ld-2.17.so[7f9cacc0b000+22000] 9[31517]: segfault at 0 ip 0000000000403e5f sp 00007ffd08c783d0 error 6 in 9[400000+6000] Lustre: 10872:0:(osd_internal.h:1332:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 558, rollback = 7 Lustre: 10872:0:(osd_internal.h:1332:osd_trans_exec_op()) Skipped 93 previous similar messages Lustre: 15394:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 15394:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 95 previous similar messages Lustre: 15394:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 15394:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 95 previous similar messages Lustre: 15394:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 15394:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 95 previous similar messages Lustre: 15394:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 15394:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 95 previous similar messages Lustre: 15394:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 15394:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 95 previous similar messages 4[2523]: segfault at 8 ip 00007f48955057e8 sp 00007ffcd7055f70 error 4 in ld-2.17.so[7f48954fa000+22000] Lustre: mdt00_001: service thread pid 9734 was inactive for 40.055 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 9734, comm: mdt00_001 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock+0x88/0x1c0 [mdt] [<0>] mdt_object_find_lock+0x54/0x170 [mdt] [<0>] mdt_reint_setxattr+0x133/0x10e0 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x73c/0xbb0 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x257/0xcd0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 15157, comm: mdt00_007 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock_try+0xa0/0x250 [mdt] [<0>] mdt_getattr_name_lock+0x1729/0x2cc0 [mdt] [<0>] mdt_intent_getattr+0x2cc/0x4e0 [mdt] [<0>] mdt_intent_opc.constprop.74+0x211/0xc60 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x397/0x980 [ptlrpc] [<0>] ldlm_handle_enqueue+0x547/0x1900 [ptlrpc] [<0>] tgt_enqueue+0x68/0x240 [ptlrpc] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x257/0xcd0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 18026, comm: mdt00_010 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock_try+0xa0/0x250 [mdt] [<0>] mdt_getattr_name_lock+0x1729/0x2cc0 [mdt] [<0>] mdt_intent_getattr+0x2cc/0x4e0 [mdt] [<0>] mdt_intent_opc.constprop.74+0x211/0xc60 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x397/0x980 [ptlrpc] [<0>] ldlm_handle_enqueue+0x547/0x1900 [ptlrpc] [<0>] tgt_enqueue+0x68/0x240 [ptlrpc] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x257/0xcd0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: mdt00_004: service thread pid 13895 was inactive for 40.124 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: mdt00_002: service thread pid 9735 was inactive for 40.054 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 2 previous similar messages Lustre: mdt00_018: service thread pid 4534 was inactive for 40.090 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 4 previous similar messages Lustre: mdt00_003: service thread pid 11526 was inactive for 40.136 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 4 previous similar messages LustreError: 9409:0:(ldlm_lockd.c:257:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0000_UUID lock: ffff8800a7185a40/0xb74af72bb2bc33fb lrc: 3/0,0 mode: PR/PR res: [0x200000402:0x129b:0x0].0x0 bits 0x1b/0x0 rrc: 23 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0xb74af72bb2bc33df expref: 616 pid: 9732 timeout: 324 lvb_type: 0 Lustre: mdt00_001: service thread pid 9734 completed after 100.493s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 3001:0:(ldlm_lockd.c:1447:ldlm_handle_enqueue()) ### lock on destroyed export ffff8802ece75d28 ns: mdt-lustre-MDT0000_UUID lock: ffff8802dd2a2200/0xb74af72bb2bc4ce4 lrc: 3/0,0 mode: PR/PR res: [0x200000402:0x129b:0x0].0x0 bits 0x1b/0x0 rrc: 21 type: IBT gid 0 flags: 0x50200400000020 nid: 0@lo remote: 0xb74af72bb2bc4cc8 expref: 18 pid: 3001 timeout: 0 lvb_type: 0 Lustre: mdt00_013: service thread pid 3001 completed after 99.745s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_009: service thread pid 15895 completed after 99.613s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: lustre-MDT0000-mdc-ffff8802ecc6b7e8: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-MDT0000-mdc-ffff8802ecc6b7e8: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: lustre-MDT0000-mdc-ffff8802ecc6b7e8: This client was evicted by lustre-MDT0000; in progress operations using this service will fail. Lustre: mdt00_010: service thread pid 18026 completed after 100.310s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_004: service thread pid 13895 completed after 100.190s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_007: service thread pid 15157 completed after 100.412s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_002: service thread pid 9735 completed after 99.611s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 4626:0:(file.c:6143:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000402:0x129b:0x0] error: rc = -5 LustreError: 4687:0:(llite_lib.c:2023:ll_md_setattr()) md_setattr fails: rc = -108 Lustre: mdt00_012: service thread pid 18961 completed after 99.639s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 4979:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802ecc6b7e8: inode [0x200000401:0x129d:0x0] mdc close failed: rc = -108 LustreError: 4944:0:(symlink.c:75:ll_readlink_internal()) lustre: inode [0x200000401:0x12a2:0x0]: rc = -108 LustreError: 4085:0:(vvp_io.c:1905:vvp_io_init()) lustre: refresh file layout [0x200000402:0x129b:0x0] error -108. Lustre: mdt00_005: service thread pid 15116 completed after 99.443s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_014: service thread pid 15132 completed after 99.263s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_011: service thread pid 18526 completed after 98.133s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_018: service thread pid 4534 completed after 98.637s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_021: service thread pid 4798 completed after 96.294s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_015: service thread pid 4476 completed after 98.133s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_019: service thread pid 4657 completed after 97.470s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_000: service thread pid 9732 completed after 99.198s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_003: service thread pid 11526 completed after 96.530s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: lustre-MDT0000-mdc-ffff8802ecc6b7e8: Connection restored to 192.168.123.77@tcp (at 0@lo) Lustre: mdt00_008: service thread pid 15893 completed after 98.420s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: 16365:0:(osd_internal.h:1332:osd_trans_exec_op()) lustre-OST0000: opcode 7: before 516 < left 618, rollback = 7 Lustre: 16365:0:(osd_internal.h:1332:osd_trans_exec_op()) Skipped 13 previous similar messages 2[8383]: segfault at 8 ip 00007f16cce0b7e8 sp 00007ffef1bcc810 error 4 in ld-2.17.so[7f16cce00000+22000] Lustre: 10868:0:(osd_handler.c:1966:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 10868:0:(osd_handler.c:1966:osd_trans_dump_creds()) Skipped 15 previous similar messages Lustre: 10868:0:(osd_handler.c:1973:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 10868:0:(osd_handler.c:1973:osd_trans_dump_creds()) Skipped 15 previous similar messages Lustre: 11526:0:(lod_lov.c:1414:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000401:0x1435:0x0] with magic=0xbd60bd0 Lustre: 11526:0:(lod_lov.c:1414:lod_parse_striping()) Skipped 1 previous similar message Lustre: 10868:0:(osd_handler.c:1983:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 10868:0:(osd_handler.c:1983:osd_trans_dump_creds()) Skipped 15 previous similar messages Lustre: 10868:0:(osd_handler.c:1990:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 10868:0:(osd_handler.c:1990:osd_trans_dump_creds()) Skipped 15 previous similar messages Lustre: 10868:0:(osd_handler.c:1997:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 10868:0:(osd_handler.c:1997:osd_trans_dump_creds()) Skipped 15 previous similar messages 4[11439]: segfault at 8 ip 00007f2c1822b7e8 sp 00007ffeda81b9f0 error 4 in ld-2.17.so[7f2c18220000+22000] traps: 0[12615] general protection ip:404e66 sp:7fffffb52f58 error:0 in 0[400000+6000] traps: 2[12645] general protection ip:404f28 sp:7fff4bf10a78 error:0 in 2[400000+6000] 7[21250]: segfault at 8 ip 00007f1bd6f9f7e8 sp 00007ffd7701ebb0 error 4 in ld-2.17.so[7f1bd6f94000+22000] LustreError: 21250:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802ef489bf8: inode [0x200000403:0x4a0:0x0] mdc close failed: rc = -13 LustreError: 21250:0:(file.c:247:ll_close_inode_openhandle()) Skipped 9 previous similar messages hrtimer: interrupt took 47167480 ns Lustre: 4798:0:(lod_lov.c:1414:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000401:0x1839:0x0] with magic=0xbd60bd0 Lustre: 4798:0:(lod_lov.c:1414:lod_parse_striping()) Skipped 1 previous similar message 3[2529]: segfault at 8 ip 00007fe0fd02a7e8 sp 00007fffe793aa70 error 4 in ld-2.17.so[7fe0fd01f000+22000] LustreError: 8939:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802ef489bf8: inode [0x200000403:0xba8:0x0] mdc close failed: rc = -13 Lustre: 9732:0:(lod_lov.c:1414:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0xc0b:0x0] with magic=0xbd60bd0 Lustre: 9732:0:(lod_lov.c:1414:lod_parse_striping()) Skipped 1 previous similar message 14[12265]: segfault at 8 ip 00007fa9b47707e8 sp 00007ffef35e71d0 error 4 in ld-2.17.so[7fa9b4765000+22000] 19[14574]: segfault at 0 ip 0000000000403e5f sp 00007ffd1c549920 error 6 in 10[400000+6000] | Link to test |
racer test 1: racer on clients: oleg249-client.virtnet DURATION=300 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffff813f40ac>] _atomic_dec_and_lock+0xc/0x70 PGD a5fe1067 PUD 960b2067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) osp(OE) ofd(OE) lod(OE) mdt(OE) mdd(OE) mgs(OE) lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) crc_t10dif crct10dif_generic rpcsec_gss_krb5 sb_edac edac_core iosf_mbi kvm_intel kvm irqbypass crc32_pclmul ghash_clmulni_intel aesni_intel squashfs lrw gf128mul glue_helper ablk_helper cryptd pcspkr i2c_piix4 i2c_core binfmt_misc ip_tables ext4 mbcache jbd2 ata_generic pata_acpi crct10dif_pclmul crct10dif_common crc32c_intel serio_raw ata_piix libata CPU: 3 PID: 13048 Comm: ll_sa_12987 Kdump: loaded Tainted: G OE ------------ 3.10.0-7.9-debug #1 Hardware name: QEMU Standard PC (i440FX + PIIX, 1996), BIOS 1.16.2-1.fc38 04/01/2014 task: ffff88009c23e660 ti: ffff8800a37d4000 task.ti: ffff8800a37d4000 RIP: 0010:[<ffffffff813f40ac>] [<ffffffff813f40ac>] _atomic_dec_and_lock+0xc/0x70 RSP: 0018:ffff8800a37d7d78 EFLAGS: 00010296 RAX: ffff88009331c8d0 RBX: 0000000000000008 RCX: 00000000000ca149 RDX: 00000000000ca148 RSI: ffff88009331cc00 RDI: 0000000000000008 RBP: ffff8800a37d7d88 R08: ffff8800a37d7fd8 R09: 002dffd3a255376d R10: 0000000000000006 R11: ffff8800a8762a80 R12: 0000000000000000 R13: ffff88009d679110 R14: ffff88009331cc00 R15: ffff88009376d200 FS: 0000000000000000(0000) GS:ffff88013e380000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 CR2: 0000000000000008 CR3: 000000009377e000 CR4: 0000000000160fe0 Call Trace: [<ffffffffa10c3481>] ll_statahead_thread+0xdb1/0x2000 [lustre] [<ffffffff810df6e2>] ? pick_next_task_fair+0x702/0x8a0 [<ffffffff81410a79>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e8dce>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffff810c834d>] ? finish_task_switch+0x5d/0x1b0 [<ffffffffa10c26d0>] ? ll_statahead_handle.constprop.28+0x1c0/0x1c0 [lustre] [<ffffffff810ba114>] kthread+0xe4/0xf0 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817f3e5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 | random: crng init done LustreError: 14982:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800b7c16800: inode [0x200000401:0xa2:0x0] mdc close failed: rc = -116 dd (15338) used greatest stack depth: 10592 bytes left LustreError: 23523:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800b7c16800: inode [0x200000402:0x523:0x0] mdc close failed: rc = -13 LustreError: 23523:0:(file.c:247:ll_close_inode_openhandle()) Skipped 1 previous similar message 19[29462]: segfault at 0 ip 0000000000403e5f sp 00007ffeed1681b0 error 6 in 19[400000+6000] LustreError: 31250:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800b3783800: inode [0x200000401:0x858:0x0] mdc close failed: rc = -13 15[10619]: segfault at 0 ip 0000000000403e5f sp 00007ffc8e6b31e0 error 6 in 15[400000+6000] 3[20551]: segfault at 8 ip 00007f37f67827e8 sp 00007ffcdef94430 error 4 in ld-2.17.so[7f37f6777000+22000] 19[22003]: segfault at 8 ip 00007fd3540d07e8 sp 00007fffe202f950 error 4 in ld-2.17.so[7fd3540c5000+22000] LustreError: 32049:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800b3783800: inode [0x200000401:0x17b9:0x0] mdc close failed: rc = -116 LustreError: 32049:0:(file.c:247:ll_close_inode_openhandle()) Skipped 1 previous similar message 19[10809]: segfault at 0 ip 0000000000403e5f sp 00007ffdc27adb10 error 6 in 17[400000+6000] LustreError: 15884:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800b7c16800: inode [0x200000402:0x1cec:0x0] mdc close failed: rc = -116 0[26100]: segfault at 8 ip 00007f9468b957e8 sp 00007fffafb9cbb0 error 4 in ld-2.17.so[7f9468b8a000+22000] 16[26484]: segfault at 8 ip 00007f07e07757e8 sp 00007ffdd6aeea10 error 4 in ld-2.17.so[7f07e076a000+22000] LustreError: 26484:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800b7c16800: inode [0x200000402:0x2230:0x0] mdc close failed: rc = -13 LustreError: 26484:0:(file.c:247:ll_close_inode_openhandle()) Skipped 1 previous similar message | Link to test |
racer test 1: racer on clients: centos-80.localnet DURATION=2700 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 PGD 8c021067 PUD 8c022067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) lustre(OE) osp(OE) ofd(OE) lod(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) jbd2 mbcache lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) crc32_generic lnet(OE) dm_flakey dm_mod libcfs(OE) crc_t10dif crct10dif_generic crct10dif_common virtio_balloon virtio_console pcspkr i2c_piix4 ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm ata_piix drm_panel_orientation_quirks serio_raw virtio_blk libata i2c_core floppy CPU: 6 PID: 12405 Comm: ll_sa_12282 Kdump: loaded Tainted: P OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.16.0-3.module_el8.7.0+1218+f626c2ff 04/01/2014 task: ffff88029ae5b760 ti: ffff88027e070000 task.ti: ffff88027e070000 RIP: 0010:[<ffffffff813f502c>] [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 RSP: 0018:ffff88027e073d78 EFLAGS: 00010296 RAX: ffff88026bf73090 RBX: 0000000000000008 RCX: 000000010026001c RDX: 0000000000000026 RSI: ffff88026bf733c0 RDI: 0000000000000008 RBP: ffff88027e073d88 R08: ffff8802a910d978 R09: 0000000000000000 R10: ffff8802a910b858 R11: ffff8802a910d978 R12: 0000000000000000 R13: ffff880299c049d0 R14: ffff88026bf733c0 R15: ffff8802a910d978 FS: 0000000000000000(0000) GS:ffff880331b80000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 00000000a5408000 CR4: 00000000000007e0 Call Trace: [<ffffffffa110e431>] ll_statahead_thread+0xdb1/0x2000 [lustre] [<ffffffff810df6e2>] ? pick_next_task_fair+0x702/0x8a0 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e329e>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffff810c834d>] ? finish_task_switch+0x5d/0x1b0 [<ffffffffa110d680>] ? ll_statahead_handle.constprop.28+0x1c0/0x1c0 [lustre] [<ffffffff810ba114>] kthread+0xe4/0xf0 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 | Lustre: 18113:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 509 < left 699, rollback = 2 Lustre: 18113:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 3/12/2, destroy: 0/0/0 Lustre: 18113:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 2/2/0, xattr_set: 9/699/0 cat (18350) used greatest stack depth: 10112 bytes left Lustre: 18113:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 5/55/0, punch: 0/0/0, quota 1/3/0 Lustre: 18113:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 10/203/4, delete: 0/0/0 Lustre: 18113:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 4/4/0, ref_del: 0/0/0 getfattr (18513) used greatest stack depth: 9936 bytes left Lustre: 18113:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 508 < left 816, rollback = 2 Lustre: 18113:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 7 previous similar messages Lustre: 18113:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 3/12/4, destroy: 0/0/0 Lustre: 18113:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 7 previous similar messages Lustre: 18113:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 4/4/0, xattr_set: 12/816/0 Lustre: 18113:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 7 previous similar messages Lustre: 18113:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 5/55/0, punch: 0/0/0, quota 1/3/0 Lustre: 18113:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 7 previous similar messages Lustre: 18113:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 12/215/3, delete: 1/1/0 Lustre: 18113:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 7 previous similar messages Lustre: 18113:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 6/6/0, ref_del: 1/1/0 Lustre: 18113:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 7 previous similar messages Lustre: 18581:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 511 < left 699, rollback = 2 Lustre: 18581:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 9 previous similar messages Lustre: 18581:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 3/12/2, destroy: 0/0/0 Lustre: 18581:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 9 previous similar messages Lustre: 18581:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 2/2/0, xattr_set: 9/699/0 Lustre: 18581:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 9 previous similar messages Lustre: 18581:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 5/55/0, punch: 0/0/0, quota 4/54/0 Lustre: 18581:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 9 previous similar messages Lustre: 18581:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 11/239/2, delete: 0/0/0 Lustre: 18581:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 9 previous similar messages Lustre: 18581:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 5/5/0, ref_del: 0/0/0 Lustre: 18581:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 9 previous similar messages Lustre: 18113:0:(mdd_dir.c:4822:mdd_migrate_object()) lustre-MDD0000: [0x200000403:0x1:0x0]/7 is open, migrate only dentry LustreError: 11497:0:(mdt_reint.c:2540:mdt_reint_migrate()) lustre-MDT0000: migrate [0x200000404:0x87:0x0]/11 failed: rc = -16 Lustre: 11497:0:(mdd_dir.c:4822:mdd_migrate_object()) lustre-MDD0001: [0x240000403:0x1:0x0]/5 is open, migrate only dentry 16[20642]: segfault at 8 ip 00007f4fb3eea7e8 sp 00007ffe01ba1070 error 4 in ld-2.17.so[7f4fb3edf000+22000] 13[21291]: segfault at 0 ip 0000000000403e5f sp 00007ffd2e2234e0 error 6 in 13[400000+6000] LustreError: 21152:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000403:0x40:0x0]: rc = -5 LustreError: 21152:0:(llite_lib.c:3700:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 Lustre: 11496:0:(mdd_dir.c:4822:mdd_migrate_object()) lustre-MDD0002: [0x280000403:0x1:0x0]/2 is open, migrate only dentry Lustre: 11496:0:(mdd_dir.c:4822:mdd_migrate_object()) Skipped 1 previous similar message LustreError: 21424:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000403:0x40:0x0]: rc = -5 LustreError: 21424:0:(lcommon_cl.c:188:cl_file_inode_init()) Skipped 6 previous similar messages LustreError: 21424:0:(llite_lib.c:3700:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 21424:0:(llite_lib.c:3700:ll_prep_inode()) Skipped 6 previous similar messages 1[22156]: segfault at 8 ip 00007f8e1440b7e8 sp 00007ffd286c2360 error 4 in ld-2.17.so[7f8e14400000+22000] LustreError: 22141:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000403:0x40:0x0]: rc = -5 LustreError: 22141:0:(lcommon_cl.c:188:cl_file_inode_init()) Skipped 4 previous similar messages LustreError: 22141:0:(llite_lib.c:3700:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 22141:0:(llite_lib.c:3700:ll_prep_inode()) Skipped 4 previous similar messages Lustre: 11497:0:(mdd_dir.c:4822:mdd_migrate_object()) lustre-MDD0000: [0x200000403:0x1:0x0]/18 is open, migrate only dentry Lustre: 11497:0:(mdd_dir.c:4822:mdd_migrate_object()) Skipped 3 previous similar messages Lustre: 18581:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 502 < left 877, rollback = 2 Lustre: 18581:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 7 previous similar messages Lustre: 18581:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 3/12/9, destroy: 0/0/0 Lustre: 18581:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 7 previous similar messages Lustre: 18581:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 2/2/0, xattr_set: 11/877/0 Lustre: 18581:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 7 previous similar messages Lustre: 18581:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 5/55/0, punch: 0/0/0, quota 4/54/0 Lustre: 18581:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 7 previous similar messages Lustre: 18581:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 12/275/4, delete: 0/0/0 Lustre: 18581:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 7 previous similar messages Lustre: 18581:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 6/6/0, ref_del: 0/0/0 Lustre: 18581:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 7 previous similar messages Lustre: dir [0x200000403:0x80:0x0] stripe 2 readdir failed: -2, directory is partially accessed! Lustre: 17407:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff8800a0759440 x1831405848290560/t4294968508(0) o101->ffc3eb26-de74-4fa5-a5e7-64173a3e2d92@0@lo:639/0 lens 376/840 e 0 to 0 dl 1746564789 ref 1 fl Interpret:H/602/0 rc 0/0 job:'dd.0' uid:0 gid:0 projid:0 LustreError: 23084:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000403:0x40:0x0]: rc = -5 LustreError: 23084:0:(lcommon_cl.c:188:cl_file_inode_init()) Skipped 6 previous similar messages LustreError: 23084:0:(llite_lib.c:3700:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 23084:0:(llite_lib.c:3700:ll_prep_inode()) Skipped 6 previous similar messages Lustre: 20991:0:(mdt_reint.c:2460:mdt_reint_migrate()) lustre-MDT0001: [0x240000403:0x1:0x0]/1 is open, migrate only dentry LustreError: 11486:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0001: failed to get lu_attr of [0x240000404:0xa:0x0]: rc = -2 LustreError: 24023:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a731a548: inode [0x240000404:0xa:0x0] mdc close failed: rc = -2 7[23866]: segfault at 8 ip 00007f2acb2667e8 sp 00007ffcf89f8d20 error 4 in ld-2.17.so[7f2acb25b000+22000] Lustre: 20991:0:(mdd_dir.c:4822:mdd_migrate_object()) lustre-MDD0000: [0x200000403:0x1:0x0]/18 is open, migrate only dentry Lustre: 20991:0:(mdd_dir.c:4822:mdd_migrate_object()) Skipped 2 previous similar messages Lustre: dir [0x280000403:0x6e:0x0] stripe 3 readdir failed: -2, directory is partially accessed! Lustre: 22465:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 498 < left 727, rollback = 2 Lustre: 22465:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 113 previous similar messages Lustre: 22465:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 3/12/12, destroy: 0/0/0 Lustre: 22465:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 113 previous similar messages Lustre: 22465:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 4/4/0, xattr_set: 11/727/0 Lustre: 22465:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 113 previous similar messages Lustre: 22465:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 5/55/0, punch: 0/0/0, quota 4/54/1 Lustre: 22465:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 113 previous similar messages Lustre: 22465:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 14/287/4, delete: 1/1/0 Lustre: 22465:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 113 previous similar messages Lustre: 22465:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 8/8/0, ref_del: 1/1/0 Lustre: 22465:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 113 previous similar messages LustreError: 26055:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000403:0x40:0x0]: rc = -5 LustreError: 26055:0:(lcommon_cl.c:188:cl_file_inode_init()) Skipped 8 previous similar messages LustreError: 26055:0:(llite_lib.c:3700:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 26055:0:(llite_lib.c:3700:ll_prep_inode()) Skipped 8 previous similar messages Lustre: 20340:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff8802ce662d40 x1831405850289408/t4294969639(0) o101->ffc3eb26-de74-4fa5-a5e7-64173a3e2d92@0@lo:645/0 lens 376/864 e 0 to 0 dl 1746564795 ref 1 fl Interpret:H/602/0 rc 0/0 job:'dd.0' uid:0 gid:0 projid:0 LustreError: 18581:0:(mdd_dir.c:4743:mdd_migrate_cmd_check()) lustre-MDD0001: '15' migration was interrupted, run 'lfs migrate -m 2 -c 3 -H crush 15' to finish migration: rc = -1 LustreError: 18581:0:(mdt_reint.c:2540:mdt_reint_migrate()) lustre-MDT0001: migrate [0x240000403:0x1:0x0]/15 failed: rc = -1 Lustre: dir [0x240000403:0xb8:0x0] stripe 3 readdir failed: -2, directory is partially accessed! Lustre: Skipped 2 previous similar messages Lustre: 13518:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0002: opcode 7: before 516 < left 618, rollback = 7 Lustre: 13517:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0002: opcode 7: before 516 < left 618, rollback = 7 Lustre: 13517:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 1 previous similar message 6[26841]: segfault at 8 ip 00007f2bf17097e8 sp 00007ffd61b70840 error 4 in ld-2.17.so[7f2bf16fe000+22000] Lustre: 17806:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0002: opcode 7: before 516 < left 618, rollback = 7 Lustre: 17806:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 2 previous similar messages LustreError: 20991:0:(mdt_reint.c:2540:mdt_reint_migrate()) lustre-MDT0000: migrate [0x280000404:0xdc:0x0]/16 failed: rc = -116 6[28401]: segfault at 8 ip 00007f3e25a237e8 sp 00007ffeaad833c0 error 4 in ld-2.17.so[7f3e25a18000+22000] LustreError: 28401:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a731a548: inode [0x240000404:0x151:0x0] mdc close failed: rc = -13 LustreError: 11487:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0002: failed to get lu_attr of [0x280000403:0x6e:0x0]: rc = -2 Lustre: 11496:0:(mdd_dir.c:4822:mdd_migrate_object()) lustre-MDD0002: [0x280000403:0x1:0x0]/12 is open, migrate only dentry Lustre: 11496:0:(mdd_dir.c:4822:mdd_migrate_object()) Skipped 7 previous similar messages Lustre: 28020:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 509 < left 1784, rollback = 2 Lustre: 28020:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 196 previous similar messages Lustre: 28020:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 3/12/3, destroy: 1/4/0 Lustre: 28020:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 206 previous similar messages Lustre: 28020:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 2/2/0, xattr_set: 21/1784/0 Lustre: 28020:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 206 previous similar messages Lustre: 28020:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 5/55/0, punch: 0/0/0, quota 7/129/0 Lustre: 28020:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 206 previous similar messages Lustre: 28020:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 14/287/3, delete: 2/5/0 Lustre: 28020:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 206 previous similar messages Lustre: 28020:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 7/7/0, ref_del: 2/2/0 Lustre: 28020:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 206 previous similar messages Lustre: 17301:0:(lod_lov.c:1414:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000403:0xf9:0x0] with magic=0xbd60bd0 Lustre: 18247:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0000: opcode 7: before 516 < left 618, rollback = 7 Lustre: 18247:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 4 previous similar messages LustreError: 29147:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000403:0x40:0x0]: rc = -5 LustreError: 29147:0:(lcommon_cl.c:188:cl_file_inode_init()) Skipped 47 previous similar messages LustreError: 29332:0:(llite_lib.c:3700:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 29332:0:(llite_lib.c:3700:ll_prep_inode()) Skipped 50 previous similar messages LustreError: 11496:0:(mdt_reint.c:2540:mdt_reint_migrate()) lustre-MDT0000: migrate [0x200000404:0x153:0x0]/14 failed: rc = -2 LustreError: 11496:0:(mdt_reint.c:2540:mdt_reint_migrate()) Skipped 1 previous similar message LustreError: 20991:0:(mdd_dir.c:4743:mdd_migrate_cmd_check()) lustre-MDD0002: '0' migration was interrupted, run 'lfs migrate -m 0 -c 2 -H crush 0' to finish migration: rc = -1 LustreError: 11486:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0002: failed to get lu_attr of [0x280000404:0x115:0x0]: rc = -2 LustreError: 29383:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a731a548: inode [0x280000404:0x115:0x0] mdc close failed: rc = -2 LustreError: 29383:0:(file.c:247:ll_close_inode_openhandle()) Skipped 1 previous similar message LustreError: 28020:0:(mdd_dir.c:4743:mdd_migrate_cmd_check()) lustre-MDD0002: '10' migration was interrupted, run 'lfs migrate -m 1 -c 1 -H crush 10' to finish migration: rc = -1 LustreError: 17405:0:(mdt_open.c:1302:mdt_cross_open()) lustre-MDT0000: [0x200000403:0x26:0x0] doesn't exist!: rc = -14 0[6907]: segfault at 8 ip 00007f0d34b057e8 sp 00007ffdc1633610 error 4 in ld-2.17.so[7f0d34afa000+22000] 12[11952]: segfault at 0 ip (null) sp 00007ffc913883f8 error 14 in 12[400000+6000] LustreError: 10927:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a70f4138: inode [0x280000403:0x211:0x0] mdc close failed: rc = -13 Lustre: 22465:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 510 < left 2140, rollback = 2 Lustre: 22465:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 131 previous similar messages Lustre: 22465:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 3/12/3, destroy: 2/8/0 Lustre: 22465:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 139 previous similar messages Lustre: 22465:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 2/2/0, xattr_set: 25/2140/0 Lustre: 22465:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 139 previous similar messages Lustre: 22465:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 5/55/0, punch: 0/0/0, quota 7/129/0 Lustre: 22465:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 139 previous similar messages Lustre: 22465:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 15/323/1, delete: 9/15/1 Lustre: 22465:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 139 previous similar messages Lustre: 22465:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 8/8/0, ref_del: 9/9/0 Lustre: 22465:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 139 previous similar messages Lustre: 22465:0:(mdd_dir.c:4822:mdd_migrate_object()) lustre-MDD0000: [0x200000403:0x2:0x0]/1 is open, migrate only dentry Lustre: 22465:0:(mdd_dir.c:4822:mdd_migrate_object()) Skipped 3 previous similar messages LustreError: 12056:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000403:0x117:0x0]: rc = -5 LustreError: 12056:0:(lcommon_cl.c:188:cl_file_inode_init()) Skipped 34 previous similar messages LustreError: 12056:0:(llite_lib.c:3700:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 12056:0:(llite_lib.c:3700:ll_prep_inode()) Skipped 31 previous similar messages 19[13531]: segfault at 8 ip 00007eff5dca17e8 sp 00007ffda643b5b0 error 4 in ld-2.17.so[7eff5dc96000+22000] LustreError: 29262:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000403:0x2ae:0x0]: rc = -2 LustreError: 14676:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a70f4138: inode [0x200000403:0x2ae:0x0] mdc close failed: rc = -2 14[14585]: segfault at 8 ip 00007f89ea8f57e8 sp 00007fff38244170 error 4 in ld-2.17.so[7f89ea8ea000+22000] LustreError: 29262:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000404:0x1d1:0x0]: rc = -2 4[16913]: segfault at 4017b0 ip 00000000004017b0 sp 00007ffef12f59f0 error 14 LustreError: 26491:0:(mdt_reint.c:2540:mdt_reint_migrate()) lustre-MDT0000: migrate [0x240000404:0x3ad:0x0]/18 failed: rc = -16 LustreError: 26491:0:(mdt_reint.c:2540:mdt_reint_migrate()) Skipped 3 previous similar messages LustreError: 15253:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a731a548: inode [0x240000404:0x167:0x0] mdc close failed: rc = -2 LustreError: 15253:0:(file.c:247:ll_close_inode_openhandle()) Skipped 1 previous similar message Lustre: 17806:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 515 < left 618, rollback = 7 Lustre: 17806:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 7 previous similar messages LustreError: 22465:0:(mdd_dir.c:4743:mdd_migrate_cmd_check()) lustre-MDD0001: '19' migration was interrupted, run 'lfs migrate -m 2 -c 3 -H crush 19' to finish migration: rc = -1 LustreError: 22465:0:(mdt_reint.c:2540:mdt_reint_migrate()) lustre-MDT0001: migrate [0x200000403:0x1:0x0]/19 failed: rc = -1 LustreError: 22465:0:(mdt_reint.c:2540:mdt_reint_migrate()) Skipped 4 previous similar messages 5[20610]: segfault at 8 ip 00007efeb400d7e8 sp 00007ffd9c8d1aa0 error 4 in ld-2.17.so[7efeb4002000+22000] Lustre: 25367:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0000: opcode 7: before 515 < left 618, rollback = 7 Lustre: 25367:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 5 previous similar messages LustreError: 13888:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0001: [0x240000404:0x3ad:0x0] migrate mdt count mismatch 1 != 2 14[21757]: segfault at 0 ip 0000000000403cf0 sp 00007ffd7d113678 error 6 in 14[400000+6000] 14[22970]: segfault at 8 ip 00007fd63c0607e8 sp 00007ffcdefc7d80 error 4 in ld-2.17.so[7fd63c055000+22000] LustreError: 27342:0:(mdd_dir.c:4743:mdd_migrate_cmd_check()) lustre-MDD0000: '19' migration was interrupted, run 'lfs migrate -m 2 -c 3 -H crush 19' to finish migration: rc = -1 Lustre: 11497:0:(mdd_dir.c:4822:mdd_migrate_object()) lustre-MDD0002: [0x200000403:0x1:0x0]/2 is open, migrate only dentry Lustre: 11497:0:(mdd_dir.c:4822:mdd_migrate_object()) Skipped 15 previous similar messages Lustre: dir [0x240000404:0x46b:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: 26491:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 507 < left 877, rollback = 2 Lustre: 26491:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 569 previous similar messages Lustre: 26491:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 3/12/5, destroy: 0/0/0 Lustre: 26491:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 583 previous similar messages Lustre: 26491:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 2/2/0, xattr_set: 11/877/0 Lustre: 26491:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 583 previous similar messages Lustre: 26491:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 5/55/0, punch: 0/0/0, quota 4/54/0 Lustre: 26491:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 583 previous similar messages Lustre: 26491:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 10/203/3, delete: 0/0/0 Lustre: 26491:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 583 previous similar messages Lustre: 26491:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 4/4/0, ref_del: 0/0/0 Lustre: 26491:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 583 previous similar messages LustreError: 21742:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000404:0x64b:0x0]: rc = -2 LustreError: 21742:0:(mdd_object.c:3901:mdd_close()) Skipped 1 previous similar message LustreError: 24598:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a70f4138: inode [0x200000404:0x64b:0x0] mdc close failed: rc = -2 LustreError: 16657:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000404:0x66d:0x0]: rc = -5 LustreError: 16657:0:(lcommon_cl.c:188:cl_file_inode_init()) Skipped 16 previous similar messages LustreError: 16657:0:(llite_lib.c:3700:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 16657:0:(llite_lib.c:3700:ll_prep_inode()) Skipped 16 previous similar messages LustreError: 21034:0:(lov_object.c:1348:lov_layout_change()) lustre-clilov-ffff8800a70f4138: cannot apply new layout on [0x200000404:0x66d:0x0] : rc = -5 LustreError: 21034:0:(vvp_io.c:1905:vvp_io_init()) lustre: refresh file layout [0x200000404:0x66d:0x0] error -5. Lustre: dir [0x200000404:0x68e:0x0] stripe 3 readdir failed: -2, directory is partially accessed! Lustre: Skipped 1 previous similar message Lustre: 18247:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 516 < left 618, rollback = 7 Lustre: 18247:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 9 previous similar messages LustreError: 26020:0:(mdt_reint.c:2540:mdt_reint_migrate()) lustre-MDT0002: migrate [0x200000404:0x177:0x0]/3 failed: rc = -2 LustreError: 26020:0:(mdt_reint.c:2540:mdt_reint_migrate()) Skipped 3 previous similar messages 2[28406]: segfault at 8 ip 00007fb9b55427e8 sp 00007ffeef7cfce0 error 4 in ld-2.17.so[7fb9b5537000+22000] 3[29113]: segfault at 8 ip 00007ff489c0e7e8 sp 00007ffc6476b930 error 4 in ld-2.17.so[7ff489c03000+22000] 5[28807]: segfault at 8 ip 00007f2b60e197e8 sp 00007ffc60149a00 error 4 in ld-2.17.so[7f2b60e0e000+22000] LustreError: 11486:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000404:0x177:0x0]: rc = -2 LustreError: 11486:0:(mdd_object.c:3901:mdd_close()) Skipped 2 previous similar messages 9[30175]: segfault at 8 ip 00007f090311b2fc sp 00007ffdabfe0a30 error 4 in ld-2.17.so[7f090310e000+22000] Lustre: dir [0x280000403:0x4e7:0x0] stripe 1 readdir failed: -2, directory is partially accessed! LustreError: 26491:0:(mdd_dir.c:4743:mdd_migrate_cmd_check()) lustre-MDD0002: '10' migration was interrupted, run 'lfs migrate -m 1 -c 2 -H crush 10' to finish migration: rc = -1 LustreError: 26491:0:(mdd_dir.c:4743:mdd_migrate_cmd_check()) Skipped 1 previous similar message LustreError: 11498:0:(lustre_lmv.h:500:lmv_is_sane()) unknown layout LMV: magic=0xcd40cd0 count=3 index=2 hash=crush:0x82000003 version=1 migrate_offset=2 migrate_hash=fnv_1a_64:2 pool= LustreError: 746:0:(lov_object.c:1348:lov_layout_change()) lustre-clilov-ffff8800a731a548: cannot apply new layout on [0x240000403:0x63f:0x0] : rc = -5 LustreError: 746:0:(vvp_io.c:1905:vvp_io_init()) lustre: refresh file layout [0x240000403:0x63f:0x0] error -5. LustreError: 1866:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a70f4138: inode [0x240000404:0x70a:0x0] mdc close failed: rc = -2 LustreError: 1866:0:(file.c:247:ll_close_inode_openhandle()) Skipped 5 previous similar messages Lustre: 21775:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 618, rollback = 7 Lustre: 21775:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 7 previous similar messages 8[4895]: segfault at 8 ip 00007fd796de17e8 sp 00007ffdc504d410 error 4 in ld-2.17.so[7fd796dd6000+22000] LustreError: 27342:0:(mdt_reint.c:2540:mdt_reint_migrate()) lustre-MDT0000: migrate [0x200000403:0x1:0x0]/1 failed: rc = -1 LustreError: 27342:0:(mdt_reint.c:2540:mdt_reint_migrate()) Skipped 10 previous similar messages 8[5083]: segfault at 8 ip 00007fb5094e77e8 sp 00007ffd99f30800 error 4 in ld-2.17.so[7fb5094dc000+22000] LustreError: 27432:0:(mdd_dir.c:4743:mdd_migrate_cmd_check()) lustre-MDD0001: '17' migration was interrupted, run 'lfs migrate -m 2 -c 2 -H crush 17' to finish migration: rc = -1 LustreError: 27432:0:(mdd_dir.c:4743:mdd_migrate_cmd_check()) Skipped 4 previous similar messages Lustre: 20350:0:(lod_lov.c:1414:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0xa8d:0x0] with magic=0xbd60bd0 Lustre: 20350:0:(lod_lov.c:1414:lod_parse_striping()) Skipped 3 previous similar messages Lustre: 17265:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff8802bd6a6940 x1831405902623872/t4294994795(0) o101->ffc3eb26-de74-4fa5-a5e7-64173a3e2d92@0@lo:57/0 lens 376/816 e 0 to 0 dl 1746564962 ref 1 fl Interpret:H/602/0 rc 0/0 job:'dd.0' uid:0 gid:0 projid:0 Lustre: 28020:0:(mdd_dir.c:4822:mdd_migrate_object()) lustre-MDD0002: [0x280000404:0x6d7:0x0]/3 is open, migrate only dentry Lustre: 28020:0:(mdd_dir.c:4822:mdd_migrate_object()) Skipped 26 previous similar messages Lustre: 28020:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 508 < left 788, rollback = 2 Lustre: 28020:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 1233 previous similar messages Lustre: 28020:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 3/12/4, destroy: 0/0/0 Lustre: 28020:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 1265 previous similar messages Lustre: 28020:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 2/2/0, xattr_set: 10/788/0 Lustre: 28020:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 1265 previous similar messages Lustre: 28020:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 5/55/0, punch: 0/0/0, quota 4/30/0 Lustre: 28020:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 1265 previous similar messages Lustre: 28020:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 12/275/3, delete: 0/0/0 Lustre: 28020:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 1265 previous similar messages Lustre: 28020:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 6/6/0, ref_del: 0/0/0 Lustre: 28020:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 1265 previous similar messages Lustre: mdt00_035: service thread pid 12836 was inactive for 40.069 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 12836, comm: mdt00_035 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_fini+0xabc/0xfd0 [ptlrpc] [<0>] ldlm_cli_enqueue+0x461/0xb00 [ptlrpc] [<0>] osp_md_object_lock+0x151/0x2f0 [osp] [<0>] lod_object_lock+0xdb/0x7c0 [lod] [<0>] mdd_object_lock+0x2d/0xd0 [mdd] [<0>] mdt_remote_object_lock_try+0x14c/0x189 [mdt] [<0>] mdt_object_lock_internal+0x3c4/0x470 [mdt] [<0>] mdt_object_lock+0x88/0x1c0 [mdt] [<0>] mdt_reint_link+0x8d4/0xf40 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x73c/0xbb0 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x257/0xcd0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 17291, comm: mdt00_011 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock+0x88/0x1c0 [mdt] [<0>] mdt_intent_getxattr+0x80/0x330 [mdt] [<0>] mdt_intent_opc.constprop.74+0x211/0xc60 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x397/0x980 [ptlrpc] [<0>] ldlm_handle_enqueue+0x547/0x1900 [ptlrpc] [<0>] tgt_enqueue+0x68/0x240 [ptlrpc] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x257/0xcd0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 12786, comm: mdt00_003 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock_try+0xa0/0x250 [mdt] [<0>] mdt_getattr_name_lock+0x1729/0x2cc0 [mdt] [<0>] mdt_intent_getattr+0x2cc/0x4e0 [mdt] [<0>] mdt_intent_opc.constprop.74+0x211/0xc60 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x397/0x980 [ptlrpc] [<0>] ldlm_handle_enqueue+0x547/0x1900 [ptlrpc] [<0>] tgt_enqueue+0x68/0x240 [ptlrpc] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x257/0xcd0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: mdt00_017: service thread pid 17405 was inactive for 40.142 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: mdt00_028: service thread pid 24463 was inactive for 40.144 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: mdt00_007: service thread pid 17237 was inactive for 40.086 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 4 previous similar messages Lustre: mdt_io00_012: service thread pid 28020 was inactive for 40.123 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 2 previous similar messages Lustre: mdt00_012: service thread pid 17301 was inactive for 72.040 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. LustreError: 11132:0:(ldlm_lockd.c:257:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0002_UUID lock: ffff8802a4953100/0xf9603e4943377c56 lrc: 3/0,0 mode: PR/PR res: [0x280000403:0x70c:0x0].0x0 bits 0x1b/0x0 rrc: 14 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0xf9603e4943377c10 expref: 434 pid: 24461 timeout: 318 lvb_type: 0 Lustre: mdt00_032: service thread pid 11997 completed after 99.592s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_007: service thread pid 17237 completed after 99.494s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_035: service thread pid 12836 completed after 100.375s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_011: service thread pid 17291 completed after 100.324s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_028: service thread pid 24463 completed after 100.067s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: lustre-MDT0002-mdc-ffff8800a731a548: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-MDT0002-mdc-ffff8800a731a548: Connection to lustre-MDT0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: lustre-MDT0002-mdc-ffff8800a731a548: This client was evicted by lustre-MDT0002; in progress operations using this service will fail. Lustre: mdt00_003: service thread pid 12786 completed after 100.135s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_017: service thread pid 17405 completed after 100.084s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_031: service thread pid 11996 completed after 99.648s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_002: service thread pid 11485 completed after 100.040s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_015: service thread pid 17391 completed after 98.994s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_005: service thread pid 16978 completed after 99.399s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_038: service thread pid 20854 completed after 99.887s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 13845:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a731a548: inode [0x280000404:0xdc:0x0] mdc close failed: rc = -108 LustreError: 8219:0:(file.c:6143:ll_inode_revalidate_fini()) lustre: revalidate FID [0x280000403:0x70c:0x0] error: rc = -5 Lustre: lustre-MDT0002-mdc-ffff8800a731a548: Connection restored to 192.168.123.82@tcp (at 0@lo) Lustre: dir [0x280000404:0x502:0x0] stripe 2 readdir failed: -2, directory is partially accessed! Lustre: Skipped 8 previous similar messages Lustre: mdt_io00_012: service thread pid 28020 completed after 88.139s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_036: service thread pid 13888 completed after 87.823s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_010: service thread pid 17265 completed after 87.739s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_033: service thread pid 12018 completed after 87.777s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_030: service thread pid 5363 completed after 87.832s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_034: service thread pid 12029 completed after 87.819s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_012: service thread pid 17301 completed after 87.822s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_021: service thread pid 20321 completed after 87.748s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_014: service thread pid 17318 completed after 87.774s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 10320:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000404:0x9f1:0x0]: rc = -5 LustreError: 10320:0:(lcommon_cl.c:188:cl_file_inode_init()) Skipped 21 previous similar messages LustreError: 10320:0:(llite_lib.c:3700:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 10320:0:(llite_lib.c:3700:ll_prep_inode()) Skipped 21 previous similar messages 3[13921]: segfault at 8 ip 00007fa1d65fd7e8 sp 00007ffc77523a60 error 4 in ld-2.17.so[7fa1d65f2000+22000] LustreError: 26020:0:(mdt_reint.c:2540:mdt_reint_migrate()) lustre-MDT0001: migrate [0x280000404:0x798:0x0]/12 failed: rc = -2 LustreError: 26020:0:(mdt_reint.c:2540:mdt_reint_migrate()) Skipped 10 previous similar messages LustreError: 17922:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0002: failed to get lu_attr of [0x280000403:0x6d6:0x0]: rc = -2 LustreError: 17922:0:(mdd_object.c:3901:mdd_close()) Skipped 3 previous similar messages Lustre: 23563:0:(lod_lov.c:1414:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000403:0x8e8:0x0] with magic=0xbd60bd0 Lustre: 23563:0:(lod_lov.c:1414:lod_parse_striping()) Skipped 1 previous similar message Lustre: 13833:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0002: opcode 7: before 516 < left 618, rollback = 7 Lustre: 13833:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 29 previous similar messages Lustre: 24461:0:(lod_lov.c:1414:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0xaf3:0x0] with magic=0xbd60bd0 Lustre: 24461:0:(lod_lov.c:1414:lod_parse_striping()) Skipped 1 previous similar message traps: 8[17805] trap invalid opcode ip:404c26 sp:7ffeaffad6f8 error:0 in 8[400000+6000] LustreError: 18581:0:(mdd_dir.c:4743:mdd_migrate_cmd_check()) lustre-MDD0000: '15' migration was interrupted, run 'lfs migrate -m 2 -c 3 -H crush 15' to finish migration: rc = -1 LustreError: 18581:0:(mdd_dir.c:4743:mdd_migrate_cmd_check()) Skipped 3 previous similar messages Lustre: 20858:0:(lod_lov.c:1414:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000404:0xb48:0x0] with magic=0xbd60bd0 Lustre: 20858:0:(lod_lov.c:1414:lod_parse_striping()) Skipped 3 previous similar messages LustreError: 17391:0:(mdt_open.c:1302:mdt_cross_open()) lustre-MDT0000: [0x200000403:0x26:0x0] doesn't exist!: rc = -14 3[24909]: segfault at 406000 ip 0000000000406000 sp 00007ffd9b5ad0d8 error 14 in 3[606000+1000] Lustre: 5649:0:(out_handler.c:879:out_tx_end()) lustre-MDT0002-osd: error during execution of #1 from /home/green/git/lustre-release/lustre/ptlrpc/../../lustre/target/out_handler.c:433: rc = -61 LustreError: 5649:0:(out_handler.c:889:out_tx_end()) lustre-MDT0002-osd: undo for /home/green/git/lustre-release/lustre/ptlrpc/../../lustre/target/out_handler.c:433: rc = -524 1[25307]: segfault at 8 ip 00007f6fe68777e8 sp 00007ffcc7e59cf0 error 4 in ld-2.17.so[7f6fe686c000+22000] Lustre: 24505:0:(lod_lov.c:1414:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0xd00:0x0] with magic=0xbd60bd0 Lustre: 24505:0:(lod_lov.c:1414:lod_parse_striping()) Skipped 1 previous similar message Lustre: dir [0x200000403:0xd88:0x0] stripe 3 readdir failed: -2, directory is partially accessed! Lustre: Skipped 8 previous similar messages 18[31771]: segfault at 0 ip 0000000000403cf0 sp 00007ffe6d6da738 error 6 in 18[400000+6000] 16[32075]: segfault at 8 ip 00007fb3756b87e8 sp 00007fff2d8bb6e0 error 4 in ld-2.17.so[7fb3756ad000+22000] Lustre: 2687:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 504 < left 1990, rollback = 2 Lustre: 2687:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 489 previous similar messages Lustre: 2687:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 3/12/7, destroy: 1/4/0 Lustre: 2687:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 517 previous similar messages Lustre: 2687:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 4/4/0, xattr_set: 25/1990/0 Lustre: 2687:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 517 previous similar messages Lustre: 2687:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 5/55/0, punch: 0/0/0, quota 7/129/0 Lustre: 2687:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 517 previous similar messages Lustre: 2687:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 15/263/4, delete: 3/6/0 Lustre: 2687:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 517 previous similar messages Lustre: 2687:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 8/8/0, ref_del: 3/3/0 Lustre: 2687:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 517 previous similar messages Lustre: 31575:0:(mdd_dir.c:4822:mdd_migrate_object()) lustre-MDD0002: [0x280000403:0x1:0x0]/3 is open, migrate only dentry Lustre: 31575:0:(mdd_dir.c:4822:mdd_migrate_object()) Skipped 15 previous similar messages 9[3129]: segfault at 8 ip 00007f103ff1b7e8 sp 00007ffc7b77e730 error 4 in ld-2.17.so[7f103ff10000+22000] LustreError: 211:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 14 [0x200000403:0xe88:0x0] inode@0000000000000000: rc = -5 Lustre: 17405:0:(lod_lov.c:1414:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000405:0x30a:0x0] with magic=0xbd60bd0 Lustre: 17405:0:(lod_lov.c:1414:lod_parse_striping()) Skipped 5 previous similar messages 7[14829]: segfault at 8 ip 00007f8bd745f7e8 sp 00007ffdf2a8ef00 error 4 in ld-2.17.so[7f8bd7454000+22000] 2[15948]: segfault at 8 ip 00007fb2bd7f47e8 sp 00007ffd5c7edae0 error 4 in ld-2.17.so[7fb2bd7e9000+22000] 1[15973]: segfault at 8 ip 00007efcaa27c7e8 sp 00007ffec79e2990 error 4 in ld-2.17.so[7efcaa271000+22000] 14[17294]: segfault at 8 ip 00007fe6cdd902fc sp 00007ffc6dd24fe0 error 4 in ld-2.17.so[7fe6cdd83000+22000] 6[17636]: segfault at 406000 ip 0000000000406000 sp 00007ffccdc8c478 error 14 in 6[606000+1000] 0[18540]: segfault at 8 ip 00007f8f81d8f7e8 sp 00007ffcf9f25b60 error 4 in ld-2.17.so[7f8f81d84000+22000] LustreError: 10567:0:(statahead.c:2399:start_statahead_thread()) lustre: invalid pattern 0X0. 11[20757]: segfault at 8 ip 00007ffae4fa37e8 sp 00007ffe7a29f270 error 4 in ld-2.17.so[7ffae4f98000+22000] Lustre: 28020:0:(mdt_reint.c:2460:mdt_reint_migrate()) lustre-MDT0000: [0x200000403:0x1:0x0]/3 is open, migrate only dentry 8[22810]: segfault at 8 ip 00007ff6459f87e8 sp 00007fff364e6910 error 4 in ld-2.17.so[7ff6459ed000+22000] 17[23465]: segfault at 404594 ip 0000000000404c98 sp 00007ffd169bb7c8 error 7 in 17[400000+6000] LustreError: 27432:0:(mdd_dir.c:4743:mdd_migrate_cmd_check()) lustre-MDD0001: '10' migration was interrupted, run 'lfs migrate -m 2 -c 2 -H crush 10' to finish migration: rc = -1 LustreError: 27432:0:(mdd_dir.c:4743:mdd_migrate_cmd_check()) Skipped 2 previous similar messages 19[24977]: segfault at 8 ip 00007fe25b72e7e8 sp 00007ffc2e73a3d0 error 4 in ld-2.17.so[7fe25b723000+22000] 12[25298]: segfault at 8 ip 00007f4d5f94b7e8 sp 00007ffd1696f7d0 error 4 in ld-2.17.so[7f4d5f940000+22000] 14[25008]: segfault at 8 ip 00007f713c29d7e8 sp 00007ffe18fe9be0 error 4 in ld-2.17.so[7f713c292000+22000] LustreError: 24615:0:(lov_object.c:1348:lov_layout_change()) lustre-clilov-ffff8800a731a548: cannot apply new layout on [0x200000403:0x1039:0x0] : rc = -5 LustreError: 24615:0:(vvp_io.c:1905:vvp_io_init()) lustre: refresh file layout [0x200000403:0x1039:0x0] error -5. Lustre: 15364:0:(lod_lov.c:1414:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000403:0xb6f:0x0] with magic=0xbd60bd0 Lustre: 15364:0:(lod_lov.c:1414:lod_parse_striping()) Skipped 39 previous similar messages LustreError: 11487:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0002: failed to get lu_attr of [0x280000403:0xba9:0x0]: rc = -2 LustreError: 11487:0:(mdd_object.c:3901:mdd_close()) Skipped 5 previous similar messages 7[26183]: segfault at 8 ip 00007f52aa85a7e8 sp 00007fff3c8879f0 error 4 in ld-2.17.so[7f52aa84f000+22000] LustreError: 28422:0:(lov_object.c:1348:lov_layout_change()) lustre-clilov-ffff8800a70f4138: cannot apply new layout on [0x280000405:0x63a:0x0] : rc = -5 LustreError: 28422:0:(vvp_io.c:1905:vvp_io_init()) lustre: refresh file layout [0x280000405:0x63a:0x0] error -5. LustreError: 20264:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0000: [0x200000404:0x12c9:0x0] migrate mdt count mismatch 3 != 2 18[32034]: segfault at 8 ip 00007f0ee8bdf7e8 sp 00007fff15901de0 error 4 in ld-2.17.so[7f0ee8bd4000+22000] 18[31857]: segfault at 8 ip 00007f9dc96bb7e8 sp 00007ffc3b7d0e40 error 4 in ld-2.17.so[7f9dc96b0000+22000] 8[32766]: segfault at 8 ip 00007eff8f7f07e8 sp 00007ffdac8d1ec0 error 4 in ld-2.17.so[7eff8f7e5000+22000] LustreError: 20264:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0001: [0x240000403:0xe25:0x0] migrate mdt count mismatch 2 != 1 14[4324]: segfault at 8 ip 00007fb1e05a47e8 sp 00007ffcd3eae530 error 4 in ld-2.17.so[7fb1e0599000+22000] LustreError: 17318:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0002: [0x280000403:0xdde:0x0] migrate mdt count mismatch 1 != 3 14[3127]: segfault at 8 ip 00007f7cbd7df7e8 sp 00007ffc7e9aca30 error 4 in ld-2.17.so[7f7cbd7d4000+22000] Lustre: dir [0x280000403:0xe2c:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 17 previous similar messages LustreError: 6793:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000403:0xf70:0x0]: rc = -5 LustreError: 6793:0:(lcommon_cl.c:188:cl_file_inode_init()) Skipped 77 previous similar messages LustreError: 6793:0:(llite_lib.c:3700:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 6793:0:(llite_lib.c:3700:ll_prep_inode()) Skipped 77 previous similar messages LustreError: 3163:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a731a548: inode [0x200000404:0x12be:0x0] mdc close failed: rc = -2 LustreError: 3163:0:(file.c:247:ll_close_inode_openhandle()) Skipped 37 previous similar messages 6[8635]: segfault at 8 ip 00007fa65775d7e8 sp 00007ffc49b617c0 error 4 in ld-2.17.so[7fa657752000+22000] Lustre: 21775:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0000: opcode 7: before 516 < left 528, rollback = 7 Lustre: 21775:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 133 previous similar messages 6[8909]: segfault at 8 ip 00007f82ae9587e8 sp 00007ffc353ea330 error 4 in ld-2.17.so[7f82ae94d000+22000] 2[9761]: segfault at 8 ip 00007f8cfbf9d7e8 sp 00007fff0dd13280 error 4 in ld-2.17.so[7f8cfbf92000+22000] LustreError: 27432:0:(mdt_reint.c:2540:mdt_reint_migrate()) lustre-MDT0001: migrate [0x200000403:0x1:0x0]/1 failed: rc = -1 LustreError: 27432:0:(mdt_reint.c:2540:mdt_reint_migrate()) Skipped 30 previous similar messages 16[10702]: segfault at 1c ip 00000000004017b0 sp 00007ffd698503f0 error 6 in 16[400000+6000] 8[12521]: segfault at 8 ip 00007f49184f77e8 sp 00007ffee4bd22a0 error 4 in ld-2.17.so[7f49184ec000+22000] LustreError: 20343:0:(mdt_open.c:1302:mdt_cross_open()) lustre-MDT0000: [0x200000403:0x26:0x0] doesn't exist!: rc = -14 LustreError: 171:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 11 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 LustreError: 171:0:(statahead.c:825:ll_statahead_interpret_work()) Skipped 1 previous similar message 5[20013]: segfault at 8 ip 00007f80eb8117e8 sp 00007ffe680edef0 error 4 in ld-2.17.so[7f80eb806000+22000] LustreError: 20350:0:(mdt_open.c:1302:mdt_cross_open()) lustre-MDT0000: [0x200000403:0x26:0x0] doesn't exist!: rc = -14 Lustre: 17301:0:(lod_lov.c:1414:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000405:0xae3:0x0] with magic=0xbd60bd0 Lustre: 17301:0:(lod_lov.c:1414:lod_parse_striping()) Skipped 19 previous similar messages 17[21773]: segfault at 8 ip 00007f0a2d8117e8 sp 00007ffe1c70e5c0 error 4 in ld-2.17.so[7f0a2d806000+22000] LustreError: 127:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 14 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 LustreError: 127:0:(statahead.c:825:ll_statahead_interpret_work()) Skipped 2 previous similar messages LustreError: 22100:0:(lov_object.c:1348:lov_layout_change()) lustre-clilov-ffff8800a70f4138: cannot apply new layout on [0x240000404:0x1156:0x0] : rc = -5 LustreError: 22100:0:(lov_object.c:1348:lov_layout_change()) Skipped 2 previous similar messages LustreError: 22100:0:(vvp_io.c:1905:vvp_io_init()) lustre: refresh file layout [0x240000404:0x1156:0x0] error -5. LustreError: 22100:0:(vvp_io.c:1905:vvp_io_init()) Skipped 2 previous similar messages LustreError: 11483:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0001: [0x240000404:0x102f:0x0] migrate mdt count mismatch 1 != 3 0[27521]: segfault at 8 ip 00007f92dc9037e8 sp 00007fff4b1f23c0 error 4 in ld-2.17.so[7f92dc8f8000+22000] LustreError: 435:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 20 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 0[30130]: segfault at 8 ip 00007f8bdd7487e8 sp 00007ffcb3f56b70 error 4 in ld-2.17.so[7f8bdd73d000+22000] 0[31860]: segfault at 8 ip 00007f76c6ecf7e8 sp 00007fff594b8b70 error 4 in ld-2.17.so[7f76c6ec4000+22000] LustreError: 1637:0:(lov_object.c:1348:lov_layout_change()) lustre-clilov-ffff8800a731a548: cannot apply new layout on [0x200000403:0x17a5:0x0] : rc = -5 LustreError: 1637:0:(lov_object.c:1348:lov_layout_change()) Skipped 1 previous similar message LustreError: 26533:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 18 [0x200000404:0x1470:0x0] inode@0000000000000000: rc = -5 LustreError: 26533:0:(statahead.c:825:ll_statahead_interpret_work()) Skipped 1 previous similar message LustreError: 2245:0:(llite_nfs.c:430:ll_dir_get_parent_fid()) lustre: failure inode [0x240000404:0x1064:0x0] get parent: rc = -116 traps: 15[6596] trap invalid opcode ip:404bae sp:7ffed7fe1508 error:0 in 15[400000+6000] 17[7085]: segfault at 8 ip 00007f44c951b7e8 sp 00007ffce880bae0 error 4 in ld-2.17.so[7f44c9510000+22000] LustreError: 434:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 7 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 LustreError: 434:0:(statahead.c:825:ll_statahead_interpret_work()) Skipped 4 previous similar messages 16[13813]: segfault at 1c ip 00000000004017b0 sp 00007ffc7352e580 error 6 in 11[400000+6000] LustreError: 15142:0:(lov_object.c:1348:lov_layout_change()) lustre-clilov-ffff8800a731a548: cannot apply new layout on [0x200000403:0x17a5:0x0] : rc = -5 LustreError: 15142:0:(lov_object.c:1348:lov_layout_change()) Skipped 3 previous similar messages LustreError: 11497:0:(mdd_dir.c:4743:mdd_migrate_cmd_check()) lustre-MDD0000: '19' migration was interrupted, run 'lfs migrate -m 2 -c 2 -H crush 19' to finish migration: rc = -1 LustreError: 11497:0:(mdd_dir.c:4743:mdd_migrate_cmd_check()) Skipped 20 previous similar messages LustreError: 13047:0:(vvp_io.c:1905:vvp_io_init()) lustre: refresh file layout [0x240000404:0x16e2:0x0] error -5. LustreError: 13047:0:(vvp_io.c:1905:vvp_io_init()) Skipped 1 previous similar message LustreError: 20849:0:(mdt_open.c:1302:mdt_cross_open()) lustre-MDT0000: [0x200000403:0x26:0x0] doesn't exist!: rc = -14 18[18313]: segfault at 8 ip 00007f60182d97e8 sp 00007ffc98be56f0 error 4 in ld-2.17.so[7f60182ce000+22000] LustreError: 11132:0:(ldlm_lockd.c:257:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0000_UUID lock: ffff88009e1b25c0/0xf9603e49438f46fc lrc: 3/0,0 mode: PR/PR res: [0x200000404:0x1df6:0x0].0x0 bits 0x1b/0x0 rrc: 4 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0xf9603e49438f46e0 expref: 1149 pid: 17291 timeout: 643 lvb_type: 0 Lustre: 11497:0:(mdd_dir.c:4822:mdd_migrate_object()) lustre-MDD0000: [0x200000403:0x1:0x0]/15 is open, migrate only dentry Lustre: 11497:0:(mdd_dir.c:4822:mdd_migrate_object()) Skipped 95 previous similar messages LustreError: lustre-MDT0000-mdc-ffff8800a731a548: operation mds_reint to node 0@lo failed: rc = -107 Lustre: lustre-MDT0000-mdc-ffff8800a731a548: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: lustre-MDT0000-mdc-ffff8800a731a548: This client was evicted by lustre-MDT0000; in progress operations using this service will fail. LustreError: 16943:0:(file.c:6143:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000404:0x1df6:0x0] error: rc = -108 Lustre: lustre-MDT0000-mdc-ffff8800a731a548: Connection restored to 192.168.123.82@tcp (at 0@lo) Lustre: 18581:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 506 < left 699, rollback = 2 Lustre: 18581:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 3571 previous similar messages Lustre: 18581:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 3/12/5, destroy: 0/0/0 Lustre: 18581:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 3771 previous similar messages Lustre: 18581:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 2/2/0, xattr_set: 9/699/0 Lustre: 18581:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 3771 previous similar messages Lustre: 18581:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 5/55/0, punch: 0/0/0, quota 4/54/0 Lustre: 18581:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 3771 previous similar messages Lustre: 18581:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 12/275/4, delete: 0/0/0 Lustre: 18581:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 3771 previous similar messages Lustre: 18581:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 6/6/0, ref_del: 0/0/0 Lustre: 18581:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 3771 previous similar messages LustreError: 15799:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0001: failed to get lu_attr of [0x240000403:0xeab:0x0]: rc = -2 LustreError: 15799:0:(mdd_object.c:3901:mdd_close()) Skipped 13 previous similar messages 2[18461]: segfault at 8 ip 00007f40a54e37e8 sp 00007fff2704af10 error 4 in ld-2.17.so[7f40a54d8000+22000] Lustre: 28657:0:(lod_lov.c:1414:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000403:0x1490:0x0] with magic=0xbd60bd0 Lustre: 28657:0:(lod_lov.c:1414:lod_parse_striping()) Skipped 39 previous similar messages 15[21362]: segfault at 406000 ip 0000000000406000 sp 00007ffd2a3a29a8 error 14 in 15[606000+1000] LustreError: 20656:0:(lov_object.c:1348:lov_layout_change()) lustre-clilov-ffff8800a70f4138: cannot apply new layout on [0x240000404:0x16e2:0x0] : rc = -5 LustreError: 20656:0:(lov_object.c:1348:lov_layout_change()) Skipped 4 previous similar messages LustreError: 23888:0:(vvp_io.c:1905:vvp_io_init()) lustre: refresh file layout [0x280000405:0x117a:0x0] error -5. 6[24637]: segfault at 8 ip 00007fee013db7e8 sp 00007ffd70216bb0 error 4 in ld-2.17.so[7fee013d0000+22000] LustreError: 20854:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0001: [0x240000403:0x1720:0x0] migrate mdt count mismatch 2 != 3 LustreError: 20854:0:(mdt_xattr.c:402:mdt_dir_layout_update()) Skipped 1 previous similar message LustreError: 11484:0:(mdd_object.c:391:mdd_xattr_get()) lustre-MDD0002: object [0x280000403:0x14f3:0x0] not found: rc = -2 7[30499]: segfault at 8 ip 00007f68bfa2e7e8 sp 00007ffeb3def490 error 4 in ld-2.17.so[7f68bfa23000+22000] 8[31072]: segfault at 8 ip 00007fa0f2ea87e8 sp 00007ffd0adf6910 error 4 in ld-2.17.so[7fa0f2e9d000+22000] 8[31353]: segfault at 8 ip 00007faf5181f7e8 sp 00007ffd090f7890 error 4 in ld-2.17.so[7faf51814000+22000] Lustre: 13833:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0002: opcode 7: before 515 < left 594, rollback = 7 Lustre: 13833:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 129 previous similar messages Lustre: lustre-OST0002-osc-ffff8800a731a548: disconnect after 21s idle Lustre: lustre-OST0003-osc-ffff8800a70f4138: disconnect after 23s idle LustreError: 11132:0:(ldlm_lockd.c:257:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0000_UUID lock: ffff88029a4de940/0xf9603e49439bd067 lrc: 3/0,0 mode: PR/PR res: [0x200000405:0x22f:0x0].0x0 bits 0x1b/0x0 rrc: 4 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0xf9603e49439bd02f expref: 238 pid: 11483 timeout: 775 lvb_type: 0 LustreError: lustre-MDT0000-mdc-ffff8800a731a548: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: 28402:0:(lov_object.c:1348:lov_layout_change()) lustre-clilov-ffff8800a731a548: cannot apply new layout on [0x280000403:0x15a9:0x0] : rc = -5 LustreError: 28402:0:(lov_object.c:1348:lov_layout_change()) Skipped 7 previous similar messages LustreError: 28402:0:(vvp_io.c:1905:vvp_io_init()) lustre: refresh file layout [0x280000403:0x15a9:0x0] error -5. LustreError: 28402:0:(vvp_io.c:1905:vvp_io_init()) Skipped 2 previous similar messages Lustre: lustre-MDT0000-mdc-ffff8800a731a548: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: lustre-MDT0000-mdc-ffff8800a731a548: This client was evicted by lustre-MDT0000; in progress operations using this service will fail. LustreError: 29585:0:(llite_nfs.c:430:ll_dir_get_parent_fid()) lustre: failure inode [0x200000405:0x2fe:0x0] get parent: rc = -108 LustreError: 30798:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a731a548: inode [0x200000405:0x219:0x0] mdc close failed: rc = -108 LustreError: 30798:0:(file.c:247:ll_close_inode_openhandle()) Skipped 57 previous similar messages LustreError: 30664:0:(llite_lib.c:1996:ll_md_setattr()) md_setattr fails: rc = -108 LustreError: Skipped 1 previous similar message LustreError: 31128:0:(file.c:6143:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000405:0x2fe:0x0] error: rc = -107 LustreError: 31128:0:(file.c:6143:ll_inode_revalidate_fini()) Skipped 3 previous similar messages Lustre: lustre-MDT0000-mdc-ffff8800a731a548: Connection restored to 192.168.123.82@tcp (at 0@lo) LustreError: 31170:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x280000405:0x12c2:0x0]: rc = -5 LustreError: 31170:0:(lcommon_cl.c:188:cl_file_inode_init()) Skipped 203 previous similar messages LustreError: 31170:0:(llite_lib.c:3700:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 31170:0:(llite_lib.c:3700:ll_prep_inode()) Skipped 203 previous similar messages LustreError: 2004:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 19 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 Lustre: dir [0x240000404:0x1bf6:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 14 previous similar messages LustreError: 20343:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0001: [0x240000403:0x1827:0x0] migrate mdt count mismatch 1 != 2 10[4710]: segfault at 8 ip 00007f5f6f7e17e8 sp 00007ffe87ae8060 error 4 in ld-2.17.so[7f5f6f7d6000+22000] 17[4490]: segfault at 8 ip 00007ff9452f37e8 sp 00007fffa554b140 error 4 in ld-2.17.so[7ff9452e8000+22000] 9[6711]: segfault at 8 ip 00007f2abb2157e8 sp 00007ffdf8ee5900 error 4 in ld-2.17.so[7f2abb20a000+22000] 11[6919]: segfault at 1c ip 00000000004017b0 sp 00007ffc9d5921a0 error 6 in 11[400000+6000] 17[7561]: segfault at 8 ip 00007fbdbf0577e8 sp 00007ffe8a435b70 error 4 in ld-2.17.so[7fbdbf04c000+22000] 17[7582]: segfault at 8 ip 00007f8165f957e8 sp 00007ffd6a732880 error 4 in ld-2.17.so[7f8165f8a000+22000] 16[11648]: segfault at 8 ip 00007f2fad6a67e8 sp 00007ffe09867750 error 4 in ld-2.17.so[7f2fad69b000+22000] 6[17566]: segfault at 8 ip 00007f671cdca7e8 sp 00007fff99b642d0 error 4 in ld-2.17.so[7f671cdbf000+22000] 7[18921]: segfault at 8 ip 00007f1b4c8e87e8 sp 00007fff5fbf2940 error 4 in ld-2.17.so[7f1b4c8dd000+22000] 4[18819]: segfault at 8 ip 00007f637bc737e8 sp 00007ffeb8843940 error 4 in ld-2.17.so[7f637bc68000+22000] LustreError: 27432:0:(mdt_reint.c:2540:mdt_reint_migrate()) lustre-MDT0000: migrate [0x200000403:0x1:0x0]/19 failed: rc = -2 LustreError: 27432:0:(mdt_reint.c:2540:mdt_reint_migrate()) Skipped 49 previous similar messages 2[20436]: segfault at 8 ip 00007f4ded15f7e8 sp 00007ffddaaa47a0 error 4 in ld-2.17.so[7f4ded154000+22000] Lustre: 27342:0:(mdt_reint.c:2460:mdt_reint_migrate()) lustre-MDT0000: [0x200000403:0x2:0x0]/3 is open, migrate only dentry Lustre: 27342:0:(mdt_reint.c:2460:mdt_reint_migrate()) Skipped 1 previous similar message LustreError: 17302:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0001: [0x240000403:0x1c5e:0x0] migrate mdt count mismatch 2 != 1 18[25462]: segfault at 8 ip 00007f8a2ac8a7e8 sp 00007ffe540d4300 error 4 in ld-2.17.so[7f8a2ac7f000+22000] LustreError: 20991:0:(mdd_dir.c:4743:mdd_migrate_cmd_check()) lustre-MDD0001: '0' migration was interrupted, run 'lfs migrate -m 0 -c 3 -H crush 0' to finish migration: rc = -1 LustreError: 20991:0:(mdd_dir.c:4743:mdd_migrate_cmd_check()) Skipped 7 previous similar messages 19[27277]: segfault at 8 ip 00007efcb74be7e8 sp 00007ffe911512b0 error 4 in ld-2.17.so[7efcb74b3000+22000] 3[30212]: segfault at 8 ip 00007f42b75887e8 sp 00007ffef9e19410 error 4 3[30261]: segfault at 8 ip 00007f15b180f7e8 sp 00007ffeadac41f0 error 4 in ld-2.17.so[7f15b1804000+22000] in ld-2.17.so[7f42b757d000+22000] 7[30267]: segfault at 8 ip 00007fbe6a4e27e8 sp 00007fffc0d91e80 error 4 in ld-2.17.so[7fbe6a4d7000+22000] 15[32594]: segfault at 8 ip 00007f7862ab57e8 sp 00007ffd457b0da0 error 4 in ld-2.17.so[7f7862aaa000+22000] 19[3839]: segfault at 8 ip 00007f21879037e8 sp 00007ffee14ccd80 error 4 in ld-2.17.so[7f21878f8000+22000] LustreError: 12029:0:(mdt_open.c:1302:mdt_cross_open()) lustre-MDT0000: [0x200000403:0x26:0x0] doesn't exist!: rc = -14 8[5044]: segfault at 8 ip 00007f6a7358b7e8 sp 00007ffd1104c270 error 4 in ld-2.17.so[7f6a73580000+22000] LustreError: 3453:0:(vvp_io.c:1905:vvp_io_init()) lustre: refresh file layout [0x200000403:0x27be:0x0] error -5. 17[7305]: segfault at 8 ip 00007fe05979b7e8 sp 00007ffda28dc450 error 4 in ld-2.17.so[7fe059790000+22000] traps: 17[7952] trap invalid opcode ip:4059b4 sp:7ffde90be0a8 error:0 in 17[400000+6000] 12[9052]: segfault at 0 ip 0000000000403e5f sp 00007ffc0cf49620 error 6 in 10[400000+6000] 19[7707]: segfault at 0 ip 0000000000403cf0 sp 00007ffc5b4b1118 error 6 in 19[400000+6000] 7[10114]: segfault at 1c ip 0000000000401360 sp 00007ffce2b959e8 error 6 in 14[400000+6000] Lustre: 31575:0:(mdt_reint.c:2460:mdt_reint_migrate()) lustre-MDT0001: [0x240000403:0x1:0x0]/13 is open, migrate only dentry LustreError: 14474:0:(lov_object.c:1348:lov_layout_change()) lustre-clilov-ffff8800a70f4138: cannot apply new layout on [0x200000406:0x761:0x0] : rc = -5 LustreError: 14474:0:(lov_object.c:1348:lov_layout_change()) Skipped 6 previous similar messages LustreError: 17995:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0002: failed to get lu_attr of [0x280000405:0x19ba:0x0]: rc = -2 LustreError: 17995:0:(mdd_object.c:3901:mdd_close()) Skipped 19 previous similar messages LustreError: 17262:0:(mdt_open.c:1302:mdt_cross_open()) lustre-MDT0000: [0x200000403:0x26:0x0] doesn't exist!: rc = -14 LustreError: 17262:0:(mdt_open.c:1302:mdt_cross_open()) Skipped 2 previous similar messages LustreError: 11485:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0000: [0x200000406:0x6ca:0x0] migrate mdt count mismatch 1 != 3 LustreError: 11485:0:(mdt_xattr.c:402:mdt_dir_layout_update()) Skipped 1 previous similar message 7[21082]: segfault at 8 ip 00007f73e344d7e8 sp 00007fffccea57d0 error 4 in ld-2.17.so[7f73e3442000+22000] Lustre: 20340:0:(lod_lov.c:1414:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000403:0x1fb1:0x0] with magic=0xbd60bd0 Lustre: 20340:0:(lod_lov.c:1414:lod_parse_striping()) Skipped 55 previous similar messages 17[26415]: segfault at 8 ip 00007ff18621e7e8 sp 00007fff16c48540 error 4 in ld-2.17.so[7ff186213000+22000] 13[28141]: segfault at 8 ip 00007ffa51a7c7e8 sp 00007ffd653353e0 error 4 in ld-2.17.so[7ffa51a71000+22000] 1[10047]: segfault at 8 ip 00007ff9952e47e8 sp 00007ffcb7551680 error 4 in ld-2.17.so[7ff9952d9000+22000] 3[11149]: segfault at 8 ip 00007f6466a6a7e8 sp 00007ffd0b881380 error 4 in ld-2.17.so[7f6466a5f000+22000] LustreError: 241:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 17 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 1[15169]: segfault at 0 ip (null) sp 00007fffde7d3598 error 14 in 1[400000+6000] 17[18577]: segfault at 1c ip 00000000004017b0 sp 00007ffdc617f550 error 6 in 17[400000+6000] LustreError: 20407:0:(vvp_io.c:1905:vvp_io_init()) lustre: refresh file layout [0x200000403:0x3056:0x0] error -5. LustreError: 20407:0:(vvp_io.c:1905:vvp_io_init()) Skipped 6 previous similar messages 19[26446]: segfault at 8 ip 00007f2a21b687e8 sp 00007ffe7c8c0430 error 4 in ld-2.17.so[7f2a21b5d000+22000] 10[26123]: segfault at 8 ip 00007f1a445497e8 sp 00007fff065a7380 error 4 in ld-2.17.so[7f1a4453e000+22000] LustreError: 24174:0:(llite_lib.c:1845:ll_update_lsm_md()) lustre: [0x200000403:0x3161:0x0] dir layout mismatch: LustreError: 24174:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=2 count=1 index=0 hash=crush:0x2000003 max_inherit=0 max_inherit_rr=0 version=2 migrate_offset=0 migrate_hash=invalid:0 pool= LustreError: 24174:0:(lustre_lmv.h:167:lmv_stripe_object_dump()) stripe[0] [0x200000400:0xbf:0x0] LustreError: 24174:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=1 count=2 index=0 hash=crush:0x82000003 max_inherit=0 max_inherit_rr=0 version=1 migrate_offset=1 migrate_hash=fnv_1a_64:2 pool= LustreError: 24191:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=1 count=1 index=0 hash=crush:0x2000003 max_inherit=0 max_inherit_rr=0 version=2 migrate_offset=0 migrate_hash=invalid:0 pool= LustreError: 24191:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=1 count=2 index=0 hash=crush:0x82000003 max_inherit=0 max_inherit_rr=0 version=1 migrate_offset=1 migrate_hash=fnv_1a_64:2 pool= 15[27185]: segfault at 0 ip 0000000000403e5f sp 00007fff731d1600 error 6 in 15[400000+6000] LustreError: 12786:0:(mdt_open.c:1302:mdt_cross_open()) lustre-MDT0000: [0x200000403:0x26:0x0] doesn't exist!: rc = -14 3[433]: segfault at 8 ip 00007f77a8dc07e8 sp 00007fffa3852770 error 4 in ld-2.17.so[7f77a8db5000+22000] 16[594]: segfault at 8 ip 00007f66255b87e8 sp 00007ffc5a1e9d60 error 4 in ld-2.17.so[7f66255ad000+22000] 0[776]: segfault at 7ffd024cf318 ip 0000000000405c3a sp 00007ffd020cacd8 error 6 in 0[400000+6000] 3[848]: segfault at 8 ip 00007f780d6a77e8 sp 00007ffc3836b470 error 4 in ld-2.17.so[7f780d69c000+22000] 2[2082]: segfault at 0 ip 0000000000403e5f sp 00007ffc42850a10 error 6 in 2[400000+6000] 2[3674]: segfault at 0 ip (null) sp 00007ffd92016e08 error 14 in 2[400000+6000] 7[3728]: segfault at 8 ip 00007f3e62c977e8 sp 00007ffdf7427450 error 4 in ld-2.17.so[7f3e62c8c000+22000] 1[5360]: segfault at 8 ip 00007faf48c137e8 sp 00007ffea2f726a0 error 4 in ld-2.17.so[7faf48c08000+22000] LustreError: 11996:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0000: [0x200000406:0x11c1:0x0] migrate mdt count mismatch 3 != 2 LustreError: 11996:0:(mdt_xattr.c:402:mdt_dir_layout_update()) Skipped 4 previous similar messages Lustre: dir [0x200000403:0x3349:0x0] stripe 2 readdir failed: -2, directory is partially accessed! Lustre: Skipped 51 previous similar messages 9[6883]: segfault at 0 ip 0000000000403cf0 sp 00007ffeb054d768 error 6 in 9[400000+6000] getfattr (7494) used greatest stack depth: 9792 bytes left 10[13402]: segfault at 8 ip 00007fc8e4eef7e8 sp 00007ffdc0dca500 error 4 in ld-2.17.so[7fc8e4ee4000+22000] 2[14524]: segfault at 8 ip 00007f59829ed7e8 sp 00007ffcc3284f80 error 4 in ld-2.17.so[7f59829e2000+22000] 2[17680]: segfault at 8 ip 00007f04ca7e17e8 sp 00007ffc3a2b6e40 error 4 in ld-2.17.so[7f04ca7d6000+22000] traps: 18[22086] trap invalid opcode ip:4054a2 sp:7ffe3aad4588 error:0 in 18[400000+6000] 1[22872]: segfault at 8 ip 00007f7e5f5697e8 sp 00007ffec9b39c70 error 4 in ld-2.17.so[7f7e5f55e000+22000] 8[23014]: segfault at 406000 ip 0000000000406000 sp 00007ffce302f6b8 error 14 in 8[606000+1000] 1[23585]: segfault at 8 ip 00007f553c2007e8 sp 00007ffd54b8c720 error 4 in ld-2.17.so[7f553c1f5000+22000] Lustre: 26020:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 505 < left 877, rollback = 2 Lustre: 26020:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 6350 previous similar messages Lustre: 26020:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 3/12/6, destroy: 0/0/0 Lustre: 26020:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 6646 previous similar messages Lustre: 26020:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 2/2/0, xattr_set: 11/877/0 Lustre: 26020:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 6646 previous similar messages Lustre: 26020:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 5/55/0, punch: 0/0/0, quota 1/3/0 Lustre: 26020:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 6646 previous similar messages Lustre: 26020:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 11/239/4, delete: 0/0/0 Lustre: 26020:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 6646 previous similar messages Lustre: 26020:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 5/5/0, ref_del: 0/0/0 Lustre: 26020:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 6646 previous similar messages Lustre: 18113:0:(mdd_dir.c:4822:mdd_migrate_object()) lustre-MDD0000: [0x200000403:0x337c:0x0]/15 is open, migrate only dentry Lustre: 18113:0:(mdd_dir.c:4822:mdd_migrate_object()) Skipped 165 previous similar messages ptlrpc_watchdog_fire: 15 callbacks suppressed Lustre: ll_ost_out00_00: service thread pid 5649 was inactive for 40.025 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Lustre: Skipped 2 previous similar messages Pid: 5649, comm: ll_ost_out00_00 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_pdo_lock+0x2a2/0x750 [mdt] [<0>] mdt_parent_lock+0x76/0x2a0 [mdt] [<0>] mdt_getattr_name_lock+0x1844/0x2cc0 [mdt] [<0>] mdt_intent_getattr+0x2cc/0x4e0 [mdt] [<0>] mdt_intent_opc.constprop.74+0x211/0xc60 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x397/0x980 [ptlrpc] [<0>] ldlm_handle_enqueue+0x547/0x1900 [ptlrpc] [<0>] mdt_batch_getattr+0x6c/0x190 [mdt] [<0>] mdt_batch+0x465/0x195b [mdt] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x257/0xcd0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe LustreError: 11132:0:(ldlm_lockd.c:257:expired_lock_main()) ### lock callback timer expired after 101s: evicting client at 0@lo ns: mdt-lustre-MDT0000_UUID lock: ffff8802a7d4b100/0xf9603e49440884ee lrc: 3/0,0 mode: PR/PR res: [0x200000406:0x16d1:0x0].0x0 bits 0x1b/0x0 rrc: 5 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0xf9603e49440884cb expref: 1038 pid: 16978 timeout: 1263 lvb_type: 0 LustreError: 17391:0:(ldlm_lockd.c:1447:ldlm_handle_enqueue()) ### lock on destroyed export ffff8802e73bca88 ns: mdt-lustre-MDT0000_UUID lock: ffff8802cd436580/0xf9603e4944095fe6 lrc: 3/0,0 mode: PR/PR res: [0x200000403:0x337c:0x0].0x0 bits 0x12/0x0 rrc: 21 type: IBT gid 0 flags: 0x50200400000020 nid: 0@lo remote: 0xf9603e4944095fd8 expref: 817 pid: 17391 timeout: 0 lvb_type: 0 LustreError: lustre-MDT0000-mdc-ffff8800a70f4138: operation mds_getattr_lock to node 0@lo failed: rc = -107 Lustre: lustre-MDT0000-mdc-ffff8800a70f4138: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 13543:0:(ldlm_lockd.c:2550:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1746565950 with bad export cookie 17969430997586500074 LustreError: lustre-MDT0000-mdc-ffff8800a70f4138: This client was evicted by lustre-MDT0000; in progress operations using this service will fail. LustreError: Skipped 1 previous similar message LustreError: 27404:0:(file.c:6143:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000403:0x337c:0x0] error: rc = -5 Lustre: ll_ost_out00_00: service thread pid 5649 completed after 74.357s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 29147:0:(llite_lib.c:1996:ll_md_setattr()) md_setattr fails: rc = -108 LustreError: 28766:0:(mdc_request.c:1469:mdc_read_page()) lustre-MDT0000-mdc-ffff8800a70f4138: [0x200000402:0x90:0x0] lock enqueue fails: rc = -108 LustreError: 1843:0:(ldlm_resource.c:981:ldlm_resource_complain()) lustre-MDT0000-mdc-ffff8800a70f4138: namespace resource [0x200000007:0x1:0x0].0x0 (ffff8802a48f4a40) refcount nonzero (1) after lock cleanup; forcing cleanup. Lustre: lustre-MDT0000-mdc-ffff8800a70f4138: Connection restored to 192.168.123.82@tcp (at 0@lo) LustreError: 17396:0:(mdt_open.c:1302:mdt_cross_open()) lustre-MDT0000: [0x200000403:0x26:0x0] doesn't exist!: rc = -14 LustreError: 17396:0:(mdt_open.c:1302:mdt_cross_open()) Skipped 1 previous similar message Lustre: 12664:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 516 < left 618, rollback = 7 Lustre: 12664:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 275 previous similar messages LustreError: 3075:0:(lov_object.c:1348:lov_layout_change()) lustre-clilov-ffff8800a731a548: cannot apply new layout on [0x200000407:0x70:0x0] : rc = -5 LustreError: 3075:0:(lov_object.c:1348:lov_layout_change()) Skipped 34 previous similar messages 2[8040]: segfault at 8 ip 00007f88f1e297e8 sp 00007fff6601bb30 error 4 in ld-2.17.so[7f88f1e1e000+22000] LustreError: 7088:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a731a548: inode [0x280000403:0x2ca8:0x0] mdc close failed: rc = -2 LustreError: 7088:0:(file.c:247:ll_close_inode_openhandle()) Skipped 89 previous similar messages LustreError: 9802:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000403:0x2e52:0x0]: rc = -5 LustreError: 9802:0:(lcommon_cl.c:188:cl_file_inode_init()) Skipped 493 previous similar messages LustreError: 9802:0:(llite_lib.c:3700:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 9802:0:(llite_lib.c:3700:ll_prep_inode()) Skipped 493 previous similar messages LustreError: 2004:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 11 [0x240000403:0x2cc1:0x0] inode@0000000000000000: rc = -5 LustreError: 2004:0:(statahead.c:825:ll_statahead_interpret_work()) Skipped 4 previous similar messages LustreError: 9201:0:(vvp_io.c:1905:vvp_io_init()) lustre: refresh file layout [0x280000403:0x2d25:0x0] error -5. LustreError: 9201:0:(vvp_io.c:1905:vvp_io_init()) Skipped 6 previous similar messages 1[10567]: segfault at 8 ip 00007f4c574147e8 sp 00007ffd8aad58f0 error 4 in ld-2.17.so[7f4c57409000+22000] 2[10961]: segfault at 8 ip 00007f64b542c7e8 sp 00007ffc0d1b1c50 error 4 in ld-2.17.so[7f64b5421000+22000] 19[13929]: segfault at 8 ip 00007f2141c507e8 sp 00007ffcb4da8910 error 4 in ld-2.17.so[7f2141c45000+22000] 12[23051]: segfault at 8 ip 00007ff7c6cb57e8 sp 00007ffec29a58a0 error 4 in ld-2.17.so[7ff7c6caa000+22000] LustreError: 20991:0:(mdt_reint.c:2540:mdt_reint_migrate()) lustre-MDT0000: migrate [0x200000403:0x1:0x0]/5 failed: rc = -1 LustreError: 20991:0:(mdt_reint.c:2540:mdt_reint_migrate()) Skipped 148 previous similar messages 12[27137]: segfault at 8 ip 00007f370044e7e8 sp 00007ffc4cf93a20 error 4 in ld-2.17.so[7f3700443000+22000] 4[28675]: segfault at 8 ip 00007f83d096b7e8 sp 00007fff2e28c9e0 error 4 in ld-2.17.so[7f83d0960000+22000] LustreError: 20264:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0001: [0x240000404:0x566c:0x0] migrate mdt count mismatch 1 != 3 LustreError: 20264:0:(mdt_xattr.c:402:mdt_dir_layout_update()) Skipped 5 previous similar messages 0[31772]: segfault at 8 ip 00007fc7521467e8 sp 00007fff638a8080 error 4 in ld-2.17.so[7fc75213b000+22000] LustreError: 31470:0:(mdd_dir.c:4743:mdd_migrate_cmd_check()) lustre-MDD0000: '18' migration was interrupted, run 'lfs migrate -m 1 -c 1 -H crush 18' to finish migration: rc = -1 LustreError: 31470:0:(mdd_dir.c:4743:mdd_migrate_cmd_check()) Skipped 61 previous similar messages 12[7565]: segfault at 8 ip 00007f68c9c407e8 sp 00007ffcd3ff1ce0 error 4 in ld-2.17.so[7f68c9c35000+22000] traps: 9[10722] trap invalid opcode ip:404afe sp:7ffde44ddc98 error:0 in 9[400000+6000] 3[11203]: segfault at 8 ip 00007fb365f3c7e8 sp 00007fff9e503aa0 error 4 in ld-2.17.so[7fb365f31000+22000] 17[12103]: segfault at 8 ip 00007fe4816687e8 sp 00007fffb65f1340 error 4 in ld-2.17.so[7fe48165d000+22000] 13[15514]: segfault at 8 ip 00007fba88ccf7e8 sp 00007ffd4f46bd90 error 4 in ld-2.17.so[7fba88cc4000+22000] 16[17568]: segfault at 8 ip 00007f572eddf7e8 sp 00007ffc24d65be0 error 4 in ld-2.17.so[7f572edd4000+22000] LustreError: 17265:0:(mdt_open.c:1302:mdt_cross_open()) lustre-MDT0000: [0x200000403:0x26:0x0] doesn't exist!: rc = -14 LustreError: 17265:0:(mdt_open.c:1302:mdt_cross_open()) Skipped 1 previous similar message LustreError: 18113:0:(lustre_lmv.h:500:lmv_is_sane()) unknown layout LMV: magic=0xcd40cd0 count=4 index=3 hash=crush:0x82000003 version=1 migrate_offset=3 migrate_hash=fnv_1a_64:2 pool= LustreError: 11486:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000407:0xaad:0x0]: rc = -2 LustreError: 11486:0:(mdd_object.c:3901:mdd_close()) Skipped 29 previous similar messages 15[24055]: segfault at 8 ip 00007f2ac21d17e8 sp 00007fffcd1ce0c0 error 4 in ld-2.17.so[7f2ac21c6000+22000] 9[27571]: segfault at 8 ip 00007f5272ba27e8 sp 00007ffe374147e0 error 4 in ld-2.17.so[7f5272b97000+22000] 9[28501]: segfault at 8 ip 00007f90d34057e8 sp 00007ffc59a2aea0 error 4 in ld-2.17.so[7f90d33fa000+22000] 4[30782]: segfault at 0 ip 0000000000403e5f sp 00007ffd22e92100 error 6 in 4[400000+6000] Lustre: 17318:0:(lod_lov.c:1414:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000405:0x325d:0x0] with magic=0xbd60bd0 Lustre: 17318:0:(lod_lov.c:1414:lod_parse_striping()) Skipped 217 previous similar messages 6[6370]: segfault at 8 ip 00007fcd838027e8 sp 00007ffd7ceaa2d0 error 4 in ld-2.17.so[7fcd837f7000+22000] LustreError: 18113:0:(lustre_lmv.h:500:lmv_is_sane()) unknown layout LMV: magic=0xcd40cd0 count=3 index=2 hash=crush:0x82000003 version=1 migrate_offset=2 migrate_hash=fnv_1a_64:2 pool= 1[8072]: segfault at 8 ip 00007fb9ae4b47e8 sp 00007ffcc2ad5f20 error 4 in ld-2.17.so[7fb9ae4a9000+22000] LustreError: 11132:0:(ldlm_lockd.c:257:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0001_UUID lock: ffff8802a67acb40/0xf9603e49444b46a3 lrc: 3/0,0 mode: PR/PR res: [0x240000404:0x5c54:0x0].0x0 bits 0x1b/0x0 rrc: 5 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0xf9603e49444b466b expref: 1032 pid: 16978 timeout: 1555 lvb_type: 0 LustreError: 11128:0:(ldlm_lockd.c:2550:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1746566242 with bad export cookie 17969430997586500326 Lustre: lustre-MDT0001-mdc-ffff8800a731a548: Connection to lustre-MDT0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 11997:0:(client.c:1373:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802cd79bc40 x1831406271190784/t0(0) o104->lustre-MDT0001@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' uid:4294967295 gid:4294967295 projid:4294967295 LustreError: lustre-MDT0001-mdc-ffff8800a731a548: This client was evicted by lustre-MDT0001; in progress operations using this service will fail. LustreError: 10851:0:(mdc_request.c:1469:mdc_read_page()) lustre-MDT0001-mdc-ffff8800a731a548: [0x240000401:0x5f:0x0] lock enqueue fails: rc = -108 LustreError: 10851:0:(mdc_request.c:1469:mdc_read_page()) Skipped 5 previous similar messages Lustre: lustre-MDT0001-mdc-ffff8800a731a548: Connection restored to 192.168.123.82@tcp (at 0@lo) 4[32527]: segfault at 8 ip 00007f1452b3f7e8 sp 00007fff01d9bc00 error 4 in ld-2.17.so[7f1452b34000+22000] LustreError: 11132:0:(ldlm_lockd.c:257:expired_lock_main()) ### lock callback timer expired after 101s: evicting client at 0@lo ns: mdt-lustre-MDT0000_UUID lock: ffff880091dc4b40/0xf9603e4944566438 lrc: 3/0,0 mode: CR/CR res: [0x200000406:0x28c1:0x0].0x0 bits 0xa/0x0 rrc: 5 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0xf9603e494456641c expref: 1079 pid: 17265 timeout: 1596 lvb_type: 0 LustreError: 11484:0:(ldlm_lockd.c:1447:ldlm_handle_enqueue()) ### lock on destroyed export ffff8802d0b58008 ns: mdt-lustre-MDT0000_UUID lock: ffff8800a3a35680/0xf9603e4944582a74 lrc: 3/0,0 mode: PR/PR res: [0x200000406:0x2840:0x0].0x0 bits 0x1b/0x0 rrc: 4 type: IBT gid 0 flags: 0x50200000000000 nid: 0@lo remote: 0xf9603e494456a403 expref: 695 pid: 11484 timeout: 0 lvb_type: 0 LustreError: 11484:0:(ldlm_lockd.c:1447:ldlm_handle_enqueue()) Skipped 7 previous similar messages LustreError: lustre-MDT0000-mdc-ffff8800a731a548: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-MDT0000-mdc-ffff8800a731a548: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 2687:0:(client.c:1373:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802cdfe2340 x1831406278275840/t0(0) o104->lustre-MDT0000@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' uid:4294967295 gid:4294967295 projid:4294967295 LustreError: lustre-MDT0000-mdc-ffff8800a731a548: This client was evicted by lustre-MDT0000; in progress operations using this service will fail. LustreError: 11653:0:(file.c:6143:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000007:0x1:0x0] error: rc = -108 LustreError: 11653:0:(file.c:6143:ll_inode_revalidate_fini()) Skipped 221 previous similar messages LustreError: 9373:0:(mdc_request.c:1469:mdc_read_page()) lustre-MDT0000-mdc-ffff8800a731a548: [0x200000401:0x76:0x0] lock enqueue fails: rc = -108 Lustre: dir [0x240000404:0x5cea:0x0] stripe 2 readdir failed: -108, directory is partially accessed! Lustre: Skipped 32 previous similar messages Lustre: lustre-MDT0000-mdc-ffff8800a731a548: Connection restored to 192.168.123.82@tcp (at 0@lo) 0[25438]: segfault at 8 ip 00007f441f2c17e8 sp 00007ffdc20d5bc0 error 4 in ld-2.17.so[7f441f2b6000+22000] traps: 3[25881] general protection ip:404910 sp:7fffefee8fe8 error:0 in 3[400000+6000] 5[27352]: segfault at 8 ip 00007f3aa9c9a7e8 sp 00007ffc4e1f4e20 error 4 in ld-2.17.so[7f3aa9c8f000+22000] Lustre: 30680:0:(out_handler.c:879:out_tx_end()) lustre-MDT0000-osd: error during execution of #2 from /home/green/git/lustre-release/lustre/ptlrpc/../../lustre/target/out_handler.c:433: rc = -61 LustreError: 30680:0:(out_handler.c:889:out_tx_end()) lustre-MDT0000-osd: undo for /home/green/git/lustre-release/lustre/ptlrpc/../../lustre/target/out_handler.c:410: rc = -524 6[32653]: segfault at 8 ip 00007fb9b5c677e8 sp 00007ffe9a0ba970 error 4 in ld-2.17.so[7fb9b5c5c000+22000] 0[32405]: segfault at 8 ip 00007fa2088997e8 sp 00007ffffcd41e20 error 4 in ld-2.17.so[7fa20888e000+22000] 13[1709]: segfault at 8 ip 00007f4d565437e8 sp 00007ffcfde1b000 error 4 in ld-2.17.so[7f4d56538000+22000] 5[3030]: segfault at 8 ip 00007f91c3abc7e8 sp 00007ffc1e6a1d40 error 4 in ld-2.17.so[7f91c3ab1000+22000] 15[3125]: segfault at 8 ip 00007f9b4b17e7e8 sp 00007ffc2788e600 error 4 in ld-2.17.so[7f9b4b173000+22000] 10[3878]: segfault at 8 ip 00007f7eb92b07e8 sp 00007ffc5a8d69a0 error 4 in ld-2.17.so[7f7eb92a5000+22000] 13[5790]: segfault at 8 ip 00007fe28c7967e8 sp 00007ffe513e1ae0 error 4 in ld-2.17.so[7fe28c78b000+22000] 8[5687]: segfault at 8 ip 00007fad5c46c7e8 sp 00007ffd22c82730 error 4 in ld-2.17.so[7fad5c461000+22000] 13[8569]: segfault at 8 ip 00007f9488e177e8 sp 00007ffd43d48e70 error 4 in ld-2.17.so[7f9488e0c000+22000] 3[10190]: segfault at 8 ip 00007fd1ff8137e8 sp 00007ffd05afd5f0 error 4 in ld-2.17.so[7fd1ff808000+22000] 7[11688]: segfault at 8 ip 00007f11bb2987e8 sp 00007fff191a01f0 error 4 in ld-2.17.so[7f11bb28d000+22000] LustreError: 17405:0:(mdt_open.c:1302:mdt_cross_open()) lustre-MDT0001: [0x240000404:0x5d35:0x0] doesn't exist!: rc = -14 LustreError: 17405:0:(mdt_open.c:1302:mdt_cross_open()) Skipped 183 previous similar messages 9[17271]: segfault at 406000 ip 0000000000406000 sp 00007fff7d4f6b08 error 14 in 9[606000+1000] LustreError: 11132:0:(ldlm_lockd.c:257:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0001_UUID lock: ffff88009e2b9e40/0xf9603e4944725757 lrc: 3/0,0 mode: PW/PW res: [0x240000403:0x3e85:0x0].0x0 bits 0x4/0x0 rrc: 5 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0xf9603e494472572d expref: 824 pid: 11485 timeout: 1806 lvb_type: 0 Lustre: 27342:0:(mdd_dir.c:4822:mdd_migrate_object()) lustre-MDD0001: [0x240000403:0x1:0x0]/9 is open, migrate only dentry Lustre: 27342:0:(mdd_dir.c:4822:mdd_migrate_object()) Skipped 129 previous similar messages LustreError: 17391:0:(ldlm_lockd.c:1447:ldlm_handle_enqueue()) ### lock on destroyed export ffff8802e5482548 ns: mdt-lustre-MDT0001_UUID lock: ffff8802c2286580/0xf9603e4944727621 lrc: 3/0,0 mode: PR/PR res: [0x240000405:0x311:0x0].0x0 bits 0x13/0x0 rrc: 4 type: IBT gid 0 flags: 0x50200000000000 nid: 0@lo remote: 0xf9603e4944727605 expref: 559 pid: 17391 timeout: 0 lvb_type: 0 LustreError: lustre-MDT0001-mdc-ffff8800a70f4138: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-MDT0001-mdc-ffff8800a70f4138: Connection to lustre-MDT0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: lustre-MDT0001-mdc-ffff8800a70f4138: This client was evicted by lustre-MDT0001; in progress operations using this service will fail. LustreError: 17391:0:(ldlm_lockd.c:1447:ldlm_handle_enqueue()) Skipped 7 previous similar messages LustreError: 22522:0:(file.c:6143:ll_inode_revalidate_fini()) lustre: revalidate FID [0x240000403:0x1:0x0] error: rc = -5 LustreError: 22522:0:(file.c:6143:ll_inode_revalidate_fini()) Skipped 84 previous similar messages Lustre: 28020:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 513 < left 966, rollback = 2 Lustre: 28020:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 5526 previous similar messages Lustre: 28020:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 3/12/1, destroy: 0/0/0 Lustre: 28020:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 5690 previous similar messages Lustre: 28020:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 2/2/0, xattr_set: 12/966/0 Lustre: 28020:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 5690 previous similar messages Lustre: 28020:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 5/55/0, punch: 0/0/0, quota 4/54/0 Lustre: 28020:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 5690 previous similar messages Lustre: 28020:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 11/239/1, delete: 0/0/0 Lustre: 28020:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 5690 previous similar messages Lustre: 28020:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 5/5/0, ref_del: 0/0/0 Lustre: 28020:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 5690 previous similar messages LustreError: 23313:0:(mdc_request.c:1469:mdc_read_page()) lustre-MDT0001-mdc-ffff8800a70f4138: [0x240000401:0x6b:0x0] lock enqueue fails: rc = -108 LustreError: 23313:0:(mdc_request.c:1469:mdc_read_page()) Skipped 3 previous similar messages Lustre: lustre-MDT0001-mdc-ffff8800a70f4138: Connection restored to 192.168.123.82@tcp (at 0@lo) traps: 0[23965] trap invalid opcode ip:405152 sp:7ffed333d2b8 error:0 in 0[400000+6000] Lustre: 11489:0:(out_handler.c:879:out_tx_end()) lustre-MDT0000-osd: error during execution of #0 from /home/green/git/lustre-release/lustre/ptlrpc/../../lustre/target/out_handler.c:433: rc = -61 7[29654]: segfault at 8 ip 00007f766260a7e8 sp 00007ffd57678660 error 4 in ld-2.17.so[7f76625ff000+22000] 11[32130]: segfault at 0 ip 0000000000403e5f sp 00007ffe1b693920 error 6 in 4[400000+6000] LustreError: 19229:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 4 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 LustreError: 19229:0:(statahead.c:825:ll_statahead_interpret_work()) Skipped 2 previous similar messages 8[8295]: segfault at 0 ip 0000000000403cf0 sp 00007ffe0d54ae58 error 6 in 8[400000+6000] Lustre: 18247:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0002: opcode 7: before 516 < left 618, rollback = 7 Lustre: 18247:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 153 previous similar messages 11[21201]: segfault at 0 ip 0000000000403e5f sp 00007ffcec621660 error 6 in 11[400000+6000] 3[22342]: segfault at 8 ip 00007f4cb2ba47e8 sp 00007ffc03a94c60 error 4 in ld-2.17.so[7f4cb2b99000+22000] LustreError: 25396:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a731a548: inode [0x200000408:0xa07:0x0] mdc close failed: rc = -13 LustreError: 25396:0:(file.c:247:ll_close_inode_openhandle()) Skipped 175 previous similar messages 6[28204]: segfault at 8 ip 00007fd49aac97e8 sp 00007ffe1fd82440 error 4 in ld-2.17.so[7fd49aabe000+22000] LustreError: 26011:0:(statahead.c:2399:start_statahead_thread()) lustre: invalid pattern 0X0. 0[28461]: segfault at 1c ip 00000000004017b0 sp 00007ffceabbce90 error 6 in 0[400000+6000] 9[31421]: segfault at 8 ip 00007f5c4bf837e8 sp 00007ffc173990a0 error 4 in ld-2.17.so[7f5c4bf78000+22000] LustreError: 20321:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0002: [0x280000405:0x373a:0x0] migrate mdt count mismatch 2 != 3 LustreError: 20321:0:(mdt_xattr.c:402:mdt_dir_layout_update()) Skipped 7 previous similar messages LustreError: 3597:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000407:0x1cb7:0x0]: rc = -5 LustreError: 3597:0:(lcommon_cl.c:188:cl_file_inode_init()) Skipped 233 previous similar messages LustreError: 3597:0:(llite_lib.c:3700:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 3597:0:(llite_lib.c:3700:ll_prep_inode()) Skipped 233 previous similar messages LustreError: 3430:0:(lov_object.c:1348:lov_layout_change()) lustre-clilov-ffff8800a731a548: cannot apply new layout on [0x200000407:0x1cb7:0x0] : rc = -5 LustreError: 3430:0:(lov_object.c:1348:lov_layout_change()) Skipped 30 previous similar messages LustreError: 3430:0:(vvp_io.c:1905:vvp_io_init()) lustre: refresh file layout [0x200000407:0x1cb7:0x0] error -5. LustreError: 3430:0:(vvp_io.c:1905:vvp_io_init()) Skipped 9 previous similar messages | Link to test |
racer test 1: racer on clients: centos-0.localnet DURATION=2700 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 PGD 31f939067 PUD 2ee86d067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) lustre(OE) osp(OE) ofd(OE) lod(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) jbd2 mbcache lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) crc32_generic lnet(OE) dm_flakey dm_mod libcfs(OE) crc_t10dif crct10dif_generic crct10dif_common virtio_balloon i2c_piix4 virtio_console pcspkr ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm ata_piix drm_panel_orientation_quirks serio_raw virtio_blk floppy i2c_core libata CPU: 13 PID: 2258 Comm: ll_sa_2214 Kdump: loaded Tainted: P OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.16.0-3.module_el8.7.0+1218+f626c2ff 04/01/2014 task: ffff8800a6493760 ti: ffff8802c3500000 task.ti: ffff8802c3500000 RIP: 0010:[<ffffffff813f502c>] [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 RSP: 0018:ffff8802c3503d78 EFLAGS: 00010296 RAX: ffff8802c1a0ba10 RBX: 0000000000000008 RCX: 0000000100260025 RDX: 0000000000000026 RSI: ffff8802c1a0bd40 RDI: 0000000000000008 RBP: ffff8802c3503d88 R08: ffff8802db826d58 R09: 0000000000000000 R10: ffff8802db823858 R11: ffff8802db826d58 R12: 0000000000000000 R13: ffff8802e9efa4f0 R14: ffff8802c1a0bd40 R15: ffff8802db826d58 FS: 0000000000000000(0000) GS:ffff880331d40000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 00000002eb9c4000 CR4: 00000000000007e0 Call Trace: [<ffffffffa10cd431>] ll_statahead_thread+0xdb1/0x2000 [lustre] [<ffffffff810dbb51>] ? put_prev_entity+0x31/0x400 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e329e>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffff810c834d>] ? finish_task_switch+0x5d/0x1b0 [<ffffffffa10cc680>] ? ll_statahead_handle.constprop.28+0x1c0/0x1c0 [lustre] [<ffffffff810ba114>] kthread+0xe4/0xf0 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 | 14[15625]: segfault at 8 ip 00007f1a676597e8 sp 00007ffed3752100 error 4 in ld-2.17.so[7f1a6764e000+22000] Lustre: 9530:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff8802edd64b40 x1831153276370688/t4294968896(0) o101->32e3ad90-6595-406b-93f0-c86f06af6e10@0@lo:628/0 lens 376/864 e 0 to 0 dl 1746323933 ref 1 fl Interpret:H/602/0 rc 0/0 job:'dd.0' uid:0 gid:0 projid:0 Lustre: 15950:0:(lod_lov.c:1414:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000401:0xf4:0x0] with magic=0xbd60bd0 traps: 11[19017] trap invalid opcode ip:404a6d sp:7e5f3468 error:0 in 11[400000+6000] 4[20229]: segfault at 8 ip 00007f3966fba7e8 sp 00007ffd727635c0 error 4 in ld-2.17.so[7f3966faf000+22000] Lustre: 20277:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 618, rollback = 7 Lustre: 20277:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 20277:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 20277:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 20277:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 20277:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 9495:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0000: opcode 7: before 516 < left 618, rollback = 7 Lustre: 9495:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 1 previous similar message Lustre: 9495:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 9495:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 9495:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 9495:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 9495:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 9495:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 9495:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 9495:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 9495:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 9495:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 9497:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 618, rollback = 7 Lustre: 9497:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 1 previous similar message Lustre: 9497:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 9497:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 9497:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 9497:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 9497:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 9497:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 9497:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 9497:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 9497:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 9497:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 9495:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 618, rollback = 7 Lustre: 9495:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 1 previous similar message Lustre: 9495:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 9495:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 9495:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 9495:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 9495:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 9495:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 9495:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 9495:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 9495:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 9495:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 15771:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0000: opcode 7: before 516 < left 618, rollback = 7 Lustre: 15771:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 1 previous similar message Lustre: 15771:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 15771:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 15771:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 15771:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 15771:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 15771:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 15771:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 15771:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 15771:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 15771:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 20277:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 618, rollback = 7 Lustre: 20277:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 1 previous similar message Lustre: 20277:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 20277:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 20277:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 20277:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 20277:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 20277:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 20277:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 20277:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 20277:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 20277:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 9495:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 618, rollback = 7 Lustre: 9495:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 3 previous similar messages Lustre: 9495:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 9495:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 9495:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 9495:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 9495:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 9495:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 9495:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 9495:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 9495:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 9495:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 3 previous similar messages LustreError: 5191:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800aa5f4a88: inode [0x200000402:0x734:0x0] mdc close failed: rc = -13 6[5824]: segfault at 0 ip 0000000000403cf0 sp 00007ffd7143ae18 error 6 in 6[400000+6000] LustreError: 8161:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802ee7c37e8: inode [0x200000402:0x805:0x0] mdc close failed: rc = -13 0[9625]: segfault at ffffffffd9bd451f ip 0000000000404a87 sp 00007ffed9052ae8 error 7 in 0[400000+6000] Lustre: 17519:0:(lod_lov.c:1414:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000401:0x8e2:0x0] with magic=0xbd60bd0 Lustre: 17519:0:(lod_lov.c:1414:lod_parse_striping()) Skipped 1 previous similar message Lustre: 9497:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0002: opcode 7: before 516 < left 618, rollback = 7 Lustre: 9497:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 33 previous similar messages Lustre: 9497:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 9497:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 33 previous similar messages Lustre: 9497:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 9497:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 33 previous similar messages Lustre: 9497:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 9497:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 33 previous similar messages Lustre: 9497:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 9497:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 33 previous similar messages Lustre: 9497:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 9497:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 33 previous similar messages LustreError: 12658:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802ee7c37e8: inode [0x200000401:0xa41:0x0] mdc close failed: rc = -13 2[14330]: segfault at 406000 ip 0000000000406000 sp 00007ffe77f10c18 error 14 in 2[606000+1000] Lustre: 15170:0:(lod_lov.c:1414:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000401:0xb88:0x0] with magic=0xbd60bd0 Lustre: 15170:0:(lod_lov.c:1414:lod_parse_striping()) Skipped 1 previous similar message 12[17223]: segfault at 8 ip 00007fc50c2727e8 sp 00007fff3e5e6bc0 error 4 in ld-2.17.so[7fc50c267000+22000] 6[17594]: segfault at 8 ip 00007fe7fd3467e8 sp 00007ffd3c873970 error 4 in ld-2.17.so[7fe7fd33b000+22000] 3[18239]: segfault at 8 ip 00007f0e915387e8 sp 00007ffcc7437250 error 4 in ld-2.17.so[7f0e9152d000+22000] 19[18571]: segfault at 8 ip 00007fd8a8f7a7e8 sp 00007fffe94007f0 error 4 in ld-2.17.so[7fd8a8f6f000+22000] Lustre: 13564:0:(lod_lov.c:1414:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000402:0xba8:0x0] with magic=0xbd60bd0 Lustre: 13564:0:(lod_lov.c:1414:lod_parse_striping()) Skipped 1 previous similar message Lustre: 14972:0:(lod_lov.c:1414:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000401:0xd26:0x0] with magic=0xbd60bd0 Lustre: 14972:0:(lod_lov.c:1414:lod_parse_striping()) Skipped 1 previous similar message 6[23784]: segfault at 8 ip 00007f59787ca7e8 sp 00007ffe72d57800 error 4 in ld-2.17.so[7f59787bf000+22000] Lustre: 9496:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0002: opcode 7: before 516 < left 618, rollback = 7 Lustre: 9496:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 39 previous similar messages Lustre: 9496:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 9496:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 39 previous similar messages Lustre: 9496:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 9496:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 39 previous similar messages Lustre: 9496:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 9496:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 39 previous similar messages Lustre: 9496:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 9496:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 39 previous similar messages Lustre: 9496:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 9496:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 39 previous similar messages 8[27430]: segfault at 0 ip 0000000000403e5f sp 00007ffe04bd6c30 error 6 in 13[400000+6000] Lustre: mdt00_000: service thread pid 8707 was inactive for 40.100 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Lustre: mdt00_001: service thread pid 8709 was inactive for 40.104 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 8709, comm: mdt00_001 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock_try+0xa0/0x250 [mdt] [<0>] mdt_getattr_name_lock+0x1729/0x2cc0 [mdt] [<0>] mdt_intent_getattr+0x2cc/0x4e0 [mdt] [<0>] mdt_intent_opc.constprop.74+0x211/0xc60 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x397/0x980 [ptlrpc] [<0>] ldlm_handle_enqueue+0x547/0x1900 [ptlrpc] [<0>] tgt_enqueue+0x68/0x240 [ptlrpc] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x257/0xcd0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 13564, comm: mdt00_016 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock+0x88/0x1c0 [mdt] [<0>] mdt_object_find_lock+0x54/0x170 [mdt] [<0>] mdt_reint_setxattr+0x133/0x10e0 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x73c/0xbb0 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x257/0xcd0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: mdt00_011: service thread pid 17521 was inactive for 40.101 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 1 previous similar message Pid: 8707, comm: mdt00_000 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock_try+0xa0/0x250 [mdt] [<0>] mdt_getattr_name_lock+0x1729/0x2cc0 [mdt] [<0>] mdt_intent_getattr+0x2cc/0x4e0 [mdt] [<0>] mdt_intent_opc.constprop.74+0x211/0xc60 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x397/0x980 [ptlrpc] [<0>] ldlm_handle_enqueue+0x547/0x1900 [ptlrpc] [<0>] tgt_enqueue+0x68/0x240 [ptlrpc] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x257/0xcd0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe LustreError: 8525:0:(ldlm_lockd.c:257:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0000_UUID lock: ffff8800a544a5c0/0x7f1562181356b3f4 lrc: 3/0,0 mode: CR/CR res: [0x200000401:0xf5a:0x0].0x0 bits 0xa/0x0 rrc: 8 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0x7f1562181356b3e6 expref: 619 pid: 14947 timeout: 462 lvb_type: 0 Lustre: mdt00_016: service thread pid 13564 completed after 100.298s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_001: service thread pid 8709 completed after 100.293s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: lustre-MDT0000-mdc-ffff8800aa5f4a88: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-MDT0000-mdc-ffff8800aa5f4a88: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: lustre-MDT0000-mdc-ffff8800aa5f4a88: This client was evicted by lustre-MDT0000; in progress operations using this service will fail. LustreError: 29782:0:(file.c:6143:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000401:0xf5a:0x0] error: rc = -5 LustreError: 15174:0:(ldlm_lockd.c:2550:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1746324270 with bad export cookie 9157333272868661088 LustreError: 30163:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800aa5f4a88: inode [0x200000401:0xf01:0x0] mdc close failed: rc = -108 Lustre: mdt00_011: service thread pid 17521 completed after 100.312s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 8707:0:(ldlm_lockd.c:1447:ldlm_handle_enqueue()) ### lock on destroyed export ffff8800a9daefc8 ns: mdt-lustre-MDT0000_UUID lock: ffff8802db8ad680/0x7f1562181356b608 lrc: 3/0,0 mode: PR/PR res: [0x200000401:0xf5a:0x0].0x0 bits 0x1b/0x0 rrc: 4 type: IBT gid 0 flags: 0x50200000000000 nid: 0@lo remote: 0x7f1562181356b5ec expref: 4 pid: 8707 timeout: 0 lvb_type: 0 Lustre: mdt00_005: service thread pid 14972 completed after 100.175s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_000: service thread pid 8707 completed after 100.316s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 29886:0:(vvp_io.c:1905:vvp_io_init()) lustre: refresh file layout [0x200000401:0xf5a:0x0] error -108. Lustre: lustre-MDT0000-mdc-ffff8800aa5f4a88: Connection restored to 192.168.123.2@tcp (at 0@lo) Lustre: 15771:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0000: opcode 7: before 516 < left 618, rollback = 7 Lustre: 15771:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 3 previous similar messages Lustre: 15771:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 15771:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 15771:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 15771:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 15771:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 15771:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 15771:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 15771:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 15771:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 15771:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 3 previous similar messages 0[30453]: segfault at 8 ip 00007f4494ede7e8 sp 00007ffd0eb89410 error 4 in ld-2.17.so[7f4494ed3000+22000] 5[31451]: segfault at 0 ip 0000000000403e5f sp 00007ffeabc86350 error 6 in 5[400000+6000] 5[31837]: segfault at 8 ip 00007f3b689017e8 sp 00007ffe59a74050 error 4 in ld-2.17.so[7f3b688f6000+22000] | Link to test |
racer test 1: racer on clients: oleg344-client.virtnet DURATION=300 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffff813f40ac>] _atomic_dec_and_lock+0xc/0x70 PGD 9d34f067 PUD 9d2b1067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) osp(OE) ofd(OE) lod(OE) mdt(OE) mdd(OE) mgs(OE) lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) crc_t10dif crct10dif_generic rpcsec_gss_krb5 sb_edac edac_core iosf_mbi kvm_intel kvm irqbypass crc32_pclmul ghash_clmulni_intel squashfs aesni_intel lrw gf128mul i2c_piix4 glue_helper i2c_core ablk_helper cryptd pcspkr binfmt_misc ip_tables ext4 mbcache jbd2 ata_generic pata_acpi ata_piix crct10dif_pclmul crct10dif_common crc32c_intel serio_raw libata CPU: 0 PID: 2249 Comm: ll_sa_32447 Kdump: loaded Tainted: G OE ------------ 3.10.0-7.9-debug #1 Hardware name: QEMU Standard PC (i440FX + PIIX, 1996), BIOS 1.16.2-1.fc38 04/01/2014 task: ffff88009d2a4440 ti: ffff88008d1c8000 task.ti: ffff88008d1c8000 RIP: 0010:[<ffffffff813f40ac>] [<ffffffff813f40ac>] _atomic_dec_and_lock+0xc/0x70 RSP: 0018:ffff88008d1cbd78 EFLAGS: 00010296 RAX: ffff8800849d08d0 RBX: 0000000000000008 RCX: ffffea0004a9b100 RDX: 0000000000000006 RSI: ffff8800849d0c00 RDI: 0000000000000008 RBP: ffff88008d1cbd88 R08: ffff8800a44a9000 R09: 000000018010000f R10: 0000000000000001 R11: ffff8800a44a9000 R12: 0000000000000000 R13: ffff88009cb43330 R14: ffff8800849d0c00 R15: ffff8800a44a9000 FS: 0000000000000000(0000) GS:ffff88013e200000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 CR2: 0000000000000008 CR3: 000000008814c000 CR4: 0000000000160ff0 Call Trace: [<ffffffffa10dbc41>] ll_statahead_thread+0xdb1/0x2000 [lustre] [<ffffffff810df6e2>] ? pick_next_task_fair+0x702/0x8a0 [<ffffffff81410a79>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e8dce>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffff810c834d>] ? finish_task_switch+0x5d/0x1b0 [<ffffffffa10dae90>] ? ll_statahead_handle.constprop.28+0x1c0/0x1c0 [lustre] [<ffffffff810ba114>] kthread+0xe4/0xf0 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817f3e5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 | ln (11093) used greatest stack depth: 10320 bytes left random: crng init done 18[11913]: segfault at 8 ip 00007fa30ac7e7e8 sp 00007ffd453fec40 error 4 in ld-2.17.so[7fa30ac73000+22000] LustreError: 11789:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff88012a528000: inode [0x240000403:0xb:0x0] mdc close failed: rc = -13 1[13511]: segfault at 8 ip 00007f12c0c0f7e8 sp 00007ffc11f549b0 error 4 in ld-2.17.so[7f12c0c04000+22000] LustreError: 10899:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000403:0x95:0x0]: rc = -5 LustreError: 10899:0:(llite_lib.c:3697:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 13882:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff88012a528000: inode [0x200000402:0xd:0x0] mdc close failed: rc = -13 LustreError: 14063:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000403:0x93:0x0]: rc = -5 LustreError: 14063:0:(llite_lib.c:3697:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 8[14509]: segfault at 0 ip 00000000004043f0 sp 00007fffb4729eb8 error 6 in 8[400000+6000] LustreError: 17834:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000403:0x175:0x0]: rc = -5 LustreError: 17834:0:(lcommon_cl.c:188:cl_file_inode_init()) Skipped 1 previous similar message LustreError: 17834:0:(llite_lib.c:3697:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 17834:0:(llite_lib.c:3697:ll_prep_inode()) Skipped 1 previous similar message LustreError: 16832:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff88012a528000: cannot apply new layout on [0x240000403:0x175:0x0] : rc = -5 LustreError: 16832:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x240000403:0x175:0x0] error -5. Lustre: dir [0x240000403:0x14e:0x0] stripe 1 readdir failed: -2, directory is partially accessed! LustreError: 19035:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000402:0x228:0x0]: rc = -5 LustreError: 19035:0:(lcommon_cl.c:188:cl_file_inode_init()) Skipped 4 previous similar messages LustreError: 19035:0:(llite_lib.c:3697:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 19035:0:(llite_lib.c:3697:ll_prep_inode()) Skipped 4 previous similar messages LustreError: 20035:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff88012a528000: inode [0x200000403:0x2fa:0x0] mdc close failed: rc = -2 Lustre: dir [0x200000403:0x35d:0x0] stripe 2 readdir failed: -2, directory is partially accessed! LustreError: 22468:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000402:0x32c:0x0]: rc = -5 LustreError: 22468:0:(llite_lib.c:3697:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 24289:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff88012ea9c800: cannot apply new layout on [0x200000402:0x32c:0x0] : rc = -5 LustreError: 24289:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x200000402:0x32c:0x0] error -5. Lustre: dir [0x200000402:0x559:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 1 previous similar message LustreError: 27764:0:(llite_nfs.c:430:ll_dir_get_parent_fid()) lustre: failure inode [0x200000403:0x59e:0x0] get parent: rc = -116 LustreError: 28141:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff88012a528000: inode [0x240000403:0x479:0x0] mdc close failed: rc = -2 Lustre: dir [0x200000402:0x71d:0x0] stripe 1 readdir failed: -2, directory is partially accessed! LustreError: 29865:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000403:0x4cf:0x0]: rc = -5 LustreError: 29865:0:(lcommon_cl.c:188:cl_file_inode_init()) Skipped 11 previous similar messages LustreError: 29865:0:(llite_lib.c:3697:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 29865:0:(llite_lib.c:3697:ll_prep_inode()) Skipped 11 previous similar messages LustreError: 29242:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff88012a528000: cannot apply new layout on [0x200000403:0x462:0x0] : rc = -5 LustreError: 29242:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x200000403:0x462:0x0] error -5. LustreError: 32107:0:(llite_nfs.c:430:ll_dir_get_parent_fid()) lustre: failure inode [0x240000402:0x7f4:0x0] get parent: rc = -116 LustreError: 10761:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff88012a528000: cannot apply new layout on [0x200000403:0x462:0x0] : rc = -5 LustreError: 10761:0:(lov_object.c:1341:lov_layout_change()) Skipped 3 previous similar messages 8[2887]: segfault at 0 ip (null) sp 00007ffdc3a8f6d8 error 14 in 8[400000+6000] Lustre: dir [0x240000402:0x8ee:0x0] stripe 2 readdir failed: -2, directory is partially accessed! Lustre: Skipped 7 previous similar messages 1[8838]: segfault at 0 ip 0000000000403e5f sp 00007ffef8dc5870 error 6 in 1[400000+6000] LustreError: 10976:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000403:0xae0:0x0]: rc = -5 LustreError: 10976:0:(lcommon_cl.c:188:cl_file_inode_init()) Skipped 149 previous similar messages LustreError: 10976:0:(llite_lib.c:3697:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 10976:0:(llite_lib.c:3697:ll_prep_inode()) Skipped 149 previous similar messages LustreError: 10935:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff88012a528000: inode [0x240000403:0x792:0x0] mdc close failed: rc = -2 LustreError: 10935:0:(file.c:247:ll_close_inode_openhandle()) Skipped 1 previous similar message Lustre: dir [0x200000403:0xd37:0x0] stripe 2 readdir failed: -2, directory is partially accessed! Lustre: Skipped 7 previous similar messages 13[13893]: segfault at 8 ip 00007f97910db7e8 sp 00007ffd9b1e4ea0 error 4 in ld-2.17.so[7f97910d0000+22000] LustreError: 487:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 17 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 LustreError: 487:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 7 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 17[14813]: segfault at 0 ip 0000000000403e5f sp 00007ffd8d03d020 error 6 in 17[400000+6000] 14[15287]: segfault at 8 ip 00007f38fb29f7e8 sp 00007fffbbed8eb0 error 4 in ld-2.17.so[7f38fb294000+22000] LustreError: 19363:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff88012ea9c800: inode [0x200000403:0xee0:0x0] mdc close failed: rc = -2 LustreError: 19141:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff88012a528000: cannot apply new layout on [0x200000403:0x1087:0x0] : rc = -5 LustreError: 19141:0:(lov_object.c:1341:lov_layout_change()) Skipped 1 previous similar message LustreError: 19141:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x200000403:0x1087:0x0] error -5. LustreError: 19141:0:(vvp_io.c:1903:vvp_io_init()) Skipped 1 previous similar message 14[24264]: segfault at 8 ip 00007fcf129677e8 sp 00007ffe64192180 error 4 in ld-2.17.so[7fcf1295c000+22000] 11[26163]: segfault at 8 ip 00007f03beb447e8 sp 00007ffc33008790 error 4 in ld-2.17.so[7f03beb39000+22000] 3[26486]: segfault at 0 ip (null) sp 00007ffc3e0fc908 error 14 in 3[400000+6000] LustreError: 27579:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff88012a528000: cannot apply new layout on [0x240000403:0x753:0x0] : rc = -5 LustreError: 27579:0:(lov_object.c:1341:lov_layout_change()) Skipped 1 previous similar message LustreError: 27579:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x240000403:0x753:0x0] error -5. LustreError: 27579:0:(vvp_io.c:1903:vvp_io_init()) Skipped 1 previous similar message 14[29188]: segfault at 8 ip 00007fde0d28f7e8 sp 00007ffdaa1e9d60 error 4 in ld-2.17.so[7fde0d284000+22000] 6[29171]: segfault at 8 ip 00007f65241d77e8 sp 00007ffc49437eb0 error 4 in ld-2.17.so[7f65241cc000+22000] 13[31420]: segfault at 0 ip 0000000000403e5f sp 00007ffdefe334e0 error 6 in 13[400000+6000] Lustre: dir [0x240000402:0x14f7:0x0] stripe 2 readdir failed: -2, directory is partially accessed! Lustre: Skipped 10 previous similar messages 9[3297]: segfault at 8 ip 00007f98905657e8 sp 00007ffc0ed84850 error 4 in ld-2.17.so[7f989055a000+22000] LustreError: 2497:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff88012a528000: inode [0x240000402:0xafe:0x0] mdc close failed: rc = -2 LustreError: 2497:0:(file.c:247:ll_close_inode_openhandle()) Skipped 2 previous similar messages LustreError: 4863:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x200000402:0x17dc:0x0] error -5. 16[5812]: segfault at 8 ip 00007fc6f0e837e8 sp 00007ffedb44aa90 error 4 in ld-2.17.so[7fc6f0e78000+22000] 2[6832]: segfault at 8 ip 00007fa5ddcb07e8 sp 00007fff21272f50 error 4 in ld-2.17.so[7fa5ddca5000+22000] 9[7147]: segfault at 8 ip 00007f544320f7e8 sp 00007fff82fef650 error 4 in ld-2.17.so[7f5443204000+22000] LustreError: 6818:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000402:0x1817:0x0]: rc = -5 LustreError: 6818:0:(lcommon_cl.c:188:cl_file_inode_init()) Skipped 97 previous similar messages LustreError: 6818:0:(llite_lib.c:3697:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 6818:0:(llite_lib.c:3697:ll_prep_inode()) Skipped 97 previous similar messages LustreError: 14052:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff88012ea9c800: cannot apply new layout on [0x200000402:0x17dc:0x0] : rc = -5 LustreError: 14052:0:(lov_object.c:1341:lov_layout_change()) Skipped 3 previous similar messages 8[15074]: segfault at 8 ip 00007f7a3448b7e8 sp 00007fff5e788580 error 4 in ld-2.17.so[7f7a34480000+22000] 18[16759]: segfault at 406000 ip 0000000000406000 sp 00007ffefa797d98 error 14 in 18[606000+1000] 19[18967]: segfault at 8 ip 00007f55ef9f77e8 sp 00007ffe6f8b7ec0 error 4 in ld-2.17.so[7f55ef9ec000+22000] LustreError: 23661:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x240000403:0x1c1f:0x0] error -5. LustreError: 23661:0:(vvp_io.c:1903:vvp_io_init()) Skipped 4 previous similar messages 4[28247]: segfault at 0 ip (null) sp 00007fff272bbd58 error 14 in 4[400000+6000] 15[29010]: segfault at 8 ip 00007f77cd5d67e8 sp 00007ffcb28f5ef0 error 4 in ld-2.17.so[7f77cd5cb000+22000] 3[29880]: segfault at 0 ip 0000000000403e5f sp 00007ffd492433a0 error 6 in 16[400000+6000] Lustre: dir [0x240000402:0x1f95:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 13 previous similar messages LustreError: 30596:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff88012ea9c800: inode [0x200000403:0x2052:0x0] mdc close failed: rc = -2 LustreError: 30596:0:(file.c:247:ll_close_inode_openhandle()) Skipped 10 previous similar messages LustreError: 1162:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 15 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 LustreError: 1162:0:(statahead.c:825:ll_statahead_interpret_work()) Skipped 2 previous similar messages 4[4769]: segfault at 8 ip 00007ff02e3b27e8 sp 00007ffffba02170 error 4 in ld-2.17.so[7ff02e3a7000+22000] 4[5197]: segfault at 8 ip 00007f45d721d7e8 sp 00007fff27fa7860 error 4 in ld-2.17.so[7f45d7212000+22000] 14[7446]: segfault at 8 ip 00007f539cd817e8 sp 00007ffea4e204f0 error 4 in ld-2.17.so[7f539cd76000+22000] LustreError: 9753:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff88012ea9c800: cannot apply new layout on [0x240000402:0x2041:0x0] : rc = -5 LustreError: 9753:0:(lov_object.c:1341:lov_layout_change()) Skipped 39 previous similar messages 1[12262]: segfault at 8 ip 00007feec79447e8 sp 00007fff557a3e80 error 4 in ld-2.17.so[7feec7939000+22000] 0[19731]: segfault at 0 ip 00000000004016cd sp 00007fff82a10d90 error 6 in 0[400000+6000] 14[23596]: segfault at 406000 ip 0000000000406000 sp 00007ffcf52aca68 error 14 in 14[606000+1000] LustreError: 46:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 19 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 LustreError: 46:0:(statahead.c:825:ll_statahead_interpret_work()) Skipped 1 previous similar message LustreError: 24827:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x240000402:0x21ef:0x0] error -5. LustreError: 24827:0:(vvp_io.c:1903:vvp_io_init()) Skipped 13 previous similar messages LustreError: 10495:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000402:0x2a2e:0x0]: rc = -5 LustreError: 10721:0:(llite_lib.c:3697:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 10721:0:(llite_lib.c:3697:ll_prep_inode()) Skipped 270 previous similar messages LustreError: 10495:0:(lcommon_cl.c:188:cl_file_inode_init()) Skipped 271 previous similar messages 17[731]: segfault at 0 ip 0000000000403e5f sp 00007ffc4df67140 error 6 in 17[400000+6000] 8[3969]: segfault at 0 ip 0000000000403e5f sp 00007ffd1c4721a0 error 6 in 8[400000+6000] 13[7101]: segfault at 8 ip 00007f371dee67e8 sp 00007ffc71e1e090 error 4 in ld-2.17.so[7f371dedb000+22000] 2[12229]: segfault at 8 ip 00007f0f28ba67e8 sp 00007ffd7f449470 error 4 in ld-2.17.so[7f0f28b9b000+22000] 0[14964]: segfault at 0 ip 0000000000401d60 sp 00007ffdf42c5568 error 6 in 0[400000+6000] LustreError: 54:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 4 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 19[16730]: segfault at 8 ip 00007fec6960a7e8 sp 00007ffcd61b21d0 error 4 in ld-2.17.so[7fec695ff000+22000] 12[17198]: segfault at 8 ip 00007fbff82f77e8 sp 00007ffc8b50d700 error 4 in ld-2.17.so[7fbff82ec000+22000] 17[20432]: segfault at 8 ip 00007f662fe267e8 sp 00007ffdb781cca0 error 4 in ld-2.17.so[7f662fe1b000+22000] Lustre: dir [0x240000402:0x3314:0x0] stripe 2 readdir failed: -2, directory is partially accessed! Lustre: Skipped 21 previous similar messages 16[25001]: segfault at 8 ip 00007f0d76e337e8 sp 00007ffdcbf894e0 error 4 in ld-2.17.so[7f0d76e28000+22000] 6[24946]: segfault at 0 ip 0000000000403e5f sp 00007ffe2489ff40 error 6 in 6[400000+6000] LustreError: 26938:0:(llite_lib.c:1845:ll_update_lsm_md()) lustre: [0x200000403:0x33e5:0x0] dir layout mismatch: LustreError: 26938:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=2 count=2 index=0 hash=crush:0x2000003 max_inherit=0 max_inherit_rr=0 version=2 migrate_offset=0 migrate_hash=invalid:0 pool= LustreError: 26938:0:(lustre_lmv.h:167:lmv_stripe_object_dump()) stripe[0] [0x200000400:0xb1:0x0] LustreError: 26938:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=1 count=3 index=0 hash=crush:0x82000003 max_inherit=0 max_inherit_rr=0 version=1 migrate_offset=2 migrate_hash=fnv_1a_64:2 pool= LustreError: 32329:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff88012a528000: inode [0x200000402:0x3cba:0x0] mdc close failed: rc = -2 LustreError: 32329:0:(file.c:247:ll_close_inode_openhandle()) Skipped 9 previous similar messages LustreError: lustre-OST0000-osc-ffff88012a528000: operation ldlm_enqueue to node 192.168.203.144@tcp failed: rc = -107 Lustre: lustre-OST0000-osc-ffff88012a528000: Connection to lustre-OST0000 (at 192.168.203.144@tcp) was lost; in progress operations using this service will wait for recovery to complete LustreError: 2014:0:(osc_request.c:1005:osc_init_grant()) lustre-OST0000-osc-ffff88012a528000: granted 8437760 but already consumed 23216128 LustreError: lustre-OST0000-osc-ffff88012a528000: This client was evicted by lustre-OST0000; in progress operations using this service will fail. Lustre: 2017:0:(llite_lib.c:4142:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.203.144@tcp:/lustre/fid: [0x200000402:0x3ec1:0x0]/ may get corrupted (rc -5) Lustre: 2018:0:(llite_lib.c:4142:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.203.144@tcp:/lustre/fid: [0x240000402:0x34fa:0x0]// may get corrupted (rc -108) Lustre: 2018:0:(llite_lib.c:4142:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.203.144@tcp:/lustre/fid: [0x200000402:0x3f16:0x0]/ may get corrupted (rc -108) Lustre: 2018:0:(llite_lib.c:4142:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.203.144@tcp:/lustre/fid: [0x200000402:0x3f27:0x0]/ may get corrupted (rc -108) Lustre: 2018:0:(llite_lib.c:4142:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.203.144@tcp:/lustre/fid: [0x200000403:0x3cbe:0x0]/ may get corrupted (rc -108) Lustre: 2017:0:(llite_lib.c:4142:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.203.144@tcp:/lustre/fid: [0x240000403:0x352d:0x0]/ may get corrupted (rc -5) Lustre: 2017:0:(llite_lib.c:4142:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.203.144@tcp:/lustre/fid: [0x200000403:0x3c88:0x0]/ may get corrupted (rc -5) Lustre: 2018:0:(llite_lib.c:4142:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.203.144@tcp:/lustre/fid: [0x240000402:0x3672:0x0]// may get corrupted (rc -108) Lustre: 2017:0:(llite_lib.c:4142:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.203.144@tcp:/lustre/fid: [0x240000403:0x30c9:0x0]/ may get corrupted (rc -5) Lustre: 2017:0:(llite_lib.c:4142:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.203.144@tcp:/lustre/fid: [0x240000402:0x2db6:0x0]/ may get corrupted (rc -5) Lustre: 2018:0:(llite_lib.c:4142:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.203.144@tcp:/lustre/fid: [0x240000402:0x3632:0x0]/ may get corrupted (rc -108) Lustre: 2018:0:(llite_lib.c:4142:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.203.144@tcp:/lustre/fid: [0x240000403:0x368c:0x0]/ may get corrupted (rc -108) Lustre: 2018:0:(llite_lib.c:4142:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.203.144@tcp:/lustre/fid: [0x200000402:0x3f42:0x0]/ may get corrupted (rc -108) Lustre: 2018:0:(llite_lib.c:4142:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.203.144@tcp:/lustre/fid: [0x240000402:0x3530:0x0]/ may get corrupted (rc -108) Lustre: 2018:0:(llite_lib.c:4142:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.203.144@tcp:/lustre/fid: [0x200000402:0x3fac:0x0]// may get corrupted (rc -108) Lustre: 2018:0:(llite_lib.c:4142:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.203.144@tcp:/lustre/fid: [0x240000402:0x368f:0x0]// may get corrupted (rc -108) Lustre: 2018:0:(llite_lib.c:4142:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.203.144@tcp:/lustre/fid: [0x240000402:0x3694:0x0]/ may get corrupted (rc -108) Lustre: 2018:0:(llite_lib.c:4142:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.203.144@tcp:/lustre/fid: [0x200000403:0x3d0b:0x0]/ may get corrupted (rc -108) Lustre: 2017:0:(llite_lib.c:4142:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.203.144@tcp:/lustre/fid: [0x240000403:0x1330:0x0]// may get corrupted (rc -5) Lustre: 2017:0:(llite_lib.c:4142:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.203.144@tcp:/lustre/fid: [0x240000402:0x2fac:0x0]/ may get corrupted (rc -5) Lustre: 2017:0:(llite_lib.c:4142:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.203.144@tcp:/lustre/fid: [0x200000402:0x197f:0x0]/ may get corrupted (rc -5) Lustre: 2017:0:(llite_lib.c:4142:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.203.144@tcp:/lustre/fid: [0x240000403:0x2919:0x0]/ may get corrupted (rc -5) LustreError: 3132:0:(ldlm_resource.c:982:ldlm_resource_complain()) lustre-OST0000-osc-ffff88012a528000: namespace resource [0x280000400:0x822:0x0].0x0 (ffff8800a5940900) refcount nonzero (1) after lock cleanup; forcing cleanup. Lustre: lustre-OST0000-osc-ffff88012a528000: Connection restored to (at 192.168.203.144@tcp) LustreError: 3528:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff88012a528000: cannot apply new layout on [0x240000403:0x3330:0x0] : rc = -5 LustreError: 3528:0:(lov_object.c:1341:lov_layout_change()) Skipped 11 previous similar messages 4[4834]: segfault at 8 ip 00007f39a51aa7e8 sp 00007ffef22a83c0 error 4 in ld-2.17.so[7f39a519f000+22000] 14[5638]: segfault at 8 ip 00007f5652c107e8 sp 00007fff2c2b4cf0 error 4 in ld-2.17.so[7f5652c05000+22000] 16[10341]: segfault at 8 ip 00007f78ac1e27e8 sp 00007ffea71b9eb0 error 4 in ld-2.17.so[7f78ac1d7000+22000] LustreError: 20627:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x240000402:0x3bb2:0x0] error -5. LustreError: 20627:0:(vvp_io.c:1903:vvp_io_init()) Skipped 6 previous similar messages 16[21717]: segfault at 8 ip 00007fcf216f77e8 sp 00007ffc44d15c20 error 4 in ld-2.17.so[7fcf216ec000+22000] 8[23663]: segfault at 8 ip 00007f64367ee7e8 sp 00007fff388d0e10 error 4 in ld-2.17.so[7f64367e3000+22000] LustreError: 46:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 18 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 2[27154]: segfault at 8 ip 00007f40385957e8 sp 00007ffff80c0b30 error 4 in ld-2.17.so[7f403858a000+22000] 2[27626]: segfault at 8 ip 00007f0c67b437e8 sp 00007ffcc1473710 error 4 in ld-2.17.so[7f0c67b38000+22000] 0[27982]: segfault at 8 ip 00007f8b3039f7e8 sp 00007fff4b0bca70 error 4 in ld-2.17.so[7f8b30394000+22000] 3[28859]: segfault at 8 ip 00007fb2c995a7e8 sp 00007ffd05ae6210 error 4 in ld-2.17.so[7fb2c994f000+22000] 0[29844]: segfault at 8 ip 00007f563f15e7e8 sp 00007fff72ebddc0 error 4 in ld-2.17.so[7f563f153000+22000] 0[32049]: segfault at 0 ip 0000000000403e5f sp 00007ffc8610aa60 error 6 in 0[400000+6000] | Link to test |
racer test 1: racer on clients: centos-45.localnet DURATION=2700 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 PGD 2ca984067 PUD 2687f4067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) osp(OE) ofd(OE) lod(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) crc32_generic lnet(OE) libcfs(OE) crc_t10dif crct10dif_generic crct10dif_common virtio_balloon virtio_console pcspkr i2c_piix4 ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm ata_piix virtio_blk drm_panel_orientation_quirks serio_raw floppy libata i2c_core CPU: 13 PID: 30928 Comm: ll_sa_30892 Kdump: loaded Tainted: P OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.16.0-3.module_el8.7.0+1218+f626c2ff 04/01/2014 task: ffff880091868010 ti: ffff8802804c4000 task.ti: ffff8802804c4000 RIP: 0010:[<ffffffff813f502c>] [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 RSP: 0018:ffff8802804c7d78 EFLAGS: 00010296 RAX: ffff88026d88d690 RBX: 0000000000000008 RCX: 000000010026001d RDX: 0000000000000026 RSI: ffff88026d88d9c0 RDI: 0000000000000008 RBP: ffff8802804c7d88 R08: ffff880262564f88 R09: 0000000000000000 R10: ffff8802625648e8 R11: ffff880262564f88 R12: 0000000000000000 R13: ffff8800972d8010 R14: ffff88026d88d9c0 R15: ffff880262564f88 FS: 0000000000000000(0000) GS:ffff880331d40000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 000000025fe02000 CR4: 00000000000007e0 Call Trace: [<ffffffffa15b3bf1>] ll_statahead_thread+0xdb1/0x2000 [lustre] [<ffffffff817e320e>] ? _raw_spin_lock+0x1e/0x20 [<ffffffff810dbb51>] ? put_prev_entity+0x31/0x400 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e329e>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffff810c834d>] ? finish_task_switch+0x5d/0x1b0 [<ffffffffa15b2e40>] ? ll_statahead_handle.constprop.28+0x1c0/0x1c0 [lustre] [<ffffffff810ba114>] kthread+0xe4/0xf0 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 | 11[25698]: segfault at 0 ip (null) sp 00007fffbed052e8 error 14 in 11[400000+6000] Lustre: 24186:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff8802d6f87340 x1830163809158144/t4294969685(0) o101->5d557860-6452-4659-85e9-4540b219faa6@0@lo:5/0 lens 376/816 e 0 to 0 dl 1745380315 ref 1 fl Interpret:H/602/0 rc 0/0 job:'dd.0' uid:0 gid:0 projid:0 Lustre: 24186:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000401:0xc8:0x0] with magic=0xbd60bd0 15[30936]: segfault at 8 ip 00007fdd73cf87e8 sp 00007fff7fabf2d0 error 4 in ld-2.17.so[7fdd73ced000+22000] LustreError: 8084:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a2eb4138: inode [0x200000401:0x269:0x0] mdc close failed: rc = -13 Lustre: 22371:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000401:0x3c9:0x0] with magic=0xbd60bd0 Lustre: 22371:0:(lod_lov.c:1417:lod_parse_striping()) Skipped 1 previous similar message LustreError: 29058:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a2eb4138: inode [0x200000402:0x59a:0x0] mdc close failed: rc = -13 19[30849]: segfault at 8 ip 00007f1cfba507e8 sp 00007fff5c014370 error 4 in ld-2.17.so[7f1cfba45000+22000] Lustre: 23071:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000402:0x771:0x0] with magic=0xbd60bd0 Lustre: 23071:0:(lod_lov.c:1417:lod_parse_striping()) Skipped 1 previous similar message Lustre: 11845:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000401:0x768:0x0] with magic=0xbd60bd0 Lustre: 11845:0:(lod_lov.c:1417:lod_parse_striping()) Skipped 1 previous similar message 4[11357]: segfault at 8 ip 00007fd7720007e8 sp 00007ffcce67f290 error 4 in ld-2.17.so[7fd771ff5000+22000] 14[14444]: segfault at 8 ip 00007f7948d4f7e8 sp 00007fff8a0c0fe0 error 4 in ld-2.17.so[7f7948d44000+22000] 18[16343]: segfault at 0 ip 0000000000403e5f sp 00007ffdf694ea80 error 6 in 18[400000+6000] LustreError: 25744:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a2eb4138: inode [0x200000401:0x966:0x0] mdc close failed: rc = -13 Lustre: 2992:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000402:0x9da:0x0] with magic=0xbd60bd0 Lustre: 2992:0:(lod_lov.c:1417:lod_parse_striping()) Skipped 1 previous similar message LustreError: 22287:0:(mdt_handler.c:746:mdt_pack_acl2body()) lustre-MDT0000: unable to read [0x200000401:0x946:0x0] ACL: rc = -2 Lustre: 11844:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000401:0xfc8:0x0] with magic=0xbd60bd0 Lustre: 11844:0:(lod_lov.c:1417:lod_parse_striping()) Skipped 1 previous similar message Lustre: 2995:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000401:0x10d3:0x0] with magic=0xbd60bd0 Lustre: 2995:0:(lod_lov.c:1417:lod_parse_striping()) Skipped 1 previous similar message LustreError: 2995:0:(mdt_handler.c:746:mdt_pack_acl2body()) lustre-MDT0000: unable to read [0x200000402:0x10f6:0x0] ACL: rc = -2 13[25983]: segfault at 8 ip 00007f87a96517e8 sp 00007ffc701e0300 error 4 in ld-2.17.so[7f87a9646000+22000] LustreError: 13400:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802de7e4a88: inode [0x200000402:0x1515:0x0] mdc close failed: rc = -13 Lustre: 25147:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000402:0x172c:0x0] with magic=0xbd60bd0 Lustre: 25147:0:(lod_lov.c:1417:lod_parse_striping()) Skipped 3 previous similar messages LustreError: 4613:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802de7e4a88: inode [0x200000401:0x186e:0x0] mdc close failed: rc = -13 5[22635]: segfault at 8 ip 00007fb16b8c37e8 sp 00007ffd6becc4c0 error 4 in ld-2.17.so[7fb16b8b8000+22000] LustreError: 2997:0:(mdt_handler.c:746:mdt_pack_acl2body()) lustre-MDT0000: unable to read [0x200000402:0x1b38:0x0] ACL: rc = -2 LustreError: 2997:0:(mdt_handler.c:746:mdt_pack_acl2body()) Skipped 1 previous similar message LustreError: 17380:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802de7e4a88: inode [0x200000401:0x1e95:0x0] mdc close failed: rc = -13 LustreError: 17380:0:(file.c:247:ll_close_inode_openhandle()) Skipped 1 previous similar message 9[19198]: segfault at 8 ip 00007f2e2ad2a7e8 sp 00007fff1c7c6fd0 error 4 in ld-2.17.so[7f2e2ad1f000+22000] Lustre: 22514:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000401:0x1e8d:0x0] with magic=0xbd60bd0 Lustre: 22514:0:(lod_lov.c:1417:lod_parse_striping()) Skipped 3 previous similar messages LustreError: 21445:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802de7e4a88: inode [0x200000401:0x2360:0x0] mdc close failed: rc = -13 LustreError: 21445:0:(file.c:247:ll_close_inode_openhandle()) Skipped 1 previous similar message 14[27964]: segfault at 8 ip 00007f693d13a7e8 sp 00007ffef9a7e3a0 error 4 in ld-2.17.so[7f693d12f000+22000] Lustre: 14278:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000401:0x2410:0x0] with magic=0xbd60bd0 Lustre: 14278:0:(lod_lov.c:1417:lod_parse_striping()) Skipped 3 previous similar messages 9[29671]: segfault at 0 ip 0000000000403e5f sp 00007ffca0d2f730 error 6 in 9[400000+6000] LustreError: 14278:0:(mdt_handler.c:746:mdt_pack_acl2body()) lustre-MDT0000: unable to read [0x200000401:0x24dc:0x0] ACL: rc = -2 17[10172]: segfault at 8 ip 00007fb8b56ee7e8 sp 00007ffce936b430 error 4 in ld-2.17.so[7fb8b56e3000+22000] 17[10701]: segfault at 8 ip 00007fcdddbfa7e8 sp 00007ffdf2249850 error 4 in ld-2.17.so[7fcdddbef000+22000] LustreError: 14633:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a2eb4138: inode [0x200000402:0x25fa:0x0] mdc close failed: rc = -13 0[20514]: segfault at 8 ip 00007fb74fb237e8 sp 00007fffe1b56a30 error 4 in ld-2.17.so[7fb74fb18000+22000] | Link to test |
racer test 1: racer on clients: oleg447-client.virtnet DURATION=300 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffff813f40ac>] _atomic_dec_and_lock+0xc/0x70 PGD 12d51e067 PUD 136521067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) osp(OE) ofd(OE) lod(OE) mdt(OE) mdd(OE) mgs(OE) lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) crc_t10dif crct10dif_generic rpcsec_gss_krb5 sb_edac edac_core iosf_mbi kvm_intel kvm irqbypass crc32_pclmul ghash_clmulni_intel squashfs aesni_intel lrw gf128mul glue_helper ablk_helper cryptd i2c_piix4 pcspkr i2c_core binfmt_misc ip_tables ext4 mbcache jbd2 ata_generic pata_acpi crct10dif_pclmul crct10dif_common crc32c_intel ata_piix serio_raw libata CPU: 2 PID: 27412 Comm: ll_sa_24787 Kdump: loaded Tainted: G OE ------------ 3.10.0-7.9-debug #1 Hardware name: QEMU Standard PC (i440FX + PIIX, 1996), BIOS 1.16.2-1.fc38 04/01/2014 task: ffff88009db80000 ti: ffff88008bad8000 task.ti: ffff88008bad8000 RIP: 0010:[<ffffffff813f40ac>] [<ffffffff813f40ac>] _atomic_dec_and_lock+0xc/0x70 RSP: 0018:ffff88008badbd78 EFLAGS: 00010296 RAX: ffff88008b19a8d0 RBX: 0000000000000008 RCX: 000000018010000e RDX: 000000018010000f RSI: ffff88008b19ac00 RDI: 0000000000000008 RBP: ffff88008badbd88 R08: ffff88009421e600 R09: 000000018010000e R10: 000000009421fe01 R11: ffff88009421e600 R12: 0000000000000000 R13: ffff8800a4c16660 R14: ffff88008b19ac00 R15: ffff88009421e600 FS: 0000000000000000(0000) GS:ffff88013e300000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 CR2: 0000000000000008 CR3: 0000000130e56000 CR4: 0000000000160fe0 Call Trace: [<ffffffffa10d3c51>] ll_statahead_thread+0xdb1/0x2000 [lustre] [<ffffffff810df6e2>] ? pick_next_task_fair+0x702/0x8a0 [<ffffffff81410a79>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e8dce>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffff810c834d>] ? finish_task_switch+0x5d/0x1b0 [<ffffffffa10d2ea0>] ? ll_statahead_handle.constprop.28+0x1c0/0x1c0 [lustre] [<ffffffff810ba114>] kthread+0xe4/0xf0 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817f3e5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 | cat (11140) used greatest stack depth: 10608 bytes left random: crng init done LustreError: 12268:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff88012bf2e800: inode [0x200000402:0x2c:0x0] mdc close failed: rc = -116 LustreError: 12563:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff88012bf2e800: inode [0x200000402:0x57:0x0] mdc close failed: rc = -116 LustreError: 12563:0:(file.c:247:ll_close_inode_openhandle()) Skipped 1 previous similar message LustreError: 13796:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800b6d34800: inode [0x240000402:0x39:0x0] mdc close failed: rc = -116 LustreError: 13796:0:(file.c:247:ll_close_inode_openhandle()) Skipped 1 previous similar message Lustre: dir [0x200000403:0x122:0x0] stripe 1 readdir failed: -2, directory is partially accessed! 10[15435]: segfault at 0 ip 0000000000403e5f sp 00007ffde1b08de0 error 6 in 10[400000+6000] LustreError: 16093:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800b6d34800: inode [0x200000403:0x152:0x0] mdc close failed: rc = -116 LustreError: 16093:0:(file.c:247:ll_close_inode_openhandle()) Skipped 3 previous similar messages LustreError: 16179:0:(lcommon_cl.c:179:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000402:0x95:0x0]: rc = -5 LustreError: 16179:0:(llite_lib.c:3698:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 Lustre: dir [0x200000403:0x125:0x0] stripe 2 readdir failed: -2, directory is partially accessed! 4[17052]: segfault at 0 ip 0000000000403e5f sp 00007ffd225a4930 error 6 in 4[400000+6000] LustreError: 17629:0:(lcommon_cl.c:179:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000402:0x95:0x0]: rc = -5 LustreError: 17629:0:(llite_lib.c:3698:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 17598:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff88012bf2e800: cannot apply new layout on [0x240000402:0x1be:0x0] : rc = -5 LustreError: 17598:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x240000402:0x1be:0x0] error -5. LustreError: 16862:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800b6d34800: inode [0x200000402:0x1a3:0x0] mdc close failed: rc = -2 LustreError: 16862:0:(file.c:247:ll_close_inode_openhandle()) Skipped 15 previous similar messages LustreError: 19634:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff88012bf2e800: cannot apply new layout on [0x240000402:0x95:0x0] : rc = -5 LustreError: 19634:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x240000402:0x95:0x0] error -5. 14[22218]: segfault at 8 ip 00007fd9c8f107e8 sp 00007ffd834b7da0 error 4 in ld-2.17.so[7fd9c8f05000+22000] LustreError: 22372:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff88012bf2e800: cannot apply new layout on [0x200000402:0x359:0x0] : rc = -5 LustreError: 22372:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x200000402:0x359:0x0] error -5. Lustre: dir [0x240000402:0x2db:0x0] stripe 2 readdir failed: -2, directory is partially accessed! Lustre: Skipped 2 previous similar messages 17[23726]: segfault at 8 ip 00007f1f6bd527e8 sp 00007ffc5be9a4b0 error 4 in ld-2.17.so[7f1f6bd47000+22000] LustreError: 25039:0:(lcommon_cl.c:179:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000403:0x29a:0x0]: rc = -5 LustreError: 25039:0:(lcommon_cl.c:179:cl_file_inode_init()) Skipped 2 previous similar messages LustreError: 25039:0:(llite_lib.c:3698:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 25039:0:(llite_lib.c:3698:ll_prep_inode()) Skipped 2 previous similar messages Lustre: dir [0x240000403:0x3ca:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 5 previous similar messages LustreError: 25483:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff88012bf2e800: cannot apply new layout on [0x200000402:0x3de:0x0] : rc = -5 LustreError: 25483:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x200000402:0x3de:0x0] error -5. LustreError: 26588:0:(lcommon_cl.c:179:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000402:0x315:0x0]: rc = -5 LustreError: 26588:0:(lcommon_cl.c:179:cl_file_inode_init()) Skipped 6 previous similar messages LustreError: 26588:0:(llite_lib.c:3698:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 26588:0:(llite_lib.c:3698:ll_prep_inode()) Skipped 6 previous similar messages LustreError: 27481:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff88012bf2e800: inode [0x200000403:0x427:0x0] mdc close failed: rc = -116 LustreError: 27481:0:(file.c:247:ll_close_inode_openhandle()) Skipped 34 previous similar messages 3[28566]: segfault at 8 ip 00007f1420bf07e8 sp 00007ffe5e2e9370 error 4 in ld-2.17.so[7f1420be5000+22000] 19[29312]: segfault at 8 ip 00007f1fb32bd7e8 sp 00007ffcc1afaa90 error 4 in ld-2.17.so[7f1fb32b2000+22000] LustreError: 31372:0:(lcommon_cl.c:179:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000402:0x55d:0x0]: rc = -5 LustreError: 31372:0:(lcommon_cl.c:179:cl_file_inode_init()) Skipped 1 previous similar message LustreError: 31372:0:(llite_lib.c:3698:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 31372:0:(llite_lib.c:3698:ll_prep_inode()) Skipped 1 previous similar message 2[31258]: segfault at 8 ip 00007f2eb91f97e8 sp 00007fff7044e790 error 4 in ld-2.17.so[7f2eb91ee000+22000] Lustre: lustre-MDT0000-mdc-ffff88012bf2e800: Connection to lustre-MDT0000 (at 192.168.204.147@tcp) was lost; in progress operations using this service will wait for recovery to complete LustreError: lustre-MDT0000-mdc-ffff88012bf2e800: operation ldlm_enqueue to node 192.168.204.147@tcp failed: rc = -107 LustreError: lustre-MDT0000-mdc-ffff88012bf2e800: This client was evicted by lustre-MDT0000; in progress operations using this service will fail. LustreError: 28610:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff88012bf2e800: inode [0x200000403:0x427:0x0] mdc close failed: rc = -108 LustreError: 27080:0:(mdc_request.c:1464:mdc_read_page()) lustre-MDT0000-mdc-ffff88012bf2e800: [0x200000400:0x22:0x0] lock enqueue fails: rc = -5 Lustre: dir [0x200000403:0x5aa:0x0] stripe 0 readdir failed: -5, directory is partially accessed! LustreError: 31411:0:(file.c:6138:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000402:0x7a9:0x0] error: rc = -5 LustreError: 28610:0:(file.c:247:ll_close_inode_openhandle()) Skipped 9 previous similar messages LustreError: 31669:0:(ldlm_resource.c:982:ldlm_resource_complain()) lustre-MDT0000-mdc-ffff88012bf2e800: namespace resource [0x200000007:0x1:0x0].0x0 (ffff88009dbe8600) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 32203:0:(file.c:6138:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000007:0x1:0x0] error: rc = -108 LustreError: 32203:0:(file.c:6138:ll_inode_revalidate_fini()) Skipped 636 previous similar messages Lustre: lustre-MDT0000-mdc-ffff88012bf2e800: Connection restored to (at 192.168.204.147@tcp) 9[412]: segfault at 8 ip 00007f7ac407d7e8 sp 00007ffe52934a50 error 4 in ld-2.17.so[7f7ac4072000+22000] LustreError: 46:0:(lcommon_cl.c:179:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000402:0x5cb:0x0]: rc = -5 LustreError: 46:0:(lcommon_cl.c:179:cl_file_inode_init()) Skipped 2 previous similar messages LustreError: 46:0:(llite_lib.c:3698:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 46:0:(llite_lib.c:3698:ll_prep_inode()) Skipped 2 previous similar messages LustreError: 46:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 9 [0x200000402:0x5cb:0x0] inode@0000000000000000: rc = -5 10[901]: segfault at 8 ip 00007f92eee8b7e8 sp 00007ffd4105cee0 error 4 in ld-2.17.so[7f92eee80000+22000] 19[3804]: segfault at 0 ip (null) sp 00007fffc75d04a8 error 14 in 19[400000+6000] 10[5782]: segfault at 1c ip 00000000004017b0 sp 00007ffc1e51cfa0 error 6 in 10[400000+6000] Lustre: dir [0x240000402:0x355:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 5 previous similar messages 16[7752]: segfault at 8 ip 00007f5dfbda87e8 sp 00007ffcce401230 error 4 in ld-2.17.so[7f5dfbd9d000+22000] 6[7833]: segfault at 8 ip 00007f0f4ffc97e8 sp 00007ffec5a8f6e0 error 4 in ld-2.17.so[7f0f4ffbe000+22000] 3[8062]: segfault at 8 ip 00007ff1494907e8 sp 00007ffcad45b9f0 error 4 in ld-2.17.so[7ff149485000+22000] 9[12135]: segfault at 8 ip 00007f42ecc317e8 sp 00007ffc08000d60 error 4 in ld-2.17.so[7f42ecc26000+22000] LustreError: 11615:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff88012bf2e800: cannot apply new layout on [0x240000402:0xb8b:0x0] : rc = -5 LustreError: 11615:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x240000402:0xb8b:0x0] error -5. 4[13826]: segfault at 8 ip 00007f23687297e8 sp 00007ffc248a3df0 error 4 in ld-2.17.so[7f236871e000+22000] LustreError: 11838:0:(lcommon_cl.c:179:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000403:0xa0a:0x0]: rc = -5 LustreError: 11838:0:(lcommon_cl.c:179:cl_file_inode_init()) Skipped 56 previous similar messages LustreError: 11838:0:(llite_lib.c:3698:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 11838:0:(llite_lib.c:3698:ll_prep_inode()) Skipped 56 previous similar messages 5[15028]: segfault at 406000 ip 0000000000406000 sp 00007ffd92c15418 error 14 in 5[606000+1000] 8[16414]: segfault at 0 ip 0000000000403e5f sp 00007ffc1830cd10 error 6 in 8[400000+6000] 2[18340]: segfault at 8 ip 00007ff9dabde7e8 sp 00007ffcedd896b0 error 4 in ld-2.17.so[7ff9dabd3000+22000] 2[18399]: segfault at 8 ip 00007f2825fad7e8 sp 00007ffdfade06d0 error 4 in ld-2.17.so[7f2825fa2000+22000] 18[18681]: segfault at 8 ip 00007f2fa1e5e7e8 sp 00007ffd471bdd90 error 4 in ld-2.17.so[7f2fa1e53000+22000] 8[20673]: segfault at 8 ip 00007f8658e5f7e8 sp 00007ffefe0e8db0 error 4 in ld-2.17.so[7f8658e54000+22000] Lustre: dir [0x240000403:0xdca:0x0] stripe 2 readdir failed: -2, directory is partially accessed! Lustre: Skipped 4 previous similar messages LustreError: 20731:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff88012bf2e800: cannot apply new layout on [0x200000402:0x96c:0x0] : rc = -5 LustreError: 20731:0:(lov_object.c:1341:lov_layout_change()) Skipped 5 previous similar messages LustreError: 24593:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x200000404:0x6e2:0x0] error -5. LustreError: 24593:0:(vvp_io.c:1903:vvp_io_init()) Skipped 3 previous similar messages LustreError: 26570:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800b6d34800: inode [0x200000404:0x9b9:0x0] mdc close failed: rc = -116 LustreError: 26570:0:(file.c:247:ll_close_inode_openhandle()) Skipped 70 previous similar messages | Link to test |
racer test 1: racer on clients: centos-20.localnet DURATION=2700 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 PGD 289a07067 PUD 2c9ebc067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) lustre(OE) osp(OE) ofd(OE) lod(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) jbd2 mbcache lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) crc32_generic lnet(OE) dm_flakey dm_mod libcfs(OE) crc_t10dif crct10dif_generic crct10dif_common virtio_console i2c_piix4 pcspkr virtio_balloon ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm ata_piix serio_raw drm_panel_orientation_quirks virtio_blk i2c_core libata floppy CPU: 3 PID: 22239 Comm: ll_sa_22015 Kdump: loaded Tainted: P OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.16.0-3.module_el8.7.0+1218+f626c2ff 04/01/2014 task: ffff8802d0bfdc40 ti: ffff8802bb0c0000 task.ti: ffff8802bb0c0000 RIP: 0010:[<ffffffff813f502c>] [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 RSP: 0018:ffff8802bb0c3d78 EFLAGS: 00010296 RAX: ffff8802794d3090 RBX: 0000000000000008 RCX: 0000000100260022 RDX: 0000000000000026 RSI: ffff8802794d33c0 RDI: 0000000000000008 RBP: ffff8802bb0c3d88 R08: ffff8800a06b2128 R09: 0000000000000000 R10: ffff8800a06b52d8 R11: ffff8800a06b2128 R12: 0000000000000000 R13: ffff8802c12724f0 R14: ffff8802794d33c0 R15: ffff8800a06b2128 FS: 0000000000000000(0000) GS:ffff880331ac0000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 00000002bcf3e000 CR4: 00000000000007e0 Call Trace: [<ffffffffa1114461>] ll_statahead_thread+0xdb1/0x2000 [lustre] [<ffffffff810df6e2>] ? pick_next_task_fair+0x702/0x8a0 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e329e>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffff810c834d>] ? finish_task_switch+0x5d/0x1b0 [<ffffffffa11136b0>] ? ll_statahead_handle.constprop.28+0x1c0/0x1c0 [lustre] [<ffffffff810ba114>] kthread+0xe4/0xf0 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 | Lustre: 12468:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 618, rollback = 7 Lustre: 12468:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 12468:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 12468:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 12468:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 12468:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 traps: 0[18237] general protection ip:404b70 sp:7ffe1ec03168 error:0 in 9[400000+6000] touch (17565) used greatest stack depth: 10080 bytes left Lustre: 18501:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 511 < left 699, rollback = 2 Lustre: 18501:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 3/12/1, destroy: 0/0/0 Lustre: 18501:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 18501:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 2/2/0, xattr_set: 9/699/0 Lustre: 18501:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 18501:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 5/55/0, punch: 0/0/0, quota 1/3/0 Lustre: 18501:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 18501:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 12/275/3, delete: 0/0/0 Lustre: 18501:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 18501:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 6/6/0, ref_del: 0/0/0 Lustre: 18501:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 12470:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 618, rollback = 7 Lustre: 12470:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 1 previous similar message Lustre: 12470:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 12470:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 15 previous similar messages Lustre: 12470:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 12470:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 15 previous similar messages Lustre: 12470:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 12470:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 15 previous similar messages Lustre: 12470:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 12470:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 15 previous similar messages Lustre: 12470:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 12470:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 15 previous similar messages Lustre: 18501:0:(mdd_dir.c:4741:mdd_migrate_object()) lustre-MDD0000: [0x200000403:0x2:0x0]/0 is open, migrate only dentry Lustre: 18195:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 618, rollback = 7 Lustre: 18195:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 1 previous similar message Lustre: 18195:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 18195:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 18195:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 18195:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 18195:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 18195:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 18195:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 18195:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 18195:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 18195:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 1 previous similar message 9[19215]: segfault at 8 ip 00007f657265f7e8 sp 00007ffcf3e1d040 error 4 in ld-2.17.so[7f6572654000+22000] Lustre: 10851:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 508 < left 699, rollback = 2 Lustre: 10851:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 15 previous similar messages Lustre: 10831:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0xc:0x0] with magic=0xbd60bd0 7[19911]: segfault at 8 ip 00007fe7b7bd77e8 sp 00007ffceccd6640 error 4 in ld-2.17.so[7fe7b7bcc000+22000] Lustre: 17903:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff8802d2b48f40 x1829451541470208/t4294968147(0) o101->4c8b5d62-e3d7-411b-8f5a-4e94e2e0e8e8@0@lo:238/0 lens 376/864 e 0 to 0 dl 1744701048 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 Lustre: 18195:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 618, rollback = 7 Lustre: 18195:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 1 previous similar message Lustre: 18195:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 18195:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 17 previous similar messages Lustre: 18195:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 18195:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 17 previous similar messages Lustre: 18195:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 18195:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 17 previous similar messages Lustre: 18195:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 18195:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 17 previous similar messages Lustre: 18195:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 18195:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 17 previous similar messages Lustre: dir [0x200000403:0x48:0x0] stripe 3 readdir failed: -2, directory is partially accessed! Lustre: 10848:0:(mdd_dir.c:4741:mdd_migrate_object()) lustre-MDD0001: [0x240000404:0x2a:0x0]/13 is open, migrate only dentry Lustre: 17168:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff8802c105a340 x1829451542002048/t4294968218(0) o101->4c8b5d62-e3d7-411b-8f5a-4e94e2e0e8e8@0@lo:241/0 lens 376/864 e 0 to 0 dl 1744701051 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 4[20254]: segfault at 8 ip 00007f95a3ec57e8 sp 00007ffc4bfe4e00 error 4 in ld-2.17.so[7f95a3eba000+22000] Lustre: 10848:0:(mdd_dir.c:4741:mdd_migrate_object()) lustre-MDD0000: [0x200000403:0x2:0x0]/17 is open, migrate only dentry Lustre: 12468:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 618, rollback = 7 Lustre: 12468:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 1 previous similar message Lustre: 20664:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 509 < left 788, rollback = 2 Lustre: 20664:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 15 previous similar messages Lustre: 17343:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 17343:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 32 previous similar messages Lustre: 17343:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 17343:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 32 previous similar messages Lustre: 17343:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 17343:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 32 previous similar messages Lustre: 17343:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 17343:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 32 previous similar messages Lustre: 17343:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 17343:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 32 previous similar messages Lustre: 12468:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 618, rollback = 7 Lustre: 12468:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 3 previous similar messages 8[24671]: segfault at 406000 ip 0000000000406000 sp 00007ffc6b50f738 error 14 in 8[606000+1000] 12[25544]: segfault at 0 ip 0000000000403e5f sp 00007fff57027260 error 6 in 12[400000+6000] Lustre: 12469:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 12469:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 12469:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 12469:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 12469:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/594/0, punch: 0/0/0, quota 1/3/0 Lustre: 12469:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 12469:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 12469:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 12469:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 12469:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 10851:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 509 < left 1695, rollback = 2 Lustre: 10851:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 28 previous similar messages Lustre: 12909:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0x2dc:0x0] with magic=0xbd60bd0 Lustre: 12909:0:(lod_lov.c:1417:lod_parse_striping()) Skipped 1 previous similar message Lustre: 18195:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 618, rollback = 7 Lustre: 18195:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 3 previous similar messages Lustre: 18195:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 18195:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 49 previous similar messages Lustre: 18195:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 18195:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 49 previous similar messages Lustre: 18195:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 18195:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 49 previous similar messages Lustre: 18195:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 18195:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 49 previous similar messages Lustre: 18195:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 18195:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 49 previous similar messages Lustre: 18501:0:(mdd_dir.c:4741:mdd_migrate_object()) lustre-MDD0000: [0x200000403:0x1:0x0]/4 is open, migrate only dentry Lustre: 18501:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 506 < left 877, rollback = 2 Lustre: 18501:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 47 previous similar messages Lustre: dir [0x200000403:0x475:0x0] stripe 2 readdir failed: -2, directory is partially accessed! Lustre: Skipped 2 previous similar messages 10[12016]: segfault at 0 ip 00000000004011b0 sp 00007ffd63dfaff0 error 6 in 10[400000+6000] Lustre: 10852:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 510 < left 1500, rollback = 2 Lustre: 10852:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 23 previous similar messages 3[13539]: segfault at 8 ip 00007f64255f17e8 sp 00007fff57907490 error 4 in ld-2.17.so[7f64255e6000+22000] LustreError: 13695:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000404:0x51d:0x0]: rc = -5 LustreError: 13695:0:(llite_lib.c:3696:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 1[16911]: segfault at 0 ip 0000000000403e5f sp 00007ffe175476a0 error 6 in 1[400000+6000] Lustre: 15348:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0x56c:0x0] with magic=0xbd60bd0 Lustre: 15348:0:(lod_lov.c:1417:lod_parse_striping()) Skipped 1 previous similar message Lustre: 18090:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff8802e6059e40 x1829451557099648/t4294979173(0) o101->67fa0521-646f-41f6-aaa0-db2cd751d950@0@lo:340/0 lens 376/816 e 0 to 0 dl 1744701150 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 Lustre: 17343:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 516 < left 564, rollback = 7 Lustre: 17343:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 1 previous similar message INFO: task file_chmod.sh:16310 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. file_chmod.sh D ffff8802e626b760 12696 16310 16270 0x00000080 Call Trace: [<ffffffff817e19d9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df7ea>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817e324e>] ? _raw_spin_unlock+0xe/0x20 [<ffffffff812512e3>] ? unlazy_walk+0x153/0x160 [<ffffffff817dfa1d>] mutex_lock+0x2d/0x40 [<ffffffff817d5b92>] lookup_slow+0x33/0xa7 [<ffffffff81256dee>] path_lookupat+0x89e/0x8d0 [<ffffffff8121f176>] ? kmem_cache_alloc+0x126/0x2e0 [<ffffffff81257c1f>] ? getname_flags+0x4f/0x1a0 [<ffffffff81256e4b>] filename_lookup+0x2b/0xc0 [<ffffffff81258d47>] user_path_at_empty+0x67/0xc0 [<ffffffff81258db1>] user_path_at+0x11/0x20 [<ffffffff8124b733>] vfs_fstatat+0x63/0xc0 [<ffffffff8124bafe>] SYSC_newstat+0x2e/0x60 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff8124bfce>] SyS_newstat+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a INFO: task file_mknod.sh:16312 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. file_mknod.sh D ffff8802dd558040 12832 16312 16270 0x00000080 Call Trace: [<ffffffff817e19d9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df7ea>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817e324e>] ? _raw_spin_unlock+0xe/0x20 [<ffffffff812512e3>] ? unlazy_walk+0x153/0x160 [<ffffffff817dfa1d>] mutex_lock+0x2d/0x40 [<ffffffff817d5b92>] lookup_slow+0x33/0xa7 [<ffffffff81256dee>] path_lookupat+0x89e/0x8d0 [<ffffffff8121f176>] ? kmem_cache_alloc+0x126/0x2e0 [<ffffffff81257c1f>] ? getname_flags+0x4f/0x1a0 [<ffffffff81256e4b>] filename_lookup+0x2b/0xc0 [<ffffffff81258d47>] user_path_at_empty+0x67/0xc0 [<ffffffff81258db1>] user_path_at+0x11/0x20 [<ffffffff8124b733>] vfs_fstatat+0x63/0xc0 [<ffffffff8124bafe>] SYSC_newstat+0x2e/0x60 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff8124bfce>] SyS_newstat+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a INFO: task file_delxattr.s:16355 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. file_delxattr.s D ffff8802dce2ba00 12800 16355 16270 0x00000080 Call Trace: [<ffffffff817e19d9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df7ea>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817e324e>] ? _raw_spin_unlock+0xe/0x20 [<ffffffff812512e3>] ? unlazy_walk+0x153/0x160 [<ffffffff817dfa1d>] mutex_lock+0x2d/0x40 [<ffffffff817d5b92>] lookup_slow+0x33/0xa7 [<ffffffff81256dee>] path_lookupat+0x89e/0x8d0 [<ffffffff8121f176>] ? kmem_cache_alloc+0x126/0x2e0 [<ffffffff81257c1f>] ? getname_flags+0x4f/0x1a0 [<ffffffff81256e4b>] filename_lookup+0x2b/0xc0 [<ffffffff81258d47>] user_path_at_empty+0x67/0xc0 [<ffffffff81258db1>] user_path_at+0x11/0x20 [<ffffffff8124b733>] vfs_fstatat+0x63/0xc0 [<ffffffff8124bafe>] SYSC_newstat+0x2e/0x60 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff8124bfce>] SyS_newstat+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a INFO: task file_setxattr.s:16359 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. file_setxattr.s D ffff8802e4be8608 12832 16359 16270 0x00000080 Call Trace: [<ffffffff817e19d9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df7ea>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817e324e>] ? _raw_spin_unlock+0xe/0x20 [<ffffffff812512e3>] ? unlazy_walk+0x153/0x160 [<ffffffff817dfa1d>] mutex_lock+0x2d/0x40 [<ffffffff817d5b92>] lookup_slow+0x33/0xa7 [<ffffffff81256dee>] path_lookupat+0x89e/0x8d0 [<ffffffff8121f176>] ? kmem_cache_alloc+0x126/0x2e0 [<ffffffff81257c1f>] ? getname_flags+0x4f/0x1a0 [<ffffffff81256e4b>] filename_lookup+0x2b/0xc0 [<ffffffff81258d47>] user_path_at_empty+0x67/0xc0 [<ffffffff81258db1>] user_path_at+0x11/0x20 [<ffffffff8124b733>] vfs_fstatat+0x63/0xc0 [<ffffffff8124bafe>] SYSC_newstat+0x2e/0x60 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff8124bfce>] SyS_newstat+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a INFO: task file_chmod.sh:16362 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. file_chmod.sh D ffff88031acb9900 12832 16362 16262 0x00000080 Call Trace: [<ffffffff817e19d9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df7ea>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817e324e>] ? _raw_spin_unlock+0xe/0x20 [<ffffffff812512e3>] ? unlazy_walk+0x153/0x160 [<ffffffff817dfa1d>] mutex_lock+0x2d/0x40 [<ffffffff817d5b92>] lookup_slow+0x33/0xa7 [<ffffffff81256dee>] path_lookupat+0x89e/0x8d0 [<ffffffff8121f176>] ? kmem_cache_alloc+0x126/0x2e0 [<ffffffff81257c1f>] ? getname_flags+0x4f/0x1a0 [<ffffffff81256e4b>] filename_lookup+0x2b/0xc0 [<ffffffff81258d47>] user_path_at_empty+0x67/0xc0 [<ffffffff81258db1>] user_path_at+0x11/0x20 [<ffffffff8124b733>] vfs_fstatat+0x63/0xc0 [<ffffffff8124bafe>] SYSC_newstat+0x2e/0x60 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff8124bfce>] SyS_newstat+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a INFO: task file_chmod.sh:16719 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. file_chmod.sh D ffff88031fbb2980 12800 16719 16292 0x00000080 Call Trace: [<ffffffff817e19d9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df7ea>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817e324e>] ? _raw_spin_unlock+0xe/0x20 [<ffffffff812512e3>] ? unlazy_walk+0x153/0x160 [<ffffffff817dfa1d>] mutex_lock+0x2d/0x40 [<ffffffff817d5b92>] lookup_slow+0x33/0xa7 [<ffffffff81256dee>] path_lookupat+0x89e/0x8d0 [<ffffffff8121f176>] ? kmem_cache_alloc+0x126/0x2e0 [<ffffffff81257c1f>] ? getname_flags+0x4f/0x1a0 [<ffffffff81256e4b>] filename_lookup+0x2b/0xc0 [<ffffffff81258d47>] user_path_at_empty+0x67/0xc0 [<ffffffff81258db1>] user_path_at+0x11/0x20 [<ffffffff8124b733>] vfs_fstatat+0x63/0xc0 [<ffffffff8124bafe>] SYSC_newstat+0x2e/0x60 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff8124bfce>] SyS_newstat+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a INFO: task file_chown.sh:16733 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. file_chown.sh D ffff8802dd55ca80 12832 16733 16274 0x00000080 Call Trace: [<ffffffff817e19d9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df7ea>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817e324e>] ? _raw_spin_unlock+0xe/0x20 [<ffffffff812512e3>] ? unlazy_walk+0x153/0x160 [<ffffffff817dfa1d>] mutex_lock+0x2d/0x40 [<ffffffff817d5b92>] lookup_slow+0x33/0xa7 [<ffffffff81256dee>] path_lookupat+0x89e/0x8d0 [<ffffffff8121f176>] ? kmem_cache_alloc+0x126/0x2e0 [<ffffffff81257c1f>] ? getname_flags+0x4f/0x1a0 [<ffffffff81256e4b>] filename_lookup+0x2b/0xc0 [<ffffffff81258d47>] user_path_at_empty+0x67/0xc0 [<ffffffff81258db1>] user_path_at+0x11/0x20 [<ffffffff8124b733>] vfs_fstatat+0x63/0xc0 [<ffffffff8124bafe>] SYSC_newstat+0x2e/0x60 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff8124bfce>] SyS_newstat+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a INFO: task file_chmod.sh:16734 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. file_chmod.sh D ffff8802dcd890c0 11424 16734 16274 0x00000080 Call Trace: [<ffffffff817e19d9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df7ea>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817e324e>] ? _raw_spin_unlock+0xe/0x20 [<ffffffff812512e3>] ? unlazy_walk+0x153/0x160 [<ffffffff817dfa1d>] mutex_lock+0x2d/0x40 [<ffffffff817d5b92>] lookup_slow+0x33/0xa7 [<ffffffff81256dee>] path_lookupat+0x89e/0x8d0 [<ffffffff8121f176>] ? kmem_cache_alloc+0x126/0x2e0 [<ffffffff81257c1f>] ? getname_flags+0x4f/0x1a0 [<ffffffff81256e4b>] filename_lookup+0x2b/0xc0 [<ffffffff81258d47>] user_path_at_empty+0x67/0xc0 [<ffffffff81258db1>] user_path_at+0x11/0x20 [<ffffffff8124b733>] vfs_fstatat+0x63/0xc0 [<ffffffff8124bafe>] SYSC_newstat+0x2e/0x60 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff8124bfce>] SyS_newstat+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a INFO: task file_chmod.sh:16737 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. file_chmod.sh D ffff8800a50de340 12696 16737 16278 0x00000080 Call Trace: [<ffffffff817e19d9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df7ea>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817e324e>] ? _raw_spin_unlock+0xe/0x20 [<ffffffff812512e3>] ? unlazy_walk+0x153/0x160 [<ffffffff817dfa1d>] mutex_lock+0x2d/0x40 [<ffffffff817d5b92>] lookup_slow+0x33/0xa7 [<ffffffff81256dee>] path_lookupat+0x89e/0x8d0 [<ffffffff8121f176>] ? kmem_cache_alloc+0x126/0x2e0 [<ffffffff81257c1f>] ? getname_flags+0x4f/0x1a0 [<ffffffff81256e4b>] filename_lookup+0x2b/0xc0 [<ffffffff81258d47>] user_path_at_empty+0x67/0xc0 [<ffffffff81258db1>] user_path_at+0x11/0x20 [<ffffffff8124b733>] vfs_fstatat+0x63/0xc0 [<ffffffff8124bafe>] SYSC_newstat+0x2e/0x60 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff8124bfce>] SyS_newstat+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a INFO: task file_mknod.sh:16739 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. file_mknod.sh D ffff8800a785b760 12800 16739 16278 0x00000080 Call Trace: [<ffffffff817e19d9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df7ea>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817e324e>] ? _raw_spin_unlock+0xe/0x20 [<ffffffff812512e3>] ? unlazy_walk+0x153/0x160 [<ffffffff817dfa1d>] mutex_lock+0x2d/0x40 [<ffffffff817d5b92>] lookup_slow+0x33/0xa7 [<ffffffff81256dee>] path_lookupat+0x89e/0x8d0 [<ffffffff8121f176>] ? kmem_cache_alloc+0x126/0x2e0 [<ffffffff81257c1f>] ? getname_flags+0x4f/0x1a0 [<ffffffff81256e4b>] filename_lookup+0x2b/0xc0 [<ffffffff81258d47>] user_path_at_empty+0x67/0xc0 [<ffffffff81258db1>] user_path_at+0x11/0x20 [<ffffffff8124b733>] vfs_fstatat+0x63/0xc0 [<ffffffff8124bafe>] SYSC_newstat+0x2e/0x60 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff8124bfce>] SyS_newstat+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a 17[24380]: segfault at 1c ip 00000000004017b0 sp 00007ffd57572d90 error 6 in 17[400000+6000] 15[25222]: segfault at 0 ip (null) sp 00007ffe10357f78 error 14 in 15[400000+6000] 9[26460]: segfault at 0 ip (null) sp 00007ffdb5bf2ca8 error 14 in 9[400000+6000] Lustre: 12469:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 12469:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 70 previous similar messages Lustre: 12469:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 12469:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 70 previous similar messages Lustre: 12469:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 12469:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 70 previous similar messages Lustre: 12469:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 12469:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 70 previous similar messages Lustre: 12469:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 12469:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 70 previous similar messages Lustre: 18055:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0x810:0x0] with magic=0xbd60bd0 Lustre: 18055:0:(lod_lov.c:1417:lod_parse_striping()) Skipped 1 previous similar message LustreError: 15449:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e739ca88: inode [0x200000404:0x543:0x0] mdc close failed: rc = -13 Lustre: 10848:0:(mdd_dir.c:4741:mdd_migrate_object()) lustre-MDD0002: [0x280000403:0x1:0x0]/0 is open, migrate only dentry 7[30845]: segfault at 1c ip 00000000004017b0 sp 00007ffe3bda4e60 error 6 in 7[400000+6000] 9[30815]: segfault at 8 ip 00007f7ca002f7e8 sp 00007ffdf4b50c30 error 4 in ld-2.17.so[7f7ca0024000+22000] traps: 5[32570] trap invalid opcode ip:405352 sp:7ffe9d0d6af8 error:0 in 5[400000+6000] Lustre: 18501:0:(mdd_dir.c:4741:mdd_migrate_object()) lustre-MDD0002: [0x280000403:0x1:0x0]/5 is open, migrate only dentry Lustre: 18501:0:(mdd_dir.c:4741:mdd_migrate_object()) Skipped 3 previous similar messages Lustre: 10852:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 504 < left 1100, rollback = 2 Lustre: 10852:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 38 previous similar messages Lustre: 18195:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 516 < left 618, rollback = 7 Lustre: 18195:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 19 previous similar messages Lustre: dir [0x280000404:0x501:0x0] stripe 1 readdir failed: -2, directory is partially accessed! 1[11088]: segfault at 8 ip 00007f11a305c7e8 sp 00007fffa7fa9c30 error 4 in ld-2.17.so[7f11a3051000+22000] traps: 7[12176] trap invalid opcode ip:405b8a sp:7ffd5c2090b8 error:0 in 7[400000+6000] Lustre: 18501:0:(mdd_dir.c:4741:mdd_migrate_object()) lustre-MDD0002: [0x280000403:0x1:0x0]/3 is open, migrate only dentry traps: 7[15124] trap invalid opcode ip:404aad sp:fdd7f468 error:0 traps: 7[15158] trap invalid opcode ip:404aad sp:c2091468 error:0 in 7[400000+6000] in 7[400000+6000] 15[17575]: segfault at 8 ip 00007f96016ad7e8 sp 00007ffdbece8060 error 4 in ld-2.17.so[7f96016a2000+22000] Lustre: 18501:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 501 < left 877, rollback = 2 Lustre: 18501:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 62 previous similar messages 4[16927]: segfault at 8 ip 00007f2433d9e7e8 sp 00007fff1f2f8190 error 4 in ld-2.17.so[7f2433d93000+22000] 17[20506]: segfault at 8 ip 00007f29021327e8 sp 00007ffeb1fd2af0 error 4 in ld-2.17.so[7f2902127000+22000] 3[20826]: segfault at 8 ip 00007fd8b2b327e8 sp 00007fff3b836fd0 error 4 in ld-2.17.so[7fd8b2b27000+22000] LustreError: 17794:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0001: failed to get lu_attr of [0x240000403:0x632:0x0]: rc = -2 LustreError: 19102:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a8378958: inode [0x240000403:0x632:0x0] mdc close failed: rc = -2 9[21222]: segfault at 8 ip 00007f9d207457e8 sp 00007ffc24e37ae0 error 4 in ld-2.17.so[7f9d2073a000+22000] Lustre: 19946:0:(mdt_reint.c:2460:mdt_reint_migrate()) lustre-MDT0001: [0x240000403:0x1:0x0]/9 is open, migrate only dentry Lustre: dir [0x240000404:0x709:0x0] stripe 3 readdir failed: -2, directory is partially accessed! Lustre: Skipped 1 previous similar message 14[21895]: segfault at 8 ip 00007f289f7177e8 sp 00007ffc855d3bc0 error 4 in ld-2.17.so[7f289f70c000+22000] Lustre: 10452:0:(mdt_reint.c:2460:mdt_reint_migrate()) lustre-MDT0000: [0x200000403:0x1:0x0]/13 is open, migrate only dentry LustreError: 18859:0:(mdt_reint.c:2540:mdt_reint_migrate()) lustre-MDT0002: migrate [0x240000404:0x5a2:0x0]/14 failed: rc = -2 Lustre: 10851:0:(mdd_dir.c:4741:mdd_migrate_object()) lustre-MDD0000: [0x200000403:0x2:0x0]/4 is open, migrate only dentry Lustre: 10851:0:(mdd_dir.c:4741:mdd_migrate_object()) Skipped 12 previous similar messages LustreError: 20961:0:(mdt_reint.c:2540:mdt_reint_migrate()) lustre-MDT0001: migrate [0x280000403:0x41d:0x0]/15 failed: rc = -2 LustreError: 73:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000403:0x722:0x0]: rc = -5 LustreError: 73:0:(llite_lib.c:3696:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 73:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep sleep [0x240000403:0x722:0x0] inode@0000000000000000: rc = -5 6[23690]: segfault at 8 ip 00007f1d3a7ff7e8 sp 00007ffc5d528b50 error 4 in ld-2.17.so[7f1d3a7f4000+22000] 11[23429]: segfault at 8 ip 00007f2203f497e8 sp 00007ffffeb3e200 error 4 in ld-2.17.so[7f2203f3e000+22000] LustreError: 10851:0:(mdt_reint.c:2540:mdt_reint_migrate()) lustre-MDT0000: migrate [0x200000403:0x1:0x0]/15 failed: rc = -2 5[24466]: segfault at 1c ip 00000000004017b0 sp 00007fffc95d1740 error 6 in 19[400000+6000] Lustre: 17168:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000403:0x456:0x0] with magic=0xbd60bd0 Lustre: 17168:0:(lod_lov.c:1417:lod_parse_striping()) Skipped 1 previous similar message LustreError: 10851:0:(mdt_reint.c:2540:mdt_reint_migrate()) lustre-MDT0002: migrate [0x280000403:0x1:0x0]/16 failed: rc = -2 LustreError: 26929:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x280000403:0x456:0x0]: rc = -5 LustreError: 26929:0:(lcommon_cl.c:188:cl_file_inode_init()) Skipped 1 previous similar message LustreError: 26929:0:(llite_lib.c:3696:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 26929:0:(llite_lib.c:3696:ll_prep_inode()) Skipped 1 previous similar message LustreError: 27732:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x280000403:0x456:0x0]: rc = -5 LustreError: 27732:0:(lcommon_cl.c:188:cl_file_inode_init()) Skipped 4 previous similar messages LustreError: 27732:0:(llite_lib.c:3696:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 27732:0:(llite_lib.c:3696:ll_prep_inode()) Skipped 4 previous similar messages LustreError: 25255:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8802e739ca88: cannot apply new layout on [0x280000403:0x4de:0x0] : rc = -5 LustreError: 25255:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x280000403:0x4de:0x0] error -5. LustreError: 13319:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0002: [0x280000403:0x44a:0x0] migrate mdt count mismatch 2 != 1 Lustre: 18068:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000404:0xb98:0x0] with magic=0xbd60bd0 Lustre: 18068:0:(lod_lov.c:1417:lod_parse_striping()) Skipped 1 previous similar message LustreError: 23015:0:(mdt_reint.c:2540:mdt_reint_migrate()) lustre-MDT0001: migrate [0x200000404:0xcb1:0x0]/20 failed: rc = -2 LustreError: 28470:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x280000403:0x456:0x0]: rc = -5 LustreError: 28470:0:(lcommon_cl.c:188:cl_file_inode_init()) Skipped 9 previous similar messages LustreError: 28470:0:(llite_lib.c:3696:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 28470:0:(llite_lib.c:3696:ll_prep_inode()) Skipped 9 previous similar messages Lustre: dir [0x200000404:0xcb1:0x0] stripe 3 readdir failed: -2, directory is partially accessed! LustreError: 22669:0:(mdd_dir.c:4662:mdd_migrate_cmd_check()) lustre-MDD0001: '16' migration was interrupted, run 'lfs migrate -m 0 -c 3 -H crush 16' to finish migration: rc = -1 LustreError: 29615:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a8378958: inode [0x240000404:0x7ff:0x0] mdc close failed: rc = -116 LustreError: 30092:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8802e739ca88: cannot apply new layout on [0x280000403:0x4de:0x0] : rc = -5 LustreError: 18501:0:(mdt_reint.c:2540:mdt_reint_migrate()) lustre-MDT0002: migrate [0x280000403:0x41f:0x0]/9 failed: rc = -2 LustreError: 18501:0:(mdt_reint.c:2540:mdt_reint_migrate()) Skipped 2 previous similar messages LustreError: 31831:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000404:0xdeb:0x0]: rc = -5 LustreError: 31831:0:(lcommon_cl.c:188:cl_file_inode_init()) Skipped 7 previous similar messages LustreError: 31831:0:(llite_lib.c:3696:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 31831:0:(llite_lib.c:3696:ll_prep_inode()) Skipped 7 previous similar messages Lustre: 22763:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 3/12/3, destroy: 1/4/0 Lustre: 22763:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 595 previous similar messages Lustre: 22763:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 2/2/0, xattr_set: 28/2407/0 Lustre: 22763:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 595 previous similar messages Lustre: 22763:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 5/55/0, punch: 0/0/0, quota 7/129/0 Lustre: 22763:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 595 previous similar messages Lustre: 22763:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 14/287/3, delete: 2/5/0 Lustre: 22763:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 595 previous similar messages Lustre: 22763:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 7/7/0, ref_del: 1/1/0 Lustre: 22763:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 595 previous similar messages 3[31488]: segfault at 8 ip 00007f1a0d3b27e8 sp 00007ffec590bea0 error 4 in ld-2.17.so[7f1a0d3a7000+22000] LustreError: 17794:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000403:0x500:0x0]: rc = -2 LustreError: 26720:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a8378958: inode [0x200000403:0x500:0x0] mdc close failed: rc = -2 2[32402]: segfault at 406000 ip 0000000000406000 sp 00007ffcdb319de8 error 14 in 2[606000+1000] LustreError: 22763:0:(mdd_dir.c:4662:mdd_migrate_cmd_check()) lustre-MDD0000: '19' migration was interrupted, run 'lfs migrate -m 1 -c 3 -H crush 19' to finish migration: rc = -1 LustreError: 18197:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000403:0xe17:0x0]: rc = -2 LustreError: 31023:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e739ca88: inode [0x200000403:0xe17:0x0] mdc close failed: rc = -2 Lustre: 10848:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 499 < left 1678, rollback = 2 Lustre: 10848:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 752 previous similar messages LustreError: 19721:0:(mdt_reint.c:2540:mdt_reint_migrate()) lustre-MDT0000: migrate [0x200000403:0x2:0x0]/1 failed: rc = -2 LustreError: 19721:0:(mdt_reint.c:2540:mdt_reint_migrate()) Skipped 6 previous similar messages 16[2640]: segfault at 8 ip 00007f09a18347e8 sp 00007ffe67947470 error 4 in ld-2.17.so[7f09a1829000+22000] LustreError: 3528:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e739ca88: inode [0x200000403:0xeee:0x0] mdc close failed: rc = -13 LustreError: 3528:0:(file.c:247:ll_close_inode_openhandle()) Skipped 1 previous similar message LustreError: 4419:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x280000403:0x519:0x0]: rc = -5 LustreError: 4419:0:(llite_lib.c:3696:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 18501:0:(mdd_dir.c:4662:mdd_migrate_cmd_check()) lustre-MDD0001: '5' migration was interrupted, run 'lfs migrate -m 2 -c 3 -H crush 5' to finish migration: rc = -1 LustreError: 3502:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8802e739ca88: cannot apply new layout on [0x240000403:0x9d9:0x0] : rc = -5 LustreError: 3502:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x240000403:0x9d9:0x0] error -5. 2[4469]: segfault at 0 ip 0000000000403cf0 sp 00007ffe3b18ea28 error 6 in 2[400000+6000] Lustre: 10852:0:(mdd_dir.c:4741:mdd_migrate_object()) lustre-MDD0000: [0x200000403:0x1:0x0]/19 is open, migrate only dentry Lustre: 10852:0:(mdd_dir.c:4741:mdd_migrate_object()) Skipped 19 previous similar messages Lustre: dir [0x280000403:0x4ca:0x0] stripe 3 readdir failed: -2, directory is partially accessed! LustreError: 3513:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8802e739ca88: cannot apply new layout on [0x240000403:0x9d9:0x0] : rc = -5 LustreError: 3513:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x240000403:0x9d9:0x0] error -5. LustreError: 10835:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0001: failed to get lu_attr of [0x240000404:0x658:0x0]: rc = -2 LustreError: 10835:0:(mdd_object.c:3901:mdd_close()) Skipped 1 previous similar message Lustre: 17027:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0xd31:0x0] with magic=0xbd60bd0 Lustre: 17027:0:(lod_lov.c:1417:lod_parse_striping()) Skipped 1 previous similar message LustreError: 18197:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0001: failed to get lu_attr of [0x240000404:0x658:0x0]: rc = -2 LustreError: 18197:0:(mdd_object.c:3901:mdd_close()) Skipped 1 previous similar message LustreError: 3744:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a8378958: inode [0x240000404:0x658:0x0] mdc close failed: rc = -2 LustreError: 3744:0:(file.c:247:ll_close_inode_openhandle()) Skipped 2 previous similar messages Lustre: 12470:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 618, rollback = 7 Lustre: 12470:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 65 previous similar messages Lustre: dir [0x240000404:0x88a:0x0] stripe 3 readdir failed: -2, directory is partially accessed! LustreError: 26705:0:(mdd_dir.c:4662:mdd_migrate_cmd_check()) lustre-MDD0001: '17' migration was interrupted, run 'lfs migrate -m 2 -c 3 -H crush 17' to finish migration: rc = -1 LustreError: 26705:0:(mdt_reint.c:2540:mdt_reint_migrate()) lustre-MDT0001: migrate [0x240000403:0x1:0x0]/17 failed: rc = -1 LustreError: 26705:0:(mdt_reint.c:2540:mdt_reint_migrate()) Skipped 7 previous similar messages LustreError: 23015:0:(mdd_dir.c:4662:mdd_migrate_cmd_check()) lustre-MDD0001: '15' migration was interrupted, run 'lfs migrate -m 2 -c 3 -H crush 15' to finish migration: rc = -1 LustreError: 11216:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x280000403:0x749:0x0]: rc = -5 LustreError: 11216:0:(lcommon_cl.c:188:cl_file_inode_init()) Skipped 8 previous similar messages LustreError: 11216:0:(llite_lib.c:3696:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 11216:0:(llite_lib.c:3696:ll_prep_inode()) Skipped 8 previous similar messages LustreError: 17166:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0001: [0x240000403:0xaa7:0x0] migrate mdt count mismatch 3 != 1 LustreError: 18324:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0001: failed to get lu_attr of [0x240000404:0x709:0x0]: rc = -2 LustreError: 18324:0:(mdd_object.c:3901:mdd_close()) Skipped 1 previous similar message LustreError: 12666:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a8378958: inode [0x240000404:0x709:0x0] mdc close failed: rc = -2 LustreError: 12666:0:(file.c:247:ll_close_inode_openhandle()) Skipped 1 previous similar message Lustre: 30595:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000403:0xbad:0x0] with magic=0xbd60bd0 Lustre: 30595:0:(lod_lov.c:1417:lod_parse_striping()) Skipped 3 previous similar messages LustreError: 10848:0:(mdd_dir.c:4662:mdd_migrate_cmd_check()) lustre-MDD0000: '17' migration was interrupted, run 'lfs migrate -m 2 -c 3 -H crush 17' to finish migration: rc = -1 LustreError: 10848:0:(mdd_dir.c:4662:mdd_migrate_cmd_check()) Skipped 1 previous similar message Lustre: dir [0x240000403:0xb7d:0x0] stripe 2 readdir failed: -2, directory is partially accessed! Lustre: Skipped 8 previous similar messages 8[17381]: segfault at 8 ip 00007f5a325187e8 sp 00007ffd1f4f98d0 error 4 in ld-2.17.so[7f5a3250d000+22000] 12[19151]: segfault at 8 ip 00007f5fa5c0c7e8 sp 00007ffda5b27e30 error 4 in ld-2.17.so[7f5fa5c01000+22000] LustreError: 14204:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8802e739ca88: cannot apply new layout on [0x280000403:0x749:0x0] : rc = -5 LustreError: 14204:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x280000403:0x749:0x0] error -5. LustreError: 10852:0:(lustre_lmv.h:500:lmv_is_sane()) unknown layout LMV: magic=0xcd40cd0 count=3 index=2 hash=crush:0x82000003 version=1 migrate_offset=2 migrate_hash=fnv_1a_64:2 pool= LustreError: 19244:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x240000404:0xaa2:0x0] error -5. LustreError: 18063:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0000: [0x200000404:0x1284:0x0] migrate mdt count mismatch 3 != 2 LustreError: 10848:0:(mdd_dir.c:4662:mdd_migrate_cmd_check()) lustre-MDD0001: '6' migration was interrupted, run 'lfs migrate -m 2 -c 1 -H crush 6' to finish migration: rc = -1 LustreError: 10848:0:(mdd_dir.c:4662:mdd_migrate_cmd_check()) Skipped 2 previous similar messages LustreError: 20664:0:(mdt_reint.c:2540:mdt_reint_migrate()) lustre-MDT0001: migrate [0x240000403:0xc2c:0x0]/10 failed: rc = -2 LustreError: 20664:0:(mdt_reint.c:2540:mdt_reint_migrate()) Skipped 20 previous similar messages LustreError: 18063:0:(mdt_open.c:1302:mdt_cross_open()) lustre-MDT0000: [0x200000404:0x12c9:0x0] doesn't exist!: rc = -14 LustreError: 25413:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000403:0x11e2:0x0]: rc = -5 LustreError: 25413:0:(lcommon_cl.c:188:cl_file_inode_init()) Skipped 29 previous similar messages LustreError: 25413:0:(llite_lib.c:3696:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 25413:0:(llite_lib.c:3696:ll_prep_inode()) Skipped 29 previous similar messages LustreError: 17184:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0002: [0x280000403:0x9b8:0x0] migrate mdt count mismatch 2 != 1 Lustre: 19946:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 512 < left 788, rollback = 2 Lustre: 19946:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 1502 previous similar messages LustreError: 26244:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a8378958: inode [0x200000404:0x13b0:0x0] mdc close failed: rc = -13 LustreError: 26244:0:(file.c:247:ll_close_inode_openhandle()) Skipped 1 previous similar message Lustre: 18058:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0x1253:0x0] with magic=0xbd60bd0 Lustre: 18058:0:(lod_lov.c:1417:lod_parse_striping()) Skipped 27 previous similar messages 2[28267]: segfault at 0 ip (null) sp 00007fffa9b7df68 error 14 in 2[400000+6000] 16[29930]: segfault at 8 ip 00007f091ae1a7e8 sp 00007ffdd6aca560 error 4 in ld-2.17.so[7f091ae0f000+22000] LustreError: 18324:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0001: failed to get lu_attr of [0x240000404:0xac3:0x0]: rc = -2 LustreError: 18324:0:(mdd_object.c:3901:mdd_close()) Skipped 1 previous similar message 19[32093]: segfault at 8 ip 00007f63fd5f77e8 sp 00007ffe3e7cafb0 error 4 in ld-2.17.so[7f63fd5ec000+22000] Lustre: 22676:0:(mdd_dir.c:4741:mdd_migrate_object()) lustre-MDD0000: [0x200000403:0x1:0x0]/7 is open, migrate only dentry Lustre: 22676:0:(mdd_dir.c:4741:mdd_migrate_object()) Skipped 35 previous similar messages 2[32488]: segfault at 0 ip (null) sp 00007ffd7d7ba378 error 14 in 2[400000+6000] LustreError: 31577:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8800a8378958: cannot apply new layout on [0x240000404:0xaa2:0x0] : rc = -5 LustreError: 31577:0:(lov_object.c:1341:lov_layout_change()) Skipped 1 previous similar message LustreError: 26705:0:(mdd_dir.c:4662:mdd_migrate_cmd_check()) lustre-MDD0002: '12' migration was interrupted, run 'lfs migrate -m 0 -c 2 -H crush 12' to finish migration: rc = -1 LustreError: 26705:0:(mdd_dir.c:4662:mdd_migrate_cmd_check()) Skipped 3 previous similar messages 0[11096]: segfault at 8 ip 00007ff34cb847e8 sp 00007ffe71c38eb0 error 4 in ld-2.17.so[7ff34cb79000+22000] 2[13486]: segfault at 8 ip 00007fa228afe7e8 sp 00007ffd797b03b0 error 4 in ld-2.17.so[7fa228af3000+22000] 15[15943]: segfault at 0 ip 0000000000403e5f sp 00007fff31f55e40 error 6 in 15[400000+6000] 8[19828]: segfault at 8 ip 00007ff3c8c3d7e8 sp 00007ffc59da90b0 error 4 in ld-2.17.so[7ff3c8c32000+22000] traps: 14[17426] general protection ip:40576f sp:7fff28b4f0e0 error:0 in 14[400000+6000] Lustre: 12470:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 12470:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 2564 previous similar messages Lustre: 12470:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 12470:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 2564 previous similar messages Lustre: 12470:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 12470:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 2564 previous similar messages Lustre: 12470:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 12470:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 2564 previous similar messages Lustre: 12470:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 12470:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 2564 previous similar messages traps: 15[2561] general protection ip:40582f sp:7ffe5f2f8a28 error:0 in 15[400000+6000] traps: 6[8675] trap invalid opcode ip:40537c sp:7ffd400af898 error:0 in 6[400000+6000] Lustre: 6914:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 618, rollback = 7 Lustre: 6914:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 149 previous similar messages 4[15568]: segfault at 8 ip 00007f16803d17e8 sp 00007ffd65e27f50 error 4 in ld-2.17.so[7f16803c6000+22000] 15[15720]: segfault at 8 ip 00007fca73e637e8 sp 00007fff42e334b0 error 4 in ld-2.17.so[7fca73e58000+22000] 18[16869]: segfault at 0 ip 0000000000403cf0 sp 00007fffbd47a888 error 6 in 18[400000+6000] 10[23126]: segfault at 8 ip 00007f0977b777e8 sp 00007fff26baf810 error 4 in ld-2.17.so[7f0977b6c000+22000] 18[24745]: segfault at 8 ip 00007faa40e497e8 sp 00007ffe0f290be0 error 4 in ld-2.17.so[7faa40e3e000+22000] 5[24897]: segfault at 0 ip (null) sp 00007fffd7ca9158 error 14 in 5[400000+6000] 8[29762]: segfault at 8 ip 00007f16116457e8 sp 00007ffebd4da8b0 error 4 in ld-2.17.so[7f161163a000+22000] LustreError: 23190:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e739ca88: inode [0x240000404:0x12b1:0x0] mdc close failed: rc = -13 LustreError: 23190:0:(file.c:247:ll_close_inode_openhandle()) Skipped 6 previous similar messages Lustre: mdt00_016: service thread pid 18090 was inactive for 40.113 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 18090, comm: mdt00_016 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock+0x88/0x1c0 [mdt] [<0>] mdt_reint_setattr+0x1324/0x15f0 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x73c/0xbb0 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x257/0xcd0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 17027, comm: mdt00_029 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock_try+0xa0/0x250 [mdt] [<0>] mdt_getattr_name_lock+0x1729/0x2cc0 [mdt] [<0>] mdt_intent_getattr+0x2cc/0x4e0 [mdt] [<0>] mdt_intent_opc.constprop.74+0x211/0xc60 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x397/0x980 [ptlrpc] [<0>] ldlm_handle_enqueue+0x547/0x1900 [ptlrpc] [<0>] tgt_enqueue+0x68/0x240 [ptlrpc] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x257/0xcd0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: lustre-OST0000-osc-ffff8802e739ca88: disconnect after 20s idle Lustre: mdt_io00_011: service thread pid 22676 was inactive for 72.206 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Lustre: Skipped 1 previous similar message Pid: 22676, comm: mdt_io00_011 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock+0x88/0x1c0 [mdt] [<0>] mdt_rename_source_lock+0x57/0xf0 [mdt] [<0>] mdt_reint_migrate+0x1832/0x24b0 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x73c/0xbb0 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x257/0xcd0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: mdt_io00_000: service thread pid 10848 was inactive for 72.173 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: mdt00_036: service thread pid 18546 was inactive for 72.066 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: mdt00_022: service thread pid 32027 was inactive for 72.166 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 7 previous similar messages Lustre: mdt_io00_005: service thread pid 20664 was inactive for 72.234 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 13 previous similar messages Lustre: lustre-OST0001-osc-ffff8800a8378958: disconnect after 22s idle LustreError: 10629:0:(ldlm_lockd.c:257:expired_lock_main()) ### lock callback timer expired after 101s: evicting client at 0@lo ns: mdt-lustre-MDT0002_UUID lock: ffff8802cc115e00/0x89c6027d5ed9f30f lrc: 3/0,0 mode: PW/PW res: [0x280000404:0x161b:0x0].0x0 bits 0x4/0x0 rrc: 8 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0x89c6027d5ed9f301 expref: 752 pid: 16990 timeout: 930 lvb_type: 0 Lustre: 22676:0:(mdd_dir.c:4741:mdd_migrate_object()) lustre-MDD0002: [0x280000403:0x1:0x0]/17 is open, migrate only dentry Lustre: 22676:0:(mdd_dir.c:4741:mdd_migrate_object()) Skipped 6 previous similar messages LustreError: 17018:0:(ldlm_lockd.c:1447:ldlm_handle_enqueue()) ### lock on destroyed export ffff8800a794d3d8 ns: mdt-lustre-MDT0002_UUID lock: ffff8802cea46d00/0x89c6027d5edc4b09 lrc: 3/0,0 mode: PR/PR res: [0x280000403:0x1:0x0].0x0 bits 0x12/0x0 rrc: 23 type: IBT gid 0 flags: 0x50200000000000 nid: 0@lo remote: 0x89c6027d5edc4ad8 expref: 232 pid: 17018 timeout: 0 lvb_type: 0 LustreError: 22676:0:(client.c:1282:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802c322cb40 x1829451678399360/t0(0) o105->lustre-MDT0002@0@lo:15/16 lens 336/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' uid:4294967295 gid:4294967295 Lustre: mdt00_031: service thread pid 20979 completed after 89.657s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_040: service thread pid 29873 completed after 89.751s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_038: service thread pid 18563 completed after 90.106s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: lustre-MDT0002-mdc-ffff8800a8378958: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-MDT0002-mdc-ffff8800a8378958: Connection to lustre-MDT0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: mdt00_012: service thread pid 18058 completed after 89.644s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_000: service thread pid 10830 completed after 90.144s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_039: service thread pid 29862 completed after 89.671s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_025: service thread pid 9562 completed after 90.126s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: lustre-MDT0002-mdc-ffff8800a8378958: This client was evicted by lustre-MDT0002; in progress operations using this service will fail. Lustre: mdt00_022: service thread pid 32027 completed after 90.123s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_005: service thread pid 15348 completed after 89.668s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_001: service thread pid 10831 completed after 90.128s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_028: service thread pid 17018 completed after 89.630s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_004: service thread pid 13319 completed after 89.694s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_006: service thread pid 17166 completed after 89.718s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 6984:0:(file.c:6136:ll_inode_revalidate_fini()) lustre: revalidate FID [0x280000403:0x1:0x0] error: rc = -5 Lustre: mdt00_023: service thread pid 32068 completed after 90.148s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_029: service thread pid 17027 completed after 100.347s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_011: service thread pid 22676 completed after 100.433s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_011: service thread pid 18055 completed after 90.225s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 10053:0:(mdc_request.c:1464:mdc_read_page()) lustre-MDT0002-mdc-ffff8800a8378958: [0x280000403:0x1:0x0] lock enqueue fails: rc = -5 LustreError: 10016:0:(statahead.c:1801:is_first_dirent()) lustre: reading dir [0x280000403:0x1:0x0] at 0 stat_pid = 6105 : rc = -5 LustreError: 10053:0:(mdc_request.c:1464:mdc_read_page()) Skipped 3 previous similar messages Lustre: mdt00_016: service thread pid 18090 completed after 100.567s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: dir [0x240000403:0x16de:0x0] stripe 1 readdir failed: -108, directory is partially accessed! Lustre: Skipped 2 previous similar messages LustreError: 17406:0:(ldlm_resource.c:982:ldlm_resource_complain()) lustre-MDT0002-mdc-ffff8800a8378958: namespace resource [0x280000403:0x1:0x0].0x0 (ffff88009b0f2340) refcount nonzero (1) after lock cleanup; forcing cleanup. Lustre: lustre-MDT0002-mdc-ffff8800a8378958: Connection restored to 192.168.123.22@tcp (at 0@lo) 18[18234]: segfault at 1c ip 00000000004017b0 sp 00007ffdd3d9fa80 error 6 in 18[400000+6000] LustreError: 10629:0:(ldlm_lockd.c:257:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0000_UUID lock: ffff880098c87840/0x89c6027d5edbfb2a lrc: 3/0,0 mode: PR/PR res: [0x200000404:0x243c:0x0].0x0 bits 0x1b/0x0 rrc: 7 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0x89c6027d5edbfb0e expref: 1199 pid: 17184 timeout: 940 lvb_type: 0 Lustre: mdt00_021: service thread pid 2469 completed after 100.282s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_036: service thread pid 18546 completed after 100.089s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_014: service thread pid 18066 completed after 98.657s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_000: service thread pid 10848 completed after 100.730s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: lustre-MDT0000-mdc-ffff8800a8378958: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 6 previous similar messages Lustre: lustre-MDT0000-mdc-ffff8800a8378958: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 17363:0:(ldlm_lockd.c:2550:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1744701844 with bad export cookie 9927625166062191770 Lustre: mdt_io00_013: service thread pid 22763 completed after 99.892s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_005: service thread pid 20664 completed after 97.238s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: lustre-MDT0000-mdc-ffff8800a8378958: This client was evicted by lustre-MDT0000; in progress operations using this service will fail. Lustre: mdt_io00_001: service thread pid 10851 completed after 97.264s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 16036:0:(file.c:6136:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000007:0x1:0x0] error: rc = -108 Lustre: mdt_io00_014: service thread pid 23015 completed after 97.204s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 16036:0:(file.c:6136:ll_inode_revalidate_fini()) Skipped 62 previous similar messages Lustre: mdt_io00_007: service thread pid 19721 completed after 97.258s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_010: service thread pid 22669 completed after 97.316s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_017: service thread pid 26705 completed after 97.317s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 7295:0:(mdc_request.c:1464:mdc_read_page()) lustre-MDT0000-mdc-ffff8800a8378958: [0x200000403:0x1:0x0] lock enqueue fails: rc = -108 LustreError: 7295:0:(mdc_request.c:1464:mdc_read_page()) Skipped 12 previous similar messages Lustre: mdt_io00_003: service thread pid 18501 completed after 96.977s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: lustre-MDT0000-mdc-ffff8800a8378958: Connection restored to 192.168.123.22@tcp (at 0@lo) Lustre: mdt_io00_009: service thread pid 20961 completed after 96.884s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: 19946:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 508 < left 1339, rollback = 2 Lustre: 19946:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 641 previous similar messages hrtimer: interrupt took 61122440 ns LustreError: 19946:0:(mdt_reint.c:2540:mdt_reint_migrate()) lustre-MDT0001: migrate [0x240000404:0x1688:0x0]/10 failed: rc = -2 LustreError: 19946:0:(mdt_reint.c:2540:mdt_reint_migrate()) Skipped 13 previous similar messages 1[20565]: segfault at 8 ip 00007f8d4be3a7e8 sp 00007fff009f97c0 error 4 in ld-2.17.so[7f8d4be2f000+22000] LustreError: 10452:0:(mdd_dir.c:4662:mdd_migrate_cmd_check()) lustre-MDD0000: '5' migration was interrupted, run 'lfs migrate -m 2 -c 2 -H crush 5' to finish migration: rc = -1 LustreError: 10452:0:(mdd_dir.c:4662:mdd_migrate_cmd_check()) Skipped 1 previous similar message 6[25620]: segfault at 8 ip 00007f43ae3557e8 sp 00007ffdae09d350 error 4 in ld-2.17.so[7f43ae34a000+22000] LustreError: 17147:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0002: failed to get lu_attr of [0x280000405:0x8f:0x0]: rc = -2 LustreError: 17147:0:(mdd_object.c:3901:mdd_close()) Skipped 3 previous similar messages 7[27876]: segfault at 8 ip 00007fbb023457e8 sp 00007ffc19835870 error 4 in ld-2.17.so[7fbb0233a000+22000] 5[31842]: segfault at 8 ip 00007f1a659357e8 sp 00007fffb71fa250 error 4 in ld-2.17.so[7f1a6592a000+22000] 2[3098]: segfault at 8 ip 00007fd0584457e8 sp 00007ffc16623b70 error 4 in ld-2.17.so[7fd05843a000+22000] LustreError: 3481:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000405:0x236:0x0]: rc = -5 LustreError: 3481:0:(lcommon_cl.c:188:cl_file_inode_init()) Skipped 31 previous similar messages LustreError: 3481:0:(llite_lib.c:3696:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 3481:0:(llite_lib.c:3696:ll_prep_inode()) Skipped 31 previous similar messages 6[7005]: segfault at 0 ip 0000000000403cf0 sp 00007ffcfcc463e8 error 6 in 6[400000+6000] Lustre: 18058:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000403:0x1998:0x0] with magic=0xbd60bd0 Lustre: 18058:0:(lod_lov.c:1417:lod_parse_striping()) Skipped 125 previous similar messages 7[7128]: segfault at 0 ip (null) sp 00007ffe9b802178 error 14 in 7[400000+6000] 12[7565]: segfault at 0 ip 0000000000401d60 sp 00007ffe6dbd0b88 error 6 in 12[400000+6000] 16[10365]: segfault at 10 ip 00007fd95d685b43 sp 00007ffe196b7aa0 error 4 in ld-2.17.so[7fd95d67a000+22000] Lustre: 20961:0:(mdt_reint.c:2460:mdt_reint_migrate()) lustre-MDT0001: [0x240000403:0x1:0x0]/13 is open, migrate only dentry 19[15087]: segfault at 0 ip (null) sp 00007ffea8e97ca8 error 14 in 5[400000+6000] LustreError: 10852:0:(lustre_lmv.h:500:lmv_is_sane()) unknown layout LMV: magic=0xcd40cd0 count=4 index=3 hash=crush:0x82000003 version=1 migrate_offset=3 migrate_hash=fnv_1a_64:2 pool= LustreError: 16990:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0002: [0x280000405:0x3d3:0x0] migrate mdt count mismatch 3 != 1 LustreError: 30770:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0001: failed to get lu_attr of [0x240000404:0x1800:0x0]: rc = -2 LustreError: 30770:0:(mdd_object.c:3901:mdd_close()) Skipped 3 previous similar messages LustreError: 18101:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e739ca88: inode [0x240000404:0x1800:0x0] mdc close failed: rc = -2 LustreError: 18101:0:(file.c:247:ll_close_inode_openhandle()) Skipped 46 previous similar messages 2[22232]: segfault at 8 ip 00007f819da217e8 sp 00007ffc99f13f40 error 4 in ld-2.17.so[7f819da16000+22000] LustreError: 30611:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0002: [0x280000405:0x466:0x0] migrate mdt count mismatch 1 != 3 17[26923]: segfault at 0 ip 0000000000403e5f sp 00007ffe2f2588b0 error 6 in 17[400000+6000] Lustre: dir [0x200000404:0x2bd5:0x0] stripe 3 readdir failed: -2, directory is partially accessed! Lustre: Skipped 22 previous similar messages LustreError: 10629:0:(ldlm_lockd.c:257:expired_lock_main()) ### lock callback timer expired after 101s: evicting client at 0@lo ns: mdt-lustre-MDT0002_UUID lock: ffff8802be9e3880/0x89c6027d5eea8015 lrc: 3/0,0 mode: PW/PW res: [0x280000405:0x20f:0x0].0x0 bits 0x4/0x0 rrc: 8 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0x89c6027d5eea7ff2 expref: 278 pid: 18099 timeout: 1090 lvb_type: 0 LustreError: lustre-MDT0002-mdc-ffff8800a8378958: operation mds_getxattr to node 0@lo failed: rc = -107 Lustre: lustre-MDT0002-mdc-ffff8800a8378958: Connection to lustre-MDT0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: Skipped 1 previous similar message LustreError: lustre-MDT0002-mdc-ffff8800a8378958: This client was evicted by lustre-MDT0002; in progress operations using this service will fail. LustreError: 31798:0:(mdc_request.c:1464:mdc_read_page()) lustre-MDT0002-mdc-ffff8800a8378958: [0x280000401:0x62:0x0] lock enqueue fails: rc = -108 LustreError: 31798:0:(mdc_request.c:1464:mdc_read_page()) Skipped 1 previous similar message Lustre: lustre-MDT0002-mdc-ffff8800a8378958: Connection restored to 192.168.123.22@tcp (at 0@lo) LustreError: 30098:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8802e739ca88: cannot apply new layout on [0x280000405:0x496:0x0] : rc = -5 LustreError: 30098:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x280000405:0x496:0x0] error -5. LustreError: 10852:0:(mdd_dir.c:4662:mdd_migrate_cmd_check()) lustre-MDD0002: '8' migration was interrupted, run 'lfs migrate -m 2 -c 3 -H crush 8' to finish migration: rc = -1 LustreError: 10852:0:(mdd_dir.c:4662:mdd_migrate_cmd_check()) Skipped 10 previous similar messages 18[770]: segfault at 8 ip 00007f44c2b877e8 sp 00007ffd2892b8e0 error 4 in ld-2.17.so[7f44c2b7c000+22000] 17[882]: segfault at 8 ip 00007f4024f0a7e8 sp 00007fffe2bc24f0 error 4 in ld-2.17.so[7f4024eff000+22000] LustreError: 17903:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0001: [0x240000403:0x1892:0x0] migrate mdt count mismatch 3 != 1 LustreError: 419:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 20 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 LustreError: 2314:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x200000405:0xa84:0x0] error -5. 15[5949]: segfault at 8 ip 00007f66f2e1e7e8 sp 00007ffe8387f7f0 error 4 in ld-2.17.so[7f66f2e13000+22000] 14[7580]: segfault at 8 ip 00007f572719e7e8 sp 00007ffeb42b2570 error 4 in ld-2.17.so[7f5727193000+22000] 5[7987]: segfault at 8 ip 00007f0f8ffca7e8 sp 00007ffe8a1203f0 error 4 in ld-2.17.so[7f0f8ffbf000+22000] LustreError: 9499:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8802e739ca88: cannot apply new layout on [0x240000403:0x1f2a:0x0] : rc = -5 LustreError: 9499:0:(lov_object.c:1341:lov_layout_change()) Skipped 2 previous similar messages 9[12332]: segfault at 8 ip 00007f2c3f6237e8 sp 00007ffd05f68e10 error 4 in ld-2.17.so[7f2c3f618000+22000] 12[12346]: segfault at 8 ip 00007fe2356687e8 sp 00007ffe06608a50 error 4 in ld-2.17.so[7fe23565d000+22000] LustreError: 18005:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0001: [0x240000404:0x1f88:0x0] migrate mdt count mismatch 1 != 3 LustreError: 18005:0:(mdt_xattr.c:402:mdt_dir_layout_update()) Skipped 1 previous similar message 5[15197]: segfault at 8 ip 00007fefb66cb7e8 sp 00007ffd00880e50 error 4 in ld-2.17.so[7fefb66c0000+22000] 14[17493]: segfault at 8 ip 00007f5876c037e8 sp 00007ffd1d034790 error 4 in ld-2.17.so[7f5876bf8000+22000] 10[18013]: segfault at 8 ip 00007f88c24657e8 sp 00007ffcef0f8750 error 4 in ld-2.17.so[7f88c245a000+22000] LustreError: 10851:0:(lustre_lmv.h:500:lmv_is_sane()) unknown layout LMV: magic=0xcd40cd0 count=2 index=1 hash=crush:0x82000003 version=1 migrate_offset=1 migrate_hash=fnv_1a_64:2 pool= Lustre: 18501:0:(mdt_reint.c:2460:mdt_reint_migrate()) lustre-MDT0002: [0x280000403:0x1:0x0]/0 is open, migrate only dentry 5[19605]: segfault at 8 ip 00007f649e1187e8 sp 00007ffc7855e8b0 error 4 in ld-2.17.so[7f649e10d000+22000] Lustre: 6914:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 6914:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 3628 previous similar messages Lustre: 6914:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 6914:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 3628 previous similar messages Lustre: 6914:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 6914:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 3628 previous similar messages Lustre: 6914:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 6914:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 3628 previous similar messages Lustre: 6914:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 6914:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 3628 previous similar messages 4[24909]: segfault at 8 ip 00007f143fc337e8 sp 00007fff1ca54cf0 error 4 in ld-2.17.so[7f143fc28000+22000] LustreError: 20750:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x240000403:0x221f:0x0] error -5. LustreError: 20750:0:(vvp_io.c:1903:vvp_io_init()) Skipped 1 previous similar message 8[26850]: segfault at 8 ip 00007efc9d57e7e8 sp 00007ffd5f8d2230 error 4 in ld-2.17.so[7efc9d573000+22000] LustreError: 18378:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0001: failed to get lu_attr of [0x240000404:0x217c:0x0]: rc = -2 LustreError: 18378:0:(mdd_object.c:3901:mdd_close()) Skipped 10 previous similar messages 14[29241]: segfault at 8 ip 00007fa5d43057e8 sp 00007ffd29897460 error 4 in ld-2.17.so[7fa5d42fa000+22000] LustreError: 20664:0:(mdt_reint.c:2540:mdt_reint_migrate()) lustre-MDT0000: migrate [0x200000404:0x3327:0x0]/13 failed: rc = -2 LustreError: 20664:0:(mdt_reint.c:2540:mdt_reint_migrate()) Skipped 77 previous similar messages 7[31289]: segfault at 8 ip 00007f877a1657e8 sp 00007fff3f3d58d0 error 4 in ld-2.17.so[7f877a15a000+22000] LustreError: 131:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 2 [0x280000404:0x20f3:0x0] inode@0000000000000000: rc = -5 14[32089]: segfault at 8 ip 00007f03b4af67e8 sp 00007ffd5e274050 error 4 in ld-2.17.so[7f03b4aeb000+22000] 14[32071]: segfault at 8 ip 00007fd8eeecb7e8 sp 00007ffc74c18010 error 4 in ld-2.17.so[7fd8eeec0000+22000] LustreError: 29268:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8802e739ca88: cannot apply new layout on [0x200000405:0x1001:0x0] : rc = -5 LustreError: 29268:0:(lov_object.c:1341:lov_layout_change()) Skipped 1 previous similar message LustreError: 9562:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0000: [0x200000405:0xfbe:0x0] migrate mdt count mismatch 3 != 1 LustreError: 9562:0:(mdt_xattr.c:402:mdt_dir_layout_update()) Skipped 1 previous similar message 16[1840]: segfault at 8 ip 00007f12d244a7e8 sp 00007ffd7dfbced0 error 4 in ld-2.17.so[7f12d243f000+22000] LustreError: 10851:0:(lustre_lmv.h:500:lmv_is_sane()) unknown layout LMV: magic=0xcd40cd0 count=4 index=3 hash=crush:0x82000003 version=1 migrate_offset=3 migrate_hash=fnv_1a_64:2 pool= 4[1372]: segfault at 8 ip 00007f245075e7e8 sp 00007ffded4fa070 error 4 in ld-2.17.so[7f2450753000+22000] 1[11399]: segfault at 1c ip 00000000004017b0 sp 00007ffe337a2a20 error 6 in 1[400000+6000] 14[8578]: segfault at 0 ip (null) sp 00007ffebfbf16c8 error 14 in 14 (deleted)[400000+6000] 19[12552]: segfault at 8 ip 00007fcf97e967e8 sp 00007ffdbb8ff5a0 error 4 in ld-2.17.so[7fcf97e8b000+22000] LustreError: 13991:0:(lcommon_cl.c:188:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000405:0x1092:0x0]: rc = -5 LustreError: 13991:0:(lcommon_cl.c:188:cl_file_inode_init()) Skipped 115 previous similar messages LustreError: 13991:0:(llite_lib.c:3696:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 13991:0:(llite_lib.c:3696:ll_prep_inode()) Skipped 115 previous similar messages Lustre: 21316:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0000: opcode 7: before 516 < left 618, rollback = 7 Lustre: 21316:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 343 previous similar messages 15[21880]: segfault at 8 ip 00007fb64bd337e8 sp 00007fff9f9b12e0 error 4 in ld-2.17.so[7fb64bd28000+22000] 16[23082]: segfault at 8 ip 00007fee14d8b7e8 sp 00007fff6334b120 error 4 in ld-2.17.so[7fee14d80000+22000] 6[24462]: segfault at 8 ip 00007f8c08e277e8 sp 00007ffc2c50e500 error 4 in ld-2.17.so[7f8c08e1c000+22000] 10[24874]: segfault at 1c ip 00000000004017b0 sp 00007ffefac31540 error 6 in 10[400000+6000] 1[398]: segfault at 8 ip 00007f6f45cc57e8 sp 00007fff6ad53760 error 4 in ld-2.17.so[7f6f45cba000+22000] 10[32061]: segfault at 8 ip 00007f2aa23f27e8 sp 00007ffcb70c9840 error 4 in ld-2.17.so[7f2aa23e7000+22000] 13[596]: segfault at 8 ip 00007f715f4997e8 sp 00007fff5f5f9be0 error 4 in ld-2.17.so[7f715f48e000+22000] 15[3407]: segfault at 8 ip 00007fbc16cf77e8 sp 00007fff9f717ec0 error 4 in ld-2.17.so[7fbc16cec000+22000] Lustre: 31070:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000405:0x1619:0x0] with magic=0xbd60bd0 Lustre: 31070:0:(lod_lov.c:1417:lod_parse_striping()) Skipped 101 previous similar messages 16[7658]: segfault at 1c ip 00000000004017b0 sp 00007ffcba58bf90 error 6 in 16[400000+6000] Lustre: dir [0x280000406:0x9dc:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 22 previous similar messages 1[12768]: segfault at 8 ip 00007f443a5a97e8 sp 00007ffd151a4450 error 4 in ld-2.17.so[7f443a59e000+22000] LustreError: 31145:0:(mdt_open.c:1302:mdt_cross_open()) lustre-MDT0000: [0x200000404:0x39fb:0x0] doesn't exist!: rc = -14 0[14536]: segfault at 0 ip 0000000000403e5f sp 00007ffc5aa09300 error 6 in 0[400000+6000] LustreError: 14552:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8802e739ca88: cannot apply new layout on [0x240000403:0x2974:0x0] : rc = -5 LustreError: 14552:0:(lov_object.c:1341:lov_layout_change()) Skipped 8 previous similar messages LustreError: 14552:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x240000403:0x2974:0x0] error -5. LustreError: 14552:0:(vvp_io.c:1903:vvp_io_init()) Skipped 2 previous similar messages LustreError: 31109:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0002: [0x280000406:0xbd0:0x0] migrate mdt count mismatch 2 != 1 LustreError: 31109:0:(mdt_xattr.c:402:mdt_dir_layout_update()) Skipped 2 previous similar messages LustreError: 20664:0:(mdd_dir.c:4662:mdd_migrate_cmd_check()) lustre-MDD0002: '14' migration was interrupted, run 'lfs migrate -m 0 -c 3 -H crush 14' to finish migration: rc = -1 LustreError: 20664:0:(mdd_dir.c:4662:mdd_migrate_cmd_check()) Skipped 18 previous similar messages 9[20130]: segfault at 8 ip 00007f04f49f47e8 sp 00007ffc62a725a0 error 4 in ld-2.17.so[7f04f49e9000+22000] 5[25773]: segfault at 8 ip 00007f0fca7177e8 sp 00007fff6458d620 error 4 in ld-2.17.so[7f0fca70c000+22000] 0[26183]: segfault at 8 ip 00007f261f4087e8 sp 00007ffc61d24ba0 error 4 in ld-2.17.so[7f261f3fd000+22000] 5[31465]: segfault at 8 ip 00007f5a67ae47e8 sp 00007ffdeb92dda0 error 4 in ld-2.17.so[7f5a67ad9000+22000] dd (428) used greatest stack depth: 10016 bytes left 9[2171]: segfault at 8 ip 00007f2e277137e8 sp 00007fff903fa7f0 error 4 in ld-2.17.so[7f2e27708000+22000] Lustre: 26631:0:(mdd_dir.c:4741:mdd_migrate_object()) lustre-MDD0001: [0x240000403:0x1:0x0]/14 is open, migrate only dentry Lustre: 26631:0:(mdd_dir.c:4741:mdd_migrate_object()) Skipped 175 previous similar messages LustreError: 10629:0:(ldlm_lockd.c:257:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0002_UUID lock: ffff880098ebd680/0x89c6027d5f563562 lrc: 3/0,0 mode: PR/PR res: [0x280000406:0xe67:0x0].0x0 bits 0x1b/0x0 rrc: 5 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0x89c6027d5f563546 expref: 732 pid: 12909 timeout: 1543 lvb_type: 0 LustreError: 32101:0:(ldlm_lockd.c:1447:ldlm_handle_enqueue()) ### lock on destroyed export ffff8802dcc0dd28 ns: mdt-lustre-MDT0002_UUID lock: ffff8802a8a8f480/0x89c6027d5f566519 lrc: 3/0,0 mode: PR/PR res: [0x280000403:0x1:0x0].0x0 bits 0x12/0x0 rrc: 21 type: IBT gid 0 flags: 0x50200000000000 nid: 0@lo remote: 0x89c6027d5f56650b expref: 217 pid: 32101 timeout: 0 lvb_type: 0 LustreError: 32101:0:(ldlm_lockd.c:1447:ldlm_handle_enqueue()) Skipped 7 previous similar messages LustreError: 19078:0:(ldlm_lockd.c:2550:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1744702447 with bad export cookie 9927625166073124937 LustreError: lustre-MDT0002-mdc-ffff8800a8378958: operation mds_reint to node 0@lo failed: rc = -107 Lustre: lustre-MDT0002-mdc-ffff8800a8378958: Connection to lustre-MDT0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: lustre-MDT0002-mdc-ffff8800a8378958: This client was evicted by lustre-MDT0002; in progress operations using this service will fail. LustreError: 19078:0:(ldlm_lockd.c:2550:ldlm_cancel_handler()) Skipped 1 previous similar message LustreError: 4892:0:(file.c:6136:ll_inode_revalidate_fini()) lustre: revalidate FID [0x280000403:0x1:0x0] error: rc = -5 LustreError: 4892:0:(file.c:6136:ll_inode_revalidate_fini()) Skipped 68 previous similar messages LustreError: 4405:0:(mdc_request.c:1464:mdc_read_page()) lustre-MDT0002-mdc-ffff8800a8378958: [0x280000401:0x84:0x0] lock enqueue fails: rc = -108 Lustre: lustre-MDT0002-mdc-ffff8800a8378958: Connection restored to 192.168.123.22@tcp (at 0@lo) Lustre: 19946:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 513 < left 2168, rollback = 2 Lustre: 19946:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 5533 previous similar messages traps: 17[9297] trap invalid opcode ip:404850 sp:7fff85486fc8 error:0 in 17[400000+6000] 19[11069]: segfault at 8 ip 00007f404d2367e8 sp 00007ffd4e29b7f0 error 4 in ld-2.17.so[7f404d22b000+22000] 4[13827]: segfault at 406000 ip 0000000000406000 sp 00007ffce787b918 error 14 in 4[606000+1000] 3[21089]: segfault at 8 ip 00007fd759f5f7e8 sp 00007ffd20ef43e0 error 4 in ld-2.17.so[7fd759f54000+22000] | Link to test |
racer test 1: racer on clients: centos-90.localnet DURATION=2700 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 PGD 2bee00067 PUD 291a0a067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) lustre(OE) osp(OE) ofd(OE) lod(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) jbd2 mbcache lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) crc32_generic lnet(OE) dm_flakey dm_mod libcfs(OE) crc_t10dif crct10dif_generic crct10dif_common pcspkr virtio_balloon virtio_console i2c_piix4 ip_tables rpcsec_gss_krb5 ata_generic drm_kms_helper ttm pata_acpi drm drm_panel_orientation_quirks ata_piix serio_raw virtio_blk i2c_core floppy libata CPU: 2 PID: 23034 Comm: ll_sa_22779 Kdump: loaded Tainted: P OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.16.0-3.module_el8.7.0+1218+f626c2ff 04/01/2014 task: ffff8802ded224f0 ti: ffff880094bf0000 task.ti: ffff880094bf0000 RIP: 0010:[<ffffffff813f502c>] [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 RSP: 0018:ffff880094bf3d78 EFLAGS: 00010296 RAX: ffff8802879d2710 RBX: 0000000000000008 RCX: 0000000100260024 RDX: 0000000000000026 RSI: ffff8802879d2a40 RDI: 0000000000000008 RBP: ffff880094bf3d88 R08: ffff8802dd8da7c8 R09: 0000000000000000 R10: ffff8802dd8df0a8 R11: ffff8802dd8da7c8 R12: 0000000000000000 R13: ffff8802abf48010 R14: ffff8802879d2a40 R15: ffff8802dd8da7c8 FS: 0000000000000000(0000) GS:ffff880331a80000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 000000029e4f8000 CR4: 00000000000007e0 Call Trace: [<ffffffffa1108bf1>] ll_statahead_thread+0xdb1/0x2000 [lustre] [<ffffffff810df6e2>] ? pick_next_task_fair+0x702/0x8a0 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e329e>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffff810c834d>] ? finish_task_switch+0x5d/0x1b0 [<ffffffffa1107e40>] ? ll_statahead_handle.constprop.28+0x1c0/0x1c0 [lustre] [<ffffffff810ba114>] kthread+0xe4/0xf0 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 | Lustre: 11593:0:(mdd_dir.c:4701:mdd_migrate_object()) lustre-MDD0000: [0x200000403:0x1:0x0]/2 is open, migrate only dentry LustreError: 11593:0:(mdt_reint.c:2523:mdt_reint_migrate()) lustre-MDT0000: migrate [0x200000403:0x1:0x0]/2 failed: rc = -114 LustreError: 18434:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e6792e98: inode [0x200000403:0x6:0x0] mdc close failed: rc = -116 Lustre: 13330:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0002: opcode 7: before 516 < left 618, rollback = 7 Lustre: 13330:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 13330:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 13330:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 13330:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 13330:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 LustreError: 18849:0:(mdt_reint.c:2523:mdt_reint_migrate()) lustre-MDT0000: migrate [0x200000403:0x1:0x0]/1 failed: rc = -114 LustreError: 18849:0:(mdt_reint.c:2523:mdt_reint_migrate()) Skipped 3 previous similar messages Lustre: 13331:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0002: opcode 7: before 516 < left 618, rollback = 7 Lustre: 13331:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 3 previous similar messages Lustre: 13331:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 13331:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 13331:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 13331:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 13331:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 13331:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 13331:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 13331:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 13331:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 13331:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 3 previous similar messages LustreError: 19327:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e6792e98: inode [0x200000403:0x7:0x0] mdc close failed: rc = -116 LustreError: 19327:0:(file.c:247:ll_close_inode_openhandle()) Skipped 3 previous similar messages LustreError: 11592:0:(mdt_reint.c:2523:mdt_reint_migrate()) lustre-MDT0000: migrate [0x200000403:0x2:0x0]/11 failed: rc = -114 LustreError: 11592:0:(mdt_reint.c:2523:mdt_reint_migrate()) Skipped 2 previous similar messages Lustre: 18918:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0002: opcode 7: before 516 < left 618, rollback = 7 Lustre: 18918:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 7 previous similar messages Lustre: 18918:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 18918:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 7 previous similar messages Lustre: 18918:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 18918:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 7 previous similar messages Lustre: 18918:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 18918:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 7 previous similar messages Lustre: 18918:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 18918:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 7 previous similar messages Lustre: 18918:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 18918:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 7 previous similar messages Lustre: 18849:0:(mdd_dir.c:4701:mdd_migrate_object()) lustre-MDD0001: [0x240000403:0x1:0x0]/10 is open, migrate only dentry LustreError: 20268:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a6e7e678: inode [0x240000403:0x5:0x0] mdc close failed: rc = -116 LustreError: 20268:0:(file.c:247:ll_close_inode_openhandle()) Skipped 2 previous similar messages LustreError: 18849:0:(mdt_reint.c:2523:mdt_reint_migrate()) lustre-MDT0000: migrate [0x200000403:0x2:0x0]/10 failed: rc = -114 LustreError: 18849:0:(mdt_reint.c:2523:mdt_reint_migrate()) Skipped 2 previous similar messages Lustre: 13330:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0002: opcode 7: before 516 < left 618, rollback = 7 Lustre: 13330:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 11 previous similar messages Lustre: 18290:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 18290:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 13 previous similar messages Lustre: 18290:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 18290:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 13 previous similar messages Lustre: 18290:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 18290:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 13 previous similar messages Lustre: 18290:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 18290:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 13 previous similar messages Lustre: 18290:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 18290:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 13 previous similar messages Lustre: mdt00_008: service thread pid 17628 was inactive for 40.039 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 17628, comm: mdt00_008 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock_try+0xa0/0x250 [mdt] [<0>] mdt_object_open_lock+0x6b9/0xc10 [mdt] [<0>] mdt_reint_open+0x2401/0x2d70 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x73c/0xbb0 [mdt] [<0>] mdt_intent_open+0x93/0x480 [mdt] [<0>] mdt_intent_opc.constprop.74+0x211/0xc60 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x397/0x980 [ptlrpc] [<0>] ldlm_handle_enqueue+0x547/0x18d0 [ptlrpc] [<0>] tgt_enqueue+0x68/0x240 [ptlrpc] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 11592, comm: mdt_io00_001 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock+0x88/0x1c0 [mdt] [<0>] mdt_rename_source_lock+0xa9/0xd6 [mdt] [<0>] mdt_reint_migrate+0x1832/0x24b0 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x73c/0xbb0 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 18849, comm: mdt_io00_003 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_rename_lock+0xc3/0x2d0 [mdt] [<0>] mdt_reint_migrate+0x879/0x24b0 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x73c/0xbb0 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: mdt_io00_000: service thread pid 11591 was inactive for 40.041 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: mdt_io00_006: service thread pid 21343 was inactive for 40.113 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: mdt00_005: service thread pid 14892 was inactive for 40.013 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 3 previous similar messages Lustre: mdt00_000: service thread pid 11571 was inactive for 40.121 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 15 previous similar messages LustreError: 11400:0:(ldlm_lockd.c:252:expired_lock_main()) ### lock callback timer expired after 101s: evicting client at 0@lo ns: mdt-lustre-MDT0001_UUID lock: ffff88009cd06d00/0x75aaaf0721fe1c lrc: 3/0,0 mode: PW/PW res: [0x240000403:0x5:0x0].0x0 bits 0x4/0x0 rrc: 6 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0x75aaaf0721fdeb expref: 60 pid: 20530 timeout: 194 lvb_type: 0 LustreError: 22307:0:(ldlm_lockd.c:1492:ldlm_handle_enqueue()) ### lock on destroyed export ffff8802e5764a88 ns: mdt-lustre-MDT0001_UUID lock: ffff8802d5afe580/0x75aaaf0722946b lrc: 3/0,0 mode: PR/PR res: [0x240000403:0x1:0x0].0x0 bits 0x12/0x0 rrc: 19 type: IBT gid 0 flags: 0x50200000000000 nid: 0@lo remote: 0x75aaaf07229433 expref: 17 pid: 22307 timeout: 0 lvb_type: 0 Lustre: mdt00_009: service thread pid 17870 completed after 96.013s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_016: service thread pid 21711 completed after 96.017s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: lustre-MDT0001-mdc-ffff8802e6792e98: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-MDT0001-mdc-ffff8802e6792e98: Connection to lustre-MDT0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: mdt00_008: service thread pid 17628 completed after 100.342s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_001: service thread pid 11574 completed after 95.921s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 11592:0:(mdt_reint.c:2523:mdt_reint_migrate()) lustre-MDT0001: migrate [0x240000403:0x1:0x0]/10 failed: rc = -114 LustreError: 11592:0:(mdt_reint.c:2523:mdt_reint_migrate()) Skipped 1 previous similar message Lustre: mdt_io00_001: service thread pid 11592 completed after 100.334s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_018: service thread pid 22307 completed after 95.913s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_012: service thread pid 18712 completed after 96.744s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: lustre-MDT0001-mdc-ffff8802e6792e98: This client was evicted by lustre-MDT0001; in progress operations using this service will fail. Lustre: mdt00_004: service thread pid 13286 completed after 96.045s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_000: service thread pid 11571 completed after 96.085s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_013: service thread pid 18851 completed after 96.608s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_014: service thread pid 20507 completed after 96.731s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_007: service thread pid 17440 completed after 97.504s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_003: service thread pid 12712 completed after 97.970s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_006: service thread pid 16907 completed after 96.964s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_005: service thread pid 14892 completed after 98.156s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_015: service thread pid 20530 completed after 97.896s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_003: service thread pid 18849 completed after 100.170s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 17192:0:(file.c:6138:ll_inode_revalidate_fini()) lustre: revalidate FID [0x240000403:0x1:0x0] error: rc = -5 Lustre: mdt00_011: service thread pid 18340 completed after 96.047s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 20488:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e6792e98: inode [0x240000403:0x5:0x0] mdc close failed: rc = -108 LustreError: 20488:0:(file.c:247:ll_close_inode_openhandle()) Skipped 2 previous similar messages LustreError: 22650:0:(ldlm_resource.c:982:ldlm_resource_complain()) lustre-MDT0001-mdc-ffff8802e6792e98: namespace resource [0x200000403:0x1:0x0].0x0 (ffff8802c5f72d40) refcount nonzero (2) after lock cleanup; forcing cleanup. Lustre: mdt_io00_000: service thread pid 11591 completed after 99.868s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: lustre-MDT0001-mdc-ffff8802e6792e98: Connection restored to 192.168.123.92@tcp (at 0@lo) Lustre: mdt_io00_005: service thread pid 21319 completed after 99.379s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: 21343:0:(mdd_dir.c:4701:mdd_migrate_object()) lustre-MDD0000: [0x200000403:0x2:0x0]/10 is open, migrate only dentry Lustre: mdt_io00_007: service thread pid 21522 completed after 98.781s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_006: service thread pid 21343 completed after 99.365s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_008: service thread pid 21552 completed after 98.515s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_009: service thread pid 21902 completed after 97.656s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_010: service thread pid 21921 completed after 97.571s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_011: service thread pid 22016 completed after 97.171s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: 22075:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 509 < left 788, rollback = 2 Lustre: 22075:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 3/12/3, destroy: 0/0/0 Lustre: 22075:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 22075:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 2/2/0, xattr_set: 10/788/0 Lustre: 22075:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 22075:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 5/55/0, punch: 0/0/0, quota 1/3/0 Lustre: 22075:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 22075:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 10/203/3, delete: 0/0/0 Lustre: 22075:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 22075:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 4/4/0, ref_del: 0/0/0 Lustre: 22075:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: mdt_io00_012: service thread pid 22075 completed after 97.137s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: 21902:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 511 < left 788, rollback = 2 Lustre: 21902:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 9 previous similar messages Lustre: 18290:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0000: opcode 7: before 516 < left 618, rollback = 7 Lustre: 18290:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 3 previous similar messages Lustre: 18889:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 18889:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 27 previous similar messages Lustre: 18889:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 18889:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 27 previous similar messages Lustre: 18889:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 18889:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 27 previous similar messages Lustre: 18889:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 18889:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 27 previous similar messages Lustre: 18889:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 18889:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 27 previous similar messages LustreError: 29025:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e6792e98: inode [0x200000403:0x1c1:0x0] mdc close failed: rc = -13 LustreError: 29025:0:(file.c:247:ll_close_inode_openhandle()) Skipped 8 previous similar messages LustreError: 21921:0:(mdt_reint.c:2523:mdt_reint_migrate()) lustre-MDT0000: migrate [0x200000403:0x1:0x0]/15 failed: rc = -114 LustreError: 21921:0:(mdt_reint.c:2523:mdt_reint_migrate()) Skipped 16 previous similar messages Lustre: 21902:0:(mdd_dir.c:4701:mdd_migrate_object()) lustre-MDD0002: [0x280000403:0x1:0x0]/15 is open, migrate only dentry Lustre: 21902:0:(mdd_dir.c:4701:mdd_migrate_object()) Skipped 2 previous similar messages Lustre: 16907:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff8802d59b5a40 x1828673368618240/t4294971816(0) o101->a3b76864-4c04-4e3c-90f7-fc9f612b4670@0@lo:492/0 lens 376/816 e 0 to 0 dl 1743959137 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 Lustre: 22016:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 510 < left 699, rollback = 2 Lustre: 22016:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 9 previous similar messages traps: 10[31770] general protection ip:40576f sp:7ffd7bfc4740 error:0 in 10[400000+6000] LustreError: 11591:0:(mdd_dir.c:4621:mdd_migrate_cmd_check()) lustre-MDD0002: '19' migration was interrupted, run 'lfs migrate -m 1 -c 1 -H crush 19' to finish migration: rc = -1 Lustre: 13334:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 516 < left 618, rollback = 7 Lustre: 13334:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 9 previous similar messages Lustre: 21902:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 506 < left 1144, rollback = 2 Lustre: 21902:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 42 previous similar messages LustreError: 744:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a6e7e678: inode [0x200000404:0x204:0x0] mdc close failed: rc = -116 LustreError: 744:0:(file.c:247:ll_close_inode_openhandle()) Skipped 8 previous similar messages Lustre: 21552:0:(mdd_dir.c:4701:mdd_migrate_object()) lustre-MDD0002: [0x280000403:0x1:0x0]/4 is open, migrate only dentry Lustre: 21552:0:(mdd_dir.c:4701:mdd_migrate_object()) Skipped 3 previous similar messages Lustre: 18290:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 18290:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 96 previous similar messages Lustre: 18290:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 18290:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 96 previous similar messages Lustre: 18290:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 18290:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 96 previous similar messages Lustre: 18290:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 18290:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 96 previous similar messages Lustre: 18290:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 18290:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 96 previous similar messages LustreError: 2372:0:(lcommon_cl.c:179:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000403:0x2f4:0x0]: rc = -5 LustreError: 2372:0:(llite_lib.c:3698:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 21902:0:(mdd_dir.c:4621:mdd_migrate_cmd_check()) lustre-MDD0002: '18' migration was interrupted, run 'lfs migrate -m 1 -c 1 -H crush 18' to finish migration: rc = -1 LustreError: 21902:0:(mdt_reint.c:2523:mdt_reint_migrate()) lustre-MDT0002: migrate [0x200000403:0x2:0x0]/18 failed: rc = -1 LustreError: 21902:0:(mdt_reint.c:2523:mdt_reint_migrate()) Skipped 33 previous similar messages Lustre: 21902:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 509 < left 2585, rollback = 2 Lustre: 21902:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 43 previous similar messages Lustre: 18290:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 528, rollback = 7 Lustre: 18290:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 35 previous similar messages LustreError: 6009:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a6e7e678: inode [0x240000403:0x19a:0x0] mdc close failed: rc = -116 LustreError: 6009:0:(file.c:247:ll_close_inode_openhandle()) Skipped 8 previous similar messages 4[6917]: segfault at 0 ip 0000000000403e5f sp 00007ffff4fa64c0 error 6 in 4[400000+6000] 12[8045]: segfault at 0 ip (null) sp 00007fff6095f888 error 14 in 12[400000+6000] Lustre: 17870:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff8800a9cb5f40 x1828673375340928/t4294973008(0) o101->a3b76864-4c04-4e3c-90f7-fc9f612b4670@0@lo:518/0 lens 376/840 e 0 to 0 dl 1743959163 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 9[8492]: segfault at 8 ip 00007f3d7fd137e8 sp 00007ffe84106180 error 4 in ld-2.17.so[7f3d7fd08000+22000] Lustre: 22075:0:(mdd_dir.c:4701:mdd_migrate_object()) lustre-MDD0001: [0x200000403:0x2:0x0]/16 is open, migrate only dentry Lustre: 22075:0:(mdd_dir.c:4701:mdd_migrate_object()) Skipped 2 previous similar messages Lustre: 21552:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 503 < left 1011, rollback = 2 Lustre: 21552:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 152 previous similar messages Lustre: 18340:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000404:0x3e6:0x0] with magic=0xbd60bd0 Lustre: dir [0x280000403:0x26e:0x0] stripe 3 readdir failed: -2, directory is partially accessed! LustreError: 18849:0:(mdd_dir.c:4621:mdd_migrate_cmd_check()) lustre-MDD0002: '0' migration was interrupted, run 'lfs migrate -m 0 -c 3 -H crush 0' to finish migration: rc = -1 Lustre: 21552:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 3/12/11, destroy: 0/0/0 Lustre: 21552:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 294 previous similar messages Lustre: 21552:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 4/4/0, xattr_set: 19/1439/0 Lustre: 21552:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 294 previous similar messages Lustre: 21552:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 5/55/0, punch: 0/0/0, quota 4/54/0 Lustre: 21552:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 294 previous similar messages Lustre: 21552:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 13/251/4, delete: 1/1/0 Lustre: 21552:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 294 previous similar messages Lustre: 18375:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 18375:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 294 previous similar messages Lustre: 22075:0:(mdd_dir.c:4701:mdd_migrate_object()) lustre-MDD0001: [0x240000403:0x1:0x0]/10 is open, migrate only dentry Lustre: 22075:0:(mdd_dir.c:4701:mdd_migrate_object()) Skipped 3 previous similar messages LustreError: 12819:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8800a6e7e678: cannot apply new layout on [0x200000404:0x48f:0x0] : rc = -5 LustreError: 12819:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x200000404:0x48f:0x0] error -5. LustreError: 20501:0:(mdd_object.c:3902:mdd_close()) lustre-MDD0001: failed to get lu_attr of [0x240000403:0x1c3:0x0]: rc = -2 LustreError: 1954:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0000: [0x200000403:0x46c:0x0] migrate mdt count mismatch 2 != 3 Lustre: 1954:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff88009b2e2840 x1828673380259328/t4294973779(0) o101->a3b76864-4c04-4e3c-90f7-fc9f612b4670@0@lo:538/0 lens 376/840 e 0 to 0 dl 1743959183 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 LustreError: 21921:0:(mdt_reint.c:2523:mdt_reint_migrate()) lustre-MDT0002: migrate [0x280000404:0x2ca:0x0]/1 failed: rc = -114 LustreError: 21921:0:(mdt_reint.c:2523:mdt_reint_migrate()) Skipped 47 previous similar messages 18[14403]: segfault at 8 ip 00007f139a4807e8 sp 00007ffc071a6bb0 error 4 in ld-2.17.so[7f139a475000+22000] 9[14280]: segfault at 8 ip 00007f122e3fe7e8 sp 00007ffcc7ab8d90 error 4 in ld-2.17.so[7f122e3f3000+22000] 9[21934]: segfault at 8 ip 00007f00821037e8 sp 00007ffcc8ac9090 error 4 in ld-2.17.so[7f00820f8000+22000] 19[25811]: segfault at 8 ip 00007fcb8f3817e8 sp 00007ffec0baaea0 error 4 in ld-2.17.so[7fcb8f376000+22000] Lustre: 18849:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 507 < left 994, rollback = 2 Lustre: 18849:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 216 previous similar messages 17[26393]: segfault at 8 ip 00007f0b34aff7e8 sp 00007ffd5b64baf0 error 4 in ld-2.17.so[7f0b34af4000+22000] LustreError: 15309:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a6e7e678: inode [0x200000404:0x6a8:0x0] mdc close failed: rc = -116 LustreError: 15309:0:(file.c:247:ll_close_inode_openhandle()) Skipped 19 previous similar messages Lustre: 18889:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0000: opcode 7: before 516 < left 618, rollback = 7 Lustre: 18889:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 39 previous similar messages 9[28594]: segfault at 8 ip 00007f80a83687e8 sp 00007ffc1aa12990 error 4 in ld-2.17.so[7f80a835d000+22000] LustreError: 27155:0:(lcommon_cl.c:179:cl_file_inode_init()) lustre: failed to initialize cl_object [0x280000403:0x3f8:0x0]: rc = -5 LustreError: 27155:0:(llite_lib.c:3698:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 27155:0:(lcommon_cl.c:179:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000403:0x40b:0x0]: rc = -5 LustreError: 27155:0:(llite_lib.c:3698:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 21921:0:(mdd_dir.c:4621:mdd_migrate_cmd_check()) lustre-MDD0001: '6' migration was interrupted, run 'lfs migrate -m 2 -c 3 -H crush 6' to finish migration: rc = -1 Lustre: 21902:0:(mdd_dir.c:4701:mdd_migrate_object()) lustre-MDD0000: [0x200000403:0x1:0x0]/6 is open, migrate only dentry Lustre: 21902:0:(mdd_dir.c:4701:mdd_migrate_object()) Skipped 5 previous similar messages 12[31175]: segfault at 8 ip 00007f8a591727e8 sp 00007fff47276a40 error 4 in ld-2.17.so[7f8a59167000+22000] Lustre: 21711:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000404:0x778:0x0] with magic=0xbd60bd0 Lustre: 21711:0:(lod_lov.c:1417:lod_parse_striping()) Skipped 3 previous similar messages 9[32617]: segfault at 8 ip 00007f1af3c3f7e8 sp 00007ffec091f9b0 error 4 in ld-2.17.so[7f1af3c34000+22000] LustreError: 1672:0:(lcommon_cl.c:179:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000403:0x756:0x0]: rc = -5 LustreError: 1672:0:(lcommon_cl.c:179:cl_file_inode_init()) Skipped 1 previous similar message LustreError: 1672:0:(llite_lib.c:3698:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 1672:0:(llite_lib.c:3698:ll_prep_inode()) Skipped 1 previous similar message Lustre: dir [0x240000403:0x4a9:0x0] stripe 2 readdir failed: -2, directory is partially accessed! Lustre: Skipped 12 previous similar messages hrtimer: interrupt took 11733232 ns Lustre: 11579:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 513 < left 728, rollback = 2 Lustre: 11579:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 315 previous similar messages 1[4238]: segfault at 8 ip 00007f9d031d67e8 sp 00007ffdac5a0c80 error 4 in ld-2.17.so[7f9d031cb000+22000] LustreError: 4392:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8802e6792e98: cannot apply new layout on [0x240000403:0x505:0x0] : rc = -5 LustreError: 4392:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x240000403:0x505:0x0] error -5. Lustre: 21552:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 3/12/7, destroy: 0/0/0 Lustre: 21552:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 584 previous similar messages Lustre: 21552:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 4/4/0, xattr_set: 11/727/0 Lustre: 21552:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 584 previous similar messages Lustre: 21552:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 5/55/0, punch: 0/0/0, quota 4/78/0 Lustre: 21552:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 584 previous similar messages Lustre: 21552:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 12/215/4, delete: 0/0/0 Lustre: 21552:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 582 previous similar messages Lustre: 21552:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 6/6/0, ref_del: 0/0/0 Lustre: 21552:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 584 previous similar messages LustreError: 21343:0:(mdt_reint.c:2523:mdt_reint_migrate()) lustre-MDT0000: migrate [0x280000403:0x536:0x0]/4 failed: rc = -114 LustreError: 21343:0:(mdt_reint.c:2523:mdt_reint_migrate()) Skipped 66 previous similar messages LustreError: 7129:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8802e6792e98: cannot apply new layout on [0x240000403:0x505:0x0] : rc = -5 LustreError: 7129:0:(lcommon_cl.c:179:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000403:0x505:0x0]: rc = -5 LustreError: 7129:0:(llite_lib.c:3698:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 8[7381]: segfault at 8 ip 00007f35e132f7e8 sp 00007fff163ff290 error 4 in ld-2.17.so[7f35e1324000+22000] LustreError: 15135:0:(lcommon_cl.c:179:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000405:0x504:0x0]: rc = -5 LustreError: 15135:0:(lcommon_cl.c:179:cl_file_inode_init()) Skipped 7 previous similar messages LustreError: 15135:0:(llite_lib.c:3698:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 15135:0:(llite_lib.c:3698:ll_prep_inode()) Skipped 7 previous similar messages LustreError: 21552:0:(mdd_dir.c:4621:mdd_migrate_cmd_check()) lustre-MDD0001: '9' migration was interrupted, run 'lfs migrate -m 1 -c 1 -H crush 9' to finish migration: rc = -1 LustreError: 21552:0:(mdd_dir.c:4621:mdd_migrate_cmd_check()) Skipped 2 previous similar messages Lustre: 2147:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000404:0x5b7:0x0] with magic=0xbd60bd0 Lustre: 2147:0:(lod_lov.c:1417:lod_parse_striping()) Skipped 1 previous similar message 16[17124]: segfault at 8 ip 00007f1de39007e8 sp 00007fff34e2f480 error 4 in ld-2.17.so[7f1de38f5000+22000] LustreError: 17488:0:(mdd_object.c:3902:mdd_close()) lustre-MDD0001: failed to get lu_attr of [0x240000405:0x56b:0x0]: rc = -2 LustreError: 15112:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a6e7e678: inode [0x240000405:0x56b:0x0] mdc close failed: rc = -2 LustreError: 15112:0:(file.c:247:ll_close_inode_openhandle()) Skipped 18 previous similar messages Lustre: dir [0x240000405:0x554:0x0] stripe 3 readdir failed: -2, directory is partially accessed! LustreError: 22969:0:(mdd_object.c:3902:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000403:0x6dc:0x0]: rc = -2 Lustre: 25440:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 618, rollback = 7 Lustre: 25440:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 103 previous similar messages LustreError: 17488:0:(mdd_object.c:3902:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000403:0x9eb:0x0]: rc = -2 LustreError: 22075:0:(lustre_lmv.h:500:lmv_is_sane()) unknown layout LMV: magic=0xcd40cd0 count=3 index=2 hash=crush:0x82000003 version=1 migrate_offset=2 migrate_hash=fnv_1a_64:2 pool= Lustre: dir [0x240000403:0x6dc:0x0] stripe 1 readdir failed: -2, directory is partially accessed! LustreError: 21522:0:(mdd_dir.c:4621:mdd_migrate_cmd_check()) lustre-MDD0002: '5' migration was interrupted, run 'lfs migrate -m 0 -c 3 -H crush 5' to finish migration: rc = -1 LustreError: 22307:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0001: [0x240000405:0x645:0x0] migrate mdt count mismatch 2 != 1 LustreError: 1928:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0001: [0x240000403:0x62e:0x0] migrate mdt count mismatch 1 != 3 LustreError: 22577:0:(lcommon_cl.c:179:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000405:0x504:0x0]: rc = -5 LustreError: 22577:0:(lcommon_cl.c:179:cl_file_inode_init()) Skipped 9 previous similar messages LustreError: 22577:0:(llite_lib.c:3698:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 22577:0:(llite_lib.c:3698:ll_prep_inode()) Skipped 9 previous similar messages Lustre: 20574:0:(mdd_dir.c:4701:mdd_migrate_object()) lustre-MDD0002: [0x280000403:0x1:0x0]/10 is open, migrate only dentry Lustre: 20574:0:(mdd_dir.c:4701:mdd_migrate_object()) Skipped 28 previous similar messages Lustre: dir [0x240000405:0x645:0x0] stripe 2 readdir failed: -2, directory is partially accessed! Lustre: Skipped 4 previous similar messages Lustre: 21343:0:(mdt_reint.c:2443:mdt_reint_migrate()) lustre-MDT0002: [0x280000403:0x1:0x0]/13 is open, migrate only dentry 14[24452]: segfault at 8 ip 00007f9a707387e8 sp 00007ffe667f3420 error 4 in ld-2.17.so[7f9a7072d000+22000] LustreError: 11577:0:(mdd_object.c:3902:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000403:0xa84:0x0]: rc = -2 12[25943]: segfault at 8 ip 00007ff595afd7e8 sp 00007ffcbe18d920 error 4 in ld-2.17.so[7ff595af2000+22000] Lustre: 706:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000405:0x64b:0x0] with magic=0xbd60bd0 Lustre: 706:0:(lod_lov.c:1417:lod_parse_striping()) Skipped 1 previous similar message Lustre: dir [0x200000403:0xbfd:0x0] stripe 3 readdir failed: -2, directory is partially accessed! 12[26838]: segfault at 8 ip 00007fe4ef1337e8 sp 00007ffec2007ec0 error 4 in ld-2.17.so[7fe4ef128000+22000] LustreError: 18849:0:(mdd_dir.c:4621:mdd_migrate_cmd_check()) lustre-MDD0000: '16' migration was interrupted, run 'lfs migrate -m 2 -c 2 -H crush 16' to finish migration: rc = -1 LustreError: 12712:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0000: [0x200000404:0xc09:0x0] migrate mdt count mismatch 1 != 2 Lustre: 21319:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 501 < left 1144, rollback = 2 Lustre: 21319:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 1136 previous similar messages 3[30516]: segfault at 8 ip 00007f5c7ac777e8 sp 00007ffcf5d3e2d0 error 4 in ld-2.17.so[7f5c7ac6c000+22000] LustreError: 17488:0:(mdd_object.c:3902:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000404:0xc09:0x0]: rc = -2 LustreError: 17488:0:(mdd_object.c:3902:mdd_close()) Skipped 4 previous similar messages LustreError: 31136:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8800a6e7e678: cannot apply new layout on [0x240000405:0x8e9:0x0] : rc = -5 LustreError: 31136:0:(lov_object.c:1341:lov_layout_change()) Skipped 1 previous similar message LustreError: 31136:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x240000405:0x8e9:0x0] error -5. Lustre: dir [0x240000405:0x98f:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 17 previous similar messages LustreError: 4513:0:(lcommon_cl.c:179:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000404:0xd23:0x0]: rc = -5 LustreError: 4513:0:(lcommon_cl.c:179:cl_file_inode_init()) Skipped 16 previous similar messages LustreError: 4513:0:(llite_lib.c:3698:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 4513:0:(llite_lib.c:3698:ll_prep_inode()) Skipped 16 previous similar messages LustreError: 30802:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8800a6e7e678: cannot apply new layout on [0x240000405:0x504:0x0] : rc = -5 LustreError: 30802:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x240000405:0x504:0x0] error -5. LustreError: 1926:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0001: [0x240000405:0x4ea:0x0] migrate mdt count mismatch 2 != 1 Lustre: 1936:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000405:0x97e:0x0] with magic=0xbd60bd0 Lustre: 1936:0:(lod_lov.c:1417:lod_parse_striping()) Skipped 3 previous similar messages LustreError: 22434:0:(mdd_dir.c:4621:mdd_migrate_cmd_check()) lustre-MDD0000: '3' migration was interrupted, run 'lfs migrate -m 2 -c 3 -H crush 3' to finish migration: rc = -1 LustreError: 22434:0:(mdd_dir.c:4621:mdd_migrate_cmd_check()) Skipped 6 previous similar messages LustreError: 22969:0:(mdd_object.c:3902:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000403:0xd33:0x0]: rc = -2 2[7974]: segfault at 8 ip 00007f8d270267e8 sp 00007ffeb6fc1580 error 4 in ld-2.17.so[7f8d2701b000+22000] LustreError: 16613:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8800a6e7e678: cannot apply new layout on [0x240000405:0x8e9:0x0] : rc = -5 4[11779]: segfault at 8 ip 00007f888a1f37e8 sp 00007ffe6a70f590 error 4 in ld-2.17.so[7f888a1e8000+22000] 5[12872]: segfault at 8 ip 00007fddd1d367e8 sp 00007ffc124fad00 error 4 in ld-2.17.so[7fddd1d2b000+22000] Lustre: 1939:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000403:0xa3c:0x0] with magic=0xbd60bd0 Lustre: 1939:0:(lod_lov.c:1417:lod_parse_striping()) Skipped 1 previous similar message 7[19560]: segfault at 8 ip 00007fe49f0737e8 sp 00007fff775fff50 error 4 in ld-2.17.so[7fe49f068000+22000] Lustre: dir [0x280000403:0xb6f:0x0] stripe 2 readdir failed: -2, directory is partially accessed! 17[20280]: segfault at 1c ip 00000000004017b0 sp 00007fff24d94570 error 6 in 17[400000+6000] Lustre: 25717:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000404:0x1226:0x0] with magic=0xbd60bd0 Lustre: 25717:0:(lod_lov.c:1417:lod_parse_striping()) Skipped 15 previous similar messages LustreError: 16762:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8800a6e7e678: cannot apply new layout on [0x240000405:0x8e9:0x0] : rc = -5 LustreError: 16762:0:(lov_object.c:1341:lov_layout_change()) Skipped 5 previous similar messages LustreError: 22493:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x280000403:0xd96:0x0] error -5. traps: 8[22447] trap invalid opcode ip:40571a sp:7ffe8a2f5db8 error:0 in 8[400000+6000] LustreError: 234:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 2 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 Lustre: 22016:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 3/12/11, destroy: 0/0/0 Lustre: 22016:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 2550 previous similar messages Lustre: 22016:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 2/2/0, xattr_set: 15/1233/0 Lustre: 22016:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 2550 previous similar messages Lustre: 22016:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 5/55/0, punch: 0/0/0, quota 4/54/0 Lustre: 22016:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 2550 previous similar messages Lustre: 22016:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 10/203/4, delete: 0/0/0 Lustre: 22016:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 2550 previous similar messages Lustre: 22016:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 4/4/0, ref_del: 0/0/0 Lustre: 22016:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 2550 previous similar messages LustreError: 20795:0:(mdt_reint.c:2523:mdt_reint_migrate()) lustre-MDT0000: migrate [0x240000403:0xd0d:0x0]/16 failed: rc = -2 LustreError: 20795:0:(mdt_reint.c:2523:mdt_reint_migrate()) Skipped 305 previous similar messages LustreError: 134:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 9 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 16[519]: segfault at 406000 ip 0000000000406000 sp 00007fffe900a508 error 14 in 16[606000+1000] LustreError: 17488:0:(mdd_object.c:3902:mdd_close()) lustre-MDD0002: failed to get lu_attr of [0x280000404:0xb6d:0x0]: rc = -2 LustreError: 17488:0:(mdd_object.c:3902:mdd_close()) Skipped 1 previous similar message 3[1479]: segfault at 8 ip 00007ff36c0d27e8 sp 00007ffc8a5cc9b0 error 4 in ld-2.17.so[7ff36c0c7000+22000] LustreError: 3082:0:(lcommon_cl.c:179:cl_file_inode_init()) lustre: failed to initialize cl_object [0x280000403:0xd82:0x0]: rc = -5 LustreError: 3082:0:(lcommon_cl.c:179:cl_file_inode_init()) Skipped 90 previous similar messages LustreError: 3082:0:(llite_lib.c:3698:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 3082:0:(llite_lib.c:3698:ll_prep_inode()) Skipped 90 previous similar messages LustreError: 3508:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x240000403:0xb0d:0x0] error -5. LustreError: 17628:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0002: [0x280000403:0x10ab:0x0] migrate mdt count mismatch 1 != 3 LustreError: 5992:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a6e7e678: inode [0x240000405:0xf07:0x0] mdc close failed: rc = -116 LustreError: 5992:0:(file.c:247:ll_close_inode_openhandle()) Skipped 52 previous similar messages 18[6148]: segfault at 8 ip 00007f13fb4587e8 sp 00007ffd6cdd5370 error 4 in ld-2.17.so[7f13fb44d000+22000] Lustre: 18889:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 618, rollback = 7 Lustre: 18889:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 161 previous similar messages LustreError: 171:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 2 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 LustreError: 20198:0:(mdd_dir.c:4621:mdd_migrate_cmd_check()) lustre-MDD0002: '7' migration was interrupted, run 'lfs migrate -m 0 -c 2 -H crush 7' to finish migration: rc = -1 LustreError: 20198:0:(mdd_dir.c:4621:mdd_migrate_cmd_check()) Skipped 12 previous similar messages Lustre: 20198:0:(mdd_dir.c:4701:mdd_migrate_object()) lustre-MDD0001: [0x200000403:0x1016:0x0]/3 is open, migrate only dentry Lustre: 20198:0:(mdd_dir.c:4701:mdd_migrate_object()) Skipped 78 previous similar messages 16[9648]: segfault at 8 ip 00007f67ecc4b7e8 sp 00007ffe5f2a57a0 error 4 in ld-2.17.so[7f67ecc40000+22000] 16[9961]: segfault at 8 ip 00007f52d947e7e8 sp 00007fff5ee702f0 error 4 in ld-2.17.so[7f52d9473000+22000] Lustre: 21711:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0x1530:0x0] with magic=0xbd60bd0 Lustre: 21711:0:(lod_lov.c:1417:lod_parse_striping()) Skipped 15 previous similar messages LustreError: 10849:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8802e6792e98: cannot apply new layout on [0x240000403:0xb0d:0x0] : rc = -5 LustreError: 10849:0:(lov_object.c:1341:lov_layout_change()) Skipped 5 previous similar messages LustreError: 244:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 2 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 LustreError: 244:0:(statahead.c:825:ll_statahead_interpret_work()) Skipped 2 previous similar messages 17[13876]: segfault at 8 ip 00007f66ede007e8 sp 00007fffe8d670f0 error 4 in ld-2.17.so[7f66eddf5000+22000] Lustre: 18849:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 513 < left 1456, rollback = 2 Lustre: 18849:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 2564 previous similar messages Lustre: 11579:0:(osd_io.c:1951:osd_ldiskfs_write()) lustre-MDT0001: adding bh without locking off 1245184 (block 304, size 3744, offs 1245184) mrename (21461) used greatest stack depth: 10096 bytes left LustreError: 17103:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x240000403:0x1056:0x0] error -5. LustreError: 17103:0:(vvp_io.c:1903:vvp_io_init()) Skipped 1 previous similar message Lustre: dir [0x200000404:0x1654:0x0] stripe 2 readdir failed: -2, directory is partially accessed! Lustre: Skipped 39 previous similar messages | Link to test |
racer test 2: racer rename: centos-105.localnet DURATION=2700 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 PGD 2dc1ff067 PUD 2e0303067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) lustre(OE) osp(OE) ofd(OE) lod(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) jbd2 mbcache lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) crc32_generic lnet(OE) dm_flakey dm_mod libcfs(OE) crc_t10dif crct10dif_generic crct10dif_common virtio_console virtio_balloon i2c_piix4 pcspkr ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm ata_piix drm_panel_orientation_quirks virtio_blk libata serio_raw i2c_core floppy CPU: 14 PID: 28472 Comm: ll_sa_28462 Kdump: loaded Tainted: P OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.16.0-3.module_el8.7.0+1218+f626c2ff 04/01/2014 task: ffff880273d98010 ti: ffff880222264000 task.ti: ffff880222264000 RIP: 0010:[<ffffffff813f502c>] [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 RSP: 0018:ffff880222267d78 EFLAGS: 00010296 RAX: ffff8801fcb13a10 RBX: 0000000000000008 RCX: 000000010026001b RDX: 0000000000000026 RSI: ffff8801fcb13d40 RDI: 0000000000000008 RBP: ffff880222267d88 R08: ffff880224cea7c8 R09: 0000000000000000 R10: ffff880224cea478 R11: ffff880224cea7c8 R12: 0000000000000000 R13: ffff88027ede0010 R14: ffff8801fcb13d40 R15: ffff880224cea7c8 FS: 0000000000000000(0000) GS:ffff880331d80000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 000000031d0e8000 CR4: 00000000000007e0 Call Trace: [<ffffffffa10ecbf1>] ll_statahead_thread+0xdb1/0x2000 [lustre] [<ffffffff817e320e>] ? _raw_spin_lock+0x1e/0x20 [<ffffffff810dbb51>] ? put_prev_entity+0x31/0x400 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e329e>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffff810c834d>] ? finish_task_switch+0x5d/0x1b0 [<ffffffffa10ebe40>] ? ll_statahead_handle.constprop.28+0x1c0/0x1c0 [lustre] [<ffffffff810ba114>] kthread+0xe4/0xf0 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 | Lustre: 14200:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff8802d34eb740 x1828476471451136/t4295242305(0) o101->d30d3017-d020-4b0b-8fed-ed809fbf4bbb@0@lo:500/0 lens 376/3056 e 0 to 0 dl 1743773415 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 Lustre: 18117:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff880094c49440 x1828476472028672/t4295158644(0) o101->d30d3017-d020-4b0b-8fed-ed809fbf4bbb@0@lo:504/0 lens 376/5408 e 0 to 0 dl 1743773419 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 Lustre: 984:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff88028079e440 x1828476472434176/t4295159509(0) o101->d30d3017-d020-4b0b-8fed-ed809fbf4bbb@0@lo:505/0 lens 376/33440 e 0 to 0 dl 1743773420 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 Lustre: 984:0:(mdt_recovery.c:102:mdt_req_from_lrd()) Skipped 3 previous similar messages Lustre: 14191:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff88026f87f840 x1828476473401216/t4295242838(0) o101->35173acf-7059-4a76-9534-ed215803bb38@0@lo:510/0 lens 376/35456 e 0 to 0 dl 1743773425 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 Lustre: 14151:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff880287d37840 x1828476475362688/t4295243093(0) o101->35173acf-7059-4a76-9534-ed215803bb38@0@lo:517/0 lens 376/35792 e 0 to 0 dl 1743773432 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 Lustre: 14151:0:(mdt_recovery.c:102:mdt_req_from_lrd()) Skipped 3 previous similar messages Lustre: 23986:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff880085645540 x1828476482706688/t4295160004(0) o101->35173acf-7059-4a76-9534-ed215803bb38@0@lo:573/0 lens 376/39544 e 0 to 0 dl 1743773488 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 Lustre: 14151:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff88025f5fee40 x1828476490334720/t4295244353(0) o101->35173acf-7059-4a76-9534-ed215803bb38@0@lo:590/0 lens 376/35792 e 0 to 0 dl 1743773505 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 Lustre: 14151:0:(mdt_recovery.c:102:mdt_req_from_lrd()) Skipped 2 previous similar messages ptlrpc_watchdog_fire: 22 callbacks suppressed Lustre: mdt_out00_004: service thread pid 17133 was inactive for 40.154 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 17133, comm: mdt_out00_004 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_pdo_lock+0x369/0x750 [mdt] [<0>] mdt_parent_lock+0x76/0x2a0 [mdt] [<0>] mdt_getattr_name_lock+0x1844/0x2cc0 [mdt] [<0>] mdt_intent_getattr+0x2cc/0x4e0 [mdt] [<0>] mdt_intent_opc.constprop.74+0x211/0xc60 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x397/0x980 [ptlrpc] [<0>] ldlm_handle_enqueue+0x547/0x18d0 [ptlrpc] [<0>] mdt_batch_getattr+0x6c/0x190 [mdt] [<0>] mdt_batch+0x465/0x195b [mdt] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 14104, comm: mdt_out00_009 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_pdo_lock+0x369/0x750 [mdt] [<0>] mdt_parent_lock+0x76/0x2a0 [mdt] [<0>] mdt_getattr_name_lock+0x1844/0x2cc0 [mdt] [<0>] mdt_intent_getattr+0x2cc/0x4e0 [mdt] [<0>] mdt_intent_opc.constprop.74+0x211/0xc60 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x397/0x980 [ptlrpc] [<0>] ldlm_handle_enqueue+0x547/0x18d0 [ptlrpc] [<0>] mdt_batch_getattr+0x6c/0x190 [mdt] [<0>] mdt_batch+0x465/0x195b [mdt] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 28191, comm: mdt00_040 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_pdo_lock+0x369/0x750 [mdt] [<0>] mdt_parent_lock+0x76/0x2a0 [mdt] [<0>] mdt_getattr_name_lock+0x1844/0x2cc0 [mdt] [<0>] mdt_intent_getattr+0x2cc/0x4e0 [mdt] [<0>] mdt_intent_opc.constprop.74+0x211/0xc60 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x397/0x980 [ptlrpc] [<0>] ldlm_handle_enqueue+0x547/0x18d0 [ptlrpc] [<0>] tgt_enqueue+0x68/0x240 [ptlrpc] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: 18563:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff880287d09940 x1828476502221440/t4295162839(0) o101->35173acf-7059-4a76-9534-ed215803bb38@0@lo:686/0 lens 376/46144 e 0 to 0 dl 1743773601 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 Lustre: 18563:0:(mdt_recovery.c:102:mdt_req_from_lrd()) Skipped 1 previous similar message LustreError: 10209:0:(ldlm_lockd.c:252:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0001_UUID lock: ffff88025ff23100/0x29b04b8d24ad2723 lrc: 3/0,0 mode: PR/PR res: [0x240000403:0x4afb:0x0].0x0 bits 0x13/0x0 rrc: 3 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0x29b04b8d24ad2284 expref: 770 pid: 17133 timeout: 3005 lvb_type: 0 Lustre: mdt00_040: service thread pid 28191 completed after 99.799s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 17133:0:(ldlm_lockd.c:1492:ldlm_handle_enqueue()) ### lock on destroyed export ffff8803249a53d8 ns: mdt-lustre-MDT0001_UUID lock: ffff8802771b3c40/0x29b04b8d24c39fe1 lrc: 3/0,0 mode: PR/PR res: [0x240000403:0x4afb:0x0].0x0 bits 0x13/0x0 rrc: 4 type: IBT gid 0 flags: 0x50200000000000 nid: 0@lo remote: 0x29b04b8d24ad2292 expref: 208 pid: 17133 timeout: 0 lvb_type: 0 LustreError: 17133:0:(ldlm_lockd.c:1492:ldlm_handle_enqueue()) Skipped 2 previous similar messages LustreError: lustre-MDT0001-mdc-ffff8802e80e8008: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: mdt_out00_009: service thread pid 14104 completed after 100.090s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: lustre-MDT0001-mdc-ffff8802e80e8008: Connection to lustre-MDT0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: lustre-MDT0001-mdc-ffff8802e80e8008: This client was evicted by lustre-MDT0001; in progress operations using this service will fail. Lustre: mdt_out00_004: service thread pid 17133 completed after 100.145s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: lustre-MDT0001-mdc-ffff8802e80e8008: Connection restored to 192.168.123.107@tcp (at 0@lo) Lustre: 14172:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 6/24/0, destroy: 1/4/0 Lustre: 14172:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 1287 previous similar messages Lustre: 14172:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1787/1787/0, xattr_set: 2680/25124/0 Lustre: 14172:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 1287 previous similar messages Lustre: 14172:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 28/157/0, punch: 0/0/0, quota 1/3/0 Lustre: 14172:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 1287 previous similar messages Lustre: 14172:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 7/118/0, delete: 2/5/1 Lustre: 14172:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 1287 previous similar messages Lustre: 14172:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 1/1/0, ref_del: 2/2/1 Lustre: 14172:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 1287 previous similar messages Lustre: 14172:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff88008f536e40 x1828476528417664/t4295162055(0) o101->d30d3017-d020-4b0b-8fed-ed809fbf4bbb@0@lo:40/0 lens 376/45112 e 0 to 0 dl 1743773710 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 Lustre: 14172:0:(mdt_recovery.c:102:mdt_req_from_lrd()) Skipped 6 previous similar messages sched: RT throttling activated LustreError: 10209:0:(ldlm_lockd.c:252:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0001_UUID lock: ffff88024ec51a80/0x29b04b8d24c82e2c lrc: 3/0,0 mode: PR/PR res: [0x240000403:0x4b06:0x0].0x0 bits 0x13/0x0 rrc: 8 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0x29b04b8d24c82dc3 expref: 581 pid: 9426 timeout: 3127 lvb_type: 0 LustreError: lustre-MDT0001-mdc-ffff8802e7a75d28: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-MDT0001-mdc-ffff8802e7a75d28: Connection to lustre-MDT0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: lustre-MDT0001-mdc-ffff8802e7a75d28: This client was evicted by lustre-MDT0001; in progress operations using this service will fail. LustreError: 12490:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e7a75d28: inode [0x240000403:0x1:0x0] mdc close failed: rc = -108 LustreError: 12490:0:(file.c:247:ll_close_inode_openhandle()) Skipped 113 previous similar messages Lustre: lustre-MDT0001-mdc-ffff8802e7a75d28: Connection restored to 192.168.123.107@tcp (at 0@lo) Lustre: 984:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0001: opcode 0: before 515 < left 2909, rollback = 0 Lustre: 984:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 1189 previous similar messages Lustre: 25750:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff8802458ccb40 x1828476571828736/t4295164354(0) o101->d30d3017-d020-4b0b-8fed-ed809fbf4bbb@0@lo:168/0 lens 376/45976 e 0 to 0 dl 1743773838 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 Lustre: 25750:0:(mdt_recovery.c:102:mdt_req_from_lrd()) Skipped 7 previous similar messages LustreError: 10209:0:(ldlm_lockd.c:252:expired_lock_main()) ### lock callback timer expired after 101s: evicting client at 0@lo ns: filter-lustre-OST0003_UUID lock: ffff88029c9987c0/0x29b04b8d24d05043 lrc: 3/0,0 mode: PW/PW res: [0x380000402:0x2146:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->1048575) gid 0 flags: 0x60000400030020 nid: 0@lo remote: 0x29b04b8d24d04a5b expref: 3814 pid: 5861 timeout: 3177 lvb_type: 0 LustreError: lustre-OST0003-osc-ffff8802e7a75d28: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-OST0003-osc-ffff8802e7a75d28: Connection to lustre-OST0003 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: lustre-OST0003-osc-ffff8802e7a75d28: This client was evicted by lustre-OST0003; in progress operations using this service will fail. LustreError: 5865:0:(ldlm_lockd.c:1492:ldlm_handle_enqueue()) ### lock on destroyed export ffff8802d4fb4a88 ns: filter-lustre-OST0003_UUID lock: ffff88024402ed00/0x29b04b8d24e6721b lrc: 3/0,0 mode: --/PW res: [0x380000401:0x2893:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x50000000020000 nid: 0@lo remote: 0x29b04b8d24e6720d expref: 3320 pid: 5865 timeout: 0 lvb_type: 0 LustreError: 32198:0:(client.c:1282:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff880074dd3740 x1828476575121920/t0(0) o104->lustre-OST0003@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' uid:4294967295 gid:4294967295 LustreError: 5865:0:(ldlm_lockd.c:1492:ldlm_handle_enqueue()) Skipped 1 previous similar message LustreError: 740:0:(client.c:1282:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88007502d040 x1828476575559680/t0(0) o104->lustre-OST0003@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' uid:4294967295 gid:4294967295 LustreError: 740:0:(client.c:1282:ptlrpc_import_delay_req()) Skipped 5 previous similar messages Lustre: 3994:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x200000405:0x1084:0x0]/ may get corrupted (rc -108) Lustre: 3984:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x200000404:0x8033:0x0]/ may get corrupted (rc -108) Lustre: 3984:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x240000407:0xd3:0x0]// may get corrupted (rc -108) Lustre: 3982:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x240000407:0xc2:0x0]/ may get corrupted (rc -108) Lustre: 3982:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x200000405:0x1092:0x0]/ may get corrupted (rc -108) Lustre: 3997:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x280000405:0x3892:0x0]/ may get corrupted (rc -108) Lustre: 3986:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x240000406:0x157:0x0]/ may get corrupted (rc -108) Lustre: 3998:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x240000407:0xab:0x0]/ may get corrupted (rc -108) LustreError: 15933:0:(ldlm_resource.c:982:ldlm_resource_complain()) lustre-OST0003-osc-ffff8802e7a75d28: namespace resource [0x380000402:0x231e:0x0].0x0 (ffff8802562e9e40) refcount nonzero (1) after lock cleanup; forcing cleanup. Lustre: lustre-OST0003-osc-ffff8802e7a75d28: Connection restored to 192.168.123.107@tcp (at 0@lo) LustreError: 10209:0:(ldlm_lockd.c:252:expired_lock_main()) ### lock callback timer expired after 101s: evicting client at 0@lo ns: filter-lustre-OST0003_UUID lock: ffff88026d9a1300/0x29b04b8d24ce6ea1 lrc: 3/0,0 mode: PW/PW res: [0x380000402:0x1e00:0x0].0x0 rrc: 2 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000480020020 nid: 0@lo remote: 0x29b04b8d24ce6e8c expref: 4324 pid: 8557 timeout: 3254 lvb_type: 0 LustreError: lustre-OST0003-osc-ffff8802e80e8008: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-OST0003-osc-ffff8802e80e8008: Connection to lustre-OST0003 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: lustre-OST0003-osc-ffff8802e80e8008: This client was evicted by lustre-OST0003; in progress operations using this service will fail. LustreError: Skipped 1 previous similar message LustreError: 7472:0:(client.c:1282:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff880259bd1940 x1828476602023936/t0(0) o104->lustre-OST0003@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' uid:4294967295 gid:4294967295 LustreError: 32205:0:(ldlm_lockd.c:1492:ldlm_handle_enqueue()) ### lock on destroyed export ffff8802e541b7e8 ns: filter-lustre-OST0003_UUID lock: ffff88024ceec000/0x29b04b8d24f94e23 lrc: 3/0,0 mode: PW/PW res: [0x380000400:0x40a3:0x0].0x0 rrc: 2 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x50000080020000 nid: 0@lo remote: 0x29b04b8d24f94e1c expref: 4228 pid: 32205 timeout: 0 lvb_type: 0 Lustre: 3988:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x240000406:0x1bf:0x0]/ may get corrupted (rc -108) Lustre: 3987:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x200000404:0x80c7:0x0]// may get corrupted (rc -108) Lustre: 3991:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x280000404:0x5435:0x0]// may get corrupted (rc -108) Lustre: 3991:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x280000404:0x5444:0x0]/ may get corrupted (rc -108) Lustre: 3992:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x280000404:0x5418:0x0]/ may get corrupted (rc -108) Lustre: 3992:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x200000405:0x1151:0x0]/ may get corrupted (rc -108) Lustre: 3996:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x280000404:0x547a:0x0]/ may get corrupted (rc -108) Lustre: 3996:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x240000407:0x17a:0x0]/ may get corrupted (rc -108) LustreError: 21061:0:(ldlm_resource.c:982:ldlm_resource_complain()) lustre-OST0003-osc-ffff8802e80e8008: namespace resource [0x380000400:0x3860:0x0].0x0 (ffff880083d27740) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 21061:0:(ldlm_resource.c:982:ldlm_resource_complain()) Skipped 137 previous similar messages Lustre: lustre-OST0003-osc-ffff8802e80e8008: Connection restored to 192.168.123.107@tcp (at 0@lo) LustreError: 10209:0:(ldlm_lockd.c:252:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0001_UUID lock: ffff88025f932980/0x29b04b8d24e4a91c lrc: 3/0,0 mode: PW/PW res: [0x300000401:0x2e03:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000480020020 nid: 0@lo remote: 0x29b04b8d24e4a915 expref: 5228 pid: 8435 timeout: 3292 lvb_type: 0 LustreError: 12310:0:(client.c:1282:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff880287eb8040 x1828476618856704/t0(0) o105->lustre-OST0001@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' uid:4294967295 gid:4294967295 LustreError: 12310:0:(client.c:1282:ptlrpc_import_delay_req()) Skipped 11 previous similar messages LustreError: 2359:0:(ldlm_lockd.c:1492:ldlm_handle_enqueue()) ### lock on destroyed export ffff8802e76c6fc8 ns: filter-lustre-OST0001_UUID lock: ffff88026bcd61c0/0x29b04b8d25053863 lrc: 3/0,0 mode: --/PW res: [0x300000400:0x440a:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x50000000020000 nid: 0@lo remote: 0x29b04b8d25053855 expref: 5219 pid: 2359 timeout: 0 lvb_type: 0 LustreError: 19327:0:(ldlm_lockd.c:2594:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1743773812 with bad export cookie 3003984021002781946 LustreError: 19327:0:(ldlm_lockd.c:2594:ldlm_cancel_handler()) Skipped 2 previous similar messages Lustre: lustre-OST0001-osc-ffff8802e7a75d28: Connection to lustre-OST0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: lustre-OST0001-osc-ffff8802e7a75d28: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: lustre-OST0001-osc-ffff8802e7a75d28: This client was evicted by lustre-OST0001; in progress operations using this service will fail. Lustre: 3984:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x200000405:0x1220:0x0]/ may get corrupted (rc -108) Lustre: 3984:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x280000405:0x391e:0x0]/ may get corrupted (rc -108) Lustre: 3992:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x200000405:0x11fd:0x0]/ may get corrupted (rc -108) Lustre: 3991:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x240000407:0x203:0x0]// may get corrupted (rc -108) Lustre: 3998:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x280000405:0x3a48:0x0]/ may get corrupted (rc -108) Lustre: 3991:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x280000405:0x39b3:0x0]/ may get corrupted (rc -108) Lustre: 3983:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x280000405:0x3a39:0x0]/ may get corrupted (rc -108) Lustre: 3990:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x240000407:0x219:0x0]/ may get corrupted (rc -108) Lustre: 3989:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x200000405:0x11f2:0x0]/ may get corrupted (rc -108) Lustre: 3989:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x200000405:0x11ec:0x0]/ may get corrupted (rc -108) Lustre: 3987:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x200000405:0x11c3:0x0]/ may get corrupted (rc -108) Lustre: 3986:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x280000404:0x54a8:0x0]/ may get corrupted (rc -108) LustreError: 24120:0:(ldlm_resource.c:982:ldlm_resource_complain()) lustre-OST0001-osc-ffff8802e7a75d28: namespace resource [0x300000402:0x359e:0x0].0x0 (ffff88006fb3f240) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 24120:0:(ldlm_resource.c:982:ldlm_resource_complain()) Skipped 338 previous similar messages LustreError: 24120:0:(ldlm_resource.c:982:ldlm_resource_complain()) lustre-OST0001-osc-ffff8802e7a75d28: namespace resource [0x300000400:0x5d40:0x0].0x0 (ffff8800952620c0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 24120:0:(ldlm_resource.c:982:ldlm_resource_complain()) Skipped 145 previous similar messages Lustre: lustre-OST0001-osc-ffff8802e7a75d28: Connection restored to 192.168.123.107@tcp (at 0@lo) LustreError: 18563:0:(mdt_open.c:1302:mdt_cross_open()) lustre-MDT0000: [0x200000404:0x8266:0x0] doesn't exist!: rc = -14 Lustre: 14217:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff8802f0486940 x1828476686682496/t4295262828(0) o101->35173acf-7059-4a76-9534-ed215803bb38@0@lo:429/0 lens 376/47984 e 0 to 0 dl 1743774099 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 Lustre: 14217:0:(mdt_recovery.c:102:mdt_req_from_lrd()) Skipped 22 previous similar messages Lustre: lustre-OST0002-osc-MDT0002: update sequence from 0x340000400 to 0x340000403 Lustre: lustre-OST0000-osc-MDT0001: update sequence from 0x2c0000401 to 0x2c0000403 Lustre: lustre-OST0001-osc-MDT0002: update sequence from 0x300000400 to 0x300000403 LustreError: 10209:0:(ldlm_lockd.c:252:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0002_UUID lock: ffff88027436b100/0x29b04b8d251ce81c lrc: 3/0,0 mode: PW/PW res: [0x340000401:0x6680:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->393215) gid 0 flags: 0x60000400020020 nid: 0@lo remote: 0x29b04b8d251ce815 expref: 6260 pid: 20676 timeout: 3470 lvb_type: 0 LustreError: 6796:0:(client.c:1282:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff880267de6940 x1828476705510912/t0(0) o104->lustre-OST0002@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' uid:4294967295 gid:4294967295 LustreError: 6796:0:(client.c:1282:ptlrpc_import_delay_req()) Skipped 42 previous similar messages LustreError: 24855:0:(ldlm_lockd.c:1492:ldlm_handle_enqueue()) ### lock on destroyed export ffff8802e541ae98 ns: filter-lustre-OST0002_UUID lock: ffff8802726f52c0/0x29b04b8d253b3710 lrc: 3/0,0 mode: PW/PW res: [0x340000403:0x4e1:0x0].0x0 rrc: 2 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x50000000020000 nid: 0@lo remote: 0x29b04b8d253b3709 expref: 5836 pid: 24855 timeout: 0 lvb_type: 0 LustreError: lustre-OST0002-osc-ffff8802e80e8008: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-OST0002-osc-ffff8802e80e8008: Connection to lustre-OST0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 18716:0:(ldlm_lockd.c:2594:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1743773990 with bad export cookie 3003984021002782198 LustreError: 18716:0:(ldlm_lockd.c:2594:ldlm_cancel_handler()) Skipped 3 previous similar messages LustreError: lustre-OST0002-osc-ffff8802e80e8008: This client was evicted by lustre-OST0002; in progress operations using this service will fail. Lustre: lustre-OST0003-osc-MDT0001: update sequence from 0x380000401 to 0x380000403 Lustre: 3984:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x280000404:0x58a9:0x0]// may get corrupted (rc -108) Lustre: 3998:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x280000405:0x3d71:0x0]/ may get corrupted (rc -108) Lustre: 3998:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x280000404:0x585d:0x0]/ may get corrupted (rc -108) Lustre: 3987:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x280000404:0x58be:0x0]// may get corrupted (rc -108) Lustre: 3984:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x280000404:0x58af:0x0]/ may get corrupted (rc -108) Lustre: 3996:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x280000404:0x5825:0x0]/ may get corrupted (rc -108) Lustre: 3996:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x200000404:0x83fb:0x0]/ may get corrupted (rc -108) Lustre: 3982:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x240000406:0x53f:0x0]// may get corrupted (rc -108) Lustre: 3982:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x200000404:0x8440:0x0]/ may get corrupted (rc -108) Lustre: 3991:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x280000404:0x58c9:0x0]/ may get corrupted (rc -108) Lustre: 3991:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x200000405:0x14f5:0x0]/ may get corrupted (rc -108) LustreError: 8699:0:(ldlm_resource.c:982:ldlm_resource_complain()) lustre-OST0002-osc-ffff8802e80e8008: namespace resource [0x340000401:0x6bcf:0x0].0x0 (ffff8802e3c5ad40) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 8699:0:(ldlm_resource.c:982:ldlm_resource_complain()) Skipped 31 previous similar messages Lustre: lustre-OST0002-osc-ffff8802e80e8008: Connection restored to 192.168.123.107@tcp (at 0@lo) Lustre: 9136:0:(out_handler.c:911:out_tx_end()) lustre-MDT0002-osd: error during execution of #2 from /home/green/git/lustre-release/lustre/ptlrpc/../../lustre/target/out_handler.c:555: rc = -17 LustreError: 9136:0:(out_lib.c:1168:out_tx_index_delete_undo()) lustre-MDT0002-osd: Oops, can not rollback index_delete yet: rc = -524 LustreError: lustre-OST0003-osc-ffff8802e7a75d28: operation ost_setattr to node 0@lo failed: rc = -107 LustreError: Skipped 1 previous similar message LustreError: lustre-OST0003-osc-ffff8802e7a75d28: This client was evicted by lustre-OST0003; in progress operations using this service will fail. Lustre: 3994:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x280000405:0x3e1f:0x0]/ may get corrupted (rc -108) Lustre: 3982:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x240000406:0x56a:0x0]/ may get corrupted (rc -108) Lustre: 3998:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x240000407:0x4cd:0x0]/ may get corrupted (rc -108) Lustre: 3998:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x280000404:0x589b:0x0]/ may get corrupted (rc -108) LustreError: 10277:0:(ldlm_resource.c:982:ldlm_resource_complain()) lustre-OST0003-osc-ffff8802e7a75d28: namespace resource [0x380000401:0x7d5f:0x0].0x0 (ffff8802699c4a40) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 10277:0:(ldlm_resource.c:982:ldlm_resource_complain()) Skipped 733 previous similar messages Lustre: lustre-OST0003-osc-MDT0002: update sequence from 0x380000400 to 0x380000404 Lustre: lustre-OST0002-osc-MDT0001: update sequence from 0x340000401 to 0x340000404 Lustre: lustre-OST0001-osc-MDT0001: update sequence from 0x300000401 to 0x300000404 Lustre: lustre-OST0000-osc-MDT0002: update sequence from 0x2c0000400 to 0x2c0000404 Lustre: lustre-OST0003-osc-MDT0000: update sequence from 0x380000402 to 0x380000405 Lustre: lustre-OST0002-osc-MDT0000: update sequence from 0x340000402 to 0x340000405 LustreError: 10209:0:(ldlm_lockd.c:252:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0003_UUID lock: ffff88029702ad40/0x29b04b8d2543f2d3 lrc: 3/0,0 mode: PW/PW res: [0x380000402:0x72d3:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 262144->18446744073709551615) gid 0 flags: 0x60000480020020 nid: 0@lo remote: 0x29b04b8d2543e7f1 expref: 4981 pid: 12497 timeout: 3604 lvb_type: 0 LustreError: 10209:0:(ldlm_lockd.c:252:expired_lock_main()) Skipped 1 previous similar message LustreError: 7711:0:(client.c:1282:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff880212b48f40 x1828476758368384/t0(0) o104->lustre-OST0003@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' uid:4294967295 gid:4294967295 LustreError: 7711:0:(client.c:1282:ptlrpc_import_delay_req()) Skipped 17 previous similar messages LustreError: lustre-OST0003-osc-ffff8802e7a75d28: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 306 previous similar messages Lustre: lustre-OST0003-osc-ffff8802e7a75d28: Connection to lustre-OST0003 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 1 previous similar message LustreError: lustre-OST0003-osc-ffff8802e7a75d28: This client was evicted by lustre-OST0003; in progress operations using this service will fail. Lustre: 3989:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x200000405:0x1611:0x0]/ may get corrupted (rc -108) Lustre: 3990:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x200000405:0x161b:0x0]// may get corrupted (rc -108) Lustre: 3990:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x200000405:0x1645:0x0]// may get corrupted (rc -108) Lustre: 3989:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x240000407:0x5e5:0x0]// may get corrupted (rc -108) LustreError: 16635:0:(ldlm_resource.c:982:ldlm_resource_complain()) lustre-OST0003-osc-ffff8802e7a75d28: namespace resource [0x380000402:0x7dee:0x0].0x0 (ffff8802233c0cc0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 16635:0:(ldlm_resource.c:982:ldlm_resource_complain()) Skipped 14 previous similar messages LustreError: lustre-OST0000-osc-ffff8802e80e8008: This client was evicted by lustre-OST0000; in progress operations using this service will fail. Lustre: lustre-OST0001-osc-MDT0000: update sequence from 0x300000402 to 0x300000405 Lustre: 3983:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x240000406:0x64f:0x0]/ may get corrupted (rc -108) Lustre: 3983:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x200000404:0x85a9:0x0]/ may get corrupted (rc -108) Lustre: 3997:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x240000406:0x5f2:0x0]/ may get corrupted (rc -108) Lustre: 3997:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x240000406:0x631:0x0]/ may get corrupted (rc -108) Lustre: 3987:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x200000404:0x85cb:0x0]/ may get corrupted (rc -108) Lustre: 3985:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x200000404:0x85c5:0x0]/ may get corrupted (rc -108) Lustre: 3984:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x280000404:0x5a3d:0x0]/ may get corrupted (rc -108) Lustre: lustre-OST0003-osc-ffff8802e7a75d28: Connection restored to 192.168.123.107@tcp (at 0@lo) Lustre: Skipped 1 previous similar message Lustre: 18526:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 6/24/0, destroy: 1/4/0 Lustre: 18526:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 1635 previous similar messages Lustre: 18526:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1577/1577/0, xattr_set: 2365/22184/0 Lustre: 18526:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 1635 previous similar messages Lustre: 18526:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 28/157/0, punch: 0/0/0, quota 1/3/0 Lustre: 18526:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 1635 previous similar messages Lustre: 18526:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 7/118/0, delete: 2/5/0 Lustre: 18526:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 1635 previous similar messages Lustre: 18526:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 1/1/0, ref_del: 2/2/0 Lustre: 18526:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 1635 previous similar messages Lustre: lustre-OST0000-osc-MDT0000: update sequence from 0x2c0000402 to 0x2c0000405 Lustre: lustre-OST0000: haven't heard from client 35173acf-7059-4a76-9534-ed215803bb38 (at 0@lo) in 31 seconds. I think it's dead, and I am evicting it. exp ffff880260206678, cur 1743774159 expire 1743774129 last 1743774128 LustreError: 17082:0:(ldlm_resource.c:982:ldlm_resource_complain()) lustre-OST0000-osc-ffff8802e80e8008: namespace resource [0x2c0000403:0x615d:0x0].0x0 (ffff880207257240) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 17082:0:(ldlm_resource.c:982:ldlm_resource_complain()) Skipped 917 previous similar messages Lustre: lustre-OST0000-osc-ffff8802e80e8008: Connection restored to 192.168.123.107@tcp (at 0@lo) LustreError: lustre-OST0000-osc-ffff8802e80e8008: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 1 previous similar message Lustre: lustre-OST0000-osc-ffff8802e80e8008: Connection to lustre-OST0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 1 previous similar message LustreError: lustre-OST0000-osc-ffff8802e80e8008: This client was evicted by lustre-OST0000; in progress operations using this service will fail. Lustre: 18489:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0000: opcode 0: before 515 < left 1203, rollback = 0 Lustre: 18489:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 1931 previous similar messages Lustre: 29546:0:(client.c:2346:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1743774255/real 1743774255] req@ffff880233cc1440 x1828476816600832/t0(0) o101->lustre-MDT0001-mdc-ffff8802e80e8008@0@lo:12/10 lens 376/27792 e 0 to 1 dl 1743774310 ref 2 fl Rpc:XQr/202/ffffffff rc 0/-1 job:'lfs.0' uid:0 gid:0 Lustre: lustre-MDT0001: Client 35173acf-7059-4a76-9534-ed215803bb38 (at 0@lo) reconnecting Lustre: lustre-OST0000-osc-MDT0001: update sequence from 0x2c0000403 to 0x2c0000406 LustreError: 14217:0:(mdt_open.c:1302:mdt_cross_open()) lustre-MDT0001: [0x240000406:0xa4e:0x0] doesn't exist!: rc = -14 Lustre: lustre-OST0002-osc-MDT0002: update sequence from 0x340000403 to 0x340000406 Lustre: 14097:0:(out_handler.c:911:out_tx_end()) lustre-MDT0000-osd: error during execution of #2 from /home/green/git/lustre-release/lustre/ptlrpc/../../lustre/target/out_handler.c:585: rc = -2 LustreError: 14097:0:(out_lib.c:1168:out_tx_index_delete_undo()) lustre-MDT0000-osd: Oops, can not rollback index_delete yet: rc = -524 LustreError: 24212:0:(mdt_open.c:1302:mdt_cross_open()) lustre-MDT0002: [0x280000405:0x442b:0x0] doesn't exist!: rc = -14 LustreError: 10209:0:(ldlm_lockd.c:252:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0002_UUID lock: ffff880096c2d680/0x29b04b8d25827049 lrc: 3/0,0 mode: PW/PW res: [0x340000405:0x4ebd:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000480020020 nid: 0@lo remote: 0x29b04b8d25823635 expref: 24172 pid: 8556 timeout: 3892 lvb_type: 0 LustreError: 10209:0:(ldlm_lockd.c:252:expired_lock_main()) Skipped 1 previous similar message LustreError: lustre-OST0002-osc-ffff8802e7a75d28: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: 3619:0:(ldlm_lockd.c:2594:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1743774412 with bad export cookie 3003984021002781960 Lustre: lustre-OST0002-osc-ffff8802e7a75d28: Connection to lustre-OST0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 1 previous similar message LustreError: 6801:0:(client.c:1282:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88005b765040 x1828476878374784/t0(0) o104->lustre-OST0002@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' uid:4294967295 gid:4294967295 LustreError: 6801:0:(client.c:1282:ptlrpc_import_delay_req()) Skipped 23 previous similar messages LustreError: lustre-OST0002-osc-ffff8802e7a75d28: This client was evicted by lustre-OST0002; in progress operations using this service will fail. LustreError: 6796:0:(ldlm_lockd.c:1492:ldlm_handle_enqueue()) ### lock on destroyed export ffff8802e82f6678 ns: filter-lustre-OST0002_UUID lock: ffff88027ed76d00/0x29b04b8d25b181cd lrc: 3/0,0 mode: PW/PW res: [0x340000405:0x62a1:0x0].0x0 rrc: 2 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x50000080020000 nid: 0@lo remote: 0x29b04b8d25b181c6 expref: 23651 pid: 6796 timeout: 0 lvb_type: 0 Lustre: 3997:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x240000407:0xa6d:0x0]/ may get corrupted (rc -108) Lustre: 3994:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x240000407:0xa23:0x0]/ may get corrupted (rc -108) Lustre: 3998:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x240000407:0xa28:0x0]/ may get corrupted (rc -108) Lustre: 3997:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x200000405:0x1a60:0x0]// may get corrupted (rc -108) Lustre: 3994:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x280000405:0x4436:0x0]// may get corrupted (rc -108) LustreError: 18489:0:(mdt_open.c:1302:mdt_cross_open()) lustre-MDT0002: [0x280000405:0x442b:0x0] doesn't exist!: rc = -14 LustreError: 9309:0:(ldlm_resource.c:982:ldlm_resource_complain()) lustre-OST0002-osc-ffff8802e7a75d28: namespace resource [0x340000405:0x7150:0x0].0x0 (ffff880209dea340) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 9309:0:(ldlm_resource.c:982:ldlm_resource_complain()) Skipped 35 previous similar messages Lustre: lustre-OST0002-osc-ffff8802e7a75d28: Connection restored to 192.168.123.107@tcp (at 0@lo) Lustre: Skipped 2 previous similar messages LustreError: 18117:0:(mdt_open.c:1302:mdt_cross_open()) lustre-MDT0002: [0x280000405:0x442b:0x0] doesn't exist!: rc = -14 LustreError: 14185:0:(mdt_open.c:1302:mdt_cross_open()) lustre-MDT0002: [0x280000405:0x442b:0x0] doesn't exist!: rc = -14 LustreError: 18545:0:(mdt_open.c:1302:mdt_cross_open()) lustre-MDT0002: [0x280000405:0x442b:0x0] doesn't exist!: rc = -14 Lustre: lustre-OST0003-osc-MDT0000: update sequence from 0x380000405 to 0x380000406 Lustre: lustre-OST0002-osc-MDT0000: update sequence from 0x340000405 to 0x340000407 Lustre: 18510:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff8801f675ee40 x1828476906640512/t4295194858(0) o101->35173acf-7059-4a76-9534-ed215803bb38@0@lo:123/0 lens 376/48232 e 0 to 0 dl 1743774548 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 Lustre: 18510:0:(mdt_recovery.c:102:mdt_req_from_lrd()) Skipped 50 previous similar messages LustreError: 14190:0:(mdt_open.c:1302:mdt_cross_open()) lustre-MDT0002: [0x280000405:0x442b:0x0] doesn't exist!: rc = -14 LustreError: 14190:0:(mdt_open.c:1302:mdt_cross_open()) Skipped 3 previous similar messages Lustre: lustre-OST0002-osc-MDT0001: update sequence from 0x340000404 to 0x340000408 LustreError: 10209:0:(ldlm_lockd.c:252:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0003_UUID lock: ffff88006046b880/0x29b04b8d25c55862 lrc: 3/0,0 mode: PW/PW res: [0x380000405:0x7f00:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000080030020 nid: 0@lo remote: 0x29b04b8d25c4cfb8 expref: 10681 pid: 7497 timeout: 4093 lvb_type: 0 LustreError: 10595:0:(client.c:1282:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88005a20c140 x1828476952433024/t0(0) o104->lustre-OST0003@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' uid:4294967295 gid:4294967295 LustreError: 10595:0:(client.c:1282:ptlrpc_import_delay_req()) Skipped 88 previous similar messages LustreError: lustre-OST0003-osc-ffff8802e80e8008: This client was evicted by lustre-OST0003; in progress operations using this service will fail. LustreError: 17098:0:(mdt_open.c:1302:mdt_cross_open()) lustre-MDT0002: [0x280000405:0x442b:0x0] doesn't exist!: rc = -14 LustreError: 17098:0:(mdt_open.c:1302:mdt_cross_open()) Skipped 10 previous similar messages Lustre: 3998:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x200000404:0x8bc9:0x0]/ may get corrupted (rc -108) Lustre: 3986:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x280000404:0x61b0:0x0]// may get corrupted (rc -108) Lustre: 3982:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x280000404:0x60eb:0x0]/ may get corrupted (rc -108) Lustre: 3982:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x240000406:0xdd4:0x0]// may get corrupted (rc -108) Lustre: 3984:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x280000404:0x6197:0x0]// may get corrupted (rc -108) Lustre: 3984:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x240000406:0xdee:0x0]/ may get corrupted (rc -108) Lustre: 3989:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x240000406:0xe17:0x0]// may get corrupted (rc -108) Lustre: 3998:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x240000406:0xcc0:0x0]/ may get corrupted (rc -108) LustreError: 24031:0:(ldlm_resource.c:982:ldlm_resource_complain()) lustre-OST0003-osc-ffff8802e80e8008: namespace resource [0x380000405:0x7f0b:0x0].0x0 (ffff88024c18b9c0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 24031:0:(ldlm_resource.c:982:ldlm_resource_complain()) Skipped 1 previous similar message Lustre: lustre-OST0003-osc-MDT0001: update sequence from 0x380000403 to 0x380000407 Lustre: lustre-OST0003: haven't heard from client 35173acf-7059-4a76-9534-ed215803bb38 (at 0@lo) in 31 seconds. I think it's dead, and I am evicting it. exp ffff880063e8c138, cur 1743774645 expire 1743774615 last 1743774614 Lustre: lustre-OST0001-osc-MDT0002: update sequence from 0x300000403 to 0x300000406 LustreError: lustre-OST0003-osc-ffff8802e80e8008: This client was evicted by lustre-OST0003; in progress operations using this service will fail. Lustre: lustre-OST0000-osc-MDT0000: update sequence from 0x2c0000405 to 0x2c0000407 Lustre: lustre-OST0001-osc-MDT0000: update sequence from 0x300000405 to 0x300000407 LustreError: 14199:0:(mdt_open.c:1302:mdt_cross_open()) lustre-MDT0002: [0x280000405:0x442b:0x0] doesn't exist!: rc = -14 Lustre: 11276:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 6/24/0, destroy: 1/4/1 Lustre: 11276:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 1703 previous similar messages Lustre: 11276:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 984/984/0, xattr_set: 1476/13776/0 Lustre: 11276:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 1703 previous similar messages Lustre: 11276:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 28/157/12, punch: 0/0/0, quota 1/3/0 Lustre: 11276:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 1703 previous similar messages Lustre: 11276:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 6/102/0, delete: 2/5/1 Lustre: 11276:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 1703 previous similar messages Lustre: 11276:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 1/1/0 Lustre: 11276:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 1703 previous similar messages LustreError: 14199:0:(mdt_open.c:1302:mdt_cross_open()) Skipped 15 previous similar messages Lustre: lustre-OST0001-osc-MDT0001: update sequence from 0x300000404 to 0x300000408 Lustre: lustre-OST0003-osc-MDT0002: update sequence from 0x380000404 to 0x380000408 hrtimer: interrupt took 4918919 ns Lustre: 14175:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0002: opcode 0: before 513 < left 2609, rollback = 0 Lustre: 14175:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 1418 previous similar messages Lustre: lustre-OST0000-osc-MDT0002: update sequence from 0x2c0000404 to 0x2c0000408 LustreError: 10209:0:(ldlm_lockd.c:252:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0002_UUID lock: ffff8802721e8040/0x29b04b8d261116cb lrc: 3/0,0 mode: PW/PW res: [0x340000408:0x18e1:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000480020020 nid: 0@lo remote: 0x29b04b8d261116c4 expref: 17618 pid: 9026 timeout: 4396 lvb_type: 0 LustreError: 7472:0:(client.c:1282:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff880051e81e40 x1828477063875968/t0(0) o104->lustre-OST0002@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' uid:4294967295 gid:4294967295 LustreError: 9175:0:(ldlm_lockd.c:2594:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1743774917 with bad export cookie 3003984021043239090 LustreError: 9175:0:(ldlm_lockd.c:2594:ldlm_cancel_handler()) Skipped 1 previous similar message Lustre: lustre-OST0002-osc-ffff8802e80e8008: Connection to lustre-OST0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 2 previous similar messages LustreError: 7472:0:(client.c:1282:ptlrpc_import_delay_req()) Skipped 3 previous similar messages LustreError: lustre-OST0002-osc-ffff8802e80e8008: This client was evicted by lustre-OST0002; in progress operations using this service will fail. LustreError: lustre-OST0002-osc-ffff8802e80e8008: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 10 previous similar messages Lustre: 3985:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x200000404:0x9182:0x0]/ may get corrupted (rc -108) Lustre: 3985:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x200000404:0x91ac:0x0]/ may get corrupted (rc -108) Lustre: 3983:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x280000404:0x6630:0x0]/ may get corrupted (rc -108) Lustre: 3988:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x240000406:0x1288:0x0]// may get corrupted (rc -108) Lustre: 3988:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x200000404:0x91dd:0x0]// may get corrupted (rc -108) Lustre: 3987:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x280000405:0x4d35:0x0]/ may get corrupted (rc -108) Lustre: 3987:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x240000406:0x126e:0x0]/ may get corrupted (rc -108) Lustre: 3988:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x280000404:0x6656:0x0]/ may get corrupted (rc -108) Lustre: lustre-OST0002-osc-ffff8802e80e8008: Connection restored to 192.168.123.107@tcp (at 0@lo) Lustre: Skipped 2 previous similar messages Lustre: lustre-OST0002-osc-MDT0002: update sequence from 0x340000406 to 0x340000409 LustreError: 30531:0:(mdt_open.c:1302:mdt_cross_open()) lustre-MDT0002: [0x280000405:0x442b:0x0] doesn't exist!: rc = -14 LustreError: 30531:0:(mdt_open.c:1302:mdt_cross_open()) Skipped 14 previous similar messages LustreError: 8045:0:(client.c:1282:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff880061f25540 x1828477118740736/t0(0) o104->lustre-OST0003@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' uid:4294967295 gid:4294967295 LustreError: 3619:0:(ldlm_lockd.c:2594:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1743775081 with bad export cookie 3003984021055062391 LustreError: 3619:0:(ldlm_lockd.c:2594:ldlm_cancel_handler()) Skipped 1 previous similar message LustreError: 8045:0:(client.c:1282:ptlrpc_import_delay_req()) Skipped 58 previous similar messages LustreError: lustre-OST0003-osc-ffff8802e80e8008: This client was evicted by lustre-OST0003; in progress operations using this service will fail. LustreError: 8556:0:(ldlm_lockd.c:1492:ldlm_handle_enqueue()) ### lock on destroyed export ffff880209322548 ns: filter-lustre-OST0003_UUID lock: ffff880074131300/0x29b04b8d2648c0da lrc: 3/0,0 mode: --/PW res: [0x380000406:0x648c:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x50000000020000 nid: 0@lo remote: 0x29b04b8d2648be72 expref: 11720 pid: 8556 timeout: 0 lvb_type: 0 LustreError: lustre-OST0002-osc-ffff8802e7a75d28: This client was evicted by lustre-OST0002; in progress operations using this service will fail. LustreError: lustre-OST0003-osc-ffff8802e7a75d28: This client was evicted by lustre-OST0003; in progress operations using this service will fail. Lustre: 3996:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x280000404:0x685d:0x0]// may get corrupted (rc -108) Lustre: 3990:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x200000404:0x9419:0x0]/ may get corrupted (rc -108) Lustre: 3988:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x280000404:0x6844:0x0]/ may get corrupted (rc -108) Lustre: 3990:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x200000404:0x944c:0x0]/ may get corrupted (rc -108) Lustre: 3996:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x280000404:0x67cd:0x0]/ may get corrupted (rc -108) Lustre: 3994:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x280000405:0x4fda:0x0]// may get corrupted (rc -108) Lustre: 3982:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x240000407:0x1362:0x0]// may get corrupted (rc -108) Lustre: 3991:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x280000404:0x6743:0x0]/ may get corrupted (rc -108) Lustre: 3991:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x280000405:0x4f5e:0x0]/ may get corrupted (rc -108) Lustre: 3997:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x280000405:0x5006:0x0]// may get corrupted (rc -108) Lustre: 3997:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x280000405:0x5019:0x0]// may get corrupted (rc -108) Lustre: 3989:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x280000405:0x4f63:0x0]/ may get corrupted (rc -108) Lustre: 3987:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x280000404:0x684c:0x0]/ may get corrupted (rc -108) Lustre: 3993:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x240000406:0x1445:0x0]// may get corrupted (rc -108) Lustre: 3993:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x280000405:0x4f2e:0x0]/ may get corrupted (rc -108) Lustre: 3992:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x200000404:0x943e:0x0]/ may get corrupted (rc -108) Lustre: 3992:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x280000405:0x4ed1:0x0]/ may get corrupted (rc -108) Lustre: 3991:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x280000405:0x4e0f:0x0]/ may get corrupted (rc -108) Lustre: 3985:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x280000405:0x4c7a:0x0]/ may get corrupted (rc -108) Lustre: 3994:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x280000405:0x4ae3:0x0]/ may get corrupted (rc -108) Lustre: 3992:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x280000405:0x4940:0x0]/ may get corrupted (rc -108) Lustre: 3982:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x280000405:0x4b0c:0x0]/ may get corrupted (rc -108) Lustre: 3998:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x280000405:0x4b37:0x0]/ may get corrupted (rc -108) Lustre: 3991:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x200000405:0x1e4d:0x0]/ may get corrupted (rc -108) Lustre: 3997:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x280000404:0x648d:0x0]/ may get corrupted (rc -108) Lustre: 3991:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x280000405:0x4c11:0x0]/ may get corrupted (rc -108) Lustre: 3985:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x280000404:0x65ea:0x0]/ may get corrupted (rc -108) Lustre: 3985:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x280000405:0x4c07:0x0]/ may get corrupted (rc -108) Lustre: 3985:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x280000405:0x4c8d:0x0]/ may get corrupted (rc -108) Lustre: 3985:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x280000404:0x6654:0x0]/ may get corrupted (rc -108) Lustre: 3985:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x280000405:0x4d35:0x0]/ may get corrupted (rc -108) Lustre: 3988:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x240000407:0x1304:0x0]/ may get corrupted (rc -108) Lustre: 3991:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x280000405:0x4f49:0x0]/ may get corrupted (rc -108) Lustre: 3987:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x280000405:0x4cf7:0x0]/ may get corrupted (rc -108) Lustre: 3987:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x280000405:0x4ea1:0x0]/ may get corrupted (rc -108) Lustre: 3987:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x280000405:0x4920:0x0]/ may get corrupted (rc -108) Lustre: 3987:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x280000405:0x4e2a:0x0]/ may get corrupted (rc -108) Lustre: 3996:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x280000405:0x4b04:0x0]/ may get corrupted (rc -108) Lustre: 3983:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x280000405:0x4c3b:0x0]/ may get corrupted (rc -108) Lustre: 3993:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x280000405:0x49e3:0x0]/ may get corrupted (rc -108) Lustre: 3993:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x280000404:0x67d3:0x0]/ may get corrupted (rc -108) Lustre: 3993:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x280000405:0x4fd2:0x0]/ may get corrupted (rc -108) Lustre: 3997:0:(llite_lib.c:4120:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.107@tcp:/lustre/fid: [0x280000405:0x4c1e:0x0]/ may get corrupted (rc -108) LustreError: 23233:0:(ldlm_resource.c:982:ldlm_resource_complain()) lustre-OST0002-osc-ffff8802e7a75d28: namespace resource [0x340000409:0x567:0x0].0x0 (ffff8801ec644cc0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 23233:0:(ldlm_resource.c:982:ldlm_resource_complain()) Skipped 298 previous similar messages Lustre: 14199:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff880095e68040 x1828477123894528/t4295216589(0) o101->d30d3017-d020-4b0b-8fed-ed809fbf4bbb@0@lo:689/0 lens 376/48232 e 0 to 0 dl 1743775114 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 Lustre: 14199:0:(mdt_recovery.c:102:mdt_req_from_lrd()) Skipped 41 previous similar messages | Link to test |
racer test 1: racer on clients: centos-0.localnet DURATION=2700 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 PGD 2e75a7067 PUD 2a1749067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) lustre(OE) osp(OE) ofd(OE) lod(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) jbd2 mbcache lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) crc32_generic lnet(OE) dm_flakey dm_mod libcfs(OE) crc_t10dif crct10dif_generic crct10dif_common virtio_balloon virtio_console pcspkr i2c_piix4 ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm drm_panel_orientation_quirks ata_piix floppy i2c_core virtio_blk serio_raw libata CPU: 9 PID: 2784 Comm: ll_sa_2078 Kdump: loaded Tainted: P OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.16.0-3.module_el8.7.0+1218+f626c2ff 04/01/2014 task: ffff8800a55149d0 ti: ffff8802b66d0000 task.ti: ffff8802b66d0000 RIP: 0010:[<ffffffff813f502c>] [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 RSP: 0018:ffff8802b66d3d78 EFLAGS: 00010296 RAX: ffff8802715b8a90 RBX: 0000000000000008 RCX: 0000000100260020 RDX: 0000000000000026 RSI: ffff8802715b8dc0 RDI: 0000000000000008 RBP: ffff8802b66d3d88 R08: ffff8802bc84b858 R09: 0000000000000000 R10: ffff8802bc84cc38 R11: ffff8802bc84b858 R12: 0000000000000000 R13: ffff8800a1820010 R14: ffff8802715b8dc0 R15: ffff8802bc84b858 FS: 0000000000000000(0000) GS:ffff880331c40000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 00000002adcc6000 CR4: 00000000000007e0 Call Trace: [<ffffffffa10d6bf1>] ll_statahead_thread+0xdb1/0x2000 [lustre] [<ffffffff810df6e2>] ? pick_next_task_fair+0x702/0x8a0 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e329e>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffff810c834d>] ? finish_task_switch+0x5d/0x1b0 [<ffffffffa10d5e40>] ? ll_statahead_handle.constprop.28+0x1c0/0x1c0 [lustre] [<ffffffff810ba114>] kthread+0xe4/0xf0 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 | Lustre: lustre-OST0001-osc-ffff8802e8cfa548: disconnect after 20s idle LustreError: 11091:0:(mdt_reint.c:2523:mdt_reint_migrate()) lustre-MDT0001: migrate [0x240000403:0x1:0x0]/1 failed: rc = -114 LustreError: 11091:0:(mdt_reint.c:2523:mdt_reint_migrate()) lustre-MDT0002: migrate [0x280000403:0x1:0x0]/17 failed: rc = -114 LustreError: 17809:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e8cfa548: inode [0x280000403:0xe:0x0] mdc close failed: rc = -116 LustreError: 11087:0:(mdt_reint.c:2523:mdt_reint_migrate()) lustre-MDT0002: migrate [0x280000403:0x1:0x0]/3 failed: rc = -114 LustreError: 11087:0:(mdt_reint.c:2523:mdt_reint_migrate()) Skipped 2 previous similar messages Lustre: 11087:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 512 < left 610, rollback = 2 Lustre: 11087:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 3/12/0, destroy: 0/0/0 Lustre: 11087:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 2/2/0, xattr_set: 8/610/0 Lustre: 11087:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 5/55/0, punch: 0/0/0, quota 1/3/0 Lustre: 11087:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 12/275/3, delete: 0/0/0 Lustre: 11087:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 6/6/0, ref_del: 0/0/0 Lustre: 11090:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 513 < left 788, rollback = 2 Lustre: 11090:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 11 previous similar messages Lustre: 11090:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 3/12/0, destroy: 0/0/0 Lustre: 11090:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 11 previous similar messages Lustre: 11090:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 2/2/0, xattr_set: 10/788/0 Lustre: 11090:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 11 previous similar messages Lustre: 11090:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 5/55/0, punch: 0/0/0, quota 4/54/0 Lustre: 11090:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 11 previous similar messages Lustre: 11090:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 10/203/2, delete: 0/0/0 Lustre: 11090:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 11 previous similar messages Lustre: 11090:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 4/4/0, ref_del: 0/0/0 Lustre: 11090:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 11 previous similar messages Lustre: 11087:0:(mdd_dir.c:4701:mdd_migrate_object()) lustre-MDD0001: [0x240000403:0x1:0x0]/0 is open, migrate only dentry LustreError: 11087:0:(mdt_reint.c:2523:mdt_reint_migrate()) lustre-MDT0001: migrate [0x240000403:0x1:0x0]/0 failed: rc = -114 LustreError: 11087:0:(mdt_reint.c:2523:mdt_reint_migrate()) Skipped 1 previous similar message Lustre: 19500:0:(mdd_dir.c:4701:mdd_migrate_object()) lustre-MDD0001: [0x240000403:0x1:0x0]/18 is open, migrate only dentry Lustre: 19500:0:(mdd_dir.c:4701:mdd_migrate_object()) Skipped 1 previous similar message Lustre: 17160:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff8802e77de940 x1828417540260096/t4294968001(0) o101->6f521cf2-7402-4a46-80c8-0ce00df9943b@0@lo:162/0 lens 376/816 e 0 to 0 dl 1743714942 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 Lustre: 19500:0:(mdd_dir.c:4701:mdd_migrate_object()) lustre-MDD0002: [0x280000403:0x1:0x0]/4 is open, migrate only dentry LustreError: 20025:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a7ba2548: inode [0x280000404:0x40:0x0] mdc close failed: rc = -116 LustreError: 20025:0:(file.c:247:ll_close_inode_openhandle()) Skipped 1 previous similar message 2[20075]: segfault at 8 ip 00007f5e5ac917e8 sp 00007ffcee5c5aa0 error 4 in ld-2.17.so[7f5e5ac86000+22000] Lustre: 17891:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 502 < left 922, rollback = 2 Lustre: 17891:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 9 previous similar messages Lustre: 17891:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 3/12/7, destroy: 2/8/0 Lustre: 17891:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 9 previous similar messages Lustre: 17891:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 4/4/0, xattr_set: 13/922/0 Lustre: 17891:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 9 previous similar messages Lustre: 17891:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 5/55/0, punch: 0/0/0, quota 1/3/0 Lustre: 17891:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 9 previous similar messages Lustre: 17891:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 18/371/4, delete: 10/16/1 Lustre: 17891:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 9 previous similar messages Lustre: 17891:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 11/11/0, ref_del: 10/10/1 Lustre: 17891:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 9 previous similar messages LustreError: 20075:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a7ba2548: inode [0x200000404:0x47:0x0] mdc close failed: rc = -13 LustreError: 11091:0:(mdt_reint.c:2523:mdt_reint_migrate()) lustre-MDT0002: migrate [0x240000403:0x11:0x0]/13 failed: rc = -2 LustreError: 11091:0:(mdt_reint.c:2523:mdt_reint_migrate()) Skipped 9 previous similar messages Lustre: 21297:0:(mdd_dir.c:4701:mdd_migrate_object()) lustre-MDD0000: [0x200000403:0x2:0x0]/3 is open, migrate only dentry LustreError: 17118:0:(mdd_object.c:3902:mdd_close()) lustre-MDD0001: failed to get lu_attr of [0x240000404:0x14:0x0]: rc = -2 LustreError: 19923:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e8cfa548: inode [0x240000404:0x14:0x0] mdc close failed: rc = -2 LustreError: 19923:0:(file.c:247:ll_close_inode_openhandle()) Skipped 1 previous similar message Lustre: dir [0x280000404:0x51:0x0] stripe 3 readdir failed: -2, directory is partially accessed! LustreError: 19633:0:(mdd_object.c:3902:mdd_close()) lustre-MDD0001: failed to get lu_attr of [0x240000404:0x14:0x0]: rc = -2 19[22969]: segfault at 8 ip 00007f5c6e25d7e8 sp 00007ffdf834ded0 error 4 in ld-2.17.so[7f5c6e252000+22000] Lustre: 21297:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 509 < left 610, rollback = 2 Lustre: 21297:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 45 previous similar messages Lustre: 21297:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 3/12/2, destroy: 0/0/0 Lustre: 21297:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 45 previous similar messages Lustre: 21297:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 2/2/0, xattr_set: 8/610/0 Lustre: 21297:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 45 previous similar messages Lustre: 21297:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 5/55/0, punch: 0/0/0, quota 1/3/0 Lustre: 21297:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 45 previous similar messages Lustre: 21297:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 12/275/4, delete: 0/0/0 Lustre: 21297:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 45 previous similar messages Lustre: 21297:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 6/6/0, ref_del: 0/0/0 Lustre: 21297:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 45 previous similar messages Lustre: 21160:0:(mdd_dir.c:4701:mdd_migrate_object()) lustre-MDD0002: [0x280000403:0x1:0x0]/0 is open, migrate only dentry Lustre: 21160:0:(mdd_dir.c:4701:mdd_migrate_object()) Skipped 2 previous similar messages LustreError: 17891:0:(mdt_reint.c:2523:mdt_reint_migrate()) lustre-MDT0000: migrate [0x200000403:0x1:0x0]/15 failed: rc = -114 LustreError: 17891:0:(mdt_reint.c:2523:mdt_reint_migrate()) Skipped 14 previous similar messages Lustre: 17891:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 510 < left 638, rollback = 2 Lustre: 17891:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 116 previous similar messages Lustre: 17891:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 3/12/2, destroy: 0/0/0 Lustre: 17891:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 116 previous similar messages Lustre: 17891:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 4/4/0, xattr_set: 10/638/0 Lustre: 17891:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 116 previous similar messages Lustre: 17891:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 5/55/0, punch: 0/0/0, quota 1/3/0 Lustre: 17891:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 116 previous similar messages Lustre: 17891:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 14/287/3, delete: 0/0/0 Lustre: 17891:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 116 previous similar messages Lustre: 17891:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 8/8/0, ref_del: 0/0/0 Lustre: 17891:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 116 previous similar messages Lustre: dir [0x280000404:0xb0:0x0] stripe 3 readdir failed: -2, directory is partially accessed! Lustre: Skipped 2 previous similar messages Lustre: 19375:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff8802dca4cb40 x1828417544213760/t4294969281(0) o101->6f521cf2-7402-4a46-80c8-0ce00df9943b@0@lo:176/0 lens 376/816 e 0 to 0 dl 1743714956 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 Lustre: 19375:0:(mdt_recovery.c:102:mdt_req_from_lrd()) Skipped 2 previous similar messages Lustre: 17920:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0002: opcode 7: before 516 < left 618, rollback = 7 LustreError: 25318:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a7ba2548: inode [0x280000404:0x3:0x0] mdc close failed: rc = -116 LustreError: 25318:0:(file.c:247:ll_close_inode_openhandle()) Skipped 4 previous similar messages 19[26536]: segfault at 8 ip 00007f50b017d7e8 sp 00007ffdccb3de70 error 4 in ld-2.17.so[7f50b0172000+22000] Lustre: dir [0x280000404:0x7e:0x0] stripe 3 readdir failed: -2, directory is partially accessed! Lustre: 19484:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0x1c1:0x0] with magic=0xbd60bd0 LustreError: 27076:0:(lcommon_cl.c:179:cl_file_inode_init()) lustre: failed to initialize cl_object [0x280000403:0xb0:0x0]: rc = -5 LustreError: 27076:0:(llite_lib.c:3698:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 Lustre: 11090:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 508 < left 816, rollback = 2 Lustre: 11090:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 53 previous similar messages Lustre: 11090:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 3/12/4, destroy: 0/0/0 Lustre: 11090:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 55 previous similar messages Lustre: 11090:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 4/4/0, xattr_set: 12/816/0 Lustre: 11090:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 55 previous similar messages Lustre: 11090:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 5/55/0, punch: 0/0/0, quota 7/105/0 Lustre: 11090:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 55 previous similar messages Lustre: 11090:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 14/287/3, delete: 0/0/0 Lustre: 11090:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 55 previous similar messages Lustre: 11090:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 8/8/0, ref_del: 0/0/0 Lustre: 11090:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 55 previous similar messages Lustre: 24777:0:(mdd_dir.c:4701:mdd_migrate_object()) lustre-MDD0000: [0x200000404:0x34:0x0]/17 is open, migrate only dentry LustreError: 27019:0:(lcommon_cl.c:179:cl_file_inode_init()) lustre: failed to initialize cl_object [0x280000403:0xf7:0x0]: rc = -5 LustreError: 27019:0:(lcommon_cl.c:179:cl_file_inode_init()) Skipped 1 previous similar message LustreError: 27019:0:(llite_lib.c:3698:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 27019:0:(llite_lib.c:3698:ll_prep_inode()) Skipped 1 previous similar message Lustre: 17160:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000403:0x13c:0x0] with magic=0xbd60bd0 Lustre: 17160:0:(lod_lov.c:1417:lod_parse_striping()) Skipped 1 previous similar message LustreError: 11075:0:(mdd_object.c:3902:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000403:0x35:0x0]: rc = -2 LustreError: 23662:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e8cfa548: inode [0x200000403:0x35:0x0] mdc close failed: rc = -2 LustreError: 23662:0:(file.c:247:ll_close_inode_openhandle()) Skipped 1 previous similar message Lustre: 14321:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000403:0xfd:0x0] with magic=0xbd60bd0 Lustre: 14321:0:(lod_lov.c:1417:lod_parse_striping()) Skipped 1 previous similar message Lustre: 17263:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000403:0xe2:0x0] with magic=0xbd60bd0 Lustre: 17263:0:(lod_lov.c:1417:lod_parse_striping()) Skipped 3 previous similar messages Lustre: 18988:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0002: opcode 7: before 516 < left 618, rollback = 7 Lustre: 18988:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 1 previous similar message Lustre: 17182:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000404:0x17e:0x0] with magic=0xbd60bd0 Lustre: 17182:0:(lod_lov.c:1417:lod_parse_striping()) Skipped 17 previous similar messages Lustre: 12513:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0000: opcode 7: before 516 < left 618, rollback = 7 Lustre: 12513:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 7 previous similar messages LustreError: 11090:0:(mdt_reint.c:2523:mdt_reint_migrate()) lustre-MDT0000: migrate [0x200000403:0x2:0x0]/14 failed: rc = -114 LustreError: 11090:0:(mdt_reint.c:2523:mdt_reint_migrate()) Skipped 6 previous similar messages 13[32640]: segfault at 8 ip 00007faf2d4767e8 sp 00007ffdd49bb9d0 error 4 in ld-2.17.so[7faf2d46b000+22000] Lustre: 21160:0:(mdd_dir.c:4701:mdd_migrate_object()) lustre-MDD0000: [0x200000403:0x1:0x0]/15 is open, migrate only dentry 9[5500]: segfault at 8 ip 00007fc8f58397e8 sp 00007fff281eee10 error 4 in ld-2.17.so[7fc8f582e000+22000] LustreError: 5469:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e8cfa548: inode [0x240000404:0x255:0x0] mdc close failed: rc = -116 LustreError: 5469:0:(file.c:247:ll_close_inode_openhandle()) Skipped 3 previous similar messages 9[5713]: segfault at 8 ip 00007fdf876e07e8 sp 00007ffca0cd1970 error 4 in ld-2.17.so[7fdf876d5000+22000] 7[6274]: segfault at 8 ip 00007f24431227e8 sp 00007ffd664a9950 error 4 in ld-2.17.so[7f2443117000+22000] Lustre: 23362:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 509 < left 1428, rollback = 2 Lustre: 23362:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 59 previous similar messages Lustre: 23362:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 3/12/3, destroy: 1/4/0 Lustre: 23362:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 69 previous similar messages Lustre: 23362:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 2/2/0, xattr_set: 17/1428/0 Lustre: 23362:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 69 previous similar messages Lustre: 23362:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 5/55/0, punch: 0/0/0, quota 7/129/0 Lustre: 23362:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 69 previous similar messages Lustre: 23362:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 14/287/3, delete: 2/5/0 Lustre: 23362:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 69 previous similar messages Lustre: 23362:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 7/7/0, ref_del: 2/2/0 Lustre: 23362:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 69 previous similar messages Lustre: 12513:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0000: opcode 7: before 515 < left 618, rollback = 7 Lustre: 12513:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 1 previous similar message LustreError: 20186:0:(mdd_object.c:3902:mdd_close()) lustre-MDD0002: failed to get lu_attr of [0x280000403:0x6f:0x0]: rc = -2 9[8890]: segfault at 8 ip 00007fb1fd66b7e8 sp 00007fff27ed8c60 error 4 in ld-2.17.so[7fb1fd660000+22000] Lustre: 12514:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 516 < left 606, rollback = 7 Lustre: 12514:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 5 previous similar messages LustreError: 24777:0:(mdd_dir.c:4621:mdd_migrate_cmd_check()) lustre-MDD0000: '13' migration was interrupted, run 'lfs migrate -m 2 -c 2 -H crush 13' to finish migration: rc = -1 LustreError: 24777:0:(mdd_dir.c:4621:mdd_migrate_cmd_check()) lustre-MDD0001: '13' migration was interrupted, run 'lfs migrate -m 2 -c 2 -H crush 13' to finish migration: rc = -1 LustreError: 11428:0:(mdd_dir.c:4621:mdd_migrate_cmd_check()) lustre-MDD0002: '19' migration was interrupted, run 'lfs migrate -m 1 -c 2 -H crush 19' to finish migration: rc = -1 18[12209]: segfault at 8 ip 00007f4e7bad47e8 sp 00007ffd7b3d4500 error 4 in ld-2.17.so[7f4e7bac9000+22000] LustreError: 23362:0:(mdt_reint.c:2523:mdt_reint_migrate()) lustre-MDT0001: migrate [0x240000404:0x311:0x0]/0 failed: rc = -114 LustreError: 23362:0:(mdt_reint.c:2523:mdt_reint_migrate()) Skipped 37 previous similar messages Lustre: 17820:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 618, rollback = 7 Lustre: 17820:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 3 previous similar messages LustreError: 23362:0:(mdd_dir.c:4621:mdd_migrate_cmd_check()) lustre-MDD0001: '14' migration was interrupted, run 'lfs migrate -m 2 -c 2 -H crush 14' to finish migration: rc = -1 1[14516]: segfault at 0 ip 0000000000403e5f sp 00007ffc3a79c4e0 error 6 in 1[400000+6000] 3[14787]: segfault at 8 ip 00007fdbf48cd7e8 sp 00007fffea3857d0 error 4 in ld-2.17.so[7fdbf48c2000+22000] 7[14352]: segfault at 8 ip 00007fd7965137e8 sp 00007ffe666435c0 error 4 in ld-2.17.so[7fd796508000+22000] Lustre: 12842:0:(mdd_dir.c:4701:mdd_migrate_object()) lustre-MDD0002: [0x280000404:0x22e:0x0]/17 is open, migrate only dentry Lustre: 12842:0:(mdd_dir.c:4701:mdd_migrate_object()) Skipped 11 previous similar messages Lustre: 18364:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 18364:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 377 previous similar messages Lustre: 18364:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 18364:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 376 previous similar messages Lustre: 18364:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 18364:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 377 previous similar messages Lustre: 18364:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 18364:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 377 previous similar messages Lustre: 18364:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 18364:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 377 previous similar messages Lustre: 11090:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 500 < left 699, rollback = 2 Lustre: 11090:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 365 previous similar messages LustreError: 15067:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a7ba2548: inode [0x240000404:0x381:0x0] mdc close failed: rc = -116 LustreError: 15067:0:(file.c:247:ll_close_inode_openhandle()) Skipped 14 previous similar messages LustreError: 21297:0:(mdd_dir.c:4621:mdd_migrate_cmd_check()) lustre-MDD0000: '9' migration was interrupted, run 'lfs migrate -m 1 -c 3 -H crush 9' to finish migration: rc = -1 LustreError: 21297:0:(mdd_dir.c:4621:mdd_migrate_cmd_check()) Skipped 1 previous similar message Lustre: 26223:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0x649:0x0] with magic=0xbd60bd0 Lustre: 26223:0:(lod_lov.c:1417:lod_parse_striping()) Skipped 1 previous similar message LustreError: 15706:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8802e8cfa548: cannot apply new layout on [0x200000404:0x537:0x0] : rc = -5 LustreError: 15706:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x200000404:0x537:0x0] error -5. Lustre: 18364:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0000: opcode 7: before 516 < left 618, rollback = 7 Lustre: 18364:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 3 previous similar messages LustreError: 23362:0:(mdd_dir.c:4621:mdd_migrate_cmd_check()) lustre-MDD0001: '10' migration was interrupted, run 'lfs migrate -m 0 -c 3 -H crush 10' to finish migration: rc = -1 LustreError: 23362:0:(mdd_dir.c:4621:mdd_migrate_cmd_check()) Skipped 1 previous similar message LustreError: 19125:0:(lcommon_cl.c:179:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000404:0x46e:0x0]: rc = -5 LustreError: 19125:0:(llite_lib.c:3698:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 13[21661]: segfault at 8 ip 00007f71ca6027e8 sp 00007fff15010420 error 4 in ld-2.17.so[7f71ca5f7000+22000] 18[21548]: segfault at 8 ip 00007f253aa897e8 sp 00007ffc28eaf630 error 4 in ld-2.17.so[7f253aa7e000+22000] Lustre: 21297:0:(mdt_reint.c:2443:mdt_reint_migrate()) lustre-MDT0002: [0x280000403:0x1:0x0]/11 is open, migrate only dentry Lustre: 12514:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0002: opcode 7: before 516 < left 618, rollback = 7 Lustre: 12514:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 9 previous similar messages LustreError: 23644:0:(lcommon_cl.c:179:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000404:0x46e:0x0]: rc = -5 LustreError: 23644:0:(llite_lib.c:3698:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 Lustre: mdt00_012: service thread pid 17260 was inactive for 40.276 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 17260, comm: mdt00_012 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock_try+0xa0/0x250 [mdt] [<0>] mdt_object_open_lock+0x6b9/0xc10 [mdt] [<0>] mdt_open_by_fid_lock+0x4ec/0xc10 [mdt] [<0>] mdt_reint_open+0x9d8/0x2d70 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x73c/0xbb0 [mdt] [<0>] mdt_intent_open+0x93/0x480 [mdt] [<0>] mdt_intent_opc.constprop.74+0x211/0xc60 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x397/0x980 [ptlrpc] [<0>] ldlm_handle_enqueue+0x547/0x18d0 [ptlrpc] [<0>] tgt_enqueue+0x68/0x240 [ptlrpc] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe LustreError: 24631:0:(mdt_reint.c:2523:mdt_reint_migrate()) lustre-MDT0001: migrate [0x200000404:0x49e:0x0]/16 failed: rc = -114 LustreError: 24631:0:(mdt_reint.c:2523:mdt_reint_migrate()) Skipped 80 previous similar messages Lustre: 11073:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000403:0x4ed:0x0] with magic=0xbd60bd0 Lustre: 11073:0:(lod_lov.c:1417:lod_parse_striping()) Skipped 1 previous similar message Lustre: 24777:0:(mdd_dir.c:4701:mdd_migrate_object()) lustre-MDD0000: [0x200000403:0x1:0x0]/8 is open, migrate only dentry Lustre: 24777:0:(mdd_dir.c:4701:mdd_migrate_object()) Skipped 16 previous similar messages 7[25806]: segfault at 8 ip 00007fe999a417e8 sp 00007fffbe491fe0 error 4 in ld-2.17.so[7fe999a36000+22000] Lustre: 23362:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 507 < left 2346, rollback = 2 Lustre: 23362:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 394 previous similar messages Lustre: 23362:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 3/12/4, destroy: 1/4/0 Lustre: 23362:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 410 previous similar messages Lustre: 23362:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 4/4/0, xattr_set: 29/2346/0 Lustre: 23362:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 410 previous similar messages Lustre: 23362:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 5/55/0, punch: 0/0/0, quota 7/129/0 Lustre: 23362:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 410 previous similar messages Lustre: 23362:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 17/335/4, delete: 3/6/0 Lustre: 23362:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 410 previous similar messages Lustre: 23362:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 10/10/0, ref_del: 2/2/0 Lustre: 23362:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 410 previous similar messages LustreError: 22944:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a7ba2548: inode [0x280000404:0x365:0x0] mdc close failed: rc = -116 LustreError: 22944:0:(file.c:247:ll_close_inode_openhandle()) Skipped 10 previous similar messages LustreError: 26508:0:(lcommon_cl.c:179:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000403:0x7e1:0x0]: rc = -5 LustreError: 26508:0:(llite_lib.c:3698:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 19633:0:(mdd_object.c:3902:mdd_close()) lustre-MDD0002: failed to get lu_attr of [0x280000403:0x4ae:0x0]: rc = -2 LustreError: 19633:0:(mdd_object.c:3902:mdd_close()) Skipped 2 previous similar messages LustreError: 26518:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8800a7ba2548: cannot apply new layout on [0x280000403:0x564:0x0] : rc = -5 LustreError: 26518:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x280000403:0x564:0x0] error -5. LustreError: 29951:0:(lcommon_cl.c:179:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000404:0x662:0x0]: rc = -5 LustreError: 29951:0:(lcommon_cl.c:179:cl_file_inode_init()) Skipped 2 previous similar messages LustreError: 29951:0:(llite_lib.c:3698:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 29951:0:(llite_lib.c:3698:ll_prep_inode()) Skipped 2 previous similar messages Lustre: 26191:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000404:0x730:0x0] with magic=0xbd60bd0 Lustre: 26191:0:(lod_lov.c:1417:lod_parse_striping()) Skipped 3 previous similar messages LustreError: 10821:0:(ldlm_lockd.c:252:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0000_UUID lock: ffff8802bc8e8b80/0x4762192c4e9b8378 lrc: 3/0,0 mode: PW/PW res: [0x200000403:0x53a:0x0].0x0 bits 0x4/0x0 rrc: 5 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0x4762192c4e9b836a expref: 516 pid: 17004 timeout: 317 lvb_type: 0 LustreError: 17933:0:(ldlm_lockd.c:2594:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1743715119 with bad export cookie 5143701402497145161 LustreError: lustre-MDT0000-mdc-ffff8802e8cfa548: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-MDT0000-mdc-ffff8802e8cfa548: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: lustre-MDT0000-mdc-ffff8802e8cfa548: This client was evicted by lustre-MDT0000; in progress operations using this service will fail. LustreError: 26223:0:(client.c:1282:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8800a3eb1940 x1828417580373248/t0(0) o104->lustre-MDT0000@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' uid:4294967295 gid:4294967295 LustreError: 4502:0:(file.c:6138:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000403:0x2:0x0] error: rc = -108 LustreError: 4502:0:(file.c:6138:ll_inode_revalidate_fini()) Skipped 2 previous similar messages Lustre: mdt00_012: service thread pid 17260 completed after 100.530s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: lustre-MDT0000-mdc-ffff8802e8cfa548: Connection restored to (at 0@lo) LustreError: 2082:0:(lcommon_cl.c:179:cl_file_inode_init()) lustre: failed to initialize cl_object [0x280000403:0x57f:0x0]: rc = -5 LustreError: 2082:0:(lcommon_cl.c:179:cl_file_inode_init()) Skipped 5 previous similar messages LustreError: 2082:0:(llite_lib.c:3698:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 2082:0:(llite_lib.c:3698:ll_prep_inode()) Skipped 5 previous similar messages 16[11495]: segfault at 8 ip 00007fabc8ffb7e8 sp 00007fff70173de0 error 4 in ld-2.17.so[7fabc8ff0000+22000] Lustre: 18364:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0002: opcode 7: before 516 < left 558, rollback = 7 Lustre: 18364:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 17 previous similar messages 11[12900]: segfault at 8 ip 00007f434e47a7e8 sp 00007ffc82b095f0 error 4 in ld-2.17.so[7f434e46f000+22000] LustreError: 15353:0:(statahead.c:2447:start_statahead_thread()) lustre: unsupported statahead pattern 0X0. 3[16973]: segfault at 8 ip 00007f0525fd07e8 sp 00007ffc4d3702f0 error 4 in ld-2.17.so[7f0525fc5000+22000] 19[16993]: segfault at 8 ip 00007f1c87f6b7e8 sp 00007ffdd7409d60 error 4 in ld-2.17.so[7f1c87f60000+22000] LustreError: 22669:0:(lcommon_cl.c:179:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000405:0x1d8:0x0]: rc = -5 LustreError: 22669:0:(llite_lib.c:3698:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 15[27048]: segfault at 8 ip 00007f809255c7e8 sp 00007ffdfe6f4150 error 4 in ld-2.17.so[7f8092551000+22000] 5[27323]: segfault at 0 ip 00000000004043f0 sp 00007ffd881c1e68 error 6 in 5[400000+6000] 15[27681]: segfault at 8 ip 00007fc13160d7e8 sp 00007ffe34160120 error 4 in ld-2.17.so[7fc131602000+22000] 10[28780]: segfault at 8 ip 00007f350cea47e8 sp 00007ffc48040700 error 4 in ld-2.17.so[7f350ce99000+22000] Lustre: 18364:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 18364:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 293 previous similar messages Lustre: 18364:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/1, xattr_set: 2/15/0 Lustre: 18364:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 293 previous similar messages Lustre: 18364:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 18364:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 293 previous similar messages Lustre: 18364:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 18364:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 293 previous similar messages Lustre: 18364:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 18364:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 293 previous similar messages Lustre: mdt_io00_011: service thread pid 11428 was inactive for 40.069 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 11428, comm: mdt_io00_011 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_pdo_lock+0x369/0x750 [mdt] [<0>] mdt_parent_lock+0x76/0x2a0 [mdt] [<0>] mdt_reint_migrate+0xdf1/0x24b0 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x73c/0xbb0 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 30675, comm: mdt_out00_004 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_pdo_lock+0x369/0x750 [mdt] [<0>] mdt_parent_lock+0x76/0x2a0 [mdt] [<0>] mdt_getattr_name_lock+0x1844/0x2cc0 [mdt] [<0>] mdt_intent_getattr+0x2cc/0x4e0 [mdt] [<0>] mdt_intent_opc.constprop.74+0x211/0xc60 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x397/0x980 [ptlrpc] [<0>] ldlm_handle_enqueue+0x547/0x18d0 [ptlrpc] [<0>] mdt_batch_getattr+0x6c/0x190 [mdt] [<0>] mdt_batch+0x465/0x195b [mdt] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: mdt_io00_009: service thread pid 24631 was inactive for 40.069 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: mdt_io00_000: service thread pid 11087 was inactive for 40.040 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 1 previous similar message Lustre: mdt_io00_006: service thread pid 21297 was inactive for 40.062 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. LustreError: 10821:0:(ldlm_lockd.c:252:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0000_UUID lock: ffff8802adf6f480/0x4762192c4eb61969 lrc: 3/0,0 mode: PW/PW res: [0x200000405:0x28e:0x0].0x0 bits 0x4/0x0 rrc: 6 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0x4762192c4eb6195b expref: 243 pid: 13801 timeout: 487 lvb_type: 0 LustreError: lustre-MDT0000-mdc-ffff8802e8cfa548: operation mds_close to node 0@lo failed: rc = -107 Lustre: lustre-MDT0000-mdc-ffff8802e8cfa548: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: Skipped 2 previous similar messages LustreError: lustre-MDT0000-mdc-ffff8802e8cfa548: This client was evicted by lustre-MDT0000; in progress operations using this service will fail. LustreError: 11428:0:(mdt_reint.c:2523:mdt_reint_migrate()) lustre-MDT0000: migrate [0x200000403:0x1:0x0]/17 failed: rc = -114 LustreError: 11428:0:(mdt_reint.c:2523:mdt_reint_migrate()) Skipped 83 previous similar messages LustreError: 17182:0:(ldlm_lockd.c:1492:ldlm_handle_enqueue()) ### lock on destroyed export ffff8802db006678 ns: mdt-lustre-MDT0000_UUID lock: ffff8802c3b56580/0x4762192c4eb637ae lrc: 3/0,0 mode: PR/PR res: [0x200000403:0x1:0x0].0x0 bits 0x13/0x0 rrc: 21 type: IBT gid 0 flags: 0x50200000000000 nid: 0@lo remote: 0x4762192c4eb63799 expref: 15 pid: 17182 timeout: 0 lvb_type: 0 Lustre: mdt_io00_011: service thread pid 11428 completed after 100.124s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_010: service thread pid 24777 completed after 99.372s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_out00_004: service thread pid 30675 completed after 99.566s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 27104:0:(file.c:6138:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000403:0x1:0x0] error: rc = -5 LustreError: 27987:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e8cfa548: inode [0x200000405:0x2b6:0x0] mdc close failed: rc = -108 LustreError: 27987:0:(file.c:247:ll_close_inode_openhandle()) Skipped 54 previous similar messages LustreError: 27917:0:(mdc_request.c:1464:mdc_read_page()) lustre-MDT0000-mdc-ffff8802e8cfa548: [0x200000401:0xe:0x0] lock enqueue fails: rc = -108 Lustre: dir [0x240000403:0x905:0x0] stripe 2 readdir failed: -108, directory is partially accessed! Lustre: Skipped 5 previous similar messages LustreError: 27104:0:(file.c:6138:ll_inode_revalidate_fini()) Skipped 58 previous similar messages Lustre: mdt_io00_009: service thread pid 24631 completed after 99.419s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: 11087:0:(mdd_dir.c:4701:mdd_migrate_object()) lustre-MDD0002: [0x280000403:0x1:0x0]/0 is open, migrate only dentry Lustre: 11087:0:(mdd_dir.c:4701:mdd_migrate_object()) Skipped 11 previous similar messages Lustre: mdt_io00_000: service thread pid 11087 completed after 96.836s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: lustre-MDT0000-mdc-ffff8802e8cfa548: Connection restored to (at 0@lo) Lustre: mdt_io00_006: service thread pid 21297 completed after 91.142s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_005: service thread pid 21160 completed after 90.600s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: 12513:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0000: opcode 7: before 516 < left 618, rollback = 7 Lustre: 12513:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 25 previous similar messages 13[31424]: segfault at 0 ip 0000000000403e5f sp 00007fffeb5a6310 error 6 in 13[400000+6000] 19[2126]: segfault at 1c ip 00000000004017b0 sp 00007ffc9e5ca090 error 6 in 19[400000+6000] 13[2138]: segfault at 8 ip 00007ff96120b7e8 sp 00007fffa49e2e00 error 4 in ld-2.17.so[7ff961200000+22000] 15[6111]: segfault at 8 ip 00007f36c2f1b7e8 sp 00007ffd7a6763b0 error 4 in ld-2.17.so[7f36c2f10000+22000] 10[18620]: segfault at 8 ip 00007f0e2796a7e8 sp 00007ffc2e80e1b0 error 4 in ld-2.17.so[7f0e2795f000+22000] 6[21117]: segfault at 8 ip 00007f064b57c7e8 sp 00007ffece98b9a0 error 4 in ld-2.17.so[7f064b571000+22000] 4[29619]: segfault at 8 ip 00007f19f7ee27e8 sp 00007ffd861a95f0 error 4 in ld-2.17.so[7f19f7ed7000+22000] 1[31930]: segfault at 0 ip (null) sp 00007fff80934a38 error 14 in 1[400000+6000] 4[2702]: segfault at 8 ip 00007f24d1a977e8 sp 00007ffe8f732d90 error 4 in ld-2.17.so[7f24d1a8c000+22000] 9[3171]: segfault at 0 ip (null) sp 00007fff8a729f38 error 14 in 9[400000+6000] 6[8433]: segfault at 0 ip (null) sp 00007ffe011c3c68 error 14 in 6[400000+6000] 0[9026]: segfault at 406000 ip 0000000000406000 sp 00007ffeae8f01f8 error 14 in 12 (deleted)[606000+1000] traps: 1[9377] general protection ip:404739 sp:7ffdaf147678 error:0 in 1[400000+6000] 12[14406]: segfault at 8 ip 00007fb46dfbb7e8 sp 00007ffd23aac0f0 error 4 in ld-2.17.so[7fb46dfb0000+22000] Lustre: 19436:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000404:0x16b7:0x0] with magic=0xbd60bd0 Lustre: 19436:0:(lod_lov.c:1417:lod_parse_striping()) Skipped 5 previous similar messages 13[16156]: segfault at 8 ip 00007fc914d8a7e8 sp 00007ffeab23dea0 error 4 in ld-2.17.so[7fc914d7f000+22000] 6[21111]: segfault at 8 ip 00007f5a684e77e8 sp 00007fffaabef2a0 error 4 in ld-2.17.so[7f5a684dc000+22000] 16[24831]: segfault at 8 ip 00007ff219dec7e8 sp 00007fff8f2dddd0 error 4 in ld-2.17.so[7ff219de1000+22000] Lustre: 17820:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 17820:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 35 previous similar messages Lustre: 17820:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 17820:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 35 previous similar messages Lustre: 17820:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 17820:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 35 previous similar messages Lustre: 17820:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 17820:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 35 previous similar messages Lustre: 17820:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 17820:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 35 previous similar messages 17[31829]: segfault at 8 ip 00007f044f5597e8 sp 00007fff97cadbd0 error 4 in ld-2.17.so[7f044f54e000+22000] 1[1429]: segfault at 8 ip 00007f8e6bd7b7e8 sp 00007ffcbfa2c920 error 4 in ld-2.17.so[7f8e6bd70000+22000] 0[6972]: segfault at 8 ip 00007ff2056107e8 sp 00007ffd93e82680 error 4 in ld-2.17.so[7ff205605000+22000] INFO: task mkdir:19602 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mkdir D ffff88009a739280 11616 19602 16512 0x00000080 Call Trace: [<ffffffff817e19d9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df7ea>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa1d>] mutex_lock+0x2d/0x40 [<ffffffff81256f65>] filename_create+0x85/0x180 [<ffffffff81257c94>] ? getname_flags+0xc4/0x1a0 [<ffffffff81257f41>] user_path_create+0x41/0x60 [<ffffffff81259286>] SyS_mkdirat+0x46/0xe0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff81259339>] SyS_mkdir+0x19/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a INFO: task mkdir:20435 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mkdir D ffff8802d9aea980 12512 20435 16338 0x00000080 Call Trace: [<ffffffff817e19d9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df7ea>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa1d>] mutex_lock+0x2d/0x40 [<ffffffff81256f65>] filename_create+0x85/0x180 [<ffffffff81257c94>] ? getname_flags+0xc4/0x1a0 [<ffffffff81257f41>] user_path_create+0x41/0x60 [<ffffffff81259286>] SyS_mkdirat+0x46/0xe0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff81259339>] SyS_mkdir+0x19/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a INFO: task mkdir:20508 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mkdir D ffff8802b5a0a140 13056 20508 16365 0x00000080 Call Trace: [<ffffffff817e19d9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df7ea>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa1d>] mutex_lock+0x2d/0x40 [<ffffffff81256f65>] filename_create+0x85/0x180 [<ffffffff81257c94>] ? getname_flags+0xc4/0x1a0 [<ffffffff81257f41>] user_path_create+0x41/0x60 [<ffffffff81259286>] SyS_mkdirat+0x46/0xe0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff81259339>] SyS_mkdir+0x19/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a INFO: task mkdir:20519 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mkdir D ffff8802adcadb00 11616 20519 16554 0x00000080 Call Trace: [<ffffffff817e19d9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df7ea>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa1d>] mutex_lock+0x2d/0x40 [<ffffffff81256f65>] filename_create+0x85/0x180 [<ffffffff81257c94>] ? getname_flags+0xc4/0x1a0 [<ffffffff81257f41>] user_path_create+0x41/0x60 [<ffffffff81259286>] SyS_mkdirat+0x46/0xe0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff81259339>] SyS_mkdir+0x19/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a INFO: task mkdir:29165 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mkdir D ffff8802b8918608 13056 29165 16584 0x00000080 Call Trace: [<ffffffff817e19d9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df7ea>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa1d>] mutex_lock+0x2d/0x40 [<ffffffff81256f65>] filename_create+0x85/0x180 [<ffffffff81257c94>] ? getname_flags+0xc4/0x1a0 [<ffffffff81257f41>] user_path_create+0x41/0x60 [<ffffffff81259286>] SyS_mkdirat+0x46/0xe0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff81259339>] SyS_mkdir+0x19/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a INFO: task mkdir:31534 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mkdir D ffff8802b8a7a980 13056 31534 16444 0x00000080 Call Trace: [<ffffffff817e19d9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df7ea>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa1d>] mutex_lock+0x2d/0x40 [<ffffffff81256f65>] filename_create+0x85/0x180 [<ffffffff81257c94>] ? getname_flags+0xc4/0x1a0 [<ffffffff81257f41>] user_path_create+0x41/0x60 [<ffffffff81259286>] SyS_mkdirat+0x46/0xe0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff81259339>] SyS_mkdir+0x19/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a INFO: task mkdir:31803 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mkdir D ffff8802e2bf10c0 13056 31803 16348 0x00000080 Call Trace: [<ffffffff817e19d9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df7ea>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa1d>] mutex_lock+0x2d/0x40 [<ffffffff81256f65>] filename_create+0x85/0x180 [<ffffffff81257c94>] ? getname_flags+0xc4/0x1a0 [<ffffffff81257f41>] user_path_create+0x41/0x60 [<ffffffff81259286>] SyS_mkdirat+0x46/0xe0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff81259339>] SyS_mkdir+0x19/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a Lustre: 11428:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 511 < left 983, rollback = 2 Lustre: 11428:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 255 previous similar messages 1[12807]: segfault at 406000 ip 0000000000406000 sp 00007fffaaa29098 error 14 in 1[606000+1000] 17[14226]: segfault at 0 ip (null) sp 00007ffdf2466208 error 14 in 17[400000+6000] Lustre: 21297:0:(mdd_dir.c:4701:mdd_migrate_object()) lustre-MDD0002: [0x280000403:0x1:0x0]/9 is open, migrate only dentry Lustre: 21297:0:(mdd_dir.c:4701:mdd_migrate_object()) Skipped 8 previous similar messages LustreError: 21297:0:(mdt_reint.c:2523:mdt_reint_migrate()) lustre-MDT0002: migrate [0x280000403:0x1:0x0]/9 failed: rc = -114 LustreError: 21297:0:(mdt_reint.c:2523:mdt_reint_migrate()) Skipped 59 previous similar messages 1[22950]: segfault at 0 ip (null) sp 00007fffd4b557c8 error 14 in 1[400000+6000] 13[24406]: segfault at 0 ip 0000000000403e5f sp 00007ffd6d1fc960 error 6 in 13[400000+6000] 2[29435]: segfault at 8 ip 00007fc2ecf1c7e8 sp 00007ffc1b2fc120 error 4 in ld-2.17.so[7fc2ecf11000+22000] LustreError: 17242:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e8cfa548: inode [0x280000404:0x112d:0x0] mdc close failed: rc = -116 LustreError: 17242:0:(file.c:247:ll_close_inode_openhandle()) Skipped 50 previous similar messages Lustre: 18364:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 515 < left 618, rollback = 7 Lustre: 18364:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 63 previous similar messages LustreError: 10821:0:(ldlm_lockd.c:252:expired_lock_main()) ### lock callback timer expired after 101s: evicting client at 0@lo ns: mdt-lustre-MDT0001_UUID lock: ffff8802a054a200/0x4762192c4eeb6a51 lrc: 3/0,0 mode: PR/PR res: [0x240000403:0x138e:0x0].0x0 bits 0x1b/0x0 rrc: 5 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0x4762192c4eeb69ef expref: 743 pid: 11819 timeout: 886 lvb_type: 0 LustreError: 17587:0:(ldlm_lockd.c:1492:ldlm_handle_enqueue()) ### lock on destroyed export ffff8802e7c88008 ns: mdt-lustre-MDT0001_UUID lock: ffff88029d144780/0x4762192c4eed012a lrc: 3/0,0 mode: PR/PR res: [0x240000403:0x1:0x0].0x0 bits 0x12/0x0 rrc: 20 type: IBT gid 0 flags: 0x50200000000000 nid: 0@lo remote: 0x4762192c4eed011c expref: 22 pid: 17587 timeout: 0 lvb_type: 0 LustreError: 17587:0:(ldlm_lockd.c:1492:ldlm_handle_enqueue()) Skipped 6 previous similar messages LustreError: lustre-MDT0001-mdc-ffff8800a7ba2548: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 7 previous similar messages Lustre: lustre-MDT0001-mdc-ffff8800a7ba2548: Connection to lustre-MDT0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: lustre-MDT0001-mdc-ffff8800a7ba2548: This client was evicted by lustre-MDT0001; in progress operations using this service will fail. LustreError: 26066:0:(file.c:6138:ll_inode_revalidate_fini()) lustre: revalidate FID [0x240000403:0x1:0x0] error: rc = -5 LustreError: 26066:0:(file.c:6138:ll_inode_revalidate_fini()) Skipped 23 previous similar messages LustreError: 4195:0:(llite_lib.c:1996:ll_md_setattr()) md_setattr fails: rc = -108 Lustre: lustre-MDT0001-mdc-ffff8800a7ba2548: Connection restored to (at 0@lo) 8[17154]: segfault at c0cfcf ip 00000000004047ee sp 00007ffd0fc2a378 error 6 in 8[400000+6000] Lustre: 17182:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000404:0x2228:0x0] with magic=0xbd60bd0 Lustre: 17182:0:(lod_lov.c:1417:lod_parse_striping()) Skipped 1 previous similar message Lustre: dir [0x240000405:0xaf:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 4 previous similar messages LustreError: 24036:0:(statahead.c:2447:start_statahead_thread()) lustre: unsupported statahead pattern 0X0. 7[8833]: segfault at 0 ip 00000000004016cd sp 00007fffa06fe2a0 error 6 in 7[400000+6000] 0[8973]: segfault at 406000 ip 0000000000406000 sp 00007ffe4fd33af8 error 14 in 0[606000+1000] 16[29004]: segfault at 8 ip 00007f1b61d337e8 sp 00007ffd0ed022b0 error 4 in ld-2.17.so[7f1b61d28000+22000] 19[30960]: segfault at 40452e ip 0000000000405390 sp 00007fffb22858d8 error 7 in 19[400000+6000] 12[5537]: segfault at 0 ip (null) sp 00007fff310a9cb8 error 14 in 12 (deleted)[400000+6000] INFO: task mkdir:24474 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mkdir D ffff8802c9703d58 11616 24474 16785 0x00000080 Call Trace: [<ffffffff817e19d9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df7ea>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa1d>] mutex_lock+0x2d/0x40 [<ffffffff81256f65>] filename_create+0x85/0x180 [<ffffffff81257c94>] ? getname_flags+0xc4/0x1a0 [<ffffffff81257f41>] user_path_create+0x41/0x60 [<ffffffff81259286>] SyS_mkdirat+0x46/0xe0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff81259339>] SyS_mkdir+0x19/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a INFO: task mkdir:24850 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mkdir D ffff8800a5e8c240 11616 24850 16670 0x00000080 Call Trace: [<ffffffff817e19d9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df7ea>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa1d>] mutex_lock+0x2d/0x40 [<ffffffff81256f65>] filename_create+0x85/0x180 [<ffffffff81257c94>] ? getname_flags+0xc4/0x1a0 [<ffffffff81257f41>] user_path_create+0x41/0x60 [<ffffffff81259286>] SyS_mkdirat+0x46/0xe0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff81259339>] SyS_mkdir+0x19/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a INFO: task mkdir:25613 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mkdir D ffff8802d9a173c0 12512 25613 16342 0x00000080 Call Trace: [<ffffffff817e19d9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df7ea>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa1d>] mutex_lock+0x2d/0x40 [<ffffffff81256f65>] filename_create+0x85/0x180 [<ffffffff81257c94>] ? getname_flags+0xc4/0x1a0 [<ffffffff81257f41>] user_path_create+0x41/0x60 [<ffffffff81259286>] SyS_mkdirat+0x46/0xe0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff81259339>] SyS_mkdir+0x19/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a 12[13076]: segfault at 406000 ip 0000000000406000 sp 00007ffcd6059fe8 error 14 in 12[606000+1000] 17[18948]: segfault at 0 ip (null) sp 00007ffd3e6ea8c8 error 14 in 17[400000+6000] 17[18953]: segfault at 0 ip (null) sp 00007ffe611168e8 error 14 in 17[400000+6000] 6[20839]: segfault at 8 ip 00007fd4c42077e8 sp 00007ffd79cc0070 error 4 in ld-2.17.so[7fd4c41fc000+22000] Lustre: 12514:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 12514:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 321 previous similar messages Lustre: 12514:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 12514:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 321 previous similar messages Lustre: 12514:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 12514:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 321 previous similar messages Lustre: 12514:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 12514:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 321 previous similar messages Lustre: 12514:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 12514:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 321 previous similar messages Lustre: 17604:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000406:0x262a:0x0] with magic=0xbd60bd0 Lustre: 17604:0:(lod_lov.c:1417:lod_parse_striping()) Skipped 3 previous similar messages 1[8370]: segfault at 0 ip 0000000000403e5f sp 00007ffdf5af63c0 error 6 in 1[400000+6000] Lustre: 24777:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 509 < left 1072, rollback = 2 Lustre: 24777:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 211 previous similar messages LustreError: 11428:0:(mdt_reint.c:2523:mdt_reint_migrate()) lustre-MDT0002: migrate [0x280000403:0x1:0x0]/11 failed: rc = -114 LustreError: 11428:0:(mdt_reint.c:2523:mdt_reint_migrate()) Skipped 120 previous similar messages Lustre: 19500:0:(mdd_dir.c:4701:mdd_migrate_object()) lustre-MDD0002: [0x280000403:0x1:0x0]/8 is open, migrate only dentry Lustre: 19500:0:(mdd_dir.c:4701:mdd_migrate_object()) Skipped 19 previous similar messages 8[23204]: segfault at 8 ip 00007f45b5dcf7e8 sp 00007ffdf4165fd0 error 4 in ld-2.17.so[7f45b5dc4000+22000] LustreError: 21160:0:(mdd_dir.c:4621:mdd_migrate_cmd_check()) lustre-MDD0001: '5' migration was interrupted, run 'lfs migrate -m 1 -c 3 -H crush 5' to finish migration: rc = -1 Lustre: dir [0x200000404:0x35b4:0x0] stripe 2 readdir failed: -2, directory is partially accessed! Lustre: Skipped 56 previous similar messages LustreError: 23192:0:(llite_lib.c:1845:ll_update_lsm_md()) lustre: [0x280000404:0x2ad8:0x0] dir layout mismatch: LustreError: 23192:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=1 count=1 index=2 hash=crush:0x2000003 max_inherit=0 max_inherit_rr=0 version=2 migrate_offset=0 migrate_hash=invalid:0 pool= LustreError: 23192:0:(lustre_lmv.h:167:lmv_stripe_object_dump()) stripe[0] [0x280000400:0x52:0x0] LustreError: 23192:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=1 count=2 index=2 hash=crush:0x82000003 max_inherit=0 max_inherit_rr=0 version=1 migrate_offset=1 migrate_hash=fnv_1a_64:2 pool= LustreError: 23192:0:(llite_lib.c:3698:ll_prep_inode()) lustre: new_inode - fatal error: rc = -22 LustreError: 23192:0:(llite_lib.c:3698:ll_prep_inode()) Skipped 1 previous similar message 14[25257]: segfault at 8 ip 00007f9d45a8b7e8 sp 00007ffed45b5ee0 error 4 in ld-2.17.so[7f9d45a80000+22000] LustreError: 25177:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8802e8cfa548: cannot apply new layout on [0x280000404:0x2af0:0x0] : rc = -5 LustreError: 25177:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x280000404:0x2af0:0x0] error -5. LustreError: 21160:0:(mdd_dir.c:4621:mdd_migrate_cmd_check()) lustre-MDD0001: '9' migration was interrupted, run 'lfs migrate -m 0 -c 1 -H crush 9' to finish migration: rc = -1 5[26877]: segfault at 406000 ip 0000000000406000 sp 00007ffd24363c88 error 14 in 5[606000+1000] LustreError: 28139:0:(lcommon_cl.c:179:cl_file_inode_init()) lustre: failed to initialize cl_object [0x280000404:0x2b05:0x0]: rc = -5 LustreError: 28139:0:(lcommon_cl.c:179:cl_file_inode_init()) Skipped 1 previous similar message LustreError: 28139:0:(llite_lib.c:3698:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 24574:0:(mdd_dir.c:4621:mdd_migrate_cmd_check()) lustre-MDD0001: '12' migration was interrupted, run 'lfs migrate -m 2 -c 2 -H crush 12' to finish migration: rc = -1 LustreError: 24574:0:(mdd_dir.c:4621:mdd_migrate_cmd_check()) Skipped 1 previous similar message Lustre: 18046:0:(out_handler.c:911:out_tx_end()) lustre-MDT0000-osd: error during execution of #8 from /home/green/git/lustre-release/lustre/ptlrpc/../../lustre/target/out_handler.c:585: rc = -2 LustreError: 18046:0:(out_handler.c:921:out_tx_end()) lustre-MDT0000-osd: undo for /home/green/git/lustre-release/lustre/ptlrpc/../../lustre/target/out_handler.c:425: rc = -524 LustreError: 30048:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8802e8cfa548: cannot apply new layout on [0x280000404:0x2af0:0x0] : rc = -5 LustreError: 30048:0:(lcommon_cl.c:179:cl_file_inode_init()) lustre: failed to initialize cl_object [0x280000404:0x2af0:0x0]: rc = -5 LustreError: 30048:0:(lcommon_cl.c:179:cl_file_inode_init()) Skipped 1 previous similar message Lustre: dir [0x240000403:0x1deb:0x0] stripe 3 readdir failed: -2, directory is partially accessed! Lustre: Skipped 2 previous similar messages 7[31239]: segfault at 8 ip 00007fb94913f7e8 sp 00007fff99846de0 error 4 in ld-2.17.so[7fb949134000+22000] LustreError: 29941:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e8cfa548: inode [0x240000403:0x1ed9:0x0] mdc close failed: rc = -116 LustreError: 29941:0:(file.c:247:ll_close_inode_openhandle()) Skipped 113 previous similar messages Lustre: 12514:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 618, rollback = 7 LustreError: 21160:0:(mdd_dir.c:4621:mdd_migrate_cmd_check()) lustre-MDD0000: '0' migration was interrupted, run 'lfs migrate -m 2 -c 1 -H crush 0' to finish migration: rc = -1 Lustre: 12514:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 175 previous similar messages 2[1325]: segfault at 8 ip 00007f736364c7e8 sp 00007ffecb2fb060 error 4 in ld-2.17.so[7f7363641000+22000] LustreError: 3831:0:(statahead.c:2399:start_statahead_thread()) lustre: invalid pattern 0X0. LustreError: 21333:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0000: [0x200000404:0x3840:0x0] migrate mdt count mismatch 1 != 3 LustreError: 4162:0:(lcommon_cl.c:179:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000404:0x386b:0x0]: rc = -5 LustreError: 4162:0:(lcommon_cl.c:179:cl_file_inode_init()) Skipped 8 previous similar messages LustreError: 4162:0:(llite_lib.c:3698:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 4162:0:(llite_lib.c:3698:ll_prep_inode()) Skipped 10 previous similar messages 7[5631]: segfault at 8 ip 00007f96041b27e8 sp 00007ffffdf35050 error 4 in ld-2.17.so[7f96041a7000+22000] LustreError: 17160:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0001: [0x240000403:0x1f5c:0x0] migrate mdt count mismatch 1 != 2 LustreError: 17924:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0001: [0x240000405:0xc17:0x0] migrate mdt count mismatch 3 != 1 15[6094]: segfault at 8 ip 00007feef1c5f7e8 sp 00007ffc1f3a7ed0 error 4 in ld-2.17.so[7feef1c54000+22000] LustreError: 11091:0:(mdd_dir.c:4621:mdd_migrate_cmd_check()) lustre-MDD0002: '8' migration was interrupted, run 'lfs migrate -m 1 -c 3 -H crush 8' to finish migration: rc = -1 LustreError: 11091:0:(mdd_dir.c:4621:mdd_migrate_cmd_check()) Skipped 2 previous similar messages LustreError: 19633:0:(mdd_object.c:3902:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000404:0x36e3:0x0]: rc = -2 LustreError: 20191:0:(mdd_object.c:3902:mdd_close()) lustre-MDD0001: failed to get lu_attr of [0x240000405:0xc7f:0x0]: rc = -2 LustreError: 17986:0:(mdd_object.c:3902:mdd_close()) lustre-MDD0002: failed to get lu_attr of [0x280000404:0x2cfe:0x0]: rc = -2 11[9160]: segfault at 8 ip 00007fd8383c57e8 sp 00007ffc4fa05760 error 4 in ld-2.17.so[7fd8383ba000+22000] Lustre: dir [0x240000403:0x1ff3:0x0] stripe 2 readdir failed: -2, directory is partially accessed! Lustre: Skipped 6 previous similar messages LustreError: 12602:0:(lcommon_cl.c:179:cl_file_inode_init()) lustre: failed to initialize cl_object [0x280000403:0x22e2:0x0]: rc = -5 LustreError: 12602:0:(lcommon_cl.c:179:cl_file_inode_init()) Skipped 30 previous similar messages LustreError: 12602:0:(llite_lib.c:3698:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 12602:0:(llite_lib.c:3698:ll_prep_inode()) Skipped 30 previous similar messages LustreError: 12602:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 14 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 16[11826]: segfault at 8 ip 00007fe6b5f647e8 sp 00007ffd6072fca0 error 4 in ld-2.17.so[7fe6b5f59000+22000] LustreError: 94:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 14 [0x280000403:0x22e2:0x0] inode@0000000000000000: rc = -5 13[14599]: segfault at 8 ip 00007f72e99b37e8 sp 00007ffd33ab1ae0 error 4 in ld-2.17.so[7f72e99a8000+22000] LustreError: 12842:0:(mdd_dir.c:4621:mdd_migrate_cmd_check()) lustre-MDD0002: '5' migration was interrupted, run 'lfs migrate -m 1 -c 1 -H crush 5' to finish migration: rc = -1 LustreError: 12842:0:(mdd_dir.c:4621:mdd_migrate_cmd_check()) Skipped 6 previous similar messages LustreError: 3301:0:(mdd_object.c:3902:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000404:0x39de:0x0]: rc = -2 LustreError: 16259:0:(llite_lib.c:1845:ll_update_lsm_md()) lustre: [0x280000403:0x24ad:0x0] dir layout mismatch: LustreError: 16259:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=2 count=2 index=2 hash=crush:0x2000003 max_inherit=0 max_inherit_rr=0 version=2 migrate_offset=0 migrate_hash=invalid:0 pool= LustreError: 16259:0:(lustre_lmv.h:167:lmv_stripe_object_dump()) stripe[0] [0x280000400:0x63:0x0] LustreError: 16259:0:(lustre_lmv.h:167:lmv_stripe_object_dump()) Skipped 2 previous similar messages LustreError: 16259:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=1 count=3 index=2 hash=crush:0x82000003 max_inherit=0 max_inherit_rr=0 version=1 migrate_offset=2 migrate_hash=crush:2000003 pool= 10[16999]: segfault at 8 ip 00007f2a8ac067e8 sp 00007ffc45f0e180 error 4 in ld-2.17.so[7f2a8abfb000+22000] 18[17636]: segfault at 8 ip 00007f89c2ae27e8 sp 00007ffca8355740 error 4 in ld-2.17.so[7f89c2ad7000+22000] LustreError: 14163:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8800a7ba2548: cannot apply new layout on [0x200000406:0x2e13:0x0] : rc = -5 LustreError: 14163:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x200000406:0x2e13:0x0] error -5. LustreError: 3319:0:(mdd_object.c:3902:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000406:0x2ae3:0x0]: rc = -2 18[19225]: segfault at 8 ip 00007f50f7a2c7e8 sp 00007ffd1bbcb1c0 error 4 in ld-2.17.so[7f50f7a21000+22000] LustreError: 7359:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0001: [0x240000403:0x209c:0x0] migrate mdt count mismatch 3 != 1 LustreError: 19649:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8800a7ba2548: cannot apply new layout on [0x200000406:0x2e13:0x0] : rc = -5 LustreError: 17505:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x200000404:0x3b46:0x0] error -5. LustreError: 15436:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8800a7ba2548: cannot apply new layout on [0x200000406:0x2e9a:0x0] : rc = -5 LustreError: 15436:0:(lov_object.c:1341:lov_layout_change()) Skipped 1 previous similar message LustreError: 15436:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x200000406:0x2e9a:0x0] error -5. 18[20383]: segfault at 8 ip 00007f090c5027e8 sp 00007ffc02ef5040 error 4 in ld-2.17.so[7f090c4f7000+22000] LustreError: 21062:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8802e8cfa548: cannot apply new layout on [0x200000404:0x3b46:0x0] : rc = -5 LustreError: 21122:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x240000405:0xe3d:0x0] error -5. LustreError: 22530:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8802e8cfa548: cannot apply new layout on [0x280000404:0x3045:0x0] : rc = -5 LustreError: 22530:0:(lov_object.c:1341:lov_layout_change()) Skipped 1 previous similar message LustreError: 22530:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x280000404:0x3045:0x0] error -5. LustreError: 25607:0:(lcommon_cl.c:179:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000403:0x207e:0x0]: rc = -5 LustreError: 25607:0:(lcommon_cl.c:179:cl_file_inode_init()) Skipped 10 previous similar messages LustreError: 25607:0:(llite_lib.c:3698:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 25607:0:(llite_lib.c:3698:ll_prep_inode()) Skipped 10 previous similar messages Lustre: dir [0x240000403:0x22c4:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 4 previous similar messages LustreError: 11091:0:(mdd_dir.c:4621:mdd_migrate_cmd_check()) lustre-MDD0000: '19' migration was interrupted, run 'lfs migrate -m 1 -c 2 -H crush 19' to finish migration: rc = -1 LustreError: 11091:0:(mdd_dir.c:4621:mdd_migrate_cmd_check()) Skipped 3 previous similar messages LustreError: 3319:0:(mdd_object.c:3902:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000404:0x3794:0x0]: rc = -2 LustreError: 11075:0:(mdd_orphans.c:261:mdd_orphan_delete()) lustre-MDD0001: could not delete orphan object [0x240000405:0x1040:0x0]: rc = -2 LustreError: 11075:0:(mdd_object.c:3956:mdd_close()) lustre-MDD0001: unable to delete [0x240000405:0x1040:0x0] from orphan list: rc = -2 LustreError: 28738:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8800a7ba2548: cannot apply new layout on [0x240000403:0x2360:0x0] : rc = -5 LustreError: 28738:0:(lov_object.c:1341:lov_layout_change()) Skipped 2 previous similar messages LustreError: 28738:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x240000403:0x2360:0x0] error -5. LustreError: 11073:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0001: [0x240000405:0x1012:0x0] migrate mdt count mismatch 1 != 2 3[8694]: segfault at 66 ip 0000000000000066 sp 00007ffe6d4ba878 error 14 in 3[400000+6000] LustreError: 7814:0:(llite_lib.c:1845:ll_update_lsm_md()) lustre: [0x200000406:0x32aa:0x0] dir layout mismatch: LustreError: 7814:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=1 count=1 index=0 hash=crush:0x2000003 max_inherit=0 max_inherit_rr=0 version=2 migrate_offset=0 migrate_hash=invalid:0 pool= LustreError: 7814:0:(lustre_lmv.h:167:lmv_stripe_object_dump()) stripe[0] [0x200000400:0x6a:0x0] LustreError: 7814:0:(lustre_lmv.h:167:lmv_stripe_object_dump()) Skipped 4 previous similar messages LustreError: 7814:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=1 count=4 index=0 hash=crush:0x82000003 max_inherit=0 max_inherit_rr=0 version=1 migrate_offset=1 migrate_hash=crush:3 pool= LustreError: 94:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 18 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 11[10692]: segfault at 0 ip (null) sp 00007ffc9a763948 error 14 in 11[400000+6000] LustreError: 19633:0:(mdd_object.c:3902:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000406:0x32aa:0x0]: rc = -2 LustreError: 19633:0:(mdd_object.c:3902:mdd_close()) Skipped 1 previous similar message 10[12680]: segfault at 8 ip 00007f6654e297e8 sp 00007ffd52072190 error 4 in ld-2.17.so[7f6654e1e000+22000] LustreError: 12797:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x240000405:0x12bd:0x0] error -5. LustreError: 14322:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0000: [0x200000404:0x4084:0x0] migrate mdt count mismatch 2 != 3 LustreError: 17172:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8802e8cfa548: cannot apply new layout on [0x240000405:0x12bd:0x0] : rc = -5 LustreError: 17172:0:(lov_object.c:1341:lov_layout_change()) Skipped 4 previous similar messages LustreError: 19396:0:(mdt_open.c:1302:mdt_cross_open()) lustre-MDT0002: [0x280000403:0x2691:0x0] doesn't exist!: rc = -14 1[19380]: segfault at 8 ip 00007f80ca5a67e8 sp 00007ffe8af4e400 error 4 in ld-2.17.so[7f80ca59b000+22000] 15[21451]: segfault at 8 ip 00007fe4a98bb7e8 sp 00007fffa6ab44b0 error 4 in ld-2.17.so[7fe4a98b0000+22000] 15[28882]: segfault at 8 ip 00007ff171f4e7e8 sp 00007fffd2f97660 error 4 in ld-2.17.so[7ff171f43000+22000] LustreError: 285:0:(lcommon_cl.c:179:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000404:0x4187:0x0]: rc = -5 LustreError: 285:0:(lcommon_cl.c:179:cl_file_inode_init()) Skipped 102 previous similar messages LustreError: 285:0:(llite_lib.c:3698:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 285:0:(llite_lib.c:3698:ll_prep_inode()) Skipped 102 previous similar messages LustreError: 285:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 19 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 2[31398]: segfault at 8 ip 00007fe007add7e8 sp 00007ffdd4c4a7f0 error 4 in ld-2.17.so[7fe007ad2000+22000] 16[3256]: segfault at 8 ip 00007f8e59af77e8 sp 00007ffdac0663d0 error 4 in ld-2.17.so[7f8e59aec000+22000] LustreError: 20186:0:(mdd_object.c:3902:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000404:0x4084:0x0]: rc = -2 LustreError: 20186:0:(mdd_object.c:3902:mdd_close()) Skipped 1 previous similar message LustreError: 3637:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x240000405:0x12bd:0x0] error -5. LustreError: 675:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 19 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 LustreError: 675:0:(statahead.c:825:ll_statahead_interpret_work()) Skipped 4 previous similar messages Lustre: dir [0x200000406:0x397e:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 12 previous similar messages traps: 2[10551] trap invalid opcode ip:405124 sp:7ffe376df118 error:0 in 2[400000+6000] 18[10593]: segfault at 8 ip 00007fe1345bb7e8 sp 00007ffcd27776f0 error 4 in ld-2.17.so[7fe1345b0000+22000] 18[10937]: segfault at 8 ip 00007f87f51a97e8 sp 00007ffc4efe65f0 error 4 in ld-2.17.so[7f87f519e000+22000] LustreError: 17891:0:(mdd_dir.c:4621:mdd_migrate_cmd_check()) lustre-MDD0002: '3' migration was interrupted, run 'lfs migrate -m 0 -c 2 -H crush 3' to finish migration: rc = -1 LustreError: 17891:0:(mdd_dir.c:4621:mdd_migrate_cmd_check()) Skipped 10 previous similar messages LustreError: 17169:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0001: [0x240000403:0x2811:0x0] migrate mdt count mismatch 3 != 2 7[15344]: segfault at 1c ip 00000000004017b0 sp 00007fff6ec12140 error 6 in 7[400000+6000] ptlrpc_watchdog_fire: 5 callbacks suppressed Lustre: mdt00_006: service thread pid 17010 was inactive for 40.059 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Lustre: Skipped 1 previous similar message Pid: 17010, comm: mdt00_006 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_check_lock+0xec/0x3c0 [mdt] [<0>] mdt_object_stripes_lock+0xba/0x660 [mdt] [<0>] mdt_reint_unlink+0x79a/0x15b0 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x73c/0xbb0 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: mdt00_001: service thread pid 11069 was inactive for 40.008 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 11069, comm: mdt00_001 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_pdo_lock+0x2a2/0x750 [mdt] [<0>] mdt_parent_lock+0x76/0x2a0 [mdt] [<0>] mdt_create+0xb44/0x1960 [mdt] [<0>] mdt_reint_create+0x2eb/0x470 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x73c/0xbb0 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: mdt_io00_015: service thread pid 25341 was inactive for 72.008 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 25341, comm: mdt_io00_015 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_fini+0xabc/0xfd0 [ptlrpc] [<0>] ldlm_cli_enqueue+0x461/0xb00 [ptlrpc] [<0>] osp_md_object_lock+0x151/0x2f0 [osp] [<0>] lod_object_lock+0xdb/0x7c0 [lod] [<0>] mdd_object_lock+0x2d/0xd0 [mdd] [<0>] mdt_remote_object_lock_try+0x14c/0x189 [mdt] [<0>] mdt_object_pdo_lock+0x729/0x750 [mdt] [<0>] mdt_parent_lock+0x76/0x2a0 [mdt] [<0>] mdt_reint_migrate+0xdf1/0x24b0 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x73c/0xbb0 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: mdt_io00_005: service thread pid 21160 was inactive for 72.092 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 1 previous similar message Lustre: mdt_io00_007: service thread pid 23362 was inactive for 72.151 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 1 previous similar message Lustre: mdt_io00_016: service thread pid 26746 was inactive for 72.116 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 4 previous similar messages Lustre: mdt_io00_020: service thread pid 17452 was inactive for 72.232 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 9 previous similar messages LustreError: 10821:0:(ldlm_lockd.c:252:expired_lock_main()) ### lock callback timer expired after 101s: evicting client at 0@lo ns: mdt-lustre-MDT0002_UUID lock: ffff88029cab7c00/0x4762192c4f793c3a lrc: 3/0,0 mode: PR/PR res: [0x280000403:0x2cca:0x0].0x0 bits 0x1b/0x0 rrc: 3 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0x4762192c4f793c1e expref: 950 pid: 17890 timeout: 1737 lvb_type: 0 LustreError: lustre-MDT0002-mdc-ffff8800a7ba2548: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 5 previous similar messages Lustre: mdt00_006: service thread pid 17010 completed after 100.501s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: lustre-MDT0002-mdc-ffff8800a7ba2548: Connection to lustre-MDT0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: lustre-MDT0002-mdc-ffff8800a7ba2548: This client was evicted by lustre-MDT0002; in progress operations using this service will fail. LustreError: 15087:0:(file.c:6138:ll_inode_revalidate_fini()) lustre: revalidate FID [0x280000403:0x2cca:0x0] error: rc = -5 LustreError: 15087:0:(file.c:6138:ll_inode_revalidate_fini()) Skipped 74 previous similar messages Lustre: mdt00_001: service thread pid 11069 completed after 99.997s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). 6[17214]: segfault at 406000 ip 0000000000406000 sp 00007ffe8f49ba38 error 14 in 6[606000+1000] Lustre: lustre-MDT0002-mdc-ffff8800a7ba2548: Connection restored to (at 0@lo) Lustre: mdt_io00_015: service thread pid 25341 completed after 98.285s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_005: service thread pid 21160 completed after 98.116s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_013: service thread pid 29610 completed after 97.870s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_007: service thread pid 23362 completed after 97.739s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_006: service thread pid 21297 completed after 97.495s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_012: service thread pid 12842 completed after 97.252s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_003: service thread pid 17891 completed after 96.937s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_017: service thread pid 26772 completed after 97.068s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_016: service thread pid 26746 completed after 97.044s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_008: service thread pid 24574 completed after 97.075s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_014: service thread pid 29677 completed after 96.990s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_000: service thread pid 11087 completed after 97.011s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_009: service thread pid 24631 completed after 96.942s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_001: service thread pid 11090 completed after 96.979s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_002: service thread pid 11091 completed after 96.719s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_004: service thread pid 19500 completed after 96.617s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_011: service thread pid 11428 completed after 96.458s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_019: service thread pid 17324 completed after 95.989s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_020: service thread pid 17452 completed after 95.483s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). 19[19146]: segfault at 8 ip 00007f6338b397e8 sp 00007ffff76b7fa0 error 4 in ld-2.17.so[7f6338b2e000+22000] LustreError: 3330:0:(mdd_object.c:3902:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000404:0x4694:0x0]: rc = -2 LustreError: 3330:0:(mdd_object.c:3902:mdd_close()) Skipped 5 previous similar messages Lustre: 17604:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000403:0x2a21:0x0] with magic=0xbd60bd0 Lustre: 17604:0:(lod_lov.c:1417:lod_parse_striping()) Skipped 83 previous similar messages Lustre: 17452:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 500 < left 1322, rollback = 2 Lustre: 17452:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 3918 previous similar messages Lustre: 26926:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 1/4/4, destroy: 0/0/0 Lustre: 26926:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 4800 previous similar messages Lustre: 26926:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 8/653/0 Lustre: 26926:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 4800 previous similar messages Lustre: 26926:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 5/21/0, punch: 0/0/0, quota 4/54/4 Lustre: 26926:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 4800 previous similar messages Lustre: 26926:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 3/69/1, delete: 0/0/0 Lustre: 26926:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 4800 previous similar messages Lustre: 23362:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 6/6/0, ref_del: 0/0/0 Lustre: 23362:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 4807 previous similar messages 14[31617]: segfault at 8 ip 00007f22e494f7e8 sp 00007fff1ce99dd0 error 4 in ld-2.17.so[7f22e4944000+22000] 8[3796]: segfault at 8 ip 00007f54af8607e8 sp 00007ffebf619170 error 4 in ld-2.17.so[7f54af855000+22000] LustreError: 4419:0:(lcommon_cl.c:179:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000404:0x4a29:0x0]: rc = -5 LustreError: 4419:0:(lcommon_cl.c:179:cl_file_inode_init()) Skipped 96 previous similar messages LustreError: 4419:0:(llite_lib.c:3698:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 4419:0:(llite_lib.c:3698:ll_prep_inode()) Skipped 96 previous similar messages 16[4245]: segfault at 8 ip 00007f51b18da7e8 sp 00007ffeb7bee2f0 error 4 in ld-2.17.so[7f51b18cf000+22000] 11[4103]: segfault at 8 ip 00007f9b9ea027e8 sp 00007ffcec8666a0 error 4 in ld-2.17.so[7f9b9e9f7000+22000] 19[7916]: segfault at 0 ip 0000000000403e5f sp 00007ffefe3359c0 error 6 in 19[400000+6000] 1[9133]: segfault at 0 ip (null) sp 00007ffddaefd158 error 14 in 1[400000+6000] LustreError: 25341:0:(mdt_reint.c:2523:mdt_reint_migrate()) lustre-MDT0001: migrate [0x240000403:0x289d:0x0]/8 failed: rc = -114 LustreError: 25341:0:(mdt_reint.c:2523:mdt_reint_migrate()) Skipped 511 previous similar messages Lustre: lustre-OST0000-osc-ffff8800a7ba2548: disconnect after 23s idle LustreError: 10821:0:(ldlm_lockd.c:252:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0001_UUID lock: ffff8802c208e1c0/0x4762192c4f90aec1 lrc: 3/0,0 mode: PR/PR res: [0x240000405:0x1b92:0x0].0x0 bits 0x1b/0x0 rrc: 3 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0x4762192c4f90aea5 expref: 735 pid: 24747 timeout: 1957 lvb_type: 0 Lustre: 11087:0:(mdd_dir.c:4701:mdd_migrate_object()) lustre-MDD0001: [0x240000403:0x1:0x0]/13 is open, migrate only dentry Lustre: 11087:0:(mdd_dir.c:4701:mdd_migrate_object()) Skipped 141 previous similar messages LustreError: 13801:0:(ldlm_lockd.c:1492:ldlm_handle_enqueue()) ### lock on destroyed export ffff880094b312a8 ns: mdt-lustre-MDT0001_UUID lock: ffff88029cc1b100/0x4762192c4f90c025 lrc: 3/0,0 mode: PR/PR res: [0x240000403:0x1:0x0].0x0 bits 0x12/0x0 rrc: 11 type: IBT gid 0 flags: 0x50200000000000 nid: 0@lo remote: 0x4762192c4f90be5e expref: 23 pid: 13801 timeout: 0 lvb_type: 0 LustreError: 13801:0:(ldlm_lockd.c:1492:ldlm_handle_enqueue()) Skipped 7 previous similar messages LustreError: lustre-MDT0001-mdc-ffff8800a7ba2548: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-MDT0001-mdc-ffff8800a7ba2548: Connection to lustre-MDT0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: lustre-MDT0001-mdc-ffff8800a7ba2548: This client was evicted by lustre-MDT0001; in progress operations using this service will fail. LustreError: 10863:0:(file.c:6138:ll_inode_revalidate_fini()) lustre: revalidate FID [0x240000403:0x1:0x0] error: rc = -5 LustreError: 10863:0:(file.c:6138:ll_inode_revalidate_fini()) Skipped 3 previous similar messages LustreError: 9647:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a7ba2548: inode [0x240000405:0x1b92:0x0] mdc close failed: rc = -5 LustreError: 9647:0:(file.c:247:ll_close_inode_openhandle()) Skipped 127 previous similar messages LustreError: 10356:0:(llite_lib.c:1996:ll_md_setattr()) md_setattr fails: rc = -108 LustreError: 10356:0:(llite_lib.c:1996:ll_md_setattr()) Skipped 4 previous similar messages LustreError: 5642:0:(mdc_request.c:1464:mdc_read_page()) lustre-MDT0001-mdc-ffff8800a7ba2548: [0x240000400:0x87:0x0] lock enqueue fails: rc = -108 Lustre: dir [0x240000403:0x289d:0x0] stripe 0 readdir failed: -108, directory is partially accessed! Lustre: Skipped 11 previous similar messages Lustre: lustre-MDT0001-mdc-ffff8800a7ba2548: Connection restored to (at 0@lo) Lustre: 17608:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 618, rollback = 7 Lustre: 17608:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 203 previous similar messages LustreError: 17592:0:(lustre_lmv.h:500:lmv_is_sane()) unknown layout LMV: magic=0xcd40cd0 count=2 index=1 hash=crush:0x82000003 version=1 migrate_offset=1 migrate_hash=fnv_1a_64:2 pool= LustreError: 21160:0:(lustre_lmv.h:500:lmv_is_sane()) unknown layout LMV: magic=0xcd40cd0 count=3 index=2 hash=crush:0x82000003 version=1 migrate_offset=2 migrate_hash=fnv_1a_64:2 pool= LustreError: 17891:0:(lustre_lmv.h:500:lmv_is_sane()) unknown layout LMV: magic=0xcd40cd0 count=3 index=2 hash=crush:0x82000003 version=1 migrate_offset=2 migrate_hash=fnv_1a_64:2 pool= LustreError: 19406:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0000: [0x200000404:0x4ce3:0x0] migrate mdt count mismatch 2 != 3 0[13062]: segfault at 8 ip 00007f42da8a97e8 sp 00007ffd81dd4b40 error 4 in ld-2.17.so[7f42da89e000+22000] 0[13075]: segfault at 8 ip 00007f2c8ec127e8 sp 00007fff48560510 error 4 in ld-2.17.so[7f2c8ec07000+22000] LustreError: 11091:0:(mdd_dir.c:4621:mdd_migrate_cmd_check()) lustre-MDD0001: '8' migration was interrupted, run 'lfs migrate -m 2 -c 3 -H crush 8' to finish migration: rc = -1 LustreError: 11091:0:(mdd_dir.c:4621:mdd_migrate_cmd_check()) Skipped 9 previous similar messages LustreError: 12010:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8802e8cfa548: cannot apply new layout on [0x240000406:0x112:0x0] : rc = -5 LustreError: 12010:0:(lov_object.c:1341:lov_layout_change()) Skipped 7 previous similar messages LustreError: 12010:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x240000406:0x112:0x0] error -5. 15[15510]: segfault at 8 ip 00007f5567ce47e8 sp 00007ffc7b361320 error 4 in ld-2.17.so[7f5567cd9000+22000] LustreError: 19633:0:(mdd_object.c:3902:mdd_close()) lustre-MDD0002: failed to get lu_attr of [0x280000403:0x3018:0x0]: rc = -2 LustreError: 19633:0:(mdd_object.c:3902:mdd_close()) Skipped 15 previous similar messages 17[24206]: segfault at 8 ip 00007ff188d477e8 sp 00007ffc9c86d240 error 4 in ld-2.17.so[7ff188d3c000+22000] 2[25269]: segfault at 8 ip 00007f88f9d1a7e8 sp 00007fff007bc6c0 error 4 in ld-2.17.so[7f88f9d0f000+22000] 14[25642]: segfault at 8 ip 00007fde1334a7e8 sp 00007ffe81488ed0 error 4 in ld-2.17.so[7fde1333f000+22000] 19[27788]: segfault at 8 ip 00007f73ddebe7e8 sp 00007ffdeac6a620 error 4 in ld-2.17.so[7f73ddeb3000+22000] Lustre: 30675:0:(out_handler.c:911:out_tx_end()) lustre-MDT0001-osd: error during execution of #0 from /home/green/git/lustre-release/lustre/ptlrpc/../../lustre/target/out_handler.c:448: rc = -61 LustreError: 24631:0:(lustre_lmv.h:500:lmv_is_sane()) unknown layout LMV: magic=0xcd40cd0 count=2 index=1 hash=crush:0x82000003 version=1 migrate_offset=1 migrate_hash=fnv_1a_64:2 pool= LustreError: 17573:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0000: [0x200000406:0x4517:0x0] migrate mdt count mismatch 1 != 3 LustreError: 17573:0:(mdt_xattr.c:402:mdt_dir_layout_update()) Skipped 4 previous similar messages 14[29819]: segfault at 8 ip 00007f06f10527e8 sp 00007fff57a0e110 error 4 in ld-2.17.so[7f06f1047000+22000] LustreError: 30113:0:(llite_nfs.c:430:ll_dir_get_parent_fid()) lustre: failure inode [0x200000404:0x4f37:0x0] get parent: rc = -2 16[32744]: segfault at 8 ip 00007f4a540d97e8 sp 00007ffd5445e100 error 4 in ld-2.17.so[7f4a540ce000+22000] | Link to test |
racer test 1: racer on clients: centos-110.localnet DURATION=2700 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 PGD af26c067 PUD 96c12067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) osp(OE) ofd(OE) lod(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) crc32_generic lnet(OE) libcfs(OE) crc_t10dif crct10dif_generic crct10dif_common pcspkr i2c_piix4 virtio_balloon virtio_console ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm drm_panel_orientation_quirks ata_piix floppy serio_raw virtio_blk i2c_core libata CPU: 1 PID: 8674 Comm: ll_sa_8517 Kdump: loaded Tainted: P OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.16.0-3.module_el8.7.0+1218+f626c2ff 04/01/2014 task: ffff88009066dc40 ti: ffff88028aeec000 task.ti: ffff88028aeec000 RIP: 0010:[<ffffffff813f502c>] [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 RSP: 0018:ffff88028aeefd78 EFLAGS: 00010296 RAX: ffff880322796c48 RBX: 0000000000000008 RCX: 000000010026001d RDX: 0000000000000026 RSI: ffff880322796f78 RDI: 0000000000000008 RBP: ffff88028aeefd88 R08: ffff8802acdb9738 R09: 0000000000000000 R10: ffff8802acdbe018 R11: ffff8802acdb9738 R12: 0000000000000000 R13: ffff8802b538dc40 R14: ffff880322796f78 R15: ffff8802acdb9738 FS: 0000000000000000(0000) GS:ffff880331a40000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 00000000af63a000 CR4: 00000000000007e0 Call Trace: [<ffffffffa158f621>] ll_statahead_thread+0xe01/0x22c0 [lustre] [<ffffffff810df6e2>] ? pick_next_task_fair+0x702/0x8a0 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e329e>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffff810c834d>] ? finish_task_switch+0x5d/0x1b0 [<ffffffffa158e820>] ? ll_statahead_handle.constprop.29+0x1c0/0x1c0 [lustre] [<ffffffff810ba114>] kthread+0xe4/0xf0 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 | Lustre: 28681:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff880098af6940 x1828046749670784/t4294967481(0) o101->c29ed6b6-c75d-4fe2-a69a-b3cbe32f9093@0@lo:638/0 lens 376/864 e 0 to 0 dl 1743361323 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 LustreError: 16203:0:(mdt_reint.c:2523:mdt_reint_migrate()) lustre-MDT0001: migrate [0x240000403:0x1:0x0]/8 failed: rc = -114 LustreError: 29085:0:(mdt_reint.c:2523:mdt_reint_migrate()) lustre-MDT0000: migrate [0x200000403:0x1:0x0]/0 failed: rc = -114 Lustre: 29776:0:(mdd_dir.c:4701:mdd_migrate_object()) lustre-MDD0002: [0x280000403:0x1:0x0]/13 is open, migrate only dentry LustreError: 29848:0:(mdd_dir.c:4621:mdd_migrate_cmd_check()) lustre-MDD0002: '16' migration was interrupted, run 'lfs migrate -m 1 -c 2 -H crush 16' to finish migration: rc = -1 LustreError: 29848:0:(mdt_reint.c:2523:mdt_reint_migrate()) lustre-MDT0002: migrate [0x240000403:0x1:0x0]/16 failed: rc = -1 LustreError: 29848:0:(mdt_reint.c:2523:mdt_reint_migrate()) Skipped 4 previous similar messages 8[28762]: segfault at 0 ip (null) sp 00007ffc0fba7ce8 error 14 in 8[400000+6000] LustreError: 16201:0:(mdt_reint.c:2523:mdt_reint_migrate()) lustre-MDT0001: migrate [0x240000403:0x1:0x0]/5 failed: rc = -114 LustreError: 16201:0:(mdt_reint.c:2523:mdt_reint_migrate()) Skipped 4 previous similar messages Lustre: 31001:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff88029b0bad40 x1828046751251584/t4294968128(0) o101->c35f4da9-5237-4343-a995-05da625dfd37@0@lo:646/0 lens 376/816 e 0 to 0 dl 1743361331 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 Lustre: 29848:0:(mdd_dir.c:4701:mdd_migrate_object()) lustre-MDD0002: [0x280000403:0x7:0x0]/11 is open, migrate only dentry LustreError: 16201:0:(mdt_reint.c:2523:mdt_reint_migrate()) lustre-MDT0000: migrate [0x200000403:0x2:0x0]/12 failed: rc = -114 LustreError: 16201:0:(mdt_reint.c:2523:mdt_reint_migrate()) Skipped 2 previous similar messages Lustre: 16203:0:(mdd_dir.c:4701:mdd_migrate_object()) lustre-MDD0000: [0x200000403:0x1:0x0]/5 is open, migrate only dentry LustreError: 16203:0:(mdd_dir.c:4621:mdd_migrate_cmd_check()) lustre-MDD0002: '12' migration was interrupted, run 'lfs migrate -m 1 -c 2 -H crush 12' to finish migration: rc = -1 LustreError: 2517:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802c886c138: inode [0x200000404:0xd8:0x0] mdc close failed: rc = -116 Lustre: 31671:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff8802a922d540 x1828046752971264/t4294968683(0) o101->c35f4da9-5237-4343-a995-05da625dfd37@0@lo:657/0 lens 376/816 e 0 to 0 dl 1743361342 ref 1 fl Interpret:H/202/0 rc 0/0 job:'cat.0' uid:0 gid:0 LustreError: 16203:0:(mdt_reint.c:2523:mdt_reint_migrate()) lustre-MDT0001: migrate [0x240000403:0x1:0x0]/11 failed: rc = -114 LustreError: 16203:0:(mdt_reint.c:2523:mdt_reint_migrate()) Skipped 8 previous similar messages LustreError: 2528:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802cd8012a8: inode [0x240000403:0x66:0x0] mdc close failed: rc = -116 Lustre: 29848:0:(mdd_dir.c:4701:mdd_migrate_object()) lustre-MDD0001: [0x240000403:0x1:0x0]/15 is open, migrate only dentry Lustre: 29848:0:(mdd_dir.c:4701:mdd_migrate_object()) Skipped 1 previous similar message Lustre: 29913:0:(mdd_dir.c:4701:mdd_migrate_object()) lustre-MDD0001: [0x240000403:0x1:0x0]/11 is open, migrate only dentry Lustre: dir [0x240000404:0x2f:0x0] stripe 2 readdir failed: -2, directory is partially accessed! LustreError: 8389:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802cd8012a8: inode [0x280000403:0x171:0x0] mdc close failed: rc = -116 LustreError: 16202:0:(mdt_reint.c:2523:mdt_reint_migrate()) lustre-MDT0002: migrate [0x280000403:0x14d:0x0]/9 failed: rc = -114 LustreError: 16202:0:(mdt_reint.c:2523:mdt_reint_migrate()) Skipped 13 previous similar messages Lustre: 4739:0:(mdd_dir.c:4701:mdd_migrate_object()) lustre-MDD0000: [0x200000403:0x2:0x0]/13 is open, migrate only dentry Lustre: 4739:0:(mdd_dir.c:4701:mdd_migrate_object()) Skipped 3 previous similar messages LustreError: 29553:0:(lustre_lmv.h:500:lmv_is_sane()) unknown layout LMV: magic=0xcd40cd0 count=4 index=3 hash=crush:0x82000003 version=1 migrate_offset=3 migrate_hash=fnv_1a_64:2 pool= LustreError: 133:0:(lcommon_cl.c:179:cl_file_inode_init()) lustre: failed to initialize cl_object [0x280000404:0x38:0x0]: rc = -5 LustreError: 133:0:(llite_lib.c:3695:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 133:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 13 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 LustreError: 11089:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8802cd8012a8: cannot apply new layout on [0x280000404:0x12c:0x0] : rc = -5 LustreError: 11089:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x280000404:0x12c:0x0] error -5. Lustre: 29630:0:(mdd_dir.c:4701:mdd_migrate_object()) lustre-MDD0002: [0x240000403:0x47:0x0]/3 is open, migrate only dentry Lustre: 29630:0:(mdd_dir.c:4701:mdd_migrate_object()) Skipped 4 previous similar messages LustreError: 4722:0:(mdd_dir.c:4621:mdd_migrate_cmd_check()) lustre-MDD0000: '15' migration was interrupted, run 'lfs migrate -m 2 -c 1 -H crush 15' to finish migration: rc = -1 LustreError: 15042:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802c886c138: inode [0x200000404:0x239:0x0] mdc close failed: rc = -2 LustreError: 15042:0:(file.c:247:ll_close_inode_openhandle()) Skipped 1 previous similar message LustreError: 28695:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0001: [0x240000404:0x15c:0x0] migrate mdt count mismatch 2 != 3 LustreError: 17382:0:(lcommon_cl.c:179:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000404:0x1d3:0x0]: rc = -5 LustreError: 17382:0:(lcommon_cl.c:179:cl_file_inode_init()) Skipped 8 previous similar messages LustreError: 17382:0:(llite_lib.c:3695:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 17382:0:(llite_lib.c:3695:ll_prep_inode()) Skipped 8 previous similar messages LustreError: 17382:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8802cd8012a8: cannot apply new layout on [0x280000404:0x12c:0x0] : rc = -5 LustreError: 19965:0:(lcommon_cl.c:179:cl_file_inode_init()) lustre: failed to initialize cl_object [0x280000404:0x1a6:0x0]: rc = -5 LustreError: 19965:0:(lcommon_cl.c:179:cl_file_inode_init()) Skipped 3 previous similar messages LustreError: 19965:0:(llite_lib.c:3695:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 19965:0:(llite_lib.c:3695:ll_prep_inode()) Skipped 3 previous similar messages LustreError: 4722:0:(mdd_dir.c:4621:mdd_migrate_cmd_check()) lustre-MDD0001: '15' migration was interrupted, run 'lfs migrate -m 2 -c 1 -H crush 15' to finish migration: rc = -1 LustreError: 4722:0:(mdt_reint.c:2523:mdt_reint_migrate()) lustre-MDT0001: migrate [0x200000403:0x2:0x0]/15 failed: rc = -1 LustreError: 4722:0:(mdt_reint.c:2523:mdt_reint_migrate()) Skipped 17 previous similar messages LustreError: 32391:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0002: [0x280000403:0x21c:0x0] migrate mdt count mismatch 2 != 3 LustreError: 24491:0:(lcommon_cl.c:179:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000404:0x1d3:0x0]: rc = -5 LustreError: 24491:0:(lcommon_cl.c:179:cl_file_inode_init()) Skipped 1 previous similar message LustreError: 24491:0:(llite_lib.c:3695:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 24491:0:(llite_lib.c:3695:ll_prep_inode()) Skipped 1 previous similar message LustreError: 24491:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8802cd8012a8: cannot apply new layout on [0x280000404:0x12c:0x0] : rc = -5 LustreError: 415:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 17 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 LustreError: 415:0:(statahead.c:825:ll_statahead_interpret_work()) Skipped 8 previous similar messages LustreError: 19304:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802c886c138: inode [0x240000403:0x68:0x0] mdc close failed: rc = -116 LustreError: 28131:0:(lcommon_cl.c:179:cl_file_inode_init()) lustre: failed to initialize cl_object [0x280000404:0x38:0x0]: rc = -5 LustreError: 28131:0:(lcommon_cl.c:179:cl_file_inode_init()) Skipped 6 previous similar messages LustreError: 28131:0:(llite_lib.c:3695:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 28131:0:(llite_lib.c:3695:ll_prep_inode()) Skipped 6 previous similar messages Lustre: dir [0x240000404:0x1ba:0x0] stripe 3 readdir failed: -2, directory is partially accessed! Lustre: 27988:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000403:0x29e:0x0] with magic=0xbd60bd0 LustreError: 5116:0:(mdd_dir.c:4621:mdd_migrate_cmd_check()) lustre-MDD0001: '16' migration was interrupted, run 'lfs migrate -m 2 -c 2 -H crush 16' to finish migration: rc = -1 Lustre: 22282:0:(mdd_dir.c:4701:mdd_migrate_object()) lustre-MDD0000: [0x200000403:0x2:0x0]/13 is open, migrate only dentry Lustre: 22282:0:(mdd_dir.c:4701:mdd_migrate_object()) Skipped 11 previous similar messages 3[30312]: segfault at 8 ip 00007fac188f37e8 sp 00007ffea89060a0 error 4 in ld-2.17.so[7fac188e8000+22000] LustreError: 230:0:(lcommon_cl.c:179:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000404:0x1a0:0x0]: rc = -5 LustreError: 230:0:(lcommon_cl.c:179:cl_file_inode_init()) Skipped 11 previous similar messages LustreError: 230:0:(llite_lib.c:3695:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 230:0:(llite_lib.c:3695:ll_prep_inode()) Skipped 11 previous similar messages LustreError: 230:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 17 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 LustreError: 230:0:(statahead.c:825:ll_statahead_interpret_work()) Skipped 2 previous similar messages 3[1978]: segfault at 8 ip 00007fd507bde7e8 sp 00007ffcae048870 error 4 in ld-2.17.so[7fd507bd3000+22000] LustreError: 2892:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802c886c138: inode [0x240000404:0x2cf:0x0] mdc close failed: rc = -116 LustreError: 2892:0:(file.c:247:ll_close_inode_openhandle()) Skipped 2 previous similar messages LustreError: 2084:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8802cd8012a8: cannot apply new layout on [0x280000404:0x12c:0x0] : rc = -5 Lustre: 10465:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0x30e:0x0] with magic=0xbd60bd0 Lustre: 10465:0:(lod_lov.c:1417:lod_parse_striping()) Skipped 3 previous similar messages LustreError: 29776:0:(mdd_dir.c:4621:mdd_migrate_cmd_check()) lustre-MDD0002: '14' migration was interrupted, run 'lfs migrate -m 1 -c 3 -H crush 14' to finish migration: rc = -1 Lustre: dir [0x240000404:0x283:0x0] stripe 3 readdir failed: -2, directory is partially accessed! LustreError: 4991:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8802cd8012a8: cannot apply new layout on [0x200000404:0x32b:0x0] : rc = -5 LustreError: 4991:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x200000404:0x32b:0x0] error -5. LustreError: 7773:0:(lcommon_cl.c:179:cl_file_inode_init()) lustre: failed to initialize cl_object [0x280000404:0x371:0x0]: rc = -5 LustreError: 7773:0:(lcommon_cl.c:179:cl_file_inode_init()) Skipped 15 previous similar messages LustreError: 7773:0:(llite_lib.c:3695:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 7773:0:(llite_lib.c:3695:ll_prep_inode()) Skipped 15 previous similar messages LustreError: 401:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 20 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 LustreError: 401:0:(statahead.c:825:ll_statahead_interpret_work()) Skipped 1 previous similar message LustreError: 29776:0:(mdt_reint.c:2523:mdt_reint_migrate()) lustre-MDT0000: migrate [0x280000404:0x282:0x0]/13 failed: rc = -2 LustreError: 29776:0:(mdt_reint.c:2523:mdt_reint_migrate()) Skipped 44 previous similar messages Lustre: dir [0x280000404:0x282:0x0] stripe 1 readdir failed: -2, directory is partially accessed! LustreError: 150:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 20 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 LustreError: 17093:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802c886c138: inode [0x240000403:0x331:0x0] mdc close failed: rc = -116 LustreError: 17093:0:(file.c:247:ll_close_inode_openhandle()) Skipped 3 previous similar messages LustreError: 16202:0:(mdd_dir.c:4621:mdd_migrate_cmd_check()) lustre-MDD0000: '15' migration was interrupted, run 'lfs migrate -m 2 -c 1 -H crush 15' to finish migration: rc = -1 LustreError: 16202:0:(mdd_dir.c:4621:mdd_migrate_cmd_check()) Skipped 4 previous similar messages LustreError: 18044:0:(lcommon_cl.c:179:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000404:0x2f9:0x0]: rc = -5 LustreError: 18044:0:(lcommon_cl.c:179:cl_file_inode_init()) Skipped 13 previous similar messages LustreError: 18044:0:(llite_lib.c:3695:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 18044:0:(llite_lib.c:3695:ll_prep_inode()) Skipped 13 previous similar messages Lustre: 32126:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000403:0x375:0x0] with magic=0xbd60bd0 Lustre: 32126:0:(lod_lov.c:1417:lod_parse_striping()) Skipped 1 previous similar message Lustre: 2146:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000403:0x36b:0x0] with magic=0xbd60bd0 Lustre: 2146:0:(lod_lov.c:1417:lod_parse_striping()) Skipped 5 previous similar messages LustreError: 127:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 17 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 Lustre: 29776:0:(mdd_dir.c:4701:mdd_migrate_object()) lustre-MDD0000: [0x200000403:0x2:0x0]/13 is open, migrate only dentry Lustre: 29776:0:(mdd_dir.c:4701:mdd_migrate_object()) Skipped 7 previous similar messages 17[25783]: segfault at 8 ip 00007fc2e44677e8 sp 00007ffe216667d0 error 4 in ld-2.17.so[7fc2e445c000+22000] LustreError: 18350:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802cd8012a8: inode [0x280000404:0x18c:0x0] mdc close failed: rc = -2 LustreError: 18350:0:(file.c:247:ll_close_inode_openhandle()) Skipped 1 previous similar message LustreError: 10395:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0000: [0x200000403:0x371:0x0] migrate mdt count mismatch 3 != 1 Lustre: dir [0x240000404:0x522:0x0] stripe 3 readdir failed: -2, directory is partially accessed! LustreError: 190:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 20 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 Lustre: 28088:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000404:0x4f2:0x0] with magic=0xbd60bd0 Lustre: 28088:0:(lod_lov.c:1417:lod_parse_striping()) Skipped 11 previous similar messages LustreError: 28088:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0001: [0x240000403:0x302:0x0] migrate mdt count mismatch 2 != 3 Lustre: lustre-MDT0001: trigger partial OI scrub for RPC inconsistency, checking FID [0x240000403:0x302:0x0]/0xa): rc = 0 Lustre: 2126:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000404:0x6cb:0x0] with magic=0xbd60bd0 Lustre: 2126:0:(lod_lov.c:1417:lod_parse_striping()) Skipped 19 previous similar messages LustreError: 5116:0:(mdd_dir.c:4621:mdd_migrate_cmd_check()) lustre-MDD0002: '13' migration was interrupted, run 'lfs migrate -m 0 -c 3 -H crush 13' to finish migration: rc = -1 LustreError: 5116:0:(mdd_dir.c:4621:mdd_migrate_cmd_check()) Skipped 3 previous similar messages Lustre: dir [0x280000403:0x4aa:0x0] stripe 1 readdir failed: -2, directory is partially accessed! LustreError: 6845:0:(lcommon_cl.c:179:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000403:0x4b8:0x0]: rc = -5 LustreError: 6845:0:(lcommon_cl.c:179:cl_file_inode_init()) Skipped 45 previous similar messages LustreError: 6845:0:(llite_lib.c:3695:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 6845:0:(llite_lib.c:3695:ll_prep_inode()) Skipped 45 previous similar messages Lustre: 2211:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000404:0x77b:0x0] with magic=0xbd60bd0 Lustre: 2211:0:(lod_lov.c:1417:lod_parse_striping()) Skipped 11 previous similar messages LustreError: 16150:0:(llite_lib.c:1845:ll_update_lsm_md()) lustre: [0x240000404:0x522:0x0] dir layout mismatch: LustreError: 16150:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=1 count=3 index=1 hash=crush:0x2000003 max_inherit=0 max_inherit_rr=0 version=2 migrate_offset=0 migrate_hash=invalid:0 pool= LustreError: 16150:0:(lustre_lmv.h:167:lmv_stripe_object_dump()) stripe[0] [0x240000400:0x1d:0x0] LustreError: 16150:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=1 count=3 index=1 hash=crush:0x2000003 max_inherit=0 max_inherit_rr=0 version=2 migrate_offset=0 migrate_hash=invalid:0 pool= LustreError: 16139:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=1 count=3 index=1 hash=crush:0x2000003 max_inherit=0 max_inherit_rr=0 version=2 migrate_offset=0 migrate_hash=invalid:0 pool= LustreError: 16139:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=1 count=3 index=1 hash=crush:0x2000003 max_inherit=0 max_inherit_rr=0 version=2 migrate_offset=0 migrate_hash=invalid:0 pool= 12[13567]: segfault at 8 ip 00007f26fc0a37e8 sp 00007fffa2b41720 error 4 in ld-2.17.so[7f26fc098000+22000] Lustre: dir [0x200000404:0x7e6:0x0] stripe 2 readdir failed: -2, directory is partially accessed! Lustre: Skipped 3 previous similar messages LustreError: 17366:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8802c886c138: cannot apply new layout on [0x200000403:0x4b8:0x0] : rc = -5 LustreError: 17366:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x200000403:0x4b8:0x0] error -5. 4[22164]: segfault at 0 ip (null) sp 00007ffe2028d938 error 14 in 4[400000+6000] LustreError: 24520:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802c886c138: inode [0x280000404:0x797:0x0] mdc close failed: rc = -116 LustreError: 24520:0:(file.c:247:ll_close_inode_openhandle()) Skipped 7 previous similar messages LustreError: 29221:0:(mdt_reint.c:2523:mdt_reint_migrate()) lustre-MDT0002: migrate [0x280000404:0x4e5:0x0]/18 failed: rc = -114 LustreError: 29221:0:(mdt_reint.c:2523:mdt_reint_migrate()) Skipped 100 previous similar messages Lustre: 10328:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0x7fa:0x0] with magic=0xbd60bd0 Lustre: 10328:0:(lod_lov.c:1417:lod_parse_striping()) Skipped 39 previous similar messages LustreError: 16202:0:(mdd_dir.c:4621:mdd_migrate_cmd_check()) lustre-MDD0001: '13' migration was interrupted, run 'lfs migrate -m 0 -c 3 -H crush 13' to finish migration: rc = -1 LustreError: 16202:0:(mdd_dir.c:4621:mdd_migrate_cmd_check()) Skipped 4 previous similar messages Lustre: dir [0x240000404:0x860:0x0] stripe 3 readdir failed: -2, directory is partially accessed! Lustre: 4739:0:(mdd_dir.c:4701:mdd_migrate_object()) lustre-MDD0002: [0x280000404:0x4e5:0x0]/4 is open, migrate only dentry Lustre: 4739:0:(mdd_dir.c:4701:mdd_migrate_object()) Skipped 24 previous similar messages LustreError: 1594:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8802c886c138: cannot apply new layout on [0x240000403:0x6ac:0x0] : rc = -5 LustreError: 1594:0:(lov_object.c:1341:lov_layout_change()) Skipped 1 previous similar message LustreError: 1594:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x240000403:0x6ac:0x0] error -5. 19[11200]: segfault at 0 ip 0000000000403e5f sp 00007ffce0da9350 error 6 in 19[400000+6000] LustreError: 20747:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0000: [0x200000403:0x7c2:0x0] migrate mdt count mismatch 2 != 3 Lustre: lustre-MDT0000: trigger partial OI scrub for RPC inconsistency, checking FID [0x200000404:0xa92:0x0]/0xa): rc = 0 LustreError: 28802:0:(lcommon_cl.c:179:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000403:0x795:0x0]: rc = -5 LustreError: 28802:0:(lcommon_cl.c:179:cl_file_inode_init()) Skipped 37 previous similar messages LustreError: 28802:0:(llite_lib.c:3695:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 28802:0:(llite_lib.c:3695:ll_prep_inode()) Skipped 37 previous similar messages 9[28972]: segfault at 8 ip 00007f649a1a37e8 sp 00007ffd13461b20 error 4 in ld-2.17.so[7f649a198000+22000] LustreError: 232:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 19 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 LustreError: 31306:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8802c886c138: cannot apply new layout on [0x200000403:0x948:0x0] : rc = -5 LustreError: 31306:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x200000403:0x948:0x0] error -5. LustreError: 3492:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802cd8012a8: inode [0x280000404:0x7ff:0x0] mdc close failed: rc = -2 LustreError: 3492:0:(file.c:247:ll_close_inode_openhandle()) Skipped 9 previous similar messages 1[7989]: segfault at 8 ip 00007fcc069a77e8 sp 00007ffc29c85310 error 4 in ld-2.17.so[7fcc0699c000+22000] LustreError: 21889:0:(mdd_dir.c:4621:mdd_migrate_cmd_check()) lustre-MDD0001: '3' migration was interrupted, run 'lfs migrate -m 2 -c 1 -H crush 3' to finish migration: rc = -1 LustreError: 21889:0:(mdd_dir.c:4621:mdd_migrate_cmd_check()) Skipped 8 previous similar messages LustreError: 17924:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x240000404:0xa8f:0x0] error -5. Lustre: 20747:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000404:0xbc5:0x0] with magic=0xbd60bd0 Lustre: 20747:0:(lod_lov.c:1417:lod_parse_striping()) Skipped 5 previous similar messages LustreError: 414:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 20 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 LustreError: 414:0:(statahead.c:825:ll_statahead_interpret_work()) Skipped 3 previous similar messages Lustre: dir [0x280000404:0xa23:0x0] stripe 3 readdir failed: -2, directory is partially accessed! Lustre: Skipped 1 previous similar message LustreError: 658:0:(statahead.c:2447:start_statahead_thread()) lustre: unsupported statahead pattern 0X0. LustreError: 8795:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8802cd8012a8: cannot apply new layout on [0x240000404:0xa8f:0x0] : rc = -5 LustreError: 8795:0:(lov_object.c:1341:lov_layout_change()) Skipped 2 previous similar messages 10[23078]: segfault at 8 ip 00007f196eec37e8 sp 00007fff65aff4d0 error 4 in ld-2.17.so[7f196eeb8000+22000] LustreError: 29952:0:(mdt_reint.c:2523:mdt_reint_migrate()) lustre-MDT0000: migrate [0x280000404:0xd48:0x0]/14 failed: rc = -114 LustreError: 29952:0:(mdt_reint.c:2523:mdt_reint_migrate()) Skipped 203 previous similar messages LustreError: 23495:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0001: [0x240000403:0xcfb:0x0] migrate mdt count mismatch 3 != 1 LustreError: 8100:0:(mdd_object.c:3902:mdd_close()) lustre-MDD0002: failed to get lu_attr of [0x280000403:0xc84:0x0]: rc = -2 Lustre: 16189:0:(out_handler.c:911:out_tx_end()) lustre-MDT0001-osd: error during execution of #0 from /home/green/git/lustre-release/lustre/ptlrpc/../../lustre/target/out_handler.c:585: rc = -2 Lustre: 16203:0:(mdd_dir.c:4701:mdd_migrate_object()) lustre-MDD0000: [0x200000402:0x1c:0x0]/14 is open, migrate only dentry Lustre: 16203:0:(mdd_dir.c:4701:mdd_migrate_object()) Skipped 42 previous similar messages Lustre: dir [0x200000403:0x13d7:0x0] stripe 2 readdir failed: -2, directory is partially accessed! Lustre: Skipped 15 previous similar messages LustreError: 17443:0:(mdt_open.c:1302:mdt_cross_open()) lustre-MDT0002: [0x280000404:0xc7d:0x0] doesn't exist!: rc = -14 LustreError: 416:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 10 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 17[25604]: segfault at 8 ip 00007fb54ff2f7e8 sp 00007ffd0a1df220 error 4 in ld-2.17.so[7fb54ff24000+22000] LustreError: 22127:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x200000403:0x11ff:0x0] error -5. Lustre: 27580:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000404:0xfde:0x0] with magic=0xbd60bd0 Lustre: 27580:0:(lod_lov.c:1417:lod_parse_striping()) Skipped 45 previous similar messages LustreError: 28573:0:(lcommon_cl.c:179:cl_file_inode_init()) lustre: failed to initialize cl_object [0x280000403:0xf38:0x0]: rc = -5 LustreError: 28573:0:(lcommon_cl.c:179:cl_file_inode_init()) Skipped 70 previous similar messages LustreError: 28573:0:(llite_lib.c:3695:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 28573:0:(llite_lib.c:3695:ll_prep_inode()) Skipped 70 previous similar messages LustreError: 17569:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0002: [0x280000404:0x113a:0x0] migrate mdt count mismatch 1 != 2 LustreError: 23784:0:(mdt_open.c:1302:mdt_cross_open()) lustre-MDT0002: [0x280000404:0xc7d:0x0] doesn't exist!: rc = -14 4[15760]: segfault at 8 ip 00007f67c10be7e8 sp 00007ffd0b2e0ff0 error 4 in ld-2.17.so[7f67c10b3000+22000] LustreError: 19947:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802cd8012a8: inode [0x240000404:0x1278:0x0] mdc close failed: rc = -116 LustreError: 19947:0:(file.c:247:ll_close_inode_openhandle()) Skipped 22 previous similar messages LustreError: 31671:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0000: [0x200000403:0x1703:0x0] migrate mdt count mismatch 2 != 1 12[542]: segfault at 8 ip 00007fc34156b7e8 sp 00007ffc3e7dff90 error 4 in ld-2.17.so[7fc341560000+22000] 5[2165]: segfault at 8 ip 00007f7db86d67e8 sp 00007ffe70337ab0 error 4 in ld-2.17.so[7f7db86cb000+22000] LustreError: 4030:0:(mdd_dir.c:4621:mdd_migrate_cmd_check()) lustre-MDD0001: '17' migration was interrupted, run 'lfs migrate -m 2 -c 1 -H crush 17' to finish migration: rc = -1 LustreError: 4030:0:(mdd_dir.c:4621:mdd_migrate_cmd_check()) Skipped 16 previous similar messages Lustre: lustre-MDT0002: trigger partial OI scrub for RPC inconsistency, checking FID [0x280000403:0x133b:0x0]/0xa): rc = 0 LustreError: 12305:0:(osd_index.c:204:__osd_xattr_load_by_oid()) lustre-MDT0002: can't get bonus, rc = -2 LustreError: 22282:0:(lustre_lmv.h:500:lmv_is_sane()) unknown layout LMV: magic=0xcd40cd0 count=2 index=1 hash=crush:0x82000003 version=1 migrate_offset=1 migrate_hash=fnv_1a_64:2 pool= LustreError: 2103:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0001: [0x240000403:0x15fc:0x0] migrate mdt count mismatch 1 != 3 LustreError: 2103:0:(mdt_xattr.c:402:mdt_dir_layout_update()) Skipped 2 previous similar messages 3[6824]: segfault at 8 ip 00007f2008db57e8 sp 00007fffa67a4dc0 error 4 in ld-2.17.so[7f2008daa000+22000] Lustre: 20575:0:(out_handler.c:911:out_tx_end()) lustre-MDT0000-osd: error during execution of #16 from /home/green/git/lustre-release/lustre/ptlrpc/../../lustre/target/out_handler.c:585: rc = -2 LustreError: 20575:0:(out_handler.c:921:out_tx_end()) lustre-MDT0000-osd: undo for /home/green/git/lustre-release/lustre/ptlrpc/../../lustre/target/out_handler.c:425: rc = -524 LustreError: 20575:0:(out_lib.c:1168:out_tx_index_delete_undo()) lustre-MDT0000-osd: Oops, can not rollback index_delete yet: rc = -524 LustreError: 16773:0:(llite_lib.c:1996:ll_md_setattr()) md_setattr fails: rc = -5 Lustre: mdt00_001: service thread pid 16185 was inactive for 40.037 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 16185, comm: mdt00_001 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock+0x88/0x1c0 [mdt] [<0>] mdt_object_find_lock+0x54/0x170 [mdt] [<0>] mdt_reint_setxattr+0x133/0x10e0 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x73c/0xbb0 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: mdt00_102: service thread pid 17388 was inactive for 40.027 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 17388, comm: mdt00_102 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock_try+0xa0/0x250 [mdt] [<0>] mdt_getattr_name_lock+0x1729/0x2cc0 [mdt] [<0>] mdt_intent_getattr+0x2cc/0x4e0 [mdt] [<0>] mdt_intent_opc.constprop.75+0x211/0xc60 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x397/0x980 [ptlrpc] [<0>] ldlm_handle_enqueue+0x547/0x18d0 [ptlrpc] [<0>] tgt_enqueue+0x68/0x240 [ptlrpc] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe LustreError: 414:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 11 [0x0:0x0:0x0] inode@0000000000000000: rc = -1 LustreError: 414:0:(statahead.c:825:ll_statahead_interpret_work()) Skipped 20 previous similar messages Lustre: dir [0x280000403:0x176e:0x0] stripe 2 readdir failed: -2, directory is partially accessed! Lustre: Skipped 89 previous similar messages LustreError: 18004:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8802c886c138: cannot apply new layout on [0x240000403:0x18a4:0x0] : rc = -5 LustreError: 18004:0:(lov_object.c:1341:lov_layout_change()) Skipped 2 previous similar messages LustreError: 18004:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x240000403:0x18a4:0x0] error -5. LustreError: 15726:0:(ldlm_lockd.c:252:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0002_UUID lock: ffff8802ad974780/0xf77df020e7507350 lrc: 3/0,0 mode: PR/PR res: [0x280000403:0x1567:0x0].0x0 bits 0x13/0x0 rrc: 5 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0xf77df020e7507318 expref: 596 pid: 30594 timeout: 991 lvb_type: 0 LustreError: lustre-MDT0002-mdc-ffff8802cd8012a8: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-MDT0002-mdc-ffff8802cd8012a8: Connection to lustre-MDT0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: mdt00_102: service thread pid 17388 completed after 98.191s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_001: service thread pid 16185 completed after 101.146s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: lustre-MDT0002-mdc-ffff8802cd8012a8: This client was evicted by lustre-MDT0002; in progress operations using this service will fail. LustreError: 25341:0:(llite_lib.c:1996:ll_md_setattr()) md_setattr fails: rc = -5 LustreError: 31331:0:(mdc_request.c:1464:mdc_read_page()) lustre-MDT0002-mdc-ffff8802cd8012a8: [0x280000403:0x1565:0x0] lock enqueue fails: rc = -108 LustreError: 27767:0:(ldlm_resource.c:982:ldlm_resource_complain()) lustre-MDT0002-mdc-ffff8802cd8012a8: namespace resource [0x240000404:0x16da:0x0].0x0 (ffff8802981a4cc0) refcount nonzero (1) after lock cleanup; forcing cleanup. Lustre: lustre-MDT0002-mdc-ffff8802cd8012a8: Connection restored to (at 0@lo) Lustre: 30980:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000403:0x18e7:0x0] with magic=0xbd60bd0 Lustre: 30980:0:(lod_lov.c:1417:lod_parse_striping()) Skipped 79 previous similar messages 12[1466]: segfault at 8 ip 00007fb8b018c7e8 sp 00007fff30b48790 error 4 in ld-2.17.so[7fb8b0181000+22000] Lustre: mdt_io00_010: service thread pid 29913 was inactive for 74.117 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Lustre: mdt_io00_002: service thread pid 16203 was inactive for 74.254 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Pid: 29913, comm: mdt_io00_010 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_fini+0xabc/0xfd0 [ptlrpc] [<0>] ldlm_cli_enqueue+0x461/0xb00 [ptlrpc] [<0>] osp_md_object_lock+0x151/0x2f0 [osp] [<0>] lod_object_lock+0xdb/0x7c0 [lod] [<0>] mdd_object_lock+0x2d/0xd0 [mdd] [<0>] mdt_remote_object_lock_try+0x14c/0x189 [mdt] [<0>] mdt_object_lock_internal+0x3c4/0x470 [mdt] [<0>] mdt_rename_lock+0xc3/0x2d0 [mdt] [<0>] mdt_reint_rename+0x129d/0x2bf0 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x73c/0xbb0 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: mdt_io00_012: service thread pid 4030 was inactive for 74.249 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: mdt_io00_004: service thread pid 29221 was inactive for 74.260 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 1 previous similar message Lustre: mdt_io00_018: service thread pid 21749 was inactive for 74.025 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 2 previous similar messages Lustre: mdt_io00_013: service thread pid 4722 was inactive for 74.170 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 3 previous similar messages Lustre: mdt_io00_017: service thread pid 15748 was inactive for 74.097 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 3 previous similar messages LustreError: 15726:0:(ldlm_lockd.c:252:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0000_UUID lock: ffff880090fd0040/0xf77df020e7632cba lrc: 3/0,0 mode: PR/PR res: [0x200000404:0x208c:0x0].0x0 bits 0x1b/0x0 rrc: 5 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0xf77df020e7632c35 expref: 755 pid: 24152 timeout: 1101 lvb_type: 0 Lustre: mdt_io00_002: service thread pid 16203 completed after 105.235s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: lustre-MDT0000-mdc-ffff8802cd8012a8: operation mds_close to node 0@lo failed: rc = -107 Lustre: lustre-MDT0000-mdc-ffff8802cd8012a8: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: lustre-MDT0000-mdc-ffff8802cd8012a8: This client was evicted by lustre-MDT0000; in progress operations using this service will fail. Lustre: mdt_io00_010: service thread pid 29913 completed after 105.142s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 31950:0:(file.c:6138:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000404:0x208c:0x0] error: rc = -108 LustreError: 904:0:(llite_lib.c:1996:ll_md_setattr()) md_setattr fails: rc = -108 LustreError: 904:0:(llite_lib.c:1996:ll_md_setattr()) Skipped 1 previous similar message Lustre: mdt_io00_012: service thread pid 4030 completed after 104.794s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 6386:0:(ldlm_resource.c:982:ldlm_resource_complain()) lustre-MDT0000-mdc-ffff8802cd8012a8: namespace resource [0x200000403:0x1:0x0].0x0 (ffff8802afd5fc40) refcount nonzero (2) after lock cleanup; forcing cleanup. LustreError: 6386:0:(ldlm_resource.c:982:ldlm_resource_complain()) Skipped 1 previous similar message Lustre: lustre-MDT0000-mdc-ffff8802cd8012a8: Connection restored to (at 0@lo) Lustre: mdt_io00_015: service thread pid 5116 completed after 104.653s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_004: service thread pid 29221 completed after 103.279s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_019: service thread pid 21889 completed after 102.893s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_008: service thread pid 29776 completed after 103.030s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_018: service thread pid 21749 completed after 101.803s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_000: service thread pid 16201 completed after 101.123s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_020: service thread pid 22282 completed after 100.099s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_001: service thread pid 16202 completed after 99.653s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_013: service thread pid 4722 completed after 97.771s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_006: service thread pid 29553 completed after 97.607s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_021: service thread pid 13009 completed after 97.555s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_011: service thread pid 29952 completed after 96.437s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_017: service thread pid 15748 completed after 89.141s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_009: service thread pid 29848 completed after 88.242s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_014: service thread pid 4739 completed after 87.643s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). | Link to test |
racer test 1: racer on clients: centos-30.localnet DURATION=2700 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 PGD 2bd66b067 PUD 2b0662067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) osp(OE) ofd(OE) lod(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) crc32_generic lnet(OE) libcfs(OE) crc_t10dif crct10dif_generic crct10dif_common virtio_console virtio_balloon i2c_piix4 pcspkr ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm drm_panel_orientation_quirks ata_piix serio_raw virtio_blk libata i2c_core floppy CPU: 3 PID: 11329 Comm: ll_sa_11312 Kdump: loaded Tainted: P OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.16.0-3.module_el8.7.0+1218+f626c2ff 04/01/2014 task: ffff8802d6295c40 ti: ffff88028b2a0000 task.ti: ffff88028b2a0000 RIP: 0010:[<ffffffff813f502c>] [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 RSP: 0018:ffff88028b2a3d78 EFLAGS: 00010296 RAX: ffff880270a14548 RBX: 0000000000000008 RCX: 0000000100260022 RDX: 0000000000000026 RSI: ffff880270a14878 RDI: 0000000000000008 RBP: ffff88028b2a3d88 R08: ffff8802d47eb508 R09: 0000000000000000 R10: ffff8802d47ee018 R11: ffff8802d47eb508 R12: 0000000000000000 R13: ffff8802c5ed5c40 R14: ffff880270a14878 R15: ffff8802d47eb508 FS: 0000000000000000(0000) GS:ffff880331ac0000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 00000002aa184000 CR4: 00000000000007e0 Call Trace: [<ffffffffa15b5621>] ll_statahead_thread+0xe01/0x22c0 [lustre] [<ffffffff817e320e>] ? _raw_spin_lock+0x1e/0x20 [<ffffffff810dbb51>] ? put_prev_entity+0x31/0x400 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e329e>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffff810c834d>] ? finish_task_switch+0x5d/0x1b0 [<ffffffffa15b4820>] ? ll_statahead_handle.constprop.29+0x1c0/0x1c0 [lustre] [<ffffffff810ba114>] kthread+0xe4/0xf0 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 | 17[22578]: segfault at 8 ip 00007fcb3e56c7e8 sp 00007ffde3705cf0 error 4 in ld-2.17.so[7fcb3e561000+22000] 18[24245]: segfault at 8 ip 00007f014ce767e8 sp 00007fffe8aa4980 error 4 in ld-2.17.so[7f014ce6b000+22000] 12[24474]: segfault at 8 ip 00007f84fe0f57e8 sp 00007ffe44cdc930 error 4 in ld-2.17.so[7f84fe0ea000+22000] Lustre: 26452:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff8802d69bad40 x1827881836261376/t4294971306(0) o101->ca3d9020-391f-43c1-88d4-b5d54c6de812@0@lo:416/0 lens 376/840 e 0 to 0 dl 1743204061 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 LustreError: 26485:0:(mdt_handler.c:746:mdt_pack_acl2body()) lustre-MDT0000: unable to read [0x200000401:0x150:0x0] ACL: rc = -2 9[17718]: segfault at 8 ip 00007f18f5ce07e8 sp 00007ffc8dd097e0 error 4 in ld-2.17.so[7f18f5cd5000+22000] Lustre: 22780:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000402:0x989:0x0] with magic=0xbd60bd0 4[23921]: segfault at 8 ip 00007f6d488f47e8 sp 00007ffcfbe007c0 error 4 in ld-2.17.so[7f6d488e9000+22000] 3[24387]: segfault at 8 ip 00007f0ebda387e8 sp 00007fff7edd7d20 error 4 in ld-2.17.so[7f0ebda2d000+22000] LustreError: 23921:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a34b2548: inode [0x200000402:0x9d5:0x0] mdc close failed: rc = -13 Lustre: 22784:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000401:0xba6:0x0] with magic=0xbd60bd0 Lustre: 22784:0:(lod_lov.c:1417:lod_parse_striping()) Skipped 1 previous similar message LustreError: 23351:0:(osd_io.c:898:osd_grow_blocksize()) object [0x2c0000400:0xf3:0x0]: change block size4096 -> 1048576 error rc = -95 Lustre: 26452:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000401:0xd9f:0x0] with magic=0xbd60bd0 Lustre: 26452:0:(lod_lov.c:1417:lod_parse_striping()) Skipped 1 previous similar message LustreError: 15965:0:(statahead.c:2447:start_statahead_thread()) lustre: unsupported statahead pattern 0X0. LustreError: 26265:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a34b2548: inode [0x200000401:0xfb8:0x0] mdc close failed: rc = -13 19[13935]: segfault at 8 ip 00007f810d01e7e8 sp 00007fff20d48e90 error 4 in ld-2.17.so[7f810d013000+22000] LustreError: 22784:0:(mdt_handler.c:746:mdt_pack_acl2body()) lustre-MDT0000: unable to read [0x200000401:0x12c4:0x0] ACL: rc = -2 LustreError: 22976:0:(mdt_handler.c:746:mdt_pack_acl2body()) lustre-MDT0000: unable to read [0x200000401:0x1321:0x0] ACL: rc = -2 8[5177]: segfault at 8 ip 00007f0f239947e8 sp 00007fff024a9ab0 error 4 in ld-2.17.so[7f0f23989000+22000] 8[6387]: segfault at 8 ip 00007f631babd7e8 sp 00007fff20ed5520 error 4 in ld-2.17.so[7f631bab2000+22000] 11[10591]: segfault at 8 ip 00007fae223eb7e8 sp 00007ffd94ddda90 error 4 in ld-2.17.so[7fae223e0000+22000] Lustre: mdt00_012: service thread pid 22984 was inactive for 40.059 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 22984, comm: mdt00_012 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock+0x88/0x1c0 [mdt] [<0>] mdt_object_find_lock+0x54/0x170 [mdt] [<0>] mdt_reint_setxattr+0x133/0x10e0 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x73c/0xbb0 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 26452, comm: mdt00_016 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock_try+0xa0/0x250 [mdt] [<0>] mdt_getattr_name_lock+0x1729/0x2cc0 [mdt] [<0>] mdt_intent_getattr+0x2cc/0x4e0 [mdt] [<0>] mdt_intent_opc.constprop.75+0x211/0xc60 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x397/0x980 [ptlrpc] [<0>] ldlm_handle_enqueue+0x547/0x18d0 [ptlrpc] [<0>] tgt_enqueue+0x68/0x240 [ptlrpc] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 13553, comm: mdt00_003 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] Lustre: mdt00_004: service thread pid 21716 was inactive for 40.141 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. [<0>] mdt_object_lock_try+0xa0/0x250 [mdt] [<0>] mdt_getattr_name_lock+0x1729/0x2cc0 [mdt] [<0>] mdt_intent_getattr+0x2cc/0x4e0 [mdt] [<0>] mdt_intent_opc.constprop.75+0x211/0xc60 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x397/0x980 [ptlrpc] [<0>] ldlm_handle_enqueue+0x547/0x18d0 [ptlrpc] [<0>] tgt_enqueue+0x68/0x240 [ptlrpc] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: mdt00_010: service thread pid 22959 was inactive for 40.091 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 5 previous similar messages Lustre: mdt00_021: service thread pid 12765 was inactive for 40.038 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 2 previous similar messages LustreError: 10992:0:(ldlm_lockd.c:252:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0000_UUID lock: ffff88008f1161c0/0x8a412acd73357cb8 lrc: 3/0,0 mode: PR/PR res: [0x200000402:0x172b:0x0].0x0 bits 0x13/0x0 rrc: 39 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0x8a412acd73357c95 expref: 703 pid: 11268 timeout: 567 lvb_type: 0 LustreError: 22780:0:(ldlm_lockd.c:1492:ldlm_handle_enqueue()) ### lock on destroyed export ffff8802dd00d3d8 ns: mdt-lustre-MDT0000_UUID lock: ffff88008ae63880/0x8a412acd73359da4 lrc: 3/0,0 mode: PR/PR res: [0x200000402:0x172b:0x0].0x0 bits 0x1b/0x0 rrc: 36 type: IBT gid 0 flags: 0x50200400000020 nid: 0@lo remote: 0x8a412acd73359d96 expref: 583 pid: 22780 timeout: 0 lvb_type: 0 Lustre: mdt00_016: service thread pid 26452 completed after 99.847s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_007: service thread pid 22780 completed after 99.651s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_012: service thread pid 22984 completed after 100.097s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: lustre-MDT0000-mdc-ffff8800a34b2548: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-MDT0000-mdc-ffff8800a34b2548: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: mdt00_015: service thread pid 26450 completed after 99.343s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_011: service thread pid 22976 completed after 99.676s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_004: service thread pid 21716 completed after 99.793s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_017: service thread pid 26485 completed after 99.638s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_014: service thread pid 26431 completed after 99.392s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: lustre-MDT0000-mdc-ffff8800a34b2548: This client was evicted by lustre-MDT0000; in progress operations using this service will fail. LustreError: 11864:0:(file.c:6138:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000402:0x172b:0x0] error: rc = -5 Lustre: mdt00_003: service thread pid 13553 completed after 99.810s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 14248:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a34b2548: inode [0x200000401:0x161f:0x0] mdc close failed: rc = -108 LustreError: 13484:0:(mdc_request.c:1464:mdc_read_page()) lustre-MDT0000-mdc-ffff8800a34b2548: [0x200000401:0x1:0x0] lock enqueue fails: rc = -108 Lustre: mdt00_010: service thread pid 22959 completed after 99.028s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 13659:0:(llite_lib.c:1996:ll_md_setattr()) md_setattr fails: rc = -108 Lustre: mdt00_019: service thread pid 6581 completed after 97.336s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_013: service thread pid 24905 completed after 96.782s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_000: service thread pid 11266 completed after 98.946s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_021: service thread pid 12765 completed after 97.922s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_020: service thread pid 12734 completed after 96.788s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_006: service thread pid 22070 completed after 97.651s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_002: service thread pid 11268 completed after 99.014s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_005: service thread pid 21941 completed after 96.805s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 11473:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x200000402:0x172b:0x0] error -108. Lustre: lustre-MDT0000-mdc-ffff8800a34b2548: Connection restored to 192.168.123.32@tcp (at 0@lo) 11[16099]: segfault at 8 ip 00007f76105e87e8 sp 00007ffe18fc7fb0 error 4 in ld-2.17.so[7f76105dd000+22000] 14[20364]: segfault at 8 ip 00007ff17099d7e8 sp 00007fff2c404dd0 error 4 in ld-2.17.so[7ff170992000+22000] LustreError: 26286:0:(statahead.c:2447:start_statahead_thread()) lustre: unsupported statahead pattern 0X0. Lustre: 22984:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0x18f:0x0] with magic=0xbd60bd0 Lustre: 22984:0:(lod_lov.c:1417:lod_parse_striping()) Skipped 1 previous similar message LustreError: 28358:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a34b12a8: inode [0x200000403:0x130:0x0] mdc close failed: rc = -13 LustreError: 28358:0:(file.c:247:ll_close_inode_openhandle()) Skipped 9 previous similar messages 8[13418]: segfault at 8 ip 00007f633c0cf7e8 sp 00007ffde5ebdff0 error 4 in ld-2.17.so[7f633c0c4000+22000] Lustre: 22788:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0x550:0x0] with magic=0xbd60bd0 Lustre: 22788:0:(lod_lov.c:1417:lod_parse_striping()) Skipped 1 previous similar message LustreError: 26431:0:(mdt_handler.c:746:mdt_pack_acl2body()) lustre-MDT0000: unable to read [0x200000403:0x5ea:0x0] ACL: rc = -2 LustreError: 485:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a34b2548: inode [0x200000402:0x1ddf:0x0] mdc close failed: rc = -13 2[28519]: segfault at 8 ip 00007f53533f67e8 sp 00007ffff4bc9f80 error 4 in ld-2.17.so[7f53533eb000+22000] Lustre: 22788:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0xbde:0x0] with magic=0xbd60bd0 Lustre: 22788:0:(lod_lov.c:1417:lod_parse_striping()) Skipped 1 previous similar message 19[6979]: segfault at 4017b0 ip 00000000004017b0 sp 00007ffefe64e200 error 14 ptlrpc_watchdog_fire: 14 callbacks suppressed Lustre: mdt_out00_005: service thread pid 366 was inactive for 40.107 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 5 previous similar messages Lustre: mdt_io00_003: service thread pid 23054 was inactive for 40.106 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Lustre: Skipped 2 previous similar messages Pid: 23054, comm: mdt_io00_003 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock+0x88/0x1c0 [mdt] [<0>] mdt_rename_source_lock+0xa9/0xd6 [mdt] [<0>] mdt_reint_rename+0x1da3/0x2bf0 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x73c/0xbb0 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe LustreError: 10992:0:(ldlm_lockd.c:252:expired_lock_main()) ### lock callback timer expired after 101s: evicting client at 0@lo ns: mdt-lustre-MDT0000_UUID lock: ffff8802aa7852c0/0x8a412acd735b1595 lrc: 3/0,0 mode: PR/PR res: [0x200000402:0x2785:0x0].0x0 bits 0x13/0x0 rrc: 3 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0x8a412acd735b153a expref: 575 pid: 366 timeout: 960 lvb_type: 0 Lustre: mdt_io00_003: service thread pid 23054 completed after 100.297s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 22984:0:(ldlm_lockd.c:1492:ldlm_handle_enqueue()) ### lock on destroyed export ffff8802c31a1bf8 ns: mdt-lustre-MDT0000_UUID lock: ffff8802b197ad40/0x8a412acd735b1b76 lrc: 3/0,0 mode: PR/PR res: [0x200000402:0x2785:0x0].0x0 bits 0x13/0x0 rrc: 2 type: IBT gid 0 flags: 0x50200000000000 nid: 0@lo remote: 0x8a412acd735b1659 expref: 5 pid: 22984 timeout: 0 lvb_type: 0 LustreError: 22984:0:(ldlm_lockd.c:1492:ldlm_handle_enqueue()) Skipped 7 previous similar messages Lustre: mdt_out00_005: service thread pid 366 completed after 100.312s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: lustre-MDT0000-mdc-ffff8800a34b2548: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: lustre-MDT0000-mdc-ffff8800a34b2548: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 2 previous similar messages LustreError: lustre-MDT0000-mdc-ffff8800a34b2548: This client was evicted by lustre-MDT0000; in progress operations using this service will fail. LustreError: 12991:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a34b2548: inode [0x200000401:0x1:0x0] mdc close failed: rc = -108 LustreError: 12991:0:(file.c:247:ll_close_inode_openhandle()) Skipped 2 previous similar messages LustreError: 11685:0:(file.c:6138:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000007:0x1:0x0] error: rc = -108 LustreError: 11685:0:(file.c:6138:ll_inode_revalidate_fini()) Skipped 16 previous similar messages Lustre: lustre-MDT0000-mdc-ffff8800a34b2548: Connection restored to 192.168.123.32@tcp (at 0@lo) Lustre: 26450:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000402:0x2983:0x0] with magic=0xbd60bd0 Lustre: 26450:0:(lod_lov.c:1417:lod_parse_striping()) Skipped 1 previous similar message Lustre: 13553:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000404:0x339:0x0] with magic=0xbd60bd0 Lustre: 13553:0:(lod_lov.c:1417:lod_parse_striping()) Skipped 3 previous similar messages LustreError: 28991:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a34b2548: inode [0x200000404:0x6cc:0x0] mdc close failed: rc = -13 LustreError: 28991:0:(file.c:247:ll_close_inode_openhandle()) Skipped 5 previous similar messages 14[2070]: segfault at 8 ip 00007fb2b03d87e8 sp 00007ffe85a4b330 error 4 in ld-2.17.so[7fb2b03cd000+22000] Lustre: 11266:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000404:0x7e4:0x0] with magic=0xbd60bd0 Lustre: 11266:0:(lod_lov.c:1417:lod_parse_striping()) Skipped 1 previous similar message 17[23119]: segfault at 8 ip 00007f3538b3a7e8 sp 00007ffc3b7b72d0 error 4 in ld-2.17.so[7f3538b2f000+22000] LustreError: 23119:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a34b2548: inode [0x200000402:0x31ce:0x0] mdc close failed: rc = -13 LustreError: 23119:0:(file.c:247:ll_close_inode_openhandle()) Skipped 1 previous similar message LustreError: 22976:0:(mdt_handler.c:746:mdt_pack_acl2body()) lustre-MDT0000: unable to read [0x200000404:0xac1:0x0] ACL: rc = -2 LustreError: 10992:0:(ldlm_lockd.c:252:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0000_UUID lock: ffff88028c6434c0/0x8a412acd7380a58f lrc: 3/0,0 mode: PR/PR res: [0x200000404:0x1048:0x0].0x0 bits 0x13/0x0 rrc: 13 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0x8a412acd7380a581 expref: 494 pid: 22780 timeout: 1330 lvb_type: 0 LustreError: 22788:0:(ldlm_lockd.c:1492:ldlm_handle_enqueue()) ### lock on destroyed export ffff8802b2275d28 ns: mdt-lustre-MDT0000_UUID lock: ffff8802b4b187c0/0x8a412acd7380b771 lrc: 3/0,0 mode: PR/PR res: [0x200000404:0x1048:0x0].0x0 bits 0x1b/0x0 rrc: 9 type: IBT gid 0 flags: 0x50200400000020 nid: 0@lo remote: 0x8a412acd7380b763 expref: 213 pid: 22788 timeout: 0 lvb_type: 0 LustreError: lustre-MDT0000-mdc-ffff8800a34b2548: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-MDT0000-mdc-ffff8800a34b2548: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: lustre-MDT0000-mdc-ffff8800a34b2548: This client was evicted by lustre-MDT0000; in progress operations using this service will fail. LustreError: 8642:0:(file.c:6138:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000404:0x1048:0x0] error: rc = -5 LustreError: 8642:0:(file.c:6138:ll_inode_revalidate_fini()) Skipped 10 previous similar messages LustreError: 8462:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x200000404:0x1048:0x0] error -5. LustreError: 8462:0:(vvp_io.c:1903:vvp_io_init()) Skipped 2 previous similar messages LustreError: 9662:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a34b2548: inode [0x200000402:0x391e:0x0] mdc close failed: rc = -108 LustreError: 22788:0:(ldlm_lockd.c:1492:ldlm_handle_enqueue()) Skipped 1 previous similar message Lustre: lustre-MDT0000-mdc-ffff8800a34b2548: Connection restored to 192.168.123.32@tcp (at 0@lo) Lustre: 22070:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000405:0x119:0x0] with magic=0xbd60bd0 Lustre: 22070:0:(lod_lov.c:1417:lod_parse_striping()) Skipped 7 previous similar messages 12[24968]: segfault at 8 ip 00007f694dc727e8 sp 00007ffcd6c70730 error 4 in ld-2.17.so[7f694dc67000+22000] 9[27990]: segfault at 4017b0 ip 00000000004017b0 sp 00007ffc7b625fc0 error 14 LustreError: 32401:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a34b12a8: inode [0x200000402:0x40a2:0x0] mdc close failed: rc = -13 LustreError: 32401:0:(file.c:247:ll_close_inode_openhandle()) Skipped 17 previous similar messages 7[2001]: segfault at 8 ip 00007f8a166e97e8 sp 00007fff87d19f90 error 4 in ld-2.17.so[7f8a166de000+22000] 17[1055]: segfault at 8 ip 00007fa9fda427e8 sp 00007ffc412c3970 error 4 in ld-2.17.so[7fa9fda37000+22000] Lustre: 22959:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000405:0x101e:0x0] with magic=0xbd60bd0 Lustre: 22959:0:(lod_lov.c:1417:lod_parse_striping()) Skipped 9 previous similar messages | Link to test |
racer test 2: racer rename: centos-15.localnet DURATION=2700 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 PGD 2db8d0067 PUD 2d6ded067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) lustre(OE) osp(OE) ofd(OE) lod(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) jbd2 mbcache lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) crc32_generic lnet(OE) dm_flakey dm_mod libcfs(OE) crc_t10dif crct10dif_generic crct10dif_common virtio_balloon virtio_console pcspkr i2c_piix4 ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm ata_piix drm_panel_orientation_quirks floppy serio_raw virtio_blk libata i2c_core CPU: 14 PID: 9742 Comm: ll_sa_9657 Kdump: loaded Tainted: P OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.16.0-3.module_el8.7.0+1218+f626c2ff 04/01/2014 task: ffff880254095c40 ti: ffff8802147d8000 task.ti: ffff8802147d8000 RIP: 0010:[<ffffffff813f502c>] [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 RSP: 0018:ffff8802147dbd78 EFLAGS: 00010296 RAX: ffff8800535f0ac8 RBX: 0000000000000008 RCX: dead000000000200 RDX: 000000000000000e RSI: ffff8800535f0df8 RDI: 0000000000000008 RBP: ffff8802147dbd88 R08: ffff880331403660 R09: 0000000000000000 R10: ffff880331403640 R11: ffff880205335628 R12: 0000000000000000 R13: ffff8802d4b98010 R14: ffff8800535f0df8 R15: ffff880205335628 FS: 0000000000000000(0000) GS:ffff880331d80000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 00000002d6dc8000 CR4: 00000000000007e0 Call Trace: [<ffffffffa10eb631>] ll_statahead_thread+0xe01/0x22c0 [lustre] [<ffffffff810df6e2>] ? pick_next_task_fair+0x702/0x8a0 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e329e>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffff810c834d>] ? finish_task_switch+0x5d/0x1b0 [<ffffffffa10ea830>] ? ll_statahead_handle.constprop.29+0x1c0/0x1c0 [lustre] [<ffffffff810ba114>] kthread+0xe4/0xf0 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 | Lustre: 7910:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff8802b2f4b740 x1827562232325760/t4295079582(0) o101->e3f32f46-7f2b-47b9-b6d8-4f9caea418d0@0@lo:240/0 lens 376/27440 e 0 to 0 dl 1742901885 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 Lustre: 17248:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff8800965ca340 x1827562233476864/t4295071653(0) o101->e3f32f46-7f2b-47b9-b6d8-4f9caea418d0@0@lo:244/0 lens 376/37448 e 0 to 0 dl 1742901889 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 Lustre: 17248:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff8802e20d0540 x1827562239235712/t4295080334(0) o101->19d5422d-541e-4da0-803c-0e442c636d68@0@lo:264/0 lens 376/35344 e 0 to 0 dl 1742901909 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 Lustre: 17248:0:(mdt_recovery.c:102:mdt_req_from_lrd()) Skipped 1 previous similar message Lustre: 7894:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff88028687d540 x1827562239908224/t4295184183(0) o101->e3f32f46-7f2b-47b9-b6d8-4f9caea418d0@0@lo:268/0 lens 376/39272 e 0 to 0 dl 1742901913 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 Lustre: 17296:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff880298eb6440 x1827562242486016/t4295080716(0) o101->e3f32f46-7f2b-47b9-b6d8-4f9caea418d0@0@lo:276/0 lens 376/35344 e 0 to 0 dl 1742901921 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 Lustre: 17296:0:(mdt_recovery.c:102:mdt_req_from_lrd()) Skipped 1 previous similar message Lustre: 17170:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff8802d0f8a340 x1827562247414528/t4295184988(0) o101->19d5422d-541e-4da0-803c-0e442c636d68@0@lo:292/0 lens 376/44128 e 0 to 0 dl 1742901937 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 Lustre: 17170:0:(mdt_recovery.c:102:mdt_req_from_lrd()) Skipped 2 previous similar messages Lustre: 16620:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff88025ddc1440 x1827562257702656/t4295081951(0) o101->e3f32f46-7f2b-47b9-b6d8-4f9caea418d0@0@lo:318/0 lens 376/42496 e 0 to 0 dl 1742901963 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 Lustre: 16620:0:(mdt_recovery.c:102:mdt_req_from_lrd()) Skipped 3 previous similar messages Lustre: 7907:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff8802546fda40 x1827562271840384/t4295075051(0) o101->19d5422d-541e-4da0-803c-0e442c636d68@0@lo:356/0 lens 376/47656 e 0 to 0 dl 1742902001 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 Lustre: 7907:0:(mdt_recovery.c:102:mdt_req_from_lrd()) Skipped 3 previous similar messages Lustre: 1192:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0000: opcode 0: before 513 < left 1509, rollback = 0 Lustre: 1192:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 1638 previous similar messages LustreError: 9779:0:(ldlm_lockd.c:252:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0003_UUID lock: ffff880263a83100/0xeaec6d3c1d267644 lrc: 3/0,0 mode: PW/PW res: [0x380000401:0x13ca:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->3145727) gid 0 flags: 0x60000400000020 nid: 0@lo remote: 0xeaec6d3c1d2675d4 expref: 3264 pid: 29854 timeout: 3094 lvb_type: 0 LustreError: lustre-OST0003-osc-ffff8802e6e9a548: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 1 previous similar message Lustre: lustre-OST0003-osc-ffff8802e6e9a548: Connection to lustre-OST0003 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 19523:0:(client.c:1282:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802e609e940 x1827562289524864/t0(0) o105->lustre-OST0003@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' uid:4294967295 gid:4294967295 LustreError: lustre-OST0003-osc-ffff8802e6e9a548: This client was evicted by lustre-OST0003; in progress operations using this service will fail. Lustre: 3954:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x280000407:0x33b1:0x0]/ may get corrupted (rc -108) Lustre: 3954:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x280000407:0x3406:0x0]/ may get corrupted (rc -108) Lustre: 3961:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x240000408:0x11db:0x0]/ may get corrupted (rc -108) Lustre: 3961:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x280000407:0x3444:0x0]/ may get corrupted (rc -108) Lustre: 3951:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x280000407:0x3434:0x0]/ may get corrupted (rc -108) Lustre: 3959:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x240000408:0x11a9:0x0]/ may get corrupted (rc -108) Lustre: 3959:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x240000407:0x2282:0x0]// may get corrupted (rc -108) Lustre: 3956:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x240000407:0x2254:0x0]/ may get corrupted (rc -108) Lustre: 3956:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x280000407:0x341e:0x0]/ may get corrupted (rc -108) Lustre: lustre-OST0003-osc-ffff8802e6e9a548: Connection restored to (at 0@lo) Lustre: 7888:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff880299863240 x1827562296792832/t4295077271(0) o101->e3f32f46-7f2b-47b9-b6d8-4f9caea418d0@0@lo:429/0 lens 376/47656 e 0 to 0 dl 1742902074 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 Lustre: 7888:0:(mdt_recovery.c:102:mdt_req_from_lrd()) Skipped 6 previous similar messages Lustre: 17779:0:(osd_io.c:1951:osd_ldiskfs_write()) lustre-MDT0000: adding bh without locking off 679936 (block 166, size 1976, offs 679112) Lustre: 17960:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 6/24/0, destroy: 1/4/0 Lustre: 17960:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 1261 previous similar messages Lustre: 17960:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1994/1994/0, xattr_set: 2991/27916/0 Lustre: 17960:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 1261 previous similar messages Lustre: 17960:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 28/157/7, punch: 0/0/0, quota 1/3/0 Lustre: 17960:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 1261 previous similar messages Lustre: 17960:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 6/102/0, delete: 2/5/1 Lustre: 17960:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 1261 previous similar messages Lustre: 17960:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 1/1/1 Lustre: 17960:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 1261 previous similar messages LustreError: 9779:0:(ldlm_lockd.c:252:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0000_UUID lock: ffff880260fc3c40/0xeaec6d3c1d45a4eb lrc: 3/0,0 mode: PW/PW res: [0x2c0000400:0x36cb:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000080030020 nid: 0@lo remote: 0xeaec6d3c1d45a4dd expref: 4923 pid: 29800 timeout: 3231 lvb_type: 0 LustreError: 12861:0:(ldlm_lockd.c:2591:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1742902051 with bad export cookie 16928025204311134711 Lustre: lustre-OST0000-osc-ffff8802e6e9a548: Connection to lustre-OST0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 9777:0:(client.c:1282:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802b2c08a40 x1827562340803200/t0(0) o105->lustre-OST0000@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' uid:4294967295 gid:4294967295 LustreError: 9777:0:(client.c:1282:ptlrpc_import_delay_req()) Skipped 1 previous similar message LustreError: 2613:0:(ldlm_lockd.c:1492:ldlm_handle_enqueue()) ### lock on destroyed export ffff8802e6278008 ns: filter-lustre-OST0000_UUID lock: ffff88025577b880/0xeaec6d3c1d5f60c4 lrc: 3/0,0 mode: PW/PW res: [0x2c0000400:0x4ac1:0x0].0x0 rrc: 2 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x50000000020000 nid: 0@lo remote: 0xeaec6d3c1d5f60bd expref: 4778 pid: 2613 timeout: 0 lvb_type: 0 LustreError: 2613:0:(ldlm_lockd.c:1492:ldlm_handle_enqueue()) Skipped 8 previous similar messages LustreError: lustre-OST0000-osc-ffff8802e6e9a548: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: lustre-OST0000-osc-ffff8802e6e9a548: This client was evicted by lustre-OST0000; in progress operations using this service will fail. LustreError: Skipped 1 previous similar message LustreError: 29777:0:(client.c:1282:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88024b001440 x1827562340965120/t0(0) o104->lustre-OST0003@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' uid:4294967295 gid:4294967295 LustreError: 29777:0:(client.c:1282:ptlrpc_import_delay_req()) Skipped 15 previous similar messages LustreError: lustre-OST0003-osc-ffff8802e6e9a548: This client was evicted by lustre-OST0003; in progress operations using this service will fail. Lustre: 3953:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x280000407:0x3545:0x0]/ may get corrupted (rc -108) Lustre: 3960:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x200000407:0x43e:0x0]/ may get corrupted (rc -108) Lustre: 3953:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x240000407:0x23ec:0x0]/ may get corrupted (rc -108) Lustre: 3954:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x280000407:0x356b:0x0]/ may get corrupted (rc -108) Lustre: 3957:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x240000407:0x24dc:0x0]/ may get corrupted (rc -108) Lustre: 3957:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x280000407:0x35d2:0x0]// may get corrupted (rc -108) Lustre: 3959:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x200000406:0x4cd7:0x0]/ may get corrupted (rc -108) Lustre: 3959:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x240000407:0x24f2:0x0]// may get corrupted (rc -108) Lustre: 3955:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x280000407:0x35c1:0x0]/ may get corrupted (rc -108) Lustre: 3957:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x28000040a:0x4fe:0x0]/ may get corrupted (rc -5) Lustre: 3965:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x28000040a:0x46d:0x0]/ may get corrupted (rc -5) Lustre: 3962:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x240000407:0x2361:0x0]// may get corrupted (rc -5) Lustre: 3967:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x28000040a:0x475:0x0]/ may get corrupted (rc -5) Lustre: 3962:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x200000407:0x46b:0x0]// may get corrupted (rc -108) Lustre: 3959:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x200000407:0x443:0x0]/ may get corrupted (rc -108) Lustre: 3962:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x240000407:0x2512:0x0]// may get corrupted (rc -108) Lustre: 3961:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x240000407:0x250b:0x0]/ may get corrupted (rc -108) Lustre: 3961:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x200000407:0x42d:0x0]/ may get corrupted (rc -108) Lustre: 3964:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x200000407:0x474:0x0]/ may get corrupted (rc -108) Lustre: 3964:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x200000407:0x43b:0x0]// may get corrupted (rc -108) Lustre: 3961:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x240000408:0x13d1:0x0]/ may get corrupted (rc -108) Lustre: 3965:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x200000406:0x4ce7:0x0]/ may get corrupted (rc -108) LustreError: 29303:0:(client.c:1282:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88008395b240 x1827562341894400/t0(0) o104->lustre-OST0000@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' uid:4294967295 gid:4294967295 LustreError: 29303:0:(client.c:1282:ptlrpc_import_delay_req()) Skipped 30 previous similar messages LustreError: 18784:0:(ldlm_resource.c:982:ldlm_resource_complain()) lustre-OST0003-osc-ffff8802e6e9a548: namespace resource [0x380000401:0x366f:0x0].0x0 (ffff8802c37b8cc0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 18784:0:(ldlm_resource.c:982:ldlm_resource_complain()) Skipped 3 previous similar messages Lustre: lustre-OST0000-osc-ffff8802e6e9a548: Connection restored to (at 0@lo) Lustre: 17603:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff88008004ee40 x1827562347781248/t4295089746(0) o101->e3f32f46-7f2b-47b9-b6d8-4f9caea418d0@0@lo:559/0 lens 376/47528 e 0 to 0 dl 1742902204 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 Lustre: 17603:0:(mdt_recovery.c:102:mdt_req_from_lrd()) Skipped 7 previous similar messages mrename (24179) used greatest stack depth: 9648 bytes left LustreError: 9779:0:(ldlm_lockd.c:252:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0003_UUID lock: ffff8802b50c0f40/0xeaec6d3c1d6adc15 lrc: 3/0,0 mode: PW/PW res: [0x380000400:0x606e:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000400000020 nid: 0@lo remote: 0xeaec6d3c1d6a5bd7 expref: 5014 pid: 2571 timeout: 3372 lvb_type: 0 LustreError: 9779:0:(ldlm_lockd.c:252:expired_lock_main()) Skipped 1 previous similar message LustreError: 2267:0:(ldlm_lockd.c:2591:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1742902192 with bad export cookie 16928025204337119054 LustreError: 2267:0:(ldlm_lockd.c:2591:ldlm_cancel_handler()) Skipped 1 previous similar message Lustre: lustre-OST0003-osc-ffff8802e6e9a548: Connection to lustre-OST0003 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 1 previous similar message LustreError: 4897:0:(client.c:1282:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88027af5ee40 x1827562383205120/t0(0) o104->lustre-OST0003@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' uid:4294967295 gid:4294967295 LustreError: 4897:0:(client.c:1282:ptlrpc_import_delay_req()) Skipped 4 previous similar messages LustreError: lustre-OST0003-osc-ffff8802e6e9a548: This client was evicted by lustre-OST0003; in progress operations using this service will fail. Lustre: 3966:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x200000407:0x51e:0x0]/ may get corrupted (rc -108) Lustre: 3965:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x200000407:0x5e1:0x0]// may get corrupted (rc -108) Lustre: 3954:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x240000407:0x267a:0x0]/ may get corrupted (rc -108) Lustre: 3954:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x200000406:0x4eb1:0x0]/ may get corrupted (rc -108) Lustre: 3951:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x280000407:0x36c3:0x0]/ may get corrupted (rc -108) Lustre: 3951:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x200000406:0x4e02:0x0]/ may get corrupted (rc -108) Lustre: 3966:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x200000407:0x581:0x0]/ may get corrupted (rc -108) Lustre: lustre-OST0003-osc-ffff8802e6e9a548: Connection restored to (at 0@lo) Lustre: Skipped 1 previous similar message LustreError: 17603:0:(mdt_open.c:1303:mdt_cross_open()) lustre-MDT0002: [0x280000407:0x3799:0x0] doesn't exist!: rc = -14 Lustre: lustre-OST0003-osc-MDT0002: update sequence from 0x380000400 to 0x380000403 Lustre: 9873:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff8802a0c48040 x1827562427832576/t4295198725(0) o101->e3f32f46-7f2b-47b9-b6d8-4f9caea418d0@0@lo:3/0 lens 376/47656 e 0 to 0 dl 1742902403 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 Lustre: 9873:0:(mdt_recovery.c:102:mdt_req_from_lrd()) Skipped 12 previous similar messages LustreError: 5132:0:(client.c:1282:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff880230f5b740 x1827562430574336/t0(0) o104->lustre-OST0002@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' uid:4294967295 gid:4294967295 LustreError: lustre-OST0001-osc-ffff8802e6e9a548: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: lustre-OST0001-osc-ffff8802e6e9a548: This client was evicted by lustre-OST0001; in progress operations using this service will fail. LustreError: lustre-OST0002-osc-ffff8802e6e9a548: This client was evicted by lustre-OST0002; in progress operations using this service will fail. Lustre: 3963:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x28000040a:0x665:0x0]/ may get corrupted (rc -108) Lustre: 3960:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x28000040a:0x69e:0x0]/ may get corrupted (rc -108) Lustre: 3951:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x200000407:0x665:0x0]/ may get corrupted (rc -108) Lustre: 3951:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x280000407:0x3833:0x0]/ may get corrupted (rc -108) Lustre: 3960:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x280000407:0x36c5:0x0]/ may get corrupted (rc -108) Lustre: 3956:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x240000407:0x2770:0x0]/ may get corrupted (rc -108) Lustre: 3956:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x200000407:0x71b:0x0]// may get corrupted (rc -108) Lustre: 3955:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x200000407:0x6f3:0x0]/ may get corrupted (rc -108) Lustre: 3955:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x240000407:0x273d:0x0]/ may get corrupted (rc -108) Lustre: 3963:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x280000407:0x371e:0x0]/ may get corrupted (rc -108) Lustre: 3964:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x280000407:0x382a:0x0]/ may get corrupted (rc -108) Lustre: 3964:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x280000407:0x3709:0x0]/ may get corrupted (rc -108) Lustre: 3951:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x240000407:0x26cd:0x0]/ may get corrupted (rc -108) Lustre: 3951:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x28000040a:0x6fb:0x0]/ may get corrupted (rc -108) Lustre: 3958:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x280000407:0x3849:0x0]// may get corrupted (rc -108) Lustre: 3958:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x240000407:0x2749:0x0]/ may get corrupted (rc -108) Lustre: 3956:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x280000407:0x37c4:0x0]/ may get corrupted (rc -108) Lustre: 3956:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x280000407:0x37ff:0x0]// may get corrupted (rc -108) Lustre: 3954:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x240000407:0x26da:0x0]/ may get corrupted (rc -108) Lustre: 3954:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x280000407:0x3794:0x0]/ may get corrupted (rc -108) Lustre: 3966:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x280000407:0x3712:0x0]/ may get corrupted (rc -108) LustreError: lustre-OST0002-osc-ffff8802e8311bf8: This client was evicted by lustre-OST0002; in progress operations using this service will fail. LustreError: 5474:0:(ldlm_lockd.c:1492:ldlm_handle_enqueue()) ### lock on destroyed export ffff8802bb2937e8 ns: filter-lustre-OST0002_UUID lock: ffff88027a71f0c0/0xeaec6d3c1d96d35f lrc: 3/0,0 mode: PW/PW res: [0x340000402:0x6bbb:0x0].0x0 rrc: 2 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x50000000020000 nid: 0@lo remote: 0xeaec6d3c1d96d358 expref: 6829 pid: 5474 timeout: 0 lvb_type: 0 Lustre: 3966:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x240000407:0x26c4:0x0]/ may get corrupted (rc -108) LustreError: 2072:0:(ldlm_resource.c:982:ldlm_resource_complain()) lustre-OST0001-osc-ffff8802e6e9a548: namespace resource [0x300000400:0x75e0:0x0].0x0 (ffff88025137b240) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 2072:0:(ldlm_resource.c:982:ldlm_resource_complain()) Skipped 26 previous similar messages Lustre: 3961:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x240000407:0x265e:0x0]/ may get corrupted (rc -108) Lustre: 3960:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x200000407:0x6db:0x0]/ may get corrupted (rc -108) LustreError: 2075:0:(ldlm_resource.c:982:ldlm_resource_complain()) lustre-OST0002-osc-ffff8802e6e9a548: namespace resource [0x340000402:0x6efe:0x0].0x0 (ffff88006e0d1e40) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 2075:0:(ldlm_resource.c:982:ldlm_resource_complain()) Skipped 34 previous similar messages sched: RT throttling activated Lustre: 3967:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x240000408:0x1613:0x0]/ may get corrupted (rc -108) Lustre: 3965:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x28000040a:0x5ea:0x0]/ may get corrupted (rc -108) Lustre: 3953:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x200000406:0x4fde:0x0]/ may get corrupted (rc -108) Lustre: 3953:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x28000040a:0x801:0x0]/ may get corrupted (rc -108) Lustre: 3965:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x280000407:0x36e7:0x0]/ may get corrupted (rc -108) Lustre: 3963:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x200000406:0x4fd0:0x0]// may get corrupted (rc -108) Lustre: lustre-OST0000-osc-MDT0002: update sequence from 0x2c0000400 to 0x2c0000403 Lustre: lustre-OST0002-osc-MDT0002: update sequence from 0x340000400 to 0x340000403 Lustre: lustre-OST0001-osc-MDT0002: update sequence from 0x300000400 to 0x300000403 LustreError: 11120:0:(osd_handler.c:4211:osd_create_local_agent_inode()) lustre-MDT0001: create local error -1 Lustre: 11120:0:(out_handler.c:911:out_tx_end()) lustre-MDT0001-osd: error during execution of #0 from /home/green/git/lustre-release/lustre/ptlrpc/../../lustre/target/out_handler.c:555: rc = -1 Lustre: lustre-OST0002-osc-MDT0001: update sequence from 0x340000401 to 0x340000404 Lustre: lustre-OST0000-osc-MDT0000: update sequence from 0x2c0000402 to 0x2c0000404 Lustre: lustre-OST0002-osc-MDT0000: update sequence from 0x340000402 to 0x340000405 Lustre: 18261:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0001: opcode 0: before 509 < left 3876, rollback = 0 Lustre: 18261:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 1488 previous similar messages ptlrpc_watchdog_fire: 30 callbacks suppressed Lustre: mdt_io00_001: service thread pid 9893 was inactive for 40.091 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Lustre: Skipped 2 previous similar messages Pid: 9893, comm: mdt_io00_001 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock+0x88/0x1c0 [mdt] [<0>] mdt_rename_source_lock+0xa9/0xd6 [mdt] [<0>] mdt_reint_rename+0x1749/0x2bf0 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x73c/0xbb0 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 11786, comm: ll_ost_out00_00 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_pdo_lock+0x369/0x750 [mdt] [<0>] mdt_parent_lock+0x76/0x2a0 [mdt] [<0>] mdt_getattr_name_lock+0x1844/0x2cc0 [mdt] [<0>] mdt_intent_getattr+0x2cc/0x4e0 [mdt] [<0>] mdt_intent_opc.constprop.75+0x211/0xc60 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x397/0x980 [ptlrpc] [<0>] ldlm_handle_enqueue+0x547/0x18d0 [ptlrpc] [<0>] mdt_batch_getattr+0x6c/0x190 [mdt] [<0>] mdt_batch+0x465/0x195b [mdt] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: lustre-OST0001-osc-MDT0000: update sequence from 0x300000402 to 0x300000404 Lustre: mdt00_045: service thread pid 7908 was inactive for 72.199 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Lustre: Skipped 1 previous similar message Pid: 7908, comm: mdt00_045 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_pdo_lock+0x369/0x750 [mdt] [<0>] mdt_parent_lock+0x76/0x2a0 [mdt] [<0>] mdt_create+0xb44/0x1960 [mdt] [<0>] mdt_reint_create+0x2eb/0x470 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x73c/0xbb0 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: lustre-OST0003-osc-MDT0000: update sequence from 0x380000402 to 0x380000404 LustreError: 9779:0:(ldlm_lockd.c:252:expired_lock_main()) ### lock callback timer expired after 102s: evicting client at 0@lo ns: mdt-lustre-MDT0002_UUID lock: ffff88027ad12980/0xeaec6d3c1dade97a lrc: 3/0,0 mode: PR/PR res: [0x280000407:0x393e:0x0].0x0 bits 0x13/0x0 rrc: 3 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0xeaec6d3c1dadd88d expref: 367 pid: 11784 timeout: 3733 lvb_type: 0 LustreError: 9779:0:(ldlm_lockd.c:252:expired_lock_main()) Skipped 3 previous similar messages Lustre: mdt_io00_001: service thread pid 9893 completed after 101.922s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: lustre-MDT0002-mdc-ffff8802e8311bf8: Connection to lustre-MDT0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 3 previous similar messages LustreError: lustre-MDT0002-mdc-ffff8802e8311bf8: This client was evicted by lustre-MDT0002; in progress operations using this service will fail. Lustre: mdt00_045: service thread pid 7908 completed after 102.074s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 11786:0:(ldlm_lockd.c:1492:ldlm_handle_enqueue()) ### lock on destroyed export ffff880296829bf8 ns: mdt-lustre-MDT0002_UUID lock: ffff880262a3e940/0xeaec6d3c1dc3248a lrc: 3/0,0 mode: PR/PR res: [0x280000407:0x393e:0x0].0x0 bits 0x13/0x0 rrc: 2 type: IBT gid 0 flags: 0x50200000000000 nid: 0@lo remote: 0xeaec6d3c1dadd89b expref: 30 pid: 11786 timeout: 0 lvb_type: 0 LustreError: 13629:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e8311bf8: inode [0x28000040a:0x9cd:0x0] mdc close failed: rc = -5 LustreError: 13629:0:(file.c:247:ll_close_inode_openhandle()) Skipped 59 previous similar messages Lustre: ll_ost_out00_00: service thread pid 11786 completed after 102.045s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: lustre-MDT0002-mdc-ffff8802e8311bf8: Connection restored to (at 0@lo) Lustre: Skipped 3 previous similar messages LustreError: 12864:0:(ldlm_lockd.c:2591:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1742902580 with bad export cookie 16928025204340718279 LustreError: 12864:0:(ldlm_lockd.c:2591:ldlm_cancel_handler()) Skipped 2 previous similar messages LustreError: lustre-OST0002-osc-ffff8802e6e9a548: This client was evicted by lustre-OST0002; in progress operations using this service will fail. LustreError: 2594:0:(client.c:1282:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff880298eb6940 x1827562506750080/t0(0) o104->lustre-OST0002@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' uid:4294967295 gid:4294967295 LustreError: 2594:0:(client.c:1282:ptlrpc_import_delay_req()) Skipped 91 previous similar messages Lustre: 3962:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x28000040a:0x936:0x0]/ may get corrupted (rc -108) Lustre: 3966:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x240000407:0x2ab9:0x0]// may get corrupted (rc -108) Lustre: 3964:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x280000407:0x3a4c:0x0]// may get corrupted (rc -108) Lustre: 3964:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x280000407:0x3a58:0x0]/ may get corrupted (rc -108) Lustre: 3964:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x280000407:0x3a12:0x0]/ may get corrupted (rc -108) Lustre: lustre-OST0003-osc-MDT0001: update sequence from 0x380000401 to 0x380000405 Lustre: lustre-OST0001-osc-MDT0001: update sequence from 0x300000401 to 0x300000405 Lustre: 17603:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 6/24/0, destroy: 1/4/0 Lustre: 17603:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 1510 previous similar messages Lustre: 17603:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 3025/3025/0, xattr_set: 4537/42456/0 Lustre: 17603:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 1510 previous similar messages Lustre: 17603:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 28/157/0, punch: 0/0/0, quota 1/3/0 Lustre: 17603:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 1510 previous similar messages Lustre: 17603:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 7/118/0, delete: 2/5/0 Lustre: 17603:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 1510 previous similar messages Lustre: 17603:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 1/1/0, ref_del: 2/2/1 Lustre: 17603:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 1510 previous similar messages LustreError: 9777:0:(client.c:1282:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff880085390040 x1827562549827968/t0(0) o105->lustre-OST0000@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' uid:4294967295 gid:4294967295 LustreError: 9777:0:(client.c:1282:ptlrpc_import_delay_req()) Skipped 37 previous similar messages LustreError: lustre-OST0000-osc-ffff8802e8311bf8: operation ost_setattr to node 0@lo failed: rc = -107 LustreError: Skipped 6 previous similar messages LustreError: 26757:0:(ldlm_lockd.c:1408:ldlm_handle_enqueue()) ### lock on disconnected export ffff8802ce5ca548 ns: filter-lustre-OST0000_UUID lock: ffff880262498f40/0xeaec6d3c1de47500 lrc: 2/0,0 mode: --/PW res: [0x2c0000401:0x6771:0x0].0x0 rrc: 4 type: EXT [0->0] (req 0->0) gid 0 flags: 0x40000000000000 nid: local remote: 0xeaec6d3c1de42472 expref: -99 pid: 26757 timeout: 0 lvb_type: 0 LustreError: lustre-OST0000-osc-ffff8802e8311bf8: This client was evicted by lustre-OST0000; in progress operations using this service will fail. Lustre: 3957:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x28000040b:0x181:0x0]/ may get corrupted (rc -108) Lustre: 3957:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x28000040b:0x169:0x0]/ may get corrupted (rc -108) Lustre: 3958:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x200000406:0x54dd:0x0]/ may get corrupted (rc -108) Lustre: 3958:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x28000040b:0x1ac:0x0]/ may get corrupted (rc -108) Lustre: 3963:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x240000408:0x1b8c:0x0]/ may get corrupted (rc -108) Lustre: 3961:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x240000407:0x2c0e:0x0]/ may get corrupted (rc -108) Lustre: 3955:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x200000407:0xb18:0x0]/ may get corrupted (rc -108) Lustre: 3955:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x28000040b:0x8a:0x0]/ may get corrupted (rc -108) LustreError: 24203:0:(ldlm_resource.c:982:ldlm_resource_complain()) lustre-OST0000-osc-ffff8802e8311bf8: namespace resource [0x2c0000404:0x236c:0x0].0x0 (ffff8802401b8540) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 24203:0:(ldlm_resource.c:982:ldlm_resource_complain()) Skipped 121 previous similar messages Lustre: ll_ost00_024: service thread pid 28884 was inactive for 40.079 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 10 previous similar messages Lustre: ll_ost00_024: service thread pid 28884 completed after 50.937s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: lustre-OST0000-osc-MDT0001: update sequence from 0x2c0000401 to 0x2c0000405 Lustre: lustre-OST0000: haven't heard from client 19d5422d-541e-4da0-803c-0e442c636d68 (at 0@lo) in 32 seconds. I think it's dead, and I am evicting it. exp ffff88024ccbe678, cur 1742902727 expire 1742902697 last 1742902695 LustreError: lustre-OST0000-osc-ffff8802e8311bf8: This client was evicted by lustre-OST0000; in progress operations using this service will fail. LustreError: 27112:0:(ldlm_resource.c:982:ldlm_resource_complain()) lustre-OST0000-osc-ffff8802e8311bf8: namespace resource [0x2c0000405:0x2bb:0x0].0x0 (ffff8802c7aac7c0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 27112:0:(ldlm_resource.c:982:ldlm_resource_complain()) Skipped 309 previous similar messages Lustre: lustre-OST0003-osc-MDT0002: update sequence from 0x380000403 to 0x380000406 Lustre: lustre-OST0000-osc-MDT0002: update sequence from 0x2c0000403 to 0x2c0000406 Lustre: lustre-OST0002-osc-MDT0001: update sequence from 0x340000404 to 0x340000406 Lustre: 17267:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff880212c49940 x1827562621466496/t4295099919(0) o101->19d5422d-541e-4da0-803c-0e442c636d68@0@lo:645/0 lens 376/47728 e 0 to 0 dl 1742903045 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 Lustre: 17267:0:(mdt_recovery.c:102:mdt_req_from_lrd()) Skipped 41 previous similar messages Lustre: lustre-OST0002-osc-MDT0002: update sequence from 0x340000403 to 0x340000407 Lustre: lustre-OST0001-osc-MDT0002: update sequence from 0x300000403 to 0x300000406 Lustre: 3964:0:(client.c:2346:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1742902792/real 1742902792] req@ffff8802227c7840 x1827562579757952/t0(0) o63->lustre-MDT0001-mdc-ffff8802e8311bf8@0@lo:24/4 lens 2624/54000 e 0 to 1 dl 1742902959 ref 1 fl Rpc:XQr/202/ffffffff rc 0/-1 job:'ll_sa.0' uid:0 gid:0 Lustre: lustre-MDT0001: Client 19d5422d-541e-4da0-803c-0e442c636d68 (at 0@lo) reconnecting Lustre: 17582:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0001: opcode 0: before 513 < left 2565, rollback = 0 Lustre: 17582:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 1469 previous similar messages LustreError: 9779:0:(ldlm_lockd.c:252:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0002_UUID lock: ffff8802555e9300/0xeaec6d3c1e226e72 lrc: 3/0,0 mode: PW/PW res: [0x340000407:0x110f:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000080030020 nid: 0@lo remote: 0xeaec6d3c1e226e6b expref: 13587 pid: 4526 timeout: 4282 lvb_type: 0 LustreError: 9779:0:(ldlm_lockd.c:252:expired_lock_main()) Skipped 2 previous similar messages Lustre: lustre-OST0002-osc-ffff8802e8311bf8: Connection to lustre-OST0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 4 previous similar messages LustreError: 4971:0:(client.c:1282:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8801fc6f7340 x1827562681335552/t0(0) o104->lustre-OST0002@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' uid:4294967295 gid:4294967295 LustreError: 4971:0:(client.c:1282:ptlrpc_import_delay_req()) Skipped 8 previous similar messages LustreError: 5177:0:(ldlm_lockd.c:1492:ldlm_handle_enqueue()) ### lock on destroyed export ffff880070f10008 ns: filter-lustre-OST0002_UUID lock: ffff88026be8d2c0/0xeaec6d3c1e37c78f lrc: 3/0,0 mode: PW/PW res: [0x340000407:0x1199:0x0].0x0 rrc: 2 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x50000080020000 nid: 0@lo remote: 0xeaec6d3c1e37c788 expref: 13268 pid: 5177 timeout: 0 lvb_type: 0 LustreError: lustre-OST0002-osc-ffff8802e8311bf8: This client was evicted by lustre-OST0002; in progress operations using this service will fail. Lustre: 3966:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x240000408:0x1f6f:0x0]/ may get corrupted (rc -108) Lustre: 3966:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x240000408:0x201e:0x0]/ may get corrupted (rc -108) Lustre: 3957:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x200000406:0x5a80:0x0]/ may get corrupted (rc -108) Lustre: 3957:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x200000406:0x5a6f:0x0]/ may get corrupted (rc -108) Lustre: 3951:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x28000040b:0x639:0x0]/ may get corrupted (rc -108) Lustre: 3951:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x240000408:0x206a:0x0]/ may get corrupted (rc -108) Lustre: 3958:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x28000040b:0x5f4:0x0]/ may get corrupted (rc -108) Lustre: 3954:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x240000408:0x2098:0x0]/ may get corrupted (rc -108) Lustre: 3954:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x200000406:0x5a5a:0x0]/ may get corrupted (rc -108) Lustre: 3965:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x240000408:0x1f6a:0x0]/ may get corrupted (rc -108) Lustre: 3965:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x28000040b:0x66f:0x0]/ may get corrupted (rc -108) Lustre: lustre-OST0002-osc-ffff8802e8311bf8: Connection restored to (at 0@lo) Lustre: Skipped 4 previous similar messages LustreError: lustre-OST0003-osc-ffff8802e6e9a548: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 101 previous similar messages LustreError: 29748:0:(client.c:1282:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8801f5436440 x1827562733876736/t0(0) o104->lustre-OST0003@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' uid:4294967295 gid:4294967295 LustreError: 29748:0:(client.c:1282:ptlrpc_import_delay_req()) Skipped 84 previous similar messages LustreError: lustre-OST0003-osc-ffff8802e6e9a548: This client was evicted by lustre-OST0003; in progress operations using this service will fail. Lustre: 3960:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x200000407:0x12db:0x0]/ may get corrupted (rc -108) Lustre: 3959:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x28000040b:0x8ad:0x0]/ may get corrupted (rc -108) Lustre: 3962:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x240000407:0x324c:0x0]/ may get corrupted (rc -108) Lustre: 3965:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x28000040b:0x8a4:0x0]/ may get corrupted (rc -108) Lustre: 3962:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x240000407:0x326e:0x0]// may get corrupted (rc -108) LustreError: 31222:0:(ldlm_resource.c:982:ldlm_resource_complain()) lustre-OST0003-osc-ffff8802e6e9a548: namespace resource [0x380000406:0x455a:0x0].0x0 (ffff88025f1e3c40) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 31222:0:(ldlm_resource.c:982:ldlm_resource_complain()) Skipped 71 previous similar messages Lustre: 11120:0:(out_handler.c:911:out_tx_end()) lustre-MDT0001-osd: error during execution of #2 from /home/green/git/lustre-release/lustre/ptlrpc/../../lustre/target/out_handler.c:585: rc = -2 LustreError: 11120:0:(out_lib.c:1168:out_tx_index_delete_undo()) lustre-MDT0001-osd: Oops, can not rollback index_delete yet: rc = -524 Lustre: 17267:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 5/20/0, destroy: 1/4/0 Lustre: 17267:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 1486 previous similar messages Lustre: 17267:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 3193/3193/0, xattr_set: 4789/44808/0 Lustre: 17267:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 1486 previous similar messages Lustre: 17267:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 24/133/0, punch: 0/0/0, quota 1/3/0 Lustre: 17267:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 1486 previous similar messages Lustre: 17267:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 6/101/0, delete: 2/5/0 Lustre: 17267:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 1486 previous similar messages Lustre: 17267:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 1/1/0, ref_del: 2/2/0 Lustre: 17267:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 1486 previous similar messages LustreError: 32395:0:(ldlm_lockd.c:2591:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1742903258 with bad export cookie 16928025204311134494 LustreError: 32395:0:(ldlm_lockd.c:2591:ldlm_cancel_handler()) Skipped 1 previous similar message LustreError: lustre-OST0003-osc-ffff8802e8311bf8: This client was evicted by lustre-OST0003; in progress operations using this service will fail. Lustre: 3958:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x28000040b:0x8b6:0x0]/ may get corrupted (rc -108) Lustre: 3964:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x200000406:0x5d15:0x0]/ may get corrupted (rc -108) Lustre: 3962:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x200000406:0x5cd4:0x0]/ may get corrupted (rc -108) Lustre: 3963:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x200000406:0x5d04:0x0]// may get corrupted (rc -108) Lustre: 3960:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x240000408:0x228b:0x0]/ may get corrupted (rc -108) Lustre: 3960:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x240000408:0x2294:0x0]/ may get corrupted (rc -108) Lustre: 3957:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x28000040b:0x946:0x0]/ may get corrupted (rc -108) Lustre: 3957:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x28000040b:0x920:0x0]// may get corrupted (rc -108) Lustre: 3951:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x28000040b:0x8c7:0x0]/ may get corrupted (rc -108) Lustre: 3967:0:(client.c:2346:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1742903034/real 1742903034] req@ffff88006d311e40 x1827562661161600/t0(0) o63->lustre-MDT0000-mdc-ffff8802e6e9a548@0@lo:24/4 lens 664/53480 e 0 to 1 dl 1742903201 ref 1 fl Rpc:XQr/202/ffffffff rc 0/-1 job:'ll_sa.0' uid:0 gid:0 Lustre: lustre-MDT0000: Client e3f32f46-7f2b-47b9-b6d8-4f9caea418d0 (at 0@lo) reconnecting LustreError: 7907:0:(mdt_open.c:1303:mdt_cross_open()) lustre-MDT0002: [0x280000407:0x4287:0x0] doesn't exist!: rc = -14 Lustre: 1232:0:(client.c:2346:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1742903268/real 1742903268] req@ffff8801ffdfb740 x1827562743781376/t0(0) o101->lustre-MDT0001-mdc-ffff8802e8311bf8@0@lo:12/10 lens 576/47752 e 0 to 1 dl 1742903284 ref 2 fl Rpc:XQr/202/ffffffff rc 0/-1 job:'mrename.0' uid:0 gid:0 Lustre: lustre-MDT0001: Client 19d5422d-541e-4da0-803c-0e442c636d68 (at 0@lo) reconnecting Lustre: lustre-OST0003: haven't heard from client 19d5422d-541e-4da0-803c-0e442c636d68 (at 0@lo) in 31 seconds. I think it's dead, and I am evicting it. exp ffff8802d5216678, cur 1742903289 expire 1742903259 last 1742903258 Lustre: lustre-OST0003-osc-MDT0001: update sequence from 0x380000405 to 0x380000407 LustreError: 9873:0:(mdt_open.c:1303:mdt_cross_open()) lustre-MDT0002: [0x280000407:0x4287:0x0] doesn't exist!: rc = -14 Lustre: lustre-OST0001-osc-MDT0001: update sequence from 0x300000405 to 0x300000407 LustreError: 17217:0:(mdt_open.c:1303:mdt_cross_open()) lustre-MDT0002: [0x280000407:0x4287:0x0] doesn't exist!: rc = -14 ptlrpc_watchdog_fire: 1 callbacks suppressed Lustre: mdt00_035: service thread pid 7888 was inactive for 40.112 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 7888, comm: mdt00_035 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] osp_precreate_reserve+0x58f/0xa80 [osp] [<0>] osp_declare_create+0x18b/0x6a0 [osp] [<0>] lod_sub_declare_create+0xed/0x250 [lod] [<0>] lod_qos_declare_object_on+0x103/0x430 [lod] [<0>] lod_ost_alloc_rr+0xa4c/0x12c0 [lod] [<0>] lod_qos_prep_create+0x168c/0x1dd0 [lod] [<0>] lod_prepare_create+0x241/0x330 [lod] [<0>] lod_declare_striped_create+0xe5/0xad0 [lod] [<0>] lod_declare_create+0x216/0x6c0 [lod] [<0>] mdd_declare_create_object_internal+0xd2/0x390 [mdd] [<0>] mdd_declare_create_object.isra.34+0x52/0x8f0 [mdd] [<0>] mdd_declare_create+0x66/0x500 [mdd] [<0>] mdd_create+0x631/0x1bc0 [mdd] [<0>] mdt_reint_open+0x20ee/0x2d70 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x73c/0xbb0 [mdt] [<0>] mdt_intent_open+0x93/0x480 [mdt] [<0>] mdt_intent_opc.constprop.75+0x211/0xc60 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x397/0x980 [ptlrpc] [<0>] ldlm_handle_enqueue+0x547/0x18d0 [ptlrpc] [<0>] tgt_enqueue+0x68/0x240 [ptlrpc] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 17780, comm: mdt00_029 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] osp_precreate_reserve+0x58f/0xa80 [osp] [<0>] osp_declare_create+0x18b/0x6a0 [osp] [<0>] lod_sub_declare_create+0xed/0x250 [lod] [<0>] lod_qos_declare_object_on+0x103/0x430 [lod] [<0>] lod_ost_alloc_rr+0xa4c/0x12c0 [lod] [<0>] lod_qos_prep_create+0x168c/0x1dd0 [lod] [<0>] lod_prepare_create+0x241/0x330 [lod] [<0>] lod_declare_striped_create+0xe5/0xad0 [lod] [<0>] lod_declare_xattr_set+0x172/0x15d0 [lod] [<0>] dt_declare_xattr_set+0x54/0x160 [mdd] [<0>] mdd_create_data+0x3b1/0x670 [mdd] [<0>] mdt_mfd_open+0xcfc/0xf70 [mdt] [<0>] mdt_finish_open+0x50b/0xa00 [mdt] [<0>] mdt_open_by_fid_lock+0x56b/0xc10 [mdt] [<0>] mdt_reint_open+0x9d8/0x2d70 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x73c/0xbb0 [mdt] [<0>] mdt_intent_open+0x93/0x480 [mdt] [<0>] mdt_intent_opc.constprop.75+0x211/0xc60 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x397/0x980 [ptlrpc] [<0>] ldlm_handle_enqueue+0x547/0x18d0 [ptlrpc] [<0>] tgt_enqueue+0x68/0x240 [ptlrpc] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: mdt00_041: service thread pid 7901 was inactive for 40.099 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Lustre: Skipped 1 previous similar message Pid: 7901, comm: mdt00_041 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] osp_precreate_reserve+0x58f/0xa80 [osp] [<0>] osp_declare_create+0x18b/0x6a0 [osp] [<0>] lod_sub_declare_create+0xed/0x250 [lod] [<0>] lod_qos_declare_object_on+0x103/0x430 [lod] [<0>] lod_ost_alloc_rr+0xa4c/0x12c0 [lod] [<0>] lod_qos_prep_create+0x168c/0x1dd0 [lod] [<0>] lod_prepare_create+0x241/0x330 [lod] [<0>] lod_declare_striped_create+0xe5/0xad0 [lod] [<0>] lod_declare_xattr_set+0x172/0x15d0 [lod] [<0>] dt_declare_xattr_set+0x54/0x160 [mdd] [<0>] mdd_create_data+0x3b1/0x670 [mdd] [<0>] mdt_mfd_open+0xcfc/0xf70 [mdt] [<0>] mdt_finish_open+0x50b/0xa00 [mdt] [<0>] mdt_open_by_fid_lock+0x56b/0xc10 [mdt] [<0>] mdt_reint_open+0x9d8/0x2d70 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x73c/0xbb0 [mdt] [<0>] mdt_intent_open+0x93/0x480 [mdt] [<0>] mdt_intent_opc.constprop.75+0x211/0xc60 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x397/0x980 [ptlrpc] [<0>] ldlm_handle_enqueue+0x547/0x18d0 [ptlrpc] [<0>] tgt_enqueue+0x68/0x240 [ptlrpc] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: mdt00_007: service thread pid 16620 was inactive for 40.080 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: mdt00_029: service thread pid 17780 completed after 42.418s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_021: service thread pid 17551 completed after 40.178s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_041: service thread pid 7901 completed after 40.936s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_035: service thread pid 7888 completed after 42.642s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_007: service thread pid 16620 completed after 40.329s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: lustre-OST0001-osc-MDT0000: update sequence from 0x300000404 to 0x300000408 Lustre: lustre-OST0000-osc-MDT0001: update sequence from 0x2c0000405 to 0x2c0000407 LustreError: 17603:0:(mdt_open.c:1303:mdt_cross_open()) lustre-MDT0002: [0x280000407:0x4287:0x0] doesn't exist!: rc = -14 LustreError: 17603:0:(mdt_open.c:1303:mdt_cross_open()) Skipped 2 previous similar messages hrtimer: interrupt took 65683547 ns LustreError: 17601:0:(mdt_open.c:1303:mdt_cross_open()) lustre-MDT0002: [0x280000407:0x4287:0x0] doesn't exist!: rc = -14 LustreError: 17601:0:(mdt_open.c:1303:mdt_cross_open()) Skipped 3 previous similar messages Lustre: lustre-OST0002-osc-MDT0000: update sequence from 0x340000405 to 0x340000408 Lustre: lustre-OST0000-osc-MDT0000: update sequence from 0x2c0000404 to 0x2c0000408 LustreError: lustre-OST0003-osc-ffff8802e8311bf8: This client was evicted by lustre-OST0003; in progress operations using this service will fail. LustreError: 7888:0:(mdt_open.c:1303:mdt_cross_open()) lustre-MDT0002: [0x280000407:0x4287:0x0] doesn't exist!: rc = -14 LustreError: 7888:0:(mdt_open.c:1303:mdt_cross_open()) Skipped 1 previous similar message Lustre: lustre-OST0003-osc-MDT0000: update sequence from 0x380000404 to 0x380000408 LustreError: 7909:0:(mdt_open.c:1303:mdt_cross_open()) lustre-MDT0002: [0x280000407:0x4287:0x0] doesn't exist!: rc = -14 LustreError: 7909:0:(mdt_open.c:1303:mdt_cross_open()) Skipped 1 previous similar message Lustre: lustre-OST0002-osc-MDT0001: update sequence from 0x340000406 to 0x340000409 LustreError: lustre-OST0000-osc-ffff8802e6e9a548: This client was evicted by lustre-OST0000; in progress operations using this service will fail. Lustre: 3963:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x240000407:0x343b:0x0]// may get corrupted (rc -108) Lustre: 3957:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x240000407:0x33bf:0x0]/ may get corrupted (rc -108) Lustre: 3957:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x240000408:0x2413:0x0]/ may get corrupted (rc -108) Lustre: 3958:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x240000407:0x3410:0x0]/ may get corrupted (rc -108) Lustre: 3958:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x200000406:0x5e31:0x0]/ may get corrupted (rc -108) Lustre: 3961:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x200000407:0x145d:0x0]// may get corrupted (rc -108) Lustre: 3961:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x28000040b:0xa1f:0x0]/ may get corrupted (rc -108) Lustre: 3954:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x200000407:0x1467:0x0]/ may get corrupted (rc -108) LustreError: 6729:0:(ldlm_resource.c:982:ldlm_resource_complain()) lustre-OST0000-osc-ffff8802e6e9a548: namespace resource [0x2c0000406:0x564a:0x0].0x0 (ffff880202460cc0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 6729:0:(ldlm_resource.c:982:ldlm_resource_complain()) Skipped 409 previous similar messages LustreError: 17603:0:(mdt_open.c:1303:mdt_cross_open()) lustre-MDT0002: [0x280000407:0x4287:0x0] doesn't exist!: rc = -14 LustreError: 17603:0:(mdt_open.c:1303:mdt_cross_open()) Skipped 1 previous similar message LustreError: lustre-OST0001-osc-ffff8802e8311bf8: This client was evicted by lustre-OST0001; in progress operations using this service will fail. Lustre: 3951:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x28000040b:0xb32:0x0]// may get corrupted (rc -108) Lustre: 3963:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x200000406:0x5f2d:0x0]// may get corrupted (rc -108) Lustre: 3956:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x200000406:0x5f11:0x0]/ may get corrupted (rc -108) Lustre: 3964:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x280000407:0x4470:0x0]/ may get corrupted (rc -108) Lustre: 3954:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x240000407:0x353d:0x0]/ may get corrupted (rc -108) Lustre: 3959:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x28000040b:0xb07:0x0]// may get corrupted (rc -108) Lustre: 3967:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x240000408:0x2505:0x0]// may get corrupted (rc -108) LustreError: 9874:0:(mdt_open.c:1303:mdt_cross_open()) lustre-MDT0002: [0x280000407:0x4287:0x0] doesn't exist!: rc = -14 LustreError: 9874:0:(mdt_open.c:1303:mdt_cross_open()) Skipped 7 previous similar messages LustreError: 13641:0:(ldlm_resource.c:982:ldlm_resource_complain()) lustre-OST0001-osc-ffff8802e8311bf8: namespace resource [0x300000405:0x7f1b:0x0].0x0 (ffff88020f9d7c40) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 13641:0:(ldlm_resource.c:982:ldlm_resource_complain()) Skipped 338 previous similar messages Lustre: 17780:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff880210820040 x1827562826692608/t4295114041(0) o101->19d5422d-541e-4da0-803c-0e442c636d68@0@lo:436/0 lens 376/47728 e 0 to 0 dl 1742903591 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 Lustre: 17780:0:(mdt_recovery.c:102:mdt_req_from_lrd()) Skipped 35 previous similar messages Lustre: lustre-OST0001: haven't heard from client 19d5422d-541e-4da0-803c-0e442c636d68 (at 0@lo) in 31 seconds. I think it's dead, and I am evicting it. exp ffff8801f78a6678, cur 1742903516 expire 1742903486 last 1742903485 Lustre: lustre-OST0003-osc-MDT0002: update sequence from 0x380000406 to 0x380000409 LustreError: lustre-OST0001-osc-ffff8802e8311bf8: This client was evicted by lustre-OST0001; in progress operations using this service will fail. Lustre: lustre-OST0000-osc-MDT0002: update sequence from 0x2c0000406 to 0x2c0000409 LustreError: 4910:0:(client.c:1282:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8801e13c1e40 x1827562861880832/t0(0) o104->lustre-OST0002@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' uid:4294967295 gid:4294967295 LustreError: 4910:0:(client.c:1282:ptlrpc_import_delay_req()) Skipped 392 previous similar messages LustreError: lustre-OST0002-osc-ffff8802e8311bf8: This client was evicted by lustre-OST0002; in progress operations using this service will fail. Lustre: 3964:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x240000408:0x2648:0x0]/ may get corrupted (rc -108) Lustre: 3964:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x28000040b:0xce1:0x0]/ may get corrupted (rc -108) Lustre: 3953:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x240000408:0x269e:0x0]/ may get corrupted (rc -108) Lustre: 3953:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x240000407:0x3701:0x0]/ may get corrupted (rc -108) Lustre: 3963:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x240000408:0x268e:0x0]/ may get corrupted (rc -108) Lustre: 3966:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x280000407:0x4607:0x0]/ may get corrupted (rc -108) Lustre: 3967:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x28000040b:0xd08:0x0]/ may get corrupted (rc -108) Lustre: lustre-OST0001-osc-MDT0002: update sequence from 0x300000406 to 0x300000409 Lustre: lustre-OST0002-osc-MDT0002: update sequence from 0x340000407 to 0x34000040a LustreError: 7908:0:(mdt_open.c:1303:mdt_cross_open()) lustre-MDT0002: [0x280000407:0x4287:0x0] doesn't exist!: rc = -14 LustreError: 7908:0:(mdt_open.c:1303:mdt_cross_open()) Skipped 18 previous similar messages Lustre: 16526:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0000: opcode 0: before 514 < left 591, rollback = 0 Lustre: 16526:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 1725 previous similar messages LustreError: 9779:0:(ldlm_lockd.c:252:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0003_UUID lock: ffff880260e9bc40/0xeaec6d3c1ec3903e lrc: 3/0,0 mode: PW/PW res: [0x380000407:0x52a5:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000480000020 nid: 0@lo remote: 0xeaec6d3c1ec38fa4 expref: 16768 pid: 4913 timeout: 4928 lvb_type: 0 LustreError: 9779:0:(ldlm_lockd.c:252:expired_lock_main()) Skipped 5 previous similar messages LustreError: 11046:0:(ldlm_lockd.c:2591:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1742903748 with bad export cookie 16928025204353632201 LustreError: 11046:0:(ldlm_lockd.c:2591:ldlm_cancel_handler()) Skipped 1 previous similar message Lustre: lustre-OST0003-osc-ffff8802e6e9a548: Connection to lustre-OST0003 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 9 previous similar messages LustreError: lustre-OST0003-osc-ffff8802e6e9a548: This client was evicted by lustre-OST0003; in progress operations using this service will fail. Lustre: 3964:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x200000406:0x6203:0x0]/ may get corrupted (rc -108) Lustre: 3966:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x240000408:0x27d7:0x0]/ may get corrupted (rc -108) Lustre: 3966:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x200000407:0x1881:0x0]/ may get corrupted (rc -108) Lustre: 3955:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x200000407:0x18d2:0x0]/ may get corrupted (rc -108) Lustre: 3953:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x280000407:0x47d3:0x0]/ may get corrupted (rc -108) Lustre: 3953:0:(llite_lib.c:4117:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.17@tcp:/lustre/fid: [0x280000407:0x4800:0x0]// may get corrupted (rc -108) LustreError: 31928:0:(ldlm_resource.c:982:ldlm_resource_complain()) lustre-OST0003-osc-ffff8802e6e9a548: namespace resource [0x380000407:0x554a:0x0].0x0 (ffff88026f3307c0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 31928:0:(ldlm_resource.c:982:ldlm_resource_complain()) Skipped 299 previous similar messages Lustre: lustre-OST0003: haven't heard from client e3f32f46-7f2b-47b9-b6d8-4f9caea418d0 (at 0@lo) in 31 seconds. I think it's dead, and I am evicting it. exp ffff8801f78a2e98, cur 1742903779 expire 1742903749 last 1742903748 Lustre: lustre-OST0003-osc-ffff8802e6e9a548: Connection restored to (at 0@lo) Lustre: Skipped 9 previous similar messages LustreError: lustre-OST0003-osc-ffff8802e6e9a548: This client was evicted by lustre-OST0003; in progress operations using this service will fail. Lustre: 1220:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 6/24/0, destroy: 1/4/0 Lustre: 1220:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 1727 previous similar messages Lustre: 1220:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 2165/2165/0, xattr_set: 3247/30416/0 Lustre: 1220:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 1727 previous similar messages Lustre: 1220:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 28/157/0, punch: 0/0/0, quota 1/3/0 Lustre: 1220:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 1727 previous similar messages Lustre: 1220:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 7/118/0, delete: 2/5/1 Lustre: 1220:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 1727 previous similar messages Lustre: 1220:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 1/1/0, ref_del: 2/2/1 Lustre: 1220:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 1727 previous similar messages Lustre: lustre-OST0001-osc-MDT0000: update sequence from 0x300000408 to 0x30000040a Lustre: lustre-OST0001-osc-MDT0001: update sequence from 0x300000407 to 0x30000040b Lustre: lustre-OST0003-osc-MDT0001: update sequence from 0x380000407 to 0x38000040a Lustre: lustre-OST0003-osc-MDT0000: update sequence from 0x380000408 to 0x38000040b Lustre: lustre-OST0002-osc-MDT0000: update sequence from 0x340000408 to 0x34000040b LustreError: 17265:0:(mdt_open.c:1303:mdt_cross_open()) lustre-MDT0002: [0x280000407:0x4287:0x0] doesn't exist!: rc = -14 LustreError: 17265:0:(mdt_open.c:1303:mdt_cross_open()) Skipped 24 previous similar messages Lustre: lustre-OST0000-osc-MDT0001: update sequence from 0x2c0000407 to 0x2c000040a | Link to test |
racer test 1: racer on clients: centos-50.localnet DURATION=2700 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 PGD 2c4dfc067 PUD 2ad9e6067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) lustre(OE) osp(OE) ofd(OE) lod(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) jbd2 mbcache lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) crc32_generic lnet(OE) dm_flakey dm_mod libcfs(OE) crc_t10dif crct10dif_generic crct10dif_common virtio_balloon virtio_console i2c_piix4 pcspkr ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm ata_piix drm_panel_orientation_quirks floppy virtio_blk libata serio_raw i2c_core CPU: 2 PID: 2351 Comm: ll_sa_2127 Kdump: loaded Tainted: P OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.16.0-3.module_el8.7.0+1218+f626c2ff 04/01/2014 task: ffff8802e7a59280 ti: ffff88031be08000 task.ti: ffff88031be08000 RIP: 0010:[<ffffffff813f502c>] [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 RSP: 0018:ffff88031be0bd78 EFLAGS: 00010296 RAX: ffff88029e350ac8 RBX: 0000000000000008 RCX: 0000000100260021 RDX: 0000000000000026 RSI: ffff88029e350df8 RDI: 0000000000000008 RBP: ffff88031be0bd88 R08: ffff8802cf55ed58 R09: 0000000000000000 R10: ffff8802cf55e6b8 R11: ffff8802cf55ed58 R12: 0000000000000000 R13: ffff8802ebe1a4f0 R14: ffff88029e350df8 R15: ffff8802cf55ed58 FS: 0000000000000000(0000) GS:ffff880331a80000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 00000002de39c000 CR4: 00000000000007e0 Call Trace: [<ffffffffa11048a1>] ll_statahead_thread+0xe01/0x22b0 [lustre] [<ffffffff817e320e>] ? _raw_spin_lock+0x1e/0x20 [<ffffffff810dbb51>] ? put_prev_entity+0x31/0x400 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e329e>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffff810c834d>] ? finish_task_switch+0x5d/0x1b0 [<ffffffffa1103aa0>] ? ll_statahead_handle.constprop.29+0x1c0/0x1c0 [lustre] [<ffffffff810ba114>] kthread+0xe4/0xf0 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 | 3[15273]: segfault at 8 ip 00007f494457a7e8 sp 00007ffc2593bfa0 error 4 in ld-2.17.so[7f494456f000+22000] Lustre: 15285:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0000: opcode 7: before 516 < left 618, rollback = 7 Lustre: 15285:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 15285:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 15285:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 15285:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 15285:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 15291:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0000: opcode 7: before 516 < left 540, rollback = 7 Lustre: 15291:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 7 previous similar messages Lustre: 15291:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 15291:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 7 previous similar messages Lustre: 15291:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 15291:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 7 previous similar messages Lustre: 15291:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/540/0, punch: 0/0/0, quota 4/150/0 Lustre: 15291:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 7 previous similar messages Lustre: 15291:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 15291:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 7 previous similar messages Lustre: 15291:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 15291:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 7 previous similar messages Lustre: 14712:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000402:0xba:0x0] with magic=0xbd60bd0 Lustre: 15285:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0002: opcode 7: before 516 < left 618, rollback = 7 Lustre: 15285:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 1 previous similar message Lustre: 15285:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 15285:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 15285:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 15285:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 15285:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 15285:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 15285:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 15285:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 15285:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 15285:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 1 previous similar message LustreError: 20697:0:(statahead.c:2447:start_statahead_thread()) lustre: unsupported statahead pattern 0X0. 12[20895]: segfault at 0 ip 0000000000403cf0 sp 00007ffdefa763a8 error 6 in 12[400000+6000] 4[21332]: segfault at 8 ip 00007fa55763b7e8 sp 00007ffc7ed83650 error 4 in ld-2.17.so[7fa557630000+22000] Lustre: 15291:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 516 < left 618, rollback = 7 Lustre: 15291:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 1 previous similar message Lustre: 15291:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 15291:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 15291:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 15291:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 15291:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 15291:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 15291:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 15291:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 15291:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 15291:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 14669:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff8800a75ab240 x1826900775456000/t4294975099(0) o101->5130003b-f326-456d-9dcf-49672dd843ff@0@lo:239/0 lens 376/816 e 0 to 0 dl 1742268439 ref 1 fl Interpret:H/202/0 rc 0/0 job:'cat.0' uid:0 gid:0 Lustre: 15291:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 618, rollback = 7 Lustre: 15291:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 3 previous similar messages Lustre: 15291:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 15291:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 15291:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 15291:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 15291:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 15291:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 15291:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 15291:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 15291:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 15291:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 9015:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0000: opcode 7: before 516 < left 618, rollback = 7 Lustre: 9015:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 7 previous similar messages Lustre: 9015:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 9015:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 7 previous similar messages Lustre: 9015:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 9015:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 7 previous similar messages Lustre: 9015:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 9015:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 7 previous similar messages Lustre: 9015:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 9015:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 7 previous similar messages Lustre: 9015:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 9015:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 7 previous similar messages Lustre: 9018:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 516 < left 618, rollback = 7 Lustre: 9018:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 13 previous similar messages Lustre: 9018:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 9018:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 13 previous similar messages Lustre: 9018:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 9018:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 13 previous similar messages Lustre: 9018:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 9018:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 13 previous similar messages Lustre: 9018:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 9018:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 13 previous similar messages Lustre: 9018:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 9018:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 13 previous similar messages 11[31300]: segfault at 8 ip 00007ff66aff07e8 sp 00007ffedbc9b980 error 4 in ld-2.17.so[7ff66afe5000+22000] LustreError: 1223:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802eee11bf8: inode [0x200000402:0x6b8:0x0] mdc close failed: rc = -13 Lustre: 9668:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000401:0x686:0x0] with magic=0xbd60bd0 Lustre: 9668:0:(lod_lov.c:1417:lod_parse_striping()) Skipped 1 previous similar message Lustre: 14733:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000402:0x75f:0x0] with magic=0xbd60bd0 Lustre: 14733:0:(lod_lov.c:1417:lod_parse_striping()) Skipped 1 previous similar message 17[3459]: segfault at 8 ip 00007fd50cfea7e8 sp 00007ffde6a5b6d0 error 4 in ld-2.17.so[7fd50cfdf000+22000] LustreError: 3459:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802f0deae98: inode [0x200000401:0x69d:0x0] mdc close failed: rc = -13 Lustre: 8320:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000402:0x7c4:0x0] with magic=0xbd60bd0 Lustre: 8320:0:(lod_lov.c:1417:lod_parse_striping()) Skipped 1 previous similar message 17[7912]: segfault at 8 ip 00007f097958e7e8 sp 00007ffc60297c60 error 4 in ld-2.17.so[7f0979583000+22000] Lustre: 15291:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0000: opcode 7: before 516 < left 618, rollback = 7 Lustre: 15291:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 13 previous similar messages Lustre: 15291:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 15291:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 13 previous similar messages Lustre: 15291:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 15291:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 13 previous similar messages Lustre: 15291:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 15291:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 13 previous similar messages Lustre: 15291:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 15291:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 13 previous similar messages Lustre: 15291:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 15291:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 13 previous similar messages 0[8671]: segfault at 406000 ip 0000000000406000 sp 00007ffe48876528 error 14 in 0[606000+1000] 12[8801]: segfault at 8 ip 00007f67f02267e8 sp 00007ffef92c0dd0 error 4 in ld-2.17.so[7f67f021b000+22000] 11[17936]: segfault at 8 ip 00007f15919d97e8 sp 00007ffd3894bb80 error 4 in ld-2.17.so[7f15919ce000+22000] 10[19369]: segfault at 8 ip 00007f90936b17e8 sp 00007ffcfd427b60 error 4 in ld-2.17.so[7f90936a6000+22000] Lustre: 9018:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0000: opcode 7: before 516 < left 618, rollback = 7 Lustre: 9018:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 29 previous similar messages Lustre: 9018:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 9018:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 29 previous similar messages Lustre: 9018:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 9018:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 29 previous similar messages Lustre: 9018:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 9018:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 29 previous similar messages Lustre: 9018:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 9018:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 29 previous similar messages Lustre: 9018:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 9018:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 29 previous similar messages 11[19725]: segfault at 1c ip 00000000004017b0 sp 00007ffd3ee59cf0 error 6 in 11[400000+6000] Lustre: 9668:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000402:0xcaa:0x0] with magic=0xbd60bd0 Lustre: 9668:0:(lod_lov.c:1417:lod_parse_striping()) Skipped 1 previous similar message LustreError: 21127:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802f0deae98: inode [0x200000402:0xc7e:0x0] mdc close failed: rc = -13 0[21597]: segfault at 8 ip 00007f26eb6ab7e8 sp 00007ffcfebd4c20 error 4 in ld-2.17.so[7f26eb6a0000+22000] 3[22443]: segfault at 8 ip 00007fcbaeb6f7e8 sp 00007ffebff75280 error 4 in ld-2.17.so[7fcbaeb64000+22000] 11[25040]: segfault at 8 ip 00007fad2c26a7e8 sp 00007fff3f1a5bd0 error 4 in ld-2.17.so[7fad2c25f000+22000] Lustre: 14669:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000402:0xee5:0x0] with magic=0xbd60bd0 Lustre: 14669:0:(lod_lov.c:1417:lod_parse_striping()) Skipped 1 previous similar message 17[26864]: segfault at 0 ip 0000000000403e5f sp 00007fffd0b923c0 error 6 in 17[400000+6000] 5[484]: segfault at 8 ip 00007f941a4b87e8 sp 00007fff1e69c290 error 4 in ld-2.17.so[7f941a4ad000+22000] LustreError: 2321:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802f0deae98: inode [0x200000402:0x1126:0x0] mdc close failed: rc = -13 1[5723]: segfault at 8 ip 00007fd76da5e7e8 sp 00007ffca6a94580 error 4 in ld-2.17.so[7fd76da53000+22000] 5[9228]: segfault at 406000 ip 0000000000406000 sp 00007fff903294d8 error 14 in 5[606000+1000] Lustre: 9019:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 618, rollback = 7 Lustre: 9019:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 37 previous similar messages Lustre: 9019:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 9019:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 37 previous similar messages Lustre: 9019:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 9019:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 37 previous similar messages Lustre: 9019:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 9019:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 37 previous similar messages Lustre: 9019:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 9019:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 37 previous similar messages Lustre: 9019:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 9019:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 37 previous similar messages Lustre: 31382:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000402:0x1534:0x0] with magic=0xbd60bd0 Lustre: 31382:0:(lod_lov.c:1417:lod_parse_striping()) Skipped 3 previous similar messages 2[15983]: segfault at 8 ip 00007f97b02677e8 sp 00007fff57d24800 error 4 in ld-2.17.so[7f97b025c000+22000] 3[21349]: segfault at 1c ip 00000000004017b0 sp 00007ffd07fd4ce0 error 6 in 3[400000+6000] Lustre: 14712:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000401:0x1a21:0x0] with magic=0xbd60bd0 Lustre: 14712:0:(lod_lov.c:1417:lod_parse_striping()) Skipped 1 previous similar message 2[2812]: segfault at 1c ip 00000000004017b0 sp 00007ffd5b8e6d20 error 6 in 2[400000+6000] LustreError: 6520:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802f0deae98: inode [0x200000401:0x1d05:0x0] mdc close failed: rc = -13 3[7484]: segfault at 8 ip 00007feb9ce027e8 sp 00007ffd9fd5f370 error 4 in ld-2.17.so[7feb9cdf7000+22000] 13[12402]: segfault at 8 ip 00007f34471e27e8 sp 00007ffd6be3a550 error 4 in ld-2.17.so[7f34471d7000+22000] Lustre: 8321:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000401:0x1f64:0x0] with magic=0xbd60bd0 Lustre: 8321:0:(lod_lov.c:1417:lod_parse_striping()) Skipped 3 previous similar messages 0[15321]: segfault at 8 ip 00007f351c96a7e8 sp 00007fff5da30b30 error 4 in ld-2.17.so[7f351c95f000+22000] LustreError: 18081:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802eee11bf8: inode [0x200000401:0x2052:0x0] mdc close failed: rc = -13 14[20057]: segfault at 0 ip 0000000000403e5f sp 00007ffd6394e970 error 6 in 14[400000+6000] traps: 6[21223] trap invalid opcode ip:404f60 sp:7fffd4d68828 error:0 in 6[400000+6000] 17[23581]: segfault at 8 ip 00007fa58ecaa7e8 sp 00007ffed566c040 error 4 in ld-2.17.so[7fa58ec9f000+22000] 10[24068]: segfault at 1c ip 00000000004017b0 sp 00007fff54401410 error 6 in 10[400000+6000] 8[26564]: segfault at 8 ip 00007f0af75d67e8 sp 00007ffe73297c80 error 4 in ld-2.17.so[7f0af75cb000+22000] LustreError: 26566:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802eee11bf8: inode [0x200000402:0x237b:0x0] mdc close failed: rc = -13 Lustre: 15285:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0000: opcode 7: before 516 < left 618, rollback = 7 Lustre: 15285:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 213 previous similar messages Lustre: 15285:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 15285:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 213 previous similar messages Lustre: 15285:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 15285:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 213 previous similar messages Lustre: 15285:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 15285:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 213 previous similar messages Lustre: 15285:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 15285:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 213 previous similar messages Lustre: 15285:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 15285:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 213 previous similar messages 16[4106]: segfault at 8 ip 00007f2c86d3c7e8 sp 00007ffe002c1460 error 4 in ld-2.17.so[7f2c86d31000+22000] LustreError: 5626:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802f0deae98: inode [0x200000402:0x26fa:0x0] mdc close failed: rc = -13 LustreError: 5626:0:(file.c:247:ll_close_inode_openhandle()) Skipped 1 previous similar message 13[10065]: segfault at 8 ip 00007f590d8e07e8 sp 00007ffdf7eab430 error 4 in ld-2.17.so[7f590d8d5000+22000] 15[21009]: segfault at 8 ip 00007ff683d6d7e8 sp 00007ffc391fcdb0 error 4 in ld-2.17.so[7ff683d62000+22000] 13[22718]: segfault at 8 ip 00007f1bb06a47e8 sp 00007ffc62bcd2c0 error 4 in ld-2.17.so[7f1bb0699000+22000] LustreError: 30988:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802eee11bf8: inode [0x200000402:0x2fcb:0x0] mdc close failed: rc = -13 LustreError: 30988:0:(file.c:247:ll_close_inode_openhandle()) Skipped 3 previous similar messages LustreError: 7146:0:(statahead.c:2399:start_statahead_thread()) lustre: invalid pattern 0X0. 15[9407]: segfault at 8 ip 00007feb125fc7e8 sp 00007ffd54b1db70 error 4 in ld-2.17.so[7feb125f1000+22000] 11[12722]: segfault at 8 ip 00007f134a0dd7e8 sp 00007ffc45960720 error 4 in ld-2.17.so[7f134a0d2000+22000] 11[17081]: segfault at 8 ip 00007f032aee37e8 sp 00007ffd4029ecc0 error 4 in ld-2.17.so[7f032aed8000+22000] Lustre: 9668:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000402:0x3883:0x0] with magic=0xbd60bd0 Lustre: 9668:0:(lod_lov.c:1417:lod_parse_striping()) Skipped 7 previous similar messages 8[26686]: segfault at 8 ip 00007f9e859bf7e8 sp 00007fff69b7aba0 error 4 in ld-2.17.so[7f9e859b4000+22000] 7[28244]: segfault at 0 ip 0000000000403e5f sp 00007ffd1061c310 error 6 in 7[400000+6000] 2[29798]: segfault at 0 ip (null) sp 00007ffe1464d688 error 14 in 2[400000+6000] 10[4932]: segfault at 0 ip 0000000000403cf0 sp 00007fff87646968 error 6 in 10[400000+6000] 12[5433]: segfault at 8 ip 00007fa6a7b087e8 sp 00007ffeac964e90 error 4 in ld-2.17.so[7fa6a7afd000+22000] 8[10971]: segfault at 0 ip 0000000000403e5f sp 00007ffca5993c50 error 6 in 8[400000+6000] 18[21195]: segfault at 8 ip 00007f2794dfa7e8 sp 00007fffe7e50750 error 4 in ld-2.17.so[7f2794def000+22000] LustreError: 28119:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802f0deae98: inode [0x200000402:0x44d7:0x0] mdc close failed: rc = -13 LustreError: 28119:0:(file.c:247:ll_close_inode_openhandle()) Skipped 2 previous similar messages 8[30160]: segfault at 8 ip 00007f57a339f7e8 sp 00007fff91405fe0 error 4 in ld-2.17.so[7f57a3394000+22000] 9[32436]: segfault at 401ac0 ip 0000000000401ac0 sp 00007ffdc2b743c8 error 7 in 9[400000+6000] 4[4609]: segfault at 1c ip 00000000004017b0 sp 00007fffcf921470 error 6 in 4[400000+6000] 15[4871]: segfault at 8 ip 00007fce2c3017e8 sp 00007ffe86955420 error 4 in ld-2.17.so[7fce2c2f6000+22000] 7[16935]: segfault at 8 ip 00007f9d435487e8 sp 00007ffc4e026750 error 4 in ld-2.17.so[7f9d4353d000+22000] 4[17173]: segfault at 8 ip 00007fc0c07377e8 sp 00007ffd22dc30c0 error 4 in ld-2.17.so[7fc0c072c000+22000] 16[17989]: segfault at 8 ip 00007efe2f4e37e8 sp 00007fffb1d42c60 error 4 in ld-2.17.so[7efe2f4d8000+22000] 6[18524]: segfault at 8 ip 00007ff4b42067e8 sp 00007fff1a0fe180 error 4 in ld-2.17.so[7ff4b41fb000+22000] Lustre: 8319:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000402:0x4d45:0x0] with magic=0xbd60bd0 Lustre: 8319:0:(lod_lov.c:1417:lod_parse_striping()) Skipped 9 previous similar messages 3[20328]: segfault at 0 ip 00000000004016cd sp 00007fff096ea710 error 6 in 3[400000+6000] 8[25701]: segfault at 8 ip 00007eff347457e8 sp 00007ffdf97bae00 error 4 in ld-2.17.so[7eff3473a000+22000] Lustre: 15291:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 516 < left 612, rollback = 7 Lustre: 15291:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 379 previous similar messages Lustre: 15291:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 15291:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 379 previous similar messages Lustre: 15291:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 15291:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 379 previous similar messages Lustre: 15291:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/612/0, punch: 0/0/0, quota 4/150/0 Lustre: 15291:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 379 previous similar messages Lustre: 15291:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 15291:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 379 previous similar messages Lustre: 15291:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 15291:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 379 previous similar messages 16[30732]: segfault at 0 ip (null) sp 00007ffd9a8f7b18 error 14 in 16[400000+6000] 7[4376]: segfault at 4045bc ip 00000000004045bc sp 00007ffffcafaa58 error 7 in 7[400000+6000] 7[5474]: segfault at 8 ip 00007f49a72f57e8 sp 00007ffc9970cec0 error 4 in ld-2.17.so[7f49a72ea000+22000] 13[7139]: segfault at 8 ip 00007fcb27b6e7e8 sp 00007ffff36c86b0 error 4 in ld-2.17.so[7fcb27b63000+22000] 6[9268]: segfault at 0 ip 0000000000403e5f sp 00007ffff3712d50 error 6 in 5[400000+6000] LustreError: 15664:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802eee11bf8: inode [0x200000401:0x574d:0x0] mdc close failed: rc = -13 LustreError: 15664:0:(file.c:247:ll_close_inode_openhandle()) Skipped 6 previous similar messages 18[19535]: segfault at 8 ip 00007f234020f7e8 sp 00007ffe61f2b760 error 4 in ld-2.17.so[7f2340204000+22000] 18[20548]: segfault at 8 ip 00007f4dca8fb7e8 sp 00007ffc6fe8cc60 error 4 in ld-2.17.so[7f4dca8f0000+22000] 11[24736]: segfault at 8 ip 00007f6161ca37e8 sp 00007ffd4419d3f0 error 4 in ld-2.17.so[7f6161c98000+22000] 4[30477]: segfault at 0 ip 0000000000403e5f sp 00007ffc014878a0 error 6 in 4[400000+6000] 17[32404]: segfault at 1c ip 00000000004017b0 sp 00007ffcabd3cb10 error 6 in 19[400000+6000] 15[3926]: segfault at 7ffe44946ff8 ip 00000000004032d8 sp 00007ffe44947000 error 6 in 15[400000+6000] 6[8597]: segfault at 8 ip 00007fbbb1dbd7e8 sp 00007ffd5673f500 error 4 in ld-2.17.so[7fbbb1db2000+22000] 9[9996]: segfault at 0 ip 0000000000403e5f sp 00007ffe689ea6a0 error 6 in 8 (deleted)[400000+6000] Lustre: mdt00_001: service thread pid 8320 was inactive for 40.098 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 8320, comm: mdt00_001 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock+0x88/0x1c0 [mdt] [<0>] mdt_object_find_lock+0x54/0x170 [mdt] [<0>] mdt_reint_setxattr+0x133/0x10e0 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x84f/0x13c0 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 20124, comm: mdt00_017 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock_try+0xa0/0x250 [mdt] [<0>] mdt_getattr_name_lock+0x173b/0x2d10 [mdt] [<0>] mdt_intent_getattr+0x2cc/0x4e0 [mdt] [<0>] mdt_intent_opc.constprop.76+0x211/0xc60 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x397/0x980 [ptlrpc] [<0>] ldlm_handle_enqueue+0x547/0x18d0 [ptlrpc] [<0>] tgt_enqueue+0x68/0x240 [ptlrpc] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 14736, comm: mdt00_011 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock+0x88/0x1c0 [mdt] [<0>] mdt_intent_getxattr+0x80/0x330 [mdt] [<0>] mdt_intent_opc.constprop.76+0x211/0xc60 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x397/0x980 [ptlrpc] [<0>] ldlm_handle_enqueue+0x547/0x18d0 [ptlrpc] [<0>] tgt_enqueue+0x68/0x240 [ptlrpc] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: mdt00_018: service thread pid 19570 was inactive for 40.121 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: mdt00_007: service thread pid 14676 was inactive for 40.096 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 1 previous similar message LustreError: 8189:0:(ldlm_lockd.c:252:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0000_UUID lock: ffff8802df2b5e00/0x11926b7eadb31884 lrc: 3/0,0 mode: PR/PR res: [0x200000402:0x60ea:0x0].0x0 bits 0x1b/0x0 rrc: 8 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0x11926b7eadb31868 expref: 1265 pid: 14669 timeout: 1532 lvb_type: 0 Lustre: mdt00_011: service thread pid 14736 completed after 100.159s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_001: service thread pid 8320 completed after 100.299s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_018: service thread pid 19570 completed after 100.061s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 8184:0:(ldlm_lockd.c:2591:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1742269825 with bad export cookie 1266192637054693604 Lustre: lustre-MDT0000-mdc-ffff8802eee11bf8: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: lustre-MDT0000-mdc-ffff8802eee11bf8: This client was evicted by lustre-MDT0000; in progress operations using this service will fail. Lustre: mdt00_017: service thread pid 20124 completed after 100.210s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_000: service thread pid 8338 completed after 99.783s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 14676:0:(ldlm_lockd.c:1492:ldlm_handle_enqueue()) ### lock on destroyed export ffff8800a918a548 ns: mdt-lustre-MDT0000_UUID lock: ffff8802c42c25c0/0x11926b7eadb3348b lrc: 3/0,0 mode: PR/PR res: [0x200000402:0x60ea:0x0].0x0 bits 0x1b/0x0 rrc: 3 type: IBT gid 0 flags: 0x50200000000000 nid: 0@lo remote: 0x11926b7eadb33001 expref: 17 pid: 14676 timeout: 0 lvb_type: 0 Lustre: mdt00_007: service thread pid 14676 completed after 99.544s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: lustre-MDT0000-mdc-ffff8802eee11bf8: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: 11903:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x200000402:0x60ea:0x0] error -108. LustreError: 11105:0:(file.c:6147:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000007:0x1:0x0] error: rc = -108 Lustre: lustre-MDT0000-mdc-ffff8802eee11bf8: Connection restored to 192.168.123.52@tcp (at 0@lo) 12[19038]: segfault at 8 ip 00007fac9a61a7e8 sp 00007ffda7f20190 error 4 in ld-2.17.so[7fac9a60f000+22000] 12[19063]: segfault at 8 ip 00007f79749817e8 sp 00007fff0ba7e430 error 4 in ld-2.17.so[7f7974976000+22000] 18[23502]: segfault at 0 ip (null) sp 00007ffd66d1a5f8 error 14 in 18[400000+6000] 9[26637]: segfault at 8 ip 00007fc0062ed7e8 sp 00007ffc5bcc9f50 error 4 in ld-2.17.so[7fc0062e2000+22000] 7[28737]: segfault at 1401420 ip 0000000001401420 sp 00007ffe91a122c8 error 14 9[29220]: segfault at 406000 ip 0000000000406000 sp 00007ffcd4555978 error 14 in 9[606000+1000] 16[29523]: segfault at 8 ip 00007fb870a5e7e8 sp 00007ffc6f4b28e0 error 4 in ld-2.17.so[7fb870a53000+22000] 18[30805]: segfault at 8 ip 00007fc1368437e8 sp 00007fff3d628b40 error 4 in ld-2.17.so[7fc136838000+22000] 6[3102]: segfault at 8 ip 00007fa3b9dc57e8 sp 00007ffc9931fe20 error 4 in ld-2.17.so[7fa3b9dba000+22000] 17[5863]: segfault at 8 ip 00007f4bc44147e8 sp 00007fffcc84eba0 error 4 in ld-2.17.so[7f4bc4409000+22000] 11[10629]: segfault at 8 ip 00007f608d6667e8 sp 00007ffffd55b940 error 4 in ld-2.17.so[7f608d65b000+22000] 16[23198]: segfault at 406000 ip 0000000000406000 sp 00007ffe0d8bfc28 error 14 in 16[606000+1000] 9[23163]: segfault at 406000 ip 0000000000406000 sp 00007fff4ae70358 error 14 in 9[606000+1000] 7[26947]: segfault at 0 ip 00000000004016cd sp 00007fffdbc3b6b0 error 6 in 7[400000+6000] 0[27895]: segfault at 8 ip 00007fad8d1ed7e8 sp 00007fff8ae7ef00 error 4 in ld-2.17.so[7fad8d1e2000+22000] 7[1395]: segfault at 8 ip 00007f79d7af37e8 sp 00007ffcc5858a70 error 4 in ld-2.17.so[7f79d7ae8000+22000] 18[3166]: segfault at 8 ip 00007fa7b87a97e8 sp 00007ffd91e0e260 error 4 in ld-2.17.so[7fa7b879e000+22000] 17[3878]: segfault at 8 ip 00007f4690c427e8 sp 00007ffdc1468a30 error 4 in ld-2.17.so[7f4690c37000+22000] Lustre: 14733:0:(lod_lov.c:1417:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000401:0x75e3:0x0] with magic=0xbd60bd0 Lustre: 14733:0:(lod_lov.c:1417:lod_parse_striping()) Skipped 13 previous similar messages 10[17142]: segfault at 0 ip (null) sp 00007ffd1522d4c8 error 14 in 10[400000+6000] Lustre: 9019:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0000: opcode 7: before 516 < left 618, rollback = 7 Lustre: 9019:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 623 previous similar messages Lustre: 9019:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 9019:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 623 previous similar messages Lustre: 9019:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 9019:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 623 previous similar messages Lustre: 9019:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 9019:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 623 previous similar messages Lustre: 9019:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 9019:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 623 previous similar messages Lustre: 9019:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 9019:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 623 previous similar messages LustreError: 23863:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802f0deae98: inode [0x200000401:0x7a8b:0x0] mdc close failed: rc = -13 LustreError: 23863:0:(file.c:247:ll_close_inode_openhandle()) Skipped 23 previous similar messages 10[32450]: segfault at 8 ip 00007f8e61f877e8 sp 00007fff7f1a0cb0 error 4 in ld-2.17.so[7f8e61f7c000+22000] 19[1652]: segfault at 8 ip 00007f1a49b057e8 sp 00007ffe8520c750 error 4 in ld-2.17.so[7f1a49afa000+22000] | Link to test |
racer test 1: racer on clients: centos-30.localnet DURATION=2700 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 PGD 2a67b3067 PUD 2b3043067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) osp(OE) ofd(OE) lod(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) crc32_generic lnet(OE) libcfs(OE) crc_t10dif crct10dif_generic crct10dif_common pcspkr i2c_piix4 virtio_balloon virtio_console ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm ata_piix drm_panel_orientation_quirks serio_raw virtio_blk floppy i2c_core libata CPU: 7 PID: 31048 Comm: ll_sa_30681 Kdump: loaded Tainted: P OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.16.0-3.module_el8.7.0+1218+f626c2ff 04/01/2014 task: ffff88025ebaa4f0 ti: ffff880276ba4000 task.ti: ffff880276ba4000 RIP: 0010:[<ffffffff813f502c>] [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 RSP: 0018:ffff880276ba7d80 EFLAGS: 00010282 RAX: ffff88007808f608 RBX: 0000000000000008 RCX: 0000000100260025 RDX: 0000000000000026 RSI: ffff88007808f938 RDI: 0000000000000008 RBP: ffff880276ba7d90 R08: ffff8802a7219098 R09: 0000000000000000 R10: ffff8802a72193e8 R11: ffff8802a7219098 R12: 0000000000000000 R13: ffff88025eba8010 R14: ffff88007808f938 R15: ffff8802a7219098 FS: 0000000000000000(0000) GS:ffff880331bc0000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 00000002827f2000 CR4: 00000000000007e0 Call Trace: [<ffffffffa15c55d1>] ll_statahead_thread+0xde1/0x2290 [lustre] [<ffffffff810df6e2>] ? pick_next_task_fair+0x702/0x8a0 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e329e>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffff810c834d>] ? finish_task_switch+0x5d/0x1b0 [<ffffffffa15c47f0>] ? ll_statahead_interpret+0x4a0/0x4a0 [lustre] [<ffffffff810ba114>] kthread+0xe4/0xf0 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 | Lustre: 22048:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000402:0x10:0x0] with magic=0xbd60bd0 dd (22578) used greatest stack depth: 9856 bytes left Lustre: 22048:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff88030f03ad40 x1826456177498368/t4294967970(0) o101->d40d7219-ec60-4e23-b14d-8593b1e0b504@0@lo:533/0 lens 376/864 e 0 to 0 dl 1741844423 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 Lustre: 26848:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000402:0x130:0x0] with magic=0xbd60bd0 Lustre: 26848:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message 18[3120]: segfault at 8 ip 00007f500a13d7e8 sp 00007ffc5e380300 error 4 in ld-2.17.so[7f500a132000+22000] 18[3585]: segfault at 8 ip 00007f7cd99ee7e8 sp 00007ffece28d1b0 error 4 in ld-2.17.so[7f7cd99e3000+22000] 13[13071]: segfault at 8 ip 00007fb8bb7437e8 sp 00007ffc593d5b90 error 4 in ld-2.17.so[7fb8bb738000+22000] LustreError: 22143:0:(mdt_handler.c:746:mdt_pack_acl2body()) lustre-MDT0000: unable to read [0x200000401:0x23f:0x0] ACL: rc = -2 Lustre: 22647:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000401:0x752:0x0] with magic=0xbd60bd0 Lustre: 22647:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message 0[16313]: segfault at 0 ip 0000000000403e5f sp 00007fffaf77ab70 error 6 in 0[400000+6000] 15[18375]: segfault at 0 ip 0000000000403e5f sp 00007ffe9e846ae0 error 6 in 15[400000+6000] LustreError: 22129:0:(mdt_handler.c:746:mdt_pack_acl2body()) lustre-MDT0000: unable to read [0x200000402:0x913:0x0] ACL: rc = -2 Lustre: 32421:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000401:0xa96:0x0] with magic=0xbd60bd0 Lustre: 32421:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message Lustre: 22138:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000402:0xc07:0x0] with magic=0xbd60bd0 Lustre: 22138:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message LustreError: 21687:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a63e37e8: inode [0x200000401:0xcc8:0x0] mdc close failed: rc = -13 LustreError: 29125:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a63e37e8: inode [0x200000401:0xdc6:0x0] mdc close failed: rc = -13 LustreError: 12828:0:(mdt_handler.c:746:mdt_pack_acl2body()) lustre-MDT0000: unable to read [0x200000401:0xefc:0x0] ACL: rc = -2 10[9073]: segfault at 8 ip 00007f27e9c857e8 sp 00007ffc02ddfc80 error 4 in ld-2.17.so[7f27e9c7a000+22000] 10[13200]: segfault at 0 ip 0000000000403e5f sp 00007fffe4e8c120 error 6 in 10[400000+6000] LustreError: 17619:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a63e37e8: inode [0x200000401:0x1062:0x0] mdc close failed: rc = -13 Lustre: 22138:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000402:0x10fd:0x0] with magic=0xbd60bd0 Lustre: 22138:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message LustreError: 27747:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802de412548: inode [0x200000402:0x120f:0x0] mdc close failed: rc = -13 LustreError: 12210:0:(ldlm_lockd.c:252:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0002_UUID lock: ffff88009f3016c0/0x7eafe8a3e9a6ef19 lrc: 3/0,0 mode: PW/PW res: [0x2c0000400:0x101:0x0].0x0 rrc: 4 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000000020020 nid: 0@lo remote: 0x7eafe8a3e9a6ef12 expref: 26 pid: 16017 timeout: 390 lvb_type: 0 LustreError: lustre-OST0002-osc-ffff8800a63e37e8: operation ost_read to node 0@lo failed: rc = -107 Lustre: lustre-OST0002-osc-ffff8800a63e37e8: Connection to lustre-OST0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: Skipped 1 previous similar message LustreError: 3657:0:(osc_request.c:1027:osc_init_grant()) lustre-OST0002-osc-ffff8800a63e37e8: granted 3407872 but already consumed 5111808 LustreError: lustre-OST0002-osc-ffff8800a63e37e8: This client was evicted by lustre-OST0002; in progress operations using this service will fail. Lustre: 3660:0:(llite_lib.c:4136:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.32@tcp:/lustre/fid: [0x200000401:0xbdb:0x0]// may get corrupted (rc -108) Lustre: 3663:0:(llite_lib.c:4136:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.32@tcp:/lustre/fid: [0x200000401:0x124d:0x0]// may get corrupted (rc -108) Lustre: lustre-OST0002-osc-ffff8800a63e37e8: Connection restored to 192.168.123.32@tcp (at 0@lo) 0[8832]: segfault at 8 ip 00007f2b2901d7e8 sp 00007ffc76703500 error 4 in ld-2.17.so[7f2b29012000+22000] 4[17774]: segfault at 8 ip 00007f7eb766b7e8 sp 00007ffe013ff4b0 error 4 in ld-2.17.so[7f7eb7660000+22000] 9[6863]: segfault at 8 ip 00007f193cb727e8 sp 00007ffd3caa8220 error 4 in ld-2.17.so[7f193cb67000+22000] LustreError: 23837:0:(osd_io.c:898:osd_grow_blocksize()) object [0x280000400:0x21d:0x0]: change block size65536 -> 1048576 error rc = -95 Lustre: 22129:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000401:0x1a88:0x0] with magic=0xbd60bd0 Lustre: 22129:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message 3[24059]: segfault at 8 ip 00007ff4d50227e8 sp 00007ffd32ac12c0 error 4 in ld-2.17.so[7ff4d5017000+22000] LustreError: 397:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a63e37e8: inode [0x200000402:0x2042:0x0] mdc close failed: rc = -13 18[3678]: segfault at 8 ip 00007f0a111357e8 sp 00007ffdc2d068e0 error 4 in ld-2.17.so[7f0a1112a000+22000] Lustre: 26848:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000402:0x2092:0x0] with magic=0xbd60bd0 Lustre: 26848:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message 16[18585]: segfault at 8 ip 00007f81e39c17e8 sp 00007fff792c6d60 error 4 in ld-2.17.so[7f81e39b6000+22000] 8[24413]: segfault at 0 ip (null) sp 00007ffe81bbef08 error 14 in 8[400000+6000] Lustre: 7951:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000401:0x25a0:0x0] with magic=0xbd60bd0 Lustre: 7951:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 5 previous similar messages 9[13597]: segfault at 8 ip 00007fb44ee3c7e8 sp 00007ffe7013e3b0 error 4 in ld-2.17.so[7fb44ee31000+22000] Lustre: mdt00_010: service thread pid 22629 was inactive for 40.152 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 12851, comm: mdt_io00_000 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_check_lock+0xec/0x3c0 [mdt] [<0>] mdt_reint_rename+0x1da6/0x2c30 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x74c/0xbc0 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 12828, comm: mdt00_002 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_pdo_lock+0x369/0x750 [mdt] [<0>] mdt_parent_lock+0x76/0x2a0 [mdt] [<0>] mdt_getattr_name_lock+0x1850/0x2d10 [mdt] [<0>] mdt_intent_getattr+0x2cc/0x4e0 [mdt] [<0>] mdt_intent_opc.constprop.73+0x211/0xc60 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x397/0x980 [ptlrpc] [<0>] ldlm_handle_enqueue+0x547/0x18d0 [ptlrpc] [<0>] tgt_enqueue+0x68/0x240 [ptlrpc] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: Skipped 2 previous similar messages Pid: 22629, comm: mdt00_010 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock+0x88/0x1c0 [mdt] [<0>] mdt_object_stripes_lock+0x126/0x660 [mdt] [<0>] mdt_reint_setattr+0x73b/0x15f0 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x74c/0xbc0 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: mdt00_016: service thread pid 26848 was inactive for 40.084 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. LustreError: 12210:0:(ldlm_lockd.c:252:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0000_UUID lock: ffff8802c5ff52c0/0x7eafe8a3e9de0b06 lrc: 3/0,0 mode: PR/PR res: [0x200000401:0x25f2:0x0].0x0 bits 0x13/0x0 rrc: 7 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0x7eafe8a3e9de0ae3 expref: 1012 pid: 22647 timeout: 763 lvb_type: 0 LustreError: 12851:0:(client.c:1282:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8800a4680a40 x1826456271889920/t0(0) o104->lustre-MDT0000@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' uid:4294967295 gid:4294967295 Lustre: mdt00_010: service thread pid 22629 completed after 100.383s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_016: service thread pid 26848 completed after 100.189s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 21206:0:(ldlm_lockd.c:2591:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1741845075 with bad export cookie 9128770760398893655 Lustre: lustre-MDT0000-mdc-ffff8802de412548: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: lustre-MDT0000-mdc-ffff8802de412548: This client was evicted by lustre-MDT0000; in progress operations using this service will fail. LustreError: 15319:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802de412548: inode [0x200000401:0x25ff:0x0] mdc close failed: rc = -108 LustreError: 15319:0:(file.c:247:ll_close_inode_openhandle()) Skipped 1 previous similar message Lustre: mdt_io00_000: service thread pid 12851 completed after 100.411s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_002: service thread pid 12828 completed after 100.387s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: lustre-MDT0000-mdc-ffff8800a63e37e8: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: lustre-MDT0000-mdc-ffff8800a63e37e8: This client was evicted by lustre-MDT0000; in progress operations using this service will fail. LustreError: 14342:0:(file.c:6081:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000007:0x1:0x0] error: rc = -108 Lustre: lustre-MDT0000-mdc-ffff8802de412548: Connection restored to 192.168.123.32@tcp (at 0@lo) 9[23063]: segfault at 4017b0 ip 00000000004017b0 sp 00007ffc34465a80 error 14 5[4259]: segfault at 0 ip (null) sp 00007fff8ca9a678 error 14 in 5[400000+6000] LustreError: 6299:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802de412548: inode [0x200000404:0x2c6:0x0] mdc close failed: rc = -13 LustreError: 6299:0:(file.c:247:ll_close_inode_openhandle()) Skipped 11 previous similar messages Lustre: 7951:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0x38a:0x0] with magic=0xbd60bd0 Lustre: 7951:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 3 previous similar messages 9[19348]: segfault at 4045bc ip 00000000004045bc sp 00007ffc49b7f368 error 7 in 8 (deleted)[400000+6000] 7[31318]: segfault at 8 ip 00007fc7a24987e8 sp 00007fffabe5d670 error 4 in ld-2.17.so[7fc7a248d000+22000] 7[31410]: segfault at 8 ip 00007f0ffb25b7e8 sp 00007ffe9561d430 error 4 in ld-2.17.so[7f0ffb250000+22000] LustreError: 3940:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802de412548: inode [0x200000403:0x744:0x0] mdc close failed: rc = -13 LustreError: 3940:0:(file.c:247:ll_close_inode_openhandle()) Skipped 2 previous similar messages 9[5528]: segfault at 8 ip 00007fb4295ad7e8 sp 00007ffe53c23010 error 4 in ld-2.17.so[7fb4295a2000+22000] 7[7253]: segfault at 8 ip 00007fe5aabc37e8 sp 00007ffd49160ec0 error 4 in ld-2.17.so[7fe5aabb8000+22000] LustreError: 24823:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802de412548: inode [0x200000403:0xf02:0x0] mdc close failed: rc = -13 LustreError: 24823:0:(file.c:247:ll_close_inode_openhandle()) Skipped 2 previous similar messages 13[8911]: segfault at 8 ip 00007fd1ff81a7e8 sp 00007ffc26624070 error 4 in ld-2.17.so[7fd1ff80f000+22000] 12[29341]: segfault at 0 ip 0000000000403e5f sp 00007ffe7067e5b0 error 6 in 12[400000+6000] 11[1162]: segfault at 0 ip 0000000000403e5f sp 00007ffedc36fbe0 error 6 in 11[400000+6000] LustreError: 22138:0:(mdt_handler.c:746:mdt_pack_acl2body()) lustre-MDT0000: unable to read [0x200000404:0x1496:0x0] ACL: rc = -2 LustreError: 22138:0:(mdt_handler.c:746:mdt_pack_acl2body()) Skipped 1 previous similar message 16[4644]: segfault at 8 ip 00007f260bd0d7e8 sp 00007ffd1f8ee4e0 error 4 in ld-2.17.so[7f260bd02000+22000] Lustre: 22027:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0x1a22:0x0] with magic=0xbd60bd0 Lustre: 22027:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 9 previous similar messages 10[12907]: segfault at 8 ip 00007fa1d93037e8 sp 00007ffe53266200 error 4 in ld-2.17.so[7fa1d92f8000+22000] 12[28668]: segfault at 8 ip 00007f5d3f03e7e8 sp 00007ffe4844f7f0 error 4 in ld-2.17.so[7f5d3f033000+22000] LustreError: 4295:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a63e37e8: inode [0x200000403:0x1e10:0x0] mdc close failed: rc = -13 4[28915]: segfault at 8 ip 00007fedb99417e8 sp 00007ffcda504c70 error 4 in ld-2.17.so[7fedb9936000+22000] 13[4124]: segfault at 8 ip 00007f76e5c397e8 sp 00007ffcc632e4d0 error 4 in ld-2.17.so[7f76e5c2e000+22000] 19[4126]: segfault at 8 ip 00007f7f57f867e8 sp 00007ffeea6255e0 error 4 in ld-2.17.so[7f7f57f7b000+22000] LustreError: 12210:0:(ldlm_lockd.c:252:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0000_UUID lock: ffff8802b56661c0/0x7eafe8a3ea2a2f5f lrc: 3/0,0 mode: PR/PR res: [0x200000403:0x23e3:0x0].0x0 bits 0x1b/0x0 rrc: 11 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0x7eafe8a3ea2a2f35 expref: 971 pid: 11179 timeout: 1325 lvb_type: 0 LustreError: 12210:0:(ldlm_lockd.c:252:expired_lock_main()) Skipped 1 previous similar message LustreError: 22629:0:(ldlm_lockd.c:1492:ldlm_handle_enqueue()) ### lock on destroyed export ffff8802d503c138 ns: mdt-lustre-MDT0000_UUID lock: ffff8802740ad680/0x7eafe8a3ea2a3e9a lrc: 3/0,0 mode: PR/PR res: [0x200000403:0x23e3:0x0].0x0 bits 0x1b/0x0 rrc: 9 type: IBT gid 0 flags: 0x50200000000000 nid: 0@lo remote: 0x7eafe8a3ea2a3e7e expref: 493 pid: 22629 timeout: 0 lvb_type: 0 LustreError: lustre-MDT0000-mdc-ffff8802de412548: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-MDT0000-mdc-ffff8802de412548: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 1 previous similar message LustreError: Skipped 2 previous similar messages LustreError: lustre-MDT0000-mdc-ffff8802de412548: This client was evicted by lustre-MDT0000; in progress operations using this service will fail. LustreError: 14490:0:(file.c:6081:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000403:0x23e3:0x0] error: rc = -5 LustreError: 14196:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x200000403:0x23e3:0x0] error -108. LustreError: 15206:0:(llite_lib.c:1995:ll_md_setattr()) md_setattr fails: rc = -108 LustreError: 14490:0:(file.c:6081:ll_inode_revalidate_fini()) Skipped 27 previous similar messages LustreError: 15035:0:(mdc_request.c:1464:mdc_read_page()) lustre-MDT0000-mdc-ffff8802de412548: [0x200000401:0x1:0x0] lock enqueue fails: rc = -108 Lustre: lustre-MDT0000-mdc-ffff8802de412548: Connection restored to 192.168.123.32@tcp (at 0@lo) Lustre: Skipped 1 previous similar message 0[16913]: segfault at 8 ip 00007f5eb2e107e8 sp 00007fff29b87b10 error 4 in ld-2.17.so[7f5eb2e05000+22000] 2[4966]: segfault at 0 ip (null) sp 00007ffcb8a206b8 error 14 in 2[400000+6000] 17[28864]: segfault at 0 ip 0000000000403e5f sp 00007ffc79cc5370 error 6 in 17[400000+6000] 1[22327]: segfault at 8 ip 00007f5106ad57e8 sp 00007fffa9739490 error 4 in ld-2.17.so[7f5106aca000+22000] LustreError: 32280:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a63e37e8: inode [0x200000405:0xbef:0x0] mdc close failed: rc = -13 LustreError: 32280:0:(file.c:247:ll_close_inode_openhandle()) Skipped 12 previous similar messages 17[3794]: segfault at 0 ip (null) sp 00007ffe834588d8 error 14 in 17[400000+6000] 13[8829]: segfault at 8 ip 00007f55f87b87e8 sp 00007ffd3cb17890 error 4 in ld-2.17.so[7f55f87ad000+22000] 1[11922]: segfault at 8 ip 00007f42adfc17e8 sp 00007ffcdd5b46e0 error 4 in ld-2.17.so[7f42adfb6000+22000] LustreError: 17526:0:(statahead.c:2394:start_statahead_thread()) lustre: invalid pattern 0X0. 14[22325]: segfault at 0 ip (null) sp 00007ffffeb1fbd8 error 14 in 14[400000+6000] Lustre: 11179:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000404:0x3c15:0x0] with magic=0xbd60bd0 Lustre: 11179:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 17 previous similar messages | Link to test |
racer test 1: racer on clients: centos-30.localnet DURATION=2700 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 PGD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) osp(OE) ofd(OE) lod(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) crc32_generic lnet(OE) libcfs(OE) crc_t10dif crct10dif_generic crct10dif_common pcspkr virtio_balloon virtio_console i2c_piix4 ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm ata_piix drm_panel_orientation_quirks virtio_blk serio_raw i2c_core libata floppy CPU: 10 PID: 23829 Comm: ll_sa_23624 Kdump: loaded Tainted: P OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.16.0-3.module_el8.7.0+1218+f626c2ff 04/01/2014 task: ffff8802896c49d0 ti: ffff8802c1994000 task.ti: ffff8802c1994000 RIP: 0010:[<ffffffff813f502c>] [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 RSP: 0018:ffff8802c1997d80 EFLAGS: 00010282 RAX: ffff88024c386c48 RBX: 0000000000000008 RCX: dead000000000200 RDX: 000000000000000a RSI: ffff88024c386f78 RDI: 0000000000000008 RBP: ffff8802c1997d90 R08: ffff880331403660 R09: 0000000000000000 R10: ffff880331403640 R11: ffff8802d99a0358 R12: 0000000000000000 R13: ffff8800af6324f0 R14: ffff88024c386f78 R15: ffff8802d99a0358 FS: 0000000000000000(0000) GS:ffff880331c80000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 0000000001c10000 CR4: 00000000000007e0 Call Trace: [<ffffffffa15a05d1>] ll_statahead_thread+0xde1/0x2290 [lustre] [<ffffffff817e320e>] ? _raw_spin_lock+0x1e/0x20 [<ffffffff810dbb51>] ? put_prev_entity+0x31/0x400 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e329e>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffff810c834d>] ? finish_task_switch+0x5d/0x1b0 [<ffffffffa159f7f0>] ? ll_statahead_interpret+0x4a0/0x4a0 [lustre] [<ffffffff810ba114>] kthread+0xe4/0xf0 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 | 16[22914]: segfault at 8 ip 00007efd85c797e8 sp 00007ffcde95dca0 error 4 in ld-2.17.so[7efd85c6e000+22000] Lustre: 22429:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff88009e340040 x1826416481004416/t4294968743(0) o101->b05aaa4a-979e-4097-add0-6ba3b7bb9047@0@lo:424/0 lens 376/840 e 0 to 0 dl 1741806564 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 17[28770]: segfault at 8 ip 00007f58a5f2f7e8 sp 00007fff5ff1ba80 error 4 in ld-2.17.so[7f58a5f24000+22000] LustreError: 28770:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a4522e98: inode [0x200000401:0xb4:0x0] mdc close failed: rc = -13 Lustre: 22347:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000401:0xfa:0x0] with magic=0xbd60bd0 Lustre: 22909:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000401:0xfa:0x0] with magic=0xbd60bd0 Lustre: 22909:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message Lustre: mdt00_012: service thread pid 22909 was inactive for 40.040 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 22909, comm: mdt00_012 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock_try+0xa0/0x250 [mdt] [<0>] mdt_getattr_name_lock+0x173b/0x2d10 [mdt] [<0>] mdt_intent_getattr+0x2cc/0x4e0 [mdt] [<0>] mdt_intent_opc.constprop.73+0x211/0xc60 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x397/0x980 [ptlrpc] [<0>] ldlm_handle_enqueue+0x547/0x18d0 [ptlrpc] [<0>] tgt_enqueue+0x68/0x240 [ptlrpc] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 12768, comm: mdt00_001 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock+0x88/0x1c0 [mdt] [<0>] mdt_object_find_lock+0x54/0x170 [mdt] [<0>] mdt_reint_setxattr+0x133/0x10e0 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x74c/0xbc0 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 22436, comm: mdt00_006 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock_try+0xa0/0x250 [mdt] [<0>] mdt_getattr_name_lock+0x173b/0x2d10 [mdt] [<0>] mdt_intent_getattr+0x2cc/0x4e0 [mdt] [<0>] mdt_intent_opc.constprop.73+0x211/0xc60 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x397/0x980 [ptlrpc] [<0>] ldlm_handle_enqueue+0x547/0x18d0 [ptlrpc] [<0>] tgt_enqueue+0x68/0x240 [ptlrpc] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe LustreError: 12423:0:(ldlm_lockd.c:252:expired_lock_main()) ### lock callback timer expired after 101s: evicting client at 0@lo ns: mdt-lustre-MDT0000_UUID lock: ffff8802d613de00/0x9f9c330d55c30a25 lrc: 3/0,0 mode: CR/CR res: [0x200000401:0x13f:0x0].0x0 bits 0xa/0x0 rrc: 6 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0x9f9c330d55c30a17 expref: 94 pid: 22850 timeout: 206 lvb_type: 0 Lustre: mdt00_001: service thread pid 12768 completed after 100.476s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 22436:0:(ldlm_lockd.c:1492:ldlm_handle_enqueue()) ### lock on destroyed export ffff8802da8e9bf8 ns: mdt-lustre-MDT0000_UUID lock: ffff8802d6138b80/0x9f9c330d55c32514 lrc: 3/0,0 mode: PR/PR res: [0x200000401:0x13f:0x0].0x0 bits 0x1b/0x0 rrc: 4 type: IBT gid 0 flags: 0x50200000000000 nid: 0@lo remote: 0x9f9c330d55c324e3 expref: 4 pid: 22436 timeout: 0 lvb_type: 0 Lustre: mdt00_006: service thread pid 22436 completed after 100.243s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: lustre-MDT0000-mdc-ffff8802da8ec138: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-MDT0000-mdc-ffff8802da8ec138: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: mdt00_012: service thread pid 22909 completed after 100.507s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: lustre-MDT0000-mdc-ffff8802da8ec138: This client was evicted by lustre-MDT0000; in progress operations using this service will fail. LustreError: 32317:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802da8ec138: inode [0x200000401:0x100:0x0] mdc close failed: rc = -108 Lustre: lustre-MDT0000-mdc-ffff8802da8ec138: Connection restored to (at 0@lo) LustreError: 24623:0:(mdt_handler.c:746:mdt_pack_acl2body()) lustre-MDT0000: unable to read [0x200000402:0xf3:0x0] ACL: rc = -2 9[3091]: segfault at 0 ip (null) sp 00007ffd7d5d1128 error 14 in 10 (deleted)[400000+6000] LustreError: 22852:0:(mdt_handler.c:746:mdt_pack_acl2body()) lustre-MDT0000: unable to read [0x200000403:0x20c:0x0] ACL: rc = -2 8[14338]: segfault at 8 ip 00007f0a05c9a7e8 sp 00007ffcdbcdfa00 error 4 in ld-2.17.so[7f0a05c8f000+22000] 0[15067]: segfault at 8 ip 00007f025f2797e8 sp 00007fffc6ee4aa0 error 4 in ld-2.17.so[7f025f26e000+22000] 12[2602]: segfault at 8 ip 00007f5d388607e8 sp 00007ffe3a582030 error 4 in ld-2.17.so[7f5d38855000+22000] LustreError: 6162:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a4522e98: inode [0x200000403:0x972:0x0] mdc close failed: rc = -13 LustreError: 6162:0:(file.c:247:ll_close_inode_openhandle()) Skipped 8 previous similar messages 8[10868]: segfault at 8 ip 00007f9838ee47e8 sp 00007ffec599ef10 error 4 in ld-2.17.so[7f9838ed9000+22000] LustreError: 31209:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802da8ec138: inode [0x200000402:0xec9:0x0] mdc close failed: rc = -13 LustreError: 1157:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802da8ec138: inode [0x200000402:0xf14:0x0] mdc close failed: rc = -13 LustreError: 19191:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802da8ec138: inode [0x200000402:0x10d5:0x0] mdc close failed: rc = -13 LustreError: 22436:0:(mdt_handler.c:746:mdt_pack_acl2body()) lustre-MDT0000: unable to read [0x200000402:0x129c:0x0] ACL: rc = -2 LustreError: 28595:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a4522e98: inode [0x200000402:0x12fc:0x0] mdc close failed: rc = -13 7[3773]: segfault at 8 ip 00007fd7074fe7e8 sp 00007ffdf88ee3c0 error 4 in ld-2.17.so[7fd7074f3000+22000] Lustre: 22436:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000402:0x14f8:0x0] with magic=0xbd60bd0 Lustre: 22436:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message LustreError: 13056:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802da8ec138: inode [0x200000402:0x149b:0x0] mdc close failed: rc = -13 LustreError: 13056:0:(file.c:247:ll_close_inode_openhandle()) Skipped 1 previous similar message 12[26149]: segfault at 8 ip 00007f3dbcb007e8 sp 00007ffc8fb5c2d0 error 4 in ld-2.17.so[7f3dbcaf5000+22000] 16[5253]: segfault at 8 ip 00007fe7b835f7e8 sp 00007ffce6de1d90 error 4 in ld-2.17.so[7fe7b8354000+22000] Lustre: 22429:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0x1793:0x0] with magic=0xbd60bd0 Lustre: 22429:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message 1[23789]: segfault at 8 ip 00007fdff2b787e8 sp 00007ffc16916550 error 4 in ld-2.17.so[7fdff2b6d000+22000] 15[1417]: segfault at 8 ip 00007fdb579d77e8 sp 00007ffeae8ca410 error 4 in ld-2.17.so[7fdb579cc000+22000] Lustre: 15240:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0x1b04:0x0] with magic=0xbd60bd0 Lustre: 15240:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message 15[5970]: segfault at 8 ip 00007f32d14407e8 sp 00007ffe08644110 error 4 in ld-2.17.so[7f32d1435000+22000] 12[8252]: segfault at 1456 ip 0000000000001456 sp 00007ffecd0a4468 error 14 in 12[400000+6000] 1[13013]: segfault at 8 ip 00007fbcd8c967e8 sp 00007ffff643f350 error 4 in ld-2.17.so[7fbcd8c8b000+22000] LustreError: 13013:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802da8ec138: inode [0x200000403:0x1bd7:0x0] mdc close failed: rc = -13 8[24163]: segfault at 8 ip 00007fe1277d37e8 sp 00007ffc03a94d80 error 4 in ld-2.17.so[7fe1277c8000+22000] 3[25997]: segfault at 8 ip 00007fe22763f7e8 sp 00007ffd7ecdebc0 error 4 in ld-2.17.so[7fe227634000+22000] 15[31282]: segfault at 8 ip 00007f0c200097e8 sp 00007fffa11fe650 error 4 in ld-2.17.so[7f0c1fffe000+22000] 1[8453]: segfault at 0 ip (null) sp 00007fff47673d08 error 14 in 1[400000+6000] 11[12385]: segfault at 8 ip 00007fa78738f7e8 sp 00007ffdfa51e730 error 4 in ld-2.17.so[7fa787384000+22000] 18[20639]: segfault at 66 ip 0000000000000066 sp 00007ffd61a36698 error 14 in 18[400000+6000] Lustre: 22628:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000402:0x22da:0x0] with magic=0xbd60bd0 Lustre: 22628:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message 18[26035]: segfault at 8 ip 00007fa9093aa7e8 sp 00007ffd2f6e64a0 error 4 in ld-2.17.so[7fa90939f000+22000] 16[28840]: segfault at 8 ip 00007f69e025b7e8 sp 00007ffc1be280e0 error 4 in ld-2.17.so[7f69e0250000+22000] 1[29453]: segfault at 0 ip (null) sp 00007ffcfb6343c8 error 14 in 1[400000+6000] LustreError: 6654:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802da8ec138: inode [0x200000403:0x239c:0x0] mdc close failed: rc = -13 LustreError: 6654:0:(file.c:247:ll_close_inode_openhandle()) Skipped 2 previous similar messages LustreError: 22909:0:(mdt_handler.c:746:mdt_pack_acl2body()) lustre-MDT0000: unable to read [0x200000403:0x23e9:0x0] ACL: rc = -2 6[13502]: segfault at 8 ip 00007f6a8abf37e8 sp 00007fff34e7e3c0 error 4 in ld-2.17.so[7f6a8abe8000+22000] 7[2707]: segfault at 14b6 ip 00000000000014b6 sp 00007ffe0c32bd88 error 14 in 7[400000+6000] Lustre: mdt00_000: service thread pid 12763 was inactive for 40.087 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Lustre: Skipped 2 previous similar messages Pid: 12763, comm: mdt00_000 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock+0x88/0x1c0 [mdt] [<0>] mdt_object_find_lock+0x54/0x170 [mdt] [<0>] mdt_reint_setxattr+0x133/0x10e0 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x74c/0xbc0 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 22852, comm: mdt00_009 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock_try+0xa0/0x250 [mdt] [<0>] mdt_getattr_name_lock+0x173b/0x2d10 [mdt] [<0>] mdt_intent_getattr+0x2cc/0x4e0 [mdt] [<0>] mdt_intent_opc.constprop.73+0x211/0xc60 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x397/0x980 [ptlrpc] [<0>] ldlm_handle_enqueue+0x547/0x18d0 [ptlrpc] [<0>] tgt_enqueue+0x68/0x240 [ptlrpc] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: mdt00_001: service thread pid 12768 was inactive for 40.127 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Lustre: Skipped 1 previous similar message Pid: 12768, comm: mdt00_001 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock_try+0xa0/0x250 [mdt] [<0>] mdt_getattr_name_lock+0x173b/0x2d10 [mdt] [<0>] mdt_intent_getattr+0x2cc/0x4e0 [mdt] [<0>] mdt_intent_opc.constprop.73+0x211/0xc60 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x397/0x980 [ptlrpc] [<0>] ldlm_handle_enqueue+0x547/0x18d0 [ptlrpc] [<0>] tgt_enqueue+0x68/0x240 [ptlrpc] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: lustre-OST0000-osc-ffff8800a4522e98: disconnect after 23s idle Lustre: lustre-OST0001-osc-ffff8802da8ec138: disconnect after 21s idle LustreError: 12423:0:(ldlm_lockd.c:252:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0000_UUID lock: ffff8802c19243c0/0x9f9c330d562142ff lrc: 3/0,0 mode: CR/CR res: [0x200000402:0x2d3c:0x0].0x0 bits 0xa/0x0 rrc: 8 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0x9f9c330d562142f1 expref: 865 pid: 15240 timeout: 1074 lvb_type: 0 Lustre: mdt00_000: service thread pid 12763 completed after 100.390s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_001: service thread pid 12768 completed after 99.404s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: lustre-MDT0000-mdc-ffff8800a4522e98: operation mds_close to node 0@lo failed: rc = -107 Lustre: lustre-MDT0000-mdc-ffff8800a4522e98: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: lustre-MDT0000-mdc-ffff8800a4522e98: This client was evicted by lustre-MDT0000; in progress operations using this service will fail. LustreError: 21675:0:(file.c:6081:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000402:0x2d3c:0x0] error: rc = -5 LustreError: 22022:0:(llite_lib.c:1995:ll_md_setattr()) md_setattr fails: rc = -108 LustreError: 14931:0:(ldlm_lockd.c:2591:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1741807532 with bad export cookie 11501123680763632517 LustreError: 22852:0:(ldlm_lockd.c:1492:ldlm_handle_enqueue()) ### lock on destroyed export ffff8800a4521bf8 ns: mdt-lustre-MDT0000_UUID lock: ffff88009d8a7840/0x9f9c330d56214410 lrc: 3/0,0 mode: PR/PR res: [0x200000402:0x2d3c:0x0].0x0 bits 0x1b/0x0 rrc: 3 type: IBT gid 0 flags: 0x50200400000020 nid: 0@lo remote: 0x9f9c330d562143f4 expref: 151 pid: 22852 timeout: 0 lvb_type: 0 Lustre: mdt00_009: service thread pid 22852 completed after 100.372s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: lustre-MDT0000-mdc-ffff8800a4522e98: Connection restored to (at 0@lo) Lustre: 17095:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000404:0x299:0x0] with magic=0xbd60bd0 Lustre: 17095:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message LustreError: 16498:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a4522e98: inode [0x200000404:0x80a:0x0] mdc close failed: rc = -13 LustreError: 16498:0:(file.c:247:ll_close_inode_openhandle()) Skipped 8 previous similar messages Lustre: 15240:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0x3c10:0x0] with magic=0xbd60bd0 Lustre: 15240:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 3 previous similar messages LustreError: 15457:0:(statahead.c:2442:start_statahead_thread()) lustre: unsupported statahead pattern 0X0. LustreError: 12769:0:(mdt_handler.c:746:mdt_pack_acl2body()) lustre-MDT0000: unable to read [0x200000404:0x184a:0x0] ACL: rc = -2 5[23781]: segfault at 8 ip 00007f12ee7467e8 sp 00007fffed690980 error 4 in ld-2.17.so[7f12ee73b000+22000] Lustre: 12768:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0x49b6:0x0] with magic=0xbd60bd0 Lustre: 12768:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 7 previous similar messages 9[9107]: segfault at 8 ip 00007f8a9bb8b7e8 sp 00007ffe777b61e0 error 4 in ld-2.17.so[7f8a9bb80000+22000] LustreError: 22628:0:(mdt_handler.c:746:mdt_pack_acl2body()) lustre-MDT0000: unable to read [0x200000404:0x2180:0x0] ACL: rc = -2 17[26992]: segfault at 8 ip 00007f5f38b477e8 sp 00007fffecf2fa30 error 4 in ld-2.17.so[7f5f38b3c000+22000] Lustre: mdt00_017: service thread pid 15291 was inactive for 40.054 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 15291, comm: mdt00_017 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock+0x88/0x1c0 [mdt] [<0>] mdt_object_find_lock+0x54/0x170 [mdt] [<0>] mdt_reint_setxattr+0x133/0x10e0 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x74c/0xbc0 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 12768, comm: mdt00_001 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock_try+0xa0/0x250 [mdt] [<0>] mdt_getattr_name_lock+0x173b/0x2d10 [mdt] [<0>] mdt_intent_getattr+0x2cc/0x4e0 [mdt] [<0>] mdt_intent_opc.constprop.73+0x211/0xc60 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x397/0x980 [ptlrpc] [<0>] ldlm_handle_enqueue+0x547/0x18d0 [ptlrpc] [<0>] tgt_enqueue+0x68/0x240 [ptlrpc] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: mdt00_013: service thread pid 22915 was inactive for 40.075 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Pid: 12769, comm: mdt00_002 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock+0x88/0x1c0 [mdt] [<0>] mdt_object_stripes_lock+0x126/0x660 [mdt] [<0>] mdt_reint_setattr+0x73b/0x15f0 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x74c/0xbc0 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: mdt00_006: service thread pid 22436 was inactive for 40.048 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 2 previous similar messages Lustre: mdt00_020: service thread pid 12637 was inactive for 40.008 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 3 previous similar messages LustreError: 12423:0:(ldlm_lockd.c:252:expired_lock_main()) ### lock callback timer expired after 101s: evicting client at 0@lo ns: mdt-lustre-MDT0000_UUID lock: ffff8800a2bae580/0x9f9c330d566f3471 lrc: 3/0,0 mode: PR/PR res: [0x200000403:0x4d29:0x0].0x0 bits 0x13/0x0 rrc: 19 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0x9f9c330d566f3455 expref: 922 pid: 12768 timeout: 1743 lvb_type: 0 Lustre: mdt00_013: service thread pid 22915 completed after 100.250s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_017: service thread pid 15291 completed after 100.503s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 12768:0:(ldlm_lockd.c:1492:ldlm_handle_enqueue()) ### lock on destroyed export ffff8800a2bb2548 ns: mdt-lustre-MDT0000_UUID lock: ffff8802b7b807c0/0x9f9c330d566f41ec lrc: 3/0,0 mode: PR/PR res: [0x200000403:0x4d29:0x0].0x0 bits 0x1b/0x0 rrc: 15 type: IBT gid 0 flags: 0x50200400000020 nid: 0@lo remote: 0x9f9c330d566f418a expref: 278 pid: 12768 timeout: 0 lvb_type: 0 LustreError: lustre-MDT0000-mdc-ffff8802da8ec138: operation mds_reint to node 0@lo failed: rc = -107 Lustre: lustre-MDT0000-mdc-ffff8802da8ec138: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 12419:0:(ldlm_lockd.c:2591:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1741808201 with bad export cookie 11501123680763849146 LustreError: lustre-MDT0000-mdc-ffff8802da8ec138: This client was evicted by lustre-MDT0000; in progress operations using this service will fail. LustreError: 31187:0:(llite_lib.c:1995:ll_md_setattr()) md_setattr fails: rc = -5 LustreError: 31143:0:(file.c:6081:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000403:0x4d29:0x0] error: rc = -5 LustreError: 31143:0:(file.c:6081:ll_inode_revalidate_fini()) Skipped 48 previous similar messages LustreError: 31187:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802da8ec138: inode [0x200000403:0x4d29:0x0] mdc close failed: rc = -108 LustreError: 31187:0:(file.c:247:ll_close_inode_openhandle()) Skipped 5 previous similar messages Lustre: mdt00_001: service thread pid 12768 completed after 100.354s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_002: service thread pid 12769 completed after 100.229s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_011: service thread pid 22891 completed after 100.147s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_000: service thread pid 12763 completed after 99.792s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_006: service thread pid 22436 completed after 99.267s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_021: service thread pid 1296 completed after 99.197s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_019: service thread pid 17095 completed after 99.300s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_020: service thread pid 12637 completed after 98.206s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_018: service thread pid 15313 completed after 98.967s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_016: service thread pid 15240 completed after 98.258s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: lustre-MDT0000-mdc-ffff8802da8ec138: Connection restored to (at 0@lo) Lustre: 22879:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000404:0x26f9:0x0] with magic=0xbd60bd0 Lustre: 22879:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message Lustre: ll_ost_out00_00: service thread pid 12527 was inactive for 40.080 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. LustreError: 12423:0:(ldlm_lockd.c:252:expired_lock_main()) ### lock callback timer expired after 101s: evicting client at 0@lo ns: mdt-lustre-MDT0000_UUID lock: ffff8802c2950040/0x9f9c330d5684ba58 lrc: 3/0,0 mode: PR/PR res: [0x200000404:0x2db8:0x0].0x0 bits 0x1b/0x0 rrc: 5 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0x9f9c330d5684b9be expref: 845 pid: 12527 timeout: 1989 lvb_type: 0 Lustre: ll_ost_out00_00: service thread pid 14704 completed after 100.392s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: lustre-MDT0000-mdc-ffff8800a4522e98: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 1 previous similar message Lustre: lustre-MDT0000-mdc-ffff8800a4522e98: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: lustre-MDT0000-mdc-ffff8800a4522e98: This client was evicted by lustre-MDT0000; in progress operations using this service will fail. Lustre: ll_ost_out00_00: service thread pid 12527 completed after 100.410s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 8247:0:(file.c:6081:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000007:0x1:0x0] error: rc = -108 LustreError: 8247:0:(file.c:6081:ll_inode_revalidate_fini()) Skipped 5 previous similar messages Lustre: lustre-MDT0000-mdc-ffff8800a4522e98: Connection restored to (at 0@lo) 14[12417]: segfault at 0 ip (null) sp 00007ffe6d37f5f8 error 14 in 14[400000+6000] 19[30394]: segfault at 8 ip 00007fafcbd037e8 sp 00007fffed093390 error 4 in ld-2.17.so[7fafcbcf8000+22000] Lustre: 22850:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000405:0xd82:0x0] with magic=0xbd60bd0 Lustre: 22850:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 5 previous similar messages LustreError: 12423:0:(ldlm_lockd.c:252:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0000_UUID lock: ffff8802b4e36940/0x9f9c330d569ae6cc lrc: 3/0,0 mode: PR/PR res: [0x200000405:0x1394:0x0].0x0 bits 0x13/0x0 rrc: 4 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0x9f9c330d569ae69b expref: 601 pid: 15240 timeout: 2223 lvb_type: 0 LustreError: lustre-MDT0000-mdc-ffff8802da8ec138: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-MDT0000-mdc-ffff8802da8ec138: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 12420:0:(ldlm_lockd.c:2591:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1741808681 with bad export cookie 11501123680775142414 LustreError: lustre-MDT0000-mdc-ffff8802da8ec138: This client was evicted by lustre-MDT0000; in progress operations using this service will fail. LustreError: 19354:0:(file.c:6081:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000007:0x1:0x0] error: rc = -108 LustreError: 19354:0:(file.c:6081:ll_inode_revalidate_fini()) Skipped 9 previous similar messages LustreError: 24330:0:(ldlm_resource.c:982:ldlm_resource_complain()) lustre-MDT0000-mdc-ffff8802da8ec138: namespace resource [0x200000007:0x1:0x0].0x0 (ffff8802d73f0f40) refcount nonzero (2) after lock cleanup; forcing cleanup. Lustre: lustre-MDT0000-mdc-ffff8802da8ec138: Connection restored to (at 0@lo) LustreError: 30669:0:(statahead.c:2442:start_statahead_thread()) lustre: unsupported statahead pattern 0X0. 0[6472]: segfault at 8 ip 00007f003ca337e8 sp 00007ffeb7eace50 error 4 in ld-2.17.so[7f003ca28000+22000] 16[9176]: segfault at 8 ip 00007f2aefc117e8 sp 00007ffd75cbc280 error 4 in ld-2.17.so[7f2aefc06000+22000] 14[9613]: segfault at 0 ip 0000000000403e5f sp 00007ffda955a4b0 error 6 in 14[400000+6000] 11[25133]: segfault at a6 ip 00000000000000a6 sp 00007ffe6ab2bf08 error 14 in 11[400000+6000] 12[15503]: segfault at 8 ip 00007f26ab6057e8 sp 00007fff641b4c90 error 4 in ld-2.17.so[7f26ab5fa000+22000] LustreError: 15850:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802da8ec138: inode [0x200000407:0xc14:0x0] mdc close failed: rc = -13 LustreError: 15850:0:(file.c:247:ll_close_inode_openhandle()) Skipped 28 previous similar messages LustreError: 22879:0:(mdt_handler.c:746:mdt_pack_acl2body()) lustre-MDT0000: unable to read [0x200000407:0xc90:0x0] ACL: rc = -2 11[21067]: segfault at 8 ip 00007f6c9616e7e8 sp 00007ffeed4b12d0 error 4 in ld-2.17.so[7f6c96163000+22000] 4[25647]: segfault at 8 ip 00007f9a743167e8 sp 00007fff88dfb9f0 error 4 in ld-2.17.so[7f9a7430b000+22000] 0[29197]: segfault at 8 ip 00007facb0d4a7e8 sp 00007ffda0761130 error 4 in ld-2.17.so[7facb0d3f000+22000] LustreError: 22436:0:(mdt_handler.c:746:mdt_pack_acl2body()) lustre-MDT0000: unable to read [0x200000407:0xf87:0x0] ACL: rc = -2 2[27952]: segfault at 8 ip 00007f89217a07e8 sp 00007ffc04148660 error 4 in ld-2.17.so[7f8921795000+22000] 2[9652]: segfault at 8 ip 00007f1b922db7e8 sp 00007ffe6ece1a40 error 4 2[9639]: segfault at 8 ip 00007f20b9b407e8 sp 00007ffe2fc5b180 error 4 in ld-2.17.so[7f20b9b35000+22000] in ld-2.17.so[7f1b922d0000+22000] 14[28306]: segfault at 0 ip (null) sp 00007ffcdeab68a8 error 14 in 14[400000+6000] 17[19374]: segfault at 0 ip (null) sp 00007fff64536e58 error 14 in 17[400000+6000] 19[26271]: segfault at 8 ip 00007fbe3fd577e8 sp 00007fff95ca28c0 error 4 in ld-2.17.so[7fbe3fd4c000+22000] 14[13382]: segfault at 0 ip (null) sp 00007ffd2bb163e8 error 14 in 14[400000+6000] 18[15182]: segfault at 8 ip 00007f08188e37e8 sp 00007fffd27cfcd0 error 4 in ld-2.17.so[7f08188d8000+22000] 13[31034]: segfault at 0 ip 0000000000403e5f sp 00007fff41f19700 error 6 in 13[400000+6000] | Link to test |
racer test 1: racer on clients: oleg146-client.virtnet DURATION=300 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffff813f40ac>] _atomic_dec_and_lock+0xc/0x70 PGD 902ca067 PUD 9c42e067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) osp(OE) ofd(OE) lod(OE) mdt(OE) mdd(OE) mgs(OE) lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) crc_t10dif crct10dif_generic rpcsec_gss_krb5 sb_edac edac_core iosf_mbi kvm_intel kvm irqbypass crc32_pclmul ghash_clmulni_intel aesni_intel lrw gf128mul i2c_piix4 glue_helper squashfs ablk_helper cryptd i2c_core pcspkr binfmt_misc ip_tables ext4 mbcache jbd2 ata_generic pata_acpi crct10dif_pclmul ata_piix crct10dif_common crc32c_intel serio_raw libata CPU: 1 PID: 3235 Comm: ll_sa_30445 Kdump: loaded Tainted: G OE ------------ 3.10.0-7.9-debug #1 Hardware name: QEMU Standard PC (i440FX + PIIX, 1996), BIOS 1.16.3-2.fc40 04/01/2014 task: ffff88009c46a220 ti: ffff88012b1b0000 task.ti: ffff88012b1b0000 RIP: 0010:[<ffffffff813f40ac>] [<ffffffff813f40ac>] _atomic_dec_and_lock+0xc/0x70 RSP: 0018:ffff88012b1b3d80 EFLAGS: 00010282 RAX: ffff88008d78e0c8 RBX: 0000000000000008 RCX: ffffea0002932880 RDX: 0000000000000004 RSI: ffff88008d78e3f8 RDI: 0000000000000008 RBP: ffff88012b1b3d90 R08: ffff88009037d000 R09: 000000018010000f R10: 0000000000000001 R11: ffff88009037d000 R12: 0000000000000000 R13: ffff8800a4709110 R14: ffff88008d78e3f8 R15: ffff88009037d000 FS: 0000000000000000(0000) GS:ffff88013e280000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 CR2: 0000000000000008 CR3: 00000000a1ef8000 CR4: 0000000000160fe0 Call Trace: [<ffffffffa10d2501>] ll_statahead_thread+0xde1/0x2290 [lustre] [<ffffffff810df6e2>] ? pick_next_task_fair+0x702/0x8a0 [<ffffffff81410a79>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e8dce>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffff810c834d>] ? finish_task_switch+0x5d/0x1b0 [<ffffffffa10d1720>] ? ll_statahead_interpret+0x4a0/0x4a0 [lustre] [<ffffffff810ba114>] kthread+0xe4/0xf0 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817f3e5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 | LustreError: 12014:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800b599b000: inode [0x240000403:0x4:0x0] mdc close failed: rc = -116 LustreError: 12266:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800b599b000: inode [0x200000403:0xa:0x0] mdc close failed: rc = -116 LustreError: 12266:0:(file.c:247:ll_close_inode_openhandle()) Skipped 2 previous similar messages 17[13114]: segfault at 8 ip 00007fe5c98f97e8 sp 00007ffc39042230 error 4 in ld-2.17.so[7fe5c98ee000+22000] LustreError: 13340:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800b599b000: inode [0x240000403:0x14:0x0] mdc close failed: rc = -116 Lustre: dir [0x240000402:0xb9:0x0] stripe 1 readdir failed: -2, directory is partially accessed! LustreError: 15507:0:(lcommon_cl.c:177:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000403:0xdf:0x0]: rc = -5 LustreError: 15507:0:(llite_lib.c:3712:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 15809:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800b599b000: inode [0x200000403:0x9e:0x0] mdc close failed: rc = -116 LustreError: 15809:0:(file.c:247:ll_close_inode_openhandle()) Skipped 1 previous similar message LustreError: 15507:0:(lcommon_cl.c:177:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000403:0xc5:0x0]: rc = -5 LustreError: 15507:0:(llite_lib.c:3712:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 0[16045]: segfault at 8 ip 00007f0a500597e8 sp 00007ffc18b60790 error 4 in ld-2.17.so[7f0a5004e000+22000] Lustre: dir [0x240000402:0x11b:0x0] stripe 2 readdir failed: -2, directory is partially accessed! Lustre: Skipped 6 previous similar messages LustreError: 15070:0:(llite_lib.c:1846:ll_update_lsm_md()) lustre: [0x240000403:0x1c4:0x0] dir layout mismatch: LustreError: 15070:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=2 count=2 index=1 hash=crush:0x2000003 max_inherit=0 max_inherit_rr=0 version=2 migrate_offset=0 migrate_hash=invalid:0 pool= LustreError: 15070:0:(lustre_lmv.h:167:lmv_stripe_object_dump()) stripe[0] [0x240000400:0x12:0x0] LustreError: 15070:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=1 count=3 index=1 hash=crush:0x82000003 max_inherit=0 max_inherit_rr=0 version=1 migrate_offset=2 migrate_hash=fnv_1a_64:2 pool= LustreError: 16476:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800b599b000: inode [0x200000402:0x157:0x0] mdc close failed: rc = -116 LustreError: 16476:0:(file.c:247:ll_close_inode_openhandle()) Skipped 5 previous similar messages LustreError: 17120:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff88012aae3800: cannot apply new layout on [0x200000403:0xdf:0x0] : rc = -5 LustreError: 17120:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x200000403:0xdf:0x0] error -5. LustreError: 17554:0:(lcommon_cl.c:177:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000403:0x188:0x0]: rc = -5 LustreError: 17554:0:(llite_lib.c:3712:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 12[18453]: segfault at 8 ip 00007f52120537e8 sp 00007ffde0d4a6d0 error 4 in ld-2.17.so[7f5212048000+22000] LustreError: 10651:0:(lcommon_cl.c:177:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000402:0x1c0:0x0]: rc = -5 LustreError: 10651:0:(lcommon_cl.c:177:cl_file_inode_init()) Skipped 1 previous similar message LustreError: 10651:0:(llite_lib.c:3712:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 10651:0:(llite_lib.c:3712:ll_prep_inode()) Skipped 1 previous similar message LustreError: 19624:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff88012aae3800: cannot apply new layout on [0x200000403:0xdf:0x0] : rc = -5 LustreError: 19624:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x200000403:0xdf:0x0] error -5. LustreError: 20955:0:(lcommon_cl.c:177:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000403:0x207:0x0]: rc = -5 LustreError: 20955:0:(lcommon_cl.c:177:cl_file_inode_init()) Skipped 5 previous similar messages LustreError: 20955:0:(llite_lib.c:3712:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 20955:0:(llite_lib.c:3712:ll_prep_inode()) Skipped 5 previous similar messages Lustre: dir [0x200000402:0x251:0x0] stripe 2 readdir failed: -2, directory is partially accessed! Lustre: Skipped 4 previous similar messages LustreError: 19109:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff88012aae3800: inode [0x240000402:0xe2:0x0] mdc close failed: rc = -2 LustreError: 19109:0:(file.c:247:ll_close_inode_openhandle()) Skipped 9 previous similar messages Lustre: dir [0x240000403:0x3d3:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 1 previous similar message Lustre: dir [0x240000403:0x376:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 2 previous similar messages LustreError: 22021:0:(lcommon_cl.c:177:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000402:0x3c8:0x0]: rc = -5 LustreError: 22021:0:(lcommon_cl.c:177:cl_file_inode_init()) Skipped 18 previous similar messages LustreError: 22021:0:(llite_lib.c:3712:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 22021:0:(llite_lib.c:3712:ll_prep_inode()) Skipped 18 previous similar messages LustreError: 486:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 9 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 LustreError: 25859:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff88012aae3800: cannot apply new layout on [0x240000403:0x207:0x0] : rc = -5 LustreError: 25859:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x240000403:0x207:0x0] error -5. LustreError: 29284:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff88012aae3800: inode [0x240000402:0x233:0x0] mdc close failed: rc = -116 LustreError: 29284:0:(file.c:247:ll_close_inode_openhandle()) Skipped 11 previous similar messages Lustre: dir [0x200000402:0x620:0x0] stripe 2 readdir failed: -2, directory is partially accessed! Lustre: Skipped 7 previous similar messages 15[30382]: segfault at 8 ip 00007f8458b147e8 sp 00007ffd57f314d0 error 4 in ld-2.17.so[7f8458b09000+22000] LustreError: 32169:0:(lcommon_cl.c:177:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000403:0x647:0x0]: rc = -5 LustreError: 32169:0:(lcommon_cl.c:177:cl_file_inode_init()) Skipped 61 previous similar messages LustreError: 32169:0:(llite_lib.c:3712:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 32169:0:(llite_lib.c:3712:ll_prep_inode()) Skipped 61 previous similar messages 16[32758]: segfault at 8 ip 00007f95fa66a7e8 sp 00007ffee7e89480 error 4 in ld-2.17.so[7f95fa65f000+22000] LustreError: 30058:0:(llite_lib.c:1846:ll_update_lsm_md()) lustre: [0x240000402:0x7a9:0x0] dir layout mismatch: LustreError: 30058:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=2 count=2 index=1 hash=crush:0x2000003 max_inherit=0 max_inherit_rr=0 version=2 migrate_offset=0 migrate_hash=invalid:0 pool= LustreError: 30058:0:(lustre_lmv.h:167:lmv_stripe_object_dump()) stripe[0] [0x240000400:0x31:0x0] LustreError: 30058:0:(lustre_lmv.h:167:lmv_stripe_object_dump()) Skipped 4 previous similar messages LustreError: 30058:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=1 count=3 index=1 hash=crush:0x82000003 max_inherit=0 max_inherit_rr=0 version=1 migrate_offset=2 migrate_hash=fnv_1a_64:2 pool= LustreError: 29592:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=7 count=2 index=1 hash=crush:0x2000003 max_inherit=0 max_inherit_rr=0 version=2 migrate_offset=0 migrate_hash=invalid:0 pool= LustreError: 29592:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=1 count=3 index=1 hash=crush:0x82000003 max_inherit=0 max_inherit_rr=0 version=1 migrate_offset=2 migrate_hash=fnv_1a_64:2 pool= LustreError: 32680:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff88012aae3800: cannot apply new layout on [0x200000403:0x6bd:0x0] : rc = -5 LustreError: 32680:0:(lov_object.c:1341:lov_layout_change()) Skipped 3 previous similar messages LustreError: 32680:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x200000403:0x6bd:0x0] error -5. LustreError: 29651:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=3 count=2 index=1 hash=crush:0x2000003 max_inherit=0 max_inherit_rr=0 version=2 migrate_offset=0 migrate_hash=invalid:0 pool= LustreError: 29651:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=1 count=3 index=1 hash=crush:0x82000003 max_inherit=0 max_inherit_rr=0 version=1 migrate_offset=2 migrate_hash=fnv_1a_64:2 pool= LustreError: 29598:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=4 count=2 index=1 hash=crush:0x2000003 max_inherit=0 max_inherit_rr=0 version=2 migrate_offset=0 migrate_hash=invalid:0 pool= LustreError: 29598:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=1 count=3 index=1 hash=crush:0x82000003 max_inherit=0 max_inherit_rr=0 version=1 migrate_offset=2 migrate_hash=fnv_1a_64:2 pool= 4[3683]: segfault at 8 ip 00007fa84c1d87e8 sp 00007fff71fb33c0 error 4 in ld-2.17.so[7fa84c1cd000+22000] LustreError: 3138:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff88012aae3800: cannot apply new layout on [0x240000403:0x829:0x0] : rc = -5 LustreError: 3138:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x240000403:0x829:0x0] error -5. Lustre: dir [0x240000402:0x992:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 8 previous similar messages 10[6133]: segfault at 0 ip 0000000000403e5f sp 00007ffd77a6d820 error 6 in 10[400000+6000] LustreError: 51:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 13 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 LustreError: 51:0:(statahead.c:825:ll_statahead_interpret_work()) Skipped 2 previous similar messages LustreError: 51:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 13 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 LustreError: 12356:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff88012aae3800: inode [0x200000403:0xd00:0x0] mdc close failed: rc = -116 LustreError: 12356:0:(file.c:247:ll_close_inode_openhandle()) Skipped 37 previous similar messages LustreError: 12632:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff88012aae3800: cannot apply new layout on [0x240000403:0x829:0x0] : rc = -5 LustreError: 12632:0:(lov_object.c:1341:lov_layout_change()) Skipped 3 previous similar messages LustreError: 15158:0:(lcommon_cl.c:177:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000403:0xb71:0x0]: rc = -5 LustreError: 15158:0:(lcommon_cl.c:177:cl_file_inode_init()) Skipped 42 previous similar messages LustreError: 15158:0:(llite_lib.c:3712:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 15158:0:(llite_lib.c:3712:ll_prep_inode()) Skipped 42 previous similar messages 11[15405]: segfault at 8 ip 00007f98319127e8 sp 00007fffbd199f10 error 4 in ld-2.17.so[7f9831907000+22000] 16[19513]: segfault at 406000 ip 0000000000406000 sp 00007ffebd235c28 error 14 in 16[606000+1000] 14[19860]: segfault at 8 ip 00007f5ffe5a67e8 sp 00007ffdbfd5a000 error 4 in ld-2.17.so[7f5ffe59b000+22000] LustreError: 21297:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff88012aae3800: cannot apply new layout on [0x200000403:0xf1f:0x0] : rc = -5 LustreError: 21297:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x200000403:0xf1f:0x0] error -5. LustreError: 21297:0:(vvp_io.c:1903:vvp_io_init()) Skipped 1 previous similar message 19[21759]: segfault at 8 ip 00007f0a16f5a7e8 sp 00007ffc3e6182c0 error 4 in ld-2.17.so[7f0a16f4f000+22000] Lustre: dir [0x200000403:0x1305:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 8 previous similar messages 15[24338]: segfault at 8 ip 00007f1ff68017e8 sp 00007fff815550d0 error 4 in ld-2.17.so[7f1ff67f6000+22000] 3[27538]: segfault at 8 ip 00007f786970e7e8 sp 00007ffc44e1f9e0 error 4 in ld-2.17.so[7f7869703000+22000] 16[28168]: segfault at 0 ip 0000000000403cf0 sp 00007ffd575961f8 error 6 in 16[400000+6000] LustreError: 29252:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x240000402:0x10da:0x0] error -5. LustreError: 29252:0:(vvp_io.c:1903:vvp_io_init()) Skipped 1 previous similar message LustreError: 589:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 12 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 LustreError: 589:0:(statahead.c:825:ll_statahead_interpret_work()) Skipped 3 previous similar messages LustreError: lustre-MDT0001-mdc-ffff8800b599b000: operation ldlm_enqueue to node 192.168.201.146@tcp failed: rc = -107 Lustre: lustre-MDT0001-mdc-ffff88012aae3800: Connection to lustre-MDT0001 (at 192.168.201.146@tcp) was lost; in progress operations using this service will wait for recovery to complete LustreError: lustre-MDT0001-mdc-ffff88012aae3800: This client was evicted by lustre-MDT0001; in progress operations using this service will fail. LustreError: lustre-MDT0001-mdc-ffff8800b599b000: This client was evicted by lustre-MDT0001; in progress operations using this service will fail. LustreError: 1752:0:(file.c:6007:ll_inode_revalidate_fini()) lustre: revalidate FID [0x240000402:0x13c4:0x0] error: rc = -5 LustreError: 3069:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff88012aae3800: inode [0x240000402:0x1592:0x0] mdc close failed: rc = -108 LustreError: 3069:0:(file.c:247:ll_close_inode_openhandle()) Skipped 41 previous similar messages LustreError: Skipped 13 previous similar messages LustreError: 3474:0:(llite_lib.c:1997:ll_md_setattr()) md_setattr fails: rc = -108 LustreError: lustre-MDT0000-mdc-ffff88012aae3800: This client was evicted by lustre-MDT0000; in progress operations using this service will fail. LustreError: 28836:0:(mdc_request.c:1457:mdc_read_page()) lustre-MDT0001-mdc-ffff8800b599b000: [0x240000402:0xdf5:0x0] lock enqueue fails: rc = -108 Lustre: dir [0x200000402:0x1746:0x0] stripe 1 readdir failed: -108, directory is partially accessed! Lustre: Skipped 7 previous similar messages LustreError: 5006:0:(llite_nfs.c:426:ll_dir_get_parent_fid()) lustre: failure inode [0x200000402:0x1721:0x0] get parent: rc = -108 LustreError: 5606:0:(file.c:6007:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000007:0x1:0x0] error: rc = -108 LustreError: 5606:0:(file.c:6007:ll_inode_revalidate_fini()) Skipped 494 previous similar messages LustreError: 3078:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x200000402:0x1435:0x0] error -108. Lustre: lustre-MDT0001-mdc-ffff88012aae3800: Connection restored to (at 192.168.201.146@tcp) LustreError: 6556:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8800b599b000: cannot apply new layout on [0x240000402:0x10da:0x0] : rc = -5 LustreError: 6556:0:(lov_object.c:1341:lov_layout_change()) Skipped 9 previous similar messages LustreError: 6556:0:(lcommon_cl.c:177:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000402:0x10da:0x0]: rc = -5 LustreError: 6556:0:(lcommon_cl.c:177:cl_file_inode_init()) Skipped 92 previous similar messages LustreError: 6556:0:(llite_lib.c:3712:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 6556:0:(llite_lib.c:3712:ll_prep_inode()) Skipped 92 previous similar messages LustreError: 6763:0:(llite_lib.c:1846:ll_update_lsm_md()) lustre: [0x240000405:0x49:0x0] dir layout mismatch: LustreError: 6763:0:(llite_lib.c:1846:ll_update_lsm_md()) Skipped 3 previous similar messages LustreError: 6763:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=2 count=2 index=1 hash=crush:0x2000003 max_inherit=0 max_inherit_rr=0 version=2 migrate_offset=0 migrate_hash=invalid:0 pool= LustreError: 6763:0:(lustre_lmv.h:167:lmv_stripe_object_dump()) stripe[0] [0x240000400:0x82:0x0] LustreError: 6763:0:(lustre_lmv.h:167:lmv_stripe_object_dump()) Skipped 19 previous similar messages LustreError: 6763:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=1 count=3 index=1 hash=crush:0x82000003 max_inherit=0 max_inherit_rr=0 version=1 migrate_offset=2 migrate_hash=fnv_1a_64:2 pool= 14[8575]: segfault at 8 ip 00007fab4ed457e8 sp 00007fff3a78c6a0 error 4 14[8627]: segfault at 8 ip 00007f7d3c9fb7e8 sp 00007ffc59a96940 error 4 in ld-2.17.so[7f7d3c9f0000+22000] in ld-2.17.so[7fab4ed3a000+22000] LustreError: 6916:0:(llite_lib.c:1846:ll_update_lsm_md()) lustre: [0x200000403:0x1b18:0x0] dir layout mismatch: LustreError: 6916:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=2 count=2 index=0 hash=crush:0x2000003 max_inherit=0 max_inherit_rr=0 version=2 migrate_offset=0 migrate_hash=invalid:0 pool= LustreError: 6916:0:(lustre_lmv.h:167:lmv_stripe_object_dump()) stripe[0] [0x200000400:0x7a:0x0] LustreError: 6916:0:(lustre_lmv.h:167:lmv_stripe_object_dump()) Skipped 4 previous similar messages LustreError: 6916:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=1 count=3 index=0 hash=crush:0x82000003 max_inherit=0 max_inherit_rr=0 version=1 migrate_offset=2 migrate_hash=fnv_1a_64:2 pool= LustreError: 6923:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=2 count=2 index=0 hash=crush:0x2000003 max_inherit=0 max_inherit_rr=0 version=2 migrate_offset=0 migrate_hash=invalid:0 pool= LustreError: 6923:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=1 count=3 index=0 hash=crush:0x82000003 max_inherit=0 max_inherit_rr=0 version=1 migrate_offset=2 migrate_hash=fnv_1a_64:2 pool= LustreError: 6670:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=1 count=2 index=0 hash=crush:0x2000003 max_inherit=0 max_inherit_rr=0 version=2 migrate_offset=0 migrate_hash=invalid:0 pool= LustreError: 6670:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=1 count=3 index=0 hash=crush:0x82000003 max_inherit=0 max_inherit_rr=0 version=1 migrate_offset=2 migrate_hash=fnv_1a_64:2 pool= 19[12066]: segfault at 1c ip 00000000004017b0 sp 00007ffd6780cfb0 error 6 in 19[400000+6000] LustreError: 12668:0:(llite_lib.c:1846:ll_update_lsm_md()) lustre: [0x200000404:0x15d:0x0] dir layout mismatch: LustreError: 12668:0:(llite_lib.c:1846:ll_update_lsm_md()) Skipped 2 previous similar messages LustreError: 12668:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=2 count=2 index=0 hash=crush:0x2000003 max_inherit=0 max_inherit_rr=0 version=2 migrate_offset=0 migrate_hash=invalid:0 pool= LustreError: 12668:0:(lustre_lmv.h:167:lmv_stripe_object_dump()) stripe[0] [0x200000400:0x7d:0x0] LustreError: 12668:0:(lustre_lmv.h:167:lmv_stripe_object_dump()) Skipped 14 previous similar messages LustreError: 12668:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=1 count=3 index=0 hash=crush:0x82000003 max_inherit=0 max_inherit_rr=0 version=1 migrate_offset=2 migrate_hash=fnv_1a_64:2 pool= 2[15330]: segfault at 8 ip 00007f00d9a157e8 sp 00007ffeea4eaea0 error 4 in ld-2.17.so[7f00d9a0a000+22000] 7[462]: segfault at 8 ip 00007f3002a517e8 sp 00007fff46ced7b0 error 4 in ld-2.17.so[7f3002a46000+22000] 2[697]: segfault at 8 ip 00007ffb6ab617e8 sp 00007ffd2f3bb310 error 4 in ld-2.17.so[7ffb6ab56000+22000] | Link to test |
racer test 1: racer on clients: centos-20.localnet DURATION=2700 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 PGD 2ac269067 PUD 2b42dc067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) lustre(OE) osp(OE) ofd(OE) lod(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) jbd2 mbcache lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) crc32_generic lnet(OE) dm_flakey dm_mod libcfs(OE) crc_t10dif crct10dif_generic crct10dif_common virtio_console i2c_piix4 virtio_balloon pcspkr ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm drm_panel_orientation_quirks ata_piix i2c_core floppy serio_raw libata virtio_blk CPU: 9 PID: 5769 Comm: ll_sa_5717 Kdump: loaded Tainted: P OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.16.0-3.module_el8.7.0+1218+f626c2ff 04/01/2014 task: ffff8802accca4f0 ti: ffff8802d374c000 task.ti: ffff8802d374c000 RIP: 0010:[<ffffffff813f502c>] [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 RSP: 0018:ffff8802d374fd78 EFLAGS: 00010296 RAX: ffff88008675d8c8 RBX: 0000000000000008 RCX: 000000010026001a RDX: 0000000000000026 RSI: ffff88008675dbf8 RDI: 0000000000000008 RBP: ffff8802d374fd88 R08: ffff88028350cc38 R09: 0000000000000000 R10: ffff88028350e6b8 R11: ffff88028350cc38 R12: 0000000000000000 R13: ffff8802d7838010 R14: ffff88008675dbf8 R15: ffff88028350cc38 FS: 0000000000000000(0000) GS:ffff880331c40000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 00000002de740000 CR4: 00000000000007e0 Call Trace: [<ffffffffa11151d1>] ll_statahead_thread+0xe01/0x22b0 [lustre] [<ffffffff817e320e>] ? _raw_spin_lock+0x1e/0x20 [<ffffffff810dbb51>] ? put_prev_entity+0x31/0x400 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e329e>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffff810c834d>] ? finish_task_switch+0x5d/0x1b0 [<ffffffffa11143d0>] ? ll_statahead_handle.constprop.29+0x1c0/0x1c0 [lustre] [<ffffffff810ba114>] kthread+0xe4/0xf0 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 | Lustre: 11356:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 510 < left 936, rollback = 2 Lustre: 11356:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 4/16/3, destroy: 1/4/0 Lustre: 11356:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 5/5/0, xattr_set: 14/936/0 Lustre: 11356:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 7/122/0, punch: 0/0/0, quota 1/3/0 Lustre: 11356:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 18/332/2, delete: 3/6/0 Lustre: 11356:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 9/9/0, ref_del: 3/3/0 Lustre: 17175:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 512 < left 983, rollback = 2 Lustre: 17175:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 8 previous similar messages Lustre: 17175:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 2/8/1, destroy: 1/4/0 Lustre: 17175:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 8 previous similar messages Lustre: 17175:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 2/2/0, xattr_set: 12/983/0 Lustre: 17175:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 8 previous similar messages Lustre: 17175:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 1/10/0, punch: 0/0/0, quota 1/3/0 Lustre: 17175:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 8 previous similar messages Lustre: 17175:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 12/194/2, delete: 3/6/0 Lustre: 17175:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 8 previous similar messages Lustre: 17175:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 7/7/0, ref_del: 3/3/0 Lustre: 17175:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 8 previous similar messages cp (18599) used greatest stack depth: 10192 bytes left Lustre: dir [0x200000404:0x1f:0x0] stripe 3 readdir failed: -2, directory is partially accessed! Lustre: 11356:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 505 < left 894, rollback = 2 Lustre: 11356:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 11 previous similar messages Lustre: 11356:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 3/12/6, destroy: 1/4/0 Lustre: 11356:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 11 previous similar messages Lustre: 11356:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 2/2/0, xattr_set: 11/894/0 Lustre: 11356:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 11 previous similar messages Lustre: 11356:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 5/55/0, punch: 0/0/0, quota 7/105/0 Lustre: 11356:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 11 previous similar messages Lustre: 11356:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 16/359/4, delete: 2/5/0 Lustre: 11356:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 11 previous similar messages Lustre: 11356:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 9/9/0, ref_del: 1/1/0 Lustre: 11356:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 11 previous similar messages LustreError: 11336:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0001: failed to get lu_attr of [0x240000404:0x2:0x0]: rc = -2 LustreError: 17282:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e6f3dd28: inode [0x240000404:0x2:0x0] mdc close failed: rc = -2 LustreError: 18659:0:(mdt_reint.c:2513:mdt_reint_migrate()) lustre-MDT0000: migrate [0x240000404:0x3e:0x0]/20 failed: rc = -2 Lustre: dir [0x240000404:0x3e:0x0] stripe 3 readdir failed: -2, directory is partially accessed! Lustre: 12717:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 516 < left 618, rollback = 7 Lustre: 12717:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 12717:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 36 previous similar messages Lustre: 12717:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 12717:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 36 previous similar messages Lustre: 12717:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 12717:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 36 previous similar messages Lustre: 12717:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 12717:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 36 previous similar messages Lustre: 12717:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 12717:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 36 previous similar messages Lustre: 22676:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 516 < left 618, rollback = 7 Lustre: 22676:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 5 previous similar messages Lustre: 17339:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 516 < left 618, rollback = 7 Lustre: 17339:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 3 previous similar messages Lustre: 23085:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 516 < left 618, rollback = 7 Lustre: 23085:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 5 previous similar messages Lustre: 23085:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 23085:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 15 previous similar messages Lustre: 23085:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 23085:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 15 previous similar messages Lustre: 23085:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 23085:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 15 previous similar messages Lustre: 23085:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 23085:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 15 previous similar messages Lustre: 23085:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 23085:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 15 previous similar messages 4[26118]: segfault at 8 ip 00007fbd1e2157e8 sp 00007fff52d7e530 error 4 in ld-2.17.so[7fbd1e20a000+22000] 1[26336]: segfault at 8 ip 00007f76c2ba87e8 sp 00007ffd878d6440 error 4 in ld-2.17.so[7f76c2b9d000+22000] 0[26620]: segfault at 0 ip 0000000000401d60 sp 00007ffeb30f9a08 error 6 in 0[400000+6000] 14[31286]: segfault at 8 ip 00007f7f1647d7e8 sp 00007fff3a2a31b0 error 4 in ld-2.17.so[7f7f16472000+22000] 3[2362]: segfault at 8 ip 00007f12c41b87e8 sp 00007ffc7a8710a0 error 4 in ld-2.17.so[7f12c41ad000+22000] Lustre: 22676:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0002: opcode 7: before 516 < left 618, rollback = 7 Lustre: 22676:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 1 previous similar message Lustre: 22676:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 22676:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 22676:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 22676:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 22676:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 22676:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 22676:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 22676:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 22676:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 22676:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 1 previous similar message 1[5017]: segfault at 8 ip 00007fe21fae07e8 sp 00007ffcbeff8790 error 4 in ld-2.17.so[7fe21fad5000+22000] 12[10364]: segfault at 8 ip 00007f16f1afd7e8 sp 00007ffe0e375470 error 4 in ld-2.17.so[7f16f1af2000+22000] 17[10297]: segfault at 8 ip 00007fe1e05617e8 sp 00007ffdf82e5b60 error 4 in ld-2.17.so[7fe1e0556000+22000] 8[15270]: segfault at 8 ip 00007f8c1e0f67e8 sp 00007ffc07d9ae30 error 4 in ld-2.17.so[7f8c1e0eb000+22000] LustreError: 11355:0:(mdt_reint.c:2513:mdt_reint_migrate()) lustre-MDT0000: migrate [0x200000403:0x1:0x0]/14 failed: rc = -16 LustreError: 6267:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e6f3dd28: inode [0x200000403:0x53a:0x0] mdc close failed: rc = -116 LustreError: 11353:0:(mdt_reint.c:2513:mdt_reint_migrate()) lustre-MDT0001: migrate [0x240000403:0x1:0x0]/3 failed: rc = -114 LustreError: 11353:0:(mdt_reint.c:2513:mdt_reint_migrate()) Skipped 8 previous similar messages LustreError: 10614:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e6f3dd28: inode [0x240000404:0x2ff:0x0] mdc close failed: rc = -116 LustreError: 10614:0:(file.c:247:ll_close_inode_openhandle()) Skipped 4 previous similar messages Lustre: 12718:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0002: opcode 7: before 516 < left 540, rollback = 7 Lustre: 12718:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 5 previous similar messages Lustre: 12718:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 12718:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 12718:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 12718:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 12718:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/540/0, punch: 0/0/0, quota 1/3/0 Lustre: 12718:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 12718:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 12718:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 12718:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 12718:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 5 previous similar messages LustreError: 18146:0:(mdt_reint.c:2513:mdt_reint_migrate()) lustre-MDT0000: migrate [0x200000403:0x2:0x0]/11 failed: rc = -16 LustreError: 18146:0:(mdt_reint.c:2513:mdt_reint_migrate()) Skipped 1 previous similar message LustreError: 14907:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e6f3dd28: inode [0x200000403:0x6cc:0x0] mdc close failed: rc = -116 LustreError: 14907:0:(file.c:247:ll_close_inode_openhandle()) Skipped 1 previous similar message LustreError: 31734:0:(mdt_reint.c:2513:mdt_reint_migrate()) lustre-MDT0000: migrate [0x200000403:0x2:0x0]/11 failed: rc = -114 LustreError: 31734:0:(mdt_reint.c:2513:mdt_reint_migrate()) Skipped 2 previous similar messages LustreError: 14907:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e6f3dd28: inode [0x200000403:0x6e1:0x0] mdc close failed: rc = -116 LustreError: 14907:0:(file.c:247:ll_close_inode_openhandle()) Skipped 2 previous similar messages Lustre: 19138:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff8802c7ef3240 x1825723161782144/t4294978940(0) o101->9f96f94b-79e7-4e34-983d-2dc9620efa89@0@lo:705/0 lens 376/840 e 0 to 0 dl 1741145465 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 Lustre: 12717:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 516 < left 618, rollback = 7 Lustre: 12717:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 5 previous similar messages Lustre: 12717:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 12717:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 12717:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 12717:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 12717:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 12717:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 12717:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 12717:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 12717:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 12717:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 5 previous similar messages LustreError: 17175:0:(mdt_reint.c:2513:mdt_reint_migrate()) lustre-MDT0002: migrate [0x280000403:0x1:0x0]/5 failed: rc = -114 LustreError: 30968:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e6f3dd28: inode [0x280000403:0x55f:0x0] mdc close failed: rc = -116 Lustre: 12720:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 618, rollback = 7 Lustre: 12720:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 1 previous similar message 3[12273]: segfault at 8 ip 00007f058a9427e8 sp 00007ffdcd518ed0 error 4 in ld-2.17.so[7f058a937000+22000] LustreError: 18146:0:(mdt_reint.c:2513:mdt_reint_migrate()) lustre-MDT0001: migrate [0x240000403:0x1:0x0]/6 failed: rc = -114 LustreError: 18146:0:(mdt_reint.c:2513:mdt_reint_migrate()) Skipped 3 previous similar messages LustreError: 11747:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e8bd5d28: inode [0x200000404:0xbac:0x0] mdc close failed: rc = -116 LustreError: 11747:0:(file.c:247:ll_close_inode_openhandle()) Skipped 2 previous similar messages 0[19871]: segfault at 8 ip 00007f65693b67e8 sp 00007ffce33be880 error 4 in ld-2.17.so[7f65693ab000+22000] Lustre: 23085:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 23085:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 11 previous similar messages Lustre: 23085:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 23085:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 11 previous similar messages Lustre: 23085:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 23085:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 11 previous similar messages Lustre: 23085:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 23085:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 11 previous similar messages Lustre: 23085:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 23085:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 11 previous similar messages 17[22963]: segfault at 8 ip 00007fddcd10c7e8 sp 00007ffdf9728710 error 4 in ld-2.17.so[7fddcd101000+22000] 9[26210]: segfault at 8 ip 00007f33dbf277e8 sp 00007ffe04c32c10 error 4 in ld-2.17.so[7f33dbf1c000+22000] Lustre: 18146:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 514 < left 1072, rollback = 2 Lustre: 18146:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 36 previous similar messages LustreError: 31176:0:(mdt_reint.c:2513:mdt_reint_migrate()) lustre-MDT0002: migrate [0x280000403:0x7a3:0x0]/6 failed: rc = -114 LustreError: 31176:0:(mdt_reint.c:2513:mdt_reint_migrate()) Skipped 10 previous similar messages LustreError: 30231:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e6f3dd28: inode [0x200000403:0xefe:0x0] mdc close failed: rc = -116 LustreError: 30231:0:(file.c:247:ll_close_inode_openhandle()) Skipped 2 previous similar messages 8[32643]: segfault at 8 ip 00007f7b13feb7e8 sp 00007fffc2f724d0 error 4 in ld-2.17.so[7f7b13fe0000+22000] Lustre: 16970:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff8802cb52b240 x1825723182922752/t4294980320(0) o101->9f96f94b-79e7-4e34-983d-2dc9620efa89@0@lo:71/0 lens 376/840 e 0 to 0 dl 1741145586 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 Lustre: 16904:0:(mdt_recovery.c:102:mdt_req_from_lrd()) @@@ restoring transno req@ffff8802eaaa0a40 x1825723184025216/t4294981029(0) o101->a7e32c7a-bae3-4b7f-ad03-c9e7d3f17000@0@lo:77/0 lens 376/816 e 0 to 0 dl 1741145592 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 19[3450]: segfault at 8 ip 00007fe388d987e8 sp 00007fffa67bac00 error 4 in ld-2.17.so[7fe388d8d000+22000] 13[2652]: segfault at 0 ip 0000000000403f00 sp 00007ffc6a34eea8 error 6 in 13[400000+6000] Lustre: 31176:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 511 < left 894, rollback = 2 Lustre: 31176:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 19 previous similar messages 19[4885]: segfault at 0 ip (null) sp 00007ffeb60a4cd8 error 14 in 19[400000+6000] 19[4972]: segfault at 0 ip (null) sp 00007fffc8210938 error 14 in 19[400000+6000] Lustre: 12720:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 618, rollback = 7 Lustre: 12720:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 57 previous similar messages 15[14411]: segfault at 0 ip 0000000000403cf0 sp 00007ffe53040128 error 6 in 15[400000+6000] 2[17535]: segfault at 8 ip 00007f88915097e8 sp 00007fff9a690f80 error 4 in ld-2.17.so[7f88914fe000+22000] 19[21583]: segfault at 8 ip 00007f35f7eeb7e8 sp 00007ffdc5f09a20 error 4 in ld-2.17.so[7f35f7ee0000+22000] Lustre: 30331:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000404:0x9e9:0x0] with magic=0xbd60bd0 LustreError: 3179:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e8bd5d28: inode [0x200000404:0xe59:0x0] mdc close failed: rc = -13 LustreError: 3179:0:(file.c:247:ll_close_inode_openhandle()) Skipped 5 previous similar messages LustreError: 11356:0:(mdt_reint.c:2513:mdt_reint_migrate()) lustre-MDT0001: migrate [0x240000403:0x1:0x0]/19 failed: rc = -114 LustreError: 11356:0:(mdt_reint.c:2513:mdt_reint_migrate()) Skipped 6 previous similar messages 8[12520]: segfault at 8 ip 00007ff1079747e8 sp 00007ffcb8406080 error 4 in ld-2.17.so[7ff107969000+22000] Lustre: 31734:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 511 < left 699, rollback = 2 Lustre: 31734:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 14 previous similar messages Lustre: 12718:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 12718:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 158 previous similar messages Lustre: 12718:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 12718:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 158 previous similar messages Lustre: 12718:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 12718:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 158 previous similar messages Lustre: 12718:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 12718:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 158 previous similar messages Lustre: 12718:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 12718:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 158 previous similar messages INFO: task mkdir:3717 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mkdir D ffff8800a3dc1280 13056 3717 16285 0x00000080 Call Trace: [<ffffffff817e19d9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df7ea>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa1d>] mutex_lock+0x2d/0x40 [<ffffffff81256f65>] filename_create+0x85/0x180 [<ffffffff81257c94>] ? getname_flags+0xc4/0x1a0 [<ffffffff81257f41>] user_path_create+0x41/0x60 [<ffffffff81259286>] SyS_mkdirat+0x46/0xe0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff81259339>] SyS_mkdir+0x19/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a INFO: task mkdir:4509 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mkdir D ffff8802c136e238 13056 4509 16641 0x00000080 Call Trace: [<ffffffff817e19d9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df7ea>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa1d>] mutex_lock+0x2d/0x40 [<ffffffff81256f65>] filename_create+0x85/0x180 [<ffffffff81257c94>] ? getname_flags+0xc4/0x1a0 [<ffffffff81257f41>] user_path_create+0x41/0x60 [<ffffffff81259286>] SyS_mkdirat+0x46/0xe0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff81259339>] SyS_mkdir+0x19/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a Lustre: 23085:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 516 < left 618, rollback = 7 Lustre: 23085:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 65 previous similar messages Lustre: 2462:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0x1a8c:0x0] with magic=0xbd60bd0 Lustre: 2462:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message 3[1477]: segfault at 8 ip 00007f785c6d37e8 sp 00007ffd08f4d030 error 4 in ld-2.17.so[7f785c6c8000+22000] 12[2771]: segfault at 8 ip 00007f783a8a97e8 sp 00007ffca02b03e0 error 4 in ld-2.17.so[7f783a89e000+22000] 2[11282]: segfault at 8 ip 00007f2d9271a7e8 sp 00007ffec60d8a20 error 4 in ld-2.17.so[7f2d9270f000+22000] LustreError: 28228:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e6f3dd28: inode [0x200000403:0x1aca:0x0] mdc close failed: rc = -13 LustreError: 28228:0:(file.c:247:ll_close_inode_openhandle()) Skipped 13 previous similar messages 14[17574]: segfault at 1c ip 00000000004017b0 sp 00007ffc9e40ff70 error 6 in 14[400000+6000] Lustre: 25484:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000404:0x1009:0x0] with magic=0xbd60bd0 Lustre: 25484:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message LustreError: 18146:0:(mdt_reint.c:2513:mdt_reint_migrate()) lustre-MDT0002: migrate [0x200000403:0x2:0x0]/15 failed: rc = -16 LustreError: 18146:0:(mdt_reint.c:2513:mdt_reint_migrate()) Skipped 18 previous similar messages Lustre: 1300:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 507 < left 1517, rollback = 2 Lustre: 1300:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 15 previous similar messages LustreError: 21228:0:(lcommon_cl.c:179:cl_file_inode_init()) lustre: failed to initialize cl_object [0x280000404:0x1131:0x0]: rc = -5 LustreError: 21228:0:(llite_lib.c:3716:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 23784:0:(lcommon_cl.c:179:cl_file_inode_init()) lustre: failed to initialize cl_object [0x280000404:0x1131:0x0]: rc = -5 LustreError: 23784:0:(llite_lib.c:3716:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 202:0:(lcommon_cl.c:179:cl_file_inode_init()) lustre: failed to initialize cl_object [0x280000404:0x1131:0x0]: rc = -5 LustreError: 202:0:(llite_lib.c:3716:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 202:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 15 [0x280000404:0x1131:0x0] inode@0000000000000000: rc = -5 LustreError: 28638:0:(lcommon_cl.c:179:cl_file_inode_init()) lustre: failed to initialize cl_object [0x280000404:0x1131:0x0]: rc = -5 LustreError: 28638:0:(lcommon_cl.c:179:cl_file_inode_init()) Skipped 2 previous similar messages LustreError: 28638:0:(llite_lib.c:3716:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 28638:0:(llite_lib.c:3716:ll_prep_inode()) Skipped 2 previous similar messages 3[31385]: segfault at 8 ip 00007ff23a9c97e8 sp 00007fff4cc39800 error 4 in ld-2.17.so[7ff23a9be000+22000] 16[32447]: segfault at 8 ip 00007faa62d197e8 sp 00007fffbabae900 error 4 in ld-2.17.so[7faa62d0e000+22000] INFO: task mkdir:17217 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mkdir D ffff8802d8931900 13056 17217 16461 0x00000080 Call Trace: [<ffffffff817e19d9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df7ea>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa1d>] mutex_lock+0x2d/0x40 [<ffffffff81256f65>] filename_create+0x85/0x180 [<ffffffff81257c94>] ? getname_flags+0xc4/0x1a0 [<ffffffff81257f41>] user_path_create+0x41/0x60 [<ffffffff81259286>] SyS_mkdirat+0x46/0xe0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff81259339>] SyS_mkdir+0x19/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a INFO: task mkdir:17723 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mkdir D ffff8802d89a9280 12496 17723 16249 0x00000080 Call Trace: [<ffffffff817e19d9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df7ea>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa1d>] mutex_lock+0x2d/0x40 [<ffffffff81256f65>] filename_create+0x85/0x180 [<ffffffff81257c94>] ? getname_flags+0xc4/0x1a0 [<ffffffff81257f41>] user_path_create+0x41/0x60 [<ffffffff81259286>] SyS_mkdirat+0x46/0xe0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff81259339>] SyS_mkdir+0x19/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a INFO: task mkdir:18488 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mkdir D ffff8802db9a31c0 11200 18488 16333 0x00000080 Call Trace: [<ffffffff817e19d9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df7ea>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa1d>] mutex_lock+0x2d/0x40 [<ffffffff81256f65>] filename_create+0x85/0x180 [<ffffffff81257c94>] ? getname_flags+0xc4/0x1a0 [<ffffffff81257f41>] user_path_create+0x41/0x60 [<ffffffff81259286>] SyS_mkdirat+0x46/0xe0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff81259339>] SyS_mkdir+0x19/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a INFO: task mkdir:18710 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mkdir D ffff8802d344e340 11200 18710 16674 0x00000080 Call Trace: [<ffffffff817e19d9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df7ea>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa1d>] mutex_lock+0x2d/0x40 [<ffffffff81256f65>] filename_create+0x85/0x180 [<ffffffff81257c94>] ? getname_flags+0xc4/0x1a0 [<ffffffff81257f41>] user_path_create+0x41/0x60 [<ffffffff81259286>] SyS_mkdirat+0x46/0xe0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff81259339>] SyS_mkdir+0x19/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a INFO: task mkdir:20366 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mkdir D ffff88009f8552c0 13024 20366 16270 0x00000080 Call Trace: [<ffffffff817e19d9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df7ea>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa1d>] mutex_lock+0x2d/0x40 [<ffffffff81256f65>] filename_create+0x85/0x180 [<ffffffff81257c94>] ? getname_flags+0xc4/0x1a0 [<ffffffff81257f41>] user_path_create+0x41/0x60 [<ffffffff81259286>] SyS_mkdirat+0x46/0xe0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff81259339>] SyS_mkdir+0x19/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a INFO: task mkdir:20487 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mkdir D ffff8800a40e1878 13024 20487 16371 0x00000080 Call Trace: [<ffffffff817e19d9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df7ea>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa1d>] mutex_lock+0x2d/0x40 [<ffffffff81256f65>] filename_create+0x85/0x180 [<ffffffff81257c94>] ? getname_flags+0xc4/0x1a0 [<ffffffff81257f41>] user_path_create+0x41/0x60 [<ffffffff81259286>] SyS_mkdirat+0x46/0xe0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff81259339>] SyS_mkdir+0x19/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a INFO: task mkdir:24198 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mkdir D ffff88009b32a980 13056 24198 16620 0x00000080 Call Trace: [<ffffffff817e19d9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df7ea>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa1d>] mutex_lock+0x2d/0x40 [<ffffffff81256f65>] filename_create+0x85/0x180 [<ffffffff81257c94>] ? getname_flags+0xc4/0x1a0 [<ffffffff81257f41>] user_path_create+0x41/0x60 [<ffffffff81259286>] SyS_mkdirat+0x46/0xe0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff81259339>] SyS_mkdir+0x19/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a INFO: task mkdir:25420 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mkdir D ffff8802cf070040 13024 25420 16595 0x00000080 Call Trace: [<ffffffff817e19d9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df7ea>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa1d>] mutex_lock+0x2d/0x40 [<ffffffff81256f65>] filename_create+0x85/0x180 [<ffffffff81257c94>] ? getname_flags+0xc4/0x1a0 [<ffffffff81257f41>] user_path_create+0x41/0x60 [<ffffffff81259286>] SyS_mkdirat+0x46/0xe0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff81259339>] SyS_mkdir+0x19/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a Lustre: 18146:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 501 < left 877, rollback = 2 Lustre: 18146:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 34 previous similar messages Lustre: 30331:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0x2347:0x0] with magic=0xbd60bd0 Lustre: 30331:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message traps: 18[8933] general protection ip:40494a sp:7ffee9233c58 error:0 in 18[400000+6000] Lustre: 7286:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000404:0x12cf:0x0] with magic=0xbd60bd0 Lustre: 7286:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message Lustre: ll_ost_out00_00: service thread pid 12723 was inactive for 40.082 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 16969, comm: mdt00_008 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_pdo_lock+0x369/0x750 [mdt] [<0>] mdt_parent_lock+0x76/0x2a0 [mdt] [<0>] mdt_getattr_name_lock+0x1850/0x2d10 [mdt] [<0>] mdt_intent_getattr+0x2cc/0x4e0 [mdt] [<0>] mdt_intent_opc.constprop.76+0x211/0xc60 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x397/0x980 [ptlrpc] [<0>] ldlm_handle_enqueue+0x547/0x18d0 [ptlrpc] [<0>] tgt_enqueue+0x68/0x240 [ptlrpc] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 18659, comm: mdt_io00_005 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock+0x88/0x1c0 [mdt] [<0>] mdt_rename_source_lock+0xa9/0xd6 [mdt] [<0>] mdt_reint_rename+0x1de3/0x2c30 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x84f/0x13c0 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: Skipped 2 previous similar messages Pid: 12723, comm: ll_ost_out00_00 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x943/0xd80 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x810 [ptlrpc] [<0>] mdt_object_pdo_lock+0x369/0x750 [mdt] [<0>] mdt_parent_lock+0x76/0x2a0 [mdt] [<0>] mdt_getattr_name_lock+0x1850/0x2d10 [mdt] [<0>] mdt_intent_getattr+0x2cc/0x4e0 [mdt] [<0>] mdt_intent_opc.constprop.76+0x211/0xc60 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x397/0x980 [ptlrpc] [<0>] ldlm_handle_enqueue+0x547/0x18d0 [ptlrpc] [<0>] mdt_batch_getattr+0x6c/0x190 [mdt] [<0>] mdt_batch+0x465/0x195b [mdt] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: mdt00_024: service thread pid 2332 was inactive for 40.143 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: mdt_io00_007: service thread pid 31734 was inactive for 72.207 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 14 previous similar messages LustreError: 11087:0:(ldlm_lockd.c:252:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0001_UUID lock: ffff8802accb2200/0xd20e60d0fbefc lrc: 3/0,0 mode: PR/PR res: [0x240000404:0x116e:0x0].0x0 bits 0x1b/0x0 rrc: 5 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0xd20e60d0fbbb4 expref: 1019 pid: 12723 timeout: 699 lvb_type: 0 Lustre: mdt00_028: service thread pid 7250 completed after 94.963s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_014: service thread pid 19153 completed after 94.986s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 19253:0:(ldlm_lockd.c:1492:ldlm_handle_enqueue()) ### lock on destroyed export ffff8802e6f3a548 ns: mdt-lustre-MDT0001_UUID lock: ffff88009b42de00/0xd20e60d111ea7 lrc: 3/0,0 mode: PR/PR res: [0x240000403:0x1:0x0].0x0 bits 0x12/0x0 rrc: 22 type: IBT gid 0 flags: 0x50200000000000 nid: 0@lo remote: 0xd20e60d111e84 expref: 146 pid: 19253 timeout: 0 lvb_type: 0 Lustre: mdt_io00_005: service thread pid 18659 completed after 100.261s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_010: service thread pid 16973 completed after 94.987s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: ll_ost_out00_00: service thread pid 12723 completed after 100.271s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_020: service thread pid 20275 completed after 94.989s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_013: service thread pid 19150 completed after 94.990s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_002: service thread pid 11334 completed after 94.999s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_026: service thread pid 14145 completed after 95.040s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: lustre-MDT0001-mdc-ffff8802e8bd5d28: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-MDT0001-mdc-ffff8802e8bd5d28: Connection to lustre-MDT0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: mdt00_003: service thread pid 13190 completed after 95.066s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_000: service thread pid 11332 completed after 95.059s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_015: service thread pid 19170 completed after 95.305s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_008: service thread pid 16969 completed after 100.225s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: lustre-MDT0001-mdc-ffff8802e8bd5d28: This client was evicted by lustre-MDT0001; in progress operations using this service will fail. Lustre: mdt00_024: service thread pid 2332 completed after 95.482s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_012: service thread pid 19138 completed after 95.018s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_027: service thread pid 7248 completed after 94.989s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_018: service thread pid 19245 completed after 95.048s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 25119:0:(file.c:6073:ll_inode_revalidate_fini()) lustre: revalidate FID [0x240000403:0x1:0x0] error: rc = -5 Lustre: mdt00_019: service thread pid 19253 completed after 95.325s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 11737:0:(mdc_request.c:1464:mdc_read_page()) lustre-MDT0001-mdc-ffff8802e8bd5d28: [0x240000402:0x18:0x0] lock enqueue fails: rc = -108 Lustre: dir [0x200000403:0x25cd:0x0] stripe 1 readdir failed: -108, directory is partially accessed! Lustre: Skipped 6 previous similar messages LustreError: 11701:0:(llite_lib.c:1996:ll_md_setattr()) md_setattr fails: rc = -108 Lustre: 18659:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 511 < left 816, rollback = 2 Lustre: 18659:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 151 previous similar messages Lustre: lustre-MDT0001-mdc-ffff8802e8bd5d28: Connection restored to 192.168.123.22@tcp (at 0@lo) Lustre: mdt_io00_007: service thread pid 31734 completed after 78.361s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). 9[4511]: segfault at 8 ip 00007f3ad49147e8 sp 00007fff1b9474a0 error 4 in ld-2.17.so[7f3ad4909000+22000] traps: 12[15349] general protection ip:404908 sp:7ffddb0b03e8 error:0 in 12[400000+6000] 17[17704]: segfault at 406000 ip 0000000000406000 sp 00007ffea4ea5638 error 14 in 17[606000+1000] LustreError: 19525:0:(lcommon_cl.c:179:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000404:0x11b5:0x0]: rc = -5 LustreError: 19525:0:(llite_lib.c:3716:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 5[22280]: segfault at 404572 ip 0000000000405012 sp 00007ffc205806b8 error 7 in 5[400000+6000] Lustre: 31176:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 3/12/3, destroy: 0/0/0 Lustre: 31176:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 356 previous similar messages Lustre: 31176:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 2/2/0, xattr_set: 17/1411/0 Lustre: 31176:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 356 previous similar messages Lustre: 31176:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 5/55/0, punch: 0/0/0, quota 4/54/0 Lustre: 31176:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 356 previous similar messages Lustre: 31176:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 12/275/3, delete: 0/0/0 Lustre: 31176:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 356 previous similar messages Lustre: 31176:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 6/6/0, ref_del: 0/0/0 Lustre: 31176:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 356 previous similar messages LustreError: 11335:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0001: failed to get lu_attr of [0x240000403:0x1250:0x0]: rc = -2 Lustre: 12720:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 618, rollback = 7 Lustre: 12720:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 145 previous similar messages 3[303]: segfault at 52 ip 0000000000404d28 sp 00007fffee046dc8 error 6 in 3[400000+6000] 10[471]: segfault at 8 ip 00007f8ed15167e8 sp 00007fff03efe4a0 error 4 in ld-2.17.so[7f8ed150b000+22000] Lustre: 25484:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0x2a8a:0x0] with magic=0xbd60bd0 Lustre: 25484:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 3 previous similar messages 6[10395]: segfault at 8 ip 00007f859363f7e8 sp 00007fffdb650990 error 4 in ld-2.17.so[7f8593634000+22000] 4[10438]: segfault at 8 ip 00007f2c692807e8 sp 00007ffec81028a0 error 4 in ld-2.17.so[7f2c69275000+22000] 10[11078]: segfault at 0 ip (null) sp 00007fff66b4f448 error 14 in 10[400000+6000] 4[15436]: segfault at 8 ip 00007fc5be75f7e8 sp 00007fff124efbc0 error 4 in ld-2.17.so[7fc5be754000+22000] Lustre: dir [0x240000405:0x2ea:0x0] stripe 3 readdir failed: -2, directory is partially accessed! Lustre: Skipped 2 previous similar messages LustreError: 11353:0:(mdd_dir.c:4619:mdd_migrate_cmd_check()) lustre-MDD0001: '15' migration was interrupted, run 'lfs migrate -m 2 -c 1 -H crush 15' to finish migration: rc = -1 LustreError: 17485:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0001: failed to get lu_attr of [0x240000405:0x2ea:0x0]: rc = -2 Lustre: dir [0x200000404:0x29c0:0x0] stripe 3 readdir failed: -2, directory is partially accessed! Lustre: Skipped 2 previous similar messages 10[19482]: segfault at 8 ip 00007f6e1d11c7e8 sp 00007fffee74e270 error 4 in ld-2.17.so[7f6e1d111000+22000] LustreError: 11087:0:(ldlm_lockd.c:252:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0001_UUID lock: ffff8802b3b6a200/0xd20e60d27ad61 lrc: 3/0,0 mode: CR/CR res: [0x240000405:0x2de:0x0].0x0 bits 0xa/0x0 rrc: 3 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0xd20e60d27a0b1 expref: 219 pid: 7365 timeout: 885 lvb_type: 0 LustreError: lustre-MDT0001-mdc-ffff8802e8bd5d28: operation mds_reint to node 0@lo failed: rc = -107 LustreError: Skipped 7 previous similar messages LustreError: 25453:0:(ldlm_lockd.c:2591:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1741146117 with bad export cookie 3695347131201909 Lustre: lustre-MDT0001-mdc-ffff8802e8bd5d28: Connection to lustre-MDT0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: lustre-MDT0001-mdc-ffff8802e8bd5d28: This client was evicted by lustre-MDT0001; in progress operations using this service will fail. LustreError: 19770:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e8bd5d28: inode [0x240000405:0x306:0x0] mdc close failed: rc = -108 LustreError: 19770:0:(file.c:247:ll_close_inode_openhandle()) Skipped 57 previous similar messages LustreError: 15461:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x240000405:0x2de:0x0] error -5. LustreError: 13999:0:(mdc_request.c:1464:mdc_read_page()) lustre-MDT0001-mdc-ffff8802e8bd5d28: [0x240000403:0x1:0x0] lock enqueue fails: rc = -108 LustreError: 13999:0:(mdc_request.c:1464:mdc_read_page()) Skipped 2 previous similar messages Lustre: lustre-MDT0001-mdc-ffff8802e8bd5d28: Connection restored to 192.168.123.22@tcp (at 0@lo) LustreError: 11087:0:(ldlm_lockd.c:252:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0000_UUID lock: ffff8802c3cb43c0/0xd20e60d2af6e9 lrc: 3/0,0 mode: PR/PR res: [0x200000404:0x2966:0x0].0x0 bits 0x1b/0x0 rrc: 5 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0xd20e60d2af6cd expref: 1834 pid: 2326 timeout: 902 lvb_type: 0 LustreError: 11355:0:(mdt_reint.c:2513:mdt_reint_migrate()) lustre-MDT0000: migrate [0x200000403:0x2:0x0]/3 failed: rc = -114 LustreError: 11355:0:(mdt_reint.c:2513:mdt_reint_migrate()) Skipped 73 previous similar messages LustreError: lustre-MDT0000-mdc-ffff8802e8bd5d28: operation mds_close to node 0@lo failed: rc = -107 Lustre: lustre-MDT0000-mdc-ffff8802e8bd5d28: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: lustre-MDT0000-mdc-ffff8802e8bd5d28: This client was evicted by lustre-MDT0000; in progress operations using this service will fail. LustreError: 11353:0:(client.c:1282:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802adadda40 x1825723258829952/t0(0) o104->lustre-MDT0000@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' uid:4294967295 gid:4294967295 LustreError: 20096:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x200000404:0x2966:0x0] error -108. LustreError: 19871:0:(mdc_request.c:1464:mdc_read_page()) lustre-MDT0000-mdc-ffff8802e8bd5d28: [0x200000403:0x2:0x0] lock enqueue fails: rc = -5 LustreError: 19496:0:(statahead.c:1800:is_first_dirent()) lustre: reading dir [0x200000403:0x2:0x0] at 0 stat_pid = 19873 : rc = -5 LustreError: 19871:0:(mdc_request.c:1464:mdc_read_page()) Skipped 4 previous similar messages LustreError: 19697:0:(file.c:6073:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000007:0x1:0x0] error: rc = -108 LustreError: 19697:0:(file.c:6073:ll_inode_revalidate_fini()) Skipped 130 previous similar messages Lustre: lustre-MDT0000-mdc-ffff8802e8bd5d28: Connection restored to 192.168.123.22@tcp (at 0@lo) Lustre: 20944:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 510 < left 922, rollback = 2 Lustre: 20944:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 242 previous similar messages LustreError: 22675:0:(lcommon_cl.c:179:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000404:0x2aa2:0x0]: rc = -5 LustreError: 22675:0:(llite_lib.c:3716:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 17175:0:(mdd_dir.c:4619:mdd_migrate_cmd_check()) lustre-MDD0001: '4' migration was interrupted, run 'lfs migrate -m 0 -c 3 -H crush 4' to finish migration: rc = -1 3[21897]: segfault at 0 ip 0000000000401d60 sp 00007fff5e28ac38 error 6 in 3[400000+6000] 18[23003]: segfault at 0 ip 0000000000401d60 sp 00007fffdb0dcb78 error 6 in 3[400000+6000] LustreError: 18146:0:(mdd_dir.c:4619:mdd_migrate_cmd_check()) lustre-MDD0002: '4' migration was interrupted, run 'lfs migrate -m 0 -c 3 -H crush 4' to finish migration: rc = -1 LustreError: 18146:0:(mdd_dir.c:4619:mdd_migrate_cmd_check()) Skipped 1 previous similar message LustreError: 31734:0:(mdd_dir.c:4619:mdd_migrate_cmd_check()) lustre-MDD0000: '1' migration was interrupted, run 'lfs migrate -m 2 -c 3 -H crush 1' to finish migration: rc = -1 LustreError: 24621:0:(lcommon_cl.c:179:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000404:0x2aee:0x0]: rc = -5 LustreError: 24621:0:(lcommon_cl.c:179:cl_file_inode_init()) Skipped 1 previous similar message LustreError: 24621:0:(llite_lib.c:3716:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 24621:0:(llite_lib.c:3716:ll_prep_inode()) Skipped 1 previous similar message Lustre: dir [0x200000404:0x2b63:0x0] stripe 2 readdir failed: -2, directory is partially accessed! Lustre: Skipped 1 previous similar message Lustre: 16969:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000403:0x176a:0x0] with magic=0xbd60bd0 Lustre: 16969:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message LustreError: 31734:0:(mdd_dir.c:4619:mdd_migrate_cmd_check()) lustre-MDD0000: '1' migration was interrupted, run 'lfs migrate -m 1 -c 3 -H crush 1' to finish migration: rc = -1 1[362]: segfault at 8 ip 00007efd7f2387e8 sp 00007ffe2b7aeba0 error 4 in ld-2.17.so[7efd7f22d000+22000] 16[996]: segfault at 8 ip 00007f25425557e8 sp 00007ffcf74a4650 error 4 in ld-2.17.so[7f254254a000+22000] LustreError: 31009:0:(dir.c:833:ll_dir_get_default_layout()) lustre: unknown magic: CD40CD0: rc = -71 Lustre: dir [0x240000404:0x15f0:0x0] stripe 3 readdir failed: -2, directory is partially accessed! Lustre: Skipped 1 previous similar message Lustre: 7264:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000403:0x1865:0x0] with magic=0xbd60bd0 Lustre: 7264:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 53 previous similar messages 5[3635]: segfault at 406000 ip 0000000000406000 sp 00007ffe60acad78 error 14 in 5[606000+1000] LustreError: 2981:0:(lcommon_cl.c:179:cl_file_inode_init()) lustre: failed to initialize cl_object [0x280000403:0x186d:0x0]: rc = -5 LustreError: 2981:0:(lcommon_cl.c:179:cl_file_inode_init()) Skipped 5 previous similar messages LustreError: 2981:0:(llite_lib.c:3716:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 2981:0:(llite_lib.c:3716:ll_prep_inode()) Skipped 5 previous similar messages LustreError: 22830:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000405:0x444:0x0]: rc = -2 3[6441]: segfault at 8 ip 00007f91178257e8 sp 00007fff1c6bb9e0 error 4 in ld-2.17.so[7f911781a000+22000] LustreError: 22830:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0002: failed to get lu_attr of [0x280000404:0x18fb:0x0]: rc = -2 LustreError: 12602:0:(mdd_dir.c:4619:mdd_migrate_cmd_check()) lustre-MDD0002: '5' migration was interrupted, run 'lfs migrate -m 1 -c 1 -H crush 5' to finish migration: rc = -1 2[7826]: segfault at 406000 ip 0000000000406000 sp 00007ffd36287228 error 14 in 2[606000+1000] 0[8838]: segfault at 8 ip 00007f0bf005b7e8 sp 00007ffcd68581a0 error 4 in ld-2.17.so[7f0bf0050000+22000] LustreError: 13185:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0001: failed to get lu_attr of [0x240000404:0x15f0:0x0]: rc = -2 Lustre: dir [0x280000404:0x1936:0x0] stripe 3 readdir failed: -2, directory is partially accessed! Lustre: Skipped 11 previous similar messages LustreError: 11292:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8802e6f3dd28: cannot apply new layout on [0x200000405:0x5fc:0x0] : rc = -5 LustreError: 11292:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x200000405:0x5fc:0x0] error -5. LustreError: 14059:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8802e6f3dd28: cannot apply new layout on [0x200000405:0x5fc:0x0] : rc = -5 LustreError: 19031:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8802e6f3dd28: cannot apply new layout on [0x200000405:0x5fc:0x0] : rc = -5 3[20222]: segfault at 400e4403 ip 0000000000405c38 sp 00007ffe24969568 error 6 in 3[400000+6000] LustreError: 21877:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8802e6f3dd28: cannot apply new layout on [0x200000405:0x5fc:0x0] : rc = -5 19[26069]: segfault at 1c ip 00000000004017b0 sp 00007ffe3ee378c0 error 6 in 19[400000+6000] LustreError: 26147:0:(lcommon_cl.c:179:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000404:0x18d4:0x0]: rc = -5 LustreError: 26147:0:(lcommon_cl.c:179:cl_file_inode_init()) Skipped 30 previous similar messages LustreError: 26147:0:(llite_lib.c:3716:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 26147:0:(llite_lib.c:3716:ll_prep_inode()) Skipped 30 previous similar messages LustreError: 17175:0:(mdd_dir.c:4619:mdd_migrate_cmd_check()) lustre-MDD0002: '5' migration was interrupted, run 'lfs migrate -m 2 -c 2 -H crush 5' to finish migration: rc = -1 LustreError: 17175:0:(mdd_dir.c:4619:mdd_migrate_cmd_check()) Skipped 3 previous similar messages 0[28216]: segfault at 7faa90c96f0d ip 0000000000404db2 sp 00007ffc177df258 error 4 in 0[400000+6000] LustreError: 31005:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8802e6f3dd28: cannot apply new layout on [0x200000405:0x5fc:0x0] : rc = -5 LustreError: 31005:0:(lov_object.c:1341:lov_layout_change()) Skipped 1 previous similar message Lustre: 13190:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000404:0x1c1b:0x0] with magic=0xbd60bd0 Lustre: 13190:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 49 previous similar messages LustreError: 2179:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8802e6f3dd28: cannot apply new layout on [0x200000405:0x5fc:0x0] : rc = -5 Lustre: dir [0x240000404:0x1af7:0x0] stripe 3 readdir failed: -2, directory is partially accessed! Lustre: Skipped 3 previous similar messages LustreError: 415:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 18 [0x200000404:0x32eb:0x0] inode@0000000000000000: rc = -5 2[9237]: segfault at 8 ip 00007fe9eb6b07e8 sp 00007ffce719aa10 error 4 in ld-2.17.so[7fe9eb6a5000+22000] 5[9602]: segfault at 8 ip 00007f901a73a7e8 sp 00007ffcb28148e0 error 4 in ld-2.17.so[7f901a72f000+22000] LustreError: 18146:0:(mdd_dir.c:4619:mdd_migrate_cmd_check()) lustre-MDD0001: '17' migration was interrupted, run 'lfs migrate -m 2 -c 3 -H crush 17' to finish migration: rc = -1 LustreError: 18146:0:(mdd_dir.c:4619:mdd_migrate_cmd_check()) Skipped 5 previous similar messages LustreError: 17485:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0002: failed to get lu_attr of [0x280000403:0x1de7:0x0]: rc = -2 LustreError: 17485:0:(mdd_object.c:3901:mdd_close()) Skipped 2 previous similar messages 16[18868]: segfault at 8 ip 00007f4625ed87e8 sp 00007ffe1cab5c10 error 4 in ld-2.17.so[7f4625ecd000+22000] 14[20836]: segfault at 8 ip 00007f84291f37e8 sp 00007fff8f0b7df0 error 4 in ld-2.17.so[7f84291e8000+22000] 9[29713]: segfault at 8 ip 00007fefa01177e8 sp 00007ffde8047650 error 4 in ld-2.17.so[7fefa010c000+22000] 16[31816]: segfault at 406000 ip 0000000000406000 sp 00007fffc283f168 error 14 in 16[606000+1000] 0[9930]: segfault at 8 ip 00007f441ccd97e8 sp 00007ffed1a7a4d0 error 4 in ld-2.17.so[7f441ccce000+22000] Lustre: 31734:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 508 < left 727, rollback = 2 Lustre: 31734:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 1944 previous similar messages LustreError: 16939:0:(lcommon_cl.c:179:cl_file_inode_init()) lustre: failed to initialize cl_object [0x280000403:0x2177:0x0]: rc = -5 LustreError: 16939:0:(lcommon_cl.c:179:cl_file_inode_init()) Skipped 71 previous similar messages LustreError: 16939:0:(llite_lib.c:3716:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 16939:0:(llite_lib.c:3716:ll_prep_inode()) Skipped 71 previous similar messages Lustre: dir [0x200000404:0x3d5d:0x0] stripe 2 readdir failed: -2, directory is partially accessed! Lustre: Skipped 6 previous similar messages 5[30313]: segfault at 0 ip 0000000000401d60 sp 00007ffd5abf8618 error 6 in 5[400000+6000] Lustre: 7365:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000405:0x137e:0x0] with magic=0xbd60bd0 Lustre: 7365:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 23 previous similar messages LustreError: 21124:0:(mdd_dir.c:4619:mdd_migrate_cmd_check()) lustre-MDD0002: '0' migration was interrupted, run 'lfs migrate -m 0 -c 1 -H crush 0' to finish migration: rc = -1 LustreError: 21124:0:(mdd_dir.c:4619:mdd_migrate_cmd_check()) Skipped 3 previous similar messages LustreError: 18830:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0001: failed to get lu_attr of [0x240000404:0x21e9:0x0]: rc = -2 Lustre: 31734:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 3/12/6, destroy: 0/0/0 Lustre: 31734:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 2890 previous similar messages Lustre: 31734:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 4/4/0, xattr_set: 13/905/0 Lustre: 31734:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 2890 previous similar messages Lustre: 31734:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 5/55/0, punch: 0/0/0, quota 4/54/0 Lustre: 22676:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 22676:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 2890 previous similar messages Lustre: 22676:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 22676:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 2890 previous similar messages Lustre: 31734:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 2892 previous similar messages 16[6470]: segfault at 8 ip 00007f803fc407e8 sp 00007ffe01892e50 error 4 in ld-2.17.so[7f803fc35000+22000] 13[7836]: segfault at 8 ip 00007f35213d27e8 sp 00007ffc8a8a0990 error 4 in ld-2.17.so[7f35213c7000+22000] Lustre: 17339:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 618, rollback = 7 Lustre: 17339:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 325 previous similar messages LustreError: 11335:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0001: failed to get lu_attr of [0x240000406:0xdfc:0x0]: rc = -2 LustreError: 11335:0:(mdd_object.c:3901:mdd_close()) Skipped 2 previous similar messages 0[12262]: segfault at 8 ip 00007fe2b85d07e8 sp 00007ffcd1447fe0 error 4 in ld-2.17.so[7fe2b85c5000+22000] LustreError: 15126:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 0 [0x280000403:0x253b:0x0] inode@0000000000000000: rc = -5 LustreError: 9722:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 20 [0x280000404:0x24e0:0x0] inode@0000000000000000: rc = -5 5[9988]: segfault at 8 ip 00007ff9bf4a37e8 sp 00007ffd07c124f0 error 4 in ld-2.17.so[7ff9bf498000+22000] 4[27457]: segfault at 8 ip 00007f62d9cd17e8 sp 00007ffce6108830 error 4 in ld-2.17.so[7f62d9cc6000+22000] LustreError: 18146:0:(mdt_reint.c:2513:mdt_reint_migrate()) lustre-MDT0001: migrate [0x240000403:0x1:0x0]/18 failed: rc = -114 LustreError: 18146:0:(mdt_reint.c:2513:mdt_reint_migrate()) Skipped 444 previous similar messages LustreError: 32580:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e8bd5d28: inode [0x240000406:0x16af:0x0] mdc close failed: rc = -116 LustreError: 32580:0:(file.c:247:ll_close_inode_openhandle()) Skipped 130 previous similar messages 10[6780]: segfault at 8 ip 00007fd83d7037e8 sp 00007fffc6ca8430 error 4 in ld-2.17.so[7fd83d6f8000+22000] 17[7104]: segfault at 8 ip 00007fded620f7e8 sp 00007ffe8c1f2620 error 4 in ld-2.17.so[7fded6204000+22000] 6[30169]: segfault at 8 ip 00007f3c3d5037e8 sp 00007ffdec4ffd10 error 4 in ld-2.17.so[7f3c3d4f8000+22000] 6[30725]: segfault at 8 ip 00007f6946eb77e8 sp 00007ffc227bc530 error 4 in ld-2.17.so[7f6946eac000+22000] 12[9092]: segfault at 0 ip (null) sp 00007ffd47070fe8 error 14 in 12[400000+6000] 18[11830]: segfault at 8 ip 00007f97713377e8 sp 00007ffc08224950 error 4 in ld-2.17.so[7f977132c000+22000] 18[14306]: segfault at 8 ip 00007fd0326117e8 sp 00007ffd97fb5900 error 4 in ld-2.17.so[7fd032606000+22000] 11[19997]: segfault at 8 ip 00007f4f456207e8 sp 00007ffdda61ed40 error 4 in ld-2.17.so[7f4f45615000+22000] 16[20546]: segfault at 8 ip 00007fc1d4fa77e8 sp 00007fff30332410 error 4 in ld-2.17.so[7fc1d4f9c000+22000] 3[22869]: segfault at 0 ip (null) sp 00007ffcfb3eea18 error 14 in 3[400000+6000] 18[23163]: segfault at 8 ip 00007efdbd63a7e8 sp 00007ffe629aa8c0 error 4 in ld-2.17.so[7efdbd62f000+22000] Lustre: dir [0x200000404:0x53f4:0x0] stripe 3 readdir failed: -2, directory is partially accessed! Lustre: Skipped 16 previous similar messages 1[28133]: segfault at 8 ip 00007fc69ff3b7e8 sp 00007ffc58555580 error 4 in ld-2.17.so[7fc69ff30000+22000] LustreError: 15793:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000405:0x2c86:0x0]: rc = -2 LustreError: 15793:0:(mdd_object.c:3901:mdd_close()) Skipped 3 previous similar messages 18[490]: segfault at 8 ip 00007f3ce49987e8 sp 00007ffedcc69de0 error 4 in ld-2.17.so[7f3ce498d000+22000] Lustre: 19253:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000404:0x319f:0x0] with magic=0xbd60bd0 Lustre: 19253:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 71 previous similar messages LustreError: 32059:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8802e8bd5d28: cannot apply new layout on [0x240000404:0x302c:0x0] : rc = -5 LustreError: 32059:0:(lov_object.c:1341:lov_layout_change()) Skipped 1 previous similar message LustreError: 32059:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x240000404:0x302c:0x0] error -5. 1[11426]: segfault at 8 ip 00007f24d54ca7e8 sp 00007ffc3ac644e0 error 4 in ld-2.17.so[7f24d54bf000+22000] 1[13062]: segfault at 8 ip 00007f763a9ab7e8 sp 00007ffdee6e5d60 error 4 in ld-2.17.so[7f763a9a0000+22000] LustreError: 23202:0:(lcommon_cl.c:179:cl_file_inode_init()) lustre: failed to initialize cl_object [0x280000404:0x33d3:0x0]: rc = -5 LustreError: 23202:0:(lcommon_cl.c:179:cl_file_inode_init()) Skipped 103 previous similar messages LustreError: 23202:0:(llite_lib.c:3716:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 23202:0:(llite_lib.c:3716:ll_prep_inode()) Skipped 103 previous similar messages 0[26171]: segfault at 8 ip 00007f43016027e8 sp 00007ffc13e01d80 error 4 in ld-2.17.so[7f43015f7000+22000] 14[28135]: segfault at 8 ip 00007f095e9f77e8 sp 00007ffeb4ef88d0 error 4 in ld-2.17.so[7f095e9ec000+22000] 18[7829]: segfault at 8 ip 00007f34cf6cf7e8 sp 00007ffe0d0b1230 error 4 in ld-2.17.so[7f34cf6c4000+22000] 14[15604]: segfault at 8 ip 00007f3f0d07d7e8 sp 00007ffcaed9d8a0 error 4 in ld-2.17.so[7f3f0d072000+22000] ptlrpc_watchdog_fire: 16 callbacks suppressed Lustre: ll_ost00_020: service thread pid 13456 was inactive for 40.077 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: ll_ost00_002: service thread pid 12713 was inactive for 40.060 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 12713, comm: ll_ost00_002 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] wait_transaction_locked+0x93/0xe0 [jbd2] [<0>] add_transaction_credits+0x270/0x310 [jbd2] [<0>] start_this_handle+0x1cc/0x460 [jbd2] [<0>] jbd2__journal_start+0xf3/0x200 [jbd2] [<0>] __ldiskfs_journal_start_sb+0x6d/0x100 [ldiskfs] [<0>] osd_trans_start+0x279/0x6b0 [osd_ldiskfs] [<0>] ofd_trans_start+0x68/0xe0 [ofd] [<0>] ofd_attr_set+0x4e2/0xaa0 [ofd] [<0>] ofd_setattr_hdl+0x325/0x9e0 [ofd] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 22676, comm: ll_ost_io00_004 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] wait_transaction_locked+0x93/0xe0 [jbd2] [<0>] add_transaction_credits+0x270/0x310 [jbd2] [<0>] start_this_handle+0x1cc/0x460 [jbd2] [<0>] jbd2__journal_start+0xf3/0x200 [jbd2] [<0>] __ldiskfs_journal_start_sb+0x6d/0x100 [ldiskfs] [<0>] osd_trans_start+0x279/0x6b0 [osd_ldiskfs] [<0>] ofd_trans_start+0x68/0xe0 [ofd] [<0>] ofd_object_punch+0x637/0xdb0 [ofd] [<0>] ofd_punch_hdl+0x4ff/0xb50 [ofd] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 1941, comm: ll_ost00_013 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] wait_transaction_locked+0x93/0xe0 [jbd2] [<0>] add_transaction_credits+0x270/0x310 [jbd2] [<0>] start_this_handle+0x1cc/0x460 [jbd2] [<0>] jbd2__journal_start+0xf3/0x200 [jbd2] [<0>] __ldiskfs_journal_start_sb+0x6d/0x100 [ldiskfs] [<0>] osd_trans_start+0x279/0x6b0 [osd_ldiskfs] [<0>] ofd_trans_start+0x68/0xe0 [ofd] [<0>] ofd_destroy+0x3a0/0xd10 [ofd] [<0>] ofd_destroy_by_fid+0x33c/0x610 [ofd] [<0>] ofd_destroy_hdl+0x20c/0xaf0 [ofd] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: ll_ost00_005: service thread pid 19137 was inactive for 40.048 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 6 previous similar messages Lustre: ll_ost00_004: service thread pid 14528 was inactive for 40.065 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 4 previous similar messages Lustre: ll_ost_io00_007: service thread pid 6212 was inactive for 40.045 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 12 previous similar messages Lustre: mdt00_026: service thread pid 14145 was inactive for 40.101 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 3 previous similar messages Lustre: ll_ost00_022: service thread pid 5826 was inactive for 72.125 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 17 previous similar messages Lustre: ll_ost00_039: service thread pid 23756 was inactive for 72.153 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 22 previous similar messages LustreError: 11087:0:(ldlm_lockd.c:252:expired_lock_main()) ### lock callback timer expired after 101s: evicting client at 0@lo ns: filter-lustre-OST0001_UUID lock: ffff8802c4615680/0xd20e60de22700 lrc: 3/0,0 mode: PW/PW res: [0x300000401:0x473:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 28672->53247) gid 0 flags: 0x60000400020020 nid: 0@lo remote: 0xd20e60de226eb expref: 53 pid: 13456 timeout: 1806 lvb_type: 0 LustreError: 12731:0:(client.c:1282:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88009f9f8a40 x1825723452542080/t0(0) o105->lustre-OST0001@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' uid:4294967295 gid:4294967295 Lustre: lustre-OST0001-osc-ffff8802e8bd5d28: Connection to lustre-OST0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: 3988:0:(client.c:2346:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1741146946/real 1741146946] req@ffff880091610540 x1825723450739328/t0(0) o2->lustre-OST0001-osc-ffff8802e8bd5d28@0@lo:28/4 lens 440/432 e 2 to 1 dl 1741147051 ref 1 fl Rpc:XQr/200/ffffffff rc 0/-1 job:'cp.0' uid:0 gid:0 Lustre: lustre-OST0001: Export ffff8802cc66b7e8 already connecting from 0@lo Lustre: 3984:0:(client.c:2346:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1741146961/real 1741146961] req@ffff8802cbf53740 x1825723452215808/t0(0) o4->lustre-OST0001-osc-ffff8802e8bd5d28@0@lo:6/4 lens 504/448 e 3 to 1 dl 1741147052 ref 2 fl Rpc:XQr/200/ffffffff rc 0/-1 job:'dd.0' uid:0 gid:0 Lustre: 3995:0:(client.c:2346:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1741146938/real 1741146938] req@ffff880093868040 x1825723450079616/t0(0) o4->lustre-OST0001-osc-ffff8802e8bd5d28@0@lo:6/4 lens 488/448 e 4 to 1 dl 1741147054 ref 2 fl Rpc:XQr/200/ffffffff rc 0/-1 job:'dd.0' uid:0 gid:0 Lustre: 3995:0:(client.c:2346:ptlrpc_expire_one_request()) Skipped 3 previous similar messages Lustre: 3981:0:(client.c:2346:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1741146940/real 1741146940] req@ffff88028ef03c40 x1825723450308736/t0(0) o4->lustre-OST0001-osc-ffff8802e8bd5d28@0@lo:6/4 lens 488/448 e 4 to 1 dl 1741147056 ref 2 fl Rpc:XQr/200/ffffffff rc 0/-1 job:'dd.0' uid:0 gid:0 Lustre: 3981:0:(client.c:2346:ptlrpc_expire_one_request()) Skipped 7 previous similar messages Lustre: lustre-OST0001: Export ffff8802cc66b7e8 already connecting from 0@lo Lustre: 3989:0:(client.c:2346:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1741146944/real 1741146944] req@ffff88009ea66940 x1825723450640256/t0(0) o2->lustre-OST0001-osc-ffff8802e8bd5d28@0@lo:28/4 lens 440/432 e 4 to 1 dl 1741147060 ref 1 fl Rpc:XQr/200/ffffffff rc 0/-1 job:'chown.0' uid:0 gid:0 Lustre: 3989:0:(client.c:2346:ptlrpc_expire_one_request()) Skipped 2 previous similar messages Lustre: lustre-OST0001: Export ffff8802cc66b7e8 already connecting from 0@lo Lustre: lustre-OST0001: Export ffff8802cc66b7e8 already connecting from 0@lo Lustre: 3988:0:(client.c:2346:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1741146955/real 1741146955] req@ffff8802cbf50f40 x1825723451610624/t0(0) o22->lustre-OST0001-osc-ffff8802e8bd5d28@0@lo:28/4 lens 440/432 e 2 to 1 dl 1741147071 ref 1 fl Rpc:XQr/200/ffffffff rc 0/-1 job:'fallocate.0' uid:0 gid:0 Lustre: 3988:0:(client.c:2346:ptlrpc_expire_one_request()) Skipped 2 previous similar messages Lustre: lustre-OST0001: Export ffff8802cc66b7e8 already connecting from 0@lo Lustre: lustre-OST0001: haven't heard from client a7e32c7a-bae3-4b7f-ad03-c9e7d3f17000 (at <unknown>) in 34 seconds. I think it's dead, and I am evicting it. exp ffff8802cc66b7e8, cur 1741147076 expire 1741147046 last 1741147042 Lustre: lustre-OST0001: Export ffff8802d90a2548 already connecting from 0@lo Lustre: lustre-OST0001: Export ffff8802d90a2548 already connecting from 0@lo Lustre: Skipped 3 previous similar messages Lustre: lustre-OST0001: Export ffff8802d90a2548 already connecting from 0@lo Lustre: Skipped 6 previous similar messages Lustre: lustre-OST0001: Export ffff8802d90a2548 already connecting from 0@lo Lustre: Skipped 12 previous similar messages LustreError: 1300:0:(ldlm_request.c:104:ldlm_expired_completion_wait()) ### lock timed out (enqueued at 1741146954, 300s ago); not entering recovery in server code, just going back to sleep ns: mdt-lustre-MDT0000_UUID lock: ffff8802ad65cb40/0xd20e60de428a7 lrc: 3/0,1 mode: --/PW res: [0x200000403:0x2:0x0].0x32 bits 0x2/0x0 rrc: 4 type: IBT flags: 0x40210400000020 pid: 1300 initiator: MDT0 LustreError: dumping log to /tmp/lustre-log.1741147254.1300 LustreError: 12676:0:(ldlm_request.c:104:ldlm_expired_completion_wait()) ### lock timed out (enqueued at 1741146955, 300s ago); not entering recovery in server code, just going back to sleep ns: mdt-lustre-MDT0000_UUID lock: ffff880094c00f40/0xd20e60de4566d lrc: 3/1,0 mode: --/PR res: [0x200000403:0x2:0x0].0x32 bits 0x2/0x0 rrc: 4 type: IBT flags: 0x40210000000000 pid: 12676 initiator: MDT0 LustreError: 12676:0:(ldlm_request.c:104:ldlm_expired_completion_wait()) Skipped 7 previous similar messages LustreError: 22408:0:(ldlm_request.c:104:ldlm_expired_completion_wait()) ### lock timed out (enqueued at 1741146957, 300s ago); not entering recovery in server code, just going back to sleep ns: filter-lustre-OST0001_UUID lock: ffff880094565a40/0xd20e60de49948 lrc: 3/0,1 mode: --/PW res: [0x300000401:0x472:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x40010080000000 nid: local remote: 0x0 expref: -99 pid: 22408 timeout: 0 lvb_type: 0 LustreError: 22408:0:(ldlm_request.c:104:ldlm_expired_completion_wait()) Skipped 9 previous similar messages Lustre: ll_ost_io00_024: service thread pid 24531 was inactive for 222.188 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 5 previous similar messages ptlrpc_watchdog_fire: 80 callbacks suppressed Lustre: ll_ost00_043: service thread pid 24275 was inactive for 250.360 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Lustre: Skipped 2 previous similar messages Pid: 24275, comm: ll_ost00_043 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] wait_transaction_locked+0x93/0xe0 [jbd2] [<0>] add_transaction_credits+0x270/0x310 [jbd2] [<0>] start_this_handle+0x1cc/0x460 [jbd2] [<0>] jbd2__journal_start+0xf3/0x200 [jbd2] [<0>] __ldiskfs_journal_start_sb+0x6d/0x100 [ldiskfs] [<0>] osd_trans_start+0x279/0x6b0 [osd_ldiskfs] [<0>] tgt_client_data_update+0x30d/0x930 [ptlrpc] [<0>] tgt_client_new+0x481/0x920 [ptlrpc] [<0>] ofd_obd_connect+0x425/0x550 [ofd] [<0>] target_handle_connect+0xe76/0x3210 [ptlrpc] [<0>] tgt_request_handle+0x6ba/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: lustre-OST0001: Export ffff8802d90a2548 already connecting from 0@lo Lustre: Skipped 25 previous similar messages Lustre: ll_ost00_042: service thread pid 24102 was inactive for 322.812 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 24102, comm: ll_ost00_042 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] wait_transaction_locked+0x93/0xe0 [jbd2] [<0>] add_transaction_credits+0x270/0x310 [jbd2] [<0>] start_this_handle+0x1cc/0x460 [jbd2] [<0>] jbd2__journal_start+0xf3/0x200 [jbd2] [<0>] __ldiskfs_journal_start_sb+0x6d/0x100 [ldiskfs] [<0>] osd_trans_start+0x279/0x6b0 [osd_ldiskfs] [<0>] tgt_client_data_update+0x30d/0x930 [ptlrpc] [<0>] tgt_client_new+0x481/0x920 [ptlrpc] [<0>] ofd_obd_connect+0x425/0x550 [ofd] [<0>] target_handle_connect+0xe76/0x3210 [ptlrpc] [<0>] tgt_request_handle+0x6ba/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: lustre-OST0003-osc-ffff8802e6f3dd28: disconnect after 20s idle Lustre: 24532:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0000: opcode 7: before 515 < left 618, rollback = 7 Lustre: 24532:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 365 previous similar messages Lustre: 24532:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 24532:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 1895 previous similar messages Lustre: 24532:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/1, xattr_set: 2/15/0 Lustre: 24532:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 1895 previous similar messages Lustre: 24532:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 24532:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 1893 previous similar messages Lustre: 24532:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 24532:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 1895 previous similar messages Lustre: 24532:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 24532:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 1895 previous similar messages Lustre: 23023:0:(service.c:1438:ptlrpc_at_send_early_reply()) @@@ Could not add any time (5/5), not sending early reply req@ffff88028f865a40 x1825723450127232/t0(0) o22->9f96f94b-79e7-4e34-983d-2dc9620efa89@0@lo:513/0 lens 440/432 e 24 to 0 dl 1741147538 ref 2 fl Interpret:/200/0 rc 0/0 job:'fallocate.0' uid:0 gid:0 Lustre: 23023:0:(service.c:1438:ptlrpc_at_send_early_reply()) @@@ Could not add any time (5/5), not sending early reply req@ffff8802c21f6940 x1825723450285056/t0(0) o6->lustre-MDT0000-mdtlov_UUID@0@lo:514/0 lens 544/432 e 24 to 0 dl 1741147539 ref 2 fl Interpret:/200/0 rc 0/0 job:'osp-syn-1-0.0' uid:0 gid:0 Lustre: 23023:0:(service.c:1438:ptlrpc_at_send_early_reply()) Skipped 4 previous similar messages Lustre: 23023:0:(service.c:1438:ptlrpc_at_send_early_reply()) @@@ Could not add any time (5/5), not sending early reply req@ffff88029bdc0a40 x1825723450377344/t0(0) o2->9f96f94b-79e7-4e34-983d-2dc9620efa89@0@lo:515/0 lens 440/432 e 24 to 0 dl 1741147540 ref 2 fl Interpret:/200/0 rc 0/0 job:'chmod.0' uid:0 gid:0 Lustre: 23023:0:(service.c:1438:ptlrpc_at_send_early_reply()) Skipped 7 previous similar messages Lustre: 24532:0:(service.c:1438:ptlrpc_at_send_early_reply()) @@@ Could not add any time (5/5), not sending early reply req@ffff8802f17f0040 x1825723450550528/t0(0) o10->9f96f94b-79e7-4e34-983d-2dc9620efa89@0@lo:518/0 lens 440/432 e 24 to 0 dl 1741147543 ref 2 fl Interpret:/200/0 rc 0/0 job:'cp.0' uid:0 gid:0 Lustre: 24532:0:(service.c:1438:ptlrpc_at_send_early_reply()) Skipped 5 previous similar messages LustreError: 11086:0:(osc_cache.c:922:osc_extent_wait()) extent ffff8802c3cf94a8@{[1616 -> 2047/2047], [3|0|+|rpc|wihY|ffff88009165fb08], [1794048|432|+|-|ffff8802aca770c0|1024|ffff88031b0249d0]} lustre-OST0001-osc-ffff8802e8bd5d28: wait ext to 0 timedout, recovery in progress? LustreError: 11086:0:(osc_cache.c:922:osc_extent_wait()) ### extent: ffff8802c3cf94a8 ns: lustre-OST0001-osc-ffff8802e8bd5d28 lock: ffff8802aca770c0/0xd20e60de21f82 lrc: 7/0,0 mode: PW/PW res: [0x300000401:0x472:0x0].0x0 rrc: 3 type: EXT [6619136->18446744073709551615] (req 6619136->6684671) gid 0 flags: 0x800429400020000 nid: local remote: 0xd20e60de21f89 expref: -99 pid: 17454 timeout: 0 lvb_type: 1 Lustre: 3983:0:(client.c:2346:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1741146938/real 1741146938] req@ffff8802a166cb40 x1825723450127232/t0(0) o22->lustre-OST0001-osc-ffff8802e6f3dd28@0@lo:28/4 lens 440/432 e 24 to 1 dl 1741147543 ref 1 fl Rpc:XQr/200/ffffffff rc 0/-1 job:'fallocate.0' uid:0 gid:0 Lustre: 3983:0:(client.c:2346:ptlrpc_expire_one_request()) Skipped 1 previous similar message Lustre: lustre-OST0001-osc-ffff8802e6f3dd28: Connection to lustre-OST0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: lustre-OST0001: Client lustre-MDT0001-mdtlov_UUID (at 0@lo) reconnecting Lustre: lustre-OST0001-osc-MDT0001: Connection restored to 192.168.123.22@tcp (at 0@lo) Lustre: lustre-OST0001-osc-MDT0002: Connection to lustre-OST0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 1 previous similar message Lustre: lustre-OST0001: Client lustre-MDT0002-mdtlov_UUID (at 0@lo) reconnecting Lustre: Skipped 1 previous similar message Lustre: lustre-OST0001-osc-MDT0002: Connection restored to 192.168.123.22@tcp (at 0@lo) Lustre: Skipped 1 previous similar message Lustre: 19170:0:(service.c:1438:ptlrpc_at_send_early_reply()) @@@ Could not add any time (5/5), not sending early reply req@ffff8800af308a40 x1825723451340160/t0(0) o101->a7e32c7a-bae3-4b7f-ad03-c9e7d3f17000@0@lo:528/0 lens 1072/3912 e 24 to 0 dl 1741147553 ref 2 fl Interpret:/200/0 rc 0/0 job:'fallocate.0' uid:0 gid:0 Lustre: 19170:0:(service.c:1438:ptlrpc_at_send_early_reply()) Skipped 2 previous similar messages LustreError: 18987:0:(osc_cache.c:922:osc_extent_wait()) extent ffff8802e5700638@{[1024 -> 1615/2047], [4|0|+|rpc|wiuY|ffff8802adb54e08], [2449408|592|+|+|ffff8802c126d680|1024|ffff88031f2c49d0]} lustre-OST0001-osc-ffff8802e6f3dd28: wait ext to 0 timedout, recovery in progress? LustreError: 18987:0:(osc_cache.c:922:osc_extent_wait()) ### extent: ffff8802e5700638 ns: lustre-OST0001-osc-ffff8802e6f3dd28 lock: ffff8802c126d680/0xd20e60ddccf67 lrc: 3/0,0 mode: PW/PW res: [0x300000400:0x45a:0x0].0x0 rrc: 2 type: EXT [0->18446744073709551615] (req 0->65535) gid 0 flags: 0x800029400000000 nid: local remote: 0xd20e60ddccf6e expref: -99 pid: 7890 timeout: 0 lvb_type: 1 Lustre: lustre-MDT0000-mdc-ffff8802e8bd5d28: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 1 previous similar message Lustre: lustre-MDT0000: Client a7e32c7a-bae3-4b7f-ad03-c9e7d3f17000 (at 0@lo) reconnecting Lustre: Skipped 1 previous similar message Lustre: lustre-MDT0000-mdc-ffff8802e8bd5d28: Connection restored to 192.168.123.22@tcp (at 0@lo) Lustre: Skipped 1 previous similar message LustreError: 14145:0:(osp_precreate.c:1735:osp_object_truncate()) can't punch object: -107 LustreError: lustre-MDT0000-mdc-ffff8802e8bd5d28: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: 1300:0:(service.c:2350:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (600/5s); client may timeout req@ffff88031ee56940 x1825723451507584/t4295150435(0) o36->9f96f94b-79e7-4e34-983d-2dc9620efa89@0@lo:529/0 lens 504/456 e 24 to 0 dl 1741147554 ref 1 fl Complete:/200/0 rc 0/0 job:'mrename.0' uid:0 gid:0 Lustre: mdt_io00_008: service thread pid 1300 completed after 604.124s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_031: service thread pid 7271 completed after 602.369s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_030: service thread pid 7264 completed after 604.021s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_036: service thread pid 13625 completed after 604.021s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_032: service thread pid 7286 completed after 604.115s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_005: service thread pid 16904 completed after 604.022s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_026: service thread pid 14145 completed after 605.080s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_007: service thread pid 16938 completed after 602.442s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_022: service thread pid 30331 completed after 603.892s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_009: service thread pid 16970 completed after 602.492s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_039: service thread pid 14991 completed after 602.457s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_020: service thread pid 20275 completed after 602.484s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_014: service thread pid 19153 completed after 604.024s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_003: service thread pid 13190 completed after 604.029s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_out00_002: service thread pid 12676 completed after 603.246s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_033: service thread pid 7365 completed after 602.512s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_035: service thread pid 25658 completed after 602.443s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_021: service thread pid 25484 completed after 602.464s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 1300:0:(mdt_reint.c:2513:mdt_reint_migrate()) lustre-MDT0001: migrate [0x240000403:0x1:0x0]/7 failed: rc = -114 LustreError: 1300:0:(mdt_reint.c:2513:mdt_reint_migrate()) Skipped 67 previous similar messages LustreError: 24499:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e8bd5d28: inode [0x240000406:0x21d6:0x0] mdc close failed: rc = -116 LustreError: 24499:0:(file.c:247:ll_close_inode_openhandle()) Skipped 50 previous similar messages LustreError: 412:0:(osc_cache.c:922:osc_extent_wait()) extent ffff8802c3cfa738@{[2000 -> 2047/2047], [3|0|+|rpc|wiuY|ffff88028fbe8308], [221184|48|+|-|ffff88009c2d07c0|1024|ffff88031ab53760]} lustre-OST0001-osc-ffff8802e8bd5d28: wait ext to 0 timedout, recovery in progress? LustreError: 412:0:(osc_cache.c:922:osc_extent_wait()) Skipped 1 previous similar message LustreError: 412:0:(osc_cache.c:922:osc_extent_wait()) ### extent: ffff8802c3cfa738 ns: lustre-OST0001-osc-ffff8802e8bd5d28 lock: ffff88009c2d07c0/0xd20e60de21e86 lrc: 4/0,0 mode: PW/PW res: [0x300000400:0x46a:0x0].0x0 rrc: 2 type: EXT [0->18446744073709551615] (req 0->67108863) gid 0 flags: 0x800020000000000 nid: local remote: 0xd20e60de21e8d expref: -99 pid: 17474 timeout: 0 lvb_type: 1 LustreError: 412:0:(osc_cache.c:922:osc_extent_wait()) Skipped 1 previous similar message Lustre: 20944:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 509 < left 894, rollback = 2 Lustre: 20944:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 1924 previous similar messages Lustre: lustre-OST0001: Export ffff8802d90a2548 already connecting from 0@lo Lustre: Skipped 53 previous similar messages LustreError: 20822:0:(statahead.c:2446:start_statahead_thread()) lustre: unsupported statahead pattern 0X0. Lustre: lustre-MDT0000-mdc-ffff8802e8bd5d28: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 1 previous similar message LustreError: 25173:0:(ldlm_request.c:123:ldlm_expired_completion_wait()) ### lock timed out (enqueued at 1741147560, 300s ago), entering recovery for lustre-MDT0000_UUID@192.168.123.22@tcp ns: lustre-MDT0000-mdc-ffff8802e8bd5d28 lock: ffff8802cb589300/0xd20e60de5d35a lrc: 4/1,0 mode: --/CR res: [0x200000405:0x3614:0x0].0x0 bits 0x8/0x0 rrc: 2 type: IBT gid 0 flags: 0x0 nid: local remote: 0xd20e60de5d361 expref: -99 pid: 25173 timeout: 0 lvb_type: 3 Lustre: lustre-MDT0000: Client a7e32c7a-bae3-4b7f-ad03-c9e7d3f17000 (at 0@lo) reconnecting Lustre: Skipped 1 previous similar message Lustre: lustre-MDT0000-mdc-ffff8802e8bd5d28: Connection restored to 192.168.123.22@tcp (at 0@lo) Lustre: Skipped 1 previous similar message LustreError: 19079:0:(ldlm_request.c:104:ldlm_expired_completion_wait()) ### lock timed out (enqueued at 1741147566, 300s ago); not entering recovery in server code, just going back to sleep ns: mdt-lustre-MDT0000_UUID lock: ffff8802a2926940/0xd20e60de71602 lrc: 3/0,1 mode: --/PW res: [0x200000403:0x1:0x0].0x34 bits 0x2/0x0 rrc: 3 type: IBT flags: 0x40210000000000 pid: 19079 initiator: MDT0 LustreError: 19185:0:(ldlm_request.c:104:ldlm_expired_completion_wait()) ### lock timed out (enqueued at 1741147576, 300s ago); not entering recovery in server code, just going back to sleep ns: mdt-lustre-MDT0000_UUID lock: ffff8802cdacd680/0xd20e60de95949 lrc: 3/1,0 mode: --/PR res: [0x200000403:0x1:0x0].0x0 bits 0x13/0x0 rrc: 14 type: IBT flags: 0x40210000000000 pid: 19185 initiator: MDT0 | Link to test |
racer test 1: racer on clients: centos-120.localnet DURATION=2700 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 PGD 2b8335067 PUD 2d4990067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) lustre(OE) osp(OE) ofd(OE) lod(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) jbd2 mbcache lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) crc32_generic lnet(OE) dm_flakey dm_mod libcfs(OE) crc_t10dif crct10dif_generic crct10dif_common pcspkr virtio_balloon virtio_console i2c_piix4 ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm ata_piix drm_panel_orientation_quirks libata virtio_blk serio_raw i2c_core floppy CPU: 1 PID: 10747 Comm: ll_sa_10254 Kdump: loaded Tainted: P OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.16.0-3.module_el8.7.0+1218+f626c2ff 04/01/2014 task: ffff8802c67e5c40 ti: ffff8802c549c000 task.ti: ffff8802c549c000 RIP: 0010:[<ffffffff813f502c>] [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 RSP: 0018:ffff8802c549fd80 EFLAGS: 00010282 RAX: ffff8802b3bc1488 RBX: 0000000000000008 RCX: 0000000100260023 RDX: 0000000000000026 RSI: ffff8802b3bc17b8 RDI: 0000000000000008 RBP: ffff8802c549fd90 R08: ffff8802dc6fae68 R09: 0000000000000000 R10: ffff8802dc6f8d48 R11: ffff8802dc6fae68 R12: 0000000000000000 R13: ffff8802d0185c40 R14: ffff8802b3bc17b8 R15: ffff8802dc6fae68 FS: 0000000000000000(0000) GS:ffff880331a40000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 00000002db01e000 CR4: 00000000000007e0 Call Trace: [<ffffffffa1113491>] ll_statahead_thread+0xde1/0x2290 [lustre] [<ffffffff810df6e2>] ? pick_next_task_fair+0x702/0x8a0 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e329e>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffff810c834d>] ? finish_task_switch+0x5d/0x1b0 [<ffffffffa11126b0>] ? ll_statahead_interpret+0x4a0/0x4a0 [lustre] [<ffffffff810ba114>] kthread+0xe4/0xf0 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 | Lustre: 11576:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 509 < left 1161, rollback = 2 Lustre: 11576:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 2/8/3, destroy: 1/4/0 Lustre: 11576:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 2/2/0, xattr_set: 14/1161/0 Lustre: 11576:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 1/10/0, punch: 0/0/0, quota 1/3/0 Lustre: 11576:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 12/194/3, delete: 3/6/0 Lustre: 11576:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 7/7/0, ref_del: 3/3/0 LustreError: 11575:0:(mdt_reint.c:2513:mdt_reint_migrate()) lustre-MDT0000: migrate [0x200000403:0x2:0x0]/6 failed: rc = -114 LustreError: 18476:0:(mdt_reint.c:2513:mdt_reint_migrate()) lustre-MDT0001: migrate [0x240000403:0x1:0x0]/10 failed: rc = -114 Lustre: 18476:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 512 < left 610, rollback = 2 Lustre: 18476:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 20 previous similar messages Lustre: 18476:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 3/12/1, destroy: 0/0/0 Lustre: 18476:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 20 previous similar messages Lustre: 18476:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 2/2/0, xattr_set: 8/610/0 Lustre: 18476:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 20 previous similar messages Lustre: 18476:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 5/55/0, punch: 0/0/0, quota 1/3/0 Lustre: 18476:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 20 previous similar messages Lustre: 18476:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 12/275/2, delete: 0/0/0 Lustre: 18476:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 20 previous similar messages Lustre: 18476:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 6/6/0, ref_del: 0/0/0 Lustre: 18476:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 20 previous similar messages LustreError: 11575:0:(mdt_reint.c:2513:mdt_reint_migrate()) lustre-MDT0002: migrate [0x280000403:0x1:0x0]/10 failed: rc = -114 LustreError: 17558:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e674ca88: inode [0x280000403:0xe:0x0] mdc close failed: rc = -116 Lustre: 18299:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0000: opcode 7: before 516 < left 618, rollback = 7 Lustre: 18299:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 18299:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 11 previous similar messages Lustre: 18299:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 18299:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 11 previous similar messages Lustre: 18299:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 18299:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 11 previous similar messages Lustre: 18299:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 18299:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 11 previous similar messages Lustre: 18299:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 18299:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 11 previous similar messages Lustre: 18889:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 512 < left 788, rollback = 2 Lustre: 18889:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 11 previous similar messages LustreError: 18809:0:(mdt_reint.c:2513:mdt_reint_migrate()) lustre-MDT0000: migrate [0x200000403:0x1:0x0]/10 failed: rc = -114 LustreError: 18809:0:(mdt_reint.c:2513:mdt_reint_migrate()) Skipped 4 previous similar messages LustreError: 19960:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e50a37e8: inode [0x200000403:0x4:0x0] mdc close failed: rc = -116 LustreError: 19689:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e674ca88: inode [0x200000403:0xf:0x0] mdc close failed: rc = -116 Lustre: 18746:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 509 < left 788, rollback = 2 Lustre: 18746:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 19 previous similar messages Lustre: 18746:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 3/12/2, destroy: 0/0/0 Lustre: 18746:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 21 previous similar messages Lustre: 18746:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 2/2/0, xattr_set: 10/788/0 Lustre: 18746:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 21 previous similar messages Lustre: 18746:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 5/55/0, punch: 0/0/0, quota 1/3/0 Lustre: 18746:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 21 previous similar messages Lustre: 18746:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 11/239/4, delete: 0/0/0 Lustre: 18746:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 21 previous similar messages Lustre: 18746:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 5/5/0, ref_del: 0/0/0 Lustre: 18746:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 21 previous similar messages LustreError: 18476:0:(mdt_reint.c:2513:mdt_reint_migrate()) lustre-MDT0000: migrate [0x200000403:0x1:0x0]/16 failed: rc = -114 LustreError: 18476:0:(mdt_reint.c:2513:mdt_reint_migrate()) Skipped 10 previous similar messages LustreError: 11573:0:(mdt_reint.c:2513:mdt_reint_migrate()) lustre-MDT0001: migrate [0x240000403:0x1:0x0]/7 failed: rc = -114 LustreError: 11573:0:(mdt_reint.c:2513:mdt_reint_migrate()) Skipped 9 previous similar messages 15[22443]: segfault at 8 ip 00007fab370957e8 sp 00007fff6dc440b0 error 4 in ld-2.17.so[7fab3708a000+22000] Lustre: 18753:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 505 < left 966, rollback = 2 Lustre: 18753:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 19 previous similar messages Lustre: 18753:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 3/12/6, destroy: 0/0/0 Lustre: 18753:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 19 previous similar messages Lustre: 18753:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 2/2/0, xattr_set: 12/966/0 Lustre: 18753:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 19 previous similar messages Lustre: 18753:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 5/55/0, punch: 0/0/0, quota 1/3/0 Lustre: 18753:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 19 previous similar messages Lustre: 18753:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 10/203/4, delete: 0/0/0 Lustre: 18753:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 19 previous similar messages Lustre: 18753:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 4/4/0, ref_del: 0/0/0 Lustre: 18753:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 19 previous similar messages LustreError: 23218:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e674ca88: inode [0x280000404:0x76:0x0] mdc close failed: rc = -116 Lustre: 17356:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0xe8:0x0] with magic=0xbd60bd0 Lustre: 17211:0:(mdt_recovery.c:128:mdt_req_from_lrd()) @@@ restoring transno req@ffff88009faa6e40 x1825081055971072/t4294969325(0) o101->531edc72-75a4-4798-aaaa-2f677493944f@0@lo:568/0 lens 376/816 e 0 to 0 dl 1740533023 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 Lustre: 17808:0:(mdt_recovery.c:128:mdt_req_from_lrd()) @@@ restoring transno req@ffff8802d23e3240 x1825081056662656/t4294970663(0) o101->49f0e8c4-2261-4e5b-8cd9-0ae1d2c559aa@0@lo:570/0 lens 376/840 e 0 to 0 dl 1740533025 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 Lustre: 19307:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0000: opcode 7: before 516 < left 618, rollback = 7 Lustre: 19307:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 1 previous similar message Lustre: 19307:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 19307:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 98 previous similar messages Lustre: 19307:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 19307:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 98 previous similar messages Lustre: 19307:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 19307:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 98 previous similar messages Lustre: 19307:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 19307:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 98 previous similar messages Lustre: 19307:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 19307:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 98 previous similar messages Lustre: 17704:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 516 < left 618, rollback = 7 Lustre: 17704:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 1 previous similar message 17[26827]: segfault at 8 ip 00007f6c565ed7e8 sp 00007ffc2e610430 error 4 in ld-2.17.so[7f6c565e2000+22000] LustreError: 18753:0:(mdt_reint.c:2513:mdt_reint_migrate()) lustre-MDT0002: migrate [0x280000403:0x1:0x0]/0 failed: rc = -114 LustreError: 18753:0:(mdt_reint.c:2513:mdt_reint_migrate()) Skipped 13 previous similar messages 13[27579]: segfault at 0 ip 0000000000403cf0 sp 00007ffe2c801618 error 6 in 13[400000+6000] Lustre: 18753:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 504 < left 983, rollback = 2 Lustre: 18753:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 98 previous similar messages LustreError: 15466:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0002: failed to get lu_attr of [0x280000404:0xa9:0x0]: rc = -2 LustreError: 27306:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e674ca88: inode [0x280000404:0xa9:0x0] mdc close failed: rc = -2 Lustre: 18299:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 516 < left 618, rollback = 7 Lustre: 18299:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 3 previous similar messages Lustre: dir [0x200000403:0x121:0x0] stripe 3 readdir failed: -2, directory is partially accessed! Lustre: 13432:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 618, rollback = 7 Lustre: 13432:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 1 previous similar message Lustre: 19918:0:(mdt_recovery.c:128:mdt_req_from_lrd()) @@@ restoring transno req@ffff88009a39d540 x1825081060748288/t4294971560(0) o101->49f0e8c4-2261-4e5b-8cd9-0ae1d2c559aa@0@lo:584/0 lens 376/840 e 0 to 0 dl 1740533039 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 Lustre: 19918:0:(mdt_recovery.c:128:mdt_req_from_lrd()) Skipped 1 previous similar message LustreError: 28022:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e50a37e8: inode [0x240000403:0xff:0x0] mdc close failed: rc = -116 LustreError: 28022:0:(file.c:247:ll_close_inode_openhandle()) Skipped 5 previous similar messages LustreError: 18522:0:(lustre_lmv.h:500:lmv_is_sane()) unknown layout LMV: magic=0xcd40cd0 count=4 index=3 hash=crush:0x82000003 version=1 migrate_offset=3 migrate_hash=fnv_1a_64:2 pool= Lustre: dir [0x200000403:0x1c1:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: 13432:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 13432:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 104 previous similar messages Lustre: 13432:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 13432:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 104 previous similar messages Lustre: 13432:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/78/0 Lustre: 13432:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 104 previous similar messages Lustre: 13432:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 13432:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 104 previous similar messages Lustre: 13432:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 13432:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 104 previous similar messages Lustre: dir [0x280000404:0x171:0x0] stripe 3 readdir failed: -2, directory is partially accessed! Lustre: Skipped 3 previous similar messages Lustre: 18476:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 503 < left 1322, rollback = 2 Lustre: 18476:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 181 previous similar messages LustreError: 1203:0:(mdd_dir.c:4619:mdd_migrate_cmd_check()) lustre-MDD0002: '16' migration was interrupted, run 'lfs migrate -m 0 -c 2 -H crush 16' to finish migration: rc = -1 LustreError: 1203:0:(mdd_dir.c:4619:mdd_migrate_cmd_check()) lustre-MDD0002: '16' migration was interrupted, run 'lfs migrate -m 0 -c 1 -H crush 16' to finish migration: rc = -1 Lustre: dir [0x200000403:0x1f3:0x0] stripe 1 readdir failed: -2, directory is partially accessed! LustreError: 32679:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e674ca88: inode [0x200000403:0x1e1:0x0] mdc close failed: rc = -116 LustreError: 32679:0:(file.c:247:ll_close_inode_openhandle()) Skipped 5 previous similar messages Lustre: 13432:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 552, rollback = 7 Lustre: 13432:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 19 previous similar messages LustreError: 11573:0:(mdd_dir.c:4619:mdd_migrate_cmd_check()) lustre-MDD0001: '16' migration was interrupted, run 'lfs migrate -m 0 -c 2 -H crush 16' to finish migration: rc = -1 Lustre: dir [0x200000403:0x247:0x0] stripe 3 readdir failed: -2, directory is partially accessed! Lustre: Skipped 1 previous similar message LustreError: 4284:0:(lcommon_cl.c:177:cl_file_inode_init()) lustre: failed to initialize cl_object [0x280000403:0x215:0x0]: rc = -5 LustreError: 4284:0:(llite_lib.c:3712:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 11[3871]: segfault at 8 ip 00007fb5185517e8 sp 00007ffe72a50a50 error 4 in ld-2.17.so[7fb518546000+22000] LustreError: 11573:0:(mdt_reint.c:2513:mdt_reint_migrate()) lustre-MDT0001: migrate [0x240000403:0x1:0x0]/15 failed: rc = -114 LustreError: 11573:0:(mdt_reint.c:2513:mdt_reint_migrate()) Skipped 52 previous similar messages LustreError: 26653:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000404:0xd8:0x0]: rc = -2 LustreError: 17973:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0002: failed to get lu_attr of [0x280000404:0x25:0x0]: rc = -2 LustreError: 11573:0:(mdd_dir.c:4619:mdd_migrate_cmd_check()) lustre-MDD0000: '18' migration was interrupted, run 'lfs migrate -m 1 -c 1 -H crush 18' to finish migration: rc = -1 Lustre: 26264:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 1/4/4, destroy: 0/0/0 Lustre: 26264:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 677 previous similar messages Lustre: 26264:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 7/639/0 Lustre: 26264:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 677 previous similar messages Lustre: 26264:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 5/21/0, punch: 0/0/0, quota 4/54/2 Lustre: 26264:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 677 previous similar messages Lustre: 26264:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 2/33/1, delete: 0/0/0 Lustre: 26264:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 677 previous similar messages Lustre: 26264:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 26264:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 677 previous similar messages Lustre: 13432:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 516 < left 552, rollback = 7 Lustre: 13432:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 3 previous similar messages 3[9726]: segfault at 8 ip 00007f010c4537e8 sp 00007fff98909a70 error 4 in ld-2.17.so[7f010c448000+22000] 2[9968]: segfault at 8 ip 00007f45bc4a07e8 sp 00007ffc1baeae40 error 4 in ld-2.17.so[7f45bc495000+22000] | Link to test |
racer test 1: racer on clients: oleg324-client.virtnet DURATION=3600 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffff813f40ac>] _atomic_dec_and_lock+0xc/0x70 PGD 948c2067 PUD a5778067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) osp(OE) ofd(OE) lod(OE) mdt(OE) mdd(OE) mgs(OE) lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) crc_t10dif crct10dif_generic rpcsec_gss_krb5 sb_edac edac_core iosf_mbi kvm_intel kvm irqbypass crc32_pclmul ghash_clmulni_intel squashfs aesni_intel lrw gf128mul glue_helper ablk_helper cryptd i2c_piix4 i2c_core pcspkr binfmt_misc ip_tables ext4 mbcache jbd2 ata_generic pata_acpi crct10dif_pclmul crct10dif_common crc32c_intel ata_piix serio_raw libata CPU: 0 PID: 7534 Comm: ll_sa_6020 Kdump: loaded Tainted: G OE ------------ 3.10.0-7.9-debug #1 Hardware name: QEMU Standard PC (i440FX + PIIX, 1996), BIOS 1.16.2-1.fc38 04/01/2014 task: ffff88012ba58000 ti: ffff88012fa20000 task.ti: ffff88012fa20000 RIP: 0010:[<ffffffff813f40ac>] [<ffffffff813f40ac>] _atomic_dec_and_lock+0xc/0x70 RSP: 0000:ffff88012fa23d80 EFLAGS: 00010282 RAX: ffff88008f5b18c8 RBX: 0000000000000008 RCX: ffffea0004ad9f80 RDX: 0000000000000006 RSI: ffff88008f5b1bf8 RDI: 0000000000000008 RBP: ffff88012fa23d90 R08: ffff88008b543c00 R09: 000000018010000f R10: 0000000000000001 R11: ffff88008b543c00 R12: 0000000000000000 R13: ffff88012d573330 R14: ffff88008f5b1bf8 R15: ffff88008b543c00 FS: 0000000000000000(0000) GS:ffff88013e200000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 CR2: 0000000000000008 CR3: 0000000091f78000 CR4: 0000000000160ff0 Call Trace: [<ffffffffa10d2491>] ll_statahead_thread+0xde1/0x2290 [lustre] [<ffffffff810df6e2>] ? pick_next_task_fair+0x702/0x8a0 [<ffffffff81410a79>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e8dce>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffff810c834d>] ? finish_task_switch+0x5d/0x1b0 [<ffffffffa10d16b0>] ? ll_statahead_interpret+0x4a0/0x4a0 [lustre] [<ffffffff810ba114>] kthread+0xe4/0xf0 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817f3e5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 | random: crng init done LustreError: 11976:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800b5c06800: inode [0x200000403:0x19:0x0] mdc close failed: rc = -116 LustreError: 12375:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800b5c06800: inode [0x240000403:0x1b:0x0] mdc close failed: rc = -116 LustreError: 12375:0:(file.c:247:ll_close_inode_openhandle()) Skipped 3 previous similar messages Lustre: dir [0x200000403:0x30:0x0] stripe 2 readdir failed: -2, directory is partially accessed! LustreError: 14137:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800b67fb800: inode [0x200000403:0x59:0x0] mdc close failed: rc = -116 LustreError: 14137:0:(file.c:247:ll_close_inode_openhandle()) Skipped 1 previous similar message Lustre: dir [0x200000402:0x17c:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 1 previous similar message LustreError: 16180:0:(lcommon_cl.c:177:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000403:0xa0:0x0]: rc = -5 LustreError: 16180:0:(llite_lib.c:3712:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 16231:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800b5c06800: inode [0x200000402:0x196:0x0] mdc close failed: rc = -116 LustreError: 16231:0:(file.c:247:ll_close_inode_openhandle()) Skipped 2 previous similar messages LustreError: 17005:0:(lcommon_cl.c:177:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000403:0xa0:0x0]: rc = -5 LustreError: 17005:0:(lcommon_cl.c:177:cl_file_inode_init()) Skipped 2 previous similar messages LustreError: 17005:0:(llite_lib.c:3712:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 17005:0:(llite_lib.c:3712:ll_prep_inode()) Skipped 2 previous similar messages LustreError: 10576:0:(lcommon_cl.c:177:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000402:0x195:0x0]: rc = -5 LustreError: 10576:0:(lcommon_cl.c:177:cl_file_inode_init()) Skipped 11 previous similar messages LustreError: 10576:0:(llite_lib.c:3712:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 10576:0:(llite_lib.c:3712:ll_prep_inode()) Skipped 11 previous similar messages Lustre: dir [0x240000403:0x170:0x0] stripe 2 readdir failed: -2, directory is partially accessed! Lustre: Skipped 1 previous similar message LustreError: 20393:0:(lcommon_cl.c:177:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000402:0x134:0x0]: rc = -5 LustreError: 20393:0:(lcommon_cl.c:177:cl_file_inode_init()) Skipped 5 previous similar messages LustreError: 20393:0:(llite_lib.c:3712:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 20393:0:(llite_lib.c:3712:ll_prep_inode()) Skipped 5 previous similar messages LustreError: 18966:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800b67fb800: inode [0x240000402:0x234:0x0] mdc close failed: rc = -116 LustreError: 18966:0:(file.c:247:ll_close_inode_openhandle()) Skipped 9 previous similar messages LustreError: 31:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep sleep [0x200000402:0x350:0x0] inode@0000000000000000: rc = -5 cp (23242) used greatest stack depth: 10576 bytes left LustreError: 10514:0:(lcommon_cl.c:177:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000403:0xfc:0x0]: rc = -5 LustreError: 10514:0:(lcommon_cl.c:177:cl_file_inode_init()) Skipped 20 previous similar messages LustreError: 10514:0:(llite_lib.c:3712:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 10514:0:(llite_lib.c:3712:ll_prep_inode()) Skipped 20 previous similar messages LustreError: 25305:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8800b67fb800: cannot apply new layout on [0x240000402:0x3f8:0x0] : rc = -5 LustreError: 25305:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x240000402:0x3f8:0x0] error -5. LustreError: 26536:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8800b5c06800: cannot apply new layout on [0x200000403:0x324:0x0] : rc = -5 LustreError: 26536:0:(lov_object.c:1341:lov_layout_change()) Skipped 7 previous similar messages LustreError: 29060:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8800b67fb800: cannot apply new layout on [0x240000402:0x3f8:0x0] : rc = -5 LustreError: 29060:0:(lov_object.c:1341:lov_layout_change()) Skipped 2 previous similar messages LustreError: lustre-MDT0000-mdc-ffff8800b5c06800: operation ldlm_enqueue to node 192.168.203.124@tcp failed: rc = -107 Lustre: lustre-MDT0000-mdc-ffff8800b5c06800: Connection to lustre-MDT0000 (at 192.168.203.124@tcp) was lost; in progress operations using this service will wait for recovery to complete LustreError: Skipped 7 previous similar messages LustreError: lustre-MDT0000-mdc-ffff8800b5c06800: This client was evicted by lustre-MDT0000; in progress operations using this service will fail. LustreError: 26320:0:(file.c:6007:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000403:0x36b:0x0] error: rc = -5 LustreError: 24805:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800b5c06800: inode [0x200000402:0x534:0x0] mdc close failed: rc = -108 LustreError: 24805:0:(file.c:247:ll_close_inode_openhandle()) Skipped 10 previous similar messages LustreError: 26655:0:(llite_lib.c:1997:ll_md_setattr()) md_setattr fails: rc = -108 LustreError: 24964:0:(mdc_request.c:1464:mdc_read_page()) lustre-MDT0000-mdc-ffff8800b5c06800: [0x200000400:0x1d:0x0] lock enqueue fails: rc = -108 Lustre: dir [0x200000402:0x4ea:0x0] stripe 0 readdir failed: -108, directory is partially accessed! LustreError: 25252:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x200000403:0x36b:0x0] error -108. LustreError: 25252:0:(vvp_io.c:1903:vvp_io_init()) Skipped 1 previous similar message LustreError: 29661:0:(ldlm_resource.c:983:ldlm_resource_complain()) lustre-MDT0000-mdc-ffff8800b5c06800: namespace resource [0x200000007:0x1:0x0].0x0 (ffff88009f1d2c00) refcount nonzero (1) after lock cleanup; forcing cleanup. Lustre: lustre-MDT0000-mdc-ffff8800b5c06800: Connection restored to (at 192.168.203.124@tcp) LustreError: 30426:0:(lcommon_cl.c:177:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000402:0x44d:0x0]: rc = -5 LustreError: 30426:0:(lcommon_cl.c:177:cl_file_inode_init()) Skipped 35 previous similar messages LustreError: 30426:0:(llite_lib.c:3712:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 30426:0:(llite_lib.c:3712:ll_prep_inode()) Skipped 35 previous similar messages 18[29560]: segfault at 8 ip 00007f9d68e407e8 sp 00007ffc8d9a78d0 error 4 in ld-2.17.so[7f9d68e35000+22000] LustreError: 69:0:(statahead.c:825:ll_statahead_interpret_work()) lustre: failed to prep 7 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 LustreError: 28355:0:(llite_nfs.c:426:ll_dir_get_parent_fid()) lustre: failure inode [0x240000402:0x446:0x0] get parent: rc = -2 LustreError: 29572:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8800b67fb800: cannot apply new layout on [0x240000402:0x134:0x0] : rc = -5 LustreError: 29572:0:(lov_object.c:1341:lov_layout_change()) Skipped 1 previous similar message LustreError: 29572:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x240000402:0x134:0x0] error -5. LustreError: 29572:0:(vvp_io.c:1903:vvp_io_init()) Skipped 1 previous similar message LustreError: 28542:0:(llite_lib.c:1846:ll_update_lsm_md()) lustre: [0x200000403:0x798:0x0] dir layout mismatch: LustreError: 28542:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=4 count=2 index=0 hash=crush:0x2000003 max_inherit=0 max_inherit_rr=0 version=2 migrate_offset=0 migrate_hash=invalid:0 pool= LustreError: 28542:0:(lustre_lmv.h:167:lmv_stripe_object_dump()) stripe[0] [0x200000400:0x2f:0x0] LustreError: 28542:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=1 count=3 index=0 hash=crush:0x82000003 max_inherit=0 max_inherit_rr=0 version=1 migrate_offset=2 migrate_hash=crush:2000003 pool= LustreError: 28730:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=4 count=2 index=0 hash=crush:0x2000003 max_inherit=0 max_inherit_rr=0 version=2 migrate_offset=0 migrate_hash=invalid:0 pool= LustreError: 28730:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=1 count=3 index=0 hash=crush:0x82000003 max_inherit=0 max_inherit_rr=0 version=1 migrate_offset=2 migrate_hash=crush:2000003 pool= 1[1244]: segfault at 8 ip 00007f47bdcfb7e8 sp 00007fff7fe8bf00 error 4 in ld-2.17.so[7f47bdcf0000+22000] LustreError: 4472:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8800b67fb800: cannot apply new layout on [0x240000402:0x134:0x0] : rc = -5 LustreError: 4472:0:(lov_object.c:1341:lov_layout_change()) Skipped 4 previous similar messages Lustre: dir [0x240000403:0xa5c:0x0] stripe 2 readdir failed: -2, directory is partially accessed! Lustre: Skipped 16 previous similar messages LustreError: 3701:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x200000403:0x906:0x0] error -5. LustreError: 11097:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800b5c06800: inode [0x200000403:0xc65:0x0] mdc close failed: rc = -116 LustreError: 11097:0:(file.c:247:ll_close_inode_openhandle()) Skipped 32 previous similar messages 1[11228]: segfault at 8 ip 00007ff21659b7e8 sp 00007ffee1584b60 error 4 in ld-2.17.so[7ff216590000+22000] LustreError: 11818:0:(lcommon_cl.c:177:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000403:0xb78:0x0]: rc = -5 LustreError: 11818:0:(lcommon_cl.c:177:cl_file_inode_init()) Skipped 127 previous similar messages LustreError: 11818:0:(llite_lib.c:3712:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 11818:0:(llite_lib.c:3712:ll_prep_inode()) Skipped 127 previous similar messages LustreError: 11837:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8800b5c06800: cannot apply new layout on [0x200000403:0x906:0x0] : rc = -5 LustreError: 11837:0:(lov_object.c:1341:lov_layout_change()) Skipped 34 previous similar messages 18[19010]: segfault at 8 ip 00007f2cb280d7e8 sp 00007ffcac0bb850 error 4 in ld-2.17.so[7f2cb2802000+22000] LustreError: 19582:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x240000402:0x75f:0x0] error -5. Lustre: dir [0x200000404:0x846:0x0] stripe 2 readdir failed: -2, directory is partially accessed! 16[25732]: segfault at 8 ip 00007f2d75f127e8 sp 00007ffd021e2ed0 error 4 in ld-2.17.so[7f2d75f07000+22000] LustreError: 26468:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8800b67fb800: cannot apply new layout on [0x240000402:0x1035:0x0] : rc = -5 LustreError: 26468:0:(lov_object.c:1341:lov_layout_change()) Skipped 10 previous similar messages LustreError: 28688:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x240000402:0x1035:0x0] error -5. LustreError: 28688:0:(vvp_io.c:1903:vvp_io_init()) Skipped 2 previous similar messages 5[30309]: segfault at 8 ip 00007f445a3e37e8 sp 00007ffd5a0ed7d0 error 4 in ld-2.17.so[7f445a3d8000+22000] LustreError: 31467:0:(llite_lib.c:1846:ll_update_lsm_md()) lustre: [0x240000402:0x1550:0x0] dir layout mismatch: LustreError: 31467:0:(llite_lib.c:1846:ll_update_lsm_md()) Skipped 1 previous similar message LustreError: 31467:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=2 count=1 index=1 hash=crush:0x2000003 max_inherit=0 max_inherit_rr=0 version=2 migrate_offset=0 migrate_hash=invalid:0 pool= LustreError: 31467:0:(lustre_lmv.h:167:lmv_stripe_object_dump()) stripe[0] [0x240000400:0x6d:0x0] LustreError: 31467:0:(lustre_lmv.h:167:lmv_stripe_object_dump()) Skipped 9 previous similar messages LustreError: 31467:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=1 count=2 index=1 hash=crush:0x82000003 max_inherit=0 max_inherit_rr=0 version=1 migrate_offset=1 migrate_hash=fnv_1a_64:2 pool= LustreError: 28236:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=5 count=2 index=1 hash=crush:0x2000003 max_inherit=0 max_inherit_rr=0 version=2 migrate_offset=0 migrate_hash=invalid:0 pool= LustreError: 28236:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=1 count=4 index=1 hash=crush:0x82000003 max_inherit=0 max_inherit_rr=0 version=1 migrate_offset=2 migrate_hash=crush:2000003 pool= 15[1257]: segfault at 8 ip 00007f5c4587a7e8 sp 00007ffcb09766d0 error 4 in ld-2.17.so[7f5c4586f000+22000] 15[1312]: segfault at 8 ip 00007fbee1cb67e8 sp 00007fff2db3fb30 error 4 in ld-2.17.so[7fbee1cab000+22000] Lustre: dir [0x200000403:0x150f:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 11 previous similar messages LustreError: 32609:0:(llite_nfs.c:426:ll_dir_get_parent_fid()) lustre: failure inode [0x200000403:0x1357:0x0] get parent: rc = -2 LustreError: 32609:0:(llite_nfs.c:426:ll_dir_get_parent_fid()) Skipped 2 previous similar messages 9[4246]: segfault at 8 ip 00007f79530a07e8 sp 00007ffd13441c10 error 4 in ld-2.17.so[7f7953095000+22000] LustreError: 6428:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800b67fb800: inode [0x200000404:0xe21:0x0] mdc close failed: rc = -116 LustreError: 6428:0:(file.c:247:ll_close_inode_openhandle()) Skipped 57 previous similar messages LustreError: 5843:0:(lcommon_cl.c:177:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000403:0xb51:0x0]: rc = -5 LustreError: 5843:0:(lcommon_cl.c:177:cl_file_inode_init()) Skipped 366 previous similar messages LustreError: 5843:0:(llite_lib.c:3712:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 5843:0:(llite_lib.c:3712:ll_prep_inode()) Skipped 366 previous similar messages | Link to test |
racer test 1: racer on clients: oleg254-client.virtnet DURATION=300 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffff813f40ac>] _atomic_dec_and_lock+0xc/0x70 PGD 12bf5b067 PUD 13130a067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) osp(OE) ofd(OE) lod(OE) mdt(OE) mdd(OE) mgs(OE) lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) crc_t10dif crct10dif_generic rpcsec_gss_krb5 sb_edac edac_core iosf_mbi kvm_intel kvm irqbypass crc32_pclmul ghash_clmulni_intel aesni_intel squashfs lrw gf128mul glue_helper ablk_helper cryptd i2c_piix4 pcspkr i2c_core binfmt_misc ip_tables ext4 mbcache jbd2 ata_generic pata_acpi crct10dif_pclmul crct10dif_common crc32c_intel ata_piix serio_raw libata CPU: 2 PID: 27203 Comm: ll_sa_27166 Kdump: loaded Tainted: G OE ------------ 3.10.0-7.9-debug #1 Hardware name: QEMU Standard PC (i440FX + PIIX, 1996), BIOS 1.16.2-1.fc38 04/01/2014 task: ffff8800a3416660 ti: ffff88009194c000 task.ti: ffff88009194c000 RIP: 0010:[<ffffffff813f40ac>] [<ffffffff813f40ac>] _atomic_dec_and_lock+0xc/0x70 RSP: 0018:ffff88009194fd80 EFLAGS: 00010282 RAX: ffff8800af3868c8 RBX: 0000000000000008 RCX: ffffea0002727980 RDX: 0000000000000005 RSI: ffff8800af386bf8 RDI: 0000000000000008 RBP: ffff88009194fd90 R08: ffff88013008d000 R09: 000000018010000f R10: 0000000000000001 R11: ffff88013008d000 R12: 0000000000000000 R13: ffff8800a68ae660 R14: ffff8800af386bf8 R15: ffff88013008d000 FS: 0000000000000000(0000) GS:ffff88013e300000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 CR2: 0000000000000008 CR3: 000000012e614000 CR4: 0000000000160fe0 Call Trace: [<ffffffffa10f0691>] ll_statahead_thread+0xde1/0x2290 [lustre] [<ffffffff810df6e2>] ? pick_next_task_fair+0x702/0x8a0 [<ffffffff81410a79>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e8dce>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffff810c834d>] ? finish_task_switch+0x5d/0x1b0 [<ffffffffa10ef8b0>] ? ll_statahead_interpret+0x4a0/0x4a0 [lustre] [<ffffffff810ba114>] kthread+0xe4/0xf0 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817f3e5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 | LustreError: 11398:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800b6089800: inode [0x240000402:0x4:0x0] mdc close failed: rc = -116 LustreError: 12254:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800b3e11800: inode [0x200000403:0x2e:0x0] mdc close failed: rc = -116 LustreError: 12254:0:(file.c:247:ll_close_inode_openhandle()) Skipped 1 previous similar message random: crng init done LustreError: 12631:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800b3e11800: inode [0x240000403:0x40:0x0] mdc close failed: rc = -116 LustreError: 12631:0:(file.c:247:ll_close_inode_openhandle()) Skipped 1 previous similar message LustreError: 14270:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800b6089800: inode [0x200000403:0xf2:0x0] mdc close failed: rc = -116 LustreError: 14270:0:(file.c:247:ll_close_inode_openhandle()) Skipped 1 previous similar message Lustre: dir [0x240000402:0x8b:0x0] stripe 2 readdir failed: -2, directory is partially accessed! LustreError: 16629:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800b3e11800: inode [0x240000402:0x112:0x0] mdc close failed: rc = -116 LustreError: 16629:0:(file.c:247:ll_close_inode_openhandle()) Skipped 5 previous similar messages 3[17022]: segfault at 8 ip 00007ff3df3637e8 sp 00007ffe65dba910 error 4 in ld-2.17.so[7ff3df358000+22000] Lustre: dir [0x200000402:0x158:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 1 previous similar message LustreError: 17164:0:(lcommon_cl.c:177:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000403:0x182:0x0]: rc = -5 LustreError: 17164:0:(llite_lib.c:3735:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 17710:0:(lcommon_cl.c:177:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000403:0x15d:0x0]: rc = -5 LustreError: 17710:0:(llite_lib.c:3735:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 19296:0:(lcommon_cl.c:177:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000403:0x2f9:0x0]: rc = -5 LustreError: 19296:0:(lcommon_cl.c:177:cl_file_inode_init()) Skipped 1 previous similar message LustreError: 19296:0:(llite_lib.c:3735:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 19296:0:(llite_lib.c:3735:ll_prep_inode()) Skipped 1 previous similar message LustreError: 19936:0:(lcommon_cl.c:177:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000403:0x2f9:0x0]: rc = -5 LustreError: 19936:0:(lcommon_cl.c:177:cl_file_inode_init()) Skipped 11 previous similar messages LustreError: 19936:0:(llite_lib.c:3735:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 19936:0:(llite_lib.c:3735:ll_prep_inode()) Skipped 11 previous similar messages LustreError: 20831:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8800b6089800: inode [0x200000402:0x2da:0x0] mdc close failed: rc = -116 LustreError: 20831:0:(file.c:247:ll_close_inode_openhandle()) Skipped 7 previous similar messages LustreError: 10463:0:(lcommon_cl.c:177:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000402:0x187:0x0]: rc = -5 LustreError: 10463:0:(lcommon_cl.c:177:cl_file_inode_init()) Skipped 22 previous similar messages LustreError: 10463:0:(llite_lib.c:3735:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 10463:0:(llite_lib.c:3735:ll_prep_inode()) Skipped 22 previous similar messages Lustre: dir [0x240000402:0x11:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 1 previous similar message 14[22884]: segfault at 8 ip 00007f9b5fe297e8 sp 00007ffcf6e6b740 error 4 in ld-2.17.so[7f9b5fe1e000+22000] 18[23209]: segfault at 0 ip 0000000000403e5f sp 00007fff8fbc7120 error 6 in 18[400000+6000] LustreError: 22442:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8800b6089800: cannot apply new layout on [0x200000403:0x311:0x0] : rc = -5 LustreError: 22442:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x200000403:0x311:0x0] error -5. LustreError: 24478:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8800b3e11800: cannot apply new layout on [0x240000402:0x191:0x0] : rc = -5 LustreError: 24478:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x240000402:0x191:0x0] error -5. LustreError: 25624:0:(lcommon_cl.c:177:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000403:0x15d:0x0]: rc = -5 LustreError: 24745:0:(llite_lib.c:3735:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 24745:0:(llite_lib.c:3735:ll_prep_inode()) Skipped 10 previous similar messages LustreError: 25624:0:(lcommon_cl.c:177:cl_file_inode_init()) Skipped 11 previous similar messages LustreError: 582:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8800b3e11800: cannot apply new layout on [0x240000402:0x191:0x0] : rc = -5 LustreError: 582:0:(statahead.c:817:ll_statahead_interpret_work()) lustre: failed to prep 11 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 | Link to test |
racer test 1: racer on clients: centos-65.localnet DURATION=2700 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 PGD 2f8699067 PUD 2ab378067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) osp(OE) ofd(OE) lod(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) dm_flakey dm_mod libcfs(OE) loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) jbd2 mbcache crc32_generic crc_t10dif crct10dif_generic crct10dif_common virtio_console virtio_balloon pcspkr i2c_piix4 ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm ata_piix drm drm_panel_orientation_quirks libata serio_raw virtio_blk i2c_core floppy [last unloaded: libcfs] CPU: 2 PID: 18865 Comm: ll_sa_18718 Kdump: loaded Tainted: P OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.16.0-3.module_el8.7.0+1218+f626c2ff 04/01/2014 task: ffff880320618010 ti: ffff88009a7c4000 task.ti: ffff88009a7c4000 RIP: 0010:[<ffffffff813f502c>] [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 RSP: 0018:ffff88009a7c7d80 EFLAGS: 00010282 RAX: ffff8800aa8d9e48 RBX: 0000000000000008 RCX: 0000000100260021 RDX: 0000000000000026 RSI: ffff8800aa8da178 RDI: 0000000000000008 RBP: ffff88009a7c7d90 R08: ffff8802edbae018 R09: 0000000000000000 R10: ffff8802edbaf0a8 R11: ffff8802edbae018 R12: 0000000000000000 R13: ffff88031a9d3760 R14: ffff8800aa8da178 R15: ffff8802edbae018 FS: 0000000000000000(0000) GS:ffff880331a80000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 00000003227c8000 CR4: 00000000000007e0 Call Trace: [<ffffffffa17744b1>] ll_statahead_thread+0xde1/0x2290 [lustre] [<ffffffff817e320e>] ? _raw_spin_lock+0x1e/0x20 [<ffffffff810dbb51>] ? put_prev_entity+0x31/0x400 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e329e>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffff810c834d>] ? finish_task_switch+0x5d/0x1b0 [<ffffffffa17736d0>] ? ll_statahead_interpret+0x4a0/0x4a0 [lustre] [<ffffffff810ba114>] kthread+0xe4/0xf0 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 | Lustre: 321:0:(mdt_recovery.c:128:mdt_req_from_lrd()) @@@ restoring transno req@ffff8802f40e1440 x1824093160807296/t4294967444(0) o101->a6d1e58f-670e-44b0-bc14-e4aa5302579d@0@lo:648/0 lens 376/864 e 0 to 0 dl 1739590863 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 4[3873]: segfault at 8 ip 00007f6a704037e8 sp 00007fff355d6d20 error 4 in ld-2.17.so[7f6a703f8000+22000] LustreError: 4740:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8803275b37e8: inode [0x200000401:0x4c:0x0] mdc close failed: rc = -13 Lustre: 319:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000401:0xba:0x0] with magic=0xbd60bd0 Lustre: 2855:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0000: opcode 7: before 515 < left 618, rollback = 7 Lustre: 2855:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 2855:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/1, xattr_set: 2/15/0 Lustre: 2855:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 2855:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 2855:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 2855:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0000: opcode 7: before 516 < left 618, rollback = 7 Lustre: 2855:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 1 previous similar message Lustre: 2855:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 2855:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 2855:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 2855:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 2855:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 2855:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 2855:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 2855:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 2855:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 2855:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 2855:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 618, rollback = 7 Lustre: 2855:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 1 previous similar message Lustre: 2855:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 2855:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 2855:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 2855:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 2855:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 2855:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 2855:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 2855:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 2855:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 2855:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 588:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 516 < left 618, rollback = 7 Lustre: 588:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 1 previous similar message Lustre: 588:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 588:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 588:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 588:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 588:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 588:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 588:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 588:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 588:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 588:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: mdt_out00_002: service thread pid 3386 was inactive for 40.150 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 3386, comm: mdt_out00_002 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x913/0xd50 [ptlrpc] Lustre: mdt00_005: service thread pid 2576 was inactive for 40.146 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. [<0>] ldlm_cli_enqueue_local+0x2ea/0x800 [ptlrpc] [<0>] mdt_object_pdo_lock+0x369/0x750 [mdt] [<0>] mdt_parent_lock+0x76/0x2a0 [mdt] [<0>] mdt_getattr_name_lock+0x1850/0x2d00 [mdt] [<0>] mdt_intent_getattr+0x2cc/0x4e0 [mdt] [<0>] mdt_intent_opc.constprop.73+0x211/0xc60 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x397/0x980 [ptlrpc] [<0>] ldlm_handle_enqueue+0x507/0x1870 [ptlrpc] [<0>] mdt_batch_getattr+0x6c/0x190 [mdt] [<0>] mdt_batch+0x465/0x195b [mdt] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 2727, comm: mdt00_009 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x913/0xd50 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x800 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock+0x88/0x1c0 [mdt] [<0>] mdt_getattr_name_lock+0xc6a/0x2d00 [mdt] [<0>] mdt_intent_getattr+0x2cc/0x4e0 [mdt] [<0>] mdt_intent_opc.constprop.73+0x211/0xc60 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x397/0x980 [ptlrpc] [<0>] ldlm_handle_enqueue+0x507/0x1870 [ptlrpc] [<0>] tgt_enqueue+0x68/0x240 [ptlrpc] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 849, comm: mdt00_003 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x913/0xd50 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x800 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock+0x88/0x1c0 [mdt] [<0>] mdt_getattr_name_lock+0xc6a/0x2d00 [mdt] [<0>] mdt_intent_getattr+0x2cc/0x4e0 [mdt] [<0>] mdt_intent_opc.constprop.73+0x211/0xc60 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x397/0x980 [ptlrpc] [<0>] ldlm_handle_enqueue+0x507/0x1870 [ptlrpc] [<0>] tgt_enqueue+0x68/0x240 [ptlrpc] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: mdt00_016: service thread pid 10790 was inactive for 40.089 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 13 previous similar messages Lustre: mdt00_017: service thread pid 10793 was inactive for 72.029 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. LustreError: 312:0:(ldlm_lockd.c:252:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0000_UUID lock: ffff8800ada24b40/0x90bd5ed754dd142 lrc: 3/0,0 mode: PR/PR res: [0x200000401:0xa8:0x0].0x0 bits 0x13/0x0 rrc: 6 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0x90bd5ed754dd111 expref: 206 pid: 3386 timeout: 5996 lvb_type: 0 Lustre: mdt00_005: service thread pid 2576 completed after 100.321s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_014: service thread pid 10748 completed after 100.108s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_002: service thread pid 334 completed after 100.334s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 321:0:(ldlm_lockd.c:1492:ldlm_handle_enqueue()) ### lock on destroyed export ffff880296bedd28 ns: mdt-lustre-MDT0000_UUID lock: ffff880325f007c0/0x90bd5ed754dd65f lrc: 3/0,0 mode: PR/PR res: [0x200000401:0x1:0x0].0x0 bits 0x13/0x0 rrc: 22 type: IBT gid 0 flags: 0x50200000000000 nid: 0@lo remote: 0x90bd5ed754dd651 expref: 20 pid: 321 timeout: 0 lvb_type: 0 Lustre: mdt00_002: service thread pid 321 completed after 100.238s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_out00_002: service thread pid 3386 completed after 100.344s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_018: service thread pid 10794 completed after 93.346s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_001: service thread pid 320 completed after 100.240s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_006: service thread pid 2628 completed after 100.240s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_007: service thread pid 2714 completed after 100.248s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_009: service thread pid 2727 completed after 100.233s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_017: service thread pid 10793 completed after 94.336s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_011: service thread pid 3498 completed after 100.240s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_013: service thread pid 10747 completed after 100.126s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_015: service thread pid 10750 completed after 100.103s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_016: service thread pid 10790 completed after 98.363s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_012: service thread pid 3561 completed after 100.257s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: lustre-MDT0000-mdc-ffff8803275b37e8: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-MDT0000-mdc-ffff8803275b37e8: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: mdt00_008: service thread pid 2719 completed after 100.136s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_003: service thread pid 849 completed after 100.240s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: lustre-MDT0000-mdc-ffff8803275b37e8: This client was evicted by lustre-MDT0000; in progress operations using this service will fail. Lustre: mdt00_000: service thread pid 319 completed after 100.240s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_out00_001: service thread pid 325 completed after 100.060s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 10726:0:(file.c:6007:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000401:0x1:0x0] error: rc = -5 LustreError: 10813:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8803275b37e8: inode [0x200000402:0x15e:0x0] mdc close failed: rc = -108 LustreError: 10813:0:(ldlm_resource.c:1149:ldlm_resource_complain()) lustre-MDT0000-mdc-ffff8803275b37e8: namespace resource [0x200000007:0x1:0x0].0x0 (ffff88032255d940) refcount nonzero (3) after lock cleanup; forcing cleanup. Lustre: lustre-MDT0000-mdc-ffff8803275b37e8: Connection restored to (at 0@lo) Lustre: 587:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 618, rollback = 7 Lustre: 587:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 21 previous similar messages Lustre: 587:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 587:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 21 previous similar messages Lustre: 587:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 587:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 21 previous similar messages Lustre: 587:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 587:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 21 previous similar messages Lustre: 587:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 587:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 21 previous similar messages Lustre: 587:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 587:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 21 previous similar messages LustreError: 11390:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8803275b37e8: inode [0x200000402:0x275:0x0] mdc close failed: rc = -13 LustreError: 11390:0:(file.c:247:ll_close_inode_openhandle()) Skipped 7 previous similar messages Lustre: 10800:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0000: opcode 7: before 516 < left 618, rollback = 7 Lustre: 10800:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 5 previous similar messages Lustre: 10800:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 10800:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 10800:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 10800:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 10800:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 10800:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 10800:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 10800:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 10800:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 10800:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 5 previous similar messages | Link to test |
racer test 1: racer on clients: centos-95.localnet DURATION=2700 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 PGD 2c99c0067 PUD 2c7afa067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) lustre(OE) osp(OE) ofd(OE) lod(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) jbd2 mbcache lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) crc32_generic lnet(OE) dm_flakey dm_mod libcfs(OE) crc_t10dif crct10dif_generic crct10dif_common virtio_console pcspkr virtio_balloon i2c_piix4 ip_tables rpcsec_gss_krb5 drm_kms_helper ttm ata_generic pata_acpi drm drm_panel_orientation_quirks ata_piix floppy virtio_blk libata serio_raw i2c_core CPU: 2 PID: 17908 Comm: ll_sa_17859 Kdump: loaded Tainted: P OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.16.0-3.module_el8.7.0+1218+f626c2ff 04/01/2014 task: ffff8800a3783760 ti: ffff8802c9b28000 task.ti: ffff8802c9b28000 RIP: 0010:[<ffffffff813f502c>] [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 RSP: 0018:ffff8802c9b2bd80 EFLAGS: 00010282 RAX: ffff8802bc191e48 RBX: 0000000000000008 RCX: 0000000100260020 RDX: 0000000000000026 RSI: ffff8802bc192178 RDI: 0000000000000008 RBP: ffff8802c9b2bd90 R08: ffff8800b50e0008 R09: 0000000000000000 R10: ffff8800b50e5628 R11: ffff8800b50e0008 R12: 0000000000000000 R13: ffff8800aacb5c40 R14: ffff8802bc192178 R15: ffff8800b50e0008 FS: 0000000000000000(0000) GS:ffff880331a80000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 00000002f5168000 CR4: 00000000000007e0 Call Trace: [<ffffffffa1120511>] ll_statahead_thread+0xde1/0x2290 [lustre] [<ffffffff817e320e>] ? _raw_spin_lock+0x1e/0x20 [<ffffffff810dbb51>] ? put_prev_entity+0x31/0x400 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e329e>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffff810c834d>] ? finish_task_switch+0x5d/0x1b0 [<ffffffffa111f730>] ? ll_statahead_interpret+0x4a0/0x4a0 [lustre] [<ffffffff810ba114>] kthread+0xe4/0xf0 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 | Lustre: 14384:0:(mdt_recovery.c:128:mdt_req_from_lrd()) @@@ restoring transno req@ffff8802e7566e40 x1823939345207424/t4294969004(0) o101->3f680f08-7d03-43a6-9e22-a61693af60b2@0@lo:454/0 lens 392/864 e 0 to 0 dl 1739444199 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 Lustre: 15465:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0000: opcode 7: before 516 < left 618, rollback = 7 Lustre: 15465:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 15465:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 15465:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 15465:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 15465:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 11[19826]: segfault at 8 ip 00007f75273277e8 sp 00007fffe8c53310 error 4 in ld-2.17.so[7f752731c000+22000] Lustre: 9409:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0000: opcode 7: before 516 < left 618, rollback = 7 Lustre: 9409:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 1 previous similar message Lustre: 9409:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 9409:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 9409:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 9409:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 9409:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 9409:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 9409:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 9409:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 9409:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 9409:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 9409:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0000: opcode 7: before 516 < left 618, rollback = 7 Lustre: 9409:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 1 previous similar message Lustre: 9409:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 9409:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 9409:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 9409:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 9409:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 9409:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 9409:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 9409:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 9409:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 9409:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 14657:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0000: opcode 7: before 516 < left 612, rollback = 7 Lustre: 14657:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 1 previous similar message Lustre: 14657:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 14657:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 14657:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 14657:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 14657:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/612/0, punch: 0/0/0, quota 1/3/0 Lustre: 14657:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 14657:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 14657:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 14657:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 14657:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 15465:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 618, rollback = 7 Lustre: 15465:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 5 previous similar messages Lustre: 15465:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 15465:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 15465:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 15465:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 15465:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 15465:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 15465:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 15465:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 15465:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 15465:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 5 previous similar messages 17[26803]: segfault at 8 ip 00007fc9f621c7e8 sp 00007ffff85dc3d0 error 4 in ld-2.17.so[7fc9f6211000+22000] Lustre: 15465:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0000: opcode 7: before 516 < left 618, rollback = 7 Lustre: 15465:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 1 previous similar message Lustre: 15465:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 15465:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 15465:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 15465:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 15465:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 15465:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 15465:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 15465:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 15465:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 15465:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 1 previous similar message 12[32110]: segfault at 8 ip 00007f6e7a2757e8 sp 00007ffc02f83b40 error 4 in ld-2.17.so[7f6e7a26a000+22000] 8[1136]: segfault at 8 ip 00007f891ab0e7e8 sp 00007ffef357c210 error 4 in ld-2.17.so[7f891ab03000+22000] Lustre: 9409:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 618, rollback = 7 Lustre: 9409:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 7 previous similar messages Lustre: 9409:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 9409:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 7 previous similar messages Lustre: 9409:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 9409:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 7 previous similar messages Lustre: 9409:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 9409:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 7 previous similar messages Lustre: 9409:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 9409:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 7 previous similar messages Lustre: 9409:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 9409:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 7 previous similar messages 12[5346]: segfault at 8 ip 00007fae367437e8 sp 00007ffd5f6127d0 error 4 in ld-2.17.so[7fae36738000+22000] LustreError: 8468:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802ed3a92a8: inode [0x200000402:0x78f:0x0] mdc close failed: rc = -13 Lustre: 8742:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000401:0x7d3:0x0] with magic=0xbd60bd0 Lustre: 9409:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0000: opcode 7: before 516 < left 618, rollback = 7 Lustre: 9409:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 19 previous similar messages Lustre: 9409:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 9409:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 19 previous similar messages Lustre: 9409:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 9409:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 19 previous similar messages Lustre: 9409:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 9409:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 19 previous similar messages Lustre: 9409:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 9409:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 19 previous similar messages Lustre: 9409:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 9409:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 19 previous similar messages 17[14542]: segfault at 8 ip 00007f8b015dc7e8 sp 00007ffe159da380 error 4 in ld-2.17.so[7f8b015d1000+22000] 8[16648]: segfault at 401da1 ip 000000000040294e sp 00007fff2a87e390 error 7 in 8[400000+6000] 7[17075]: segfault at 8 ip 00007f07ea0817e8 sp 00007ffe58a88ac0 error 4 in ld-2.17.so[7f07ea076000+22000] Lustre: 14679:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000401:0xbed:0x0] with magic=0xbd60bd0 Lustre: 14679:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message Lustre: 15465:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0000: opcode 7: before 516 < left 618, rollback = 7 Lustre: 15465:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 25 previous similar messages Lustre: 15465:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 15465:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 25 previous similar messages Lustre: 15465:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 15465:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 25 previous similar messages Lustre: 15465:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 15465:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 25 previous similar messages Lustre: 15465:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 15465:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 25 previous similar messages Lustre: 15465:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 15465:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 25 previous similar messages 19[25051]: segfault at 8 ip 00007f7cbb5987e8 sp 00007ffd29286e90 error 4 in ld-2.17.so[7f7cbb58d000+22000] 2[32527]: segfault at 8 ip 00007f8658ad37e8 sp 00007fff0ff23aa0 error 4 in ld-2.17.so[7f8658ac8000+22000] 14[7735]: segfault at 8 ip 00007f62f5b4e7e8 sp 00007fffb71b3390 error 4 in ld-2.17.so[7f62f5b43000+22000] 2[11495]: segfault at 8 ip 00007f24c32877e8 sp 00007ffddcb27150 error 4 in ld-2.17.so[7f24c327c000+22000] LustreError: 12612:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802eecea548: inode [0x200000401:0x13df:0x0] mdc close failed: rc = -13 16[16626]: segfault at 8 ip 00007f2519c137e8 sp 00007ffc6aea5a50 error 4 in ld-2.17.so[7f2519c08000+22000] LustreError: 16626:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802eecea548: inode [0x200000401:0x15e9:0x0] mdc close failed: rc = -13 13[17289]: segfault at 8 ip 00007f3d41cde7e8 sp 00007ffe60017e90 error 4 in ld-2.17.so[7f3d41cd3000+22000] 8[25610]: segfault at 0 ip (null) sp 00007ffeee16a068 error 14 in 8[400000+6000] Lustre: 15465:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 516 < left 618, rollback = 7 Lustre: 15465:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 131 previous similar messages Lustre: 15465:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 15465:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 131 previous similar messages Lustre: 15465:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 15465:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 131 previous similar messages Lustre: 15465:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 15465:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 131 previous similar messages Lustre: 15465:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 15465:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 131 previous similar messages Lustre: 15465:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 15465:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 131 previous similar messages Lustre: 14384:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000401:0x1bf0:0x0] with magic=0xbd60bd0 Lustre: 14384:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message LustreError: 31870:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802eecea548: inode [0x200000402:0x1a07:0x0] mdc close failed: rc = -13 1[32526]: segfault at 8 ip 00007f52f0b5d7e8 sp 00007ffd77b2a690 error 4 in ld-2.17.so[7f52f0b52000+22000] Lustre: lustre-OST0002-osc-ffff8802ed3a92a8: disconnect after 21s idle Lustre: mdt00_017: service thread pid 3041 was inactive for 40.030 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 11504, comm: mdt00_003 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x913/0xd50 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x800 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock+0x88/0x1c0 [mdt] [<0>] mdt_object_find_lock+0x54/0x170 [mdt] [<0>] mdt_reint_setxattr+0x133/0x10e0 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x74c/0xbc0 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: Skipped 1 previous similar message Pid: 3041, comm: mdt00_017 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x913/0xd50 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x800 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock_try+0xa0/0x250 [mdt] [<0>] mdt_getattr_name_lock+0x173b/0x2d00 [mdt] [<0>] mdt_intent_getattr+0x2cc/0x4e0 [mdt] [<0>] mdt_intent_opc.constprop.73+0x211/0xc60 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x397/0x980 [ptlrpc] [<0>] ldlm_handle_enqueue+0x507/0x1870 [ptlrpc] [<0>] tgt_enqueue+0x68/0x240 [ptlrpc] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 8739, comm: mdt00_000 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x913/0xd50 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x800 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock_try+0xa0/0x250 [mdt] [<0>] mdt_getattr_name_lock+0x173b/0x2d00 [mdt] [<0>] mdt_intent_getattr+0x2cc/0x4e0 [mdt] [<0>] mdt_intent_opc.constprop.73+0x211/0xc60 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x397/0x980 [ptlrpc] [<0>] ldlm_handle_enqueue+0x507/0x1870 [ptlrpc] [<0>] tgt_enqueue+0x68/0x240 [ptlrpc] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: mdt00_015: service thread pid 23275 was inactive for 40.099 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: mdt00_009: service thread pid 15876 was inactive for 40.069 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 1 previous similar message Lustre: mdt00_010: service thread pid 20121 was inactive for 40.115 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 3 previous similar messages LustreError: 8577:0:(ldlm_lockd.c:252:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0000_UUID lock: ffff8800a33b61c0/0xfc087fc77dd529ba lrc: 3/0,0 mode: PR/PR res: [0x200000402:0x1c92:0x0].0x0 bits 0x13/0x0 rrc: 15 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0xfc087fc77dd529ac expref: 1093 pid: 11504 timeout: 626 lvb_type: 0 Lustre: mdt00_003: service thread pid 11504 completed after 100.349s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 23275:0:(ldlm_lockd.c:1492:ldlm_handle_enqueue()) ### lock on destroyed export ffff8802eda02e98 ns: mdt-lustre-MDT0000_UUID lock: ffff8802cd2a0400/0xfc087fc77dd5400a lrc: 3/0,0 mode: PR/PR res: [0x200000402:0x1c92:0x0].0x0 bits 0x1b/0x0 rrc: 12 type: IBT gid 0 flags: 0x50200000000000 nid: 0@lo remote: 0xfc087fc77dd53ffc expref: 408 pid: 23275 timeout: 0 lvb_type: 0 Lustre: mdt00_015: service thread pid 23275 completed after 100.027s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_004: service thread pid 14384 completed after 98.538s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_006: service thread pid 14679 completed after 98.116s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: lustre-MDT0000-mdc-ffff8802eecea548: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-MDT0000-mdc-ffff8802eecea548: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: mdt00_000: service thread pid 8739 completed after 100.308s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_016: service thread pid 32667 completed after 99.731s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: lustre-MDT0000-mdc-ffff8802eecea548: This client was evicted by lustre-MDT0000; in progress operations using this service will fail. Lustre: mdt00_008: service thread pid 15769 completed after 98.177s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_009: service thread pid 15876 completed after 98.860s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_017: service thread pid 3041 completed after 100.362s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_010: service thread pid 20121 completed after 97.153s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 6917:0:(file.c:5986:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000402:0x1c92:0x0] error: rc = -5 LustreError: 7731:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802eecea548: inode [0x200000402:0x1c1e:0x0] mdc close failed: rc = -108 LustreError: 7535:0:(llite_lib.c:1997:ll_md_setattr()) md_setattr fails: rc = -108 Lustre: lustre-MDT0000-mdc-ffff8802eecea548: Connection restored to (at 0@lo) LustreError: 14183:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802ed3a92a8: inode [0x200000403:0x1c3:0x0] mdc close failed: rc = -13 LustreError: 14183:0:(file.c:247:ll_close_inode_openhandle()) Skipped 14 previous similar messages Lustre: 14384:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000402:0x1ea9:0x0] with magic=0xbd60bd0 Lustre: 14384:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message 6[15494]: segfault at 8 ip 00007f0e952a37e8 sp 00007ffe882924f0 error 4 in ld-2.17.so[7f0e95298000+22000] | Link to test |
racer test 1: racer on clients: centos-70.localnet DURATION=2700 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 PGD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) lustre(OE) osp(OE) ofd(OE) lod(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) jbd2 mbcache lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) crc32_generic lnet(OE) dm_flakey dm_mod libcfs(OE) crc_t10dif crct10dif_generic crct10dif_common virtio_balloon virtio_console i2c_piix4 pcspkr ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm ata_piix drm_panel_orientation_quirks serio_raw i2c_core libata virtio_blk floppy CPU: 8 PID: 17419 Comm: ll_sa_17396 Kdump: loaded Tainted: P OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.16.0-3.module_el8.7.0+1218+f626c2ff 04/01/2014 task: ffff88032971c9d0 ti: ffff880291c44000 task.ti: ffff880291c44000 RIP: 0010:[<ffffffff813f502c>] [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 RSP: 0018:ffff880291c47d80 EFLAGS: 00010282 RAX: ffff880297946288 RBX: 0000000000000008 RCX: 0000000100260025 RDX: 0000000000000026 RSI: ffff8802979465b8 RDI: 0000000000000008 RBP: ffff880291c47d90 R08: ffff8802a79c13e8 R09: 0000000000000000 R10: ffff8802a79c0008 R11: ffff8802a79c13e8 R12: 0000000000000000 R13: ffff8802dc141280 R14: ffff8802979465b8 R15: ffff8802a79c13e8 FS: 0000000000000000(0000) GS:ffff880331c00000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 0000000001c10000 CR4: 00000000000007e0 Call Trace: [<ffffffffa10e6641>] ll_statahead_thread+0xde1/0x2290 [lustre] [<ffffffff810df6e2>] ? pick_next_task_fair+0x702/0x8a0 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e329e>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffff810c834d>] ? finish_task_switch+0x5d/0x1b0 [<ffffffffa10e5860>] ? ll_statahead_interpret+0x4a0/0x4a0 [lustre] [<ffffffff810ba114>] kthread+0xe4/0xf0 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 | LustreError: 10761:0:(mdt_reint.c:2513:mdt_reint_migrate()) lustre-MDT0000: migrate [0x200000403:0x1:0x0]/7 failed: rc = -114 LustreError: 17240:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e7f4efc8: inode [0x200000403:0xf:0x0] mdc close failed: rc = -116 Lustre: 14557:0:(mdt_recovery.c:128:mdt_req_from_lrd()) @@@ restoring transno req@ffff8802f0830040 x1822963478687232/t4294967448(0) o101->14689558-0526-4148-8dfa-d216f0afe710@0@lo:691/0 lens 376/864 e 0 to 0 dl 1738513521 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 cp (17168) used greatest stack depth: 10192 bytes left LustreError: 10761:0:(mdt_reint.c:2513:mdt_reint_migrate()) lustre-MDT0001: migrate [0x240000403:0x1:0x0]/10 failed: rc = -16 Lustre: 17771:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 507 < left 699, rollback = 2 Lustre: 17771:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 3/12/4, destroy: 0/0/0 Lustre: 17771:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 2/2/0, xattr_set: 9/699/0 Lustre: 17771:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 5/55/0, punch: 0/0/0, quota 1/3/0 Lustre: 17771:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 10/203/4, delete: 0/0/0 Lustre: 17771:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 4/4/0, ref_del: 0/0/0 LustreError: 18029:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e58c5d28: inode [0x200000403:0x1f:0x0] mdc close failed: rc = -116 LustreError: 10759:0:(mdt_reint.c:2513:mdt_reint_migrate()) lustre-MDT0002: migrate [0x280000403:0x1:0x0]/15 failed: rc = -16 LustreError: 10759:0:(mdt_reint.c:2513:mdt_reint_migrate()) Skipped 3 previous similar messages LustreError: 10761:0:(mdd_dir.c:4619:mdd_migrate_cmd_check()) lustre-MDD0001: '5' migration was interrupted, run 'lfs migrate -m 2 -c 1 -H crush 5' to finish migration: rc = -1 LustreError: 18757:0:(mdt_reint.c:2513:mdt_reint_migrate()) lustre-MDT0002: migrate [0x280000403:0x1:0x0]/15 failed: rc = -16 LustreError: 18757:0:(mdt_reint.c:2513:mdt_reint_migrate()) Skipped 2 previous similar messages Lustre: 10761:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 505 < left 610, rollback = 2 Lustre: 10761:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 7 previous similar messages Lustre: 10761:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 3/12/6, destroy: 0/0/0 Lustre: 10761:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 7 previous similar messages Lustre: 10761:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 2/2/0, xattr_set: 8/610/0 Lustre: 10761:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 7 previous similar messages Lustre: 10761:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 5/55/0, punch: 0/0/0, quota 1/3/0 Lustre: 10761:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 7 previous similar messages Lustre: 10761:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 11/239/4, delete: 0/0/0 Lustre: 10761:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 7 previous similar messages Lustre: 10761:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 5/5/0, ref_del: 0/0/0 Lustre: 10761:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 7 previous similar messages LustreError: 19346:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e58c5d28: inode [0x200000404:0x30:0x0] mdc close failed: rc = -116 LustreError: 10760:0:(mdt_reint.c:2513:mdt_reint_migrate()) lustre-MDT0002: migrate [0x280000403:0x1:0x0]/9 failed: rc = -114 LustreError: 10760:0:(mdt_reint.c:2513:mdt_reint_migrate()) Skipped 4 previous similar messages Lustre: 20439:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 509 < left 922, rollback = 2 Lustre: 20439:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 13 previous similar messages Lustre: 20439:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 3/12/3, destroy: 1/4/0 Lustre: 20439:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 13 previous similar messages Lustre: 20439:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 4/4/0, xattr_set: 13/922/0 Lustre: 20439:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 13 previous similar messages Lustre: 20439:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 5/55/0, punch: 0/0/0, quota 7/105/0 Lustre: 20439:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 13 previous similar messages Lustre: 20439:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 15/263/3, delete: 3/6/0 Lustre: 20439:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 13 previous similar messages Lustre: 20439:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 8/8/0, ref_del: 3/3/0 Lustre: 20439:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 13 previous similar messages Lustre: 18976:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 509 < left 638, rollback = 2 Lustre: 18976:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 17 previous similar messages Lustre: 18976:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 3/12/2, destroy: 0/0/0 Lustre: 18976:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 17 previous similar messages Lustre: 18976:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 4/4/0, xattr_set: 10/638/0 Lustre: 18976:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 17 previous similar messages Lustre: 18976:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 5/55/0, punch: 0/0/0, quota 4/54/0 Lustre: 18976:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 17 previous similar messages Lustre: 18976:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 13/251/4, delete: 0/0/0 Lustre: 18976:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 17 previous similar messages Lustre: 18976:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 7/7/0, ref_del: 0/0/0 Lustre: 18976:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 17 previous similar messages LustreError: 22889:0:(lcommon_cl.c:177:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000403:0xf9:0x0]: rc = -5 LustreError: 22889:0:(llite_lib.c:3735:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 18757:0:(mdt_reint.c:2513:mdt_reint_migrate()) lustre-MDT0000: migrate [0x200000403:0x1:0x0]/14 failed: rc = -114 LustreError: 22966:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e7f4efc8: inode [0x200000403:0x1a5:0x0] mdc close failed: rc = -116 Lustre: 20341:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 509 < left 922, rollback = 2 Lustre: 20341:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 5 previous similar messages Lustre: 20341:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 3/12/2, destroy: 1/4/0 Lustre: 20341:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 20341:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 4/4/0, xattr_set: 13/922/0 Lustre: 20341:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 20341:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 5/55/0, punch: 0/0/0, quota 1/3/0 Lustre: 20341:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 20341:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 16/299/4, delete: 3/6/0 Lustre: 20341:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 20341:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 9/9/0, ref_del: 3/3/0 Lustre: 20341:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 5 previous similar messages LustreError: 31629:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e7f4efc8: inode [0x200000403:0x209:0x0] mdc close failed: rc = -116 Lustre: 12623:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0002: opcode 7: before 516 < left 618, rollback = 7 Lustre: 20118:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0002: opcode 7: before 516 < left 618, rollback = 7 Lustre: 20118:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 3 previous similar messages Lustre: 20118:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 20118:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 64 previous similar messages Lustre: 10760:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 508 < left 610, rollback = 2 Lustre: 10760:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 60 previous similar messages Lustre: 10760:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 2/2/0, xattr_set: 8/610/0 Lustre: 10760:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 64 previous similar messages Lustre: 10760:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 5/55/0, punch: 0/0/0, quota 4/54/0 Lustre: 10760:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 64 previous similar messages Lustre: 10760:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 11/239/3, delete: 0/0/0 Lustre: 10760:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 64 previous similar messages Lustre: 10760:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 5/5/0, ref_del: 0/0/0 Lustre: 10760:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 64 previous similar messages 10[32031]: segfault at 0 ip 0000000000403cf0 sp 00007ffe1d91b728 error 6 in 10[400000+6000] LustreError: 10760:0:(mdd_dir.c:4619:mdd_migrate_cmd_check()) lustre-MDD0000: '18' migration was interrupted, run 'lfs migrate -m 1 -c 2 -H crush 18' to finish migration: rc = -1 LustreError: 1782:0:(mdt_reint.c:2513:mdt_reint_migrate()) lustre-MDT0001: migrate [0x240000403:0x1:0x0]/8 failed: rc = -2 LustreError: 1782:0:(mdt_reint.c:2513:mdt_reint_migrate()) Skipped 25 previous similar messages Lustre: 20032:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 515 < left 618, rollback = 7 Lustre: 20032:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 3 previous similar messages LustreError: 10761:0:(mdd_dir.c:4619:mdd_migrate_cmd_check()) lustre-MDD0000: '4' migration was interrupted, run 'lfs migrate -m 1 -c 3 -H crush 4' to finish migration: rc = -1 LustreError: 3070:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e58c5d28: inode [0x200000404:0x307:0x0] mdc close failed: rc = -13 LustreError: 3070:0:(file.c:247:ll_close_inode_openhandle()) Skipped 3 previous similar messages Lustre: 20118:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 516 < left 618, rollback = 7 Lustre: 20118:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 1 previous similar message Lustre: 10761:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 500 < left 1055, rollback = 2 Lustre: 10761:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 143 previous similar messages Lustre: 10761:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 3/12/11, destroy: 0/0/0 Lustre: 10761:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 149 previous similar messages Lustre: 10761:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 2/2/0, xattr_set: 13/1055/0 Lustre: 10761:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 161 previous similar messages Lustre: 10761:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 5/55/0, punch: 0/0/0, quota 4/54/0 Lustre: 10761:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 161 previous similar messages Lustre: 10761:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 12/275/4, delete: 0/0/0 Lustre: 10761:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 161 previous similar messages Lustre: 10761:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 6/6/0, ref_del: 0/0/0 Lustre: 10761:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 161 previous similar messages Lustre: 17734:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0000: opcode 7: before 516 < left 570, rollback = 7 Lustre: 17734:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 11 previous similar messages LustreError: 10761:0:(mdd_dir.c:4619:mdd_migrate_cmd_check()) lustre-MDD0000: '12' migration was interrupted, run 'lfs migrate -m 2 -c 3 -H crush 12' to finish migration: rc = -1 LustreError: 7490:0:(lcommon_cl.c:177:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000404:0x354:0x0]: rc = -5 LustreError: 7490:0:(llite_lib.c:3735:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 3[7331]: segfault at 8 ip 00007f191b6357e8 sp 00007ffd0b0a6c70 error 4 in ld-2.17.so[7f191b62a000+22000] LustreError: 7639:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e58c5d28: inode [0x200000403:0x36d:0x0] mdc close failed: rc = -116 LustreError: 7639:0:(file.c:247:ll_close_inode_openhandle()) Skipped 4 previous similar messages LustreError: 8825:0:(lcommon_cl.c:177:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000404:0x354:0x0]: rc = -5 LustreError: 8825:0:(llite_lib.c:3735:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 16274:0:(lcommon_cl.c:177:cl_file_inode_init()) lustre: failed to initialize cl_object [0x280000404:0x208:0x0]: rc = -5 LustreError: 16274:0:(lcommon_cl.c:177:cl_file_inode_init()) Skipped 1 previous similar message LustreError: 16274:0:(llite_lib.c:3735:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 16274:0:(llite_lib.c:3735:ll_prep_inode()) Skipped 1 previous similar message Lustre: 20148:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 515 < left 618, rollback = 7 Lustre: 20148:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 15 previous similar messages LustreError: 9842:0:(mdt_reint.c:2513:mdt_reint_migrate()) lustre-MDT0000: migrate [0x200000403:0x2:0x0]/17 failed: rc = -16 LustreError: 9842:0:(mdt_reint.c:2513:mdt_reint_migrate()) Skipped 41 previous similar messages Lustre: 10738:0:(mdt_recovery.c:128:mdt_req_from_lrd()) @@@ restoring transno req@ffff88009f34c140 x1822963497851264/t4294978126(0) o101->14689558-0526-4148-8dfa-d216f0afe710@0@lo:33/0 lens 376/864 e 0 to 0 dl 1738513618 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 LustreError: 10186:0:(lcommon_cl.c:177:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000403:0x27d:0x0]: rc = -5 LustreError: 10186:0:(llite_lib.c:3735:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 19[11082]: segfault at 8 ip 00007f30d227a7e8 sp 00007fffb5803d80 error 4 in ld-2.17.so[7f30d226f000+22000] Lustre: dir [0x240000404:0x261:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: 19416:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 540, rollback = 7 Lustre: 19416:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 5 previous similar messages Lustre: 19416:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 19416:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 314 previous similar messages Lustre: 19416:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 19416:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 314 previous similar messages Lustre: 19416:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/540/0, punch: 0/0/0, quota 1/3/0 Lustre: 19416:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 314 previous similar messages Lustre: 19416:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 19416:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 314 previous similar messages Lustre: 19416:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 19416:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 314 previous similar messages Lustre: 17206:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 507 < left 727, rollback = 2 Lustre: 17206:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 292 previous similar messages LustreError: 16307:0:(lcommon_cl.c:177:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000404:0x2ea:0x0]: rc = -5 LustreError: 16307:0:(llite_lib.c:3735:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 2705:0:(mdd_dir.c:4619:mdd_migrate_cmd_check()) lustre-MDD0001: '5' migration was interrupted, run 'lfs migrate -m 2 -c 2 -H crush 5' to finish migration: rc = -1 LustreError: 17622:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0001: failed to get lu_attr of [0x240000404:0x25a:0x0]: rc = -2 Lustre: dir [0x240000403:0x32e:0x0] stripe 1 readdir failed: -2, directory is partially accessed! LustreError: 17622:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0002: failed to get lu_attr of [0x280000403:0x2d1:0x0]: rc = -2 LustreError: 10745:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0001: failed to get lu_attr of [0x240000404:0x25a:0x0]: rc = -2 LustreError: 12981:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e7f4efc8: inode [0x240000404:0x25a:0x0] mdc close failed: rc = -2 LustreError: 12981:0:(file.c:247:ll_close_inode_openhandle()) Skipped 3 previous similar messages LustreError: 10759:0:(mdd_dir.c:4619:mdd_migrate_cmd_check()) lustre-MDD0000: '11' migration was interrupted, run 'lfs migrate -m 1 -c 2 -H crush 11' to finish migration: rc = -1 LustreError: 10759:0:(mdd_dir.c:4619:mdd_migrate_cmd_check()) Skipped 1 previous similar message Lustre: 18790:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000404:0x630:0x0] with magic=0xbd60bd0 LustreError: 19427:0:(lcommon_cl.c:177:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000403:0x27d:0x0]: rc = -5 LustreError: 19427:0:(lcommon_cl.c:177:cl_file_inode_init()) Skipped 5 previous similar messages LustreError: 19427:0:(llite_lib.c:3735:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 19427:0:(llite_lib.c:3735:ll_prep_inode()) Skipped 5 previous similar messages Lustre: dir [0x280000403:0x472:0x0] stripe 1 readdir failed: -2, directory is partially accessed! LustreError: 3571:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0002: failed to get lu_attr of [0x280000404:0x1f2:0x0]: rc = -2 LustreError: 3571:0:(mdd_object.c:3901:mdd_close()) Skipped 2 previous similar messages LustreError: 16155:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8802e7f4efc8: cannot apply new layout on [0x280000404:0x208:0x0] : rc = -5 LustreError: 16155:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x280000404:0x208:0x0] error -5. LustreError: 20298:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8802e7f4efc8: cannot apply new layout on [0x280000404:0x208:0x0] : rc = -5 LustreError: 20075:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8802e7f4efc8: cannot apply new layout on [0x280000404:0x208:0x0] : rc = -5 LustreError: 18857:0:(mdd_dir.c:4619:mdd_migrate_cmd_check()) lustre-MDD0002: '12' migration was interrupted, run 'lfs migrate -m 1 -c 3 -H crush 12' to finish migration: rc = -1 LustreError: 18857:0:(mdd_dir.c:4619:mdd_migrate_cmd_check()) Skipped 1 previous similar message LustreError: 22632:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8802e58c5d28: cannot apply new layout on [0x240000403:0x27d:0x0] : rc = -5 LustreError: 22632:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x240000403:0x27d:0x0] error -5. Lustre: 20118:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 516 < left 618, rollback = 7 Lustre: 20118:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 15 previous similar messages Lustre: 29499:0:(mdt_recovery.c:128:mdt_req_from_lrd()) @@@ restoring transno req@ffff8802d21d8f40 x1822963514369536/t4294982763(0) o101->14689558-0526-4148-8dfa-d216f0afe710@0@lo:86/0 lens 376/840 e 0 to 0 dl 1738513671 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 5[24852]: segfault at 8 ip 00007f00c0bdb7e8 sp 00007ffcefd1c230 error 4 in ld-2.17.so[7f00c0bd0000+22000] 5[24930]: segfault at 8 ip 00007f5a57eb97e8 sp 00007ffd1c529b70 error 4 in ld-2.17.so[7f5a57eae000+22000] LustreError: 17622:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0002: failed to get lu_attr of [0x280000404:0x1de:0x0]: rc = -2 13[25074]: segfault at 8 ip 00007f0d22aaf7e8 sp 00007ffcef6b1530 error 4 in ld-2.17.so[7f0d22aa4000+22000] LustreError: 209:0:(statahead.c:817:ll_statahead_interpret_work()) lustre: failed to prep 18 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 Lustre: 10738:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000404:0x736:0x0] with magic=0xbd60bd0 Lustre: 10738:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message LustreError: 127:0:(statahead.c:817:ll_statahead_interpret_work()) lustre: failed to prep 18 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 LustreError: 2846:0:(mdt_reint.c:2513:mdt_reint_migrate()) lustre-MDT0002: migrate [0x200000403:0x2:0x0]/3 failed: rc = -1 LustreError: 2846:0:(mdt_reint.c:2513:mdt_reint_migrate()) Skipped 76 previous similar messages LustreError: 28299:0:(lcommon_cl.c:177:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000403:0x301:0x0]: rc = -5 LustreError: 28299:0:(lcommon_cl.c:177:cl_file_inode_init()) Skipped 36 previous similar messages LustreError: 28299:0:(llite_lib.c:3735:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 28299:0:(llite_lib.c:3735:ll_prep_inode()) Skipped 36 previous similar messages Lustre: 12656:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000404:0x6d4:0x0] with magic=0xbd60bd0 Lustre: 12656:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message Lustre: 18864:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 499 < left 905, rollback = 2 Lustre: 18864:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 1231 previous similar messages Lustre: 18864:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 3/12/12, destroy: 0/0/0 Lustre: 18864:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 1265 previous similar messages Lustre: 18864:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 4/4/0, xattr_set: 13/905/0 Lustre: 18864:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 1265 previous similar messages Lustre: 18864:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 5/55/0, punch: 0/0/0, quota 4/54/0 Lustre: 18864:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 1265 previous similar messages Lustre: 18864:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 13/251/4, delete: 1/1/0 Lustre: 18864:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 1265 previous similar messages Lustre: 18864:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 7/7/0, ref_del: 1/1/0 Lustre: 18864:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 1265 previous similar messages LustreError: 30808:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000403:0x624:0x0]: rc = -2 Lustre: 3651:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000404:0x510:0x0] with magic=0xbd60bd0 Lustre: 3651:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 5 previous similar messages LustreError: 32201:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e7f4efc8: inode [0x280000403:0x4d6:0x0] mdc close failed: rc = -116 LustreError: 32201:0:(file.c:247:ll_close_inode_openhandle()) Skipped 13 previous similar messages Lustre: dir [0x280000404:0x58c:0x0] stripe 3 readdir failed: -2, directory is partially accessed! Lustre: Skipped 1 previous similar message Lustre: 18791:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0x664:0x0] with magic=0xbd60bd0 Lustre: 18791:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message LustreError: 10745:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0001: failed to get lu_attr of [0x240000404:0x4bd:0x0]: rc = -2 LustreError: 4244:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8802e7f4efc8: cannot apply new layout on [0x280000404:0x494:0x0] : rc = -5 LustreError: 4244:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x280000404:0x494:0x0] error -5. Lustre: dir [0x240000404:0x644:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 1 previous similar message Lustre: 20148:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 552, rollback = 7 Lustre: 20148:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 23 previous similar messages 13[9245]: segfault at 8 ip 00007f850a58e7e8 sp 00007ffd271ffa90 error 4 in ld-2.17.so[7f850a583000+22000] 17[9750]: segfault at 8 ip 00007f095ce6f7e8 sp 00007ffddd2b5e20 error 4 in ld-2.17.so[7f095ce64000+22000] Lustre: dir [0x240000403:0x7f0:0x0] stripe 1 readdir failed: -2, directory is partially accessed! LustreError: 11727:0:(lcommon_cl.c:177:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000403:0x6bf:0x0]: rc = -5 LustreError: 11727:0:(lcommon_cl.c:177:cl_file_inode_init()) Skipped 18 previous similar messages LustreError: 11727:0:(llite_lib.c:3735:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 11727:0:(llite_lib.c:3735:ll_prep_inode()) Skipped 18 previous similar messages traps: 16[11767] trap invalid opcode ip:40585e sp:7ffddbc8ee68 error:0 in 16[400000+6000] 12[12521]: segfault at 8 ip 00007f2244d767e8 sp 00007ffc7e7a8ae0 error 4 in ld-2.17.so[7f2244d6b000+22000] Lustre: 14557:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000403:0x849:0x0] with magic=0xbd60bd0 Lustre: 14557:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 5 previous similar messages traps: 8[12963] trap invalid opcode ip:40469e sp:7ffdd5851ae8 error:0 in 8[400000+6000] traps: 8[14015] trap invalid opcode ip:40469e sp:7ffc1ead73e8 error:0 in 8[400000+6000] 13[14808]: segfault at 8 ip 00007ffa32ced7e8 sp 00007fffb486ce90 error 4 in ld-2.17.so[7ffa32ce2000+22000] LustreError: 2705:0:(mdd_dir.c:4619:mdd_migrate_cmd_check()) lustre-MDD0002: '1' migration was interrupted, run 'lfs migrate -m 0 -c 1 -H crush 1' to finish migration: rc = -1 LustreError: 2705:0:(mdd_dir.c:4619:mdd_migrate_cmd_check()) Skipped 7 previous similar messages 16[15527]: segfault at 0 ip (null) sp 00007fff7bd5b998 error 14 in 16[400000+6000] Lustre: 13602:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000403:0x6f3:0x0] with magic=0xbd60bd0 Lustre: 13602:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 5 previous similar messages Lustre: dir [0x280000404:0x922:0x0] stripe 2 readdir failed: -2, directory is partially accessed! 8[20256]: segfault at 8 ip 00007fa2e336e7e8 sp 00007ffceb5caa70 error 4 in ld-2.17.so[7fa2e3363000+22000] LustreError: 787:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0002: failed to get lu_attr of [0x280000404:0x922:0x0]: rc = -2 LustreError: 787:0:(mdd_object.c:3901:mdd_close()) Skipped 6 previous similar messages LustreError: 22665:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8802e7f4efc8: cannot apply new layout on [0x200000404:0xd67:0x0] : rc = -5 LustreError: 22665:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x200000404:0xd67:0x0] error -5. LustreError: 2818:0:(mdt_reint.c:2513:mdt_reint_migrate()) lustre-MDT0000: migrate [0x280000403:0x1:0x0]/12 failed: rc = -16 LustreError: 2818:0:(mdt_reint.c:2513:mdt_reint_migrate()) Skipped 135 previous similar messages Lustre: 32427:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0xd5e:0x0] with magic=0xbd60bd0 Lustre: 32427:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 37 previous similar messages Lustre: 1782:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 504 < left 1189, rollback = 2 Lustre: 1782:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 2143 previous similar messages Lustre: 1782:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 3/12/7, destroy: 1/4/0 Lustre: 1782:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 2169 previous similar messages Lustre: 1782:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 4/4/0, xattr_set: 16/1189/0 Lustre: 1782:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 2169 previous similar messages Lustre: 1782:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 5/55/0, punch: 0/0/0, quota 9/107/0 Lustre: 1782:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 2169 previous similar messages Lustre: 1782:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 17/335/4, delete: 2/5/0 Lustre: 1782:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 2169 previous similar messages Lustre: 1782:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 10/10/0, ref_del: 1/1/0 Lustre: 1782:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 2169 previous similar messages LustreError: 17082:0:(mdt_xattr.c:402:mdt_dir_layout_update()) lustre-MDT0002: [0x280000404:0xa8a:0x0] migrate mdt count mismatch 1 != 2 LustreError: 2705:0:(mdd_dir.c:4619:mdd_migrate_cmd_check()) lustre-MDD0000: '2' migration was interrupted, run 'lfs migrate -m 2 -c 3 -H crush 2' to finish migration: rc = -1 LustreError: 2705:0:(mdd_dir.c:4619:mdd_migrate_cmd_check()) Skipped 7 previous similar messages LustreError: 31319:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e58c5d28: inode [0x240000403:0xc23:0x0] mdc close failed: rc = -116 LustreError: 31319:0:(file.c:247:ll_close_inode_openhandle()) Skipped 20 previous similar messages 17[490]: segfault at 8 ip 00007f70538227e8 sp 00007ffea6120fd0 error 4 in ld-2.17.so[7f7053817000+22000] Lustre: dir [0x200000404:0xfdb:0x0] stripe 2 readdir failed: -2, directory is partially accessed! Lustre: Skipped 5 previous similar messages LustreError: 1615:0:(llite_lib.c:1869:ll_update_lsm_md()) lustre: [0x280000404:0xcb4:0x0] dir layout mismatch: LustreError: 1615:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=2 count=2 index=2 hash=crush:0x2000003 max_inherit=0 max_inherit_rr=0 version=2 migrate_offset=0 migrate_hash=invalid:0 pool= LustreError: 1615:0:(lustre_lmv.h:167:lmv_stripe_object_dump()) stripe[0] [0x280000400:0x57:0x0] LustreError: 1615:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=1 count=3 index=2 hash=crush:0x82000003 max_inherit=0 max_inherit_rr=0 version=1 migrate_offset=2 migrate_hash=fnv_1a_64:2 pool= LustreError: 17624:0:(mdd_object.c:3901:mdd_close()) lustre-MDD0001: failed to get lu_attr of [0x240000403:0xceb:0x0]: rc = -2 LustreError: 17624:0:(mdd_object.c:3901:mdd_close()) Skipped 6 previous similar messages Lustre: 17266:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000404:0xd43:0x0] with magic=0xbd60bd0 Lustre: 17266:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 13 previous similar messages Lustre: 12623:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0000: opcode 7: before 516 < left 618, rollback = 7 Lustre: 12623:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 55 previous similar messages LustreError: 16371:0:(lcommon_cl.c:177:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000403:0x804:0x0]: rc = -5 LustreError: 16371:0:(lcommon_cl.c:177:cl_file_inode_init()) Skipped 51 previous similar messages LustreError: 16371:0:(llite_lib.c:3735:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 16371:0:(llite_lib.c:3735:ll_prep_inode()) Skipped 51 previous similar messages LustreError: 25058:0:(statahead.c:817:ll_statahead_interpret_work()) lustre: failed to prep 5 [0x280000404:0xc3b:0x0] inode@0000000000000000: rc = -5 LustreError: 25058:0:(statahead.c:817:ll_statahead_interpret_work()) Skipped 1 previous similar message | Link to test |
racer test 1: racer on clients: oleg108-client.virtnet DURATION=300 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffff813f40ac>] _atomic_dec_and_lock+0xc/0x70 PGD 88e79067 PUD 86466067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) osp(OE) ofd(OE) lod(OE) mdt(OE) mdd(OE) mgs(OE) lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) crc_t10dif crct10dif_generic rpcsec_gss_krb5 sb_edac edac_core iosf_mbi kvm_intel kvm irqbypass crc32_pclmul ghash_clmulni_intel squashfs aesni_intel lrw gf128mul glue_helper ablk_helper cryptd pcspkr i2c_piix4 i2c_core binfmt_misc ip_tables ext4 mbcache jbd2 ata_generic pata_acpi crct10dif_pclmul crct10dif_common crc32c_intel ata_piix libata serio_raw CPU: 2 PID: 19815 Comm: ll_sa_19718 Kdump: loaded Tainted: G OE ------------ 3.10.0-7.9-debug #1 Hardware name: QEMU Standard PC (i440FX + PIIX, 1996), BIOS 1.16.3-2.fc40 04/01/2014 task: ffff8800a38c4440 ti: ffff880089460000 task.ti: ffff880089460000 RIP: 0010:[<ffffffff813f40ac>] [<ffffffff813f40ac>] _atomic_dec_and_lock+0xc/0x70 RSP: 0018:ffff880089463d80 EFLAGS: 00010282 RAX: ffff88008a9878c8 RBX: 0000000000000008 RCX: 0000000100100009 RDX: 000000010010000a RSI: ffff88008a987bf8 RDI: 0000000000000008 RBP: ffff880089463d90 R08: ffff880092855400 R09: 0000000100100009 R10: 0000000092854201 R11: ffff880092855400 R12: 0000000000000000 R13: ffff88008bff8000 R14: ffff88008a987bf8 R15: ffff880092855400 FS: 0000000000000000(0000) GS:ffff88013e300000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 CR2: 0000000000000008 CR3: 0000000086f6c000 CR4: 0000000000160fe0 Call Trace: [<ffffffffa1088f41>] ll_statahead_thread+0xde1/0x2290 [lustre] [<ffffffff810df6e2>] ? pick_next_task_fair+0x702/0x8a0 [<ffffffff81410a79>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e8dce>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffff810c834d>] ? finish_task_switch+0x5d/0x1b0 [<ffffffffa1088160>] ? ll_statahead_handle.constprop.29+0x1c0/0x1c0 [lustre] [<ffffffff810ba114>] kthread+0xe4/0xf0 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817f3e5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 | LustreError: 11815:0:(file.c:267:ll_close_inode_openhandle()) lustre-clilmv-ffff8800b58e7800: inode [0x240000402:0x13:0x0] mdc close failed: rc = -116 random: crng init done 13[11941]: segfault at 8 ip 00007ff2bef417e8 sp 00007ffd84fc5970 error 4 in ld-2.17.so[7ff2bef36000+22000] LustreError: 11832:0:(file.c:267:ll_close_inode_openhandle()) lustre-clilmv-ffff8800b58e7800: inode [0x200000403:0xa:0x0] mdc close failed: rc = -116 LustreError: 11832:0:(file.c:267:ll_close_inode_openhandle()) Skipped 7 previous similar messages LustreError: 12874:0:(lcommon_cl.c:195:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000402:0x93:0x0]: rc = -5 LustreError: 12874:0:(llite_lib.c:3745:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 Lustre: dir [0x240000403:0x21:0x0] stripe 2 readdir failed: -2, directory is partially accessed! LustreError: 14062:0:(file.c:267:ll_close_inode_openhandle()) lustre-clilmv-ffff8800b58e7800: inode [0x200000403:0x65:0x0] mdc close failed: rc = -116 LustreError: 14062:0:(file.c:267:ll_close_inode_openhandle()) Skipped 2 previous similar messages Lustre: dir [0x240000402:0x137:0x0] stripe 1 readdir failed: -2, directory is partially accessed! LustreError: 15356:0:(lcommon_cl.c:195:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000403:0x1ff:0x0]: rc = -5 LustreError: 15356:0:(lcommon_cl.c:195:cl_file_inode_init()) Skipped 1 previous similar message LustreError: 15356:0:(llite_lib.c:3745:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 15356:0:(llite_lib.c:3745:ll_prep_inode()) Skipped 1 previous similar message LustreError: 16328:0:(file.c:267:ll_close_inode_openhandle()) lustre-clilmv-ffff88012bf47000: inode [0x200000402:0x142:0x0] mdc close failed: rc = -116 LustreError: 16328:0:(file.c:267:ll_close_inode_openhandle()) Skipped 4 previous similar messages 5[19073]: segfault at 8 ip 00007f490a73f87a sp 00007ffc96f46200 error 4 in ld-2.17.so[7f490a734000+22000] Lustre: dir [0x240000402:0x2e3:0x0] stripe 1 readdir failed: -2, directory is partially accessed! LustreError: 20055:0:(lcommon_cl.c:195:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000403:0x2a3:0x0]: rc = -5 LustreError: 20055:0:(lcommon_cl.c:195:cl_file_inode_init()) Skipped 2 previous similar messages LustreError: 20055:0:(llite_lib.c:3745:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 20055:0:(llite_lib.c:3745:ll_prep_inode()) Skipped 2 previous similar messages LustreError: 20015:0:(file.c:267:ll_close_inode_openhandle()) lustre-clilmv-ffff88012bf47000: inode [0x200000403:0x13a:0x0] mdc close failed: rc = -116 LustreError: 20015:0:(file.c:267:ll_close_inode_openhandle()) Skipped 10 previous similar messages 18[21916]: segfault at 406000 ip 0000000000406000 sp 00007fff5b2d65f8 error 14 in 18[606000+1000] Lustre: dir [0x240000403:0x4ae:0x0] stripe 2 readdir failed: -2, directory is partially accessed! Lustre: Skipped 2 previous similar messages LustreError: 11073:0:(lcommon_cl.c:195:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000403:0x3a5:0x0]: rc = -5 LustreError: 11073:0:(lcommon_cl.c:195:cl_file_inode_init()) Skipped 1 previous similar message LustreError: 11073:0:(llite_lib.c:3745:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 11073:0:(llite_lib.c:3745:ll_prep_inode()) Skipped 1 previous similar message 16[27779]: segfault at 8 ip 00007f1b410ac7e8 sp 00007fffcf27f410 error 4 in ld-2.17.so[7f1b410a1000+22000] LustreError: 10980:0:(lcommon_cl.c:195:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000403:0x5bf:0x0]: rc = -5 LustreError: 10980:0:(lcommon_cl.c:195:cl_file_inode_init()) Skipped 6 previous similar messages LustreError: 10980:0:(llite_lib.c:3745:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 10980:0:(llite_lib.c:3745:ll_prep_inode()) Skipped 6 previous similar messages LustreError: 27672:0:(file.c:267:ll_close_inode_openhandle()) lustre-clilmv-ffff8800b58e7800: inode [0x200000402:0x4d3:0x0] mdc close failed: rc = -116 LustreError: 27672:0:(file.c:267:ll_close_inode_openhandle()) Skipped 19 previous similar messages 14[30499]: segfault at 8 ip 00007f9aee86d7e8 sp 00007ffd59e4d290 error 4 in ld-2.17.so[7f9aee862000+22000] LustreError: 776:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8800b58e7800: cannot apply new layout on [0x240000403:0x679:0x0] : rc = -5 LustreError: 776:0:(vvp_io.c:1921:vvp_io_init()) lustre: refresh file layout [0x240000403:0x679:0x0] error -5. LustreError: 2509:0:(llite_nfs.c:446:ll_dir_get_parent_fid()) lustre: failure inode [0x240000403:0x976:0x0] get parent: rc = -116 LustreError: 10980:0:(lcommon_cl.c:195:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000403:0x6cb:0x0]: rc = -5 LustreError: 10980:0:(lcommon_cl.c:195:cl_file_inode_init()) Skipped 27 previous similar messages LustreError: 10980:0:(llite_lib.c:3745:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 10980:0:(llite_lib.c:3745:ll_prep_inode()) Skipped 27 previous similar messages Lustre: dir [0x240000403:0x93c:0x0] stripe 1 readdir failed: -2, directory is partially accessed! LustreError: 4663:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff88012bf47000: cannot apply new layout on [0x240000402:0x67b:0x0] : rc = -5 LustreError: 4663:0:(vvp_io.c:1921:vvp_io_init()) lustre: refresh file layout [0x240000402:0x67b:0x0] error -5. LustreError: 10798:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff88012bf47000: cannot apply new layout on [0x200000402:0xadb:0x0] : rc = -5 LustreError: 10798:0:(vvp_io.c:1921:vvp_io_init()) lustre: refresh file layout [0x200000402:0xadb:0x0] error -5. 13[13115]: segfault at 8 ip 00007f43eebca7e8 sp 00007ffcdcfa8360 error 4 in ld-2.17.so[7f43eebbf000+22000] 9[13434]: segfault at 8 ip 00007f22ea4b67e8 sp 00007ffe76c6df20 error 4 in ld-2.17.so[7f22ea4ab000+22000] LustreError: 13824:0:(file.c:267:ll_close_inode_openhandle()) lustre-clilmv-ffff8800b58e7800: inode [0x200000402:0x9cb:0x0] mdc close failed: rc = -116 LustreError: 13824:0:(file.c:267:ll_close_inode_openhandle()) Skipped 29 previous similar messages Lustre: dir [0x240000403:0xc91:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 2 previous similar messages LustreError: 17:0:(statahead.c:836:ll_statahead_interpret_work()) lustre: failed to prep 17 [0x240000403:0xb84:0x0] inode@0000000000000000: rc = -5 LustreError: 15043:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8800b58e7800: cannot apply new layout on [0x240000403:0xb84:0x0] : rc = -5 LustreError: 15043:0:(lov_object.c:1341:lov_layout_change()) Skipped 2 previous similar messages LustreError: 15043:0:(vvp_io.c:1921:vvp_io_init()) lustre: refresh file layout [0x240000403:0xb84:0x0] error -5. LustreError: 15043:0:(vvp_io.c:1921:vvp_io_init()) Skipped 2 previous similar messages LustreError: 10401:0:(lcommon_cl.c:195:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000402:0xd85:0x0]: rc = -5 LustreError: 10401:0:(lcommon_cl.c:195:cl_file_inode_init()) Skipped 86 previous similar messages LustreError: 10401:0:(llite_lib.c:3745:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 10401:0:(llite_lib.c:3745:ll_prep_inode()) Skipped 86 previous similar messages 17[21277]: segfault at 8 ip 00007f56400af7e8 sp 00007ffed788ce50 error 4 in ld-2.17.so[7f56400a4000+22000] LustreError: 23474:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff88012bf47000: cannot apply new layout on [0x200000402:0xadb:0x0] : rc = -5 LustreError: 23474:0:(lov_object.c:1341:lov_layout_change()) Skipped 4 previous similar messages 15[23944]: segfault at 8 ip 00007f78419377e8 sp 00007ffe6f3c1220 error 4 in ld-2.17.so[7f784192c000+22000] 4[26831]: segfault at 8 ip 00007f4e37aa47e8 sp 00007ffd2d6b68a0 error 4 in ld-2.17.so[7f4e37a99000+22000] 4[27252]: segfault at 8 ip 00007f7cd1cf97e8 sp 00007ffc50e0bb20 error 4 in ld-2.17.so[7f7cd1cee000+22000] LustreError: 46:0:(statahead.c:836:ll_statahead_interpret_work()) lustre: failed to prep 19 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 LustreError: 28747:0:(vvp_io.c:1921:vvp_io_init()) lustre: refresh file layout [0x200000403:0x12fe:0x0] error -5. LustreError: 28747:0:(vvp_io.c:1921:vvp_io_init()) Skipped 2 previous similar messages LustreError: 31016:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8800b58e7800: cannot apply new layout on [0x200000403:0xc34:0x0] : rc = -5 LustreError: 31016:0:(lov_object.c:1341:lov_layout_change()) Skipped 7 previous similar messages Lustre: dir [0x240000402:0xaf1:0x0] stripe 2 readdir failed: -2, directory is partially accessed! Lustre: Skipped 27 previous similar messages fallocate (2593) used greatest stack depth: 10288 bytes left 9[9125]: segfault at 8 ip 00007fcb915887e8 sp 00007ffe55a38200 error 4 in ld-2.17.so[7fcb9157d000+22000] LustreError: 9874:0:(vvp_io.c:1921:vvp_io_init()) lustre: refresh file layout [0x200000403:0x1acf:0x0] error -5. LustreError: 9874:0:(vvp_io.c:1921:vvp_io_init()) Skipped 1 previous similar message 12[11181]: segfault at 0 ip (null) sp 00007ffed0dfa0e8 error 14 in 12[400000+6000] LustreError: 11410:0:(file.c:267:ll_close_inode_openhandle()) lustre-clilmv-ffff88012bf47000: inode [0x240000402:0x1606:0x0] mdc close failed: rc = -2 LustreError: 11410:0:(file.c:267:ll_close_inode_openhandle()) Skipped 49 previous similar messages LustreError: 17037:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8800b58e7800: cannot apply new layout on [0x200000403:0x1833:0x0] : rc = -5 LustreError: 17037:0:(lov_object.c:1341:lov_layout_change()) Skipped 20 previous similar messages 1[19493]: segfault at 8 ip 00007f3a494717e8 sp 00007ffeea5b09a0 error 4 in ld-2.17.so[7f3a49466000+22000] LustreError: 20104:0:(lcommon_cl.c:195:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000403:0x1833:0x0]: rc = -5 LustreError: 20104:0:(lcommon_cl.c:195:cl_file_inode_init()) Skipped 263 previous similar messages LustreError: 20104:0:(llite_lib.c:3745:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 20104:0:(llite_lib.c:3745:ll_prep_inode()) Skipped 263 previous similar messages 6[20141]: segfault at 8 ip 00007f6cdc9647e8 sp 00007ffe304e0c20 error 4 in ld-2.17.so[7f6cdc959000+22000] LustreError: 32755:0:(vvp_io.c:1921:vvp_io_init()) lustre: refresh file layout [0x240000403:0x1a9d:0x0] error -5. LustreError: 32755:0:(vvp_io.c:1921:vvp_io_init()) Skipped 7 previous similar messages 12[3875]: segfault at 1c ip 00000000004017b0 sp 00007ffe15a71b50 error 6 in 3[400000+6000] 17[7794]: segfault at 8 ip 00007f4a080bc7e8 sp 00007fff6fdf5ec0 error 4 in ld-2.17.so[7f4a080b1000+22000] LustreError: 49:0:(statahead.c:836:ll_statahead_interpret_work()) lustre: failed to prep 16 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 LustreError: 21518:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8800b58e7800: cannot apply new layout on [0x240000403:0x1a9d:0x0] : rc = -5 LustreError: 21518:0:(lov_object.c:1341:lov_layout_change()) Skipped 125 previous similar messages LustreError: 17:0:(statahead.c:836:ll_statahead_interpret_work()) lustre: failed to prep 12 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 LustreError: 17:0:(statahead.c:836:ll_statahead_interpret_work()) Skipped 3 previous similar messages Lustre: dir [0x240000402:0x2cd4:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 18 previous similar messages LustreError: 46:0:(statahead.c:836:ll_statahead_interpret_work()) lustre: failed to prep 12 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 LustreError: 5058:0:(vvp_io.c:1921:vvp_io_init()) lustre: refresh file layout [0x240000402:0x25ea:0x0] error -5. LustreError: 5058:0:(vvp_io.c:1921:vvp_io_init()) Skipped 7 previous similar messages LustreError: 46:0:(statahead.c:836:ll_statahead_interpret_work()) lustre: failed to prep 12 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 17[6458]: segfault at 8 ip 00007f14049657e8 sp 00007ffc08f6f5a0 error 4 in ld-2.17.so[7f140495a000+22000] 0[9943]: segfault at 8 ip 00007fa2bbfd37e8 sp 00007ffe37a1bdb0 error 4 in ld-2.17.so[7fa2bbfc8000+22000] LustreError: 18559:0:(file.c:267:ll_close_inode_openhandle()) lustre-clilmv-ffff8800b58e7800: inode [0x200000402:0x569f:0x0] mdc close failed: rc = -116 LustreError: 18559:0:(file.c:267:ll_close_inode_openhandle()) Skipped 275 previous similar messages traps: 8[20243] trap invalid opcode ip:405162 sp:7fffd50d6c70 error:0 in 8[400000+6000] LustreError: 26674:0:(lcommon_cl.c:195:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000403:0x1a9d:0x0]: rc = -5 LustreError: 26674:0:(lcommon_cl.c:195:cl_file_inode_init()) Skipped 571 previous similar messages LustreError: 26674:0:(llite_lib.c:3745:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 26674:0:(llite_lib.c:3745:ll_prep_inode()) Skipped 571 previous similar messages LustreError: lustre-MDT0000-mdc-ffff88012bf47000: operation ldlm_enqueue to node 192.168.201.108@tcp failed: rc = -107 Lustre: lustre-MDT0000-mdc-ffff88012bf47000: Connection to lustre-MDT0000 (at 192.168.201.108@tcp) was lost; in progress operations using this service will wait for recovery to complete LustreError: lustre-MDT0000-mdc-ffff88012bf47000: This client was evicted by lustre-MDT0000; in progress operations using this service will fail. LustreError: 17863:0:(file.c:5979:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000403:0x1d8d:0x0] error: rc = -5 LustreError: Skipped 5 previous similar messages LustreError: 18268:0:(llite_lib.c:2032:ll_md_setattr()) md_setattr fails: rc = -108 Lustre: dir [0x240000402:0x165a:0x0] stripe 1 readdir failed: -108, directory is partially accessed! Lustre: Skipped 26 previous similar messages LustreError: 18428:0:(mdc_request.c:1454:mdc_read_page()) lustre-MDT0000-mdc-ffff88012bf47000: [0x200000401:0x3d:0x0] lock enqueue fails: rc = -108 LustreError: 17457:0:(file.c:5979:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000007:0x1:0x0] error: rc = -108 LustreError: 17457:0:(file.c:5979:ll_inode_revalidate_fini()) Skipped 830 previous similar messages LustreError: 16625:0:(ldlm_resource.c:1149:ldlm_resource_complain()) lustre-MDT0000-mdc-ffff88012bf47000: namespace resource [0x200000007:0x1:0x0].0x0 (ffff880092bc9000) refcount nonzero (1) after lock cleanup; forcing cleanup. Lustre: lustre-MDT0000-mdc-ffff88012bf47000: Connection restored to (at 192.168.201.108@tcp) LustreError: lustre-MDT0000-mdc-ffff8800b58e7800: operation ldlm_enqueue to node 192.168.201.108@tcp failed: rc = -107 Lustre: lustre-MDT0000-mdc-ffff8800b58e7800: Connection to lustre-MDT0000 (at 192.168.201.108@tcp) was lost; in progress operations using this service will wait for recovery to complete LustreError: lustre-MDT0000-mdc-ffff8800b58e7800: This client was evicted by lustre-MDT0000; in progress operations using this service will fail. LustreError: Skipped 8 previous similar messages LustreError: 27411:0:(file.c:5979:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000402:0x1:0x0] error: rc = -5 LustreError: 28868:0:(mdc_request.c:1454:mdc_read_page()) lustre-MDT0000-mdc-ffff8800b58e7800: [0x200000402:0x1:0x0] lock enqueue fails: rc = -108 LustreError: 30164:0:(statahead.c:1806:is_first_dirent()) lustre: reading dir [0x200000402:0x1:0x0] at 0 stat_pid = 27154 : rc = -108 LustreError: 16301:0:(llite_lib.c:2032:ll_md_setattr()) md_setattr fails: rc = -108 LustreError: 16301:0:(llite_lib.c:2032:ll_md_setattr()) Skipped 5 previous similar messages LustreError: 28868:0:(mdc_request.c:1454:mdc_read_page()) Skipped 14 previous similar messages Lustre: lustre-MDT0000-mdc-ffff8800b58e7800: Connection restored to (at 192.168.201.108@tcp) | Link to test |
racer test 1: racer on clients: centos-100.localnet DURATION=2700 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 PGD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) osp(OE) ofd(OE) lod(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) dm_flakey dm_mod libcfs(OE) loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) jbd2 mbcache crc32_generic crc_t10dif crct10dif_generic crct10dif_common i2c_piix4 pcspkr virtio_console virtio_balloon ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm ata_piix drm_panel_orientation_quirks floppy virtio_blk serio_raw libata i2c_core [last unloaded: libcfs] CPU: 9 PID: 6428 Comm: ll_sa_6413 Kdump: loaded Tainted: P OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.16.0-3.module_el8.7.0+1218+f626c2ff 04/01/2014 task: ffff880321209280 ti: ffff880325a24000 task.ti: ffff880325a24000 RIP: 0010:[<ffffffff813f502c>] [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 RSP: 0018:ffff880325a27d80 EFLAGS: 00010282 RAX: ffff8802a21d9e48 RBX: 0000000000000008 RCX: 000000010026001c RDX: 0000000000000026 RSI: ffff8802a21da178 RDI: 0000000000000008 RBP: ffff880325a27d90 R08: ffff8800ac2b6018 R09: 0000000000000000 R10: ffff8800ac2b0008 R11: ffff8800ac2b6018 R12: 0000000000000000 R13: ffff8800a88b1280 R14: ffff8802a21da178 R15: ffff8800ac2b6018 FS: 0000000000000000(0000) GS:ffff880331c40000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 0000000001c10000 CR4: 00000000000007e0 Call Trace: [<ffffffffa1769641>] ll_statahead_thread+0xde1/0x2290 [lustre] [<ffffffff817e320e>] ? _raw_spin_lock+0x1e/0x20 [<ffffffff810dbb51>] ? put_prev_entity+0x31/0x400 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e329e>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffff810c834d>] ? finish_task_switch+0x5d/0x1b0 [<ffffffffa1768860>] ? ll_statahead_interpret+0x4a0/0x4a0 [lustre] [<ffffffff810ba114>] kthread+0xe4/0xf0 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 | Lustre: 4639:0:(mdt_recovery.c:128:mdt_req_from_lrd()) @@@ restoring transno req@ffff8800a7f22d40 x1822417943169664/t4294967441(0) o101->d6f9c9ac-aab9-430c-9d3c-d259548a8cef@0@lo:609/0 lens 376/816 e 0 to 0 dl 1737993244 ref 1 fl Interpret:H/202/0 rc 0/0 job:'cat.0' uid:0 gid:0 18[8399]: segfault at 0 ip 0000000000403cf0 sp 00007ffc2acc9718 error 6 in 18[400000+6000] Lustre: 4395:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 516 < left 618, rollback = 7 Lustre: 4395:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 4395:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 4395:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 4395:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 4395:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 traps: 3[9993] trap invalid opcode ip:405250 sp:7ffe59f78d88 error:0 in 3[400000+6000] Lustre: 4394:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0002: opcode 7: before 516 < left 618, rollback = 7 Lustre: 4394:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 1 previous similar message Lustre: 4394:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 4394:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 4394:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 4394:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 4394:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 4394:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 4394:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 4394:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 4394:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 4394:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 4395:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 516 < left 618, rollback = 7 Lustre: 4395:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 3 previous similar messages Lustre: 4395:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 4395:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 4395:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 4395:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 4395:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 4395:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 4395:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 4395:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 4395:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 4395:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 8417:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 618, rollback = 7 Lustre: 8417:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 5 previous similar messages Lustre: 8417:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 8417:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 8417:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 8417:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 8417:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 8417:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 8417:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 8417:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 8417:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 8417:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 6312:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 516 < left 618, rollback = 7 Lustre: 6312:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 5 previous similar messages Lustre: 6312:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 6312:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 6312:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 6312:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 6312:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 6312:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 6312:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 6312:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 6312:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 6312:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 8417:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0000: opcode 7: before 516 < left 618, rollback = 7 Lustre: 8417:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 3 previous similar messages Lustre: 8417:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 8417:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 8417:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 8417:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 8417:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 8417:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 8417:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 8417:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 8417:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 8417:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 3 previous similar messages LustreError: 23878:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff88009c8d8008: inode [0x200000401:0x670:0x0] mdc close failed: rc = -13 Lustre: 30429:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 516 < left 618, rollback = 7 Lustre: 30429:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 11 previous similar messages Lustre: 30429:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 30429:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 11 previous similar messages Lustre: 30429:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 30429:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 11 previous similar messages Lustre: 30429:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 30429:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 11 previous similar messages Lustre: 30429:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 30429:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 11 previous similar messages Lustre: 30429:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 30429:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 11 previous similar messages 6[691]: segfault at 0 ip (null) sp 00007ffd33783f88 error 14 in 6[400000+6000] 8[908]: segfault at 8 ip 00007f5159e557e8 sp 00007ffec1d37b50 error 4 in ld-2.17.so[7f5159e4a000+22000] Lustre: 19781:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0002: opcode 7: before 516 < left 618, rollback = 7 Lustre: 19781:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 37 previous similar messages Lustre: 19781:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 19781:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 37 previous similar messages Lustre: 19781:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 19781:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 37 previous similar messages Lustre: 19781:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 19781:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 37 previous similar messages Lustre: 19781:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 19781:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 37 previous similar messages Lustre: 19781:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 19781:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 37 previous similar messages Lustre: mdt00_015: service thread pid 10286 was inactive for 40.106 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 10286, comm: mdt00_015 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x913/0xd50 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x800 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock+0x88/0x1c0 [mdt] [<0>] mdt_object_find_lock+0x54/0x170 [mdt] [<0>] mdt_reint_setxattr+0x133/0x10e0 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x74c/0xbc0 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 4639, comm: mdt00_003 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x913/0xd50 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x800 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock_try+0xa0/0x250 [mdt] [<0>] mdt_getattr_name_lock+0x173b/0x2d00 [mdt] Lustre: mdt00_000: service thread pid 4137 was inactive for 40.091 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. [<0>] mdt_intent_getattr+0x2cc/0x4e0 [mdt] [<0>] mdt_intent_opc.constprop.73+0x211/0xc60 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x397/0x980 [ptlrpc] [<0>] ldlm_handle_enqueue+0x507/0x1870 [ptlrpc] [<0>] tgt_enqueue+0x68/0x240 [ptlrpc] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 10281, comm: mdt00_014 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x913/0xd50 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x800 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock_try+0xa0/0x250 [mdt] [<0>] mdt_getattr_name_lock+0x173b/0x2d00 [mdt] [<0>] mdt_intent_getattr+0x2cc/0x4e0 [mdt] [<0>] mdt_intent_opc.constprop.73+0x211/0xc60 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x397/0x980 [ptlrpc] [<0>] ldlm_handle_enqueue+0x507/0x1870 [ptlrpc] [<0>] tgt_enqueue+0x68/0x240 [ptlrpc] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: mdt00_006: service thread pid 6355 was inactive for 40.086 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 9 previous similar messages LustreError: 4130:0:(ldlm_lockd.c:252:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0000_UUID lock: ffff8800a7d46940/0x6b149fc605b4d4c7 lrc: 3/0,0 mode: PR/PR res: [0x200000401:0x9fe:0x0].0x0 bits 0x1b/0x0 rrc: 8 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0x6b149fc605b4d4b9 expref: 338 pid: 4639 timeout: 3123 lvb_type: 0 Lustre: mdt00_015: service thread pid 10286 completed after 100.408s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 4137:0:(ldlm_lockd.c:1492:ldlm_handle_enqueue()) ### lock on destroyed export ffff8802eb0337e8 ns: mdt-lustre-MDT0000_UUID lock: ffff8802d017e580/0x6b149fc605b4eac3 lrc: 3/0,0 mode: PR/PR res: [0x200000401:0x9fe:0x0].0x0 bits 0x1b/0x0 rrc: 5 type: IBT gid 0 flags: 0x50200400000020 nid: 0@lo remote: 0x6b149fc605b4ea92 expref: 18 pid: 4137 timeout: 0 lvb_type: 0 Lustre: mdt00_000: service thread pid 4137 completed after 100.132s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: lustre-MDT0000-mdc-ffff8802f5c112a8: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-MDT0000-mdc-ffff8802f5c112a8: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: lustre-MDT0000-mdc-ffff8802f5c112a8: This client was evicted by lustre-MDT0000; in progress operations using this service will fail. LustreError: 2128:0:(file.c:5984:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000401:0x1:0x0] error: rc = -5 Lustre: mdt00_003: service thread pid 4639 completed after 100.116s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_014: service thread pid 10281 completed after 100.118s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 1997:0:(mdc_request.c:1454:mdc_read_page()) lustre-MDT0000-mdc-ffff8802f5c112a8: [0x200000401:0x1:0x0] lock enqueue fails: rc = -5 Lustre: mdt_io00_003: service thread pid 8651 completed after 100.022s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_002: service thread pid 4139 completed after 99.524s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_008: service thread pid 6778 completed after 99.935s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_006: service thread pid 6355 completed after 99.530s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_001: service thread pid 4138 completed after 99.527s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_017: service thread pid 2171 completed after 99.519s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_018: service thread pid 2172 completed after 99.514s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_004: service thread pid 6162 completed after 99.926s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_005: service thread pid 6249 completed after 99.933s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_012: service thread pid 10270 completed after 99.947s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_016: service thread pid 10289 completed after 99.943s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_010: service thread pid 8183 completed after 99.952s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_007: service thread pid 6512 completed after 99.540s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_013: service thread pid 10278 completed after 99.939s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 1921:0:(vvp_io.c:1903:vvp_io_init()) lustre: refresh file layout [0x200000401:0x9fe:0x0] error -108. Lustre: lustre-MDT0000-mdc-ffff8802f5c112a8: Connection restored to 192.168.123.102@tcp (at 0@lo) Lustre: 4395:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0000: opcode 7: before 516 < left 618, rollback = 7 Lustre: 4395:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 3 previous similar messages Lustre: 4395:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 4395:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 4395:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 4395:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 4395:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 4395:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 4395:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 4395:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 4395:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 4395:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 10281:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0x10:0x0] with magic=0xbd60bd0 10[4029]: segfault at 0 ip (null) sp 00007ffc53091bc8 error 14 in 10[400000+6000] 14[8174]: segfault at 8 ip 00007f8ca69bb7e8 sp 00007ffc09925d00 error 4 in ld-2.17.so[7f8ca69b0000+22000] LustreError: 12057:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802f5c112a8: inode [0x200000402:0xd36:0x0] mdc close failed: rc = -13 3[13290]: segfault at 8 ip 00007f8ffe9d47e8 sp 00007ffe67125a60 error 4 in ld-2.17.so[7f8ffe9c9000+22000] Lustre: 2195:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0x411:0x0] with magic=0xbd60bd0 Lustre: 2195:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message Lustre: 8183:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0x43b:0x0] with magic=0xbd60bd0 Lustre: 8183:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message 16[20183]: segfault at 0 ip (null) sp 00007ffdac364018 error 14 in 16[400000+6000] 15[25878]: segfault at 8 ip 00007f245ce1a7e8 sp 00007ffd2d4fed80 error 4 in ld-2.17.so[7f245ce0f000+22000] Lustre: 6249:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0x8e2:0x0] with magic=0xbd60bd0 Lustre: 6249:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message 5[26910]: segfault at 8 ip 00007f3d5c4177e8 sp 00007fffec7c7730 error 4 in ld-2.17.so[7f3d5c40c000+22000] traps: 10[28376] trap stack segment ip:404ddb sp:7ffe69931ca8 error:0 in 10[400000+6000] Lustre: 4396:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 618, rollback = 7 Lustre: 4396:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 105 previous similar messages Lustre: 4396:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 4396:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 105 previous similar messages Lustre: 4396:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 4396:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 105 previous similar messages Lustre: 4396:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 4396:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 105 previous similar messages Lustre: 4396:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 4396:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 105 previous similar messages Lustre: 4396:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 4396:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 105 previous similar messages 10[5547]: segfault at 8 ip 00007f249c1667e8 sp 00007ffe6e61e230 error 4 in ld-2.17.so[7f249c15b000+22000] LustreError: 29133:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff88009c8d8008: inode [0x200000403:0x1352:0x0] mdc close failed: rc = -13 Lustre: 4639:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0x178c:0x0] with magic=0xbd60bd0 Lustre: 4639:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message LustreError: 10470:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802f5c112a8: inode [0x200000403:0x1732:0x0] mdc close failed: rc = -13 LustreError: 14562:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802f5c112a8: inode [0x200000402:0x241b:0x0] mdc close failed: rc = -13 5[17218]: segfault at 8 ip 00007f0d66c377e8 sp 00007ffda3fcad90 error 4 in ld-2.17.so[7f0d66c2c000+22000] 19[19901]: segfault at 8 ip 00007ff7a0ece7e8 sp 00007ffcbe120f70 error 4 in ld-2.17.so[7ff7a0ec3000+22000] Lustre: 2172:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000402:0x2649:0x0] with magic=0xbd60bd0 Lustre: 2172:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message 11[23328]: segfault at 8 ip 00007f0f03b857e8 sp 00007ffdc27b6bb0 error 4 in ld-2.17.so[7f0f03b7a000+22000] 14[2951]: segfault at 406000 ip 0000000000406000 sp 00007ffe4d9c14c8 error 14 in 14[606000+1000] 2[3184]: segfault at 8 ip 00007f6a4be107e8 sp 00007fff0b23c860 error 4 in ld-2.17.so[7f6a4be05000+22000] LustreError: 3184:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802f5c112a8: inode [0x200000403:0x226d:0x0] mdc close failed: rc = -13 2[3476]: segfault at 8 ip 00007f693e4de7e8 sp 00007fff6cd48180 error 4 in ld-2.17.so[7f693e4d3000+22000] Lustre: 30429:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0002: opcode 7: before 516 < left 618, rollback = 7 Lustre: 30429:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 265 previous similar messages Lustre: 30429:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 30429:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 265 previous similar messages Lustre: 30429:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 30429:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 265 previous similar messages Lustre: 30429:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 30429:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 265 previous similar messages Lustre: 30429:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 30429:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 265 previous similar messages Lustre: 30429:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 30429:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 265 previous similar messages Lustre: 2197:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000402:0x2cd5:0x0] with magic=0xbd60bd0 Lustre: 2197:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message 19[8896]: segfault at 4017b0 ip 00000000004017b0 sp 00007ffcf7924c90 error 14 5[9080]: segfault at 0 ip 0000000000403cf0 sp 00007ffc926e2298 error 6 in 5[400000+6000] 9[11376]: segfault at 0 ip 0000000000403cf0 sp 00007ffe70c5cbb8 error 6 in 9[400000+6000] LustreError: 11365:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff88009c8d8008: inode [0x200000402:0x2df6:0x0] mdc close failed: rc = -13 Lustre: 10286:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000402:0x3121:0x0] with magic=0xbd60bd0 Lustre: 10286:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 3 previous similar messages 1[22364]: segfault at 8 ip 00007fc0100c97e8 sp 00007ffd2d4c6990 error 4 in ld-2.17.so[7fc0100be000+22000] 8[22703]: segfault at 0 ip (null) sp 00007ffcf31f9848 error 14 in 8[400000+6000] LustreError: 23420:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802f5c112a8: inode [0x200000403:0x28ac:0x0] mdc close failed: rc = -13 LustreError: 23420:0:(file.c:247:ll_close_inode_openhandle()) Skipped 1 previous similar message 18[2530]: segfault at 8 ip 00007efdaf1c67e8 sp 00007fffd2ed0400 error 4 in ld-2.17.so[7efdaf1bb000+22000] 16[6185]: segfault at 8 ip 00007f2ec0b457e8 sp 00007fff63787aa0 error 4 in ld-2.17.so[7f2ec0b3a000+22000] 18[8984]: segfault at 8 ip 00007fc9e749b7e8 sp 00007fff8ae13540 error 4 in ld-2.17.so[7fc9e7490000+22000] LustreError: 11630:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802f5c112a8: inode [0x200000402:0x38e2:0x0] mdc close failed: rc = -13 LustreError: 11630:0:(file.c:247:ll_close_inode_openhandle()) Skipped 5 previous similar messages 13[14348]: segfault at 8 ip 00007feeca3e57e8 sp 00007ffc4582c1f0 error 4 in ld-2.17.so[7feeca3da000+22000] Lustre: 7153:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000402:0x3a5c:0x0] with magic=0xbd60bd0 Lustre: 7153:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message 14[6205]: segfault at 406000 ip 0000000000406000 sp 00007fff0b3bd358 error 14 in 14[606000+1000] 13[6270]: segfault at 8 ip 00007f9a09c1b7e8 sp 00007ffcc82fbbe0 error 4 in ld-2.17.so[7f9a09c10000+22000] 4[16644]: segfault at 8 ip 00007fc18c4317e8 sp 00007ffe9e400b00 error 4 in ld-2.17.so[7fc18c426000+22000] 12[16916]: segfault at 8 ip 00007efcb04817e8 sp 00007ffd55d937c0 error 4 in ld-2.17.so[7efcb0476000+22000] 13[22897]: segfault at 8 ip 00007f99c95ff7e8 sp 00007ffc4ba82a30 error 4 in ld-2.17.so[7f99c95f4000+22000] 15[23862]: segfault at 1c ip 00000000004017b0 sp 00007ffead4dd0d0 error 6 in 15[400000+6000] 3[24108]: segfault at 1c ip 00000000004017b0 sp 00007fff21673cf0 error 6 in 3[400000+6000] LustreError: 24108:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff88009c8d8008: inode [0x200000403:0x3df2:0x0] mdc close failed: rc = -13 LustreError: 24108:0:(file.c:247:ll_close_inode_openhandle()) Skipped 1 previous similar message 1[30522]: segfault at 8 ip 00007f49ea9577e8 sp 00007ffd177d79f0 error 4 in ld-2.17.so[7f49ea94c000+22000] 1[30695]: segfault at 8 ip 00007fccd611d7e8 sp 00007ffdaa8ff760 error 4 in ld-2.17.so[7fccd6112000+22000] 1[31382]: segfault at 8 ip 00007fbd774f17e8 sp 00007ffc9a428920 error 4 in ld-2.17.so[7fbd774e6000+22000] 5[31947]: segfault at 8 ip 00007ff549f4e7e8 sp 00007fffdf879e80 error 4 in ld-2.17.so[7ff549f43000+22000] Lustre: 6249:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000402:0x4b2f:0x0] with magic=0xbd60bd0 Lustre: 6249:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 7 previous similar messages 3[6376]: segfault at 8 ip 00007f5c9ce317e8 sp 00007ffe73616290 error 4 in ld-2.17.so[7f5c9ce26000+22000] 18[15293]: segfault at 8 ip 00007f64e33747e8 sp 00007ffcea484c30 error 4 in ld-2.17.so[7f64e3369000+22000] LustreError: 18540:0:(statahead.c:2382:start_statahead_thread()) lustre: invalid pattern 0X0. 4[30830]: segfault at 8 ip 00007f3c10fa37e8 sp 00007fffca8d4980 error 4 in ld-2.17.so[7f3c10f98000+22000] 2[447]: segfault at 8 ip 00007fe61a1d77e8 sp 00007ffdaa55dc80 error 4 in ld-2.17.so[7fe61a1cc000+22000] 18[7982]: segfault at 1c ip 00000000004017b0 sp 00007ffcedf49ea0 error 6 in 18[400000+6000] 1[9834]: segfault at 8 ip 00007f1fa34397e8 sp 00007fff97fd4880 error 4 in ld-2.17.so[7f1fa342e000+22000] 16[10026]: segfault at 8 ip 00007fed181d27e8 sp 00007ffcf4354590 error 4 in ld-2.17.so[7fed181c7000+22000] 6[10212]: segfault at 8 ip 00007f564a1c47e8 sp 00007ffc2cd99fd0 error 4 in ld-2.17.so[7f564a1b9000+22000] Lustre: 19781:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0002: opcode 7: before 516 < left 618, rollback = 7 Lustre: 19781:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 589 previous similar messages Lustre: 19781:0:(osd_handler.c:1967:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 19781:0:(osd_handler.c:1967:osd_trans_dump_creds()) Skipped 589 previous similar messages Lustre: 19781:0:(osd_handler.c:1974:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 19781:0:(osd_handler.c:1974:osd_trans_dump_creds()) Skipped 589 previous similar messages Lustre: 19781:0:(osd_handler.c:1984:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 19781:0:(osd_handler.c:1984:osd_trans_dump_creds()) Skipped 589 previous similar messages Lustre: 19781:0:(osd_handler.c:1991:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 19781:0:(osd_handler.c:1991:osd_trans_dump_creds()) Skipped 589 previous similar messages Lustre: 19781:0:(osd_handler.c:1998:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 19781:0:(osd_handler.c:1998:osd_trans_dump_creds()) Skipped 589 previous similar messages 8[30540]: segfault at 8 ip 00007f3b7dfe87e8 sp 00007fffed827a40 error 4 in ld-2.17.so[7f3b7dfdd000+22000] 0[30762]: segfault at 8 ip 00007fb09f63e7e8 sp 00007ffc6dcafd60 error 4 in ld-2.17.so[7fb09f633000+22000] 5[31261]: segfault at 8 ip 00007f98e5daa7e8 sp 00007ffde32e77b0 error 4 in ld-2.17.so[7f98e5d9f000+22000] 6[1485]: segfault at 0 ip 0000000000403e5f sp 00007ffcdd324a50 error 6 in 6[400000+6000] 10[10688]: segfault at 8 ip 00007f1c9245a7e8 sp 00007ffc0b17d8a0 error 4 in ld-2.17.so[7f1c9244f000+22000] 1[15080]: segfault at 8 ip 00007f648cf4d7e8 sp 00007fff79d62610 error 4 in ld-2.17.so[7f648cf42000+22000] 2[23113]: segfault at 8 ip 00007f1e5ad547e8 sp 00007ffd0e1a8af0 error 4 in ld-2.17.so[7f1e5ad49000+22000] Lustre: 2197:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0x6157:0x0] with magic=0xbd60bd0 Lustre: 2197:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 25 previous similar messages LustreError: 1979:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802f5c112a8: inode [0x200000403:0x64b8:0x0] mdc close failed: rc = -13 LustreError: 1979:0:(file.c:247:ll_close_inode_openhandle()) Skipped 5 previous similar messages 3[18907]: segfault at 8 ip 00007f2292bdc7e8 sp 00007ffe4f8b09d0 error 4 in ld-2.17.so[7f2292bd1000+22000] 14[19749]: segfault at 8 ip 00007faad14387e8 sp 00007ffd1d651de0 error 4 in ld-2.17.so[7faad142d000+22000] 8[7193]: segfault at 8 ip 00007fc8d6cee7e8 sp 00007fff22767200 error 4 in ld-2.17.so[7fc8d6ce3000+22000] 3[13661]: segfault at 8 ip 00007f8b285e67e8 sp 00007ffe4ac38140 error 4 in ld-2.17.so[7f8b285db000+22000] 8[15165]: segfault at 8 ip 00007fd57a83d7e8 sp 00007ffe89255810 error 4 in ld-2.17.so[7fd57a832000+22000] traps: 4[25869] trap invalid opcode ip:405b84 sp:7ffdf16efda8 error:0 in 4[400000+6000] 9[3869]: segfault at 8 ip 00007f1645e9b7e8 sp 00007ffed0f57f30 error 4 in ld-2.17.so[7f1645e90000+22000] 9[4652]: segfault at 8 ip 00007f8d74db97e8 sp 00007fffcc03ddd0 error 4 in ld-2.17.so[7f8d74dae000+22000] | Link to test |
racer test 1: racer on clients: centos-5.localnet DURATION=2700 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 PGD 2d72d7067 PUD 2c8f4c067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) lustre(OE) osp(OE) ofd(OE) lod(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) jbd2 mbcache lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) crc32_generic lnet(OE) dm_flakey dm_mod libcfs(OE) crc_t10dif crct10dif_generic crct10dif_common virtio_balloon pcspkr virtio_console i2c_piix4 ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm ata_piix drm drm_panel_orientation_quirks libata serio_raw i2c_core virtio_blk floppy CPU: 14 PID: 12959 Comm: ll_sa_12939 Kdump: loaded Tainted: P OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.16.0-3.module_el8.7.0+1218+f626c2ff 04/01/2014 task: ffff8802ae6c24f0 ti: ffff8802abbb0000 task.ti: ffff8802abbb0000 RIP: 0010:[<ffffffff813f502c>] [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 RSP: 0018:ffff8802abbb3d80 EFLAGS: 00010282 RAX: ffff88029db44f08 RBX: 0000000000000008 RCX: 0000000100260023 RDX: 0000000000000026 RSI: ffff88029db45238 RDI: 0000000000000008 RBP: ffff8802abbb3d90 R08: ffff8800a7a527c8 R09: 0000000000000000 R10: ffff8800a7a53508 R11: ffff8800a7a527c8 R12: 0000000000000000 R13: ffff8802df7b8010 R14: ffff88029db45238 R15: ffff8800a7a527c8 FS: 0000000000000000(0000) GS:ffff880331d80000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 00000002ab542000 CR4: 00000000000007e0 Call Trace: [<ffffffffa10f6171>] ll_statahead_thread+0xde1/0x2290 [lustre] [<ffffffff810df6e2>] ? pick_next_task_fair+0x702/0x8a0 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e329e>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffff810c834d>] ? finish_task_switch+0x5d/0x1b0 [<ffffffffa10f5390>] ? ll_statahead_handle.constprop.29+0x1c0/0x1c0 [lustre] [<ffffffff810ba114>] kthread+0xe4/0xf0 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 | Lustre: 11560:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 510 < left 699, rollback = 2 Lustre: 11560:0:(osd_handler.c:1951:osd_trans_dump_creds()) create: 3/12/2, destroy: 0/0/0 Lustre: 11560:0:(osd_handler.c:1958:osd_trans_dump_creds()) attr_set: 2/2/0, xattr_set: 9/699/0 Lustre: 11560:0:(osd_handler.c:1968:osd_trans_dump_creds()) write: 5/55/0, punch: 0/0/0, quota 1/3/0 Lustre: 11560:0:(osd_handler.c:1975:osd_trans_dump_creds()) insert: 10/203/3, delete: 0/0/0 Lustre: 11560:0:(osd_handler.c:1982:osd_trans_dump_creds()) ref_add: 4/4/0, ref_del: 0/0/0 LustreError: 11560:0:(mdt_reint.c:2533:mdt_reint_migrate()) lustre-MDT0000: migrate [0x200000403:0x2:0x0]/17 failed: rc = -114 Lustre: 17972:0:(mdt_recovery.c:148:mdt_req_from_lrd()) @@@ restoring transno req@ffff8802df63c140 x1821918876998016/t4294967391(0) o101->e10def5c-7ac9-4980-8ea5-de2ed0387bd1@0@lo:325/0 lens 376/864 e 0 to 0 dl 1737517310 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 LustreError: 11560:0:(mdt_reint.c:2533:mdt_reint_migrate()) lustre-MDT0001: migrate [0x240000403:0x1:0x0]/1 failed: rc = -114 LustreError: 18234:0:(file.c:267:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e68953d8: inode [0x240000404:0x4:0x0] mdc close failed: rc = -116 5[17719]: segfault at 8 ip 00007f557911c7e8 sp 00007ffff0a0aab0 error 4 in ld-2.17.so[7f5579111000+22000] Lustre: 18087:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 510 < left 727, rollback = 2 Lustre: 18087:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 7 previous similar messages Lustre: 18087:0:(osd_handler.c:1951:osd_trans_dump_creds()) create: 3/12/2, destroy: 0/0/0 Lustre: 18087:0:(osd_handler.c:1951:osd_trans_dump_creds()) Skipped 7 previous similar messages Lustre: 18087:0:(osd_handler.c:1958:osd_trans_dump_creds()) attr_set: 4/4/0, xattr_set: 11/727/0 Lustre: 18087:0:(osd_handler.c:1958:osd_trans_dump_creds()) Skipped 7 previous similar messages Lustre: 18087:0:(osd_handler.c:1968:osd_trans_dump_creds()) write: 5/55/0, punch: 0/0/0, quota 1/3/0 Lustre: 18087:0:(osd_handler.c:1968:osd_trans_dump_creds()) Skipped 7 previous similar messages Lustre: 18087:0:(osd_handler.c:1975:osd_trans_dump_creds()) insert: 12/215/3, delete: 0/0/0 Lustre: 18087:0:(osd_handler.c:1975:osd_trans_dump_creds()) Skipped 7 previous similar messages Lustre: 18087:0:(osd_handler.c:1982:osd_trans_dump_creds()) ref_add: 6/6/0, ref_del: 0/0/0 Lustre: 18087:0:(osd_handler.c:1982:osd_trans_dump_creds()) Skipped 7 previous similar messages LustreError: 11560:0:(mdt_reint.c:2533:mdt_reint_migrate()) lustre-MDT0000: migrate [0x200000403:0x1:0x0]/8 failed: rc = -114 LustreError: 11560:0:(mdt_reint.c:2533:mdt_reint_migrate()) Skipped 2 previous similar messages Lustre: 18087:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 513 < left 788, rollback = 2 Lustre: 18087:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 7 previous similar messages Lustre: 18087:0:(osd_handler.c:1951:osd_trans_dump_creds()) create: 3/12/0, destroy: 0/0/0 Lustre: 18087:0:(osd_handler.c:1951:osd_trans_dump_creds()) Skipped 7 previous similar messages Lustre: 18087:0:(osd_handler.c:1958:osd_trans_dump_creds()) attr_set: 2/2/0, xattr_set: 10/788/0 Lustre: 18087:0:(osd_handler.c:1958:osd_trans_dump_creds()) Skipped 7 previous similar messages Lustre: 18087:0:(osd_handler.c:1968:osd_trans_dump_creds()) write: 5/55/0, punch: 0/0/0, quota 1/3/0 Lustre: 18087:0:(osd_handler.c:1968:osd_trans_dump_creds()) Skipped 7 previous similar messages Lustre: 18087:0:(osd_handler.c:1975:osd_trans_dump_creds()) insert: 11/239/2, delete: 0/0/0 Lustre: 18087:0:(osd_handler.c:1975:osd_trans_dump_creds()) Skipped 7 previous similar messages Lustre: 18087:0:(osd_handler.c:1982:osd_trans_dump_creds()) ref_add: 5/5/0, ref_del: 0/0/0 Lustre: 18087:0:(osd_handler.c:1982:osd_trans_dump_creds()) Skipped 7 previous similar messages LustreError: 11560:0:(mdt_reint.c:2533:mdt_reint_migrate()) lustre-MDT0001: migrate [0x240000403:0x1:0x0]/8 failed: rc = -16 LustreError: 11560:0:(mdt_reint.c:2533:mdt_reint_migrate()) Skipped 1 previous similar message 11[19030]: segfault at 8 ip 00007ff3b7a847e8 sp 00007ffd2ab00aa0 error 4 in ld-2.17.so[7ff3b7a79000+22000] 10[19050]: segfault at 8 ip 00007fc4908e57e8 sp 00007fffafdd6310 error 4 in ld-2.17.so[7fc4908da000+22000] Lustre: 18866:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 511 < left 1278, rollback = 2 Lustre: 18866:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 31 previous similar messages Lustre: 18866:0:(osd_handler.c:1951:osd_trans_dump_creds()) create: 3/12/2, destroy: 1/4/0 Lustre: 18866:0:(osd_handler.c:1951:osd_trans_dump_creds()) Skipped 31 previous similar messages Lustre: 18866:0:(osd_handler.c:1958:osd_trans_dump_creds()) attr_set: 4/4/0, xattr_set: 17/1278/0 Lustre: 18866:0:(osd_handler.c:1958:osd_trans_dump_creds()) Skipped 31 previous similar messages Lustre: 18866:0:(osd_handler.c:1968:osd_trans_dump_creds()) write: 5/55/0, punch: 0/0/0, quota 7/105/0 Lustre: 18866:0:(osd_handler.c:1968:osd_trans_dump_creds()) Skipped 31 previous similar messages Lustre: 18866:0:(osd_handler.c:1975:osd_trans_dump_creds()) insert: 15/263/2, delete: 3/6/0 Lustre: 18866:0:(osd_handler.c:1975:osd_trans_dump_creds()) Skipped 31 previous similar messages Lustre: 18866:0:(osd_handler.c:1982:osd_trans_dump_creds()) ref_add: 8/8/0, ref_del: 3/3/0 Lustre: 18866:0:(osd_handler.c:1982:osd_trans_dump_creds()) Skipped 31 previous similar messages LustreError: 18087:0:(mdt_reint.c:2533:mdt_reint_migrate()) lustre-MDT0001: migrate [0x240000403:0x1:0x0]/8 failed: rc = -114 LustreError: 18087:0:(mdt_reint.c:2533:mdt_reint_migrate()) Skipped 9 previous similar messages Lustre: 17468:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 515 < left 618, rollback = 7 Lustre: 20169:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 501 < left 699, rollback = 2 Lustre: 20169:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 14 previous similar messages Lustre: 20169:0:(osd_handler.c:1951:osd_trans_dump_creds()) create: 3/12/10, destroy: 0/0/0 Lustre: 20169:0:(osd_handler.c:1951:osd_trans_dump_creds()) Skipped 16 previous similar messages Lustre: 20169:0:(osd_handler.c:1958:osd_trans_dump_creds()) attr_set: 2/2/0, xattr_set: 9/699/0 Lustre: 20169:0:(osd_handler.c:1958:osd_trans_dump_creds()) Skipped 16 previous similar messages Lustre: 20169:0:(osd_handler.c:1968:osd_trans_dump_creds()) write: 5/55/0, punch: 0/0/0, quota 4/54/0 Lustre: 20169:0:(osd_handler.c:1968:osd_trans_dump_creds()) Skipped 16 previous similar messages Lustre: 20169:0:(osd_handler.c:1975:osd_trans_dump_creds()) insert: 11/239/4, delete: 0/0/0 Lustre: 20169:0:(osd_handler.c:1975:osd_trans_dump_creds()) Skipped 16 previous similar messages Lustre: 20169:0:(osd_handler.c:1982:osd_trans_dump_creds()) ref_add: 5/5/0, ref_del: 0/0/0 Lustre: 20169:0:(osd_handler.c:1982:osd_trans_dump_creds()) Skipped 16 previous similar messages LustreError: 21302:0:(file.c:267:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e68953d8: inode [0x200000403:0x6f:0x0] mdc close failed: rc = -116 LustreError: 21302:0:(file.c:267:ll_close_inode_openhandle()) Skipped 1 previous similar message Lustre: 18845:0:(mdt_recovery.c:148:mdt_req_from_lrd()) @@@ restoring transno req@ffff8802d0ce9e40 x1821918880584704/t4294968939(0) o101->e10def5c-7ac9-4980-8ea5-de2ed0387bd1@0@lo:339/0 lens 376/864 e 0 to 0 dl 1737517324 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 LustreError: 21863:0:(file.c:267:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e5f5dd28: inode [0x200000404:0xa8:0x0] mdc close failed: rc = -116 Lustre: 17994:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000404:0x8e:0x0] with magic=0xbd60bd0 LustreError: 23383:0:(lcommon_cl.c:195:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000403:0x82:0x0]: rc = -5 LustreError: 23383:0:(llite_lib.c:3755:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 11559:0:(mdt_reint.c:2533:mdt_reint_migrate()) lustre-MDT0000: migrate [0x200000403:0x1:0x0]/8 failed: rc = -16 LustreError: 11559:0:(mdt_reint.c:2533:mdt_reint_migrate()) Skipped 23 previous similar messages Lustre: dir [0x280000404:0xb6:0x0] stripe 1 readdir failed: -2, directory is partially accessed! LustreError: 19018:0:(mdd_object.c:3861:mdd_close()) lustre-MDD0001: failed to get lu_attr of [0x240000404:0x5e:0x0]: rc = -2 LustreError: 22119:0:(file.c:267:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e5f5dd28: inode [0x240000404:0x5e:0x0] mdc close failed: rc = -2 LustreError: 22119:0:(file.c:267:ll_close_inode_openhandle()) Skipped 1 previous similar message 6[24292]: segfault at 8 ip 00007f9bb2ce27e8 sp 00007ffee6f42e70 error 4 in ld-2.17.so[7f9bb2cd7000+22000] LustreError: 25059:0:(lcommon_cl.c:195:cl_file_inode_init()) lustre: failed to initialize cl_object [0x280000404:0x13d:0x0]: rc = -5 LustreError: 25059:0:(llite_lib.c:3755:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 12[25263]: segfault at 8 ip 00007f78a8c2c7e8 sp 00007ffdc6697810 error 4 in ld-2.17.so[7f78a8c21000+22000] Lustre: 20183:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 509 < left 638, rollback = 2 Lustre: 20183:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 196 previous similar messages Lustre: 20183:0:(osd_handler.c:1951:osd_trans_dump_creds()) create: 3/12/2, destroy: 0/0/0 Lustre: 20183:0:(osd_handler.c:1951:osd_trans_dump_creds()) Skipped 196 previous similar messages Lustre: 20183:0:(osd_handler.c:1958:osd_trans_dump_creds()) attr_set: 4/4/0, xattr_set: 10/638/0 Lustre: 20183:0:(osd_handler.c:1958:osd_trans_dump_creds()) Skipped 196 previous similar messages Lustre: 20183:0:(osd_handler.c:1968:osd_trans_dump_creds()) write: 5/55/0, punch: 0/0/0, quota 7/81/0 Lustre: 20183:0:(osd_handler.c:1968:osd_trans_dump_creds()) Skipped 196 previous similar messages Lustre: 20183:0:(osd_handler.c:1975:osd_trans_dump_creds()) insert: 13/251/4, delete: 0/0/0 Lustre: 20183:0:(osd_handler.c:1975:osd_trans_dump_creds()) Skipped 196 previous similar messages Lustre: 20183:0:(osd_handler.c:1982:osd_trans_dump_creds()) ref_add: 7/7/0, ref_del: 0/0/0 Lustre: 20183:0:(osd_handler.c:1982:osd_trans_dump_creds()) Skipped 196 previous similar messages Lustre: 17662:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000404:0x135:0x0] with magic=0xbd60bd0 Lustre: 17662:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message LustreError: 24487:0:(llite_nfs.c:446:ll_dir_get_parent_fid()) lustre: failure inode [0x240000404:0x168:0x0] get parent: rc = -116 LustreError: 26631:0:(lcommon_cl.c:195:cl_file_inode_init()) lustre: failed to initialize cl_object [0x280000404:0x13d:0x0]: rc = -5 LustreError: 26631:0:(lcommon_cl.c:195:cl_file_inode_init()) Skipped 2 previous similar messages LustreError: 26631:0:(llite_lib.c:3755:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 26631:0:(llite_lib.c:3755:ll_prep_inode()) Skipped 2 previous similar messages Lustre: 17785:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000404:0x187:0x0] with magic=0xbd60bd0 Lustre: 17785:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message Lustre: 17468:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 516 < left 618, rollback = 7 Lustre: 17468:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 1 previous similar message 19[28432]: segfault at 8 ip 00007f8bee3c67e8 sp 00007fff8f97f6e0 error 4 in ld-2.17.so[7f8bee3bb000+22000] LustreError: 28725:0:(lcommon_cl.c:195:cl_file_inode_init()) lustre: failed to initialize cl_object [0x280000404:0x13d:0x0]: rc = -5 LustreError: 28725:0:(lcommon_cl.c:195:cl_file_inode_init()) Skipped 2 previous similar messages LustreError: 28725:0:(llite_lib.c:3755:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 28725:0:(llite_lib.c:3755:ll_prep_inode()) Skipped 2 previous similar messages 6[29013]: segfault at 8 ip 00007f7fb94de7e8 sp 00007ffeea913540 error 4 in ld-2.17.so[7f7fb94d3000+22000] Lustre: 18424:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0002: opcode 7: before 516 < left 618, rollback = 7 Lustre: 18424:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 9 previous similar messages Lustre: 20883:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 618, rollback = 7 Lustre: 20883:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 7 previous similar messages 3[32699]: segfault at 8 ip 00007fd5dd1e17e8 sp 00007fff8658c5a0 error 4 in ld-2.17.so[7fd5dd1d6000+22000] Lustre: 13553:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0000: opcode 7: before 516 < left 618, rollback = 7 Lustre: 13553:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 9 previous similar messages Lustre: 13553:0:(osd_handler.c:1951:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 13553:0:(osd_handler.c:1951:osd_trans_dump_creds()) Skipped 77 previous similar messages Lustre: 13553:0:(osd_handler.c:1958:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 13553:0:(osd_handler.c:1958:osd_trans_dump_creds()) Skipped 77 previous similar messages Lustre: 13553:0:(osd_handler.c:1968:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 13553:0:(osd_handler.c:1968:osd_trans_dump_creds()) Skipped 77 previous similar messages Lustre: 13553:0:(osd_handler.c:1975:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 13553:0:(osd_handler.c:1975:osd_trans_dump_creds()) Skipped 77 previous similar messages Lustre: 13553:0:(osd_handler.c:1982:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 13553:0:(osd_handler.c:1982:osd_trans_dump_creds()) Skipped 77 previous similar messages Lustre: 11540:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000404:0x13d:0x0] with magic=0xbd60bd0 Lustre: 11540:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 11 previous similar messages LustreError: 21502:0:(mdt_reint.c:2533:mdt_reint_migrate()) lustre-MDT0001: migrate [0x240000403:0x1:0x0]/13 failed: rc = -114 LustreError: 21502:0:(mdt_reint.c:2533:mdt_reint_migrate()) Skipped 14 previous similar messages LustreError: 25654:0:(file.c:267:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e5f5dd28: inode [0x240000403:0x192:0x0] mdc close failed: rc = -116 LustreError: 25654:0:(file.c:267:ll_close_inode_openhandle()) Skipped 3 previous similar messages Lustre: 11560:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 510 < left 1500, rollback = 2 Lustre: 11560:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 49 previous similar messages 18[872]: segfault at 0 ip 0000000000401870 sp 00007fff20b5f9c8 error 6 in 18[400000+6000] LustreError: 4505:0:(lcommon_cl.c:195:cl_file_inode_init()) lustre: failed to initialize cl_object [0x280000403:0x20e:0x0]: rc = -5 LustreError: 4505:0:(lcommon_cl.c:195:cl_file_inode_init()) Skipped 2 previous similar messages LustreError: 4505:0:(llite_lib.c:3755:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 4505:0:(llite_lib.c:3755:ll_prep_inode()) Skipped 2 previous similar messages Lustre: 14402:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0x2ce:0x0] with magic=0xbd60bd0 Lustre: 14402:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 3 previous similar messages Lustre: 17521:0:(mdt_recovery.c:148:mdt_req_from_lrd()) @@@ restoring transno req@ffff8802ded49e40 x1821918893226368/t4294975221(0) o101->7c954870-36d3-41b9-9abe-a1502ef09ad8@0@lo:374/0 lens 376/816 e 0 to 0 dl 1737517359 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 Lustre: 18424:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0000: opcode 7: before 516 < left 618, rollback = 7 Lustre: 18424:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 13 previous similar messages 0[6503]: segfault at 8 ip 00007f3b36afa7e8 sp 00007ffc3fdd7af0 error 4 in ld-2.17.so[7f3b36aef000+22000] LustreError: 9199:0:(lcommon_cl.c:195:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000404:0x261:0x0]: rc = -5 LustreError: 9199:0:(lcommon_cl.c:195:cl_file_inode_init()) Skipped 1 previous similar message LustreError: 9199:0:(llite_lib.c:3755:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 9199:0:(llite_lib.c:3755:ll_prep_inode()) Skipped 1 previous similar message LustreError: 872:0:(file.c:267:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e68953d8: inode [0x200000404:0x275:0x0] mdc close failed: rc = -13 LustreError: 872:0:(file.c:267:ll_close_inode_openhandle()) Skipped 3 previous similar messages Lustre: 20883:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0002: opcode 7: before 516 < left 618, rollback = 7 Lustre: 20883:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 9 previous similar messages 16[12531]: segfault at 8 ip 00007f2b7bd8b7e8 sp 00007ffee9b8d130 error 4 in ld-2.17.so[7f2b7bd80000+22000] Lustre: 24435:0:(osd_handler.c:1951:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 24435:0:(osd_handler.c:1951:osd_trans_dump_creds()) Skipped 296 previous similar messages Lustre: 24435:0:(osd_handler.c:1958:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 24435:0:(osd_handler.c:1958:osd_trans_dump_creds()) Skipped 296 previous similar messages Lustre: 24435:0:(osd_handler.c:1968:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 24435:0:(osd_handler.c:1968:osd_trans_dump_creds()) Skipped 296 previous similar messages Lustre: 24435:0:(osd_handler.c:1975:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 24435:0:(osd_handler.c:1975:osd_trans_dump_creds()) Skipped 296 previous similar messages Lustre: 24435:0:(osd_handler.c:1982:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 24435:0:(osd_handler.c:1982:osd_trans_dump_creds()) Skipped 296 previous similar messages 3[15792]: segfault at 1c ip 00000000004017b0 sp 00007ffdf6569590 error 6 in 10[400000+6000] LustreError: 18953:0:(lcommon_cl.c:195:cl_file_inode_init()) lustre: failed to initialize cl_object [0x280000403:0x20e:0x0]: rc = -5 LustreError: 18953:0:(lcommon_cl.c:195:cl_file_inode_init()) Skipped 1 previous similar message LustreError: 18953:0:(llite_lib.c:3755:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 18953:0:(llite_lib.c:3755:ll_prep_inode()) Skipped 1 previous similar message Lustre: 13553:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 618, rollback = 7 Lustre: 13553:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 17 previous similar messages 4[25666]: segfault at 1c ip 00000000004017b0 sp 00007ffc827bd730 error 6 in 4[400000+6000] LustreError: 21992:0:(mdt_reint.c:2533:mdt_reint_migrate()) lustre-MDT0002: migrate [0x280000403:0x1:0x0]/6 failed: rc = -16 LustreError: 21992:0:(mdt_reint.c:2533:mdt_reint_migrate()) Skipped 10 previous similar messages LustreError: 27151:0:(file.c:267:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e5f5dd28: inode [0x280000404:0x423:0x0] mdc close failed: rc = -116 Lustre: 11540:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000404:0x423:0x0] with magic=0xbd60bd0 Lustre: 11540:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message Lustre: 21992:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 499 < left 966, rollback = 2 Lustre: 21992:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 262 previous similar messages 5[29578]: segfault at 406000 ip 0000000000406000 sp 00007fffe1f507b8 error 14 in 5[606000+1000] LustreError: 11541:0:(mdd_object.c:3861:mdd_close()) lustre-MDD0001: failed to get lu_attr of [0x240000403:0x1c0:0x0]: rc = -2 LustreError: 30902:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8802e5f5dd28: cannot apply new layout on [0x280000403:0x20e:0x0] : rc = -5 LustreError: 30902:0:(vvp_io.c:1921:vvp_io_init()) lustre: refresh file layout [0x280000403:0x20e:0x0] error -5. 6[31159]: segfault at 8 ip 00007f96332b57e8 sp 00007ffd7fb4c800 error 4 in ld-2.17.so[7f96332aa000+22000] 5[32384]: segfault at 406000 ip 0000000000406000 sp 00007ffc4825b218 error 14 in 5[606000+1000] LustreError: 11541:0:(mdd_object.c:3861:mdd_close()) lustre-MDD0001: failed to get lu_attr of [0x240000404:0x42c:0x0]: rc = -2 Lustre: 18129:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000404:0x566:0x0] with magic=0xbd60bd0 Lustre: 18129:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message Lustre: 21992:0:(osd_handler.c:1951:osd_trans_dump_creds()) create: 3/12/4, destroy: 0/0/0 Lustre: 21992:0:(osd_handler.c:1951:osd_trans_dump_creds()) Skipped 435 previous similar messages Lustre: 21992:0:(osd_handler.c:1958:osd_trans_dump_creds()) attr_set: 4/4/0, xattr_set: 12/816/0 Lustre: 21992:0:(osd_handler.c:1958:osd_trans_dump_creds()) Skipped 435 previous similar messages Lustre: 21992:0:(osd_handler.c:1968:osd_trans_dump_creds()) write: 5/55/0, punch: 0/0/0, quota 4/54/0 Lustre: 21992:0:(osd_handler.c:1968:osd_trans_dump_creds()) Skipped 435 previous similar messages Lustre: 21992:0:(osd_handler.c:1975:osd_trans_dump_creds()) insert: 12/215/3, delete: 0/0/0 Lustre: 21992:0:(osd_handler.c:1975:osd_trans_dump_creds()) Skipped 435 previous similar messages Lustre: 21992:0:(osd_handler.c:1982:osd_trans_dump_creds()) ref_add: 6/6/0, ref_del: 0/0/0 Lustre: 21992:0:(osd_handler.c:1982:osd_trans_dump_creds()) Skipped 435 previous similar messages LustreError: 4989:0:(lcommon_cl.c:195:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000404:0x453:0x0]: rc = -5 LustreError: 4989:0:(lcommon_cl.c:195:cl_file_inode_init()) Skipped 2 previous similar messages LustreError: 4989:0:(llite_lib.c:3755:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 4989:0:(llite_lib.c:3755:ll_prep_inode()) Skipped 2 previous similar messages LustreError: 2398:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8802e68953d8: cannot apply new layout on [0x240000404:0x453:0x0] : rc = -5 LustreError: 2398:0:(vvp_io.c:1921:vvp_io_init()) lustre: refresh file layout [0x240000404:0x453:0x0] error -5. Lustre: dir [0x280000403:0x537:0x0] stripe 3 readdir failed: -2, directory is partially accessed! Lustre: Skipped 5 previous similar messages LustreError: 5759:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8802e68953d8: cannot apply new layout on [0x240000404:0x453:0x0] : rc = -5 LustreError: 5065:0:(file.c:267:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e5f5dd28: inode [0x240000404:0x4c4:0x0] mdc close failed: rc = -116 LustreError: 5065:0:(file.c:267:ll_close_inode_openhandle()) Skipped 14 previous similar messages LustreError: 19018:0:(mdd_object.c:3861:mdd_close()) lustre-MDD0001: failed to get lu_attr of [0x240000403:0x40a:0x0]: rc = -2 LustreError: 17392:0:(mdd_object.c:384:mdd_xattr_get()) lustre-MDD0002: object [0x280000403:0x5cd:0x0] not found: rc = -2 LustreError: 10872:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8802e5f5dd28: cannot apply new layout on [0x280000403:0x20e:0x0] : rc = -5 LustreError: 11541:0:(mdd_object.c:3861:mdd_close()) lustre-MDD0002: failed to get lu_attr of [0x280000404:0x6a4:0x0]: rc = -2 Lustre: 24435:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0000: opcode 7: before 516 < left 618, rollback = 7 Lustre: 24435:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 97 previous similar messages Lustre: 16876:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000403:0x5e5:0x0] with magic=0xbd60bd0 Lustre: 16876:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 27 previous similar messages 8[16730]: segfault at 8 ip 00007f706eae97e8 sp 00007ffca3012190 error 4 in ld-2.17.so[7f706eade000+22000] LustreError: 20183:0:(mdt_reint.c:2533:mdt_reint_migrate()) lustre-MDT0000: migrate [0x280000403:0x636:0x0]/sleep failed: rc = -114 LustreError: 20183:0:(mdt_reint.c:2533:mdt_reint_migrate()) Skipped 88 previous similar messages Lustre: 21479:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 508 < left 816, rollback = 2 Lustre: 21479:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 865 previous similar messages Lustre: dir [0x280000404:0x6c4:0x0] stripe 2 readdir failed: -2, directory is partially accessed! LustreError: 18851:0:(mdt_xattr.c:422:mdt_dir_layout_update()) lustre-MDT0001: [0x240000403:0x644:0x0] migrate mdt count mismatch 1 != 2 LustreError: 14828:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8802e68953d8: cannot apply new layout on [0x200000404:0x8a3:0x0] : rc = -5 LustreError: 14828:0:(lov_object.c:1341:lov_layout_change()) Skipped 2 previous similar messages LustreError: 14828:0:(vvp_io.c:1921:vvp_io_init()) lustre: refresh file layout [0x200000404:0x8a3:0x0] error -5. LustreError: 20169:0:(mdd_dir.c:4619:mdd_migrate_cmd_check()) lustre-MDD0002: '10' migration was interrupted, run 'lfs migrate -m 2 -c 1 -H crush 10' to finish migration: rc = -1 Lustre: dir [0x200000404:0xb61:0x0] stripe 3 readdir failed: -2, directory is partially accessed! Lustre: Skipped 2 previous similar messages LustreError: 22153:0:(lcommon_cl.c:195:cl_file_inode_init()) lustre: failed to initialize cl_object [0x280000403:0x6a1:0x0]: rc = -5 LustreError: 22153:0:(lcommon_cl.c:195:cl_file_inode_init()) Skipped 17 previous similar messages LustreError: 22153:0:(llite_lib.c:3755:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 22153:0:(llite_lib.c:3755:ll_prep_inode()) Skipped 17 previous similar messages 4[22079]: segfault at 8 ip 00007fea58eb07e8 sp 00007ffdf2a77a50 error 4 in ld-2.17.so[7fea58ea5000+22000] 5[21301]: segfault at 8 ip 00007f51af2267e8 sp 00007ffd9fc17890 error 4 in ld-2.17.so[7f51af21b000+22000] LustreError: 22645:0:(file.c:267:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e5f5dd28: inode [0x200000403:0xac5:0x0] mdc close failed: rc = -116 LustreError: 22645:0:(file.c:267:ll_close_inode_openhandle()) Skipped 15 previous similar messages 13[27045]: segfault at 0 ip (null) sp 00007fff0ccc07d8 error 14 in 13[400000+6000] 2[27360]: segfault at 8 ip 00007f89eee0d7e8 sp 00007ffeaf3e9e60 error 4 in ld-2.17.so[7f89eee02000+22000] 10[27619]: segfault at 8 ip 00007fd6435597e8 sp 00007ffd6b85ddd0 error 4 in ld-2.17.so[7fd64354e000+22000] 15[29460]: segfault at 8 ip 00007fd728e8e7e8 sp 00007ffc41006b60 error 4 in ld-2.17.so[7fd728e83000+22000] traps: 2[2098] trap invalid opcode ip:4051aa sp:7ffecbc8d658 error:0 in 2[400000+6000] 12[5229]: segfault at 0 ip 0000000000403e5f sp 00007fffba04db30 error 6 in 12[400000+6000] Lustre: 24435:0:(osd_handler.c:1951:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 24435:0:(osd_handler.c:1951:osd_trans_dump_creds()) Skipped 742 previous similar messages Lustre: 24435:0:(osd_handler.c:1958:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 24435:0:(osd_handler.c:1958:osd_trans_dump_creds()) Skipped 742 previous similar messages Lustre: 24435:0:(osd_handler.c:1968:osd_trans_dump_creds()) write: 2/546/0, punch: 0/0/0, quota 1/3/0 Lustre: 24435:0:(osd_handler.c:1968:osd_trans_dump_creds()) Skipped 742 previous similar messages Lustre: 24435:0:(osd_handler.c:1975:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 24435:0:(osd_handler.c:1975:osd_trans_dump_creds()) Skipped 742 previous similar messages Lustre: 24435:0:(osd_handler.c:1982:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 24435:0:(osd_handler.c:1982:osd_trans_dump_creds()) Skipped 742 previous similar messages 3[20713]: segfault at 8 ip 00007fd4238947e8 sp 00007ffe53b8d000 error 4 in ld-2.17.so[7fd423889000+22000] Lustre: 20883:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 516 < left 618, rollback = 7 Lustre: 20883:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 47 previous similar messages LustreError: 19694:0:(mdt_reint.c:2533:mdt_reint_migrate()) lustre-MDT0000: migrate [0x200000403:0x1:0x0]/5 failed: rc = -114 LustreError: 19694:0:(mdt_reint.c:2533:mdt_reint_migrate()) Skipped 44 previous similar messages Lustre: 18850:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000403:0x89b:0x0] with magic=0xbd60bd0 Lustre: 18850:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 7 previous similar messages 16[21508]: segfault at 8 ip 00007ff98ccd37e8 sp 00007fff304abc20 error 4 in ld-2.17.so[7ff98ccc8000+22000] Lustre: 20169:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 502 < left 1011, rollback = 2 Lustre: 20169:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 160 previous similar messages 4[24505]: segfault at 8 ip 00007f58b18ac7e8 sp 00007ffc6110a7b0 error 4 in ld-2.17.so[7f58b18a1000+22000] LustreError: 19444:0:(mdd_object.c:3861:mdd_close()) lustre-MDD0001: failed to get lu_attr of [0x240000404:0xab2:0x0]: rc = -2 LustreError: 1631:0:(file.c:267:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e68953d8: inode [0x240000404:0xb26:0x0] mdc close failed: rc = -116 LustreError: 1631:0:(file.c:267:ll_close_inode_openhandle()) Skipped 19 previous similar messages Lustre: mdt00_025: service thread pid 18145 was inactive for 40.166 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 18145, comm: mdt00_025 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x913/0xd50 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x800 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock+0x88/0x1c0 [mdt] [<0>] mdt_object_find_lock+0x54/0x170 [mdt] [<0>] mdt_reint_setxattr+0x133/0x10e0 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x74c/0xbc0 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 18129, comm: mdt00_023 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x913/0xd50 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x800 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock_try+0xa0/0x250 [mdt] [<0>] mdt_getattr_name_lock+0x173b/0x2d00 [mdt] [<0>] mdt_intent_getattr+0x2cc/0x4e0 [mdt] [<0>] mdt_intent_opc.constprop.73+0x211/0xc60 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x397/0x980 [ptlrpc] [<0>] ldlm_handle_enqueue+0x507/0x1870 [ptlrpc] [<0>] tgt_enqueue+0x68/0x240 [ptlrpc] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: mdt00_033: service thread pid 1413 was inactive for 40.083 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Lustre: Skipped 1 previous similar message Pid: 1413, comm: mdt00_033 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x913/0xd50 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x800 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock_try+0xa0/0x250 [mdt] [<0>] mdt_getattr_name_lock+0x173b/0x2d00 [mdt] [<0>] mdt_intent_getattr+0x2cc/0x4e0 [mdt] [<0>] mdt_intent_opc.constprop.73+0x211/0xc60 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x397/0x980 [ptlrpc] [<0>] ldlm_handle_enqueue+0x507/0x1870 [ptlrpc] [<0>] tgt_enqueue+0x68/0x240 [ptlrpc] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe LustreError: 19743:0:(mdd_dir.c:4619:mdd_migrate_cmd_check()) lustre-MDD0000: '19' migration was interrupted, run 'lfs migrate -m 1 -c 3 -H crush 19' to finish migration: rc = -1 LustreError: 5599:0:(lcommon_cl.c:195:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000404:0x1417:0x0]: rc = -5 LustreError: 5599:0:(lcommon_cl.c:195:cl_file_inode_init()) Skipped 11 previous similar messages LustreError: 5599:0:(llite_lib.c:3755:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 5599:0:(llite_lib.c:3755:ll_prep_inode()) Skipped 11 previous similar messages Lustre: dir [0x280000403:0xc48:0x0] stripe 3 readdir failed: -2, directory is partially accessed! Lustre: Skipped 2 previous similar messages LustreError: 19743:0:(mdd_dir.c:4619:mdd_migrate_cmd_check()) lustre-MDD0000: '5' migration was interrupted, run 'lfs migrate -m 1 -c 1 -H crush 5' to finish migration: rc = -1 Lustre: dir [0x280000403:0xd50:0x0] stripe 2 readdir failed: -2, directory is partially accessed! Lustre: Skipped 2 previous similar messages LustreError: 19694:0:(mdd_dir.c:4619:mdd_migrate_cmd_check()) lustre-MDD0000: '5' migration was interrupted, run 'lfs migrate -m 1 -c 1 -H crush 5' to finish migration: rc = -1 LustreError: 21992:0:(mdd_dir.c:4619:mdd_migrate_cmd_check()) lustre-MDD0000: '5' migration was interrupted, run 'lfs migrate -m 1 -c 1 -H crush 5' to finish migration: rc = -1 LustreError: 21992:0:(mdd_dir.c:4619:mdd_migrate_cmd_check()) Skipped 1 previous similar message LustreError: 18284:0:(mdd_object.c:3861:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000403:0x12ac:0x0]: rc = -2 | Link to test |
racer test 1: racer on clients: centos-115.localnet DURATION=2700 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 PGD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) lustre(OE) osp(OE) ofd(OE) lod(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) jbd2 mbcache lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) crc32_generic lnet(OE) dm_flakey dm_mod libcfs(OE) crc_t10dif crct10dif_generic crct10dif_common virtio_balloon pcspkr virtio_console i2c_piix4 ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm drm_panel_orientation_quirks ata_piix i2c_core serio_raw virtio_blk libata floppy CPU: 6 PID: 22151 Comm: ll_sa_22143 Kdump: loaded Tainted: P OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.16.0-3.module_el8.7.0+1218+f626c2ff 04/01/2014 task: ffff8802dbda8010 ti: ffff88008ffe8000 task.ti: ffff88008ffe8000 RIP: 0010:[<ffffffff813f502c>] [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 RSP: 0018:ffff88008ffebd80 EFLAGS: 00010282 RAX: ffff880296e8ec48 RBX: 0000000000000008 RCX: 0000000100260021 RDX: 0000000000000026 RSI: ffff880296e8ef78 RDI: 0000000000000008 RBP: ffff88008ffebd90 R08: ffff8802e16b4598 R09: 0000000000000000 R10: ffff8802e16b2b18 R11: ffff8802e16b4598 R12: 0000000000000000 R13: ffff8802dbdadc40 R14: ffff880296e8ef78 R15: ffff8802e16b4598 FS: 0000000000000000(0000) GS:ffff880331b80000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 0000000001c10000 CR4: 00000000000007e0 Call Trace: [<ffffffffa1107171>] ll_statahead_thread+0xde1/0x2290 [lustre] [<ffffffff817e320e>] ? _raw_spin_lock+0x1e/0x20 [<ffffffff810dbb51>] ? put_prev_entity+0x31/0x400 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e329e>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffff810c834d>] ? finish_task_switch+0x5d/0x1b0 [<ffffffffa1106390>] ? ll_statahead_handle.constprop.29+0x1c0/0x1c0 [lustre] [<ffffffff810ba114>] kthread+0xe4/0xf0 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 | 2[16421]: segfault at 406000 ip 0000000000406000 sp 00007ffc71b642d8 error 14 in 2[606000+1000] 3[16652]: segfault at 8 ip 00007f2ebbdb97e8 sp 00007fff11e18690 error 4 in ld-2.17.so[7f2ebbdae000+22000] Lustre: 16157:0:(mdt_recovery.c:148:mdt_req_from_lrd()) @@@ restoring transno req@ffff8800a8cf1940 x1820362059208192/t4294969182(0) o101->da79ba20-8a77-475c-9cc8-aee2b44c43dd@0@lo:718/0 lens 376/816 e 0 to 0 dl 1736032618 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 7[18506]: segfault at 0 ip 0000000000401d60 sp 00007ffe3bac2458 error 6 in 7[400000+6000] 7[18682]: segfault at 0 ip 0000000000401d60 sp 00007fff55c085f8 error 6 in 7[400000+6000] Lustre: 16155:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000402:0x199:0x0] with magic=0xbd60bd0 Lustre: 15234:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0000: opcode 7: before 516 < left 618, rollback = 7 Lustre: 15234:0:(osd_handler.c:1951:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 15234:0:(osd_handler.c:1958:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 15234:0:(osd_handler.c:1968:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 15234:0:(osd_handler.c:1975:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 15234:0:(osd_handler.c:1982:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 15234:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0000: opcode 7: before 516 < left 618, rollback = 7 Lustre: 15234:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 1 previous similar message Lustre: 15234:0:(osd_handler.c:1951:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 15234:0:(osd_handler.c:1951:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 15234:0:(osd_handler.c:1958:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 15234:0:(osd_handler.c:1958:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 15234:0:(osd_handler.c:1968:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 15234:0:(osd_handler.c:1968:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 15234:0:(osd_handler.c:1975:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 15234:0:(osd_handler.c:1975:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 15234:0:(osd_handler.c:1982:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 15234:0:(osd_handler.c:1982:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 15234:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0002: opcode 7: before 516 < left 618, rollback = 7 Lustre: 15234:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 1 previous similar message Lustre: 15234:0:(osd_handler.c:1951:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 15234:0:(osd_handler.c:1951:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 15234:0:(osd_handler.c:1958:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 15234:0:(osd_handler.c:1958:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 15234:0:(osd_handler.c:1968:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 15234:0:(osd_handler.c:1968:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 15234:0:(osd_handler.c:1975:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 15234:0:(osd_handler.c:1975:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 15234:0:(osd_handler.c:1982:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 15234:0:(osd_handler.c:1982:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 18544:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0000: opcode 7: before 516 < left 618, rollback = 7 Lustre: 18544:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 3 previous similar messages Lustre: 18544:0:(osd_handler.c:1951:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 18544:0:(osd_handler.c:1951:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 18544:0:(osd_handler.c:1958:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 18544:0:(osd_handler.c:1958:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 18544:0:(osd_handler.c:1968:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 18544:0:(osd_handler.c:1968:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 18544:0:(osd_handler.c:1975:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 18544:0:(osd_handler.c:1975:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 18544:0:(osd_handler.c:1982:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 18544:0:(osd_handler.c:1982:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 15030:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000401:0x38e:0x0] with magic=0xbd60bd0 Lustre: 15030:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message Lustre: 15234:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 516 < left 618, rollback = 7 Lustre: 15234:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 7 previous similar messages Lustre: 15234:0:(osd_handler.c:1951:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 15234:0:(osd_handler.c:1951:osd_trans_dump_creds()) Skipped 7 previous similar messages Lustre: 15234:0:(osd_handler.c:1958:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 15234:0:(osd_handler.c:1958:osd_trans_dump_creds()) Skipped 7 previous similar messages Lustre: 15234:0:(osd_handler.c:1968:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 15234:0:(osd_handler.c:1968:osd_trans_dump_creds()) Skipped 7 previous similar messages Lustre: 15234:0:(osd_handler.c:1975:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 15234:0:(osd_handler.c:1975:osd_trans_dump_creds()) Skipped 7 previous similar messages Lustre: 15234:0:(osd_handler.c:1982:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 15234:0:(osd_handler.c:1982:osd_trans_dump_creds()) Skipped 7 previous similar messages Lustre: 15234:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 516 < left 618, rollback = 7 Lustre: 15234:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 9 previous similar messages Lustre: 15234:0:(osd_handler.c:1951:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 15234:0:(osd_handler.c:1951:osd_trans_dump_creds()) Skipped 9 previous similar messages Lustre: 15234:0:(osd_handler.c:1958:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 15234:0:(osd_handler.c:1958:osd_trans_dump_creds()) Skipped 9 previous similar messages Lustre: 15234:0:(osd_handler.c:1968:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 15234:0:(osd_handler.c:1968:osd_trans_dump_creds()) Skipped 9 previous similar messages Lustre: 15234:0:(osd_handler.c:1975:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 15234:0:(osd_handler.c:1975:osd_trans_dump_creds()) Skipped 9 previous similar messages Lustre: 15234:0:(osd_handler.c:1982:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 15234:0:(osd_handler.c:1982:osd_trans_dump_creds()) Skipped 9 previous similar messages LustreError: 31847:0:(file.c:267:ll_close_inode_openhandle()) lustre-clilmv-ffff8802ee724138: inode [0x200000402:0x590:0x0] mdc close failed: rc = -13 4[954]: segfault at 8 ip 00007f356fbf77e8 sp 00007ffc93d7a520 error 4 in ld-2.17.so[7f356fbec000+22000] Lustre: 18544:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 516 < left 618, rollback = 7 Lustre: 18544:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 23 previous similar messages Lustre: 18544:0:(osd_handler.c:1951:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 18544:0:(osd_handler.c:1951:osd_trans_dump_creds()) Skipped 23 previous similar messages Lustre: 18544:0:(osd_handler.c:1958:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 18544:0:(osd_handler.c:1958:osd_trans_dump_creds()) Skipped 23 previous similar messages Lustre: 18544:0:(osd_handler.c:1968:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 18544:0:(osd_handler.c:1968:osd_trans_dump_creds()) Skipped 23 previous similar messages Lustre: 18544:0:(osd_handler.c:1975:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 18544:0:(osd_handler.c:1975:osd_trans_dump_creds()) Skipped 23 previous similar messages Lustre: 18544:0:(osd_handler.c:1982:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 18544:0:(osd_handler.c:1982:osd_trans_dump_creds()) Skipped 23 previous similar messages Lustre: 9093:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000402:0xa61:0x0] with magic=0xbd60bd0 Lustre: 9093:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message LustreError: 15788:0:(file.c:267:ll_close_inode_openhandle()) lustre-clilmv-ffff8802ee724138: inode [0x200000402:0xb18:0x0] mdc close failed: rc = -13 Lustre: 9830:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 516 < left 618, rollback = 7 Lustre: 9830:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 25 previous similar messages Lustre: 9830:0:(osd_handler.c:1951:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 9830:0:(osd_handler.c:1951:osd_trans_dump_creds()) Skipped 25 previous similar messages Lustre: 9830:0:(osd_handler.c:1958:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 9830:0:(osd_handler.c:1958:osd_trans_dump_creds()) Skipped 25 previous similar messages Lustre: 9830:0:(osd_handler.c:1968:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 9830:0:(osd_handler.c:1968:osd_trans_dump_creds()) Skipped 25 previous similar messages Lustre: 9830:0:(osd_handler.c:1975:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 9830:0:(osd_handler.c:1975:osd_trans_dump_creds()) Skipped 25 previous similar messages Lustre: 9830:0:(osd_handler.c:1982:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 9830:0:(osd_handler.c:1982:osd_trans_dump_creds()) Skipped 25 previous similar messages Lustre: 14994:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000401:0xd3f:0x0] with magic=0xbd60bd0 Lustre: 14994:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message 14[30288]: segfault at 8 ip 00007f35c431c7e8 sp 00007ffc101f7c40 error 4 in ld-2.17.so[7f35c4311000+22000] traps: 2[2206] trap invalid opcode ip:404b3e sp:7ffef47f1488 error:0 in 2[400000+6000] Lustre: 28291:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0002: opcode 7: before 516 < left 618, rollback = 7 Lustre: 28291:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 82 previous similar messages Lustre: 28291:0:(osd_handler.c:1951:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 28291:0:(osd_handler.c:1951:osd_trans_dump_creds()) Skipped 82 previous similar messages Lustre: 28291:0:(osd_handler.c:1958:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 28291:0:(osd_handler.c:1958:osd_trans_dump_creds()) Skipped 82 previous similar messages Lustre: 28291:0:(osd_handler.c:1968:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 28291:0:(osd_handler.c:1968:osd_trans_dump_creds()) Skipped 82 previous similar messages Lustre: 9829:0:(osd_handler.c:1975:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 9829:0:(osd_handler.c:1975:osd_trans_dump_creds()) Skipped 83 previous similar messages Lustre: 9829:0:(osd_handler.c:1982:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 9829:0:(osd_handler.c:1982:osd_trans_dump_creds()) Skipped 83 previous similar messages 9[9895]: segfault at 0 ip (null) sp 00007ffe771b52d8 error 14 in 9[400000+6000] 15[21986]: segfault at 8 ip 00007f75825547e8 sp 00007ffc1f4c1420 error 4 in ld-2.17.so[7f7582549000+22000] Lustre: 14874:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000402:0x19d3:0x0] with magic=0xbd60bd0 Lustre: 14874:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message 17[13042]: segfault at 406000 ip 0000000000406000 sp 00007ffd15bf8ab8 error 14 in 17[606000+1000] Lustre: 9830:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0000: opcode 7: before 516 < left 618, rollback = 7 Lustre: 9830:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 78 previous similar messages Lustre: 9830:0:(osd_handler.c:1951:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 9830:0:(osd_handler.c:1951:osd_trans_dump_creds()) Skipped 78 previous similar messages Lustre: 9830:0:(osd_handler.c:1958:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 9830:0:(osd_handler.c:1958:osd_trans_dump_creds()) Skipped 78 previous similar messages Lustre: 9830:0:(osd_handler.c:1968:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 9830:0:(osd_handler.c:1968:osd_trans_dump_creds()) Skipped 78 previous similar messages Lustre: 9830:0:(osd_handler.c:1975:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 9830:0:(osd_handler.c:1975:osd_trans_dump_creds()) Skipped 77 previous similar messages Lustre: 9830:0:(osd_handler.c:1982:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 9830:0:(osd_handler.c:1982:osd_trans_dump_creds()) Skipped 77 previous similar messages Lustre: 16155:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000401:0x22e9:0x0] with magic=0xbd60bd0 Lustre: 16155:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message LustreError: 21700:0:(file.c:267:ll_close_inode_openhandle()) lustre-clilmv-ffff8802ef1ddd28: inode [0x200000401:0x22ec:0x0] mdc close failed: rc = -13 16[22977]: segfault at 8 ip 00007f0fbc9c77e8 sp 00007fffd2a263f0 error 4 in ld-2.17.so[7f0fbc9bc000+22000] LustreError: 22977:0:(file.c:267:ll_close_inode_openhandle()) lustre-clilmv-ffff8802ee724138: inode [0x200000401:0x2398:0x0] mdc close failed: rc = -13 9[26975]: segfault at 8 ip 00007fcd795217e8 sp 00007ffd62095170 error 4 in ld-2.17.so[7fcd79516000+22000] Lustre: 15034:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000401:0x25b7:0x0] with magic=0xbd60bd0 Lustre: 15034:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message 9[30648]: segfault at 8 ip 00007fe45149b7e8 sp 00007fffa4e2cf50 error 4 in ld-2.17.so[7fe451490000+22000] 8[7766]: segfault at 8 ip 00007f4eef3ec7e8 sp 00007ffd27196cd0 error 4 in ld-2.17.so[7f4eef3e1000+22000] 12[8571]: segfault at 8 ip 00007f6d3c83d7e8 sp 00007ffe2e0b4c90 error 4 in ld-2.17.so[7f6d3c832000+22000] 9[18051]: segfault at 8 ip 00007f59fe9837e8 sp 00007ffce7331a80 error 4 in ld-2.17.so[7f59fe978000+22000] 9[19374]: segfault at 0 ip (null) sp 00007ffc26fc5558 error 14 in 9[400000+6000] LustreError: 22026:0:(file.c:267:ll_close_inode_openhandle()) lustre-clilmv-ffff8802ee724138: inode [0x200000401:0x2dc6:0x0] mdc close failed: rc = -13 LustreError: 24975:0:(file.c:267:ll_close_inode_openhandle()) lustre-clilmv-ffff8802ee724138: inode [0x200000401:0x2eed:0x0] mdc close failed: rc = -13 11[26691]: segfault at 8 ip 00007f13cef167e8 sp 00007ffc831e4340 error 4 in ld-2.17.so[7f13cef0b000+22000] 16[2892]: segfault at 8 ip 00007f5ba28bc7e8 sp 00007ffc971e7d10 error 4 in ld-2.17.so[7f5ba28b1000+22000] Lustre: 16155:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000402:0x33ec:0x0] with magic=0xbd60bd0 Lustre: 16155:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message 8[9947]: segfault at 8 ip 00007f35f25887e8 sp 00007fffc97c1cc0 error 4 in ld-2.17.so[7f35f257d000+22000] Lustre: 9829:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 618, rollback = 7 Lustre: 9829:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 249 previous similar messages Lustre: 18544:0:(osd_handler.c:1951:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 18544:0:(osd_handler.c:1951:osd_trans_dump_creds()) Skipped 251 previous similar messages Lustre: 18544:0:(osd_handler.c:1958:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 18544:0:(osd_handler.c:1958:osd_trans_dump_creds()) Skipped 251 previous similar messages Lustre: 18544:0:(osd_handler.c:1968:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 18544:0:(osd_handler.c:1968:osd_trans_dump_creds()) Skipped 251 previous similar messages Lustre: 18544:0:(osd_handler.c:1975:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 18544:0:(osd_handler.c:1975:osd_trans_dump_creds()) Skipped 251 previous similar messages Lustre: 18544:0:(osd_handler.c:1982:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 18544:0:(osd_handler.c:1982:osd_trans_dump_creds()) Skipped 251 previous similar messages 5[18193]: segfault at 0 ip (null) sp 00007ffd46f0a678 error 14 in 5[400000+6000] LustreError: 22114:0:(file.c:267:ll_close_inode_openhandle()) lustre-clilmv-ffff8802ef1ddd28: inode [0x200000402:0x3950:0x0] mdc close failed: rc = -13 7[23247]: segfault at 8 ip 00007faba76717e8 sp 00007ffcb5fc74f0 error 4 in ld-2.17.so[7faba7666000+22000] 1[27284]: segfault at 8 ip 00007faaf19a37e8 sp 00007ffce1c830b0 error 4 in ld-2.17.so[7faaf1998000+22000] LustreError: 32283:0:(file.c:267:ll_close_inode_openhandle()) lustre-clilmv-ffff8802ee724138: inode [0x200000401:0x3cd7:0x0] mdc close failed: rc = -13 Lustre: 9093:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000401:0x3d60:0x0] with magic=0xbd60bd0 Lustre: 9093:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 3 previous similar messages 4[2401]: segfault at 8 ip 00007f3e03eca7e8 sp 00007ffec414dac0 error 4 in ld-2.17.so[7f3e03ebf000+22000] 5[5087]: segfault at 8 ip 00007f714d9c07e8 sp 00007fff98bb4380 error 4 in ld-2.17.so[7f714d9b5000+22000] 4[21992]: segfault at d ip 0000000000405c90 sp 00007ffdb48c4e18 error 6 in 4[400000+6000] LustreError: 21992:0:(file.c:267:ll_close_inode_openhandle()) lustre-clilmv-ffff8802ee724138: inode [0x200000402:0x44c2:0x0] mdc close failed: rc = -13 LustreError: 26135:0:(statahead.c:2399:start_statahead_thread()) lustre: invalid pattern 0X10. traps: 15[32088] trap invalid opcode ip:4053d0 sp:7fff88cb4888 error:0 in 15[400000+6000] Lustre: 15036:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000402:0x4cc3:0x0] with magic=0xbd60bd0 Lustre: 15036:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 5 previous similar messages 17[22262]: segfault at 1c ip 00000000004017b0 sp 00007ffd9009e860 error 6 in 17[400000+6000] 13[24843]: segfault at 8 ip 00007f97c8af47e8 sp 00007fffbe1a4f70 error 4 in ld-2.17.so[7f97c8ae9000+22000] 6[29684]: segfault at 8 ip 00007f9a004fd7e8 sp 00007ffc67631f60 error 4 in ld-2.17.so[7f9a004f2000+22000] Lustre: mdt00_000: service thread pid 9091 was inactive for 40.036 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 9091, comm: mdt00_000 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x913/0xd50 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x800 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock+0x88/0x1c0 [mdt] [<0>] mdt_intent_getxattr+0x80/0x330 [mdt] [<0>] mdt_intent_opc.constprop.73+0x211/0xc60 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x397/0x980 [ptlrpc] [<0>] ldlm_handle_enqueue+0x507/0x1870 [ptlrpc] [<0>] tgt_enqueue+0x68/0x240 [ptlrpc] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 9093, comm: mdt00_002 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x913/0xd50 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x800 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock+0x88/0x1c0 [mdt] [<0>] mdt_object_find_lock+0x54/0x170 [mdt] [<0>] mdt_reint_setxattr+0x133/0x10e0 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x74c/0xbc0 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 23225, comm: mdt00_016 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x913/0xd50 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x800 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock_try+0xa0/0x250 [mdt] [<0>] mdt_getattr_name_lock+0x173b/0x2d00 [mdt] [<0>] mdt_intent_getattr+0x2cc/0x4e0 [mdt] [<0>] mdt_intent_opc.constprop.73+0x211/0xc60 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x397/0x980 [ptlrpc] [<0>] ldlm_handle_enqueue+0x507/0x1870 [ptlrpc] [<0>] tgt_enqueue+0x68/0x240 [ptlrpc] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: mdt00_008: service thread pid 15014 was inactive for 40.148 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: mdt00_013: service thread pid 15038 was inactive for 72.244 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 1 previous similar message LustreError: 8975:0:(ldlm_lockd.c:252:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0000_UUID lock: ffff8802dcf387c0/0x31d6d6c8d736a838 lrc: 3/0,0 mode: PR/PR res: [0x200000402:0x54bf:0x0].0x0 bits 0x1b/0x0 rrc: 13 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0x31d6d6c8d736a81c expref: 1748 pid: 14989 timeout: 1037 lvb_type: 0 LustreError: 9092:0:(ldlm_lockd.c:1492:ldlm_handle_enqueue()) ### lock on destroyed export ffff8800aa1b5d28 ns: mdt-lustre-MDT0000_UUID lock: ffff8802d1672d40/0x31d6d6c8d736be11 lrc: 3/0,0 mode: PR/PR res: [0x200000402:0x54bf:0x0].0x0 bits 0x1b/0x0 rrc: 9 type: IBT gid 0 flags: 0x50200000000000 nid: 0@lo remote: 0x31d6d6c8d736bdcb expref: 68 pid: 9092 timeout: 0 lvb_type: 0 Lustre: mdt00_008: service thread pid 15014 completed after 100.108s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_001: service thread pid 9092 completed after 99.833s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_002: service thread pid 9093 completed after 100.407s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: lustre-MDT0000-mdc-ffff8802ee724138: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-MDT0000-mdc-ffff8802ee724138: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 15380:0:(ldlm_lockd.c:2591:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1736033547 with bad export cookie 3591293910955207416 Lustre: mdt00_013: service thread pid 15038 completed after 79.997s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_000: service thread pid 9091 completed after 100.404s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_016: service thread pid 23225 completed after 100.160s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: lustre-MDT0000-mdc-ffff8802ee724138: This client was evicted by lustre-MDT0000; in progress operations using this service will fail. LustreError: 31329:0:(file.c:5979:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000402:0x54bf:0x0] error: rc = -5 LustreError: 4930:0:(llite_lib.c:2040:ll_md_setattr()) md_setattr fails: rc = -108 LustreError: 15024:0:(file.c:267:ll_close_inode_openhandle()) lustre-clilmv-ffff8802ee724138: inode [0x200000401:0x561f:0x0] mdc close failed: rc = -108 LustreError: 15024:0:(file.c:267:ll_close_inode_openhandle()) Skipped 2 previous similar messages LustreError: 723:0:(mdc_request.c:1454:mdc_read_page()) lustre-MDT0000-mdc-ffff8802ee724138: [0x200000402:0x54be:0x0] lock enqueue fails: rc = -108 LustreError: 15024:0:(ldlm_resource.c:1221:ldlm_resource_complain()) lustre-MDT0000-mdc-ffff8802ee724138: namespace resource [0x200000401:0x1:0x0].0x0 (ffff8802d800fc40) refcount nonzero (1) after lock cleanup; forcing cleanup. Lustre: lustre-MDT0000-mdc-ffff8802ee724138: Connection restored to (at 0@lo) 9[12864]: segfault at 0 ip (null) sp 00007ffeb4b442a8 error 14 in 9[400000+6000] Lustre: 14979:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0x10bd:0x0] with magic=0xbd60bd0 Lustre: 14979:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 15 previous similar messages Lustre: 28291:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0000: opcode 7: before 516 < left 594, rollback = 7 Lustre: 28291:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 501 previous similar messages Lustre: 28291:0:(osd_handler.c:1951:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 28291:0:(osd_handler.c:1951:osd_trans_dump_creds()) Skipped 499 previous similar messages Lustre: 28291:0:(osd_handler.c:1958:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 28291:0:(osd_handler.c:1958:osd_trans_dump_creds()) Skipped 499 previous similar messages Lustre: 28291:0:(osd_handler.c:1968:osd_trans_dump_creds()) write: 2/594/0, punch: 0/0/0, quota 4/150/0 Lustre: 28291:0:(osd_handler.c:1968:osd_trans_dump_creds()) Skipped 499 previous similar messages Lustre: 28291:0:(osd_handler.c:1975:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 28291:0:(osd_handler.c:1975:osd_trans_dump_creds()) Skipped 499 previous similar messages Lustre: 28291:0:(osd_handler.c:1982:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 28291:0:(osd_handler.c:1982:osd_trans_dump_creds()) Skipped 499 previous similar messages 14[21294]: segfault at 8 ip 00007f37666f97e8 sp 00007ffcf40aaa40 error 4 in ld-2.17.so[7f37666ee000+22000] 10[28666]: segfault at 8 ip 00007fee030b07e8 sp 00007ffc5a686050 error 4 in ld-2.17.so[7fee030a5000+22000] 4[5788]: segfault at 8 ip 00007f4dfdfb57e8 sp 00007ffc245797c0 error 4 in ld-2.17.so[7f4dfdfaa000+22000] 1[11959]: segfault at 8 ip 00007fc33ebda2fc sp 00007ffd11e33b50 error 4 in ld-2.17.so[7fc33ebcd000+22000] 7[20542]: segfault at 8 ip 00007fe20a8777e8 sp 00007ffe64522690 error 4 in ld-2.17.so[7fe20a86c000+22000] 19[24681]: segfault at 8 ip 00007f83e535e7e8 sp 00007ffe1fac5d70 error 4 in ld-2.17.so[7f83e5353000+22000] LustreError: 24681:0:(file.c:267:ll_close_inode_openhandle()) lustre-clilmv-ffff8802ee724138: inode [0x200000403:0x266c:0x0] mdc close failed: rc = -13 LustreError: 24681:0:(file.c:267:ll_close_inode_openhandle()) Skipped 9 previous similar messages 4[6041]: segfault at 8 ip 00007f2d8d0cf7e8 sp 00007ffe43984fc0 error 4 in ld-2.17.so[7f2d8d0c4000+22000] 2[25035]: segfault at 8 ip 00007fc547f617e8 sp 00007ffd327b3cc0 error 4 in ld-2.17.so[7fc547f56000+22000] 14[27215]: segfault at 8 ip 00007f80284237e8 sp 00007ffe65886120 error 4 in ld-2.17.so[7f8028418000+22000] 13[3283]: segfault at 8 ip 00007f134b33b7e8 sp 00007fffac604660 error 4 in ld-2.17.so[7f134b330000+22000] 15[6466]: segfault at 8 ip 00007f5355fbb7e8 sp 00007ffc4dd74700 error 4 in ld-2.17.so[7f5355fb0000+22000] Lustre: 28085:0:(osd_internal.h:1326:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 618, rollback = 7 Lustre: 28085:0:(osd_internal.h:1326:osd_trans_exec_op()) Skipped 563 previous similar messages Lustre: 28085:0:(osd_handler.c:1951:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 28085:0:(osd_handler.c:1951:osd_trans_dump_creds()) Skipped 563 previous similar messages Lustre: 28085:0:(osd_handler.c:1958:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 28085:0:(osd_handler.c:1958:osd_trans_dump_creds()) Skipped 563 previous similar messages Lustre: 28085:0:(osd_handler.c:1968:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 28085:0:(osd_handler.c:1968:osd_trans_dump_creds()) Skipped 563 previous similar messages Lustre: 28085:0:(osd_handler.c:1975:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 28085:0:(osd_handler.c:1975:osd_trans_dump_creds()) Skipped 563 previous similar messages Lustre: 28085:0:(osd_handler.c:1982:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 28085:0:(osd_handler.c:1982:osd_trans_dump_creds()) Skipped 563 previous similar messages Lustre: 9091:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000401:0x9d91:0x0] with magic=0xbd60bd0 Lustre: 9091:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 21 previous similar messages 11[4796]: segfault at 8 ip 00007fafd49fe7e8 sp 00007ffeda4ba9b0 error 4 in ld-2.17.so[7fafd49f3000+22000] 1[5874]: segfault at 406000 ip 0000000000406000 sp 00007ffe1b721848 error 14 in 1[606000+1000] LustreError: 8465:0:(statahead.c:2399:start_statahead_thread()) lustre: invalid pattern 0X0. LustreError: 8975:0:(ldlm_lockd.c:252:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0003_UUID lock: ffff8802e17cc000/0x31d6d6c8d7f03b4e lrc: 3/0,0 mode: PW/PW res: [0x300000400:0xd0e:0x0].0x0 rrc: 5 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000000020020 nid: 0@lo remote: 0x31d6d6c8d7f03b47 expref: 14 pid: 9821 timeout: 2072 lvb_type: 0 LustreError: lustre-OST0003-osc-ffff8802ef1ddd28: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 2 previous similar messages Lustre: lustre-OST0003-osc-ffff8802ef1ddd28: Connection to lustre-OST0003 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: lustre-OST0003-osc-ffff8802ef1ddd28: This client was evicted by lustre-OST0003; in progress operations using this service will fail. Lustre: 3823:0:(llite_lib.c:4166:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.117@tcp:/lustre/fid: [0x200000401:0xb2ce:0x0]/ may get corrupted (rc -108) Lustre: 3823:0:(llite_lib.c:4166:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.117@tcp:/lustre/fid: [0x200000401:0xaa35:0x0]// may get corrupted (rc -108) LustreError: 20244:0:(ldlm_resource.c:1221:ldlm_resource_complain()) lustre-OST0003-osc-ffff8802ef1ddd28: namespace resource [0x300000400:0xd0e:0x0].0x0 (ffff8802afd25440) refcount nonzero (1) after lock cleanup; forcing cleanup. Lustre: lustre-OST0003-osc-ffff8802ef1ddd28: Connection restored to (at 0@lo) LustreError: 29794:0:(file.c:267:ll_close_inode_openhandle()) lustre-clilmv-ffff8802ee724138: inode [0x200000401:0xb5cc:0x0] mdc close failed: rc = -13 LustreError: 29794:0:(file.c:267:ll_close_inode_openhandle()) Skipped 4 previous similar messages | Link to test |
racer test 1: racer on clients: centos-55.localnet DURATION=2700 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 PGD 2a75fd067 PUD 2b4f9b067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) osp(OE) ofd(OE) lod(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) crc32_generic lnet(OE) libcfs(OE) crc_t10dif crct10dif_generic crct10dif_common virtio_console i2c_piix4 virtio_balloon pcspkr ip_tables rpcsec_gss_krb5 ata_generic drm_kms_helper pata_acpi ttm drm drm_panel_orientation_quirks ata_piix serio_raw virtio_blk i2c_core floppy libata CPU: 1 PID: 12885 Comm: ll_sa_12878 Kdump: loaded Tainted: P OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.16.0-3.module_el8.7.0+1218+f626c2ff 04/01/2014 task: ffff8802b411dc40 ti: ffff88027e30c000 task.ti: ffff88027e30c000 RIP: 0010:[<ffffffff813f502c>] [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 RSP: 0018:ffff88027e30fd80 EFLAGS: 00010282 RAX: ffff880278dfe288 RBX: 0000000000000008 RCX: dead000000000200 RDX: 0000000000000001 RSI: ffff880278dfe5b8 RDI: 0000000000000008 RBP: ffff88027e30fd90 R08: ffff880331403660 R09: 0000000000000000 R10: ffff880331403640 R11: ffff8802ad9cd628 R12: 0000000000000000 R13: ffff8802b0f849d0 R14: ffff880278dfe5b8 R15: ffff8802ad9cd628 FS: 0000000000000000(0000) GS:ffff880331a40000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 00000002a71d0000 CR4: 00000000000007e0 Call Trace: [<ffffffffa15a1f01>] ll_statahead_thread+0xde1/0x2290 [lustre] [<ffffffff817e320e>] ? _raw_spin_lock+0x1e/0x20 [<ffffffff810dbb51>] ? put_prev_entity+0x31/0x400 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e329e>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffff810c834d>] ? finish_task_switch+0x5d/0x1b0 [<ffffffffa15a1120>] ? ll_statahead_handle.constprop.29+0x1c0/0x1c0 [lustre] [<ffffffff810ba114>] kthread+0xe4/0xf0 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 | Lustre: 20060:0:(mdt_recovery.c:148:mdt_req_from_lrd()) @@@ restoring transno req@ffff8802d55e0f40 x1819866874143616/t4294968958(0) o101->b01e8869-f8f6-4fdb-bfdd-da6d5306436b@0@lo:393/0 lens 376/840 e 0 to 0 dl 1735560418 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 Lustre: 20055:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000401:0x505:0x0] with magic=0xbd60bd0 19[10454]: segfault at 8 ip 00007f667964a7e8 sp 00007ffe7d115240 error 4 in ld-2.17.so[7f667963f000+22000] Lustre: 20184:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000402:0x76f:0x0] with magic=0xbd60bd0 Lustre: 20184:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message 4[15370]: segfault at 8 ip 00007f5c6849e7e8 sp 00007fff3cb24ac0 error 4 in ld-2.17.so[7f5c68493000+22000] Lustre: 12857:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000402:0x7b9:0x0] with magic=0xbd60bd0 Lustre: 12857:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message 6[19924]: segfault at 8 ip 00007f27a08c67e8 sp 00007ffe8207e010 error 4 in ld-2.17.so[7f27a08bb000+22000] Lustre: 10624:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000401:0x82a:0x0] with magic=0xbd60bd0 Lustre: 10624:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message Lustre: 20142:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000401:0x830:0x0] with magic=0xbd60bd0 Lustre: 20142:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message 17[26029]: segfault at 8 ip 00007f613cd497e8 sp 00007fffa57eec20 error 4 in ld-2.17.so[7f613cd3e000+22000] Lustre: 20052:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000401:0x910:0x0] with magic=0xbd60bd0 Lustre: 20052:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message 10[2999]: segfault at 8 ip 00007f5ff22d77e8 sp 00007fffcb704d70 error 4 in ld-2.17.so[7f5ff22cc000+22000] 7[3971]: segfault at 8 ip 00007ff8f716c7e8 sp 00007ffc8ec28530 error 4 in ld-2.17.so[7ff8f7161000+22000] 4[15320]: segfault at 8 ip 00007f00008ef7e8 sp 00007ffda011cc30 error 4 in ld-2.17.so[7f00008e4000+22000] 3[15153]: segfault at 8 ip 00007f1a92e7b7e8 sp 00007ffdba83a350 error 4 in ld-2.17.so[7f1a92e70000+22000] LustreError: 15153:0:(file.c:267:ll_close_inode_openhandle()) lustre-clilmv-ffff8802dcfc92a8: inode [0x200000401:0xb80:0x0] mdc close failed: rc = -13 LustreError: 29622:0:(file.c:267:ll_close_inode_openhandle()) lustre-clilmv-ffff8802dcfc92a8: inode [0x200000401:0xce5:0x0] mdc close failed: rc = -13 17[10246]: segfault at 8 ip 00007ff0b98a27e8 sp 00007ffc33ad7a30 error 4 in ld-2.17.so[7ff0b9897000+22000] Lustre: 20079:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000402:0xf24:0x0] with magic=0xbd60bd0 Lustre: 20079:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message LustreError: 32655:0:(file.c:267:ll_close_inode_openhandle()) lustre-clilmv-ffff8802dd4992a8: inode [0x200000402:0x1261:0x0] mdc close failed: rc = -13 15[2043]: segfault at 8 ip 00007f47c94047e8 sp 00007fffcc4a8b30 error 4 in ld-2.17.so[7f47c93f9000+22000] Lustre: 20055:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000401:0x128a:0x0] with magic=0xbd60bd0 Lustre: 20055:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 3 previous similar messages 19[12942]: segfault at 8 ip 00007f379e5b87e8 sp 00007ffd7bba8d70 error 4 in ld-2.17.so[7f379e5ad000+22000] 19[14277]: segfault at 8 ip 00007f43b215c7e8 sp 00007ffdb37dc690 error 4 in ld-2.17.so[7f43b2151000+22000] Lustre: mdt00_002: service thread pid 10625 was inactive for 40.122 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 10625, comm: mdt00_002 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x913/0xd50 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x800 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock+0x88/0x1c0 [mdt] [<0>] mdt_getattr_name_lock+0xc6a/0x2d00 [mdt] [<0>] mdt_intent_getattr+0x2cc/0x4e0 [mdt] [<0>] mdt_intent_opc.constprop.73+0x211/0xc60 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x397/0x980 [ptlrpc] [<0>] ldlm_handle_enqueue+0x507/0x1870 [ptlrpc] [<0>] tgt_enqueue+0x68/0x240 [ptlrpc] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 26324, comm: mdt_io00_003 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x913/0xd50 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x800 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock+0x88/0x1c0 [mdt] [<0>] mdt_rename_source_lock+0xa9/0xd6 [mdt] [<0>] mdt_reint_rename+0x1789/0x2c30 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x74c/0xbc0 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 16459, comm: ll_ost_out00_00 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x913/0xd50 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x800 [ptlrpc] [<0>] mdt_object_pdo_lock+0x4d9/0x7e0 [mdt] [<0>] mdt_parent_lock+0x76/0x2a0 [mdt] [<0>] mdt_getattr_name_lock+0x1850/0x2d00 [mdt] [<0>] mdt_intent_getattr+0x2cc/0x4e0 [mdt] [<0>] mdt_intent_opc.constprop.73+0x211/0xc60 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x397/0x980 [ptlrpc] [<0>] ldlm_handle_enqueue+0x507/0x1870 [ptlrpc] [<0>] mdt_batch_getattr+0x6c/0x190 [mdt] [<0>] mdt_batch+0x465/0x195b [mdt] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: ll_ost_out00_00: service thread pid 25426 was inactive for 40.183 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: mdt00_018: service thread pid 12857 was inactive for 40.060 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 1 previous similar message LustreError: 10317:0:(ldlm_lockd.c:252:expired_lock_main()) ### lock callback timer expired after 101s: evicting client at 0@lo ns: mdt-lustre-MDT0000_UUID lock: ffff8802a3e79300/0x2a43fcc0d7e5b52f lrc: 3/0,0 mode: PR/PR res: [0x200000401:0x1876:0x0].0x0 bits 0x13/0x0 rrc: 5 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0x2a43fcc0d7e5b4e9 expref: 755 pid: 25426 timeout: 749 lvb_type: 0 Lustre: mdt_io00_003: service thread pid 26324 completed after 100.174s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 10624:0:(ldlm_lockd.c:1492:ldlm_handle_enqueue()) ### lock on destroyed export ffff8802dcb4e678 ns: mdt-lustre-MDT0000_UUID lock: ffff8802c24cf480/0x2a43fcc0d7e5ba29 lrc: 3/0,0 mode: PR/PR res: [0x200000401:0x1:0x0].0x0 bits 0x13/0x0 rrc: 19 type: IBT gid 0 flags: 0x50200000000000 nid: 0@lo remote: 0x2a43fcc0d7e5ba1b expref: 288 pid: 10624 timeout: 0 lvb_type: 0 Lustre: mdt00_001: service thread pid 10624 completed after 99.355s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_007: service thread pid 20051 completed after 99.350s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: lustre-MDT0000-mdc-ffff8802dcfc92a8: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-MDT0000-mdc-ffff8802dcfc92a8: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: mdt00_004: service thread pid 19907 completed after 99.364s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_002: service thread pid 10625 completed after 100.182s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_015: service thread pid 20212 completed after 99.362s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_005: service thread pid 19997 completed after 100.087s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_003: service thread pid 19898 completed after 99.368s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: lustre-MDT0000-mdc-ffff8802dcfc92a8: This client was evicted by lustre-MDT0000; in progress operations using this service will fail. Lustre: ll_ost_out00_00: service thread pid 12270 completed after 98.926s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_009: service thread pid 20055 completed after 99.367s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: ll_ost_out00_00: service thread pid 25426 completed after 100.196s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_018: service thread pid 12857 completed after 99.365s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 11104:0:(file.c:5979:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000401:0x1:0x0] error: rc = -5 Lustre: ll_ost_out00_00: service thread pid 16459 completed after 100.195s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 12003:0:(file.c:267:ll_close_inode_openhandle()) lustre-clilmv-ffff8802dcfc92a8: inode [0x200000401:0x16c5:0x0] mdc close failed: rc = -108 Lustre: lustre-MDT0000-mdc-ffff8802dcfc92a8: Connection restored to 192.168.123.57@tcp (at 0@lo) LustreError: 13649:0:(file.c:267:ll_close_inode_openhandle()) lustre-clilmv-ffff8802dd4992a8: inode [0x200000402:0x17fe:0x0] mdc close failed: rc = -13 LustreError: 13649:0:(file.c:267:ll_close_inode_openhandle()) Skipped 7 previous similar messages | Link to test |
racer test 1: racer on clients: centos-60.localnet DURATION=2700 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 PGD 20e168067 PUD 22bc4e067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) osp(OE) ofd(OE) lod(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) crc32_generic lnet(OE) libcfs(OE) crc_t10dif crct10dif_generic crct10dif_common virtio_console virtio_balloon i2c_piix4 pcspkr ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm floppy ata_piix drm_panel_orientation_quirks serio_raw libata i2c_core virtio_blk CPU: 9 PID: 27403 Comm: ll_sa_25995 Kdump: loaded Tainted: P OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.16.0-3.module_el8.7.0+1218+f626c2ff 04/01/2014 task: ffff88029ffa0010 ti: ffff8802d4768000 task.ti: ffff8802d4768000 RIP: 0010:[<ffffffff813f502c>] [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 RSP: 0018:ffff8802d476bd80 EFLAGS: 00010282 RAX: ffff8801e56f8ac8 RBX: 0000000000000008 RCX: 0000000100260023 RDX: 0000000000000026 RSI: ffff8801e56f8df8 RDI: 0000000000000008 RBP: ffff8802d476bd90 R08: ffff8802f2be6368 R09: 0000000000000000 R10: ffff8802f2be1738 R11: ffff8802f2be6368 R12: 0000000000000000 R13: ffff8802c7818010 R14: ffff8801e56f8df8 R15: ffff8802f2be6368 FS: 0000000000000000(0000) GS:ffff880331c40000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 00000002905b8000 CR4: 00000000000007e0 Call Trace: [<ffffffffa158af01>] ll_statahead_thread+0xde1/0x2290 [lustre] [<ffffffff810df6e2>] ? pick_next_task_fair+0x702/0x8a0 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e329e>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffff810c834d>] ? finish_task_switch+0x5d/0x1b0 [<ffffffffa158a120>] ? ll_statahead_handle.constprop.29+0x1c0/0x1c0 [lustre] [<ffffffff810ba114>] kthread+0xe4/0xf0 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 | LustreError: 15373:0:(mdt_reint.c:2533:mdt_reint_migrate()) lustre-MDT0001: migrate [0x240000403:0x1:0x0]/15 failed: rc = -114 Lustre: 29027:0:(mdt_recovery.c:148:mdt_req_from_lrd()) @@@ restoring transno req@ffff88009f2c1e40 x1819630269905024/t4294967933(0) o101->087df90b-edf0-4570-a583-e2f355d5988a@0@lo:458/0 lens 376/840 e 0 to 0 dl 1735334738 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 LustreError: 29180:0:(mdt_reint.c:2533:mdt_reint_migrate()) lustre-MDT0000: migrate [0x200000403:0x2:0x0]/7 failed: rc = -16 cp (28975) used greatest stack depth: 10192 bytes left LustreError: 29024:0:(file.c:267:ll_close_inode_openhandle()) lustre-clilmv-ffff8802cba4d3d8: inode [0x200000404:0x3:0x0] mdc close failed: rc = -13 LustreError: 15373:0:(mdt_reint.c:2533:mdt_reint_migrate()) lustre-MDT0000: migrate [0x200000403:0x2:0x0]/17 failed: rc = -114 LustreError: 15373:0:(mdt_reint.c:2533:mdt_reint_migrate()) Skipped 4 previous similar messages Lustre: 28560:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000404:0xf:0x0] with magic=0xbd60bd0 LustreError: 29480:0:(file.c:267:ll_close_inode_openhandle()) lustre-clilmv-ffff8800b97c2548: inode [0x280000403:0x5:0x0] mdc close failed: rc = -116 LustreError: 29480:0:(file.c:267:ll_close_inode_openhandle()) Skipped 1 previous similar message ls (27736) used greatest stack depth: 10112 bytes left Lustre: 28141:0:(mdt_recovery.c:148:mdt_req_from_lrd()) @@@ restoring transno req@ffff8802ea650040 x1819630270612736/t4294968051(0) o101->087df90b-edf0-4570-a583-e2f355d5988a@0@lo:461/0 lens 376/840 e 0 to 0 dl 1735334741 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 LustreError: 28614:0:(mdt_handler.c:776:mdt_pack_acl2body()) lustre-MDT0002: unable to read [0x280000403:0x55:0x0] ACL: rc = -2 LustreError: 15373:0:(mdt_reint.c:2533:mdt_reint_migrate()) lustre-MDT0001: migrate [0x200000403:0x5e:0x0]/sleep failed: rc = -114 LustreError: 15373:0:(mdt_reint.c:2533:mdt_reint_migrate()) Skipped 4 previous similar messages LustreError: 29652:0:(mdd_dir.c:4619:mdd_migrate_cmd_check()) lustre-MDD0002: '1' migration was interrupted, run 'lfs migrate -m 0 -c 2 -H crush 1' to finish migration: rc = -1 LustreError: 3709:0:(file.c:267:ll_close_inode_openhandle()) lustre-clilmv-ffff8800b97c2548: inode [0x200000403:0xa9:0x0] mdc close failed: rc = -116 LustreError: 3709:0:(file.c:267:ll_close_inode_openhandle()) Skipped 1 previous similar message LustreError: 1819:0:(mdt_reint.c:2533:mdt_reint_migrate()) lustre-MDT0001: migrate [0x240000403:0x1:0x0]/5 failed: rc = -114 LustreError: 1819:0:(mdt_reint.c:2533:mdt_reint_migrate()) Skipped 9 previous similar messages Lustre: 8096:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000404:0xd5:0x0] with magic=0xbd60bd0 Lustre: 8096:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message LustreError: 15373:0:(mdt_reint.c:2533:mdt_reint_migrate()) lustre-MDT0000: migrate [0x200000403:0x2:0x0]/7 failed: rc = -114 LustreError: 15373:0:(mdt_reint.c:2533:mdt_reint_migrate()) Skipped 4 previous similar messages LustreError: 10005:0:(mdd_dir.c:4619:mdd_migrate_cmd_check()) lustre-MDD0001: '11' migration was interrupted, run 'lfs migrate -m 0 -c 2 -H crush 11' to finish migration: rc = -1 LustreError: 1819:0:(mdd_dir.c:4619:mdd_migrate_cmd_check()) lustre-MDD0001: '2' migration was interrupted, run 'lfs migrate -m 0 -c 3 -H crush 2' to finish migration: rc = -1 LustreError: 12304:0:(file.c:267:ll_close_inode_openhandle()) lustre-clilmv-ffff8802cba4d3d8: inode [0x200000403:0x129:0x0] mdc close failed: rc = -116 LustreError: 12304:0:(file.c:267:ll_close_inode_openhandle()) Skipped 1 previous similar message Lustre: dir [0x200000403:0xc0:0x0] stripe 2 readdir failed: -2, directory is partially accessed! LustreError: 15373:0:(mdt_reint.c:2533:mdt_reint_migrate()) lustre-MDT0001: migrate [0x240000403:0x50:0x0]/7 failed: rc = -114 LustreError: 15373:0:(mdt_reint.c:2533:mdt_reint_migrate()) Skipped 22 previous similar messages Lustre: 27434:0:(mdt_recovery.c:148:mdt_req_from_lrd()) @@@ restoring transno req@ffff88009bfccb40 x1819630277479040/t4294970974(0) o101->bdd24f4e-cccb-4a15-a038-3723eeb5b7d7@0@lo:498/0 lens 376/840 e 0 to 0 dl 1735334778 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 Lustre: 6424:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000403:0xbf:0x0] with magic=0xbd60bd0 Lustre: 6424:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message LustreError: 15373:0:(mdd_dir.c:4619:mdd_migrate_cmd_check()) lustre-MDD0002: '19' migration was interrupted, run 'lfs migrate -m 0 -c 3 -H crush 19' to finish migration: rc = -1 LustreError: 1907:0:(mdd_dir.c:4619:mdd_migrate_cmd_check()) lustre-MDD0001: '6' migration was interrupted, run 'lfs migrate -m 2 -c 2 -H crush 6' to finish migration: rc = -1 LustreError: 27585:0:(lcommon_cl.c:195:cl_file_inode_init()) lustre: failed to initialize cl_object [0x280000404:0x19a:0x0]: rc = -5 LustreError: 27585:0:(llite_lib.c:3745:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 29156:0:(file.c:267:ll_close_inode_openhandle()) lustre-clilmv-ffff8800b97c2548: inode [0x200000403:0x84:0x0] mdc close failed: rc = -116 LustreError: 29156:0:(file.c:267:ll_close_inode_openhandle()) Skipped 6 previous similar messages LustreError: 30279:0:(lcommon_cl.c:195:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000403:0x1e8:0x0]: rc = -5 LustreError: 30279:0:(lcommon_cl.c:195:cl_file_inode_init()) Skipped 1 previous similar message LustreError: 30279:0:(llite_lib.c:3745:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 30279:0:(llite_lib.c:3745:ll_prep_inode()) Skipped 1 previous similar message LustreError: 9652:0:(mdt_reint.c:2533:mdt_reint_migrate()) lustre-MDT0001: migrate [0x240000403:0x1:0x0]/4 failed: rc = -16 LustreError: 9652:0:(mdt_reint.c:2533:mdt_reint_migrate()) Skipped 33 previous similar messages LustreError: 26313:0:(mdd_dir.c:4619:mdd_migrate_cmd_check()) lustre-MDD0002: '19' migration was interrupted, run 'lfs migrate -m 0 -c 3 -H crush 19' to finish migration: rc = -1 LustreError: 26313:0:(mdd_dir.c:4619:mdd_migrate_cmd_check()) Skipped 2 previous similar messages Lustre: dir [0x200000403:0x183:0x0] stripe 1 readdir failed: -2, directory is partially accessed! LustreError: 931:0:(file.c:267:ll_close_inode_openhandle()) lustre-clilmv-ffff8802cba4d3d8: inode [0x200000404:0x22:0x0] mdc close failed: rc = -116 LustreError: 931:0:(file.c:267:ll_close_inode_openhandle()) Skipped 1 previous similar message LustreError: 701:0:(lcommon_cl.c:195:cl_file_inode_init()) lustre: failed to initialize cl_object [0x280000404:0x24d:0x0]: rc = -5 LustreError: 701:0:(llite_lib.c:3745:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 6294:0:(lcommon_cl.c:195:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000403:0x1e8:0x0]: rc = -5 LustreError: 6294:0:(lcommon_cl.c:195:cl_file_inode_init()) Skipped 1 previous similar message LustreError: 6294:0:(llite_lib.c:3745:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 6294:0:(llite_lib.c:3745:ll_prep_inode()) Skipped 1 previous similar message LustreError: 701:0:(lcommon_cl.c:195:cl_file_inode_init()) lustre: failed to initialize cl_object [0x280000404:0x278:0x0]: rc = -5 LustreError: 701:0:(lcommon_cl.c:195:cl_file_inode_init()) Skipped 4 previous similar messages LustreError: 701:0:(llite_lib.c:3745:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 701:0:(llite_lib.c:3745:ll_prep_inode()) Skipped 4 previous similar messages Lustre: lustre-MDT0002: trigger partial OI scrub for RPC inconsistency, checking FID [0x280000404:0x1e7:0x0]/0xa): rc = 0 Lustre: dir [0x280000404:0x1e7:0x0] stripe 2 readdir failed: -2, directory is partially accessed! Lustre: Skipped 3 previous similar messages LustreError: 29652:0:(mdd_dir.c:4619:mdd_migrate_cmd_check()) lustre-MDD0002: '18' migration was interrupted, run 'lfs migrate -m 1 -c 1 -H crush 18' to finish migration: rc = -1 LustreError: 29652:0:(mdd_dir.c:4619:mdd_migrate_cmd_check()) Skipped 1 previous similar message LustreError: 11393:0:(lcommon_cl.c:195:cl_file_inode_init()) lustre: failed to initialize cl_object [0x280000403:0x1aa:0x0]: rc = -5 LustreError: 11393:0:(lcommon_cl.c:195:cl_file_inode_init()) Skipped 33 previous similar messages LustreError: 11393:0:(llite_lib.c:3745:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 11393:0:(llite_lib.c:3745:ll_prep_inode()) Skipped 33 previous similar messages LustreError: 10080:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8800b97c2548: cannot apply new layout on [0x280000404:0x24d:0x0] : rc = -5 LustreError: 10080:0:(vvp_io.c:1921:vvp_io_init()) lustre: refresh file layout [0x280000404:0x24d:0x0] error -5. Lustre: lustre-MDT0001: trigger partial OI scrub for RPC inconsistency, checking FID [0x240000404:0x67:0x0]/0xa): rc = 0 LustreError: 19601:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8800b97c2548: cannot apply new layout on [0x280000404:0x24d:0x0] : rc = -5 LustreError: 20054:0:(lcommon_cl.c:195:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000403:0x21a:0x0]: rc = -5 LustreError: 20054:0:(lcommon_cl.c:195:cl_file_inode_init()) Skipped 20 previous similar messages LustreError: 20054:0:(llite_lib.c:3745:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 20054:0:(llite_lib.c:3745:ll_prep_inode()) Skipped 20 previous similar messages Lustre: 22524:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0x42d:0x0] with magic=0xbd60bd0 Lustre: 22524:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message LustreError: 28947:0:(mdt_handler.c:776:mdt_pack_acl2body()) lustre-MDT0000: unable to read [0x200000403:0x2f2:0x0] ACL: rc = -2 Lustre: dir [0x240000403:0x385:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 2 previous similar messages Lustre: dir [0x280000404:0x2ec:0x0] stripe 1 readdir failed: -2, directory is partially accessed! traps: 10[24527] general protection ip:401360 sp:7ffd0ff0b818 error:0 in 10[400000+6000] LustreError: 26249:0:(mdt_reint.c:2533:mdt_reint_migrate()) lustre-MDT0001: migrate [0x280000404:0x3a4:0x0]/1 failed: rc = -114 LustreError: 26249:0:(mdt_reint.c:2533:mdt_reint_migrate()) Skipped 64 previous similar messages LustreError: 28861:0:(mdd_dir.c:4619:mdd_migrate_cmd_check()) lustre-MDD0001: '6' migration was interrupted, run 'lfs migrate -m 2 -c 2 -H crush 6' to finish migration: rc = -1 LustreError: 28861:0:(mdd_dir.c:4619:mdd_migrate_cmd_check()) Skipped 5 previous similar messages LustreError: 22286:0:(file.c:267:ll_close_inode_openhandle()) lustre-clilmv-ffff8802cba4d3d8: inode [0x280000403:0x1da:0x0] mdc close failed: rc = -2 LustreError: 22286:0:(file.c:267:ll_close_inode_openhandle()) Skipped 1 previous similar message Lustre: 28110:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000404:0x208:0x0] with magic=0xbd60bd0 Lustre: 28110:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message LustreError: 1943:0:(lcommon_cl.c:195:cl_file_inode_init()) lustre: failed to initialize cl_object [0x280000403:0x4df:0x0]: rc = -5 LustreError: 1943:0:(lcommon_cl.c:195:cl_file_inode_init()) Skipped 42 previous similar messages LustreError: 1943:0:(llite_lib.c:3745:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 1943:0:(llite_lib.c:3745:ll_prep_inode()) Skipped 42 previous similar messages Lustre: dir [0x240000404:0x58c:0x0] stripe 1 readdir failed: -2, directory is partially accessed! LustreError: 391:0:(statahead.c:836:ll_statahead_interpret_work()) lustre: failed to prep 15 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 LustreError: 173:0:(statahead.c:836:ll_statahead_interpret_work()) lustre: failed to prep 14 [0x280000403:0x4df:0x0] inode@0000000000000000: rc = -5 Lustre: dir [0x280000404:0x20:0x0] stripe 2 readdir failed: -2, directory is partially accessed! LustreError: 11526:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8800b97c2548: cannot apply new layout on [0x280000404:0x24d:0x0] : rc = -5 Lustre: 6569:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0x77d:0x0] with magic=0xbd60bd0 Lustre: 6569:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 13 previous similar messages LustreError: 12473:0:(file.c:267:ll_close_inode_openhandle()) lustre-clilmv-ffff8802cba4d3d8: inode [0x200000404:0x5b0:0x0] mdc close failed: rc = -2 LustreError: 12473:0:(file.c:267:ll_close_inode_openhandle()) Skipped 9 previous similar messages LustreError: 17380:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8800b97c2548: cannot apply new layout on [0x280000404:0x24d:0x0] : rc = -5 LustreError: 17380:0:(lov_object.c:1341:lov_layout_change()) Skipped 3 previous similar messages LustreError: 17380:0:(vvp_io.c:1921:vvp_io_init()) lustre: refresh file layout [0x280000404:0x24d:0x0] error -5. LustreError: 87:0:(statahead.c:836:ll_statahead_interpret_work()) lustre: failed to prep 14 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 LustreError: 9652:0:(mdd_dir.c:4619:mdd_migrate_cmd_check()) lustre-MDD0002: '16' migration was interrupted, run 'lfs migrate -m 0 -c 1 -H crush 16' to finish migration: rc = -1 LustreError: 9652:0:(mdd_dir.c:4619:mdd_migrate_cmd_check()) Skipped 7 previous similar messages Lustre: dir [0x280000403:0x5c2:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 2 previous similar messages LustreError: 31362:0:(lcommon_cl.c:195:cl_file_inode_init()) lustre: failed to initialize cl_object [0x280000403:0x4df:0x0]: rc = -5 LustreError: 31362:0:(lcommon_cl.c:195:cl_file_inode_init()) Skipped 108 previous similar messages LustreError: 31362:0:(llite_lib.c:3745:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 31362:0:(llite_lib.c:3745:ll_prep_inode()) Skipped 108 previous similar messages Lustre: 2161:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000403:0x4df:0x0] with magic=0xbd60bd0 Lustre: 2161:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 29 previous similar messages LustreError: 399:0:(statahead.c:836:ll_statahead_interpret_work()) lustre: failed to prep 11 [0x280000403:0x73f:0x0] inode@0000000000000000: rc = -5 LustreError: 399:0:(statahead.c:836:ll_statahead_interpret_work()) Skipped 4 previous similar messages 8[6915]: segfault at 8 ip 00007f37c31377e8 sp 00007fff733b0a20 error 4 in ld-2.17.so[7f37c312c000+22000] Lustre: 28560:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000403:0x924:0x0] with magic=0xbd60bd0 Lustre: 28560:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 7 previous similar messages LustreError: 2268:0:(mdt_xattr.c:415:mdt_dir_layout_update()) lustre-MDT0001: [0x240000404:0x9c5:0x0] migrate mdt count mismatch 2 != 1 14[17678]: segfault at 8 ip 00007f02099237e8 sp 00007ffffb41d190 error 4 in ld-2.17.so[7f0209918000+22000] LustreError: 17678:0:(file.c:267:ll_close_inode_openhandle()) lustre-clilmv-ffff8802cba4d3d8: inode [0x240000404:0x980:0x0] mdc close failed: rc = -13 LustreError: 17678:0:(file.c:267:ll_close_inode_openhandle()) Skipped 4 previous similar messages LustreError: 9659:0:(mdt_reint.c:2533:mdt_reint_migrate()) lustre-MDT0002: migrate [0x240000404:0x9c5:0x0]/4 failed: rc = -16 LustreError: 9659:0:(mdt_reint.c:2533:mdt_reint_migrate()) Skipped 140 previous similar messages 14[23733]: segfault at 8 ip 00007f88955e87e8 sp 00007ffc34918960 error 4 in ld-2.17.so[7f88955dd000+22000] 14[28501]: segfault at 8 ip 00007fb7e21567e8 sp 00007ffd2add4390 error 4 in ld-2.17.so[7fb7e214b000+22000] Lustre: lustre-MDT0000: trigger partial OI scrub for RPC inconsistency, checking FID [0x200000403:0xa56:0x0]/0xa): rc = 0 7[1185]: segfault at 8 ip 00007f4a45c9a7e8 sp 00007ffefcbab3a0 error 4 in ld-2.17.so[7f4a45c8f000+22000] 5[6891]: segfault at 8 ip 00007f8d4bc077e8 sp 00007ffe29d9bdf0 error 4 in ld-2.17.so[7f8d4bbfc000+22000] LustreError: 2902:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8802cba4d3d8: cannot apply new layout on [0x200000403:0xc06:0x0] : rc = -5 LustreError: 2902:0:(lov_object.c:1341:lov_layout_change()) Skipped 1 previous similar message LustreError: 2902:0:(vvp_io.c:1921:vvp_io_init()) lustre: refresh file layout [0x200000403:0xc06:0x0] error -5. Lustre: dir [0x200000404:0xc6c:0x0] stripe 3 readdir failed: -2, directory is partially accessed! Lustre: Skipped 13 previous similar messages LustreError: 12835:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8800b97c2548: cannot apply new layout on [0x200000404:0xdc8:0x0] : rc = -5 LustreError: 12835:0:(vvp_io.c:1921:vvp_io_init()) lustre: refresh file layout [0x200000404:0xdc8:0x0] error -5. LustreError: 22982:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8800b97c2548: cannot apply new layout on [0x200000404:0xdc8:0x0] : rc = -5 Lustre: 28502:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000404:0xc30:0x0] with magic=0xbd60bd0 Lustre: 28502:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 25 previous similar messages LustreError: 26313:0:(mdd_dir.c:4619:mdd_migrate_cmd_check()) lustre-MDD0002: '0' migration was interrupted, run 'lfs migrate -m 0 -c 3 -H crush 0' to finish migration: rc = -1 LustreError: 26313:0:(mdd_dir.c:4619:mdd_migrate_cmd_check()) Skipped 10 previous similar messages LustreError: 29736:0:(lcommon_cl.c:195:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000403:0xf0e:0x0]: rc = -5 LustreError: 29736:0:(lcommon_cl.c:195:cl_file_inode_init()) Skipped 74 previous similar messages LustreError: 29736:0:(llite_lib.c:3745:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 29736:0:(llite_lib.c:3745:ll_prep_inode()) Skipped 74 previous similar messages LustreError: 10022:0:(file.c:267:ll_close_inode_openhandle()) lustre-clilmv-ffff8800b97c2548: inode [0x200000404:0xef0:0x0] mdc close failed: rc = -116 LustreError: 10022:0:(file.c:267:ll_close_inode_openhandle()) Skipped 7 previous similar messages LustreError: 15471:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8800b97c2548: cannot apply new layout on [0x200000404:0xdc8:0x0] : rc = -5 LustreError: 15359:0:(mdd_object.c:3866:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000403:0xf6e:0x0]: rc = -2 LustreError: 15359:0:(mdd_object.c:3866:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000403:0xf6e:0x0]: rc = -2 3[28511]: segfault at 8 ip 00007efcddae17e8 sp 00007fff19030110 error 4 in ld-2.17.so[7efcddad6000+22000] LustreError: 7668:0:(vvp_io.c:1921:vvp_io_init()) lustre: refresh file layout [0x280000404:0xd09:0x0] error -5. Lustre: dir [0x200000404:0x1374:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 19 previous similar messages Lustre: 6549:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000403:0x1045:0x0] with magic=0xbd60bd0 Lustre: 6549:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 37 previous similar messages LustreError: 384:0:(statahead.c:836:ll_statahead_interpret_work()) lustre: failed to prep 7 [0x200000403:0x13d7:0x0] inode@0000000000000000: rc = -5 Lustre: mdt00_061: service thread pid 4282 was inactive for 40.076 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 4282, comm: mdt00_061 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x913/0xd50 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x800 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_check_lock+0xec/0x3c0 [mdt] [<0>] mdt_object_stripes_lock+0xba/0x660 [mdt] [<0>] mdt_reint_unlink+0x7aa/0x15e0 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x74c/0xbc0 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 22407, comm: mdt00_088 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x913/0xd50 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x2ea/0x800 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock+0x88/0x1c0 [mdt] [<0>] mdt_getattr_name_lock+0x105/0x2d00 [mdt] [<0>] mdt_intent_getattr+0x2cc/0x4e0 [mdt] [<0>] mdt_intent_opc.constprop.73+0x211/0xc60 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x397/0x980 [ptlrpc] [<0>] ldlm_handle_enqueue+0x507/0x1870 [ptlrpc] [<0>] tgt_enqueue+0x68/0x240 [ptlrpc] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe 16[29221]: segfault at 0 ip (null) sp 00007ffe524034b8 error 14 in 16[400000+6000] Lustre: mdt00_045: service thread pid 2257 was inactive for 40.164 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Lustre: Skipped 1 previous similar message Pid: 2257, comm: mdt00_045 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x913/0xd50 [ptlrpc] [<0>] ldlm_cli_enqueue_fini+0xa54/0xf50 [ptlrpc] [<0>] ldlm_cli_enqueue+0x472/0xac0 [ptlrpc] [<0>] osp_md_object_lock+0x151/0x2f0 [osp] [<0>] lod_object_lock+0x7bc/0x7d0 [lod] [<0>] mdd_object_lock+0x2d/0xd0 [mdd] [<0>] mdt_object_stripes_lock+0x371/0x660 [mdt] [<0>] mdt_dir_layout_update+0x389/0x1080 [mdt] [<0>] mdt_reint_setxattr+0xa00/0x1020 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x74c/0xbc0 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc61/0x1640 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: mdt00_097: service thread pid 22544 was inactive for 40.087 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: mdt00_042: service thread pid 2161 was inactive for 40.163 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. LustreError: 15005:0:(ldlm_lockd.c:252:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0000_UUID lock: ffff8802dda816c0/0x708f3de7b963284 lrc: 3/0,0 mode: CR/CR res: [0x200000403:0x1395:0x0].0x0 bits 0xa/0x0 rrc: 3 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0x708f3de7b963276 expref: 909 pid: 28564 timeout: 654 lvb_type: 0 LustreError: lustre-MDT0000-mdc-ffff8800b97c2548: operation mds_reint to node 0@lo failed: rc = -107 Lustre: lustre-MDT0000-mdc-ffff8800b97c2548: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: mdt00_061: service thread pid 4282 completed after 100.273s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_088: service thread pid 22407 completed after 100.267s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_045: service thread pid 2257 completed after 96.290s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: lustre-MDT0000-mdc-ffff8800b97c2548: This client was evicted by lustre-MDT0000; in progress operations using this service will fail. LustreError: 6757:0:(mdc_request.c:1454:mdc_read_page()) lustre-MDT0000-mdc-ffff8800b97c2548: [0x200000404:0x1495:0x0] lock enqueue fails: rc = -108 LustreError: 5890:0:(vvp_io.c:1921:vvp_io_init()) lustre: refresh file layout [0x200000403:0x1395:0x0] error -5. LustreError: 9144:0:(llite_lib.c:2032:ll_md_setattr()) md_setattr fails: rc = -108 LustreError: 1832:0:(file.c:5979:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000007:0x1:0x0] error: rc = -108 LustreError: 26154:0:(llite_nfs.c:446:ll_dir_get_parent_fid()) lustre: failure inode [0x200000403:0x1760:0x0] get parent: rc = -108 Lustre: lustre-MDT0000-mdc-ffff8800b97c2548: Connection restored to 192.168.123.62@tcp (at 0@lo) Lustre: mdt00_097: service thread pid 22544 completed after 96.777s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_042: service thread pid 2161 completed after 96.335s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_008: service thread pid 1819 was inactive for 76.176 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: mdt_io00_021: service thread pid 30038 was inactive for 76.135 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 5 previous similar messages Lustre: mdt_io00_016: service thread pid 26171 was inactive for 76.115 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 5 previous similar messages LustreError: 15005:0:(ldlm_lockd.c:252:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0001_UUID lock: ffff8800a4be0b80/0x708f3de7ba0f7f1 lrc: 3/0,0 mode: PR/PR res: [0x240000403:0x12bc:0x0].0x0 bits 0x1b/0x0 rrc: 4 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0x708f3de7ba0f7b9 expref: 571 pid: 4136 timeout: 710 lvb_type: 0 LustreError: 9652:0:(mdt_reint.c:2533:mdt_reint_migrate()) lustre-MDT0001: migrate [0x240000403:0x1:0x0]/3 failed: rc = -16 LustreError: 9652:0:(mdt_reint.c:2533:mdt_reint_migrate()) Skipped 222 previous similar messages Lustre: mdt_io00_013: service thread pid 9652 completed after 101.355s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: lustre-MDT0001-mdc-ffff8800b97c2548: operation mds_close to node 0@lo failed: rc = -107 Lustre: lustre-MDT0001-mdc-ffff8800b97c2548: Connection to lustre-MDT0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: mdt_io00_008: service thread pid 1819 completed after 101.366s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: lustre-MDT0001-mdc-ffff8800b97c2548: This client was evicted by lustre-MDT0001; in progress operations using this service will fail. Lustre: mdt_io00_009: service thread pid 1907 completed after 99.855s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 6561:0:(file.c:5979:ll_inode_revalidate_fini()) lustre: revalidate FID [0x240000403:0x12bc:0x0] error: rc = -108 LustreError: 6561:0:(file.c:5979:ll_inode_revalidate_fini()) Skipped 16 previous similar messages LustreError: 8315:0:(mdc_request.c:1454:mdc_read_page()) lustre-MDT0001-mdc-ffff8800b97c2548: [0x240000401:0x18:0x0] lock enqueue fails: rc = -108 LustreError: 8315:0:(mdc_request.c:1454:mdc_read_page()) Skipped 8 previous similar messages Lustre: lustre-MDT0001-mdc-ffff8800b97c2548: Connection restored to 192.168.123.62@tcp (at 0@lo) Lustre: mdt_io00_018: service thread pid 26313 completed after 99.962s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_002: service thread pid 15373 completed after 99.409s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_021: service thread pid 30038 completed after 99.346s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_020: service thread pid 30029 completed after 99.423s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_015: service thread pid 10005 completed after 99.300s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_003: service thread pid 28861 completed after 99.050s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_006: service thread pid 29652 completed after 98.061s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_012: service thread pid 9317 completed after 99.820s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_000: service thread pid 15371 completed after 99.872s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_016: service thread pid 26171 completed after 99.753s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_010: service thread pid 1915 completed after 99.295s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). INFO: task mrename:766 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mrename D ffff88028717dc40 10912 766 26898 0x00000080 Call Trace: [<ffffffff817e19d9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df7ea>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa1d>] mutex_lock+0x2d/0x40 [<ffffffff81252061>] lock_rename+0x31/0xe0 [<ffffffff812586af>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed462>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff812597de>] SyS_renameat2+0xe/0x10 [<ffffffff8125981e>] SyS_rename+0x1e/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a INFO: task mrename:8289 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mrename D ffff8802b27ac9d0 12504 8289 27101 0x00000080 Call Trace: [<ffffffff817e19d9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df7ea>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa1d>] mutex_lock+0x2d/0x40 [<ffffffff81252061>] lock_rename+0x31/0xe0 [<ffffffff812586af>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed462>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff812597de>] SyS_renameat2+0xe/0x10 [<ffffffff8125981e>] SyS_rename+0x1e/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a INFO: task mrename:10407 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mrename D ffff8802adbde238 11744 10407 27031 0x00000080 Call Trace: [<ffffffff817e19d9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df7ea>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa1d>] mutex_lock+0x2d/0x40 [<ffffffff81252061>] lock_rename+0x31/0xe0 [<ffffffff812586af>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed462>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff812597de>] SyS_renameat2+0xe/0x10 [<ffffffff8125981e>] SyS_rename+0x1e/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a INFO: task mrename:11768 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mrename D ffff8802c8e90010 11744 11768 27127 0x00000080 Call Trace: [<ffffffff817e19d9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df7ea>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa1d>] mutex_lock+0x2d/0x40 [<ffffffff81252061>] lock_rename+0x31/0xe0 [<ffffffff812586af>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed462>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff812597de>] SyS_renameat2+0xe/0x10 [<ffffffff8125981e>] SyS_rename+0x1e/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a INFO: task mrename:12770 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mrename D ffff8802ce0cdc40 11744 12770 26870 0x00000080 Call Trace: [<ffffffff817e19d9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df7ea>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa1d>] mutex_lock+0x2d/0x40 [<ffffffff81252061>] lock_rename+0x31/0xe0 [<ffffffff812586af>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed462>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff812597de>] SyS_renameat2+0xe/0x10 [<ffffffff8125981e>] SyS_rename+0x1e/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a INFO: task mrename:13159 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mrename D ffff880099bb6b80 11040 13159 26977 0x00000080 Call Trace: [<ffffffff817e19d9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df7ea>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa1d>] mutex_lock+0x2d/0x40 [<ffffffff81252061>] lock_rename+0x31/0xe0 [<ffffffff812586af>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed462>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff812597de>] SyS_renameat2+0xe/0x10 [<ffffffff8125981e>] SyS_rename+0x1e/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a INFO: task mrename:16291 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mrename D ffff8802b2029878 12504 16291 26833 0x00000080 Call Trace: [<ffffffff817e19d9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df7ea>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa1d>] mutex_lock+0x2d/0x40 [<ffffffff81252061>] lock_rename+0x31/0xe0 [<ffffffff812586af>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed462>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff812597de>] SyS_renameat2+0xe/0x10 [<ffffffff8125981e>] SyS_rename+0x1e/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a INFO: task mrename:16912 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mrename D ffff8802bdfa8040 11744 16912 26968 0x00000080 Call Trace: [<ffffffff817e19d9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df7ea>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa1d>] mutex_lock+0x2d/0x40 [<ffffffff81252061>] lock_rename+0x31/0xe0 [<ffffffff812586af>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed462>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff812597de>] SyS_renameat2+0xe/0x10 [<ffffffff8125981e>] SyS_rename+0x1e/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a INFO: task mrename:20855 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mrename D ffff8802c8e93d58 11744 20855 27001 0x00000080 Call Trace: [<ffffffff817e19d9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df7ea>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa1d>] mutex_lock+0x2d/0x40 [<ffffffff81252061>] lock_rename+0x31/0xe0 [<ffffffff812586af>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed462>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff812597de>] SyS_renameat2+0xe/0x10 [<ffffffff8125981e>] SyS_rename+0x1e/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a INFO: task mrename:20896 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mrename D ffff8802b969a140 11744 20896 26891 0x00000080 Call Trace: [<ffffffff817e19d9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df7ea>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa1d>] mutex_lock+0x2d/0x40 [<ffffffff81252061>] lock_rename+0x31/0xe0 [<ffffffff812586af>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed462>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff812597de>] SyS_renameat2+0xe/0x10 [<ffffffff8125981e>] SyS_rename+0x1e/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a 11[13713]: segfault at 8 ip 00007f3a174037e8 sp 00007ffdd9f65b10 error 4 in ld-2.17.so[7f3a173f8000+22000] LustreError: 1907:0:(mdd_dir.c:4619:mdd_migrate_cmd_check()) lustre-MDD0001: '9' migration was interrupted, run 'lfs migrate -m 2 -c 2 -H crush 9' to finish migration: rc = -1 LustreError: 1907:0:(mdd_dir.c:4619:mdd_migrate_cmd_check()) Skipped 16 previous similar messages LustreError: 26647:0:(lcommon_cl.c:195:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000403:0x1099:0x0]: rc = -5 LustreError: 26647:0:(lcommon_cl.c:195:cl_file_inode_init()) Skipped 80 previous similar messages LustreError: 26647:0:(llite_lib.c:3745:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 26647:0:(llite_lib.c:3745:ll_prep_inode()) Skipped 80 previous similar messages LustreError: 26494:0:(file.c:267:ll_close_inode_openhandle()) lustre-clilmv-ffff8802cba4d3d8: inode [0x280000403:0x12b9:0x0] mdc close failed: rc = -2 LustreError: 26494:0:(file.c:267:ll_close_inode_openhandle()) Skipped 111 previous similar messages 16[5959]: segfault at 8 ip 00007fec947bb7e8 sp 00007ffdc58c5ea0 error 4 in ld-2.17.so[7fec947b0000+22000] LustreError: 3967:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8802cba4d3d8: cannot apply new layout on [0x240000405:0x214:0x0] : rc = -5 LustreError: 3967:0:(lov_object.c:1341:lov_layout_change()) Skipped 3 previous similar messages LustreError: 3967:0:(vvp_io.c:1921:vvp_io_init()) lustre: refresh file layout [0x240000405:0x214:0x0] error -5. LustreError: 127:0:(statahead.c:836:ll_statahead_interpret_work()) lustre: failed to prep 19 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 LustreError: 15604:0:(mdd_object.c:3866:mdd_close()) lustre-MDD0002: failed to get lu_attr of [0x280000404:0x12e6:0x0]: rc = -2 Lustre: 29016:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000403:0x1602:0x0] with magic=0xbd60bd0 Lustre: 29016:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 49 previous similar messages Lustre: dir [0x200000404:0x298a:0x0] stripe 3 readdir failed: -2, directory is partially accessed! Lustre: Skipped 29 previous similar messages LustreError: 127:0:(statahead.c:836:ll_statahead_interpret_work()) lustre: failed to prep 13 [0x200000405:0xba2:0x0] inode@0000000000000000: rc = -5 LustreError: 127:0:(statahead.c:836:ll_statahead_interpret_work()) Skipped 1 previous similar message LustreError: 11241:0:(vvp_io.c:1921:vvp_io_init()) lustre: refresh file layout [0x280000403:0x1671:0x0] error -5. 4[27720]: segfault at 8 ip 00007f7a3026a7e8 sp 00007fff49972610 error 4 in ld-2.17.so[7f7a3025f000+22000] 0[27732]: segfault at 8 ip 00007f58eda3b7e8 sp 00007ffd9bfa99f0 error 4 in ld-2.17.so[7f58eda30000+22000] 4[28719]: segfault at 8 ip 00007fa4a4b3f7e8 sp 00007ffd2fe67070 error 4 in ld-2.17.so[7fa4a4b34000+22000] LustreError: 28141:0:(mdt_xattr.c:415:mdt_dir_layout_update()) lustre-MDT0002: [0x280000404:0x16f5:0x0] migrate mdt count mismatch 1 != 3 LustreError: 391:0:(statahead.c:836:ll_statahead_interpret_work()) lustre: failed to prep 15 [0x280000404:0x18fa:0x0] inode@0000000000000000: rc = -5 LustreError: 28967:0:(mdt_handler.c:776:mdt_pack_acl2body()) lustre-MDT0000: unable to read [0x200000404:0x2bca:0x0] ACL: rc = -2 LustreError: 16790:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8802cba4d3d8: cannot apply new layout on [0x280000403:0x1671:0x0] : rc = -5 LustreError: 16790:0:(lov_object.c:1341:lov_layout_change()) Skipped 4 previous similar messages 9[18351]: segfault at 8 ip 00007f0caad077e8 sp 00007ffd82cfc3e0 error 4 in ld-2.17.so[7f0caacfc000+22000] LustreError: 6623:0:(mdt_handler.c:776:mdt_pack_acl2body()) lustre-MDT0001: unable to read [0x240000405:0x7b0:0x0] ACL: rc = -2 LustreError: 6623:0:(mdt_handler.c:776:mdt_pack_acl2body()) Skipped 1 previous similar message LustreError: 15005:0:(ldlm_lockd.c:252:expired_lock_main()) ### lock callback timer expired after 101s: evicting client at 0@lo ns: mdt-lustre-MDT0002_UUID lock: ffff8802ebaded00/0x708f3de7bd9116d lrc: 3/0,0 mode: PR/PR res: [0x280000404:0x1a85:0x0].0x0 bits 0x1b/0x0 rrc: 4 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0x708f3de7bd91151 expref: 498 pid: 29044 timeout: 1089 lvb_type: 0 LustreError: lustre-MDT0002-mdc-ffff8800b97c2548: operation mds_close to node 0@lo failed: rc = -107 Lustre: lustre-MDT0002-mdc-ffff8800b97c2548: Connection to lustre-MDT0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: lustre-MDT0002-mdc-ffff8800b97c2548: This client was evicted by lustre-MDT0002; in progress operations using this service will fail. LustreError: 27680:0:(file.c:5979:ll_inode_revalidate_fini()) lustre: revalidate FID [0x280000404:0x1a85:0x0] error: rc = -108 Lustre: lustre-MDT0002-mdc-ffff8800b97c2548: Connection restored to 192.168.123.62@tcp (at 0@lo) LustreError: 19481:0:(vvp_io.c:1921:vvp_io_init()) lustre: refresh file layout [0x280000403:0x1893:0x0] error -5. 11[30509]: segfault at 8 ip 00007f3eb99637e8 sp 00007ffea9db7a90 error 4 in ld-2.17.so[7f3eb9958000+22000] LustreError: 174:0:(statahead.c:836:ll_statahead_interpret_work()) lustre: failed to prep 13 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 LustreError: 174:0:(statahead.c:836:ll_statahead_interpret_work()) Skipped 4 previous similar messages 5[7625]: segfault at 8 ip 00007f211743f7e8 sp 00007ffeb21d8840 error 4 in ld-2.17.so[7f2117434000+22000] 6[14373]: segfault at 8 ip 00007feabd26f7e8 sp 00007ffdb9e4c680 error 4 in ld-2.17.so[7feabd264000+22000] LustreError: 1907:0:(mdt_reint.c:2533:mdt_reint_migrate()) lustre-MDT0000: migrate [0x200000405:0x1785:0x0]/14 failed: rc = -16 LustreError: 1907:0:(mdt_reint.c:2533:mdt_reint_migrate()) Skipped 387 previous similar messages LustreError: 15005:0:(ldlm_lockd.c:252:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0000_UUID lock: ffff88025c257840/0x708f3de7be2697c lrc: 3/0,0 mode: PR/PR res: [0x200000405:0x152d:0x0].0x0 bits 0x1b/0x0 rrc: 4 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0x708f3de7be26905 expref: 598 pid: 1973 timeout: 1230 lvb_type: 0 LustreError: 5239:0:(ldlm_lockd.c:2591:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1735335848 with bad export cookie 506923094955836358 Lustre: lustre-MDT0000-mdc-ffff8802cba4d3d8: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: lustre-MDT0000-mdc-ffff8802cba4d3d8: This client was evicted by lustre-MDT0000; in progress operations using this service will fail. LustreError: 26862:0:(llite_lib.c:2032:ll_md_setattr()) md_setattr fails: rc = -5 LustreError: 26862:0:(llite_lib.c:2032:ll_md_setattr()) Skipped 1 previous similar message LustreError: 28508:0:(file.c:5979:ll_inode_revalidate_fini()) lustre: revalidate FID [0x280000404:0x1fc1:0x0] error: rc = -5 LustreError: 28895:0:(ldlm_resource.c:1149:ldlm_resource_complain()) lustre-MDT0000-mdc-ffff8802cba4d3d8: namespace resource [0x280000404:0x1adc:0x0].0x0 (ffff8802c87ee5c0) refcount nonzero (1) after lock cleanup; forcing cleanup. Lustre: lustre-MDT0000-mdc-ffff8802cba4d3d8: Connection restored to 192.168.123.62@tcp (at 0@lo) LustreError: 4198:0:(ldlm_lockd.c:1492:ldlm_handle_enqueue()) ### lock on destroyed export ffff8802cbaf1bf8 ns: mdt-lustre-MDT0000_UUID lock: ffff8800a3c82d40/0x708f3de7bef16ce lrc: 3/0,0 mode: PR/PR res: [0x200000402:0x7c:0x0].0x0 bits 0x12/0x0 rrc: 4 type: IBT gid 0 flags: 0x50200000000000 nid: 0@lo remote: 0x708f3de7bef16b9 expref: 4 pid: 4198 timeout: 0 lvb_type: 0 LustreError: 13648:0:(ldlm_lockd.c:1492:ldlm_handle_enqueue()) ### lock on destroyed export ffff8802cbaf1bf8 ns: mdt-lustre-MDT0000_UUID lock: ffff8800a5346580/0x708f3de7bef197c lrc: 3/0,0 mode: PR/PR res: [0x200000402:0x7b:0x0].0x0 bits 0x12/0x0 rrc: 5 type: IBT gid 0 flags: 0x50200000000000 nid: 0@lo remote: 0x708f3de7bef1967 expref: 2 pid: 13648 timeout: 0 lvb_type: 0 0[412]: segfault at 8 ip 00007f788ba937e8 sp 00007ffe4da85fd0 error 4 in ld-2.17.so[7f788ba88000+22000] LustreError: 29115:0:(mdt_xattr.c:415:mdt_dir_layout_update()) lustre-MDT0000: [0x200000405:0x1a09:0x0] migrate mdt count mismatch 3 != 2 LustreError: 26781:0:(lcommon_cl.c:195:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000405:0x1a03:0x0]: rc = -5 LustreError: 26781:0:(lcommon_cl.c:195:cl_file_inode_init()) Skipped 204 previous similar messages LustreError: 26781:0:(llite_lib.c:3745:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 26781:0:(llite_lib.c:3745:ll_prep_inode()) Skipped 204 previous similar messages LustreError: 15371:0:(mdd_dir.c:4619:mdd_migrate_cmd_check()) lustre-MDD0001: '13' migration was interrupted, run 'lfs migrate -m 2 -c 2 -H crush 13' to finish migration: rc = -1 LustreError: 15371:0:(mdd_dir.c:4619:mdd_migrate_cmd_check()) Skipped 40 previous similar messages LustreError: 9533:0:(mdd_object.c:3866:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000404:0x345a:0x0]: rc = -2 LustreError: 9533:0:(mdd_object.c:3866:mdd_close()) Skipped 1 previous similar message LustreError: 12876:0:(file.c:267:ll_close_inode_openhandle()) lustre-clilmv-ffff8802cba4d3d8: inode [0x200000404:0x345a:0x0] mdc close failed: rc = -2 LustreError: 12876:0:(file.c:267:ll_close_inode_openhandle()) Skipped 100 previous similar messages LustreError: 29380:0:(mdd_object.c:3866:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000406:0x1ca:0x0]: rc = -2 LustreError: 29380:0:(mdd_object.c:3866:mdd_close()) Skipped 1 previous similar message LustreError: 150:0:(statahead.c:836:ll_statahead_interpret_work()) lustre: failed to prep 16 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 LustreError: 150:0:(statahead.c:836:ll_statahead_interpret_work()) Skipped 15 previous similar messages LustreError: 1730:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8800b97c2548: cannot apply new layout on [0x240000404:0x26e7:0x0] : rc = -5 LustreError: 1730:0:(lov_object.c:1341:lov_layout_change()) Skipped 6 previous similar messages LustreError: 1730:0:(vvp_io.c:1921:vvp_io_init()) lustre: refresh file layout [0x240000404:0x26e7:0x0] error -5. Lustre: 15354:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000404:0x2959:0x0] with magic=0xbd60bd0 Lustre: 15354:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 167 previous similar messages LustreError: 15005:0:(ldlm_lockd.c:252:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0000_UUID lock: ffff8800a3f02d40/0x708f3de7c05a36d lrc: 3/0,0 mode: PR/PR res: [0x200000406:0x61f:0x0].0x0 bits 0x1b/0x0 rrc: 6 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0x708f3de7c05a33c expref: 509 pid: 2188 timeout: 1491 lvb_type: 0 LustreError: lustre-MDT0000-mdc-ffff8800b97c2548: operation mds_close to node 0@lo failed: rc = -107 Lustre: lustre-MDT0000-mdc-ffff8800b97c2548: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: lustre-MDT0000-mdc-ffff8800b97c2548: This client was evicted by lustre-MDT0000; in progress operations using this service will fail. LustreError: 15403:0:(file.c:5979:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000007:0x1:0x0] error: rc = -108 LustreError: 15403:0:(file.c:5979:ll_inode_revalidate_fini()) Skipped 12 previous similar messages Lustre: lustre-MDT0000-mdc-ffff8800b97c2548: Connection restored to 192.168.123.62@tcp (at 0@lo) Lustre: dir [0x240000404:0x3652:0x0] stripe 3 readdir failed: -2, directory is partially accessed! Lustre: Skipped 28 previous similar messages LustreError: 17508:0:(mdd_object.c:3866:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000405:0x1b7a:0x0]: rc = -2 LustreError: 17508:0:(mdd_object.c:3866:mdd_close()) Skipped 1 previous similar message LustreError: 15005:0:(ldlm_lockd.c:252:expired_lock_main()) ### lock callback timer expired after 101s: evicting client at 0@lo ns: mdt-lustre-MDT0002_UUID lock: ffff88026da88b80/0x708f3de7c15bdda lrc: 3/0,0 mode: PR/PR res: [0x280000404:0x2710:0x0].0x0 bits 0x1b/0x0 rrc: 4 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0x708f3de7c15bdbe expref: 462 pid: 8246 timeout: 1654 lvb_type: 0 LustreError: lustre-MDT0002-mdc-ffff8800b97c2548: operation mds_close to node 0@lo failed: rc = -107 Lustre: lustre-MDT0002-mdc-ffff8800b97c2548: Connection to lustre-MDT0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: lustre-MDT0002-mdc-ffff8800b97c2548: This client was evicted by lustre-MDT0002; in progress operations using this service will fail. LustreError: 19817:0:(llite_lib.c:2032:ll_md_setattr()) md_setattr fails: rc = -108 LustreError: 19817:0:(llite_lib.c:2032:ll_md_setattr()) Skipped 1 previous similar message LustreError: 23687:0:(file.c:5979:ll_inode_revalidate_fini()) lustre: revalidate FID [0x280000404:0x2710:0x0] error: rc = -108 Lustre: lustre-MDT0002-mdc-ffff8800b97c2548: Connection restored to 192.168.123.62@tcp (at 0@lo) LustreError: 87:0:(statahead.c:836:ll_statahead_interpret_work()) lustre: failed to prep 6 [0x240000404:0x35c2:0x0] inode@0000000000000000: rc = -5 LustreError: 87:0:(statahead.c:836:ll_statahead_interpret_work()) Skipped 6 previous similar messages LustreError: 9317:0:(lustre_lmv.h:500:lmv_is_sane()) unknown layout LMV: magic=0xcd40cd0 count=3 index=2 hash=crush:0x82000003 version=1 migrate_offset=2 migrate_hash=fnv_1a_64:2 pool= LustreError: 6617:0:(mdd_object.c:384:mdd_xattr_get()) lustre-MDD0000: object [0x200000407:0x77d:0x0] not found: rc = -2 LustreError: 22070:0:(statahead.c:2447:start_statahead_thread()) lustre: unsupported statahead pattern 0X0. LustreError: 13913:0:(vvp_io.c:1921:vvp_io_init()) lustre: refresh file layout [0x200000406:0xb74:0x0] error -5. LustreError: 13913:0:(vvp_io.c:1921:vvp_io_init()) Skipped 2 previous similar messages LustreError: 2329:0:(mdt_xattr.c:415:mdt_dir_layout_update()) lustre-MDT0002: [0x280000406:0x2fc:0x0] migrate mdt count mismatch 1 != 2 LustreError: 10005:0:(mdt_reint.c:2533:mdt_reint_migrate()) lustre-MDT0001: migrate [0x240000404:0x3e1d:0x0]/20 failed: rc = -2 LustreError: 10005:0:(mdt_reint.c:2533:mdt_reint_migrate()) Skipped 397 previous similar messages LustreError: 30768:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8800b97c2548: cannot apply new layout on [0x280000404:0x2ea7:0x0] : rc = -5 LustreError: 30768:0:(lov_object.c:1341:lov_layout_change()) Skipped 21 previous similar messages LustreError: 174:0:(lcommon_cl.c:195:cl_file_inode_init()) lustre: failed to initialize cl_object [0x280000404:0x2e92:0x0]: rc = -5 LustreError: 174:0:(lcommon_cl.c:195:cl_file_inode_init()) Skipped 348 previous similar messages LustreError: 174:0:(llite_lib.c:3745:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 174:0:(llite_lib.c:3745:ll_prep_inode()) Skipped 348 previous similar messages LustreError: 1907:0:(mdd_dir.c:4619:mdd_migrate_cmd_check()) lustre-MDD0000: '6' migration was interrupted, run 'lfs migrate -m 2 -c 1 -H crush 6' to finish migration: rc = -1 LustreError: 1907:0:(mdd_dir.c:4619:mdd_migrate_cmd_check()) Skipped 36 previous similar messages LustreError: 9785:0:(file.c:267:ll_close_inode_openhandle()) lustre-clilmv-ffff8800b97c2548: inode [0x200000407:0xc85:0x0] mdc close failed: rc = -116 LustreError: 9785:0:(file.c:267:ll_close_inode_openhandle()) Skipped 91 previous similar messages 11[12840]: segfault at 8 ip 00007fbcea3a87e8 sp 00007ffc8b120bd0 error 4 in ld-2.17.so[7fbcea39d000+22000] LustreError: 1511:0:(llite_lib.c:1881:ll_update_lsm_md()) lustre: [0x200000407:0xe74:0x0] dir layout mismatch: LustreError: 1511:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=2 count=2 index=0 hash=crush:0x2000003 max_inherit=0 max_inherit_rr=0 version=2 migrate_offset=0 migrate_hash=invalid:0 pool= LustreError: 1511:0:(lustre_lmv.h:167:lmv_stripe_object_dump()) stripe[0] [0x200000400:0xed:0x0] LustreError: 1511:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=1 count=3 index=0 hash=crush:0x82000003 max_inherit=0 max_inherit_rr=0 version=1 migrate_offset=2 migrate_hash=fnv_1a_64:2 pool= LustreError: 15358:0:(mdd_object.c:3866:mdd_close()) lustre-MDD0002: failed to get lu_attr of [0x280000404:0x2fb4:0x0]: rc = -2 LustreError: 15358:0:(mdd_object.c:3866:mdd_close()) Skipped 1 previous similar message 18[9262]: segfault at 8 ip 00007f77703547e8 sp 00007fffd8a69070 error 4 in ld-2.17.so[7f7770349000+22000] Lustre: 13671:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000405:0x2334:0x0] with magic=0xbd60bd0 Lustre: 13671:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 111 previous similar messages LustreError: 15005:0:(ldlm_lockd.c:252:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0000_UUID lock: ffff8802c00fa5c0/0x708f3de7c4d855c lrc: 3/0,0 mode: PR/PR res: [0x200000406:0x16a2:0x0].0x0 bits 0x1b/0x0 rrc: 3 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0x708f3de7c4d8539 expref: 651 pid: 22565 timeout: 2072 lvb_type: 0 LustreError: lustre-MDT0000-mdc-ffff8800b97c2548: operation mds_close to node 0@lo failed: rc = -107 Lustre: lustre-MDT0000-mdc-ffff8800b97c2548: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: lustre-MDT0000-mdc-ffff8800b97c2548: This client was evicted by lustre-MDT0000; in progress operations using this service will fail. LustreError: 19684:0:(file.c:5979:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000406:0x16a2:0x0] error: rc = -108 Lustre: lustre-MDT0000-mdc-ffff8800b97c2548: Connection restored to 192.168.123.62@tcp (at 0@lo) LustreError: 22544:0:(mdt_xattr.c:415:mdt_dir_layout_update()) lustre-MDT0002: [0x280000406:0xa70:0x0] migrate mdt count mismatch 2 != 3 12[13200]: segfault at 8 ip 00007f3a2c8a77e8 sp 00007ffc9ab45dd0 error 4 in ld-2.17.so[7f3a2c89c000+22000] LustreError: 8083:0:(mdt_handler.c:776:mdt_pack_acl2body()) lustre-MDT0000: unable to read [0x200000407:0xf73:0x0] ACL: rc = -2 Lustre: dir [0x280000406:0xe7d:0x0] stripe 3 readdir failed: -2, directory is partially accessed! Lustre: Skipped 36 previous similar messages LustreError: 29105:0:(mdt_xattr.c:415:mdt_dir_layout_update()) lustre-MDT0001: [0x240000404:0x4197:0x0] migrate mdt count mismatch 2 != 3 9[9395]: segfault at 8 ip 00007ff7e42517e8 sp 00007ffc20b5a4a0 error 4 in ld-2.17.so[7ff7e4246000+22000] LustreError: 87:0:(statahead.c:836:ll_statahead_interpret_work()) lustre: failed to prep 19 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 LustreError: 87:0:(statahead.c:836:ll_statahead_interpret_work()) Skipped 11 previous similar messages LustreError: 15005:0:(ldlm_lockd.c:252:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0001_UUID lock: ffff88025ac64b40/0x708f3de7c5b1857 lrc: 3/0,0 mode: PW/PW res: [0x240000405:0x2646:0x0].0x0 bits 0x4/0x0 rrc: 3 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0x708f3de7c5b1842 expref: 658 pid: 22407 timeout: 2244 lvb_type: 0 LustreError: 28141:0:(ldlm_lockd.c:1492:ldlm_handle_enqueue()) ### lock on destroyed export ffff8802a07c12a8 ns: mdt-lustre-MDT0001_UUID lock: ffff8802cb1dde00/0x708f3de7c6c175c lrc: 3/0,0 mode: PR/PR res: [0x240000405:0x29a3:0x0].0x0 bits 0x1b/0x0 rrc: 2 type: IBT gid 0 flags: 0x50200000000000 nid: 0@lo remote: 0x708f3de7c6be7f2 expref: 11 pid: 28141 timeout: 0 lvb_type: 0 LustreError: lustre-MDT0001-mdc-ffff8800b97c2548: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-MDT0001-mdc-ffff8800b97c2548: Connection to lustre-MDT0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: lustre-MDT0001-mdc-ffff8800b97c2548: This client was evicted by lustre-MDT0001; in progress operations using this service will fail. LustreError: 31493:0:(llite_lib.c:2032:ll_md_setattr()) md_setattr fails: rc = -108 LustreError: 31493:0:(llite_lib.c:2032:ll_md_setattr()) Skipped 1 previous similar message LustreError: 5297:0:(mdc_request.c:1454:mdc_read_page()) lustre-MDT0001-mdc-ffff8800b97c2548: [0x240000405:0x2942:0x0] lock enqueue fails: rc = -108 LustreError: 5297:0:(statahead.c:1806:is_first_dirent()) lustre: reading dir [0x240000405:0x2942:0x0] at 0 stat_pid = 5297 : rc = -108 Lustre: lustre-MDT0001-mdc-ffff8800b97c2548: Connection restored to 192.168.123.62@tcp (at 0@lo) 4[16950]: segfault at 8 ip 00007fc4cffcd7e8 sp 00007ffc13483050 error 4 in ld-2.17.so[7fc4cffc2000+22000] 7[21090]: segfault at 8 ip 00007f21ef3307e8 sp 00007fff6d32e050 error 4 in ld-2.17.so[7f21ef325000+22000] | Link to test |
racer test 1: racer on clients: oleg405-client.virtnet DURATION=300 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffff813f40ac>] _atomic_dec_and_lock+0xc/0x70 PGD a5d49067 PUD a5d4a067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) osp(OE) ofd(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) crc_t10dif crct10dif_generic rpcsec_gss_krb5 sb_edac edac_core iosf_mbi kvm_intel kvm irqbypass crc32_pclmul ghash_clmulni_intel squashfs aesni_intel lrw i2c_piix4 gf128mul glue_helper i2c_core ablk_helper cryptd pcspkr binfmt_misc ip_tables ext4 mbcache jbd2 ata_generic pata_acpi crct10dif_pclmul crct10dif_common crc32c_intel ata_piix serio_raw libata CPU: 0 PID: 18784 Comm: ll_sa_18742 Kdump: loaded Tainted: G OE ------------ 3.10.0-7.9-debug #1 Hardware name: QEMU Standard PC (i440FX + PIIX, 1996), BIOS 1.16.2-1.fc38 04/01/2014 task: ffff8800a278a220 ti: ffff880089f08000 task.ti: ffff880089f08000 RIP: 0010:[<ffffffff813f40ac>] [<ffffffff813f40ac>] _atomic_dec_and_lock+0xc/0x70 RSP: 0018:ffff880089f0bd80 EFLAGS: 00010282 RAX: ffff8800892fe0c8 RBX: 0000000000000008 RCX: 00000000000e3d02 RDX: 00000000000e3d01 RSI: ffff8800892fe3f8 RDI: 0000000000000008 RBP: ffff880089f0bd90 R08: ffff880089f0bfd8 R09: 002dffd0ce3dec73 R10: 0000000000000009 R11: 0000000000004978 R12: 0000000000000000 R13: ffff8800a5e40000 R14: ffff8800892fe3f8 R15: ffff88009ec73c00 FS: 0000000000000000(0000) GS:ffff88013e200000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 CR2: 0000000000000008 CR3: 00000000a5d60000 CR4: 0000000000160ff0 Call Trace: [<ffffffffa1091a61>] ll_statahead_thread+0xde1/0x2290 [lustre] [<ffffffff810df6e2>] ? pick_next_task_fair+0x702/0x8a0 [<ffffffff81410a79>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e8dce>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffff810c834d>] ? finish_task_switch+0x5d/0x1b0 [<ffffffffa1090c80>] ? ll_statahead_handle.constprop.29+0x1c0/0x1c0 [lustre] [<ffffffff810ba114>] kthread+0xe4/0xf0 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817f3e5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 | random: crng init done LustreError: 11713:0:(file.c:262:ll_close_inode_openhandle()) lustre-clilmv-ffff88012e144000: inode [0x200000403:0x6:0x0] mdc close failed: rc = -116 LustreError: 12479:0:(file.c:262:ll_close_inode_openhandle()) lustre-clilmv-ffff88012e144000: inode [0x200000402:0x5:0x0] mdc close failed: rc = -116 LustreError: 12479:0:(file.c:262:ll_close_inode_openhandle()) Skipped 4 previous similar messages 5[13529]: segfault at 8 ip 00007f64f99727e8 sp 00007ffcf1ece6f0 error 4 in ld-2.17.so[7f64f9967000+22000] LustreError: 12986:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8800b67e1800: cannot apply new layout on [0x200000403:0xb6:0x0] : rc = -5 LustreError: 12986:0:(vvp_io.c:1921:vvp_io_init()) lustre: refresh file layout [0x200000403:0xb6:0x0] error -5. LustreError: 12979:0:(file.c:262:ll_close_inode_openhandle()) lustre-clilmv-ffff8800b67e1800: inode [0x240000402:0xb6:0x0] mdc close failed: rc = -116 LustreError: 12979:0:(file.c:262:ll_close_inode_openhandle()) Skipped 1 previous similar message Lustre: dir [0x240000402:0xde:0x0] stripe 2 readdir failed: -2, directory is partially accessed! LustreError: 14778:0:(lcommon_cl.c:195:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000402:0x11f:0x0]: rc = -5 LustreError: 14778:0:(llite_lib.c:3737:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 16121:0:(file.c:262:ll_close_inode_openhandle()) lustre-clilmv-ffff88012e144000: inode [0x200000402:0x103:0x0] mdc close failed: rc = -116 LustreError: 16121:0:(file.c:262:ll_close_inode_openhandle()) Skipped 3 previous similar messages LustreError: 15712:0:(lcommon_cl.c:195:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000402:0x11f:0x0]: rc = -5 LustreError: 15712:0:(lcommon_cl.c:195:cl_file_inode_init()) Skipped 2 previous similar messages LustreError: 16453:0:(llite_lib.c:3737:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 16453:0:(llite_lib.c:3737:ll_prep_inode()) Skipped 2 previous similar messages LustreError: 16777:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8800b67e1800: cannot apply new layout on [0x240000403:0xcb:0x0] : rc = -5 LustreError: 16777:0:(vvp_io.c:1921:vvp_io_init()) lustre: refresh file layout [0x240000403:0xcb:0x0] error -5. LustreError: 17188:0:(lcommon_cl.c:195:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000403:0xec:0x0]: rc = -5 LustreError: 17188:0:(lcommon_cl.c:195:cl_file_inode_init()) Skipped 4 previous similar messages LustreError: 17188:0:(llite_lib.c:3737:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 17188:0:(llite_lib.c:3737:ll_prep_inode()) Skipped 4 previous similar messages LustreError: 18991:0:(file.c:262:ll_close_inode_openhandle()) lustre-clilmv-ffff8800b67e1800: inode [0x200000402:0x103:0x0] mdc close failed: rc = -116 LustreError: 18991:0:(file.c:262:ll_close_inode_openhandle()) Skipped 6 previous similar messages LustreError: 10533:0:(lcommon_cl.c:195:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000403:0xec:0x0]: rc = -5 LustreError: 10533:0:(lcommon_cl.c:195:cl_file_inode_init()) Skipped 2 previous similar messages LustreError: 10533:0:(llite_lib.c:3737:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 10533:0:(llite_lib.c:3737:ll_prep_inode()) Skipped 2 previous similar messages 10[19618]: segfault at 1c ip 00000000004017b0 sp 00007ffd8f5ecba0 error 6 in 10[400000+6000] Lustre: dir [0x200000403:0x2ef:0x0] stripe 2 readdir failed: -2, directory is partially accessed! Lustre: Skipped 1 previous similar message Lustre: dir [0x240000402:0x318:0x0] stripe 2 readdir failed: -2, directory is partially accessed! 4[23192]: segfault at 8 ip 00007f98d38427e8 sp 00007fff144845f0 error 4 in ld-2.17.so[7f98d3837000+22000] LustreError: 23207:0:(lcommon_cl.c:195:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000402:0x434:0x0]: rc = -5 LustreError: 23207:0:(lcommon_cl.c:195:cl_file_inode_init()) Skipped 4 previous similar messages LustreError: 23207:0:(llite_lib.c:3737:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 23207:0:(llite_lib.c:3737:ll_prep_inode()) Skipped 4 previous similar messages Lustre: dir [0x200000402:0x53b:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 2 previous similar messages 7[25582]: segfault at 8 ip 00007f448534e7e8 sp 00007fffce339d70 error 4 in ld-2.17.so[7f4485343000+22000] LustreError: 26267:0:(file.c:262:ll_close_inode_openhandle()) lustre-clilmv-ffff8800b67e1800: inode [0x200000403:0x6a2:0x0] mdc close failed: rc = -116 LustreError: 26267:0:(file.c:262:ll_close_inode_openhandle()) Skipped 18 previous similar messages dd (26523) used greatest stack depth: 10352 bytes left Lustre: dir [0x240000403:0x30e:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 1 previous similar message LustreError: 29512:0:(lcommon_cl.c:195:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000402:0x458:0x0]: rc = -5 LustreError: 29512:0:(lcommon_cl.c:195:cl_file_inode_init()) Skipped 64 previous similar messages LustreError: 29512:0:(llite_lib.c:3737:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 29512:0:(llite_lib.c:3737:ll_prep_inode()) Skipped 64 previous similar messages LustreError: 30389:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8800b67e1800: cannot apply new layout on [0x240000403:0x436:0x0] : rc = -5 LustreError: 30389:0:(vvp_io.c:1921:vvp_io_init()) lustre: refresh file layout [0x240000403:0x436:0x0] error -5. 0[32264]: segfault at 0 ip 0000000000403cf0 sp 00007fff297c7ce8 error 6 in 0[400000+6000] LustreError: 591:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8800b67e1800: cannot apply new layout on [0x200000403:0xb45:0x0] : rc = -5 LustreError: 591:0:(lov_object.c:1341:lov_layout_change()) Skipped 2 previous similar messages LustreError: 591:0:(vvp_io.c:1921:vvp_io_init()) lustre: refresh file layout [0x200000403:0xb45:0x0] error -5. LustreError: 591:0:(vvp_io.c:1921:vvp_io_init()) Skipped 1 previous similar message Lustre: dir [0x200000403:0x8d6:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 63 previous similar messages LustreError: 2179:0:(llite_lib.c:1873:ll_update_lsm_md()) lustre: [0x200000402:0xb03:0x0] dir layout mismatch: LustreError: 2179:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=1 count=1 index=0 hash=crush:0x2000003 max_inherit=0 max_inherit_rr=0 version=2 migrate_offset=0 migrate_hash=invalid:0 pool= LustreError: 2179:0:(lustre_lmv.h:167:lmv_stripe_object_dump()) stripe[0] [0x200000400:0x2b:0x0] LustreError: 2179:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=1 count=3 index=0 hash=crush:0x82000003 max_inherit=0 max_inherit_rr=0 version=1 migrate_offset=1 migrate_hash=crush:2000003 pool= LustreError: 2110:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=4 count=1 index=0 hash=crush:0x2000003 max_inherit=0 max_inherit_rr=0 version=2 migrate_offset=0 migrate_hash=invalid:0 pool= LustreError: 2110:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=1 count=3 index=0 hash=crush:0x82000003 max_inherit=0 max_inherit_rr=0 version=1 migrate_offset=1 migrate_hash=crush:2000003 pool= LustreError: 4883:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8800b67e1800: cannot apply new layout on [0x240000403:0x436:0x0] : rc = -5 LustreError: 4883:0:(lov_object.c:1341:lov_layout_change()) Skipped 3 previous similar messages LustreError: 5223:0:(vvp_io.c:1921:vvp_io_init()) lustre: refresh file layout [0x240000403:0x599:0x0] error -5. LustreError: 5623:0:(mdc_request.c:1469:mdc_read_page()) lustre-MDT0001-mdc-ffff8800b67e1800: dir page locate: [0x240000401:0x15:0x0] at 0: rc -5 LustreError: 7001:0:(file.c:262:ll_close_inode_openhandle()) lustre-clilmv-ffff88012e144000: inode [0x200000403:0xd47:0x0] mdc close failed: rc = -116 LustreError: 7001:0:(file.c:262:ll_close_inode_openhandle()) Skipped 18 previous similar messages LustreError: 11464:0:(lcommon_cl.c:195:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000402:0x72a:0x0]: rc = -5 LustreError: 11464:0:(lcommon_cl.c:195:cl_file_inode_init()) Skipped 44 previous similar messages LustreError: 11464:0:(llite_lib.c:3737:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 11464:0:(llite_lib.c:3737:ll_prep_inode()) Skipped 44 previous similar messages LustreError: 10540:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8800b67e1800: cannot apply new layout on [0x240000403:0x436:0x0] : rc = -5 LustreError: 10540:0:(lov_object.c:1341:lov_layout_change()) Skipped 10 previous similar messages LustreError: 14827:0:(vvp_io.c:1921:vvp_io_init()) lustre: refresh file layout [0x240000403:0x436:0x0] error -5. Lustre: dir [0x200000403:0x8d6:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 185 previous similar messages 0[23082]: segfault at 8 ip 00007f0aabebf7e8 sp 00007ffdf17eaf80 error 4 in ld-2.17.so[7f0aabeb4000+22000] LustreError: 83:0:(statahead.c:836:ll_statahead_interpret_work()) lustre: failed to prep 5 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 LustreError: 83:0:(statahead.c:836:ll_statahead_interpret_work()) lustre: failed to prep 5 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 LustreError: 924:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8800b67e1800: cannot apply new layout on [0x240000403:0xf8a:0x0] : rc = -5 LustreError: 924:0:(lov_object.c:1341:lov_layout_change()) Skipped 7 previous similar messages LustreError: 924:0:(vvp_io.c:1921:vvp_io_init()) lustre: refresh file layout [0x240000403:0xf8a:0x0] error -5. LustreError: 924:0:(vvp_io.c:1921:vvp_io_init()) Skipped 1 previous similar message LustreError: 5618:0:(file.c:262:ll_close_inode_openhandle()) lustre-clilmv-ffff8800b67e1800: inode [0x200000403:0x1d1e:0x0] mdc close failed: rc = -116 LustreError: 5618:0:(file.c:262:ll_close_inode_openhandle()) Skipped 52 previous similar messages LustreError: 483:0:(statahead.c:836:ll_statahead_interpret_work()) lustre: failed to prep 15 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 LustreError: 7409:0:(lcommon_cl.c:195:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000402:0x1fe1:0x0]: rc = -5 LustreError: 7409:0:(lcommon_cl.c:195:cl_file_inode_init()) Skipped 99 previous similar messages LustreError: 7409:0:(llite_lib.c:3737:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 7409:0:(llite_lib.c:3737:ll_prep_inode()) Skipped 99 previous similar messages 1[12114]: segfault at 8 ip 00007fb6a3c747e8 sp 00007ffd63a48080 error 4 in ld-2.17.so[7fb6a3c69000+22000] Lustre: dir [0x200000403:0x25e5:0x0] stripe 2 readdir failed: -2, directory is partially accessed! Lustre: Skipped 57 previous similar messages LustreError: lustre-OST0001-osc-ffff8800b67e1800: BAD READ CHECKSUM: from 192.168.204.105@tcp inode [0x200000402:0x2624:0x0] object 0x2c0000401:1190 extent [0-4194303], client 971a59ac/971a59ac, server 8d02ccf7, cksum_type 4 LustreError: 2006:0:(osc_request.c:2455:osc_brw_redo_request()) @@@ redo for recoverable error -11 req@ffff88009da78a80 x1818611910897408/t0(0) o3->lustre-OST0001-osc-ffff8800b67e1800@192.168.204.105@tcp:6/4 lens 488/440 e 0 to 0 dl 1734363576 ref 2 fl Interpret:RMQU/200/0 rc 24576/24576 job:'cat.0' uid:0 gid:0 3[16872]: segfault at 8 ip 00007fc4bea927e8 sp 00007ffca0a3c920 error 4 in ld-2.17.so[7fc4bea87000+22000] 8[21811]: segfault at 8 ip 00007f20ba4867e8 sp 00007ffe11b67ec0 error 4 in ld-2.17.so[7f20ba47b000+22000] 10[26212]: segfault at 8 ip 00007ff5765847e8 sp 00007ffc52dc4b80 error 4 in ld-2.17.so[7ff576579000+22000] LustreError: 32058:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff8800b67e1800: cannot apply new layout on [0x240000403:0x1327:0x0] : rc = -5 LustreError: 32058:0:(lov_object.c:1341:lov_layout_change()) Skipped 38 previous similar messages LustreError: 46:0:(statahead.c:836:ll_statahead_interpret_work()) lustre: failed to prep 12 [0x200000403:0x28df:0x0] inode@0000000000000000: rc = -5 LustreError: 46:0:(statahead.c:836:ll_statahead_interpret_work()) Skipped 1 previous similar message Lustre: lustre-MDT0001-mdc-ffff8800b67e1800: Connection to lustre-MDT0001 (at 192.168.204.105@tcp) was lost; in progress operations using this service will wait for recovery to complete LustreError: lustre-MDT0001-mdc-ffff8800b67e1800: operation mds_getxattr to node 192.168.204.105@tcp failed: rc = -107 LustreError: lustre-MDT0001-mdc-ffff8800b67e1800: This client was evicted by lustre-MDT0001; in progress operations using this service will fail. LustreError: 14949:0:(mdc_request.c:1454:mdc_read_page()) lustre-MDT0001-mdc-ffff8800b67e1800: [0x240000402:0x1:0x0] lock enqueue fails: rc = -108 LustreError: 2317:0:(llite_lib.c:2024:ll_md_setattr()) md_setattr fails: rc = -108 Lustre: lustre-MDT0001-mdc-ffff8800b67e1800: Connection restored to (at 192.168.204.105@tcp) LustreError: lustre-MDT0000-mdc-ffff8800b67e1800: operation ldlm_enqueue to node 192.168.204.105@tcp failed: rc = -107 LustreError: Skipped 2 previous similar messages Lustre: lustre-MDT0000-mdc-ffff8800b67e1800: Connection to lustre-MDT0000 (at 192.168.204.105@tcp) was lost; in progress operations using this service will wait for recovery to complete LustreError: lustre-MDT0000-mdc-ffff8800b67e1800: This client was evicted by lustre-MDT0000; in progress operations using this service will fail. LustreError: 4578:0:(file.c:262:ll_close_inode_openhandle()) lustre-clilmv-ffff8800b67e1800: inode [0x200000402:0x3398:0x0] mdc close failed: rc = -108 LustreError: 2178:0:(vvp_io.c:1921:vvp_io_init()) lustre: refresh file layout [0x200000402:0x3136:0x0] error -108. LustreError: 2178:0:(vvp_io.c:1921:vvp_io_init()) Skipped 10 previous similar messages LustreError: 4578:0:(file.c:262:ll_close_inode_openhandle()) Skipped 95 previous similar messages LustreError: 3489:0:(file.c:5972:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000007:0x1:0x0] error: rc = -108 LustreError: 3489:0:(file.c:5972:ll_inode_revalidate_fini()) Skipped 7 previous similar messages LustreError: 2652:0:(mdc_request.c:1454:mdc_read_page()) lustre-MDT0000-mdc-ffff8800b67e1800: [0x200000402:0x2:0x0] lock enqueue fails: rc = -108 LustreError: 2652:0:(mdc_request.c:1454:mdc_read_page()) Skipped 13 previous similar messages LustreError: 5542:0:(file.c:5972:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000007:0x1:0x0] error: rc = -108 LustreError: 5542:0:(file.c:5972:ll_inode_revalidate_fini()) Skipped 928 previous similar messages Lustre: lustre-MDT0000-mdc-ffff8800b67e1800: Connection restored to (at 192.168.204.105@tcp) LustreError: 6801:0:(lcommon_cl.c:195:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000402:0x11a3:0x0]: rc = -5 LustreError: 6801:0:(lcommon_cl.c:195:cl_file_inode_init()) Skipped 290 previous similar messages LustreError: 6801:0:(llite_lib.c:3737:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 6801:0:(llite_lib.c:3737:ll_prep_inode()) Skipped 290 previous similar messages LustreError: 83:0:(statahead.c:836:ll_statahead_interpret_work()) lustre: failed to prep 6 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 LustreError: 83:0:(statahead.c:836:ll_statahead_interpret_work()) Skipped 3 previous similar messages Lustre: dir [0x200000402:0x2f82:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 91 previous similar messages 1[7780]: segfault at 8 ip 00007fa396a767e8 sp 00007ffcf3f52700 error 4 in ld-2.17.so[7fa396a6b000+22000] LustreError: 83:0:(statahead.c:836:ll_statahead_interpret_work()) lustre: failed to prep 11 [0x200000403:0x34ac:0x0] inode@0000000000000000: rc = -5 13[20912]: segfault at 8 ip 00007f41b74767e8 sp 00007ffc9b4459a0 error 4 in ld-2.17.so[7f41b746b000+22000] 2[22156]: segfault at 8 ip 00007f4898ba87e8 sp 00007ffd32bd8de0 error 4 in ld-2.17.so[7f4898b9d000+22000] LustreError: 22635:0:(lov_object.c:1341:lov_layout_change()) lustre-clilov-ffff88012e144000: cannot apply new layout on [0x240000404:0x32c:0x0] : rc = -5 LustreError: 22635:0:(lov_object.c:1341:lov_layout_change()) Skipped 1 previous similar message 0[27287]: segfault at 8 ip 00007f29b2a527e8 sp 00007ffdbcc26ce0 error 4 in ld-2.17.so[7f29b2a47000+22000] 18[27861]: segfault at 8 ip 00007f0b7d03f7e8 sp 00007ffce2c99a80 error 4 in ld-2.17.so[7f0b7d034000+22000] 4[542]: segfault at 8 ip 00007f7316ca87e8 sp 00007ffd3308fcb0 error 4 in ld-2.17.so[7f7316c9d000+22000] 14[6672]: segfault at 8 ip 00007f2d9411c7e8 sp 00007ffe2cd93330 error 4 in ld-2.17.so[7f2d94111000+22000] 0[7939]: segfault at 8 ip 00007fe0d5ff77e8 sp 00007ffe00f9a320 error 4 in ld-2.17.so[7fe0d5fec000+22000] 12[8517]: segfault at 8 ip 00007f31c03567e8 sp 00007fff90feb5b0 error 4 in ld-2.17.so[7f31c034b000+22000] LustreError: 7763:0:(llite_lib.c:1873:ll_update_lsm_md()) lustre: [0x200000404:0x1116:0x0] dir layout mismatch: LustreError: 7763:0:(llite_lib.c:1873:ll_update_lsm_md()) Skipped 1 previous similar message LustreError: 7763:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=2 count=1 index=0 hash=crush:0x2000003 max_inherit=0 max_inherit_rr=0 version=2 migrate_offset=0 migrate_hash=invalid:0 pool= LustreError: 7763:0:(lustre_lmv.h:167:lmv_stripe_object_dump()) stripe[0] [0x200000400:0xd0:0x0] LustreError: 7763:0:(lustre_lmv.h:167:lmv_stripe_object_dump()) Skipped 7 previous similar messages LustreError: 7763:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=1 count=2 index=0 hash=crush:0x82000003 max_inherit=0 max_inherit_rr=0 version=1 migrate_offset=1 migrate_hash=crush:2000003 pool= LustreError: 7380:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=1 count=1 index=0 hash=crush:0x2000003 max_inherit=0 max_inherit_rr=0 version=2 migrate_offset=0 migrate_hash=invalid:0 pool= LustreError: 7380:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: magic=0xcd20cd0 refs=1 count=2 index=0 hash=crush:0x82000003 max_inherit=0 max_inherit_rr=0 version=1 migrate_offset=1 migrate_hash=crush:2000003 pool= 7[21526]: segfault at 8 ip 00007feaccee67e8 sp 00007ffd4a127500 error 4 in ld-2.17.so[7feaccedb000+22000] 16[23383]: segfault at 8 ip 00007f34e556c7e8 sp 00007fff165f2150 error 4 in ld-2.17.so[7f34e5561000+22000] LustreError: 30025:0:(vvp_io.c:1921:vvp_io_init()) lustre: refresh file layout [0x200000403:0x4c73:0x0] error -5. LustreError: 30025:0:(vvp_io.c:1921:vvp_io_init()) Skipped 8 previous similar messages 8[31635]: segfault at 8 ip 00007f907763e7e8 sp 00007ffff4b09570 error 4 in ld-2.17.so[7f9077633000+22000] LustreError: 10234:0:(statahead.c:836:ll_statahead_interpret_work()) lustre: failed to prep 15 [0x0:0x0:0x0] inode@0000000000000000: rc = -5 7[14088]: segfault at 8 ip 00007f02484d77e8 sp 00007ffe51df7b80 error 4 in ld-2.17.so[7f02484cc000+22000] 5[14723]: segfault at 8 ip 00007f865512f7e8 sp 00007ffe13d723a0 error 4 in ld-2.17.so[7f8655124000+22000] | Link to test |
racer test 2: racer rename: centos-60.localnet DURATION=2700 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 PGD 2cb261067 PUD 2e4b35067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) lustre(OE) osp(OE) ofd(OE) lod(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) jbd2 mbcache lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) crc32_generic lnet(OE) dm_flakey dm_mod libcfs(OE) crc_t10dif crct10dif_generic crct10dif_common virtio_balloon virtio_console pcspkr i2c_piix4 ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm ata_piix drm_panel_orientation_quirks virtio_blk serio_raw i2c_core floppy libata CPU: 12 PID: 24003 Comm: ll_sa_23479 Kdump: loaded Tainted: P OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.16.0-3.module_el8.7.0+1218+f626c2ff 04/01/2014 task: ffff88006e8adc40 ti: ffff880069f60000 task.ti: ffff880069f60000 RIP: 0010:[<ffffffff813f502c>] [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 RSP: 0018:ffff880069f63d80 EFLAGS: 00010282 RAX: ffff88006c4a6c48 RBX: 0000000000000008 RCX: 000000010026001c RDX: 0000000000000026 RSI: ffff88006c4a6f78 RDI: 0000000000000008 RBP: ffff880069f63d90 R08: ffff8802bdd71738 R09: 0000000000000000 R10: ffff8802bdd73ef8 R11: ffff8802bdd71738 R12: 0000000000000000 R13: ffff88006e8ac9d0 R14: ffff88006c4a6f78 R15: ffff8802bdd71738 FS: 0000000000000000(0000) GS:ffff880331d00000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 000000027d0f4000 CR4: 00000000000007e0 Call Trace: [<ffffffffa10ce7b1>] ll_statahead_thread+0xde1/0x2290 [lustre] [<ffffffff810df6e2>] ? pick_next_task_fair+0x702/0x8a0 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e329e>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffff810c834d>] ? finish_task_switch+0x5d/0x1b0 [<ffffffffa10cd9d0>] ? ll_statahead_handle.constprop.29+0x1c0/0x1c0 [lustre] [<ffffffff810ba114>] kthread+0xe4/0xf0 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 | Lustre: 18050:0:(mdt_recovery.c:148:mdt_req_from_lrd()) @@@ restoring transno req@ffff8800a4a23240 x1818245412834176/t4295198408(0) o101->5c2412f7-8fbc-46dc-93f9-3a52836c262f@0@lo:477/0 lens 376/1904 e 0 to 0 dl 1734016527 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 Lustre: 18050:0:(mdt_recovery.c:148:mdt_req_from_lrd()) @@@ restoring transno req@ffff88028b1bd540 x1818245413280640/t4295115195(0) o101->5c2412f7-8fbc-46dc-93f9-3a52836c262f@0@lo:481/0 lens 376/19840 e 0 to 0 dl 1734016531 ref 1 fl Interpret:H/202/0 rc 0/0 job:'lfs.0' uid:0 gid:0 Lustre: 16754:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x20000040a:0x2:0x0] with magic=0xbd60bd0 Lustre: 16754:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 59 previous similar messages Lustre: 17772:0:(mdt_recovery.c:148:mdt_req_from_lrd()) @@@ restoring transno req@ffff8802b9ca0f40 x1818245416005504/t4295144033(0) o101->8cd39cb4-9077-46d4-b3fb-a96c96036464@0@lo:495/0 lens 376/48016 e 0 to 0 dl 1734016545 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 Lustre: 16754:0:(mdt_recovery.c:148:mdt_req_from_lrd()) @@@ restoring transno req@ffff8802947f6940 x1818245418375296/t4295199218(0) o101->5c2412f7-8fbc-46dc-93f9-3a52836c262f@0@lo:499/0 lens 376/37456 e 0 to 0 dl 1734016549 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 Lustre: 12842:0:(mdt_recovery.c:148:mdt_req_from_lrd()) @@@ restoring transno req@ffff8802d103f340 x1818245420351744/t4295144292(0) o101->5c2412f7-8fbc-46dc-93f9-3a52836c262f@0@lo:506/0 lens 376/48016 e 0 to 0 dl 1734016556 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 Lustre: 12842:0:(mdt_recovery.c:148:mdt_req_from_lrd()) Skipped 1 previous similar message Lustre: 28853:0:(mdt_recovery.c:148:mdt_req_from_lrd()) @@@ restoring transno req@ffff880259489e40 x1818245427179008/t4295144795(0) o101->5c2412f7-8fbc-46dc-93f9-3a52836c262f@0@lo:525/0 lens 376/48016 e 0 to 0 dl 1734016575 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 Lustre: 28853:0:(mdt_recovery.c:148:mdt_req_from_lrd()) Skipped 1 previous similar message Lustre: 12844:0:(mdt_recovery.c:148:mdt_req_from_lrd()) @@@ restoring transno req@ffff8800867a8f40 x1818245441615232/t4295200991(0) o101->8cd39cb4-9077-46d4-b3fb-a96c96036464@0@lo:566/0 lens 376/48016 e 0 to 0 dl 1734016616 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 Lustre: 12844:0:(mdt_recovery.c:148:mdt_req_from_lrd()) Skipped 1 previous similar message Lustre: 23636:0:(mdt_recovery.c:148:mdt_req_from_lrd()) @@@ restoring transno req@ffff8802c08c4140 x1818245451357440/t4295118319(0) o101->5c2412f7-8fbc-46dc-93f9-3a52836c262f@0@lo:608/0 lens 376/45112 e 0 to 0 dl 1734016658 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 Lustre: 23636:0:(mdt_recovery.c:148:mdt_req_from_lrd()) Skipped 4 previous similar messages LustreError: 10259:0:(ldlm_lockd.c:241:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0000_UUID lock: ffff88009d322980/0xd14284908cb92921 lrc: 3/0,0 mode: PW/PW res: [0x2c0000401:0x167b:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->524287) gid 0 flags: 0x60000400000020 nid: 0@lo remote: 0xd14284908cb926c0 expref: 5569 pid: 12630 timeout: 3031 lvb_type: 0 LustreError: 6176:0:(client.c:1300:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88009f6dc140 x1818245455845120/t0(0) o104->lustre-OST0000@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' uid:4294967295 gid:4294967295 LustreError: lustre-OST0000-osc-ffff8802e75e8958: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 6 previous similar messages Lustre: lustre-OST0000-osc-ffff8802e75e8958: Connection to lustre-OST0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: lustre-OST0000-osc-ffff8802e75e8958: This client was evicted by lustre-OST0000; in progress operations using this service will fail. LustreError: 10530:0:(ldlm_lockd.c:2575:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1734016535 with bad export cookie 15078760258734082917 LustreError: 4996:0:(client.c:1300:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff880294fec140 x1818245456003712/t0(0) o104->lustre-OST0000@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' uid:4294967295 gid:4294967295 LustreError: 4996:0:(client.c:1300:ptlrpc_import_delay_req()) Skipped 9 previous similar messages Lustre: 3854:0:(llite_lib.c:4127:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.62@tcp:/lustre/fid: [0x280000408:0x16e:0x0]// may get corrupted (rc -108) LustreError: 14285:0:(client.c:1300:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802c6d24640 x1818245456633088/t0(0) o104->lustre-OST0000@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' uid:4294967295 gid:4294967295 LustreError: 14285:0:(client.c:1300:ptlrpc_import_delay_req()) Skipped 11 previous similar messages LustreError: 10509:0:(ldlm_resource.c:1149:ldlm_resource_complain()) lustre-OST0000-osc-ffff8802e75e8958: namespace resource [0x2c0000401:0x1cbf:0x0].0x0 (ffff88024932eac0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 10509:0:(ldlm_resource.c:1149:ldlm_resource_complain()) Skipped 17 previous similar messages LustreError: 10509:0:(ldlm_resource.c:1149:ldlm_resource_complain()) lustre-OST0000-osc-ffff8802e75e8958: namespace resource [0x2c0000400:0x43b2:0x0].0x0 (ffff880295d6e340) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 10509:0:(ldlm_resource.c:1149:ldlm_resource_complain()) Skipped 17 previous similar messages Lustre: lustre-OST0000-osc-ffff8802e75e8958: Connection restored to 192.168.123.62@tcp (at 0@lo) Lustre: 23260:0:(mdt_recovery.c:148:mdt_req_from_lrd()) @@@ restoring transno req@ffff8802bf028040 x1818245480729344/t4295121192(0) o101->8cd39cb4-9077-46d4-b3fb-a96c96036464@0@lo:703/0 lens 376/45232 e 0 to 0 dl 1734016753 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 Lustre: 23260:0:(mdt_recovery.c:148:mdt_req_from_lrd()) Skipped 4 previous similar messages Lustre: 17194:0:(osd_handler.c:1961:osd_trans_dump_creds()) create: 6/24/0, destroy: 1/4/0 Lustre: 17194:0:(osd_handler.c:1961:osd_trans_dump_creds()) Skipped 131703 previous similar messages Lustre: 17194:0:(osd_handler.c:1968:osd_trans_dump_creds()) attr_set: 2241/2241/0, xattr_set: 3361/31480/0 Lustre: 17194:0:(osd_handler.c:1968:osd_trans_dump_creds()) Skipped 131706 previous similar messages Lustre: 17194:0:(osd_handler.c:1978:osd_trans_dump_creds()) write: 28/157/0, punch: 0/0/0, quota 1/3/0 Lustre: 17194:0:(osd_handler.c:1978:osd_trans_dump_creds()) Skipped 131706 previous similar messages Lustre: 17194:0:(osd_handler.c:1985:osd_trans_dump_creds()) insert: 7/118/0, delete: 2/5/1 Lustre: 17194:0:(osd_handler.c:1985:osd_trans_dump_creds()) Skipped 131705 previous similar messages Lustre: 17194:0:(osd_handler.c:1992:osd_trans_dump_creds()) ref_add: 1/1/0, ref_del: 2/2/1 Lustre: 17194:0:(osd_handler.c:1992:osd_trans_dump_creds()) Skipped 131703 previous similar messages Lustre: 11545:0:(client.c:2364:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1734016558/real 1734016558] req@ffff8802bf6ead40 x1818245460719616/t0(0) o101->lustre-MDT0000-mdc-ffff8802e75e8958@0@lo:12/10 lens 576/35944 e 0 to 1 dl 1734016694 ref 2 fl Rpc:XQr/202/ffffffff rc 0/-1 job:'mrename.0' uid:0 gid:0 Lustre: lustre-MDT0000-mdc-ffff8802e75e8958: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: lustre-MDT0000: Client 8cd39cb4-9077-46d4-b3fb-a96c96036464 (at 0@lo) reconnecting Lustre: lustre-MDT0000-mdc-ffff8802e75e8958: Connection restored to 192.168.123.62@tcp (at 0@lo) LustreError: 10259:0:(ldlm_lockd.c:241:expired_lock_main()) ### lock callback timer expired after 122s: evicting client at 0@lo ns: filter-lustre-OST0000_UUID lock: ffff8802d120c780/0xd14284908cd504bc lrc: 3/0,0 mode: PW/PW res: [0x2c0000400:0x4005:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000480030020 nid: 0@lo remote: 0xd14284908cd504b5 expref: 3764 pid: 29727 timeout: 3169 lvb_type: 0 LustreError: 18751:0:(client.c:1300:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88007866ee40 x1818245509944192/t0(0) o105->lustre-OST0000@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' uid:4294967295 gid:4294967295 LustreError: 18751:0:(client.c:1300:ptlrpc_import_delay_req()) Skipped 11 previous similar messages LustreError: lustre-OST0000-osc-ffff8800ba80ca88: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: lustre-OST0000-osc-ffff8800ba80ca88: This client was evicted by lustre-OST0000; in progress operations using this service will fail. Lustre: 3863:0:(llite_lib.c:4127:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.62@tcp:/lustre/fid: [0x280000406:0x2d75:0x0]// may get corrupted (rc -108) Lustre: 3860:0:(llite_lib.c:4127:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.62@tcp:/lustre/fid: [0x280000408:0x4ae:0x0]/ may get corrupted (rc -108) Lustre: 3860:0:(llite_lib.c:4127:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.62@tcp:/lustre/fid: [0x240000405:0x3e09:0x0]// may get corrupted (rc -108) LustreError: 19302:0:(ldlm_resource.c:1149:ldlm_resource_complain()) lustre-OST0000-osc-ffff8800ba80ca88: namespace resource [0x2c0000401:0x403f:0x0].0x0 (ffff8802a6bba840) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 19302:0:(ldlm_resource.c:1149:ldlm_resource_complain()) Skipped 2 previous similar messages Lustre: 18056:0:(osd_internal.h:1330:osd_trans_exec_op()) lustre-MDT0000: opcode 0: before 514 < left 1773, rollback = 0 Lustre: 18056:0:(osd_internal.h:1330:osd_trans_exec_op()) Skipped 175992 previous similar messages Lustre: lustre-OST0000: haven't heard from client 5c2412f7-8fbc-46dc-93f9-3a52836c262f (at 0@lo) in 31 seconds. I think it's dead, and I am evicting it. exp ffff8800b41bb7e8, cur 1734016727 expire 1734016697 last 1734016696 LustreError: lustre-OST0000-osc-ffff8800ba80ca88: This client was evicted by lustre-OST0000; in progress operations using this service will fail. Lustre: 17772:0:(mdt_recovery.c:148:mdt_req_from_lrd()) @@@ restoring transno req@ffff880071af9940 x1818245524660864/t4295204994(0) o101->8cd39cb4-9077-46d4-b3fb-a96c96036464@0@lo:154/0 lens 376/48016 e 0 to 0 dl 1734016959 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 Lustre: 17772:0:(mdt_recovery.c:148:mdt_req_from_lrd()) Skipped 9 previous similar messages Lustre: lustre-OST0002-osc-MDT0002: update sequence from 0x340000400 to 0x340000403 | Link to test |
racer test 1: racer on clients: centos-5.localnet DURATION=2700 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 PGD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) osp(OE) ofd(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) dm_flakey dm_mod libcfs(OE) loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) jbd2 mbcache crc32_generic crc_t10dif crct10dif_generic crct10dif_common i2c_piix4 virtio_balloon virtio_console pcspkr ip_tables rpcsec_gss_krb5 drm_kms_helper ttm ata_generic pata_acpi drm ata_piix drm_panel_orientation_quirks floppy virtio_blk serio_raw libata i2c_core [last unloaded: libcfs] CPU: 4 PID: 16119 Comm: ll_sa_16050 Kdump: loaded Tainted: P OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.16.0-3.module_el8.7.0+1218+f626c2ff 04/01/2014 task: ffff8802d50a5c40 ti: ffff8800acabc000 task.ti: ffff8800acabc000 RIP: 0010:[<ffffffff813f502c>] [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 RSP: 0018:ffff8800acabfd80 EFLAGS: 00010282 RAX: ffff8802f88d31c8 RBX: 0000000000000008 RCX: 0000000100260022 RDX: 0000000000000026 RSI: ffff8802f88d34f8 RDI: 0000000000000008 RBP: ffff8800acabfd90 R08: ffff88032757c248 R09: 0000000000000000 R10: ffff88032757bef8 R11: ffff88032757c248 R12: 0000000000000000 R13: ffff8802d50a49d0 R14: ffff8802f88d34f8 R15: ffff88032757c248 FS: 0000000000000000(0000) GS:ffff880331b00000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 0000000001c10000 CR4: 00000000000007e0 Call Trace: [<ffffffffa172ba61>] ll_statahead_thread+0xde1/0x2290 [lustre] [<ffffffff810dbb51>] ? put_prev_entity+0x31/0x400 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e329e>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffff810c834d>] ? finish_task_switch+0x5d/0x1b0 [<ffffffffa172ac80>] ? ll_statahead_handle.constprop.29+0x1c0/0x1c0 [lustre] [<ffffffff810ba114>] kthread+0xe4/0xf0 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 | Lustre: 9733:0:(osd_internal.h:1314:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 516 < left 618, rollback = 7 Lustre: 9733:0:(osd_handler.c:1955:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 9733:0:(osd_handler.c:1962:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 9733:0:(osd_handler.c:1972:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 9733:0:(osd_handler.c:1979:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 9733:0:(osd_handler.c:1986:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 7831:0:(osd_internal.h:1314:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 516 < left 618, rollback = 7 Lustre: 7831:0:(osd_internal.h:1314:osd_trans_exec_op()) Skipped 1 previous similar message Lustre: 7831:0:(osd_handler.c:1955:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 7831:0:(osd_handler.c:1955:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 7831:0:(osd_handler.c:1962:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 7831:0:(osd_handler.c:1962:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 7831:0:(osd_handler.c:1972:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 7831:0:(osd_handler.c:1972:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 7831:0:(osd_handler.c:1979:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 7831:0:(osd_handler.c:1979:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 7831:0:(osd_handler.c:1986:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 7831:0:(osd_handler.c:1986:osd_trans_dump_creds()) Skipped 1 previous similar message 11[10031]: segfault at 406000 ip 0000000000406000 sp 00007fff84bcf448 error 14 in 11[606000+1000] Lustre: 7833:0:(osd_internal.h:1314:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 516 < left 618, rollback = 7 Lustre: 7833:0:(osd_internal.h:1314:osd_trans_exec_op()) Skipped 1 previous similar message Lustre: 7833:0:(osd_handler.c:1955:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 7833:0:(osd_handler.c:1955:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 7833:0:(osd_handler.c:1962:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 7833:0:(osd_handler.c:1962:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 7833:0:(osd_handler.c:1972:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 7833:0:(osd_handler.c:1972:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 7833:0:(osd_handler.c:1979:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 7833:0:(osd_handler.c:1979:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 7833:0:(osd_handler.c:1986:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 7833:0:(osd_handler.c:1986:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 9692:0:(mdt_recovery.c:148:mdt_req_from_lrd()) @@@ restoring transno req@ffff8802c9d91440 x1817407971125120/t4294969276(0) o101->6695c4fa-811f-4d29-abad-42976bf4a389@0@lo:455/0 lens 376/840 e 0 to 0 dl 1733215450 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 Lustre: 7831:0:(osd_internal.h:1314:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 516 < left 618, rollback = 7 Lustre: 7831:0:(osd_internal.h:1314:osd_trans_exec_op()) Skipped 1 previous similar message Lustre: 7831:0:(osd_handler.c:1955:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 7831:0:(osd_handler.c:1955:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 7831:0:(osd_handler.c:1962:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 7831:0:(osd_handler.c:1962:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 7831:0:(osd_handler.c:1972:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 7831:0:(osd_handler.c:1972:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 7831:0:(osd_handler.c:1979:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 7831:0:(osd_handler.c:1979:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 7831:0:(osd_handler.c:1986:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 7831:0:(osd_handler.c:1986:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 7833:0:(osd_internal.h:1314:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 516 < left 618, rollback = 7 Lustre: 7833:0:(osd_internal.h:1314:osd_trans_exec_op()) Skipped 1 previous similar message Lustre: 7833:0:(osd_handler.c:1955:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 7833:0:(osd_handler.c:1955:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 7833:0:(osd_handler.c:1962:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 7833:0:(osd_handler.c:1962:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 7833:0:(osd_handler.c:1972:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 7833:0:(osd_handler.c:1972:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 7833:0:(osd_handler.c:1979:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 7833:0:(osd_handler.c:1979:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 7833:0:(osd_handler.c:1986:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 7833:0:(osd_handler.c:1986:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 12510:0:(osd_internal.h:1314:osd_trans_exec_op()) lustre-OST0000: opcode 7: before 516 < left 618, rollback = 7 Lustre: 12510:0:(osd_internal.h:1314:osd_trans_exec_op()) Skipped 3 previous similar messages Lustre: 12510:0:(osd_handler.c:1955:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 12510:0:(osd_handler.c:1955:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 12510:0:(osd_handler.c:1962:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 12510:0:(osd_handler.c:1962:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 12510:0:(osd_handler.c:1972:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 12510:0:(osd_handler.c:1972:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 12510:0:(osd_handler.c:1979:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 12510:0:(osd_handler.c:1979:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 12510:0:(osd_handler.c:1986:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 12510:0:(osd_handler.c:1986:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 7832:0:(osd_internal.h:1314:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 618, rollback = 7 Lustre: 7832:0:(osd_internal.h:1314:osd_trans_exec_op()) Skipped 1 previous similar message Lustre: 7832:0:(osd_handler.c:1955:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 7832:0:(osd_handler.c:1955:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 7832:0:(osd_handler.c:1962:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 7832:0:(osd_handler.c:1962:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 7832:0:(osd_handler.c:1972:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 7832:0:(osd_handler.c:1972:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 7832:0:(osd_handler.c:1979:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 7832:0:(osd_handler.c:1979:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 7832:0:(osd_handler.c:1986:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 7832:0:(osd_handler.c:1986:osd_trans_dump_creds()) Skipped 1 previous similar message 18[17706]: segfault at 8 ip 00007f8eb6c497e8 sp 00007ffea7c4d2c0 error 4 in ld-2.17.so[7f8eb6c3e000+22000] Lustre: 9733:0:(osd_internal.h:1314:osd_trans_exec_op()) lustre-OST0000: opcode 7: before 516 < left 618, rollback = 7 Lustre: 9733:0:(osd_internal.h:1314:osd_trans_exec_op()) Skipped 7 previous similar messages Lustre: 9733:0:(osd_handler.c:1955:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 9733:0:(osd_handler.c:1955:osd_trans_dump_creds()) Skipped 7 previous similar messages Lustre: 9733:0:(osd_handler.c:1962:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 9733:0:(osd_handler.c:1962:osd_trans_dump_creds()) Skipped 7 previous similar messages Lustre: 9733:0:(osd_handler.c:1972:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 9733:0:(osd_handler.c:1972:osd_trans_dump_creds()) Skipped 7 previous similar messages Lustre: 9733:0:(osd_handler.c:1979:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 9733:0:(osd_handler.c:1979:osd_trans_dump_creds()) Skipped 7 previous similar messages Lustre: 9733:0:(osd_handler.c:1986:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 9733:0:(osd_handler.c:1986:osd_trans_dump_creds()) Skipped 7 previous similar messages LustreError: 19817:0:(file.c:262:ll_close_inode_openhandle()) lustre-clilmv-ffff8802c9479bf8: inode [0x200000401:0x32d:0x0] mdc close failed: rc = -13 Lustre: 13477:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000402:0x34e:0x0] with magic=0xbd60bd0 5[21818]: segfault at 8 ip 00007f4f3acf07e8 sp 00007ffccb1f2f70 error 4 in ld-2.17.so[7f4f3ace5000+22000] 12[23184]: segfault at 8 ip 00007f9a13d337e8 sp 00007ffc53988ea0 error 4 in ld-2.17.so[7f9a13d28000+22000] Lustre: 12510:0:(osd_internal.h:1314:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 516 < left 618, rollback = 7 Lustre: 12510:0:(osd_internal.h:1314:osd_trans_exec_op()) Skipped 9 previous similar messages Lustre: 12510:0:(osd_handler.c:1955:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 12510:0:(osd_handler.c:1955:osd_trans_dump_creds()) Skipped 9 previous similar messages Lustre: 12510:0:(osd_handler.c:1962:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 12510:0:(osd_handler.c:1962:osd_trans_dump_creds()) Skipped 9 previous similar messages Lustre: 12510:0:(osd_handler.c:1972:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 12510:0:(osd_handler.c:1972:osd_trans_dump_creds()) Skipped 9 previous similar messages Lustre: 12510:0:(osd_handler.c:1979:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 12510:0:(osd_handler.c:1979:osd_trans_dump_creds()) Skipped 9 previous similar messages Lustre: 12510:0:(osd_handler.c:1986:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 12510:0:(osd_handler.c:1986:osd_trans_dump_creds()) Skipped 9 previous similar messages LustreError: 25218:0:(file.c:262:ll_close_inode_openhandle()) lustre-clilmv-ffff8802c9479bf8: inode [0x200000402:0x44e:0x0] mdc close failed: rc = -13 8[26183]: segfault at 8 ip 00007fd6e36487e8 sp 00007ffdc636d000 error 4 in ld-2.17.so[7fd6e363d000+22000] 8[26126]: segfault at 8 ip 00007fce4a3de7e8 sp 00007ffd0257a130 error 4 in ld-2.17.so[7fce4a3d3000+22000] 11[26489]: segfault at 8 ip 00007f0219b727e8 sp 00007fffa40cf270 error 4 in ld-2.17.so[7f0219b67000+22000] 6[31197]: segfault at 8 ip 00007f61c1ce97e8 sp 00007ffead9cb090 error 4 in ld-2.17.so[7f61c1cde000+22000] Lustre: 7833:0:(osd_internal.h:1314:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 618, rollback = 7 Lustre: 7833:0:(osd_internal.h:1314:osd_trans_exec_op()) Skipped 25 previous similar messages Lustre: 7833:0:(osd_handler.c:1955:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 7833:0:(osd_handler.c:1955:osd_trans_dump_creds()) Skipped 25 previous similar messages Lustre: 7833:0:(osd_handler.c:1962:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 7833:0:(osd_handler.c:1962:osd_trans_dump_creds()) Skipped 25 previous similar messages Lustre: 7833:0:(osd_handler.c:1972:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 7833:0:(osd_handler.c:1972:osd_trans_dump_creds()) Skipped 25 previous similar messages Lustre: 7833:0:(osd_handler.c:1979:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 7833:0:(osd_handler.c:1979:osd_trans_dump_creds()) Skipped 25 previous similar messages Lustre: 7833:0:(osd_handler.c:1986:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 7833:0:(osd_handler.c:1986:osd_trans_dump_creds()) Skipped 25 previous similar messages 14[3334]: segfault at 0 ip 0000000000403cf0 sp 00007ffd5ddcf898 error 6 in 14[400000+6000] 19[3406]: segfault at 8 ip 00007fa6bcaab7e8 sp 00007ffce2883e80 error 4 in ld-2.17.so[7fa6bcaa0000+22000] 1[6649]: segfault at 8 ip 00007ff60c9e87e8 sp 00007ffc70495850 error 4 1[6424]: segfault at 8 ip 00007fdc934277e8 sp 00007ffc4287a3d0 error 4 in ld-2.17.so[7fdc9341c000+22000] in ld-2.17.so[7ff60c9dd000+22000] 11[7854]: segfault at 8 ip 00007fa6f91267e8 sp 00007ffe4f095b10 error 4 in ld-2.17.so[7fa6f911b000+22000] Lustre: 7575:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000401:0x8a0:0x0] with magic=0xbd60bd0 Lustre: 7575:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message 4[9262]: segfault at 8 ip 00007f5adeb997e8 sp 00007ffd02f12860 error 4 in ld-2.17.so[7f5adeb8e000+22000] LustreError: lustre-MDT0000-mdc-ffff8802e512ca88: operation ost_read to node 0@lo failed: rc = -116 | Link to test |
racer test 1: racer on clients: centos-110.localnet DURATION=2700 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 PGD 2972b1067 PUD 2972b2067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) osp(OE) ofd(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) crc32_generic crc_t10dif crct10dif_generic crct10dif_common i2c_piix4 pcspkr virtio_balloon virtio_console ip_tables rpcsec_gss_krb5 drm_kms_helper ata_generic ttm pata_acpi drm drm_panel_orientation_quirks ata_piix floppy virtio_blk serio_raw i2c_core libata [last unloaded: libcfs] CPU: 0 PID: 11958 Comm: ll_sa_11882 Kdump: loaded Tainted: P OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.16.0-3.module_el8.7.0+1218+f626c2ff 04/01/2014 task: ffff88009c20a4f0 ti: ffff8802a2bc8000 task.ti: ffff8802a2bc8000 RIP: 0010:[<ffffffff813f502c>] [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 RSP: 0018:ffff8802a2bcbd80 EFLAGS: 00010282 RAX: ffff8802713ef608 RBX: 0000000000000008 RCX: 0000000100260023 RDX: 0000000000000026 RSI: ffff8802713ef938 RDI: 0000000000000008 RBP: ffff8802a2bcbd90 R08: ffff8800a71cfa98 R09: 0000000000000000 R10: ffff8800a71c9a88 R11: ffff8800a71cfa98 R12: 0000000000000000 R13: ffff88009c209280 R14: ffff8802713ef938 R15: ffff8800a71cfa98 FS: 0000000000000000(0000) GS:ffff880331a00000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 00000002ede7c000 CR4: 00000000000007f0 Call Trace: [<ffffffffa156ce21>] ll_statahead_thread+0xde1/0x2290 [lustre] [<ffffffff810dbb51>] ? put_prev_entity+0x31/0x400 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e329e>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffff810c834d>] ? finish_task_switch+0x5d/0x1b0 [<ffffffffa156c040>] ? ll_statahead_handle.constprop.29+0x1c0/0x1c0 [lustre] [<ffffffff810ba114>] kthread+0xe4/0xf0 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 | Lustre: 14922:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000402:0x20:0x0] with magic=0xbd60bd0 LustreError: 14750:0:(mdt_handler.c:776:mdt_pack_acl2body()) lustre-MDT0000: unable to read [0x200000402:0x1c:0x0] ACL: rc = -2 16[20780]: segfault at 8 ip 00007f24163157e8 sp 00007ffdbe12f0e0 error 4 in ld-2.17.so[7f241630a000+22000] 12[22500]: segfault at 8 ip 00007fb74d31f7e8 sp 00007fff1c942340 error 4 in ld-2.17.so[7fb74d314000+22000] Lustre: 14921:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000401:0xc7:0x0] with magic=0xbd60bd0 Lustre: 14921:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message Lustre: 9018:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000401:0x165:0x0] with magic=0xbd60bd0 Lustre: 9018:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message Lustre: 14884:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000401:0x1ce:0x0] with magic=0xbd60bd0 Lustre: 14884:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message Lustre: 14921:0:(mdt_recovery.c:148:mdt_req_from_lrd()) @@@ restoring transno req@ffff8802a4c36440 x1817163548858880/t4294977692(0) o101->a2ccc6b4-1383-4b7f-bea8-13e9f8d8fb27@0@lo:56/0 lens 376/840 e 0 to 0 dl 1732982511 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 Lustre: 14750:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000402:0x250:0x0] with magic=0xbd60bd0 Lustre: 14750:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message 1[1833]: segfault at 8 ip 00007fc9e8e8a7e8 sp 00007ffecc261560 error 4 in ld-2.17.so[7fc9e8e7f000+22000] LustreError: 18321:0:(mdt_handler.c:776:mdt_pack_acl2body()) lustre-MDT0000: unable to read [0x200000401:0x2a4:0x0] ACL: rc = -2 3[12282]: segfault at 8 ip 00007f160263e7e8 sp 00007ffe6f851bb0 error 4 in ld-2.17.so[7f1602633000+22000] 11[14814]: segfault at 8 ip 00007f7ed558c7e8 sp 00007fff79cdbde0 error 4 in ld-2.17.so[7f7ed5581000+22000] LustreError: 14814:0:(file.c:262:ll_close_inode_openhandle()) lustre-clilmv-ffff8802725892a8: inode [0x200000401:0x397:0x0] mdc close failed: rc = -13 1[15744]: segfault at 8 ip 00007fad78dc27e8 sp 00007ffffd1f49f0 error 4 in ld-2.17.so[7fad78db7000+22000] 13[15906]: segfault at 8 ip 00007fcb88c0a7e8 sp 00007ffce46896d0 error 4 in ld-2.17.so[7fcb88bff000+22000] LustreError: 17807:0:(file.c:262:ll_close_inode_openhandle()) lustre-clilmv-ffff8802725892a8: inode [0x200000401:0x402:0x0] mdc close failed: rc = -13 LustreError: 14932:0:(mdt_handler.c:776:mdt_pack_acl2body()) lustre-MDT0000: unable to read [0x200000402:0x3ac:0x0] ACL: rc = -2 LustreError: 24255:0:(file.c:262:ll_close_inode_openhandle()) lustre-clilmv-ffff8802c4a8ca88: inode [0x200000401:0x4c8:0x0] mdc close failed: rc = -13 9[31482]: segfault at 0 ip (null) sp 00007ffc65dc1f98 error 14 in 9[400000+6000] 17[32016]: segfault at 8 ip 00007f7d6b9c27e8 sp 00007ffd7fd7e550 error 4 in ld-2.17.so[7f7d6b9b7000+22000] 17[32027]: segfault at 8 ip 00007f2637fdf7e8 sp 00007ffd073e7030 error 4 in ld-2.17.so[7f2637fd4000+22000] LustreError: 14925:0:(mdt_handler.c:776:mdt_pack_acl2body()) lustre-MDT0000: unable to read [0x200000401:0xc0b:0x0] ACL: rc = -2 Lustre: 9015:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000401:0xd22:0x0] with magic=0xbd60bd0 Lustre: 9015:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message LustreError: 14884:0:(mdt_handler.c:776:mdt_pack_acl2body()) lustre-MDT0000: unable to read [0x200000401:0xe0b:0x0] ACL: rc = -2 9[8370]: segfault at 8 ip 00007f08a3ac87e8 sp 00007ffdabb49780 error 4 in ld-2.17.so[7f08a3abd000+22000] LustreError: 20748:0:(file.c:262:ll_close_inode_openhandle()) lustre-clilmv-ffff8802725892a8: inode [0x200000402:0xed0:0x0] mdc close failed: rc = -13 LustreError: 14884:0:(mdt_handler.c:776:mdt_pack_acl2body()) lustre-MDT0000: unable to read [0x200000401:0x102e:0x0] ACL: rc = -2 LustreError: 29402:0:(file.c:262:ll_close_inode_openhandle()) lustre-clilmv-ffff8802725892a8: inode [0x200000402:0xfd6:0x0] mdc close failed: rc = -13 LustreError: 31855:0:(file.c:262:ll_close_inode_openhandle()) lustre-clilmv-ffff8802c4a8ca88: inode [0x200000402:0xfd6:0x0] mdc close failed: rc = -13 LustreError: 31855:0:(file.c:262:ll_close_inode_openhandle()) Skipped 1 previous similar message LustreError: 32568:0:(file.c:262:ll_close_inode_openhandle()) lustre-clilmv-ffff8802725892a8: inode [0x200000402:0xfd6:0x0] mdc close failed: rc = -13 Lustre: 27817:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000401:0x11d3:0x0] with magic=0xbd60bd0 Lustre: 27817:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message 17[15351]: segfault at 8 ip 00007f97aaa9d7e8 sp 00007ffd952cfde0 error 4 in ld-2.17.so[7f97aaa92000+22000] Lustre: 9015:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000401:0x127f:0x0] with magic=0xbd60bd0 Lustre: 9015:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message Lustre: 9015:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000401:0x12b1:0x0] with magic=0xbd60bd0 Lustre: 9015:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message Lustre: 18336:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000401:0x1440:0x0] with magic=0xbd60bd0 Lustre: 18336:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message Lustre: 32126:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000401:0x151d:0x0] with magic=0xbd60bd0 Lustre: 32126:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message LustreError: 9015:0:(mdt_handler.c:776:mdt_pack_acl2body()) lustre-MDT0000: unable to read [0x200000401:0x1609:0x0] ACL: rc = -2 Lustre: 14593:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000402:0x1670:0x0] with magic=0xbd60bd0 Lustre: 14593:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message 19[31861]: segfault at 8 ip 00007f69dcf397e8 sp 00007ffdf49c3240 error 4 in ld-2.17.so[7f69dcf2e000+22000] LustreError: 2718:0:(file.c:262:ll_close_inode_openhandle()) lustre-clilmv-ffff8802725892a8: inode [0x200000402:0x179c:0x0] mdc close failed: rc = -13 LustreError: 6890:0:(file.c:262:ll_close_inode_openhandle()) lustre-clilmv-ffff8802c4a8ca88: inode [0x200000401:0x1946:0x0] mdc close failed: rc = -13 LustreError: 7580:0:(file.c:262:ll_close_inode_openhandle()) lustre-clilmv-ffff8802c4a8ca88: inode [0x200000401:0x1946:0x0] mdc close failed: rc = -13 | Link to test |
racer test 2: racer rename: centos-35.localnet DURATION=2700 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 PGD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) osp(OE) ofd(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) crc32_generic crc_t10dif crct10dif_generic crct10dif_common virtio_console i2c_piix4 pcspkr virtio_balloon ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm ata_piix drm_panel_orientation_quirks virtio_blk serio_raw libata i2c_core floppy [last unloaded: libcfs] CPU: 9 PID: 21584 Comm: ll_sa_20623 Kdump: loaded Tainted: P OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.16.0-3.module_el8.7.0+1218+f626c2ff 04/01/2014 task: ffff88026d199280 ti: ffff8802c5ea0000 task.ti: ffff8802c5ea0000 RIP: 0010:[<ffffffff813f502c>] [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 RSP: 0018:ffff8802c5ea3d80 EFLAGS: 00010282 RAX: ffff8801c9190ac8 RBX: 0000000000000008 RCX: 0000000100260022 RDX: 0000000000000026 RSI: ffff8801c9190df8 RDI: 0000000000000008 RBP: ffff8802c5ea3d90 R08: ffff8800468586a8 R09: 0000000000000000 R10: ffff880046858008 R11: ffff8800468586a8 R12: ffff8801c9190df8 R13: 0000000000000000 R14: ffff8800468586a8 R15: ffff8801c9190df8 FS: 0000000000000000(0000) GS:ffff880331c40000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 000000031fdc8000 CR4: 00000000000007e0 Call Trace: [<ffffffffa1552ad1>] ll_statahead_thread+0x741/0x2200 [lustre] [<ffffffff817eb9d0>] ? kprobe_flush_task+0xd0/0x170 [<ffffffffa1552390>] ? ll_statahead_handle.constprop.28+0x1c0/0x1c0 [lustre] [<ffffffff810ba114>] kthread+0xe4/0xf0 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 | Lustre: 620:0:(mdt_recovery.c:148:mdt_req_from_lrd()) @@@ restoring transno req@ffff880128cce940 x1813668073830016/t4295096108(0) o101->fadea5df-3cf5-4919-8c06-f8779c894561@0@lo:653/0 lens 376/7208 e 0 to 0 dl 1729651293 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 Lustre: 1889:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000407:0x2588:0x0] with magic=0xbd60bd0 Lustre: 1889:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 47 previous similar messages Lustre: 1086:0:(mdt_recovery.c:148:mdt_req_from_lrd()) @@@ restoring transno req@ffff8801c7ad6440 x1813668073845760/t4295127766(0) o101->46e6ab48-fc78-48fd-b04c-e29831ff7110@0@lo:654/0 lens 376/15304 e 0 to 0 dl 1729651294 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 Lustre: 1086:0:(mdt_recovery.c:148:mdt_req_from_lrd()) Skipped 1 previous similar message | Link to test |
racer test 1: racer on clients: centos-55.localnet DURATION=2700 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 PGD 2e439b067 PUD 3229cd067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) osp(OE) ofd(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) dm_flakey dm_mod libcfs(OE) loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) jbd2 mbcache crc32_generic crc_t10dif crct10dif_generic crct10dif_common i2c_piix4 virtio_balloon pcspkr virtio_console ip_tables rpcsec_gss_krb5 ata_generic drm_kms_helper pata_acpi ttm drm drm_panel_orientation_quirks ata_piix virtio_blk serio_raw i2c_core floppy libata [last unloaded: libcfs] CPU: 9 PID: 18276 Comm: ll_sa_18181 Kdump: loaded Tainted: P OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.16.0-3.module_el8.7.0+1218+f626c2ff 04/01/2014 task: ffff8802a0a11280 ti: ffff880321070000 task.ti: ffff880321070000 RIP: 0010:[<ffffffff813f502c>] [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 RSP: 0018:ffff880321073d80 EFLAGS: 00010282 RAX: ffff880321b56c48 RBX: 0000000000000008 RCX: 0000000100260012 RDX: 0000000000000026 RSI: ffff880321b56f78 RDI: 0000000000000008 RBP: ffff880321073d90 R08: ffff8803210386a8 R09: 0000000000000000 R10: ffff88032103e018 R11: ffff8803210386a8 R12: ffff880321b56f78 R13: 0000000000000000 R14: ffff8803210386a8 R15: ffff880321b56f78 FS: 0000000000000000(0000) GS:ffff880331c40000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 00000002ed75a000 CR4: 00000000000007e0 Call Trace: [<ffffffffa172cad1>] ll_statahead_thread+0x741/0x2200 [lustre] [<ffffffff817e320e>] ? _raw_spin_lock+0x1e/0x20 [<ffffffff810dbb51>] ? put_prev_entity+0x31/0x400 [<ffffffffa172c390>] ? ll_statahead_handle.constprop.28+0x1c0/0x1c0 [lustre] [<ffffffff810ba114>] kthread+0xe4/0xf0 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 | Lustre: 16981:0:(mdt_recovery.c:148:mdt_req_from_lrd()) @@@ restoring transno req@ffff8802ea46e440 x1813581529872512/t4294968266(0) o101->1a1f3f26-70eb-4436-8bf8-4930be24ef34@0@lo:101/0 lens 376/864 e 0 to 0 dl 1729566181 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 Lustre: 14639:0:(osd_internal.h:1314:osd_trans_exec_op()) lustre-OST0002: opcode 7: before 516 < left 618, rollback = 7 Lustre: 14639:0:(osd_handler.c:1955:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 14639:0:(osd_handler.c:1962:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 14639:0:(osd_handler.c:1972:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 14639:0:(osd_handler.c:1979:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 14639:0:(osd_handler.c:1986:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 14640:0:(osd_internal.h:1314:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 618, rollback = 7 Lustre: 14640:0:(osd_internal.h:1314:osd_trans_exec_op()) Skipped 1 previous similar message Lustre: 14640:0:(osd_handler.c:1955:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 14640:0:(osd_handler.c:1955:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 14640:0:(osd_handler.c:1962:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 14640:0:(osd_handler.c:1962:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 14640:0:(osd_handler.c:1972:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 14640:0:(osd_handler.c:1972:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 14640:0:(osd_handler.c:1979:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 14640:0:(osd_handler.c:1979:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 14640:0:(osd_handler.c:1986:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 14640:0:(osd_handler.c:1986:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 14637:0:(osd_internal.h:1314:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 516 < left 618, rollback = 7 Lustre: 14637:0:(osd_internal.h:1314:osd_trans_exec_op()) Skipped 1 previous similar message Lustre: 14637:0:(osd_handler.c:1955:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 14637:0:(osd_handler.c:1955:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 14637:0:(osd_handler.c:1962:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 14637:0:(osd_handler.c:1962:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 14637:0:(osd_handler.c:1972:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 14637:0:(osd_handler.c:1972:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 14637:0:(osd_handler.c:1979:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 14637:0:(osd_handler.c:1979:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 14637:0:(osd_handler.c:1986:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 14637:0:(osd_handler.c:1986:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 14640:0:(osd_internal.h:1314:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 618, rollback = 7 Lustre: 14640:0:(osd_internal.h:1314:osd_trans_exec_op()) Skipped 1 previous similar message Lustre: 14640:0:(osd_handler.c:1955:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 14640:0:(osd_handler.c:1955:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 14640:0:(osd_handler.c:1962:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 14640:0:(osd_handler.c:1962:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 14640:0:(osd_handler.c:1972:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 14640:0:(osd_handler.c:1972:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 14640:0:(osd_handler.c:1979:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 14640:0:(osd_handler.c:1979:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 14640:0:(osd_handler.c:1986:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 14640:0:(osd_handler.c:1986:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 14630:0:(osd_internal.h:1314:osd_trans_exec_op()) lustre-OST0000: opcode 7: before 516 < left 618, rollback = 7 Lustre: 14630:0:(osd_internal.h:1314:osd_trans_exec_op()) Skipped 7 previous similar messages Lustre: 14630:0:(osd_handler.c:1955:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 14630:0:(osd_handler.c:1955:osd_trans_dump_creds()) Skipped 7 previous similar messages Lustre: 14630:0:(osd_handler.c:1962:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 14630:0:(osd_handler.c:1962:osd_trans_dump_creds()) Skipped 7 previous similar messages Lustre: 14630:0:(osd_handler.c:1972:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 14630:0:(osd_handler.c:1972:osd_trans_dump_creds()) Skipped 7 previous similar messages Lustre: 14630:0:(osd_handler.c:1979:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 14630:0:(osd_handler.c:1979:osd_trans_dump_creds()) Skipped 7 previous similar messages Lustre: 14630:0:(osd_handler.c:1986:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 14630:0:(osd_handler.c:1986:osd_trans_dump_creds()) Skipped 7 previous similar messages Lustre: 16937:0:(osd_internal.h:1314:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 618, rollback = 7 Lustre: 16937:0:(osd_internal.h:1314:osd_trans_exec_op()) Skipped 7 previous similar messages Lustre: 16937:0:(osd_handler.c:1955:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 16937:0:(osd_handler.c:1955:osd_trans_dump_creds()) Skipped 7 previous similar messages Lustre: 16937:0:(osd_handler.c:1962:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 16937:0:(osd_handler.c:1962:osd_trans_dump_creds()) Skipped 7 previous similar messages Lustre: 16937:0:(osd_handler.c:1972:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 16937:0:(osd_handler.c:1972:osd_trans_dump_creds()) Skipped 7 previous similar messages Lustre: 16937:0:(osd_handler.c:1979:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 16937:0:(osd_handler.c:1979:osd_trans_dump_creds()) Skipped 7 previous similar messages Lustre: 16937:0:(osd_handler.c:1986:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 16937:0:(osd_handler.c:1986:osd_trans_dump_creds()) Skipped 7 previous similar messages Lustre: 14636:0:(osd_internal.h:1314:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 516 < left 618, rollback = 7 Lustre: 14636:0:(osd_internal.h:1314:osd_trans_exec_op()) Skipped 5 previous similar messages Lustre: 14636:0:(osd_handler.c:1955:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 14636:0:(osd_handler.c:1955:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 14636:0:(osd_handler.c:1962:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 14636:0:(osd_handler.c:1962:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 14636:0:(osd_handler.c:1972:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/78/0 Lustre: 14636:0:(osd_handler.c:1972:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 14636:0:(osd_handler.c:1979:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 14636:0:(osd_handler.c:1979:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 14636:0:(osd_handler.c:1986:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 14636:0:(osd_handler.c:1986:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: mdt02_005: service thread pid 17353 was inactive for 40.013 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 17353, comm: mdt02_005 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x913/0xd50 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x890 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock+0x88/0x1c0 [mdt] [<0>] mdt_object_find_lock+0x54/0x170 [mdt] [<0>] mdt_reint_setxattr+0x133/0x1020 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x74c/0xbc0 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc7e/0x1690 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 16385, comm: mdt01_003 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x913/0xd50 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x890 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock_try+0xa0/0x250 [mdt] [<0>] mdt_getattr_name_lock+0x1699/0x2bd0 [mdt] [<0>] mdt_intent_getattr+0x2cc/0x4e0 [mdt] [<0>] mdt_intent_opc.constprop.73+0x211/0xc50 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x34f/0x930 [ptlrpc] [<0>] ldlm_handle_enqueue+0x507/0x1850 [ptlrpc] [<0>] tgt_enqueue+0x68/0x240 [ptlrpc] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc7e/0x1690 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: mdt04_006: service thread pid 16981 was inactive for 40.063 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Lustre: Skipped 1 previous similar message Pid: 16981, comm: mdt04_006 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x913/0xd50 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x890 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock_try+0xa0/0x250 [mdt] [<0>] mdt_getattr_name_lock+0x1699/0x2bd0 [mdt] [<0>] mdt_intent_getattr+0x2cc/0x4e0 [mdt] [<0>] mdt_intent_opc.constprop.73+0x211/0xc50 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x34f/0x930 [ptlrpc] [<0>] ldlm_handle_enqueue+0x507/0x1850 [ptlrpc] [<0>] tgt_enqueue+0x68/0x240 [ptlrpc] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc7e/0x1690 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe LustreError: 14251:0:(ldlm_lockd.c:241:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0000_UUID lock: ffff8802e922d680/0x495b3d5f7f855f0f lrc: 3/0,0 mode: CR/CR res: [0x200000402:0x2da:0x0].0x0 bits 0xa/0x0 rrc: 6 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0x495b3d5f7f855f01 expref: 163 pid: 14268 timeout: 3069 lvb_type: 0 Lustre: mdt02_005: service thread pid 17353 completed after 100.190s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 16981:0:(ldlm_lockd.c:1478:ldlm_handle_enqueue()) ### lock on destroyed export ffff8802e3f8efc8 ns: mdt-lustre-MDT0000_UUID lock: ffff8800aa180f40/0x495b3d5f7f856c7c lrc: 3/0,0 mode: PR/PR res: [0x200000402:0x2da:0x0].0x0 bits 0x1b/0x0 rrc: 3 type: IBT gid 0 flags: 0x50200000000000 nid: 0@lo remote: 0x495b3d5f7f856c60 expref: 5 pid: 16981 timeout: 0 lvb_type: 0 Lustre: mdt04_006: service thread pid 16981 completed after 99.729s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: lustre-MDT0000-mdc-ffff8802e3e12548: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-MDT0000-mdc-ffff8802e3e12548: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: lustre-MDT0000-mdc-ffff8802e3e12548: This client was evicted by lustre-MDT0000; in progress operations using this service will fail. Lustre: mdt01_003: service thread pid 16385 completed after 99.922s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 24240:0:(file.c:262:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e3e12548: inode [0x200000401:0x9f:0x0] mdc close failed: rc = -108 LustreError: 24240:0:(ldlm_resource.c:1149:ldlm_resource_complain()) lustre-MDT0000-mdc-ffff8802e3e12548: namespace resource [0x200000402:0x2da:0x0].0x0 (ffff8800a7e22340) refcount nonzero (2) after lock cleanup; forcing cleanup. LustreError: 23987:0:(vvp_io.c:1921:vvp_io_init()) lustre: refresh file layout [0x200000401:0x242:0x0] error -108. Lustre: lustre-MDT0000-mdc-ffff8802e3e12548: Connection restored to 192.168.123.57@tcp (at 0@lo) Lustre: 14643:0:(osd_internal.h:1314:osd_trans_exec_op()) lustre-OST0000: opcode 7: before 516 < left 618, rollback = 7 Lustre: 14643:0:(osd_internal.h:1314:osd_trans_exec_op()) Skipped 3 previous similar messages Lustre: 14643:0:(osd_handler.c:1955:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 14643:0:(osd_handler.c:1955:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 14643:0:(osd_handler.c:1962:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 14643:0:(osd_handler.c:1962:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 14643:0:(osd_handler.c:1972:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 14643:0:(osd_handler.c:1972:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 14643:0:(osd_handler.c:1979:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 14643:0:(osd_handler.c:1979:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 14643:0:(osd_handler.c:1986:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 14643:0:(osd_handler.c:1986:osd_trans_dump_creds()) Skipped 3 previous similar messages 11[26970]: segfault at 8 ip 00007f5f5e49b7e8 sp 00007ffe739f2540 error 4 in ld-2.17.so[7f5f5e490000+22000] Lustre: 19879:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000401:0x342:0x0] with magic=0xbd60bd0 Lustre: 15767:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000401:0x406:0x0] with magic=0xbd60bd0 Lustre: 15767:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message 7[30375]: segfault at 8 ip 00007fe192af27e8 sp 00007ffd478df3e0 error 4 in ld-2.17.so[7fe192ae7000+22000] Lustre: 24527:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000401:0x50a:0x0] with magic=0xbd60bd0 Lustre: 24527:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message LustreError: 4187:0:(file.c:262:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a5890958: inode [0x200000403:0x3fc:0x0] mdc close failed: rc = -13 LustreError: 4187:0:(file.c:262:ll_close_inode_openhandle()) Skipped 9 previous similar messages 11[4326]: segfault at 8 ip 00007f0dfcacc7e8 sp 00007ffe53306150 error 4 in ld-2.17.so[7f0dfcac1000+22000] Lustre: 25128:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0x4a0:0x0] with magic=0xbd60bd0 Lustre: 25128:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message 14[6338]: segfault at 8 ip 00007fd7493d67e8 sp 00007fff474cb220 error 4 in ld-2.17.so[7fd7493cb000+22000] 10[7439]: segfault at 8 ip 00007f6efd0a57e8 sp 00007ffc02771020 error 4 in ld-2.17.so[7f6efd09a000+22000] Lustre: 14621:0:(osd_internal.h:1314:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 516 < left 618, rollback = 7 Lustre: 14621:0:(osd_internal.h:1314:osd_trans_exec_op()) Skipped 65 previous similar messages Lustre: 14621:0:(osd_handler.c:1955:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 14621:0:(osd_handler.c:1955:osd_trans_dump_creds()) Skipped 65 previous similar messages Lustre: 14621:0:(osd_handler.c:1962:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 14621:0:(osd_handler.c:1962:osd_trans_dump_creds()) Skipped 65 previous similar messages Lustre: 14621:0:(osd_handler.c:1972:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 14621:0:(osd_handler.c:1972:osd_trans_dump_creds()) Skipped 65 previous similar messages Lustre: 14621:0:(osd_handler.c:1979:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 14621:0:(osd_handler.c:1979:osd_trans_dump_creds()) Skipped 65 previous similar messages Lustre: 14621:0:(osd_handler.c:1986:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 14621:0:(osd_handler.c:1986:osd_trans_dump_creds()) Skipped 65 previous similar messages LustreError: 9967:0:(file.c:262:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e3e12548: inode [0x200000403:0x567:0x0] mdc close failed: rc = -13 Lustre: 25128:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0x754:0x0] with magic=0xbd60bd0 Lustre: 25128:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 3 previous similar messages 16[13452]: segfault at 8 ip 00007f0e03a327e8 sp 00007ffffefb7910 error 4 in ld-2.17.so[7f0e03a27000+22000] 6[16279]: segfault at 8 ip 00007fe6364df7e8 sp 00007ffd257d3250 error 4 in ld-2.17.so[7fe6364d4000+22000] 12[18997]: segfault at 8 ip 00007f3395e347e8 sp 00007ffc975bb900 error 4 in ld-2.17.so[7f3395e29000+22000] 17[23875]: segfault at 8 ip 00007f518d9c87e8 sp 00007ffd935f7170 error 4 in ld-2.17.so[7f518d9bd000+22000] Lustre: 20409:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000401:0x107d:0x0] with magic=0xbd60bd0 Lustre: 20409:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 3 previous similar messages 12[31943]: segfault at 8 ip 00007f7b452e57e8 sp 00007ffe9817e4a0 error 4 in ld-2.17.so[7f7b452da000+22000] Lustre: 14644:0:(osd_internal.h:1314:osd_trans_exec_op()) lustre-OST0000: opcode 7: before 516 < left 618, rollback = 7 Lustre: 14644:0:(osd_internal.h:1314:osd_trans_exec_op()) Skipped 71 previous similar messages Lustre: 14644:0:(osd_handler.c:1955:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 14644:0:(osd_handler.c:1955:osd_trans_dump_creds()) Skipped 71 previous similar messages Lustre: 14644:0:(osd_handler.c:1962:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 14644:0:(osd_handler.c:1962:osd_trans_dump_creds()) Skipped 71 previous similar messages Lustre: 14644:0:(osd_handler.c:1972:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 14644:0:(osd_handler.c:1972:osd_trans_dump_creds()) Skipped 71 previous similar messages Lustre: 14644:0:(osd_handler.c:1979:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 14644:0:(osd_handler.c:1979:osd_trans_dump_creds()) Skipped 71 previous similar messages Lustre: 14644:0:(osd_handler.c:1986:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 14644:0:(osd_handler.c:1986:osd_trans_dump_creds()) Skipped 71 previous similar messages 5[13858]: segfault at 8 ip 00007f876ab457e8 sp 00007ffea738dfa0 error 4 in ld-2.17.so[7f876ab3a000+22000] 1[16455]: segfault at 8 ip 00007f478aebc7e8 sp 00007ffd79a48ed0 error 4 in ld-2.17.so[7f478aeb1000+22000] Lustre: 16686:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0x17be:0x0] with magic=0xbd60bd0 Lustre: 16686:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message LustreError: 28435:0:(file.c:262:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e3e12548: inode [0x200000401:0x1b5a:0x0] mdc close failed: rc = -13 Lustre: 14266:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000401:0x1d2f:0x0] with magic=0xbd60bd0 Lustre: 14266:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 5 previous similar messages Lustre: mdt07_001: service thread pid 14280 was inactive for 40.150 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 14280, comm: mdt07_001 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x913/0xd50 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x890 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock+0x88/0x1c0 [mdt] [<0>] mdt_object_find_lock+0x54/0x170 [mdt] [<0>] mdt_reint_setxattr+0x133/0x1020 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x74c/0xbc0 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc7e/0x1690 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 19344, comm: mdt00_005 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x913/0xd50 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x890 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock_try+0xa0/0x250 [mdt] [<0>] mdt_getattr_name_lock+0x1699/0x2bd0 [mdt] [<0>] mdt_intent_getattr+0x2cc/0x4e0 [mdt] [<0>] mdt_intent_opc.constprop.73+0x211/0xc50 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x34f/0x930 [ptlrpc] [<0>] ldlm_handle_enqueue+0x507/0x1850 [ptlrpc] [<0>] tgt_enqueue+0x68/0x240 [ptlrpc] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc7e/0x1690 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 14259, comm: mdt00_001 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x913/0xd50 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x890 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock_try+0xa0/0x250 [mdt] [<0>] mdt_getattr_name_lock+0x1699/0x2bd0 [mdt] [<0>] mdt_intent_getattr+0x2cc/0x4e0 [mdt] [<0>] mdt_intent_opc.constprop.73+0x211/0xc50 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x34f/0x930 [ptlrpc] [<0>] ldlm_handle_enqueue+0x507/0x1850 [ptlrpc] [<0>] tgt_enqueue+0x68/0x240 [ptlrpc] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc7e/0x1690 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe LustreError: 14251:0:(ldlm_lockd.c:241:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0000_UUID lock: ffff8802e8d52d40/0x495b3d5f7fbc14db lrc: 3/0,0 mode: CR/CR res: [0x200000403:0x1aa6:0x0].0x0 bits 0xa/0x0 rrc: 5 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0x495b3d5f7fbc14cd expref: 840 pid: 19344 timeout: 3453 lvb_type: 0 Lustre: mdt07_001: service thread pid 14280 completed after 100.084s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 14227:0:(ldlm_lockd.c:2575:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1729566684 with bad export cookie 5285886068019682503 Lustre: lustre-MDT0000-mdc-ffff8800a5890958: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: lustre-MDT0000-mdc-ffff8800a5890958: This client was evicted by lustre-MDT0000; in progress operations using this service will fail. LustreError: 19344:0:(ldlm_lockd.c:1478:ldlm_handle_enqueue()) ### lock on destroyed export ffff8802e3dfe678 ns: mdt-lustre-MDT0000_UUID lock: ffff8800a99e0400/0x495b3d5f7fbc1d39 lrc: 3/0,0 mode: PR/PR res: [0x200000403:0x1aa6:0x0].0x0 bits 0x1b/0x0 rrc: 3 type: IBT gid 0 flags: 0x50200000000000 nid: 0@lo remote: 0x495b3d5f7fbc1d0f expref: 4 pid: 19344 timeout: 0 lvb_type: 0 Lustre: mdt00_001: service thread pid 14259 completed after 99.419s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_005: service thread pid 19344 completed after 99.888s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: lustre-MDT0000-mdc-ffff8800a5890958: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: 1841:0:(vvp_io.c:1921:vvp_io_init()) lustre: refresh file layout [0x200000403:0x1aa6:0x0] error -108. LustreError: 2151:0:(file.c:262:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a5890958: inode [0x200000403:0x1a9f:0x0] mdc close failed: rc = -108 LustreError: 1933:0:(file.c:5733:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000007:0x1:0x0] error: rc = -108 LustreError: 1975:0:(mdc_request.c:1454:mdc_read_page()) lustre-MDT0000-mdc-ffff8800a5890958: [0x200000401:0x1:0x0] lock enqueue fails: rc = -108 Lustre: lustre-MDT0000-mdc-ffff8800a5890958: Connection restored to 192.168.123.57@tcp (at 0@lo) 19[5023]: segfault at 8 ip 00007fead973e7e8 sp 00007fffba5e71c0 error 4 in ld-2.17.so[7fead9733000+22000] 2[5555]: segfault at 8 ip 00007fb89a2517e8 sp 00007fff20a56fc0 error 4 in ld-2.17.so[7fb89a246000+22000] 9[11088]: segfault at 8 ip 00007f8b0d4b07e8 sp 00007ffeee979ba0 error 4 in ld-2.17.so[7f8b0d4a5000+22000] LustreError: 11235:0:(file.c:262:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e3e12548: inode [0x200000404:0x35f:0x0] mdc close failed: rc = -13 LustreError: 11235:0:(file.c:262:ll_close_inode_openhandle()) Skipped 4 previous similar messages 11[11355]: segfault at 8 ip 00007f64dd6527e8 sp 00007ffd9ab669d0 error 4 in ld-2.17.so[7f64dd647000+22000] Lustre: 17151:0:(osd_internal.h:1314:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 516 < left 618, rollback = 7 Lustre: 17151:0:(osd_internal.h:1314:osd_trans_exec_op()) Skipped 109 previous similar messages Lustre: 17151:0:(osd_handler.c:1955:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 17151:0:(osd_handler.c:1955:osd_trans_dump_creds()) Skipped 109 previous similar messages Lustre: 17151:0:(osd_handler.c:1962:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 17151:0:(osd_handler.c:1962:osd_trans_dump_creds()) Skipped 109 previous similar messages Lustre: 17151:0:(osd_handler.c:1972:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 17151:0:(osd_handler.c:1972:osd_trans_dump_creds()) Skipped 109 previous similar messages Lustre: 17151:0:(osd_handler.c:1979:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 17151:0:(osd_handler.c:1979:osd_trans_dump_creds()) Skipped 109 previous similar messages Lustre: 17151:0:(osd_handler.c:1986:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 17151:0:(osd_handler.c:1986:osd_trans_dump_creds()) Skipped 109 previous similar messages 3[26044]: segfault at 8 ip 00007fa3446e27e8 sp 00007ffebe906770 error 4 in ld-2.17.so[7fa3446d7000+22000] Lustre: 18054:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000404:0x943:0x0] with magic=0xbd60bd0 Lustre: 18054:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 3 previous similar messages 15[31997]: segfault at 0 ip 00000000004043f0 sp 00007ffe598bd678 error 6 in 17 (deleted)[400000+6000] 8[8427]: segfault at 8 ip 00007fee87bee7e8 sp 00007ffc21e1b3a0 error 4 in ld-2.17.so[7fee87be3000+22000] 6[14885]: segfault at 8 ip 00007f394187d7e8 sp 00007ffd88688da0 error 4 in ld-2.17.so[7f3941872000+22000] 6[19155]: segfault at 8 ip 00007f05078287e8 sp 00007fff1977cc90 error 4 in ld-2.17.so[7f050781d000+22000] 6[21149]: segfault at 8 ip 00007fc5d2ef57e8 sp 00007fffd906b3e0 error 4 in ld-2.17.so[7fc5d2eea000+22000] Lustre: 19879:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0x2f49:0x0] with magic=0xbd60bd0 Lustre: 19879:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 13 previous similar messages LustreError: 1288:0:(file.c:262:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e3e12548: inode [0x200000404:0x164b:0x0] mdc close failed: rc = -13 16[11981]: segfault at 8 ip 00007f0ac9fcc7e8 sp 00007fff1ff4afa0 error 4 in ld-2.17.so[7f0ac9fc1000+22000] 2[12310]: segfault at 8 ip 00007f983d7bc7e8 sp 00007fffedb1f570 error 4 in ld-2.17.so[7f983d7b1000+22000] 7[15080]: segfault at 8 ip 00007fdc935f77e8 sp 00007ffdd4f659b0 error 4 in ld-2.17.so[7fdc935ec000+22000] 7[14938]: segfault at 8 ip 00007f18bb6b97e8 sp 00007ffd24429c20 error 4 in ld-2.17.so[7f18bb6ae000+22000] 11[17767]: segfault at 8 ip 00007fad574297e8 sp 00007fff2cd55f00 error 4 in ld-2.17.so[7fad5741e000+22000] traps: 0[18277] trap invalid opcode ip:40510d sp:7ffd22bbbe88 error:0 in 0[400000+6000] 2[28561]: segfault at 8 ip 00007f2f95c757e8 sp 00007fffbbda2300 error 4 in ld-2.17.so[7f2f95c6a000+22000] LustreError: 28561:0:(file.c:262:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e3e12548: inode [0x200000404:0x1fa9:0x0] mdc close failed: rc = -13 LustreError: 28561:0:(file.c:262:ll_close_inode_openhandle()) Skipped 1 previous similar message 8[29107]: segfault at 8 ip 00007f9d125897e8 sp 00007ffebc08af90 error 4 in ld-2.17.so[7f9d1257e000+22000] 12[3103]: segfault at 8 ip 00007f4891f177e8 sp 00007ffdc9f88ba0 error 4 in ld-2.17.so[7f4891f0c000+22000] 8[5662]: segfault at 7f0d896ed4ce ip 0000000000405dbf sp 00007ffd63ff5418 error 6 in 8[400000+6000] 16[8717]: segfault at 1c ip 00000000004017b0 sp 00007fffbddda330 error 6 in 14[400000+6000] 11[13934]: segfault at 8 ip 00007f10fc2777e8 sp 00007ffcbcb840d0 error 4 in ld-2.17.so[7f10fc26c000+22000] 16[16823]: segfault at 8 ip 00007fab523f67e8 sp 00007ffd822a3ae0 error 4 in ld-2.17.so[7fab523eb000+22000] LustreError: 17428:0:(file.c:262:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a5890958: inode [0x200000404:0x265d:0x0] mdc close failed: rc = -13 LustreError: 17428:0:(file.c:262:ll_close_inode_openhandle()) Skipped 1 previous similar message Lustre: 14271:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000404:0x2672:0x0] with magic=0xbd60bd0 Lustre: 14271:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 9 previous similar messages 5[17861]: segfault at 8 ip 00007f3dd3b137e8 sp 00007ffc6e2dfd20 error 4 in ld-2.17.so[7f3dd3b08000+22000] Lustre: 14640:0:(osd_internal.h:1314:osd_trans_exec_op()) lustre-OST0001: opcode 7: before 516 < left 618, rollback = 7 Lustre: 14640:0:(osd_internal.h:1314:osd_trans_exec_op()) Skipped 377 previous similar messages Lustre: 14640:0:(osd_handler.c:1955:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 14640:0:(osd_handler.c:1955:osd_trans_dump_creds()) Skipped 377 previous similar messages Lustre: 14640:0:(osd_handler.c:1962:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 14640:0:(osd_handler.c:1962:osd_trans_dump_creds()) Skipped 377 previous similar messages Lustre: 14640:0:(osd_handler.c:1972:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 1/3/0 Lustre: 14640:0:(osd_handler.c:1972:osd_trans_dump_creds()) Skipped 377 previous similar messages Lustre: 14640:0:(osd_handler.c:1979:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 14640:0:(osd_handler.c:1979:osd_trans_dump_creds()) Skipped 377 previous similar messages Lustre: 14640:0:(osd_handler.c:1986:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 14640:0:(osd_handler.c:1986:osd_trans_dump_creds()) Skipped 377 previous similar messages 17[5731]: segfault at 8 ip 00007f50680617e8 sp 00007ffdee447160 error 4 in ld-2.17.so[7f5068056000+22000] 2[7311]: segfault at 8 ip 00007f04e55d07e8 sp 00007fffcc331ab0 error 4 in ld-2.17.so[7f04e55c5000+22000] 9[9099]: segfault at 8 ip 00007fdb951877e8 sp 00007fffc1b7c390 error 4 in ld-2.17.so[7fdb9517c000+22000] 11[15572]: segfault at 8 ip 00007f97df23e7e8 sp 00007ffec80da180 error 4 in ld-2.17.so[7f97df233000+22000] LustreError: 17063:0:(file.c:262:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e3e12548: inode [0x200000404:0x30d0:0x0] mdc close failed: rc = -13 LustreError: 17063:0:(file.c:262:ll_close_inode_openhandle()) Skipped 4 previous similar messages 3[19981]: segfault at 8 ip 00007f094e92c7e8 sp 00007ffc4447b9f0 error 4 in ld-2.17.so[7f094e921000+22000] 2[26957]: segfault at 8 ip 00007f87a8d507e8 sp 00007ffd594b3250 error 4 in ld-2.17.so[7f87a8d45000+22000] 5[6049]: segfault at 8 ip 00007fa1f6e997e8 sp 00007ffd60db5be0 error 4 in ld-2.17.so[7fa1f6e8e000+22000] 10[7686]: segfault at 8 ip 00007ff97a9987e8 sp 00007ffd04b87b20 error 4 in ld-2.17.so[7ff97a98d000+22000] 1[13940]: segfault at 8 ip 00007f866fb787e8 sp 00007ffd6fe8a5f0 error 4 in ld-2.17.so[7f866fb6d000+22000] 16[22447]: segfault at 8 ip 00007ffbd2c5f7e8 sp 00007ffc681f1e30 error 4 in ld-2.17.so[7ffbd2c54000+22000] 2[22751]: segfault at 8 ip 00007f968353b7e8 sp 00007fff3e90d8b0 error 4 in ld-2.17.so[7f9683530000+22000] 14[23824]: segfault at 8 ip 00007f7c819ac7e8 sp 00007ffefdfc3360 error 4 in ld-2.17.so[7f7c819a1000+22000] 0[24414]: segfault at 8 ip 00007faa84c8e7e8 sp 00007ffccff8f940 error 4 in ld-2.17.so[7faa84c83000+22000] 12[24633]: segfault at 8 ip 00007f6bce1e57e8 sp 00007fffb93ee660 error 4 in ld-2.17.so[7f6bce1da000+22000] 7[25652]: segfault at 8 ip 00007f7d90abc7e8 sp 00007ffd085882f0 error 4 in ld-2.17.so[7f7d90ab1000+22000] 4[26968]: segfault at 8 ip 00007f3a65b5d7e8 sp 00007fffa0d4e8a0 error 4 in ld-2.17.so[7f3a65b52000+22000] 7[31268]: segfault at 8 ip 00007fce3db167e8 sp 00007ffde31aa570 error 4 in ld-2.17.so[7fce3db0b000+22000] 12[1965]: segfault at 8 ip 00007fd5f4db57e8 sp 00007ffc888ca340 error 4 in ld-2.17.so[7fd5f4daa000+22000] LustreError: 6643:0:(file.c:262:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a5890958: inode [0x200000403:0x5d93:0x0] mdc close failed: rc = -13 7[13803]: segfault at 8 ip 00007f02c304f7e8 sp 00007ffe8df743d0 error 4 in ld-2.17.so[7f02c3044000+22000] 19[19054]: segfault at 8 ip 00007ffb710e07e8 sp 00007ffeaa325830 error 4 in ld-2.17.so[7ffb710d5000+22000] Lustre: mdt03_007: service thread pid 15321 was inactive for 40.134 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Lustre: Skipped 2 previous similar messages Pid: 15321, comm: mdt03_007 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x913/0xd50 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x890 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock+0x88/0x1c0 [mdt] [<0>] mdt_reint_link+0x8ef/0xf10 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x74c/0xbc0 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc7e/0x1690 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 14264, comm: mdt02_000 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x913/0xd50 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x890 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock+0x88/0x1c0 [mdt] [<0>] mdt_getattr_name_lock+0xbf3/0x2bd0 [mdt] [<0>] mdt_intent_getattr+0x2cc/0x4e0 [mdt] [<0>] mdt_intent_opc.constprop.73+0x211/0xc50 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x34f/0x930 [ptlrpc] [<0>] ldlm_handle_enqueue+0x507/0x1850 [ptlrpc] [<0>] tgt_enqueue+0x68/0x240 [ptlrpc] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc7e/0x1690 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: mdt03_005: service thread pid 24527 was inactive for 40.086 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Lustre: Skipped 1 previous similar message Pid: 24527, comm: mdt03_005 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x913/0xd50 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x890 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock+0x88/0x1c0 [mdt] [<0>] mdt_intent_getxattr+0x78/0x320 [mdt] [<0>] mdt_intent_opc.constprop.73+0x211/0xc50 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x34f/0x930 [ptlrpc] [<0>] ldlm_handle_enqueue+0x507/0x1850 [ptlrpc] [<0>] tgt_enqueue+0x68/0x240 [ptlrpc] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc7e/0x1690 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: lustre-OST0000-osc-ffff8802e3e12548: disconnect after 20s idle LustreError: 14251:0:(ldlm_lockd.c:241:expired_lock_main()) ### lock callback timer expired after 101s: evicting client at 0@lo ns: mdt-lustre-MDT0000_UUID lock: ffff8802f41670c0/0x495b3d5f805763ae lrc: 3/0,0 mode: PR/PR res: [0x200000403:0x61fd:0x0].0x0 bits 0x1b/0x0 rrc: 6 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0x495b3d5f8057638b expref: 1100 pid: 16656 timeout: 4590 lvb_type: 0 Lustre: mdt03_007: service thread pid 15321 completed after 100.559s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt03_005: service thread pid 24527 completed after 100.004s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: lustre-MDT0000-mdc-ffff8802e3e12548: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-MDT0000-mdc-ffff8802e3e12548: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: mdt02_000: service thread pid 14264 completed after 100.275s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: lustre-MDT0000-mdc-ffff8802e3e12548: This client was evicted by lustre-MDT0000; in progress operations using this service will fail. LustreError: 20130:0:(vvp_io.c:1921:vvp_io_init()) lustre: refresh file layout [0x200000403:0x61fd:0x0] error -5. LustreError: 19992:0:(file.c:5733:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000007:0x1:0x0] error: rc = -108 LustreError: 19992:0:(file.c:5733:ll_inode_revalidate_fini()) Skipped 1 previous similar message Lustre: lustre-MDT0000-mdc-ffff8802e3e12548: Connection restored to 192.168.123.57@tcp (at 0@lo) 15[22984]: segfault at 8 ip 00007f37577077e8 sp 00007ffd814e5fe0 error 4 in ld-2.17.so[7f37576fc000+22000] 9[27408]: segfault at 8 ip 00007fc1013537e8 sp 00007ffca8ab8560 error 4 in ld-2.17.so[7fc101348000+22000] Lustre: 16937:0:(osd_internal.h:1314:osd_trans_exec_op()) lustre-OST0002: opcode 7: before 515 < left 618, rollback = 7 Lustre: 16937:0:(osd_internal.h:1314:osd_trans_exec_op()) Skipped 329 previous similar messages Lustre: 16937:0:(osd_handler.c:1955:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 16937:0:(osd_handler.c:1955:osd_trans_dump_creds()) Skipped 329 previous similar messages Lustre: 16937:0:(osd_handler.c:1962:osd_trans_dump_creds()) attr_set: 1/1/1, xattr_set: 2/15/0 Lustre: 16937:0:(osd_handler.c:1962:osd_trans_dump_creds()) Skipped 329 previous similar messages Lustre: 16937:0:(osd_handler.c:1972:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 16937:0:(osd_handler.c:1972:osd_trans_dump_creds()) Skipped 329 previous similar messages Lustre: 16937:0:(osd_handler.c:1979:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 16937:0:(osd_handler.c:1979:osd_trans_dump_creds()) Skipped 329 previous similar messages Lustre: 16937:0:(osd_handler.c:1986:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 16937:0:(osd_handler.c:1986:osd_trans_dump_creds()) Skipped 329 previous similar messages Lustre: 14268:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000405:0x358:0x0] with magic=0xbd60bd0 Lustre: 14268:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 31 previous similar messages 14[7370]: segfault at 8 ip 00007f2b75bcd7e8 sp 00007fff205a6be0 error 4 in ld-2.17.so[7f2b75bc2000+22000] 5[8041]: segfault at 8 ip 00007fda16ab27e8 sp 00007ffdf7596070 error 4 in ld-2.17.so[7fda16aa7000+22000] 2[9331]: segfault at 8 ip 00007fb7786ca7e8 sp 00007fff6f116670 error 4 in ld-2.17.so[7fb7786bf000+22000] 7[10202]: segfault at 74 ip 0000000000404ca6 sp 00007ffeacce37f8 error 6 in 7[400000+6000] 18[13269]: segfault at 8 ip 00007fd42ee8e7e8 sp 00007ffe031dec00 error 4 in ld-2.17.so[7fd42ee83000+22000] 7[18493]: segfault at 8 ip 00007f54791057e8 sp 00007ffcd7385520 error 4 in ld-2.17.so[7f54790fa000+22000] 18[18685]: segfault at 8 ip 00007efc0fb7b7e8 sp 00007fff583cd840 error 4 in ld-2.17.so[7efc0fb70000+22000] 6[24203]: segfault at 406000 ip 0000000000406000 sp 00007ffd5cee2578 error 14 in 6[606000+1000] 8[1893]: segfault at 8 ip 00007f6b0797d7e8 sp 00007ffe527d36c0 error 4 in ld-2.17.so[7f6b07972000+22000] 18[6278]: segfault at 8 ip 00007f4ecde987e8 sp 00007fff66bb8790 error 4 in ld-2.17.so[7f4ecde8d000+22000] 18[7488]: segfault at 8 ip 00007f4d485f67e8 sp 00007ffc4d5875c0 error 4 in ld-2.17.so[7f4d485eb000+22000] 4[12140]: segfault at 8 ip 00007f4457cf17e8 sp 00007ffec612eff0 error 4 in ld-2.17.so[7f4457ce6000+22000] 10[27751]: segfault at 8 ip 00007f5d870607e8 sp 00007ffca5cee310 error 4 in ld-2.17.so[7f5d87055000+22000] 0[505]: segfault at 4017b0 ip 00000000004017b0 sp 00007ffc09018480 error 14 1[1704]: segfault at 8 ip 00007ff6aeba87e8 sp 00007fffc28b1120 error 4 in ld-2.17.so[7ff6aeb9d000+22000] 9[1697]: segfault at 8 ip 00007f498addf7e8 sp 00007ffd2c5bc4e0 error 4 in ld-2.17.so[7f498add4000+22000] 5[2007]: segfault at 8 ip 00007fe510df37e8 sp 00007ffdd145fd80 error 4 in ld-2.17.so[7fe510de8000+22000] 10[5453]: segfault at 8 ip 00007feb30aec7e8 sp 00007fffcc8aa040 error 4 in ld-2.17.so[7feb30ae1000+22000] 12[6999]: segfault at 8 ip 00007f9950db77e8 sp 00007ffe66b3c950 error 4 in ld-2.17.so[7f9950dac000+22000] 0[9512]: segfault at 4017b0 ip 00000000004017b0 sp 00007ffed4dd7ba0 error 14 11[12336]: segfault at 8 ip 00007fecf7d3f7e8 sp 00007fff13667de0 error 4 in ld-2.17.so[7fecf7d34000+22000] 17[15003]: segfault at 8 ip 00007fc4fd40a7e8 sp 00007ffedab42630 error 4 in ld-2.17.so[7fc4fd3ff000+22000] 1[15388]: segfault at 8 ip 00007f5d119697e8 sp 00007ffc2dbb67e0 error 4 in ld-2.17.so[7f5d1195e000+22000] 17[17102]: segfault at 8 ip 00007f488804a7e8 sp 00007ffe131871e0 error 4 in ld-2.17.so[7f488803f000+22000] 10[20420]: segfault at 8 ip 00007fd1f63827e8 sp 00007ffeec206520 error 4 in ld-2.17.so[7fd1f6377000+22000] 12[23331]: segfault at 8 ip 00007f81b38627e8 sp 00007ffcf7c182b0 error 4 in ld-2.17.so[7f81b3857000+22000] 17[24694]: segfault at 8 ip 00007fcb0c6797e8 sp 00007ffde51883d0 error 4 in ld-2.17.so[7fcb0c66e000+22000] 11[25517]: segfault at 8 ip 00007ff53ae927e8 sp 00007fffdb1564c0 error 4 in ld-2.17.so[7ff53ae87000+22000] 14[28870]: segfault at 40452e ip 0000000000405384 sp 00007ffe458ee0d8 error 7 in 14[400000+6000] traps: 17[4220] trap invalid opcode ip:4058ba sp:7ffe7ca94f88 error:0 in 17[400000+6000] 11[4738]: segfault at 8 ip 00007f648874f7e8 sp 00007ffe72816e10 error 4 in ld-2.17.so[7f6488744000+22000] 13[4792]: segfault at 8 ip 00007f0d4971d7e8 sp 00007ffdaf259470 error 4 in ld-2.17.so[7f0d49712000+22000] 13[6262]: segfault at 8 ip 00007f049ccb37e8 sp 00007ffd1dcfc390 error 4 in ld-2.17.so[7f049cca8000+22000] 6[6341]: segfault at 8 ip 00007f9c5082b7e8 sp 00007ffe2680d820 error 4 in ld-2.17.so[7f9c50820000+22000] 10[7356]: segfault at 8 ip 00007f34ab5157e8 sp 00007ffdc9e7b210 error 4 in ld-2.17.so[7f34ab50a000+22000] Lustre: 2139:0:(osd_internal.h:1314:osd_trans_exec_op()) lustre-OST0003: opcode 7: before 516 < left 618, rollback = 7 Lustre: 2139:0:(osd_internal.h:1314:osd_trans_exec_op()) Skipped 465 previous similar messages Lustre: 2139:0:(osd_handler.c:1955:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 2139:0:(osd_handler.c:1955:osd_trans_dump_creds()) Skipped 465 previous similar messages Lustre: 2139:0:(osd_handler.c:1962:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 2/15/0 Lustre: 2139:0:(osd_handler.c:1962:osd_trans_dump_creds()) Skipped 465 previous similar messages Lustre: 2139:0:(osd_handler.c:1972:osd_trans_dump_creds()) write: 2/618/0, punch: 0/0/0, quota 4/150/0 Lustre: 2139:0:(osd_handler.c:1972:osd_trans_dump_creds()) Skipped 465 previous similar messages Lustre: 2139:0:(osd_handler.c:1979:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 2139:0:(osd_handler.c:1979:osd_trans_dump_creds()) Skipped 465 previous similar messages Lustre: 2139:0:(osd_handler.c:1986:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 2139:0:(osd_handler.c:1986:osd_trans_dump_creds()) Skipped 465 previous similar messages 13[19249]: segfault at 8 ip 00007f8d67e577e8 sp 00007ffd53f895d0 error 4 in ld-2.17.so[7f8d67e4c000+22000] 14[20754]: segfault at 0 ip (null) sp 00007fffc7d05d58 error 14 in 14[400000+6000] 6[20998]: segfault at 8 ip 00007f09756e27e8 sp 00007ffcc3f94f20 error 4 in ld-2.17.so[7f09756d7000+22000] 4[24837]: segfault at 8 ip 00007f411f7b17e8 sp 00007fff883f4460 error 4 in ld-2.17.so[7f411f7a6000+22000] 13[25708]: segfault at 8 ip 00007ff47ed5b7e8 sp 00007ffcc4d43dc0 error 4 in ld-2.17.so[7ff47ed50000+22000] 10[26166]: segfault at 8 ip 00007f37b1e797e8 sp 00007ffe649d2e60 error 4 in ld-2.17.so[7f37b1e6e000+22000] Lustre: 14258:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000405:0x2f77:0x0] with magic=0xbd60bd0 Lustre: 14258:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 7 previous similar messages 19[10392]: segfault at 8 ip 00007f32c4c317e8 sp 00007ffd3d862710 error 4 in ld-2.17.so[7f32c4c26000+22000] LustreError: 11082:0:(file.c:262:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e3e12548: inode [0x200000404:0x7a01:0x0] mdc close failed: rc = -13 LustreError: 11082:0:(file.c:262:ll_close_inode_openhandle()) Skipped 12 previous similar messages 9[14857]: segfault at 0 ip (null) sp 00007ffc2bf8f368 error 14 in 19[400000+6000] 9[15216]: segfault at 8 ip 00007fb9885c07e8 sp 00007ffedd35b5f0 error 4 in ld-2.17.so[7fb9885b5000+22000] 14[15656]: segfault at 8 ip 00007f81958447e8 sp 00007fff0fffc0f0 error 4 in ld-2.17.so[7f8195839000+22000] 18[16227]: segfault at 8 ip 00007fec5106a7e8 sp 00007ffe6b6e22f0 error 4 in ld-2.17.so[7fec5105f000+22000] 13[19932]: segfault at 8 ip 00007f712d8dc7e8 sp 00007ffc1221de80 error 4 in ld-2.17.so[7f712d8d1000+22000] 1[21366]: segfault at 8 ip 00007ff4f40547e8 sp 00007ffdb190ff00 error 4 in ld-2.17.so[7ff4f4049000+22000] 17[22497]: segfault at 8 ip 00007f742ed567e8 sp 00007ffff31ae8b0 error 4 in ld-2.17.so[7f742ed4b000+22000] 16[25688]: segfault at 8 ip 00007fb8789f87e8 sp 00007ffd3531bca0 error 4 in ld-2.17.so[7fb8789ed000+22000] Lustre: ll_ost00_007: service thread pid 5184 was inactive for 40.135 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 5184, comm: ll_ost00_007 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] call_rwsem_down_write_failed+0x17/0x30 [<0>] osd_write_lock+0x5f/0xc0 [osd_ldiskfs] [<0>] ofd_attr_set+0x552/0xad0 [ofd] [<0>] ofd_setattr_hdl+0x325/0x9d0 [ofd] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc7e/0x1690 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 7563, comm: ll_ost00_008 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] wait_transaction_locked+0x93/0xe0 [jbd2] [<0>] add_transaction_credits+0x270/0x310 [jbd2] [<0>] start_this_handle+0x1cc/0x460 [jbd2] [<0>] jbd2__journal_restart+0x10f/0x170 [jbd2] [<0>] jbd2_journal_restart+0x13/0x20 [jbd2] [<0>] osd_fallocate+0x4c4/0xac0 [osd_ldiskfs] [<0>] ofd_object_fallocate+0x6cc/0xac0 [ofd] [<0>] ofd_fallocate_hdl+0x36a/0xc70 [ofd] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc7e/0x1690 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 14581, comm: ll_ost00_000 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Lustre: ll_ost00_005: service thread pid 28195 was inactive for 40.192 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Call Trace: [<0>] wait_transaction_locked+0x93/0xe0 [jbd2] [<0>] add_transaction_credits+0x270/0x310 [jbd2] [<0>] start_this_handle+0x1cc/0x460 [jbd2] [<0>] jbd2__journal_start+0xf3/0x200 [jbd2] [<0>] __ldiskfs_journal_start_sb+0x6d/0x100 [ldiskfs] [<0>] osd_trans_start+0x279/0x6f0 [osd_ldiskfs] [<0>] ofd_trans_start+0x68/0xe0 [ofd] [<0>] ofd_destroy+0x3a0/0xad0 [ofd] [<0>] ofd_destroy_by_fid+0x33c/0x610 [ofd] [<0>] ofd_destroy_hdl+0x20c/0xae0 [ofd] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc7e/0x1690 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: ll_ost04_003: service thread pid 17163 was inactive for 40.083 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 3 previous similar messages Lustre: ll_ost05_002: service thread pid 14598 was inactive for 40.035 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 5 previous similar messages Lustre: ll_ost_io00_003: service thread pid 18255 was inactive for 40.047 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 4 previous similar messages Lustre: ll_ost_io02_005: service thread pid 2139 was inactive for 40.154 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 1 previous similar message Lustre: ll_ost_io04_001: service thread pid 14634 was inactive for 40.090 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 2 previous similar messages Lustre: ll_ost_io07_002: service thread pid 14644 was inactive for 72.251 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. | Link to test |
racer test 1: racer on clients: centos-95.localnet DURATION=2700 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 PGD 266714067 PUD 25ef5e067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) osp(OE) ofd(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) crc32_generic crc_t10dif crct10dif_generic crct10dif_common pcspkr virtio_console i2c_piix4 virtio_balloon ip_tables rpcsec_gss_krb5 drm_kms_helper ttm drm ata_generic pata_acpi drm_panel_orientation_quirks ata_piix serio_raw i2c_core virtio_blk floppy libata [last unloaded: libcfs] CPU: 3 PID: 17658 Comm: ll_sa_17600 Kdump: loaded Tainted: P OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.16.0-3.module_el8.7.0+1218+f626c2ff 04/01/2014 task: ffff880250899280 ti: ffff8800773a8000 task.ti: ffff8800773a8000 RIP: 0010:[<ffffffff813f502c>] [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 RSP: 0018:ffff8800773abd80 EFLAGS: 00010282 RAX: ffff880127568108 RBX: 0000000000000008 RCX: 000000010026001e RDX: 0000000000000026 RSI: ffff880127568438 RDI: 0000000000000008 RBP: ffff8800773abd90 R08: ffff88014a0f93e8 R09: 0000000000000000 R10: ffff88014a0f8d48 R11: ffff88014a0f93e8 R12: ffff880127568438 R13: 0000000000000000 R14: ffff88014a0f93e8 R15: ffff880127568438 FS: 0000000000000000(0000) GS:ffff880331ac0000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 00000002945e8000 CR4: 00000000000007e0 Call Trace: [<ffffffffa1559a31>] ll_statahead_thread+0x741/0x2200 [lustre] [<ffffffff817e320e>] ? _raw_spin_lock+0x1e/0x20 [<ffffffff810dbb51>] ? put_prev_entity+0x31/0x400 [<ffffffffa15592f0>] ? ll_statahead_handle.constprop.28+0x1c0/0x1c0 [lustre] [<ffffffff810ba114>] kthread+0xe4/0xf0 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 | LustreError: 29479:0:(mdt_handler.c:776:mdt_pack_acl2body()) lustre-MDT0000: unable to read [0x200000404:0x24:0x0] ACL: rc = -2 LustreError: 19523:0:(mdt_reint.c:2533:mdt_reint_migrate()) lustre-MDT0000: migrate [0x200000403:0x1:0x0]/5 failed: rc = -114 LustreError: 661:0:(file.c:262:ll_close_inode_openhandle()) lustre-clilmv-ffff88027a688958: inode [0x200000403:0x14:0x0] mdc close failed: rc = -116 LustreError: 19517:0:(mdt_reint.c:2533:mdt_reint_migrate()) lustre-MDT0000: migrate [0x200000403:0x2:0x0]/2 failed: rc = -114 LustreError: 1517:0:(file.c:262:ll_close_inode_openhandle()) lustre-clilmv-ffff8801e178b7e8: inode [0x200000403:0x2c:0x0] mdc close failed: rc = -116 LustreError: 19526:0:(mdt_reint.c:2533:mdt_reint_migrate()) lustre-MDT0001: migrate [0x240000403:0x1:0x0]/14 failed: rc = -114 LustreError: 19526:0:(mdt_reint.c:2533:mdt_reint_migrate()) Skipped 1 previous similar message LustreError: 3052:0:(file.c:262:ll_close_inode_openhandle()) lustre-clilmv-ffff8801e178b7e8: inode [0x240000403:0x2c:0x0] mdc close failed: rc = -116 LustreError: 19531:0:(mdt_reint.c:2533:mdt_reint_migrate()) lustre-MDT0000: migrate [0x200000403:0x2:0x0]/8 failed: rc = -114 LustreError: 19531:0:(mdt_reint.c:2533:mdt_reint_migrate()) Skipped 2 previous similar messages LustreError: 3940:0:(file.c:262:ll_close_inode_openhandle()) lustre-clilmv-ffff8801e178b7e8: inode [0x280000404:0x31:0x0] mdc close failed: rc = -116 LustreError: 3940:0:(file.c:262:ll_close_inode_openhandle()) Skipped 2 previous similar messages Lustre: 19462:0:(mdt_recovery.c:148:mdt_req_from_lrd()) @@@ restoring transno req@ffff880155e23c40 x1811500712581504/t4294968709(0) o101->6c94fc14-2ae2-4b26-8b2f-bfc62ce1ad62@0@lo:641/0 lens 376/816 e 0 to 0 dl 1727581826 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 LustreError: 3452:0:(mdt_reint.c:2533:mdt_reint_migrate()) lustre-MDT0000: migrate [0x200000403:0x2:0x0]/11 failed: rc = -16 LustreError: 3452:0:(mdt_reint.c:2533:mdt_reint_migrate()) Skipped 5 previous similar messages LustreError: 4569:0:(file.c:262:ll_close_inode_openhandle()) lustre-clilmv-ffff88027a688958: inode [0x280000404:0x54:0x0] mdc close failed: rc = -116 LustreError: 19523:0:(mdt_reint.c:2533:mdt_reint_migrate()) lustre-MDT0001: migrate [0x240000403:0x92:0x0]/1 failed: rc = -114 LustreError: 19523:0:(mdt_reint.c:2533:mdt_reint_migrate()) Skipped 7 previous similar messages LustreError: 3064:0:(file.c:262:ll_close_inode_openhandle()) lustre-clilmv-ffff8801e178b7e8: inode [0x240000404:0x53:0x0] mdc close failed: rc = -116 LustreError: 3064:0:(file.c:262:ll_close_inode_openhandle()) Skipped 4 previous similar messages Lustre: 19456:0:(mdt_recovery.c:148:mdt_req_from_lrd()) @@@ restoring transno req@ffff880155e24b40 x1811500715176704/t4294970538(0) o101->cfdd0331-5323-4375-9e7e-4f2f34ecbe02@0@lo:667/0 lens 376/816 e 0 to 0 dl 1727581852 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 LustreError: 3452:0:(mdt_reint.c:2533:mdt_reint_migrate()) lustre-MDT0002: migrate [0x280000403:0x1:0x0]/19 failed: rc = -114 LustreError: 3452:0:(mdt_reint.c:2533:mdt_reint_migrate()) Skipped 18 previous similar messages LustreError: 11168:0:(file.c:262:ll_close_inode_openhandle()) lustre-clilmv-ffff88027a688958: inode [0x280000404:0x7c:0x0] mdc close failed: rc = -116 LustreError: 11168:0:(file.c:262:ll_close_inode_openhandle()) Skipped 3 previous similar messages Lustre: 620:0:(mdt_recovery.c:148:mdt_req_from_lrd()) @@@ restoring transno req@ffff88006efbf840 x1811500719270144/t4294971533(0) o101->6c94fc14-2ae2-4b26-8b2f-bfc62ce1ad62@0@lo:703/0 lens 376/864 e 0 to 0 dl 1727581888 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 LustreError: 14932:0:(mdt_reint.c:2533:mdt_reint_migrate()) lustre-MDT0001: migrate [0x240000403:0x1:0x0]/13 failed: rc = -114 LustreError: 14932:0:(mdt_reint.c:2533:mdt_reint_migrate()) Skipped 22 previous similar messages LustreError: 17088:0:(lcommon_cl.c:195:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000404:0x17f:0x0]: rc = -5 LustreError: 17088:0:(llite_lib.c:3731:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 17407:0:(file.c:262:ll_close_inode_openhandle()) lustre-clilmv-ffff88027a688958: inode [0x200000403:0x15e:0x0] mdc close failed: rc = -116 LustreError: 17407:0:(file.c:262:ll_close_inode_openhandle()) Skipped 2 previous similar messages LustreError: 18207:0:(lcommon_cl.c:195:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000404:0x17f:0x0]: rc = -5 LustreError: 18207:0:(llite_lib.c:3731:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 23070:0:(lcommon_cl.c:195:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000404:0x17f:0x0]: rc = -5 LustreError: 23070:0:(llite_lib.c:3731:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 23342:0:(lcommon_cl.c:195:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000404:0x17f:0x0]: rc = -5 LustreError: 23342:0:(llite_lib.c:3731:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 Lustre: lustre-MDT0000: trigger partial OI scrub for RPC inconsistency, checking FID [0x200000404:0x120:0x0]/0xa): rc = 0 5[26223]: segfault at 8 ip 00007feab5a8d7e8 sp 00007fff6989b010 error 4 in ld-2.17.so[7feab5a82000+22000] Lustre: dir [0x280000404:0xa9:0x0] stripe 2 readdir failed: -2, directory is partially accessed! Lustre: dir [0x200000404:0x1b8:0x0] stripe 1 readdir failed: -2, directory is partially accessed! LustreError: 3452:0:(mdd_dir.c:4457:mdd_migrate_cmd_check()) lustre-MDD0001: '19' migration was interrupted, run 'lfs migrate -m 0 -c 1 -H crush 19' to finish migration: rc = -1 Lustre: 3450:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000404:0x164:0x0] with magic=0xbd60bd0 LustreError: 24071:0:(lcommon_cl.c:195:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000404:0x1a0:0x0]: rc = -5 LustreError: 24071:0:(llite_lib.c:3731:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 Lustre: 19459:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000403:0x269:0x0] with magic=0xbd60bd0 Lustre: 19459:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message LustreError: 12588:0:(mdt_reint.c:2533:mdt_reint_migrate()) lustre-MDT0001: migrate [0x240000403:0x1:0x0]/16 failed: rc = -16 LustreError: 12588:0:(mdt_reint.c:2533:mdt_reint_migrate()) Skipped 32 previous similar messages LustreError: 4976:0:(lcommon_cl.c:195:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000404:0x1a0:0x0]: rc = -5 LustreError: 4976:0:(lcommon_cl.c:195:cl_file_inode_init()) Skipped 1 previous similar message LustreError: 4976:0:(llite_lib.c:3731:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 4976:0:(llite_lib.c:3731:ll_prep_inode()) Skipped 1 previous similar message LustreError: 21592:0:(file.c:262:ll_close_inode_openhandle()) lustre-clilmv-ffff88027a688958: inode [0x280000404:0xbe:0x0] mdc close failed: rc = -2 LustreError: 21592:0:(file.c:262:ll_close_inode_openhandle()) Skipped 6 previous similar messages LustreError: 6462:0:(lcommon_cl.c:195:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000403:0x272:0x0]: rc = -5 LustreError: 6462:0:(lcommon_cl.c:195:cl_file_inode_init()) Skipped 2 previous similar messages LustreError: 6462:0:(llite_lib.c:3731:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 6462:0:(llite_lib.c:3731:ll_prep_inode()) Skipped 2 previous similar messages 6[15458]: segfault at 8 ip 00007fe96f78e7e8 sp 00007fffb0aa49c0 error 4 in ld-2.17.so[7fe96f783000+22000] LustreError: 18343:0:(lcommon_cl.c:195:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000403:0x31a:0x0]: rc = -5 LustreError: 18343:0:(lcommon_cl.c:195:cl_file_inode_init()) Skipped 2 previous similar messages LustreError: 18343:0:(llite_lib.c:3731:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 18343:0:(llite_lib.c:3731:ll_prep_inode()) Skipped 2 previous similar messages 8[16083]: segfault at 8 ip 00007f70773f97e8 sp 00007ffe6e6d7470 error 4 in ld-2.17.so[7f70773ee000+22000] Lustre: dir [0x240000404:0x227:0x0] stripe 1 readdir failed: -2, directory is partially accessed! LustreError: 30631:0:(mdd_dir.c:4457:mdd_migrate_cmd_check()) lustre-MDD0001: '9' migration was interrupted, run 'lfs migrate -m 2 -c 3 -H crush 9' to finish migration: rc = -1 11[25438]: segfault at 8 ip 00007fa5667c47e8 sp 00007ffc05154b80 error 4 in ld-2.17.so[7fa5667b9000+22000] Lustre: 19454:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000404:0x2fa:0x0] with magic=0xbd60bd0 Lustre: 19454:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message LustreError: 19525:0:(mdd_dir.c:4457:mdd_migrate_cmd_check()) lustre-MDD0002: '18' migration was interrupted, run 'lfs migrate -m 1 -c 1 -H crush 18' to finish migration: rc = -1 Lustre: 29590:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0x40a:0x0] with magic=0xbd60bd0 Lustre: 29590:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message LustreError: 19531:0:(mdt_reint.c:2533:mdt_reint_migrate()) lustre-MDT0000: migrate [0x280000403:0x33b:0x0]/14 failed: rc = -114 LustreError: 19531:0:(mdt_reint.c:2533:mdt_reint_migrate()) Skipped 61 previous similar messages Lustre: 2963:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000404:0x47b:0x0] with magic=0xbd60bd0 Lustre: 2963:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message Lustre: lustre-MDT0001: trigger partial OI scrub for RPC inconsistency, checking FID [0x240000404:0x4a2:0x0]/0xa): rc = 0 Lustre: dir [0x240000404:0x4a2:0x0] stripe 2 readdir failed: -2, directory is partially accessed! LustreError: 1078:0:(lcommon_cl.c:195:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000403:0x2cc:0x0]: rc = -5 LustreError: 1078:0:(lcommon_cl.c:195:cl_file_inode_init()) Skipped 11 previous similar messages LustreError: 1078:0:(llite_lib.c:3731:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 1078:0:(llite_lib.c:3731:ll_prep_inode()) Skipped 11 previous similar messages LustreError: 29053:0:(lov_object.c:1340:lov_layout_change()) lustre-clilov-ffff88027a688958: cannot apply new layout on [0x280000404:0x42f:0x0] : rc = -5 LustreError: 29053:0:(vvp_io.c:1921:vvp_io_init()) lustre: refresh file layout [0x280000404:0x42f:0x0] error -5. 15[7719]: segfault at 0 ip (null) sp 00007ffd8dc53428 error 14 in 15[400000+6000] Lustre: 19525:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000404:0x4dc:0x0] with magic=0xbd60bd0 Lustre: 19525:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 9 previous similar messages LustreError: 12485:0:(lov_object.c:1340:lov_layout_change()) lustre-clilov-ffff88027a688958: cannot apply new layout on [0x280000404:0x42f:0x0] : rc = -5 LustreError: 1576:0:(statahead.c:830:ll_statahead_interpret_work()) lustre: getattr callback for 18 [0x240000404:0x4a5:0x0]: rc = -5 LustreError: 20187:0:(lov_object.c:1340:lov_layout_change()) lustre-clilov-ffff88027a688958: cannot apply new layout on [0x280000404:0x42f:0x0] : rc = -5 LustreError: 13725:0:(mdd_dir.c:4457:mdd_migrate_cmd_check()) lustre-MDD0001: '1' migration was interrupted, run 'lfs migrate -m 2 -c 1 -H crush 1' to finish migration: rc = -1 LustreError: 8694:0:(file.c:262:ll_close_inode_openhandle()) lustre-clilmv-ffff8801e178b7e8: inode [0x240000404:0x392:0x0] mdc close failed: rc = -2 LustreError: 8694:0:(file.c:262:ll_close_inode_openhandle()) Skipped 12 previous similar messages Lustre: 29395:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000404:0x6fc:0x0] with magic=0xbd60bd0 Lustre: 29395:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message Lustre: mdt03_014: service thread pid 2946 was inactive for 40.079 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 2946, comm: mdt03_014 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x913/0xd50 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x890 [ptlrpc] [<0>] mdt_object_pdo_lock+0x4d9/0x7e0 [mdt] [<0>] mdt_parent_lock+0x76/0x2a0 [mdt] [<0>] mdt_reint_open+0x125d/0x2df0 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x74c/0xbc0 [mdt] [<0>] mdt_intent_open+0x93/0x480 [mdt] [<0>] mdt_intent_opc.constprop.73+0x211/0xc50 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x34f/0x930 [ptlrpc] [<0>] ldlm_handle_enqueue+0x507/0x1850 [ptlrpc] [<0>] tgt_enqueue+0x68/0x240 [ptlrpc] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc7e/0x1690 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: mdt01_009: service thread pid 4154 was inactive for 40.119 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 4154, comm: mdt01_009 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x913/0xd50 [ptlrpc] [<0>] ldlm_cli_enqueue_fini+0xa24/0xf20 [ptlrpc] [<0>] ldlm_cli_enqueue+0x461/0xa70 [ptlrpc] [<0>] osp_md_object_lock+0x151/0x2f0 [osp] [<0>] lod_object_lock+0xdb/0x7d0 [lod] [<0>] mdd_object_lock+0x2d/0xd0 [mdd] [<0>] mdt_remote_object_lock_try+0x14c/0x189 [mdt] [<0>] mdt_object_lock_internal+0x3c4/0x470 [mdt] [<0>] mdt_object_check_lock+0x332/0x3c0 [mdt] [<0>] mdt_object_stripes_lock+0xba/0x660 [mdt] [<0>] mdt_reint_unlink+0x7aa/0x15e0 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x74c/0xbc0 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc7e/0x1690 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: mdt05_008: service thread pid 1314 was inactive for 72.155 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 1314, comm: mdt05_008 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x913/0xd50 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x890 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock+0x88/0x1c0 [mdt] [<0>] mdt_reint_setattr+0x1324/0x15f0 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x74c/0xbc0 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc7e/0x1690 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: mdt_io00_005: service thread pid 18038 was inactive for 74.017 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: mdt_io02_005: service thread pid 14462 was inactive for 74.224 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 1 previous similar message Lustre: mdt_io01_004: service thread pid 10662 was inactive for 74.089 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 1 previous similar message Lustre: mdt_io04_002: service thread pid 19529 was inactive for 76.109 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 6 previous similar messages Lustre: mdt_io02_004: service thread pid 13825 was inactive for 74.280 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 6 previous similar messages LustreError: 19365:0:(ldlm_lockd.c:241:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0001_UUID lock: ffff880280d3e940/0xa91f70f6d520885d lrc: 3/0,0 mode: PW/PW res: [0x240000404:0x637:0x0].0x0 bits 0x4/0x0 rrc: 6 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0xa91f70f6d5208848 expref: 333 pid: 19455 timeout: 29860 lvb_type: 0 Lustre: mdt05_008: service thread pid 1314 completed after 100.093s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt00_002: service thread pid 19455 completed after 100.166s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_005: service thread pid 18038 completed after 100.163s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: lustre-MDT0001-mdc-ffff88027a688958: operation mds_reint to node 0@lo failed: rc = -107 Lustre: lustre-MDT0001-mdc-ffff88027a688958: Connection to lustre-MDT0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: lustre-MDT0001-mdc-ffff88027a688958: This client was evicted by lustre-MDT0001; in progress operations using this service will fail. Lustre: mdt05_013: service thread pid 9391 completed after 95.745s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt03_014: service thread pid 2946 completed after 98.230s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 27403:0:(llite_lib.c:2018:ll_md_setattr()) md_setattr fails: rc = -108 LustreError: 19030:0:(mdc_request.c:1454:mdc_read_page()) lustre-MDT0001-mdc-ffff88027a688958: [0x240000400:0x30:0x0] lock enqueue fails: rc = -108 Lustre: dir [0x240000403:0x591:0x0] stripe 0 readdir failed: -108, directory is partially accessed! Lustre: Skipped 3 previous similar messages Lustre: mdt_io07_005: service thread pid 29764 completed after 99.651s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt01_009: service thread pid 4154 completed after 91.202s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io06_002: service thread pid 19535 completed after 99.580s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: lustre-MDT0001-mdc-ffff88027a688958: Connection restored to 192.168.123.97@tcp (at 0@lo) Lustre: mdt_io04_004: service thread pid 8256 completed after 99.678s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io02_005: service thread pid 14462 completed after 99.958s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io01_003: service thread pid 3816 completed after 99.715s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io06_004: service thread pid 5943 completed after 99.432s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_004: service thread pid 18023 completed after 101.361s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io01_004: service thread pid 10662 completed after 101.515s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io04_002: service thread pid 19529 completed after 99.739s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io03_007: service thread pid 4654 completed after 98.799s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io07_003: service thread pid 7073 completed after 98.767s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io01_006: service thread pid 14932 completed after 98.132s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io07_000: service thread pid 19536 completed after 96.468s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_003: service thread pid 12588 completed after 95.683s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io02_003: service thread pid 13725 completed after 95.561s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io02_004: service thread pid 13825 completed after 93.633s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io04_006: service thread pid 8359 completed after 94.937s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 30631:0:(mdd_dir.c:4457:mdd_migrate_cmd_check()) lustre-MDD0001: '4' migration was interrupted, run 'lfs migrate -m 2 -c 2 -H crush 4' to finish migration: rc = -1 touch (15030) used greatest stack depth: 9872 bytes left Lustre: dir [0x200000403:0x673:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 16 previous similar messages LustreError: 12154:0:(mdt_xattr.c:415:mdt_dir_layout_update()) lustre-MDT0001: [0x240000404:0x7cb:0x0] migrate mdt count mismatch 2 != 1 Lustre: 19529:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000403:0x449:0x0] with magic=0xbd60bd0 Lustre: 19529:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 1 previous similar message 12[4918]: segfault at 8 ip 00007f5ffd6c37e8 sp 00007ffe3b39bba0 error 4 in ld-2.17.so[7f5ffd6b8000+22000] 16[6967]: segfault at 8 ip 00007fa8388fb7e8 sp 00007ffe871fc7f0 error 4 in ld-2.17.so[7fa8388f0000+22000] LustreError: 28642:0:(lcommon_cl.c:195:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000403:0x6f7:0x0]: rc = -5 LustreError: 28642:0:(lcommon_cl.c:195:cl_file_inode_init()) Skipped 14 previous similar messages LustreError: 28642:0:(llite_lib.c:3731:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 28642:0:(llite_lib.c:3731:ll_prep_inode()) Skipped 14 previous similar messages LustreError: 19529:0:(mdd_dir.c:4457:mdd_migrate_cmd_check()) lustre-MDD0000: '3' migration was interrupted, run 'lfs migrate -m 1 -c 3 -H crush 3' to finish migration: rc = -1 LustreError: 19529:0:(mdd_dir.c:4457:mdd_migrate_cmd_check()) Skipped 1 previous similar message LustreError: 1880:0:(lov_object.c:1340:lov_layout_change()) lustre-clilov-ffff8801e178b7e8: cannot apply new layout on [0x200000403:0x6f7:0x0] : rc = -5 LustreError: 1880:0:(vvp_io.c:1921:vvp_io_init()) lustre: refresh file layout [0x200000403:0x6f7:0x0] error -5. 7[7978]: segfault at 8 ip 00007f495483f7e8 sp 00007ffd6cdddf40 error 4 in ld-2.17.so[7f4954834000+22000] Lustre: dir [0x200000404:0x777:0x0] stripe 3 readdir failed: -2, directory is partially accessed! LustreError: 28642:0:(lov_object.c:1340:lov_layout_change()) lustre-clilov-ffff8801e178b7e8: cannot apply new layout on [0x200000403:0x6f7:0x0] : rc = -5 LustreError: 2038:0:(statahead.c:830:ll_statahead_interpret_work()) lustre: getattr callback for 6 [0x200000403:0x50f:0x0]: rc = -5 LustreError: 13825:0:(mdd_dir.c:4457:mdd_migrate_cmd_check()) lustre-MDD0002: '3' migration was interrupted, run 'lfs migrate -m 0 -c 2 -H crush 3' to finish migration: rc = -1 LustreError: 16728:0:(lov_object.c:1340:lov_layout_change()) lustre-clilov-ffff8801e178b7e8: cannot apply new layout on [0x200000403:0x6f7:0x0] : rc = -5 Lustre: mdt_io05_000: service thread pid 19530 was inactive for 74.146 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 1 previous similar message LustreError: 19365:0:(ldlm_lockd.c:241:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0000_UUID lock: ffff880221251e40/0xa91f70f6d52e5216 lrc: 3/0,0 mode: PR/PR res: [0x200000403:0x942:0x0].0x0 bits 0x1b/0x0 rrc: 7 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0xa91f70f6d52e51fa expref: 381 pid: 19464 timeout: 30040 lvb_type: 0 Lustre: mdt02_012: service thread pid 11718 completed after 100.230s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 19536:0:(mdt_reint.c:2533:mdt_reint_migrate()) lustre-MDT0000: migrate [0x200000403:0x2:0x0]/15 failed: rc = -114 LustreError: 19536:0:(mdt_reint.c:2533:mdt_reint_migrate()) Skipped 94 previous similar messages LustreError: lustre-MDT0000-mdc-ffff88027a688958: operation mds_close to node 0@lo failed: rc = -107 Lustre: lustre-MDT0000-mdc-ffff88027a688958: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: mdt_io05_000: service thread pid 19530 completed after 103.399s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: lustre-MDT0000-mdc-ffff88027a688958: This client was evicted by lustre-MDT0000; in progress operations using this service will fail. LustreError: 18793:0:(file.c:262:ll_close_inode_openhandle()) lustre-clilmv-ffff88027a688958: inode [0x200000403:0x942:0x0] mdc close failed: rc = -5 LustreError: 18793:0:(file.c:262:ll_close_inode_openhandle()) Skipped 32 previous similar messages LustreError: 18782:0:(file.c:5733:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000403:0x2:0x0] error: rc = -108 Lustre: mdt_io05_001: service thread pid 19531 completed after 102.064s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io05_002: service thread pid 19532 completed after 102.082s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io05_007: service thread pid 17046 completed after 101.812s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: lustre-MDT0000-mdc-ffff88027a688958: Connection restored to 192.168.123.97@tcp (at 0@lo) Lustre: mdt_io05_006: service thread pid 24638 completed after 97.339s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 8850:0:(statahead.c:830:ll_statahead_interpret_work()) lustre: getattr callback for 6 [0x200000403:0x50f:0x0]: rc = -5 LustreError: 29764:0:(mdd_dir.c:4457:mdd_migrate_cmd_check()) lustre-MDD0001: '7' migration was interrupted, run 'lfs migrate -m 2 -c 2 -H crush 7' to finish migration: rc = -1 7[26964]: segfault at 8 ip 00007f9a1badc7e8 sp 00007ffc4f8964d0 error 4 in ld-2.17.so[7f9a1bad1000+22000] Lustre: 8940:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0x8d0:0x0] with magic=0xbd60bd0 Lustre: 8940:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 13 previous similar messages LustreError: 12017:0:(mdt_xattr.c:415:mdt_dir_layout_update()) lustre-MDT0001: [0x240000404:0xa70:0x0] migrate mdt count mismatch 2 != 1 LustreError: 3760:0:(statahead.c:830:ll_statahead_interpret_work()) lustre: getattr callback for 10 [0x200000404:0x2105:0x0]: rc = -5 LustreError: 24406:0:(llite_nfs.c:446:ll_dir_get_parent_fid()) lustre: failure inode [0x280000404:0x9c3:0x0] get parent: rc = -2 Lustre: dir [0x200000403:0x85e:0x0] stripe 2 readdir failed: -2, directory is partially accessed! Lustre: Skipped 20 previous similar messages LustreError: 2064:0:(mdd_dir.c:4457:mdd_migrate_cmd_check()) lustre-MDD0000: '19' migration was interrupted, run 'lfs migrate -m 0 -c 1 -H crush 19' to finish migration: rc = -1 LustreError: 2064:0:(mdd_dir.c:4457:mdd_migrate_cmd_check()) Skipped 1 previous similar message 2[20700]: segfault at 8 ip 00007fdbe05e27e8 sp 00007ffe8db39e40 error 4 in ld-2.17.so[7fdbe05d7000+22000] LustreError: 24857:0:(lcommon_cl.c:195:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000404:0x2522:0x0]: rc = -5 LustreError: 24857:0:(lcommon_cl.c:195:cl_file_inode_init()) Skipped 24 previous similar messages LustreError: 24857:0:(llite_lib.c:3731:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 24857:0:(llite_lib.c:3731:ll_prep_inode()) Skipped 24 previous similar messages LustreError: 2952:0:(mdt_xattr.c:415:mdt_dir_layout_update()) lustre-MDT0001: [0x240000404:0xc32:0x0] migrate mdt count mismatch 3 != 2 0[29687]: segfault at 8 ip 00007f304b9197e8 sp 00007ffeff342880 error 4 in ld-2.17.so[7f304b90e000+22000] Lustre: dir [0x200000404:0x2485:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 10 previous similar messages LustreError: 3885:0:(llite_nfs.c:446:ll_dir_get_parent_fid()) lustre: failure inode [0x280000404:0xbf4:0x0] get parent: rc = -116 LustreError: 3885:0:(llite_nfs.c:446:ll_dir_get_parent_fid()) Skipped 4 previous similar messages Lustre: 12154:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000404:0x275e:0x0] with magic=0xbd60bd0 Lustre: 12154:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 17 previous similar messages LustreError: 5688:0:(llite_lib.c:1867:ll_update_lsm_md()) lustre: [0x200000404:0x277e:0x0] dir layout mismatch: LustreError: 5688:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: refs 215092432 magic=0x1 count=1 index=0 hash=crush:0x2000003 max_inherit=0 max_inherit_rr=0 version=2 migrate_offset=0 migrate_hash=invalid:0 pool= LustreError: 5688:0:(lustre_lmv.h:167:lmv_stripe_object_dump()) stripe[0] [0x200000400:0x52:0x0] LustreError: 5688:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: refs 215092432 magic=0x1 count=2 index=0 hash=crush:0x82000003 max_inherit=0 max_inherit_rr=0 version=1 migrate_offset=1 migrate_hash=fnv_1a_64:2 pool= LustreError: 13368:0:(lov_object.c:1340:lov_layout_change()) lustre-clilov-ffff88027a688958: cannot apply new layout on [0x200000404:0x26e3:0x0] : rc = -5 LustreError: 13368:0:(vvp_io.c:1921:vvp_io_init()) lustre: refresh file layout [0x200000404:0x26e3:0x0] error -5. LustreError: 18415:0:(mdd_dir.c:4457:mdd_migrate_cmd_check()) lustre-MDD0002: '16' migration was interrupted, run 'lfs migrate -m 2 -c 1 -H crush 16' to finish migration: rc = -1 LustreError: 18415:0:(mdd_dir.c:4457:mdd_migrate_cmd_check()) Skipped 2 previous similar messages 17[22532]: segfault at 8 ip 00007ff6230837e8 sp 00007ffcbea25230 error 4 in ld-2.17.so[7ff623078000+22000] ptlrpc_watchdog_fire: 26 callbacks suppressed Lustre: mdt06_016: service thread pid 11165 was inactive for 40.094 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 11165, comm: mdt06_016 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x913/0xd50 [ptlrpc] [<0>] ldlm_cli_enqueue_fini+0xa24/0xf20 [ptlrpc] [<0>] ldlm_cli_enqueue+0x461/0xa70 [ptlrpc] [<0>] osp_md_object_lock+0x151/0x2f0 [osp] [<0>] lod_object_lock+0xdb/0x7d0 [lod] [<0>] mdd_object_lock+0x2d/0xd0 [mdd] [<0>] mdt_remote_object_lock_try+0x14c/0x189 [mdt] [<0>] mdt_object_lock_internal+0x3c4/0x470 [mdt] [<0>] mdt_object_lock+0x88/0x1c0 [mdt] [<0>] mdt_reint_link+0x8ef/0xf10 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x74c/0xbc0 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc7e/0x1690 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 2466, comm: mdt04_008 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x913/0xd50 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x890 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock+0x88/0x1c0 [mdt] [<0>] mdt_getattr_name_lock+0xf7/0x2bd0 [mdt] [<0>] mdt_intent_getattr+0x2cc/0x4e0 [mdt] [<0>] mdt_intent_opc.constprop.73+0x211/0xc50 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x34f/0x930 [ptlrpc] [<0>] ldlm_handle_enqueue+0x507/0x1850 [ptlrpc] [<0>] tgt_enqueue+0x68/0x240 [ptlrpc] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc7e/0x1690 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe LustreError: 22710:0:(lov_object.c:1340:lov_layout_change()) lustre-clilov-ffff88027a688958: cannot apply new layout on [0x200000404:0x26e3:0x0] : rc = -5 LustreError: 22710:0:(lov_object.c:1340:lov_layout_change()) Skipped 1 previous similar message Pid: 19466, comm: mdt04_001 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x913/0xd50 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x890 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock_try+0xa0/0x250 [mdt] [<0>] mdt_getattr_name_lock+0x1699/0x2bd0 [mdt] [<0>] mdt_intent_getattr+0x2cc/0x4e0 [mdt] [<0>] mdt_intent_opc.constprop.73+0x211/0xc50 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x34f/0x930 [ptlrpc] [<0>] ldlm_handle_enqueue+0x507/0x1850 [ptlrpc] [<0>] tgt_enqueue+0x68/0x240 [ptlrpc] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc7e/0x1690 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: lustre-MDT0002: trigger partial OI scrub for RPC inconsistency, checking FID [0x280000403:0xbc4:0x0]/0xa): rc = 0 LustreError: 3760:0:(statahead.c:830:ll_statahead_interpret_work()) lustre: getattr callback for 13 [0x200000404:0x26e3:0x0]: rc = -5 LustreError: 19365:0:(ldlm_lockd.c:241:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0002_UUID lock: ffff88032481b100/0xa91f70f6d541e807 lrc: 3/0,0 mode: CR/CR res: [0x280000403:0xb02:0x0].0x0 bits 0xa/0x0 rrc: 5 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0xa91f70f6d541e7f9 expref: 367 pid: 19466 timeout: 30324 lvb_type: 0 Lustre: mdt06_016: service thread pid 11165 completed after 100.055s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt04_008: service thread pid 2466 completed after 99.946s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: lustre-MDT0002-mdc-ffff8801e178b7e8: operation mds_getxattr to node 0@lo failed: rc = -107 Lustre: lustre-MDT0002-mdc-ffff8801e178b7e8: Connection to lustre-MDT0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: lustre-MDT0002-mdc-ffff8801e178b7e8: This client was evicted by lustre-MDT0002; in progress operations using this service will fail. Lustre: mdt04_001: service thread pid 19466 completed after 98.876s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 14192:0:(llite_lib.c:2018:ll_md_setattr()) md_setattr fails: rc = -5 LustreError: 14192:0:(llite_lib.c:2018:ll_md_setattr()) Skipped 2 previous similar messages LustreError: 12479:0:(file.c:5733:ll_inode_revalidate_fini()) lustre: revalidate FID [0x280000403:0xb02:0x0] error: rc = -108 LustreError: 12479:0:(file.c:5733:ll_inode_revalidate_fini()) Skipped 7 previous similar messages LustreError: 10654:0:(mdc_request.c:1454:mdc_read_page()) lustre-MDT0002-mdc-ffff8801e178b7e8: [0x280000400:0x4c:0x0] lock enqueue fails: rc = -108 LustreError: 10654:0:(mdc_request.c:1454:mdc_read_page()) Skipped 2 previous similar messages Lustre: dir [0x280000404:0xd70:0x0] stripe 0 readdir failed: -108, directory is partially accessed! Lustre: Skipped 39 previous similar messages LustreError: 19469:0:(ldlm_lockd.c:1478:ldlm_handle_enqueue()) ### lock on destroyed export ffff88027a689bf8 ns: mdt-lustre-MDT0002_UUID lock: ffff88013c0fe940/0xa91f70f6d541e92d lrc: 3/0,0 mode: PR/PR res: [0x280000403:0xb02:0x0].0x0 bits 0x1b/0x0 rrc: 3 type: IBT gid 0 flags: 0x50200000000000 nid: 0@lo remote: 0xa91f70f6d541e911 expref: 8 pid: 19469 timeout: 0 lvb_type: 0 Lustre: lustre-MDT0002-mdc-ffff8801e178b7e8: Connection restored to 192.168.123.97@tcp (at 0@lo) Lustre: mdt03_018: service thread pid 12640 was inactive for 40.076 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 5 previous similar messages Lustre: mdt_io00_007: service thread pid 18778 was inactive for 72.019 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. LustreError: 19365:0:(ldlm_lockd.c:241:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0001_UUID lock: ffff880259b81300/0xa91f70f6d54c22fc lrc: 3/0,0 mode: PR/PR res: [0x240000404:0xfcc:0x0].0x0 bits 0x1b/0x0 rrc: 3 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0xa91f70f6d54c22d2 expref: 305 pid: 3006 timeout: 30430 lvb_type: 0 Lustre: mdt03_018: service thread pid 12640 completed after 98.490s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: lustre-MDT0001-mdc-ffff88027a688958: operation mds_close to node 0@lo failed: rc = -107 Lustre: lustre-MDT0001-mdc-ffff88027a688958: Connection to lustre-MDT0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: lustre-MDT0001-mdc-ffff88027a688958: This client was evicted by lustre-MDT0001; in progress operations using this service will fail. LustreError: 14215:0:(llite_lib.c:2018:ll_md_setattr()) md_setattr fails: rc = -108 LustreError: 11507:0:(vvp_io.c:1921:vvp_io_init()) lustre: refresh file layout [0x240000404:0xfcc:0x0] error -108. Lustre: lustre-MDT0001-mdc-ffff88027a688958: Connection restored to 192.168.123.97@tcp (at 0@lo) LustreError: 19487:0:(mdd_object.c:3864:mdd_close()) lustre-MDD0002: failed to get lu_attr of [0x280000403:0x879:0x0]: rc = -2 Lustre: mdt_io00_004: service thread pid 18023 completed after 97.931s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io00_007: service thread pid 18778 completed after 97.989s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 3061:0:(mdt_handler.c:776:mdt_pack_acl2body()) lustre-MDT0002: unable to read [0x280000403:0xb70:0x0] ACL: rc = -2 LustreError: 21348:0:(statahead.c:830:ll_statahead_interpret_work()) lustre: getattr callback for sleep [0x200000404:0x26ca:0x0]: rc = -5 LustreError: 2823:0:(mdt_handler.c:776:mdt_pack_acl2body()) lustre-MDT0001: unable to read [0x240000404:0x1086:0x0] ACL: rc = -2 Lustre: 1297:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000403:0xee9:0x0] with magic=0xbd60bd0 Lustre: 1297:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 55 previous similar messages 7[1547]: segfault at 8 ip 00007f6473aac7e8 sp 00007ffe3850b370 error 4 in ld-2.17.so[7f6473aa1000+22000] 9[30721]: segfault at 8 ip 00007fbfd1a747e8 sp 00007ffeb13be130 error 4 in ld-2.17.so[7fbfd1a69000+22000] 2[6375]: segfault at 8 ip 00007f3bf9c9e7e8 sp 00007ffd33605bb0 error 4 in ld-2.17.so[7f3bf9c93000+22000] LustreError: 8359:0:(mdt_reint.c:2533:mdt_reint_migrate()) lustre-MDT0000: migrate [0x280000405:0x139:0x0]/14 failed: rc = -114 LustreError: 8359:0:(mdt_reint.c:2533:mdt_reint_migrate()) Skipped 289 previous similar messages LustreError: 19477:0:(mdd_object.c:3864:mdd_close()) lustre-MDD0002: failed to get lu_attr of [0x280000405:0xd4:0x0]: rc = -2 LustreError: 19477:0:(mdd_object.c:3864:mdd_close()) Skipped 1 previous similar message LustreError: 30185:0:(file.c:262:ll_close_inode_openhandle()) lustre-clilmv-ffff88027a688958: inode [0x280000405:0xd4:0x0] mdc close failed: rc = -2 LustreError: 30185:0:(file.c:262:ll_close_inode_openhandle()) Skipped 99 previous similar messages LustreError: 30967:0:(mdd_object.c:3864:mdd_close()) lustre-MDD0002: failed to get lu_attr of [0x280000405:0xd4:0x0]: rc = -2 LustreError: 19522:0:(mdd_dir.c:4457:mdd_migrate_cmd_check()) lustre-MDD0000: '8' migration was interrupted, run 'lfs migrate -m 1 -c 3 -H crush 8' to finish migration: rc = -1 LustreError: 19522:0:(mdd_dir.c:4457:mdd_migrate_cmd_check()) Skipped 9 previous similar messages 6[17130]: segfault at 8 ip 00007f999dc367e8 sp 00007ffe2a7ae8f0 error 4 in ld-2.17.so[7f999dc2b000+22000] LustreError: 11418:0:(mdt_open.c:1280:mdt_cross_open()) lustre-MDT0002: [0x280000403:0xff6:0x0] doesn't exist!: rc = -14 LustreError: 29509:0:(mdt_xattr.c:415:mdt_dir_layout_update()) lustre-MDT0001: [0x240000404:0x131a:0x0] migrate mdt count mismatch 1 != 2 Lustre: dir [0x240000404:0x131a:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 21 previous similar messages LustreError: 19365:0:(ldlm_lockd.c:241:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0000_UUID lock: ffff88029a628400/0xa91f70f6d55f48b5 lrc: 3/0,0 mode: PR/PR res: [0x200000405:0xc34:0x0].0x0 bits 0x1b/0x0 rrc: 4 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0xa91f70f6d55f4899 expref: 517 pid: 19476 timeout: 30700 lvb_type: 0 LustreError: lustre-MDT0000-mdc-ffff88027a688958: operation mds_close to node 0@lo failed: rc = -107 Lustre: lustre-MDT0000-mdc-ffff88027a688958: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: lustre-MDT0000-mdc-ffff88027a688958: This client was evicted by lustre-MDT0000; in progress operations using this service will fail. LustreError: 2488:0:(mdc_request.c:1454:mdc_read_page()) lustre-MDT0000-mdc-ffff88027a688958: [0x200000402:0x50:0x0] lock enqueue fails: rc = -108 LustreError: 1085:0:(llite_lib.c:2018:ll_md_setattr()) md_setattr fails: rc = -108 LustreError: 30226:0:(file.c:5733:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000007:0x1:0x0] error: rc = -108 LustreError: 3939:0:(ldlm_resource.c:1149:ldlm_resource_complain()) lustre-MDT0000-mdc-ffff88027a688958: namespace resource [0x200000405:0xc34:0x0].0x0 (ffff88028e60ad40) refcount nonzero (1) after lock cleanup; forcing cleanup. Lustre: lustre-MDT0000-mdc-ffff88027a688958: Connection restored to 192.168.123.97@tcp (at 0@lo) Lustre: 7073:0:(service.c:2368:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (104/4s); client may timeout req@ffff88024ff92d40 x1811500838237824/t4295049253(0) o36->cfdd0331-5323-4375-9e7e-4f2f34ecbe02@0@lo:388/0 lens 688/456 e 1 to 0 dl 1727583083 ref 1 fl Complete:/200/0 rc 0/0 job:'lfs.0' uid:0 gid:0 19[30700]: segfault at 8 ip 00007f7a6c5707e8 sp 00007fffa7cafec0 error 4 in ld-2.17.so[7f7a6c565000+22000] LustreError: 6278:0:(lcommon_cl.c:195:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000404:0x1301:0x0]: rc = -5 LustreError: 6278:0:(lcommon_cl.c:195:cl_file_inode_init()) Skipped 113 previous similar messages LustreError: 6278:0:(llite_lib.c:3731:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 6278:0:(llite_lib.c:3731:ll_prep_inode()) Skipped 113 previous similar messages 5[18991]: segfault at 0 ip 0000000000403e5f sp 00007ffce548a790 error 6 in 5[400000+6000] 5[18999]: segfault at 0 ip 0000000000403e5f sp 00007ffc78ff24f0 error 6 in 5[400000+6000] LustreError: 31783:0:(mdd_object.c:3864:mdd_close()) lustre-MDD0002: failed to get lu_attr of [0x280000403:0x1359:0x0]: rc = -2 LustreError: 6776:0:(mdd_object.c:3864:mdd_close()) lustre-MDD0002: failed to get lu_attr of [0x280000403:0x1359:0x0]: rc = -2 LustreError: 23979:0:(lov_object.c:1340:lov_layout_change()) lustre-clilov-ffff88027a688958: cannot apply new layout on [0x240000404:0x179a:0x0] : rc = -5 LustreError: 23979:0:(lov_object.c:1340:lov_layout_change()) Skipped 8 previous similar messages LustreError: 23979:0:(vvp_io.c:1921:vvp_io_init()) lustre: refresh file layout [0x240000404:0x179a:0x0] error -5. 7[4366]: segfault at 8 ip 00007f56b79317e8 sp 00007ffe327cd9d0 error 4 in ld-2.17.so[7f56b7926000+22000] 1[2058]: segfault at 0 ip (null) sp 00007ffd8e3e0878 error 14 in 1[400000+6000] LustreError: 19480:0:(mdd_object.c:3864:mdd_close()) lustre-MDD0001: failed to get lu_attr of [0x240000404:0x18e6:0x0]: rc = -2 12[13463]: segfault at 8 ip 00007fe9abd427e8 sp 00007ffc9bd4bfb0 error 4 in ld-2.17.so[7fe9abd37000+22000] ptlrpc_watchdog_fire: 3 callbacks suppressed Lustre: mdt05_011: service thread pid 2992 was inactive for 40.118 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Lustre: Skipped 2 previous similar messages Pid: 2992, comm: mdt05_011 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x913/0xd50 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x890 [ptlrpc] [<0>] mdt_object_pdo_lock+0x4d9/0x7e0 [mdt] [<0>] mdt_parent_lock+0x76/0x2a0 [mdt] [<0>] mdt_getattr_name_lock+0x17b4/0x2bd0 [mdt] [<0>] mdt_intent_getattr+0x2cc/0x4e0 [mdt] [<0>] mdt_intent_opc.constprop.73+0x211/0xc50 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x34f/0x930 [ptlrpc] [<0>] ldlm_handle_enqueue+0x507/0x1850 [ptlrpc] [<0>] tgt_enqueue+0x68/0x240 [ptlrpc] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc7e/0x1690 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe LustreError: 19365:0:(ldlm_lockd.c:241:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0001_UUID lock: ffff880190f6a200/0xa91f70f6d5837021 lrc: 3/0,0 mode: PR/PR res: [0x240000406:0xbbd:0x0].0x0 bits 0x1b/0x0 rrc: 3 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0xa91f70f6d5837005 expref: 388 pid: 9376 timeout: 31098 lvb_type: 0 Lustre: mdt05_011: service thread pid 2992 completed after 94.917s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: lustre-MDT0001-mdc-ffff88027a688958: operation mds_close to node 0@lo failed: rc = -107 Lustre: lustre-MDT0001-mdc-ffff88027a688958: Connection to lustre-MDT0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: lustre-MDT0001-mdc-ffff88027a688958: This client was evicted by lustre-MDT0001; in progress operations using this service will fail. LustreError: 9200:0:(llite_lib.c:2018:ll_md_setattr()) md_setattr fails: rc = -108 LustreError: 9200:0:(llite_lib.c:2018:ll_md_setattr()) Skipped 5 previous similar messages Lustre: lustre-MDT0001-mdc-ffff88027a688958: Connection restored to 192.168.123.97@tcp (at 0@lo) LustreError: 8256:0:(mdd_dir.c:4457:mdd_migrate_cmd_check()) lustre-MDD0002: '0' migration was interrupted, run 'lfs migrate -m 2 -c 3 -H crush 0' to finish migration: rc = -1 LustreError: 8256:0:(mdd_dir.c:4457:mdd_migrate_cmd_check()) Skipped 26 previous similar messages Lustre: 29554:0:(lod_lov.c:1418:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000405:0x9a2:0x0] with magic=0xbd60bd0 Lustre: 29554:0:(lod_lov.c:1418:lod_parse_striping()) Skipped 87 previous similar messages Lustre: dir [0x280000403:0x1abc:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 43 previous similar messages 4[31747]: segfault at 8 ip 00007fbc28a6f7e8 sp 00007ffc82e37260 error 4 in ld-2.17.so[7fbc28a64000+22000] LustreError: 2961:0:(mdt_xattr.c:415:mdt_dir_layout_update()) lustre-MDT0002: [0x280000405:0xba9:0x0] migrate mdt count mismatch 1 != 2 LustreError: 8256:0:(mdt_reint.c:2533:mdt_reint_migrate()) lustre-MDT0000: migrate [0x200000403:0x1:0x0]/1 failed: rc = -114 LustreError: 8256:0:(mdt_reint.c:2533:mdt_reint_migrate()) Skipped 311 previous similar messages LustreError: 4221:0:(file.c:262:ll_close_inode_openhandle()) lustre-clilmv-ffff88027a688958: inode [0x240000404:0x1de2:0x0] mdc close failed: rc = -116 LustreError: 4221:0:(file.c:262:ll_close_inode_openhandle()) Skipped 94 previous similar messages LustreError: 2175:0:(llite_lib.c:1867:ll_update_lsm_md()) lustre: [0x280000405:0xc03:0x0] dir layout mismatch: LustreError: 2175:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: refs 215092432 magic=0x4 count=2 index=2 hash=crush:0x2000003 max_inherit=0 max_inherit_rr=0 version=2 migrate_offset=0 migrate_hash=invalid:0 pool= LustreError: 2175:0:(lustre_lmv.h:167:lmv_stripe_object_dump()) stripe[0] [0x280000400:0xa3:0x0] LustreError: 2175:0:(lustre_lmv.h:167:lmv_stripe_object_dump()) Skipped 2 previous similar messages LustreError: 2175:0:(lustre_lmv.h:160:lmv_stripe_object_dump()) dump LMV: refs 215092432 magic=0x1 count=3 index=2 hash=crush:0x82000003 max_inherit=0 max_inherit_rr=0 version=1 migrate_offset=2 migrate_hash=fnv_1a_64:2 pool= LustreError: 4404:0:(lov_object.c:1340:lov_layout_change()) lustre-clilov-ffff8801e178b7e8: cannot apply new layout on [0x280000403:0x1b14:0x0] : rc = -5 LustreError: 4404:0:(lov_object.c:1340:lov_layout_change()) Skipped 1 previous similar message LustreError: 4404:0:(vvp_io.c:1921:vvp_io_init()) lustre: refresh file layout [0x280000403:0x1b14:0x0] error -5. LustreError: 4404:0:(vvp_io.c:1921:vvp_io_init()) Skipped 1 previous similar message LustreError: 19481:0:(mdd_object.c:3864:mdd_close()) lustre-MDD0001: failed to get lu_attr of [0x240000404:0x1b57:0x0]: rc = -2 LustreError: 19481:0:(mdd_object.c:3864:mdd_close()) Skipped 1 previous similar message LustreError: 15169:0:(vvp_io.c:1921:vvp_io_init()) lustre: refresh file layout [0x280000403:0x1d28:0x0] error -5. | Link to test |
racer test 2: racer rename: centos-85.localnet DURATION=2700 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 PGD 25f70e067 PUD 24cdf4067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) osp(OE) ofd(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) crc32_generic crc_t10dif crct10dif_generic crct10dif_common i2c_piix4 virtio_balloon virtio_console pcspkr ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm drm_panel_orientation_quirks ata_piix virtio_blk floppy serio_raw i2c_core libata [last unloaded: libcfs] CPU: 3 PID: 13846 Comm: ll_sa_13786 Kdump: loaded Tainted: P OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.16.0-3.module_el8.7.0+1218+f626c2ff 04/01/2014 task: ffff8802eb725c40 ti: ffff8802046f0000 task.ti: ffff8802046f0000 RIP: 0010:[<ffffffff813f502c>] [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 RSP: 0018:ffff8802046f3d80 EFLAGS: 00010282 RAX: ffff88019f64a808 RBX: 0000000000000008 RCX: 0000000100260023 RDX: 0000000000000026 RSI: ffff88019f64ab38 RDI: 0000000000000008 RBP: ffff8802046f3d90 R08: ffff880121ff3ba8 R09: 0000000000000000 R10: ffff880121ff09f8 R11: ffff880121ff3ba8 R12: ffff88019f64ab38 R13: 0000000000000000 R14: ffff880121ff3ba8 R15: ffff88019f64ab38 FS: 0000000000000000(0000) GS:ffff880331ac0000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 0000000254570000 CR4: 00000000000007e0 Call Trace: [<ffffffffa1562a31>] ll_statahead_thread+0x741/0x2200 [lustre] [<ffffffff817e320e>] ? _raw_spin_lock+0x1e/0x20 [<ffffffff810dbb51>] ? put_prev_entity+0x31/0x400 [<ffffffffa15622f0>] ? ll_statahead_handle.constprop.28+0x1c0/0x1c0 [lustre] [<ffffffff810ba114>] kthread+0xe4/0xf0 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 | Lustre: 19913:0:(mdt_recovery.c:148:mdt_req_from_lrd()) @@@ restoring transno req@ffff8800ae502840 x1811467183297408/t4295100705(0) o101->c18410c5-3a99-4ab9-8efa-c40ccb6a7232@0@lo:617/0 lens 376/7784 e 0 to 0 dl 1727552357 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 Lustre: 19913:0:(mdt_recovery.c:148:mdt_req_from_lrd()) Skipped 1 previous similar message Lustre: 28960:0:(mdt_recovery.c:148:mdt_req_from_lrd()) @@@ restoring transno req@ffff880177442340 x1811467184705280/t4295101226(0) o101->c18410c5-3a99-4ab9-8efa-c40ccb6a7232@0@lo:642/0 lens 376/33712 e 0 to 0 dl 1727552382 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 Lustre: 8301:0:(mdt_recovery.c:148:mdt_req_from_lrd()) @@@ restoring transno req@ffff880282b5a840 x1811467184211712/t4295101227(0) o101->13209b22-8e41-4095-8360-02b9175f2c7e@0@lo:0/0 lens 376/33712 e 0 to 0 dl 1727552495 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 Lustre: 30629:0:(mdt_recovery.c:148:mdt_req_from_lrd()) @@@ restoring transno req@ffff8802cc5b9e40 x1811467186877056/t4295111593(0) o101->13209b22-8e41-4095-8360-02b9175f2c7e@0@lo:655/0 lens 376/41176 e 0 to 0 dl 1727552395 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 Lustre: 30629:0:(mdt_recovery.c:148:mdt_req_from_lrd()) Skipped 1 previous similar message Lustre: 30629:0:(mdt_recovery.c:148:mdt_req_from_lrd()) @@@ restoring transno req@ffff88009a9fc140 x1811467186031616/t4295089690(0) o101->c18410c5-3a99-4ab9-8efa-c40ccb6a7232@0@lo:736/0 lens 376/48128 e 0 to 0 dl 1727552476 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 ptlrpc_watchdog_fire: 16 callbacks suppressed Lustre: mdt03_003: service thread pid 5988 was inactive for 40.051 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Lustre: Skipped 2 previous similar messages Pid: 5988, comm: mdt03_003 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] osp_precreate_reserve+0x58f/0xa80 [osp] [<0>] osp_declare_create+0x18b/0x6a0 [osp] [<0>] lod_sub_declare_create+0xed/0x250 [lod] [<0>] lod_qos_declare_object_on+0x103/0x430 [lod] [<0>] lod_ost_alloc_rr+0xa4c/0x12c0 [lod] [<0>] lod_qos_prep_create+0x168c/0x1dd0 [lod] [<0>] lod_prepare_create+0x241/0x330 [lod] [<0>] lod_declare_striped_create+0xe5/0xad0 [lod] [<0>] lod_declare_xattr_set+0x180/0x2390 [lod] [<0>] dt_declare_xattr_set+0x54/0x160 [mdd] [<0>] mdd_create_data+0x3b1/0x670 [mdd] [<0>] mdt_mfd_open+0xcfc/0xf70 [mdt] [<0>] mdt_finish_open+0x50b/0xa00 [mdt] [<0>] mdt_open_by_fid_lock+0x56b/0xc10 [mdt] [<0>] mdt_reint_open+0x9d8/0x2df0 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x74c/0xbc0 [mdt] [<0>] mdt_intent_open+0x93/0x480 [mdt] [<0>] mdt_intent_opc.constprop.73+0x211/0xc50 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x34f/0x930 [ptlrpc] [<0>] ldlm_handle_enqueue+0x507/0x1850 [ptlrpc] [<0>] tgt_enqueue+0x68/0x240 [ptlrpc] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc7e/0x1690 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: mdt03_003: service thread pid 5988 completed after 51.867s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: 13961:0:(mdt_recovery.c:148:mdt_req_from_lrd()) @@@ restoring transno req@ffff880165907340 x1811467196669824/t4295089800(0) o101->c18410c5-3a99-4ab9-8efa-c40ccb6a7232@0@lo:752/0 lens 376/48128 e 0 to 0 dl 1727552492 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 Lustre: 25286:0:(mdt_recovery.c:148:mdt_req_from_lrd()) @@@ restoring transno req@ffff880273fd1e40 x1811467204684672/t4295112385(0) o101->13209b22-8e41-4095-8360-02b9175f2c7e@0@lo:126/0 lens 376/41176 e 0 to 0 dl 1727552621 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 Lustre: 28404:0:(mdt_recovery.c:148:mdt_req_from_lrd()) @@@ restoring transno req@ffff88032695c140 x1811467207661184/t4295112572(0) o101->13209b22-8e41-4095-8360-02b9175f2c7e@0@lo:87/0 lens 376/47480 e 0 to 0 dl 1727552582 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 Lustre: 28404:0:(mdt_recovery.c:148:mdt_req_from_lrd()) Skipped 2 previous similar messages Lustre: 4243:0:(mdt_recovery.c:148:mdt_req_from_lrd()) @@@ restoring transno req@ffff88001849c140 x1811467216424448/t4295102108(0) o101->13209b22-8e41-4095-8360-02b9175f2c7e@0@lo:183/0 lens 376/48136 e 0 to 0 dl 1727552678 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 Lustre: 4243:0:(mdt_recovery.c:148:mdt_req_from_lrd()) Skipped 1 previous similar message Lustre: 12782:0:(mdt_recovery.c:148:mdt_req_from_lrd()) @@@ restoring transno req@ffff880019ac3c40 x1811467244911232/t4295102803(0) o101->c18410c5-3a99-4ab9-8efa-c40ccb6a7232@0@lo:344/0 lens 376/48136 e 0 to 0 dl 1727552839 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 Lustre: 12782:0:(mdt_recovery.c:148:mdt_req_from_lrd()) Skipped 3 previous similar messages LustreError: 28962:0:(mdt_handler.c:776:mdt_pack_acl2body()) lustre-MDT0001: unable to read [0x240000404:0x32cf:0x0] ACL: rc = -2 Lustre: 8301:0:(mdt_recovery.c:148:mdt_req_from_lrd()) @@@ restoring transno req@ffff88024c96b240 x1811467307884032/t4295104993(0) o101->c18410c5-3a99-4ab9-8efa-c40ccb6a7232@0@lo:547/0 lens 376/48136 e 0 to 0 dl 1727553042 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 Lustre: 8301:0:(mdt_recovery.c:148:mdt_req_from_lrd()) Skipped 7 previous similar messages Lustre: mdt_io00_004: service thread pid 4943 was inactive for 40.029 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 4943, comm: mdt_io00_004 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x913/0xd50 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x890 [ptlrpc] [<0>] mdt_object_pdo_lock+0x4d9/0x7e0 [mdt] [<0>] mdt_parent_lock+0x76/0x2a0 [mdt] [<0>] mdt_lock_two_dirs+0x31/0x1b0 [mdt] [<0>] mdt_reint_rename+0x12be/0x2c30 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x74c/0xbc0 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc7e/0x1690 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: mdt_io00_004: service thread pid 4943 completed after 41.661s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: ll_ost03_016: service thread pid 31000 was inactive for 40.094 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 31000, comm: ll_ost03_016 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] cv_wait_common+0xb0/0x150 [spl] [<0>] __cv_wait_io+0x18/0x20 [spl] [<0>] txg_wait_synced_impl+0xdd/0x130 [zfs] [<0>] txg_wait_synced+0x10/0x50 [zfs] [<0>] osd_trans_stop+0x56d/0x5e0 [osd_zfs] [<0>] ofd_trans_stop+0x22/0x60 [ofd] [<0>] ofd_precreate_objects+0x424/0x2100 [ofd] [<0>] ofd_create_hdl+0x424/0x2400 [ofd] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc7e/0x1690 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: ll_ost03_016: service thread pid 31000 completed after 51.385s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: lustre-OST0001-osc-MDT0001: update sequence from 0x300000401 to 0x300000403 Lustre: lustre-OST0000-osc-MDT0001: update sequence from 0x2c0000401 to 0x2c0000403 Lustre: lustre-OST0002-osc-MDT0001: update sequence from 0x340000401 to 0x340000403 Lustre: 19927:0:(mdt_recovery.c:148:mdt_req_from_lrd()) @@@ restoring transno req@ffff880169c65540 x1811467406899712/t4295118962(0) o101->c18410c5-3a99-4ab9-8efa-c40ccb6a7232@0@lo:409/0 lens 376/48112 e 0 to 0 dl 1727553659 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 Lustre: 19927:0:(mdt_recovery.c:148:mdt_req_from_lrd()) Skipped 27 previous similar messages LustreError: 6397:0:(mdt_handler.c:776:mdt_pack_acl2body()) lustre-MDT0001: unable to read [0x240000404:0x3509:0x0] ACL: rc = -2 Lustre: mdt05_018: service thread pid 17938 was inactive for 134.135 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 17938, comm: mdt05_018 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] osp_precreate_reserve+0x58f/0xa80 [osp] [<0>] osp_declare_create+0x18b/0x6a0 [osp] [<0>] lod_sub_declare_create+0xed/0x250 [lod] [<0>] lod_qos_declare_object_on+0x103/0x430 [lod] [<0>] lod_ost_alloc_rr+0xa4c/0x12c0 [lod] [<0>] lod_qos_prep_create+0x168c/0x1dd0 [lod] [<0>] lod_prepare_create+0x241/0x330 [lod] [<0>] lod_declare_striped_create+0xe5/0xad0 [lod] [<0>] lod_declare_create+0x217/0x680 [lod] [<0>] mdd_declare_create_object_internal+0xd2/0x390 [mdd] [<0>] mdd_declare_create_object.isra.34+0x52/0x8f0 [mdd] [<0>] mdd_declare_create+0x66/0x500 [mdd] [<0>] mdd_create+0x631/0x1bc0 [mdd] [<0>] mdt_reint_open+0x2129/0x2df0 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x74c/0xbc0 [mdt] [<0>] mdt_intent_open+0x93/0x480 [mdt] [<0>] mdt_intent_opc.constprop.73+0x211/0xc50 [mdt] [<0>] mdt_intent_policy+0x10f/0x460 [mdt] [<0>] ldlm_lock_enqueue+0x34f/0x930 [ptlrpc] [<0>] ldlm_handle_enqueue+0x507/0x1850 [ptlrpc] [<0>] tgt_enqueue+0x68/0x240 [ptlrpc] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc7e/0x1690 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: mdt05_018: service thread pid 17938 completed after 148.976s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: lustre-OST0003-osc-MDT0001: update sequence from 0x380000401 to 0x380000403 Lustre: ll_ost01_018: service thread pid 20375 was inactive for 40.110 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 20375, comm: ll_ost01_018 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] cv_wait_common+0xb0/0x150 [spl] [<0>] __cv_wait_io+0x18/0x20 [spl] [<0>] txg_wait_synced_impl+0xdd/0x130 [zfs] [<0>] txg_wait_synced+0x10/0x50 [zfs] [<0>] dmu_tx_wait+0x26d/0x3a0 [zfs] [<0>] dmu_tx_assign+0x91/0x480 [zfs] [<0>] osd_trans_start+0x1a7/0x480 [osd_zfs] [<0>] ofd_precreate_objects+0xa1b/0x2100 [ofd] [<0>] ofd_create_hdl+0x424/0x2400 [ofd] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc7e/0x1690 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: lustre-OST0002-osc-MDT0000: update sequence from 0x340000402 to 0x340000404 Lustre: ll_ost01_018: service thread pid 20375 completed after 51.781s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt_io05_002: service thread pid 29038 was inactive for 40.081 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 29038, comm: mdt_io05_002 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x913/0xd50 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x890 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_check_lock+0xec/0x3c0 [mdt] [<0>] mdt_reint_rename+0x295e/0x2c30 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x74c/0xbc0 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc7e/0x1690 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: mdt_io05_002: service thread pid 29038 completed after 73.488s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: lustre-OST0003-osc-MDT0000: update sequence from 0x380000400 to 0x380000404 Lustre: lustre-OST0003-osc-MDT0002: update sequence from 0x380000402 to 0x380000405 Lustre: lustre-OST0000-osc-MDT0002: update sequence from 0x2c0000400 to 0x2c0000404 Lustre: mdt_io04_006: service thread pid 2112 was inactive for 40.075 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 2 previous similar messages Lustre: mdt_io04_006: service thread pid 2112 completed after 44.291s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: lustre-OST0001-osc-MDT0002: update sequence from 0x300000400 to 0x300000404 Lustre: lustre-OST0002-osc-MDT0002: update sequence from 0x340000400 to 0x340000405 Lustre: lustre-OST0000-osc-MDT0000: update sequence from 0x2c0000402 to 0x2c0000405 Lustre: lustre-OST0001-osc-MDT0000: update sequence from 0x300000402 to 0x300000405 Lustre: 13159:0:(mdt_recovery.c:148:mdt_req_from_lrd()) @@@ restoring transno req@ffff8802858a0040 x1811467508797056/t4295121504(0) o101->c18410c5-3a99-4ab9-8efa-c40ccb6a7232@0@lo:346/0 lens 376/48112 e 0 to 0 dl 1727554351 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 Lustre: 13159:0:(mdt_recovery.c:148:mdt_req_from_lrd()) Skipped 27 previous similar messages LustreError: 13407:0:(mdt_open.c:1280:mdt_cross_open()) lustre-MDT0000: [0x200000404:0x3cfc:0x0] doesn't exist!: rc = -14 LustreError: 13408:0:(mdt_handler.c:776:mdt_pack_acl2body()) lustre-MDT0000: unable to read [0x200000404:0x3d0a:0x0] ACL: rc = -2 Lustre: 28962:0:(mdt_recovery.c:148:mdt_req_from_lrd()) @@@ restoring transno req@ffff88024aaa5540 x1811467609277056/t4295116427(0) o101->13209b22-8e41-4095-8360-02b9175f2c7e@0@lo:132/0 lens 376/48416 e 0 to 0 dl 1727554892 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 Lustre: 28962:0:(mdt_recovery.c:148:mdt_req_from_lrd()) Skipped 24 previous similar messages ptlrpc_watchdog_fire: 1 callbacks suppressed Lustre: mdt_io06_005: service thread pid 30040 was inactive for 40.082 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 30040, comm: mdt_io06_005 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x913/0xd50 [ptlrpc] [<0>] ldlm_cli_enqueue_fini+0xa24/0xf20 [ptlrpc] [<0>] ldlm_cli_enqueue+0x461/0xa70 [ptlrpc] [<0>] osp_md_object_lock+0x151/0x2f0 [osp] [<0>] lod_object_lock+0xdb/0x7d0 [lod] [<0>] mdd_object_lock+0x2d/0xd0 [mdd] [<0>] mdt_remote_object_lock_try+0x14c/0x189 [mdt] [<0>] mdt_object_pdo_lock+0x70e/0x7e0 [mdt] [<0>] mdt_parent_lock+0x76/0x2a0 [mdt] [<0>] mdt_lock_two_dirs+0x31/0x1b0 [mdt] [<0>] mdt_reint_rename+0x12be/0x2c30 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x74c/0xbc0 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x74e/0x1a60 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x281/0xce0 [ptlrpc] [<0>] ptlrpc_main+0xc7e/0x1690 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: mdt_io06_005: service thread pid 30040 completed after 48.161s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 8439:0:(mdt_open.c:1280:mdt_cross_open()) lustre-MDT0001: [0x240000406:0x7ac:0x0] doesn't exist!: rc = -14 | Link to test |
racer test 2: racer rename: oleg229-client.virtnet DURATION=300 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffff813f40ac>] _atomic_dec_and_lock+0xc/0x70 PGD 13419c067 PUD 11e5d4067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) osp(OE) ofd(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) crc_t10dif crct10dif_generic rpcsec_gss_krb5 sb_edac edac_core iosf_mbi kvm_intel kvm irqbypass crc32_pclmul ghash_clmulni_intel squashfs aesni_intel lrw gf128mul glue_helper ablk_helper cryptd i2c_piix4 i2c_core pcspkr binfmt_misc ip_tables ext4 mbcache jbd2 ata_generic pata_acpi crct10dif_pclmul crct10dif_common crc32c_intel ata_piix serio_raw libata CPU: 0 PID: 29691 Comm: ll_sa_29667 Kdump: loaded Tainted: G OE ------------ 3.10.0-7.9-debug #1 Hardware name: QEMU Standard PC (i440FX + PIIX, 1996), BIOS 1.16.2-1.fc38 04/01/2014 task: ffff880090018000 ti: ffff88007f2c0000 task.ti: ffff88007f2c0000 RIP: 0010:[<ffffffff813f40ac>] [<ffffffff813f40ac>] _atomic_dec_and_lock+0xc/0x70 RSP: 0018:ffff88007f2c3d80 EFLAGS: 00010282 RAX: ffff88006c5a38c8 RBX: 0000000000000008 RCX: 000000018010000f RDX: 0000000180100010 RSI: ffff88006c5a3bf8 RDI: 0000000000000008 RBP: ffff88007f2c3d90 R08: ffff8800a7cd7a00 R09: 000000018010000f R10: 0000000000000001 R11: ffff8800a7cd7a00 R12: ffff88006c5a3bf8 R13: 0000000000000000 R14: ffff8800a7cd7a00 R15: ffff88006c5a3bf8 FS: 0000000000000000(0000) GS:ffff88013e200000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 CR2: 0000000000000008 CR3: 000000011e414000 CR4: 0000000000160ff0 Call Trace: [<ffffffffa105cab1>] ll_statahead_thread+0x741/0x2200 [lustre] [<ffffffff810df6e2>] ? pick_next_task_fair+0x702/0x8a0 [<ffffffffa105c370>] ? ll_statahead_handle.constprop.28+0x1c0/0x1c0 [lustre] [<ffffffff810ba114>] kthread+0xe4/0xf0 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817f3e5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 | Link to test | |
racer test 1: racer on clients: centos-105.localnet DURATION=2700 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 PGD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) osp(OE) ofd(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) crc32_generic lnet(OE) libcfs(OE) crc_t10dif crct10dif_generic crct10dif_common virtio_console virtio_balloon pcspkr i2c_piix4 ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm drm_panel_orientation_quirks ata_piix serio_raw floppy virtio_blk libata i2c_core CPU: 4 PID: 2597 Comm: ll_sa_31228 Kdump: loaded Tainted: P OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.16.0-3.module_el8.7.0+1218+f626c2ff 04/01/2014 task: ffff88026a445c40 ti: ffff8800552b0000 task.ti: ffff8800552b0000 RIP: 0010:[<ffffffff813f502c>] [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 RSP: 0018:ffff8800552b3d80 EFLAGS: 00010282 RAX: ffff8801f9d50108 RBX: 0000000000000008 RCX: dead000000000200 RDX: 0000000000000004 RSI: ffff8801f9d50438 RDI: 0000000000000008 RBP: ffff8800552b3d90 R08: ffff880331403660 R09: 0000000000000000 R10: ffff880331403640 R11: ffff880263d52b18 R12: 0000000000000000 R13: ffff8802655549d0 R14: ffff8801f9d50438 R15: ffff880263d52b18 FS: 0000000000000000(0000) GS:ffff880331b00000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 0000000001c10000 CR4: 00000000000007e0 Call Trace: [<ffffffffa1544661>] ll_statahead_thread+0xde1/0x2270 [lustre] [<ffffffff817e320e>] ? _raw_spin_lock+0x1e/0x20 [<ffffffff810dbb51>] ? put_prev_entity+0x31/0x400 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e329e>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffff810c834d>] ? finish_task_switch+0x5d/0x1b0 [<ffffffffa1543880>] ? ll_statahead_handle.constprop.29+0x1c0/0x1c0 [lustre] [<ffffffff810ba114>] kthread+0xe4/0xf0 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 | Lustre: 26486:0:(mdt_recovery.c:148:mdt_req_from_lrd()) @@@ restoring transno req@ffff880262c86e40 x1806339018510080/t4294967582(0) o101->2dee6b44-6f06-4758-a74f-c98b9371b7eb@0@lo:678/0 lens 376/816 e 0 to 0 dl 1722659263 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 Lustre: 25819:0:(lod_lov.c:1438:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000404:0xa:0x0] with magic=0xbd60bd0 LustreError: 13632:0:(mdt_reint.c:2533:mdt_reint_migrate()) lustre-MDT0001: migrate [0x240000403:0x1:0x0]/4 failed: rc = -114 LustreError: 13646:0:(mdt_reint.c:2533:mdt_reint_migrate()) lustre-MDT0000: migrate [0x200000403:0x2:0x0]/14 failed: rc = -16 LustreError: 13646:0:(mdt_reint.c:2533:mdt_reint_migrate()) Skipped 2 previous similar messages LustreError: 28791:0:(file.c:262:ll_close_inode_openhandle()) lustre-clilmv-ffff880280804a88: inode [0x200000403:0x6:0x0] mdc close failed: rc = -116 LustreError: 13626:0:(mdt_reint.c:2533:mdt_reint_migrate()) lustre-MDT0000: migrate [0x200000403:0x2:0x0]/16 failed: rc = -16 LustreError: 13642:0:(mdt_reint.c:2533:mdt_reint_migrate()) lustre-MDT0001: migrate [0x240000403:0x1:0x0]/11 failed: rc = -16 LustreError: 27808:0:(file.c:262:ll_close_inode_openhandle()) lustre-clilmv-ffff880280804a88: inode [0x240000403:0x2f:0x0] mdc close failed: rc = -116 LustreError: 30491:0:(mdt_reint.c:2533:mdt_reint_migrate()) lustre-MDT0002: migrate [0x280000403:0x1:0x0]/3 failed: rc = -16 LustreError: 30491:0:(mdt_reint.c:2533:mdt_reint_migrate()) Skipped 7 previous similar messages LustreError: 13640:0:(mdt_reint.c:2533:mdt_reint_migrate()) lustre-MDT0000: migrate [0x240000403:0x1:0x0]/2 failed: rc = -16 LustreError: 28647:0:(file.c:262:ll_close_inode_openhandle()) lustre-clilmv-ffff880284826678: inode [0x200000403:0x3a:0x0] mdc close failed: rc = -13 LustreError: 30062:0:(mdt_reint.c:2533:mdt_reint_migrate()) lustre-MDT0001: migrate [0x240000403:0x1:0x0]/7 failed: rc = -114 LustreError: 30062:0:(mdt_reint.c:2533:mdt_reint_migrate()) Skipped 4 previous similar messages LustreError: 5348:0:(file.c:262:ll_close_inode_openhandle()) lustre-clilmv-ffff880284826678: inode [0x240000404:0x31:0x0] mdc close failed: rc = -116 LustreError: 5341:0:(file.c:262:ll_close_inode_openhandle()) lustre-clilmv-ffff880280804a88: inode [0x240000404:0x85:0x0] mdc close failed: rc = -116 LustreError: 5341:0:(file.c:262:ll_close_inode_openhandle()) Skipped 1 previous similar message LustreError: 30932:0:(mdt_reint.c:2533:mdt_reint_migrate()) lustre-MDT0001: migrate [0x240000403:0x1:0x0]/15 failed: rc = -16 LustreError: 30932:0:(mdt_reint.c:2533:mdt_reint_migrate()) Skipped 10 previous similar messages LustreError: 14338:0:(file.c:262:ll_close_inode_openhandle()) lustre-clilmv-ffff880280804a88: inode [0x240000403:0x108:0x0] mdc close failed: rc = -116 LustreError: 14338:0:(file.c:262:ll_close_inode_openhandle()) Skipped 3 previous similar messages 10[28573]: segfault at 8 ip 00007fbf5f0c57e8 sp 00007ffc0b9b6f50 error 4 in ld-2.17.so[7fbf5f0ba000+22000] INFO: task mkdir:7453 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mkdir D ffff88027a1a6340 13056 7453 25518 0x00000080 Call Trace: [<ffffffff817e19d9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df7ea>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa1d>] mutex_lock+0x2d/0x40 [<ffffffff81256f65>] filename_create+0x85/0x180 [<ffffffff81257c94>] ? getname_flags+0xc4/0x1a0 [<ffffffff81257f41>] user_path_create+0x41/0x60 [<ffffffff81259286>] SyS_mkdirat+0x46/0xe0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a [<ffffffff81259339>] SyS_mkdir+0x19/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a LustreError: 13635:0:(mdt_reint.c:2533:mdt_reint_migrate()) lustre-MDT0002: migrate [0x280000403:0x1:0x0]/14 failed: rc = -114 LustreError: 13635:0:(mdt_reint.c:2533:mdt_reint_migrate()) Skipped 1 previous similar message LustreError: 20958:0:(file.c:262:ll_close_inode_openhandle()) lustre-clilmv-ffff880284826678: inode [0x280000404:0x21e:0x0] mdc close failed: rc = -116 LustreError: 20958:0:(file.c:262:ll_close_inode_openhandle()) Skipped 1 previous similar message Lustre: 27116:0:(mdt_recovery.c:148:mdt_req_from_lrd()) @@@ restoring transno req@ffff88024c3e1440 x1806339030367488/t4294972354(0) o101->0efc0a80-5612-433e-9b63-c568555382ef@0@lo:109/0 lens 376/864 e 0 to 0 dl 1722659449 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 LustreError: 11592:0:(file.c:262:ll_close_inode_openhandle()) lustre-clilmv-ffff880280804a88: inode [0x240000404:0x24f:0x0] mdc close failed: rc = -116 LustreError: 11592:0:(file.c:262:ll_close_inode_openhandle()) Skipped 4 previous similar messages Lustre: mdt03_005: service thread pid 27880 was inactive for 40.063 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 27880, comm: mdt03_005 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x913/0xd50 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x890 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock_try+0xa0/0x250 [mdt] [<0>] mdt_getattr_name_lock+0x173b/0x2d00 [mdt] [<0>] mdt_intent_getattr+0x2cc/0x4e0 [mdt] [<0>] mdt_intent_opc.constprop.75+0x211/0xc50 [mdt] [<0>] mdt_intent_policy+0x10d/0x470 [mdt] [<0>] ldlm_lock_enqueue+0x34f/0x930 [ptlrpc] [<0>] ldlm_handle_enqueue+0x507/0x1850 [ptlrpc] [<0>] tgt_enqueue+0x68/0x240 [ptlrpc] [<0>] tgt_request_handle+0x74e/0x1a50 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x273/0xcc0 [ptlrpc] [<0>] ptlrpc_main+0xc7e/0x1690 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: mdt00_001: service thread pid 13542 was inactive for 74.188 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 13542, comm: mdt00_001 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x913/0xd50 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x890 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock+0x88/0x1c0 [mdt] [<0>] mdt_object_find_lock+0x54/0x170 [mdt] [<0>] mdt_reint_setxattr+0x133/0x1020 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x74c/0xbc0 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x74e/0x1a50 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x273/0xcc0 [ptlrpc] [<0>] ptlrpc_main+0xc7e/0x1690 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 27126, comm: mdt02_007 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022 Call Trace: [<0>] ldlm_completion_ast+0x913/0xd50 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x890 [ptlrpc] [<0>] mdt_object_lock_internal+0x1b3/0x470 [mdt] [<0>] mdt_object_lock_try+0xa0/0x250 [mdt] [<0>] mdt_getattr_name_lock+0x173b/0x2d00 [mdt] [<0>] mdt_intent_getattr+0x2cc/0x4e0 [mdt] [<0>] mdt_intent_opc.constprop.75+0x211/0xc50 [mdt] [<0>] mdt_intent_policy+0x10d/0x470 [mdt] [<0>] ldlm_lock_enqueue+0x34f/0x930 [ptlrpc] [<0>] ldlm_handle_enqueue+0x507/0x1850 [ptlrpc] [<0>] tgt_enqueue+0x68/0x240 [ptlrpc] [<0>] tgt_request_handle+0x74e/0x1a50 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x273/0xcc0 [ptlrpc] [<0>] ptlrpc_main+0xc7e/0x1690 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe LustreError: 13413:0:(ldlm_lockd.c:261:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0000_UUID lock: ffff8802635d87c0/0x28fef00cb3fde43b lrc: 3/0,0 mode: CR/CR res: [0x200000403:0x507:0x0].0x0 bits 0xa/0x0 rrc: 7 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0x28fef00cb3fde42d expref: 305 pid: 4531 timeout: 517 lvb_type: 0 LustreError: 13638:0:(mdt_reint.c:2533:mdt_reint_migrate()) lustre-MDT0000: migrate [0x200000403:0x1:0x0]/19 failed: rc = -16 LustreError: 13638:0:(mdt_reint.c:2533:mdt_reint_migrate()) Skipped 21 previous similar messages LustreError: lustre-MDT0000-mdc-ffff880284826678: operation mds_close to node 0@lo failed: rc = -107 Lustre: lustre-MDT0000-mdc-ffff880284826678: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 4600:0:(ldlm_lockd.c:1498:ldlm_handle_enqueue()) ### lock on destroyed export ffff880280e0e678 ns: mdt-lustre-MDT0000_UUID lock: ffff88024e6552c0/0x28fef00cb401608a lrc: 3/0,0 mode: PR/PR res: [0x200000403:0x1:0x0].0x0 bits 0x13/0x0 rrc: 19 type: IBT gid 0 flags: 0x50200400000020 nid: 0@lo remote: 0x28fef00cb4016060 expref: 19 pid: 4600 timeout: 0 lvb_type: 0 LustreError: lustre-MDT0000-mdc-ffff880284826678: This client was evicted by lustre-MDT0000; in progress operations using this service will fail. Lustre: mdt03_005: service thread pid 27880 completed after 99.108s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). Lustre: mdt02_007: service thread pid 27126 completed after 99.189s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 24359:0:(file.c:5695:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000403:0x1:0x0] error: rc = -5 LustreError: 24210:0:(file.c:262:ll_close_inode_openhandle()) lustre-clilmv-ffff880284826678: inode [0x200000403:0x551:0x0] mdc close failed: rc = -108 LustreError: 24210:0:(file.c:262:ll_close_inode_openhandle()) Skipped 9 previous similar messages LustreError: 19899:0:(mdc_request.c:1472:mdc_read_page()) lustre-MDT0000-mdc-ffff880284826678: [0x200000402:0xd:0x0] lock enqueue fails: rc = -108 Lustre: dir [0x280000403:0x294:0x0] stripe 1 readdir failed: -108, directory is partially accessed! LustreError: 19933:0:(vvp_io.c:1921:vvp_io_init()) lustre: refresh file layout [0x200000404:0x51e:0x0] error -108. LustreError: 24359:0:(file.c:5695:ll_inode_revalidate_fini()) Skipped 10 previous similar messages Lustre: mdt00_001: service thread pid 13542 completed after 100.110s. This likely indicates the system was overloaded (too many service threads, or not enough hardware resources). LustreError: 27862:0:(ldlm_resource.c:1172:ldlm_resource_complain()) lustre-MDT0000-mdc-ffff880284826678: namespace resource [0x200000403:0x55f:0x0].0x0 (ffff8802686fc040) refcount nonzero (1) after lock cleanup; forcing cleanup. Lustre: lustre-MDT0000-mdc-ffff880284826678: Connection restored to (at 0@lo) LustreError: 13547:0:(mdt_handler.c:777:mdt_pack_acl2body()) lustre-MDT0000: unable to read [0x200000404:0x5be:0x0] ACL: rc = -2 LustreError: 4348:0:(lcommon_cl.c:195:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000403:0x393:0x0]: rc = -5 LustreError: 4348:0:(llite_lib.c:3742:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 13640:0:(mdd_dir.c:4474:mdd_migrate_cmd_check()) lustre-MDD0002: '14' migration was interrupted, run 'lfs migrate -m 1 -c 3 -H crush 14' to finish migration: rc = -1 Lustre: 4526:0:(mdt_recovery.c:148:mdt_req_from_lrd()) @@@ restoring transno req@ffff88009adf6e40 x1806339041914240/t4294977837(0) o101->0efc0a80-5612-433e-9b63-c568555382ef@0@lo:396/0 lens 376/816 e 0 to 0 dl 1722659736 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 LustreError: 13644:0:(lustre_lmv.h:517:lmv_is_sane()) unknown layout LMV: magic=0xcd40cd0 count=4 index=3 hash=crush:0x82000003 version=1 migrate_offset=3 migrate_hash=fnv_1a_64:2 pool= LustreError: 13626:0:(mdd_dir.c:4474:mdd_migrate_cmd_check()) lustre-MDD0001: '0' migration was interrupted, run 'lfs migrate -m 0 -c 3 -H crush 0' to finish migration: rc = -1 Lustre: lustre-MDT0000: trigger partial OI scrub for RPC inconsistency, checking FID [0x200000405:0x54:0x0]/0xa): rc = 0 LustreError: 21283:0:(file.c:262:ll_close_inode_openhandle()) lustre-clilmv-ffff880280804a88: inode [0x200000404:0x717:0x0] mdc close failed: rc = -116 LustreError: 21283:0:(file.c:262:ll_close_inode_openhandle()) Skipped 22 previous similar messages Lustre: dir [0x200000404:0x753:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 4 previous similar messages LustreError: 31816:0:(mdd_dir.c:4474:mdd_migrate_cmd_check()) lustre-MDD0002: '10' migration was interrupted, run 'lfs migrate -m 2 -c 2 -H crush 10' to finish migration: rc = -1 Lustre: dir [0x240000403:0x457:0x0] stripe 2 readdir failed: -2, directory is partially accessed! Lustre: Skipped 1 previous similar message LustreError: 21677:0:(mdc_request.c:1487:mdc_read_page()) lustre-MDT0000-mdc-ffff880284826678: dir page locate: [0x200000405:0x12f:0x0] at 0: rc -5 Lustre: 4532:0:(lod_lov.c:1438:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000404:0x90a:0x0] with magic=0xbd60bd0 Lustre: 4532:0:(lod_lov.c:1438:lod_parse_striping()) Skipped 1 previous similar message LustreError: 30932:0:(mdd_dir.c:4474:mdd_migrate_cmd_check()) lustre-MDD0001: '12' migration was interrupted, run 'lfs migrate -m 2 -c 3 -H crush 12' to finish migration: rc = -1 LustreError: 6510:0:(lcommon_cl.c:195:cl_file_inode_init()) lustre: failed to initialize cl_object [0x280000404:0x5d6:0x0]: rc = -5 LustreError: 6510:0:(llite_lib.c:3742:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 6510:0:(lcommon_cl.c:195:cl_file_inode_init()) lustre: failed to initialize cl_object [0x280000404:0x5b0:0x0]: rc = -5 LustreError: 6510:0:(llite_lib.c:3742:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 13723:0:(lcommon_cl.c:195:cl_file_inode_init()) lustre: failed to initialize cl_object [0x280000404:0x5b2:0x0]: rc = -5 LustreError: 13723:0:(llite_lib.c:3742:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 13723:0:(statahead.c:830:ll_statahead_interpret_work()) lustre: getattr callback for 19 [0x280000404:0x5b2:0x0]: rc = -5 LustreError: 13641:0:(mdt_reint.c:2533:mdt_reint_migrate()) lustre-MDT0000: migrate [0x280000403:0x5bc:0x0]/10 failed: rc = -16 LustreError: 13641:0:(mdt_reint.c:2533:mdt_reint_migrate()) Skipped 101 previous similar messages LustreError: 4762:0:(lcommon_cl.c:195:cl_file_inode_init()) lustre: failed to initialize cl_object [0x280000404:0x5b2:0x0]: rc = -5 LustreError: 4762:0:(lcommon_cl.c:195:cl_file_inode_init()) Skipped 1 previous similar message LustreError: 4762:0:(llite_lib.c:3742:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 4762:0:(llite_lib.c:3742:ll_prep_inode()) Skipped 1 previous similar message Lustre: lustre-MDT0001: trigger partial OI scrub for RPC inconsistency, checking FID [0x240000404:0x571:0x0]/0xa): rc = 0 LustreError: 169:0:(statahead.c:830:ll_statahead_interpret_work()) lustre: getattr callback for 19 [0x0:0x0:0x0]: rc = -5 LustreError: 31816:0:(mdd_dir.c:4474:mdd_migrate_cmd_check()) lustre-MDD0002: '16' migration was interrupted, run 'lfs migrate -m 2 -c 3 -H crush 16' to finish migration: rc = -1 LustreError: 31816:0:(mdd_dir.c:4474:mdd_migrate_cmd_check()) Skipped 1 previous similar message LustreError: 16799:0:(lcommon_cl.c:195:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000403:0x5d2:0x0]: rc = -5 LustreError: 16799:0:(lcommon_cl.c:195:cl_file_inode_init()) Skipped 3 previous similar messages LustreError: 16799:0:(llite_lib.c:3742:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 16799:0:(llite_lib.c:3742:ll_prep_inode()) Skipped 3 previous similar messages Lustre: 5297:0:(lod_lov.c:1438:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000404:0x631:0x0] with magic=0xbd60bd0 Lustre: 5297:0:(lod_lov.c:1438:lod_parse_striping()) Skipped 1 previous similar message LustreError: 21638:0:(lcommon_cl.c:195:cl_file_inode_init()) lustre: failed to initialize cl_object [0x280000403:0x660:0x0]: rc = -5 LustreError: 21638:0:(lcommon_cl.c:195:cl_file_inode_init()) Skipped 4 previous similar messages LustreError: 21638:0:(llite_lib.c:3742:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 21638:0:(llite_lib.c:3742:ll_prep_inode()) Skipped 4 previous similar messages LustreError: 13632:0:(mdd_dir.c:4474:mdd_migrate_cmd_check()) lustre-MDD0000: '10' migration was interrupted, run 'lfs migrate -m 1 -c 2 -H crush 10' to finish migration: rc = -1 LustreError: 22900:0:(lcommon_cl.c:195:cl_file_inode_init()) lustre: failed to initialize cl_object [0x280000403:0x660:0x0]: rc = -5 LustreError: 22900:0:(lcommon_cl.c:195:cl_file_inode_init()) Skipped 2 previous similar messages LustreError: 22900:0:(llite_lib.c:3742:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 22900:0:(llite_lib.c:3742:ll_prep_inode()) Skipped 2 previous similar messages Lustre: dir [0x280000403:0x53f:0x0] stripe 2 readdir failed: -2, directory is partially accessed! Lustre: Skipped 3 previous similar messages LustreError: 7392:0:(mdd_dir.c:4474:mdd_migrate_cmd_check()) lustre-MDD0002: '14' migration was interrupted, run 'lfs migrate -m 0 -c 1 -H crush 14' to finish migration: rc = -1 LustreError: 7392:0:(mdd_dir.c:4474:mdd_migrate_cmd_check()) Skipped 1 previous similar message Lustre: 25029:0:(lod_lov.c:1438:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000403:0x6cb:0x0] with magic=0xbd60bd0 Lustre: 25029:0:(lod_lov.c:1438:lod_parse_striping()) Skipped 1 previous similar message LustreError: 10514:0:(file.c:262:ll_close_inode_openhandle()) lustre-clilmv-ffff880280804a88: inode [0x240000404:0x864:0x0] mdc close failed: rc = -116 LustreError: 10514:0:(file.c:262:ll_close_inode_openhandle()) Skipped 10 previous similar messages LustreError: 11151:0:(lcommon_cl.c:195:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000404:0x7ba:0x0]: rc = -5 LustreError: 11151:0:(lcommon_cl.c:195:cl_file_inode_init()) Skipped 8 previous similar messages LustreError: 11151:0:(llite_lib.c:3742:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 11151:0:(llite_lib.c:3742:ll_prep_inode()) Skipped 8 previous similar messages LustreError: 13640:0:(mdd_dir.c:4474:mdd_migrate_cmd_check()) lustre-MDD0001: '5' migration was interrupted, run 'lfs migrate -m 0 -c 2 -H crush 5' to finish migration: rc = -1 LustreError: 13640:0:(mdd_dir.c:4474:mdd_migrate_cmd_check()) Skipped 2 previous similar messages Lustre: dir [0x200000405:0x659:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 1 previous similar message LustreError: 29069:0:(mdt_xattr.c:415:mdt_dir_layout_update()) lustre-MDT0000: [0x200000404:0xbc7:0x0] migrate mdt count mismatch 2 != 1 LustreError: 327:0:(statahead.c:830:ll_statahead_interpret_work()) lustre: getattr callback for 13 [0x280000404:0x761:0x0]: rc = -5 LustreError: 327:0:(statahead.c:830:ll_statahead_interpret_work()) Skipped 2 previous similar messages LustreError: 13650:0:(mdd_dir.c:4474:mdd_migrate_cmd_check()) lustre-MDD0000: '0' migration was interrupted, run 'lfs migrate -m 1 -c 3 -H crush 0' to finish migration: rc = -1 LustreError: 13650:0:(mdd_dir.c:4474:mdd_migrate_cmd_check()) Skipped 4 previous similar messages Lustre: dir [0x240000403:0x921:0x0] stripe 3 readdir failed: -2, directory is partially accessed! Lustre: 29086:0:(lod_lov.c:1438:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000405:0x829:0x0] with magic=0xbd60bd0 Lustre: 29086:0:(lod_lov.c:1438:lod_parse_striping()) Skipped 1 previous similar message LustreError: 22505:0:(lov_object.c:1358:lov_layout_change()) lustre-clilov-ffff880284826678: cannot apply new layout on [0x240000404:0x96c:0x0] : rc = -5 LustreError: 22505:0:(vvp_io.c:1921:vvp_io_init()) lustre: refresh file layout [0x240000404:0x96c:0x0] error -5. LustreError: 22505:0:(vvp_io.c:1921:vvp_io_init()) Skipped 1 previous similar message 16[29624]: segfault at 8 ip 00007feec63687e8 sp 00007ffeed6cafc0 error 4 in ld-2.17.so[7feec635d000+22000] LustreError: 6243:0:(lcommon_cl.c:195:cl_file_inode_init()) lustre: failed to initialize cl_object [0x280000404:0x86c:0x0]: rc = -5 LustreError: 6243:0:(lcommon_cl.c:195:cl_file_inode_init()) Skipped 20 previous similar messages LustreError: 6243:0:(llite_lib.c:3742:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 6243:0:(llite_lib.c:3742:ll_prep_inode()) Skipped 20 previous similar messages Lustre: 26929:0:(lod_lov.c:1438:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000404:0xaeb:0x0] with magic=0xbd60bd0 Lustre: 26929:0:(lod_lov.c:1438:lod_parse_striping()) Skipped 11 previous similar messages LustreError: 6383:0:(lov_object.c:1358:lov_layout_change()) lustre-clilov-ffff880284826678: cannot apply new layout on [0x280000404:0x86c:0x0] : rc = -5 LustreError: 6383:0:(vvp_io.c:1921:vvp_io_init()) lustre: refresh file layout [0x280000404:0x86c:0x0] error -5. Lustre: dir [0x200000405:0x95d:0x0] stripe 2 readdir failed: -2, directory is partially accessed! Lustre: Skipped 6 previous similar messages Lustre: 13545:0:(lod_lov.c:1438:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000404:0x1022:0x0] with magic=0xbd60bd0 Lustre: 13545:0:(lod_lov.c:1438:lod_parse_striping()) Skipped 3 previous similar messages Lustre: dir [0x240000403:0xa90:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 4 previous similar messages Lustre: 4537:0:(lod_lov.c:1438:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000403:0xb96:0x0] with magic=0xbd60bd0 Lustre: 4537:0:(lod_lov.c:1438:lod_parse_striping()) Skipped 3 previous similar messages LustreError: 10558:0:(mdd_dir.c:4474:mdd_migrate_cmd_check()) lustre-MDD0002: '16' migration was interrupted, run 'lfs migrate -m 0 -c 2 -H crush 16' to finish migration: rc = -1 LustreError: 10558:0:(mdd_dir.c:4474:mdd_migrate_cmd_check()) Skipped 5 previous similar messages Lustre: dir [0x200000404:0x11fb:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 1 previous similar message Lustre: 28825:0:(lod_lov.c:1438:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000403:0xeff:0x0] with magic=0xbd60bd0 Lustre: 28825:0:(lod_lov.c:1438:lod_parse_striping()) Skipped 7 previous similar messages LustreError: 10653:0:(mdt_reint.c:2533:mdt_reint_migrate()) lustre-MDT0001: migrate [0x200000404:0x12fe:0x0]/sleep failed: rc = -114 LustreError: 10653:0:(mdt_reint.c:2533:mdt_reint_migrate()) Skipped 252 previous similar messages LustreError: 4169:0:(lcommon_cl.c:195:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000403:0xed8:0x0]: rc = -5 LustreError: 4169:0:(lcommon_cl.c:195:cl_file_inode_init()) Skipped 56 previous similar messages LustreError: 4169:0:(llite_lib.c:3742:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 4169:0:(llite_lib.c:3742:ll_prep_inode()) Skipped 56 previous similar messages LustreError: 25102:0:(mdd_object.c:3884:mdd_close()) lustre-MDD0002: failed to get lu_attr of [0x280000403:0xf9d:0x0]: rc = -2 Lustre: dir [0x200000404:0x1535:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 10 previous similar messages Lustre: 5365:0:(lod_lov.c:1438:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000404:0xf08:0x0] with magic=0xbd60bd0 Lustre: 5365:0:(lod_lov.c:1438:lod_parse_striping()) Skipped 23 previous similar messages LustreError: 13681:0:(file.c:262:ll_close_inode_openhandle()) lustre-clilmv-ffff880284826678: inode [0x200000404:0x1748:0x0] mdc close failed: rc = -2 LustreError: 13681:0:(file.c:262:ll_close_inode_openhandle()) Skipped 20 previous similar messages LustreError: 26838:0:(mdt_handler.c:777:mdt_pack_acl2body()) lustre-MDT0002: unable to read [0x280000404:0x11c4:0x0] ACL: rc = -2 LustreError: 131:0:(statahead.c:830:ll_statahead_interpret_work()) lustre: getattr callback for 13 [0x200000404:0x1707:0x0]: rc = -5 LustreError: 29525:0:(mdd_dir.c:4474:mdd_migrate_cmd_check()) lustre-MDD0001: '2' migration was interrupted, run 'lfs migrate -m 0 -c 2 -H crush 2' to finish migration: rc = -1 LustreError: 29525:0:(mdd_dir.c:4474:mdd_migrate_cmd_check()) Skipped 10 previous similar messages LustreError: 29111:0:(mdd_object.c:3884:mdd_close()) lustre-MDD0001: failed to get lu_attr of [0x240000404:0xb9e:0x0]: rc = -2 LustreError: 29111:0:(mdd_object.c:3884:mdd_close()) Skipped 1 previous similar message LustreError: 13586:0:(mdd_object.c:3884:mdd_close()) lustre-MDD0001: failed to get lu_attr of [0x240000404:0xb9e:0x0]: rc = -2 LustreError: 19796:0:(statahead.c:830:ll_statahead_interpret_work()) lustre: getattr callback for 9 [0x0:0x0:0x0]: rc = -5 LustreError: 26623:0:(mdt_xattr.c:415:mdt_dir_layout_update()) lustre-MDT0000: [0x200000405:0x10c6:0x0] migrate mdt count mismatch 2 != 3 LustreError: 13588:0:(mdd_object.c:3884:mdd_close()) lustre-MDD0001: failed to get lu_attr of [0x240000404:0x137f:0x0]: rc = -2 LustreError: 26623:0:(mdt_xattr.c:415:mdt_dir_layout_update()) lustre-MDT0002: [0x280000404:0x13ac:0x0] migrate mdt count mismatch 1 != 3 4[32393]: segfault at 8 ip 00007fbeb0e707e8 sp 00007ffdac128580 error 4 in ld-2.17.so[7fbeb0e65000+22000] Lustre: dir [0x280000404:0x13ac:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 20 previous similar messages 3[1781]: segfault at 8 ip 00007f532c47d7e8 sp 00007ffd2ef73b70 error 4 in ld-2.17.so[7f532c472000+22000] LustreError: 10058:0:(statahead.c:830:ll_statahead_interpret_work()) lustre: getattr callback for 17 [0x0:0x0:0x0]: rc = -5 LustreError: 10058:0:(statahead.c:830:ll_statahead_interpret_work()) Skipped 5 previous similar messages Lustre: 23078:0:(lod_lov.c:1438:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000404:0x148c:0x0] with magic=0xbd60bd0 Lustre: 23078:0:(lod_lov.c:1438:lod_parse_striping()) Skipped 57 previous similar messages LustreError: 27701:0:(mdt_xattr.c:415:mdt_dir_layout_update()) lustre-MDT0002: [0x280000403:0x1434:0x0] migrate mdt count mismatch 1 != 2 LustreError: 22426:0:(lov_object.c:1358:lov_layout_change()) lustre-clilov-ffff880280804a88: cannot apply new layout on [0x280000403:0x12a8:0x0] : rc = -5 LustreError: 22426:0:(vvp_io.c:1921:vvp_io_init()) lustre: refresh file layout [0x280000403:0x12a8:0x0] error -5. LustreError: 7944:0:(lov_object.c:1358:lov_layout_change()) lustre-clilov-ffff880280804a88: cannot apply new layout on [0x280000403:0x12a8:0x0] : rc = -5 LustreError: 6501:0:(lov_object.c:1358:lov_layout_change()) lustre-clilov-ffff880280804a88: cannot apply new layout on [0x280000403:0x12a8:0x0] : rc = -5 LustreError: 6501:0:(vvp_io.c:1921:vvp_io_init()) lustre: refresh file layout [0x280000403:0x12a8:0x0] error -5. 19[9439]: segfault at 8 ip 00007f75a3b107e8 sp 00007ffeba2756f0 error 4 in ld-2.17.so[7f75a3b05000+22000] LustreError: 27366:0:(mdt_xattr.c:415:mdt_dir_layout_update()) lustre-MDT0002: [0x280000404:0x17bc:0x0] migrate mdt count mismatch 3 != 1 LustreError: 9031:0:(lov_object.c:1358:lov_layout_change()) lustre-clilov-ffff880280804a88: cannot apply new layout on [0x240000403:0x1805:0x0] : rc = -5 LustreError: 9031:0:(vvp_io.c:1921:vvp_io_init()) lustre: refresh file layout [0x240000403:0x1805:0x0] error -5. LustreError: 131:0:(statahead.c:830:ll_statahead_interpret_work()) lustre: getattr callback for 15 [0x0:0x0:0x0]: rc = -5 LustreError: 131:0:(statahead.c:830:ll_statahead_interpret_work()) Skipped 1 previous similar message LustreError: 11194:0:(lov_object.c:1358:lov_layout_change()) lustre-clilov-ffff880280804a88: cannot apply new layout on [0x240000403:0x1805:0x0] : rc = -5 LustreError: 14911:0:(lcommon_cl.c:195:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000404:0x1dc0:0x0]: rc = -5 LustreError: 14911:0:(lcommon_cl.c:195:cl_file_inode_init()) Skipped 124 previous similar messages LustreError: 14911:0:(llite_lib.c:3742:ll_prep_inode()) lustre: new_inode - fatal error: rc = -5 LustreError: 14911:0:(llite_lib.c:3742:ll_prep_inode()) Skipped 124 previous similar messages LustreError: 66:0:(statahead.c:830:ll_statahead_interpret_work()) lustre: getattr callback for 4 [0x200000404:0x1dc0:0x0]: rc = -5 LustreError: 66:0:(statahead.c:830:ll_statahead_interpret_work()) Skipped 4 previous similar messages LustreError: 14156:0:(mdt_reint.c:2533:mdt_reint_migrate()) lustre-MDT0000: migrate [0x200000403:0x2:0x0]/19 failed: rc = -16 LustreError: 14156:0:(mdt_reint.c:2533:mdt_reint_migrate()) Skipped 299 previous similar messages LustreError: 26906:0:(lov_object.c:1358:lov_layout_change()) lustre-clilov-ffff880280804a88: cannot apply new layout on [0x240000403:0x1805:0x0] : rc = -5 | Link to test |
racer test 2: racer rename: oleg254-client.virtnet DURATION=300 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffff813f40ac>] _atomic_dec_and_lock+0xc/0x70 PGD 134659067 PUD 12fa7e067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) osp(OE) ofd(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) crc_t10dif crct10dif_generic rpcsec_gss_krb5 sb_edac edac_core iosf_mbi kvm_intel kvm irqbypass crc32_pclmul ghash_clmulni_intel squashfs aesni_intel lrw gf128mul glue_helper ablk_helper cryptd i2c_piix4 i2c_core pcspkr binfmt_misc ip_tables ext4 mbcache jbd2 ata_generic pata_acpi crct10dif_pclmul crct10dif_common crc32c_intel serio_raw ata_piix libata CPU: 2 PID: 23571 Comm: ll_sa_23387 Kdump: loaded Tainted: G OE ------------ 3.10.0-7.9-debug #1 Hardware name: QEMU Standard PC (i440FX + PIIX, 1996), BIOS 1.16.2-1.fc38 04/01/2014 task: ffff8800a75b5550 ti: ffff88007b06c000 task.ti: ffff88007b06c000 RIP: 0010:[<ffffffff813f40ac>] [<ffffffff813f40ac>] _atomic_dec_and_lock+0xc/0x70 RSP: 0018:ffff88007b06fd80 EFLAGS: 00010282 RAX: ffff88007d9c08c8 RBX: 0000000000000008 RCX: ffffea0004d42400 RDX: 0000000000000005 RSI: ffff88007d9c0bf8 RDI: 0000000000000008 RBP: ffff88007b06fd90 R08: ffff8801314dd800 R09: 000000018010000f R10: 0000000000000001 R11: ffff8801314dd800 R12: ffff88007d9c0bf8 R13: 0000000000000000 R14: ffff8801314dd800 R15: ffff88007d9c0bf8 FS: 0000000000000000(0000) GS:ffff88013e300000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 CR2: 0000000000000008 CR3: 000000012f124000 CR4: 0000000000160fe0 Call Trace: [<ffffffffa103b301>] ll_statahead_thread+0x741/0x2200 [lustre] [<ffffffff810df6e2>] ? pick_next_task_fair+0x702/0x8a0 [<ffffffffa103abc0>] ? revalidate_statahead_dentry+0x1660/0x1660 [lustre] [<ffffffff810ba114>] kthread+0xe4/0xf0 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817f3e5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 | Link to test | |
racer test 1: racer on clients: oleg160-client.virtnet DURATION=300 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffff813f40ac>] _atomic_dec_and_lock+0xc/0x70 PGD a7660067 PUD a7661067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) osp(OE) ofd(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) crc_t10dif crct10dif_generic rpcsec_gss_krb5 sb_edac edac_core iosf_mbi kvm_intel kvm irqbypass crc32_pclmul ghash_clmulni_intel aesni_intel squashfs lrw gf128mul glue_helper ablk_helper cryptd i2c_piix4 pcspkr i2c_core binfmt_misc ip_tables ext4 mbcache jbd2 ata_generic pata_acpi crct10dif_pclmul crct10dif_common crc32c_intel ata_piix serio_raw libata CPU: 0 PID: 9067 Comm: ll_sa_8984 Kdump: loaded Tainted: G OE ------------ 3.10.0-7.9-debug #1 Hardware name: QEMU Standard PC (i440FX + PIIX, 1996), BIOS 1.16.3-1.fc39 04/01/2014 task: ffff8800a83c6660 ti: ffff8800a6c0c000 task.ti: ffff8800a6c0c000 RIP: 0010:[<ffffffff813f40ac>] [<ffffffff813f40ac>] _atomic_dec_and_lock+0xc/0x70 RSP: 0018:ffff8800a6c0fd80 EFLAGS: 00010282 RAX: ffff8800af4308c8 RBX: 0000000000000008 RCX: ffffea0002db8d00 RDX: 000000000000000b RSI: ffff8800af430bf8 RDI: 0000000000000008 RBP: ffff8800a6c0fd90 R08: ffff8800b7c11a00 R09: 000000018010000f R10: 0000000000000001 R11: ffff8800b7c11a00 R12: ffff8800af430bf8 R13: 0000000000000000 R14: ffff8800af430bf8 R15: ffff8800aaf74c00 FS: 0000000000000000(0000) GS:ffff88013e200000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 CR2: 0000000000000008 CR3: 00000000a76b2000 CR4: 0000000000160ff0 Call Trace: [<ffffffffa105e514>] ll_statahead_thread+0x734/0x21c0 [lustre] [<ffffffff810df6e2>] ? pick_next_task_fair+0x702/0x8a0 [<ffffffff81410a79>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffffa105dde0>] ? revalidate_statahead_dentry+0x1660/0x1660 [lustre] [<ffffffff810ba114>] kthread+0xe4/0xf0 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817f3e5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 | Link to test | |
racer test 1: racer on clients: oleg227-client.virtnet DURATION=300 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffff813f40ac>] _atomic_dec_and_lock+0xc/0x70 PGD a62b4067 PUD a6305067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) osp(OE) ofd(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) crc_t10dif crct10dif_generic rpcsec_gss_krb5 sb_edac edac_core iosf_mbi kvm_intel kvm irqbypass crc32_pclmul ghash_clmulni_intel squashfs aesni_intel lrw gf128mul glue_helper ablk_helper cryptd i2c_piix4 pcspkr i2c_core binfmt_misc ip_tables ext4 mbcache jbd2 ata_generic pata_acpi crct10dif_pclmul crct10dif_common crc32c_intel ata_piix serio_raw libata CPU: 1 PID: 32171 Comm: ll_sa_32114 Kdump: loaded Tainted: G OE ------------ 3.10.0-7.9-debug #1 Hardware name: QEMU Standard PC (i440FX + PIIX, 1996), BIOS 1.16.2-1.fc38 04/01/2014 task: ffff8800a70a2220 ti: ffff8800a433c000 task.ti: ffff8800a433c000 RIP: 0010:[<ffffffff813f40ac>] [<ffffffff813f40ac>] _atomic_dec_and_lock+0xc/0x70 RSP: 0018:ffff8800a433fd80 EFLAGS: 00010282 RAX: ffff880136f910c8 RBX: 0000000000000008 RCX: 000000018010000f RDX: 0000000180100010 RSI: ffff880136f913f8 RDI: 0000000000000008 RBP: ffff8800a433fd90 R08: ffff8800a4fdaa00 R09: 000000018010000f R10: 0000000000000001 R11: ffff8800a4fdaa00 R12: ffff880136f913f8 R13: 0000000000000000 R14: ffff880136f913f8 R15: ffff8800b3d37400 FS: 0000000000000000(0000) GS:ffff88013e280000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 CR2: 0000000000000008 CR3: 00000000a6b6a000 CR4: 0000000000160fe0 Call Trace: [<ffffffffa1078274>] ll_statahead_thread+0x734/0x21c0 [lustre] [<ffffffff810df6e2>] ? pick_next_task_fair+0x702/0x8a0 [<ffffffff81410a79>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffffa1077b40>] ? revalidate_statahead_dentry+0x1660/0x1660 [lustre] [<ffffffff810ba114>] kthread+0xe4/0xf0 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817f3e5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 | random: crng init done LustreError: 19609:0:(file.c:264:ll_close_inode_openhandle()) lustre-clilmv-ffff8800aadc0800: inode [0x200000401:0x331:0x0] mdc close failed: rc = -13 LustreError: 20222:0:(file.c:264:ll_close_inode_openhandle()) lustre-clilmv-ffff8800aadc0800: inode [0x200000401:0x331:0x0] mdc close failed: rc = -13 0[24489]: segfault at 8 ip 00007f2f768317e8 sp 00007ffc31361640 error 4 in ld-2.17.so[7f2f76826000+22000] LustreError: 11067:0:(file.c:264:ll_close_inode_openhandle()) lustre-clilmv-ffff8800a9d7f800: inode [0x200000401:0xcde:0x0] mdc close failed: rc = -13 | Link to test |
racer test 1: racer on clients: oleg236-client.virtnet DURATION=300 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffff813f40ac>] _atomic_dec_and_lock+0xc/0x70 PGD a438e067 PUD 9eabe067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) osp(OE) ofd(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) crc_t10dif crct10dif_generic rpcsec_gss_krb5 sb_edac edac_core iosf_mbi kvm_intel kvm irqbypass crc32_pclmul ghash_clmulni_intel squashfs aesni_intel lrw gf128mul glue_helper ablk_helper cryptd pcspkr i2c_piix4 i2c_core binfmt_misc ip_tables ext4 mbcache jbd2 ata_generic pata_acpi crct10dif_pclmul ata_piix crct10dif_common crc32c_intel libata serio_raw CPU: 2 PID: 4295 Comm: ll_sa_3799 Kdump: loaded Tainted: G OE ------------ 3.10.0-7.9-debug #1 Hardware name: QEMU Standard PC (i440FX + PIIX, 1996), BIOS 1.16.2-1.fc38 04/01/2014 task: ffff88009ba20000 ti: ffff88012dea0000 task.ti: ffff88012dea0000 RIP: 0010:[<ffffffff813f40ac>] [<ffffffff813f40ac>] _atomic_dec_and_lock+0xc/0x70 RSP: 0018:ffff88012dea3d80 EFLAGS: 00010282 RAX: ffff8800b02280c8 RBX: 0000000000000008 RCX: ffffea0004bf4880 RDX: 0000000000000005 RSI: ffff8800b02283f8 RDI: 0000000000000008 RBP: ffff88012dea3d90 R08: ffff88012ea8c000 R09: 000000018010000f R10: 0000000000000001 R11: ffff88012ea8c000 R12: ffff8800b02283f8 R13: 0000000000000000 R14: ffff8800b02283f8 R15: ffff8800ab70a800 FS: 0000000000000000(0000) GS:ffff88013e300000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 CR2: 0000000000000008 CR3: 000000009c8e6000 CR4: 0000000000160fe0 Call Trace: [<ffffffffa105b274>] ll_statahead_thread+0x734/0x21c0 [lustre] [<ffffffff810df6e2>] ? pick_next_task_fair+0x702/0x8a0 [<ffffffff81410a79>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffffa105ab40>] ? revalidate_statahead_dentry+0x1660/0x1660 [lustre] [<ffffffff810ba114>] kthread+0xe4/0xf0 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817f3e5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 | LustreError: 11241:0:(file.c:264:ll_close_inode_openhandle()) lustre-clilmv-ffff8800b5c2a800: inode [0x200000402:0xa:0x0] mdc close failed: rc = -116 random: crng init done LustreError: 11378:0:(file.c:264:ll_close_inode_openhandle()) lustre-clilmv-ffff8800b5c24000: inode [0x240000403:0x16:0x0] mdc close failed: rc = -116 LustreError: 11378:0:(file.c:264:ll_close_inode_openhandle()) Skipped 2 previous similar messages LustreError: 11729:0:(dir.c:829:ll_dir_get_default_layout()) lustre: unknown magic: CD40CD0: rc = -71 15[12697]: segfault at 8 ip 00007fa7b23817e8 sp 00007fff97691cf0 error 4 in ld-2.17.so[7fa7b2376000+22000] LustreError: 12784:0:(file.c:264:ll_close_inode_openhandle()) lustre-clilmv-ffff8800b5c2a800: inode [0x240000403:0xba:0x0] mdc close failed: rc = -116 LustreError: 12784:0:(file.c:264:ll_close_inode_openhandle()) Skipped 1 previous similar message LustreError: 14032:0:(lcommon_cl.c:196:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000403:0xbd:0x0]: rc = -5 LustreError: 14032:0:(llite_lib.c:3693:ll_prep_inode()) new_inode -fatal: rc -5 Lustre: dir [0x200000403:0x11c:0x0] stripe 1 readdir failed: -2, directory is partially accessed! LustreError: 16456:0:(lcommon_cl.c:196:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000403:0x160:0x0]: rc = -5 LustreError: 16456:0:(llite_lib.c:3693:ll_prep_inode()) new_inode -fatal: rc -5 LustreError: 16752:0:(file.c:264:ll_close_inode_openhandle()) lustre-clilmv-ffff8800b5c2a800: inode [0x200000402:0x162:0x0] mdc close failed: rc = -116 LustreError: 16752:0:(file.c:264:ll_close_inode_openhandle()) Skipped 3 previous similar messages Lustre: dir [0x200000403:0x19d:0x0] stripe 2 readdir failed: -2, directory is partially accessed! LustreError: 17267:0:(lcommon_cl.c:196:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000402:0x10f:0x0]: rc = -5 LustreError: 17267:0:(lcommon_cl.c:196:cl_file_inode_init()) Skipped 1 previous similar message LustreError: 17267:0:(llite_lib.c:3693:ll_prep_inode()) new_inode -fatal: rc -5 LustreError: 17267:0:(llite_lib.c:3693:ll_prep_inode()) Skipped 1 previous similar message 1[17694]: segfault at 8 ip 00007f84f5f937e8 sp 00007ffe80f6ddb0 error 4 in ld-2.17.so[7f84f5f88000+22000] LustreError: 20263:0:(file.c:264:ll_close_inode_openhandle()) lustre-clilmv-ffff8800b5c2a800: inode [0x200000402:0x1a3:0x0] mdc close failed: rc = -116 LustreError: 20263:0:(file.c:264:ll_close_inode_openhandle()) Skipped 5 previous similar messages LustreError: 20662:0:(lcommon_cl.c:196:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000403:0x1e6:0x0]: rc = -5 LustreError: 20662:0:(lcommon_cl.c:196:cl_file_inode_init()) Skipped 13 previous similar messages LustreError: 20662:0:(llite_lib.c:3693:ll_prep_inode()) new_inode -fatal: rc -5 LustreError: 20662:0:(llite_lib.c:3693:ll_prep_inode()) Skipped 13 previous similar messages 15[21601]: segfault at 4017b0 ip 00000000004017b0 sp 00007ffd79f34c30 error 14 Lustre: dir [0x240000403:0x3bb:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 4 previous similar messages LustreError: 24495:0:(lcommon_cl.c:196:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000403:0x1e6:0x0]: rc = -5 LustreError: 24495:0:(lcommon_cl.c:196:cl_file_inode_init()) Skipped 13 previous similar messages LustreError: 24495:0:(llite_lib.c:3693:ll_prep_inode()) new_inode -fatal: rc -5 LustreError: 24495:0:(llite_lib.c:3693:ll_prep_inode()) Skipped 13 previous similar messages Lustre: dir [0x200000402:0x4a8:0x0] stripe 2 readdir failed: -2, directory is partially accessed! Lustre: Skipped 4 previous similar messages 9[26843]: segfault at 0 ip (null) sp 00007ffd81d1d578 error 14 in 9[400000+6000] 5[26725]: segfault at 8 ip 00007f3d4796f7e8 sp 00007fffc171c910 error 4 in ld-2.17.so[7f3d47964000+22000] 17[27902]: segfault at 8 ip 00007fdb568037e8 sp 00007ffea8ac5290 error 4 in ld-2.17.so[7fdb567f8000+22000] LustreError: 28541:0:(file.c:264:ll_close_inode_openhandle()) lustre-clilmv-ffff8800b5c24000: inode [0x200000403:0x5bf:0x0] mdc close failed: rc = -116 LustreError: 28541:0:(file.c:264:ll_close_inode_openhandle()) Skipped 18 previous similar messages LustreError: 31147:0:(lcommon_cl.c:196:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000403:0x66d:0x0]: rc = -5 LustreError: 31147:0:(lcommon_cl.c:196:cl_file_inode_init()) Skipped 56 previous similar messages LustreError: 31147:0:(llite_lib.c:3693:ll_prep_inode()) new_inode -fatal: rc -5 LustreError: 31147:0:(llite_lib.c:3693:ll_prep_inode()) Skipped 56 previous similar messages LustreError: 30952:0:(lov_object.c:1360:lov_layout_change()) lustre-clilov-ffff8800b5c2a800: cannot apply new layout on [0x240000403:0x7cd:0x0] : rc = -5 LustreError: 30952:0:(vvp_io.c:1920:vvp_io_init()) lustre: refresh file layout [0x240000403:0x7cd:0x0] error -5. Lustre: dir [0x240000403:0x7f1:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 4 previous similar messages LustreError: 1804:0:(lov_object.c:1360:lov_layout_change()) lustre-clilov-ffff8800b5c2a800: cannot apply new layout on [0x240000403:0x7cd:0x0] : rc = -5 5[1271]: segfault at 8 ip 00007faa30e6c7e8 sp 00007ffe625fb2d0 error 4 in ld-2.17.so[7faa30e61000+22000] | Link to test |
racer test 1: racer on clients: centos-120.localnet DURATION=2700 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 PGD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) osp(OE) ofd(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) crc32_generic crc_t10dif crct10dif_generic crct10dif_common virtio_balloon virtio_console i2c_piix4 pcspkr ip_tables rpcsec_gss_krb5 drm_kms_helper ata_generic ttm pata_acpi drm ata_piix drm_panel_orientation_quirks floppy serio_raw virtio_blk libata i2c_core [last unloaded: libcfs] CPU: 8 PID: 28767 Comm: ll_sa_27944 Kdump: loaded Tainted: P OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.16.0-3.module_el8.7.0+1218+f626c2ff 04/01/2014 task: ffff8801fc873760 ti: ffff8800b3a94000 task.ti: ffff8800b3a94000 RIP: 0010:[<ffffffff813f502c>] [<ffffffff813f502c>] _atomic_dec_and_lock+0xc/0x70 RSP: 0018:ffff8800b3a97d80 EFLAGS: 00010282 RAX: ffff880097a1e988 RBX: 0000000000000008 RCX: 0000000100260021 RDX: 0000000000000026 RSI: ffff880097a1ecb8 RDI: 0000000000000008 RBP: ffff8800b3a97d90 R08: ffff88028cb13ef8 R09: 0000000000000000 R10: ffff88028cb14248 R11: ffff88028cb13ef8 R12: ffff880097a1ecb8 R13: 0000000000000000 R14: ffff880097a1ecb8 R15: ffff8802cb2f5a58 FS: 0000000000000000(0000) GS:ffff880331c00000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 0000000001c10000 CR4: 00000000000007e0 Call Trace: [<ffffffffa1588314>] ll_statahead_thread+0x734/0x21c0 [lustre] [<ffffffff810dbb51>] ? put_prev_entity+0x31/0x400 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffffa1587be0>] ? revalidate_statahead_dentry+0x1660/0x1660 [lustre] [<ffffffff810ba114>] kthread+0xe4/0xf0 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 | Lustre: 29104:0:(mdt_recovery.c:148:mdt_req_from_lrd()) @@@ restoring transno req@ffff8800893f1440 x1798499710911936/t4294972000(0) o101->25cdf247-10ca-4900-8ea4-956f33c93a7d@0@lo:535/0 lens 376/840 e 0 to 0 dl 1715183110 ref 1 fl Interpret:H/202/0 rc 0/0 job:'dd.0' uid:0 gid:0 9[14193]: segfault at 8 ip 00007fd38cbff7e8 sp 00007fff7b9ac9b0 error 4 in ld-2.17.so[7fd38cbf4000+22000] 19[19618]: segfault at 8 ip 00007fe7de4807e8 sp 00007ffcbc74db00 error 4 in ld-2.17.so[7fe7de475000+22000] 7[22922]: segfault at 8 ip 00007f3e1d78b7e8 sp 00007ffcf42e6ec0 error 4 in ld-2.17.so[7f3e1d780000+22000] | Link to test |
racer test 1: racer on clients: oleg407-client.virtnet DURATION=300 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffff813f40ac>] _atomic_dec_and_lock+0xc/0x70 PGD 12bf8e067 PUD 12bfef067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) crc32_generic libcfs(OE) crc_t10dif crct10dif_generic crct10dif_common rpcsec_gss_krb5 squashfs i2c_piix4 i2c_core pcspkr binfmt_misc ip_tables ext4 mbcache jbd2 ata_generic pata_acpi ata_piix serio_raw libata CPU: 1 PID: 9736 Comm: ll_sa_9562 Kdump: loaded Tainted: G OE ------------ 3.10.0-7.9-debug #1 Hardware name: QEMU Standard PC (i440FX + PIIX, 1996), BIOS 1.16.2-1.fc38 04/01/2014 task: ffff88012e292220 ti: ffff8800a6fc4000 task.ti: ffff8800a6fc4000 RIP: 0010:[<ffffffff813f40ac>] [<ffffffff813f40ac>] _atomic_dec_and_lock+0xc/0x70 RSP: 0018:ffff8800a6fc7d78 EFLAGS: 00010296 RAX: ffff8800b0a900c8 RBX: 0000000000000008 RCX: 000000018010000e RDX: 000000018010000f RSI: ffff8800b0a903f8 RDI: 0000000000000008 RBP: ffff8800a6fc7d88 R08: ffff88012d4b1600 R09: 000000018010000e R10: 000000002d4b1401 R11: ffff88012d4b1600 R12: 0000000000000000 R13: ffff88012d4b1600 R14: ffff8800b0a903f8 R15: ffff88012d4b16a8 FS: 0000000000000000(0000) GS:ffff88013e280000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 CR2: 0000000000000008 CR3: 000000012e20e000 CR4: 00000000000006e0 Call Trace: [<ffffffffa0ec7ff1>] ll_statahead_thread+0x5b1/0x21d0 [lustre] [<ffffffff810df6e2>] ? pick_next_task_fair+0x702/0x8a0 [<ffffffff817e8dce>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffff810c834d>] ? finish_task_switch+0x5d/0x1b0 [<ffffffffa0ec7a40>] ? ll_statahead_handle.constprop.29+0x1c0/0x1c0 [lustre] [<ffffffff810ba114>] kthread+0xe4/0xf0 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817f3e5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba030>] ? kthread_create_on_node+0x140/0x140 | Link to test |