Editing crashreport #67398

ReasonCrashing FunctionWhere to cut BacktraceReports Count
LBUGllog_osd_write_recllog_osd_write_rec
llog_write_rec
llog_cat_add_rec
llog_add
sub_updates_write
top_trans_stop
lod_trans_stop
mdd_trans_stop
mdd_xattr_del
mdt_reint_setxattr
mdt_reint_rec
mdt_reint_internal
mdt_reint
tgt_request_handle
ptlrpc_server_handle_request
ptlrpc_main
kthread
1

Added fields:

Match messages in logs
(every line would be required to be present in log output
Copy from "Messages before crash" column below):
Match messages in full crash
(every line would be required to be present in crash log output
Copy from "Full Crash" column below):
Limit to a test:
(Copy from below "Failing text"):
Delete these reports as invalid (real bug in review or some such)
Bug or comment:
Extra info:

Failures list (last 100):

Failing TestFull CrashMessages before crashComment
racer test 1: racer on clients: centos-105.localnet DURATION=2700
LustreError: 11715:0:(llog_osd.c:628:llog_osd_write_rec()) LBUG
Pid: 11715, comm: mdt07_013 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022
Call Trace:
[<0>] libcfs_call_trace+0x90/0xf0 [libcfs]
[<0>] lbug_with_loc+0x4c/0xa0 [libcfs]
[<0>] llog_osd_write_rec+0x1a85/0x1ac0 [obdclass]
[<0>] llog_write_rec+0x290/0x590 [obdclass]
[<0>] llog_cat_add_rec+0x1e1/0x990 [obdclass]
[<0>] llog_add+0x17f/0x1f0 [obdclass]
[<0>] sub_updates_write+0x303/0xe32 [ptlrpc]
[<0>] top_trans_stop+0x4a2/0xfb0 [ptlrpc]
[<0>] lod_trans_stop+0x25c/0x340 [lod]
[<0>] mdd_trans_stop+0x28/0x16e [mdd]
[<0>] mdd_xattr_del+0x239/0x670 [mdd]
[<0>] mdt_reint_setxattr+0xc9f/0x1210 [mdt]
[<0>] mdt_reint_rec+0x87/0x240 [mdt]
[<0>] mdt_reint_internal+0x76c/0xb50 [mdt]
[<0>] mdt_reint+0x67/0x150 [mdt]
[<0>] tgt_request_handle+0x93a/0x19c0 [ptlrpc]
[<0>] ptlrpc_server_handle_request+0x250/0xc30 [ptlrpc]
[<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc]
[<0>] kthread+0xe4/0xf0
[<0>] ret_from_fork_nospec_begin+0x7/0x21
[<0>] 0xfffffffffffffffe
LustreError: 11070:0:(file.c:243:ll_close_inode_openhandle()) lustre-clilmv-ffff88022f5bae98: inode [0x240000403:0x17:0x0] mdc close failed: rc = -116
Lustre: 7185:0:(mdt_recovery.c:200:mdt_req_from_lrd()) @@@ restoring transno req@ffff8802602c70c0 x1729879002943040/t4294967874(0) o101->25b6622e-f7b0-41a7-afdd-bcb2a74f3477@0@lo:643/0 lens 376/840 e 0 to 0 dl 1649741328 ref 1 fl Interpret:H/2/0 rc 0/0 job:'dd.0'
LustreError: 11160:0:(file.c:243:ll_close_inode_openhandle()) lustre-clilmv-ffff88028d824a88: inode [0x240000404:0x53:0x0] mdc close failed: rc = -116
LustreError: 11160:0:(file.c:243:ll_close_inode_openhandle()) Skipped 1 previous similar message
Lustre: 31961:0:(mdt_recovery.c:200:mdt_req_from_lrd()) @@@ restoring transno req@ffff88005d4e0cc0 x1729879003102208/t4294967809(0) o101->e69a90e6-91b0-4a42-8c91-e9bbac9fbe7a@0@lo:649/0 lens 376/816 e 0 to 0 dl 1649741334 ref 1 fl Interpret:H/2/0 rc 0/0 job:'dd.0'
LustreError: 14561:0:(file.c:243:ll_close_inode_openhandle()) lustre-clilmv-ffff88022f5bae98: inode [0x240000404:0x1:0x0] mdc close failed: rc = -116
LustreError: 31968:0:(mdd_object.c:3465:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000403:0x42:0x0]: rc = -2
LustreError: 13262:0:(file.c:243:ll_close_inode_openhandle()) lustre-clilmv-ffff88028d824a88: inode [0x200000403:0x42:0x0] mdc close failed: rc = -2
Lustre: 9912:0:(mdt_recovery.c:200:mdt_req_from_lrd()) @@@ restoring transno req@ffff8802e8933240 x1729879003796224/t4294968932(0) o101->25b6622e-f7b0-41a7-afdd-bcb2a74f3477@0@lo:663/0 lens 376/840 e 0 to 0 dl 1649741348 ref 1 fl Interpret:H/2/0 rc 0/0 job:'dd.0'
LustreError: 21358:0:(file.c:243:ll_close_inode_openhandle()) lustre-clilmv-ffff88022f5bae98: inode [0x240000403:0xbb:0x0] mdc close failed: rc = -116
Lustre: mdt00_006: service thread pid 8760 was inactive for 40.009 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes:
Pid: 8760, comm: mdt00_006 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022
Call Trace:
[<0>] ldlm_completion_ast+0x7f7/0xa50 [ptlrpc]
[<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc]
[<0>] mdt_object_local_lock+0x527/0xb90 [mdt]
[<0>] mdt_object_lock_internal+0x70/0x390 [mdt]
[<0>] mdt_getattr_name_lock+0x12e/0x2c50 [mdt]
[<0>] mdt_intent_getattr+0x2d5/0x4b0 [mdt]
[<0>] mdt_intent_opc+0x1dd/0xc10 [mdt]
[<0>] mdt_intent_policy+0x1a1/0x360 [mdt]
[<0>] ldlm_lock_enqueue+0x3c2/0xb40 [ptlrpc]
[<0>] ldlm_handle_enqueue0+0x8c6/0x1780 [ptlrpc]
[<0>] tgt_enqueue+0x64/0x240 [ptlrpc]
[<0>] tgt_request_handle+0x93a/0x19c0 [ptlrpc]
[<0>] ptlrpc_server_handle_request+0x250/0xc30 [ptlrpc]
[<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc]
[<0>] kthread+0xe4/0xf0
[<0>] ret_from_fork_nospec_begin+0x7/0x21
[<0>] 0xfffffffffffffffe
LustreError: 9379:0:(mdt_handler.c:754:mdt_pack_acl2body()) lustre-MDT0001: unable to read [0x240000404:0xee:0x0] ACL: rc = -2
LustreError: 23761:0:(file.c:243:ll_close_inode_openhandle()) lustre-clilmv-ffff88022f5bae98: inode [0x240000404:0xad:0x0] mdc close failed: rc = -116
LustreError: 9978:0:(mdd_dir.c:4213:mdd_migrate_cmd_check()) lustre-MDD0000: '9' migration was interrupted, run 'lfs migrate -m 1 -c 1 -H crush 9' to finish migration.
Lustre: dir [0x240000403:0xb2:0x0] stripe 1 readdir failed: -2, directory is partially accessed!
Lustre: mdt07_004: service thread pid 8633 was inactive for 62.067 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes:
Pid: 8633, comm: mdt07_004 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022
Call Trace:
[<0>] ldlm_completion_ast+0x7f7/0xa50 [ptlrpc]
[<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc]
[<0>] mdt_object_local_lock+0x527/0xb90 [mdt]
[<0>] mdt_object_lock_internal+0x70/0x390 [mdt]
[<0>] mdt_object_find_lock+0x6a/0x1a0 [mdt]
[<0>] mdt_reint_setxattr+0x342/0x1210 [mdt]
[<0>] mdt_reint_rec+0x87/0x240 [mdt]
[<0>] mdt_reint_internal+0x76c/0xb50 [mdt]
[<0>] mdt_reint+0x67/0x150 [mdt]
[<0>] tgt_request_handle+0x93a/0x19c0 [ptlrpc]
[<0>] ptlrpc_server_handle_request+0x250/0xc30 [ptlrpc]
[<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc]
[<0>] kthread+0xe4/0xf0
[<0>] ret_from_fork_nospec_begin+0x7/0x21
[<0>] 0xfffffffffffffffe
Pid: 31950, comm: mdt04_000 3.10.0-7.9-debug #2 SMP Tue Feb 1 18:17:58 EST 2022
Call Trace:
[<0>] ldlm_completion_ast+0x7f7/0xa50 [ptlrpc]
[<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc]
[<0>] mdt_object_local_lock+0x527/0xb90 [mdt]
[<0>] mdt_object_lock_internal+0x70/0x390 [mdt]
[<0>] mdt_getattr_name_lock+0xdb6/0x2c50 [mdt]
[<0>] mdt_intent_getattr+0x2d5/0x4b0 [mdt]
[<0>] mdt_intent_opc+0x1dd/0xc10 [mdt]
[<0>] mdt_intent_policy+0x1a1/0x360 [mdt]
[<0>] ldlm_lock_enqueue+0x3c2/0xb40 [ptlrpc]
[<0>] ldlm_handle_enqueue0+0x8c6/0x1780 [ptlrpc]
[<0>] tgt_enqueue+0x64/0x240 [ptlrpc]
[<0>] tgt_request_handle+0x93a/0x19c0 [ptlrpc]
[<0>] ptlrpc_server_handle_request+0x250/0xc30 [ptlrpc]
[<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc]
[<0>] kthread+0xe4/0xf0
[<0>] ret_from_fork_nospec_begin+0x7/0x21
[<0>] 0xfffffffffffffffe
Lustre: mdt01_006: service thread pid 9030 was inactive for 62.157 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one.
Lustre: mdt05_005: service thread pid 9053 was inactive for 62.075 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one.
Lustre: Skipped 3 previous similar messages
LustreError: 551:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0000-osp-MDT0001: fail to cancel 1 llog-records: rc = -116
LustreError: 551:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0000-osp-MDT0001: fail to cancel 1 of 1 llog-records: rc = -116
LustreError: 8699:0:(llog_cat.c:602:llog_cat_add_rec()) llog_write_rec -116: lh=ffff88025f00b858
LustreError: 8699:0:(update_trans.c:1062:top_trans_stop()) lustre-MDT0000-osp-MDT0001: write updates failed: rc = -116
LustreError: 21560:0:(llite_lib.c:1871:ll_md_setattr()) md_setattr fails: rc = -116
LustreError: 551:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0000-osp-MDT0001: fail to cancel 1 llog-records: rc = -2
LustreError: 551:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) Skipped 1 previous similar message
LustreError: 551:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0000-osp-MDT0001: fail to cancel 1 of 1 llog-records: rc = -2
LustreError: 551:0:(llog_cat.c:787:llog_cat_cancel_records()) Skipped 1 previous similar message
LustreError: 7885:0:(file.c:243:ll_close_inode_openhandle()) lustre-clilmv-ffff88028d824a88: inode [0x200000403:0x119:0x0] mdc close failed: rc = -116
Lustre: dir [0x240000403:0x24b:0x0] stripe 1 readdir failed: -2, directory is partially accessed!
LustreError: 31846:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0000_UUID lock: ffff880243dd9e40/0xb56a1a02826c6563 lrc: 3/0,0 mode: PR/PR res: [0x200000403:0x6:0x0].0x0 bits 0x12/0x0 rrc: 16 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0xb56a1a02826c613b expref: 39 pid: 31938 timeout: 18156 lvb_type: 0
LustreError: 31846:0:(mdd_object.c:3465:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000403:0x42:0x0]: rc = -2
LustreError: 31938:0:(ldlm_lockd.c:1425:ldlm_handle_enqueue0()) ### lock on destroyed export ffff88025838c138 ns: mdt-lustre-MDT0000_UUID lock: ffff8800a783bc40/0xb56a1a02826c8839 lrc: 3/0,0 mode: PR/PR res: [0x200000403:0x6:0x0].0x0 bits 0x12/0x0 rrc: 13 type: IBT gid 0 flags: 0x50200000000000 nid: 0@lo remote: 0xb56a1a02826c7cc4 expref: 22 pid: 31938 timeout: 0 lvb_type: 0
LustreError: 11-0: lustre-MDT0000-mdc-ffff88022f5bae98: operation ldlm_enqueue to node 0@lo failed: rc = -107
Lustre: lustre-MDT0000-mdc-ffff88022f5bae98: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete
LustreError: 167-0: lustre-MDT0000-mdc-ffff88022f5bae98: This client was evicted by lustre-MDT0000; in progress operations using this service will fail.
LustreError: 8311:0:(file.c:5115:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000403:0x6:0x0] error: rc = -5
LustreError: 11654:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-MDT0000-mdc-ffff88022f5bae98: namespace resource [0x200000400:0x4:0x0].0x0 (ffff88031ea99940) refcount nonzero (1) after lock cleanup; forcing cleanup.
Lustre: lustre-MDT0000-mdc-ffff88022f5bae98: Connection restored to 192.168.123.107@tcp (at 0@lo)
LustreError: 14867:0:(mdd_dir.c:4213:mdd_migrate_cmd_check()) lustre-MDD0000: '3' migration was interrupted, run 'lfs migrate -m 1 -c 1 -H crush 3' to finish migration.
Lustre: lustre-MDT0001: trigger partial OI scrub for RPC inconsistency, checking FID [0x240000403:0xa5:0x0]/0xa): rc = 0
Lustre: 6788:0:(lod_lov.c:1315:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000405:0x4d:0x0] with magic=0xbd60bd0
Lustre: 9019:0:(lod_lov.c:1315:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000405:0x57:0x0] with magic=0xbd60bd0
Lustre: 9019:0:(lod_lov.c:1315:lod_parse_striping()) Skipped 1 previous similar message
Lustre: 31938:0:(lod_lov.c:1315:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000405:0x27:0x0] with magic=0xbd60bd0
Lustre: 31938:0:(lod_lov.c:1315:lod_parse_striping()) Skipped 3 previous similar messages
LustreError: 25798:0:(file.c:243:ll_close_inode_openhandle()) lustre-clilmv-ffff88028d824a88: inode [0x200000403:0x177:0x0] mdc close failed: rc = -2
LustreError: 25798:0:(file.c:243:ll_close_inode_openhandle()) Skipped 1 previous similar message
LustreError: 15057:0:(mdd_dir.c:4213:mdd_migrate_cmd_check()) lustre-MDD0000: '1' migration was interrupted, run 'lfs migrate -m 2 -c 1 -H crush 1' to finish migration.
Lustre: mdt03_003: service thread pid 3912 was inactive for 62.114 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one.
Lustre: Skipped 4 previous similar messages
Lustre: mdt02_002: service thread pid 31946 was inactive for 66.116 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one.
Lustre: dir [0x200000403:0x191:0x0] stripe 1 readdir failed: -2, directory is partially accessed!
Lustre: lustre-MDT0000: trigger partial OI scrub for RPC inconsistency, checking FID [0x200000403:0x191:0x0]/0xa): rc = 0
LustreError: 15293:0:(osd_index.c:217:__osd_xattr_load_by_oid()) lustre-MDT0000: can't get bonus, rc = -2
LustreError: 18697:0:(mdd_dir.c:4213:mdd_migrate_cmd_check()) lustre-MDD0000: '11' migration was interrupted, run 'lfs migrate -m 2 -c 1 -H crush 11' to finish migration.
LustreError: 31951:0:(mdd_dir.c:4213:mdd_migrate_cmd_check()) lustre-MDD0001: '2' migration was interrupted, run 'lfs migrate -m 2 -c 1 -H crush 2' to finish migration.
LustreError: 31951:0:(mdd_dir.c:4213:mdd_migrate_cmd_check()) Skipped 1 previous similar message
LustreError: 31846:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0000_UUID lock: ffff8800a7838040/0xb56a1a028271c17f lrc: 3/0,0 mode: PR/PR res: [0x200000403:0x6:0x0].0x0 bits 0x12/0x0 rrc: 6 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0xb56a1a02826c7de3 expref: 128 pid: 15057 timeout: 18256 lvb_type: 0
LustreError: 14887:0:(ldlm_lockd.c:1425:ldlm_handle_enqueue0()) ### lock on destroyed export ffff880290dd6678 ns: mdt-lustre-MDT0000_UUID lock: ffff880284090040/0xb56a1a028271d22d lrc: 3/0,0 mode: PR/PR res: [0x200000403:0x6:0x0].0x0 bits 0x12/0x0 rrc: 4 type: IBT gid 0 flags: 0x50200000000000 nid: 0@lo remote: 0xb56a1a028271c57d expref: 24 pid: 14887 timeout: 0 lvb_type: 0
LustreError: 14887:0:(ldlm_lockd.c:1425:ldlm_handle_enqueue0()) Skipped 7 previous similar messages
LustreError: 11-0: lustre-MDT0000-mdc-ffff88022f5bae98: operation mds_reint to node 0@lo failed: rc = -107
LustreError: Skipped 7 previous similar messages
Lustre: lustre-MDT0000-mdc-ffff88022f5bae98: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete
LustreError: 167-0: lustre-MDT0000-mdc-ffff88022f5bae98: This client was evicted by lustre-MDT0000; in progress operations using this service will fail.
LustreError: 11477:0:(file.c:5115:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000404:0x12:0x0] error: rc = -5
LustreError: 11477:0:(file.c:5115:ll_inode_revalidate_fini()) Skipped 38 previous similar messages
LustreError: 20117:0:(mdc_request.c:1474:mdc_read_page()) lustre-MDT0000-mdc-ffff88022f5bae98: [0x200000403:0x1:0x0] lock enqueue fails: rc = -5
LustreError: 20117:0:(mdc_request.c:1474:mdc_read_page()) Skipped 1 previous similar message
LustreError: 24183:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-MDT0000-mdc-ffff88022f5bae98: namespace resource [0x200000404:0x12:0x0].0x0 (ffff88028606e840) refcount nonzero (1) after lock cleanup; forcing cleanup.
LustreError: 24183:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 21 previous similar messages
Lustre: lustre-MDT0000-mdc-ffff88022f5bae98: Connection restored to 192.168.123.107@tcp (at 0@lo)
LustreError: 20664:0:(file.c:243:ll_close_inode_openhandle()) lustre-clilmv-ffff88022f5bae98: inode [0x280000403:0x225:0x0] mdc close failed: rc = -116
LustreError: 20664:0:(file.c:243:ll_close_inode_openhandle()) Skipped 13 previous similar messages
Lustre: 26268:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff8802911be440 x1729879013188096/t0(0) o10->lustre-OST0001-osc-ffff88028d824a88@0@lo:6/4 lens 440/432 e 0 to 0 dl 1649741552 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dir_create.sh.0'
1[7372]: segfault at 8 ip 00007ff7e64c67e8 sp 00007ffc80e683d0 error 4 in ld-2.17.so[7ff7e64bb000+22000]
LustreError: 7820:0:(mdd_dir.c:4213:mdd_migrate_cmd_check()) lustre-MDD0001: '17' migration was interrupted, run 'lfs migrate -m 2 -c 1 -H crush 17' to finish migration.
Lustre: dir [0x280000404:0xbc8:0x0] stripe 1 readdir failed: -2, directory is partially accessed!
Lustre: Skipped 7 previous similar messages
9[24304]: segfault at 8 ip 00007f3e100177e8 sp 00007fff2bfbc670 error 4 in ld-2.17.so[7f3e1000c000+22000]
LustreError: 8566:0:(mdd_dir.c:4213:mdd_migrate_cmd_check()) lustre-MDD0001: '0' migration was interrupted, run 'lfs migrate -m 0 -c 1 -H crush 0' to finish migration.
LustreError: 7808:0:(llite_nfs.c:338:ll_dir_get_parent_fid()) lustre: failure inode [0x280000404:0x68f:0x0] get parent: rc = -2
Lustre: dir [0x240000404:0x686:0x0] stripe 1 readdir failed: -2, directory is partially accessed!
Lustre: lustre-MDT0002: trigger partial OI scrub for RPC inconsistency, checking FID [0x280000403:0x3be:0x0]/0xa): rc = 0
Lustre: 8699:0:(lod_lov.c:1315:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000403:0x3e9:0x0] with magic=0xbd60bd0
Lustre: 8699:0:(lod_lov.c:1315:lod_parse_striping()) Skipped 3 previous similar messages
LustreError: 31846:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0001_UUID lock: ffff880295c39300/0xb56a1a02827baa7a lrc: 3/0,0 mode: PR/PR res: [0x240000403:0x50c:0x0].0x0 bits 0x13/0x0 rrc: 4 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0xb56a1a02827baa26 expref: 152 pid: 9019 timeout: 18365 lvb_type: 0
LustreError: 11-0: lustre-MDT0001-mdc-ffff88028d824a88: operation ldlm_enqueue to node 0@lo failed: rc = -107
LustreError: Skipped 6 previous similar messages
Lustre: lustre-MDT0001-mdc-ffff88028d824a88: Connection to lustre-MDT0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete
LustreError: 167-0: lustre-MDT0001-mdc-ffff88028d824a88: This client was evicted by lustre-MDT0001; in progress operations using this service will fail.
LustreError: 31839:0:(ldlm_lockd.c:2500:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1649741586 with bad export cookie 13072289466476663552
LustreError: 4374:0:(llite_lib.c:1871:ll_md_setattr()) md_setattr fails: rc = -5
LustreError: 27839:0:(file.c:5115:ll_inode_revalidate_fini()) lustre: revalidate FID [0x240000403:0x50c:0x0] error: rc = -5
LustreError: 27839:0:(file.c:5115:ll_inode_revalidate_fini()) Skipped 8 previous similar messages
LustreError: 1966:0:(mdc_request.c:1474:mdc_read_page()) lustre-MDT0001-mdc-ffff88028d824a88: [0x240000403:0x569:0x0] lock enqueue fails: rc = -108
Lustre: lustre-MDT0001-mdc-ffff88028d824a88: Connection restored to 192.168.123.107@tcp (at 0@lo)
14[10466]: segfault at 8 ip 00007f58976667e8 sp 00007ffd4696fd00 error 4 in ld-2.17.so[7f589765b000+22000]
LustreError: 11975:0:(mdd_dir.c:4213:mdd_migrate_cmd_check()) lustre-MDD0001: '19' migration was interrupted, run 'lfs migrate -m 0 -c 1 -H crush 19' to finish migration.
LustreError: 11975:0:(mdd_dir.c:4213:mdd_migrate_cmd_check()) Skipped 3 previous similar messages
8[12793]: segfault at 8 ip 00007fa0a285f7e8 sp 00007fff0209fd30 error 4 in ld-2.17.so[7fa0a2854000+22000]
Lustre: 26274:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff88027ec53ec0 x1729879018771136/t0(0) o10->lustre-OST0000-osc-ffff88028d824a88@0@lo:6/4 lens 440/432 e 0 to 0 dl 1649741655 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'truncate.0'
LustreError: 17113:0:(file.c:243:ll_close_inode_openhandle()) lustre-clilmv-ffff88028d824a88: inode [0x240000404:0x719:0x0] mdc close failed: rc = -2
LustreError: 17113:0:(file.c:243:ll_close_inode_openhandle()) Skipped 20 previous similar messages
LustreError: 16580:0:(mdd_object.c:3465:mdd_close()) lustre-MDD0002: failed to get lu_attr of [0x280000403:0x5fe:0x0]: rc = -2
LustreError: 16580:0:(mdd_object.c:3465:mdd_close()) lustre-MDD0002: failed to get lu_attr of [0x280000403:0x5fe:0x0]: rc = -2
Lustre: dir [0x200000403:0x60b:0x0] stripe 1 readdir failed: -2, directory is partially accessed!
Lustre: Skipped 10 previous similar messages
LustreError: 25365:0:(mdd_dir.c:4213:mdd_migrate_cmd_check()) lustre-MDD0000: '11' migration was interrupted, run 'lfs migrate -m 2 -c 1 -H crush 11' to finish migration.
LustreError: 25365:0:(mdd_dir.c:4213:mdd_migrate_cmd_check()) Skipped 1 previous similar message
LustreError: 31846:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0000_UUID lock: ffff880243dda200/0xb56a1a0282922b11 lrc: 3/0,0 mode: PR/PR res: [0x200000403:0x5e7:0x0].0x0 bits 0x11/0x0 rrc: 5 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0xb56a1a0282922af5 expref: 136 pid: 2673 timeout: 18600 lvb_type: 0
LustreError: 11-0: lustre-MDT0000-mdc-ffff88022f5bae98: operation mds_close to node 0@lo failed: rc = -107
LustreError: Skipped 1 previous similar message
Lustre: lustre-MDT0000-mdc-ffff88022f5bae98: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete
LustreError: 167-0: lustre-MDT0000-mdc-ffff88022f5bae98: This client was evicted by lustre-MDT0000; in progress operations using this service will fail.
LustreError: 21444:0:(llite_lib.c:3119:ll_prep_inode()) new_inode -fatal: rc -2
LustreError: 21803:0:(file.c:5115:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000403:0x1:0x0] error: rc = -108
LustreError: 21803:0:(file.c:5115:ll_inode_revalidate_fini()) Skipped 1 previous similar message
LustreError: 28354:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-MDT0000-mdc-ffff88022f5bae98: namespace resource [0x200000403:0x5e7:0x0].0x0 (ffff880293694040) refcount nonzero (1) after lock cleanup; forcing cleanup.
LustreError: 28354:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 1 previous similar message
Lustre: lustre-MDT0000-mdc-ffff88022f5bae98: Connection restored to 192.168.123.107@tcp (at 0@lo)
17[27582]: segfault at 8 ip 00007f4e0b9377e8 sp 00007ffd73ead0a0 error 4 in ld-2.17.so[7f4e0b92c000+22000]
LustreError: 32622:0:(mdd_dir.c:4213:mdd_migrate_cmd_check()) lustre-MDD0002: '13' migration was interrupted, run 'lfs migrate -m 1 -c 1 -H crush 13' to finish migration.
LustreError: 32622:0:(mdd_dir.c:4213:mdd_migrate_cmd_check()) Skipped 3 previous similar messages
LustreError: 31846:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0002_UUID lock: ffff8802822c1300/0xb56a1a02829345c6 lrc: 3/0,0 mode: PR/PR res: [0x280000403:0x914:0x0].0x0 bits 0x13/0x0 rrc: 4 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0xb56a1a02829345aa expref: 216 pid: 31938 timeout: 18612 lvb_type: 0
LustreError: 11-0: lustre-MDT0002-mdc-ffff88028d824a88: operation ldlm_enqueue to node 0@lo failed: rc = -107
Lustre: lustre-MDT0002-mdc-ffff88028d824a88: Connection to lustre-MDT0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete
LustreError: 167-0: lustre-MDT0002-mdc-ffff88028d824a88: This client was evicted by lustre-MDT0002; in progress operations using this service will fail.
LustreError: 3005:0:(llite_lib.c:1871:ll_md_setattr()) md_setattr fails: rc = -5
LustreError: 3365:0:(mdc_request.c:1474:mdc_read_page()) lustre-MDT0002-mdc-ffff88028d824a88: [0x280000400:0x2e:0x0] lock enqueue fails: rc = -108
LustreError: 3365:0:(mdc_request.c:1474:mdc_read_page()) Skipped 2 previous similar messages
Lustre: dir [0x280000404:0xf6d:0x0] stripe 0 readdir failed: -108, directory is partially accessed!
Lustre: Skipped 7 previous similar messages
LustreError: 26268:0:(mdc_locks.c:1389:mdc_intent_getattr_async_interpret()) lustre-MDT0002-mdc-ffff88028d824a88: ldlm_cli_enqueue_fini() failed: rc = -5
LustreError: 333:0:(file.c:5115:ll_inode_revalidate_fini()) lustre: revalidate FID [0x280000403:0x914:0x0] error: rc = -5
LustreError: 333:0:(file.c:5115:ll_inode_revalidate_fini()) Skipped 6 previous similar messages
Lustre: lustre-MDT0002-mdc-ffff88028d824a88: Connection restored to 192.168.123.107@tcp (at 0@lo)
LustreError: 32138:0:(ldlm_lockd.c:1425:ldlm_handle_enqueue0()) ### lock on destroyed export ffff88008ab8e678 ns: mdt-lustre-MDT0002_UUID lock: ffff8802f4a1d2c0/0xb56a1a0282978972 lrc: 3/0,0 mode: PR/PR res: [0x280000403:0x896:0x0].0x0 bits 0x1b/0x0 rrc: 2 type: IBT gid 0 flags: 0x50200000000000 nid: 0@lo remote: 0xb56a1a0282976b0a expref: 12 pid: 32138 timeout: 0 lvb_type: 0
LustreError: 32138:0:(ldlm_lockd.c:1425:ldlm_handle_enqueue0()) Skipped 6 previous similar messages
Lustre: 18542:0:(lod_lov.c:1315:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000404:0x10f3:0x0] with magic=0xbd60bd0
Lustre: 18542:0:(lod_lov.c:1315:lod_parse_striping()) Skipped 1 previous similar message
LustreError: 17709:0:(file.c:243:ll_close_inode_openhandle()) lustre-clilmv-ffff88022f5bae98: inode [0x280000403:0x8db:0x0] mdc close failed: rc = -13
LustreError: 17709:0:(file.c:243:ll_close_inode_openhandle()) Skipped 38 previous similar messages
7[26150]: segfault at 8 ip 00007f265d7127e8 sp 00007fff8b44de70 error 4 in ld-2.17.so[7f265d707000+22000]
LustreError: 1495:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0000-osp-MDT0002: fail to cancel 1 llog-records: rc = -2
LustreError: 1495:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) Skipped 1 previous similar message
LustreError: 1495:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0000-osp-MDT0002: fail to cancel 1 of 1 llog-records: rc = -2
LustreError: 1495:0:(llog_cat.c:787:llog_cat_cancel_records()) Skipped 1 previous similar message
Lustre: dir [0x240000404:0xe05:0x0] stripe 1 readdir failed: -2, directory is partially accessed!
Lustre: 11689:0:(lod_lov.c:1315:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000404:0xeba:0x0] with magic=0xbd60bd0
Lustre: 11689:0:(lod_lov.c:1315:lod_parse_striping()) Skipped 5 previous similar messages
LustreError: 31966:0:(mdd_orphans.c:282:mdd_orphan_delete()) lustre-MDD0001: could not delete orphan object [0x240000404:0xe4e:0x0]: rc = -2
LustreError: 31966:0:(mdd_object.c:3517:mdd_close()) lustre-MDD0001: unable to delete [0x240000404:0xe4e:0x0] from orphan list: rc = -2
LustreError: 8566:0:(mdd_object.c:401:mdd_xattr_get()) lustre-MDD0002: object [0x280000405:0x30c:0x0] not found: rc = -2
LustreError: 31846:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0000_UUID lock: ffff88009173de00/0xb56a1a02829f56e1 lrc: 3/0,0 mode: PR/PR res: [0x200000407:0x11e:0x0].0x0 bits 0x12/0x0 rrc: 5 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0xb56a1a02829f56d3 expref: 195 pid: 31960 timeout: 18803 lvb_type: 0
LustreError: 11-0: lustre-MDT0000-mdc-ffff88022f5bae98: operation ldlm_enqueue to node 0@lo failed: rc = -107
Lustre: lustre-MDT0000-mdc-ffff88022f5bae98: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete
LustreError: 167-0: lustre-MDT0000-mdc-ffff88022f5bae98: This client was evicted by lustre-MDT0000; in progress operations using this service will fail.
LustreError: 30750:0:(llite_lib.c:1871:ll_md_setattr()) md_setattr fails: rc = -5
LustreError: 30750:0:(llite_lib.c:1871:ll_md_setattr()) Skipped 1 previous similar message
LustreError: 31131:0:(llite_lib.c:3119:ll_prep_inode()) new_inode -fatal: rc -2
Lustre: lustre-MDT0000-mdc-ffff88022f5bae98: Connection restored to 192.168.123.107@tcp (at 0@lo)
LustreError: 9437:0:(ldlm_lockd.c:1425:ldlm_handle_enqueue0()) ### lock on destroyed export ffff880274232e98 ns: mdt-lustre-MDT0000_UUID lock: ffff8802cc2b5680/0xb56a1a0282a9d918 lrc: 3/0,0 mode: PR/PR res: [0x200000407:0x314:0x0].0x0 bits 0x13/0x0 rrc: 4 type: IBT gid 0 flags: 0x50200000000000 nid: 0@lo remote: 0xb56a1a0282a9d8f5 expref: 4 pid: 9437 timeout: 0 lvb_type: 0
LustreError: 9437:0:(ldlm_lockd.c:1425:ldlm_handle_enqueue0()) Skipped 2 previous similar messages
LustreError: 26435:0:(llite_lib.c:1708:ll_update_lsm_md()) lustre: [0x280000405:0x1c6:0x0] dir layout mismatch:
LustreError: 26435:0:(lustre_lmv.h:142:lsm_md_dump()) magic 0xcd20cd0 stripe count 1 master mdt 2 hash type crush:0x2000003 max-inherit 0 max-inherit-rr 0 version 1 migrate offset 0 migrate hash 0x0 pool
LustreError: 26435:0:(lustre_lmv.h:149:lsm_md_dump()) stripe[0] [0x280000400:0x3a:0x0]
LustreError: 26435:0:(lustre_lmv.h:142:lsm_md_dump()) magic 0xcd20cd0 stripe count 2 master mdt 2 hash type crush:0x82000003 max-inherit 0 max-inherit-rr 0 version 0 migrate offset 1 migrate hash 0x2 pool
LustreError: 26435:0:(lustre_lmv.h:149:lsm_md_dump()) stripe[0] [0x280000400:0x3a:0x0]
LustreError: 26433:0:(lustre_lmv.h:142:lsm_md_dump()) magic 0xcd20cd0 stripe count 1 master mdt 2 hash type crush:0x2000003 max-inherit 0 max-inherit-rr 0 version 1 migrate offset 0 migrate hash 0x0 pool
LustreError: 26433:0:(lustre_lmv.h:149:lsm_md_dump()) stripe[0] [0x280000400:0x3a:0x0]
LustreError: 26433:0:(lustre_lmv.h:142:lsm_md_dump()) magic 0xcd20cd0 stripe count 2 master mdt 2 hash type crush:0x82000003 max-inherit 0 max-inherit-rr 0 version 0 migrate offset 1 migrate hash 0x2 pool
LustreError: 26433:0:(lustre_lmv.h:149:lsm_md_dump()) stripe[0] [0x280000400:0x3a:0x0]
LustreError: 26433:0:(lustre_lmv.h:149:lsm_md_dump()) stripe[1] [0x240000404:0xdcb:0x0]
LustreError: 26435:0:(lustre_lmv.h:149:lsm_md_dump()) stripe[1] [0x240000404:0xdcb:0x0]
Lustre: dir [0x280000405:0x1c6:0x0] stripe 1 readdir failed: -2, directory is partially accessed!
Lustre: Skipped 13 previous similar messages
Lustre: 9948:0:(lod_lov.c:1315:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000408:0x5e:0x0] with magic=0xbd60bd0
Lustre: 9948:0:(lod_lov.c:1315:lod_parse_striping()) Skipped 11 previous similar messages
LustreError: 31980:0:(out_handler.c:910:out_tx_end()) lustre-MDT0000-osd: undo for /home/green/git/lustre-release/lustre/ptlrpc/../../lustre/target/out_handler.c:445: rc = -524
LustreError: 1495:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0000-osp-MDT0002: fail to cancel 1 llog-records: rc = -116
LustreError: 1495:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0000-osp-MDT0002: fail to cancel 1 of 1 llog-records: rc = -116
Lustre: 9437:0:(lod_lov.c:1315:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000408:0xec:0x0] with magic=0xbd60bd0
Lustre: 9437:0:(lod_lov.c:1315:lod_parse_striping()) Skipped 3 previous similar messages
LustreError: 11715:0:(llog_osd.c:626:llog_osd_write_rec()) lustre-MDT0000-osp-MDT0002: index 1523 already set in llog bitmap [0x200000402:0x2:0x0]
Link to test
Return to new crashes list