Match messages in logs (every line would be required to be present in log output Copy from "Messages before crash" column below): | |
Match messages in full crash (every line would be required to be present in crash log output Copy from "Full Crash" column below): | |
Limit to a test: (Copy from below "Failing text"): | |
Delete these reports as invalid (real bug in review or some such) | |
Bug or comment: | |
Extra info: |
Failing Test | Full Crash | Messages before crash | Comment |
---|---|---|---|
sanity test 319: lost lease lock on migrate error | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa15fc11c>] ll_migrate+0x9ec/0xf50 [lustre] PGD 27cb02067 PUD 292c13067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) osp(OE) ofd(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) lquota(OE) lfsck(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) ext4 mbcache jbd2 loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) crc32_generic crc_t10dif crct10dif_generic crct10dif_common virtio_console virtio_balloon pcspkr i2c_piix4 ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm ata_piix drm_panel_orientation_quirks virtio_blk serio_raw floppy libata i2c_core [last unloaded: obdecho] CPU: 0 PID: 27808 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.16.0-3.module_el8.7.0+1218+f626c2ff 04/01/2014 task: ffff880239fe49d0 ti: ffff8802d6208000 task.ti: ffff8802d6208000 RIP: 0010:[<ffffffffa15fc11c>] [<ffffffffa15fc11c>] ll_migrate+0x9ec/0xf50 [lustre] RSP: 0018:ffff8802d620bbd0 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff88013623e6b8 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff8802b6addfc8 RDI: ffff8802b6addfa8 RBP: ffff8802d620bc48 R08: ffff880276e316f8 R09: ffff8802d620ba68 R10: ffff88013623ae68 R11: ffff8801362389f8 R12: ffff88018074e718 R13: 0000000000000000 R14: ffff8801e1fd0a58 R15: ffff880089bce738 FS: 00007f0c68299740(0000) GS:ffff880331a00000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 0000000320b36000 CR4: 00000000000007f0 Call Trace: [<ffffffffa15e8bef>] ll_dir_ioctl+0x5d0f/0x6fe0 [lustre] [<ffffffff811b630d>] ? find_get_pages_tag+0x10d/0x260 [<ffffffff811c36c1>] ? pagevec_lookup_tag+0x21/0x30 [<ffffffff811b403e>] ? __filemap_fdatawait_range+0xbe/0x1a0 [<ffffffff810b10ca>] ? __queue_delayed_work+0xaa/0x1a0 [<ffffffff817df54c>] ? __mutex_unlock_slowpath+0xdc/0x190 [<ffffffff817df60e>] ? mutex_unlock+0xe/0x10 [<ffffffff81276d80>] ? sync_inodes_sb+0x160/0x430 [<ffffffff810b10ca>] ? __queue_delayed_work+0xaa/0x1a0 [<ffffffff810b1a51>] ? try_to_grab_pending+0xb1/0x180 [<ffffffff8125b47d>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff8125b7d1>] SyS_ioctl+0xa1/0xc0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a | LustreError: 27792:0:(llite_lib.c:1937:ll_md_setattr()) md_setattr fails: rc = -28 LustreError: 27792:0:(file.c:242:ll_close_inode_openhandle()) lustre-clilmv-ffff88031dd2efc8: inode [0x2000059f3:0x92f0:0x0] mdc close failed: rc = -28 LustreError: 27792:0:(file.c:242:ll_close_inode_openhandle()) Skipped 1 previous similar message LustreError: 27808:0:(fail.c:138:__cfs_fail_timeout_set()) cfs_fail_timeout id 32c sleeping for 5000ms LustreError: 27808:0:(fail.c:149:__cfs_fail_timeout_set()) cfs_fail_timeout id 32c awake | Link to test |
sanity test 300uf: migrate with too many local locks | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa1705e65>] ll_migrate+0xa45/0xff0 [lustre] PGD 1c203e067 PUD 1d5228067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) osp(OE) ofd(OE) lod(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) lquota(OE) lfsck(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) nfsd ext4 mbcache jbd2 loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) crc32_generic crc_t10dif crct10dif_generic crct10dif_common virtio_balloon pcspkr virtio_console i2c_piix4 ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm ata_piix drm_panel_orientation_quirks floppy serio_raw virtio_blk libata i2c_core [last unloaded: obdecho] CPU: 3 PID: 23150 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.16.0-3.module_el8.7.0+1218+f626c2ff 04/01/2014 task: ffff880127b4a4f0 ti: ffff8801a26f4000 task.ti: ffff8801a26f4000 RIP: 0010:[<ffffffffa1705e65>] [<ffffffffa1705e65>] ll_migrate+0xa45/0xff0 [lustre] RSP: 0018:ffff8801a26f7bc0 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff8801046c9e48 RCX: 0000000000000000 RDX: 0000000000000000 RSI: 0000000000000000 RDI: ffff8801d1a10ad8 RBP: ffff8801a26f7c40 R08: 000000000000ffff R09: 000000000000ffff R10: 0000000000000000 R11: 3d63722820676e69 R12: ffff8800a2d84968 R13: 0000000000000000 R14: ffff8800a5350d48 R15: ffff8801046ce288 FS: 00007fc05791d740(0000) GS:ffff880331ac0000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 0000000204426000 CR4: 00000000000007e0 Call Trace: [<ffffffffa16eeff7>] ll_dir_ioctl+0x6bc7/0x7860 [lustre] [<ffffffff81402ab9>] ? snprintf+0x49/0x70 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff810a2009>] ? mod_timer+0x119/0x260 [<ffffffff810a2168>] ? add_timer+0x18/0x20 [<ffffffff810b10ab>] ? __queue_delayed_work+0x8b/0x1a0 [<ffffffff817e32dc>] ? _raw_spin_unlock_bh+0x1c/0x20 [<ffffffff811d92ef>] ? bdi_wakeup_thread_delayed+0x5f/0x70 [<ffffffff81277f01>] ? __mark_inode_dirty+0x161/0x2c0 [<ffffffff8125b47d>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff817df60e>] ? mutex_unlock+0xe/0x10 [<ffffffff8125b89a>] ? iterate_dir+0xaa/0x120 [<ffffffff8125b7d1>] SyS_ioctl+0xa1/0xc0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a | Lustre: DEBUG MARKER: touch/create LustreError: 799:0:(osd_handler.c:207:osd_trans_start()) lustre-MDT0001: failed to start transaction due to ENOSPC: rc = -28 LustreError: 799:0:(osd_handler.c:207:osd_trans_start()) Skipped 4 previous similar messages LustreError: 23041:0:(llite_lib.c:1996:ll_md_setattr()) md_setattr fails: rc = -28 LustreError: 23041:0:(file.c:247:ll_close_inode_openhandle()) lustre-clilmv-ffff8802cfa4e678: inode [0x240002b15:0x94fd:0x0] mdc close failed: rc = -28 Lustre: DEBUG MARKER: hardlinks Lustre: DEBUG MARKER: cancel lru Lustre: DEBUG MARKER: migrate | Link to test |
sanity test 300uf: migrate with too many local locks | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa1690315>] ll_migrate+0xa45/0xff0 [lustre] PGD 2f555c067 PUD 308aff067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) osp(OE) ofd(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) lquota(OE) lfsck(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) nfsd ext4 mbcache jbd2 loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) crc32_generic crc_t10dif crct10dif_generic crct10dif_common virtio_console virtio_balloon pcspkr i2c_piix4 ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm drm_panel_orientation_quirks ata_piix serio_raw floppy libata virtio_blk i2c_core [last unloaded: obdecho] CPU: 13 PID: 27986 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.16.0-3.module_el8.7.0+1218+f626c2ff 04/01/2014 task: ffff880301621280 ti: ffff8802e8c34000 task.ti: ffff8802e8c34000 RIP: 0010:[<ffffffffa1690315>] [<ffffffffa1690315>] ll_migrate+0xa45/0xff0 [lustre] RSP: 0018:ffff8802e8c37bc0 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff880249ffc548 RCX: 0000000000000000 RDX: 0000000000000000 RSI: 0000000000000000 RDI: ffff880077da8fd8 RBP: ffff8802e8c37c40 R08: 000000000000ffff R09: 000000000000ffff R10: 0000000000000000 R11: 3d63722820676e69 R12: ffff8802846e5138 R13: 0000000000000000 R14: ffff8802464c1738 R15: ffff880249ffcf08 FS: 00007f95c305e740(0000) GS:ffff880331d40000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 0000000308938000 CR4: 00000000000007e0 Call Trace: [<ffffffffa167ada4>] ll_dir_ioctl+0x7184/0x7ed0 [lustre] [<ffffffff81402ab9>] ? snprintf+0x49/0x70 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffffa022fc17>] ? cfs_trace_unlock_tcd+0x27/0x70 [libcfs] [<ffffffffa02304a4>] ? libcfs_debug_msg+0x6d4/0xc20 [libcfs] [<ffffffffa03b5be9>] ? lprocfs_counter_add+0xe9/0x180 [obdclass] [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e324e>] ? _raw_spin_unlock+0xe/0x20 [<ffffffff81277f48>] ? __mark_inode_dirty+0x1a8/0x2c0 [<ffffffff8125b47d>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff817df60e>] ? mutex_unlock+0xe/0x10 [<ffffffff8125b89a>] ? iterate_dir+0xaa/0x120 [<ffffffff8125b7d1>] SyS_ioctl+0xa1/0xc0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a | Lustre: DEBUG MARKER: touch/create Lustre: DEBUG MARKER: hardlinks LustreError: 14261:0:(osd_handler.c:207:osd_trans_start()) lustre-MDT0001: failed to start transaction due to ENOSPC: rc = -28 LustreError: 14261:0:(osd_handler.c:207:osd_trans_start()) Skipped 2 previous similar messages Lustre: DEBUG MARKER: cancel lru Lustre: DEBUG MARKER: migrate | Link to test |
sanity test 300uf: migrate with too many local locks | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa1697315>] ll_migrate+0xa45/0xff0 [lustre] PGD 28b850067 PUD 175548067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) osp(OE) ofd(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) lquota(OE) lfsck(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) nfsd ext4 mbcache jbd2 loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) crc32_generic crc_t10dif crct10dif_generic crct10dif_common pcspkr virtio_console i2c_piix4 virtio_balloon ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm drm_panel_orientation_quirks ata_piix libata floppy virtio_blk serio_raw i2c_core [last unloaded: obdecho] CPU: 3 PID: 22885 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.16.0-3.module_el8.7.0+1218+f626c2ff 04/01/2014 task: ffff88028d5b3760 ti: ffff88024565c000 task.ti: ffff88024565c000 RIP: 0010:[<ffffffffa1697315>] [<ffffffffa1697315>] ll_migrate+0xa45/0xff0 [lustre] RSP: 0018:ffff88024565fbc0 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff880002eecf08 RCX: 0000000000000000 RDX: 0000000000000000 RSI: 0000000000000000 RDI: ffff8801bfd35558 RBP: ffff88024565fc40 R08: 000000000000ffff R09: 000000000000ffff R10: 0000000000000000 R11: 3d63722820676e69 R12: ffff8801273c1c28 R13: 0000000000000000 R14: ffff88022df4e6b8 R15: ffff880002eed8c8 FS: 00007fbb23449740(0000) GS:ffff880331ac0000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 000000018b1d0000 CR4: 00000000000007e0 Call Trace: [<ffffffffa1681da4>] ll_dir_ioctl+0x7184/0x7ed0 [lustre] [<ffffffff81402ab9>] ? snprintf+0x49/0x70 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffffa0255c17>] ? cfs_trace_unlock_tcd+0x27/0x70 [libcfs] [<ffffffffa02564a4>] ? libcfs_debug_msg+0x6d4/0xc20 [libcfs] [<ffffffffa03b5be9>] ? lprocfs_counter_add+0xe9/0x180 [obdclass] [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e324e>] ? _raw_spin_unlock+0xe/0x20 [<ffffffff81277f48>] ? __mark_inode_dirty+0x1a8/0x2c0 [<ffffffff8125b47d>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff817df60e>] ? mutex_unlock+0xe/0x10 [<ffffffff8125b89a>] ? iterate_dir+0xaa/0x120 [<ffffffff8125b7d1>] SyS_ioctl+0xa1/0xc0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a | Lustre: DEBUG MARKER: touch/create Lustre: DEBUG MARKER: hardlinks LustreError: 2645:0:(osd_handler.c:207:osd_trans_start()) lustre-MDT0001: failed to start transaction due to ENOSPC: rc = -28 Lustre: DEBUG MARKER: cancel lru Lustre: DEBUG MARKER: migrate | Link to test |
sanity test 300uf: migrate with too many local locks | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa16b1fb5>] ll_migrate+0xa45/0xff0 [lustre] PGD 1fb40067 PUD 1fb41067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) osp(OE) ofd(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) lquota(OE) lfsck(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) ext4 mbcache jbd2 loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) crc32_generic crc_t10dif crct10dif_generic crct10dif_common pcspkr virtio_console virtio_balloon i2c_piix4 ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm ata_piix drm_panel_orientation_quirks serio_raw virtio_blk i2c_core libata floppy [last unloaded: obdecho] CPU: 15 PID: 15257 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.16.0-3.module_el8.7.0+1218+f626c2ff 04/01/2014 task: ffff88003c53b760 ti: ffff88002b060000 task.ti: ffff88002b060000 RIP: 0010:[<ffffffffa16b1fb5>] [<ffffffffa16b1fb5>] ll_migrate+0xa45/0xff0 [lustre] RSP: 0018:ffff88002b063bc0 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff8802e592cd08 RCX: 0000000000000000 RDX: 0000000000000000 RSI: 0000000000000000 RDI: ffff880087d648b8 RBP: ffff88002b063c40 R08: 000000000000ffff R09: 000000000000ffff R10: 0000000000000000 R11: 3d63722820676e69 R12: ffff880025243078 R13: 0000000000000000 R14: ffff88009dfccf88 R15: ffff8802e5933a08 FS: 00007f5cd4f2d740(0000) GS:ffff880331dc0000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 0000000010f3e000 CR4: 00000000000007e0 Call Trace: [<ffffffffa169ce0c>] ll_dir_ioctl+0x6b9c/0x7750 [lustre] [<ffffffff81402ab9>] ? snprintf+0x49/0x70 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff810a2009>] ? mod_timer+0x119/0x260 [<ffffffff810a2168>] ? add_timer+0x18/0x20 [<ffffffff810b10ab>] ? __queue_delayed_work+0x8b/0x1a0 [<ffffffffa03bea09>] ? lprocfs_counter_add+0xe9/0x180 [obdclass] [<ffffffff817e32dc>] ? _raw_spin_unlock_bh+0x1c/0x20 [<ffffffff811d92ef>] ? bdi_wakeup_thread_delayed+0x5f/0x70 [<ffffffff81277f01>] ? __mark_inode_dirty+0x161/0x2c0 [<ffffffff8125b47d>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff817df60e>] ? mutex_unlock+0xe/0x10 [<ffffffff8125b89a>] ? iterate_dir+0xaa/0x120 [<ffffffff8125b7d1>] SyS_ioctl+0xa1/0xc0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a | Lustre: DEBUG MARKER: touch/create LustreError: 27010:0:(osd_handler.c:226:osd_trans_start()) lustre-MDT0001: failed to start transaction due to ENOSPC: rc = -28 LustreError: 27010:0:(osd_handler.c:226:osd_trans_start()) Skipped 6 previous similar messages LustreError: 15086:0:(llite_lib.c:2021:ll_md_setattr()) md_setattr fails: rc = -28 LustreError: 15086:0:(file.c:262:ll_close_inode_openhandle()) lustre-clilmv-ffff88027339ae98: inode [0x240002b15:0x918a:0x0] mdc close failed: rc = -28 Lustre: DEBUG MARKER: hardlinks Lustre: DEBUG MARKER: cancel lru Lustre: DEBUG MARKER: migrate | Link to test |
sanity test 300uf: migrate with too many local locks | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa1696fb5>] ll_migrate+0xa45/0xff0 [lustre] PGD 2efed5067 PUD 1fecea067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) osp(OE) ofd(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) lquota(OE) lfsck(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) ext4 mbcache jbd2 loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) crc32_generic crc_t10dif crct10dif_generic crct10dif_common virtio_console pcspkr i2c_piix4 virtio_balloon ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm ata_piix floppy drm_panel_orientation_quirks virtio_blk serio_raw libata i2c_core [last unloaded: obdecho] CPU: 2 PID: 12871 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.16.0-3.module_el8.7.0+1218+f626c2ff 04/01/2014 task: ffff8802758c24f0 ti: ffff88029cf04000 task.ti: ffff88029cf04000 RIP: 0010:[<ffffffffa1696fb5>] [<ffffffffa1696fb5>] ll_migrate+0xa45/0xff0 [lustre] RSP: 0018:ffff88029cf07bc0 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff8801cb836988 RCX: 0000000000000000 RDX: 0000000000000000 RSI: 0000000000000000 RDI: ffff8803270f0a58 RBP: ffff88029cf07c40 R08: 000000000000ffff R09: 000000000000ffff R10: 0000000000000000 R11: 3d63722820676e69 R12: ffff88001b2f7848 R13: 0000000000000000 R14: ffff88018d181738 R15: ffff8801cb830a88 FS: 00007ff0cb46e740(0000) GS:ffff880331a80000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 000000010b378000 CR4: 00000000000007e0 Call Trace: [<ffffffffa1681e0c>] ll_dir_ioctl+0x6b9c/0x7750 [lustre] [<ffffffff81402ab9>] ? snprintf+0x49/0x70 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff810a2009>] ? mod_timer+0x119/0x260 [<ffffffff810a2168>] ? add_timer+0x18/0x20 [<ffffffff810b10ab>] ? __queue_delayed_work+0x8b/0x1a0 [<ffffffffa03d5a09>] ? lprocfs_counter_add+0xe9/0x180 [obdclass] [<ffffffff817e32dc>] ? _raw_spin_unlock_bh+0x1c/0x20 [<ffffffff811d92ef>] ? bdi_wakeup_thread_delayed+0x5f/0x70 [<ffffffff81277f01>] ? __mark_inode_dirty+0x161/0x2c0 [<ffffffff8125b47d>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff817df60e>] ? mutex_unlock+0xe/0x10 [<ffffffff8125b89a>] ? iterate_dir+0xaa/0x120 [<ffffffff8125b7d1>] SyS_ioctl+0xa1/0xc0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a | Lustre: DEBUG MARKER: touch/create LustreError: 10713:0:(osd_handler.c:226:osd_trans_start()) lustre-MDT0001: failed to start transaction due to ENOSPC: rc = -28 LustreError: 12771:0:(llite_lib.c:2021:ll_md_setattr()) md_setattr fails: rc = -28 LustreError: 12771:0:(file.c:262:ll_close_inode_openhandle()) lustre-clilmv-ffff8802c573d3d8: inode [0x240002b15:0x8bfd:0x0] mdc close failed: rc = -28 Lustre: DEBUG MARKER: hardlinks Lustre: DEBUG MARKER: cancel lru Lustre: DEBUG MARKER: migrate LustreError: 10508:0:(osd_handler.c:226:osd_trans_start()) lustre-MDT0001: failed to start transaction due to ENOSPC: rc = -28 LustreError: 10508:0:(osd_handler.c:226:osd_trans_start()) Skipped 19 previous similar messages | Link to test |
sanity test 300uf: migrate with too many local locks | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa168c0a5>] ll_migrate+0xa45/0xff0 [lustre] PGD 28d087067 PUD 1d9d81067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) osp(OE) ofd(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) lquota(OE) lfsck(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) ext4 mbcache jbd2 loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) crc32_generic crc_t10dif crct10dif_generic crct10dif_common virtio_balloon virtio_console i2c_piix4 pcspkr ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm drm_panel_orientation_quirks floppy ata_piix virtio_blk serio_raw libata i2c_core [last unloaded: obdecho] CPU: 1 PID: 17794 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.16.0-3.module_el8.7.0+1218+f626c2ff 04/01/2014 task: ffff880040e10010 ti: ffff8802757c8000 task.ti: ffff8802757c8000 RIP: 0010:[<ffffffffa168c0a5>] [<ffffffffa168c0a5>] ll_migrate+0xa45/0xff0 [lustre] RSP: 0018:ffff8802757cbbc0 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff88005165b088 RCX: 0000000000000000 RDX: 0000000000000000 RSI: 0000000000000000 RDI: ffff88026a7ea5c8 RBP: ffff8802757cbc40 R08: 000000000000ffff R09: 000000000000ffff R10: 0000000000000000 R11: 3d63722820676e69 R12: ffff88029aacb208 R13: 0000000000000000 R14: ffff8802a5192128 R15: ffff88005165cd08 FS: 00007f422cb3c740(0000) GS:ffff880331a40000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 00000002d2256000 CR4: 00000000000007e0 Call Trace: [<ffffffffa1676e0c>] ll_dir_ioctl+0x6b9c/0x7750 [lustre] [<ffffffff81402ab9>] ? snprintf+0x49/0x70 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff810a2009>] ? mod_timer+0x119/0x260 [<ffffffff810a2168>] ? add_timer+0x18/0x20 [<ffffffff810b10ab>] ? __queue_delayed_work+0x8b/0x1a0 [<ffffffffa0472a09>] ? lprocfs_counter_add+0xe9/0x180 [obdclass] [<ffffffff817e32dc>] ? _raw_spin_unlock_bh+0x1c/0x20 [<ffffffff811d92ef>] ? bdi_wakeup_thread_delayed+0x5f/0x70 [<ffffffff81277f01>] ? __mark_inode_dirty+0x161/0x2c0 [<ffffffff8125b47d>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff817df60e>] ? mutex_unlock+0xe/0x10 [<ffffffff8125b89a>] ? iterate_dir+0xaa/0x120 [<ffffffff8125b7d1>] SyS_ioctl+0xa1/0xc0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a | Lustre: DEBUG MARKER: touch/create LustreError: 28452:0:(osd_handler.c:226:osd_trans_start()) lustre-MDT0001: failed to start transaction due to ENOSPC: rc = -28 LustreError: 28452:0:(osd_handler.c:226:osd_trans_start()) Skipped 1 previous similar message LustreError: 17610:0:(llite_lib.c:2028:ll_md_setattr()) md_setattr fails: rc = -28 LustreError: 17610:0:(file.c:264:ll_close_inode_openhandle()) lustre-clilmv-ffff88008844a548: inode [0x240002b15:0x8ca5:0x0] mdc close failed: rc = -28 Lustre: DEBUG MARKER: hardlinks Lustre: DEBUG MARKER: cancel lru Lustre: DEBUG MARKER: migrate | Link to test |
sanity test 300uf: migrate with too many local locks | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa168b0a5>] ll_migrate+0xa45/0xff0 [lustre] PGD 2bf03e067 PUD 2ccd1c067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) osp(OE) ofd(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) lquota(OE) lfsck(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) ext4 mbcache jbd2 loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) crc32_generic crc_t10dif crct10dif_generic crct10dif_common virtio_balloon i2c_piix4 virtio_console pcspkr ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm ata_piix drm_panel_orientation_quirks libata virtio_blk serio_raw floppy i2c_core [last unloaded: obdecho] CPU: 1 PID: 9068 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.16.0-3.module_el8.7.0+1218+f626c2ff 04/01/2014 task: ffff880123083760 ti: ffff88019bb48000 task.ti: ffff88019bb48000 RIP: 0010:[<ffffffffa168b0a5>] [<ffffffffa168b0a5>] ll_migrate+0xa45/0xff0 [lustre] RSP: 0018:ffff88019bb4bbc0 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff8802f83bba08 RCX: 0000000000000000 RDX: 0000000000000000 RSI: 0000000000000000 RDI: ffff8802bb5f7048 RBP: ffff88019bb4bc40 R08: 000000000000ffff R09: 000000000000ffff R10: 0000000000000000 R11: 3d63722820676e69 R12: ffff8801d4109138 R13: 0000000000000000 R14: ffff8802b2b131b8 R15: ffff8802f83bd688 FS: 00007f3d159ce740(0000) GS:ffff880331a40000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 00000001af866000 CR4: 00000000000007e0 Call Trace: [<ffffffffa1675e0c>] ll_dir_ioctl+0x6b9c/0x7750 [lustre] [<ffffffff81402ab9>] ? snprintf+0x49/0x70 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffffa024ec07>] ? cfs_trace_unlock_tcd+0x27/0x70 [libcfs] [<ffffffffa024f494>] ? libcfs_debug_msg+0x6d4/0xc20 [libcfs] [<ffffffffa03afa19>] ? lprocfs_counter_add+0xe9/0x180 [obdclass] [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e324e>] ? _raw_spin_unlock+0xe/0x20 [<ffffffff81277f48>] ? __mark_inode_dirty+0x1a8/0x2c0 [<ffffffff8125b47d>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff817df60e>] ? mutex_unlock+0xe/0x10 [<ffffffff8125b89a>] ? iterate_dir+0xaa/0x120 [<ffffffff8125b7d1>] SyS_ioctl+0xa1/0xc0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a | Lustre: DEBUG MARKER: touch/create Lustre: DEBUG MARKER: hardlinks LustreError: 17623:0:(osd_handler.c:226:osd_trans_start()) lustre-MDT0001: failed to start transaction due to ENOSPC: rc = -28 Lustre: DEBUG MARKER: cancel lru Lustre: DEBUG MARKER: migrate | Link to test |
sanity test 316: lfs migrate of file with large_xattr enabled | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa164a04c>] ll_migrate+0xa4c/0xf90 [lustre] PGD 6b2c2067 PUD 53ac9067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) lquota(OE) lfsck(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) nfsd ext4 mbcache jbd2 loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) crc32_generic crc_t10dif crct10dif_generic crct10dif_common pcspkr virtio_balloon virtio_console i2c_piix4 ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm drm_panel_orientation_quirks ata_piix serio_raw virtio_blk i2c_core libata floppy [last unloaded: obdecho] CPU: 8 PID: 11727 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.16.0-3.module_el8.7.0+1218+f626c2ff 04/01/2014 task: ffff8802a012a4f0 ti: ffff880072550000 task.ti: ffff880072550000 RIP: 0010:[<ffffffffa164a04c>] [<ffffffffa164a04c>] ll_migrate+0xa4c/0xf90 [lustre] RSP: 0018:ffff880072553bc0 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff8801d6b7c248 RCX: 0000000000000000 RDX: 0000000000000000 RSI: 0000000000000000 RDI: ffff88026f7aea78 RBP: ffff880072553c40 R08: 645f636f695f6c6c R09: 737265765f617461 R10: 617461645f636f69 R11: 6e6f69737265765f R12: ffff8800a6b7f078 R13: 0000000000000000 R14: ffff8802657a8108 R15: ffff8802ba8d0248 FS: 00007f168b73a740(0000) GS:ffff880331c00000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 0000000004408000 CR4: 00000000000007e0 Call Trace: [<ffffffffa1635bcb>] ll_dir_ioctl+0x6a0b/0x7590 [lustre] [<ffffffff81402544>] ? vsnprintf+0x234/0x6a0 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e324e>] ? _raw_spin_unlock+0xe/0x20 [<ffffffffa0246487>] ? cfs_trace_unlock_tcd+0x27/0x70 [libcfs] [<ffffffffa0246d30>] ? libcfs_debug_msg+0x6f0/0xa90 [libcfs] [<ffffffff81416998>] ? debug_check_no_obj_freed+0x178/0x260 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e324e>] ? _raw_spin_unlock+0xe/0x20 [<ffffffff81277f48>] ? __mark_inode_dirty+0x1a8/0x2c0 [<ffffffff8125b47d>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff817df60e>] ? mutex_unlock+0xe/0x10 [<ffffffff8125b89a>] ? iterate_dir+0xaa/0x120 [<ffffffff8125b7d1>] SyS_ioctl+0xa1/0xc0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a | LustreError: 4858:0:(osd_handler.c:221:osd_trans_start()) lustre-MDT0001: failed to start transaction due to ENOSPC LustreError: 4858:0:(osd_handler.c:221:osd_trans_start()) Skipped 2 previous similar messages | Link to test |
sanity test 319: lost lease lock on migrate error | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa160f13c>] ll_migrate+0xa1c/0xfa0 [lustre] PGD 1b1ce067 PUD 93b2f067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC LustreError: 26656:0:(file.c:242:ll_close_inode_openhandle()) lustre-clilmv-ffff8800668a53d8: inode [0x2000059f4:0x92f5:0x0] mdc close failed: rc = -28 Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) lquota(OE) lfsck(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) nfsd ext4 mbcache jbd2 loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) crc32_generic crc_t10dif crct10dif_generic crct10dif_common virtio_balloon virtio_console pcspkr i2c_piix4 ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm ata_piix drm_panel_orientation_quirks serio_raw floppy virtio_blk i2c_core libata [last unloaded: obdecho] CPU: 10 PID: 26641 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.15.0-1.module_el8.6.0+1087+b42c8331 04/01/2014 task: ffff880139e29280 ti: ffff8801a1ad0000 task.ti: ffff8801a1ad0000 RIP: 0010:[<ffffffffa160f13c>] [<ffffffffa160f13c>] ll_migrate+0xa1c/0xfa0 [lustre] RSP: 0018:ffff8801a1ad3bc8 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff88009e525cc8 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff8801040f3c78 RDI: ffff8801040f3c58 RBP: ffff8801a1ad3c48 R08: ffff88015c5225f8 R09: ffff8801a1ad3a60 R10: ffff88009e5266b8 R11: ffff88009e525978 R12: ffff880083ef47d8 R13: 0000000000000000 R14: ffff8802bc416158 R15: ffff8801618551f8 FS: 00007f8a6315f740(0000) GS:ffff880331c80000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 0000000099eee000 CR4: 00000000000007e0 Call Trace: [<ffffffff81242293>] ? __check_object_size+0x1c3/0x220 [<ffffffffa15fc1c6>] ll_dir_ioctl+0x54e6/0x6430 [lustre] [<ffffffff811b630d>] ? find_get_pages_tag+0x10d/0x260 [<ffffffff811c36c1>] ? pagevec_lookup_tag+0x21/0x30 [<ffffffff811b403e>] ? __filemap_fdatawait_range+0xbe/0x1a0 [<ffffffff810b10ca>] ? __queue_delayed_work+0xaa/0x1a0 [<ffffffff810d5999>] ? pick_next_entity+0xa9/0x190 [<ffffffff810df63e>] ? pick_next_task_fair+0x65e/0x8a0 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff8125b47d>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff8127cb80>] ? do_fsync+0xa0/0xa0 [<ffffffff8125b7d1>] SyS_ioctl+0xa1/0xc0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a | LustreError: 12062:0:(osd_handler.c:221:osd_trans_start()) lustre-MDT0000: failed to start transaction due to ENOSPC LustreError: 12062:0:(osd_handler.c:221:osd_trans_start()) Skipped 3 previous similar messages LustreError: 26632:0:(llite_lib.c:1941:ll_md_setattr()) md_setattr fails: rc = -28 LustreError: 26632:0:(file.c:242:ll_close_inode_openhandle()) lustre-clilmv-ffff8800668a53d8: inode [0x2000059f4:0x92f5:0x0] mdc close failed: rc = -28 LustreError: 26641:0:(fail.c:138:__cfs_fail_timeout_set()) cfs_fail_timeout id 32c sleeping for 5000ms LustreError: 26641:0:(fail.c:149:__cfs_fail_timeout_set()) cfs_fail_timeout id 32c awake | Link to test |
sanity test 319: lost lease lock on migrate error | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa15fc13c>] ll_migrate+0xa1c/0xfa0 [lustre] PGD 26a09067 PUD f691067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) lquota(OE) lfsck(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) ext4 mbcache jbd2 loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) crc32_generic crc_t10dif crct10dif_generic crct10dif_common pcspkr virtio_balloon virtio_console i2c_piix4 ip_tables rpcsec_gss_krb5 drm_kms_helper ttm ata_generic pata_acpi drm drm_panel_orientation_quirks ata_piix serio_raw virtio_blk i2c_core libata floppy [last unloaded: obdecho] CPU: 0 PID: 18046 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.15.0-1.module_el8.6.0+1087+b42c8331 04/01/2014 task: ffff88024ac6b760 ti: ffff8800977fc000 task.ti: ffff8800977fc000 RIP: 0010:[<ffffffffa15fc13c>] [<ffffffffa15fc13c>] ll_migrate+0xa1c/0xfa0 [lustre] RSP: 0018:ffff8800977ffbc8 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff88026d5da7c8 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff8802c3cdba98 RDI: ffff8802c3cdba78 RBP: ffff8800977ffc48 R08: ffff88006e85d158 R09: ffff8800977ffa60 R10: ffff88026d5d8358 R11: ffff88026d5dd978 R12: ffff8802798fea38 R13: 0000000000000000 R14: ffff880301946158 R15: ffff88000acb4758 FS: 00007facd0931740(0000) GS:ffff880331a00000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 000000001cdea000 CR4: 00000000000007f0 Call Trace: [<ffffffff81242293>] ? __check_object_size+0x1c3/0x220 [<ffffffffa15e91c6>] ll_dir_ioctl+0x54e6/0x6430 [lustre] [<ffffffff811b630d>] ? find_get_pages_tag+0x10d/0x260 [<ffffffff811c36c1>] ? pagevec_lookup_tag+0x21/0x30 [<ffffffff811b403e>] ? __filemap_fdatawait_range+0xbe/0x1a0 [<ffffffff810b10ca>] ? __queue_delayed_work+0xaa/0x1a0 [<ffffffff817df54c>] ? __mutex_unlock_slowpath+0xdc/0x190 [<ffffffff817df60e>] ? mutex_unlock+0xe/0x10 [<ffffffff81276d80>] ? sync_inodes_sb+0x160/0x430 [<ffffffff810b10ca>] ? __queue_delayed_work+0xaa/0x1a0 [<ffffffff810b1a51>] ? try_to_grab_pending+0xb1/0x180 [<ffffffff8125b47d>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff8125b7d1>] SyS_ioctl+0xa1/0xc0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a | LustreError: 18046:0:(fail.c:138:__cfs_fail_timeout_set()) cfs_fail_timeout id 32c sleeping for 5000ms LustreError: 24508:0:(osd_handler.c:221:osd_trans_start()) lustre-MDT0000: failed to start transaction due to ENOSPC LustreError: 18046:0:(fail.c:149:__cfs_fail_timeout_set()) cfs_fail_timeout id 32c awake LustreError: 18095:0:(file.c:242:ll_close_inode_openhandle()) lustre-clilmv-ffff88029052ca88: inode [0x2000059f4:0x913b:0x0] mdc close failed: rc = -28 | Link to test |
sanity test 319: lost lease lock on migrate error | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa160c1bc>] ll_migrate+0xa1c/0xfa0 [lustre] PGD 184318067 PUD 27bcfd067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) lquota(OE) lfsck(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) ext4 mbcache jbd2 loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) crc32_generic crc_t10dif crct10dif_generic crct10dif_common virtio_balloon pcspkr virtio_console i2c_piix4 ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm ata_piix drm_panel_orientation_quirks virtio_blk serio_raw floppy libata i2c_core [last unloaded: obdecho] CPU: 15 PID: 19719 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.15.0-1.module_el8.6.0+1087+b42c8331 04/01/2014 task: ffff8802711d5c40 ti: ffff880108f64000 task.ti: ffff880108f64000 RIP: 0010:[<ffffffffa160c1bc>] [<ffffffffa160c1bc>] ll_migrate+0xa1c/0xfa0 [lustre] RSP: 0018:ffff880108f67bc8 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff88028063bef8 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff88032763ad78 RDI: ffff88032763ad58 RBP: ffff880108f67c48 R08: ffff8802840d4258 R09: ffff880108f67a60 R10: ffff8801822c5cc8 R11: ffff8801822c7a98 R12: ffff8801c0c50af8 R13: 0000000000000000 R14: ffff880308d40a58 R15: ffff8802c5de4758 FS: 00007fbff2e7c740(0000) GS:ffff880331dc0000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 0000000153788000 CR4: 00000000000007e0 Call Trace: [<ffffffff81242293>] ? __check_object_size+0x1c3/0x220 [<ffffffffa15f91c6>] ll_dir_ioctl+0x54e6/0x6430 [lustre] [<ffffffff811b630d>] ? find_get_pages_tag+0x10d/0x260 [<ffffffff811c36c1>] ? pagevec_lookup_tag+0x21/0x30 [<ffffffff811b403e>] ? __filemap_fdatawait_range+0xbe/0x1a0 [<ffffffff810b10ca>] ? __queue_delayed_work+0xaa/0x1a0 [<ffffffff810d5999>] ? pick_next_entity+0xa9/0x190 [<ffffffff810df63e>] ? pick_next_task_fair+0x65e/0x8a0 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff8125b47d>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff8127cb80>] ? do_fsync+0xa0/0xa0 [<ffffffff8125b7d1>] SyS_ioctl+0xa1/0xc0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a | LustreError: 19708:0:(llite_lib.c:1930:ll_md_setattr()) md_setattr fails: rc = -28 LustreError: 19708:0:(file.c:242:ll_close_inode_openhandle()) lustre-clilmv-ffff880153d5b7e8: inode [0x2000059f4:0x9195:0x0] mdc close failed: rc = -28 LustreError: 19708:0:(file.c:242:ll_close_inode_openhandle()) Skipped 2 previous similar messages LustreError: 19719:0:(fail.c:138:__cfs_fail_timeout_set()) cfs_fail_timeout id 32c sleeping for 5000ms LustreError: 19719:0:(fail.c:149:__cfs_fail_timeout_set()) cfs_fail_timeout id 32c awake LustreError: 19750:0:(file.c:242:ll_close_inode_openhandle()) lustre-clilmv-ffff880153d5b7e8: inode [0x2000059f4:0x9195:0x0] mdc close failed: rc = -28 | Link to test |
sanity test 319: lost lease lock on migrate error | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa15f81bc>] ll_migrate+0xa1c/0xfa0 [lustre] PGD 17c27b067 PUD 278ef1067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) lquota(OE) lfsck(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) ext4 mbcache jbd2 loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) crc32_generic crc_t10dif crct10dif_generic crct10dif_common virtio_balloon virtio_console i2c_piix4 pcspkr ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm ata_piix drm_panel_orientation_quirks libata virtio_blk serio_raw i2c_core floppy [last unloaded: obdecho] CPU: 7 PID: 24403 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.15.0-1.module_el8.6.0+1087+b42c8331 04/01/2014 task: ffff88028d941280 ti: ffff88019d954000 task.ti: ffff88019d954000 RIP: 0010:[<ffffffffa15f81bc>] [<ffffffffa15f81bc>] ll_migrate+0xa1c/0xfa0 [lustre] RSP: 0018:ffff88019d957bc8 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff88014133ea08 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff8802fb8b1c08 RDI: ffff8802fb8b1be8 RBP: ffff88019d957c48 R08: ffff88008baaf138 R09: ffff88019d957a60 R10: ffff880141339a88 R11: ffff8801413393e8 R12: ffff8800aa318328 R13: 0000000000000000 R14: ffff8801b2fb0a58 R15: ffff88018dc5b218 FS: 00007f4dd5ce6740(0000) GS:ffff880331bc0000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 00000002e2524000 CR4: 00000000000007e0 Call Trace: [<ffffffff81242293>] ? __check_object_size+0x1c3/0x220 [<ffffffffa15e51c6>] ll_dir_ioctl+0x54e6/0x6430 [lustre] [<ffffffff8103de0f>] ? save_stack_trace+0x1f/0x30 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817d5802>] ? free_debug_processing+0x240/0x272 [<ffffffff8108982f>] ? __mmdrop+0x6f/0xa0 [<ffffffff8121e4c4>] ? __slab_free+0xa4/0x300 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff81416998>] ? debug_check_no_obj_freed+0x178/0x260 [<ffffffff8108982f>] ? __mmdrop+0x6f/0xa0 [<ffffffff8121e991>] ? kmem_cache_free+0x271/0x290 [<ffffffff8108982f>] ? __mmdrop+0x6f/0xa0 [<ffffffff810c844d>] ? finish_task_switch+0x15d/0x1b0 [<ffffffff8125b47d>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff8127cb80>] ? do_fsync+0xa0/0xa0 [<ffffffff8125b7d1>] SyS_ioctl+0xa1/0xc0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a | LustreError: 24374:0:(llite_lib.c:1930:ll_md_setattr()) md_setattr fails: rc = -28 LustreError: 24374:0:(file.c:242:ll_close_inode_openhandle()) lustre-clilmv-ffff88013d262e98: inode [0x2000059f4:0x9197:0x0] mdc close failed: rc = -28 LustreError: 24403:0:(fail.c:138:__cfs_fail_timeout_set()) cfs_fail_timeout id 32c sleeping for 5000ms LustreError: 24403:0:(fail.c:149:__cfs_fail_timeout_set()) cfs_fail_timeout id 32c awake LustreError: 24441:0:(file.c:242:ll_close_inode_openhandle()) lustre-clilmv-ffff88013d262e98: inode [0x2000059f4:0x9197:0x0] mdc close failed: rc = -28 | Link to test |
sanity test 319: lost lease lock on migrate error | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa161013c>] ll_migrate+0xa1c/0xfa0 [lustre] PGD 216080067 PUD 15a2f4067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) lquota(OE) lfsck(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) nfsd ext4 mbcache jbd2 loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) crc32_generic crc_t10dif crct10dif_generic crct10dif_common pcspkr i2c_piix4 virtio_console virtio_balloon ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm ata_piix drm_panel_orientation_quirks serio_raw virtio_blk libata i2c_core floppy [last unloaded: obdecho] CPU: 1 PID: 27642 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.15.0-1.module_el8.6.0+1087+b42c8331 04/01/2014 task: ffff8802c816dc40 ti: ffff88017c4f4000 task.ti: ffff88017c4f4000 RIP: 0010:[<ffffffffa161013c>] [<ffffffffa161013c>] ll_migrate+0xa1c/0xfa0 [lustre] RSP: 0018:ffff88017c4f7bc8 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff88009b33bba8 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff880272ffa038 RDI: ffff880272ffa018 RBP: ffff88017c4f7c48 R08: ffff880087c147f8 R09: ffff88017c4f7a60 R10: ffff88009b338008 R11: ffff88009b339dd8 R12: ffff8802592b1458 R13: 0000000000000000 R14: ffff8800a98ae158 R15: ffff880325022cc8 FS: 00007fca1599d740(0000) GS:ffff880331a40000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 0000000169dda000 CR4: 00000000000007e0 Call Trace: [<ffffffff81242293>] ? __check_object_size+0x1c3/0x220 [<ffffffffa15fd1c6>] ll_dir_ioctl+0x54e6/0x6430 [lustre] [<ffffffff810d5999>] ? pick_next_entity+0xa9/0x190 [<ffffffff810d7e3c>] ? set_next_entity+0x3c/0xe0 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e329e>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffff810d5999>] ? pick_next_entity+0xa9/0x190 [<ffffffff810df63e>] ? pick_next_task_fair+0x65e/0x8a0 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff8125b47d>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff8127cb80>] ? do_fsync+0xa0/0xa0 [<ffffffff8125b7d1>] SyS_ioctl+0xa1/0xc0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a | LustreError: 27628:0:(llite_lib.c:1927:ll_md_setattr()) md_setattr fails: rc = -28 LustreError: 27628:0:(file.c:242:ll_close_inode_openhandle()) lustre-clilmv-ffff8801e1cd4a88: inode [0x2000059f4:0x8ca4:0x0] mdc close failed: rc = -28 LustreError: 27628:0:(file.c:242:ll_close_inode_openhandle()) Skipped 2 previous similar messages LustreError: 27642:0:(fail.c:138:__cfs_fail_timeout_set()) cfs_fail_timeout id 32c sleeping for 5000ms LustreError: 27642:0:(fail.c:149:__cfs_fail_timeout_set()) cfs_fail_timeout id 32c awake LustreError: 27674:0:(file.c:242:ll_close_inode_openhandle()) lustre-clilmv-ffff8801e1cd4a88: inode [0x2000059f4:0x8ca4:0x0] mdc close failed: rc = -28 | Link to test |
sanity test 319: lost lease lock on migrate error | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa161e1bc>] ll_migrate+0xa1c/0xfa0 [lustre] PGD 2e9e36067 PUD 14f3c2067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) lquota(OE) lfsck(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) nfsd ext4 mbcache jbd2 loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) crc32_generic crc_t10dif crct10dif_generic crct10dif_common virtio_console virtio_balloon i2c_piix4 pcspkr ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm ata_piix drm_panel_orientation_quirks libata serio_raw virtio_blk i2c_core floppy [last unloaded: obdecho] CPU: 9 PID: 8571 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.15.0-1.module_el8.6.0+1087+b42c8331 04/01/2014 task: ffff88016b9e9280 ti: ffff8801f07e8000 task.ti: ffff8801f07e8000 RIP: 0010:[<ffffffffa161e1bc>] [<ffffffffa161e1bc>] ll_migrate+0xa1c/0xfa0 [lustre] RSP: 0018:ffff8801f07ebbc8 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff8802a99a1738 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff8802ca745188 RDI: ffff8802ca745168 RBP: ffff8801f07ebc48 R08: ffff8800090beb98 R09: ffff8801f07eba60 R10: ffff8802a99a6018 R11: ffff8802a99a0008 R12: ffff8802568e9c28 R13: 0000000000000000 R14: ffff88023b119cd8 R15: ffff88026fee5748 FS: 00007f499cb7d740(0000) GS:ffff880331c40000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 00000002568be000 CR4: 00000000000007e0 Call Trace: [<ffffffff81242293>] ? __check_object_size+0x1c3/0x220 [<ffffffffa160b1c6>] ll_dir_ioctl+0x54e6/0x6430 [lustre] [<ffffffff811b630d>] ? find_get_pages_tag+0x10d/0x260 [<ffffffff811c36c1>] ? pagevec_lookup_tag+0x21/0x30 [<ffffffff811b403e>] ? __filemap_fdatawait_range+0xbe/0x1a0 [<ffffffff810b10ca>] ? __queue_delayed_work+0xaa/0x1a0 [<ffffffff817df54c>] ? __mutex_unlock_slowpath+0xdc/0x190 [<ffffffff817df60e>] ? mutex_unlock+0xe/0x10 [<ffffffff81276d80>] ? sync_inodes_sb+0x160/0x430 [<ffffffff810b10ca>] ? __queue_delayed_work+0xaa/0x1a0 [<ffffffff810b1a51>] ? try_to_grab_pending+0xb1/0x180 [<ffffffff8125b47d>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff8125b7d1>] SyS_ioctl+0xa1/0xc0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a | LustreError: 8540:0:(llite_lib.c:1930:ll_md_setattr()) md_setattr fails: rc = -28 LustreError: 8540:0:(file.c:242:ll_close_inode_openhandle()) lustre-clilmv-ffff8801c9ef4a88: inode [0x2000059f4:0x924c:0x0] mdc close failed: rc = -28 LustreError: 8571:0:(fail.c:138:__cfs_fail_timeout_set()) cfs_fail_timeout id 32c sleeping for 5000ms LustreError: 24494:0:(osd_handler.c:221:osd_trans_start()) lustre-MDT0000: failed to start transaction due to ENOSPC LustreError: 24494:0:(osd_handler.c:221:osd_trans_start()) Skipped 11 previous similar messages LustreError: 8571:0:(fail.c:149:__cfs_fail_timeout_set()) cfs_fail_timeout id 32c awake LustreError: 8610:0:(file.c:242:ll_close_inode_openhandle()) lustre-clilmv-ffff8801c9ef4a88: inode [0x2000059f4:0x924c:0x0] mdc close failed: rc = -28 | Link to test |
sanity test 319: lost lease lock on migrate error | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa16001bc>] ll_migrate+0xa1c/0xfa0 [lustre] PGD 2ddf30067 PUD 28d201067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) lquota(OE) lfsck(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) ext4 mbcache jbd2 loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) crc32_generic crc_t10dif crct10dif_generic crct10dif_common i2c_piix4 pcspkr virtio_console virtio_balloon ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm ata_piix drm_panel_orientation_quirks virtio_blk i2c_core serio_raw floppy libata [last unloaded: obdecho] CPU: 14 PID: 32421 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.15.0-1.module_el8.6.0+1087+b42c8331 04/01/2014 task: ffff88011fc39280 ti: ffff880181404000 task.ti: ffff880181404000 RIP: 0010:[<ffffffffa16001bc>] [<ffffffffa16001bc>] ll_migrate+0xa1c/0xfa0 [lustre] RSP: 0018:ffff880181407bc8 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff8802f6b4c8e8 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff8802d0e3e268 RDI: ffff8802d0e3e248 RBP: ffff880181407c48 R08: ffff880064c4fc78 R09: ffff880181407a60 R10: ffff8802f6b489f8 R11: ffff8802f6b4c598 R12: ffff8800828af078 R13: 0000000000000000 R14: ffff8802dbae7018 R15: ffff88012452c208 FS: 00007f8c8ec86740(0000) GS:ffff880331d80000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 0000000273e5e000 CR4: 00000000000007e0 Call Trace: [<ffffffff81242293>] ? __check_object_size+0x1c3/0x220 [<ffffffffa15ed1c6>] ll_dir_ioctl+0x54e6/0x6430 [lustre] [<ffffffff811b630d>] ? find_get_pages_tag+0x10d/0x260 [<ffffffff811c36c1>] ? pagevec_lookup_tag+0x21/0x30 [<ffffffff811b403e>] ? __filemap_fdatawait_range+0xbe/0x1a0 [<ffffffff810d5999>] ? pick_next_entity+0xa9/0x190 [<ffffffff810df63e>] ? pick_next_task_fair+0x65e/0x8a0 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff8125b47d>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff8127cb80>] ? do_fsync+0xa0/0xa0 [<ffffffff8125b7d1>] SyS_ioctl+0xa1/0xc0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a | LustreError: 32421:0:(fail.c:138:__cfs_fail_timeout_set()) cfs_fail_timeout id 32c sleeping for 5000ms LustreError: 21510:0:(osd_handler.c:221:osd_trans_start()) lustre-MDT0000: failed to start transaction due to ENOSPC LustreError: 32421:0:(fail.c:149:__cfs_fail_timeout_set()) cfs_fail_timeout id 32c awake LustreError: 32495:0:(file.c:242:ll_close_inode_openhandle()) lustre-clilmv-ffff88031ccc9bf8: inode [0x2000059f4:0x9247:0x0] mdc close failed: rc = -28 LustreError: 32495:0:(file.c:242:ll_close_inode_openhandle()) Skipped 2 previous similar messages | Link to test |
sanity test 319: lost lease lock on migrate error | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa15ff17c>] ll_migrate+0xa1c/0xfa0 [lustre] PGD 26d6e067 PUD 81037067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) lquota(OE) lfsck(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) nfsd ext4 mbcache jbd2 loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) crc32_generic crc_t10dif crct10dif_generic crct10dif_common virtio_balloon virtio_console pcspkr i2c_piix4 ip_tables rpcsec_gss_krb5 ata_generic drm_kms_helper pata_acpi ttm drm ata_piix drm_panel_orientation_quirks serio_raw virtio_blk floppy i2c_core libata [last unloaded: obdecho] CPU: 9 PID: 15881 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.15.0-1.module_el8.6.0+1087+b42c8331 04/01/2014 task: ffff8801c4a71280 ti: ffff8800a760c000 task.ti: ffff8800a760c000 RIP: 0010:[<ffffffffa15ff17c>] [<ffffffffa15ff17c>] ll_migrate+0xa1c/0xfa0 [lustre] RSP: 0018:ffff8800a760fbc8 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff8801858a06a8 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff880318d41718 RDI: ffff880318d416f8 RBP: ffff8800a760fc48 R08: ffff88031aa10f78 R09: ffff8800a760fa60 R10: ffff8801858a3858 R11: ffff8801858a3508 R12: ffff880273bb84b8 R13: 0000000000000000 R14: ffff88021cf6b898 R15: ffff88028b1fb768 FS: 00007f5bd1f5c740(0000) GS:ffff880331c40000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 00000000187c2000 CR4: 00000000000007e0 Call Trace: [<ffffffff81242293>] ? __check_object_size+0x1c3/0x220 [<ffffffffa15ec1c6>] ll_dir_ioctl+0x54e6/0x6430 [lustre] [<ffffffff817e05ca>] ? __schedule+0x32a/0x7d0 [<ffffffff8127c9c0>] ? generic_write_sync+0x60/0x60 [<ffffffff817e0a99>] ? schedule+0x29/0x70 [<ffffffff817dd739>] ? schedule_timeout+0x209/0x290 [<ffffffff817e324e>] ? _raw_spin_unlock+0xe/0x20 [<ffffffff810b0d01>] ? __queue_work+0x151/0x420 [<ffffffff817df54c>] ? __mutex_unlock_slowpath+0xdc/0x190 [<ffffffff817df60e>] ? mutex_unlock+0xe/0x10 [<ffffffff81276d80>] ? sync_inodes_sb+0x160/0x430 [<ffffffff810b10ca>] ? __queue_delayed_work+0xaa/0x1a0 [<ffffffff810b1a51>] ? try_to_grab_pending+0xb1/0x180 [<ffffffff8125b47d>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff8125b7d1>] SyS_ioctl+0xa1/0xc0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a | LustreError: 15816:0:(llite_lib.c:1930:ll_md_setattr()) md_setattr fails: rc = -28 LustreError: 15816:0:(file.c:242:ll_close_inode_openhandle()) lustre-clilmv-ffff88005b699bf8: inode [0x200005224:0x9101:0x0] mdc close failed: rc = -28 LustreError: 15881:0:(fail.c:138:__cfs_fail_timeout_set()) cfs_fail_timeout id 32c sleeping for 5000ms LustreError: 15881:0:(fail.c:149:__cfs_fail_timeout_set()) cfs_fail_timeout id 32c awake LustreError: 15953:0:(file.c:242:ll_close_inode_openhandle()) lustre-clilmv-ffff88005b699bf8: inode [0x200005224:0x9101:0x0] mdc close failed: rc = -28 | Link to test |
sanity test 319: lost lease lock on migrate error | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa160d17c>] ll_migrate+0xa1c/0xfa0 [lustre] PGD 28acdb067 PUD 12bbcc067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) lquota(OE) lfsck(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) nfsd ext4 mbcache jbd2 loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) crc32_generic crc_t10dif crct10dif_generic crct10dif_common pcspkr virtio_balloon i2c_piix4 virtio_console ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm ata_piix drm_panel_orientation_quirks floppy serio_raw virtio_blk libata i2c_core [last unloaded: obdecho] CPU: 6 PID: 28577 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.15.0-1.module_el8.6.0+1087+b42c8331 04/01/2014 task: ffff8801fb29a4f0 ti: ffff8801f4538000 task.ti: ffff8801f4538000 RIP: 0010:[<ffffffffa160d17c>] [<ffffffffa160d17c>] ll_migrate+0xa1c/0xfa0 [lustre] RSP: 0018:ffff8801f453bbc8 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff88013942b1b8 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff88020a39d0f8 RDI: ffff88020a39d0d8 RBP: ffff8801f453bc48 R08: ffff88031e3ced78 R09: ffff8801f453ba60 R10: ffff88013942ae68 R11: ffff88013942cf88 R12: ffff8802861d2d58 R13: 0000000000000000 R14: ffff8801ba8fa618 R15: ffff880017029cd8 FS: 00007f0ce0e42740(0000) GS:ffff880331b80000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 00000001d5ca0000 CR4: 00000000000007e0 Call Trace: [<ffffffff81242293>] ? __check_object_size+0x1c3/0x220 [<ffffffffa15fa1c6>] ll_dir_ioctl+0x54e6/0x6430 [lustre] [<ffffffff811b630d>] ? find_get_pages_tag+0x10d/0x260 [<ffffffff811c36c1>] ? pagevec_lookup_tag+0x21/0x30 [<ffffffff811b403e>] ? __filemap_fdatawait_range+0xbe/0x1a0 [<ffffffff810b10ca>] ? __queue_delayed_work+0xaa/0x1a0 [<ffffffff817df54c>] ? __mutex_unlock_slowpath+0xdc/0x190 [<ffffffff817df60e>] ? mutex_unlock+0xe/0x10 [<ffffffff81276d80>] ? sync_inodes_sb+0x160/0x430 [<ffffffff8125b47d>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff8125b7d1>] SyS_ioctl+0xa1/0xc0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a | LustreError: 11977:0:(osd_handler.c:221:osd_trans_start()) lustre-MDT0000: failed to start transaction due to ENOSPC LustreError: 11977:0:(osd_handler.c:221:osd_trans_start()) Skipped 13 previous similar messages LustreError: 28536:0:(llite_lib.c:1930:ll_md_setattr()) md_setattr fails: rc = -28 LustreError: 28536:0:(file.c:242:ll_close_inode_openhandle()) lustre-clilmv-ffff88009ef84138: inode [0x2000059f4:0x8b59:0x0] mdc close failed: rc = -28 LustreError: 28577:0:(fail.c:138:__cfs_fail_timeout_set()) cfs_fail_timeout id 32c sleeping for 5000ms LustreError: 28577:0:(fail.c:149:__cfs_fail_timeout_set()) cfs_fail_timeout id 32c awake LustreError: 28610:0:(file.c:242:ll_close_inode_openhandle()) lustre-clilmv-ffff88009ef84138: inode [0x2000059f4:0x8b59:0x0] mdc close failed: rc = -28 | Link to test |
sanity test 319: lost lease lock on migrate error | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa15f358c>] ll_migrate+0xa1c/0xfa0 [lustre] PGD a7fc4067 PUD 8b662067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) lquota(OE) lfsck(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) ext4 mbcache jbd2 loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) crc32_generic crc_t10dif crct10dif_generic crct10dif_common virtio_console virtio_balloon pcspkr i2c_piix4 ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm drm_panel_orientation_quirks ata_piix virtio_blk serio_raw i2c_core floppy libata [last unloaded: obdecho] CPU: 2 PID: 29619 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.15.0-1.module_el8.6.0+1087+b42c8331 04/01/2014 task: ffff88031d23a4f0 ti: ffff880048d74000 task.ti: ffff880048d74000 RIP: 0010:[<ffffffffa15f358c>] [<ffffffffa15f358c>] ll_migrate+0xa1c/0xfa0 [lustre] RSP: 0018:ffff880048d77bc8 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff8801a13a9dd8 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff8800b46f2068 RDI: ffff8800b46f2048 RBP: ffff880048d77c48 R08: ffff880085745518 R09: ffff880048d77a60 R10: ffff88011b0fbef8 R11: ffff88011b0fd628 R12: ffff8802ea8a95e8 R13: 0000000000000000 R14: ffff880038ca1398 R15: ffff8801708e9788 FS: 00007f5565026740(0000) GS:ffff880331a80000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 000000009b006000 CR4: 00000000000007e0 Call Trace: [<ffffffff81242293>] ? __check_object_size+0x1c3/0x220 [<ffffffffa15e01c6>] ll_dir_ioctl+0x54e6/0x6430 [lustre] [<ffffffff811b630d>] ? find_get_pages_tag+0x10d/0x260 [<ffffffff811c36c1>] ? pagevec_lookup_tag+0x21/0x30 [<ffffffff811b403e>] ? __filemap_fdatawait_range+0xbe/0x1a0 [<ffffffff810b10ca>] ? __queue_delayed_work+0xaa/0x1a0 [<ffffffff810d5999>] ? pick_next_entity+0xa9/0x190 [<ffffffff810df63e>] ? pick_next_task_fair+0x65e/0x8a0 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff8125b47d>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff8127cb80>] ? do_fsync+0xa0/0xa0 [<ffffffff8125b7d1>] SyS_ioctl+0xa1/0xc0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a | LustreError: 29599:0:(llite_lib.c:1927:ll_md_setattr()) md_setattr fails: rc = -28 LustreError: 29599:0:(file.c:242:ll_close_inode_openhandle()) lustre-clilmv-ffff88027279efc8: inode [0x2000059f4:0x92f8:0x0] mdc close failed: rc = -28 LustreError: 29619:0:(fail.c:138:__cfs_fail_timeout_set()) cfs_fail_timeout id 32c sleeping for 5000ms LustreError: 29619:0:(fail.c:149:__cfs_fail_timeout_set()) cfs_fail_timeout id 32c awake LustreError: 20381:0:(osd_handler.c:221:osd_trans_start()) lustre-MDT0000: failed to start transaction due to ENOSPC LustreError: 20381:0:(osd_handler.c:221:osd_trans_start()) Skipped 10 previous similar messages LustreError: 29639:0:(file.c:242:ll_close_inode_openhandle()) lustre-clilmv-ffff88027279efc8: inode [0x2000059f4:0x92f8:0x0] mdc close failed: rc = -28 | Link to test |
sanity test 319: lost lease lock on migrate error | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa160d20c>] ll_migrate+0xa1c/0xfa0 [lustre] PGD 8b033067 PUD a2427067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) lquota(OE) lfsck(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) ext4 mbcache jbd2 loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) crc32_generic crc_t10dif crct10dif_generic crct10dif_common virtio_balloon pcspkr virtio_console i2c_piix4 ip_tables rpcsec_gss_krb5 drm_kms_helper ata_generic ttm pata_acpi drm drm_panel_orientation_quirks ata_piix virtio_blk serio_raw i2c_core floppy libata [last unloaded: obdecho] CPU: 15 PID: 2830 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.15.0-1.module_el8.6.0+1087+b42c8331 04/01/2014 task: ffff88024bc45c40 ti: ffff8800854c8000 task.ti: ffff8800854c8000 RIP: 0010:[<ffffffffa160d20c>] [<ffffffffa160d20c>] ll_migrate+0xa1c/0xfa0 [lustre] RSP: 0018:ffff8800854cbbc8 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff88011706a128 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff88027c101d88 RDI: ffff88027c101d68 RBP: ffff8800854cbc48 R08: ffff88028b057318 R09: ffff8800854cba60 R10: ffff880117068358 R11: ffff88011706b1b8 R12: ffff8800828e5458 R13: 0000000000000000 R14: ffff880184e07018 R15: ffff8802cd258ce8 FS: 00007fc1c7dc6740(0000) GS:ffff880331dc0000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 000000000e6ce000 CR4: 00000000000007e0 Call Trace: [<ffffffff81242293>] ? __check_object_size+0x1c3/0x220 [<ffffffffa15fa1c6>] ll_dir_ioctl+0x54e6/0x6430 [lustre] [<ffffffff811b630d>] ? find_get_pages_tag+0x10d/0x260 [<ffffffff811c36c1>] ? pagevec_lookup_tag+0x21/0x30 [<ffffffff811b403e>] ? __filemap_fdatawait_range+0xbe/0x1a0 [<ffffffff817df54c>] ? __mutex_unlock_slowpath+0xdc/0x190 [<ffffffff817df60e>] ? mutex_unlock+0xe/0x10 [<ffffffff81276d80>] ? sync_inodes_sb+0x160/0x430 [<ffffffff8125b47d>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff8125b7d1>] SyS_ioctl+0xa1/0xc0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a | LustreError: 2792:0:(llite_lib.c:1922:ll_md_setattr()) md_setattr fails: rc = -28 LustreError: 2792:0:(file.c:242:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e63ba548: inode [0x2000059f4:0x93be:0x0] mdc close failed: rc = -28 LustreError: 2792:0:(file.c:242:ll_close_inode_openhandle()) Skipped 2 previous similar messages LustreError: 2830:0:(fail.c:138:__cfs_fail_timeout_set()) cfs_fail_timeout id 32c sleeping for 5000ms LustreError: 2830:0:(fail.c:149:__cfs_fail_timeout_set()) cfs_fail_timeout id 32c awake LustreError: 2880:0:(file.c:242:ll_close_inode_openhandle()) lustre-clilmv-ffff8802e63ba548: inode [0x2000059f4:0x93be:0x0] mdc close failed: rc = -28 | Link to test |
sanity test 319: lost lease lock on migrate error | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa160f20c>] ll_migrate+0xa1c/0xfa0 [lustre] PGD 15b52a067 PUD 27904f067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) lquota(OE) lfsck(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) ext4 mbcache jbd2 loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) crc32_generic crc_t10dif crct10dif_generic crct10dif_common pcspkr i2c_piix4 virtio_balloon virtio_console ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm ata_piix drm_panel_orientation_quirks virtio_blk serio_raw i2c_core libata floppy [last unloaded: obdecho] CPU: 9 PID: 5261 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.15.0-1.module_el8.6.0+1087+b42c8331 04/01/2014 task: ffff88024c739280 ti: ffff88011eab4000 task.ti: ffff88011eab4000 RIP: 0010:[<ffffffffa160f20c>] [<ffffffffa160f20c>] ll_migrate+0xa1c/0xfa0 [lustre] RSP: 0018:ffff88011eab7bc8 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff8801199ccc38 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff88027b23f168 RDI: ffff88027b23f148 RBP: ffff88011eab7c48 R08: ffff880287d75518 R09: ffff88011eab7a60 R10: ffff8801269b3ba8 R11: ffff8801269b1738 R12: ffff88015a6ee268 R13: 0000000000000000 R14: ffff8801f18a5d98 R15: ffff880275995748 FS: 00007f0784fbf740(0000) GS:ffff880331c40000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 00000002efed6000 CR4: 00000000000007e0 Call Trace: [<ffffffff81242293>] ? __check_object_size+0x1c3/0x220 [<ffffffffa15fc1c6>] ll_dir_ioctl+0x54e6/0x6430 [lustre] [<ffffffff817e05ca>] ? __schedule+0x32a/0x7d0 [<ffffffff8127c9c0>] ? generic_write_sync+0x60/0x60 [<ffffffff817e0a99>] ? schedule+0x29/0x70 [<ffffffff817dd739>] ? schedule_timeout+0x209/0x290 [<ffffffff817e324e>] ? _raw_spin_unlock+0xe/0x20 [<ffffffff810b0d01>] ? __queue_work+0x151/0x420 [<ffffffff810d5999>] ? pick_next_entity+0xa9/0x190 [<ffffffff810df63e>] ? pick_next_task_fair+0x65e/0x8a0 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff8125b47d>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff8127cb80>] ? do_fsync+0xa0/0xa0 [<ffffffff8125b7d1>] SyS_ioctl+0xa1/0xc0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a | LustreError: 5261:0:(fail.c:138:__cfs_fail_timeout_set()) cfs_fail_timeout id 32c sleeping for 5000ms LustreError: 5261:0:(fail.c:149:__cfs_fail_timeout_set()) cfs_fail_timeout id 32c awake LustreError: 5328:0:(file.c:242:ll_close_inode_openhandle()) lustre-clilmv-ffff8802cd2c8958: inode [0x200005224:0x9231:0x0] mdc close failed: rc = -28 LustreError: 5328:0:(file.c:242:ll_close_inode_openhandle()) Skipped 2 previous similar messages | Link to test |
sanity test 319: lost lease lock on migrate error | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa161041c>] ll_migrate+0xa1c/0xfa0 [lustre] PGD 2928ad067 PUD 186730067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) lquota(OE) lfsck(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) nfsd ext4 mbcache jbd2 loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) crc32_generic crc_t10dif crct10dif_generic crct10dif_common virtio_balloon i2c_piix4 pcspkr virtio_console ip_tables rpcsec_gss_krb5 drm_kms_helper ata_generic pata_acpi ttm drm ata_piix drm_panel_orientation_quirks serio_raw i2c_core virtio_blk libata floppy [last unloaded: obdecho] CPU: 15 PID: 21803 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.15.0-1.module_el8.6.0+1087+b42c8331 04/01/2014 task: ffff88028805c9d0 ti: ffff88025e4ac000 task.ti: ffff88025e4ac000 RIP: 0010:[<ffffffffa161041c>] [<ffffffffa161041c>] ll_migrate+0xa1c/0xfa0 [lustre] RSP: 0018:ffff88025e4afbc8 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff8802f14a3508 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff8802be6f4808 RDI: ffff8802be6f47e8 RBP: ffff88025e4afc48 R08: ffff8800890ebe58 R09: ffff88025e4afa60 R10: ffff8802f14a4c38 R11: ffff8802f14a5628 R12: ffff880280999458 R13: 0000000000000000 R14: ffff8800786d1398 R15: ffff88028aa96c88 FS: 00007f385a83d740(0000) GS:ffff880331dc0000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 000000028df44000 CR4: 00000000000007e0 Call Trace: [<ffffffff81242293>] ? __check_object_size+0x1c3/0x220 [<ffffffffa15fccf1>] ll_dir_ioctl+0x5e71/0x74a0 [lustre] [<ffffffff811b630d>] ? find_get_pages_tag+0x10d/0x260 [<ffffffff811c36c1>] ? pagevec_lookup_tag+0x21/0x30 [<ffffffff811b403e>] ? __filemap_fdatawait_range+0xbe/0x1a0 [<ffffffff810b10ca>] ? __queue_delayed_work+0xaa/0x1a0 [<ffffffff817df54c>] ? __mutex_unlock_slowpath+0xdc/0x190 [<ffffffff817df60e>] ? mutex_unlock+0xe/0x10 [<ffffffff81276d80>] ? sync_inodes_sb+0x160/0x430 [<ffffffff810b10ca>] ? __queue_delayed_work+0xaa/0x1a0 [<ffffffff810b1a51>] ? try_to_grab_pending+0xb1/0x180 [<ffffffff8125b47d>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff8125b7d1>] SyS_ioctl+0xa1/0xc0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a | LustreError: 21803:0:(fail.c:138:__cfs_fail_timeout_set()) cfs_fail_timeout id 32c sleeping for 5000ms LustreError: 8190:0:(osd_handler.c:221:osd_trans_start()) lustre-MDT0000: failed to start transaction due to ENOSPC LustreError: 8190:0:(osd_handler.c:221:osd_trans_start()) Skipped 4 previous similar messages LustreError: 21803:0:(fail.c:149:__cfs_fail_timeout_set()) cfs_fail_timeout id 32c awake LustreError: 21837:0:(file.c:242:ll_close_inode_openhandle()) lustre-clilmv-ffff880283048008: inode [0x2000059f4:0x91fa:0x0] mdc close failed: rc = -28 | Link to test |
sanity test 319: lost lease lock on migrate error | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa161135c>] ll_migrate+0xa1c/0xfa0 [lustre] PGD 2ce69e067 PUD 23cad9067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) lquota(OE) lfsck(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) nfsd ext4 mbcache jbd2 loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) crc32_generic crc_t10dif crct10dif_generic crct10dif_common pcspkr virtio_console virtio_balloon i2c_piix4 ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm ata_piix drm_panel_orientation_quirks i2c_core floppy serio_raw virtio_blk libata [last unloaded: obdecho] CPU: 2 PID: 10765 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.15.0-1.module_el8.6.0+1087+b42c8331 04/01/2014 task: ffff8802cfe1dc40 ti: ffff88020b0d0000 task.ti: ffff88020b0d0000 RIP: 0010:[<ffffffffa161135c>] [<ffffffffa161135c>] ll_migrate+0xa1c/0xfa0 [lustre] RSP: 0018:ffff88020b0d3bc8 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff8801731ea128 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff88030f93c048 RDI: ffff88030f93c028 RBP: ffff88020b0d3c48 R08: ffff88021372d8d8 R09: ffff88020b0d3a60 R10: ffff8801731ec248 R11: ffff8801731e9dd8 R12: ffff88026981b848 R13: 0000000000000000 R14: ffff88001b85cb18 R15: ffff88028e083768 FS: 00007fb833f78740(0000) GS:ffff880331a80000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 0000000274486000 CR4: 00000000000007e0 Call Trace: [<ffffffff81242293>] ? __check_object_size+0x1c3/0x220 [<ffffffffa15fdc7f>] ll_dir_ioctl+0x5ebf/0x74a0 [lustre] [<ffffffff811b630d>] ? find_get_pages_tag+0x10d/0x260 [<ffffffff811c36c1>] ? pagevec_lookup_tag+0x21/0x30 [<ffffffff811b403e>] ? __filemap_fdatawait_range+0xbe/0x1a0 [<ffffffff810d5999>] ? pick_next_entity+0xa9/0x190 [<ffffffff810df63e>] ? pick_next_task_fair+0x65e/0x8a0 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff8125b47d>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff8127cb80>] ? do_fsync+0xa0/0xa0 [<ffffffff8125b7d1>] SyS_ioctl+0xa1/0xc0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a | LustreError: 10735:0:(llite_lib.c:1921:ll_md_setattr()) md_setattr fails: rc = -28 LustreError: 10735:0:(file.c:242:ll_close_inode_openhandle()) lustre-clilmv-ffff8802f74cd3d8: inode [0x2000059f4:0x9403:0x0] mdc close failed: rc = -28 LustreError: 10765:0:(fail.c:138:__cfs_fail_timeout_set()) cfs_fail_timeout id 32c sleeping for 5000ms LustreError: 10765:0:(fail.c:149:__cfs_fail_timeout_set()) cfs_fail_timeout id 32c awake LustreError: 30277:0:(osd_handler.c:221:osd_trans_start()) lustre-MDT0000: failed to start transaction due to ENOSPC LustreError: 30277:0:(osd_handler.c:221:osd_trans_start()) Skipped 7 previous similar messages LustreError: 10784:0:(file.c:242:ll_close_inode_openhandle()) lustre-clilmv-ffff8802f74cd3d8: inode [0x2000059f4:0x9403:0x0] mdc close failed: rc = -28 | Link to test |
sanity test 319: lost lease lock on migrate error | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa160d2ac>] ll_migrate+0xa1c/0xfa0 [lustre] PGD 2e3fec067 PUD 1553f8067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC LustreError: 9580:0:(file.c:242:ll_close_inode_openhandle()) lustre-clilmv-ffff8801bcf74a88: inode [0x2000059f4:0x9399:0x0] mdc close failed: rc = -28 Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) lquota(OE) lfsck(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) nfsd ext4 mbcache jbd2 loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) crc32_generic crc_t10dif crct10dif_generic crct10dif_common i2c_piix4 pcspkr virtio_console virtio_balloon ip_tables rpcsec_gss_krb5 ata_generic drm_kms_helper pata_acpi ttm drm ata_piix drm_panel_orientation_quirks serio_raw virtio_blk i2c_core libata floppy [last unloaded: obdecho] CPU: 1 PID: 9543 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.15.0-1.module_el8.6.0+1087+b42c8331 04/01/2014 task: ffff88031cdd1280 ti: ffff88010141c000 task.ti: ffff88010141c000 RIP: 0010:[<ffffffffa160d2ac>] [<ffffffffa160d2ac>] ll_migrate+0xa1c/0xfa0 [lustre] RSP: 0018:ffff88010141fbc8 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff8801707e31b8 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff880139b36fe8 RDI: ffff880139b36fc8 RBP: ffff88010141fc48 R08: ffff8801c0ef5c98 R09: 0000000000000000 R10: ffff8800991da7c8 R11: ffff8800991db508 R12: ffff88000c19da98 R13: 0000000000000000 R14: ffff880061f25458 R15: ffff8800240c0248 FS: 00007f07f9a88740(0000) GS:ffff880331a40000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 0000000135b6c000 CR4: 00000000000007e0 Call Trace: [<ffffffff81242293>] ? __check_object_size+0x1c3/0x220 [<ffffffffa15f9c7f>] ll_dir_ioctl+0x5ebf/0x74a0 [lustre] [<ffffffff811b630d>] ? find_get_pages_tag+0x10d/0x260 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff811c36c1>] ? pagevec_lookup_tag+0x21/0x30 [<ffffffff811b403e>] ? __filemap_fdatawait_range+0xbe/0x1a0 [<ffffffff810d5999>] ? pick_next_entity+0xa9/0x190 [<ffffffff810df63e>] ? pick_next_task_fair+0x65e/0x8a0 [<ffffffff817e329e>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff8125b47d>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff8127cb80>] ? do_fsync+0xa0/0xa0 [<ffffffff8125b7d1>] SyS_ioctl+0xa1/0xc0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a | LustreError: 9536:0:(llite_lib.c:1894:ll_md_setattr()) md_setattr fails: rc = -28 LustreError: 9536:0:(file.c:242:ll_close_inode_openhandle()) lustre-clilmv-ffff8801bcf74a88: inode [0x2000059f4:0x9399:0x0] mdc close failed: rc = -28 LustreError: 9536:0:(file.c:242:ll_close_inode_openhandle()) Skipped 2 previous similar messages LustreError: 9543:0:(fail.c:138:__cfs_fail_timeout_set()) cfs_fail_timeout id 32c sleeping for 5000ms LustreError: 9543:0:(fail.c:149:__cfs_fail_timeout_set()) cfs_fail_timeout id 32c awake | Link to test |
sanity test 319: lost lease lock on migrate error | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa16082ac>] ll_migrate+0xa1c/0xfa0 [lustre] PGD 1329d067 PUD 27453067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) lquota(OE) lfsck(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) nfsd ext4 mbcache jbd2 loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) crc32_generic crc_t10dif crct10dif_generic crct10dif_common virtio_console pcspkr virtio_balloon i2c_piix4 ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm ata_piix drm_panel_orientation_quirks libata virtio_blk floppy i2c_core serio_raw [last unloaded: obdecho] CPU: 9 PID: 7924 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.15.0-1.module_el8.6.0+1087+b42c8331 04/01/2014 task: ffff8800103649d0 ti: ffff880084120000 task.ti: ffff880084120000 RIP: 0010:[<ffffffffa16082ac>] [<ffffffffa16082ac>] ll_migrate+0xa1c/0xfa0 [lustre] RSP: 0018:ffff880084123bc8 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff88006e20cf88 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff8801020e5c28 RDI: ffff8801020e5c08 RBP: ffff880084123c48 R08: ffff880085f12d78 R09: ffff880084123a60 R10: ffff8800039b5628 R11: ffff8800039b06a8 R12: ffff88025f5f4e18 R13: 0000000000000000 R14: ffff8801ef471cd8 R15: ffff88005c190ce8 FS: 00007fa911fe1740(0000) GS:ffff880331c40000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 00000000712d0000 CR4: 00000000000007e0 Call Trace: [<ffffffff81242293>] ? __check_object_size+0x1c3/0x220 [<ffffffffa15f4c7f>] ll_dir_ioctl+0x5ebf/0x74a0 [lustre] [<ffffffff811b630d>] ? find_get_pages_tag+0x10d/0x260 [<ffffffff811c36c1>] ? pagevec_lookup_tag+0x21/0x30 [<ffffffff811b403e>] ? __filemap_fdatawait_range+0xbe/0x1a0 [<ffffffff817df54c>] ? __mutex_unlock_slowpath+0xdc/0x190 [<ffffffff817df60e>] ? mutex_unlock+0xe/0x10 [<ffffffff81276d80>] ? sync_inodes_sb+0x160/0x430 [<ffffffff810b10ca>] ? __queue_delayed_work+0xaa/0x1a0 [<ffffffff810b1a51>] ? try_to_grab_pending+0xb1/0x180 [<ffffffff8125b47d>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff8125b7d1>] SyS_ioctl+0xa1/0xc0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a | LustreError: 7924:0:(fail.c:138:__cfs_fail_timeout_set()) cfs_fail_timeout id 32c sleeping for 5000ms LustreError: 31351:0:(osd_handler.c:221:osd_trans_start()) lustre-MDT0000: failed to start transaction due to ENOSPC LustreError: 31351:0:(osd_handler.c:221:osd_trans_start()) Skipped 5 previous similar messages LustreError: 7924:0:(fail.c:149:__cfs_fail_timeout_set()) cfs_fail_timeout id 32c awake LustreError: 7978:0:(file.c:242:ll_close_inode_openhandle()) lustre-clilmv-ffff880283845d28: inode [0x2000059f4:0x9282:0x0] mdc close failed: rc = -28 | Link to test |
sanity test 319: lost lease lock on migrate error | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa15f72ac>] ll_migrate+0xa1c/0xfa0 [lustre] PGD 161a22067 PUD 195034067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) lquota(OE) lfsck(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) ext4 mbcache jbd2 loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) crc32_generic crc_t10dif crct10dif_generic crct10dif_common virtio_balloon virtio_console i2c_piix4 pcspkr ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm ata_piix drm_panel_orientation_quirks floppy virtio_blk serio_raw libata i2c_core [last unloaded: obdecho] CPU: 2 PID: 20619 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.15.0-1.module_el8.6.0+1087+b42c8331 04/01/2014 task: ffff88012caedc40 ti: ffff8801349b0000 task.ti: ffff8801349b0000 RIP: 0010:[<ffffffffa15f72ac>] [<ffffffffa15f72ac>] ll_migrate+0xa1c/0xfa0 [lustre] RSP: 0018:ffff8801349b3bc8 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff88009a433ba8 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff880271782b08 RDI: ffff880271782ae8 RBP: ffff8801349b3c48 R08: ffff88012172b138 R09: ffff8801349b3a60 R10: ffff88013ca5a128 R11: ffff88013ca5ea08 R12: ffff880146bc4648 R13: 0000000000000000 R14: ffff8801a797a618 R15: ffff8802c8fbcca8 FS: 00007f3dcac71740(0000) GS:ffff880331a80000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 000000015ccca000 CR4: 00000000000007e0 Call Trace: [<ffffffff81242293>] ? __check_object_size+0x1c3/0x220 [<ffffffffa15e3c7f>] ll_dir_ioctl+0x5ebf/0x74a0 [lustre] [<ffffffff811b630d>] ? find_get_pages_tag+0x10d/0x260 [<ffffffff811c36c1>] ? pagevec_lookup_tag+0x21/0x30 [<ffffffff811b403e>] ? __filemap_fdatawait_range+0xbe/0x1a0 [<ffffffff810b10ca>] ? __queue_delayed_work+0xaa/0x1a0 [<ffffffff817df54c>] ? __mutex_unlock_slowpath+0xdc/0x190 [<ffffffff817df60e>] ? mutex_unlock+0xe/0x10 [<ffffffff81276d80>] ? sync_inodes_sb+0x160/0x430 [<ffffffff810b10ca>] ? __queue_delayed_work+0xaa/0x1a0 [<ffffffff810b1a51>] ? try_to_grab_pending+0xb1/0x180 [<ffffffff8125b47d>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff8125b7d1>] SyS_ioctl+0xa1/0xc0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a | LustreError: 20619:0:(fail.c:138:__cfs_fail_timeout_set()) cfs_fail_timeout id 32c sleeping for 5000ms LustreError: 16098:0:(osd_handler.c:221:osd_trans_start()) lustre-MDT0000: failed to start transaction due to ENOSPC LustreError: 16098:0:(osd_handler.c:221:osd_trans_start()) Skipped 3 previous similar messages LustreError: 20619:0:(fail.c:149:__cfs_fail_timeout_set()) cfs_fail_timeout id 32c awake LustreError: 20702:0:(file.c:242:ll_close_inode_openhandle()) lustre-clilmv-ffff88027756efc8: inode [0x2000059f4:0x948b:0x0] mdc close failed: rc = -28 | Link to test |
sanity test 319: lost lease lock on migrate error | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa160a2ac>] ll_migrate+0xa1c/0xfa0 [lustre] PGD 293e05067 PUD 2d7144067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) lquota(OE) lfsck(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) nfsd ext4 mbcache jbd2 loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) crc32_generic crc_t10dif crct10dif_generic crct10dif_common i2c_piix4 pcspkr virtio_console virtio_balloon ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm drm_panel_orientation_quirks ata_piix virtio_blk i2c_core serio_raw libata floppy [last unloaded: obdecho] CPU: 12 PID: 2313 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.15.0-1.module_el8.6.0+1087+b42c8331 04/01/2014 task: ffff880129f05c40 ti: ffff8801ebf5c000 task.ti: ffff8801ebf5c000 RIP: 0010:[<ffffffffa160a2ac>] [<ffffffffa160a2ac>] ll_migrate+0xa1c/0xfa0 [lustre] RSP: 0018:ffff8801ebf5fbc8 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff88016ae927c8 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff88017b21d768 RDI: ffff88017b21d748 RBP: ffff8801ebf5fc48 R08: ffff88025dcab138 R09: ffff8801ebf5fa60 R10: ffff88026e8ec248 R11: ffff88026e8ed628 R12: ffff8801f2cf2ee8 R13: 0000000000000000 R14: ffff8802a54e4b18 R15: ffff88008c829238 FS: 00007f985c855740(0000) GS:ffff880331d00000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 0000000136d64000 CR4: 00000000000007e0 Call Trace: [<ffffffff81242293>] ? __check_object_size+0x1c3/0x220 [<ffffffffa15f6c7f>] ll_dir_ioctl+0x5ebf/0x74a0 [lustre] [<ffffffff817e05ca>] ? __schedule+0x32a/0x7d0 [<ffffffff8127c9c0>] ? generic_write_sync+0x60/0x60 [<ffffffff817e0a99>] ? schedule+0x29/0x70 [<ffffffff817dd739>] ? schedule_timeout+0x209/0x290 [<ffffffff817e324e>] ? _raw_spin_unlock+0xe/0x20 [<ffffffff810b0d01>] ? __queue_work+0x151/0x420 [<ffffffff810d5999>] ? pick_next_entity+0xa9/0x190 [<ffffffff810df63e>] ? pick_next_task_fair+0x65e/0x8a0 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff8125b47d>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff8127cb80>] ? do_fsync+0xa0/0xa0 [<ffffffff8125b7d1>] SyS_ioctl+0xa1/0xc0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a | LustreError: 2313:0:(fail.c:138:__cfs_fail_timeout_set()) cfs_fail_timeout id 32c sleeping for 5000ms LustreError: 2313:0:(fail.c:149:__cfs_fail_timeout_set()) cfs_fail_timeout id 32c awake LustreError: 13838:0:(osd_handler.c:221:osd_trans_start()) lustre-MDT0000: failed to start transaction due to ENOSPC LustreError: 13838:0:(osd_handler.c:221:osd_trans_start()) Skipped 6 previous similar messages LustreError: 2423:0:(file.c:242:ll_close_inode_openhandle()) lustre-clilmv-ffff88023acc4a88: inode [0x2000059f4:0x92e8:0x0] mdc close failed: rc = -28 | Link to test |
sanity test 319: lost lease lock on migrate error | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa15f52ac>] ll_migrate+0xa1c/0xfa0 [lustre] PGD 8a7b7067 PUD 87566067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) lquota(OE) lfsck(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) ext4 mbcache jbd2 loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) crc32_generic crc_t10dif crct10dif_generic crct10dif_common virtio_balloon i2c_piix4 virtio_console pcspkr ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm ata_piix virtio_blk drm_panel_orientation_quirks serio_raw libata i2c_core floppy [last unloaded: obdecho] CPU: 8 PID: 8337 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.15.0-1.module_el8.6.0+1087+b42c8331 04/01/2014 task: ffff88026d618010 ti: ffff88002ac7c000 task.ti: ffff88002ac7c000 RIP: 0010:[<ffffffffa15f52ac>] [<ffffffffa15f52ac>] ll_migrate+0xa1c/0xfa0 [lustre] RSP: 0018:ffff88002ac7fbc8 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff8800118248e8 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff88031d15ebe8 RDI: ffff88031d15ebc8 RBP: ffff88002ac7fc48 R08: ffff8800759807f8 R09: ffff88002ac7fa60 R10: ffff880011820d48 R11: ffff880011825628 R12: ffff88002912c648 R13: 0000000000000000 R14: ffff880307518118 R15: ffff8803272871d8 FS: 00007f365b18b740(0000) GS:ffff880331c00000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 000000002b3ae000 CR4: 00000000000007e0 Call Trace: [<ffffffff81242293>] ? __check_object_size+0x1c3/0x220 [<ffffffffa15e1c7f>] ll_dir_ioctl+0x5ebf/0x74a0 [lustre] [<ffffffff811b630d>] ? find_get_pages_tag+0x10d/0x260 [<ffffffff811c36c1>] ? pagevec_lookup_tag+0x21/0x30 [<ffffffff811b403e>] ? __filemap_fdatawait_range+0xbe/0x1a0 [<ffffffff810b10ca>] ? __queue_delayed_work+0xaa/0x1a0 [<ffffffff810d5999>] ? pick_next_entity+0xa9/0x190 [<ffffffff810df63e>] ? pick_next_task_fair+0x65e/0x8a0 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff8125b47d>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff8127cb80>] ? do_fsync+0xa0/0xa0 [<ffffffff8125b7d1>] SyS_ioctl+0xa1/0xc0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a | LustreError: 14458:0:(osd_handler.c:221:osd_trans_start()) lustre-MDT0000: failed to start transaction due to ENOSPC LustreError: 14458:0:(osd_handler.c:221:osd_trans_start()) Skipped 3 previous similar messages LustreError: 8300:0:(llite_lib.c:1894:ll_md_setattr()) md_setattr fails: rc = -28 LustreError: 8300:0:(file.c:242:ll_close_inode_openhandle()) lustre-clilmv-ffff8802bcb58958: inode [0x2000059f4:0x9367:0x0] mdc close failed: rc = -28 LustreError: 8337:0:(fail.c:138:__cfs_fail_timeout_set()) cfs_fail_timeout id 32c sleeping for 5000ms LustreError: 8337:0:(fail.c:149:__cfs_fail_timeout_set()) cfs_fail_timeout id 32c awake LustreError: 8417:0:(file.c:242:ll_close_inode_openhandle()) lustre-clilmv-ffff8802bcb58958: inode [0x2000059f4:0x9367:0x0] mdc close failed: rc = -28 | Link to test |
sanity test 319: lost lease lock on migrate error | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa165e24c>] ll_migrate+0xa1c/0xfa0 [lustre] PGD 29d36e067 PUD 324b1a067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) lquota(OE) lfsck(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) nfsd ext4 mbcache jbd2 loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) crc32_generic crc_t10dif crct10dif_generic crct10dif_common virtio_balloon virtio_console pcspkr i2c_piix4 ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm ata_piix drm_panel_orientation_quirks serio_raw virtio_blk floppy libata i2c_core [last unloaded: obdecho] CPU: 8 PID: 16761 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.15.0-1.module_el8.6.0+1087+b42c8331 04/01/2014 task: ffff8802ca0d9280 ti: ffff8801312f0000 task.ti: ffff8801312f0000 RIP: 0010:[<ffffffffa165e24c>] [<ffffffffa165e24c>] ll_migrate+0xa1c/0xfa0 [lustre] RSP: 0018:ffff8801312f3bc8 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff88010ce76a08 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff880288865158 RDI: ffff880288865138 RBP: ffff8801312f3c48 R08: ffff8802c4d229b8 R09: ffff8801312f3a60 R10: ffff88010ce71a88 R11: ffff88010ce72128 R12: ffff8801c9a6df48 R13: 0000000000000000 R14: ffff88013b661398 R15: ffff8801a9caacc8 FS: 00007fe92d52a740(0000) GS:ffff880331c00000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 000000019b800000 CR4: 00000000000007e0 Call Trace: [<ffffffff81242293>] ? __check_object_size+0x1c3/0x220 [<ffffffffa164b2a8>] ll_dir_ioctl+0x5618/0x65a0 [lustre] [<ffffffff810d5999>] ? pick_next_entity+0xa9/0x190 [<ffffffff810d7e3c>] ? set_next_entity+0x3c/0xe0 [<ffffffff810df6e2>] ? pick_next_task_fair+0x702/0x8a0 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e329e>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffff810c834d>] ? finish_task_switch+0x5d/0x1b0 [<ffffffff817e05ca>] ? __schedule+0x32a/0x7d0 [<ffffffff810d5999>] ? pick_next_entity+0xa9/0x190 [<ffffffff810df63e>] ? pick_next_task_fair+0x65e/0x8a0 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff8125b47d>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff8127cb80>] ? do_fsync+0xa0/0xa0 [<ffffffff8125b7d1>] SyS_ioctl+0xa1/0xc0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a | LustreError: 1585:0:(osd_handler.c:221:osd_trans_start()) lustre-MDT0000: failed to start transaction due to ENOSPC LustreError: 1585:0:(osd_handler.c:221:osd_trans_start()) Skipped 11 previous similar messages LustreError: 16733:0:(llite_lib.c:1894:ll_md_setattr()) md_setattr fails: rc = -28 LustreError: 16733:0:(file.c:242:ll_close_inode_openhandle()) lustre-clilmv-ffff8802be6a4a88: inode [0x2000059f3:0x9336:0x0] mdc close failed: rc = -28 LustreError: 16733:0:(file.c:242:ll_close_inode_openhandle()) Skipped 2 previous similar messages LustreError: 16761:0:(fail.c:138:__cfs_fail_timeout_set()) cfs_fail_timeout id 32c sleeping for 5000ms LustreError: 16761:0:(fail.c:149:__cfs_fail_timeout_set()) cfs_fail_timeout id 32c awake LustreError: 16822:0:(file.c:242:ll_close_inode_openhandle()) lustre-clilmv-ffff8802be6a4a88: inode [0x2000059f3:0x9336:0x0] mdc close failed: rc = -28 | Link to test |
sanity test 319: lost lease lock on migrate error | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa15fbe7c>] ll_migrate+0x9ec/0xf50 [lustre] PGD 31f8c0067 PUD 27aaaa067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) lquota(OE) lfsck(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) ext4 mbcache jbd2 loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) crc32_generic crc_t10dif crct10dif_generic crct10dif_common i2c_piix4 pcspkr virtio_console virtio_balloon ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm ata_piix drm_panel_orientation_quirks virtio_blk serio_raw floppy i2c_core libata [last unloaded: obdecho] CPU: 2 PID: 26756 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.15.0-1.module_el8.6.0+1087+b42c8331 04/01/2014 task: ffff8802c32624f0 ti: ffff88010f018000 task.ti: ffff88010f018000 RIP: 0010:[<ffffffffa15fbe7c>] [<ffffffffa15fbe7c>] ll_migrate+0x9ec/0xf50 [lustre] RSP: 0018:ffff88010f01bbd0 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff880136013508 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff8802c58b95c8 RDI: ffff8802c58b95a8 RBP: ffff88010f01bc48 R08: ffff88002a1b25f8 R09: ffff88010f01ba68 R10: ffff8800075c6368 R11: ffff8800075c3ef8 R12: ffff880111d7f208 R13: 0000000000000000 R14: ffff880302e1c1d8 R15: ffff8802c9e8e1e8 FS: 00007fdf35fa9740(0000) GS:ffff880331a80000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 00000001508b2000 CR4: 00000000000007e0 Call Trace: [<ffffffffa15e8b82>] ll_dir_ioctl+0x5d12/0x7290 [lustre] [<ffffffff811b630d>] ? find_get_pages_tag+0x10d/0x260 [<ffffffff811c36c1>] ? pagevec_lookup_tag+0x21/0x30 [<ffffffff811b403e>] ? __filemap_fdatawait_range+0xbe/0x1a0 [<ffffffff817df54c>] ? __mutex_unlock_slowpath+0xdc/0x190 [<ffffffff817df60e>] ? mutex_unlock+0xe/0x10 [<ffffffff81276d80>] ? sync_inodes_sb+0x160/0x430 [<ffffffff810b10ca>] ? __queue_delayed_work+0xaa/0x1a0 [<ffffffff810b1a51>] ? try_to_grab_pending+0xb1/0x180 [<ffffffff8125b47d>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff8125b7d1>] SyS_ioctl+0xa1/0xc0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a | LustreError: 28565:0:(osd_handler.c:221:osd_trans_start()) lustre-MDT0000: failed to start transaction due to ENOSPC LustreError: 28565:0:(osd_handler.c:221:osd_trans_start()) Skipped 5 previous similar messages LustreError: 26733:0:(llite_lib.c:1868:ll_md_setattr()) md_setattr fails: rc = -28 LustreError: 26733:0:(file.c:243:ll_close_inode_openhandle()) lustre-clilmv-ffff8801330d6fc8: inode [0x200005223:0x93e5:0x0] mdc close failed: rc = -28 LustreError: 26756:0:(fail.c:138:__cfs_fail_timeout_set()) cfs_fail_timeout id 32c sleeping for 5000ms LustreError: 26756:0:(fail.c:149:__cfs_fail_timeout_set()) cfs_fail_timeout id 32c awake | Link to test |
sanity test 319: lost lease lock on migrate error | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa15ffe7c>] ll_migrate+0x9ec/0xf50 [lustre] PGD 232b6a067 PUD 1461ed067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) lquota(OE) lfsck(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) ext4 mbcache jbd2 loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) crc32_generic crc_t10dif crct10dif_generic crct10dif_common i2c_piix4 virtio_balloon pcspkr virtio_console ip_tables rpcsec_gss_krb5 ata_generic drm_kms_helper pata_acpi ttm drm drm_panel_orientation_quirks ata_piix i2c_core serio_raw libata virtio_blk floppy [last unloaded: obdecho] CPU: 6 PID: 11690 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.15.0-1.module_el8.6.0+1087+b42c8331 04/01/2014 task: ffff8802c431b760 ti: ffff880167c50000 task.ti: ffff880167c50000 RIP: 0010:[<ffffffffa15ffe7c>] [<ffffffffa15ffe7c>] ll_migrate+0x9ec/0xf50 [lustre] RSP: 0018:ffff880167c53bd0 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff88000831b1b8 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff8802ce249bd8 RDI: ffff8802ce249bb8 RBP: ffff880167c53c48 R08: ffff88010d1d8f78 R09: ffff880167c53a68 R10: ffff88018fb1e368 R11: ffff88018fb1ab18 R12: ffff8801cfa80198 R13: 0000000000000000 R14: ffff880304dee6d8 R15: ffff880283e551f8 FS: 00007fa6b4a0e740(0000) GS:ffff880331b80000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 000000014aad2000 CR4: 00000000000007e0 Call Trace: [<ffffffffa15ecb82>] ll_dir_ioctl+0x5d12/0x7290 [lustre] [<ffffffff811b630d>] ? find_get_pages_tag+0x10d/0x260 [<ffffffff811c36c1>] ? pagevec_lookup_tag+0x21/0x30 [<ffffffff811b403e>] ? __filemap_fdatawait_range+0xbe/0x1a0 [<ffffffff817df54c>] ? __mutex_unlock_slowpath+0xdc/0x190 [<ffffffff817df60e>] ? mutex_unlock+0xe/0x10 [<ffffffff81276d80>] ? sync_inodes_sb+0x160/0x430 [<ffffffff8125b47d>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff8125b7d1>] SyS_ioctl+0xa1/0xc0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a | LustreError: 11630:0:(llite_lib.c:1871:ll_md_setattr()) md_setattr fails: rc = -28 LustreError: 11630:0:(file.c:243:ll_close_inode_openhandle()) lustre-clilmv-ffff8802c0fa8008: inode [0x200005223:0x8be6:0x0] mdc close failed: rc = -28 LustreError: 11690:0:(fail.c:138:__cfs_fail_timeout_set()) cfs_fail_timeout id 32c sleeping for 5000ms LustreError: 892:0:(osd_handler.c:221:osd_trans_start()) lustre-MDT0000: failed to start transaction due to ENOSPC LustreError: 892:0:(osd_handler.c:221:osd_trans_start()) Skipped 7 previous similar messages LustreError: 11690:0:(fail.c:149:__cfs_fail_timeout_set()) cfs_fail_timeout id 32c awake LustreError: 11800:0:(file.c:243:ll_close_inode_openhandle()) lustre-clilmv-ffff8802c0fa8008: inode [0x200005223:0x8be6:0x0] mdc close failed: rc = -28 | Link to test |
sanity test 319: lost lease lock on migrate error | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa1602e7c>] ll_migrate+0x9ec/0xf50 [lustre] PGD 9eb9b067 PUD 9cbd9067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) lquota(OE) lfsck(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) ext4 mbcache jbd2 loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) crc32_generic crc_t10dif crct10dif_generic crct10dif_common virtio_balloon virtio_console pcspkr i2c_piix4 ip_tables rpcsec_gss_krb5 drm_kms_helper ata_generic ttm pata_acpi drm drm_panel_orientation_quirks floppy ata_piix serio_raw virtio_blk i2c_core libata [last unloaded: obdecho] CPU: 0 PID: 14684 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.15.0-1.module_el8.6.0+1087+b42c8331 04/01/2014 task: ffff8801465aa4f0 ti: ffff8800104e0000 task.ti: ffff8800104e0000 RIP: 0010:[<ffffffffa1602e7c>] [<ffffffffa1602e7c>] ll_migrate+0x9ec/0xf50 [lustre] RSP: 0018:ffff8800104e3bd0 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff88013347f3f8 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff8802cb509568 RDI: ffff8802cb509548 RBP: ffff8800104e3c48 R08: ffff880238753c78 R09: ffff8800104e3a68 R10: ffff88013347a7c8 R11: ffff8801334793e8 R12: ffff8801417c5db8 R13: 0000000000000000 R14: ffff8801b0dbb898 R15: ffff8801b97ff728 FS: 00007f63f49a3740(0000) GS:ffff880331a00000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 000000009bb88000 CR4: 00000000000007f0 Call Trace: [<ffffffffa15efb82>] ll_dir_ioctl+0x5d12/0x7290 [lustre] [<ffffffff817e05ca>] ? __schedule+0x32a/0x7d0 [<ffffffff8127c9c0>] ? generic_write_sync+0x60/0x60 [<ffffffff817e0a99>] ? schedule+0x29/0x70 [<ffffffff817dd739>] ? schedule_timeout+0x209/0x290 [<ffffffff817e324e>] ? _raw_spin_unlock+0xe/0x20 [<ffffffff810b0d01>] ? __queue_work+0x151/0x420 [<ffffffff810d5268>] ? __enqueue_entity+0x78/0x80 [<ffffffff810dbbab>] ? put_prev_entity+0x8b/0x400 [<ffffffff810d5999>] ? pick_next_entity+0xa9/0x190 [<ffffffff810d7e3c>] ? set_next_entity+0x3c/0xe0 [<ffffffff810df6e2>] ? pick_next_task_fair+0x702/0x8a0 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e329e>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffff810c834d>] ? finish_task_switch+0x5d/0x1b0 [<ffffffff8125b47d>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff8127cb80>] ? do_fsync+0xa0/0xa0 [<ffffffff8125b7d1>] SyS_ioctl+0xa1/0xc0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a | LustreError: 14637:0:(llite_lib.c:1871:ll_md_setattr()) md_setattr fails: rc = -28 LustreError: 14637:0:(file.c:243:ll_close_inode_openhandle()) lustre-clilmv-ffff880137d20008: inode [0x200005223:0x9323:0x0] mdc close failed: rc = -28 LustreError: 14684:0:(fail.c:138:__cfs_fail_timeout_set()) cfs_fail_timeout id 32c sleeping for 5000ms LustreError: 14684:0:(fail.c:149:__cfs_fail_timeout_set()) cfs_fail_timeout id 32c awake LustreError: 10993:0:(osd_handler.c:221:osd_trans_start()) lustre-MDT0000: failed to start transaction due to ENOSPC LustreError: 10993:0:(osd_handler.c:221:osd_trans_start()) Skipped 11 previous similar messages LustreError: 14702:0:(file.c:243:ll_close_inode_openhandle()) lustre-clilmv-ffff880137d20008: inode [0x200005223:0x9323:0x0] mdc close failed: rc = -28 | Link to test |
sanity test 319: lost lease lock on migrate error | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa1601e7c>] ll_migrate+0x9ec/0xf50 [lustre] PGD 138bda067 PUD 16de8f067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) lquota(OE) lfsck(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) ext4 mbcache jbd2 loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) crc32_generic crc_t10dif crct10dif_generic crct10dif_common virtio_console virtio_balloon i2c_piix4 pcspkr ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm ata_piix drm_panel_orientation_quirks serio_raw virtio_blk floppy libata i2c_core [last unloaded: obdecho] CPU: 14 PID: 30030 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.15.0-1.module_el8.6.0+1087+b42c8331 04/01/2014 task: ffff88031ccd5c40 ti: ffff88017f9f4000 task.ti: ffff88017f9f4000 RIP: 0010:[<ffffffffa1601e7c>] [<ffffffffa1601e7c>] ll_migrate+0x9ec/0xf50 [lustre] RSP: 0018:ffff88017f9f7bd0 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff88014e2b8008 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff8802941802c8 RDI: ffff8802941802a8 RBP: ffff88017f9f7c48 R08: ffff8802c7ee1158 R09: ffff88017f9f7a68 R10: ffff880197ad13e8 R11: ffff880197ad48e8 R12: ffff880170285f48 R13: 0000000000000000 R14: ffff880052192618 R15: ffff88008e9871d8 FS: 00007fe300637740(0000) GS:ffff880331d80000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 0000000156ffe000 CR4: 00000000000007e0 Call Trace: [<ffffffffa15eeb82>] ll_dir_ioctl+0x5d12/0x7290 [lustre] [<ffffffff817e05ca>] ? __schedule+0x32a/0x7d0 [<ffffffff8127c9c0>] ? generic_write_sync+0x60/0x60 [<ffffffff817e0a99>] ? schedule+0x29/0x70 [<ffffffff817dd739>] ? schedule_timeout+0x209/0x290 [<ffffffff817e324e>] ? _raw_spin_unlock+0xe/0x20 [<ffffffff810b0d01>] ? __queue_work+0x151/0x420 [<ffffffff810d5268>] ? __enqueue_entity+0x78/0x80 [<ffffffff810dbbab>] ? put_prev_entity+0x8b/0x400 [<ffffffff810d5999>] ? pick_next_entity+0xa9/0x190 [<ffffffff810d7e3c>] ? set_next_entity+0x3c/0xe0 [<ffffffff810df6e2>] ? pick_next_task_fair+0x702/0x8a0 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e329e>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffff810c834d>] ? finish_task_switch+0x5d/0x1b0 [<ffffffff8125b47d>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff8127cb80>] ? do_fsync+0xa0/0xa0 [<ffffffff8125b7d1>] SyS_ioctl+0xa1/0xc0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a | LustreError: 29953:0:(llite_lib.c:1871:ll_md_setattr()) md_setattr fails: rc = -28 LustreError: 29953:0:(file.c:243:ll_close_inode_openhandle()) lustre-clilmv-ffff88001a7e9bf8: inode [0x2000059f3:0x9418:0x0] mdc close failed: rc = -28 LustreError: 29953:0:(file.c:243:ll_close_inode_openhandle()) Skipped 1 previous similar message LustreError: 30030:0:(fail.c:138:__cfs_fail_timeout_set()) cfs_fail_timeout id 32c sleeping for 5000ms LustreError: 30030:0:(fail.c:149:__cfs_fail_timeout_set()) cfs_fail_timeout id 32c awake LustreError: 30076:0:(file.c:243:ll_close_inode_openhandle()) lustre-clilmv-ffff88001a7e9bf8: inode [0x2000059f3:0x9418:0x0] mdc close failed: rc = -28 | Link to test |
sanity test 319: lost lease lock on migrate error | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa15fde7c>] ll_migrate+0x9ec/0xf50 [lustre] PGD 55691067 PUD ba234067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) lquota(OE) lfsck(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) nfsd ext4 mbcache jbd2 loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) crc32_generic crc_t10dif crct10dif_generic crct10dif_common virtio_balloon virtio_console i2c_piix4 pcspkr ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm drm_panel_orientation_quirks ata_piix floppy serio_raw virtio_blk libata i2c_core [last unloaded: obdecho] CPU: 12 PID: 7076 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.15.0-1.module_el8.6.0+1087+b42c8331 04/01/2014 task: ffff880141895c40 ti: ffff880088944000 task.ti: ffff880088944000 RIP: 0010:[<ffffffffa15fde7c>] [<ffffffffa15fde7c>] ll_migrate+0x9ec/0xf50 [lustre] RSP: 0018:ffff880088947bd0 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff880266f486a8 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff880070fd7748 RDI: ffff880070fd7728 RBP: ffff880088947c48 R08: ffff88024ad70438 R09: ffff880088947a68 R10: ffff880331403640 R11: ffff8801e929f748 R12: ffff88001e9dabc8 R13: 0000000000000000 R14: ffff8802ad58dd98 R15: ffff8803244aa228 FS: 00007f0b1565b740(0000) GS:ffff880331d00000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 0000000086182000 CR4: 00000000000007e0 Call Trace: [<ffffffffa15eab82>] ll_dir_ioctl+0x5d12/0x7290 [lustre] [<ffffffff811b518b>] ? unlock_page+0x2b/0x30 [<ffffffff811e5e5e>] ? do_read_fault.isra.64+0x12e/0x170 [<ffffffff811ea1ff>] ? handle_pte_fault+0x2ef/0xca0 [<ffffffff8103de0f>] ? save_stack_trace+0x1f/0x30 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff811ecf9d>] ? __handle_mm_fault+0x31d/0x720 [<ffffffff8125b47d>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff817e8655>] ? __do_page_fault+0x1c5/0x500 [<ffffffff8125b7d1>] SyS_ioctl+0xa1/0xc0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a | LustreError: 7049:0:(llite_lib.c:1871:ll_md_setattr()) md_setattr fails: rc = -28 LustreError: 7049:0:(file.c:243:ll_close_inode_openhandle()) lustre-clilmv-ffff880083e112a8: inode [0x2000059f3:0x92c0:0x0] mdc close failed: rc = -28 LustreError: 7049:0:(file.c:243:ll_close_inode_openhandle()) Skipped 2 previous similar messages | Link to test |
sanity test 319: lost lease lock on migrate error | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa1614e7c>] ll_migrate+0x9ec/0xf50 [lustre] PGD 2f614c067 PUD 1f8b3a067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) lquota(OE) lfsck(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) nfsd ext4 mbcache jbd2 loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) crc32_generic crc_t10dif crct10dif_generic crct10dif_common pcspkr virtio_balloon i2c_piix4 virtio_console ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm ata_piix drm_panel_orientation_quirks virtio_blk libata serio_raw floppy i2c_core [last unloaded: obdecho] CPU: 15 PID: 30795 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.15.0-1.module_el8.6.0+1087+b42c8331 04/01/2014 task: ffff8802c9bd3760 ti: ffff8801eee20000 task.ti: ffff8801eee20000 RIP: 0010:[<ffffffffa1614e7c>] [<ffffffffa1614e7c>] ll_migrate+0x9ec/0xf50 [lustre] RSP: 0018:ffff8801eee23bd0 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff8802f8d5d628 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff88004725a828 RDI: ffff88004725a808 RBP: ffff8801eee23c48 R08: ffff88028ac9ed78 R09: ffff8801eee23a68 R10: ffff8802f8d5f0a8 R11: ffff8802f8d5ed58 R12: ffff8801ac5d4c88 R13: 0000000000000000 R14: ffff8801f16ae6d8 R15: ffff8802b1d72228 FS: 00007ff303ec1740(0000) GS:ffff880331dc0000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 0000000166fce000 CR4: 00000000000007e0 Call Trace: [<ffffffffa1601b82>] ll_dir_ioctl+0x5d12/0x7290 [lustre] [<ffffffff811b518b>] ? unlock_page+0x2b/0x30 [<ffffffff811e5e5e>] ? do_read_fault.isra.64+0x12e/0x170 [<ffffffff811ea1ff>] ? handle_pte_fault+0x2ef/0xca0 [<ffffffff8103de0f>] ? save_stack_trace+0x1f/0x30 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff811ecf9d>] ? __handle_mm_fault+0x31d/0x720 [<ffffffff8125b47d>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff817e8655>] ? __do_page_fault+0x1c5/0x500 [<ffffffff8125b7d1>] SyS_ioctl+0xa1/0xc0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a | LustreError: 14396:0:(osd_handler.c:221:osd_trans_start()) lustre-MDT0000: failed to start transaction due to ENOSPC LustreError: 14396:0:(osd_handler.c:221:osd_trans_start()) Skipped 12 previous similar messages LustreError: 30780:0:(llite_lib.c:1871:ll_md_setattr()) md_setattr fails: rc = -28 LustreError: 30780:0:(file.c:243:ll_close_inode_openhandle()) lustre-clilmv-ffff88001b8392a8: inode [0x200005223:0x940e:0x0] mdc close failed: rc = -28 | Link to test |
sanity test 319: lost lease lock on migrate error | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa161ce7c>] ll_migrate+0x9ec/0xf50 [lustre] PGD 161826067 PUD 274390067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) lquota(OE) lfsck(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) nfsd ext4 mbcache jbd2 loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) crc32_generic crc_t10dif crct10dif_generic crct10dif_common i2c_piix4 virtio_console pcspkr virtio_balloon ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm drm_panel_orientation_quirks ata_piix virtio_blk floppy libata i2c_core serio_raw [last unloaded: obdecho] CPU: 5 PID: 27553 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.15.0-1.module_el8.6.0+1087+b42c8331 04/01/2014 task: ffff8802c6de1280 ti: ffff880153bd8000 task.ti: ffff880153bd8000 RIP: 0010:[<ffffffffa161ce7c>] [<ffffffffa161ce7c>] ll_migrate+0x9ec/0xf50 [lustre] RSP: 0018:ffff880153bdbbd0 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff880039610008 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff8802bd575618 RDI: ffff8802bd5755f8 RBP: ffff880153bdbc48 R08: ffff88003c8a1158 R09: ffff880153bdba68 R10: ffff880039613ba8 R11: ffff8800396152d8 R12: ffff8802b5cf7b68 R13: 0000000000000000 R14: ffff8802a4325d98 R15: ffff88008a635c98 FS: 00007f51bd85a740(0000) GS:ffff880331b40000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 00000001a5464000 CR4: 00000000000007e0 Call Trace: [<ffffffffa1609b82>] ll_dir_ioctl+0x5d12/0x7290 [lustre] [<ffffffff811b518b>] ? unlock_page+0x2b/0x30 [<ffffffff811e5e5e>] ? do_read_fault.isra.64+0x12e/0x170 [<ffffffff811ea1ff>] ? handle_pte_fault+0x2ef/0xca0 [<ffffffff8103de0f>] ? save_stack_trace+0x1f/0x30 [<ffffffff811ecf9d>] ? __handle_mm_fault+0x31d/0x720 [<ffffffff8125b47d>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff817e8655>] ? __do_page_fault+0x1c5/0x500 [<ffffffff8125b7d1>] SyS_ioctl+0xa1/0xc0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a | LustreError: 27496:0:(llite_lib.c:1871:ll_md_setattr()) md_setattr fails: rc = -28 LustreError: 27496:0:(file.c:243:ll_close_inode_openhandle()) lustre-clilmv-ffff880274de2e98: inode [0x2000059f3:0x9276:0x0] mdc close failed: rc = -28 LustreError: 27496:0:(file.c:243:ll_close_inode_openhandle()) Skipped 2 previous similar messages | Link to test |
sanity test 319: lost lease lock on migrate error | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa15f0e7c>] ll_migrate+0x9ec/0xf50 [lustre] PGD 446f2067 PUD 6d02e067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) lquota(OE) lfsck(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) nfsd ext4 mbcache jbd2 loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) crc32_generic crc_t10dif crct10dif_generic crct10dif_common virtio_balloon virtio_console pcspkr i2c_piix4 ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm ata_piix drm_panel_orientation_quirks virtio_blk serio_raw floppy i2c_core libata [last unloaded: obdecho] CPU: 15 PID: 9030 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.15.0-1.module_el8.6.0+1087+b42c8331 04/01/2014 task: ffff88026f73a4f0 ti: ffff880024820000 task.ti: ffff880024820000 RIP: 0010:[<ffffffffa15f0e7c>] [<ffffffffa15f0e7c>] ll_migrate+0x9ec/0xf50 [lustre] RSP: 0018:ffff880024823bd0 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff8802b622d978 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff880247ce1838 RDI: ffff880247ce1818 RBP: ffff880024823c48 R08: ffff88027b97bc78 R09: ffff880024823a68 R10: ffff8802b622e368 R11: ffff8802b62289f8 R12: ffff8800163b23f8 R13: 0000000000000000 R14: ffff880070ffaf58 R15: ffff880283bee1e8 FS: 00007fc6978c8740(0000) GS:ffff880331dc0000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 000000003d820000 CR4: 00000000000007e0 Call Trace: [<ffffffffa15ddb82>] ll_dir_ioctl+0x5d12/0x7290 [lustre] [<ffffffff811b518b>] ? unlock_page+0x2b/0x30 [<ffffffff811e5e5e>] ? do_read_fault.isra.64+0x12e/0x170 [<ffffffff811ea1ff>] ? handle_pte_fault+0x2ef/0xca0 [<ffffffff8103de0f>] ? save_stack_trace+0x1f/0x30 [<ffffffff811ecf9d>] ? __handle_mm_fault+0x31d/0x720 [<ffffffff8125b47d>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff817e8655>] ? __do_page_fault+0x1c5/0x500 [<ffffffff8125b7d1>] SyS_ioctl+0xa1/0xc0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a | LustreError: 18207:0:(osd_handler.c:221:osd_trans_start()) lustre-MDT0000: failed to start transaction due to ENOSPC LustreError: 8999:0:(llite_lib.c:1871:ll_md_setattr()) md_setattr fails: rc = -28 LustreError: 8999:0:(file.c:243:ll_close_inode_openhandle()) lustre-clilmv-ffff880054d537e8: inode [0x2000059f3:0x93c4:0x0] mdc close failed: rc = -28 LustreError: 8999:0:(file.c:243:ll_close_inode_openhandle()) Skipped 2 previous similar messages | Link to test |
sanity test 319: lost lease lock on migrate error | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa15fce7c>] ll_migrate+0x9ec/0xf50 [lustre] PGD 26d909067 PUD 2912cc067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) lquota(OE) lfsck(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) nfsd ext4 mbcache jbd2 loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) crc32_generic crc_t10dif crct10dif_generic crct10dif_common virtio_balloon pcspkr virtio_console i2c_piix4 ip_tables rpcsec_gss_krb5 drm_kms_helper ttm ata_generic drm pata_acpi ata_piix drm_panel_orientation_quirks serio_raw virtio_blk floppy libata i2c_core [last unloaded: obdecho] CPU: 15 PID: 13482 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.15.0-1.module_el8.6.0+1087+b42c8331 04/01/2014 task: ffff88024cf33760 ti: ffff88012a33c000 task.ti: ffff88012a33c000 RIP: 0010:[<ffffffffa15fce7c>] [<ffffffffa15fce7c>] ll_migrate+0x9ec/0xf50 [lustre] RSP: 0018:ffff88012a33fbd0 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff880105aef3f8 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff880274fee7e8 RDI: ffff880274fee7c8 RBP: ffff88012a33fc48 R08: ffff8802869da418 R09: ffff88012a33fa68 R10: ffff880105aec248 R11: ffff880105aefa98 R12: ffff880073256a38 R13: 0000000000000000 R14: ffff8801fda2c1d8 R15: ffff8802c61aec88 FS: 00007f906f903740(0000) GS:ffff880331dc0000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 000000028145c000 CR4: 00000000000007e0 Call Trace: [<ffffffffa15e9b82>] ll_dir_ioctl+0x5d12/0x7290 [lustre] [<ffffffff811b518b>] ? unlock_page+0x2b/0x30 [<ffffffff811e5e5e>] ? do_read_fault.isra.64+0x12e/0x170 [<ffffffff811ea1ff>] ? handle_pte_fault+0x2ef/0xca0 [<ffffffff8103de0f>] ? save_stack_trace+0x1f/0x30 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff811ecf9d>] ? __handle_mm_fault+0x31d/0x720 [<ffffffff8125b47d>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff817e8655>] ? __do_page_fault+0x1c5/0x500 [<ffffffff8125b7d1>] SyS_ioctl+0xa1/0xc0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a | LustreError: 13461:0:(llite_lib.c:1871:ll_md_setattr()) md_setattr fails: rc = -28 LustreError: 13461:0:(file.c:243:ll_close_inode_openhandle()) lustre-clilmv-ffff8801b1222e98: inode [0x2000059f3:0x94b3:0x0] mdc close failed: rc = -28 LustreError: 13461:0:(file.c:243:ll_close_inode_openhandle()) Skipped 1 previous similar message | Link to test |
sanity test 319: lost lease lock on migrate error | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa1606e7c>] ll_migrate+0x9ec/0xf50 [lustre] PGD 293f66067 PUD 285c80067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) lquota(OE) lfsck(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) ext4 mbcache jbd2 loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) crc32_generic crc_t10dif crct10dif_generic crct10dif_common pcspkr virtio_console virtio_balloon i2c_piix4 ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm ata_piix drm_panel_orientation_quirks floppy virtio_blk libata i2c_core serio_raw [last unloaded: obdecho] CPU: 14 PID: 30334 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.15.0-1.module_el8.6.0+1087+b42c8331 04/01/2014 task: ffff88001e02b760 ti: ffff88016aa54000 task.ti: ffff88016aa54000 RIP: 0010:[<ffffffffa1606e7c>] [<ffffffffa1606e7c>] ll_migrate+0x9ec/0xf50 [lustre] RSP: 0018:ffff88016aa57bd0 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff88012d459dd8 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff8802cdcc0d68 RDI: ffff8802cdcc0d48 RBP: ffff88016aa57c48 R08: ffff8802fac927d8 R09: ffff88016aa57a68 R10: ffff88012d458d48 R11: ffff88012d45cc38 R12: ffff8802f7221db8 R13: 0000000000000000 R14: ffff8802f7531cd8 R15: ffff880013d371d8 FS: 00007f89def31740(0000) GS:ffff880331d80000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 000000014fd46000 CR4: 00000000000007e0 Call Trace: [<ffffffffa15f3b82>] ll_dir_ioctl+0x5d12/0x7290 [lustre] [<ffffffff811b518b>] ? unlock_page+0x2b/0x30 [<ffffffff811e5e5e>] ? do_read_fault.isra.64+0x12e/0x170 [<ffffffff811ea1ff>] ? handle_pte_fault+0x2ef/0xca0 [<ffffffff8103de0f>] ? save_stack_trace+0x1f/0x30 [<ffffffff811ecf9d>] ? __handle_mm_fault+0x31d/0x720 [<ffffffff8125b47d>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff817e8655>] ? __do_page_fault+0x1c5/0x500 [<ffffffff8125b7d1>] SyS_ioctl+0xa1/0xc0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a | LustreError: 30289:0:(llite_lib.c:1871:ll_md_setattr()) md_setattr fails: rc = -28 LustreError: 30289:0:(file.c:243:ll_close_inode_openhandle()) lustre-clilmv-ffff880071b61bf8: inode [0x2000059f3:0x9456:0x0] mdc close failed: rc = -28 LustreError: 30289:0:(file.c:243:ll_close_inode_openhandle()) Skipped 1 previous similar message | Link to test |
sanity test 319: lost lease lock on migrate error | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa1601e7c>] ll_migrate+0x9ec/0xf50 [lustre] PGD 248434067 PUD 2ec1fb067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) lquota(OE) lfsck(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) nfsd ext4 mbcache jbd2 loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) crc32_generic crc_t10dif crct10dif_generic crct10dif_common virtio_balloon i2c_piix4 virtio_console pcspkr ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm ata_piix drm_panel_orientation_quirks serio_raw virtio_blk libata i2c_core floppy [last unloaded: obdecho] CPU: 10 PID: 24137 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.13.0-2.module_el8.5.0+746+bbd5d70c 04/01/2014 task: ffff880283c0b760 ti: ffff8801ea0e4000 task.ti: ffff8801ea0e4000 RIP: 0010:[<ffffffffa1601e7c>] [<ffffffffa1601e7c>] ll_migrate+0x9ec/0xf50 [lustre] RSP: 0018:ffff8801ea0e7bd0 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff8801ff762128 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff880323b391f8 RDI: ffff880323b391d8 RBP: ffff8801ea0e7c48 R08: ffff88017c0eeb98 R09: ffff8801ea0e7a68 R10: ffff8801ff7631b8 R11: ffff8801ff766d58 R12: ffff8801d78d8e18 R13: 0000000000000000 R14: ffff8802d3934b18 R15: ffff8801856bb218 FS: 00007fdb7e437740(0000) GS:ffff880331c80000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 00000001761f4000 CR4: 00000000000007e0 Call Trace: [<ffffffffa15eeb82>] ll_dir_ioctl+0x5d12/0x7290 [lustre] [<ffffffff811b518b>] ? unlock_page+0x2b/0x30 [<ffffffff811e5e5e>] ? do_read_fault.isra.64+0x12e/0x170 [<ffffffff811ea1ff>] ? handle_pte_fault+0x2ef/0xca0 [<ffffffff8103de0f>] ? save_stack_trace+0x1f/0x30 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff811ecf9d>] ? __handle_mm_fault+0x31d/0x720 [<ffffffff8125b47d>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff817e8655>] ? __do_page_fault+0x1c5/0x500 [<ffffffff8125b7d1>] SyS_ioctl+0xa1/0xc0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a | LustreError: 24107:0:(llite_lib.c:1871:ll_md_setattr()) md_setattr fails: rc = -28 LustreError: 24107:0:(file.c:243:ll_close_inode_openhandle()) lustre-clilmv-ffff88000fa237e8: inode [0x200005223:0x94a6:0x0] mdc close failed: rc = -28 LustreError: 24107:0:(file.c:243:ll_close_inode_openhandle()) Skipped 1 previous similar message | Link to test |
sanity test 319: lost lease lock on migrate error | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa1609e7c>] ll_migrate+0x9ec/0xf50 [lustre] PGD 15d162067 PUD 28e662067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) lquota(OE) lfsck(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) ext4 mbcache jbd2 loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) crc32_generic crc_t10dif crct10dif_generic crct10dif_common virtio_console pcspkr i2c_piix4 virtio_balloon ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm ata_piix drm_panel_orientation_quirks serio_raw virtio_blk libata floppy i2c_core [last unloaded: obdecho] CPU: 7 PID: 331 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.13.0-2.module_el8.5.0+746+bbd5d70c 04/01/2014 task: ffff8802709c8010 ti: ffff88018548c000 task.ti: ffff88018548c000 RIP: 0010:[<ffffffffa1609e7c>] [<ffffffffa1609e7c>] ll_migrate+0x9ec/0xf50 [lustre] RSP: 0018:ffff88018548fbd0 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff88005480a7c8 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff880289a6ebb8 RDI: ffff880289a6eb98 RBP: ffff88018548fc48 R08: ffff8802cd24b8b8 R09: ffff88018548fa68 R10: ffff88005480e368 R11: ffff8800548089f8 R12: ffff8802e5327cf8 R13: 0000000000000000 R14: ffff8800b2331398 R15: ffff880290a68ce8 FS: 00007f9dfd235740(0000) GS:ffff880331bc0000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 0000000273210000 CR4: 00000000000007e0 Call Trace: [<ffffffffa15f6b82>] ll_dir_ioctl+0x5d12/0x7290 [lustre] [<ffffffff811b518b>] ? unlock_page+0x2b/0x30 [<ffffffff811e5e5e>] ? do_read_fault.isra.64+0x12e/0x170 [<ffffffff811ea1ff>] ? handle_pte_fault+0x2ef/0xca0 [<ffffffff8103de0f>] ? save_stack_trace+0x1f/0x30 [<ffffffff811ecf9d>] ? __handle_mm_fault+0x31d/0x720 [<ffffffff8125b47d>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff817e8655>] ? __do_page_fault+0x1c5/0x500 [<ffffffff8125b7d1>] SyS_ioctl+0xa1/0xc0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a | LustreError: 305:0:(llite_lib.c:1871:ll_md_setattr()) md_setattr fails: rc = -28 LustreError: 305:0:(file.c:243:ll_close_inode_openhandle()) lustre-clilmv-ffff8800192bc138: inode [0x2000059f3:0x96da:0x0] mdc close failed: rc = -28 LustreError: 305:0:(file.c:243:ll_close_inode_openhandle()) Skipped 1 previous similar message | Link to test |
sanity test 319: lost lease lock on migrate error | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa15f7e7c>] ll_migrate+0x9ec/0xf50 [lustre] PGD 1e117b067 PUD 296f91067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) lquota(OE) lfsck(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) nfsd ext4 mbcache jbd2 loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) crc32_generic crc_t10dif crct10dif_generic crct10dif_common virtio_console pcspkr i2c_piix4 virtio_balloon ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm drm_panel_orientation_quirks ata_piix virtio_blk serio_raw floppy i2c_core libata [last unloaded: obdecho] CPU: 7 PID: 5157 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.13.0-2.module_el8.5.0+746+bbd5d70c 04/01/2014 task: ffff88014b789280 ti: ffff8801a619c000 task.ti: ffff8801a619c000 RIP: 0010:[<ffffffffa15f7e7c>] [<ffffffffa15f7e7c>] ll_migrate+0x9ec/0xf50 [lustre] RSP: 0018:ffff8801a619fbd0 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff8801ab7206a8 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff880288958178 RDI: ffff880288958158 RBP: ffff8801a619fc48 R08: ffff880277dc4bb8 R09: ffff8801a619fa68 R10: ffff880331403640 R11: ffff880014f0cc38 R12: ffff88000ec41a98 R13: 0000000000000000 R14: ffff8801e9938a58 R15: ffff8802c2e4bcb8 FS: 00007f6fc314d740(0000) GS:ffff880331bc0000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 000000010d0ca000 CR4: 00000000000007e0 Call Trace: [<ffffffffa15e4b82>] ll_dir_ioctl+0x5d12/0x7290 [lustre] [<ffffffff811b518b>] ? unlock_page+0x2b/0x30 [<ffffffff811e5e5e>] ? do_read_fault.isra.64+0x12e/0x170 [<ffffffff811ea1ff>] ? handle_pte_fault+0x2ef/0xca0 [<ffffffff8103de0f>] ? save_stack_trace+0x1f/0x30 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff811ecf9d>] ? __handle_mm_fault+0x31d/0x720 [<ffffffff8125b47d>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff817e8655>] ? __do_page_fault+0x1c5/0x500 [<ffffffff8125b7d1>] SyS_ioctl+0xa1/0xc0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a | LustreError: 5131:0:(llite_lib.c:1871:ll_md_setattr()) md_setattr fails: rc = -28 LustreError: 5131:0:(file.c:243:ll_close_inode_openhandle()) lustre-clilmv-ffff8802c7f9c138: inode [0x200005223:0x92a2:0x0] mdc close failed: rc = -28 LustreError: 5131:0:(file.c:243:ll_close_inode_openhandle()) Skipped 1 previous similar message | Link to test |
sanity test 319: lost lease lock on migrate error | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa1625e7c>] ll_migrate+0x9ec/0xf50 [lustre] PGD 2c4021067 PUD 31f1c8067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) lquota(OE) lfsck(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) nfsd ext4 mbcache jbd2 loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) crc32_generic crc_t10dif crct10dif_generic crct10dif_common virtio_console i2c_piix4 virtio_balloon pcspkr ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm ata_piix drm_panel_orientation_quirks serio_raw virtio_blk floppy i2c_core libata [last unloaded: obdecho] CPU: 5 PID: 2240 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.13.0-2.module_el8.5.0+746+bbd5d70c 04/01/2014 task: ffff880041460010 ti: ffff880125bf8000 task.ti: ffff880125bf8000 RIP: 0010:[<ffffffffa1625e7c>] [<ffffffffa1625e7c>] ll_migrate+0x9ec/0xf50 [lustre] RSP: 0018:ffff880125bfbbd0 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff8802da972128 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff8802c64e70a8 RDI: ffff8802c64e7088 RBP: ffff880125bfbc48 R08: ffff8801c8ba3138 R09: ffff880125bfba68 R10: ffff8801abc989f8 R11: ffff8801abc99dd8 R12: ffff8802076ae718 R13: 0000000000000000 R14: ffff8802527166d8 R15: ffff8800a82db768 FS: 00007fa30d893740(0000) GS:ffff880331b40000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 000000020b274000 CR4: 00000000000007e0 Call Trace: [<ffffffffa1612b82>] ll_dir_ioctl+0x5d12/0x7290 [lustre] [<ffffffff811b518b>] ? unlock_page+0x2b/0x30 [<ffffffff811e5e5e>] ? do_read_fault.isra.64+0x12e/0x170 [<ffffffff811ea1ff>] ? handle_pte_fault+0x2ef/0xca0 [<ffffffff8103de0f>] ? save_stack_trace+0x1f/0x30 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff811ecf9d>] ? __handle_mm_fault+0x31d/0x720 [<ffffffff8125b47d>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff817e8655>] ? __do_page_fault+0x1c5/0x500 [<ffffffff8125b7d1>] SyS_ioctl+0xa1/0xc0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a | LustreError: 2212:0:(llite_lib.c:1873:ll_md_setattr()) md_setattr fails: rc = -28 LustreError: 2212:0:(file.c:243:ll_close_inode_openhandle()) lustre-clilmv-ffff88023e1bdd28: inode [0x200005223:0x94c5:0x0] mdc close failed: rc = -28 | Link to test |
sanity test 319: lost lease lock on migrate error | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa1615e7c>] ll_migrate+0x9ec/0xf50 [lustre] PGD 1f8104067 PUD 2930a4067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) lquota(OE) lfsck(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) nfsd ext4 mbcache jbd2 loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) crc32_generic crc_t10dif crct10dif_generic crct10dif_common virtio_console pcspkr virtio_balloon i2c_piix4 ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm ata_piix drm_panel_orientation_quirks floppy serio_raw virtio_blk libata i2c_core [last unloaded: obdecho] CPU: 8 PID: 6392 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.13.0-2.module_el8.5.0+746+bbd5d70c 04/01/2014 task: ffff8802d7cc3760 ti: ffff880278208000 task.ti: ffff880278208000 RIP: 0010:[<ffffffffa1615e7c>] [<ffffffffa1615e7c>] ll_migrate+0x9ec/0xf50 [lustre] RSP: 0018:ffff88027820bbd0 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff8801925466b8 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff88027c7e2188 RDI: ffff88027c7e2168 RBP: ffff88027820bc48 R08: ffff8802214996f8 R09: ffff88027820ba68 R10: ffff880192546d58 R11: ffff880192545cc8 R12: ffff8801a23e5908 R13: 0000000000000000 R14: ffff88023783c1d8 R15: ffff880293d31cd8 FS: 00007fdac7e59740(0000) GS:ffff880331c00000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 0000000191650000 CR4: 00000000000007e0 Call Trace: [<ffffffffa1602b82>] ll_dir_ioctl+0x5d12/0x7290 [lustre] [<ffffffff811b518b>] ? unlock_page+0x2b/0x30 [<ffffffff811e5e5e>] ? do_read_fault.isra.64+0x12e/0x170 [<ffffffff811ea1ff>] ? handle_pte_fault+0x2ef/0xca0 [<ffffffff8103de0f>] ? save_stack_trace+0x1f/0x30 [<ffffffff811ecf9d>] ? __handle_mm_fault+0x31d/0x720 [<ffffffff8125b47d>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff817e8655>] ? __do_page_fault+0x1c5/0x500 [<ffffffff8125b7d1>] SyS_ioctl+0xa1/0xc0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a | LustreError: 6262:0:(osd_handler.c:221:osd_trans_start()) lustre-MDT0000: failed to start transaction due to ENOSPC LustreError: 6262:0:(osd_handler.c:221:osd_trans_start()) Skipped 4 previous similar messages LustreError: 6341:0:(llite_lib.c:1873:ll_md_setattr()) md_setattr fails: rc = -28 LustreError: 6341:0:(file.c:243:ll_close_inode_openhandle()) lustre-clilmv-ffff880160b512a8: inode [0x2000059f3:0x8d39:0x0] mdc close failed: rc = -28 LustreError: 6341:0:(file.c:243:ll_close_inode_openhandle()) Skipped 1 previous similar message | Link to test |
sanity test 319: lost lease lock on migrate error | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa1602e7c>] ll_migrate+0x9ec/0xf50 [lustre] PGD 11a06d067 PUD 137aa3067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) lquota(OE) lfsck(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) nfsd ext4 mbcache jbd2 loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) crc32_generic crc_t10dif crct10dif_generic crct10dif_common i2c_piix4 virtio_balloon virtio_console pcspkr ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm drm_panel_orientation_quirks ata_piix serio_raw virtio_blk libata i2c_core floppy [last unloaded: obdecho] CPU: 8 PID: 22487 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.13.0-2.module_el8.5.0+746+bbd5d70c 04/01/2014 task: ffff88028d2149d0 ti: ffff8801a1d08000 task.ti: ffff8801a1d08000 RIP: 0010:[<ffffffffa1602e7c>] [<ffffffffa1602e7c>] ll_migrate+0x9ec/0xf50 [lustre] RSP: 0018:ffff8801a1d0bbd0 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff8801d6c8b508 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff880217601198 RDI: ffff880217601178 RBP: ffff8801a1d0bc48 R08: ffff88017e422f58 R09: ffff8801a1d0ba68 R10: ffff8801d6c8c248 R11: ffff8801d6c8b1b8 R12: ffff88016367ad58 R13: 0000000000000000 R14: ffff88020bcb2618 R15: ffff88026fff61e8 FS: 00007f960d284740(0000) GS:ffff880331c00000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 00000002c06ce000 CR4: 00000000000007e0 Call Trace: [<ffffffffa15efb82>] ll_dir_ioctl+0x5d12/0x7290 [lustre] [<ffffffff811b518b>] ? unlock_page+0x2b/0x30 [<ffffffff811e5e5e>] ? do_read_fault.isra.64+0x12e/0x170 [<ffffffff811ea1ff>] ? handle_pte_fault+0x2ef/0xca0 [<ffffffff8103de0f>] ? save_stack_trace+0x1f/0x30 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff811ecf9d>] ? __handle_mm_fault+0x31d/0x720 [<ffffffff8125b47d>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff817e8655>] ? __do_page_fault+0x1c5/0x500 [<ffffffff8125b7d1>] SyS_ioctl+0xa1/0xc0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a | LustreError: 24802:0:(osd_handler.c:221:osd_trans_start()) lustre-MDT0000: failed to start transaction due to ENOSPC LustreError: 24802:0:(osd_handler.c:221:osd_trans_start()) Skipped 1 previous similar message LustreError: 22468:0:(llite_lib.c:1873:ll_md_setattr()) md_setattr fails: rc = -28 LustreError: 22468:0:(file.c:243:ll_close_inode_openhandle()) lustre-clilmv-ffff88001fb0e678: inode [0x200005223:0x9376:0x0] mdc close failed: rc = -28 | Link to test |
sanity test 319: lost lease lock on migrate error | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa1652e7c>] ll_migrate+0x9ec/0xf50 [lustre] PGD 24563e067 PUD 23cdf7067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) lquota(OE) lfsck(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) nfsd ext4 mbcache jbd2 loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) crc32_generic crc_t10dif crct10dif_generic crct10dif_common virtio_balloon i2c_piix4 virtio_console pcspkr ip_tables rpcsec_gss_krb5 drm_kms_helper ata_generic pata_acpi ttm drm ata_piix drm_panel_orientation_quirks serio_raw virtio_blk i2c_core floppy libata [last unloaded: obdecho] CPU: 7 PID: 31999 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.13.0-2.module_el8.5.0+746+bbd5d70c 04/01/2014 task: ffff8802c6d3c9d0 ti: ffff880273c28000 task.ti: ffff880273c28000 RIP: 0010:[<ffffffffa1652e7c>] [<ffffffffa1652e7c>] ll_migrate+0x9ec/0xf50 [lustre] RSP: 0018:ffff880273c2bbd0 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff88016e4e3ba8 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff88009e404bd8 RDI: ffff88009e404bb8 RBP: ffff880273c2bc48 R08: ffff8800ace36d78 R09: ffff880273c2ba68 R10: ffff88016e4e3858 R11: ffff88016e4e6018 R12: ffff8801392e7208 R13: 0000000000000000 R14: ffff88026d9ddd98 R15: ffff880143da0ce8 FS: 00007f1571f52740(0000) GS:ffff880331bc0000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 000000013970c000 CR4: 00000000000007e0 Call Trace: [<ffffffffa163fb82>] ll_dir_ioctl+0x5d12/0x7290 [lustre] [<ffffffff811b518b>] ? unlock_page+0x2b/0x30 [<ffffffff811e5e5e>] ? do_read_fault.isra.64+0x12e/0x170 [<ffffffff811ea1ff>] ? handle_pte_fault+0x2ef/0xca0 [<ffffffff8103de0f>] ? save_stack_trace+0x1f/0x30 [<ffffffff811ecf9d>] ? __handle_mm_fault+0x31d/0x720 [<ffffffff8125b47d>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff817e8655>] ? __do_page_fault+0x1c5/0x500 [<ffffffff8125b7d1>] SyS_ioctl+0xa1/0xc0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a | LustreError: 31964:0:(llite_lib.c:1873:ll_md_setattr()) md_setattr fails: rc = -28 LustreError: 31964:0:(file.c:243:ll_close_inode_openhandle()) lustre-clilmv-ffff88006df55d28: inode [0x200005223:0x9466:0x0] mdc close failed: rc = -28 LustreError: 31964:0:(file.c:243:ll_close_inode_openhandle()) Skipped 2 previous similar messages | Link to test |
sanity test 319: lost lease lock on migrate error | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa15ffe7c>] ll_migrate+0x9ec/0xf50 [lustre] PGD 10745e067 PUD 13e57a067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) lquota(OE) lfsck(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) nfsd ext4 mbcache jbd2 loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) crc32_generic crc_t10dif crct10dif_generic crct10dif_common pcspkr virtio_console virtio_balloon i2c_piix4 ip_tables rpcsec_gss_krb5 drm_kms_helper ttm ata_generic drm pata_acpi drm_panel_orientation_quirks virtio_blk ata_piix serio_raw i2c_core libata floppy [last unloaded: obdecho] CPU: 1 PID: 5293 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 1.13.0-2.module_el8.5.0+746+bbd5d70c 04/01/2014 task: ffff880089a8c9d0 ti: ffff88019047c000 task.ti: ffff88019047c000 RIP: 0010:[<ffffffffa15ffe7c>] [<ffffffffa15ffe7c>] ll_migrate+0x9ec/0xf50 [lustre] RSP: 0018:ffff88019047fbd0 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff88006493ab18 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff88027d6e9048 RDI: ffff88027d6e9028 RBP: ffff88019047fc48 R08: ffff88009bf309d8 R09: ffff88019047fa68 R10: ffff88006493bba8 R11: ffff88006493ea08 R12: ffff88027bf13398 R13: 0000000000000000 R14: ffff8801631741d8 R15: ffff880049f39238 FS: 00007fb903ff8740(0000) GS:ffff880331a40000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 00000002d8b88000 CR4: 00000000000007e0 Call Trace: [<ffffffffa15ecb82>] ll_dir_ioctl+0x5d12/0x7290 [lustre] [<ffffffff811b518b>] ? unlock_page+0x2b/0x30 [<ffffffff811e5e5e>] ? do_read_fault.isra.64+0x12e/0x170 [<ffffffff811ea1ff>] ? handle_pte_fault+0x2ef/0xca0 [<ffffffff8103de0f>] ? save_stack_trace+0x1f/0x30 [<ffffffff811ecf9d>] ? __handle_mm_fault+0x31d/0x720 [<ffffffff8125b47d>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff817e8655>] ? __do_page_fault+0x1c5/0x500 [<ffffffff8125b7d1>] SyS_ioctl+0xa1/0xc0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a | LustreError: 5251:0:(llite_lib.c:1873:ll_md_setattr()) md_setattr fails: rc = -28 LustreError: 5251:0:(file.c:243:ll_close_inode_openhandle()) lustre-clilmv-ffff8800ae51b7e8: inode [0x2000059f3:0x92e8:0x0] mdc close failed: rc = -28 | Link to test |
sanity test 319: lost lease lock on migrate error | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa15ec01c>] ll_migrate+0x9ec/0xf50 [lustre] PGD 2e8e29067 PUD 2c5a99067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) lquota(OE) lfsck(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) nfsd ext4 mbcache jbd2 loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) crc32_generic crc_t10dif crct10dif_generic crct10dif_common virtio_console i2c_piix4 virtio_balloon pcspkr ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm ata_piix drm_panel_orientation_quirks virtio_blk floppy serio_raw libata i2c_core [last unloaded: obdecho] CPU: 11 PID: 5109 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #1 Hardware name: Red Hat KVM, BIOS 1.13.0-2.module_el8.5.0+746+bbd5d70c 04/01/2014 task: ffff880016fb9280 ti: ffff8802a0248000 task.ti: ffff8802a0248000 RIP: 0010:[<ffffffffa15ec01c>] [<ffffffffa15ec01c>] ll_migrate+0x9ec/0xf50 [lustre] RSP: 0018:ffff8802a024bbd0 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff8800230470a8 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff88027843cc98 RDI: ffff88027843cc78 RBP: ffff8802a024bc48 R08: ffff880255884438 R09: ffff8802a024ba68 R10: ffff880023043858 R11: ffff880023046018 R12: ffff88014352a3f8 R13: 0000000000000000 R14: ffff880021b92f58 R15: ffff88008cb27728 FS: 00007fe07e0b3740(0000) GS:ffff880331cc0000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 000000027dd78000 CR4: 00000000000007e0 Call Trace: [<ffffffffa15d8ece>] ll_dir_ioctl+0x618e/0x7560 [lustre] [<ffffffff812341c0>] ? __mem_cgroup_commit_charge+0xc0/0x390 [<ffffffff817e324e>] ? _raw_spin_unlock+0xe/0x20 [<ffffffff811ea55d>] ? handle_pte_fault+0x64d/0xca0 [<ffffffff8103de0f>] ? save_stack_trace+0x1f/0x30 [<ffffffff811ecf9d>] ? __handle_mm_fault+0x31d/0x720 [<ffffffff8125b47d>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff817e8655>] ? __do_page_fault+0x1c5/0x500 [<ffffffff8125b7d1>] SyS_ioctl+0xa1/0xc0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a | LustreError: 5071:0:(llite_lib.c:1871:ll_md_setattr()) md_setattr fails: rc = -28 LustreError: 5071:0:(file.c:243:ll_close_inode_openhandle()) lustre-clilmv-ffff8802c3efca88: inode [0x2000059f3:0x92ef:0x0] mdc close failed: rc = -28 LustreError: 5071:0:(file.c:243:ll_close_inode_openhandle()) Skipped 1 previous similar message | Link to test |
sanity test 319: lost lease lock on migrate error | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa15e718c>] ll_migrate+0x9ec/0xf50 [lustre] PGD 98170067 PUD 12886067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) lquota(OE) lfsck(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) ext4 mbcache jbd2 loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) crc32_generic crc_t10dif crct10dif_generic crct10dif_common pcspkr i2c_piix4 virtio_balloon virtio_console ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm drm_panel_orientation_quirks ata_piix serio_raw floppy i2c_core virtio_blk libata [last unloaded: obdecho] CPU: 10 PID: 25047 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #1 Hardware name: Red Hat KVM, BIOS 1.13.0-2.module_el8.5.0+746+bbd5d70c 04/01/2014 task: ffff880138b2c9d0 ti: ffff8802eee00000 task.ti: ffff8802eee00000 RIP: 0010:[<ffffffffa15e718c>] [<ffffffffa15e718c>] ll_migrate+0x9ec/0xf50 [lustre] RSP: 0018:ffff8802eee03bd0 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff88013252c248 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff8802765021e8 RDI: ffff8802765021c8 RBP: ffff8802eee03c48 R08: ffff8802859b5338 R09: ffff8802eee03a68 R10: ffff880174aaea08 R11: ffff880174aae368 R12: ffff8802c01c6a38 R13: 0000000000000000 R14: ffff8801d70f9cd8 R15: ffff880130cd0ce8 FS: 00007f5b8f6c7740(0000) GS:ffff880331c80000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 000000003d2f2000 CR4: 00000000000007e0 Call Trace: [<ffffffffa15d40a6>] ll_dir_ioctl+0x5c46/0x7080 [lustre] [<ffffffff811b518b>] ? unlock_page+0x2b/0x30 [<ffffffff811e5e5e>] ? do_read_fault.isra.64+0x12e/0x170 [<ffffffff811ea1ff>] ? handle_pte_fault+0x2ef/0xca0 [<ffffffff8103de0f>] ? save_stack_trace+0x1f/0x30 [<ffffffff811ecf9d>] ? __handle_mm_fault+0x31d/0x720 [<ffffffff8125b47d>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff817e8655>] ? __do_page_fault+0x1c5/0x500 [<ffffffff8125b7d1>] SyS_ioctl+0xa1/0xc0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a | LustreError: 25028:0:(llite_lib.c:1871:ll_md_setattr()) md_setattr fails: rc = -28 | Link to test |
sanity test 319: lost lease lock on migrate error | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa15f99bc>] ll_migrate+0x9ec/0xf50 [lustre] PGD 31ae2f067 PUD 12de4a067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) lquota(OE) lfsck(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) ext4 mbcache jbd2 loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) crc32_generic crc_t10dif crct10dif_generic crct10dif_common i2c_piix4 virtio_balloon pcspkr virtio_console ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm ata_piix drm_panel_orientation_quirks serio_raw virtio_blk i2c_core libata floppy [last unloaded: obdecho] CPU: 0 PID: 28816 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #1 Hardware name: Red Hat KVM, BIOS 1.13.0-2.module_el8.5.0+746+bbd5d70c 04/01/2014 task: ffff880290328010 ti: ffff88019f530000 task.ti: ffff88019f530000 RIP: 0010:[<ffffffffa15f99bc>] [<ffffffffa15f99bc>] ll_migrate+0x9ec/0xf50 [lustre] RSP: 0018:ffff88019f533bd0 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff880134e4cc38 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff8802914e7508 RDI: ffff8802914e74e8 RBP: ffff88019f533c48 R08: ffff8801e8448f78 R09: ffff88019f533a68 R10: ffff880134e486a8 R11: ffff880134e4f0a8 R12: ffff880098fc8008 R13: 0000000000000000 R14: ffff88003cf20a58 R15: ffff8802c0e78ce8 FS: 00007f4906fb8740(0000) GS:ffff880331a00000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 000000013f0c0000 CR4: 00000000000007f0 Call Trace: [<ffffffffa15e6bf7>] ll_dir_ioctl+0x5dc7/0x6ee0 [lustre] [<ffffffff811b518b>] ? unlock_page+0x2b/0x30 [<ffffffff811e5e5e>] ? do_read_fault.isra.64+0x12e/0x170 [<ffffffff811ea1ff>] ? handle_pte_fault+0x2ef/0xca0 [<ffffffff8103de0f>] ? save_stack_trace+0x1f/0x30 [<ffffffff811ecf9d>] ? __handle_mm_fault+0x31d/0x720 [<ffffffff8125b47d>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff817e8655>] ? __do_page_fault+0x1c5/0x500 [<ffffffff8125b7d1>] SyS_ioctl+0xa1/0xc0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a | LustreError: 28766:0:(llite_lib.c:1871:ll_md_setattr()) md_setattr fails: rc = -28 | Link to test |
sanity test 319: lost lease lock on migrate error | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa15ff9bc>] ll_migrate+0x9ec/0xf50 [lustre] PGD 14b6b0067 PUD 2874e1067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) lquota(OE) lfsck(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) ext4 mbcache jbd2 loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) crc32_generic crc_t10dif crct10dif_generic crct10dif_common virtio_balloon i2c_piix4 virtio_console pcspkr ip_tables rpcsec_gss_krb5 drm_kms_helper ata_generic ttm pata_acpi drm drm_panel_orientation_quirks serio_raw virtio_blk i2c_core ata_piix floppy libata [last unloaded: obdecho] CPU: 15 PID: 19452 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #1 Hardware name: Red Hat KVM, BIOS 1.13.0-2.module_el8.5.0+746+bbd5d70c 04/01/2014 task: ffff88032901dc40 ti: ffff8802f4e44000 task.ti: ffff8802f4e44000 RIP: 0010:[<ffffffffa15ff9bc>] [<ffffffffa15ff9bc>] ll_migrate+0x9ec/0xf50 [lustre] RSP: 0018:ffff8802f4e47bd0 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff88011dc27a98 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff8800ae7c42b8 RDI: ffff8800ae7c4298 RBP: ffff8802f4e47c48 R08: ffff88027791ef58 R09: ffff8802f4e47a68 R10: ffff88014b0bb1b8 R11: ffff88014b0bb858 R12: ffff880164850968 R13: 0000000000000000 R14: ffff8801b5c6dd98 R15: ffff88008b497c78 FS: 00007f86fca3a740(0000) GS:ffff880331dc0000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 0000000159100000 CR4: 00000000000007e0 Call Trace: [<ffffffffa15ecbf7>] ll_dir_ioctl+0x5dc7/0x6ee0 [lustre] [<ffffffff811b518b>] ? unlock_page+0x2b/0x30 [<ffffffff811e5e5e>] ? do_read_fault.isra.64+0x12e/0x170 [<ffffffff811ea1ff>] ? handle_pte_fault+0x2ef/0xca0 [<ffffffff8103de0f>] ? save_stack_trace+0x1f/0x30 [<ffffffff811ecf9d>] ? __handle_mm_fault+0x31d/0x720 [<ffffffff8125b47d>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff817e8655>] ? __do_page_fault+0x1c5/0x500 [<ffffffff8125b7d1>] SyS_ioctl+0xa1/0xc0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a | LustreError: 7577:0:(osd_handler.c:221:osd_trans_start()) lustre-MDT0000: failed to start transaction due to ENOSPC LustreError: 7577:0:(osd_handler.c:221:osd_trans_start()) Skipped 1 previous similar message | Link to test |
sanity test 319: lost lease lock on migrate error | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa15f39bc>] ll_migrate+0x9ec/0xf50 [lustre] PGD 27982d067 PUD 2f1bf2067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) lquota(OE) lfsck(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) nfsd ext4 mbcache jbd2 loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) crc32_generic crc_t10dif crct10dif_generic crct10dif_common virtio_console virtio_balloon pcspkr i2c_piix4 ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm ata_piix drm_panel_orientation_quirks floppy serio_raw virtio_blk i2c_core libata [last unloaded: obdecho] CPU: 7 PID: 32521 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #1 Hardware name: Red Hat KVM, BIOS 1.13.0-2.module_el8.5.0+746+bbd5d70c 04/01/2014 task: ffff8802ee74b760 ti: ffff88027b65c000 task.ti: ffff88027b65c000 RIP: 0010:[<ffffffffa15f39bc>] [<ffffffffa15f39bc>] ll_migrate+0x9ec/0xf50 [lustre] RSP: 0018:ffff88027b65fbd0 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff880084382e68 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff88008a0c0c18 RDI: ffff88008a0c0bf8 RBP: ffff88027b65fc48 R08: ffff8802594fb8b8 R09: ffff88027b65fa68 R10: ffff8800843870a8 R11: ffff880084383858 R12: ffff8802beb344b8 R13: 0000000000000000 R14: ffff880069c5f018 R15: ffff880324188798 FS: 00007fafe3de8740(0000) GS:ffff880331bc0000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 00000002bb950000 CR4: 00000000000007e0 Call Trace: [<ffffffffa15e0bf7>] ll_dir_ioctl+0x5dc7/0x6ee0 [lustre] [<ffffffff811b518b>] ? unlock_page+0x2b/0x30 [<ffffffff811e5e5e>] ? do_read_fault.isra.64+0x12e/0x170 [<ffffffff811ea1ff>] ? handle_pte_fault+0x2ef/0xca0 [<ffffffff8103de0f>] ? save_stack_trace+0x1f/0x30 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff811ecf9d>] ? __handle_mm_fault+0x31d/0x720 [<ffffffff8125b47d>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff817e8655>] ? __do_page_fault+0x1c5/0x500 [<ffffffff8125b7d1>] SyS_ioctl+0xa1/0xc0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a | LustreError: 5488:0:(osd_handler.c:221:osd_trans_start()) lustre-MDT0000: failed to start transaction due to ENOSPC LustreError: 32493:0:(llite_lib.c:1848:ll_md_setattr()) md_setattr fails: rc = -28 | Link to test |
sanity test 319: lost lease lock on migrate error | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa15ff9bc>] ll_migrate+0x9ec/0xf50 [lustre] PGD 92ae067 PUD 6e93f067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) lquota(OE) lfsck(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) nfsd ext4 mbcache jbd2 loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) crc32_generic crc_t10dif crct10dif_generic crct10dif_common virtio_balloon pcspkr virtio_console i2c_piix4 ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm drm_panel_orientation_quirks ata_piix floppy virtio_blk serio_raw i2c_core libata [last unloaded: obdecho] CPU: 10 PID: 30231 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #1 Hardware name: Red Hat KVM, BIOS 1.13.0-2.module_el8.5.0+746+bbd5d70c 04/01/2014 task: ffff88027e4b24f0 ti: ffff88006804c000 task.ti: ffff88006804c000 RIP: 0010:[<ffffffffa15ff9bc>] [<ffffffffa15ff9bc>] ll_migrate+0x9ec/0xf50 [lustre] RSP: 0018:ffff88006804fbd0 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff88023b13c598 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff88009e7431d8 RDI: ffff88009e7431b8 RBP: ffff88006804fc48 R08: ffff8801ee92e7d8 R09: ffff88006804fa68 R10: ffff880331403640 R11: ffff880288419098 R12: ffff8801e9c21f48 R13: 0000000000000000 R14: ffff8803212d1cd8 R15: ffff8801498ec208 FS: 00007f411ae49740(0000) GS:ffff880331c80000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 000000004ab18000 CR4: 00000000000007e0 Call Trace: [<ffffffffa15ecbf7>] ll_dir_ioctl+0x5dc7/0x6ee0 [lustre] [<ffffffff811b518b>] ? unlock_page+0x2b/0x30 [<ffffffff811e5e5e>] ? do_read_fault.isra.64+0x12e/0x170 [<ffffffff811ea1ff>] ? handle_pte_fault+0x2ef/0xca0 [<ffffffff8103de0f>] ? save_stack_trace+0x1f/0x30 [<ffffffff811ecf9d>] ? __handle_mm_fault+0x31d/0x720 [<ffffffff8125b47d>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff817e8655>] ? __do_page_fault+0x1c5/0x500 [<ffffffff8125b7d1>] SyS_ioctl+0xa1/0xc0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a | LustreError: 4658:0:(osd_handler.c:221:osd_trans_start()) lustre-MDT0000: failed to start transaction due to ENOSPC LustreError: 30206:0:(llite_lib.c:1848:ll_md_setattr()) md_setattr fails: rc = -28 | Link to test |
sanity test 319: lost lease lock on migrate error | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa15e69bc>] ll_migrate+0x9ec/0xf50 [lustre] PGD 171c27067 PUD 2e3607067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) lquota(OE) lfsck(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) ext4 mbcache jbd2 loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) crc32_generic crc_t10dif crct10dif_generic crct10dif_common virtio_balloon virtio_console pcspkr i2c_piix4 ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm serio_raw ata_piix drm_panel_orientation_quirks i2c_core virtio_blk libata floppy [last unloaded: obdecho] CPU: 1 PID: 29062 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #1 Hardware name: Red Hat KVM, BIOS 1.13.0-2.module_el8.5.0+746+bbd5d70c 04/01/2014 task: ffff88005086a4f0 ti: ffff880176ee8000 task.ti: ffff880176ee8000 RIP: 0010:[<ffffffffa15e69bc>] [<ffffffffa15e69bc>] ll_migrate+0x9ec/0xf50 [lustre] RSP: 0018:ffff880176eebbd0 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff88013e843ef8 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff8802aef53ac8 RDI: ffff8802aef53aa8 RBP: ffff880176eebc48 R08: ffff880062546d78 R09: ffff880176eeba68 R10: ffff880331403640 R11: ffff8801227cd2d8 R12: ffff8801346fd5e8 R13: 0000000000000000 R14: ffff88016d90f018 R15: ffff88027a068248 FS: 00007fe704dbf740(0000) GS:ffff880331a40000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 0000000106b72000 CR4: 00000000000007e0 Call Trace: [<ffffffffa15d3bf7>] ll_dir_ioctl+0x5dc7/0x6ee0 [lustre] [<ffffffff811b518b>] ? unlock_page+0x2b/0x30 [<ffffffff811e5e5e>] ? do_read_fault.isra.64+0x12e/0x170 [<ffffffff811ea1ff>] ? handle_pte_fault+0x2ef/0xca0 [<ffffffff8103de0f>] ? save_stack_trace+0x1f/0x30 [<ffffffff811ecf9d>] ? __handle_mm_fault+0x31d/0x720 [<ffffffff8125b47d>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff817e8655>] ? __do_page_fault+0x1c5/0x500 [<ffffffff8125b7d1>] SyS_ioctl+0xa1/0xc0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a | LustreError: 29030:0:(llite_lib.c:1848:ll_md_setattr()) md_setattr fails: rc = -28 | Link to test |
sanity test 319: lost lease lock on migrate error | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa15d99bc>] ll_migrate+0x9ec/0xf50 [lustre] PGD 167f26067 PUD 2bc3fc067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) lquota(OE) lfsck(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) ext4 mbcache jbd2 loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) crc32_generic crc_t10dif crct10dif_generic crct10dif_common virtio_console i2c_piix4 pcspkr virtio_balloon ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm drm_panel_orientation_quirks ata_piix floppy serio_raw i2c_core virtio_blk libata [last unloaded: obdecho] CPU: 0 PID: 8955 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #1 Hardware name: Red Hat KVM, BIOS 1.13.0-2.module_el8.5.0+746+bbd5d70c 04/01/2014 task: ffff8801ab09a4f0 ti: ffff880139110000 task.ti: ffff880139110000 RIP: 0010:[<ffffffffa15d99bc>] [<ffffffffa15d99bc>] ll_migrate+0x9ec/0xf50 [lustre] RSP: 0018:ffff880139113bd0 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff880162414248 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff880284f2fa58 RDI: ffff880284f2fa38 RBP: ffff880139113c48 R08: ffff880244aeab98 R09: ffff880139113a68 R10: ffff8802b9ee5628 R11: ffff8802b9ee7a98 R12: ffff88013fcf5458 R13: 0000000000000000 R14: ffff8802de805458 R15: ffff880074520ce8 FS: 00007fcf368ff740(0000) GS:ffff880331a00000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 000000028a29c000 CR4: 00000000000007f0 Call Trace: [<ffffffffa15c6bf7>] ll_dir_ioctl+0x5dc7/0x6ee0 [lustre] [<ffffffff811b518b>] ? unlock_page+0x2b/0x30 [<ffffffff811e5e5e>] ? do_read_fault.isra.64+0x12e/0x170 [<ffffffff811ea1ff>] ? handle_pte_fault+0x2ef/0xca0 [<ffffffff8103de0f>] ? save_stack_trace+0x1f/0x30 [<ffffffff811ecf9d>] ? __handle_mm_fault+0x31d/0x720 [<ffffffff8125b47d>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff817e8655>] ? __do_page_fault+0x1c5/0x500 [<ffffffff8125b7d1>] SyS_ioctl+0xa1/0xc0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a | LustreError: 17954:0:(osd_handler.c:221:osd_trans_start()) lustre-MDT0000: failed to start transaction due to ENOSPC LustreError: 8924:0:(llite_lib.c:1848:ll_md_setattr()) md_setattr fails: rc = -28 | Link to test |
sanity test 319: lost lease lock on migrate error | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa15fbc6c>] ll_migrate+0x9ec/0xf50 [lustre] PGD 2c5da2067 PUD 170bf8067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) lquota(OE) lfsck(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) nfsd ext4 mbcache jbd2 loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) crc32_generic crc_t10dif crct10dif_generic crct10dif_common pcspkr i2c_piix4 virtio_console virtio_balloon ip_tables rpcsec_gss_krb5 drm_kms_helper ttm ata_generic pata_acpi drm drm_panel_orientation_quirks ata_piix virtio_blk floppy serio_raw i2c_core libata [last unloaded: obdecho] CPU: 10 PID: 25579 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #1 Hardware name: Red Hat KVM, BIOS 1.13.0-2.module_el8.5.0+746+bbd5d70c 04/01/2014 task: ffff8803296e9280 ti: ffff880241e10000 task.ti: ffff880241e10000 RIP: 0010:[<ffffffffa15fbc6c>] [<ffffffffa15fbc6c>] ll_migrate+0x9ec/0xf50 [lustre] RSP: 0018:ffff880241e13bd0 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff8801ff73c248 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff880282b32c88 RDI: ffff880282b32c68 RBP: ffff880241e13c48 R08: ffff880289b35e78 R09: ffff880241e13a68 R10: ffff8801ff73ae68 R11: ffff8801ff73b508 R12: ffff880175816268 R13: 0000000000000000 R14: ffff88001965cb18 R15: ffff880207706738 FS: 00007f342dcca740(0000) GS:ffff880331c80000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 0000000298c0e000 CR4: 00000000000007e0 Call Trace: [<ffffffffa15e8b86>] ll_dir_ioctl+0x5c46/0x7080 [lustre] [<ffffffff811b518b>] ? unlock_page+0x2b/0x30 [<ffffffff811e5e5e>] ? do_read_fault.isra.64+0x12e/0x170 [<ffffffff811ea1ff>] ? handle_pte_fault+0x2ef/0xca0 [<ffffffff8103de0f>] ? save_stack_trace+0x1f/0x30 [<ffffffff811ecf9d>] ? __handle_mm_fault+0x31d/0x720 [<ffffffff8125b47d>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff817e8655>] ? __do_page_fault+0x1c5/0x500 [<ffffffff8125b7d1>] SyS_ioctl+0xa1/0xc0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a | LustreError: 18534:0:(osd_handler.c:221:osd_trans_start()) lustre-MDT0000: failed to start transaction due to ENOSPC LustreError: 18534:0:(osd_handler.c:221:osd_trans_start()) Skipped 1 previous similar message LustreError: 25563:0:(llite_lib.c:1848:ll_md_setattr()) md_setattr fails: rc = -28 | Link to test |
sanity test 319: lost lease lock on migrate error | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa15fac6c>] ll_migrate+0x9ec/0xf50 [lustre] PGD 2cea23067 PUD 1a7a92067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) lquota(OE) lfsck(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) ext4 mbcache jbd2 loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) crc32_generic crc_t10dif crct10dif_generic crct10dif_common virtio_balloon i2c_piix4 virtio_console pcspkr ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm ata_piix drm_panel_orientation_quirks serio_raw libata virtio_blk floppy i2c_core [last unloaded: obdecho] CPU: 4 PID: 25266 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #1 Hardware name: Red Hat KVM, BIOS 1.13.0-2.module_el8.5.0+746+bbd5d70c 04/01/2014 task: ffff88032000c9d0 ti: ffff88016dd38000 task.ti: ffff88016dd38000 RIP: 0010:[<ffffffffa15fac6c>] [<ffffffffa15fac6c>] ll_migrate+0x9ec/0xf50 [lustre] RSP: 0018:ffff88016dd3bbd0 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff88001f38cf88 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff88027d13b0b8 RDI: ffff88027d13b098 RBP: ffff88016dd3bc48 R08: ffff88000b9f29b8 R09: ffff88016dd3ba68 R10: ffff88001f38ed58 R11: ffff88001f388d48 R12: ffff88016012ea38 R13: 0000000000000000 R14: ffff8801b5bf0a58 R15: ffff880276fe4208 FS: 00007f479acf3740(0000) GS:ffff880331b00000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 00000002c99ba000 CR4: 00000000000007e0 Call Trace: [<ffffffffa15e7b86>] ll_dir_ioctl+0x5c46/0x7080 [lustre] [<ffffffff811b518b>] ? unlock_page+0x2b/0x30 [<ffffffff811e5e5e>] ? do_read_fault.isra.64+0x12e/0x170 [<ffffffff811ea1ff>] ? handle_pte_fault+0x2ef/0xca0 [<ffffffff8103de0f>] ? save_stack_trace+0x1f/0x30 [<ffffffff811ecf9d>] ? __handle_mm_fault+0x31d/0x720 [<ffffffff8125b47d>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff817e8655>] ? __do_page_fault+0x1c5/0x500 [<ffffffff8125b7d1>] SyS_ioctl+0xa1/0xc0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a | LustreError: 30819:0:(osd_handler.c:221:osd_trans_start()) lustre-MDT0000: failed to start transaction due to ENOSPC LustreError: 30819:0:(osd_handler.c:221:osd_trans_start()) Skipped 1 previous similar message | Link to test |
sanity test 319: lost lease lock on migrate error | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa15eec6c>] ll_migrate+0x9ec/0xf50 [lustre] PGD 18e33b067 PUD 2f84b1067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) lquota(OE) lfsck(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) ext4 mbcache jbd2 loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) crc32_generic crc_t10dif crct10dif_generic crct10dif_common virtio_balloon i2c_piix4 virtio_console pcspkr ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm ata_piix drm drm_panel_orientation_quirks libata i2c_core virtio_blk serio_raw floppy [last unloaded: obdecho] CPU: 14 PID: 18832 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #1 Hardware name: Red Hat KVM, BIOS 1.13.0-2.module_el8.5.0+746+bbd5d70c 04/01/2014 task: ffff88026e64dc40 ti: ffff88011be90000 task.ti: ffff88011be90000 RIP: 0010:[<ffffffffa15eec6c>] [<ffffffffa15eec6c>] ll_migrate+0x9ec/0xf50 [lustre] RSP: 0018:ffff88011be93bd0 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff880130b7bba8 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff8802c6fced08 RDI: ffff8802c6fcece8 RBP: ffff88011be93c48 R08: ffff88026dee9338 R09: ffff88011be93a68 R10: ffff880130b789f8 R11: ffff880130b7c598 R12: ffff88019c74e718 R13: 0000000000000000 R14: ffff8801027b66d8 R15: ffff8800832d4758 FS: 00007fc7a5c16740(0000) GS:ffff880331d80000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 000000019c750000 CR4: 00000000000007e0 Call Trace: [<ffffffffa15dbb86>] ll_dir_ioctl+0x5c46/0x7080 [lustre] [<ffffffff811b518b>] ? unlock_page+0x2b/0x30 [<ffffffff811e5e5e>] ? do_read_fault.isra.64+0x12e/0x170 [<ffffffff811ea1ff>] ? handle_pte_fault+0x2ef/0xca0 [<ffffffff8103de0f>] ? save_stack_trace+0x1f/0x30 [<ffffffff811ecf9d>] ? __handle_mm_fault+0x31d/0x720 [<ffffffff8125b47d>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff817e8655>] ? __do_page_fault+0x1c5/0x500 [<ffffffff8125b7d1>] SyS_ioctl+0xa1/0xc0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a | LustreError: 19403:0:(osd_handler.c:221:osd_trans_start()) lustre-MDT0000: failed to start transaction due to ENOSPC LustreError: 19403:0:(osd_handler.c:221:osd_trans_start()) Skipped 4 previous similar messages LustreError: 18801:0:(llite_lib.c:1848:ll_md_setattr()) md_setattr fails: rc = -28 | Link to test |
sanity test 319: lost lease lock on migrate error | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa15eef9c>] ll_migrate+0x9ec/0xf50 [lustre] PGD 293e82067 PUD 13c40a067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) lquota(OE) lfsck(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) ext4 mbcache jbd2 loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) crc32_generic crc_t10dif crct10dif_generic crct10dif_common pcspkr i2c_piix4 virtio_balloon virtio_console ip_tables rpcsec_gss_krb5 ata_generic drm_kms_helper pata_acpi ttm drm drm_panel_orientation_quirks ata_piix virtio_blk serio_raw libata i2c_core floppy [last unloaded: obdecho] CPU: 9 PID: 13185 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #1 Hardware name: Red Hat KVM, BIOS 1.13.0-2.module_el8.5.0+746+bbd5d70c 04/01/2014 task: ffff8801fb4c8010 ti: ffff8802878f4000 task.ti: ffff8802878f4000 RIP: 0010:[<ffffffffa15eef9c>] [<ffffffffa15eef9c>] ll_migrate+0x9ec/0xf50 [lustre] RSP: 0018:ffff8802878f7bd0 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff88016a1d2e68 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff88009fc6dc28 RDI: ffff88009fc6dc08 RBP: ffff8802878f7c48 R08: ffff880295ebef58 R09: ffff8802878f7a68 R10: ffff880331403640 R11: ffff88018dd0cc38 R12: ffff8802e07e0c88 R13: 0000000000000000 R14: ffff8800458fdd98 R15: ffff8802daeeec88 FS: 00007f7d32473740(0000) GS:ffff880331c40000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 00000001a0678000 CR4: 00000000000007e0 Call Trace: [<ffffffffa15dca2d>] ll_dir_ioctl+0x542d/0x61e0 [lustre] [<ffffffff811b518b>] ? unlock_page+0x2b/0x30 [<ffffffff811e5e5e>] ? do_read_fault.isra.64+0x12e/0x170 [<ffffffff811ea1ff>] ? handle_pte_fault+0x2ef/0xca0 [<ffffffff8103de0f>] ? save_stack_trace+0x1f/0x30 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff811ecf9d>] ? __handle_mm_fault+0x31d/0x720 [<ffffffff8125b47d>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff817e8655>] ? __do_page_fault+0x1c5/0x500 [<ffffffff8125b7d1>] SyS_ioctl+0xa1/0xc0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a | LustreError: 21273:0:(osd_handler.c:221:osd_trans_start()) lustre-MDT0000: failed to start transaction due to ENOSPC LustreError: 21273:0:(osd_handler.c:221:osd_trans_start()) Skipped 2 previous similar messages LustreError: 13157:0:(llite_lib.c:1843:ll_md_setattr()) md_setattr fails: rc = -28 | Link to test |
sanity test 319: lost lease lock on migrate error | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa15ebf9c>] ll_migrate+0x9ec/0xf50 [lustre] PGD 11d069067 PUD 122724067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_zfs(OE) lquota(OE) lfsck(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) ext4 mbcache jbd2 loop zfs(PO) zunicode(PO) zlua(PO) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) crc32_generic crc_t10dif crct10dif_generic crct10dif_common virtio_console virtio_balloon i2c_piix4 pcspkr ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm drm_panel_orientation_quirks ata_piix floppy serio_raw i2c_core libata virtio_blk [last unloaded: obdecho] CPU: 14 PID: 11829 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #1 Hardware name: Red Hat KVM, BIOS 1.13.0-2.module_el8.5.0+746+bbd5d70c 04/01/2014 task: ffff880183e88010 ti: ffff88012aaa8000 task.ti: ffff88012aaa8000 RIP: 0010:[<ffffffffa15ebf9c>] [<ffffffffa15ebf9c>] ll_migrate+0x9ec/0xf50 [lustre] RSP: 0018:ffff88012aaabbd0 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff8800859f4c38 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff880068d4e7b8 RDI: ffff880068d4e798 RBP: ffff88012aaabc48 R08: ffff88016d2256f8 R09: ffff88012aaaba68 R10: ffff8800859f3ef8 R11: ffff8800859f27c8 R12: ffff8800202136b8 R13: 0000000000000000 R14: ffff8801640a9398 R15: ffff880199b42cc8 FS: 00007fea6d36e740(0000) GS:ffff880331d80000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000000000000008 CR3: 000000024a406000 CR4: 00000000000007e0 Call Trace: [<ffffffffa15d9a2d>] ll_dir_ioctl+0x542d/0x61e0 [lustre] [<ffffffff811b518b>] ? unlock_page+0x2b/0x30 [<ffffffff811e5e5e>] ? do_read_fault.isra.64+0x12e/0x170 [<ffffffff811ea1ff>] ? handle_pte_fault+0x2ef/0xca0 [<ffffffff8103de0f>] ? save_stack_trace+0x1f/0x30 [<ffffffff814119f9>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff811ecf9d>] ? __handle_mm_fault+0x31d/0x720 [<ffffffff8125b47d>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff817e8655>] ? __do_page_fault+0x1c5/0x500 [<ffffffff8125b7d1>] SyS_ioctl+0xa1/0xc0 [<ffffffff817edf49>] ? system_call_after_swapgs+0x96/0x13a [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 [<ffffffff817edf55>] ? system_call_after_swapgs+0xa2/0x13a | LustreError: 17085:0:(osd_handler.c:221:osd_trans_start()) lustre-MDT0000: failed to start transaction due to ENOSPC LustreError: 11797:0:(llite_lib.c:1843:ll_md_setattr()) md_setattr fails: rc = -28 | Link to test |
racer test 1: racer on clients: centos-85.localnet DURATION=2700 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa104de52>] ll_migrate+0x9b2/0xec0 [lustre] PGD 800000008c2c5067 PUD 81ad1067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: loop zfs(PO) zunicode(PO) zzstd(O) zlua(O) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) jbd2 mbcache lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) dm_flakey dm_mod libcfs(OE) crc_t10dif crct10dif_generic sb_edac edac_core iosf_mbi crc32_pclmul ghash_clmulni_intel aesni_intel lrw gf128mul glue_helper ablk_helper cryptd virtio_balloon pcspkr virtio_console i2c_piix4 ip_tables rpcsec_gss_krb5 drm_kms_helper ata_generic pata_acpi ttm crct10dif_pclmul drm crct10dif_common ata_piix crc32c_intel drm_panel_orientation_quirks virtio_blk libata serio_raw i2c_core floppy CPU: 11 PID: 30710 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011 task: ffff880326a7c9d0 ti: ffff880265bbc000 task.ti: ffff880265bbc000 RIP: 0010:[<ffffffffa104de52>] [<ffffffffa104de52>] ll_migrate+0x9b2/0xec0 [lustre] RSP: 0018:ffff880265bbfbc8 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff880242aa4968 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff880321fab758 RDI: ffff880321fab738 RBP: ffff880265bbfc48 R08: ffff88025f601158 R09: 0000000000000001 R10: 0000000000000000 R11: ffff880265bbf5e6 R12: 0000000000000000 R13: ffff880261382128 R14: ffff8800911a9cd8 R15: 0000000000000030 FS: 00007ff7995e6740(0000) GS:ffff880331cc0000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 CR2: 0000000000000008 CR3: 0000000267ec2000 CR4: 00000000001607e0 Call Trace: [<ffffffff81242213>] ? __check_object_size+0x1c3/0x220 [<ffffffffa103b561>] ll_dir_ioctl+0x5d01/0x6ed0 [lustre] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff8115260f>] ? delayacct_end+0x8f/0xb0 [<ffffffff811b62dd>] ? find_get_pages_tag+0x10d/0x260 [<ffffffff811c3691>] ? pagevec_lookup_tag+0x21/0x30 [<ffffffff811b400e>] ? __filemap_fdatawait_range+0xbe/0x1a0 [<ffffffff8125b3fd>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff81264d2b>] ? iput+0x3b/0x180 [<ffffffff8125b751>] SyS_ioctl+0xa1/0xc0 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 | Lustre: lustre-OST0000-osc-ffff880298df53d8: disconnect after 21s idle Lustre: 11097:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 8: before 510 < left 6612, rollback = 9 Lustre: 11097:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 388/1552/0, destroy: 1/4/0 Lustre: 11097:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 3/3/1, xattr_set: 392/148/0 Lustre: 11097:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1941/16694/0, punch: 0/0/0, quota 6/6/0 Lustre: 11097:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 389/6612/0, delete: 2/5/1 Lustre: 11097:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 1/1/0, ref_del: 2/2/0 Lustre: 11097:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 509 < left 16694, rollback = 7 Lustre: 11113:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 511 < left 16694, rollback = 7 Lustre: 11113:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 388/1552/0, destroy: 1/4/0 Lustre: 11113:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 11113:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 388/0/0 Lustre: 11113:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 11113:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1941/16694/0, punch: 0/0/0, quota 6/6/0 Lustre: 11113:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 11113:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 388/6596/0, delete: 2/5/0 Lustre: 11113:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 11113:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 1/1/1 Lustre: 11113:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 11113:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 510 < left 16693, rollback = 7 Lustre: 11113:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 388/1552/0, destroy: 1/4/0 Lustre: 16626:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 504 < left 582, rollback = 2 Lustre: 16626:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 6/582/0 Lustre: 16626:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 16626:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 5/136/0, punch: 0/0/0, quota 8/56/0 Lustre: 16626:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 16626:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 10/203/4, delete: 0/0/0 Lustre: 16626:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 16626:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 4/4/0, ref_del: 0/0/0 Lustre: 16626:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 11113:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 474 previous similar messages ------------[ cut here ]------------ WARNING: CPU: 2 PID: 11113 at /home/green/git/lustre-release/ldiskfs/namei.c:3331 ldiskfs_orphan_add+0x11e/0x2a0 [ldiskfs] Modules linked in: loop zfs(PO) zunicode(PO) zzstd(O) zlua(O) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) jbd2 mbcache lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) dm_flakey dm_mod libcfs(OE) crc_t10dif crct10dif_generic sb_edac edac_core iosf_mbi crc32_pclmul ghash_clmulni_intel aesni_intel lrw gf128mul glue_helper ablk_helper cryptd virtio_balloon pcspkr virtio_console i2c_piix4 ip_tables rpcsec_gss_krb5 drm_kms_helper ata_generic pata_acpi ttm crct10dif_pclmul drm crct10dif_common ata_piix crc32c_intel drm_panel_orientation_quirks virtio_blk libata serio_raw i2c_core floppy CPU: 2 PID: 11113 Comm: mdt_rdpg01_001 Tainted: P OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011 Call Trace: [<ffffffff817d9418>] dump_stack+0x19/0x1b [<ffffffff8108d538>] __warn+0xd8/0x100 [<ffffffff8108d67d>] warn_slowpath_null+0x1d/0x20 [<ffffffffa0be54ae>] ldiskfs_orphan_add+0x11e/0x2a0 [ldiskfs] [<ffffffffa0bd69eb>] ldiskfs_xattr_inode_orphan_add.constprop.17+0xbb/0x110 [ldiskfs] [<ffffffffa0bd9417>] ldiskfs_xattr_delete_inode+0x77/0x3c0 [ldiskfs] [<ffffffffa0bf9aa0>] ? ldiskfs_evict_inode+0x190/0x670 [ldiskfs] [<ffffffffa0bf9ad9>] ldiskfs_evict_inode+0x1c9/0x670 [ldiskfs] [<ffffffff812649af>] evict+0xaf/0x180 [<ffffffff81264de5>] iput+0xf5/0x180 [<ffffffffa0c8ed08>] osd_object_delete+0x1f8/0x370 [osd_ldiskfs] [<ffffffffa04d859d>] lu_object_free.isra.26+0xcd/0x1d0 [obdclass] [<ffffffffa04dccf1>] lu_object_put+0xa1/0x480 [obdclass] [<ffffffffa0e46af4>] mdt_mfd_close+0x1f4/0x870 [mdt] [<ffffffffa0e4ce14>] mdt_close_internal+0xb4/0x240 [mdt] [<ffffffffa0e4d224>] mdt_close+0x284/0x850 [mdt] [<ffffffffa07df2d4>] ? tgt_request_preprocess.isra.27+0xf4/0x850 [ptlrpc] [<ffffffffa07e026e>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<ffffffffa032d45e>] ? libcfs_nid2str_r+0xfe/0x130 [lnet] [<ffffffffa078b990>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<ffffffffa078d529>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e32be>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffffa078c950>] ? ptlrpc_wait_event+0x620/0x620 [ptlrpc] [<ffffffff810ba0f4>] kthread+0xe4/0xf0 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 ---[ end trace 28bda789721aa7ff ]--- Lustre: 4129:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff88028a469940 x1710331742644800/t0(0) o10->lustre-OST0001-osc-ffff880298df53d8@0@lo:6/4 lens 440/432 e 0 to 0 dl 1631099681 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'truncate.0' Lustre: 11089:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 505 < left 521, rollback = 2 Lustre: 11089:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 5 previous similar messages Lustre: 11089:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 3/12/3, destroy: 0/0/0 Lustre: 11089:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 2321 previous similar messages Lustre: 11089:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 2/2/0, xattr_set: 7/521/0 Lustre: 11089:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 2871 previous similar messages Lustre: 11089:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 5/136/0, punch: 0/0/0, quota 8/56/0 Lustre: 11089:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 2871 previous similar messages Lustre: 11089:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 12/215/4, delete: 1/1/0 Lustre: 11089:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 2869 previous similar messages Lustre: 11089:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 6/6/0, ref_del: 1/1/0 Lustre: 11089:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 2869 previous similar messages LustreError: 12289:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0001-osp-MDT0002: fail to cancel 1 llog-records: rc = -116 LustreError: 12289:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0001-osp-MDT0002: fail to cancel 1 of 1 llog-records: rc = -116 Lustre: 16601:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 512 < left 1903, rollback = 7 Lustre: 16601:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 2867 previous similar messages Lustre: 18306:0:(service.c:1378:ptlrpc_at_send_early_reply()) @@@ Already past deadline (-1s), not sending early reply. Consider increasing at_early_margin (5)? req@ffff880287b9d180 x1710331742999936/t0(0) o1->bc1e05b8-368a-4968-928a-1b45107d748e@0@lo:712/0 lens 440/0 e 0 to 0 dl 1631099692 ref 2 fl New:/0/ffffffff rc 0/-1 job:'lfs.0' Lustre: 11104:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 512 < left 1746, rollback = 7 Lustre: 11104:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 1893 previous similar messages Lustre: 11104:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 11104:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 1897 previous similar messages Lustre: 11104:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 1/89/0 Lustre: 11104:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 1897 previous similar messages Lustre: 11104:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1737/1746/0, punch: 0/0/0, quota 3/3/0 Lustre: 11104:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 1897 previous similar messages Lustre: 11104:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 11104:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 1897 previous similar messages Lustre: 11104:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 11104:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 1897 previous similar messages LustreError: 18301:0:(service.c:2290:ptlrpc_server_handle_request()) @@@ Dropping timed-out request from 12345-0@lo: deadline 6/5s ago req@ffff880275fc1f80 x1710331742994176/t0(0) o1->bc1e05b8-368a-4968-928a-1b45107d748e@0@lo:712/0 lens 440/0 e 0 to 0 dl 1631099692 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'lfs.0' Lustre: 18330:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (6/5s); client may timeout req@ffff88027548ea80 x1710331742994432/t0(0) o1->bc1e05b8-368a-4968-928a-1b45107d748e@0@lo:712/0 lens 440/0 e 0 to 0 dl 1631099692 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'lfs.0' LustreError: 18301:0:(service.c:2290:ptlrpc_server_handle_request()) Skipped 2 previous similar messages LustreError: 18301:0:(service.c:2290:ptlrpc_server_handle_request()) @@@ Dropping timed-out request from 12345-0@lo: deadline 6/5s ago req@ffff88028018de00 x1710331742999616/t0(0) o1->bc1e05b8-368a-4968-928a-1b45107d748e@0@lo:712/0 lens 440/0 e 0 to 0 dl 1631099692 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'lfs.0' LustreError: 18301:0:(service.c:2290:ptlrpc_server_handle_request()) Skipped 14 previous similar messages Lustre: 18301:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (6/6s); client may timeout req@ffff88028018de00 x1710331742999616/t0(0) o1->bc1e05b8-368a-4968-928a-1b45107d748e@0@lo:712/0 lens 440/0 e 0 to 0 dl 1631099692 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'lfs.0' Lustre: 18301:0:(service.c:2328:ptlrpc_server_handle_request()) Skipped 16 previous similar messages Lustre: 18421:0:(service.c:1378:ptlrpc_at_send_early_reply()) @@@ Already past deadline (-1s), not sending early reply. Consider increasing at_early_margin (5)? req@ffff880279530680 x1710331743026752/t0(0) o1->bc1e05b8-368a-4968-928a-1b45107d748e@0@lo:717/0 lens 440/0 e 0 to 0 dl 1631099697 ref 2 fl New:/0/ffffffff rc 0/-1 job:'lfs.0' Lustre: 18421:0:(service.c:1378:ptlrpc_at_send_early_reply()) Skipped 17 previous similar messages LustreError: 18350:0:(service.c:2290:ptlrpc_server_handle_request()) @@@ Dropping timed-out request from 12345-0@lo: deadline 6/7s ago req@ffff88027ec78040 x1710331743023552/t0(0) o1->bc1e05b8-368a-4968-928a-1b45107d748e@0@lo:717/0 lens 440/0 e 0 to 0 dl 1631099697 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'lfs.0' Lustre: 18315:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (6/7s); client may timeout req@ffff88027ec7bec0 x1710331743023744/t0(0) o1->bc1e05b8-368a-4968-928a-1b45107d748e@0@lo:717/0 lens 440/0 e 0 to 0 dl 1631099697 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'lfs.0' LustreError: 18350:0:(service.c:2290:ptlrpc_server_handle_request()) Skipped 17 previous similar messages Lustre: 4139:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1631099686/real 1631099686] req@ffff880277e38680 x1710331742994176/t0(0) o1->lustre-OST0003-osc-ffff880298dfae98@0@lo:28/4 lens 440/432 e 0 to 1 dl 1631099695 ref 1 fl Rpc:XQr/0/ffffffff rc 0/-1 job:'lfs.0' Lustre: lustre-OST0003-osc-ffff880298dfae98: Connection to lustre-OST0003 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: lustre-OST0003: Client bc1e05b8-368a-4968-928a-1b45107d748e (at 0@lo) reconnecting Lustre: lustre-OST0003-osc-ffff880298dfae98: Connection restored to 192.168.123.185@tcp (at 0@lo) Lustre: 4140:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1631099691/real 1631099691] req@ffff880085e31300 x1710331743023552/t0(0) o1->lustre-OST0002-osc-ffff880298dfae98@0@lo:28/4 lens 440/432 e 0 to 1 dl 1631099703 ref 1 fl Rpc:XQr/0/ffffffff rc 0/-1 job:'lfs.0' Lustre: lustre-OST0002-osc-ffff880298dfae98: Connection to lustre-OST0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: lustre-OST0000: Client bc1e05b8-368a-4968-928a-1b45107d748e (at 0@lo) reconnecting Lustre: lustre-OST0000-osc-ffff880298dfae98: Connection restored to 192.168.123.185@tcp (at 0@lo) Lustre: 17011:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1631099668/real 1631099668] req@ffff8802849b9f80 x1710331742473088/t0(0) o101->lustre-MDT0000-mdc-ffff880298df53d8@0@lo:12/10 lens 376/32368 e 0 to 1 dl 1631099712 ref 2 fl Rpc:XQr/2/ffffffff rc 0/-1 job:'dd.0' Lustre: 17011:0:(client.c:2285:ptlrpc_expire_one_request()) Skipped 17 previous similar messages Lustre: lustre-MDT0000-mdc-ffff880298df53d8: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 1 previous similar message Lustre: lustre-MDT0000: Client 1154ae0e-996a-42a4-bcbf-d44ba44315ac (at 0@lo) reconnecting Lustre: Skipped 1 previous similar message Lustre: lustre-MDT0000-mdc-ffff880298df53d8: Connection restored to 192.168.123.185@tcp (at 0@lo) Lustre: Skipped 1 previous similar message Lustre: 18407:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 505 < left 699, rollback = 2 Lustre: 18407:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 3 previous similar messages Lustre: 18407:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 3/12/5, destroy: 0/0/0 Lustre: 18407:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 1736 previous similar messages Lustre: 18407:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 2/2/0, xattr_set: 9/699/0 Lustre: 18407:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 1736 previous similar messages Lustre: 18407:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 5/136/0, punch: 0/0/0, quota 8/56/0 Lustre: 18407:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 1736 previous similar messages Lustre: 18407:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 12/215/2, delete: 1/1/0 Lustre: 18407:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 1736 previous similar messages Lustre: 18407:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 6/6/0, ref_del: 1/1/0 Lustre: 18407:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 1736 previous similar messages sched: RT throttling activated Lustre: 11089:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 8: before 509 < left 22405, rollback = 9 Lustre: 11089:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 7 previous similar messages Lustre: 11089:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 1317/5268/0, destroy: 1/4/0 Lustre: 11089:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 7 previous similar messages Lustre: 11089:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 3/3/1, xattr_set: 1321/148/0 Lustre: 11089:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 7 previous similar messages Lustre: 11089:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 6586/56641/0, punch: 0/0/0, quota 6/6/0 Lustre: 11089:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 7 previous similar messages Lustre: 11089:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 1318/22405/0, delete: 2/5/1 Lustre: 11089:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 7 previous similar messages Lustre: 11089:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 1/1/0, ref_del: 2/2/1 Lustre: 11089:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 7 previous similar messages Lustre: 11089:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 507 < left 56641, rollback = 7 Lustre: 11089:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 1736 previous similar messages LustreError: 10998:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0001_UUID lock: ffff88008ab587c0/0x4105371653f934af lrc: 3/0,0 mode: PW/PW res: [0x300000400:0x5:0x0].0x0 rrc: 4 type: EXT [0->18446744073709551615] (req 0->1048575) gid 0 flags: 0x60000400020020 nid: 0@lo remote: 0x4105371653f934a8 expref: 2784 pid: 17754 timeout: 451 lvb_type: 0 LustreError: 18382:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff880275eb7700 x1710331746390528/t0(0) o106->lustre-OST0001@0@lo:15/16 lens 328/280 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 18348:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802652e3ec0 x1710331746405440/t0(0) o104->lustre-OST0001@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 11-0: lustre-OST0001-osc-ffff880298dfae98: operation ost_setattr to node 0@lo failed: rc = -107 Lustre: lustre-OST0001-osc-ffff880298dfae98: Connection to lustre-OST0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0001-osc-ffff880298dfae98: This client was evicted by lustre-OST0001; in progress operations using this service will fail. LustreError: 17953:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802772df0c0 x1710331746466560/t0(0) o106->lustre-OST0001@0@lo:15/16 lens 328/280 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 17953:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 2 previous similar messages Lustre: 4145:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.185@tcp:/lustre/fid: [0x280000404:0x52:0x0]// may get corrupted (rc -108) LustreError: 19981:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0001-osc-ffff880298dfae98: namespace resource [0x160:0x0:0x0].0x0 (ffff88025d2851c0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 19981:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0001-osc-ffff880298dfae98: namespace resource [0x300000401:0xc6:0x0].0x0 (ffff880263e8ca40) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 19981:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 128 previous similar messages Lustre: lustre-OST0001-osc-ffff880298dfae98: Connection restored to 192.168.123.185@tcp (at 0@lo) Lustre: 16886:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 6s req@ffff880264490cc0 x1710331746747136/t0(0) o41->lustre-MDT0002-mdtlov_UUID@0@lo:0/0 lens 224/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'osp-pre-0-2.0' Lustre: 16526:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 506 < left 5170, rollback = 7 Lustre: 16526:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 1966 previous similar messages Lustre: 16526:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 120/480/0, destroy: 0/0/0 Lustre: 16526:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 1967 previous similar messages Lustre: 16526:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/1, xattr_set: 122/15/0 Lustre: 16526:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 1967 previous similar messages Lustre: 16526:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 601/5170/0, punch: 0/0/0, quota 10/42/5 Lustre: 16526:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 1967 previous similar messages Lustre: 16526:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 120/2040/0, delete: 0/0/0 Lustre: 16526:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 1967 previous similar messages Lustre: 16526:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 16526:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 1967 previous similar messages Lustre: 17721:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 6s req@ffff8802754bd180 x1710331746874112/t0(0) o106->LOV_OSC_UUID@0@lo:0/0 lens 328/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'' INFO: task mv:16414 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mv D ffff880284bca140 11344 16414 15893 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff81251fe1>] lock_rename+0x31/0xe0 [<ffffffff8125862f>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed432>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817e8a76>] ? trace_do_page_fault+0x56/0x170 [<ffffffff8125975e>] SyS_renameat2+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task mv:16890 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mv D ffff880294a490c0 11344 16890 16548 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff81251fe1>] lock_rename+0x31/0xe0 [<ffffffff8125862f>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed432>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817e8a76>] ? trace_do_page_fault+0x56/0x170 [<ffffffff8125975e>] SyS_renameat2+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 Lustre: 16553:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 8: before 511 < left 32213, rollback = 9 LustreError: 17308:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000404:0x83:0x0]: rc = -2 LustreError: 20270:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff880298dfae98: inode [0x200000403:0x3:0x0] mdc close failed: rc = -13 Lustre: 16553:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 509 < left 5170, rollback = 7 Lustre: 16553:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 8028 previous similar messages Lustre: 11084:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 502 < left 671, rollback = 2 Lustre: 11084:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 48 previous similar messages Lustre: 18159:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 6s req@ffff880294a33838 x1710331747825600/t0(0) o105->LOV_OSC_UUID@0@lo:0/0 lens 392/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'' Lustre: 17140:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 3/12/10, destroy: 0/0/0 Lustre: 17140:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 20516 previous similar messages Lustre: 17140:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 7/671/0 Lustre: 17140:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 20516 previous similar messages Lustre: 17140:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 5/136/0, punch: 0/0/0, quota 8/56/0 Lustre: 17140:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 20516 previous similar messages Lustre: 17140:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 10/203/4, delete: 0/0/0 Lustre: 17140:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 20516 previous similar messages Lustre: 17140:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 4/4/0, ref_del: 0/0/0 Lustre: 17140:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 20516 previous similar messages LustreError: 10998:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 101s: evicting client at 0@lo ns: filter-lustre-OST0001_UUID lock: ffff880284ddb880/0x4105371653f93cb2 lrc: 3/0,0 mode: PW/PW res: [0x300000400:0x5:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->1048575) gid 0 flags: 0x60000000020020 nid: 0@lo remote: 0x4105371653f93c8f expref: 3389 pid: 17420 timeout: 553 lvb_type: 0 LustreError: 10985:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1631099883 with bad export cookie 4685211556386458677 LustreError: 12673:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88008b881300 x1710331749400576/t0(0) o104->lustre-OST0001@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 12673:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 6 previous similar messages Lustre: lustre-OST0001-osc-ffff880298df53d8: Connection to lustre-OST0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0001-osc-ffff880298df53d8: This client was evicted by lustre-OST0001; in progress operations using this service will fail. LustreError: 10985:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) Skipped 2 previous similar messages Lustre: 4143:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.185@tcp:/lustre/fid: [0x280000403:0xbb:0x0]// may get corrupted (rc -108) LustreError: 22055:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0001-osc-ffff880298df53d8: namespace resource [0x300000401:0x247:0x0].0x0 (ffff880268b44540) refcount nonzero (1) after lock cleanup; forcing cleanup. Lustre: 16897:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 504 < left 21080, rollback = 7 Lustre: 16897:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 21317 previous similar messages Lustre: 17508:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 8: before 512 < left 8346, rollback = 9 Lustre: 17508:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 42 previous similar messages Lustre: lustre-OST0001: haven't heard from client 1154ae0e-996a-42a4-bcbf-d44ba44315ac (at 0@lo) in 47 seconds. I think it's dead, and I am evicting it. exp ffff880264b22e98, cur 1631099930 expire 1631099900 last 1631099883 INFO: task mv:18505 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mv D ffff88028921d2c0 11168 18505 16726 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff81251fe1>] lock_rename+0x31/0xe0 [<ffffffff8125862f>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed432>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817e8a76>] ? trace_do_page_fault+0x56/0x170 [<ffffffff8125975e>] SyS_renameat2+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 LustreError: 22055:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0001-osc-ffff880298df53d8: namespace resource [0x300000400:0x130:0x0].0x0 (ffff88007b3874c0) refcount nonzero (1) after lock cleanup; forcing cleanup. Lustre: lustre-OST0001-osc-ffff880298df53d8: Connection restored to 192.168.123.185@tcp (at 0@lo) LustreError: 11-0: lustre-OST0001-osc-ffff880298df53d8: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-OST0001-osc-ffff880298df53d8: Connection to lustre-OST0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0001-osc-ffff880298df53d8: This client was evicted by lustre-OST0001; in progress operations using this service will fail. LustreError: 11103:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0001: '1' migration was interrupted, run 'lfs migrate -m 0 -c 1 -H crush 1' to finish migration. LustreError: 11238:0:(out_handler.c:910:out_tx_end()) lustre-MDT0000-osd: undo for /home/green/git/lustre-release/lustre/ptlrpc/../../lustre/target/out_handler.c:445: rc = -524 LustreError: 12289:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0000-osp-MDT0002: fail to cancel 1 llog-records: rc = -2 LustreError: 12289:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) Skipped 1 previous similar message LustreError: 12289:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0000-osp-MDT0002: fail to cancel 1 of 1 llog-records: rc = -2 LustreError: 12289:0:(llog_cat.c:787:llog_cat_cancel_records()) Skipped 1 previous similar message Lustre: 16897:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 8: before 510 < left 7377, rollback = 9 Lustre: 16897:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 34 previous similar messages Lustre: 17103:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 378/1512/0, destroy: 0/0/0 Lustre: 17103:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 25573 previous similar messages Lustre: 17103:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/1, xattr_set: 380/15/0 Lustre: 17103:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 25573 previous similar messages Lustre: 17103:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1891/16264/0, punch: 0/0/0, quota 10/106/7 Lustre: 17103:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 25573 previous similar messages Lustre: 17103:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 378/6426/0, delete: 0/0/0 Lustre: 17103:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 25573 previous similar messages Lustre: 17103:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 17103:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 25573 previous similar messages Lustre: 4133:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff8802c5a2de00 x1710331754505344/t0(0) o10->lustre-OST0002-osc-ffff880298dfae98@0@lo:6/4 lens 440/432 e 0 to 0 dl 1631100043 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'cp.0' LustreError: 16861:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0002: failed to get lu_attr of [0x280000404:0x8b:0x0]: rc = -2 Lustre: 18804:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 510 < left 16307, rollback = 7 Lustre: 18804:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 25338 previous similar messages Lustre: dir [0x200000404:0x131:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: dir [0x200000404:0x131:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 10 previous similar messages INFO: task mv:22344 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mv D ffff88028d4d5c40 11344 22344 16062 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff81251fe1>] lock_rename+0x31/0xe0 [<ffffffff8125862f>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed432>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817e8a76>] ? trace_do_page_fault+0x56/0x170 [<ffffffff8125975e>] SyS_renameat2+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task mv:22403 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mv D ffff88008dc7a980 11216 22403 15874 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff81251fe1>] lock_rename+0x31/0xe0 [<ffffffff8125862f>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed432>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817e8a76>] ? trace_do_page_fault+0x56/0x170 [<ffffffff8125975e>] SyS_renameat2+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task mv:22804 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mv D ffff880286c26340 11216 22804 16052 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff81251fe1>] lock_rename+0x31/0xe0 [<ffffffff8125862f>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed432>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817e8a76>] ? trace_do_page_fault+0x56/0x170 [<ffffffff8125975e>] SyS_renameat2+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 LustreError: 11092:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0000: '0' migration was interrupted, run 'lfs migrate -m 2 -c 1 -H crush 0' to finish migration. LustreError: 10998:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0003_UUID lock: ffff880260bfde00/0x410537165407acea lrc: 3/0,0 mode: PW/PW res: [0x380000400:0x2a4:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->917503) gid 0 flags: 0x60000400020020 nid: 0@lo remote: 0x410537165407ac26 expref: 2318 pid: 20724 timeout: 741 lvb_type: 0 LustreError: 28212:0:(ldlm_lockd.c:1425:ldlm_handle_enqueue0()) ### lock on destroyed export ffff88028cc05d28 ns: filter-lustre-OST0003_UUID lock: ffff88007227e580/0x41053716540fbbac lrc: 3/0,0 mode: PR/PR res: [0x380000400:0x45a:0x0].0x0 rrc: 4 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x50000000000000 nid: 0@lo remote: 0x41053716540fab3d expref: 2326 pid: 28212 timeout: 0 lvb_type: 1 LustreError: 10996:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8800803570c0 x1710331757550720/t0(0) o105->lustre-OST0003@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 10996:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 43 previous similar messages LustreError: 11-0: lustre-OST0003-osc-ffff880298dfae98: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 4 previous similar messages Lustre: lustre-OST0003-osc-ffff880298dfae98: Connection to lustre-OST0003 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0003-osc-ffff880298dfae98: This client was evicted by lustre-OST0003; in progress operations using this service will fail. Lustre: 4140:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.185@tcp:/lustre/fid: [0x240000403:0xe4:0x0]// may get corrupted (rc -108) Lustre: 4140:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.185@tcp:/lustre/fid: [0x240000404:0x1f7:0x0]// may get corrupted (rc -108) LustreError: 28288:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0003-osc-ffff880298dfae98: namespace resource [0x380000400:0x430:0x0].0x0 (ffff880277b0ccc0) refcount nonzero (2) after lock cleanup; forcing cleanup. LustreError: 28288:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 306 previous similar messages Lustre: lustre-OST0003-osc-ffff880298dfae98: Connection restored to 192.168.123.185@tcp (at 0@lo) Lustre: Skipped 1 previous similar message Lustre: 16997:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 6s req@ffff88024b928cc0 x1710331758206208/t0(0) o104->LOV_OSC_UUID@0@lo:0/0 lens 328/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'' LustreError: 12289:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0001-osp-MDT0002: fail to cancel 1 llog-records: rc = -2 LustreError: 12289:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0001-osp-MDT0002: fail to cancel 1 of 1 llog-records: rc = -2 Lustre: 21203:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 499 < left 2140, rollback = 2 Lustre: 21203:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 44 previous similar messages LustreError: 10998:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 114s: evicting client at 0@lo ns: filter-lustre-OST0001_UUID lock: ffff8802790a8f40/0x4105371654093a93 lrc: 3/0,0 mode: PW/PW res: [0x300000400:0x361:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->458751) gid 0 flags: 0x60000480000020 nid: 0@lo remote: 0x4105371654093a62 expref: 4143 pid: 12670 timeout: 800 lvb_type: 0 LustreError: 10997:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff880075610680 x1710331759888704/t0(0) o105->lustre-OST0003@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 10997:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 13 previous similar messages LustreError: 19399:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1631100146 with bad export cookie 4685211556386456563 LustreError: 11-0: lustre-OST0001-osc-ffff880298df53d8: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-OST0001-osc-ffff880298df53d8: Connection to lustre-OST0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0001-osc-ffff880298df53d8: This client was evicted by lustre-OST0001; in progress operations using this service will fail. Lustre: 4144:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.185@tcp:/lustre/fid: [0x200000403:0x11c:0x0]/ may get corrupted (rc -5) Lustre: 19571:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 7s req@ffff880085f570c0 x1710331759943296/t0(0) o400->lustre-MDT0001-mdtlov_UUID@0@lo:0/0 lens 224/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'kworker/u32:8.0' Lustre: 19571:0:(service.c:2157:ptlrpc_server_handle_req_in()) Skipped 3 previous similar messages Lustre: 4143:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.185@tcp:/lustre/fid: [0x200000404:0x21b:0x0]// may get corrupted (rc -108) Lustre: 4144:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.185@tcp:/lustre/fid: [0x200000404:0x16e:0x0]// may get corrupted (rc -108) Lustre: 4129:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.185@tcp:/lustre/fid: [0x240000404:0x22e:0x0]// may get corrupted (rc -108) Lustre: 4129:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.185@tcp:/lustre/fid: [0x240000403:0xa1:0x0]// may get corrupted (rc -108) Lustre: 4130:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.185@tcp:/lustre/fid: [0x280000403:0x15e:0x0]// may get corrupted (rc -108) Lustre: 4130:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.185@tcp:/lustre/fid: [0x280000404:0x14b:0x0]/ may get corrupted (rc -108) Lustre: 4131:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff880248be0680 x1710331760077376/t0(0) o10->lustre-OST0001-osc-ffff880298dfae98@0@lo:6/4 lens 440/432 e 0 to 0 dl 1631100237 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dir_create.sh.0' LustreError: 18328:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff880261541f80 x1710331760080064/t0(0) o104->lustre-OST0003@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 18328:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 24 previous similar messages Lustre: 17863:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 7s req@ffff880249a9ac00 x1710331760073088/t0(0) o35->bc1e05b8-368a-4968-928a-1b45107d748e@0@lo:0/0 lens 392/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'ls.0' Lustre: 17863:0:(service.c:2157:ptlrpc_server_handle_req_in()) Skipped 5 previous similar messages LustreError: 29623:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0003-osc-ffff880298df53d8: namespace resource [0x470:0x0:0x0].0x0 (ffff880077794040) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 29623:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 513 previous similar messages Lustre: lustre-OST0003-osc-ffff880298df53d8: Connection restored to 192.168.123.185@tcp (at 0@lo) INFO: task mv:23853 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mv D ffff88026296c9d0 11344 23853 15935 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff81251fe1>] lock_rename+0x31/0xe0 [<ffffffff8125862f>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed432>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817e8a76>] ? trace_do_page_fault+0x56/0x170 [<ffffffff8125975e>] SyS_renameat2+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task mv:24055 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mv D ffff88029412b1c0 11040 24055 15858 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff81251fe1>] lock_rename+0x31/0xe0 [<ffffffff8125862f>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed432>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817e8a76>] ? trace_do_page_fault+0x56/0x170 [<ffffffff8125975e>] SyS_renameat2+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task mv:25295 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mv D ffff88026283c9d0 12128 25295 16464 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff81251fe1>] lock_rename+0x31/0xe0 [<ffffffff8125862f>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed432>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817e8a76>] ? trace_do_page_fault+0x56/0x170 [<ffffffff8125975e>] SyS_renameat2+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task mv:26235 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mv D ffff880289418880 11344 26235 16872 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff81251fe1>] lock_rename+0x31/0xe0 [<ffffffff8125862f>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed432>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817e8a76>] ? trace_do_page_fault+0x56/0x170 [<ffffffff8125975e>] SyS_renameat2+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 Lustre: 4130:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff88024c032c00 x1710331760763520/t0(0) o10->lustre-OST0003-osc-ffff880298df53d8@0@lo:6/4 lens 440/432 e 0 to 0 dl 1631100236 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dir_create.sh.0' Lustre: 19573:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 7s req@ffff8802404789a8 x1710331760432448/t0(0) o106->LOV_OSC_UUID@0@lo:0/0 lens 328/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'' Lustre: dir [0x280000404:0x8f:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 1 previous similar message Lustre: dir [0x280000404:0x113:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: 18267:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 51/204/0, destroy: 1/4/0 Lustre: 18267:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 76147 previous similar messages Lustre: 18267:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 3/3/1, xattr_set: 55/148/0 Lustre: 18267:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 76147 previous similar messages Lustre: 18267:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 256/2203/0, punch: 0/0/0, quota 8/56/0 Lustre: 18267:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 76147 previous similar messages Lustre: 18267:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 52/883/0, delete: 2/5/1 Lustre: 18267:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 76146 previous similar messages Lustre: 18267:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 1/1/0, ref_del: 2/2/1 Lustre: 18267:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 76147 previous similar messages Lustre: 4146:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff880265aeb240 x1710331763673280/t0(0) o10->lustre-OST0002-osc-ffff880298dfae98@0@lo:6/4 lens 440/432 e 0 to 0 dl 1631100345 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'truncate.0' LustreError: 10998:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0003_UUID lock: ffff88027d2a07c0/0x410537165414c239 lrc: 3/0,0 mode: PW/PW res: [0x380000400:0x2f6:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000480010020 nid: 0@lo remote: 0x410537165414c232 expref: 2978 pid: 19295 timeout: 941 lvb_type: 0 LustreError: 10998:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 1 previous similar message LustreError: 18497:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88024cdbf700 x1710331763697920/t0(0) o105->lustre-OST0003@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 18497:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 2 previous similar messages LustreError: 11-0: lustre-OST0003-osc-ffff880298dfae98: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 35 previous similar messages Lustre: lustre-OST0003-osc-ffff880298dfae98: Connection to lustre-OST0003 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 1 previous similar message LustreError: 167-0: lustre-OST0003-osc-ffff880298dfae98: This client was evicted by lustre-OST0003; in progress operations using this service will fail. LustreError: Skipped 1 previous similar message Lustre: 17204:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 6s req@ffff8800b2272ee8 x1710331763702784/t0(0) o35->bc1e05b8-368a-4968-928a-1b45107d748e@0@lo:0/0 lens 392/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'cp.0' LustreError: 31371:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0003-osc-ffff880298dfae98: namespace resource [0x380000400:0xc00:0x0].0x0 (ffff88026423de40) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 31371:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 379 previous similar messages Lustre: lustre-OST0003-osc-ffff880298dfae98: Connection restored to 192.168.123.185@tcp (at 0@lo) Lustre: Skipped 1 previous similar message Lustre: 16553:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 509 < left 63263, rollback = 7 Lustre: 16553:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 77134 previous similar messages Lustre: 4146:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff880265aeb240 x1710331765476160/t0(0) o10->lustre-OST0002-osc-ffff880298dfae98@0@lo:6/4 lens 440/432 e 0 to 0 dl 1631100399 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'truncate.0' LustreError: 23904:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0001: failed to get lu_attr of [0x240000403:0x1c4:0x0]: rc = -2 dd (1023) used greatest stack depth: 10080 bytes left Lustre: 4138:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff88024c7970c0 x1710331769150080/t0(0) o10->lustre-OST0001-osc-ffff880298dfae98@0@lo:6/4 lens 440/432 e 0 to 0 dl 1631100481 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dir_create.sh.0' LustreError: 10998:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 135s: evicting client at 0@lo ns: filter-lustre-OST0003_UUID lock: ffff88007ba5ed00/0x410537165418ca18 lrc: 3/0,0 mode: PW/PW res: [0x425:0x0:0x0].0x0 rrc: 4 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000000020020 nid: 0@lo remote: 0x410537165418c9fc expref: 2888 pid: 23239 timeout: 1051 lvb_type: 0 LustreError: 10997:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88026f80b880 x1710331769256384/t0(0) o105->lustre-OST0003@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 10997:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 28 previous similar messages LustreError: 11-0: lustre-OST0002-osc-ffff880298df53d8: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-OST0002-osc-ffff880298df53d8: Connection to lustre-OST0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0002-osc-ffff880298df53d8: This client was evicted by lustre-OST0002; in progress operations using this service will fail. LustreError: 10991:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1631100417 with bad export cookie 4685211556388297248 LustreError: 10991:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) Skipped 1 previous similar message Lustre: 4138:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.185@tcp:/lustre/fid: [0x280000403:0x310:0x0]/ may get corrupted (rc -108) Lustre: 4138:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.185@tcp:/lustre/fid: [0x200000403:0x1fc:0x0]/ may get corrupted (rc -108) LustreError: 5:0:(ofd_obd.c:501:ofd_destroy_export()) lustre-OST0002: cli 1154ae0e-996a-42a4-bcbf-d44ba44315ac/ffff88028da86678 has 61440 pending on destroyed export LustreError: 5:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_destroy_export: tot_granted 168046016 != fo_tot_granted 168107456 LustreError: 5:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_destroy_export: tot_pending 90112 != fo_tot_pending 151552 LustreError: 12688:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 168046016 != fo_tot_granted 168107456 LustreError: 12688:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 90112 != fo_tot_pending 151552 Lustre: 4131:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.185@tcp:/lustre/fid: [0x200000403:0x1f3:0x0]// may get corrupted (rc -108) Lustre: 4132:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.185@tcp:/lustre/fid: [0x200000403:0x1da:0x0]// may get corrupted (rc -108) Lustre: 4132:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.185@tcp:/lustre/fid: [0x280000404:0x258:0x0]// may get corrupted (rc -108) LustreError: 12691:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 168017344 != fo_tot_granted 168078784 LustreError: 12691:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 61440 != fo_tot_pending 122880 Lustre: 26609:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 501 < left 1027, rollback = 2 Lustre: 26609:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 177 previous similar messages LustreError: 12689:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 168017344 != fo_tot_granted 168078784 LustreError: 12689:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 1 previous similar message LustreError: 12689:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 61440 != fo_tot_pending 122880 LustreError: 12689:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 1 previous similar message Lustre: dir [0x200000404:0x174:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 9 previous similar messages LustreError: 12691:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 176455104 != fo_tot_granted 176516544 LustreError: 12691:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 1 previous similar message LustreError: 12691:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 61440 != fo_tot_pending 122880 LustreError: 12691:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 1 previous similar message LustreError: 1721:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0003-osc-ffff880298df53d8: namespace resource [0x430:0x0:0x0].0x0 (ffff88026552e5c0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 1721:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 439 previous similar messages Lustre: lustre-OST0003-osc-ffff880298df53d8: Connection restored to 192.168.123.185@tcp (at 0@lo) LustreError: 12689:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 176453440 != fo_tot_granted 176514880 LustreError: 12689:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 3 previous similar messages LustreError: 12689:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 61440 != fo_tot_pending 122880 LustreError: 12689:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 3 previous similar messages Lustre: 4146:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff880265aeb240 x1710331770190720/t0(0) o10->lustre-OST0002-osc-ffff880298dfae98@0@lo:6/4 lens 440/432 e 0 to 0 dl 1631100529 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'truncate.0' Lustre: 4146:0:(client.c:1485:after_reply()) Skipped 2 previous similar messages LustreError: 12682:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 192977216 != fo_tot_granted 193038656 LustreError: 12682:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 6 previous similar messages LustreError: 12682:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 61440 != fo_tot_pending 122880 LustreError: 12682:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 6 previous similar messages LustreError: 12682:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 225691968 != fo_tot_granted 225753408 LustreError: 12682:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 9 previous similar messages LustreError: 12682:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 61440 != fo_tot_pending 122880 LustreError: 12682:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 9 previous similar messages LustreError: 19055:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0000: '9' migration was interrupted, run 'lfs migrate -m 1 -c 1 -H crush 9' to finish migration. Lustre: ll_ost02_010: service thread pid 19978 was inactive for 40.050 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 22083, comm: ll_ost02_013 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] ofd_destroy_by_fid+0x1d1/0x520 [ofd] [<0>] ofd_destroy_hdl+0x27f/0xa20 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 22097, comm: ll_ost02_017 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] ofd_destroy_by_fid+0x1d1/0x520 [ofd] [<0>] ofd_destroy_hdl+0x27f/0xa20 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: Skipped 2 previous similar messages Pid: 19978, comm: ll_ost02_010 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] ofd_destroy_by_fid+0x1d1/0x520 [ofd] [<0>] ofd_destroy_hdl+0x27f/0xa20 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: ll_ost02_004: service thread pid 17944 was inactive for 40.596 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: ll_ost04_013: service thread pid 19310 was inactive for 42.033 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 4 previous similar messages LustreError: 12679:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 267418176 != fo_tot_granted 267479616 LustreError: 12679:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 26 previous similar messages LustreError: 12679:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 61440 != fo_tot_pending 122880 LustreError: 12679:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 26 previous similar messages Lustre: 4791:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (22/1s); client may timeout req@ffff880248caa5c0 x1710331774081472/t0(0) o101->1154ae0e-996a-42a4-bcbf-d44ba44315ac@0@lo:92/0 lens 328/400 e 0 to 0 dl 1631100582 ref 2 fl Complete:/0/0 rc 301/301 job:'ln.0' Lustre: 4791:0:(service.c:2328:ptlrpc_server_handle_request()) Skipped 18 previous similar messages LustreError: 4791:0:(service.c:2290:ptlrpc_server_handle_request()) @@@ Dropping timed-out request from 12345-0@lo: deadline 22/1s ago req@ffff88028ab53ec0 x1710331774083776/t0(0) o101->1154ae0e-996a-42a4-bcbf-d44ba44315ac@0@lo:92/0 lens 328/0 e 0 to 0 dl 1631100582 ref 2 fl Interpret:/0/ffffffff rc 0/-1 job:'ln.0' Lustre: 21987:0:(service.c:1378:ptlrpc_at_send_early_reply()) @@@ Already past deadline (-1s), not sending early reply. Consider increasing at_early_margin (5)? req@ffff88027e2af0c0 x1710331774073728/t0(0) o1->1154ae0e-996a-42a4-bcbf-d44ba44315ac@0@lo:92/0 lens 440/0 e 0 to 0 dl 1631100582 ref 2 fl New:/0/ffffffff rc 0/-1 job:'lfs.0' Lustre: 21987:0:(service.c:1378:ptlrpc_at_send_early_reply()) Skipped 17 previous similar messages Lustre: 17834:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (22/1s); client may timeout req@ffff88007c3f6440 x1710331774073216/t0(0) o1->1154ae0e-996a-42a4-bcbf-d44ba44315ac@0@lo:92/0 lens 440/432 e 0 to 0 dl 1631100582 ref 1 fl Complete:/0/0 rc 0/0 job:'lfs.0' Lustre: 17834:0:(service.c:2328:ptlrpc_server_handle_request()) Skipped 9 previous similar messages Lustre: 4742:0:(service.c:1378:ptlrpc_at_send_early_reply()) @@@ Already past deadline (-1s), not sending early reply. Consider increasing at_early_margin (5)? req@ffff880279ca6440 x1710331774072576/t0(0) o1->1154ae0e-996a-42a4-bcbf-d44ba44315ac@0@lo:93/0 lens 440/432 e 0 to 0 dl 1631100583 ref 2 fl Interpret:/0/0 rc 0/0 job:'lfs.0' Lustre: 4742:0:(service.c:1378:ptlrpc_at_send_early_reply()) Skipped 58 previous similar messages Lustre: 22067:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (22/3s); client may timeout req@ffff88007236ac00 x1710331774071936/t0(0) o1->1154ae0e-996a-42a4-bcbf-d44ba44315ac@0@lo:92/0 lens 440/432 e 0 to 0 dl 1631100582 ref 1 fl Complete:/0/0 rc 0/0 job:'lfs.0' Lustre: 22067:0:(service.c:2328:ptlrpc_server_handle_request()) Skipped 7 previous similar messages Lustre: 19248:0:(service.c:1378:ptlrpc_at_send_early_reply()) @@@ Already past deadline (-1s), not sending early reply. Consider increasing at_early_margin (5)? req@ffff88029954bec0 x1710331774338432/t0(0) o101->1154ae0e-996a-42a4-bcbf-d44ba44315ac@0@lo:95/0 lens 328/0 e 0 to 0 dl 1631100585 ref 2 fl New:/0/ffffffff rc 0/-1 job:'ln.0' Lustre: 19248:0:(service.c:1378:ptlrpc_at_send_early_reply()) Skipped 3 previous similar messages Lustre: ll_ost01_024: service thread pid 30499 was inactive for 64.091 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 7 previous similar messages LustreError: 21741:0:(service.c:2290:ptlrpc_server_handle_request()) @@@ Dropping timed-out request from 12345-0@lo: deadline 23/4s ago req@ffff880263b30680 x1710331774323840/t0(0) o101->1154ae0e-996a-42a4-bcbf-d44ba44315ac@0@lo:95/0 lens 328/0 e 0 to 0 dl 1631100585 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'mkdir.0' Lustre: 4948:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (23/4s); client may timeout req@ffff880241deb240 x1710331774326144/t0(0) o101->1154ae0e-996a-42a4-bcbf-d44ba44315ac@0@lo:95/0 lens 328/0 e 0 to 0 dl 1631100585 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'mkdir.0' Lustre: 4948:0:(service.c:2328:ptlrpc_server_handle_request()) Skipped 4 previous similar messages LustreError: 21741:0:(service.c:2290:ptlrpc_server_handle_request()) Skipped 25 previous similar messages Lustre: 11092:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000404:0x335:0x0] with magic=0xbd60bd0 Lustre: 18416:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000404:0x353:0x0] with magic=0xbd60bd0 Lustre: 18416:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 3 previous similar messages Lustre: ll_ost05_023: service thread pid 18331 was inactive for 90.111 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 12 previous similar messages LustreError: 10998:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0003_UUID lock: ffff8800768aed00/0x41053716541f92ab lrc: 3/0,0 mode: PW/PW res: [0x42a:0x0:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000400010020 nid: 0@lo remote: 0x41053716541f92a4 expref: 4841 pid: 12669 timeout: 1294 lvb_type: 0 LustreError: 10998:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 1 previous similar message LustreError: 11-0: lustre-OST0003-osc-ffff880298dfae98: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: 21159:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1631100622 with bad export cookie 4685211556388583989 LustreError: 21159:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) Skipped 1 previous similar message LustreError: 167-0: lustre-OST0003-osc-ffff880298dfae98: This client was evicted by lustre-OST0003; in progress operations using this service will fail. LustreError: Skipped 1 previous similar message LustreError: 21473:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802739ab880 x1710331776859264/t0(0) o104->lustre-OST0003@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 21473:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 3 previous similar messages Lustre: 4135:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1631100562/real 1631100562] req@ffff880266367700 x1710331774323840/t0(0) o101->lustre-OST0001-osc-ffff880298df53d8@0@lo:28/4 lens 328/400 e 0 to 1 dl 1631100610 ref 1 fl Rpc:XQr/0/ffffffff rc 0/-1 job:'mkdir.0' Lustre: lustre-OST0001: Client 1154ae0e-996a-42a4-bcbf-d44ba44315ac (at 0@lo) reconnecting Lustre: lustre-OST0001-osc-ffff880298df53d8: Connection restored to 192.168.123.185@tcp (at 0@lo) Lustre: Skipped 1 previous similar message LustreError: 1540:0:(ofd_obd.c:501:ofd_destroy_export()) lustre-OST0003: cli bc1e05b8-368a-4968-928a-1b45107d748e/ffff8802c4c40008 has 61440 pending on destroyed export Lustre: mdt: This server is not able to keep up with request traffic (cpu-bound). Lustre: 18416:0:(service.c:1612:ptlrpc_at_check_timed()) earlyQ=1 reqQ=0 recA=4, svcEst=58, delay=5452ms Lustre: 18416:0:(service.c:1378:ptlrpc_at_send_early_reply()) @@@ Already past deadline (-1s), not sending early reply. Consider increasing at_early_margin (5)? req@ffff88027832de00 x1710331773971904/t0(0) o36->bc1e05b8-368a-4968-928a-1b45107d748e@0@lo:144/0 lens 688/2888 e 0 to 0 dl 1631100634 ref 2 fl Interpret:/0/0 rc 0/0 job:'lfs.0' Lustre: 18416:0:(service.c:1378:ptlrpc_at_send_early_reply()) Skipped 18 previous similar messages LustreError: 10998:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0002_UUID lock: ffff8802420234c0/0x410537165425671e lrc: 3/0,0 mode: PR/PR res: [0x280000403:0x117:0x0].0x0 bits 0x13/0x0 rrc: 3 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0x41053716542566ca expref: 179 pid: 23972 timeout: 1328 lvb_type: 0 LustreError: 10998:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 3 previous similar messages LustreError: 16626:0:(ldlm_lockd.c:1425:ldlm_handle_enqueue0()) ### lock on destroyed export ffff88008fa412a8 ns: mdt-lustre-MDT0002_UUID lock: ffff8802499bc3c0/0x4105371654263566 lrc: 3/0,0 mode: PR/PR res: [0x280000403:0x1:0x0].0x0 bits 0x12/0x0 rrc: 12 type: IBT gid 0 flags: 0x50200400000020 nid: 0@lo remote: 0x410537165425b361 expref: 24 pid: 16626 timeout: 0 lvb_type: 0 LustreError: 11-0: lustre-MDT0002-mdc-ffff880298dfae98: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 6 previous similar messages LustreError: 1771:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff880298dfae98: inode [0x280000404:0x25a:0x0] mdc close failed: rc = -108 LustreError: 4753:0:(file.c:5042:ll_inode_revalidate_fini()) lustre: revalidate FID [0x280000403:0x1:0x0] error: rc = -5 LustreError: 32123:0:(mdc_request.c:1436:mdc_read_page()) lustre-MDT0002-mdc-ffff880298dfae98: [0x280000403:0xc5:0x0] lock enqueue fails: rc = -108 Lustre: dir [0x240000403:0x2cf:0x0] stripe 1 readdir failed: -108, directory is partially accessed! Lustre: Skipped 2 previous similar messages LustreError: 3730:0:(vvp_io.c:1793:vvp_io_init()) lustre: refresh file layout [0x280000404:0x391:0x0] error -108. LustreError: 3730:0:(vvp_io.c:1793:vvp_io_init()) Skipped 1 previous similar message LustreError: 4134:0:(mdc_locks.c:1386:mdc_intent_getattr_async_interpret()) lustre-MDT0002-mdc-ffff880298dfae98: ldlm_cli_enqueue_fini() failed: rc = -108 LustreError: 5386:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff880298dfae98: inode [0x280000404:0x14a:0x0] mdc close failed: rc = -108 LustreError: 5386:0:(file.c:234:ll_close_inode_openhandle()) Skipped 10 previous similar messages LustreError: 4498:0:(file.c:5042:ll_inode_revalidate_fini()) lustre: revalidate FID [0x280000403:0x389:0x0] error: rc = -108 LustreError: 4498:0:(file.c:5042:ll_inode_revalidate_fini()) Skipped 33 previous similar messages Lustre: 31384:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1631100557/real 1631100557] req@ffff880278329f80 x1710331773971904/t0(0) o36->lustre-MDT0001-mdc-ffff880298dfae98@0@lo:12/10 lens 688/4592 e 0 to 1 dl 1631100661 ref 2 fl Rpc:XQr/0/ffffffff rc 0/-1 job:'lfs.0' Lustre: 31384:0:(client.c:2285:ptlrpc_expire_one_request()) Skipped 3 previous similar messages Lustre: lustre-MDT0001: Client bc1e05b8-368a-4968-928a-1b45107d748e (at 0@lo) reconnecting Lustre: Skipped 3 previous similar messages LustreError: 12679:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 264363328 != fo_tot_granted 264424768 LustreError: 12679:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 107 previous similar messages LustreError: 12679:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 61440 != fo_tot_pending 122880 LustreError: 12679:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 107 previous similar messages LustreError: 21407:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0001: '5' migration was interrupted, run 'lfs migrate -m 0 -c 1 -H crush 5' to finish migration. LustreError: 6434:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff880298dfae98: inode [0x200000404:0x349:0x0] mdc close failed: rc = -13 LustreError: 6434:0:(file.c:234:ll_close_inode_openhandle()) Skipped 8 previous similar messages Lustre: 4139:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff88024334de00 x1710331780871040/t0(0) o10->lustre-OST0003-osc-ffff880298dfae98@0@lo:6/4 lens 440/432 e 0 to 0 dl 1631100774 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'truncate.0' Lustre: 4139:0:(client.c:1485:after_reply()) Skipped 2 previous similar messages Lustre: 28277:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0x25e:0x0] with magic=0xbd60bd0 Lustre: 28277:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 1 previous similar message Lustre: 11100:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0x24b:0x0] with magic=0xbd60bd0 Lustre: 11100:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 1 previous similar message LustreError: 7019:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff880298df53d8: inode [0x200000404:0x355:0x0] mdc close failed: rc = -13 Lustre: 23972:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0x284:0x0] with magic=0xbd60bd0 Lustre: 23972:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 1 previous similar message Lustre: 28263:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 3/12/5, destroy: 0/0/0 Lustre: 28263:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 132789 previous similar messages Lustre: 28263:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 16/1472/0 Lustre: 28263:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 132789 previous similar messages Lustre: 28263:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 5/136/0, punch: 0/0/0, quota 8/56/0 Lustre: 28263:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 132788 previous similar messages Lustre: 28263:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 10/203/3, delete: 0/0/0 Lustre: 28263:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 132790 previous similar messages Lustre: 28263:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 4/4/0, ref_del: 0/0/0 Lustre: 28263:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 132789 previous similar messages LustreError: 11123:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000404:0xd4:0x0]: rc = -2 Lustre: 23904:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 485 < left 68355, rollback = 7 Lustre: 23904:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 137446 previous similar messages LustreError: 10998:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 149s: evicting client at 0@lo ns: filter-lustre-OST0000_UUID lock: ffff88024468e1c0/0x41053716542c1aa3 lrc: 3/0,0 mode: PW/PW res: [0x2c0000401:0x7bd:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000400020020 nid: 0@lo remote: 0x41053716542c19d8 expref: 1931 pid: 2936 timeout: 1455 lvb_type: 0 LustreError: 25828:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1631100833 with bad export cookie 4685211556387025348 LustreError: 25828:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) Skipped 5 previous similar messages LustreError: 11-0: lustre-OST0001-osc-ffff880298dfae98: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-OST0001-osc-ffff880298dfae98: Connection to lustre-OST0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 9 previous similar messages LustreError: 167-0: lustre-OST0001-osc-ffff880298dfae98: This client was evicted by lustre-OST0001; in progress operations using this service will fail. LustreError: Skipped 2 previous similar messages LustreError: 9214:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0000-osc-ffff880298dfae98: namespace resource [0x7a0:0x0:0x0].0x0 (ffff880242858a40) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 9214:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 117 previous similar messages Lustre: 4133:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.185@tcp:/lustre/fid: [0x200000404:0x416:0x0]/ may get corrupted (rc -108) LustreError: 5:0:(ofd_obd.c:501:ofd_destroy_export()) lustre-OST0001: cli bc1e05b8-368a-4968-928a-1b45107d748e/ffff88027a65a548 has 28672 pending on destroyed export LustreError: 12530:0:(ofd_obd.c:501:ofd_destroy_export()) lustre-OST0002: cli bc1e05b8-368a-4968-928a-1b45107d748e/ffff88028f60a548 has 61440 pending on destroyed export LustreError: 8289:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff880298dfae98: inode [0x200000403:0x2c6:0x0] mdc close failed: rc = -108 LustreError: 7504:0:(file.c:5042:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000403:0x1:0x0] error: rc = -108 LustreError: 7504:0:(file.c:5042:ll_inode_revalidate_fini()) Skipped 1 previous similar message Lustre: mdt04_002: service thread pid 11092 was inactive for 96.175 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 2 previous similar messages LustreError: 9625:0:(file.c:5042:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000403:0x1:0x0] error: rc = -108 LustreError: 9625:0:(file.c:5042:ll_inode_revalidate_fini()) Skipped 7 previous similar messages LustreError: 9758:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -108 LustreError: 16859:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1631100863 with bad export cookie 4685211556389030134 LustreError: 16859:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) Skipped 10 previous similar messages LustreError: 9892:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -108 LustreError: 9945:0:(file.c:5042:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000403:0x1:0x0] error: rc = -108 LustreError: 9945:0:(file.c:5042:ll_inode_revalidate_fini()) Skipped 38 previous similar messages LustreError: 8547:0:(llite_lib.c:2983:ll_prep_inode()) new_inode -fatal: rc -108 LustreError: 10437:0:(file.c:5042:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000403:0x1:0x0] error: rc = -108 LustreError: 10437:0:(file.c:5042:ll_inode_revalidate_fini()) Skipped 115 previous similar messages LustreError: 9605:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff880298dfae98: inode [0x200000404:0x2fe:0x0] mdc close failed: rc = -108 LustreError: 9605:0:(file.c:234:ll_close_inode_openhandle()) Skipped 1 previous similar message LustreError: 154:0:(ofd_obd.c:501:ofd_destroy_export()) lustre-OST0003: cli 1154ae0e-996a-42a4-bcbf-d44ba44315ac/ffff8802649bc138 has 3649536 pending on destroyed export Lustre: 4145:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.185@tcp:/lustre/fid: [0x280000403:0x50c:0x0]/ may get corrupted (rc -108) Lustre: dir [0x280000405:0x89:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 4 previous similar messages Lustre: 12680:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 6s req@ffff88028a60bec0 x1710331787360768/t0(0) o13->lustre-MDT0002-mdtlov_UUID@0@lo:0/0 lens 224/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'osp-pre-3-2.0' Lustre: 12680:0:(service.c:2157:ptlrpc_server_handle_req_in()) Skipped 1 previous similar message LustreError: 11909:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0002-osp-MDT0001: fail to cancel 1 llog-records: rc = -2 LustreError: 11909:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) Skipped 1 previous similar message LustreError: 11909:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0002-osp-MDT0001: fail to cancel 1 of 1 llog-records: rc = -2 LustreError: 11909:0:(llog_cat.c:787:llog_cat_cancel_records()) Skipped 1 previous similar message Lustre: lustre-OST0003-osc-ffff880298df53d8: Connection restored to 192.168.123.185@tcp (at 0@lo) Lustre: Skipped 12 previous similar messages LustreError: 31503:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 69142208 != fo_tot_granted 69265088 LustreError: 31503:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 259 previous similar messages LustreError: 31503:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 1792 != fo_tot_pending 124672 LustreError: 31503:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 259 previous similar messages LustreError: 10998:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0002_UUID lock: ffff880278a67840/0x410537165434fb8f lrc: 3/0,0 mode: PW/PW res: [0x340000400:0x1c09:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->134217727) gid 0 flags: 0x60000400020020 nid: 0@lo remote: 0x410537165434fb88 expref: 2335 pid: 30616 timeout: 1632 lvb_type: 0 LustreError: 10998:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 12 previous similar messages LustreError: 11-0: lustre-OST0002-osc-ffff880298dfae98: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 2 previous similar messages LustreError: 28293:0:(ldlm_lockd.c:1342:ldlm_handle_enqueue0()) ### lock on disconnected export ffff8802459d8008 ns: filter-lustre-OST0002_UUID lock: ffff88027f852d40/0x41053716543ae40d lrc: 2/0,0 mode: --/PR res: [0x340000400:0x2311:0x0].0x0 rrc: 3 type: EXT [0->0] (req 0->0) gid 0 flags: 0x40000000000000 nid: local remote: 0x41053716543a327b expref: -99 pid: 28293 timeout: 0 lvb_type: 0 LustreError: 167-0: lustre-OST0002-osc-ffff880298dfae98: This client was evicted by lustre-OST0002; in progress operations using this service will fail. LustreError: Skipped 5 previous similar messages LustreError: 30398:0:(ldlm_lockd.c:1342:ldlm_handle_enqueue0()) ### lock on disconnected export ffff8802459d8008 ns: filter-lustre-OST0002_UUID lock: ffff8802624eed00/0x41053716543b2456 lrc: 2/0,0 mode: --/PR res: [0x340000400:0x2312:0x0].0x0 rrc: 2 type: EXT [0->0] (req 0->0) gid 0 flags: 0x40000000000000 nid: local remote: 0x41053716543a3297 expref: -99 pid: 30398 timeout: 0 lvb_type: 0 LustreError: 30398:0:(ldlm_lockd.c:1342:ldlm_handle_enqueue0()) Skipped 2 previous similar messages LustreError: 12670:0:(ldlm_lockd.c:1342:ldlm_handle_enqueue0()) ### lock on disconnected export ffff8802459d8008 ns: filter-lustre-OST0002_UUID lock: ffff88025d3cb4c0/0x41053716543b38d8 lrc: 2/0,0 mode: --/PR res: [0x340000400:0x1222:0x0].0x0 rrc: 3 type: EXT [0->0] (req 0->0) gid 0 flags: 0x40000000000000 nid: local remote: 0x41053716543a4e2e expref: -99 pid: 12670 timeout: 0 lvb_type: 0 LustreError: 12670:0:(ldlm_lockd.c:1342:ldlm_handle_enqueue0()) Skipped 56 previous similar messages Lustre: 4138:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.185@tcp:/lustre/fid: [0x240000404:0x3b3:0x0]/ may get corrupted (rc -108) Lustre: 11094:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 506 < left 983, rollback = 2 Lustre: 11094:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 965 previous similar messages Lustre: 4138:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff88027a61b240 x1710331791027008/t0(0) o10->lustre-OST0001-osc-ffff880298dfae98@0@lo:6/4 lens 440/432 e 0 to 0 dl 1631101057 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dir_create.sh.0' LustreError: 26609:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0002: '17' migration was interrupted, run 'lfs migrate -m 1 -c 1 -H crush 17' to finish migration. LustreError: 11684:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0002-osc-ffff880298dfae98: namespace resource [0x340000400:0x2280:0x0].0x0 (ffff880267026d40) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 11684:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 241 previous similar messages LustreError: 11654:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff880261e10cc0 x1710331793944064/t0(0) o104->lustre-OST0003@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 11654:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 231 previous similar messages LustreError: 16662:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1631101082 with bad export cookie 4685211556390509220 Lustre: 19511:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 6s req@ffff8802768ff018 x1710331793925376/t0(0) o104->LOV_OSC_UUID@0@lo:0/0 lens 328/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'' Lustre: 19511:0:(service.c:2157:ptlrpc_server_handle_req_in()) Skipped 1 previous similar message Lustre: 4144:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.185@tcp:/lustre/fid: [0x240000403:0x47a:0x0]// may get corrupted (rc -108) Lustre: 18424:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (51/1s); client may timeout req@ffff88026eafbec0 x1710331793409984/t0(0) o1->bc1e05b8-368a-4968-928a-1b45107d748e@0@lo:641/0 lens 440/432 e 1 to 0 dl 1631101131 ref 2 fl Complete:/0/0 rc 0/0 job:'lfs.0' Lustre: 18424:0:(service.c:2328:ptlrpc_server_handle_request()) Skipped 18 previous similar messages Lustre: 4903:0:(service.c:1378:ptlrpc_at_send_early_reply()) @@@ Already past deadline (-3s), not sending early reply. Consider increasing at_early_margin (5)? req@ffff8800750057c0 x1710331793409536/t0(0) o1->bc1e05b8-368a-4968-928a-1b45107d748e@0@lo:641/0 lens 440/432 e 1 to 0 dl 1631101131 ref 1 fl Complete:/0/0 rc 0/0 job:'lfs.0' Lustre: lustre-OST0003: haven't heard from client 1154ae0e-996a-42a4-bcbf-d44ba44315ac (at 0@lo) in 54 seconds. I think it's dead, and I am evicting it. exp ffff880077584a88, cur 1631101137 expire 1631101107 last 1631101083 LustreError: 30510:0:(ldlm_lockd.c:1425:ldlm_handle_enqueue0()) ### lock on destroyed export ffff8802f1ecc138 ns: filter-lustre-OST0001_UUID lock: ffff88027937c000/0x4105371654447cb2 lrc: 3/0,0 mode: PW/PW res: [0x300000401:0x1338:0x0].0x0 rrc: 2 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x50000000020000 nid: 0@lo remote: 0x4105371654447c3b expref: 4394 pid: 30510 timeout: 0 lvb_type: 0 LustreError: 20090:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1631101149 with bad export cookie 4685211556390399481 LustreError: 20090:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) Skipped 3 previous similar messages LustreError: 30510:0:(ldlm_lockd.c:1425:ldlm_handle_enqueue0()) Skipped 7 previous similar messages Lustre: 4132:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.185@tcp:/lustre/fid: [0x200000403:0x337:0x0]/ may get corrupted (rc -108) Lustre: 4131:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.185@tcp:/lustre/fid: [0x240000403:0x49b:0x0]/ may get corrupted (rc -108) LustreError: 20234:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0000: '2' migration was interrupted, run 'lfs migrate -m 1 -c 1 -H crush 2' to finish migration. LustreError: 18380:0:(lustre_lmv.h:438:lmv_is_sane()) insane LMV: magic=0xcd40cd0 count=2 index=1 hash=crush:0x82000003 version=0 migrate offset=1 migrate hash=crush:3. Lustre: 4143:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff88024ddd4500 x1710331800174208/t0(0) o10->lustre-OST0001-osc-ffff880298dfae98@0@lo:6/4 lens 440/432 e 0 to 0 dl 1631101341 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'cp.0' Lustre: 4143:0:(client.c:1485:after_reply()) Skipped 5 previous similar messages LustreError: 28215:0:(lustre_lmv.h:438:lmv_is_sane()) insane LMV: magic=0xcd40cd0 count=2 index=1 hash=crush:0x82000003 version=0 migrate offset=1 migrate hash=crush:3. LustreError: 17446:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0000: '0' migration was interrupted, run 'lfs migrate -m 2 -c 1 -H crush 0' to finish migration. LustreError: 10998:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0002_UUID lock: ffff8802643a6940/0x410537165447ebb0 lrc: 3/0,0 mode: PW/PW res: [0x804:0x0:0x0].0x0 rrc: 4 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000000030020 nid: 0@lo remote: 0x410537165447eb9b expref: 4238 pid: 4796 timeout: 1982 lvb_type: 0 LustreError: 10998:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 4 previous similar messages LustreError: 11-0: lustre-OST0002-osc-ffff880298dfae98: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 25 previous similar messages LustreError: 167-0: lustre-OST0002-osc-ffff880298dfae98: This client was evicted by lustre-OST0002; in progress operations using this service will fail. LustreError: Skipped 5 previous similar messages Lustre: 4129:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.185@tcp:/lustre/fid: [0x200000405:0xa0:0x0]/ may get corrupted (rc -108) Lustre: 4130:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.185@tcp:/lustre/fid: [0x280000405:0x1b8:0x0]// may get corrupted (rc -108) LustreError: 28215:0:(lustre_lmv.h:438:lmv_is_sane()) insane LMV: magic=0xcd40cd0 count=2 index=1 hash=crush:0x82000003 version=0 migrate offset=1 migrate hash=crush:3. LustreError: 15803:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0002-osc-ffff880298dfae98: namespace resource [0x810:0x0:0x0].0x0 (ffff880266ab96c0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 15803:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 1914 previous similar messages LustreError: 16030:0:(lov_object.c:1299:lov_layout_change()) lustre-clilov-ffff880298df53d8: cannot apply new layout on [0x200000403:0x398:0x0] : rc = -22 LustreError: 16030:0:(lcommon_cl.c:196:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000403:0x398:0x0]: rc = -22 LustreError: 16346:0:(lov_object.c:1299:lov_layout_change()) lustre-clilov-ffff880298df53d8: cannot apply new layout on [0x200000403:0x398:0x0] : rc = -22 Lustre: lustre-OST0003-osc-ffff880298dfae98: Connection to lustre-OST0003 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 12 previous similar messages LustreError: 16346:0:(lcommon_cl.c:196:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000403:0x398:0x0]: rc = -22 LustreError: 16346:0:(llite_lib.c:2983:ll_prep_inode()) new_inode -fatal: rc -22 LustreError: 16406:0:(lov_object.c:1299:lov_layout_change()) lustre-clilov-ffff880298df53d8: cannot apply new layout on [0x200000403:0x398:0x0] : rc = -22 LustreError: 16406:0:(lcommon_cl.c:196:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000403:0x398:0x0]: rc = -22 LustreError: 16406:0:(llite_lib.c:2983:ll_prep_inode()) new_inode -fatal: rc -22 Lustre: lustre-OST0002: haven't heard from client bc1e05b8-368a-4968-928a-1b45107d748e (at 0@lo) in 49 seconds. I think it's dead, and I am evicting it. exp ffff88026725efc8, cur 1631101362 expire 1631101332 last 1631101313 LustreError: 16737:0:(lov_object.c:1299:lov_layout_change()) lustre-clilov-ffff880298df53d8: cannot apply new layout on [0x200000403:0x398:0x0] : rc = -22 LustreError: 16737:0:(lcommon_cl.c:196:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000403:0x398:0x0]: rc = -22 LustreError: 16737:0:(llite_lib.c:2983:ll_prep_inode()) new_inode -fatal: rc -22 LustreError: 28195:0:(lustre_lmv.h:438:lmv_is_sane()) insane LMV: magic=0xcd40cd0 count=2 index=1 hash=crush:0x82000003 version=0 migrate offset=1 migrate hash=crush:3. LustreError: 16406:0:(lov_object.c:1299:lov_layout_change()) lustre-clilov-ffff880298df53d8: cannot apply new layout on [0x200000403:0x398:0x0] : rc = -22 LustreError: 16406:0:(lov_object.c:1299:lov_layout_change()) Skipped 5 previous similar messages LustreError: 16406:0:(lcommon_cl.c:196:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000403:0x398:0x0]: rc = -22 LustreError: 16406:0:(lcommon_cl.c:196:cl_file_inode_init()) Skipped 5 previous similar messages LustreError: 16406:0:(llite_lib.c:2983:ll_prep_inode()) new_inode -fatal: rc -22 LustreError: 16406:0:(llite_lib.c:2983:ll_prep_inode()) Skipped 5 previous similar messages LustreError: 18286:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0000: '2' migration was interrupted, run 'lfs migrate -m 1 -c 1 -H crush 2' to finish migration. LustreError: 17508:0:(lustre_lmv.h:438:lmv_is_sane()) insane LMV: magic=0xcd40cd0 count=2 index=1 hash=crush:0x82000003 version=0 migrate offset=1 migrate hash=crush:3. LustreError: 16860:0:(lov_object.c:1299:lov_layout_change()) lustre-clilov-ffff880298df53d8: cannot apply new layout on [0x200000403:0x398:0x0] : rc = -22 LustreError: 16860:0:(lov_object.c:1299:lov_layout_change()) Skipped 10 previous similar messages LustreError: 16860:0:(lcommon_cl.c:196:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000403:0x398:0x0]: rc = -22 LustreError: 16860:0:(lcommon_cl.c:196:cl_file_inode_init()) Skipped 10 previous similar messages LustreError: 16860:0:(llite_lib.c:2983:ll_prep_inode()) new_inode -fatal: rc -22 LustreError: 16860:0:(llite_lib.c:2983:ll_prep_inode()) Skipped 10 previous similar messages Lustre: 18367:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 3/12/5, destroy: 0/0/0 Lustre: 18367:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 295798 previous similar messages Lustre: 18367:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 21/1917/0 Lustre: 18367:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 295801 previous similar messages Lustre: 18367:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 5/136/0, punch: 0/0/0, quota 8/56/0 Lustre: 18367:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 295802 previous similar messages Lustre: 18367:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 10/203/5, delete: 0/0/0 Lustre: 18367:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 295799 previous similar messages Lustre: 18367:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 4/4/0, ref_del: 0/0/0 Lustre: 18367:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 295794 previous similar messages LustreError: 16882:0:(lov_object.c:1299:lov_layout_change()) lustre-clilov-ffff880298df53d8: cannot apply new layout on [0x200000403:0x398:0x0] : rc = -22 LustreError: 16882:0:(lov_object.c:1299:lov_layout_change()) Skipped 7 previous similar messages LustreError: 16882:0:(lcommon_cl.c:196:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000403:0x398:0x0]: rc = -22 LustreError: 16882:0:(lcommon_cl.c:196:cl_file_inode_init()) Skipped 7 previous similar messages LustreError: 16882:0:(llite_lib.c:2983:ll_prep_inode()) new_inode -fatal: rc -22 LustreError: 16882:0:(llite_lib.c:2983:ll_prep_inode()) Skipped 7 previous similar messages LustreError: 18804:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0002: failed to get lu_attr of [0x280000404:0x8f:0x0]: rc = -2 Lustre: 28262:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 493 < left 81007, rollback = 7 Lustre: 28262:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 290292 previous similar messages LustreError: 16861:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000405:0xa7:0x0]: rc = -2 LustreError: 17307:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000405:0xa7:0x0]: rc = -2 LustreError: 17325:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff880298dfae98: inode [0x240000404:0x483:0x0] mdc close failed: rc = -108 LustreError: 17325:0:(file.c:234:ll_close_inode_openhandle()) Skipped 15 previous similar messages LustreError: 17913:0:(lov_object.c:1299:lov_layout_change()) lustre-clilov-ffff880298df53d8: cannot apply new layout on [0x200000403:0x398:0x0] : rc = -22 LustreError: 17913:0:(lov_object.c:1299:lov_layout_change()) Skipped 13 previous similar messages LustreError: 17913:0:(lcommon_cl.c:196:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000403:0x398:0x0]: rc = -22 LustreError: 17913:0:(lcommon_cl.c:196:cl_file_inode_init()) Skipped 13 previous similar messages LustreError: 17913:0:(llite_lib.c:2983:ll_prep_inode()) new_inode -fatal: rc -22 LustreError: 17913:0:(llite_lib.c:2983:ll_prep_inode()) Skipped 13 previous similar messages 7[18271]: segfault at 8 ip 00007f8f193f27e8 sp 00007ffd45f6a010 error 4 in ld-2.17.so[7f8f193e7000+22000] LustreError: 11085:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0002: '12' migration was interrupted, run 'lfs migrate -m 1 -c 1 -H crush 12' to finish migration. LustreError: 18680:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1631101451 with bad export cookie 4685211556391565618 Lustre: 4129:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.185@tcp:/lustre/fid: [0x280000403:0x645:0x0]// may get corrupted (rc -108) LustreError: 18086:0:(vvp_io.c:1793:vvp_io_init()) lustre: refresh file layout [0x280000403:0x645:0x0] error -108. Lustre: 4130:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.185@tcp:/lustre/fid: [0x240000403:0x497:0x0]/ may get corrupted (rc -108) LustreError: 16620:0:(ofd_obd.c:501:ofd_destroy_export()) lustre-OST0001: cli 1154ae0e-996a-42a4-bcbf-d44ba44315ac/ffff88008ee48958 has 122880 pending on destroyed export Lustre: 4145:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.185@tcp:/lustre/fid: [0x240000403:0x52f:0x0]/ may get corrupted (rc -108) Lustre: 4146:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.185@tcp:/lustre/fid: [0x280000405:0x1dd:0x0]/ may get corrupted (rc -108) Lustre: 4145:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.185@tcp:/lustre/fid: [0x240000404:0x4bf:0x0]/ may get corrupted (rc -108) Lustre: 4146:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.185@tcp:/lustre/fid: [0x240000403:0x4b8:0x0]// may get corrupted (rc -108) LustreError: 11122:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000405:0xd1:0x0]: rc = -2 LustreError: 12687:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 64783680 != fo_tot_granted 64935232 LustreError: 12687:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 540 previous similar messages LustreError: 12687:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 151552 LustreError: 12687:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 540 previous similar messages Lustre: 23904:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 8s req@ffff88007bcc4b40 x1710331806824832/t0(0) o35->1154ae0e-996a-42a4-bcbf-d44ba44315ac@0@lo:0/0 lens 392/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'rm.0' Lustre: 23904:0:(service.c:2157:ptlrpc_server_handle_req_in()) Skipped 11 previous similar messages Lustre: lustre-OST0003-osc-ffff880298df53d8: Connection restored to 192.168.123.185@tcp (at 0@lo) Lustre: Skipped 11 previous similar messages LustreError: 18615:0:(lov_object.c:1299:lov_layout_change()) lustre-clilov-ffff880298dfae98: cannot apply new layout on [0x200000403:0x398:0x0] : rc = -22 LustreError: 18615:0:(lov_object.c:1299:lov_layout_change()) Skipped 2 previous similar messages LustreError: 18615:0:(lcommon_cl.c:196:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000403:0x398:0x0]: rc = -22 LustreError: 18615:0:(lcommon_cl.c:196:cl_file_inode_init()) Skipped 2 previous similar messages LustreError: 18615:0:(llite_lib.c:2983:ll_prep_inode()) new_inode -fatal: rc -22 LustreError: 18615:0:(llite_lib.c:2983:ll_prep_inode()) Skipped 2 previous similar messages Lustre: 28263:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000405:0xce:0x0] with magic=0xbd60bd0 Lustre: 28263:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 1 previous similar message Lustre: 4138:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.185@tcp:/lustre/fid: [0x240000405:0x62:0x0]// may get corrupted (rc -108) Lustre: 11099:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 8: before 511 < left 25668, rollback = 9 Lustre: 11099:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 291 previous similar messages LustreError: 11082:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0002: '10' migration was interrupted, run 'lfs migrate -m 1 -c 1 -H crush 10' to finish migration. LustreError: 22057:0:(lov_object.c:1299:lov_layout_change()) lustre-clilov-ffff880298dfae98: cannot apply new layout on [0x200000403:0x398:0x0] : rc = -22 LustreError: 22057:0:(lov_object.c:1299:lov_layout_change()) Skipped 34 previous similar messages LustreError: 22057:0:(lcommon_cl.c:196:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000403:0x398:0x0]: rc = -22 LustreError: 22057:0:(lcommon_cl.c:196:cl_file_inode_init()) Skipped 34 previous similar messages LustreError: 22057:0:(llite_lib.c:2983:ll_prep_inode()) new_inode -fatal: rc -22 LustreError: 22057:0:(llite_lib.c:2983:ll_prep_inode()) Skipped 34 previous similar messages ptlrpc_watchdog_fire: 30 callbacks suppressed Lustre: ll_ost01_043: service thread pid 18829 was inactive for 40.095 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 18829, comm: ll_ost01_043 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] tgt_extent_lock+0xea/0x2a0 [ptlrpc] [<0>] ofd_getattr_hdl+0x365/0x720 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: 4130:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.185@tcp:/lustre/fid: [0x280000403:0x6e1:0x0]/ may get corrupted (rc -108) Lustre: 4130:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.185@tcp:/lustre/fid: [0x240000404:0x47e:0x0]/ may get corrupted (rc -108) LustreError: 5:0:(ofd_obd.c:501:ofd_destroy_export()) lustre-OST0000: cli 1154ae0e-996a-42a4-bcbf-d44ba44315ac/ffff88008a34ae98 has 28672 pending on destroyed export Lustre: 4139:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.185@tcp:/lustre/fid: [0x280000403:0x737:0x0]/ may get corrupted (rc -5) LustreError: 17946:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88022e33c500 x1710331817452928/t0(0) o104->lustre-OST0001@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 17946:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 763 previous similar messages Lustre: 4134:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.185@tcp:/lustre/fid: [0x200000405:0x233:0x0]/ may get corrupted (rc -108) Lustre: 4133:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.185@tcp:/lustre/fid: [0x200000405:0x1db:0x0]/ may get corrupted (rc -108) Lustre: 4133:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.185@tcp:/lustre/fid: [0x200000403:0x4d3:0x0]/ may get corrupted (rc -108) 13[24250]: segfault at 8 ip 00007f5f9ee9f7e8 sp 00007ffc0afc5150 error 4 in ld-2.17.so[7f5f9ee94000+22000] Lustre: dir [0x280000403:0x608:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 3 previous similar messages Lustre: 4138:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff88024b97f700 x1710331819612864/t0(0) o10->lustre-OST0003-osc-ffff880298df53d8@0@lo:6/4 lens 440/432 e 0 to 0 dl 1631101888 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dd.0' Lustre: 4138:0:(client.c:1485:after_reply()) Skipped 5 previous similar messages Lustre: ll_ost02_031: service thread pid 4793 was inactive for 40.055 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 4793, comm: ll_ost02_031 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] tgt_extent_lock+0xea/0x2a0 [ptlrpc] [<0>] ofd_getattr_hdl+0x365/0x720 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe 17[25070]: segfault at 8 ip 00007f54b9a227e8 sp 00007ffdc2ed5850 error 4 in ld-2.17.so[7f54b9a17000+22000] LustreError: 26211:0:(vvp_io.c:1793:vvp_io_init()) lustre: refresh file layout [0x200000405:0x26d:0x0] error -22. LustreError: 25173:0:(lcommon_cl.c:196:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000405:0x26d:0x0]: rc = -22 LustreError: 25173:0:(lcommon_cl.c:196:cl_file_inode_init()) Skipped 57 previous similar messages LustreError: 25173:0:(llite_lib.c:2983:ll_prep_inode()) new_inode -fatal: rc -22 LustreError: 25173:0:(llite_lib.c:2983:ll_prep_inode()) Skipped 55 previous similar messages LustreError: 26468:0:(lov_object.c:1299:lov_layout_change()) lustre-clilov-ffff880298df53d8: cannot apply new layout on [0x200000403:0x398:0x0] : rc = -22 LustreError: 26468:0:(lov_object.c:1299:lov_layout_change()) Skipped 59 previous similar messages LustreError: 10998:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 101s: evicting client at 0@lo ns: filter-lustre-OST0003_UUID lock: ffff88025f628400/0x4105371654604b34 lrc: 3/0,0 mode: PW/PW res: [0x380000400:0x358a:0x0].0x0 rrc: 5 type: EXT [0->18446744073709551615] (req 0->1310719) gid 0 flags: 0x60000400020020 nid: 0@lo remote: 0x41053716546048b7 expref: 2155 pid: 8445 timeout: 2594 lvb_type: 0 LustreError: 10998:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 9 previous similar messages LustreError: 11-0: lustre-OST0003-osc-ffff880298df53d8: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 15 previous similar messages LustreError: 167-0: lustre-OST0003-osc-ffff880298df53d8: This client was evicted by lustre-OST0003; in progress operations using this service will fail. LustreError: Skipped 10 previous similar messages Lustre: 4135:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.185@tcp:/lustre/fid: [0x240000403:0x63e:0x0]/ may get corrupted (rc -108) Lustre: 4136:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.185@tcp:/lustre/fid: [0x200000405:0x1f4:0x0]// may get corrupted (rc -108) Lustre: 4135:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.185@tcp:/lustre/fid: [0x240000403:0x4b8:0x0]/ may get corrupted (rc -108) LustreError: 27577:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0003-osc-ffff880298df53d8: namespace resource [0x380000400:0x3cc0:0x0].0x0 (ffff8802521796c0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 27577:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 3035 previous similar messages LustreError: 19337:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0000: '12' migration was interrupted, run 'lfs migrate -m 1 -c 1 -H crush 12' to finish migration. LustreError: 19337:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) Skipped 2 previous similar messages LustreError: 28219:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff880298df53d8: inode [0x280000403:0x830:0x0] mdc close failed: rc = -13 LustreError: 28219:0:(file.c:234:ll_close_inode_openhandle()) Skipped 20 previous similar messages Lustre: 16920:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 3/12/2, destroy: 0/0/0 Lustre: 16920:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 172516 previous similar messages Lustre: 16920:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 2/2/0, xattr_set: 11/877/0 Lustre: 16920:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 172516 previous similar messages Lustre: 16920:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 5/136/0, punch: 0/0/0, quota 6/6/0 Lustre: 16920:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 172519 previous similar messages Lustre: 16920:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 12/215/2, delete: 1/1/0 Lustre: 16920:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 172517 previous similar messages Lustre: 16920:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 6/6/0, ref_del: 1/1/0 Lustre: 16920:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 172517 previous similar messages Lustre: 10383:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 510 < left 781, rollback = 7 Lustre: 10383:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 172982 previous similar messages Lustre: 18367:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000403:0x621:0x0] with magic=0xbd60bd0 Lustre: 18367:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 1 previous similar message Lustre: ll_ost07_054: service thread pid 11676 was inactive for 100.075 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 11676, comm: ll_ost07_054 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] tgt_extent_lock+0xea/0x2a0 [ptlrpc] [<0>] ofd_getattr_hdl+0x365/0x720 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: 28419:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000403:0x668:0x0] with magic=0xbd60bd0 Lustre: 28419:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 3 previous similar messages LustreError: 10991:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1631102033 with bad export cookie 4685211556392595059 LustreError: 10991:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) Skipped 3 previous similar messages Lustre: lustre-OST0003-osc-ffff880298dfae98: Connection to lustre-OST0003 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 10 previous similar messages Lustre: 4144:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.185@tcp:/lustre/fid: [0x240000403:0x690:0x0]// may get corrupted (rc -108) Lustre: 4143:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.185@tcp:/lustre/fid: [0x280000403:0x855:0x0]// may get corrupted (rc -108) Lustre: 4143:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.185@tcp:/lustre/fid: [0x240000405:0x228:0x0]// may get corrupted (rc -108) Lustre: 4144:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.185@tcp:/lustre/fid: [0x200000405:0x31c:0x0]// may get corrupted (rc -108) LustreError: 12685:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 261520064 != fo_tot_granted 261548736 LustreError: 12685:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 696 previous similar messages LustreError: 12685:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 4218880 != fo_tot_pending 4247552 LustreError: 12685:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 696 previous similar messages 19[32206]: segfault at 0 ip (null) sp 00007ffcca91a548 error 14 in 14[400000+6000] LustreError: 32222:0:(vvp_io.c:1793:vvp_io_init()) lustre: refresh file layout [0x200000403:0x398:0x0] error -22. Lustre: 4143:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.185@tcp:/lustre/fid: [0x200000405:0x3cd:0x0]/ may get corrupted (rc -5) LustreError: 375:0:(ofd_obd.c:501:ofd_destroy_export()) lustre-OST0002: cli 1154ae0e-996a-42a4-bcbf-d44ba44315ac/ffff8802c4cb6fc8 has 1196032 pending on destroyed export Lustre: lustre-OST0002-osc-ffff880298df53d8: Connection restored to 192.168.123.185@tcp (at 0@lo) Lustre: Skipped 8 previous similar messages ------------[ cut here ]------------ WARNING: CPU: 10 PID: 19337 at /home/green/git/lustre-release/ldiskfs/ext4_jbd2.c:266 __ldiskfs_handle_dirty_metadata+0x1d2/0x230 [ldiskfs] Modules linked in: loop zfs(PO) zunicode(PO) zzstd(O) zlua(O) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) jbd2 mbcache lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) dm_flakey dm_mod libcfs(OE) crc_t10dif crct10dif_generic sb_edac edac_core iosf_mbi crc32_pclmul ghash_clmulni_intel aesni_intel lrw gf128mul glue_helper ablk_helper cryptd virtio_balloon pcspkr virtio_console i2c_piix4 ip_tables rpcsec_gss_krb5 drm_kms_helper ata_generic pata_acpi ttm crct10dif_pclmul drm crct10dif_common ata_piix crc32c_intel drm_panel_orientation_quirks virtio_blk libata serio_raw i2c_core floppy CPU: 10 PID: 19337 Comm: mdt05_009 Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011 Call Trace: [<ffffffff817d9418>] dump_stack+0x19/0x1b [<ffffffff8108d538>] __warn+0xd8/0x100 [<ffffffff8108d67d>] warn_slowpath_null+0x1d/0x20 [<ffffffffa0c14262>] __ldiskfs_handle_dirty_metadata+0x1d2/0x230 [ldiskfs] [<ffffffffa0bf433d>] ldiskfs_getblk+0x13d/0x210 [ldiskfs] [<ffffffffa0bf4437>] ldiskfs_bread+0x27/0xe0 [ldiskfs] [<ffffffffa0cb8e97>] osd_write+0x4e7/0xcb0 [osd_ldiskfs] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffffa04de6f3>] dt_record_write+0x33/0x120 [obdclass] [<ffffffffa049d6b2>] llog_osd_write_rec+0x1052/0x19a0 [obdclass] [<ffffffffa048a760>] llog_write_rec+0x290/0x590 [obdclass] [<ffffffffa04904ee>] llog_cat_add_rec+0x23e/0x950 [obdclass] [<ffffffffa04875ff>] llog_add+0x17f/0x1f0 [obdclass] [<ffffffffa04b8a39>] ? lprocfs_counter_add+0xf9/0x160 [obdclass] [<ffffffffa082160d>] sub_updates_write+0xb6b/0xe32 [ptlrpc] [<ffffffffa07f1892>] top_trans_stop+0x7c2/0xf60 [ptlrpc] [<ffffffffa0ef3f3c>] lod_trans_stop+0x25c/0x340 [lod] [<ffffffffa0dc8408>] mdd_trans_stop+0x28/0x16e [mdd] [<ffffffffa0da9b2f>] mdd_migrate_object+0xf1f/0x1ba0 [mdd] [<ffffffffa04dd375>] ? lu_object_find_at+0x285/0xb10 [obdclass] [<ffffffffa0daab86>] mdd_migrate+0x3d6/0x8e0 [mdd] [<ffffffffa0e7c0d8>] mdo_migrate+0x4c/0x4e [mdt] [<ffffffffa0e3be53>] mdt_reint_migrate+0x1023/0x11d0 [mdt] [<ffffffffa0e3c087>] mdt_reint_rec+0x87/0x240 [mdt] [<ffffffffa0e1155c>] mdt_reint_internal+0x76c/0xb40 [mdt] [<ffffffffa0e1e657>] mdt_reint+0x67/0x150 [mdt] [<ffffffffa07e026e>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<ffffffffa032d45e>] ? libcfs_nid2str_r+0xfe/0x130 [lnet] [<ffffffffa078b990>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<ffffffffa078d529>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffffa078c950>] ? ptlrpc_wait_event+0x620/0x620 [ptlrpc] [<ffffffff810ba0f4>] kthread+0xe4/0xf0 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 ---[ end trace 28bda789721aa800 ]--- LDISKFS-fs: ldiskfs_getblk:888: aborting transaction: error 28 in __ldiskfs_handle_dirty_metadata LDISKFS-fs error (device dm-1): ldiskfs_getblk:888: inode #193: block 20973: comm mdt05_009: journal_dirty_metadata failed: handle type 0 started at line 1972, credits 512/0, errcode -28 Aborting journal on device dm-1-8. LDISKFS-fs (dm-1): Remounting filesystem read-only LustreError: 19337:0:(osd_io.c:2103:osd_ldiskfs_write_record()) lustre-MDT0001: error reading offset 659456 (block 161, size 48, offs 659448), credits 0/0: rc = -28 LustreError: 19337:0:(update_trans.c:985:top_trans_stop()) lustre-MDT0001-osd: write updates failed: rc = -28 LustreError: 19337:0:(osd_io.c:2110:osd_ldiskfs_write_record()) journal_get_write_access() returned error -30 LustreError: 19337:0:(tgt_lastrcvd.c:1253:tgt_add_reply_data()) lustre-MDT0001: can't update reply_data file: rc = -30 LustreError: 19337:0:(osd_handler.c:2075:osd_trans_stop()) lustre-MDT0001: failed in transaction hook: rc = -30 LDISKFS-fs error (device dm-1) in osd_trans_stop:2082: error 28 LustreError: 12911:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-OST0000-osc-MDT0001: fail to cancel 1 llog-records: rc = -30 | Externally reported by onyx-68 boilpot email |
racer test 1: racer on clients: centos-115.localnet DURATION=2700 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa179fe52>] ll_migrate+0x9b2/0xec0 [lustre] PGD 80000000a5c56067 PUD a5c57067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) dm_flakey dm_mod loop zfs(PO) zunicode(PO) zzstd(O) zlua(O) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) jbd2 mbcache crc_t10dif crct10dif_generic sb_edac edac_core iosf_mbi crc32_pclmul ghash_clmulni_intel aesni_intel lrw gf128mul glue_helper ablk_helper cryptd pcspkr virtio_balloon virtio_console i2c_piix4 ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm crct10dif_pclmul drm crct10dif_common drm_panel_orientation_quirks ata_piix crc32c_intel serio_raw virtio_blk i2c_core libata floppy [last unloaded: libcfs] CPU: 4 PID: 3140 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011 task: ffff8802c4199280 ti: ffff8800a368c000 task.ti: ffff8800a368c000 RIP: 0010:[<ffffffffa179fe52>] [<ffffffffa179fe52>] ll_migrate+0x9b2/0xec0 [lustre] RSP: 0018:ffff8800a368fbc8 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff8802a0ab55e8 RCX: 0000000000000000 RDX: 0000000000000000 RSI: 0000000000000247 RDI: 0000000000000247 RBP: ffff8800a368fc48 R08: ffff88008af77660 R09: 0000000000000000 R10: ffff88008af77640 R11: ffff8802d9c2c5a8 R12: 0000000000000000 R13: ffff8802e6e79738 R14: ffff880063e83898 R15: 0000000000000030 FS: 00007f1952ccc740(0000) GS:ffff880331b00000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 CR2: 0000000000000008 CR3: 00000000a045c000 CR4: 00000000001607e0 Call Trace: [<ffffffff81242213>] ? __check_object_size+0x1c3/0x220 [<ffffffffa178d561>] ll_dir_ioctl+0x5d01/0x6ed0 [lustre] [<ffffffff811b515b>] ? unlock_page+0x2b/0x30 [<ffffffff811e5e2e>] ? do_read_fault.isra.64+0x12e/0x170 [<ffffffff811ea1cf>] ? handle_pte_fault+0x2ef/0xca0 [<ffffffff810dbb8b>] ? put_prev_entity+0x8b/0x400 [<ffffffff810d5979>] ? pick_next_entity+0xa9/0x190 LustreError: 9080:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -30 LustreError: 9080:0:(llite_lib.c:1836:ll_md_setattr()) Skipped 3 previous similar messages [<ffffffff810d7e1c>] ? set_next_entity+0x3c/0xe0 [<ffffffff811ecf6d>] ? __handle_mm_fault+0x31d/0x720 [<ffffffff8125b3fd>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff81245a38>] ? vfs_write+0x178/0x230 [<ffffffff8125b751>] SyS_ioctl+0xa1/0xc0 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 | Lustre: 24261:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 512 < left 531, rollback = 7 Lustre: 24261:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 2/8/0, destroy: 0/0/0 Lustre: 24261:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 5/192/0 Lustre: 24261:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 446/531/0, punch: 0/0/0, quota 6/6/0 Lustre: 24261:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 2/34/0, delete: 0/0/0 Lustre: 24261:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 24261:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 511 < left 530, rollback = 7 Lustre: 24261:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 2/8/0, destroy: 0/0/0 Lustre: 24261:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 5/192/0 Lustre: 24261:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 446/531/1, punch: 0/0/0, quota 6/6/0 Lustre: 24261:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 2/34/0, delete: 0/0/0 Lustre: 24261:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 24261:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 500 < left 528, rollback = 7 Lustre: 24261:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 436 previous similar messages Lustre: 24261:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 2/8/0, destroy: 0/0/0 Lustre: 24261:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 436 previous similar messages Lustre: 24261:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 5/192/9 Lustre: 24261:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 436 previous similar messages Lustre: 24261:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 446/531/3, punch: 0/0/0, quota 6/6/0 Lustre: 24261:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 436 previous similar messages Lustre: 24261:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 2/34/0, delete: 0/0/0 Lustre: 24261:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 436 previous similar messages Lustre: 24261:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 24261:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 436 previous similar messages Lustre: 22308:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 508 < left 610, rollback = 2 Lustre: 22308:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 3/12/1, destroy: 0/0/0 Lustre: 22308:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 1745 previous similar messages Lustre: 22308:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 2/2/0, xattr_set: 8/610/0 Lustre: 22308:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 1745 previous similar messages Lustre: 22308:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 5/136/0, punch: 0/0/0, quota 6/6/0 Lustre: 22308:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 1745 previous similar messages Lustre: 22308:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 12/215/3, delete: 1/1/0 Lustre: 22308:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 1745 previous similar messages Lustre: 22308:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 6/6/0, ref_del: 1/1/0 Lustre: 22308:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 1745 previous similar messages Lustre: 21526:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 509 < left 7406, rollback = 7 Lustre: 21526:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 1745 previous similar messages Lustre: 24409:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 505 < left 582, rollback = 2 Lustre: 24409:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 5 previous similar messages LustreError: 21560:0:(out_handler.c:910:out_tx_end()) lustre-MDT0001-osd: undo for /home/green/git/lustre-release/lustre/ptlrpc/../../lustre/target/out_handler.c:445: rc = -524 LustreError: 21606:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0001-osp-MDT0000: fail to cancel 1 llog-records: rc = -116 LustreError: 21606:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0001-osp-MDT0000: fail to cancel 1 of 1 llog-records: rc = -116 LustreError: 21606:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0001-osp-MDT0000: fail to cancel 1 llog-records: rc = -116 LustreError: 21606:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0001-osp-MDT0000: fail to cancel 1 of 1 llog-records: rc = -116 Lustre: 23621:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 510 < left 1491, rollback = 7 Lustre: 23621:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 688 previous similar messages Lustre: 23621:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 1/4/2, destroy: 0/0/0 Lustre: 23621:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 700 previous similar messages Lustre: 23621:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 4/209/0 Lustre: 23621:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 700 previous similar messages Lustre: 23621:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1482/1491/0, punch: 0/0/0, quota 6/6/0 Lustre: 23621:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 700 previous similar messages Lustre: 23621:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 2/33/0, delete: 0/0/0 Lustre: 23621:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 700 previous similar messages Lustre: 23621:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 23621:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 700 previous similar messages LustreError: 25302:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff880071ab1bf8: inode [0x280000403:0x1b:0x0] mdc close failed: rc = -13 Lustre: 25571:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 512 < left 595, rollback = 7 Lustre: 25571:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 1481 previous similar messages Lustre: 25571:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 2/8/0, destroy: 0/0/0 Lustre: 25571:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 1481 previous similar messages Lustre: 25571:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 5/192/0 Lustre: 25571:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 1481 previous similar messages Lustre: 25571:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 510/595/0, punch: 0/0/0, quota 6/6/0 Lustre: 25571:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 1481 previous similar messages Lustre: 25571:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 2/34/0, delete: 0/0/0 Lustre: 25571:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 1481 previous similar messages Lustre: 25571:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 25571:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 1481 previous similar messages LustreError: 25553:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0001: '15' migration was interrupted, run 'lfs migrate -m 2 -c 1 -H crush 15' to finish migration. Lustre: 24830:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 8: before 509 < left 8533, rollback = 9 Lustre: 24830:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 5 previous similar messages LustreError: 26545:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff880071ab1bf8: inode [0x200000404:0x6:0x0] mdc close failed: rc = -13 Lustre: 25536:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 501 < left 610, rollback = 2 LustreError: 26897:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff880071ab1bf8: inode [0x200000404:0x6:0x0] mdc close failed: rc = -13 Lustre: 21544:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 510 < left 61457, rollback = 7 Lustre: 21544:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 12155 previous similar messages Lustre: 21544:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 1429/5716/0, destroy: 1/4/1 Lustre: 21544:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 12169 previous similar messages Lustre: 21544:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 1429/0/0 Lustre: 21544:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 12169 previous similar messages Lustre: 21544:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 7146/61457/0, punch: 0/0/0, quota 6/6/0 Lustre: 21544:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 12169 previous similar messages Lustre: 21544:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 1429/24293/0, delete: 2/5/1 Lustre: 21544:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 12169 previous similar messages Lustre: 21544:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 1/1/0 Lustre: 21544:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 12169 previous similar messages ------------[ cut here ]------------ WARNING: CPU: 14 PID: 21544 at /home/green/git/lustre-release/ldiskfs/namei.c:3331 ldiskfs_orphan_add+0x11e/0x2a0 [ldiskfs] Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) dm_flakey dm_mod loop zfs(PO) zunicode(PO) zzstd(O) zlua(O) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) jbd2 mbcache crc_t10dif crct10dif_generic sb_edac edac_core iosf_mbi crc32_pclmul ghash_clmulni_intel aesni_intel lrw gf128mul glue_helper ablk_helper cryptd pcspkr virtio_balloon virtio_console i2c_piix4 ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm crct10dif_pclmul drm crct10dif_common drm_panel_orientation_quirks ata_piix crc32c_intel serio_raw virtio_blk i2c_core libata floppy [last unloaded: libcfs] CPU: 14 PID: 21544 Comm: mdt_rdpg07_000 Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011 Call Trace: [<ffffffff817d9418>] dump_stack+0x19/0x1b [<ffffffff8108d538>] __warn+0xd8/0x100 [<ffffffff8108d67d>] warn_slowpath_null+0x1d/0x20 [<ffffffffa0d6b4ae>] ldiskfs_orphan_add+0x11e/0x2a0 [ldiskfs] [<ffffffffa0d5c9eb>] ldiskfs_xattr_inode_orphan_add.constprop.17+0xbb/0x110 [ldiskfs] [<ffffffffa0d5f417>] ldiskfs_xattr_delete_inode+0x77/0x3c0 [ldiskfs] [<ffffffffa0d7faa0>] ? ldiskfs_evict_inode+0x190/0x670 [ldiskfs] [<ffffffffa0d7fad9>] ldiskfs_evict_inode+0x1c9/0x670 [ldiskfs] [<ffffffff812649af>] evict+0xaf/0x180 [<ffffffff81264de5>] iput+0xf5/0x180 [<ffffffffa0c9bd08>] osd_object_delete+0x1f8/0x370 [osd_ldiskfs] [<ffffffffa04de59d>] lu_object_free.isra.26+0xcd/0x1d0 [obdclass] [<ffffffffa04e2cf1>] lu_object_put+0xa1/0x480 [obdclass] [<ffffffffa0efbaf4>] mdt_mfd_close+0x1f4/0x870 [mdt] [<ffffffffa0f01e14>] mdt_close_internal+0xb4/0x240 [mdt] [<ffffffffa0f02224>] mdt_close+0x284/0x850 [mdt] [<ffffffffa07e52d4>] ? tgt_request_preprocess.isra.27+0xf4/0x850 [ptlrpc] [<ffffffffa07e626e>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<ffffffffa03ca45e>] ? libcfs_nid2str_r+0xfe/0x130 [lnet] [<ffffffffa0791990>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<ffffffffa0793529>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffffa0792950>] ? ptlrpc_wait_event+0x620/0x620 [ptlrpc] [<ffffffff810ba0f4>] kthread+0xe4/0xf0 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 ---[ end trace 331e0184efa6e1eb ]--- Lustre: 21524:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 8: before 509 < left 7445, rollback = 9 Lustre: 21524:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 12 previous similar messages Lustre: 24236:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 504 < left 36904, rollback = 7 Lustre: 24236:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 20392 previous similar messages Lustre: 24236:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 858/3432/0, destroy: 0/0/0 Lustre: 24236:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 20393 previous similar messages Lustre: 24236:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/1, xattr_set: 860/15/0 Lustre: 24236:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 20393 previous similar messages Lustre: 24236:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 4291/36904/0, punch: 0/0/0, quota 14/174/7 Lustre: 24236:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 20392 previous similar messages Lustre: 24236:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 858/14586/0, delete: 0/0/0 Lustre: 24236:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 20392 previous similar messages Lustre: 24236:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 24236:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 20393 previous similar messages LustreError: 25330:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0001: '15' migration was interrupted, run 'lfs migrate -m 2 -c 1 -H crush 15' to finish migration. LustreError: 26115:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0002: '18' migration was interrupted, run 'lfs migrate -m 1 -c 1 -H crush 18' to finish migration. Lustre: 21525:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 8: before 510 < left 13497, rollback = 9 LustreError: 28322:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff88008c3a92a8: inode [0x200000404:0x6:0x0] mdc close failed: rc = -13 Lustre: 23621:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 8: before 511 < left 2973, rollback = 9 Lustre: 23621:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 1 previous similar message LustreError: 21499:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0003_UUID lock: ffff88021d029300/0x1765494a08b78931 lrc: 3/0,0 mode: PW/PW res: [0x380000401:0x2:0x0].0x0 rrc: 4 type: EXT [0->18446744073709551615] (req 0->131071) gid 0 flags: 0x60000480000020 nid: 0@lo remote: 0x1765494a08b78907 expref: 1542 pid: 22185 timeout: 3654 lvb_type: 0 LustreError: 29091:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802a7093880 x1710325411333568/t0(0) o105->lustre-OST0003@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 11-0: lustre-OST0003-osc-ffff880071ab1bf8: operation ost_setattr to node 0@lo failed: rc = -107 Lustre: lustre-OST0003-osc-ffff880071ab1bf8: Connection to lustre-OST0003 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0003-osc-ffff880071ab1bf8: This client was evicted by lustre-OST0003; in progress operations using this service will fail. Lustre: 18309:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x280000404:0x6e:0x0]// may get corrupted (rc -108) LustreError: 29571:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0003-osc-ffff880071ab1bf8: namespace resource [0x10:0x0:0x0].0x0 (ffff88024b4f3240) refcount nonzero (1) after lock cleanup; forcing cleanup. Lustre: lustre-OST0003-osc-ffff880071ab1bf8: Connection restored to 192.168.123.215@tcp (at 0@lo) Lustre: 24261:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 8: before 510 < left 15350, rollback = 9 Lustre: 24261:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 1 previous similar message Lustre: 24261:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 902/3608/0, destroy: 1/4/0 Lustre: 24261:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 17200 previous similar messages Lustre: 24261:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 3/3/1, xattr_set: 906/148/0 Lustre: 24261:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 17200 previous similar messages Lustre: 24261:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 4511/38796/0, punch: 0/0/0, quota 10/74/0 Lustre: 24261:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 17201 previous similar messages Lustre: 24261:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 903/15350/0, delete: 2/5/1 Lustre: 24261:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 17201 previous similar messages Lustre: 24261:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 1/1/0, ref_del: 2/2/0 Lustre: 24261:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 17201 previous similar messages Lustre: 24261:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 509 < left 38796, rollback = 7 Lustre: 24261:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 17197 previous similar messages LustreError: 21606:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0002-osp-MDT0000: fail to cancel 1 llog-records: rc = -2 LustreError: 21606:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) Skipped 1 previous similar message LustreError: 21606:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0002-osp-MDT0000: fail to cancel 1 of 1 llog-records: rc = -2 LustreError: 21606:0:(llog_cat.c:787:llog_cat_cancel_records()) Skipped 1 previous similar message Lustre: mdt04_006: service thread pid 24743 was inactive for 62.053 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 24743, comm: mdt04_006 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] mdt_object_local_lock+0x527/0xb90 [mdt] [<0>] mdt_object_lock_internal+0x70/0x390 [mdt] [<0>] mdt_reint_object_lock+0x2c/0x60 [mdt] [<0>] mdt_reint_striped_lock+0x89/0x5b0 [mdt] [<0>] mdt_migrate_parent_lock+0x3f/0xb0 [mdt] [<0>] mdt_reint_migrate+0x8b7/0x11d0 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x76c/0xb40 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 22693, comm: mdt00_003 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] mdt_rename_lock+0x24b/0x4d0 [mdt] [<0>] mdt_reint_migrate+0x6e8/0x11d0 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] Lustre: mdt05_003: service thread pid 23873 was inactive for 62.091 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Lustre: Skipped 1 previous similar message [<0>] mdt_reint_internal+0x76c/0xb40 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 23873, comm: mdt05_003 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] mdt_rename_lock+0x24b/0x4d0 [mdt] [<0>] mdt_reint_rename+0x31c/0x2ff0 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x76c/0xb40 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] Lustre: mdt06_004: service thread pid 24550 was inactive for 62.230 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: mdt03_006: service thread pid 25330 was inactive for 62.158 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 1 previous similar message Lustre: mdt02_010: service thread pid 25571 was inactive for 62.125 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 1 previous similar message Lustre: mdt02_004: service thread pid 24451 was inactive for 62.276 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 1 previous similar message Lustre: mdt00_009: service thread pid 24869 was inactive for 62.119 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 1 previous similar message LustreError: 21499:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0000_UUID lock: ffff8802b01b8400/0x1765494a08ba3f65 lrc: 3/0,0 mode: PW/PW res: [0x2c0000400:0x20c:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->1048575) gid 0 flags: 0x60000400000020 nid: 0@lo remote: 0x1765494a08ba3efc expref: 1481 pid: 26582 timeout: 3732 lvb_type: 0 LustreError: 30361:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802c9deb240 x1710325414736320/t0(0) o105->lustre-OST0000@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 22203:0:(ldlm_lockd.c:1425:ldlm_handle_enqueue0()) ### lock on destroyed export ffff880278215d28 ns: filter-lustre-OST0000_UUID lock: ffff88022074c780/0x1765494a08c35df3 lrc: 3/0,0 mode: --/PW res: [0x2c0000400:0x6ab:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x50000000020000 nid: 0@lo remote: 0x1765494a08c35dec expref: 1271 pid: 22203 timeout: 0 lvb_type: 0 LustreError: 21494:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1631093716 with bad export cookie 1685834217842756103 Lustre: lustre-OST0000-osc-ffff88008c3a92a8: Connection to lustre-OST0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0000-osc-ffff88008c3a92a8: This client was evicted by lustre-OST0000; in progress operations using this service will fail. LustreError: 32507:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0000-osc-ffff88008c3a92a8: namespace resource [0x2c0000400:0x70e:0x0].0x0 (ffff880097738cc0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 32507:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 41 previous similar messages Lustre: lustre-OST0000-osc-ffff88008c3a92a8: Connection restored to 192.168.123.215@tcp (at 0@lo) Lustre: mdt03_003: service thread pid 23007 was inactive for 62.264 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. LustreError: 21499:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0001_UUID lock: ffff880092dc4f00/0x1765494a08be38c6 lrc: 3/0,0 mode: PR/PR res: [0x240000404:0x14:0x0].0x0 bits 0x12/0x0 rrc: 6 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0x1765494a08be389c expref: 80 pid: 25553 timeout: 3753 lvb_type: 0 LustreError: 11-0: lustre-MDT0001-mdc-ffff880071ab1bf8: operation mds_reint to node 0@lo failed: rc = -107 Lustre: lustre-MDT0001-mdc-ffff880071ab1bf8: Connection to lustre-MDT0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-MDT0001-mdc-ffff880071ab1bf8: This client was evicted by lustre-MDT0001; in progress operations using this service will fail. LustreError: 26665:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff880071ab1bf8: inode [0x240000403:0x4c:0x0] mdc close failed: rc = -108 LustreError: 28358:0:(llite_lib.c:2983:ll_prep_inode()) new_inode -fatal: rc -108 Lustre: lustre-MDT0001-mdc-ffff880071ab1bf8: Connection restored to 192.168.123.215@tcp (at 0@lo) Lustre: 24550:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 498 < left 1055, rollback = 2 Lustre: 24550:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 2 previous similar messages LustreError: 25553:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0000: '18' migration was interrupted, run 'lfs migrate -m 1 -c 1 -H crush 18' to finish migration. LustreError: 21499:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0003_UUID lock: ffff880226d7f480/0x1765494a08c07323 lrc: 3/0,0 mode: PW/PW res: [0x380000400:0x2f5:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000400020020 nid: 0@lo remote: 0x1765494a08c072cf expref: 1703 pid: 22200 timeout: 3781 lvb_type: 0 LustreError: 22299:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff880092434b40 x1710325416144384/t0(0) o105->lustre-OST0003@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 11-0: lustre-OST0003-osc-ffff880071ab1bf8: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-OST0003-osc-ffff880071ab1bf8: Connection to lustre-OST0003 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0003-osc-ffff880071ab1bf8: This client was evicted by lustre-OST0003; in progress operations using this service will fail. Lustre: 18311:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x240000404:0xe0:0x0]/ may get corrupted (rc -108) Lustre: lustre-OST0003-osc-ffff880071ab1bf8: Connection restored to 192.168.123.215@tcp (at 0@lo) Lustre: 18315:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff8802b29cd7c0 x1710325416918976/t0(0) o10->lustre-OST0003-osc-ffff880071ab1bf8@0@lo:6/4 lens 440/432 e 0 to 0 dl 1631093831 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'cp.0' Lustre: dir [0x280000403:0x17:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: 18316:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff8802b1d16440 x1710325417333696/t0(0) o10->lustre-OST0001-osc-ffff880071ab1bf8@0@lo:6/4 lens 440/432 e 0 to 0 dl 1631093806 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dd.0' Lustre: 25055:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 3/12/3, destroy: 0/0/0 Lustre: 25055:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 32341 previous similar messages Lustre: 25055:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 7/671/0 Lustre: 25055:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 32341 previous similar messages Lustre: 25055:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 5/136/0, punch: 0/0/0, quota 6/6/0 Lustre: 25055:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 32341 previous similar messages Lustre: 25055:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 10/203/4, delete: 0/0/0 Lustre: 25055:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 32341 previous similar messages Lustre: 25055:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 4/4/0, ref_del: 0/0/0 Lustre: 25055:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 32341 previous similar messages Lustre: 21517:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 510 < left 3321, rollback = 7 Lustre: 21517:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 32275 previous similar messages Lustre: dir [0x240000404:0xda:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: 18315:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff8802b29cd7c0 x1710325420127232/t0(0) o10->lustre-OST0003-osc-ffff880071ab1bf8@0@lo:6/4 lens 440/432 e 0 to 0 dl 1631093888 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'cp.0' Lustre: 18315:0:(client.c:1485:after_reply()) Skipped 74 previous similar messages Lustre: 18315:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff8802a7038040 x1710325420400000/t0(0) o10->lustre-OST0002-osc-ffff880071ab1bf8@0@lo:6/4 lens 440/432 e 0 to 0 dl 1631093902 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dd.0' Lustre: 18316:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff8802b1d16440 x1710325420720448/t0(0) o10->lustre-OST0001-osc-ffff880071ab1bf8@0@lo:6/4 lens 440/432 e 0 to 0 dl 1631093919 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dd.0' LustreError: 2715:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff88008c3a92a8: inode [0x200000404:0x5f:0x0] mdc close failed: rc = -13 LustreError: 2715:0:(file.c:234:ll_close_inode_openhandle()) Skipped 8 previous similar messages LustreError: 21499:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0002_UUID lock: ffff8802baf41e40/0x1765494a08c543fc lrc: 3/0,0 mode: PW/PW res: [0x340000400:0x1e9:0x0].0x0 rrc: 4 type: EXT [0->18446744073709551615] (req 1048576->18446744073709551615) gid 0 flags: 0x60000000020020 nid: 0@lo remote: 0x1765494a08c543f5 expref: 2064 pid: 26572 timeout: 3883 lvb_type: 0 LustreError: 23266:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802b67a6440 x1710325421128320/t0(0) o105->lustre-OST0002@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 23266:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 27 previous similar messages LustreError: 11-0: lustre-OST0002-osc-ffff88008c3a92a8: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-OST0002-osc-ffff88008c3a92a8: Connection to lustre-OST0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0002-osc-ffff88008c3a92a8: This client was evicted by lustre-OST0002; in progress operations using this service will fail. Lustre: 18309:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x240000404:0x1b9:0x0]// may get corrupted (rc -108) Lustre: lustre-OST0002-osc-ffff88008c3a92a8: Connection restored to 192.168.123.215@tcp (at 0@lo) Lustre: 26713:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 504 < left 1294, rollback = 2 Lustre: 26713:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 72 previous similar messages LustreError: 21539:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0002: failed to get lu_attr of [0x280000404:0x1fa:0x0]: rc = -2 Lustre: 27359:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 6s req@ffff880225b0cb40 x1710325422217728/t0(0) o13->lustre-MDT0002-mdtlov_UUID@0@lo:0/0 lens 224/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'osp-pre-0-2.0' Lustre: 22206:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 6s req@ffff8802b02c9f80 x1710325422565824/t0(0) o101->16015095-046a-4c50-a5f2-954fef3e91d0@0@lo:0/0 lens 328/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'truncate.0' LustreError: 21499:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0003_UUID lock: ffff880094f2ed00/0x1765494a08ce0e0c lrc: 3/0,0 mode: PW/PW res: [0x380000401:0x53c:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000400000020 nid: 0@lo remote: 0x1765494a08ce0e05 expref: 3389 pid: 26563 timeout: 4012 lvb_type: 0 LustreError: 6722:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88009f844b40 x1710325426848384/t0(0) o104->lustre-OST0003@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 6722:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 24 previous similar messages LustreError: 11-0: lustre-OST0003-osc-ffff880071ab1bf8: operation ost_getattr to node 0@lo failed: rc = -107 Lustre: lustre-OST0003-osc-ffff880071ab1bf8: Connection to lustre-OST0003 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0003-osc-ffff880071ab1bf8: This client was evicted by lustre-OST0003; in progress operations using this service will fail. Lustre: 18315:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x280000403:0x26c:0x0]/ may get corrupted (rc -108) Lustre: 18316:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x280000403:0x27e:0x0]/ may get corrupted (rc -108) Lustre: 18320:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff8802d1859940 x1710325427022144/t0(0) o10->lustre-OST0003-osc-ffff88008c3a92a8@0@lo:6/4 lens 440/432 e 0 to 0 dl 1631094063 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dir_create.sh.0' Lustre: 18320:0:(client.c:1485:after_reply()) Skipped 27 previous similar messages LustreError: 7391:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0003-osc-ffff880071ab1bf8: namespace resource [0x380000401:0x53f:0x0].0x0 (ffff880092303c40) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 7391:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 5 previous similar messages Lustre: 29080:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000403:0x26a:0x0] with magic=0xbd60bd0 LustreError: 3625:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0001: '8' migration was interrupted, run 'lfs migrate -m 0 -c 1 -H crush 8' to finish migration. Lustre: 21528:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000403:0x26f:0x0] with magic=0xbd60bd0 Lustre: 21528:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 1 previous similar message Lustre: 6721:0:(service.c:1378:ptlrpc_at_send_early_reply()) @@@ Already past deadline (-1s), not sending early reply. Consider increasing at_early_margin (5)? req@ffff8802b9f3c500 x1710325425372672/t0(0) o1->1c8284fc-3353-43bb-bb2e-6d88a2482b3d@0@lo:317/0 lens 440/0 e 0 to 0 dl 1631094012 ref 2 fl New:/0/ffffffff rc 0/-1 job:'lfs.0' Lustre: 22694:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000403:0x276:0x0] with magic=0xbd60bd0 Lustre: 6708:0:(service.c:1378:ptlrpc_at_send_early_reply()) @@@ Already past deadline (-1s), not sending early reply. Consider increasing at_early_margin (5)? req@ffff88009f900680 x1710325425403200/t0(0) o1->1c8284fc-3353-43bb-bb2e-6d88a2482b3d@0@lo:318/0 lens 440/0 e 0 to 0 dl 1631094013 ref 2 fl New:/0/ffffffff rc 0/-1 job:'lfs.0' Lustre: 6708:0:(service.c:1378:ptlrpc_at_send_early_reply()) Skipped 237 previous similar messages Lustre: 22694:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 1 previous similar message Lustre: 28146:0:(service.c:1378:ptlrpc_at_send_early_reply()) @@@ Already past deadline (-2s), not sending early reply. Consider increasing at_early_margin (5)? req@ffff8802232b9940 x1710325425645440/t0(0) o2->1c8284fc-3353-43bb-bb2e-6d88a2482b3d@0@lo:319/0 lens 440/0 e 0 to 0 dl 1631094014 ref 2 fl New:/0/ffffffff rc 0/-1 job:'chown.0' Lustre: 28146:0:(service.c:1378:ptlrpc_at_send_early_reply()) Skipped 335 previous similar messages LustreError: 21499:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0001_UUID lock: ffff8800935007c0/0x1765494a08cd796f lrc: 3/0,0 mode: PW/PW res: [0x300000400:0x718:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000000020020 nid: 0@lo remote: 0x1765494a08cd795a expref: 3429 pid: 32326 timeout: 4033 lvb_type: 0 LustreError: 25344:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802d0ca1300 x1710325427962368/t0(0) o104->lustre-OST0001@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 26200:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1631094017 with bad export cookie 1685834217842756110 Lustre: lustre-OST0001-osc-ffff88008c3a92a8: Connection to lustre-OST0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0001-osc-ffff88008c3a92a8: This client was evicted by lustre-OST0001; in progress operations using this service will fail. LustreError: 25344:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 114 previous similar messages LustreError: 25344:0:(ldlm_lockd.c:1425:ldlm_handle_enqueue0()) ### lock on destroyed export ffff8802782153d8 ns: filter-lustre-OST0001_UUID lock: ffff8802c4cf2d40/0x1765494a08d4be42 lrc: 3/0,0 mode: --/PW res: [0x300000400:0x783:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x50000000020000 nid: 0@lo remote: 0x1765494a08d4be2d expref: 3336 pid: 25344 timeout: 0 lvb_type: 0 LustreError: 6690:0:(service.c:2290:ptlrpc_server_handle_request()) @@@ Dropping timed-out request from 12345-0@lo: deadline 43/6s ago req@ffff88022460b240 x1710325425354112/t0(0) o1->1c8284fc-3353-43bb-bb2e-6d88a2482b3d@0@lo:317/0 lens 440/0 e 0 to 0 dl 1631094012 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'lfs.0' Lustre: 6688:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (43/6s); client may timeout req@ffff8800944e9f80 x1710325425354240/t0(0) o1->1c8284fc-3353-43bb-bb2e-6d88a2482b3d@0@lo:317/0 lens 440/0 e 0 to 0 dl 1631094012 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'lfs.0' LustreError: 6690:0:(service.c:2290:ptlrpc_server_handle_request()) Skipped 65 previous similar messages Lustre: 6705:0:(service.c:1378:ptlrpc_at_send_early_reply()) @@@ Already past deadline (-1s), not sending early reply. Consider increasing at_early_margin (5)? req@ffff88029ea970c0 x1710325425938048/t0(0) o2->16015095-046a-4c50-a5f2-954fef3e91d0@0@lo:323/0 lens 440/0 e 0 to 0 dl 1631094018 ref 2 fl New:/0/ffffffff rc 0/-1 job:'chown.0' Lustre: 6705:0:(service.c:1378:ptlrpc_at_send_early_reply()) Skipped 772 previous similar messages LustreError: 6690:0:(service.c:2290:ptlrpc_server_handle_request()) @@@ Dropping timed-out request from 12345-0@lo: deadline 43/7s ago req@ffff8802cfd26440 x1710325425393472/t0(0) o1->1c8284fc-3353-43bb-bb2e-6d88a2482b3d@0@lo:318/0 lens 440/0 e 0 to 0 dl 1631094013 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'lfs.0' LustreError: 6690:0:(service.c:2290:ptlrpc_server_handle_request()) Skipped 170 previous similar messages Lustre: 6690:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (43/7s); client may timeout req@ffff8802cfd26440 x1710325425393472/t0(0) o1->1c8284fc-3353-43bb-bb2e-6d88a2482b3d@0@lo:318/0 lens 440/0 e 0 to 0 dl 1631094013 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'lfs.0' Lustre: 6690:0:(service.c:2328:ptlrpc_server_handle_request()) Skipped 257 previous similar messages LustreError: 28866:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1631094021 with bad export cookie 1685834217842755879 LustreError: 28866:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) Skipped 2 previous similar messages LustreError: 6680:0:(service.c:2290:ptlrpc_server_handle_request()) @@@ Dropping timed-out request from 12345-0@lo: deadline 43/8s ago req@ffff88009f9e8040 x1710325425601024/t0(0) o2->1c8284fc-3353-43bb-bb2e-6d88a2482b3d@0@lo:319/0 lens 440/0 e 0 to 0 dl 1631094014 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'chown.0' Lustre: 6730:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (43/8s); client may timeout req@ffff88009f867700 x1710325425601152/t0(0) o2->1c8284fc-3353-43bb-bb2e-6d88a2482b3d@0@lo:319/0 lens 440/0 e 0 to 0 dl 1631094014 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'chown.0' Lustre: 6730:0:(service.c:2328:ptlrpc_server_handle_request()) Skipped 315 previous similar messages LustreError: 6680:0:(service.c:2290:ptlrpc_server_handle_request()) Skipped 372 previous similar messages LustreError: 30106:0:(service.c:2290:ptlrpc_server_handle_request()) @@@ Dropping timed-out request from 12345-0@lo: deadline 43/2s ago req@ffff880093779f80 x1710325426100672/t0(0) o2->16015095-046a-4c50-a5f2-954fef3e91d0@0@lo:327/0 lens 440/0 e 0 to 0 dl 1631094022 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'chmod.0' LustreError: 30106:0:(service.c:2290:ptlrpc_server_handle_request()) Skipped 1697 previous similar messages Lustre: 30106:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (43/2s); client may timeout req@ffff880093779f80 x1710325426100672/t0(0) o2->16015095-046a-4c50-a5f2-954fef3e91d0@0@lo:327/0 lens 440/0 e 0 to 0 dl 1631094022 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'chmod.0' Lustre: 30106:0:(service.c:2328:ptlrpc_server_handle_request()) Skipped 1737 previous similar messages LustreError: 5563:0:(ofd_obd.c:501:ofd_destroy_export()) lustre-OST0003: cli 16015095-046a-4c50-a5f2-954fef3e91d0/ffff8802650dc138 has 28672 pending on destroyed export LustreError: 5563:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_destroy_export: tot_granted 268069696 != fo_tot_granted 268098368 LustreError: 5563:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_destroy_export: tot_pending 896 != fo_tot_pending 29568 LustreError: 7391:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0003-osc-ffff880071ab1bf8: namespace resource [0x380000401:0x5df:0x0].0x0 (ffff8802ceb474c0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 7391:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 758 previous similar messages Lustre: lustre-OST0003-osc-ffff880071ab1bf8: Connection restored to 192.168.123.215@tcp (at 0@lo) LustreError: 22212:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 268067904 != fo_tot_granted 268096576 LustreError: 22212:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 28672 Lustre: 18307:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x240000405:0xb9:0x0]/ may get corrupted (rc -108) Lustre: 18315:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x200000403:0x249:0x0]/ may get corrupted (rc -108) LustreError: 22212:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 268067904 != fo_tot_granted 268096576 LustreError: 22212:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 1 previous similar message LustreError: 22212:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 28672 LustreError: 22212:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 1 previous similar message LustreError: 7782:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0000-osc-ffff88008c3a92a8: namespace resource [0x2c0000400:0x770:0x0].0x0 (ffff880090035940) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 7782:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 228 previous similar messages LustreError: 22212:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 268067904 != fo_tot_granted 268096576 LustreError: 22212:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 1 previous similar message LustreError: 22212:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 28672 LustreError: 22212:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 1 previous similar message Lustre: 18310:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff8802c0143880 x1710325428705920/t0(0) o10->lustre-OST0001-osc-ffff88008c3a92a8@0@lo:6/4 lens 440/432 e 0 to 0 dl 1631094085 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'truncate.0' LustreError: 22211:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 271242304 != fo_tot_granted 271270976 LustreError: 22211:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 1 previous similar message LustreError: 22211:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 28672 LustreError: 22211:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 1 previous similar message LustreError: 22212:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 272688192 != fo_tot_granted 272716864 LustreError: 22212:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 2 previous similar messages LustreError: 22212:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 2998272 != fo_tot_pending 3026944 LustreError: 22212:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 2 previous similar messages Lustre: 21510:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 503 < left 85580, rollback = 7 Lustre: 21510:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 98391 previous similar messages Lustre: 21510:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 1990/7960/0, destroy: 0/0/0 Lustre: 21510:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 98572 previous similar messages Lustre: 21510:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/1, xattr_set: 1992/15/0 Lustre: 21510:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 98574 previous similar messages Lustre: 21510:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 9951/85580/0, punch: 0/0/0, quota 14/174/8 Lustre: 21510:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 98573 previous similar messages Lustre: 21510:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 1990/33830/0, delete: 0/0/0 Lustre: 21510:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 98572 previous similar messages Lustre: 21510:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 21510:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 98571 previous similar messages Lustre: 18318:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1631093975/real 1631093975] req@ffff8802b8a97700 x1710325425912768/t0(0) o2->lustre-OST0000-osc-ffff880071ab1bf8@0@lo:28/4 lens 440/432 e 0 to 1 dl 1631094038 ref 1 fl Rpc:XQr/0/ffffffff rc 0/-1 job:'chown.0' Lustre: lustre-OST0000-osc-ffff880071ab1bf8: Connection to lustre-OST0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 2 previous similar messages Lustre: lustre-OST0002: Client lustre-MDT0002-mdtlov_UUID (at 0@lo) reconnecting Lustre: lustre-OST0001-osc-MDT0000: Connection restored to 192.168.123.215@tcp (at 0@lo) Lustre: Skipped 3 previous similar messages Lustre: lustre-OST0001: deleting orphan objects from 0x0:2693 to 0x0:2721 Lustre: lustre-OST0002: deleting orphan objects from 0x340000400:2642 to 0x340000400:2657 LustreError: 3122:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 269689920 != fo_tot_granted 269718592 LustreError: 3122:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 4 previous similar messages LustreError: 3122:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 28672 LustreError: 3122:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 4 previous similar messages Lustre: lustre-OST0003: deleting orphan objects from 0x380000400:2674 to 0x380000400:2689 Lustre: 18310:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff8802c0143880 x1710325431196800/t0(0) o10->lustre-OST0001-osc-ffff88008c3a92a8@0@lo:6/4 lens 440/432 e 0 to 0 dl 1631094152 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'truncate.0' LustreError: 22212:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 271322688 != fo_tot_granted 271351360 LustreError: 22212:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 11 previous similar messages LustreError: 22212:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 28672 LustreError: 22212:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 11 previous similar messages Lustre: 26563:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1631094068/real 1631094068] req@ffff88009f9fb880 x1710325429779136/t0(0) o104->lustre-OST0001@0@lo:15/16 lens 328/224 e 0 to 1 dl 1631094118 ref 1 fl Rpc:XQr/2/ffffffff rc -11/-1 job:'' Lustre: 26563:0:(client.c:2285:ptlrpc_expire_one_request()) Skipped 422 previous similar messages Lustre: 18314:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1631094069/real 1631094069] req@ffff8802c1732c00 x1710325429217856/t0(0) o1->lustre-OST0002-osc-ffff88008c3a92a8@0@lo:28/4 lens 440/432 e 0 to 1 dl 1631094113 ref 1 fl Rpc:XQr/2/ffffffff rc -11/-1 job:'lfs.0' Lustre: 18314:0:(client.c:2285:ptlrpc_expire_one_request()) Skipped 1 previous similar message Lustre: lustre-OST0002: Client 1c8284fc-3353-43bb-bb2e-6d88a2482b3d (at 0@lo) reconnecting Lustre: Skipped 6 previous similar messages Lustre: lustre-OST0002-osc-ffff88008c3a92a8: Connection restored to 192.168.123.215@tcp (at 0@lo) Lustre: Skipped 6 previous similar messages Lustre: 18314:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1631094071/real 1631094071] req@ffff880214da8680 x1710325429290624/t0(0) o1->lustre-OST0003-osc-ffff88008c3a92a8@0@lo:28/4 lens 440/432 e 0 to 1 dl 1631094123 ref 1 fl Rpc:XQr/2/ffffffff rc -11/-1 job:'lfs.0' Lustre: 18314:0:(client.c:2285:ptlrpc_expire_one_request()) Skipped 94 previous similar messages Lustre: lustre-OST0003: Client 1c8284fc-3353-43bb-bb2e-6d88a2482b3d (at 0@lo) reconnecting Lustre: 24334:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 500 < left 521, rollback = 2 Lustre: 24334:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 181 previous similar messages LustreError: 21499:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0002_UUID lock: ffff88009ee28040/0x1765494a08d7dbc4 lrc: 3/0,0 mode: PW/PW res: [0x340000400:0x937:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000000030020 nid: 0@lo remote: 0x1765494a08d7dbbd expref: 4017 pid: 6099 timeout: 4182 lvb_type: 0 LustreError: 21499:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 2 previous similar messages LustreError: 21497:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802cf79bec0 x1710325434611520/t0(0) o105->lustre-OST0002@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 11-0: lustre-OST0002-osc-ffff880071ab1bf8: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-OST0002-osc-ffff880071ab1bf8: Connection to lustre-OST0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 8 previous similar messages LustreError: 21497:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 27 previous similar messages LustreError: 167-0: lustre-OST0002-osc-ffff880071ab1bf8: This client was evicted by lustre-OST0002; in progress operations using this service will fail. LustreError: Skipped 2 previous similar messages LustreError: 22217:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 271028032 != fo_tot_granted 271056704 LustreError: 22217:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 21 previous similar messages LustreError: 22217:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 28672 LustreError: 22217:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 21 previous similar messages LustreError: 9871:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0002-osc-ffff880071ab1bf8: namespace resource [0x340000401:0x5cf:0x0].0x0 (ffff8802be16a840) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 9871:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 2003 previous similar messages Lustre: 7748:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1631094112/real 1631094112] req@ffff880090a39f80 x1710325432446464/t0(0) o101->lustre-MDT0001-mdc-ffff880071ab1bf8@0@lo:12/10 lens 576/40248 e 0 to 1 dl 1631094181 ref 2 fl Rpc:XQr/2/ffffffff rc 0/-1 job:'file_exec.sh.0' Lustre: 7748:0:(client.c:2285:ptlrpc_expire_one_request()) Skipped 30 previous similar messages Lustre: lustre-MDT0001: Client 16015095-046a-4c50-a5f2-954fef3e91d0 (at 0@lo) reconnecting Lustre: 18318:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff8802bbbe6a80 x1710325436037632/t0(0) o10->lustre-OST0001-osc-ffff880071ab1bf8@0@lo:6/4 lens 440/432 e 0 to 0 dl 1631094270 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'truncate.0' Lustre: dir [0x240000404:0x27c:0x0] stripe 1 readdir failed: -2, directory is partially accessed! LustreError: 21499:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0001_UUID lock: ffff8802ae296d00/0x1765494a08dafa73 lrc: 4/0,0 mode: PR/PR res: [0x240000404:0x207:0x0].0x0 bits 0x1b/0x0 rrc: 4 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0x1765494a08dafa3b expref: 112 pid: 24725 timeout: 4269 lvb_type: 0 LustreError: 26566:0:(ldlm_lockd.c:1425:ldlm_handle_enqueue0()) ### lock on destroyed export ffff88009a406fc8 ns: mdt-lustre-MDT0001_UUID lock: ffff8802bb3dcf00/0x1765494a08de6da0 lrc: 3/0,0 mode: PR/PR res: [0x240000403:0x1:0x0].0x0 bits 0x12/0x0 rrc: 5 type: IBT gid 0 flags: 0x50200000000000 nid: 0@lo remote: 0x1765494a08de6d92 expref: 5 pid: 26566 timeout: 0 lvb_type: 0 LustreError: 11-0: lustre-MDT0001-mdc-ffff880071ab1bf8: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: 167-0: lustre-MDT0001-mdc-ffff880071ab1bf8: This client was evicted by lustre-MDT0001; in progress operations using this service will fail. LustreError: 10290:0:(file.c:5042:ll_inode_revalidate_fini()) lustre: revalidate FID [0x240000403:0x1:0x0] error: rc = -5 LustreError: 13299:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff880071ab1bf8: inode [0x240000405:0x125:0x0] mdc close failed: rc = -108 LustreError: 9134:0:(mdc_request.c:1436:mdc_read_page()) lustre-MDT0001-mdc-ffff880071ab1bf8: [0x240000403:0x1:0x0] lock enqueue fails: rc = -108 Lustre: lustre-MDT0001-mdc-ffff880071ab1bf8: Connection restored to 192.168.123.215@tcp (at 0@lo) Lustre: Skipped 3 previous similar messages LustreError: 30361:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff880090a3a5c0 x1710325439721664/t0(0) o105->lustre-OST0003@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 30361:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 15 previous similar messages Lustre: 18316:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x200000404:0x345:0x0]// may get corrupted (rc -5) LustreError: 13471:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0003-osc-ffff88008c3a92a8: namespace resource [0xdd5:0x0:0x0].0x0 (ffff8802a1faa5c0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 13471:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 119 previous similar messages Lustre: 8784:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 6s req@ffff8802b1d3c500 x1710325439936256/t0(0) o101->1c8284fc-3353-43bb-bb2e-6d88a2482b3d@0@lo:0/0 lens 328/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'cp.0' LustreError: 22210:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 26671808 != fo_tot_granted 26700480 LustreError: 22210:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 51 previous similar messages LustreError: 22210:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 28672 != fo_tot_pending 57344 LustreError: 22210:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 51 previous similar messages ptlrpc_watchdog_fire: 10 callbacks suppressed Lustre: mdt01_008: service thread pid 29080 was inactive for 92.264 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 29080, comm: mdt01_008 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] mdt_object_local_lock+0x527/0xb90 [mdt] [<0>] mdt_object_lock_internal+0x70/0x390 [mdt] [<0>] mdt_getattr_name_lock+0xaca/0x2590 [mdt] [<0>] mdt_intent_getattr+0x2d5/0x4b0 [mdt] [<0>] mdt_intent_opc+0x1dd/0xc10 [mdt] [<0>] mdt_intent_policy+0x1a1/0x360 [mdt] [<0>] ldlm_lock_enqueue+0x3c2/0xb40 [ptlrpc] [<0>] ldlm_handle_enqueue0+0xaa6/0x16d0 [ptlrpc] [<0>] tgt_enqueue+0x64/0x240 [ptlrpc] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: mdt04_009: service thread pid 24813 was inactive for 89.250 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 24813, comm: mdt04_009 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] mdt_rename_lock+0x24b/0x4d0 [mdt] [<0>] mdt_reint_migrate+0x6e8/0x11d0 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x76c/0xb40 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe LustreError: 11-0: lustre-MDT0002-mdc-ffff88008c3a92a8: operation mds_close to node 0@lo failed: rc = -107 LustreError: Skipped 1 previous similar message Lustre: lustre-MDT0002-mdc-ffff88008c3a92a8: Connection to lustre-MDT0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 3 previous similar messages LustreError: 9815:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff88008c3a92a8: inode [0x280000404:0x317:0x0] mdc close failed: rc = -108 LustreError: 9815:0:(file.c:234:ll_close_inode_openhandle()) Skipped 6 previous similar messages LustreError: 11043:0:(llite_lib.c:2983:ll_prep_inode()) new_inode -fatal: rc -108 LustreError: 11462:0:(file.c:5042:ll_inode_revalidate_fini()) lustre: revalidate FID [0x280000403:0x1:0x0] error: rc = -108 LustreError: 12692:0:(mdc_request.c:1436:mdc_read_page()) lustre-MDT0002-mdc-ffff88008c3a92a8: [0x280000403:0x2a5:0x0] lock enqueue fails: rc = -108 LustreError: 12692:0:(mdc_request.c:1436:mdc_read_page()) Skipped 9 previous similar messages Lustre: dir [0x280000403:0x2a0:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: dir [0x240000405:0x75:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: dir [0x200000403:0x240:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 5 previous similar messages Lustre: 26713:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0x247:0x0] with magic=0xbd60bd0 Lustre: 26713:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 1 previous similar message LustreError: 2085:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8800975325c0 x1710325443205248/t0(0) o106->lustre-OST0002@0@lo:15/16 lens 328/280 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 2085:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 1 previous similar message LustreError: 21499:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0003_UUID lock: ffff8802aa90f840/0x1765494a08e53610 lrc: 3/0,0 mode: PW/PW res: [0x380000400:0xae3:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 2097152->18446744073709551615) gid 0 flags: 0x60000400020020 nid: 0@lo remote: 0x1765494a08e53609 expref: 3414 pid: 2085 timeout: 4397 lvb_type: 0 LustreError: 21499:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 3 previous similar messages Lustre: 22265:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0x269:0x0] with magic=0xbd60bd0 Lustre: 22265:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 1 previous similar message LustreError: 167-0: lustre-OST0003-osc-ffff880071ab1bf8: This client was evicted by lustre-OST0003; in progress operations using this service will fail. LustreError: Skipped 3 previous similar messages Lustre: 3625:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 6s req@ffff8802c3c470c0 x1710325443262208/t0(0) o101->16015095-046a-4c50-a5f2-954fef3e91d0@0@lo:0/0 lens 4616/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'touch.0' Lustre: 18314:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x200000403:0x211:0x0]/ may get corrupted (rc -108) Lustre: 18314:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x200000404:0x3b8:0x0]/ may get corrupted (rc -108) LustreError: 5563:0:(ofd_obd.c:501:ofd_destroy_export()) lustre-OST0002: cli 1c8284fc-3353-43bb-bb2e-6d88a2482b3d/ffff8802a60f53d8 has 28672 pending on destroyed export Lustre: 21556:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 6s req@ffff8802ceb16440 x1710325443375616/t0(0) o101->lustre-MDT0000-mdtlov_UUID@0@lo:0/0 lens 328/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'mdt04_012.0' Lustre: 21556:0:(service.c:2157:ptlrpc_server_handle_req_in()) Skipped 1 previous similar message Lustre: 22265:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0x26d:0x0] with magic=0xbd60bd0 Lustre: 22265:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 1 previous similar message LustreError: 16315:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0002-osc-ffff88008c3a92a8: namespace resource [0x340000401:0x83f:0x0].0x0 (ffff8802b13ffc40) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 16315:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 3 previous similar messages Lustre: lustre-OST0002-osc-ffff88008c3a92a8: Connection restored to 192.168.123.215@tcp (at 0@lo) Lustre: Skipped 2 previous similar messages Lustre: 27721:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0x1cf:0x0] with magic=0xbd60bd0 Lustre: 27721:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 3 previous similar messages Lustre: 18315:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff8802aed470c0 x1710325444388032/t0(0) o10->lustre-OST0003-osc-ffff880071ab1bf8@0@lo:6/4 lens 440/432 e 0 to 0 dl 1631094433 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'truncate.0' Lustre: 18315:0:(client.c:1485:after_reply()) Skipped 3 previous similar messages Lustre: 30255:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 6s req@ffff880296e28040 x1710325444385024/t0(0) o106->LOV_OSC_UUID@0@lo:0/0 lens 328/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'' LustreError: 8361:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802b0593880 x1710325447537408/t0(0) o106->lustre-OST0000@0@lo:15/16 lens 328/280 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 21477:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1631094507 with bad export cookie 1685834217844686836 LustreError: 21477:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) Skipped 2 previous similar messages LustreError: 8361:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 38 previous similar messages LustreError: 11-0: lustre-OST0000-osc-ffff88008c3a92a8: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 3 previous similar messages LustreError: 18626:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0000-osc-ffff88008c3a92a8: namespace resource [0x2c0000401:0x81f:0x0].0x0 (ffff8802bba01bc0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 18626:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 45 previous similar messages LustreError: 2078:0:(service.c:2290:ptlrpc_server_handle_request()) @@@ Dropping timed-out request from 12345-0@lo: deadline 21/12s ago req@ffff8802c9f592f8 x1710325447484416/t0(0) o101->1c8284fc-3353-43bb-bb2e-6d88a2482b3d@0@lo:77/0 lens 328/0 e 0 to 0 dl 1631094527 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'chmod.0' Lustre: 2078:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (21/12s); client may timeout req@ffff8802c9f592f8 x1710325447484416/t0(0) o101->1c8284fc-3353-43bb-bb2e-6d88a2482b3d@0@lo:77/0 lens 328/0 e 0 to 0 dl 1631094527 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'chmod.0' LustreError: 22221:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 236193088 != fo_tot_granted 236221760 LustreError: 22221:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 135 previous similar messages LustreError: 22221:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 28672 LustreError: 22221:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 135 previous similar messages Lustre: 25087:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 476/1904/0, destroy: 1/4/0 Lustre: 25087:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 166891 previous similar messages Lustre: 25087:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 5/5/0, xattr_set: 484/283/0 Lustre: 25087:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 166891 previous similar messages Lustre: 25087:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 2381/20478/0, punch: 0/0/0, quota 6/6/0 Lustre: 25087:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 166890 previous similar messages Lustre: 25087:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 478/8124/0, delete: 3/6/1 Lustre: 25087:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 166891 previous similar messages Lustre: 25087:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 1/1/0, ref_del: 1/1/0 Lustre: 25087:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 166891 previous similar messages Lustre: 25087:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 507 < left 20478, rollback = 7 Lustre: 25087:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 166700 previous similar messages Lustre: lustre-OST0001-osc-ffff880071ab1bf8: Connection to lustre-OST0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 3 previous similar messages LustreError: 2265:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1631094626 with bad export cookie 1685834217845267542 LustreError: 2265:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) Skipped 4 previous similar messages LustreError: 21499:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0000_UUID lock: ffff880094466940/0x1765494a08f04129 lrc: 3/0,0 mode: PR/PR res: [0x200000403:0x365:0x0].0x0 bits 0x12/0x0 rrc: 3 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0x1765494a08f0406c expref: 176 pid: 27738 timeout: 4678 lvb_type: 0 LustreError: 21499:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 3 previous similar messages LustreError: 11-0: lustre-MDT0000-mdc-ffff88008c3a92a8: operation mds_close to node 0@lo failed: rc = -107 LustreError: Skipped 20 previous similar messages LustreError: 167-0: lustre-MDT0000-mdc-ffff88008c3a92a8: This client was evicted by lustre-MDT0000; in progress operations using this service will fail. LustreError: Skipped 3 previous similar messages LustreError: 18481:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff88008c3a92a8: inode [0x200000404:0x4dc:0x0] mdc close failed: rc = -108 LustreError: 18481:0:(file.c:234:ll_close_inode_openhandle()) Skipped 17 previous similar messages LustreError: 18675:0:(llite_lib.c:2983:ll_prep_inode()) new_inode -fatal: rc -108 LustreError: 20346:0:(file.c:5042:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000007:0x1:0x0] error: rc = -108 LustreError: 20334:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-MDT0000-mdc-ffff88008c3a92a8: namespace resource [0x200000404:0x4fd:0x0].0x0 (ffff8800935f82c0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 20334:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 421 previous similar messages Lustre: lustre-MDT0000-mdc-ffff88008c3a92a8: Connection restored to 192.168.123.215@tcp (at 0@lo) Lustre: Skipped 4 previous similar messages Lustre: 27790:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 496 < left 1027, rollback = 2 Lustre: 27790:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 184 previous similar messages LustreError: 24777:0:(ldlm_lockd.c:1425:ldlm_handle_enqueue0()) ### lock on destroyed export ffff880290bd6678 ns: mdt-lustre-MDT0001_UUID lock: ffff8802d8a134c0/0x1765494a08f0f603 lrc: 3/0,0 mode: PR/PR res: [0x240000405:0x11f:0x0].0x0 bits 0x12/0x0 rrc: 11 type: IBT gid 0 flags: 0x50200000000000 nid: 0@lo remote: 0x1765494a08f0e4a6 expref: 16 pid: 24777 timeout: 0 lvb_type: 0 LustreError: 16525:0:(file.c:5042:ll_inode_revalidate_fini()) lustre: revalidate FID [0x240000405:0x11f:0x0] error: rc = -5 LustreError: 19346:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -108 LustreError: 19847:0:(mdc_request.c:1436:mdc_read_page()) lustre-MDT0001-mdc-ffff88008c3a92a8: [0x240000401:0xd:0x0] lock enqueue fails: rc = -108 LustreError: 19847:0:(mdc_request.c:1436:mdc_read_page()) Skipped 13 previous similar messages Lustre: dir [0x200000404:0x4fd:0x0] stripe 1 readdir failed: -108, directory is partially accessed! LustreError: 16525:0:(file.c:5042:ll_inode_revalidate_fini()) Skipped 61 previous similar messages LustreError: 25087:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0001: '19' migration was interrupted, run 'lfs migrate -m 2 -c 1 -H crush 19' to finish migration. LustreError: 21683:0:(service.c:2290:ptlrpc_server_handle_request()) @@@ Dropping timed-out request from 12345-0@lo: deadline 33/1s ago req@ffff88008f17f018 x1710325455158848/t0(0) o1->1c8284fc-3353-43bb-bb2e-6d88a2482b3d@0@lo:286/0 lens 440/0 e 0 to 0 dl 1631094736 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'lfs.0' Lustre: 21679:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (33/1s); client may timeout req@ffff88022343f700 x1710325455159104/t0(0) o1->1c8284fc-3353-43bb-bb2e-6d88a2482b3d@0@lo:286/0 lens 440/0 e 0 to 0 dl 1631094736 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'lfs.0' LustreError: 21683:0:(service.c:2290:ptlrpc_server_handle_request()) Skipped 2 previous similar messages Lustre: ll_ost01_017: service thread pid 32330 was inactive for 42.009 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 32330, comm: ll_ost01_017 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] tgt_extent_lock+0xea/0x2a0 [ptlrpc] [<0>] ofd_getattr_hdl+0x365/0x720 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 8775, comm: ll_ost01_038 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] tgt_extent_lock+0xea/0x2a0 [ptlrpc] [<0>] ofd_getattr_hdl+0x365/0x720 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 4074, comm: ll_ost01_029 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] tgt_extent_lock+0xea/0x2a0 [ptlrpc] [<0>] ofd_getattr_hdl+0x365/0x720 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: ll_ost01_048: service thread pid 20453 was inactive for 42.189 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. LustreError: 27175:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802cc60e440 x1710325458329984/t0(0) o105->lustre-OST0003@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 27175:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 71 previous similar messages Lustre: 18322:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x200000403:0x48e:0x0]/ may get corrupted (rc -108) Lustre: 18321:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x240000406:0xe2:0x0]// may get corrupted (rc -108) Lustre: 18321:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x200000403:0x3fe:0x0]// may get corrupted (rc -108) Lustre: 27337:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (56/21s); client may timeout req@ffff88028b68e6c8 x1710325455036288/t0(0) o1->1c8284fc-3353-43bb-bb2e-6d88a2482b3d@0@lo:302/0 lens 440/432 e 1 to 0 dl 1631094752 ref 1 fl Complete:/0/0 rc 0/0 job:'lfs.0' Lustre: 27337:0:(service.c:2328:ptlrpc_server_handle_request()) Skipped 47 previous similar messages Lustre: 27247:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 6s req@ffff8800a24dbec0 x1710325461693056/t0(0) o35->16015095-046a-4c50-a5f2-954fef3e91d0@0@lo:0/0 lens 392/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'touch.0' LustreError: 32351:0:(ldlm_lockd.c:1425:ldlm_handle_enqueue0()) ### lock on destroyed export ffff8802782137e8 ns: filter-lustre-OST0002_UUID lock: ffff8802bf174000/0x1765494a0901965e lrc: 3/0,0 mode: --/PW res: [0x340000400:0x12af:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x50000000020000 nid: 0@lo remote: 0x1765494a09019657 expref: 6252 pid: 32351 timeout: 0 lvb_type: 0 LustreError: 32351:0:(ldlm_lockd.c:1425:ldlm_handle_enqueue0()) Skipped 7 previous similar messages Lustre: 18309:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x200000405:0x8f:0x0]// may get corrupted (rc -108) Lustre: 18310:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x280000405:0x224:0x0]/ may get corrupted (rc -108) Lustre: 18318:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff8802c33bc500 x1710325462241408/t0(0) o10->lustre-OST0003-osc-ffff88008c3a92a8@0@lo:6/4 lens 440/432 e 0 to 0 dl 1631094956 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'truncate.0' Lustre: 18318:0:(client.c:1485:after_reply()) Skipped 2 previous similar messages Lustre: 9316:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 9s req@ffff8802b67a4500 x1710325463289280/t0(0) o103->lustre-MDT0001-mdtlov_UUID@0@lo:0/0 lens 328/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'jbd2/dm-1-8.0' Lustre: dir [0x200000405:0x3c:0x0] stripe 1 readdir failed: -2, directory is partially accessed! LustreError: 21798:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0000-osp-MDT0001: fail to cancel 1 llog-records: rc = -116 LustreError: 21798:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0000-osp-MDT0001: fail to cancel 1 of 1 llog-records: rc = -116 LustreError: 21798:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0000-osp-MDT0001: fail to cancel 1 llog-records: rc = -2 LustreError: 21798:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0000-osp-MDT0001: fail to cancel 1 of 1 llog-records: rc = -2 Lustre: 27738:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000405:0x212:0x0] with magic=0xbd60bd0 Lustre: 27738:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 19 previous similar messages LustreError: 11-0: lustre-OST0003-osc-ffff880071ab1bf8: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: 21473:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1631095030 with bad export cookie 1685834217847157346 LustreError: Skipped 263 previous similar messages Lustre: 18316:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x200000403:0x55a:0x0]/ may get corrupted (rc -108) ptlrpc_watchdog_fire: 89 callbacks suppressed Lustre: mdt01_011: service thread pid 25124 was inactive for 106.263 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Lustre: Skipped 2 previous similar messages Pid: 25124, comm: mdt01_011 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_fini+0xa25/0xec0 [ptlrpc] [<0>] ldlm_cli_enqueue+0x436/0xa30 [ptlrpc] [<0>] osp_md_object_lock+0x160/0x300 [osp] [<0>] lod_object_lock+0xdb/0x7d0 [lod] [<0>] mdd_object_lock+0x2d/0xd0 [mdd] [<0>] mdt_remote_object_lock_try+0x1cc/0x510 [mdt] [<0>] mdt_remote_object_lock+0x2a/0x30 [mdt] [<0>] mdt_revoke_remote_lookup_lock+0x86/0x150 [mdt] [<0>] mdt_link_parents_lock.isra.52+0x6b1/0x15f0 [mdt] [<0>] mdt_reint_migrate+0x930/0x11d0 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x76c/0xb40 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe LustreError: 22213:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 153599936 != fo_tot_granted 153628608 LustreError: 22213:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 398 previous similar messages LustreError: 22213:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 28672 LustreError: 22213:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 398 previous similar messages Lustre: mdt01_002: service thread pid 21511 was inactive for 106.245 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 21511, comm: mdt01_002 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] mdt_rename_lock+0x24b/0x4d0 [mdt] [<0>] mdt_reint_migrate+0x6e8/0x11d0 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x76c/0xb40 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 25155, comm: mdt01_013 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Lustre: mdt01_007: service thread pid 25087 was inactive for 106.150 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 88 previous similar messages Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] mdt_object_local_lock+0x527/0xb90 [mdt] [<0>] mdt_object_lock_internal+0x70/0x390 [mdt] [<0>] mdt_getattr_name_lock+0xaca/0x2590 [mdt] [<0>] mdt_intent_getattr+0x2d5/0x4b0 [mdt] [<0>] mdt_intent_opc+0x1dd/0xc10 [mdt] [<0>] mdt_intent_policy+0x1a1/0x360 [mdt] [<0>] ldlm_lock_enqueue+0x3c2/0xb40 [ptlrpc] [<0>] ldlm_handle_enqueue0+0xaa6/0x16d0 [ptlrpc] [<0>] tgt_enqueue+0x64/0x240 [ptlrpc] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe LustreError: 25879:0:(mdc_request.c:1436:mdc_read_page()) lustre-MDT0002-mdc-ffff88008c3a92a8: [0x280000403:0x1:0x0] lock enqueue fails: rc = -108 LustreError: 25765:0:(file.c:5042:ll_inode_revalidate_fini()) lustre: revalidate FID [0x240000404:0x3bb:0x0] error: rc = -5 LustreError: 25765:0:(file.c:5042:ll_inode_revalidate_fini()) Skipped 8 previous similar messages LustreError: 25879:0:(mdc_request.c:1436:mdc_read_page()) Skipped 20 previous similar messages LustreError: 26721:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff88008c3a92a8: inode [0x280000405:0x1de:0x0] mdc close failed: rc = -108 LustreError: 26721:0:(file.c:234:ll_close_inode_openhandle()) Skipped 22 previous similar messages LustreError: 26721:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-MDT0002-mdc-ffff88008c3a92a8: namespace resource [0x280000403:0x1:0x0].0x0 (ffff880099bda0c0) refcount nonzero (2) after lock cleanup; forcing cleanup. LustreError: 26721:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 1315 previous similar messages LustreError: 23079:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0002: '2' migration was interrupted, run 'lfs migrate -m 0 -c 1 -H crush 2' to finish migration. LustreError: 25168:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0000: '3' migration was interrupted, run 'lfs migrate -m 2 -c 1 -H crush 3' to finish migration. Lustre: 24813:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000405:0x215:0x0] with magic=0xbd60bd0 Lustre: 24813:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 1 previous similar message Lustre: 21519:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 87/348/0, destroy: 0/0/0 Lustre: 21519:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 213503 previous similar messages Lustre: 21519:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/1, xattr_set: 89/15/0 Lustre: 21519:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 213502 previous similar messages Lustre: 21519:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 436/3751/0, punch: 0/0/0, quota 14/142/9 Lustre: 21519:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 213503 previous similar messages Lustre: 21519:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 87/1479/0, delete: 0/0/0 Lustre: 21519:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 213502 previous similar messages Lustre: 21519:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 21519:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 213501 previous similar messages Lustre: 21519:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 502 < left 3751, rollback = 7 Lustre: 21519:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 213280 previous similar messages LustreError: 21499:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 101s: evicting client at 0@lo ns: filter-lustre-OST0002_UUID lock: ffff8802d08dc3c0/0x1765494a0908a5bc lrc: 3/0,0 mode: PW/PW res: [0x340000400:0x181a:0x0].0x0 rrc: 4 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000080020020 nid: 0@lo remote: 0x1765494a0908a584 expref: 2562 pid: 25344 timeout: 5223 lvb_type: 0 LustreError: 21499:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 7 previous similar messages Lustre: lustre-OST0002-osc-ffff88008c3a92a8: Connection to lustre-OST0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 9 previous similar messages LustreError: 167-0: lustre-OST0002-osc-ffff88008c3a92a8: This client was evicted by lustre-OST0002; in progress operations using this service will fail. LustreError: Skipped 7 previous similar messages Lustre: lustre-OST0002-osc-ffff88008c3a92a8: Connection restored to 192.168.123.215@tcp (at 0@lo) Lustre: Skipped 7 previous similar messages Lustre: 17176:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 6s req@ffff8802b215a598 x1710325479499776/t0(0) o105->LOV_OSC_UUID@0@lo:0/0 lens 392/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'' Lustre: 17176:0:(service.c:2157:ptlrpc_server_handle_req_in()) Skipped 14 previous similar messages Lustre: 21513:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 496 < left 671, rollback = 2 Lustre: 21513:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 286 previous similar messages LustreError: 6724:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802d2ef5e00 x1710325481948672/t0(0) o104->lustre-OST0000@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 6724:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 112 previous similar messages Lustre: 18321:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x280000406:0xc1:0x0]/ may get corrupted (rc -108) LustreError: 21510:0:(ldlm_lockd.c:1425:ldlm_handle_enqueue0()) ### lock on destroyed export ffff8802c169b7e8 ns: mdt-lustre-MDT0001_UUID lock: ffff88009ef34000/0x1765494a09169419 lrc: 3/0,0 mode: PR/PR res: [0x240000406:0x54:0x0].0x0 bits 0x12/0x0 rrc: 19 type: IBT gid 0 flags: 0x50200000000000 nid: 0@lo remote: 0x1765494a0916936a expref: 16 pid: 21510 timeout: 0 lvb_type: 0 LustreError: 26179:0:(file.c:5042:ll_inode_revalidate_fini()) lustre: revalidate FID [0x240000406:0x54:0x0] error: rc = -5 LustreError: 26179:0:(file.c:5042:ll_inode_revalidate_fini()) Skipped 2 previous similar messages LustreError: 28565:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff88008c3a92a8: inode [0x240000406:0x54:0x0] mdc close failed: rc = -108 LustreError: 28565:0:(file.c:234:ll_close_inode_openhandle()) Skipped 10 previous similar messages Lustre: 27721:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000405:0x2af:0x0] with magic=0xbd60bd0 Lustre: 27721:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 23 previous similar messages Lustre: dir [0x200000403:0x506:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 1 previous similar message Lustre: 18322:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff88029ea63880 x1710325486842496/t0(0) o10->lustre-OST0003-osc-ffff88008c3a92a8@0@lo:6/4 lens 440/432 e 0 to 0 dl 1631095520 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dd.0' Lustre: 18322:0:(client.c:1485:after_reply()) Skipped 2 previous similar messages LustreError: 30775:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1631095452 with bad export cookie 1685834217845718538 LustreError: 16784:0:(ofd_obd.c:501:ofd_destroy_export()) lustre-OST0003: cli 1c8284fc-3353-43bb-bb2e-6d88a2482b3d/ffff8802abe28958 has 28672 pending on destroyed export Lustre: 18313:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x280000406:0x12b:0x0]// may get corrupted (rc -108) Lustre: 18314:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x200000403:0x631:0x0]/ may get corrupted (rc -108) Lustre: 18313:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x200000405:0x259:0x0]/ may get corrupted (rc -108) Lustre: 18310:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x280000406:0xdb:0x0]// may get corrupted (rc -108) ptlrpc_watchdog_fire: 2 callbacks suppressed Lustre: ll_ost06_046: service thread pid 18714 was inactive for 40.032 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Lustre: Skipped 1 previous similar message Pid: 18714, comm: ll_ost06_046 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] tgt_extent_lock+0xea/0x2a0 [ptlrpc] [<0>] ofd_getattr_hdl+0x365/0x720 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 9645, comm: ll_ost06_024 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] tgt_extent_lock+0xea/0x2a0 [ptlrpc] [<0>] ofd_getattr_hdl+0x365/0x720 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 9654, comm: ll_ost06_028 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] tgt_extent_lock+0xea/0x2a0 [ptlrpc] [<0>] ofd_getattr_hdl+0x365/0x720 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: ll_ost06_042: service thread pid 18693 was inactive for 40.933 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 1 previous similar message LustreError: 25019:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0001: failed to get lu_attr of [0x240000408:0x23:0x0]: rc = -2 LustreError: 2335:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1631095542 with bad export cookie 1685834217849066841 LustreError: 2335:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) Skipped 2 previous similar messages LustreError: 11-0: lustre-OST0002-osc-ffff880071ab1bf8: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 14 previous similar messages Lustre: 25939:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 7s req@ffff8802c82b8058 x1710325490877184/t0(0) o104->LOV_OSC_UUID@0@lo:0/0 lens 328/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'' Lustre: 25939:0:(service.c:2157:ptlrpc_server_handle_req_in()) Skipped 1 previous similar message Lustre: dir [0x240000408:0x19:0x0] stripe 1 readdir failed: -2, directory is partially accessed! LustreError: 1675:0:(ofd_obd.c:501:ofd_destroy_export()) lustre-OST0001: cli 16015095-046a-4c50-a5f2-954fef3e91d0/ffff8802a9b06678 has 28672 pending on destroyed export Lustre: 18317:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x280000406:0xc8:0x0]/ may get corrupted (rc -108) Lustre: 18318:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x240000406:0x352:0x0]/ may get corrupted (rc -108) Lustre: 18312:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x200000405:0x1f6:0x0]// may get corrupted (rc -108) LustreError: 1675:0:(ofd_obd.c:501:ofd_destroy_export()) lustre-OST0000: cli 16015095-046a-4c50-a5f2-954fef3e91d0/ffff880292196fc8 has 28672 pending on destroyed export LustreError: 25055:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0002: '8' migration was interrupted, run 'lfs migrate -m 0 -c 1 -H crush 8' to finish migration. LustreError: 22218:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 152056640 != fo_tot_granted 152085312 LustreError: 22218:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 552 previous similar messages LustreError: 22218:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 28672 != fo_tot_pending 57344 LustreError: 22218:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 552 previous similar messages LustreError: 21538:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0001: failed to get lu_attr of [0x240000407:0x144:0x0]: rc = -2 LustreError: 1675:0:(ofd_obd.c:501:ofd_destroy_export()) lustre-OST0003: cli 1c8284fc-3353-43bb-bb2e-6d88a2482b3d/ffff8802264fca88 has 61440 pending on destroyed export Lustre: ost: This server is not able to keep up with request traffic (cpu-bound). Lustre: 26618:0:(service.c:1612:ptlrpc_at_check_timed()) earlyQ=0 reqQ=0 recA=1, svcEst=49, delay=5849ms Lustre: 18321:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x200000405:0x1f6:0x0]// may get corrupted (rc -108) Lustre: 18322:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x200000403:0x746:0x0]// may get corrupted (rc -108) Lustre: 18321:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x200000405:0x316:0x0]/ may get corrupted (rc -108) Lustre: 18322:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x200000403:0x72a:0x0]// may get corrupted (rc -108) LustreError: 3545:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0003-osc-ffff88008c3a92a8: namespace resource [0x380000400:0x1f20:0x0].0x0 (ffff8802daea56c0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 3545:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 2914 previous similar messages Lustre: ll_ost02_022: service thread pid 32735 was inactive for 98.020 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 25 previous similar messages LustreError: 21484:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1631095778 with bad export cookie 1685834217849732807 LustreError: 21484:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) Skipped 4 previous similar messages LustreError: 24917:0:(ldlm_lockd.c:1425:ldlm_handle_enqueue0()) ### lock on destroyed export ffff88028e5737e8 ns: mdt-lustre-MDT0000_UUID lock: ffff8802cf04c000/0x1765494a092edb40 lrc: 3/0,0 mode: CR/CR res: [0x200000405:0x2d7:0x0].0x0 bits 0xa/0x0 rrc: 4 type: IBT gid 0 flags: 0x50200400000000 nid: 0@lo remote: 0x1765494a092edb1d expref: 195 pid: 24917 timeout: 0 lvb_type: 0 LustreError: 24917:0:(ldlm_lockd.c:1425:ldlm_handle_enqueue0()) Skipped 6 previous similar messages LustreError: 2693:0:(file.c:5042:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000403:0x741:0x0] error: rc = -5 LustreError: 4766:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -5 LustreError: 1158:0:(mdc_request.c:1436:mdc_read_page()) lustre-MDT0000-mdc-ffff880071ab1bf8: [0x200000400:0x28:0x0] lock enqueue fails: rc = -108 LustreError: 1158:0:(mdc_request.c:1436:mdc_read_page()) Skipped 6 previous similar messages LustreError: 2693:0:(file.c:5042:ll_inode_revalidate_fini()) Skipped 13 previous similar messages LustreError: 4958:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -108 LustreError: 4958:0:(llite_lib.c:1836:ll_md_setattr()) Skipped 3 previous similar messages LustreError: 3039:0:(mdc_request.c:1436:mdc_read_page()) lustre-MDT0000-mdc-ffff880071ab1bf8: [0x200000400:0x25:0x0] lock enqueue fails: rc = -108 LustreError: 23589:0:(vvp_io.c:1793:vvp_io_init()) lustre: refresh file layout [0x200000405:0x2d7:0x0] error -108. Lustre: 24261:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 509 < left 32776, rollback = 7 Lustre: 24261:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 291182 previous similar messages Lustre: 24261:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 762/3048/0, destroy: 1/4/0 Lustre: 24261:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 291582 previous similar messages Lustre: 24261:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 3/3/1, xattr_set: 766/148/0 Lustre: 24261:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 291584 previous similar messages Lustre: 24261:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 3811/32776/0, punch: 0/0/0, quota 10/74/0 Lustre: 24261:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 291587 previous similar messages Lustre: 24261:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 763/12970/0, delete: 2/5/1 Lustre: 24261:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 291586 previous similar messages Lustre: 24261:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 1/1/0, ref_del: 2/2/1 Lustre: 24261:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 291582 previous similar messages Lustre: dir [0x200000403:0x701:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 10 previous similar messages LustreError: 21499:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 154s: evicting client at 0@lo ns: filter-lustre-OST0001_UUID lock: ffff8802d5388040/0x1765494a0925e81e lrc: 3/0,0 mode: PW/PW res: [0x300000401:0x2eb6:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->786431) gid 0 flags: 0x60000400020020 nid: 0@lo remote: 0x1765494a0925e817 expref: 2502 pid: 25488 timeout: 5890 lvb_type: 0 LustreError: 21499:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 27 previous similar messages Lustre: lustre-OST0001-osc-ffff88008c3a92a8: Connection to lustre-OST0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 12 previous similar messages LustreError: 21535:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000403:0x506:0x0]: rc = -2 LustreError: 167-0: lustre-OST0001-osc-ffff88008c3a92a8: This client was evicted by lustre-OST0001; in progress operations using this service will fail. LustreError: Skipped 12 previous similar messages Lustre: 18307:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x200000403:0x708:0x0]// may get corrupted (rc -5) LustreError: 1181:0:(ofd_obd.c:501:ofd_destroy_export()) lustre-OST0001: cli 1c8284fc-3353-43bb-bb2e-6d88a2482b3d/ffff8802c82ba548 has 57344 pending on destroyed export LustreError: 27185:0:(ofd_obd.c:501:ofd_destroy_export()) lustre-OST0002: cli 1c8284fc-3353-43bb-bb2e-6d88a2482b3d/ffff8802c1e14138 has 61440 pending on destroyed export Lustre: 18320:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x200000405:0x386:0x0]/ may get corrupted (rc -108) Lustre: lustre-OST0001-osc-ffff88008c3a92a8: Connection restored to 192.168.123.215@tcp (at 0@lo) Lustre: Skipped 13 previous similar messages Lustre: 21506:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 501 < left 760, rollback = 2 Lustre: 21506:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 369 previous similar messages LustreError: 25019:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000405:0x354:0x0]: rc = -2 LustreError: 5748:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1631095929 with bad export cookie 1685834217849799601 LustreError: 5748:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) Skipped 11 previous similar messages Lustre: 18314:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x280000403:0x72d:0x0]// may get corrupted (rc -108) ------------[ cut here ]------------ WARNING: CPU: 4 PID: 5678 at /home/green/git/lustre-release/ldiskfs/ext4_jbd2.c:266 __ldiskfs_handle_dirty_metadata+0x1d2/0x230 [ldiskfs] Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) dm_flakey dm_mod loop zfs(PO) zunicode(PO) zzstd(O) zlua(O) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) jbd2 mbcache crc_t10dif crct10dif_generic sb_edac edac_core iosf_mbi crc32_pclmul ghash_clmulni_intel aesni_intel lrw gf128mul glue_helper ablk_helper cryptd pcspkr virtio_balloon virtio_console i2c_piix4 ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm crct10dif_pclmul drm crct10dif_common drm_panel_orientation_quirks ata_piix crc32c_intel serio_raw virtio_blk i2c_core libata floppy [last unloaded: libcfs] CPU: 4 PID: 5678 Comm: mdt02_015 Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011 Call Trace: [<ffffffff817d9418>] dump_stack+0x19/0x1b [<ffffffff8108d538>] __warn+0xd8/0x100 [<ffffffff8108d67d>] warn_slowpath_null+0x1d/0x20 [<ffffffffa0d9a262>] __ldiskfs_handle_dirty_metadata+0x1d2/0x230 [ldiskfs] [<ffffffffa0d7a33d>] ldiskfs_getblk+0x13d/0x210 [ldiskfs] [<ffffffffa0d7a437>] ldiskfs_bread+0x27/0xe0 [ldiskfs] [<ffffffffa0cc5e97>] osd_write+0x4e7/0xcb0 [osd_ldiskfs] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffffa04e46f3>] dt_record_write+0x33/0x120 [obdclass] [<ffffffffa04a36b2>] llog_osd_write_rec+0x1052/0x19a0 [obdclass] [<ffffffffa0490760>] llog_write_rec+0x290/0x590 [obdclass] [<ffffffffa04964ee>] llog_cat_add_rec+0x23e/0x950 [obdclass] [<ffffffffa048d5ff>] llog_add+0x17f/0x1f0 [obdclass] [<ffffffffa04bea39>] ? lprocfs_counter_add+0xf9/0x160 [obdclass] [<ffffffffa082760d>] sub_updates_write+0xb6b/0xe32 [ptlrpc] [<ffffffffa07f7892>] top_trans_stop+0x7c2/0xf60 [ptlrpc] [<ffffffffa0fa8f3c>] lod_trans_stop+0x25c/0x340 [lod] [<ffffffffa0e7d408>] mdd_trans_stop+0x28/0x16e [mdd] [<ffffffffa0e5eb2f>] mdd_migrate_object+0xf1f/0x1ba0 [mdd] [<ffffffffa04e3375>] ? lu_object_find_at+0x285/0xb10 [obdclass] [<ffffffffa0e5fb86>] mdd_migrate+0x3d6/0x8e0 [mdd] [<ffffffffa0f310d8>] mdo_migrate+0x4c/0x4e [mdt] [<ffffffffa0ef0e53>] mdt_reint_migrate+0x1023/0x11d0 [mdt] [<ffffffffa0ef1087>] mdt_reint_rec+0x87/0x240 [mdt] [<ffffffffa0ec655c>] mdt_reint_internal+0x76c/0xb40 [mdt] [<ffffffffa0ed3657>] mdt_reint+0x67/0x150 [mdt] [<ffffffffa07e626e>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<ffffffffa03ca45e>] ? libcfs_nid2str_r+0xfe/0x130 [lnet] [<ffffffffa0791990>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<ffffffffa0793529>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e32be>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffffa0792950>] ? ptlrpc_wait_event+0x620/0x620 [ptlrpc] [<ffffffff810ba0f4>] kthread+0xe4/0xf0 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 ---[ end trace 331e0184efa6e1ec ]--- LDISKFS-fs: ldiskfs_getblk:888: aborting transaction: error 28 in __ldiskfs_handle_dirty_metadata LDISKFS-fs error (device dm-1): ldiskfs_getblk:888: inode #176: block 53819: comm mdt02_015: journal_dirty_metadata failed: handle type 0 started at line 1972, credits 512/0, errcode -28 Aborting journal on device dm-1-8. LDISKFS-fs (dm-1): Remounting filesystem read-only LustreError: 5678:0:(osd_io.c:2103:osd_ldiskfs_write_record()) lustre-MDT0001: error reading offset 598016 (block 146, size 40, offs 598000), credits 0/0: rc = -28 LustreError: 5678:0:(update_trans.c:985:top_trans_stop()) lustre-MDT0001-osd: write updates failed: rc = -28 LustreError: 5678:0:(osd_io.c:2110:osd_ldiskfs_write_record()) journal_get_write_access() returned error -30 LustreError: 5678:0:(tgt_lastrcvd.c:1253:tgt_add_reply_data()) lustre-MDT0001: can't update reply_data file: rc = -30 LustreError: 5678:0:(osd_handler.c:2075:osd_trans_stop()) lustre-MDT0001: failed in transaction hook: rc = -30 LDISKFS-fs error (device dm-1) in osd_trans_stop:2082: error 28 LustreError: 5678:0:(osd_handler.c:2085:osd_trans_stop()) lustre-MDT0001: failed to stop transaction: rc = -28 LustreError: 5678:0:(update_trans.c:1011:top_trans_stop()) lustre-MDT0001-osd: stop trans failed: rc = -30 LustreError: 22736:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-OST0001-osc-MDT0001: fail to cancel 36 llog-records: rc = -30 LustreError: 22736:0:(osp_sync.c:1094:osp_sync_process_committed()) lustre-OST0001-osc-MDT0001: can't cancel 36 records: rc = -30 LustreError: 8491:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -5 LustreError: 21783:0:(osd_handler.c:1779:osd_trans_commit_cb()) transaction @0xffff8800a10fa508 commit error: 2 LustreError: 21606:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0001-osp-MDT0000: fail to cancel 1 of 1 llog-records: rc = -116 LDISKFS-fs error (device dm-1) in osd_trans_stop:2082: IO failure LustreError: 8428:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -30 LustreError: 8428:0:(llite_lib.c:1836:ll_md_setattr()) Skipped 4 previous similar messages LustreError: 22736:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-OST0001-osc-MDT0001: fail to cancel 101 llog-records: rc = -30 LustreError: 22736:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) Skipped 8 previous similar messages LustreError: 22736:0:(osp_sync.c:1094:osp_sync_process_committed()) lustre-OST0001-osc-MDT0001: can't cancel 101 records: rc = -30 LustreError: 22736:0:(osp_sync.c:1094:osp_sync_process_committed()) Skipped 5 previous similar messages LustreError: 22004:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0001-osp-MDT0002: fail to cancel 1 llog-records: rc = -116 LustreError: 22004:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0001-osp-MDT0002: fail to cancel 1 of 1 llog-records: rc = -116 LustreError: 22004:0:(llog_cat.c:787:llog_cat_cancel_records()) Skipped 2 previous similar messages LustreError: 22004:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0001-osp-MDT0002: fail to cancel 1 llog-records: rc = -2 LustreError: 22004:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) Skipped 1 previous similar message LustreError: 22004:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0001-osp-MDT0002: fail to cancel 1 of 1 llog-records: rc = -2 LustreError: 22004:0:(llog_cat.c:787:llog_cat_cancel_records()) Skipped 1 previous similar message LustreError: 23212:0:(osp_sync.c:1094:osp_sync_process_committed()) lustre-OST0003-osc-MDT0001: can't cancel 6 records: rc = -30 LustreError: 24140:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -30 LustreError: 24140:0:(llite_lib.c:1836:ll_md_setattr()) Skipped 5 previous similar messages LustreError: 21606:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0001-osp-MDT0000: fail to cancel 1 of 1 llog-records: rc = -116 LustreError: 22736:0:(osp_sync.c:1079:osp_sync_process_committed()) lustre-OST0001-osc-MDT0001: can't cancel record: rc = -30 LustreError: 22004:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0001-osp-MDT0002: fail to cancel 1 llog-records: rc = -116 LustreError: 22004:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) Skipped 6 previous similar messages LustreError: 22004:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0001-osp-MDT0002: fail to cancel 1 of 1 llog-records: rc = -116 LustreError: 22004:0:(llog_cat.c:787:llog_cat_cancel_records()) Skipped 2 previous similar messages LustreError: 22728:0:(osp_sync.c:1094:osp_sync_process_committed()) lustre-OST0000-osc-MDT0001: can't cancel 39 records: rc = -30 LustreError: 22728:0:(osp_sync.c:1094:osp_sync_process_committed()) Skipped 2 previous similar messages LustreError: 8817:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -30 LustreError: 8817:0:(llite_lib.c:1836:ll_md_setattr()) Skipped 1 previous similar message LustreError: 22728:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-OST0000-osc-MDT0001: fail to cancel 9 llog-records: rc = -30 LustreError: 22728:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) Skipped 7 previous similar messages LustreError: 22728:0:(osp_sync.c:1094:osp_sync_process_committed()) lustre-OST0000-osc-MDT0001: can't cancel 9 records: rc = -30 LustreError: 22728:0:(osp_sync.c:1094:osp_sync_process_committed()) lustre-OST0000-osc-MDT0001: can't cancel 270 records: rc = -30 LustreError: 22728:0:(osp_sync.c:1094:osp_sync_process_committed()) Skipped 1 previous similar message LustreError: 22004:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0001-osp-MDT0002: fail to cancel 1 of 1 llog-records: rc = -116 LustreError: 22004:0:(llog_cat.c:787:llog_cat_cancel_records()) Skipped 6 previous similar messages | Externally reported by onyx-68 boilpot email |
racer test 1: racer on clients: centos-105.localnet DURATION=2700 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa103ee52>] ll_migrate+0x9b2/0xec0 [lustre] PGD 800000027128c067 PUD 276898067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: loop zfs(PO) zunicode(PO) zzstd(O) zlua(O) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) jbd2 mbcache lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) dm_flakey dm_mod libcfs(OE) crc_t10dif crct10dif_generic sb_edac edac_core iosf_mbi crc32_pclmul ghash_clmulni_intel aesni_intel lrw gf128mul glue_helper ablk_helper cryptd pcspkr virtio_console virtio_balloon i2c_piix4 ip_tables rpcsec_gss_krb5 drm_kms_helper ata_generic ttm pata_acpi crct10dif_pclmul crct10dif_common drm crc32c_intel drm_panel_orientation_quirks ata_piix serio_raw virtio_blk libata i2c_core floppy CPU: 3 PID: 6763 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011 task: ffff88028ea45c40 ti: ffff880251e08000 task.ti: ffff880251e08000 RIP: 0010:[<ffffffffa103ee52>] [<ffffffffa103ee52>] ll_migrate+0x9b2/0xec0 [lustre] RSP: 0018:ffff880251e0bbc8 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff880260e876b8 RCX: 0000000000000000 RDX: 0000000100006e64 RSI: ffff880328c0b5d8 RDI: ffff880328c0b5b8 RBP: ffff880251e0bc48 R08: ffff880241c498d8 R09: ffff880251e0ba60 R10: 0000000000000003 R11: 0000000000000246 R12: 0000000000000000 R13: ffff88027c33ed58 R14: ffff88029ab10a58 R15: 0000000000000030 FS: 00007f8d1682b740(0000) GS:ffff880331ac0000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 CR2: 0000000000000008 CR3: 0000000286312000 CR4: 00000000001607e0 Call Trace: [<ffffffff81242213>] ? __check_object_size+0x1c3/0x220 [<ffffffffa102c561>] ll_dir_ioctl+0x5d01/0x6ed0 [lustre] [<ffffffff811b515b>] ? unlock_page+0x2b/0x30 [<ffffffff811e5e2e>] ? do_read_fault.isra.64+0x12e/0x170 [<ffffffff811ea1cf>] ? handle_pte_fault+0x2ef/0xca0 [<ffffffff811ecf6d>] ? __handle_mm_fault+0x31d/0x720 [<ffffffff8125b3fd>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff81245a38>] ? vfs_write+0x178/0x230 [<ffffffff8125b751>] SyS_ioctl+0xa1/0xc0 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 | Lustre: lustre-OST0000-osc-ffff8802998a8958: disconnect after 21s idle Lustre: 11143:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 508 < left 788, rollback = 2 Lustre: 11143:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 3/12/3, destroy: 0/0/0 Lustre: 11143:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 2/2/0, xattr_set: 10/788/0 Lustre: 11143:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 5/136/0, punch: 0/0/0, quota 8/56/0 Lustre: 11143:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 12/215/1, delete: 1/1/0 Lustre: 11143:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 6/6/0, ref_del: 1/1/0 Lustre: 11143:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 507 < left 788, rollback = 2 Lustre: 11143:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 3/12/3, destroy: 0/0/0 Lustre: 11143:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 2/2/0, xattr_set: 10/788/0 Lustre: 11143:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 5/136/0, punch: 0/0/0, quota 8/56/0 Lustre: 11143:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 12/215/2, delete: 1/1/0 Lustre: 11143:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 6/6/0, ref_del: 1/1/0 Lustre: 11120:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 512 < left 1193, rollback = 7 Lustre: 11120:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 2/8/0, destroy: 0/0/0 Lustre: 11120:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 8 previous similar messages Lustre: 11120:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 5/192/0 Lustre: 11120:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 8 previous similar messages Lustre: 11120:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1108/1193/0, punch: 0/0/0, quota 6/6/0 Lustre: 11120:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 8 previous similar messages Lustre: 11120:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 2/34/0, delete: 0/0/0 Lustre: 11120:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 8 previous similar messages LustreError: 11329:0:(out_handler.c:910:out_tx_end()) lustre-MDT0000-osd: undo for /home/green/git/lustre-release/lustre/ptlrpc/../../lustre/target/out_handler.c:445: rc = -524 Lustre: 11120:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 11120:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 8 previous similar messages Lustre: 11120:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 512 < left 1193, rollback = 7 Lustre: 11120:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 481 < left 1171, rollback = 7 Lustre: 11120:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 2478 previous similar messages Lustre: 11120:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 2/8/0, destroy: 0/0/0 Lustre: 11120:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 2695 previous similar messages Lustre: 11120:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 5/192/9 Lustre: 11120:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 2695 previous similar messages Lustre: 11120:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1108/1193/22, punch: 0/0/0, quota 6/6/0 Lustre: 11120:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 2695 previous similar messages Lustre: 11120:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 2/34/0, delete: 0/0/0 Lustre: 11120:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 2696 previous similar messages Lustre: 11120:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 11120:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 2697 previous similar messages Lustre: 11140:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 498 < left 671, rollback = 2 Lustre: 11140:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 8 previous similar messages Lustre: 17384:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 500 < left 47267, rollback = 7 Lustre: 17384:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 3019 previous similar messages Lustre: 11132:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 503 < left 877, rollback = 2 Lustre: 11132:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 7 previous similar messages Lustre: 11132:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 3/12/6, destroy: 0/0/0 Lustre: 11132:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 7217 previous similar messages Lustre: 11132:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 2/2/0, xattr_set: 11/877/0 Lustre: 11132:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 7217 previous similar messages Lustre: 11132:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 5/136/0, punch: 0/0/0, quota 8/56/0 Lustre: 11132:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 7217 previous similar messages Lustre: 11132:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 12/215/4, delete: 1/1/0 Lustre: 11132:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 7217 previous similar messages Lustre: 11132:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 6/6/0, ref_del: 1/1/0 Lustre: 11132:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 7217 previous similar messages LustreError: 11952:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0000-osp-MDT0001: fail to cancel 1 llog-records: rc = -2 LustreError: 11952:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0000-osp-MDT0001: fail to cancel 1 of 1 llog-records: rc = -2 Lustre: 18660:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 505 < left 521, rollback = 2 Lustre: 18660:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 17 previous similar messages Lustre: 11128:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 511 < left 798, rollback = 7 Lustre: 11128:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 4405 previous similar messages Lustre: 11128:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 1/4/0, destroy: 0/0/0 Lustre: 11128:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 21 previous similar messages Lustre: 11128:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 4/209/0 Lustre: 11128:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 21 previous similar messages Lustre: 11128:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 789/798/0, punch: 0/0/0, quota 8/80/0 Lustre: 11128:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 21 previous similar messages Lustre: 11128:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 2/33/1, delete: 0/0/0 Lustre: 11128:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 20 previous similar messages Lustre: 11128:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 11128:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 19 previous similar messages Lustre: 11120:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 505 < left 582, rollback = 2 Lustre: 11120:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 3 previous similar messages Lustre: 18660:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 510 < left 798, rollback = 7 Lustre: 18660:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 1563 previous similar messages Lustre: dir [0x200000403:0x56:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: 17393:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 8: before 511 < left 18699, rollback = 9 Lustre: 17393:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 5 previous similar messages Lustre: 17393:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 1099/4396/0, destroy: 1/4/0 Lustre: 17393:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 5419 previous similar messages Lustre: 17393:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 3/3/0, xattr_set: 1103/148/0 Lustre: 17393:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 5419 previous similar messages Lustre: 17393:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 5496/47267/0, punch: 0/0/0, quota 10/74/0 Lustre: 17393:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 5419 previous similar messages Lustre: 17393:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 1100/18699/0, delete: 2/5/0 Lustre: 17393:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 5419 previous similar messages Lustre: 17393:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 1/1/0, ref_del: 2/2/1 Lustre: 17393:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 5419 previous similar messages Lustre: 17469:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 512 < left 1179, rollback = 7 Lustre: 17469:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 3850 previous similar messages ------------[ cut here ]------------ WARNING: CPU: 10 PID: 11141 at /home/green/git/lustre-release/ldiskfs/namei.c:3331 ldiskfs_orphan_add+0x11e/0x2a0 [ldiskfs] Modules linked in: loop zfs(PO) zunicode(PO) zzstd(O) zlua(O) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) jbd2 mbcache lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) dm_flakey dm_mod libcfs(OE) crc_t10dif crct10dif_generic sb_edac edac_core iosf_mbi crc32_pclmul ghash_clmulni_intel aesni_intel lrw gf128mul glue_helper ablk_helper cryptd pcspkr virtio_console virtio_balloon i2c_piix4 ip_tables rpcsec_gss_krb5 drm_kms_helper ata_generic ttm pata_acpi crct10dif_pclmul crct10dif_common drm crc32c_intel drm_panel_orientation_quirks ata_piix serio_raw virtio_blk libata i2c_core floppy CPU: 10 PID: 11141 Comm: mdt05_001 Tainted: P OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011 Call Trace: [<ffffffff817d9418>] dump_stack+0x19/0x1b [<ffffffff8108d538>] __warn+0xd8/0x100 [<ffffffff8108d67d>] warn_slowpath_null+0x1d/0x20 [<ffffffffa0bf54ae>] ldiskfs_orphan_add+0x11e/0x2a0 [ldiskfs] [<ffffffffa0be69eb>] ldiskfs_xattr_inode_orphan_add.constprop.17+0xbb/0x110 [ldiskfs] [<ffffffffa0be9417>] ldiskfs_xattr_delete_inode+0x77/0x3c0 [ldiskfs] [<ffffffffa0c09aa0>] ? ldiskfs_evict_inode+0x190/0x670 [ldiskfs] [<ffffffffa0c09ad9>] ldiskfs_evict_inode+0x1c9/0x670 [ldiskfs] [<ffffffff812649af>] evict+0xaf/0x180 [<ffffffff81264de5>] iput+0xf5/0x180 [<ffffffffa0c9ed08>] osd_object_delete+0x1f8/0x370 [osd_ldiskfs] [<ffffffffa04e859d>] lu_object_free.isra.26+0xcd/0x1d0 [obdclass] [<ffffffffa04eccf1>] lu_object_put+0xa1/0x480 [obdclass] [<ffffffffa0e1e440>] mdt_object_put+0x30/0x110 [mdt] [<ffffffffa0e25e18>] mdt_reint_unlink+0x918/0x19b0 [mdt] [<ffffffffa050e79e>] ? lu_ucred+0x1e/0x30 [obdclass] [<ffffffffa0e15af2>] ? ucred_set_audit_enabled.isra.13+0x22/0x60 [mdt] [<ffffffffa0e2d087>] mdt_reint_rec+0x87/0x240 [mdt] [<ffffffffa0e0255c>] mdt_reint_internal+0x76c/0xb40 [mdt] [<ffffffffa0e0f657>] mdt_reint+0x67/0x150 [mdt] [<ffffffffa07f026e>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<ffffffffa033545e>] ? libcfs_nid2str_r+0xfe/0x130 [lnet] [<ffffffffa079b990>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<ffffffffa079d529>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffffa079c950>] ? ptlrpc_wait_event+0x620/0x620 [ptlrpc] [<ffffffff810ba0f4>] kthread+0xe4/0xf0 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 ---[ end trace 647798641f8bde2e ]--- LustreError: 21041:0:(dir.c:713:ll_dir_get_default_layout()) unknown magic: CD40CD0 Lustre: 18052:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 512 < left 32991, rollback = 7 Lustre: 18052:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 12119 previous similar messages Lustre: 18052:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 767/3068/0, destroy: 1/4/0 Lustre: 18052:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 12184 previous similar messages Lustre: 18052:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 767/0/0 Lustre: 18052:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 12184 previous similar messages Lustre: 18052:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 3836/32991/0, punch: 0/0/0, quota 10/74/0 Lustre: 18052:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 12184 previous similar messages Lustre: 18052:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 767/13039/0, delete: 2/5/0 Lustre: 18052:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 12184 previous similar messages Lustre: 18052:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 1/1/0 Lustre: 18052:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 12184 previous similar messages Lustre: 17384:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 507 < left 610, rollback = 2 Lustre: 17384:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 63 previous similar messages Lustre: 4145:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff880087efcb40 x1710260894533632/t0(0) o10->lustre-OST0003-osc-ffff880299ab4138@0@lo:6/4 lens 440/432 e 0 to 0 dl 1631032228 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'truncate.0' Lustre: 17441:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000403:0x80:0x0] with magic=0xbd60bd0 Lustre: 17841:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000404:0x53:0x0] with magic=0xbd60bd0 Lustre: 17841:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 1 previous similar message Lustre: ll_ost00_000: service thread pid 12702 was inactive for 62.171 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 12721, comm: ll_ost05_002 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] ofd_destroy_by_fid+0x1d1/0x520 [ofd] [<0>] ofd_destroy_hdl+0x27f/0xa20 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 12720, comm: ll_ost05_001 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] ofd_destroy_by_fid+0x1d1/0x520 [ofd] [<0>] ofd_destroy_hdl+0x27f/0xa20 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: Skipped 2 previous similar messages Pid: 12702, comm: ll_ost00_000 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] ofd_destroy_by_fid+0x1d1/0x520 [ofd] [<0>] ofd_destroy_hdl+0x27f/0xa20 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: ll_ost05_004: service thread pid 19779 was inactive for 63.006 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: ll_ost05_008: service thread pid 22124 was inactive for 62.047 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 3 previous similar messages LustreError: 11004:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0002_UUID lock: ffff880273770040/0xc6b178ed82533323 lrc: 3/0,0 mode: PW/PW res: [0x340000400:0x7c:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000000020020 nid: 0@lo remote: 0xc6b178ed8253329e expref: 986 pid: 12707 timeout: 421 lvb_type: 0 LustreError: 11-0: lustre-OST0002-osc-ffff880299ab4138: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-OST0002-osc-ffff880299ab4138: Connection to lustre-OST0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0002-osc-ffff880299ab4138: This client was evicted by lustre-OST0002; in progress operations using this service will fail. LustreError: 21747:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff880089c08040 x1710260895513280/t0(0) o104->lustre-OST0002@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' Lustre: ll_ost01_000: service thread pid 12705 was inactive for 62.114 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 2 previous similar messages LustreError: 19169:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88008af7f700 x1710260895527936/t0(0) o104->lustre-OST0002@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 19169:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 1 previous similar message LustreError: 11004:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0003_UUID lock: ffff8800852fd2c0/0xc6b178ed8252cb34 lrc: 3/0,0 mode: PW/PW res: [0x380000400:0x6:0x0].0x0 rrc: 4 type: EXT [0->18446744073709551615] (req 0->524287) gid 0 flags: 0x60000400030020 nid: 0@lo remote: 0xc6b178ed8252cb26 expref: 816 pid: 12707 timeout: 423 lvb_type: 0 LustreError: 23344:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88007b9abec0 x1710260895552384/t0(0) o104->lustre-OST0003@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 23344:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 4 previous similar messages LustreError: 13874:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1631032199 with bad export cookie 14317357651876395617 LustreError: 11-0: lustre-OST0003-osc-ffff880299ab4138: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-OST0003-osc-ffff880299ab4138: Connection to lustre-OST0003 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0003-osc-ffff880299ab4138: This client was evicted by lustre-OST0003; in progress operations using this service will fail. Lustre: 4138:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.205@tcp:/lustre/fid: [0x240000404:0xf8:0x0]// may get corrupted (rc -108) Lustre: 4138:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.205@tcp:/lustre/fid: [0x200000403:0xf9:0x0]// may get corrupted (rc -108) Lustre: 4139:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.205@tcp:/lustre/fid: [0x240000404:0x107:0x0]/ may get corrupted (rc -108) Lustre: 4144:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.205@tcp:/lustre/fid: [0x280000404:0xd7:0x0]// may get corrupted (rc -108) Lustre: 4144:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.205@tcp:/lustre/fid: [0x240000403:0x11d:0x0]/ may get corrupted (rc -108) Lustre: 4145:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.205@tcp:/lustre/fid: [0x280000404:0xe4:0x0]/ may get corrupted (rc -108) Lustre: 4139:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.205@tcp:/lustre/fid: [0x200000404:0xff:0x0]// may get corrupted (rc -108) LustreError: 25108:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0003-osc-ffff880299ab4138: namespace resource [0x380000400:0x6:0x0].0x0 (ffff88027c914f40) refcount nonzero (1) after lock cleanup; forcing cleanup. Lustre: lustre-OST0003-osc-ffff880299ab4138: Connection restored to 192.168.123.205@tcp (at 0@lo) Lustre: ll_ost01_006: service thread pid 19662 was inactive for 66.033 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 4 previous similar messages sched: RT throttling activated Lustre: 17973:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 510 < left 797, rollback = 7 Lustre: 17973:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 22378 previous similar messages Lustre: 11128:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 1293/5172/0, destroy: 1/4/1 Lustre: 11128:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 22611 previous similar messages Lustre: 11128:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 5/5/0, xattr_set: 1301/283/1 Lustre: 11128:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 22611 previous similar messages Lustre: 11128:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 6466/55609/42, punch: 0/0/0, quota 6/6/0 Lustre: 11128:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 22611 previous similar messages Lustre: 11128:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 1295/22013/1, delete: 3/6/1 Lustre: 11128:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 22611 previous similar messages Lustre: 17973:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 1/1/0, ref_del: 2/2/0 Lustre: 17973:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 22708 previous similar messages Lustre: 13165:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 8: before 509 < left 6850, rollback = 9 Lustre: 13165:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 35 previous similar messages LustreError: 11004:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0001_UUID lock: ffff880268385a40/0xc6b178ed82596505 lrc: 3/0,0 mode: PW/PW res: [0xdd:0x0:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 786432->18446744073709551615) gid 0 flags: 0x60000000030020 nid: 0@lo remote: 0xc6b178ed825964d4 expref: 2273 pid: 20106 timeout: 522 lvb_type: 0 LustreError: 11002:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8800850be440 x1710260899258432/t0(0) o105->lustre-OST0001@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 11002:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 1 previous similar message LustreError: 11-0: lustre-OST0001-osc-ffff8802998a8958: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-OST0001-osc-ffff8802998a8958: Connection to lustre-OST0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0001-osc-ffff8802998a8958: This client was evicted by lustre-OST0001; in progress operations using this service will fail. LustreError: 17526:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1631032300 with bad export cookie 14317357651876396842 Lustre: lustre-OST0003-osc-ffff8802998a8958: Connection to lustre-OST0003 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: 17894:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 6s req@ffff8800ab885428 x1710260899187328/t0(0) o35->6c066995-9548-470f-a292-69335ad47963@0@lo:0/0 lens 392/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'ls.0' LustreError: 167-0: lustre-OST0003-osc-ffff8802998a8958: This client was evicted by lustre-OST0003; in progress operations using this service will fail. LustreError: 23344:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802728a4b40 x1710260899374656/t0(0) o104->lustre-OST0003@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 23344:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 12 previous similar messages Lustre: 4152:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.205@tcp:/lustre/fid: [0x280000403:0xf0:0x0]// may get corrupted (rc -108) Lustre: 4151:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.205@tcp:/lustre/fid: [0x240000404:0x1a3:0x0]// may get corrupted (rc -108) 5[27176]: segfault at 0 ip (null) sp 00007ffe1eec3c98 error 14 in 5[400000+6000] Lustre: 4147:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.205@tcp:/lustre/fid: [0x280000403:0xcd:0x0]// may get corrupted (rc -108) LustreError: 27615:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0003-osc-ffff8802998a8958: namespace resource [0x154:0x0:0x0].0x0 (ffff880084bd1bc0) refcount nonzero (1) after lock cleanup; forcing cleanup. Lustre: lustre-OST0003-osc-ffff8802998a8958: Connection restored to 192.168.123.205@tcp (at 0@lo) Lustre: Skipped 1 previous similar message LustreError: 27561:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0001-osc-ffff8802998a8958: namespace resource [0x1e0:0x0:0x0].0x0 (ffff88027848ed40) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 27561:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 7 previous similar messages Lustre: lustre-OST0001-osc-ffff8802998a8958: Connection restored to 192.168.123.205@tcp (at 0@lo) LustreError: 11004:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0002_UUID lock: ffff880277a75680/0xc6b178ed825a7461 lrc: 3/0,0 mode: PW/PW res: [0x340000400:0x8:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000480020020 nid: 0@lo remote: 0xc6b178ed825a7430 expref: 1501 pid: 14399 timeout: 543 lvb_type: 0 LustreError: 11004:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 1 previous similar message LustreError: 11-0: lustre-OST0002-osc-ffff880299ab4138: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-OST0002-osc-ffff880299ab4138: Connection to lustre-OST0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0002-osc-ffff880299ab4138: This client was evicted by lustre-OST0002; in progress operations using this service will fail. Lustre: 4147:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.205@tcp:/lustre/fid: [0x240000404:0x1b3:0x0]// may get corrupted (rc -108) Lustre: 4147:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.205@tcp:/lustre/fid: [0x280000403:0xfa:0x0]/ may get corrupted (rc -108) Lustre: 4152:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff88028c3c1940 x1710260900241472/t0(0) o10->lustre-OST0000-osc-ffff8802998a8958@0@lo:6/4 lens 440/432 e 0 to 0 dl 1631032388 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'truncate.0' LustreError: 28118:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0002-osc-ffff880299ab4138: namespace resource [0x340000401:0xcc6:0x0].0x0 (ffff88007b823240) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 28118:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 223 previous similar messages Lustre: lustre-OST0002-osc-ffff880299ab4138: Connection restored to 192.168.123.205@tcp (at 0@lo) LustreError: 12352:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0001-osp-MDT0002: fail to cancel 1 llog-records: rc = -2 LustreError: 12352:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0001-osp-MDT0002: fail to cancel 1 of 1 llog-records: rc = -2 LustreError: 11495:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0001-osp-MDT0000: fail to cancel 1 llog-records: rc = -2 LustreError: 11495:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0001-osp-MDT0000: fail to cancel 1 of 1 llog-records: rc = -2 Lustre: 17522:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 508 < left 14630, rollback = 7 Lustre: 17522:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 64357 previous similar messages Lustre: 17522:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 340/1360/0, destroy: 1/4/1 Lustre: 17522:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 64217 previous similar messages Lustre: 17522:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 3/3/1, xattr_set: 344/148/0 Lustre: 17522:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 64217 previous similar messages Lustre: 17522:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1701/14630/0, punch: 0/0/0, quota 10/90/0 Lustre: 17522:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 64217 previous similar messages Lustre: 17522:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 341/5796/0, delete: 2/5/1 Lustre: 17522:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 64217 previous similar messages Lustre: 17522:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 1/1/0, ref_del: 2/2/1 Lustre: 17522:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 64120 previous similar messages LustreError: 11495:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0001-osp-MDT0000: fail to cancel 1 llog-records: rc = -2 LustreError: 11495:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0001-osp-MDT0000: fail to cancel 1 of 1 llog-records: rc = -2 Lustre: 11324:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 505 < left 550, rollback = 2 Lustre: 11324:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 97 previous similar messages LustreError: 30862:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff880299ab4138: inode [0x240000404:0x1c1:0x0] mdc close failed: rc = -13 LustreError: 11004:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 112s: evicting client at 0@lo ns: filter-lustre-OST0003_UUID lock: ffff88027fad6580/0xc6b178ed82606e2e lrc: 3/0,0 mode: PW/PW res: [0x152:0x0:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000000030020 nid: 0@lo remote: 0xc6b178ed82606e12 expref: 1335 pid: 25329 timeout: 670 lvb_type: 0 LustreError: 20690:0:(ldlm_lockd.c:1342:ldlm_handle_enqueue0()) ### lock on disconnected export ffff8800abf2ae98 ns: filter-lustre-OST0003_UUID lock: ffff88007da04000/0xc6b178ed8265a831 lrc: 2/0,0 mode: --/PW res: [0x179:0x0:0x0].0x0 rrc: 7 type: EXT [0->0] (req 0->0) gid 0 flags: 0x40000000000000 nid: local remote: 0xc6b178ed8265a823 expref: -99 pid: 20690 timeout: 0 lvb_type: 0 LustreError: 17529:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88027275a5c0 x1710260904858368/t0(0) o105->lustre-OST0003@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 17529:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 7 previous similar messages LustreError: 11001:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1631032457 with bad export cookie 14317357651877341793 LustreError: 11-0: lustre-OST0003-osc-ffff8802998a8958: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-OST0003-osc-ffff8802998a8958: Connection to lustre-OST0003 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0003-osc-ffff8802998a8958: This client was evicted by lustre-OST0003; in progress operations using this service will fail. Lustre: 4151:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.205@tcp:/lustre/fid: [0x200000403:0x1ea:0x0]// may get corrupted (rc -108) LustreError: 32071:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0003-osc-ffff8802998a8958: namespace resource [0x7e0:0x0:0x0].0x0 (ffff880264a816c0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 32071:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 1 previous similar message Lustre: lustre-OST0003-osc-ffff8802998a8958: Connection restored to 192.168.123.205@tcp (at 0@lo) LustreError: 27644:0:(lustre_lmv.h:438:lmv_is_sane()) insane LMV: magic=0xcd40cd0 count=2 index=1 hash=crush:0x82000003 version=0 migrate offset=1 migrate hash=crush:3. Lustre: 26051:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000403:0x234:0x0] with magic=0xbd60bd0 Lustre: 26051:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 3 previous similar messages LustreError: 32052:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0000: '4' migration was interrupted, run 'lfs migrate -m 1 -c 1 -H crush 4' to finish migration. 3[1554]: segfault at 8 ip 00007f6667ee97e8 sp 00007ffea44e0a90 error 4 in ld-2.17.so[7f6667ede000+22000] LustreError: 11004:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0000_UUID lock: ffff88007da03100/0xc6b178ed8265a8e7 lrc: 3/0,0 mode: PW/PW res: [0x17b:0x0:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000000020020 nid: 0@lo remote: 0xc6b178ed8265a8c4 expref: 1800 pid: 23336 timeout: 785 lvb_type: 0 LustreError: 20718:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff880254f88cc0 x1710260908933504/t0(0) o105->lustre-OST0000@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 20718:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 136 previous similar messages LustreError: 11-0: lustre-OST0000-osc-ffff880299ab4138: operation ost_setattr to node 0@lo failed: rc = -107 LustreError: Skipped 1 previous similar message Lustre: lustre-OST0000-osc-ffff880299ab4138: Connection to lustre-OST0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 4137:0:(osc_request.c:1035:osc_init_grant()) lustre-OST0000-osc-ffff880299ab4138: granted 8437760 but already consumed 10137600 LustreError: 167-0: lustre-OST0000-osc-ffff880299ab4138: This client was evicted by lustre-OST0000; in progress operations using this service will fail. Lustre: 4141:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.205@tcp:/lustre/fid: [0x200000403:0x1be:0x0]// may get corrupted (rc -108) Lustre: 4140:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.205@tcp:/lustre/fid: [0x280000403:0x1a1:0x0]/ may get corrupted (rc -108) Lustre: 4140:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.205@tcp:/lustre/fid: [0x200000403:0x246:0x0]/ may get corrupted (rc -108) Lustre: 4140:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.205@tcp:/lustre/fid: [0x200000403:0x201:0x0]// may get corrupted (rc -108) LustreError: 2591:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0000-osc-ffff880299ab4138: namespace resource [0xab0:0x0:0x0].0x0 (ffff88026f50afc0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 2591:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 151 previous similar messages Lustre: lustre-OST0000-osc-ffff880299ab4138: Connection restored to 192.168.123.205@tcp (at 0@lo) Lustre: dir [0x240000403:0x289:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: dir [0x280000404:0x216:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: 12337:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 512 < left 1458, rollback = 7 Lustre: 12337:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 76206 previous similar messages Lustre: 12337:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 1/4/0, destroy: 0/0/0 Lustre: 12337:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 76405 previous similar messages Lustre: 12337:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 4/209/0 Lustre: 17820:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 5/136/0, punch: 0/0/0, quota 8/56/0 Lustre: 17820:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 76406 previous similar messages Lustre: 17820:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 12/215/4, delete: 1/1/0 Lustre: 17820:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 76405 previous similar messages Lustre: 17820:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 6/6/0, ref_del: 1/1/0 Lustre: 17820:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 76406 previous similar messages Lustre: 12337:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 76408 previous similar messages Lustre: 4153:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff8802606b1f80 x1710260911800256/t0(0) o10->lustre-OST0000-osc-ffff8802998a8958@0@lo:6/4 lens 440/432 e 0 to 0 dl 1631032700 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'cp.0' LustreError: 11153:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0002: '17' migration was interrupted, run 'lfs migrate -m 1 -c 1 -H crush 17' to finish migration. Lustre: 17482:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 8: before 509 < left 10930, rollback = 9 Lustre: 17482:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 178 previous similar messages LustreError: 11004:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0003_UUID lock: ffff880272f087c0/0xc6b178ed8266eda3 lrc: 3/0,0 mode: PW/PW res: [0x146:0x0:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000000030020 nid: 0@lo remote: 0xc6b178ed8266ecfb expref: 1985 pid: 25329 timeout: 895 lvb_type: 0 LustreError: 26811:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1631032670 with bad export cookie 14317357651877800314 LustreError: 26811:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) Skipped 1 previous similar message Lustre: lustre-OST0003-osc-ffff8802998a8958: Connection to lustre-OST0003 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 26068:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802860f57c0 x1710260913360256/t0(0) o104->lustre-OST0003@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 26068:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 3 previous similar messages LustreError: 11-0: lustre-OST0003-osc-ffff8802998a8958: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 11 previous similar messages LustreError: 167-0: lustre-OST0003-osc-ffff8802998a8958: This client was evicted by lustre-OST0003; in progress operations using this service will fail. LustreError: 1670:0:(ofd_obd.c:501:ofd_destroy_export()) lustre-OST0003: cli 6c066995-9548-470f-a292-69335ad47963/ffff8802874b5d28 has 28672 pending on destroyed export LustreError: 1670:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_destroy_export: tot_granted 228813504 != fo_tot_granted 228842176 LustreError: 1670:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_destroy_export: tot_pending 28672 != fo_tot_pending 57344 LustreError: 12750:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 228813504 != fo_tot_granted 228842176 LustreError: 815:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 28672 != fo_tot_pending 57344 LustreError: 12750:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 1 previous similar message LustreError: 815:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 228814912 != fo_tot_granted 228843584 LustreError: 12750:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 28672 != fo_tot_pending 57344 LustreError: 12750:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 1 previous similar message LustreError: 815:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 1 previous similar message LustreError: 29082:0:(ldlm_lockd.c:1342:ldlm_handle_enqueue0()) ### lock on disconnected export ffff880080978958 ns: filter-lustre-OST0000_UUID lock: ffff880079de4780/0xc6b178ed82732982 lrc: 2/0,0 mode: --/PR res: [0x2c0000400:0x5f7:0x0].0x0 rrc: 5 type: EXT [0->0] (req 0->0) gid 0 flags: 0x40000000000000 nid: local remote: 0xc6b178ed8272d7ff expref: -99 pid: 29082 timeout: 0 lvb_type: 0 LustreError: 20167:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1631032685 with bad export cookie 14317357651878188569 LustreError: 20167:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) Skipped 1 previous similar message Lustre: 4140:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.205@tcp:/lustre/fid: [0x200000403:0x2d9:0x0]/ may get corrupted (rc -108) Lustre: 4141:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.205@tcp:/lustre/fid: [0x200000404:0x2bd:0x0]// may get corrupted (rc -108) LustreError: 12741:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 242895808 != fo_tot_granted 242924480 LustreError: 12741:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 5910528 != fo_tot_pending 5939200 LustreError: 12741:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 1 previous similar message Lustre: lustre-OST0003-osc-ffff8802998a8958: Connection restored to 192.168.123.205@tcp (at 0@lo) LustreError: 12752:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 237013056 != fo_tot_granted 237041728 LustreError: 12752:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 2 previous similar messages LustreError: 12752:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 28672 != fo_tot_pending 57344 LustreError: 12752:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 2 previous similar messages ptlrpc_watchdog_fire: 29 callbacks suppressed Lustre: mdt03_004: service thread pid 16892 was inactive for 42.022 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 16892, comm: mdt03_004 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] mdt_object_local_lock+0x448/0xb90 [mdt] [<0>] mdt_object_lock_internal+0x70/0x390 [mdt] [<0>] mdt_reint_object_lock+0x2c/0x60 [mdt] [<0>] mdt_object_lock_save+0x29/0x50 [mdt] [<0>] mdt_reint_rename+0x10f0/0x2ff0 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x76c/0xb40 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe LustreError: 4740:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0000-osc-ffff880299ab4138: namespace resource [0x815:0x0:0x0].0x0 (ffff88027d54efc0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 4740:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 15 previous similar messages Pid: 11135, comm: mdt03_001 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] mdt_rename_lock+0x24b/0x4d0 [mdt] [<0>] mdt_reint_migrate+0x6e8/0x11d0 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x76c/0xb40 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe LustreError: 12752:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 243438400 != fo_tot_granted 243467072 LustreError: 12752:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 2 previous similar messages LustreError: 12752:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 28672 != fo_tot_pending 57344 LustreError: 12752:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 2 previous similar messages LustreError: 4712:0:(lov_object.c:1299:lov_layout_change()) lustre-clilov-ffff8802998a8958: cannot apply new layout on [0x200000404:0x322:0x0] : rc = -22 LustreError: 4712:0:(lcommon_cl.c:196:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000404:0x322:0x0]: rc = -22 LustreError: 12734:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 243441088 != fo_tot_granted 243469760 LustreError: 12734:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 4 previous similar messages LustreError: 12734:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 29568 != fo_tot_pending 58240 LustreError: 12734:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 4 previous similar messages LustreError: 5135:0:(lov_object.c:1299:lov_layout_change()) lustre-clilov-ffff8802998a8958: cannot apply new layout on [0x200000404:0x322:0x0] : rc = -22 LustreError: 5135:0:(lcommon_cl.c:196:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000404:0x322:0x0]: rc = -22 LustreError: 5135:0:(llite_lib.c:2983:ll_prep_inode()) new_inode -fatal: rc -22 Lustre: lustre-OST0000: haven't heard from client 9da291a9-717f-4193-8b26-da6e22b22b79 (at 0@lo) in 48 seconds. I think it's dead, and I am evicting it. exp ffff8802709d1bf8, cur 1631032734 expire 1631032704 last 1631032686 LustreError: 5499:0:(lov_object.c:1299:lov_layout_change()) lustre-clilov-ffff880299ab4138: cannot apply new layout on [0x200000404:0x322:0x0] : rc = -22 LustreError: 5499:0:(lov_object.c:1299:lov_layout_change()) Skipped 1 previous similar message LustreError: 5499:0:(vvp_io.c:1793:vvp_io_init()) lustre: refresh file layout [0x200000404:0x322:0x0] error -22. LustreError: 5413:0:(lov_object.c:1299:lov_layout_change()) lustre-clilov-ffff880299ab4138: cannot apply new layout on [0x200000404:0x322:0x0] : rc = -22 LustreError: 5413:0:(lov_object.c:1299:lov_layout_change()) Skipped 1 previous similar message LustreError: 5413:0:(lcommon_cl.c:196:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000404:0x322:0x0]: rc = -22 LustreError: 5413:0:(lcommon_cl.c:196:cl_file_inode_init()) Skipped 1 previous similar message LustreError: 5413:0:(llite_lib.c:2983:ll_prep_inode()) new_inode -fatal: rc -22 LustreError: 5413:0:(llite_lib.c:2983:ll_prep_inode()) Skipped 1 previous similar message LustreError: 815:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 272436032 != fo_tot_granted 272464704 LustreError: 815:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 9 previous similar messages LustreError: 815:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 200704 != fo_tot_pending 229376 LustreError: 815:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 9 previous similar messages LustreError: 5532:0:(lov_object.c:1299:lov_layout_change()) lustre-clilov-ffff880299ab4138: cannot apply new layout on [0x200000404:0x322:0x0] : rc = -22 LustreError: 5532:0:(lcommon_cl.c:196:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000404:0x322:0x0]: rc = -22 LustreError: 5532:0:(llite_lib.c:2983:ll_prep_inode()) new_inode -fatal: rc -22 LustreError: 4740:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0000-osc-ffff880299ab4138: namespace resource [0xc30:0x0:0x0].0x0 (ffff880253186fc0) refcount nonzero (1) after lock cleanup; forcing cleanup. Lustre: lustre-OST0000-osc-ffff880299ab4138: Connection restored to 192.168.123.205@tcp (at 0@lo) LustreError: 11-0: lustre-OST0000-osc-ffff880299ab4138: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 1 previous similar message Lustre: lustre-OST0000-osc-ffff880299ab4138: Connection to lustre-OST0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 1 previous similar message LustreError: 167-0: lustre-OST0000-osc-ffff880299ab4138: This client was evicted by lustre-OST0000; in progress operations using this service will fail. LustreError: Skipped 1 previous similar message LustreError: 5643:0:(lov_object.c:1299:lov_layout_change()) lustre-clilov-ffff880299ab4138: cannot apply new layout on [0x200000404:0x322:0x0] : rc = -22 LustreError: 5643:0:(lcommon_cl.c:196:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000404:0x322:0x0]: rc = -22 LustreError: 5643:0:(llite_lib.c:2983:ll_prep_inode()) new_inode -fatal: rc -22 Lustre: 17402:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000404:0x2b7:0x0] with magic=0xbd60bd0 Lustre: 17402:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 1 previous similar message LustreError: 15984:0:(lcommon_cl.c:196:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000404:0x322:0x0]: rc = -22 LustreError: 15984:0:(lcommon_cl.c:196:cl_file_inode_init()) Skipped 1 previous similar message LustreError: 15984:0:(llite_lib.c:2983:ll_prep_inode()) new_inode -fatal: rc -22 LustreError: 15984:0:(llite_lib.c:2983:ll_prep_inode()) Skipped 1 previous similar message LustreError: 6209:0:(lov_object.c:1299:lov_layout_change()) lustre-clilov-ffff880299ab4138: cannot apply new layout on [0x200000404:0x322:0x0] : rc = -22 LustreError: 6209:0:(lov_object.c:1299:lov_layout_change()) Skipped 6 previous similar messages LustreError: 6209:0:(llite_lib.c:2983:ll_prep_inode()) new_inode -fatal: rc -22 LustreError: 6209:0:(llite_lib.c:2983:ll_prep_inode()) Skipped 4 previous similar messages Lustre: ll_ost03_001: service thread pid 12713 was inactive for 40.087 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Lustre: Skipped 1 previous similar message Pid: 12713, comm: ll_ost03_001 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] tgt_extent_lock+0xea/0x2a0 [ptlrpc] [<0>] ofd_getattr_hdl+0x365/0x720 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe LustreError: 6340:0:(lcommon_cl.c:196:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000404:0x322:0x0]: rc = -22 LustreError: 6340:0:(lcommon_cl.c:196:cl_file_inode_init()) Skipped 8 previous similar messages LustreError: 5811:0:(llite_lib.c:2983:ll_prep_inode()) new_inode -fatal: rc -22 LustreError: 5811:0:(llite_lib.c:2983:ll_prep_inode()) Skipped 8 previous similar messages LustreError: 12748:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 279790656 != fo_tot_granted 279819328 LustreError: 12748:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 25 previous similar messages LustreError: 12748:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 458752 != fo_tot_pending 487424 LustreError: 12748:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 25 previous similar messages LustreError: 6929:0:(lov_object.c:1299:lov_layout_change()) lustre-clilov-ffff880299ab4138: cannot apply new layout on [0x200000404:0x322:0x0] : rc = -22 LustreError: 6929:0:(lov_object.c:1299:lov_layout_change()) Skipped 13 previous similar messages Lustre: 27653:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000404:0x312:0x0] with magic=0xbd60bd0 Lustre: 27653:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 3 previous similar messages LustreError: 6809:0:(lcommon_cl.c:196:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000404:0x322:0x0]: rc = -22 LustreError: 6809:0:(lcommon_cl.c:196:cl_file_inode_init()) Skipped 10 previous similar messages LustreError: 11952:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0002-osp-MDT0001: fail to cancel 1 llog-records: rc = -2 LustreError: 11952:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0002-osp-MDT0001: fail to cancel 1 of 1 llog-records: rc = -2 LustreError: 11004:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 150s: evicting client at 0@lo ns: filter-lustre-OST0000_UUID lock: ffff88007dbce940/0xc6b178ed826cf061 lrc: 3/0,0 mode: PW/PW res: [0x815:0x0:0x0].0x0 rrc: 2 type: EXT [0->18446744073709551615] (req 0->524287) gid 0 flags: 0x60000080020020 nid: 0@lo remote: 0xc6b178ed826cefdc expref: 2195 pid: 31781 timeout: 1061 lvb_type: 0 LustreError: 11004:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 1 previous similar message LustreError: 12313:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff880285a16a80 x1710260919089344/t0(0) o105->lustre-OST0000@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 11001:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1631032836 with bad export cookie 14317357651876397185 LustreError: 11001:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) Skipped 5 previous similar messages LustreError: 12313:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 12 previous similar messages LustreError: 11-0: lustre-OST0000-osc-ffff8802998a8958: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 6 previous similar messages Lustre: 4144:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.205@tcp:/lustre/fid: [0x240000404:0x337:0x0]// may get corrupted (rc -108) Lustre: 4145:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.205@tcp:/lustre/fid: [0x280000403:0x2fc:0x0]/ may get corrupted (rc -108) Lustre: 11001:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 7s req@ffff880272bb9940 x1710260919108800/t0(0) o103->9da291a9-717f-4193-8b26-da6e22b22b79@0@lo:0/0 lens 328/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'ldlm_bl_07.0' LustreError: 682:0:(ofd_obd.c:501:ofd_destroy_export()) lustre-OST0000: cli 6c066995-9548-470f-a292-69335ad47963/ffff88008aeaca88 has 28672 pending on destroyed export LustreError: 17894:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0001: failed to get lu_attr of [0x240000404:0x2fe:0x0]: rc = -2 Lustre: ldlm_canceld: This server is not able to keep up with request traffic (cpu-bound). LustreError: 10998:0:(service.c:2290:ptlrpc_server_handle_request()) @@@ Dropping timed-out request from 12345-0@lo: deadline 8/2s ago req@ffff880272bb9940 x1710260919108800/t0(0) o103->9da291a9-717f-4193-8b26-da6e22b22b79@0@lo:305/0 lens 328/0 e 0 to 0 dl 1631032845 ref 2 fl Interpret:H/0/ffffffff rc 0/-1 job:'ldlm_bl_07.0' Lustre: 10998:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (8/2s); client may timeout req@ffff880272bb9940 x1710260919108800/t0(0) o103->9da291a9-717f-4193-8b26-da6e22b22b79@0@lo:305/0 lens 328/0 e 0 to 0 dl 1631032845 ref 2 fl Interpret:H/0/ffffffff rc 0/-1 job:'ldlm_bl_07.0' Lustre: 11001:0:(service.c:1612:ptlrpc_at_check_timed()) earlyQ=1 reqQ=0 recA=0, svcEst=37, delay=0ms Lustre: 11001:0:(service.c:1378:ptlrpc_at_send_early_reply()) @@@ Already past deadline (-2s), not sending early reply. Consider increasing at_early_margin (5)? req@ffff880272bb9940 x1710260919108800/t0(0) o103->9da291a9-717f-4193-8b26-da6e22b22b79@0@lo:305/0 lens 328/0 e 0 to 0 dl 1631032845 ref 1 fl Interpret:H/0/ffffffff rc 0/-1 job:'ldlm_bl_07.0' LustreError: 6819:0:(llite_lib.c:2983:ll_prep_inode()) new_inode -fatal: rc -22 LustreError: 6819:0:(llite_lib.c:2983:ll_prep_inode()) Skipped 9 previous similar messages INFO: task mv:3927 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mv D ffff880325722980 11088 3927 16093 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff81251fe1>] lock_rename+0x31/0xe0 [<ffffffff8125862f>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed432>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817e8a76>] ? trace_do_page_fault+0x56/0x170 [<ffffffff8125975e>] SyS_renameat2+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task chmod:4590 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. chmod D ffff880325721900 11344 4590 16097 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff817d5bb2>] lookup_slow+0x33/0xa7 [<ffffffff81256d6e>] path_lookupat+0x89e/0x8d0 [<ffffffff8121f126>] ? kmem_cache_alloc+0x126/0x2e0 [<ffffffff81257b9f>] ? getname_flags+0x4f/0x1a0 [<ffffffff81256dcb>] filename_lookup+0x2b/0xc0 [<ffffffff81258cc7>] user_path_at_empty+0x67/0xc0 [<ffffffff817e8655>] ? __do_page_fault+0x1c5/0x500 [<ffffffff81258d31>] user_path_at+0x11/0x20 [<ffffffff81244207>] SyS_fchmodat+0x47/0xc0 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task rm:4613 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. rm D ffff88008c701280 11344 4613 15955 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff8125828b>] do_unlinkat+0x13b/0x2b0 [<ffffffff817e8a76>] ? trace_do_page_fault+0x56/0x170 [<ffffffff812592fb>] SyS_unlinkat+0x1b/0x40 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task mv:4754 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mv D ffff88026e028010 11216 4754 15970 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff81251fe1>] lock_rename+0x31/0xe0 [<ffffffff8125862f>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed432>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817e8a76>] ? trace_do_page_fault+0x56/0x170 [<ffffffff8125975e>] SyS_renameat2+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task cat:4810 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. cat D ffff880282b331c0 11360 4810 16150 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffffa1090134>] vvp_io_write_start+0xb44/0xc80 [lustre] [<ffffffffa09d4ae9>] ? lov_lock_enqueue+0x99/0x170 [lov] [<ffffffffa04fa627>] ? cl_lock_request+0x67/0x210 [obdclass] [<ffffffffa04fc8ed>] cl_io_start+0x6d/0x150 [obdclass] [<ffffffffa04fee9f>] cl_io_loop+0x9f/0x210 [obdclass] [<ffffffffa10363cf>] ll_file_io_generic+0x33f/0xee0 [lustre] [<ffffffffa103745c>] ll_file_aio_write+0x4ec/0x800 [lustre] [<ffffffffa1037881>] ll_file_write+0x111/0x1e0 [lustre] [<ffffffffa1037fc1>] ? ll_file_read+0x111/0x1e0 [lustre] [<ffffffff81245996>] vfs_write+0xd6/0x230 [<ffffffff81245812>] ? vfs_read+0x112/0x1c0 [<ffffffff812467df>] SyS_write+0x7f/0xf0 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task file_exec.sh:4822 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. file_exec.sh D ffff88008aca52c0 11248 4822 16378 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff817d5bb2>] lookup_slow+0x33/0xa7 [<ffffffff81256d6e>] path_lookupat+0x89e/0x8d0 [<ffffffff8121f126>] ? kmem_cache_alloc+0x126/0x2e0 [<ffffffff81257b9f>] ? getname_flags+0x4f/0x1a0 [<ffffffff81256dcb>] filename_lookup+0x2b/0xc0 [<ffffffff81258cc7>] user_path_at_empty+0x67/0xc0 [<ffffffff81258d31>] user_path_at+0x11/0x20 [<ffffffff8124b6b3>] vfs_fstatat+0x63/0xc0 [<ffffffff8124ba7e>] SYSC_newstat+0x2e/0x60 [<ffffffff817e8a76>] ? trace_do_page_fault+0x56/0x170 Lustre: 4152:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.205@tcp:/lustre/fid: [0x240000403:0x37d:0x0]// may get corrupted (rc -108) [<ffffffff8124bf4e>] SyS_newstat+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task rm:4953 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. rm D ffff880295a01280 11344 4953 16217 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff812580b5>] do_rmdir+0x165/0x200 [<ffffffff810b69ad>] ? task_work_run+0xcd/0xf0 [<ffffffff81259305>] SyS_unlinkat+0x25/0x40 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task chmod:5181 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. chmod D ffff88028c1d24f0 11344 5181 15987 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff817d5bb2>] lookup_slow+0x33/0xa7 [<ffffffff81256d6e>] path_lookupat+0x89e/0x8d0 [<ffffffff8121f126>] ? kmem_cache_alloc+0x126/0x2e0 [<ffffffff81257b9f>] ? getname_flags+0x4f/0x1a0 [<ffffffff81256dcb>] filename_lookup+0x2b/0xc0 [<ffffffff81258cc7>] user_path_at_empty+0x67/0xc0 [<ffffffff817e8655>] ? __do_page_fault+0x1c5/0x500 [<ffffffff81258d31>] user_path_at+0x11/0x20 [<ffffffff81244207>] SyS_fchmodat+0x47/0xc0 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task getfattr:5245 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. getfattr D ffff88029445a140 12736 5245 16271 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff817d5bb2>] lookup_slow+0x33/0xa7 [<ffffffff81256d6e>] path_lookupat+0x89e/0x8d0 [<ffffffff8121f126>] ? kmem_cache_alloc+0x126/0x2e0 [<ffffffff81257b9f>] ? getname_flags+0x4f/0x1a0 [<ffffffff81256dcb>] filename_lookup+0x2b/0xc0 [<ffffffff81258cc7>] user_path_at_empty+0x67/0xc0 [<ffffffff81258d31>] user_path_at+0x11/0x20 [<ffffffff8124b6b3>] vfs_fstatat+0x63/0xc0 [<ffffffff8124bae1>] SYSC_newlstat+0x31/0x60 [<ffffffff817e8a76>] ? trace_do_page_fault+0x56/0x170 [<ffffffff8124bf5e>] SyS_newlstat+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task ln:5282 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. ln D ffff880293f71900 12760 5282 16051 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff817d5bb2>] lookup_slow+0x33/0xa7 [<ffffffff81256d6e>] path_lookupat+0x89e/0x8d0 [<ffffffff8121f126>] ? kmem_cache_alloc+0x126/0x2e0 [<ffffffff81257b9f>] ? getname_flags+0x4f/0x1a0 [<ffffffff81256dcb>] filename_lookup+0x2b/0xc0 [<ffffffff81258cc7>] user_path_at_empty+0x67/0xc0 [<ffffffff81258d31>] user_path_at+0x11/0x20 [<ffffffff8124b6b3>] vfs_fstatat+0x63/0xc0 [<ffffffff8124ba7e>] SYSC_newstat+0x2e/0x60 [<ffffffff817e8a76>] ? trace_do_page_fault+0x56/0x170 [<ffffffff8124bf4e>] SyS_newstat+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 LustreError: 7347:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0000-osc-ffff8802998a8958: namespace resource [0x1550:0x0:0x0].0x0 (ffff8802749b34c0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 7347:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 137 previous similar messages Lustre: lustre-OST0000: haven't heard from client 6c066995-9548-470f-a292-69335ad47963 (at 0@lo) in 47 seconds. I think it's dead, and I am evicting it. exp ffff880255ef2e98, cur 1631032884 expire 1631032854 last 1631032837 LustreError: 7707:0:(lov_object.c:1299:lov_layout_change()) lustre-clilov-ffff8802998a8958: cannot apply new layout on [0x200000404:0x322:0x0] : rc = -22 Lustre: 4153:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1631032837/real 1631032837] req@ffff88025724c500 x1710260919108800/t0(0) o103->lustre-OST0001-osc-ffff880299ab4138@0@lo:17/18 lens 328/224 e 0 to 1 dl 1631032870 ref 1 fl Rpc:XQr/0/ffffffff rc 0/-1 job:'ldlm_bl_07.0' Lustre: lustre-OST0001-osc-ffff880299ab4138: Connection to lustre-OST0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 1 previous similar message LustreError: 7707:0:(lov_object.c:1299:lov_layout_change()) Skipped 9 previous similar messages Lustre: lustre-OST0001: Client 9da291a9-717f-4193-8b26-da6e22b22b79 (at 0@lo) reconnecting Lustre: lustre-OST0001-osc-ffff880299ab4138: Connection restored to 192.168.123.205@tcp (at 0@lo) Lustre: Skipped 1 previous similar message LustreError: 7803:0:(lcommon_cl.c:196:cl_file_inode_init()) lustre: failed to initialize cl_object [0x200000404:0x322:0x0]: rc = -22 LustreError: 7803:0:(lcommon_cl.c:196:cl_file_inode_init()) Skipped 9 previous similar messages LustreError: 11149:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0001: '7' migration was interrupted, run 'lfs migrate -m 2 -c 1 -H crush 7' to finish migration. LustreError: 167-0: lustre-OST0000-osc-ffff8802998a8958: This client was evicted by lustre-OST0000; in progress operations using this service will fail. LustreError: Skipped 1 previous similar message Lustre: 26049:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000404:0x335:0x0] with magic=0xbd60bd0 Lustre: 26049:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 1 previous similar message LustreError: 7293:0:(llite_lib.c:2983:ll_prep_inode()) new_inode -fatal: rc -22 LustreError: 7293:0:(llite_lib.c:2983:ll_prep_inode()) Skipped 8 previous similar messages Lustre: ll_ost05_037: service thread pid 31787 was inactive for 40.324 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 14 previous similar messages Lustre: mdt05_004: service thread pid 17469 was inactive for 54.057 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 7 previous similar messages LustreError: 12749:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 84143552 != fo_tot_granted 84172224 LustreError: 12749:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 71 previous similar messages LustreError: 12749:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 28672 LustreError: 12749:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 71 previous similar messages ------------[ cut here ]------------ WARNING: CPU: 3 PID: 26746 at /home/green/git/lustre-release/ldiskfs/ext4_jbd2.c:266 __ldiskfs_handle_dirty_metadata+0x1d2/0x230 [ldiskfs] Modules linked in: loop zfs(PO) zunicode(PO) zzstd(O) zlua(O) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) jbd2 mbcache lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) dm_flakey dm_mod libcfs(OE) crc_t10dif crct10dif_generic sb_edac edac_core iosf_mbi crc32_pclmul ghash_clmulni_intel aesni_intel lrw gf128mul glue_helper ablk_helper cryptd pcspkr virtio_console virtio_balloon i2c_piix4 ip_tables rpcsec_gss_krb5 drm_kms_helper ata_generic ttm pata_acpi crct10dif_pclmul crct10dif_common drm crc32c_intel drm_panel_orientation_quirks ata_piix serio_raw virtio_blk libata i2c_core floppy CPU: 3 PID: 26746 Comm: mdt01_013 Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011 Call Trace: [<ffffffff817d9418>] dump_stack+0x19/0x1b [<ffffffff8108d538>] __warn+0xd8/0x100 [<ffffffff8108d67d>] warn_slowpath_null+0x1d/0x20 [<ffffffffa0c24262>] __ldiskfs_handle_dirty_metadata+0x1d2/0x230 [ldiskfs] [<ffffffffa0c0433d>] ldiskfs_getblk+0x13d/0x210 [ldiskfs] [<ffffffffa0c04437>] ldiskfs_bread+0x27/0xe0 [ldiskfs] [<ffffffffa0cc8e97>] osd_write+0x4e7/0xcb0 [osd_ldiskfs] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffffa04ee6f3>] dt_record_write+0x33/0x120 [obdclass] [<ffffffffa04ad6b2>] llog_osd_write_rec+0x1052/0x19a0 [obdclass] [<ffffffffa049a760>] llog_write_rec+0x290/0x590 [obdclass] [<ffffffffa04a04ee>] llog_cat_add_rec+0x23e/0x950 [obdclass] [<ffffffffa04975ff>] llog_add+0x17f/0x1f0 [obdclass] [<ffffffffa04c8a39>] ? lprocfs_counter_add+0xf9/0x160 [obdclass] [<ffffffffa083160d>] sub_updates_write+0xb6b/0xe32 [ptlrpc] [<ffffffffa0801892>] top_trans_stop+0x7c2/0xf60 [ptlrpc] [<ffffffffa0ee4f3c>] lod_trans_stop+0x25c/0x340 [lod] [<ffffffffa0db9408>] mdd_trans_stop+0x28/0x16e [mdd] [<ffffffffa0d9ab2f>] mdd_migrate_object+0xf1f/0x1ba0 [mdd] [<ffffffffa04ed375>] ? lu_object_find_at+0x285/0xb10 [obdclass] [<ffffffffa0d9bb86>] mdd_migrate+0x3d6/0x8e0 [mdd] [<ffffffffa0e6d0d8>] mdo_migrate+0x4c/0x4e [mdt] [<ffffffffa0e2ce53>] mdt_reint_migrate+0x1023/0x11d0 [mdt] [<ffffffffa0e2d087>] mdt_reint_rec+0x87/0x240 [mdt] [<ffffffffa0e0255c>] mdt_reint_internal+0x76c/0xb40 [mdt] [<ffffffffa0e0f657>] mdt_reint+0x67/0x150 [mdt] [<ffffffffa07f026e>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<ffffffffa033545e>] ? libcfs_nid2str_r+0xfe/0x130 [lnet] [<ffffffffa079b990>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<ffffffffa079d529>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e32be>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffffa079c950>] ? ptlrpc_wait_event+0x620/0x620 [ptlrpc] [<ffffffff810ba0f4>] kthread+0xe4/0xf0 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 ---[ end trace 647798641f8bde2f ]--- LDISKFS-fs: ldiskfs_getblk:888: aborting transaction: error 28 in __ldiskfs_handle_dirty_metadata LDISKFS-fs error (device dm-0): ldiskfs_getblk:888: inode #206: block 53380: comm mdt01_013: journal_dirty_metadata failed: handle type 0 started at line 1972, credits 512/0, errcode -28 Aborting journal on device dm-0-8. LDISKFS-fs (dm-0): Remounting filesystem read-only LustreError: 26746:0:(osd_io.c:2103:osd_ldiskfs_write_record()) lustre-MDT0000: error reading offset 204800 (block 50, size 40, offs 204784), credits 0/0: rc = -28 LustreError: 26746:0:(update_trans.c:985:top_trans_stop()) lustre-MDT0000-osd: write updates failed: rc = -28 LustreError: 26746:0:(osd_io.c:2110:osd_ldiskfs_write_record()) journal_get_write_access() returned error -30 LustreError: 26746:0:(tgt_lastrcvd.c:1253:tgt_add_reply_data()) lustre-MDT0000: can't update reply_data file: rc = -30 LustreError: 26746:0:(osd_handler.c:2075:osd_trans_stop()) lustre-MDT0000: failed in transaction hook: rc = -30 LustreError: 13739:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-OST0001-osc-MDT0000: fail to cancel 23 llog-records: rc = -30 LustreError: 8688:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -30 LDISKFS-fs error (device dm-0) in osd_trans_stop:2082: error 28 LustreError: 13739:0:(osp_sync.c:1094:osp_sync_process_committed()) lustre-OST0001-osc-MDT0000: can't cancel 23 records: rc = -30 LustreError: 13739:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-OST0001-osc-MDT0000: fail to cancel 1 of 1 llog-records: rc = -30 LustreError: 13739:0:(osp_sync.c:1079:osp_sync_process_committed()) lustre-OST0001-osc-MDT0000: can't cancel record: rc = -30 LustreError: 26746:0:(osd_handler.c:2085:osd_trans_stop()) lustre-MDT0000: failed to stop transaction: rc = -28 LustreError: 10941:0:(osd_handler.c:1779:osd_trans_commit_cb()) transaction @0xffff8802796414d8 commit error: 2 LDISKFS-fs error (device dm-0) in osd_trans_stop:2082: IO failure LustreError: 26746:0:(update_trans.c:1011:top_trans_stop()) lustre-MDT0000-osd: stop trans failed: rc = -30 LustreError: 8754:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -30 LustreError: 8754:0:(llite_lib.c:1836:ll_md_setattr()) Skipped 3 previous similar messages Lustre: ll_ost06_030: service thread pid 5603 was inactive for 86.155 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. LustreError: 8309:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -30 LustreError: 8309:0:(llite_lib.c:1836:ll_md_setattr()) Skipped 1 previous similar message Lustre: 4143:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff880251803240 x1710260924035904/t0(0) o10->lustre-OST0000-osc-ffff8802998a8958@0@lo:6/4 lens 440/432 e 0 to 0 dl 1631033034 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'cp.0' Lustre: ll_ost06_024: service thread pid 5535 was inactive for 89.494 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 1 previous similar message LustreError: 8914:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -30 LustreError: 8914:0:(llite_lib.c:1836:ll_md_setattr()) Skipped 2 previous similar messages LustreError: 11004:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 116s: evicting client at 0@lo ns: filter-lustre-OST0002_UUID lock: ffff8802778b6d00/0xc6b178ed827a4478 lrc: 3/0,0 mode: PW/PW res: [0x340000400:0x691:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 2097152->18446744073709551615) gid 0 flags: 0x60000000030020 nid: 0@lo remote: 0xc6b178ed827a4471 expref: 1779 pid: 26418 timeout: 1186 lvb_type: 0 LustreError: 20913:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff880257600040 x1710260924118080/t0(0) o105->lustre-OST0002@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 20913:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 21 previous similar messages LustreError: 11-0: lustre-OST0002-osc-ffff880299ab4138: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 14 previous similar messages LustreError: 14411:0:(osp_sync.c:1079:osp_sync_process_committed()) lustre-OST0003-osc-MDT0000: can't cancel record: rc = -30 LustreError: 14411:0:(osp_sync.c:1079:osp_sync_process_committed()) Skipped 1 previous similar message LustreError: 6083:0:(ofd_obd.c:501:ofd_destroy_export()) lustre-OST0002: cli 9da291a9-717f-4193-8b26-da6e22b22b79/ffff8800786d2e98 has 28672 pending on destroyed export LustreError: 8782:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -30 LustreError: 8782:0:(llite_lib.c:1836:ll_md_setattr()) Skipped 7 previous similar messages Lustre: 17488:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000404:0x372:0x0] with magic=0xbd60bd0 Lustre: 17488:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 1 previous similar message 6[8812]: segfault at 0 ip (null) sp 00007ffd422c0d08 error 14 in 6[400000+6000] Lustre: 4142:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1631032890/real 1631032890] req@ffff880078107700 x1710260920730944/t0(0) o103->lustre-OST0001-osc-ffff880299ab4138@0@lo:17/18 lens 328/224 e 0 to 1 dl 1631032939 ref 1 fl Rpc:XQr/2/ffffffff rc -11/-1 job:'ldlm_bl_05.0' Lustre: lustre-OST0001: Client 9da291a9-717f-4193-8b26-da6e22b22b79 (at 0@lo) reconnecting LustreError: 24467:0:(mdt_open.c:1237:mdt_cross_open()) lustre-MDT0001: [0x240000404:0x408:0x0] doesn't exist!: rc = -14 LustreError: 11119:0:(mdt_open.c:1237:mdt_cross_open()) lustre-MDT0001: [0x240000404:0x408:0x0] doesn't exist!: rc = -14 LustreError: 11119:0:(mdt_open.c:1237:mdt_cross_open()) Skipped 5 previous similar messages LustreError: 11952:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0000-osp-MDT0001: fail to cancel 1 llog-records: rc = -116 LustreError: 11952:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) Skipped 12 previous similar messages LustreError: 11952:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0000-osp-MDT0001: fail to cancel 1 of 1 llog-records: rc = -116 LustreError: 11952:0:(llog_cat.c:787:llog_cat_cancel_records()) Skipped 11 previous similar messages LustreError: 9372:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -30 LustreError: 9372:0:(llite_lib.c:1836:ll_md_setattr()) Skipped 4 previous similar messages LustreError: 19828:0:(mdt_open.c:1237:mdt_cross_open()) lustre-MDT0001: [0x240000404:0x408:0x0] doesn't exist!: rc = -14 LustreError: 19828:0:(mdt_open.c:1237:mdt_cross_open()) Skipped 1 previous similar message LustreError: 8859:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -30 Lustre: 4143:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff880251803240 x1710260925262272/t0(0) o10->lustre-OST0000-osc-ffff8802998a8958@0@lo:6/4 lens 440/432 e 1 to 0 dl 1631033061 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'cp.0' LustreError: 11952:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0000-osp-MDT0001: fail to cancel 1 llog-records: rc = -116 LustreError: 11952:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) Skipped 9 previous similar messages LustreError: 11952:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0000-osp-MDT0001: fail to cancel 1 of 1 llog-records: rc = -116 LustreError: 11952:0:(llog_cat.c:787:llog_cat_cancel_records()) Skipped 9 previous similar messages LustreError: 16071:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -30 LustreError: 16071:0:(llite_lib.c:1836:ll_md_setattr()) Skipped 4 previous similar messages | Externally reported by onyx-68 boilpot email |
racer test 1: racer on clients: centos-0.localnet DURATION=2700 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa1051e52>] ll_migrate+0x9b2/0xec0 [lustre] PGD 800000028b3a6067 PUD 28b3a7067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: loop zfs(PO) zunicode(PO) zzstd(O) zlua(O) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) jbd2 mbcache lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) dm_flakey dm_mod libcfs(OE) crc_t10dif crct10dif_generic sb_edac edac_core iosf_mbi crc32_pclmul ghash_clmulni_intel aesni_intel lrw gf128mul glue_helper ablk_helper cryptd virtio_balloon virtio_console pcspkr i2c_piix4 ip_tables rpcsec_gss_krb5 drm_kms_helper ttm ata_generic pata_acpi drm crct10dif_pclmul crct10dif_common crc32c_intel drm_panel_orientation_quirks ata_piix serio_raw virtio_blk i2c_core libata floppy CPU: 6 PID: 25076 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011 task: ffff8802912324f0 ti: ffff88024ad2c000 task.ti: ffff88024ad2c000 RIP: 0010:[<ffffffffa1051e52>] [<ffffffffa1051e52>] ll_migrate+0x9b2/0xec0 [lustre] RSP: 0018:ffff88024ad2fbc8 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff880253c01458 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff880327331138 RDI: ffff880327331118 RBP: ffff88024ad2fc48 R08: ffff8802668e2058 R09: 0000000000000001 R10: 0000000000000000 R11: ffff88024ad2f5e6 R12: 0000000000000000 R13: ffff8800848d48e8 R14: ffff88029b7166d8 R15: 0000000000000030 FS: 00007fa279316740(0000) GS:ffff880331b80000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 CR2: 0000000000000008 CR3: 0000000266662000 CR4: 00000000001607e0 Call Trace: [<ffffffff81242213>] ? __check_object_size+0x1c3/0x220 [<ffffffffa103f561>] ll_dir_ioctl+0x5d01/0x6ed0 [lustre] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff8115260f>] ? delayacct_end+0x8f/0xb0 [<ffffffff81152744>] ? __delayacct_blkio_end+0x34/0x60 [<ffffffff817e0257>] ? io_schedule_timeout+0xe7/0x130 [<ffffffff811b62dd>] ? find_get_pages_tag+0x10d/0x260 [<ffffffff811c3691>] ? pagevec_lookup_tag+0x21/0x30 [<ffffffff811b400e>] ? __filemap_fdatawait_range+0xbe/0x1a0 [<ffffffff8125b3fd>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff81264d2b>] ? iput+0x3b/0x180 [<ffffffff8125b751>] SyS_ioctl+0xa1/0xc0 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 | Lustre: lustre-OST0000-osc-ffff88029a0e8008: disconnect after 20s idle Lustre: 11112:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 512 < left 828, rollback = 7 Lustre: 11112:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 11112:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 1/89/0 Lustre: 11112:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 819/828/0, punch: 0/0/0, quota 3/3/0 Lustre: 11112:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 11112:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 11112:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 511 < left 827, rollback = 7 Lustre: 11112:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 299 previous similar messages Lustre: 11112:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 11112:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 299 previous similar messages Lustre: 11112:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 1/89/0 Lustre: 11112:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 299 previous similar messages Lustre: 11112:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 819/828/1, punch: 0/0/0, quota 3/3/0 Lustre: 11112:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 299 previous similar messages Lustre: 11112:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 11112:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 299 previous similar messages Lustre: 11112:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 11112:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 299 previous similar messages Lustre: 11112:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 511 < left 827, rollback = 7 Lustre: 11112:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 242 previous similar messages Lustre: 11112:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 11112:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 242 previous similar messages Lustre: 11112:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 1/89/0 Lustre: 11112:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 242 previous similar messages Lustre: 11112:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 819/828/1, punch: 0/0/0, quota 3/3/0 Lustre: 11112:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 242 previous similar messages Lustre: 11112:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 11112:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 242 previous similar messages Lustre: 11112:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 11112:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 242 previous similar messages Lustre: 11100:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 506 < left 894, rollback = 2 Lustre: 11100:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 2/8/2, destroy: 1/4/0 Lustre: 11100:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 401 previous similar messages Lustre: 11100:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 2/2/0, xattr_set: 11/894/0 Lustre: 11100:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 401 previous similar messages Lustre: 11100:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1/10/0, punch: 0/0/0, quota 6/6/0 Lustre: 11100:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 401 previous similar messages Lustre: 11100:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 12/194/4, delete: 3/6/1 Lustre: 11100:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 401 previous similar messages Lustre: 11100:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 7/7/0, ref_del: 3/3/0 Lustre: 11100:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 401 previous similar messages Lustre: 11100:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 505 < left 894, rollback = 2 Lustre: 11100:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 1 previous similar message Lustre: 11086:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 507 < left 610, rollback = 2 Lustre: 11086:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 6 previous similar messages cat (16892) used greatest stack depth: 10240 bytes left Lustre: 17299:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 502 < left 671, rollback = 2 Lustre: 17299:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 5 previous similar messages Lustre: 17299:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 3/12/6, destroy: 0/0/0 Lustre: 17299:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 13 previous similar messages Lustre: 17299:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 7/671/0 Lustre: 17299:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 13 previous similar messages Lustre: 17299:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 5/136/0, punch: 0/0/0, quota 8/56/0 Lustre: 17299:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 13 previous similar messages Lustre: 17299:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 10/203/4, delete: 0/0/0 Lustre: 17299:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 13 previous similar messages Lustre: 17299:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 4/4/0, ref_del: 0/0/0 Lustre: 17299:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 13 previous similar messages Lustre: 11111:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 512 < left 874, rollback = 7 Lustre: 11111:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 400 previous similar messages LustreError: 17991:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff88009e7fb7e8: inode [0x200000404:0x19:0x0] mdc close failed: rc = -13 Lustre: 17963:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 507 < left 760, rollback = 2 Lustre: 17963:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 9 previous similar messages Lustre: 17341:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 494 < left 1329, rollback = 7 Lustre: 17341:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 4771 previous similar messages Lustre: 11094:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 190/760/0, destroy: 0/0/0 Lustre: 11094:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 4933 previous similar messages Lustre: 11094:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/1, xattr_set: 192/15/0 Lustre: 11094:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 4933 previous similar messages Lustre: 11094:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 951/8180/0, punch: 0/0/0, quota 14/174/3 Lustre: 11094:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 4933 previous similar messages Lustre: 11094:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 190/3230/0, delete: 0/0/0 Lustre: 11094:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 4933 previous similar messages Lustre: 11094:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 11094:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 4933 previous similar messages Lustre: 4120:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff88008e8a3240 x1710205953078720/t0(0) o10->lustre-OST0000-osc-ffff88029a0e8008@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630979683 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'truncate.0' Lustre: 16762:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 503 < left 1250, rollback = 2 Lustre: 16762:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 15 previous similar messages Lustre: 16729:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 512 < left 784, rollback = 7 Lustre: 16729:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 1667 previous similar messages Lustre: 17637:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 1/4/4, destroy: 0/0/0 Lustre: 17637:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 6551 previous similar messages Lustre: 17637:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 4/209/0 Lustre: 17637:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 6552 previous similar messages Lustre: 17637:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1553/1562/0, punch: 0/0/0, quota 6/6/2 Lustre: 17637:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 6552 previous similar messages Lustre: 17637:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 2/33/1, delete: 0/0/0 Lustre: 17637:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 6552 previous similar messages Lustre: 17637:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 17637:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 6552 previous similar messages Lustre: 11107:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 505 < left 1562, rollback = 7 Lustre: 11107:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 12936 previous similar messages ------------[ cut here ]------------ WARNING: CPU: 11 PID: 11132 at /home/green/git/lustre-release/ldiskfs/namei.c:3331 ldiskfs_orphan_add+0x11e/0x2a0 [ldiskfs] Modules linked in: loop zfs(PO) zunicode(PO) zzstd(O) zlua(O) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) jbd2 mbcache lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) dm_flakey dm_mod libcfs(OE) crc_t10dif crct10dif_generic sb_edac edac_core iosf_mbi crc32_pclmul ghash_clmulni_intel aesni_intel lrw gf128mul glue_helper ablk_helper cryptd virtio_balloon virtio_console pcspkr i2c_piix4 ip_tables rpcsec_gss_krb5 drm_kms_helper ttm ata_generic pata_acpi drm crct10dif_pclmul crct10dif_common crc32c_intel drm_panel_orientation_quirks ata_piix serio_raw virtio_blk i2c_core libata floppy CPU: 11 PID: 11132 Comm: mdt_rdpg05_001 Tainted: P OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011 Call Trace: [<ffffffff817d9418>] dump_stack+0x19/0x1b [<ffffffff8108d538>] __warn+0xd8/0x100 [<ffffffff8108d67d>] warn_slowpath_null+0x1d/0x20 [<ffffffffa0be94ae>] ldiskfs_orphan_add+0x11e/0x2a0 [ldiskfs] [<ffffffffa0bda9eb>] ldiskfs_xattr_inode_orphan_add.constprop.17+0xbb/0x110 [ldiskfs] [<ffffffffa0bdd417>] ldiskfs_xattr_delete_inode+0x77/0x3c0 [ldiskfs] [<ffffffffa0bfdaa0>] ? ldiskfs_evict_inode+0x190/0x670 [ldiskfs] [<ffffffffa0bfdad9>] ldiskfs_evict_inode+0x1c9/0x670 [ldiskfs] [<ffffffff812649af>] evict+0xaf/0x180 [<ffffffff81264de5>] iput+0xf5/0x180 [<ffffffffa0c92d08>] osd_object_delete+0x1f8/0x370 [osd_ldiskfs] [<ffffffffa04c559d>] lu_object_free.isra.26+0xcd/0x1d0 [obdclass] [<ffffffffa04c9cf1>] lu_object_put+0xa1/0x480 [obdclass] [<ffffffffa0e4aaf4>] mdt_mfd_close+0x1f4/0x870 [mdt] [<ffffffffa0e50e14>] mdt_close_internal+0xb4/0x240 [mdt] [<ffffffffa0e51224>] mdt_close+0x284/0x850 [mdt] [<ffffffffa07cc2d4>] ? tgt_request_preprocess.isra.27+0xf4/0x850 [ptlrpc] [<ffffffffa07cd26e>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<ffffffffa03b145e>] ? libcfs_nid2str_r+0xfe/0x130 [lnet] [<ffffffffa0778990>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<ffffffffa077a529>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffffa0779950>] ? ptlrpc_wait_event+0x620/0x620 [ptlrpc] [<ffffffff810ba0f4>] kthread+0xe4/0xf0 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 ---[ end trace dc8986bfb67aba52 ]--- Lustre: 19174:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 500 < left 788, rollback = 2 Lustre: 19174:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 14 previous similar messages Lustre: 19174:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 3/12/8, destroy: 0/0/0 Lustre: 19174:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 17872 previous similar messages Lustre: 19174:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 2/2/0, xattr_set: 10/788/0 Lustre: 19174:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 17872 previous similar messages Lustre: 19174:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 5/136/0, punch: 0/0/0, quota 8/56/0 Lustre: 19174:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 17872 previous similar messages Lustre: 19174:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 12/215/4, delete: 1/1/0 Lustre: 19174:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 17872 previous similar messages Lustre: 19174:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 6/6/0, ref_del: 1/1/0 Lustre: 19174:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 17872 previous similar messages Lustre: 18995:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1630979682/real 1630979682] req@ffff88029cd79300 x1710205953123392/t0(0) o101->lustre-MDT0000-mdc-ffff88009e7fb7e8@0@lo:12/10 lens 576/17800 e 0 to 1 dl 1630979728 ref 2 fl Rpc:XQr/2/ffffffff rc 0/-1 job:'rm.0' Lustre: lustre-MDT0000-mdc-ffff88009e7fb7e8: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: lustre-MDT0000: Client 0ac4ffd4-9906-4a09-82f1-8303128482e2 (at 0@lo) reconnecting Lustre: lustre-MDT0000-mdc-ffff88009e7fb7e8: Connection restored to 192.168.123.100@tcp (at 0@lo) Lustre: 4116:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff88029d2ef0c0 x1710205955484224/t0(0) o10->lustre-OST0003-osc-ffff88009e7fb7e8@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630979793 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'cp.0' Lustre: 17285:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 506 < left 29680, rollback = 7 Lustre: 17285:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 9941 previous similar messages sched: RT throttling activated Lustre: 19174:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 8: before 509 < left 3246, rollback = 9 Lustre: 19174:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 37 previous similar messages LustreError: 10969:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0002_UUID lock: ffff88027901d2c0/0xb337f8daa20e8926 lrc: 3/0,0 mode: PW/PW res: [0x115:0x0:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000400000020 nid: 0@lo remote: 0xb337f8daa20e88cb expref: 1128 pid: 12691 timeout: 402 lvb_type: 0 LustreError: 19579:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff880273090cc0 x1710205957669120/t0(0) o106->lustre-OST0002@0@lo:15/16 lens 328/280 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 12692:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88028feeea80 x1710205957692160/t0(0) o106->lustre-OST0002@0@lo:15/16 lens 328/280 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 12692:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 2 previous similar messages LustreError: 11-0: lustre-OST0002-osc-ffff88029a0e8008: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-OST0002-osc-ffff88029a0e8008: Connection to lustre-OST0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0002-osc-ffff88029a0e8008: This client was evicted by lustre-OST0002; in progress operations using this service will fail. Lustre: 4108:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.100@tcp:/lustre/fid: [0x240000403:0xb7:0x0]// may get corrupted (rc -108) Lustre: 4109:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.100@tcp:/lustre/fid: [0x200000404:0xb5:0x0]// may get corrupted (rc -108) Lustre: 4109:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.100@tcp:/lustre/fid: [0x240000403:0x9e:0x0]// may get corrupted (rc -108) Lustre: 4108:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.100@tcp:/lustre/fid: [0x200000403:0x25:0x0]// may get corrupted (rc -108) Lustre: 4109:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.100@tcp:/lustre/fid: [0x280000404:0x51:0x0]/ may get corrupted (rc -108) LustreError: 10969:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0003_UUID lock: ffff88028de16580/0xb337f8daa20cfff2 lrc: 3/0,0 mode: PW/PW res: [0x380000400:0x6:0x0].0x0 rrc: 4 type: EXT [0->18446744073709551615] (req 0->1310719) gid 0 flags: 0x60000480020020 nid: 0@lo remote: 0xb337f8daa20cffac expref: 1323 pid: 12703 timeout: 410 lvb_type: 0 LustreError: 11-0: lustre-OST0003-osc-ffff88009e7fb7e8: operation ost_setattr to node 0@lo failed: rc = -107 LustreError: Skipped 1 previous similar message Lustre: lustre-OST0003-osc-ffff88009e7fb7e8: Connection to lustre-OST0003 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0003-osc-ffff88009e7fb7e8: This client was evicted by lustre-OST0003; in progress operations using this service will fail. Lustre: 4114:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.100@tcp:/lustre/fid: [0x200000404:0xbc:0x0]/ may get corrupted (rc -5) Lustre: 4114:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.100@tcp:/lustre/fid: [0x240000404:0xe6:0x0]// may get corrupted (rc -108) Lustre: 4107:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.100@tcp:/lustre/fid: [0x280000403:0x29:0x0]// may get corrupted (rc -5) Lustre: 4111:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.100@tcp:/lustre/fid: [0x240000404:0xb6:0x0]/ may get corrupted (rc -108) Lustre: 4111:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.100@tcp:/lustre/fid: [0x280000403:0x63:0x0]/ may get corrupted (rc -108) Lustre: 21437:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 3/12/12, destroy: 0/0/0 Lustre: 21437:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 23656 previous similar messages Lustre: 21437:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 7/671/0 Lustre: 21437:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 23656 previous similar messages Lustre: 21437:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 5/136/0, punch: 0/0/0, quota 8/56/0 Lustre: 21437:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 23656 previous similar messages Lustre: 21437:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 10/203/4, delete: 0/0/0 Lustre: 21437:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 23656 previous similar messages Lustre: 21437:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 4/4/0, ref_del: 0/0/0 Lustre: 21437:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 23655 previous similar messages Lustre: 4110:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.100@tcp:/lustre/fid: [0x280000403:0x76:0x0]// may get corrupted (rc -5) LustreError: 22220:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0002-osc-ffff88029a0e8008: namespace resource [0x340000400:0xe0:0x0].0x0 (ffff88026f3d96c0) refcount nonzero (1) after lock cleanup; forcing cleanup. Lustre: lustre-OST0002-osc-ffff88029a0e8008: Connection restored to 192.168.123.100@tcp (at 0@lo) LustreError: 22261:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0003-osc-ffff88009e7fb7e8: namespace resource [0x380000400:0xe0:0x0].0x0 (ffff880283824040) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 22261:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 13 previous similar messages Lustre: 17341:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 510 < left 29680, rollback = 7 Lustre: 17341:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 23596 previous similar messages LustreError: 18861:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0000: '1' migration was interrupted, run 'lfs migrate -m 1 -c 1 -H crush 1' to finish migration. Lustre: 14273:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 8: before 512 < left 21691, rollback = 9 Lustre: 14273:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 73 previous similar messages LustreError: 10969:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 102s: evicting client at 0@lo ns: filter-lustre-OST0000_UUID lock: ffff88007af21300/0xb337f8daa213e7f0 lrc: 3/0,0 mode: PW/PW res: [0x2c0000400:0xc7:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 2097152->18446744073709551615) gid 0 flags: 0x60000000020020 nid: 0@lo remote: 0xb337f8daa213e7cd expref: 2035 pid: 19588 timeout: 520 lvb_type: 0 LustreError: 21373:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff880270b3d7c0 x1710205961467008/t0(0) o106->lustre-OST0000@0@lo:15/16 lens 328/280 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 21373:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 6 previous similar messages LustreError: 11-0: lustre-OST0000-osc-ffff88009e7fb7e8: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 3 previous similar messages Lustre: lustre-OST0000-osc-ffff88009e7fb7e8: Connection to lustre-OST0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 20642:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88024a64f700 x1710205961541632/t0(0) o104->lustre-OST0000@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 20642:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 40 previous similar messages LustreError: 167-0: lustre-OST0000-osc-ffff88009e7fb7e8: This client was evicted by lustre-OST0000; in progress operations using this service will fail. LustreError: 24579:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0000-osc-ffff88009e7fb7e8: namespace resource [0x2c0000401:0x7cf:0x0].0x0 (ffff880265ca34c0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 24579:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 579 previous similar messages Lustre: lustre-OST0000-osc-ffff88009e7fb7e8: Connection restored to 192.168.123.100@tcp (at 0@lo) Lustre: Skipped 1 previous similar message Lustre: mdt00_006: service thread pid 17514 was inactive for 52.937 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 17514, comm: mdt00_006 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] lod_object_lock+0x351/0x7d0 [lod] [<0>] mdd_object_lock+0x2d/0xd0 [mdd] [<0>] mdt_reint_striped_lock+0x378/0x5b0 [mdt] [<0>] mdt_migrate_parent_lock+0x3f/0xb0 [mdt] [<0>] mdt_reint_migrate+0x8b7/0x11d0 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x76c/0xb40 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: 16560:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 1552/6208/0, destroy: 1/4/0 Lustre: 16560:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 32448 previous similar messages Lustre: 16560:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 3/3/1, xattr_set: 1556/148/0 Lustre: 16560:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 32448 previous similar messages Lustre: 16560:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 7761/66746/0, punch: 0/0/0, quota 8/56/0 Lustre: 16560:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 32448 previous similar messages Lustre: 16560:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 1553/26400/0, delete: 2/5/1 Lustre: 16560:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 32448 previous similar messages Lustre: 16560:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 1/1/0, ref_del: 2/2/1 Lustre: 16560:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 32448 previous similar messages Lustre: 17637:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 508 < left 15533, rollback = 7 Lustre: 17637:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 38590 previous similar messages Lustre: 4119:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff880272ef8040 x1710205963279744/t0(0) o10->lustre-OST0000-osc-ffff88009e7fb7e8@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630979991 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dir_create.sh.0' Lustre: ll_ost03_007: service thread pid 18957 was inactive for 62.113 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 18957, comm: ll_ost03_007 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] tgt_extent_lock+0xea/0x2a0 [ptlrpc] [<0>] ofd_getattr_hdl+0x365/0x720 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe ------------[ cut here ]------------ WARNING: CPU: 1 PID: 16757 at /home/green/git/lustre-release/ldiskfs/ext4_jbd2.c:266 __ldiskfs_handle_dirty_metadata+0x1d2/0x230 [ldiskfs] Modules linked in: loop zfs(PO) zunicode(PO) zzstd(O) zlua(O) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) jbd2 mbcache lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) dm_flakey dm_mod libcfs(OE) crc_t10dif crct10dif_generic sb_edac edac_core iosf_mbi crc32_pclmul ghash_clmulni_intel aesni_intel lrw gf128mul glue_helper ablk_helper cryptd virtio_balloon virtio_console pcspkr i2c_piix4 ip_tables rpcsec_gss_krb5 drm_kms_helper ttm ata_generic pata_acpi drm crct10dif_pclmul crct10dif_common crc32c_intel drm_panel_orientation_quirks ata_piix serio_raw virtio_blk i2c_core libata floppy CPU: 1 PID: 16757 Comm: mdt00_005 Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011 Call Trace: [<ffffffff817d9418>] dump_stack+0x19/0x1b [<ffffffff8108d538>] __warn+0xd8/0x100 [<ffffffff8108d67d>] warn_slowpath_null+0x1d/0x20 [<ffffffffa0c18262>] __ldiskfs_handle_dirty_metadata+0x1d2/0x230 [ldiskfs] [<ffffffffa0bf833d>] ldiskfs_getblk+0x13d/0x210 [ldiskfs] [<ffffffffa0bf8437>] ldiskfs_bread+0x27/0xe0 [ldiskfs] [<ffffffffa0cbce97>] osd_write+0x4e7/0xcb0 [osd_ldiskfs] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffffa04cb6f3>] dt_record_write+0x33/0x120 [obdclass] [<ffffffffa048a6b2>] llog_osd_write_rec+0x1052/0x19a0 [obdclass] [<ffffffffa0477760>] llog_write_rec+0x290/0x590 [obdclass] [<ffffffffa047d4ee>] llog_cat_add_rec+0x23e/0x950 [obdclass] [<ffffffffa04745ff>] llog_add+0x17f/0x1f0 [obdclass] [<ffffffffa04a5a39>] ? lprocfs_counter_add+0xf9/0x160 [obdclass] [<ffffffffa080e60d>] sub_updates_write+0xb6b/0xe32 [ptlrpc] [<ffffffffa07de892>] top_trans_stop+0x7c2/0xf60 [ptlrpc] [<ffffffffa0ef7f3c>] lod_trans_stop+0x25c/0x340 [lod] [<ffffffffa0dcc408>] mdd_trans_stop+0x28/0x16e [mdd] [<ffffffffa0dadb2f>] mdd_migrate_object+0xf1f/0x1ba0 [mdd] [<ffffffffa04ca375>] ? lu_object_find_at+0x285/0xb10 [obdclass] [<ffffffffa0daeb86>] mdd_migrate+0x3d6/0x8e0 [mdd] [<ffffffffa0e800d8>] mdo_migrate+0x4c/0x4e [mdt] [<ffffffffa0e3fe53>] mdt_reint_migrate+0x1023/0x11d0 [mdt] [<ffffffffa0e40087>] mdt_reint_rec+0x87/0x240 [mdt] [<ffffffffa0e1555c>] mdt_reint_internal+0x76c/0xb40 [mdt] [<ffffffffa0e22657>] mdt_reint+0x67/0x150 [mdt] [<ffffffffa07cd26e>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<ffffffffa03b145e>] ? libcfs_nid2str_r+0xfe/0x130 [lnet] [<ffffffffa0778990>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<ffffffffa077a529>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e32be>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffffa0779950>] ? ptlrpc_wait_event+0x620/0x620 [ptlrpc] [<ffffffff810ba0f4>] kthread+0xe4/0xf0 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 ---[ end trace dc8986bfb67aba53 ]--- LDISKFS-fs: ldiskfs_getblk:888: aborting transaction: error 28 in __ldiskfs_handle_dirty_metadata LDISKFS-fs error (device dm-1): ldiskfs_getblk:888: inode #166: block 20753: comm mdt00_005: journal_dirty_metadata failed: handle type 0 started at line 1972, credits 512/0, errcode -28 Aborting journal on device dm-1-8. LDISKFS-fs (dm-1): Remounting filesystem read-only LustreError: 16757:0:(osd_io.c:2103:osd_ldiskfs_write_record()) lustre-MDT0001: error reading offset 36864 (block 9, size 48, offs 36856), credits 0/0: rc = -28 LustreError: 16757:0:(update_trans.c:985:top_trans_stop()) lustre-MDT0001-osd: write updates failed: rc = -28 LustreError: 16757:0:(osd_io.c:2110:osd_ldiskfs_write_record()) journal_get_write_access() returned error -30 LustreError: 16757:0:(tgt_lastrcvd.c:1253:tgt_add_reply_data()) lustre-MDT0001: can't update reply_data file: rc = -30 LustreError: 16757:0:(osd_handler.c:2075:osd_trans_stop()) lustre-MDT0001: failed in transaction hook: rc = -30 LDISKFS-fs error (device dm-1) in osd_trans_stop:2082: error 28 LustreError: 16757:0:(osd_handler.c:2085:osd_trans_stop()) lustre-MDT0001: failed to stop transaction: rc = -28 LustreError: 16757:0:(update_trans.c:1011:top_trans_stop()) lustre-MDT0001-osd: stop trans failed: rc = -30 LustreError: 11881:0:(osd_handler.c:1779:osd_trans_commit_cb()) transaction @0xffff88026fc904a8 commit error: 2 LustreError: 11451:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0001-osp-MDT0000: fail to cancel 1 llog-records: rc = -116 LustreError: 11451:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0001-osp-MDT0000: fail to cancel 1 of 1 llog-records: rc = -116 LustreError: 12314:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0001-osp-MDT0002: fail to cancel 1 llog-records: rc = -116 LustreError: 12314:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) Skipped 1 previous similar message LustreError: 12314:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0001-osp-MDT0002: fail to cancel 1 of 1 llog-records: rc = -116 LustreError: 12314:0:(llog_cat.c:787:llog_cat_cancel_records()) Skipped 1 previous similar message Lustre: mdt07_004: service thread pid 16512 was inactive for 66.169 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 16512, comm: mdt07_004 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] mdt_object_local_lock+0x527/0xb90 [mdt] [<0>] mdt_object_lock_internal+0x70/0x390 [mdt] [<0>] mdt_reint_object_lock+0x2c/0x60 [mdt] [<0>] mdt_reint_striped_lock+0x89/0x5b0 [mdt] [<0>] mdt_migrate_parent_lock+0x3f/0xb0 [mdt] [<0>] mdt_reint_migrate+0x8b7/0x11d0 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x76c/0xb40 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: ll_ost01_005: service thread pid 18363 was inactive for 64.140 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. INFO: task mv:21640 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mv D ffff8802931a73c0 11216 21640 16100 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff81251fe1>] lock_rename+0x31/0xe0 [<ffffffff8125862f>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed432>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817e8a76>] ? trace_do_page_fault+0x56/0x170 [<ffffffff8125975e>] SyS_renameat2+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task mv:21645 blocked for more than 120 seconds. Lustre: ll_ost01_012: service thread pid 20019 was inactive for 64.243 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 14 previous similar messages "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mv D ffff880276068010 11920 21645 17175 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff81251fe1>] lock_rename+0x31/0xe0 [<ffffffff8125862f>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed432>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817e8a76>] ? trace_do_page_fault+0x56/0x170 [<ffffffff8125975e>] SyS_renameat2+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task mv:21812 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mv D ffff880287141900 12656 21812 16664 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff81251fe1>] lock_rename+0x31/0xe0 [<ffffffff8125862f>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed432>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817e8a76>] ? trace_do_page_fault+0x56/0x170 [<ffffffff8125975e>] SyS_renameat2+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task mv:21998 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mv D ffff880274e10608 11344 21998 16458 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff81251fe1>] lock_rename+0x31/0xe0 [<ffffffff8125862f>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed432>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817e8a76>] ? trace_do_page_fault+0x56/0x170 [<ffffffff8125975e>] SyS_renameat2+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task mv:22841 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mv D ffff88028ced8010 11216 22841 16476 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff81251fe1>] lock_rename+0x31/0xe0 [<ffffffff8125862f>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed432>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817e8a76>] ? trace_do_page_fault+0x56/0x170 [<ffffffff8125975e>] SyS_renameat2+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task mv:22867 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mv D ffff88028af93760 11216 22867 16898 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff81251fe1>] lock_rename+0x31/0xe0 [<ffffffff8125862f>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed432>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817e8a76>] ? trace_do_page_fault+0x56/0x170 [<ffffffff8125975e>] SyS_renameat2+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 Lustre: ll_ost01_029: service thread pid 24939 was inactive for 64.210 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 8 previous similar messages Lustre: ll_ost01_050: service thread pid 24995 was inactive for 64.003 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 22 previous similar messages LustreError: 10969:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0000_UUID lock: ffff880276bc3c40/0xb337f8daa2197e6d lrc: 3/0,0 mode: PW/PW res: [0x2c0000401:0x7c1:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000000030020 nid: 0@lo remote: 0xb337f8daa2197e3c expref: 1826 pid: 12694 timeout: 607 lvb_type: 0 LustreError: 17844:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88027883de00 x1710205963533056/t0(0) o105->lustre-OST0000@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' Lustre: ll_ost01_057: service thread pid 25081 was inactive for 64.052 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 8 previous similar messages LustreError: 24169:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -30 Lustre: lustre-OST0000-osc-ffff88029a0e8008: Connection to lustre-OST0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0000-osc-ffff88029a0e8008: This client was evicted by lustre-OST0000; in progress operations using this service will fail. LustreError: 25572:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0000-osc-ffff88029a0e8008: namespace resource [0x2c0000401:0x7cf:0x0].0x0 (ffff880081ef8540) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 25572:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 547 previous similar messages Lustre: 24918:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (68/1s); client may timeout req@ffff880282e29300 x1710205962501312/t0(0) o2->adc3166f-56b4-492d-93bb-9aed3ef6d28b@0@lo:306/0 lens 440/224 e 1 to 0 dl 1630979996 ref 1 fl Complete:/0/0 rc -107/-107 job:'chown.0' LustreError: 13198:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-OST0000-osc-MDT0001: fail to cancel 1 llog-records: rc = -30 LustreError: 13198:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) Skipped 1 previous similar message LustreError: 13198:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-OST0000-osc-MDT0001: fail to cancel 1 of 1 llog-records: rc = -30 LustreError: 13198:0:(llog_cat.c:787:llog_cat_cancel_records()) Skipped 1 previous similar message LustreError: 14050:0:(osp_sync.c:1094:osp_sync_process_committed()) lustre-OST0001-osc-MDT0001: can't cancel 34 records: rc = -30 LustreError: 13198:0:(osp_sync.c:1079:osp_sync_process_committed()) lustre-OST0000-osc-MDT0001: can't cancel record: rc = -30 LustreError: 14054:0:(osp_sync.c:1079:osp_sync_process_committed()) lustre-OST0002-osc-MDT0001: can't cancel record: rc = -30 LustreError: 14054:0:(osp_sync.c:1079:osp_sync_process_committed()) Skipped 3 previous similar messages LustreError: 13198:0:(osp_sync.c:1094:osp_sync_process_committed()) lustre-OST0000-osc-MDT0001: can't cancel 16 records: rc = -30 LustreError: 13198:0:(osp_sync.c:1094:osp_sync_process_committed()) Skipped 6 previous similar messages Lustre: 4119:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff880272ef8040 x1710205963739584/t0(0) o10->lustre-OST0000-osc-ffff88009e7fb7e8@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630980050 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dir_create.sh.0' LustreError: 13198:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-OST0000-osc-MDT0001: fail to cancel 42 llog-records: rc = -30 LustreError: 13198:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) Skipped 37 previous similar messages LustreError: 13198:0:(osp_sync.c:1094:osp_sync_process_committed()) lustre-OST0000-osc-MDT0001: can't cancel 42 records: rc = -30 LustreError: 13198:0:(osp_sync.c:1094:osp_sync_process_committed()) Skipped 8 previous similar messages LustreError: 13198:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-OST0000-osc-MDT0001: fail to cancel 1 of 1 llog-records: rc = -30 LustreError: 13198:0:(llog_cat.c:787:llog_cat_cancel_records()) Skipped 21 previous similar messages LustreError: 13198:0:(osp_sync.c:1079:osp_sync_process_committed()) lustre-OST0000-osc-MDT0001: can't cancel record: rc = -30 LustreError: 13198:0:(osp_sync.c:1079:osp_sync_process_committed()) Skipped 17 previous similar messages LustreError: 14648:0:(osp_sync.c:1094:osp_sync_process_committed()) lustre-OST0003-osc-MDT0001: can't cancel 113 records: rc = -30 LustreError: 14648:0:(osp_sync.c:1094:osp_sync_process_committed()) Skipped 7 previous similar messages LustreError: 14648:0:(osp_sync.c:1079:osp_sync_process_committed()) lustre-OST0003-osc-MDT0001: can't cancel record: rc = -30 LustreError: 14648:0:(osp_sync.c:1079:osp_sync_process_committed()) Skipped 1 previous similar message LustreError: 14050:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-OST0001-osc-MDT0001: fail to cancel 32 llog-records: rc = -30 LustreError: 14050:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) Skipped 12 previous similar messages LustreError: 14050:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-OST0001-osc-MDT0001: fail to cancel 1 of 1 llog-records: rc = -30 LustreError: 14050:0:(llog_cat.c:787:llog_cat_cancel_records()) Skipped 3 previous similar messages LustreError: 25573:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0002-osc-ffff88009e7fb7e8: namespace resource [0x340000401:0x7bf:0x0].0x0 (ffff8802497be840) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 25573:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 23 previous similar messages LustreError: 14648:0:(osp_sync.c:1094:osp_sync_process_committed()) lustre-OST0003-osc-MDT0001: can't cancel 279 records: rc = -30 LustreError: 14648:0:(osp_sync.c:1094:osp_sync_process_committed()) Skipped 15 previous similar messages Lustre: lustre-OST0000-osc-ffff88029a0e8008: Connection restored to 192.168.123.100@tcp (at 0@lo) LustreError: 14050:0:(osp_sync.c:1079:osp_sync_process_committed()) lustre-OST0001-osc-MDT0001: can't cancel record: rc = -30 LustreError: 14050:0:(osp_sync.c:1079:osp_sync_process_committed()) Skipped 5 previous similar messages LustreError: 25548:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -30 Lustre: mdt07_002: service thread pid 11113 was inactive for 66.027 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 6 previous similar messages LustreError: 11-0: lustre-MDT0001-mdc-ffff88009e7fb7e8: operation ldlm_enqueue to node 0@lo failed: rc = -30 LustreError: 25697:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -30 LustreError: 25697:0:(llite_lib.c:1836:ll_md_setattr()) Skipped 1 previous similar message | Externally reported by onyx-68 boilpot email |
racer test 1: racer on clients: centos-20.localnet DURATION=2700 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa17aae52>] ll_migrate+0x9b2/0xec0 [lustre] PGD 800000023e999067 PUD 29f04a067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) dm_flakey dm_mod loop zfs(PO) zunicode(PO) zzstd(O) zlua(O) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) jbd2 mbcache crc_t10dif crct10dif_generic sb_edac edac_core iosf_mbi crc32_pclmul ghash_clmulni_intel aesni_intel lrw gf128mul glue_helper ablk_helper cryptd virtio_balloon pcspkr i2c_piix4 virtio_console ip_tables rpcsec_gss_krb5 ata_generic drm_kms_helper pata_acpi ttm drm crct10dif_pclmul crct10dif_common drm_panel_orientation_quirks ata_piix crc32c_intel serio_raw virtio_blk i2c_core libata floppy [last unloaded: libcfs] CPU: 9 PID: 8531 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011 task: ffff8802a47e1280 ti: ffff8802a0200000 task.ti: ffff8802a0200000 RIP: 0010:[<ffffffffa17aae52>] [<ffffffffa17aae52>] ll_migrate+0x9b2/0xec0 [lustre] RSP: 0018:ffff8802a0203bc8 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff8802f7de1138 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff88031428ac58 RDI: ffff88031428ac38 RBP: ffff8802a0203c48 R08: ffff8802683cf4f8 R09: ffff8802a0203a60 R10: 0000000000000006 R11: ffff8802d1db0d48 R12: 0000000000000000 R13: ffff880083633ef8 R14: ffff880274ec0118 R15: 0000000000000030 FS: 00007fa315368740(0000) GS:ffff880331c40000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 CR2: 0000000000000008 CR3: 0000000245b96000 CR4: 00000000001607e0 Call Trace: [<ffffffff81242213>] ? __check_object_size+0x1c3/0x220 [<ffffffffa1798561>] ll_dir_ioctl+0x5d01/0x6ed0 [lustre] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff810d5248>] ? __enqueue_entity+0x78/0x80 [<ffffffff810dbb8b>] ? put_prev_entity+0x8b/0x400 [<ffffffff810d5979>] ? pick_next_entity+0xa9/0x190 [<ffffffff810d7e1c>] ? set_next_entity+0x3c/0xe0 [<ffffffff810df6c2>] ? pick_next_task_fair+0x702/0x8a0 [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e32be>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffff810c832d>] ? finish_task_switch+0x5d/0x1b0 [<ffffffff817e05ea>] ? __schedule+0x32a/0x7d0 [<ffffffff8125b3fd>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff81406d25>] ? lockref_put_or_lock+0x25/0x30 [<ffffffff8126a634>] ? mntput+0x24/0x40 [<ffffffff8125b751>] SyS_ioctl+0xa1/0xc0 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 | Lustre: 16480:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 512 < left 929, rollback = 7 Lustre: 16480:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 16480:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 1/89/0 Lustre: 16480:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 920/929/0, punch: 0/0/0, quota 3/3/0 Lustre: 16480:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 16480:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 16480:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 511 < left 928, rollback = 7 Lustre: 16480:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 271 previous similar messages Lustre: 16480:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 16480:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 271 previous similar messages Lustre: 16480:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 1/89/0 Lustre: 16480:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 271 previous similar messages Lustre: 16480:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 920/929/1, punch: 0/0/0, quota 3/3/0 Lustre: 16480:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 271 previous similar messages Lustre: 16480:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 16480:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 271 previous similar messages Lustre: 16480:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 16480:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 271 previous similar messages Lustre: 16480:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 511 < left 928, rollback = 7 Lustre: 16480:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 402 previous similar messages Lustre: 16480:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 16480:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 402 previous similar messages Lustre: 16480:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 1/89/0 Lustre: 16480:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 402 previous similar messages Lustre: 16480:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 920/929/1, punch: 0/0/0, quota 3/3/0 Lustre: 16480:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 530 previous similar messages Lustre: 16480:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 16480:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 530 previous similar messages Lustre: 16480:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 16480:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 530 previous similar messages Lustre: 14983:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 512 < left 728, rollback = 7 Lustre: 14983:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 244 previous similar messages Lustre: 14983:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 14983:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 244 previous similar messages Lustre: 14983:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 1/89/0 Lustre: 14983:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 244 previous similar messages Lustre: 14983:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 719/728/0, punch: 0/0/0, quota 3/3/0 Lustre: 14983:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 116 previous similar messages Lustre: 14983:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 14983:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 116 previous similar messages Lustre: 14983:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 14983:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 116 previous similar messages Lustre: 18803:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 510 < left 760, rollback = 2 Lustre: 18803:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 3/12/0, destroy: 0/0/0 Lustre: 18803:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 718 previous similar messages Lustre: 18803:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 8/760/0 Lustre: 18803:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 718 previous similar messages Lustre: 18803:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 5/136/0, punch: 0/0/0, quota 8/56/0 Lustre: 18803:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 718 previous similar messages Lustre: 18803:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 10/203/2, delete: 0/0/0 Lustre: 18803:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 718 previous similar messages Lustre: 18803:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 4/4/0, ref_del: 0/0/0 Lustre: 18803:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 718 previous similar messages Lustre: 18803:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 509 < left 760, rollback = 2 Lustre: 19946:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 512 < left 520, rollback = 7 Lustre: 19946:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 718 previous similar messages Lustre: 19946:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 19946:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 9 previous similar messages Lustre: 19946:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 1/89/0 Lustre: 19946:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 9 previous similar messages Lustre: 19946:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 511/520/0, punch: 0/0/0, quota 3/3/0 Lustre: 19946:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 9 previous similar messages Lustre: 19946:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 19946:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 9 previous similar messages Lustre: 19946:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 19946:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 9 previous similar messages Lustre: 17449:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 504 < left 788, rollback = 2 Lustre: 17449:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 8 previous similar messages Lustre: 19982:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 8: before 509 < left 3246, rollback = 9 Lustre: 19982:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 9 previous similar messages Lustre: 17539:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 512 < left 707, rollback = 7 Lustre: 17539:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 512 previous similar messages Lustre: 17471:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 503 < left 671, rollback = 2 Lustre: 17471:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 1 previous similar message Lustre: 14974:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 920/3680/0, destroy: 1/4/0 Lustre: 14974:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 3596 previous similar messages Lustre: 14974:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 3/3/1, xattr_set: 924/148/0 Lustre: 14974:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 3596 previous similar messages Lustre: 14974:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 4601/39570/0, punch: 0/0/0, quota 6/6/0 Lustre: 14974:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 3596 previous similar messages Lustre: 14974:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 921/15656/0, delete: 2/5/0 Lustre: 14974:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 3596 previous similar messages Lustre: 14974:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 1/1/0, ref_del: 2/2/0 Lustre: 14974:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 3596 previous similar messages Lustre: 19972:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 8: before 509 < left 8703, rollback = 9 Lustre: 19972:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 8 previous similar messages Lustre: 20902:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 507 < left 26369, rollback = 7 Lustre: 20902:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 3066 previous similar messages ------------[ cut here ]------------ WARNING: CPU: 14 PID: 18541 at /home/green/git/lustre-release/ldiskfs/namei.c:3331 ldiskfs_orphan_add+0x11e/0x2a0 [ldiskfs] Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) dm_flakey dm_mod loop zfs(PO) zunicode(PO) zzstd(O) zlua(O) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) jbd2 mbcache crc_t10dif crct10dif_generic sb_edac edac_core iosf_mbi crc32_pclmul ghash_clmulni_intel aesni_intel lrw gf128mul glue_helper ablk_helper cryptd virtio_balloon pcspkr i2c_piix4 virtio_console ip_tables rpcsec_gss_krb5 ata_generic drm_kms_helper pata_acpi ttm drm crct10dif_pclmul crct10dif_common drm_panel_orientation_quirks ata_piix crc32c_intel serio_raw virtio_blk i2c_core libata floppy [last unloaded: libcfs] CPU: 14 PID: 18541 Comm: mdt07_006 Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011 Call Trace: [<ffffffff817d9418>] dump_stack+0x19/0x1b [<ffffffff8108d538>] __warn+0xd8/0x100 [<ffffffff8108d67d>] warn_slowpath_null+0x1d/0x20 [<ffffffffa0d764ae>] ldiskfs_orphan_add+0x11e/0x2a0 [ldiskfs] [<ffffffffa0d679eb>] ldiskfs_xattr_inode_orphan_add.constprop.17+0xbb/0x110 [ldiskfs] [<ffffffffa0d6a417>] ldiskfs_xattr_delete_inode+0x77/0x3c0 [ldiskfs] [<ffffffffa0d8aaa0>] ? ldiskfs_evict_inode+0x190/0x670 [ldiskfs] [<ffffffffa0d8aad9>] ldiskfs_evict_inode+0x1c9/0x670 [ldiskfs] [<ffffffff812649af>] evict+0xaf/0x180 [<ffffffff81264de5>] iput+0xf5/0x180 [<ffffffffa0ca6d08>] osd_object_delete+0x1f8/0x370 [osd_ldiskfs] [<ffffffffa04d759d>] lu_object_free.isra.26+0xcd/0x1d0 [obdclass] [<ffffffffa04dbcf1>] lu_object_put+0xa1/0x480 [obdclass] [<ffffffffa0ed8e09>] mdt_getattr_name_lock+0x859/0x2590 [mdt] [<ffffffffa0778097>] ? lustre_msg_buf+0x17/0x60 [ptlrpc] [<ffffffffa07a2bbf>] ? __req_capsule_get+0x41f/0x6a0 [ptlrpc] [<ffffffffa07784cc>] ? lustre_msg_get_flags+0x2c/0xa0 [ptlrpc] [<ffffffffa0ee1785>] mdt_intent_getattr+0x2d5/0x4b0 [mdt] [<ffffffffa0ed528d>] mdt_intent_opc+0x1dd/0xc10 [mdt] [<ffffffffa0ee14b0>] ? mdt_intent_layout+0xe90/0xe90 [mdt] [<ffffffffa0ede2a1>] mdt_intent_policy+0x1a1/0x360 [mdt] [<ffffffffa0729962>] ldlm_lock_enqueue+0x3c2/0xb40 [ptlrpc] [<ffffffff817e340e>] ? _raw_read_unlock+0xe/0x20 [<ffffffffa0300d85>] ? cfs_hash_rw_unlock+0x15/0x20 [libcfs] [<ffffffffa0303fb6>] ? cfs_hash_add+0xa6/0x180 [libcfs] [<ffffffffa07537a6>] ldlm_handle_enqueue0+0xaa6/0x16d0 [ptlrpc] [<ffffffffa0778000>] ? lustre_msg_buf_v2+0x160/0x1e0 [ptlrpc] [<ffffffffa07d5494>] tgt_enqueue+0x64/0x240 [ptlrpc] [<ffffffffa07df26e>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<ffffffffa034745e>] ? libcfs_nid2str_r+0xfe/0x130 [lnet] [<ffffffffa078a990>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<ffffffffa078c529>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffffa078b950>] ? ptlrpc_wait_event+0x620/0x620 [ptlrpc] [<ffffffff810ba0f4>] kthread+0xe4/0xf0 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 ---[ end trace 5b2c3ed042866441 ]--- Lustre: 20912:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 505 < left 805, rollback = 2 Lustre: 20912:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 19 previous similar messages Lustre: 11778:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff880091163240 x1710150313961984/t0(0) o10->lustre-OST0002-osc-ffff880298c7e678@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630926661 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'truncate.0' Lustre: 18696:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 190/760/0, destroy: 1/4/1 Lustre: 18696:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 2491 previous similar messages Lustre: 18696:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 190/0/0 Lustre: 18696:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 2491 previous similar messages Lustre: 18696:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 951/8180/0, punch: 0/0/0, quota 6/6/0 Lustre: 18696:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 2491 previous similar messages Lustre: 18696:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 190/3230/0, delete: 2/5/1 Lustre: 18696:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 2491 previous similar messages Lustre: 18696:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 1/1/0 Lustre: 18696:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 2491 previous similar messages LustreError: 14966:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0000_UUID lock: ffff880091fbbc40/0xd30819a9e29976ff lrc: 3/0,0 mode: PW/PW res: [0x2:0x0:0x0].0x0 rrc: 4 type: EXT [0->18446744073709551615] (req 0->786431) gid 0 flags: 0x60000400030020 nid: 0@lo remote: 0xd30819a9e29976e3 expref: 1711 pid: 15683 timeout: 19343 lvb_type: 0 LustreError: 11-0: lustre-OST0000-osc-ffff880298c7e678: operation ost_write to node 0@lo failed: rc = -107 Lustre: lustre-OST0000-osc-ffff880298c7e678: Connection to lustre-OST0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0000-osc-ffff880298c7e678: This client was evicted by lustre-OST0000; in progress operations using this service will fail. Lustre: 11777:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.120@tcp:/lustre/fid: [0x280000403:0x35:0x0]// may get corrupted (rc -108) Lustre: 11776:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.120@tcp:/lustre/fid: [0x200000404:0xdb:0x0]// may get corrupted (rc -108) Lustre: 11776:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.120@tcp:/lustre/fid: [0x280000404:0xdc:0x0]/ may get corrupted (rc -108) Lustre: 11776:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.120@tcp:/lustre/fid: [0x280000404:0xbf:0x0]/ may get corrupted (rc -108) Lustre: 11776:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.120@tcp:/lustre/fid: [0x280000404:0xae:0x0]// may get corrupted (rc -108) LustreError: 24001:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0000-osc-ffff880298c7e678: namespace resource [0x10:0x0:0x0].0x0 (ffff880065f6de40) refcount nonzero (1) after lock cleanup; forcing cleanup. Lustre: ll_ost01_005: service thread pid 19832 was inactive for 62.115 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 19832, comm: ll_ost01_005 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] tgt_extent_lock+0xea/0x2a0 [ptlrpc] [<0>] ofd_getattr_hdl+0x365/0x720 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 20012, comm: ll_ost01_008 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] tgt_extent_lock+0xea/0x2a0 [ptlrpc] [<0>] ofd_getattr_hdl+0x365/0x720 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 19987, comm: ll_ost01_006 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] tgt_extent_lock+0xea/0x2a0 [ptlrpc] [<0>] ofd_getattr_hdl+0x365/0x720 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: ll_ost01_003: service thread pid 16371 was inactive for 62.258 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: ll_ost01_012: service thread pid 20864 was inactive for 62.040 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 5 previous similar messages Lustre: ll_ost01_024: service thread pid 20935 was inactive for 62.009 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 10 previous similar messages Lustre: ll_ost01_039: service thread pid 21091 was inactive for 62.239 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 16 previous similar messages LustreError: 14966:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0000_UUID lock: ffff88024fba5e00/0xd30819a9e29a5998 lrc: 3/0,0 mode: PW/PW res: [0x2c0000400:0x3:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->3145727) gid 0 flags: 0x60000400020020 nid: 0@lo remote: 0xd30819a9e29a598a expref: 2139 pid: 19832 timeout: 19355 lvb_type: 0 Lustre: ll_ost01_055: service thread pid 21396 was inactive for 62.199 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 13 previous similar messages LustreError: 14966:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0002_UUID lock: ffff88026d2b1a80/0xd30819a9e29c953e lrc: 3/0,0 mode: PW/PW res: [0x340000400:0x6:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 1048576->18446744073709551615) gid 0 flags: 0x60000000020020 nid: 0@lo remote: 0xd30819a9e29c9530 expref: 2429 pid: 19987 timeout: 19364 lvb_type: 0 Lustre: lustre-OST0000-osc-ffff8802cfdd12a8: Connection to lustre-OST0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0002-osc-ffff8802cfdd12a8: This client was evicted by lustre-OST0002; in progress operations using this service will fail. LustreError: 24028:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0000-osc-ffff8802cfdd12a8: namespace resource [0x2c0000400:0x130:0x0].0x0 (ffff88021ee516c0) refcount nonzero (1) after lock cleanup; forcing cleanup. Lustre: lustre-OST0002-osc-ffff8802cfdd12a8: Connection restored to 192.168.123.120@tcp (at 0@lo) LustreError: 24028:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 142 previous similar messages Lustre: 11778:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff880091163240 x1710150314540224/t0(0) o10->lustre-OST0002-osc-ffff880298c7e678@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630926755 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'truncate.0' Lustre: mdt02_001: service thread pid 14980 was inactive for 62.134 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 9 previous similar messages LustreError: 14966:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0001_UUID lock: ffff8802ea317840/0xd30819a9e29d19ea lrc: 3/0,0 mode: PW/PW res: [0x300000400:0xc7:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000400000020 nid: 0@lo remote: 0xd30819a9e29d19dc expref: 1798 pid: 15669 timeout: 19386 lvb_type: 0 LustreError: 19784:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88027d399300 x1710150314547776/t0(0) o104->lustre-OST0003@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 21428:0:(service.c:2290:ptlrpc_server_handle_request()) @@@ Dropping timed-out request from 12345-0@lo: deadline 93/6s ago req@ffff8802a37f9f80 x1710150312596992/t0(0) o1->717fdd9d-1a9e-495a-9c80-e8efeeb57655@0@lo:616/0 lens 440/0 e 2 to 0 dl 1630926701 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'lfs.0' Lustre: 21428:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (93/6s); client may timeout req@ffff8802a37f9f80 x1710150312596992/t0(0) o1->717fdd9d-1a9e-495a-9c80-e8efeeb57655@0@lo:616/0 lens 440/0 e 2 to 0 dl 1630926701 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'lfs.0' LustreError: 19784:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 2 previous similar messages LustreError: 11-0: lustre-OST0001-osc-ffff880298c7e678: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-OST0001-osc-ffff880298c7e678: Connection to lustre-OST0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 1 previous similar message LustreError: 167-0: lustre-OST0001-osc-ffff880298c7e678: This client was evicted by lustre-OST0001; in progress operations using this service will fail. LustreError: Skipped 1 previous similar message LustreError: 20935:0:(service.c:2290:ptlrpc_server_handle_request()) @@@ Dropping timed-out request from 12345-0@lo: deadline 94/5s ago req@ffff8800667fd7c0 x1710150312616448/t0(0) o1->717fdd9d-1a9e-495a-9c80-e8efeeb57655@0@lo:618/0 lens 440/0 e 2 to 0 dl 1630926703 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'lfs.0' Lustre: 19997:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (94/5s); client may timeout req@ffff8802aecc70c0 x1710150312616576/t0(0) o1->717fdd9d-1a9e-495a-9c80-e8efeeb57655@0@lo:618/0 lens 440/0 e 2 to 0 dl 1630926703 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'lfs.0' Lustre: 19997:0:(service.c:2328:ptlrpc_server_handle_request()) Skipped 98 previous similar messages LustreError: 20935:0:(service.c:2290:ptlrpc_server_handle_request()) Skipped 92 previous similar messages LustreError: 24028:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0000-osc-ffff8802cfdd12a8: namespace resource [0x2c0000400:0x130:0x0].0x0 (ffff88021ee516c0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 24028:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 12 previous similar messages Lustre: lustre-OST0000-osc-ffff8802cfdd12a8: Connection restored to 192.168.123.120@tcp (at 0@lo) LustreError: 20906:0:(service.c:2290:ptlrpc_server_handle_request()) @@@ Dropping timed-out request from 12345-0@lo: deadline 70/23s ago req@ffff8802ebec5e00 x1710150312989184/t0(0) o101->717fdd9d-1a9e-495a-9c80-e8efeeb57655@0@lo:601/0 lens 328/0 e 1 to 0 dl 1630926686 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'ln.0' LustreError: 20906:0:(service.c:2290:ptlrpc_server_handle_request()) Skipped 386 previous similar messages Lustre: 20906:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (70/23s); client may timeout req@ffff8802ebec5e00 x1710150312989184/t0(0) o101->717fdd9d-1a9e-495a-9c80-e8efeeb57655@0@lo:601/0 lens 328/0 e 1 to 0 dl 1630926686 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'ln.0' Lustre: 20906:0:(service.c:2328:ptlrpc_server_handle_request()) Skipped 403 previous similar messages LustreError: 11-0: lustre-OST0003-osc-ffff880298c7e678: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: 14954:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630926709 with bad export cookie 15206432359258286029 Lustre: lustre-OST0002-osc-ffff880298c7e678: Connection to lustre-OST0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 1 previous similar message LustreError: 14954:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) Skipped 6 previous similar messages LustreError: 167-0: lustre-OST0002-osc-ffff880298c7e678: This client was evicted by lustre-OST0002; in progress operations using this service will fail. LustreError: Skipped 1 previous similar message Lustre: lustre-OST0000-osc-ffff880298c7e678: Connection restored to 192.168.123.120@tcp (at 0@lo) Lustre: 20732:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 512 < left 658, rollback = 7 Lustre: 20732:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 3852 previous similar messages Lustre: 20732:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 20732:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 1404 previous similar messages Lustre: 20732:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 1/89/0 Lustre: 20732:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 1404 previous similar messages Lustre: 20732:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 649/658/0, punch: 0/0/0, quota 3/3/0 Lustre: 20732:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 1404 previous similar messages Lustre: 20732:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 20732:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 1404 previous similar messages Lustre: 20732:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 20732:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 1404 previous similar messages Lustre: lustre-OST0001-osc-ffff880298c7e678: Connection restored to 192.168.123.120@tcp (at 0@lo) Lustre: 14987:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 505 < left 688, rollback = 2 Lustre: 14987:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 19 previous similar messages Lustre: dir [0x280000404:0x39:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: dir [0x280000404:0xaa:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 4 previous similar messages Lustre: dir [0x200000404:0xc4:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 3 previous similar messages LustreError: 20004:0:(mdt_xattr.c:429:mdt_dir_layout_update()) lustre-MDT0002: [0x280000403:0xb9:0x0] migrate mdt index mismatch 2 != 0 LustreError: 22084:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0002: failed to get lu_attr of [0x280000403:0x39:0x0]: rc = -2 Lustre: 22084:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 511 < left 27917, rollback = 7 Lustre: 22084:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 21454 previous similar messages LustreError: 14999:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0002: failed to get lu_attr of [0x280000403:0x39:0x0]: rc = -2 Lustre: 11788:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff8802f372c500 x1710150318122368/t0(0) o10->lustre-OST0000-osc-ffff880298c7e678@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630926853 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'truncate.0' Lustre: 17906:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 8: before 509 < left 3909, rollback = 9 Lustre: 17906:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 83 previous similar messages Lustre: 17906:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 229/916/0, destroy: 1/4/0 Lustre: 17906:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 35233 previous similar messages Lustre: 17906:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 3/3/1, xattr_set: 233/148/0 Lustre: 17906:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 35235 previous similar messages Lustre: 17906:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1146/9857/0, punch: 0/0/0, quota 10/74/0 Lustre: 17906:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 35237 previous similar messages Lustre: 17906:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 230/3909/0, delete: 2/5/1 Lustre: 17906:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 35237 previous similar messages Lustre: 17906:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 1/1/0, ref_del: 2/2/1 Lustre: 17906:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 35237 previous similar messages LustreError: 14966:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0003_UUID lock: ffff8802d9ae6580/0xd30819a9e2a42248 lrc: 3/0,0 mode: PW/PW res: [0x1ee:0x0:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000000020020 nid: 0@lo remote: 0xd30819a9e2a42202 expref: 876 pid: 24507 timeout: 19530 lvb_type: 0 LustreError: 14966:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 2 previous similar messages LustreError: 11-0: lustre-OST0003-osc-ffff8802cfdd12a8: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 5 previous similar messages Lustre: lustre-OST0003-osc-ffff8802cfdd12a8: Connection to lustre-OST0003 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 19296:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff880265194500 x1710150320969280/t0(0) o105->lustre-OST0003@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 19296:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 31 previous similar messages LustreError: 167-0: lustre-OST0003-osc-ffff8802cfdd12a8: This client was evicted by lustre-OST0003; in progress operations using this service will fail. Lustre: lustre-OST0003-osc-ffff8802cfdd12a8: Connection restored to 192.168.123.120@tcp (at 0@lo) Lustre: Skipped 2 previous similar messages Lustre: 11779:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff8802a467b240 x1710150321808064/t0(0) o10->lustre-OST0001-osc-ffff880298c7e678@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630926934 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dir_create.sh.0' LustreError: 14994:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0000: '5' migration was interrupted, run 'lfs migrate -m 1 -c 1 -H crush 5' to finish migration. Lustre: 17675:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 512 < left 1439, rollback = 7 Lustre: 17675:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 26957 previous similar messages Lustre: 14933:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 7s req@ffff8802f73f7018 x1710150322771456/t0(0) o104->717fdd9d-1a9e-495a-9c80-e8efeeb57655@0@lo:0/0 lens 328/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'' LustreError: 20732:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0000: '11' migration was interrupted, run 'lfs migrate -m 2 -c 1 -H crush 11' to finish migration. LustreError: 14966:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0001_UUID lock: ffff880086d04b40/0xd30819a9e2acb1f7 lrc: 3/0,0 mode: PW/PW res: [0x1c0:0x0:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000400020020 nid: 0@lo remote: 0xd30819a9e2acb1cd expref: 3243 pid: 25291 timeout: 19640 lvb_type: 0 LustreError: 19296:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802c2630040 x1710150325686720/t0(0) o105->lustre-OST0001@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 11-0: lustre-OST0001-osc-ffff880298c7e678: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 4 previous similar messages Lustre: lustre-OST0001-osc-ffff880298c7e678: Connection to lustre-OST0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0001-osc-ffff880298c7e678: This client was evicted by lustre-OST0001; in progress operations using this service will fail. LustreError: 18794:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802e8b5e440 x1710150325803712/t0(0) o105->lustre-OST0002@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 24478:0:(ldlm_lockd.c:1425:ldlm_handle_enqueue0()) ### lock on destroyed export ffff880284500008 ns: filter-lustre-OST0002_UUID lock: ffff880272e7f480/0xd30819a9e2b4ac53 lrc: 5/0,0 mode: PW/PW res: [0x340000400:0x5b7:0x0].0x0 rrc: 4 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x50000000020000 nid: 0@lo remote: 0xd30819a9e2b4ac45 expref: 4196 pid: 24478 timeout: 0 lvb_type: 0 LustreError: 18794:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 1 previous similar message Lustre: 11785:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.120@tcp:/lustre/fid: [0x240000404:0x16a:0x0]// may get corrupted (rc -108) LustreError: 26154:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802f6f32c00 x1710150325960064/t0(0) o104->lustre-OST0002@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 26154:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 77 previous similar messages LustreError: 30141:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0001-osc-ffff880298c7e678: namespace resource [0x300000400:0x3f0:0x0].0x0 (ffff8802e30b25c0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 30141:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 89 previous similar messages Lustre: lustre-OST0001-osc-ffff880298c7e678: Connection restored to 192.168.123.120@tcp (at 0@lo) Lustre: 15011:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 6s req@ffff880243655180 x1710150325992960/t0(0) o35->4daad2b5-7c77-41c5-b493-e6876f5214e1@0@lo:0/0 lens 392/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'chown.0' LustreError: 30267:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0002-osc-ffff880298c7e678: namespace resource [0x340000400:0x58b:0x0].0x0 (ffff88024cf84f40) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 30267:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 311 previous similar messages Lustre: 18016:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 496 < left 849, rollback = 2 Lustre: 18016:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 38 previous similar messages Lustre: dir [0x240000403:0x1bd:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 2 previous similar messages LustreError: 14966:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0002_UUID lock: ffff880246c09300/0xd30819a9e2b46191 lrc: 3/0,0 mode: PW/PW res: [0x340000400:0x2b8:0x0].0x0 rrc: 4 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000000020020 nid: 0@lo remote: 0xd30819a9e2b46175 expref: 1765 pid: 16483 timeout: 19746 lvb_type: 0 LustreError: 14966:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 1 previous similar message LustreError: 21015:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802409b3240 x1710150330318848/t0(0) o104->lustre-OST0002@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 21015:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 4 previous similar messages LustreError: 11-0: lustre-OST0002-osc-ffff8802cfdd12a8: operation ost_write to node 0@lo failed: rc = -107 LustreError: Skipped 1 previous similar message Lustre: lustre-OST0002-osc-ffff8802cfdd12a8: Connection to lustre-OST0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 1 previous similar message LustreError: 167-0: lustre-OST0002-osc-ffff8802cfdd12a8: This client was evicted by lustre-OST0002; in progress operations using this service will fail. LustreError: Skipped 1 previous similar message Lustre: 11782:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.120@tcp:/lustre/fid: [0x240000403:0x204:0x0]// may get corrupted (rc -5) LustreError: 17677:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630927069 with bad export cookie 15206432359259590857 Lustre: 11781:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.120@tcp:/lustre/fid: [0x280000403:0x252:0x0]// may get corrupted (rc -108) Lustre: 11780:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.120@tcp:/lustre/fid: [0x240000404:0x149:0x0]// may get corrupted (rc -108) Lustre: 11783:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.120@tcp:/lustre/fid: [0x240000403:0x261:0x0]/ may get corrupted (rc -5) LustreError: 587:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0003-osc-ffff8802cfdd12a8: namespace resource [0x160:0x0:0x0].0x0 (ffff8802adf62840) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 587:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 69 previous similar messages Lustre: 11782:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.120@tcp:/lustre/fid: [0x200000403:0x194:0x0]/ may get corrupted (rc -108) Lustre: lustre-OST0003-osc-ffff8802cfdd12a8: Connection restored to 192.168.123.120@tcp (at 0@lo) Lustre: Skipped 1 previous similar message 14[572]: segfault at 8 ip 00007f64ea2007e8 sp 00007fff16aeb7f0 error 4 in ld-2.17.so[7f64ea1f5000+22000] Lustre: 22117:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 372/1488/0, destroy: 1/4/1 Lustre: 22117:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 110424 previous similar messages Lustre: 22117:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 3/3/0, xattr_set: 376/148/0 Lustre: 22117:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 110427 previous similar messages Lustre: 22117:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1861/16006/0, punch: 0/0/0, quota 8/56/0 Lustre: 22117:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 110427 previous similar messages Lustre: 22117:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 373/6340/0, delete: 2/5/0 Lustre: 22117:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 110423 previous similar messages Lustre: 22117:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 1/1/0, ref_del: 2/2/1 Lustre: 22117:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 110424 previous similar messages LustreError: 14966:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0003_UUID lock: ffff88026a9f0400/0xd30819a9e2bce71b lrc: 3/0,0 mode: PW/PW res: [0x380000400:0x3e7:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 786432->18446744073709551615) gid 0 flags: 0x60000000020020 nid: 0@lo remote: 0xd30819a9e2bce6e3 expref: 1914 pid: 20064 timeout: 19849 lvb_type: 0 LustreError: 14966:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 1 previous similar message LustreError: 16386:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630927170 with bad export cookie 15206432359258849249 Lustre: lustre-OST0003-osc-ffff880298c7e678: Connection to lustre-OST0003 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 1 previous similar message LustreError: 15675:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802219ebec0 x1710150334656960/t0(0) o104->lustre-OST0003@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 15675:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 47 previous similar messages LustreError: 167-0: lustre-OST0003-osc-ffff880298c7e678: This client was evicted by lustre-OST0003; in progress operations using this service will fail. LustreError: Skipped 1 previous similar message LustreError: 16386:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) Skipped 3 previous similar messages LustreError: 22314:0:(ofd_obd.c:501:ofd_destroy_export()) lustre-OST0003: cli 4daad2b5-7c77-41c5-b493-e6876f5214e1/ffff8802ea355d28 has 122880 pending on destroyed export LustreError: 22314:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_destroy_export: tot_granted 40779328 != fo_tot_granted 40902208 LustreError: 22314:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_destroy_export: tot_pending 0 != fo_tot_pending 122880 LustreError: 2844:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0003-osc-ffff880298c7e678: namespace resource [0x160:0x0:0x0].0x0 (ffff88025f232840) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 2844:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 97 previous similar messages Lustre: lustre-OST0003-osc-ffff880298c7e678: Connection restored to 192.168.123.120@tcp (at 0@lo) Lustre: Skipped 1 previous similar message Lustre: 22084:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 509 < left 36904, rollback = 7 Lustre: 22084:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 109697 previous similar messages LustreError: 11-0: lustre-MDT0002-mdc-ffff8802cfdd12a8: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 1 previous similar message LustreError: 31347:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff8802cfdd12a8: inode [0x280000403:0x240:0x0] mdc close failed: rc = -108 LustreError: 31885:0:(llite_lib.c:2983:ll_prep_inode()) new_inode -fatal: rc -5 LustreError: 3140:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff8802cfdd12a8: inode [0x280000404:0x1f7:0x0] mdc close failed: rc = -108 LustreError: 3140:0:(file.c:234:ll_close_inode_openhandle()) Skipped 20 previous similar messages LustreError: 15699:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 80710336 != fo_tot_granted 80833216 LustreError: 15699:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 90112 != fo_tot_pending 212992 LustreError: 15700:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 110729920 != fo_tot_granted 110852800 LustreError: 15700:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 122880 LustreError: 15007:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000404:0x1b0:0x0]: rc = -2 LustreError: 15699:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 110729920 != fo_tot_granted 110852800 LustreError: 15699:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 122880 LustreError: 15688:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 110731072 != fo_tot_granted 110853952 LustreError: 15688:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 1 previous similar message LustreError: 15688:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 122880 LustreError: 15688:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 1 previous similar message 6[3064]: segfault at 8 ip 00007f385bac97e8 sp 00007ffde0c3b320 error 4 in ld-2.17.so[7f385babe000+22000] LustreError: 561:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 119074240 != fo_tot_granted 119197120 LustreError: 561:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 3 previous similar messages LustreError: 561:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 122880 LustreError: 561:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 3 previous similar messages LustreError: 14987:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0000: '6' migration was interrupted, run 'lfs migrate -m 1 -c 1 -H crush 6' to finish migration. LustreError: 18761:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000403:0xf0:0x0]: rc = -2 LustreError: 15702:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 143175616 != fo_tot_granted 143298496 LustreError: 15702:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 5 previous similar messages LustreError: 15702:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 122880 LustreError: 15702:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 5 previous similar messages Lustre: 26590:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 8: before 509 < left 11202, rollback = 9 Lustre: 26590:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 248 previous similar messages Lustre: 11775:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff88026c114500 x1710150338737856/t0(0) o10->lustre-OST0000-osc-ffff880298c7e678@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630927304 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'cp.0' Lustre: 5431:0:(service.c:1378:ptlrpc_at_send_early_reply()) @@@ Already past deadline (-1s), not sending early reply. Consider increasing at_early_margin (5)? req@ffff88024d23f0c0 x1710150338310464/t0(0) o2->717fdd9d-1a9e-495a-9c80-e8efeeb57655@0@lo:434/0 lens 440/0 e 0 to 0 dl 1630927274 ref 2 fl New:/0/ffffffff rc 0/-1 job:'touch.0' LustreError: 15693:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 243346880 != fo_tot_granted 243469760 LustreError: 15693:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 12 previous similar messages LustreError: 15693:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 4277120 != fo_tot_pending 4400000 LustreError: 15693:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 12 previous similar messages Lustre: 5197:0:(service.c:1378:ptlrpc_at_send_early_reply()) @@@ Already past deadline (-1s), not sending early reply. Consider increasing at_early_margin (5)? req@ffff88008c159940 x1710150338502912/t0(0) o101->4daad2b5-7c77-41c5-b493-e6876f5214e1@0@lo:436/0 lens 328/0 e 0 to 0 dl 1630927276 ref 2 fl New:/0/ffffffff rc 0/-1 job:'ln.0' Lustre: 5197:0:(service.c:1378:ptlrpc_at_send_early_reply()) Skipped 61 previous similar messages Lustre: 11784:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1630927260/real 1630927260] req@ffff8802651b1300 x1710150338502912/t0(0) o101->lustre-OST0003-osc-ffff880298c7e678@0@lo:28/4 lens 328/400 e 0 to 1 dl 1630927284 ref 1 fl Rpc:XQr/0/ffffffff rc 0/-1 job:'ln.0' Lustre: lustre-OST0003-osc-ffff880298c7e678: Connection to lustre-OST0003 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 1 previous similar message Lustre: lustre-OST0003: Client 4daad2b5-7c77-41c5-b493-e6876f5214e1 (at 0@lo) reconnecting Lustre: lustre-OST0003-osc-ffff880298c7e678: Connection restored to 192.168.123.120@tcp (at 0@lo) Lustre: Skipped 1 previous similar message Lustre: dir [0x240000403:0x34d:0x0] stripe 1 readdir failed: -2, directory is partially accessed! LustreError: 14966:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0003_UUID lock: ffff88006b1a1a80/0xd30819a9e2c4e870 lrc: 3/0,0 mode: PW/PW res: [0x380000400:0x53d:0x0].0x0 rrc: 5 type: EXT [0->18446744073709551615] (req 0->786431) gid 0 flags: 0x60000400020020 nid: 0@lo remote: 0xd30819a9e2c4e5a6 expref: 549 pid: 21022 timeout: 19974 lvb_type: 0 LustreError: 14966:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 1 previous similar message LustreError: 11-0: lustre-OST0003-osc-ffff880298c7e678: operation ost_setattr to node 0@lo failed: rc = -107 LustreError: 167-0: lustre-OST0003-osc-ffff880298c7e678: This client was evicted by lustre-OST0003; in progress operations using this service will fail. LustreError: Skipped 1 previous similar message Lustre: 11785:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1630927257/real 1630927257] req@ffff880091928680 x1710150338423744/t0(0) o2->lustre-OST0003-osc-ffff8802cfdd12a8@0@lo:28/4 lens 440/432 e 0 to 1 dl 1630927294 ref 1 fl Rpc:XQr/0/ffffffff rc 0/-1 job:'touch.0' Lustre: lustre-OST0001: Client 717fdd9d-1a9e-495a-9c80-e8efeeb57655 (at 0@lo) reconnecting Lustre: 11786:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.120@tcp:/lustre/fid: [0x240000404:0x1f2:0x0]// may get corrupted (rc -108) Lustre: 11781:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.120@tcp:/lustre/fid: [0x280000404:0x304:0x0]/ may get corrupted (rc -108) Lustre: 11781:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.120@tcp:/lustre/fid: [0x280000404:0x316:0x0]// may get corrupted (rc -108) Lustre: 11780:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.120@tcp:/lustre/fid: [0x280000404:0x24c:0x0]// may get corrupted (rc -108) Lustre: 5197:0:(service.c:1436:ptlrpc_at_send_early_reply()) @@@ Could not add any time (44/35), not sending early reply req@ffff8802aee4de00 x1710150338422336/t0(0) o2->717fdd9d-1a9e-495a-9c80-e8efeeb57655@0@lo:462/0 lens 440/0 e 1 to 0 dl 1630927344 ref 2 fl New:/0/ffffffff rc 0/-1 job:'touch.0' LustreError: 6685:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0003-osc-ffff880298c7e678: namespace resource [0x380000400:0x540:0x0].0x0 (ffff8802d5f33c40) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 6685:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 13 previous similar messages Lustre: 11785:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1630927258/real 1630927258] req@ffff880265128cc0 x1710150338429760/t0(0) o2->lustre-OST0000-osc-ffff8802cfdd12a8@0@lo:28/4 lens 440/432 e 0 to 1 dl 1630927301 ref 1 fl Rpc:XQr/0/ffffffff rc 0/-1 job:'touch.0' Lustre: 11785:0:(client.c:2285:ptlrpc_expire_one_request()) Skipped 113 previous similar messages Lustre: lustre-OST0000: Client 717fdd9d-1a9e-495a-9c80-e8efeeb57655 (at 0@lo) reconnecting Lustre: Skipped 2 previous similar messages LustreError: 15267:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0000-osp-MDT0001: fail to cancel 1 llog-records: rc = -116 LustreError: 15267:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0000-osp-MDT0001: fail to cancel 1 of 1 llog-records: rc = -116 LustreError: 18312:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0000: '3' migration was interrupted, run 'lfs migrate -m 2 -c 1 -H crush 3' to finish migration. Lustre: 11775:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff88026c114500 x1710150340333120/t0(0) o10->lustre-OST0000-osc-ffff880298c7e678@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630927347 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'cp.0' Lustre: 11785:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1630927299/real 1630927299] req@ffff8800a605ac00 x1710150338430144/t0(0) o2->lustre-OST0002-osc-ffff8802cfdd12a8@0@lo:28/4 lens 440/432 e 0 to 1 dl 1630927320 ref 1 fl Rpc:XQr/2/ffffffff rc 0/-1 job:'touch.0' Lustre: lustre-OST0002: Client 717fdd9d-1a9e-495a-9c80-e8efeeb57655 (at 0@lo) reconnecting Lustre: 11785:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1630927307/real 1630927307] req@ffff880079319f80 x1710150338301888/t0(0) o2->lustre-OST0000-osc-ffff8802cfdd12a8@0@lo:28/4 lens 440/432 e 0 to 1 dl 1630927326 ref 1 fl Rpc:XQr/2/ffffffff rc -11/-1 job:'touch.0' Lustre: 11785:0:(client.c:2285:ptlrpc_expire_one_request()) Skipped 59 previous similar messages Lustre: lustre-OST0000: Client 717fdd9d-1a9e-495a-9c80-e8efeeb57655 (at 0@lo) reconnecting Lustre: Skipped 2 previous similar messages ptlrpc_watchdog_fire: 78 callbacks suppressed Lustre: ll_ost05_004: service thread pid 19765 was inactive for 76.001 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 20 previous similar messages Lustre: ll_ost05_025: service thread pid 1261 was inactive for 76.000 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Lustre: Skipped 2 previous similar messages Pid: 1261, comm: ll_ost05_025 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] tgt_extent_lock+0xea/0x2a0 [ptlrpc] [<0>] ofd_getattr_hdl+0x365/0x720 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 20360, comm: ll_ost05_009 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] tgt_extent_lock+0xea/0x2a0 [ptlrpc] [<0>] ofd_getattr_hdl+0x365/0x720 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 30625, comm: ll_ost05_023 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] tgt_extent_lock+0xea/0x2a0 [ptlrpc] [<0>] ofd_getattr_hdl+0x365/0x720 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe LustreError: 15694:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 185286848 != fo_tot_granted 185409728 LustreError: 15694:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 36 previous similar messages LustreError: 15694:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 122880 LustreError: 15694:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 36 previous similar messages Lustre: lustre-OST0003: haven't heard from client 4daad2b5-7c77-41c5-b493-e6876f5214e1 (at 0@lo) in 50 seconds. I think it's dead, and I am evicting it. exp ffff880066768008, cur 1630927346 expire 1630927316 last 1630927296 Lustre: 30625:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (90/9s); client may timeout req@ffff880241a40cc0 x1710150337988032/t0(0) o1->4daad2b5-7c77-41c5-b493-e6876f5214e1@0@lo:501/0 lens 440/0 e 0 to 0 dl 1630927341 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'lfs.0' Lustre: 30625:0:(service.c:2328:ptlrpc_server_handle_request()) Skipped 149 previous similar messages LustreError: 24414:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8800742b5e00 x1710150340770048/t0(0) o104->lustre-OST0000@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 24414:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 13 previous similar messages Lustre: 5424:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (62/33s); client may timeout req@ffff88023b29d180 x1710150338295488/t0(0) o2->717fdd9d-1a9e-495a-9c80-e8efeeb57655@0@lo:434/0 lens 440/0 e 0 to 0 dl 1630927318 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'touch.0' Lustre: 5424:0:(service.c:2328:ptlrpc_server_handle_request()) Skipped 57 previous similar messages LustreError: 5184:0:(service.c:2290:ptlrpc_server_handle_request()) @@@ Dropping timed-out request from 12345-0@lo: deadline 16/43s ago req@ffff880095800040 x1710150339621696/t0(0) o101->4daad2b5-7c77-41c5-b493-e6876f5214e1@0@lo:468/0 lens 328/0 e 0 to 0 dl 1630927308 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'ln.0' LustreError: 5184:0:(service.c:2290:ptlrpc_server_handle_request()) Skipped 155 previous similar messages LustreError: 7654:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0000-osc-ffff880298c7e678: namespace resource [0x130:0x0:0x0].0x0 (ffff8802f41fa0c0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 7654:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 185 previous similar messages Lustre: dir [0x240000403:0x399:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: 11785:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1630927321/real 1630927321] req@ffff880295a857c0 x1710150338312832/t0(0) o2->lustre-OST0002-osc-ffff8802cfdd12a8@0@lo:28/4 lens 440/432 e 2 to 1 dl 1630927370 ref 1 fl Rpc:XQr/2/ffffffff rc -11/-1 job:'touch.0' Lustre: 11785:0:(client.c:2285:ptlrpc_expire_one_request()) Skipped 8 previous similar messages Lustre: lustre-OST0002: Client 717fdd9d-1a9e-495a-9c80-e8efeeb57655 (at 0@lo) reconnecting LustreError: 11-0: lustre-OST0000-osc-ffff8802cfdd12a8: operation ost_getattr to node 0@lo failed: rc = -107 LustreError: Skipped 42 previous similar messages Lustre: 15671:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (93/8s); client may timeout req@ffff880099d31300 x1710150339183680/t0(0) o1->4daad2b5-7c77-41c5-b493-e6876f5214e1@0@lo:538/0 lens 440/432 e 2 to 0 dl 1630927378 ref 1 fl Complete:/0/0 rc 0/0 job:'lfs.0' Lustre: 15671:0:(service.c:2328:ptlrpc_server_handle_request()) Skipped 392 previous similar messages LustreError: 30665:0:(ldlm_lockd.c:1425:ldlm_handle_enqueue0()) ### lock on destroyed export ffff8802c3a32548 ns: filter-lustre-OST0001_UUID lock: ffff8802b4d8ed00/0xd30819a9e2cf6f54 lrc: 1/0,0 mode: PW/PW res: [0x300000400:0xac6:0x0].0x0 rrc: 2 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x54801000000000 nid: 0@lo remote: 0xd30819a9e2ced102 expref: 487 pid: 30665 timeout: 0 lvb_type: 0 LustreError: 29315:0:(ofd_obd.c:501:ofd_destroy_export()) lustre-OST0000: cli 717fdd9d-1a9e-495a-9c80-e8efeeb57655/ffff88026dd612a8 has 61440 pending on destroyed export Lustre: 30668:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (93/10s); client may timeout req@ffff88009587bec0 x1710150339183744/t0(0) o1->4daad2b5-7c77-41c5-b493-e6876f5214e1@0@lo:538/0 lens 440/432 e 2 to 0 dl 1630927378 ref 1 fl Complete:/0/0 rc 0/0 job:'lfs.0' Lustre: 30668:0:(service.c:2328:ptlrpc_server_handle_request()) Skipped 10 previous similar messages Lustre: 11781:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.120@tcp:/lustre/fid: [0x240000403:0x3b4:0x0]/ may get corrupted (rc -108) Lustre: 11780:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.120@tcp:/lustre/fid: [0x280000405:0x100:0x0]/ may get corrupted (rc -108) Lustre: 11780:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.120@tcp:/lustre/fid: [0x280000405:0x10c:0x0]// may get corrupted (rc -108) Lustre: 11779:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.120@tcp:/lustre/fid: [0x200000404:0x2a9:0x0]/ may get corrupted (rc -108) Lustre: 11779:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.120@tcp:/lustre/fid: [0x200000404:0x264:0x0]/ may get corrupted (rc -108) Lustre: 14932:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 7s req@ffff88025d7aa598 x1710150344480832/t0(0) o105->LOV_OSC_UUID@0@lo:0/0 lens 392/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'' LustreError: 18044:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0001: '10' migration was interrupted, run 'lfs migrate -m 2 -c 1 -H crush 10' to finish migration. LustreError: 15688:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 174205888 != fo_tot_granted 174267328 LustreError: 15688:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 115 previous similar messages LustreError: 15688:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 4220672 != fo_tot_pending 4282112 LustreError: 15688:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 115 previous similar messages LustreError: 21064:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0000: '10' migration was interrupted, run 'lfs migrate -m 2 -c 1 -H crush 10' to finish migration. LustreError: 14966:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0003_UUID lock: ffff8800913d2200/0xd30819a9e2d4fa08 lrc: 3/0,0 mode: PW/PW res: [0x380000400:0xcc2:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 1048576->18446744073709551615) gid 0 flags: 0x60000400000020 nid: 0@lo remote: 0xd30819a9e2d4f9f3 expref: 2038 pid: 1232 timeout: 20224 lvb_type: 0 LustreError: 14966:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 15 previous similar messages LustreError: 14965:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88006fed1300 x1710150347545344/t0(0) o105->lustre-OST0003@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 14965:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 55 previous similar messages LustreError: 11-0: lustre-OST0003-osc-ffff8802cfdd12a8: operation ost_setattr to node 0@lo failed: rc = -107 LustreError: Skipped 32 previous similar messages Lustre: lustre-OST0003-osc-ffff8802cfdd12a8: Connection to lustre-OST0003 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 17 previous similar messages LustreError: 167-0: lustre-OST0003-osc-ffff8802cfdd12a8: This client was evicted by lustre-OST0003; in progress operations using this service will fail. LustreError: Skipped 7 previous similar messages LustreError: 25099:0:(ldlm_lockd.c:1425:ldlm_handle_enqueue0()) ### lock on destroyed export ffff8802dabe53d8 ns: filter-lustre-OST0003_UUID lock: ffff8802f41f61c0/0xd30819a9e2d99469 lrc: 3/0,0 mode: PW/PW res: [0xd69:0x0:0x0].0x0 rrc: 2 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x50000080020000 nid: 0@lo remote: 0xd30819a9e2d9945b expref: 1616 pid: 25099 timeout: 0 lvb_type: 0 LustreError: 11702:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0003-osc-ffff8802cfdd12a8: namespace resource [0x190:0x0:0x0].0x0 (ffff88009584c7c0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 11702:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 655 previous similar messages Lustre: lustre-OST0003-osc-ffff8802cfdd12a8: Connection restored to 192.168.123.120@tcp (at 0@lo) Lustre: Skipped 17 previous similar messages ------------[ cut here ]------------ WARNING: CPU: 8 PID: 24026 at /home/green/git/lustre-release/ldiskfs/ext4_jbd2.c:266 __ldiskfs_handle_dirty_metadata+0x1d2/0x230 [ldiskfs] Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) dm_flakey dm_mod loop zfs(PO) zunicode(PO) zzstd(O) zlua(O) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) jbd2 mbcache crc_t10dif crct10dif_generic sb_edac edac_core iosf_mbi crc32_pclmul ghash_clmulni_intel aesni_intel lrw gf128mul glue_helper ablk_helper cryptd virtio_balloon pcspkr i2c_piix4 virtio_console ip_tables rpcsec_gss_krb5 ata_generic drm_kms_helper pata_acpi ttm drm crct10dif_pclmul crct10dif_common drm_panel_orientation_quirks ata_piix crc32c_intel serio_raw virtio_blk i2c_core libata floppy [last unloaded: libcfs] CPU: 8 PID: 24026 Comm: mdt04_010 Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011 Call Trace: [<ffffffff817d9418>] dump_stack+0x19/0x1b [<ffffffff8108d538>] __warn+0xd8/0x100 [<ffffffff8108d67d>] warn_slowpath_null+0x1d/0x20 [<ffffffffa0da5262>] __ldiskfs_handle_dirty_metadata+0x1d2/0x230 [ldiskfs] [<ffffffffa0d8533d>] ldiskfs_getblk+0x13d/0x210 [ldiskfs] [<ffffffffa0d85437>] ldiskfs_bread+0x27/0xe0 [ldiskfs] [<ffffffffa0cd0e97>] osd_write+0x4e7/0xcb0 [osd_ldiskfs] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffffa04dd6f3>] dt_record_write+0x33/0x120 [obdclass] [<ffffffffa049c6b2>] llog_osd_write_rec+0x1052/0x19a0 [obdclass] [<ffffffffa0489760>] llog_write_rec+0x290/0x590 [obdclass] [<ffffffffa048f4ee>] llog_cat_add_rec+0x23e/0x950 [obdclass] [<ffffffffa04865ff>] llog_add+0x17f/0x1f0 [obdclass] [<ffffffffa04b7a39>] ? lprocfs_counter_add+0xf9/0x160 [obdclass] [<ffffffffa082060d>] sub_updates_write+0xb6b/0xe32 [ptlrpc] [<ffffffff811beccd>] ? __free_pages+0x1d/0x30 [<ffffffff8121efe4>] ? kfree+0x1e4/0x200 [<ffffffffa07f0892>] top_trans_stop+0x7c2/0xf60 [ptlrpc] [<ffffffffa0fb3f3c>] lod_trans_stop+0x25c/0x340 [lod] [<ffffffffa0e88408>] mdd_trans_stop+0x28/0x16e [mdd] [<ffffffffa0e69b2f>] mdd_migrate_object+0xf1f/0x1ba0 [mdd] [<ffffffffa04dc375>] ? lu_object_find_at+0x285/0xb10 [obdclass] [<ffffffffa0e6ab86>] mdd_migrate+0x3d6/0x8e0 [mdd] [<ffffffffa0f3c0d8>] mdo_migrate+0x4c/0x4e [mdt] [<ffffffffa0efbe53>] mdt_reint_migrate+0x1023/0x11d0 [mdt] [<ffffffffa0efc087>] mdt_reint_rec+0x87/0x240 [mdt] [<ffffffffa0ed155c>] mdt_reint_internal+0x76c/0xb40 [mdt] [<ffffffffa0ede657>] mdt_reint+0x67/0x150 [mdt] [<ffffffffa07df26e>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<ffffffffa034745e>] ? libcfs_nid2str_r+0xfe/0x130 [lnet] [<ffffffffa078a990>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<ffffffffa078c529>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffffa078b950>] ? ptlrpc_wait_event+0x620/0x620 [ptlrpc] [<ffffffff810ba0f4>] kthread+0xe4/0xf0 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 ---[ end trace 5b2c3ed042866442 ]--- LDISKFS-fs: ldiskfs_getblk:888: aborting transaction: error 28 in __ldiskfs_handle_dirty_metadata LDISKFS-fs error (device dm-1): ldiskfs_getblk:888: inode #163: block 20469: comm mdt04_010: journal_dirty_metadata failed: handle type 0 started at line 1972, credits 512/0, errcode -28 Aborting journal on device dm-1-8. LDISKFS-fs (dm-1): Remounting filesystem read-only LustreError: 24026:0:(osd_io.c:2103:osd_ldiskfs_write_record()) lustre-MDT0001: error reading offset 270336 (block 66, size 40, offs 270320), credits 0/0: rc = -28 LustreError: 24026:0:(update_trans.c:985:top_trans_stop()) lustre-MDT0001-osd: write updates failed: rc = -28 LustreError: 24026:0:(osd_io.c:2110:osd_ldiskfs_write_record()) journal_get_write_access() returned error -30 LustreError: 24026:0:(tgt_lastrcvd.c:1253:tgt_add_reply_data()) lustre-MDT0001: can't update reply_data file: rc = -30 LustreError: 24026:0:(osd_handler.c:2075:osd_trans_stop()) lustre-MDT0001: failed in transaction hook: rc = -30 LDISKFS-fs error (device dm-1) in osd_trans_stop:2082: error 28 LustreError: 24026:0:(osd_handler.c:2085:osd_trans_stop()) lustre-MDT0001: failed to stop transaction: rc = -28 LustreError: 15250:0:(osd_handler.c:1779:osd_trans_commit_cb()) transaction @0xffff88021f417788 commit error: 2 LustreError: 24026:0:(update_trans.c:1011:top_trans_stop()) lustre-MDT0001-osd: stop trans failed: rc = -30 LustreError: 24026:0:(update_trans.c:1084:top_trans_stop()) lustre-OST0003-osc-MDT0001: stop trans failed: rc = -30 LustreError: 17067:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -30 LustreError: 15072:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0001-osp-MDT0000: fail to cancel 1 llog-records: rc = -116 LustreError: 15072:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0001-osp-MDT0000: fail to cancel 1 of 1 llog-records: rc = -116 LustreError: 12308:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -30 | Externally reported by onyx-68 boilpot email |
racer test 1: racer on clients: centos-30.localnet DURATION=2700 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa1068e52>] ll_migrate+0x9b2/0xec0 [lustre] PGD 800000029222d067 PUD 2776c1067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: loop zfs(PO) zunicode(PO) zzstd(O) zlua(O) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) jbd2 mbcache lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) dm_flakey dm_mod libcfs(OE) crc_t10dif crct10dif_generic sb_edac edac_core iosf_mbi crc32_pclmul ghash_clmulni_intel aesni_intel lrw gf128mul glue_helper ablk_helper cryptd pcspkr virtio_balloon virtio_console i2c_piix4 ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm crct10dif_pclmul ata_piix crct10dif_common drm_panel_orientation_quirks crc32c_intel serio_raw virtio_blk i2c_core libata floppy CPU: 13 PID: 5065 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011 task: ffff8800812d3760 ti: ffff88025100c000 task.ti: ffff88025100c000 RIP: 0010:[<ffffffffa1068e52>] [<ffffffffa1068e52>] ll_migrate+0x9b2/0xec0 [lustre] RSP: 0000:ffff88025100fbc8 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff88007bb81db8 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff880320c4c258 RDI: ffff880320c4c238 RBP: ffff88025100fc48 R08: ffff88024e2f67d8 R09: ffff88025100fa60 R10: 0000000000000001 R11: ffff88028593d7c0 R12: 0000000000000000 R13: ffff8802563f4598 R14: ffff880072adcb18 R15: 0000000000000030 FS: 00007fce3bc05740(0000) GS:ffff880331d40000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 CR2: 0000000000000008 CR3: 00000002704f2000 CR4: 00000000001607e0 Call Trace: [<ffffffff81242213>] ? __check_object_size+0x1c3/0x220 [<ffffffffa1056561>] ll_dir_ioctl+0x5d01/0x6ed0 [lustre] [<ffffffff811b515b>] ? unlock_page+0x2b/0x30 [<ffffffff811e5e2e>] ? do_read_fault.isra.64+0x12e/0x170 [<ffffffff811ea1cf>] ? handle_pte_fault+0x2ef/0xca0 [<ffffffff811ecf6d>] ? __handle_mm_fault+0x31d/0x720 [<ffffffff8125b3fd>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff81245a38>] ? vfs_write+0x178/0x230 [<ffffffff8125b751>] SyS_ioctl+0xa1/0xc0 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 | Lustre: lustre-OST0001-osc-ffff88029ad2a548: disconnect after 21s idle Lustre: lustre-OST0002-osc-ffff88029ad2a548: disconnect after 22s idle Lustre: 11111:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 512 < left 1172, rollback = 7 Lustre: 11111:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 11111:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 1/89/0 Lustre: 11111:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1163/1172/0, punch: 0/0/0, quota 3/3/0 Lustre: 11111:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 11111:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 11111:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 512 < left 1172, rollback = 7 Lustre: 11111:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 237 previous similar messages Lustre: 11110:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 11110:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 491 previous similar messages Lustre: 11110:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 1/89/0 Lustre: 11110:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 491 previous similar messages Lustre: 11110:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 576/585/0, punch: 0/0/0, quota 3/3/0 Lustre: 11110:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 491 previous similar messages Lustre: 11110:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 11110:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 491 previous similar messages Lustre: 11110:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 11110:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 491 previous similar messages Lustre: 11091:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 512 < left 49976, rollback = 7 Lustre: 11091:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 1500 previous similar messages Lustre: 11091:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 1162/4648/0, destroy: 0/0/0 Lustre: 11091:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 1246 previous similar messages Lustre: 11091:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 1164/15/0 Lustre: 11091:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 1246 previous similar messages Lustre: 11091:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 5811/49976/0, punch: 0/0/0, quota 10/106/0 Lustre: 11091:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 1246 previous similar messages Lustre: 11091:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 1162/19754/0, delete: 0/0/0 Lustre: 11091:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 1246 previous similar messages Lustre: 11091:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 11091:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 1246 previous similar messages Lustre: 16624:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 512 < left 1164, rollback = 7 Lustre: 16624:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 4661 previous similar messages Lustre: 16624:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 16624:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 5099 previous similar messages Lustre: 16624:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 1/89/0 Lustre: 16624:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 5099 previous similar messages Lustre: 16624:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1155/1164/1, punch: 0/0/0, quota 3/3/0 Lustre: 16624:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 5099 previous similar messages Lustre: 16624:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 16624:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 5099 previous similar messages Lustre: 16624:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 16624:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 5099 previous similar messages Lustre: 11100:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 501 < left 866, rollback = 2 Lustre: 11108:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 512 < left 1529, rollback = 7 Lustre: 11108:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 1154 previous similar messages Lustre: 11108:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 11108:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 726 previous similar messages Lustre: 11108:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 1/89/0 Lustre: 11108:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 726 previous similar messages Lustre: 11108:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1520/1529/0, punch: 0/0/0, quota 3/3/0 Lustre: 11108:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 726 previous similar messages Lustre: 11108:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 11108:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 726 previous similar messages Lustre: 11108:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 11108:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 726 previous similar messages Lustre: 17104:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 509 < left 10932, rollback = 7 Lustre: 17104:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 2486 previous similar messages Lustre: 17104:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 254/1016/0, destroy: 0/0/0 Lustre: 17104:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 2485 previous similar messages Lustre: 17104:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/1, xattr_set: 256/15/0 Lustre: 17104:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 2485 previous similar messages Lustre: 17104:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1271/10932/0, punch: 0/0/0, quota 14/174/2 Lustre: 17104:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 2485 previous similar messages Lustre: 17104:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 254/4318/0, delete: 0/0/0 Lustre: 17104:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 2485 previous similar messages Lustre: 17104:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 17104:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 2485 previous similar messages Lustre: 16824:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 501 < left 1400, rollback = 2 Lustre: 16824:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 8 previous similar messages Lustre: 16824:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 3/12/6, destroy: 1/4/0 Lustre: 16824:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 1016 previous similar messages Lustre: 16824:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 15/1400/0 Lustre: 16824:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 1016 previous similar messages Lustre: 16824:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 5/136/0, punch: 0/0/0, quota 10/106/0 Lustre: 16824:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 1016 previous similar messages Lustre: 16824:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 12/215/4, delete: 2/5/1 Lustre: 16824:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 1016 previous similar messages Lustre: 16824:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 5/5/0, ref_del: 2/2/0 Lustre: 16824:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 1016 previous similar messages sched: RT throttling activated Lustre: 14392:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 494 < left 582, rollback = 2 Lustre: 14392:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 17 previous similar messages Lustre: 15069:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 512 < left 1132, rollback = 7 Lustre: 15069:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 1016 previous similar messages LustreError: 11148:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0002: failed to get lu_attr of [0x280000404:0x36:0x0]: rc = -2 LustreError: 12814:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0001: '5' migration was interrupted, run 'lfs migrate -m 0 -c 1 -H crush 5' to finish migration. Lustre: 13882:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 8: before 511 < left 19684, rollback = 9 Lustre: 13882:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 17 previous similar messages Lustre: 18002:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 18002:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 14214 previous similar messages Lustre: 18002:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 1/89/0 Lustre: 18002:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 14214 previous similar messages Lustre: 18002:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1583/1592/0, punch: 0/0/0, quota 3/3/0 Lustre: 18002:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 14214 previous similar messages Lustre: 18002:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 18002:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 14214 previous similar messages Lustre: 18002:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 18002:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 14213 previous similar messages Lustre: 18002:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 511 < left 1591, rollback = 7 Lustre: 18002:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 14179 previous similar messages Lustre: 15069:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 505 < left 624, rollback = 2 Lustre: 11102:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 8: before 510 < left 9791, rollback = 9 Lustre: 11102:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 10 previous similar messages LustreError: 10978:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0002_UUID lock: ffff88027749e940/0xbfe3d35722953d53 lrc: 3/0,0 mode: PW/PW res: [0x340000401:0x170:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000000020020 nid: 0@lo remote: 0xbfe3d35722953d30 expref: 3530 pid: 18022 timeout: 410 lvb_type: 0 LustreError: 14919:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff880256fa4500 x1710112928228224/t0(0) o105->lustre-OST0002@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 11-0: lustre-OST0002-osc-ffff88029ad2a548: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-OST0002-osc-ffff88029ad2a548: Connection to lustre-OST0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 19504:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802843aa5c0 x1710112928253952/t0(0) o104->lustre-OST0002@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 19504:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 9 previous similar messages LustreError: 11-0: lustre-OST0002-osc-ffff88029ad2a548: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 2 previous similar messages LustreError: 11-0: lustre-OST0002-osc-ffff88029ad2a548: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: 167-0: lustre-OST0002-osc-ffff88029ad2a548: This client was evicted by lustre-OST0002; in progress operations using this service will fail. LustreError: 19536:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff880254f97700 x1710112928305088/t0(0) o104->lustre-OST0002@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' Lustre: 13087:0:(service.c:1378:ptlrpc_at_send_early_reply()) @@@ Already past deadline (-1s), not sending early reply. Consider increasing at_early_margin (5)? req@ffff8802792c5180 x1710112926604672/t0(0) o1->0bf25ca2-fb17-4410-82fe-9fe450342b30@0@lo:469/0 lens 440/0 e 0 to 0 dl 1630891069 ref 2 fl New:/0/ffffffff rc 0/-1 job:'lfs.0' Lustre: 18719:0:(service.c:1378:ptlrpc_at_send_early_reply()) @@@ Already past deadline (-1s), not sending early reply. Consider increasing at_early_margin (5)? req@ffff880262e90680 x1710112926640384/t0(0) o1->0bf25ca2-fb17-4410-82fe-9fe450342b30@0@lo:470/0 lens 440/0 e 0 to 0 dl 1630891070 ref 2 fl New:/0/ffffffff rc 0/-1 job:'lfs.0' Lustre: 18719:0:(service.c:1378:ptlrpc_at_send_early_reply()) Skipped 42 previous similar messages Lustre: 4128:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.130@tcp:/lustre/fid: [0x280000403:0x56:0x0]// may get corrupted (rc -108) Lustre: 19504:0:(service.c:1378:ptlrpc_at_send_early_reply()) @@@ Already past deadline (-1s), not sending early reply. Consider increasing at_early_margin (5)? req@ffff880287e23ec0 x1710112926866304/t0(0) o101->0bf25ca2-fb17-4410-82fe-9fe450342b30@0@lo:473/0 lens 328/0 e 0 to 0 dl 1630891073 ref 2 fl New:/0/ffffffff rc 0/-1 job:'ln.0' Lustre: 19504:0:(service.c:1378:ptlrpc_at_send_early_reply()) Skipped 175 previous similar messages LustreError: 10978:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0003_UUID lock: ffff88028aaf6d00/0xbfe3d3572293f923 lrc: 3/0,0 mode: PW/PW res: [0x380000400:0x5:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000400000020 nid: 0@lo remote: 0xbfe3d3572293f915 expref: 3266 pid: 12675 timeout: 419 lvb_type: 0 LustreError: 18324:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630891075 with bad export cookie 13827127652153247927 LustreError: 17755:0:(ldlm_lockd.c:1425:ldlm_handle_enqueue0()) ### lock on destroyed export ffff880297486fc8 ns: filter-lustre-OST0003_UUID lock: ffff8802675fcf00/0xbfe3d357229c40b7 lrc: 3/0,0 mode: --/PW res: [0x380000400:0x235:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x50000000020000 nid: 0@lo remote: 0xbfe3d357229c405c expref: 3075 pid: 17755 timeout: 0 lvb_type: 0 LustreError: 11-0: lustre-OST0003-osc-ffff880298db12a8: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 1 previous similar message Lustre: lustre-OST0003-osc-ffff880298db12a8: Connection to lustre-OST0003 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0003-osc-ffff880298db12a8: This client was evicted by lustre-OST0003; in progress operations using this service will fail. Lustre: 18719:0:(service.c:1378:ptlrpc_at_send_early_reply()) @@@ Already past deadline (-1s), not sending early reply. Consider increasing at_early_margin (5)? req@ffff8802638abec0 x1710112926991488/t0(0) o101->0bf25ca2-fb17-4410-82fe-9fe450342b30@0@lo:477/0 lens 328/0 e 0 to 0 dl 1630891077 ref 2 fl New:/0/ffffffff rc 0/-1 job:'ln.0' Lustre: 18719:0:(service.c:1378:ptlrpc_at_send_early_reply()) Skipped 440 previous similar messages LustreError: 20399:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0002-osc-ffff88029ad2a548: namespace resource [0x340000401:0xf:0x0].0x0 (ffff880081428f40) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 10978:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 103s: evicting client at 0@lo ns: filter-lustre-OST0003_UUID lock: ffff880271f18400/0xbfe3d357229547f6 lrc: 3/0,0 mode: PW/PW res: [0x380000401:0x6:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000400030020 nid: 0@lo remote: 0xbfe3d357229547ef expref: 3775 pid: 17760 timeout: 420 lvb_type: 0 LustreError: 20399:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0002-osc-ffff88029ad2a548: namespace resource [0x340000401:0x1f:0x0].0x0 (ffff880271019bc0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 12690:0:(ldlm_lockd.c:1342:ldlm_handle_enqueue0()) ### lock on disconnected export ffff880298c76fc8 ns: filter-lustre-OST0003_UUID lock: ffff88007d5dcf00/0xbfe3d357229c6c9a lrc: 2/0,0 mode: --/PR res: [0x380000400:0x5a7:0x0].0x0 rrc: 2 type: EXT [0->0] (req 0->0) gid 0 flags: 0x40000000000000 nid: local remote: 0xbfe3d357229c5e2a expref: -99 pid: 12690 timeout: 0 lvb_type: 0 LustreError: 19453:0:(service.c:2290:ptlrpc_server_handle_request()) @@@ Dropping timed-out request from 12345-0@lo: deadline 43/11s ago req@ffff880088e30cc0 x1710112926602880/t0(0) o1->0bf25ca2-fb17-4410-82fe-9fe450342b30@0@lo:469/0 lens 440/0 e 0 to 0 dl 1630891069 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'lfs.0' Lustre: 19453:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (43/11s); client may timeout req@ffff880088e30cc0 x1710112926602880/t0(0) o1->0bf25ca2-fb17-4410-82fe-9fe450342b30@0@lo:469/0 lens 440/0 e 0 to 0 dl 1630891069 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'lfs.0' LustreError: 11-0: lustre-OST0003-osc-ffff88029ad2a548: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-OST0003-osc-ffff88029ad2a548: Connection to lustre-OST0003 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete ------------[ cut here ]------------ LustreError: 10977:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff880258ab8040 x1710112928810304/t0(0) o105->lustre-OST0003@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' WARNING: CPU: 6 PID: 11133 at /home/green/git/lustre-release/ldiskfs/namei.c:3331 ldiskfs_orphan_add+0x11e/0x2a0 [ldiskfs] LustreError: 167-0: lustre-OST0003-osc-ffff88029ad2a548: This client was evicted by lustre-OST0003; in progress operations using this service will fail. Modules linked in: loop zfs(PO) zunicode(PO) zzstd(O) zlua(O) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) jbd2 mbcache lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) dm_flakey dm_mod libcfs(OE) crc_t10dif crct10dif_generic sb_edac edac_core iosf_mbi crc32_pclmul ghash_clmulni_intel aesni_intel lrw gf128mul glue_helper ablk_helper cryptd pcspkr virtio_balloon virtio_console i2c_piix4 ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm crct10dif_pclmul ata_piix crct10dif_common drm_panel_orientation_quirks crc32c_intel serio_raw virtio_blk i2c_core libata floppy CPU: 6 PID: 11133 Comm: mdt_rdpg03_000 Tainted: P OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011 Call Trace: [<ffffffff817d9418>] dump_stack+0x19/0x1b [<ffffffff8108d538>] __warn+0xd8/0x100 [<ffffffff8108d67d>] warn_slowpath_null+0x1d/0x20 [<ffffffffa0c004ae>] ldiskfs_orphan_add+0x11e/0x2a0 [ldiskfs] [<ffffffffa0bf19eb>] ldiskfs_xattr_inode_orphan_add.constprop.17+0xbb/0x110 [ldiskfs] [<ffffffffa0bf4417>] ldiskfs_xattr_delete_inode+0x77/0x3c0 [ldiskfs] [<ffffffffa0c14aa0>] ? ldiskfs_evict_inode+0x190/0x670 [ldiskfs] [<ffffffffa0c14ad9>] ldiskfs_evict_inode+0x1c9/0x670 [ldiskfs] [<ffffffff812649af>] evict+0xaf/0x180 [<ffffffff81264de5>] iput+0xf5/0x180 [<ffffffffa0ca9d08>] osd_object_delete+0x1f8/0x370 [osd_ldiskfs] [<ffffffffa04c459d>] lu_object_free.isra.26+0xcd/0x1d0 [obdclass] [<ffffffffa04c8cf1>] lu_object_put+0xa1/0x480 [obdclass] [<ffffffffa0e61af4>] mdt_mfd_close+0x1f4/0x870 [mdt] [<ffffffffa0e67e14>] mdt_close_internal+0xb4/0x240 [mdt] [<ffffffffa0e68224>] mdt_close+0x284/0x850 [mdt] [<ffffffffa07cb2d4>] ? tgt_request_preprocess.isra.27+0xf4/0x850 [ptlrpc] [<ffffffffa07cc26e>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<ffffffffa032345e>] ? libcfs_nid2str_r+0xfe/0x130 [lnet] [<ffffffffa0777990>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<ffffffff810a20ad>] ? mod_timer+0x1dd/0x260 [<ffffffffa0779529>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffffa0778950>] ? ptlrpc_wait_event+0x620/0x620 [ptlrpc] [<ffffffff810ba0f4>] kthread+0xe4/0xf0 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 ---[ end trace 58aae1581d9a010e ]--- LustreError: 20550:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0003-osc-ffff880298db12a8: namespace resource [0x380000400:0x10:0x0].0x0 (ffff880280d791c0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 20550:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 891 previous similar messages Lustre: ll_ost06_011: service thread pid 18717 was inactive for 62.178 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 18717, comm: ll_ost06_011 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] tgt_extent_lock+0xea/0x2a0 [ptlrpc] Lustre: 17263:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 505 < left 849, rollback = 2 Lustre: 17263:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 2 previous similar messages [<0>] ofd_getattr_hdl+0x365/0x720 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 18715, comm: ll_ost06_010 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] tgt_extent_lock+0xea/0x2a0 [ptlrpc] [<0>] ofd_getattr_hdl+0x365/0x720 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 18042, comm: ll_ost06_006 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] tgt_extent_lock+0xea/0x2a0 [ptlrpc] [<0>] ofd_getattr_hdl+0x365/0x720 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: ll_ost06_001: service thread pid 12685 was inactive for 62.924 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: ll_ost06_020: service thread pid 19430 was inactive for 62.144 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 9 previous similar messages LustreError: 10978:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0001_UUID lock: ffff880266e861c0/0xbfe3d357229661fc lrc: 3/0,0 mode: PW/PW res: [0x300000400:0x389:0x0].0x0 rrc: 4 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000400000020 nid: 0@lo remote: 0xbfe3d357229661bd expref: 3128 pid: 17901 timeout: 433 lvb_type: 0 LustreError: 14920:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88025a62b240 x1710112929303872/t0(0) o105->lustre-OST0001@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 10965:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630891090 with bad export cookie 13827127652153247913 LustreError: 11-0: lustre-OST0001-osc-ffff880298db12a8: operation ost_getattr to node 0@lo failed: rc = -107 Lustre: lustre-OST0001-osc-ffff880298db12a8: Connection to lustre-OST0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: 4122:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.130@tcp:/lustre/fid: [0x280000403:0x96:0x0]// may get corrupted (rc -108) LustreError: 167-0: lustre-OST0001-osc-ffff880298db12a8: This client was evicted by lustre-OST0001; in progress operations using this service will fail. Lustre: ll_ost06_053: service thread pid 19545 was inactive for 62.024 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 5 previous similar messages Lustre: 4122:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.130@tcp:/lustre/fid: [0x200000404:0x2c:0x0]/ may get corrupted (rc -108) Lustre: 19536:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (69/2s); client may timeout req@ffff8802635fb880 x1710112926612608/t0(0) o1->0bf25ca2-fb17-4410-82fe-9fe450342b30@0@lo:495/0 lens 440/432 e 1 to 0 dl 1630891095 ref 1 fl Complete:/0/0 rc 0/0 job:'lfs.0' Lustre: 19536:0:(service.c:2328:ptlrpc_server_handle_request()) Skipped 42 previous similar messages LustreError: 19504:0:(service.c:2290:ptlrpc_server_handle_request()) @@@ Dropping timed-out request from 12345-0@lo: deadline 68/6s ago req@ffff8802c626d428 x1710112926620864/t0(0) o1->0bf25ca2-fb17-4410-82fe-9fe450342b30@0@lo:494/0 lens 440/0 e 1 to 0 dl 1630891094 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'lfs.0' LustreError: 19504:0:(service.c:2290:ptlrpc_server_handle_request()) Skipped 42 previous similar messages Lustre: 19504:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (68/6s); client may timeout req@ffff8802c626d428 x1710112926620864/t0(0) o1->0bf25ca2-fb17-4410-82fe-9fe450342b30@0@lo:494/0 lens 440/0 e 1 to 0 dl 1630891094 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'lfs.0' Lustre: 19504:0:(service.c:2328:ptlrpc_server_handle_request()) Skipped 7 previous similar messages LustreError: 20691:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0003-osc-ffff88029ad2a548: namespace resource [0x380000401:0xf:0x0].0x0 (ffff8802716247c0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 20691:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 667 previous similar messages Lustre: 4126:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1630891036/real 1630891036] req@ffff880258888cc0 x1710112927241728/t0(0) o101->lustre-OST0000-osc-ffff880298db12a8@0@lo:28/4 lens 328/400 e 0 to 1 dl 1630891091 ref 1 fl Rpc:XQr/0/ffffffff rc 0/-1 job:'ln.0' Lustre: lustre-OST0000-osc-ffff880298db12a8: Connection to lustre-OST0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: lustre-OST0000: Client 0bf25ca2-fb17-4410-82fe-9fe450342b30 (at 0@lo) reconnecting Lustre: lustre-OST0000-osc-ffff88029ad2a548: Connection restored to 192.168.123.130@tcp (at 0@lo) Lustre: lustre-OST0002: deleting orphan objects from 0x340000400:1876 to 0x340000400:1921 Lustre: 19506:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (43/25s); client may timeout req@ffff8802572ff700 x1710112926874176/t0(0) o1->0bf25ca2-fb17-4410-82fe-9fe450342b30@0@lo:479/0 lens 440/0 e 0 to 0 dl 1630891079 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'lfs.0' LustreError: 19703:0:(service.c:2290:ptlrpc_server_handle_request()) @@@ Dropping timed-out request from 12345-0@lo: deadline 43/26s ago req@ffff8802572f8680 x1710112926874304/t0(0) o1->0bf25ca2-fb17-4410-82fe-9fe450342b30@0@lo:479/0 lens 440/0 e 0 to 0 dl 1630891079 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'lfs.0' LustreError: 19703:0:(service.c:2290:ptlrpc_server_handle_request()) Skipped 76 previous similar messages Lustre: 11129:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 509 < left 24735, rollback = 7 Lustre: 11129:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 11793 previous similar messages Lustre: 11129:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 575/2300/0, destroy: 1/4/1 Lustre: 11129:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 11827 previous similar messages Lustre: 11129:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 575/0/0 Lustre: 11129:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 11827 previous similar messages Lustre: 11129:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 2876/24735/0, punch: 0/0/0, quota 10/74/0 Lustre: 11129:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 11827 previous similar messages Lustre: 11129:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 575/9775/0, delete: 2/5/1 Lustre: 11129:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 11827 previous similar messages Lustre: 11129:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 1/1/1 Lustre: 11129:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 11827 previous similar messages LustreError: 20811:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0001-osc-ffff880298db12a8: namespace resource [0x300000400:0x10:0x0].0x0 (ffff880280d78540) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 20811:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 994 previous similar messages LustreError: 19469:0:(service.c:2290:ptlrpc_server_handle_request()) @@@ Dropping timed-out request from 12345-0@lo: deadline 68/14s ago req@ffff880264319300 x1710112926621632/t0(0) o1->0bf25ca2-fb17-4410-82fe-9fe450342b30@0@lo:494/0 lens 440/0 e 1 to 0 dl 1630891094 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'lfs.0' LustreError: 19469:0:(service.c:2290:ptlrpc_server_handle_request()) Skipped 429 previous similar messages Lustre: 19504:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (71/10s); client may timeout req@ffff88026b0d7700 x1710112926644992/t0(0) o1->0bf25ca2-fb17-4410-82fe-9fe450342b30@0@lo:499/0 lens 440/0 e 1 to 0 dl 1630891099 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'lfs.0' Lustre: 19504:0:(service.c:2328:ptlrpc_server_handle_request()) Skipped 720 previous similar messages Lustre: 19536:0:(service.c:1436:ptlrpc_at_send_early_reply()) @@@ Could not add any time (72/30), not sending early reply req@ffff880254907700 x1710112927606656/t0(0) o101->03033789-bf5a-4e49-a70e-41fbce44720e@0@lo:513/0 lens 328/0 e 1 to 0 dl 1630891182 ref 2 fl New:/0/ffffffff rc 0/-1 job:'ln.0' Lustre: lustre-OST0002: haven't heard from client 03033789-bf5a-4e49-a70e-41fbce44720e (at 0@lo) in 53 seconds. I think it's dead, and I am evicting it. exp ffff8800823b1bf8, cur 1630891122 expire 1630891092 last 1630891069 Lustre: lustre-OST0003: haven't heard from client 0bf25ca2-fb17-4410-82fe-9fe450342b30 (at 0@lo) in 47 seconds. I think it's dead, and I am evicting it. exp ffff8802f3898008, cur 1630891123 expire 1630891093 last 1630891076 LustreError: 20399:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0002-osc-ffff88029ad2a548: namespace resource [0x340000400:0x30:0x0].0x0 (ffff88026c7d11c0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 20399:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 753 previous similar messages Lustre: lustre-OST0002-osc-ffff88029ad2a548: Connection restored to 192.168.123.130@tcp (at 0@lo) Lustre: Skipped 4 previous similar messages Lustre: 17204:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 8: before 511 < left 25872, rollback = 9 Lustre: 17204:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 17 previous similar messages Lustre: 19483:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (93/6s); client may timeout req@ffff880297b870c0 x1710112926586368/t0(0) o1->0bf25ca2-fb17-4410-82fe-9fe450342b30@0@lo:518/0 lens 440/432 e 2 to 0 dl 1630891118 ref 1 fl Complete:/0/0 rc 0/0 job:'lfs.0' Lustre: 19483:0:(service.c:2328:ptlrpc_server_handle_request()) Skipped 1810 previous similar messages LustreError: 11-0: lustre-OST0002-osc-ffff88029ad2a548: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-OST0002-osc-ffff88029ad2a548: Connection to lustre-OST0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 4 previous similar messages LustreError: Skipped 3 previous similar messages LustreError: 167-0: lustre-OST0002-osc-ffff88029ad2a548: This client was evicted by lustre-OST0002; in progress operations using this service will fail. INFO: task lfs:17451 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. lfs D ffff88029595f3c0 11360 17451 16290 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff8139a526>] ima_file_check+0xa6/0x1b0 [<ffffffff81255397>] do_last+0x587/0x1280 [<ffffffff812546de>] ? link_path_walk+0x27e/0x8c0 [<ffffffff812570ad>] path_openat+0xcd/0x5b0 [<ffffffff81258dfd>] do_filp_open+0x4d/0xb0 [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e326e>] ? _raw_spin_unlock+0xe/0x20 [<ffffffff81267893>] ? __alloc_fd+0xc3/0x170 [<ffffffff812447e4>] do_sys_open+0x124/0x220 [<ffffffff812448fe>] SyS_open+0x1e/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task chown:18381 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. chown D ffff88027597bd58 12760 18381 16588 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff81257b2d>] ? putname+0x3d/0x60 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff81242ddd>] chown_common.isra.15+0xed/0x1b0 [<ffffffff817e8655>] ? __do_page_fault+0x1c5/0x500 [<ffffffff8124436e>] SyS_fchownat+0xce/0x120 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task touch:19035 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. touch D ffff88028d282980 12944 19035 17362 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff817d5bb2>] lookup_slow+0x33/0xa7 [<ffffffff81254bd0>] link_path_walk+0x770/0x8c0 [<ffffffff81254924>] link_path_walk+0x4c4/0x8c0 [<ffffffff8124818c>] ? get_empty_filp+0x5c/0x1f0 [<ffffffff8125708e>] path_openat+0xae/0x5b0 [<ffffffff81258dfd>] do_filp_open+0x4d/0xb0 [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e326e>] ? _raw_spin_unlock+0xe/0x20 [<ffffffff81267893>] ? __alloc_fd+0xc3/0x170 [<ffffffff812447e4>] do_sys_open+0x124/0x220 [<ffffffff812448fe>] SyS_open+0x1e/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task setfattr:19037 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. setfattr D ffff880290098040 12904 19037 16225 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff817d5bb2>] lookup_slow+0x33/0xa7 [<ffffffff81256d6e>] path_lookupat+0x89e/0x8d0 [<ffffffff8121f126>] ? kmem_cache_alloc+0x126/0x2e0 [<ffffffff81257b9f>] ? getname_flags+0x4f/0x1a0 [<ffffffff81256dcb>] filename_lookup+0x2b/0xc0 [<ffffffff81258cc7>] user_path_at_empty+0x67/0xc0 [<ffffffff817e8655>] ? __do_page_fault+0x1c5/0x500 [<ffffffff81258d31>] user_path_at+0x11/0x20 [<ffffffff81270cc3>] SyS_removexattr+0x43/0xd0 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task ln:19056 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. ln D ffff880283e6d2c0 12760 19056 16462 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff817d5bb2>] lookup_slow+0x33/0xa7 [<ffffffff81256d6e>] path_lookupat+0x89e/0x8d0 [<ffffffff8121f126>] ? kmem_cache_alloc+0x126/0x2e0 [<ffffffff81257b9f>] ? getname_flags+0x4f/0x1a0 [<ffffffff81256dcb>] filename_lookup+0x2b/0xc0 [<ffffffff81258cc7>] user_path_at_empty+0x67/0xc0 [<ffffffff81258d31>] user_path_at+0x11/0x20 [<ffffffff8124b6b3>] vfs_fstatat+0x63/0xc0 [<ffffffff8124ba7e>] SYSC_newstat+0x2e/0x60 [<ffffffff817e8a76>] ? trace_do_page_fault+0x56/0x170 [<ffffffff8124bf4e>] SyS_newstat+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task truncate:19074 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. truncate D ffff8802744fb760 12904 19074 17245 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff817d5bb2>] lookup_slow+0x33/0xa7 [<ffffffff81256d6e>] path_lookupat+0x89e/0x8d0 [<ffffffff8121f126>] ? kmem_cache_alloc+0x126/0x2e0 [<ffffffff81257b9f>] ? getname_flags+0x4f/0x1a0 [<ffffffff81256dcb>] filename_lookup+0x2b/0xc0 [<ffffffff81258cc7>] user_path_at_empty+0x67/0xc0 [<ffffffff817e8655>] ? __do_page_fault+0x1c5/0x500 [<ffffffff81258d31>] user_path_at+0x11/0x20 [<ffffffff81243914>] do_sys_truncate+0x44/0xb0 [<ffffffff81243b0e>] SyS_truncate+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task mkdir:19083 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mkdir D ffff880292f1ca80 11824 19083 16395 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff81256ee5>] filename_create+0x85/0x180 [<ffffffff8121f126>] ? kmem_cache_alloc+0x126/0x2e0 [<ffffffff81257c14>] ? getname_flags+0xc4/0x1a0 [<ffffffff81257ec1>] user_path_create+0x41/0x60 [<ffffffff81259206>] SyS_mkdirat+0x46/0xe0 [<ffffffff812592b9>] SyS_mkdir+0x19/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task file_exec.sh:19090 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. file_exec.sh D ffff88029009c240 12832 19090 17373 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff812550a6>] do_last+0x296/0x1280 [<ffffffff812546de>] ? link_path_walk+0x27e/0x8c0 [<ffffffff812570ad>] path_openat+0xcd/0x5b0 [<ffffffff810c18ca>] ? prepare_creds+0x3a/0x200 [<ffffffff8121daba>] ? ___slab_alloc+0x17a/0x560 [<ffffffff81258dfd>] do_filp_open+0x4d/0xb0 [<ffffffff810c18ca>] ? prepare_creds+0x3a/0x200 [<ffffffff810c1592>] ? creds_are_invalid.part.1+0x12/0x50 [<ffffffff810c1592>] ? creds_are_invalid.part.1+0x12/0x50 [<ffffffff810c1a20>] ? prepare_creds+0x190/0x200 [<ffffffff8124db04>] do_open_exec+0x24/0x110 [<ffffffff8124e171>] do_execve_common.isra.22+0x221/0x850 [<ffffffff81257c00>] ? getname_flags+0xb0/0x1a0 [<ffffffff8124ea49>] SyS_execve+0x29/0x30 [<ffffffff817ee5a8>] stub_execve+0x48/0x80 INFO: task rm:19108 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. Lustre: lustre-OST0001: haven't heard from client 0bf25ca2-fb17-4410-82fe-9fe450342b30 (at 0@lo) in 48 seconds. I think it's dead, and I am evicting it. exp ffff88024d5637e8, cur 1630891138 expire 1630891108 last 1630891090 Lustre: Skipped 1 previous similar message rm D ffff88029516c240 12760 19108 16009 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff8125828b>] do_unlinkat+0x13b/0x2b0 [<ffffffff817e8a76>] ? trace_do_page_fault+0x56/0x170 [<ffffffff812592fb>] SyS_unlinkat+0x1b/0x40 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 LustreError: 11119:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0000: '3' migration was interrupted, run 'lfs migrate -m 1 -c 1 -H crush 3' to finish migration. Lustre: 17842:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1630891104/real 1630891104] req@ffff880078760cc0 x1710112928249408/t0(0) o101->lustre-OST0000-osc-ffff880298db12a8@0@lo:28/4 lens 328/400 e 0 to 1 dl 1630891148 ref 2 fl Rpc:XQr/2/ffffffff rc -11/-1 job:'truncate.0' Lustre: 17842:0:(client.c:2285:ptlrpc_expire_one_request()) Skipped 153 previous similar messages Lustre: lustre-OST0000-osc-ffff880298db12a8: Connection to lustre-OST0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: lustre-OST0000: Client 0bf25ca2-fb17-4410-82fe-9fe450342b30 (at 0@lo) reconnecting Lustre: Skipped 4 previous similar messages Lustre: lustre-OST0000-osc-ffff880298db12a8: Connection restored to 192.168.123.130@tcp (at 0@lo) Lustre: Skipped 1 previous similar message Lustre: 4126:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1630891104/real 1630891104] req@ffff8802976dbec0 x1710112929364096/t0(0) o101->lustre-OST0002-osc-ffff880298db12a8@0@lo:28/4 lens 328/400 e 0 to 1 dl 1630891148 ref 1 fl Rpc:XQr/2/ffffffff rc -11/-1 job:'getfattr.0' Lustre: lustre-OST0002: Client 0bf25ca2-fb17-4410-82fe-9fe450342b30 (at 0@lo) reconnecting Lustre: lustre-OST0000-osc-ffff88029ad2a548: Connection restored to 192.168.123.130@tcp (at 0@lo) Lustre: lustre-OST0001: deleting orphan objects from 0x300000401:884 to 0x300000401:929 LustreError: 20691:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0003-osc-ffff88029ad2a548: namespace resource [0x380000400:0x30:0x0].0x0 (ffff88026c7d2ac0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 20691:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 2006 previous similar messages LustreError: 167-0: lustre-OST0003-osc-ffff880298db12a8: This client was evicted by lustre-OST0003; in progress operations using this service will fail. Lustre: 4114:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff880262fb2c00 x1710112934239296/t0(0) o10->lustre-OST0000-osc-ffff88029ad2a548@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630891240 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'cp.0' Lustre: 11141:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 12s req@ffff880257dd25c0 x1710112933969280/t0(0) o37->03033789-bf5a-4e49-a70e-41fbce44720e@0@lo:0/0 lens 448/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'ls.0' LustreError: 16706:0:(mdt_xattr.c:429:mdt_dir_layout_update()) lustre-MDT0001: [0x240000403:0xcd:0x0] migrate mdt index mismatch 1 != 0 Lustre: 16952:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 8: before 511 < left 5711, rollback = 9 Lustre: 16952:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 1 previous similar message LustreError: 14232:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0000: '11' migration was interrupted, run 'lfs migrate -m 2 -c 1 -H crush 11' to finish migration. LustreError: 22406:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff880298db12a8: inode [0x280000403:0x1d:0x0] mdc close failed: rc = -13 LustreError: 19174:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0002: '19' migration was interrupted, run 'lfs migrate -m 1 -c 1 -H crush 19' to finish migration. Lustre: dir [0x240000403:0x90:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: dir [0x240000403:0x90:0x0] stripe 2 readdir failed: -2, directory is partially accessed! Lustre: 4118:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1630891103/real 1630891103] req@ffff880263da70c0 x1710112930103552/t0(0) o101->lustre-OST0000-osc-ffff880298db12a8@0@lo:28/4 lens 328/400 e 0 to 1 dl 1630891162 ref 1 fl Rpc:XNQr/0/ffffffff rc 0/-1 job:'ln.0' Lustre: 4118:0:(client.c:2285:ptlrpc_expire_one_request()) Skipped 219 previous similar messages Lustre: 4114:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff880262fb2c00 x1710112935982784/t0(0) o10->lustre-OST0000-osc-ffff88029ad2a548@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630891263 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'cp.0' Lustre: 21323:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 3/12/9, destroy: 0/0/0 Lustre: 21323:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 37397 previous similar messages Lustre: 21323:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 2/2/0, xattr_set: 11/877/0 Lustre: 21323:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 37397 previous similar messages Lustre: 21323:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 5/136/0, punch: 0/0/0, quota 8/56/0 Lustre: 21323:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 37397 previous similar messages Lustre: 21323:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 12/215/4, delete: 1/1/0 Lustre: 21323:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 37397 previous similar messages Lustre: 21323:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 6/6/0, ref_del: 1/1/0 Lustre: 21323:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 37397 previous similar messages Lustre: 11108:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 510 < left 41548, rollback = 7 Lustre: 11108:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 37387 previous similar messages Lustre: 4118:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1630891103/real 1630891103] req@ffff880285938cc0 x1710112930099840/t0(0) o101->lustre-OST0000-osc-ffff880298db12a8@0@lo:28/4 lens 328/400 e 0 to 1 dl 1630891245 ref 1 fl Rpc:XNQr/0/ffffffff rc 0/-1 job:'ln.0' Lustre: 4118:0:(client.c:2285:ptlrpc_expire_one_request()) Skipped 4 previous similar messages INFO: task rm:19009 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. rm D ffff88025fb624f0 11648 19009 16036 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff8125828b>] do_unlinkat+0x13b/0x2b0 [<ffffffff817df62e>] ? mutex_unlock+0xe/0x10 [<ffffffff81293556>] ? dnotify_flush+0x46/0x110 [<ffffffff812592fb>] SyS_unlinkat+0x1b/0x40 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 LustreError: 11145:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0001: failed to get lu_attr of [0x240000404:0x40:0x0]: rc = -2 LustreError: 16667:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0001: '5' migration was interrupted, run 'lfs migrate -m 0 -c 1 -H crush 5' to finish migration. LustreError: 10978:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0003_UUID lock: ffff88027abfb4c0/0xbfe3d35722a44156 lrc: 3/0,0 mode: PW/PW res: [0x380000400:0xa1:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000000030020 nid: 0@lo remote: 0xbfe3d35722a44148 expref: 1664 pid: 17778 timeout: 634 lvb_type: 0 LustreError: 12663:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88025cadc500 x1710112938763968/t0(0) o104->lustre-OST0003@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 10965:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630891292 with bad export cookie 13827127652154235151 LustreError: 11-0: lustre-OST0003-osc-ffff880298db12a8: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 12 previous similar messages Lustre: lustre-OST0003-osc-ffff880298db12a8: Connection to lustre-OST0003 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 7 previous similar messages LustreError: 167-0: lustre-OST0003-osc-ffff880298db12a8: This client was evicted by lustre-OST0003; in progress operations using this service will fail. LustreError: Skipped 2 previous similar messages Lustre: lustre-OST0003-osc-ffff880298db12a8: Connection restored to 192.168.123.130@tcp (at 0@lo) Lustre: Skipped 9 previous similar messages Lustre: lustre-OST0003-osc-ffff880298db12a8: disconnect after 20s idle LustreError: 10978:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0003_UUID lock: ffff880255499a80/0xbfe3d35722a58ccc lrc: 3/0,0 mode: PW/PW res: [0x380000400:0x8fe:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000400010020 nid: 0@lo remote: 0xbfe3d35722a58c71 expref: 5206 pid: 19474 timeout: 669 lvb_type: 0 LustreError: 19549:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88025a764b40 x1710112939829184/t0(0) o106->lustre-OST0003@0@lo:15/16 lens 328/280 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 19549:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 6 previous similar messages LustreError: 167-0: lustre-OST0003-osc-ffff88029ad2a548: This client was evicted by lustre-OST0003; in progress operations using this service will fail. Lustre: 4113:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.130@tcp:/lustre/fid: [0x240000403:0x17b:0x0]// may get corrupted (rc -108) Lustre: 4114:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.130@tcp:/lustre/fid: [0x280000403:0x190:0x0]// may get corrupted (rc -108) Lustre: 19286:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 503 < left 699, rollback = 2 Lustre: 19286:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 81 previous similar messages LustreError: 24815:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0003-osc-ffff88029ad2a548: namespace resource [0x380000400:0xc9b:0x0].0x0 (ffff880275f7d1c0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 24815:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 3018 previous similar messages Lustre: lustre-OST0003-osc-ffff88029ad2a548: Connection restored to 192.168.123.130@tcp (at 0@lo) Lustre: 4116:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff880254e9a5c0 x1710112940903040/t0(0) o10->lustre-OST0003-osc-ffff880298db12a8@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630891407 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dd.0' LustreError: 16692:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0001: '5' migration was interrupted, run 'lfs migrate -m 0 -c 1 -H crush 5' to finish migration. Lustre: 26422:0:(service.c:1378:ptlrpc_at_send_early_reply()) @@@ Already past deadline (-1s), not sending early reply. Consider increasing at_early_margin (5)? req@ffff880297b91300 x1710112943369024/t4294984811(0) o10->0bf25ca2-fb17-4410-82fe-9fe450342b30@0@lo:76/0 lens 440/432 e 0 to 0 dl 1630891431 ref 1 fl Complete:/0/0 rc 0/0 job:'dir_create.sh.0' Lustre: 26422:0:(service.c:1378:ptlrpc_at_send_early_reply()) Skipped 176 previous similar messages LustreError: 10978:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0003_UUID lock: ffff8800822cbc40/0xbfe3d35722adbe79 lrc: 3/0,0 mode: PW/PW res: [0x380000401:0x630:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000000020020 nid: 0@lo remote: 0xbfe3d35722adbe4f expref: 1442 pid: 24490 timeout: 804 lvb_type: 0 LustreError: 14919:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88025a684500 x1710112945278464/t0(0) o105->lustre-OST0003@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 14919:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 58 previous similar messages LustreError: 11-0: lustre-OST0003-osc-ffff88029ad2a548: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 1 previous similar message Lustre: lustre-OST0003-osc-ffff88029ad2a548: Connection to lustre-OST0003 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 1 previous similar message LustreError: 167-0: lustre-OST0003-osc-ffff88029ad2a548: This client was evicted by lustre-OST0003; in progress operations using this service will fail. Lustre: 4121:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff88007873ea80 x1710112945368960/t0(0) o10->lustre-OST0000-osc-ffff880298db12a8@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630891497 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dd.0' LustreError: 27076:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0003-osc-ffff88029ad2a548: namespace resource [0x380000400:0x15fa:0x0].0x0 (ffff880260d9d440) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 27076:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 105 previous similar messages Lustre: lustre-OST0003-osc-ffff88029ad2a548: Connection restored to 192.168.123.130@tcp (at 0@lo) ptlrpc_watchdog_fire: 28 callbacks suppressed Lustre: ll_ost05_004: service thread pid 17755 was inactive for 62.055 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 11 previous similar messages Lustre: ll_ost05_038: service thread pid 24474 was inactive for 62.044 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Lustre: Skipped 2 previous similar messages Pid: 24474, comm: ll_ost05_038 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] tgt_extent_lock+0xea/0x2a0 [ptlrpc] [<0>] ofd_getattr_hdl+0x365/0x720 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 24603, comm: ll_ost05_055 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] tgt_extent_lock+0xea/0x2a0 [ptlrpc] [<0>] ofd_getattr_hdl+0x365/0x720 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 18398, comm: ll_ost05_011 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] tgt_extent_lock+0xea/0x2a0 [ptlrpc] [<0>] ofd_getattr_hdl+0x365/0x720 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: 14232:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 3/12/7, destroy: 0/0/0 Lustre: 14232:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 131683 previous similar messages Lustre: 14232:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 15/1383/0 Lustre: 14232:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 131683 previous similar messages Lustre: 14232:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 5/136/0, punch: 0/0/0, quota 8/56/0 Lustre: 14232:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 131683 previous similar messages Lustre: 14232:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 10/203/2, delete: 0/0/0 Lustre: 14232:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 131683 previous similar messages Lustre: 14232:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 4/4/0, ref_del: 0/0/0 Lustre: 14232:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 131683 previous similar messages Lustre: 16897:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 510 < left 80721, rollback = 7 Lustre: 16897:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 146576 previous similar messages Lustre: 4120:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff88029a09d180 x1710112947480768/t0(0) o10->lustre-OST0000-osc-ffff880298db12a8@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630891546 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dd.0' Lustre: 4120:0:(client.c:1485:after_reply()) Skipped 131 previous similar messages LustreError: 12293:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0001-osp-MDT0002: fail to cancel 1 llog-records: rc = -116 LustreError: 12293:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0001-osp-MDT0002: fail to cancel 1 of 1 llog-records: rc = -116 LustreError: 10978:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0001_UUID lock: ffff88025d7a5680/0xbfe3d35722afdf06 lrc: 3/0,0 mode: PW/PW res: [0x300000400:0x161c:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->393215) gid 0 flags: 0x60000400000020 nid: 0@lo remote: 0xbfe3d35722afdef1 expref: 3068 pid: 19703 timeout: 863 lvb_type: 0 LustreError: 10967:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630891519 with bad export cookie 13827127652154240093 LustreError: 10967:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) Skipped 1 previous similar message Lustre: 4124:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.130@tcp:/lustre/fid: [0x280000404:0x153:0x0]// may get corrupted (rc -108) Lustre: 4124:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.130@tcp:/lustre/fid: [0x240000404:0xe8:0x0]// may get corrupted (rc -108) Lustre: 4121:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff88007873ea80 x1710112947776000/t0(0) o10->lustre-OST0000-osc-ffff880298db12a8@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630891588 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dd.0' Lustre: 4121:0:(client.c:1485:after_reply()) Skipped 45 previous similar messages LustreError: 25320:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802553a8680 x1710112947977920/t0(0) o104->lustre-OST0001@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 25320:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 20 previous similar messages LustreError: 1301:0:(ofd_obd.c:501:ofd_destroy_export()) lustre-OST0001: cli 0bf25ca2-fb17-4410-82fe-9fe450342b30/ffff8802f3898958 has 1269760 pending on destroyed export LustreError: 27388:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 203407424 != fo_tot_granted 204677184 LustreError: 27388:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 1269760 LustreError: 1301:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_destroy_export: tot_granted 203404736 != fo_tot_granted 204674496 LustreError: 1301:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_destroy_export: tot_pending 0 != fo_tot_pending 1269760 LustreError: 12697:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 211816640 != fo_tot_granted 213086400 LustreError: 12697:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 896 != fo_tot_pending 1270656 LustreError: 12697:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 211814848 != fo_tot_granted 213084608 LustreError: 12697:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 896 != fo_tot_pending 1270656 LustreError: 12708:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 235987392 != fo_tot_granted 237257152 LustreError: 12708:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 2 previous similar messages LustreError: 12708:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 5120000 != fo_tot_pending 6389760 LustreError: 12708:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 2 previous similar messages LustreError: 12710:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 230864704 != fo_tot_granted 232134464 LustreError: 12710:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 2 previous similar messages LustreError: 12710:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 1269760 LustreError: 12710:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 2 previous similar messages Lustre: 4121:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.130@tcp:/lustre/fid: [0x240000404:0x66:0x0]/ may get corrupted (rc -108) Lustre: lustre-OST0001: haven't heard from client 0bf25ca2-fb17-4410-82fe-9fe450342b30 (at 0@lo) in 47 seconds. I think it's dead, and I am evicting it. exp ffff88024ade8008, cur 1630891566 expire 1630891536 last 1630891519 LustreError: 12710:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 230865728 != fo_tot_granted 232135488 LustreError: 11057:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_obd_disconnect: tot_pending 0 != fo_tot_pending 1269760 LustreError: 11057:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 6 previous similar messages LustreError: 12710:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 7 previous similar messages Lustre: 4124:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff88026c3f25c0 x1710112950350528/t0(0) o10->lustre-OST0003-osc-ffff88029ad2a548@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630891595 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dir_create.sh.0' Lustre: 4124:0:(client.c:1485:after_reply()) Skipped 71 previous similar messages Lustre: lustre-OST0003: haven't heard from client 0bf25ca2-fb17-4410-82fe-9fe450342b30 (at 0@lo) in 47 seconds. I think it's dead, and I am evicting it. exp ffff880277fe9bf8, cur 1630891590 expire 1630891560 last 1630891543 Lustre: 17390:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 6s req@ffff880258662c00 x1710112950832384/t0(0) o1000->lustre-MDT0001-mdtlov_UUID@0@lo:0/0 lens 304/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'mdt03_002.0' LustreError: 12707:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 239756736 != fo_tot_granted 241026496 LustreError: 12707:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 18 previous similar messages LustreError: 12707:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 1269760 LustreError: 12707:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 19 previous similar messages Lustre: dir [0x200000403:0xbf:0x0] stripe 1 readdir failed: -2, directory is partially accessed! LustreError: 18488:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0001: '15' migration was interrupted, run 'lfs migrate -m 2 -c 1 -H crush 15' to finish migration. Lustre: 4121:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff88007873ea80 x1710112951886912/t0(0) o10->lustre-OST0000-osc-ffff880298db12a8@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630891691 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dd.0' Lustre: 4121:0:(client.c:1485:after_reply()) Skipped 15 previous similar messages Lustre: 16694:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 498 < left 2646, rollback = 2 Lustre: 16694:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 101 previous similar messages LustreError: 10978:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 120s: evicting client at 0@lo ns: filter-lustre-OST0002_UUID lock: ffff880257fbda40/0xbfe3d35722af87f6 lrc: 3/0,0 mode: PW/PW res: [0x340000401:0x64e:0x0].0x0 rrc: 5 type: EXT [0->18446744073709551615] (req 0->1572863) gid 0 flags: 0x60000080020020 nid: 0@lo remote: 0xbfe3d35722af87ef expref: 3857 pid: 18297 timeout: 987 lvb_type: 0 LustreError: 10978:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 1 previous similar message LustreError: 10957:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630891643 with bad export cookie 13827127652154084938 LustreError: 11-0: lustre-OST0002-osc-ffff88029ad2a548: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 2 previous similar messages Lustre: lustre-OST0002-osc-ffff88029ad2a548: Connection to lustre-OST0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 2 previous similar messages LustreError: 167-0: lustre-OST0002-osc-ffff88029ad2a548: This client was evicted by lustre-OST0002; in progress operations using this service will fail. LustreError: Skipped 2 previous similar messages Lustre: 4118:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.130@tcp:/lustre/fid: [0x280000403:0x2d1:0x0]/ may get corrupted (rc -108) Lustre: 4119:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.130@tcp:/lustre/fid: [0x280000403:0x304:0x0]// may get corrupted (rc -108) Lustre: 4119:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.130@tcp:/lustre/fid: [0x200000404:0xe2:0x0]// may get corrupted (rc -108) Lustre: 16897:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 6s req@ffff8800ab324ad8 x1710112953235328/t0(0) o35->03033789-bf5a-4e49-a70e-41fbce44720e@0@lo:0/0 lens 392/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'ls.0' Lustre: 12709:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 6s req@ffff88025a5ff700 x1710112953387328/t0(0) o13->lustre-MDT0002-mdtlov_UUID@0@lo:0/0 lens 224/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'osp-pre-2-2.0' LustreError: 19692:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88026e0c2c00 x1710112953751104/t0(0) o106->lustre-OST0002@0@lo:15/16 lens 328/280 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 19692:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 186 previous similar messages Lustre: lustre-OST0002-osc-ffff88029ad2a548: Connection restored to 192.168.123.130@tcp (at 0@lo) LustreError: 12695:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 239759808 != fo_tot_granted 241029568 LustreError: 12695:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 33 previous similar messages LustreError: 12695:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 1269760 LustreError: 12695:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 33 previous similar messages Lustre: 4121:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff880078739f80 x1710112954727104/t0(0) o10->lustre-OST0000-osc-ffff880298db12a8@0@lo:6/4 lens 440/432 e 1 to 0 dl 1630891707 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dd.0' Lustre: 4121:0:(client.c:1485:after_reply()) Skipped 102 previous similar messages LustreError: 15473:0:(ldlm_request.c:123:ldlm_expired_completion_wait()) ### lock timed out (enqueued at 1630891419, 300s ago); not entering recovery in server code, just going back to sleep ns: filter-lustre-OST0000_UUID lock: ffff880265f1ed00/0xbfe3d35722afe6d1 lrc: 3/0,1 mode: --/PW res: [0x2c0000400:0x179e:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x40010000000000 nid: local remote: 0x0 expref: -99 pid: 15473 timeout: 0 lvb_type: 0 LustreError: 24477:0:(ldlm_request.c:123:ldlm_expired_completion_wait()) ### lock timed out (enqueued at 1630891419, 300s ago); not entering recovery in server code, just going back to sleep ns: filter-lustre-OST0000_UUID lock: ffff880252ab5a40/0xbfe3d35722afe97f lrc: 3/0,1 mode: --/PW res: [0x2c0000400:0x17a3:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x40010000000000 nid: local remote: 0x0 expref: -99 pid: 24477 timeout: 0 lvb_type: 0 LustreError: 24477:0:(ldlm_request.c:123:ldlm_expired_completion_wait()) Skipped 3 previous similar messages LustreError: dumping log to /tmp/lustre-log.1630891720.15473 LustreError: 24526:0:(ldlm_request.c:123:ldlm_expired_completion_wait()) ### lock timed out (enqueued at 1630891420, 300s ago); not entering recovery in server code, just going back to sleep ns: filter-lustre-OST0000_UUID lock: ffff88027b18f480/0xbfe3d35722afe9a9 lrc: 3/0,1 mode: --/PW res: [0x2c0000400:0x17a9:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x40010000000000 nid: local remote: 0x0 expref: -99 pid: 24526 timeout: 0 lvb_type: 0 LustreError: 24526:0:(ldlm_request.c:123:ldlm_expired_completion_wait()) Skipped 16 previous similar messages LustreError: 19604:0:(ldlm_request.c:123:ldlm_expired_completion_wait()) ### lock timed out (enqueued at 1630891419, 304s ago); not entering recovery in server code, just going back to sleep ns: filter-lustre-OST0000_UUID lock: ffff88026c82a200/0xbfe3d35722afe6c3 lrc: 3/0,1 mode: --/PW res: [0x2c0000400:0x179d:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x40010000000000 nid: local remote: 0x0 expref: -99 pid: 19604 timeout: 0 lvb_type: 0 LustreError: 10978:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0002_UUID lock: ffff88007a11d680/0xbfe3d35722be3be9 lrc: 3/0,0 mode: PW/PW res: [0x340000400:0x1a4d:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000000020020 nid: 0@lo remote: 0xbfe3d35722be3bdb expref: 2217 pid: 18042 timeout: 1131 lvb_type: 0 LustreError: 11138:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000403:0xbf:0x0]: rc = -2 LustreError: 12710:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 245590592 != fo_tot_granted 246860352 LustreError: 12710:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 36 previous similar messages LustreError: 12710:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 1269760 LustreError: 12710:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 36 previous similar messages Lustre: 4116:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.130@tcp:/lustre/fid: [0x200000403:0x7e:0x0]/ may get corrupted (rc -108) Lustre: 4117:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.130@tcp:/lustre/fid: [0x240000403:0x2a0:0x0]// may get corrupted (rc -108) Lustre: 4117:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.130@tcp:/lustre/fid: [0x280000403:0x389:0x0]// may get corrupted (rc -108) Lustre: 4116:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.130@tcp:/lustre/fid: [0x200000403:0x177:0x0]/ may get corrupted (rc -108) LustreError: 31975:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0002-osc-ffff88029ad2a548: namespace resource [0x340000400:0x23b0:0x0].0x0 (ffff880272268cc0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 31975:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 419 previous similar messages Lustre: lustre-OST0002-osc-ffff88029ad2a548: Connection restored to 192.168.123.130@tcp (at 0@lo) Lustre: 4119:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff8802774a7700 x1710112961831168/t0(0) o10->lustre-OST0000-osc-ffff88029ad2a548@0@lo:6/4 lens 440/432 e 1 to 0 dl 1630891907 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dd.0' Lustre: 4119:0:(client.c:1485:after_reply()) Skipped 69 previous similar messages LustreError: 16897:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0002: failed to get lu_attr of [0x280000403:0x119:0x0]: rc = -2 Lustre: dir [0x240000403:0x2e6:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 2 previous similar messages ptlrpc_watchdog_fire: 43 callbacks suppressed Lustre: mdt05_000: service thread pid 11110 was inactive for 64.137 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Lustre: Skipped 2 previous similar messages Pid: 11110, comm: mdt05_000 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] mdt_object_local_lock+0x527/0xb90 [mdt] [<0>] mdt_object_lock_internal+0x70/0x390 [mdt] [<0>] mdt_reint_object_lock+0x2c/0x60 [mdt] [<0>] mdt_reint_striped_lock+0x89/0x5b0 [mdt] [<0>] mdt_reint_unlink+0x8dd/0x19b0 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x76c/0xb40 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: dir [0x240000403:0x25f:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 1 previous similar message Lustre: mdt05_007: service thread pid 19174 was inactive for 64.245 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 19174, comm: mdt05_007 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] mdt_object_local_lock+0x527/0xb90 [mdt] [<0>] mdt_object_lock_internal+0x70/0x390 [mdt] [<0>] mdt_reint_object_lock+0x2c/0x60 [mdt] [<0>] mdt_reint_striped_lock+0x89/0x5b0 [mdt] [<0>] mdt_reint_unlink+0x8dd/0x19b0 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x76c/0xb40 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe LustreError: 11433:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0002-osp-MDT0000: fail to cancel 1 llog-records: rc = -2 LustreError: 11433:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) Skipped 1 previous similar message LustreError: 11433:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0002-osp-MDT0000: fail to cancel 1 of 1 llog-records: rc = -2 LustreError: 11433:0:(llog_cat.c:787:llog_cat_cancel_records()) Skipped 1 previous similar message Lustre: mdt02_005: service thread pid 16952 was inactive for 90.121 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Lustre: mdt02_004: service thread pid 16630 was inactive for 90.128 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 45 previous similar messages Pid: 16952, comm: mdt02_005 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] mdt_object_local_lock+0x527/0xb90 [mdt] [<0>] mdt_object_lock_internal+0x70/0x390 [mdt] [<0>] mdt_getattr_name_lock+0xaca/0x2590 [mdt] [<0>] mdt_intent_getattr+0x2d5/0x4b0 [mdt] [<0>] mdt_intent_opc+0x1dd/0xc10 [mdt] [<0>] mdt_intent_policy+0x1a1/0x360 [mdt] [<0>] ldlm_lock_enqueue+0x3c2/0xb40 [ptlrpc] [<0>] ldlm_handle_enqueue0+0xaa6/0x16d0 [ptlrpc] [<0>] tgt_enqueue+0x64/0x240 [ptlrpc] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe LustreError: 11910:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0002-osp-MDT0001: fail to cancel 1 llog-records: rc = -2 LustreError: 11910:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) Skipped 1 previous similar message LustreError: 11910:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0002-osp-MDT0001: fail to cancel 1 of 1 llog-records: rc = -2 LustreError: 11910:0:(llog_cat.c:787:llog_cat_cancel_records()) Skipped 1 previous similar message Lustre: 20785:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000403:0x449:0x0] with magic=0xbd60bd0 Lustre: 11101:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 1083/4332/0, destroy: 0/0/0 Lustre: 11101:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 243566 previous similar messages Lustre: 11101:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/1, xattr_set: 1085/15/0 Lustre: 11101:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 243564 previous similar messages Lustre: 11101:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 5416/46579/0, punch: 0/0/0, quota 14/174/10 Lustre: 11101:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 243565 previous similar messages Lustre: 11101:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 1083/18411/0, delete: 0/0/0 Lustre: 11101:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 243565 previous similar messages Lustre: 11101:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 11101:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 243567 previous similar messages Lustre: 25851:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 505 < left 1202, rollback = 7 Lustre: 25851:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 239958 previous similar messages Lustre: 26936:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000403:0x44b:0x0] with magic=0xbd60bd0 Lustre: 26936:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 1 previous similar message LustreError: 12701:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 283486016 != fo_tot_granted 284755776 LustreError: 12701:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 75 previous similar messages LustreError: 12701:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 1269760 LustreError: 12701:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 75 previous similar messages LustreError: 10978:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0002_UUID lock: ffff880253c152c0/0xbfe3d35722c9734e lrc: 3/0,0 mode: PW/PW res: [0x340000400:0x2337:0x0].0x0 rrc: 4 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000400020020 nid: 0@lo remote: 0xbfe3d35722c97347 expref: 1524 pid: 19693 timeout: 1411 lvb_type: 0 LustreError: 18063:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802612e5e00 x1710112970065408/t0(0) o104->lustre-OST0002@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 18063:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 55 previous similar messages LustreError: 11-0: lustre-OST0002-osc-ffff88029ad2a548: operation ost_setattr to node 0@lo failed: rc = -107 LustreError: Skipped 3 previous similar messages Lustre: lustre-OST0002-osc-ffff88029ad2a548: Connection to lustre-OST0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 3 previous similar messages LustreError: 2736:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff88029ad2a548: inode [0x280000403:0x1d:0x0] mdc close failed: rc = -13 LustreError: 167-0: lustre-OST0001-osc-ffff88029ad2a548: This client was evicted by lustre-OST0001; in progress operations using this service will fail. LustreError: 26991:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630892068 with bad export cookie 13827127652153255963 LustreError: 26991:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) Skipped 2 previous similar messages LustreError: Skipped 3 previous similar messages Lustre: 4123:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.130@tcp:/lustre/fid: [0x280000404:0x20a:0x0]/ may get corrupted (rc -108) Lustre: 4122:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.130@tcp:/lustre/fid: [0x280000403:0x467:0x0]// may get corrupted (rc -108) Lustre: 4122:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.130@tcp:/lustre/fid: [0x240000403:0x362:0x0]/ may get corrupted (rc -108) Lustre: 4123:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.130@tcp:/lustre/fid: [0x240000403:0x351:0x0]/ may get corrupted (rc -108) LustreError: 24678:0:(ofd_obd.c:501:ofd_destroy_export()) lustre-OST0002: cli 03033789-bf5a-4e49-a70e-41fbce44720e/ffff88007ca9c138 has 811008 pending on destroyed export LustreError: 17019:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630892091 with bad export cookie 13827127652156651355 LustreError: 2801:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0002-osc-ffff88029ad2a548: namespace resource [0x340000400:0x2340:0x0].0x0 (ffff88027720dbc0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 2801:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 995 previous similar messages Lustre: 4121:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.130@tcp:/lustre/fid: [0x280000403:0x456:0x0]/ may get corrupted (rc -108) Lustre: 4117:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.130@tcp:/lustre/fid: [0x280000403:0x3e8:0x0]// may get corrupted (rc -108) Lustre: 4120:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.130@tcp:/lustre/fid: [0x240000403:0x324:0x0]/ may get corrupted (rc -108) Lustre: 4121:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.130@tcp:/lustre/fid: [0x240000403:0x367:0x0]// may get corrupted (rc -108) Lustre: lustre-OST0001-osc-ffff88029ad2a548: Connection restored to 192.168.123.130@tcp (at 0@lo) Lustre: Skipped 4 previous similar messages Lustre: 4119:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff88027f13ac00 x1710112971279040/t0(0) o10->lustre-OST0000-osc-ffff880298db12a8@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630892179 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dir_create.sh.0' Lustre: 4119:0:(client.c:1485:after_reply()) Skipped 2 previous similar messages Lustre: 22996:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000403:0x475:0x0] with magic=0xbd60bd0 Lustre: 22996:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 1 previous similar message Lustre: 16952:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000403:0x445:0x0] with magic=0xbd60bd0 Lustre: 16952:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 1 previous similar message Lustre: 20785:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000403:0x4d1:0x0] with magic=0xbd60bd0 Lustre: 20785:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 1 previous similar message Lustre: 11111:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 499 < left 1339, rollback = 2 Lustre: 11111:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 193 previous similar messages Lustre: 20785:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000403:0x356:0x0] with magic=0xbd60bd0 Lustre: 20785:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 3 previous similar messages Lustre: 21323:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000403:0x35e:0x0] with magic=0xbd60bd0 Lustre: 21323:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 1 previous similar message Lustre: 19322:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000403:0x3ab:0x0] with magic=0xbd60bd0 Lustre: 19322:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 9 previous similar messages Lustre: 4114:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.130@tcp:/lustre/fid: [0x280000403:0x3e8:0x0]/ may get corrupted (rc -108) Lustre: 4113:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.130@tcp:/lustre/fid: [0x200000403:0x255:0x0]// may get corrupted (rc -108) Lustre: 4117:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.130@tcp:/lustre/fid: [0x200000403:0x279:0x0]/ may get corrupted (rc -108) Lustre: 4117:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.130@tcp:/lustre/fid: [0x280000403:0x54c:0x0]/ may get corrupted (rc -108) Lustre: 4116:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.130@tcp:/lustre/fid: [0x240000404:0x314:0x0]// may get corrupted (rc -108) Lustre: 4113:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.130@tcp:/lustre/fid: [0x240000403:0x42e:0x0]// may get corrupted (rc -108) ptlrpc_watchdog_fire: 15 callbacks suppressed Lustre: ll_ost00_046: service thread pid 25547 was inactive for 40.018 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 25547, comm: ll_ost00_046 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] tgt_extent_lock+0xea/0x2a0 [ptlrpc] [<0>] ofd_getattr_hdl+0x365/0x720 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 23304, comm: ll_ost00_017 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Lustre: ll_ost00_007: service thread pid 17814 was inactive for 40.114 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 14 previous similar messages Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] tgt_extent_lock+0xea/0x2a0 [ptlrpc] [<0>] ofd_getattr_hdl+0x365/0x720 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 25521, comm: ll_ost00_030 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] tgt_extent_lock+0xea/0x2a0 [ptlrpc] [<0>] ofd_getattr_hdl+0x365/0x720 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe LustreError: 21296:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0001: '14' migration was interrupted, run 'lfs migrate -m 0 -c 1 -H crush 14' to finish migration. LustreError: 5431:0:(file.c:5042:ll_inode_revalidate_fini()) lustre: revalidate FID [0x280000404:0x3a9:0x0] error: rc = -5 LustreError: 3670:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff88029ad2a548: inode [0x280000403:0x4df:0x0] mdc close failed: rc = -5 LustreError: 5431:0:(file.c:5042:ll_inode_revalidate_fini()) Skipped 12 previous similar messages LustreError: 5455:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -108 LustreError: 2683:0:(mdc_request.c:1436:mdc_read_page()) lustre-MDT0002-mdc-ffff88029ad2a548: [0x280000400:0x9:0x0] lock enqueue fails: rc = -108 Lustre: dir [0x280000404:0x13d:0x0] stripe 0 readdir failed: -108, directory is partially accessed! Lustre: Skipped 1 previous similar message Lustre: ll_ost00_040: service thread pid 25541 was inactive for 46.068 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 41 previous similar messages Lustre: lustre-OST0001: haven't heard from client 03033789-bf5a-4e49-a70e-41fbce44720e (at 0@lo) in 52 seconds. I think it's dead, and I am evicting it. exp ffff880257a35d28, cur 1630892347 expire 1630892317 last 1630892295 Lustre: 4129:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.130@tcp:/lustre/fid: [0x240000403:0x417:0x0]// may get corrupted (rc -108) Lustre: 4128:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.130@tcp:/lustre/fid: [0x280000403:0x4fe:0x0]// may get corrupted (rc -108) Lustre: 4128:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.130@tcp:/lustre/fid: [0x240000404:0x26e:0x0]// may get corrupted (rc -108) Lustre: 4117:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.130@tcp:/lustre/fid: [0x280000403:0x3da:0x0]/ may get corrupted (rc -108) LustreError: 4139:0:(vvp_io.c:1793:vvp_io_init()) lustre: refresh file layout [0x280000403:0x4fe:0x0] error -108. Lustre: 4128:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.130@tcp:/lustre/fid: [0x200000403:0x24c:0x0]/ may get corrupted (rc -108) Lustre: 4128:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.130@tcp:/lustre/fid: [0x240000403:0x3f1:0x0]/ may get corrupted (rc -108) Lustre: 4128:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.130@tcp:/lustre/fid: [0x280000403:0x55b:0x0]/ may get corrupted (rc -108) Lustre: 4128:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.130@tcp:/lustre/fid: [0x240000403:0x317:0x0]/ may get corrupted (rc -108) Lustre: 4128:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.130@tcp:/lustre/fid: [0x280000404:0x2ed:0x0]/ may get corrupted (rc -108) Lustre: 4128:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.130@tcp:/lustre/fid: [0x200000404:0x115:0x0]// may get corrupted (rc -108) Lustre: 4128:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.130@tcp:/lustre/fid: [0x200000404:0x11d:0x0]/ may get corrupted (rc -108) Lustre: 4128:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.130@tcp:/lustre/fid: [0x280000404:0x38d:0x0]/ may get corrupted (rc -108) Lustre: lustre-OST0002-osc-ffff88029ad2a548: Connection restored to 192.168.123.130@tcp (at 0@lo) Lustre: Skipped 5 previous similar messages LustreError: 30634:0:(service.c:2290:ptlrpc_server_handle_request()) @@@ Dropping timed-out request from 12345-0@lo: deadline 25/98s ago req@ffff88029bb070c0 x1710112976388544/t0(0) o1->03033789-bf5a-4e49-a70e-41fbce44720e@0@lo:196/0 lens 440/0 e 0 to 0 dl 1630892306 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'lfs.0' Lustre: 25522:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (25/98s); client may timeout req@ffff8802588557c0 x1710112976388864/t0(0) o1->03033789-bf5a-4e49-a70e-41fbce44720e@0@lo:196/0 lens 440/0 e 0 to 0 dl 1630892306 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'lfs.0' Lustre: 25522:0:(service.c:2328:ptlrpc_server_handle_request()) Skipped 6 previous similar messages LustreError: 30634:0:(service.c:2290:ptlrpc_server_handle_request()) Skipped 1888 previous similar messages LustreError: 1301:0:(ofd_obd.c:501:ofd_destroy_export()) lustre-OST0000: cli 03033789-bf5a-4e49-a70e-41fbce44720e/ffff880298c39bf8 has 122880 pending on destroyed export Lustre: 4119:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.130@tcp:/lustre/fid: [0x280000404:0x3aa:0x0]// may get corrupted (rc -108) Lustre: 4119:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.130@tcp:/lustre/fid: [0x280000404:0x3a8:0x0]/ may get corrupted (rc -108) LustreError: 11097:0:(ldlm_lockd.c:1425:ldlm_handle_enqueue0()) ### lock on destroyed export ffff88029966ca88 ns: mdt-lustre-MDT0002_UUID lock: ffff880266ee61c0/0xbfe3d35722d822e8 lrc: 3/0,0 mode: PR/PR res: [0x280000404:0x3a9:0x0].0x0 bits 0x13/0x0 rrc: 25 type: IBT gid 0 flags: 0x50200000000000 nid: 0@lo remote: 0xbfe3d35722d822cc expref: 16 pid: 11097 timeout: 0 lvb_type: 0 Lustre: 16648:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (88/28s); client may timeout req@ffff8802751dc500 x1710112976853248/t0(0) o101->03033789-bf5a-4e49-a70e-41fbce44720e@0@lo:269/0 lens 576/640 e 0 to 0 dl 1630892379 ref 1 fl Complete:/0/0 rc -107/-107 job:'ls.0' Lustre: 16648:0:(service.c:2328:ptlrpc_server_handle_request()) Skipped 13 previous similar messages LustreError: 11119:0:(ldlm_lockd.c:1425:ldlm_handle_enqueue0()) ### lock on destroyed export ffff88029966ca88 ns: mdt-lustre-MDT0002_UUID lock: ffff880286197480/0xbfe3d35722d7ffe1 lrc: 3/0,0 mode: PR/PR res: [0x280000404:0x3a9:0x0].0x0 bits 0x13/0x0 rrc: 9 type: IBT gid 0 flags: 0x50200000000000 nid: 0@lo remote: 0xbfe3d35722d7ff94 expref: 5 pid: 11119 timeout: 0 lvb_type: 0 LustreError: 11119:0:(ldlm_lockd.c:1425:ldlm_handle_enqueue0()) Skipped 4 previous similar messages Lustre: lustre-OST0003-osc-ffff88029ad2a548: disconnect after 36s idle LustreError: 11095:0:(ldlm_lockd.c:1425:ldlm_handle_enqueue0()) ### lock on destroyed export ffff88029966ca88 ns: mdt-lustre-MDT0002_UUID lock: ffff88026221cf00/0xbfe3d35722d828fa lrc: 3/0,0 mode: PR/PR res: [0x280000404:0x3a9:0x0].0x0 bits 0x13/0x0 rrc: 5 type: IBT gid 0 flags: 0x50200000000000 nid: 0@lo remote: 0xbfe3d35722d82883 expref: 2 pid: 11095 timeout: 0 lvb_type: 0 ------------[ cut here ]------------ WARNING: CPU: 11 PID: 11111 at /home/green/git/lustre-release/ldiskfs/ext4_jbd2.c:266 __ldiskfs_handle_dirty_metadata+0x1d2/0x230 [ldiskfs] Modules linked in: loop zfs(PO) zunicode(PO) zzstd(O) zlua(O) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) jbd2 mbcache lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) dm_flakey dm_mod libcfs(OE) crc_t10dif crct10dif_generic sb_edac edac_core iosf_mbi crc32_pclmul ghash_clmulni_intel aesni_intel lrw gf128mul glue_helper ablk_helper cryptd pcspkr virtio_balloon virtio_console i2c_piix4 ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm drm crct10dif_pclmul ata_piix crct10dif_common drm_panel_orientation_quirks crc32c_intel serio_raw virtio_blk i2c_core libata floppy CPU: 11 PID: 11111 Comm: mdt05_001 Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011 Call Trace: [<ffffffff817d9418>] dump_stack+0x19/0x1b [<ffffffff8108d538>] __warn+0xd8/0x100 [<ffffffff8108d67d>] warn_slowpath_null+0x1d/0x20 [<ffffffffa0c2f262>] __ldiskfs_handle_dirty_metadata+0x1d2/0x230 [ldiskfs] [<ffffffffa0c0f33d>] ldiskfs_getblk+0x13d/0x210 [ldiskfs] [<ffffffffa0c0f437>] ldiskfs_bread+0x27/0xe0 [ldiskfs] [<ffffffffa0cd3e97>] osd_write+0x4e7/0xcb0 [osd_ldiskfs] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffffa04ca6f3>] dt_record_write+0x33/0x120 [obdclass] [<ffffffffa04896b2>] llog_osd_write_rec+0x1052/0x19a0 [obdclass] [<ffffffffa0476760>] llog_write_rec+0x290/0x590 [obdclass] [<ffffffffa047c4ee>] llog_cat_add_rec+0x23e/0x950 [obdclass] [<ffffffffa04735ff>] llog_add+0x17f/0x1f0 [obdclass] [<ffffffffa04a4a39>] ? lprocfs_counter_add+0xf9/0x160 [obdclass] [<ffffffffa080d60d>] sub_updates_write+0xb6b/0xe32 [ptlrpc] [<ffffffffa07dd892>] top_trans_stop+0x7c2/0xf60 [ptlrpc] [<ffffffffa0f0ef3c>] lod_trans_stop+0x25c/0x340 [lod] [<ffffffffa0de3408>] mdd_trans_stop+0x28/0x16e [mdd] [<ffffffffa0dc4b2f>] mdd_migrate_object+0xf1f/0x1ba0 [mdd] [<ffffffffa04c9375>] ? lu_object_find_at+0x285/0xb10 [obdclass] [<ffffffffa0dc5b86>] mdd_migrate+0x3d6/0x8e0 [mdd] [<ffffffffa0e970d8>] mdo_migrate+0x4c/0x4e [mdt] [<ffffffffa0e56e53>] mdt_reint_migrate+0x1023/0x11d0 [mdt] [<ffffffffa0e57087>] mdt_reint_rec+0x87/0x240 [mdt] [<ffffffffa0e2c55c>] mdt_reint_internal+0x76c/0xb40 [mdt] [<ffffffffa0e39657>] mdt_reint+0x67/0x150 [mdt] [<ffffffffa07cc26e>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<ffffffffa032345e>] ? libcfs_nid2str_r+0xfe/0x130 [lnet] [<ffffffffa0777990>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<ffffffffa0779529>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffffa0778950>] ? ptlrpc_wait_event+0x620/0x620 [ptlrpc] [<ffffffff810ba0f4>] kthread+0xe4/0xf0 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 ---[ end trace 58aae1581d9a010f ]--- LDISKFS-fs: ldiskfs_getblk:888: aborting transaction: error 28 in __ldiskfs_handle_dirty_metadata LDISKFS-fs error (device dm-2): ldiskfs_getblk:888: inode #231: block 51071: comm mdt05_001: journal_dirty_metadata failed: handle type 0 started at line 1972, credits 512/0, errcode -28 Aborting journal on device dm-2-8. LDISKFS-fs (dm-2): Remounting filesystem read-only LustreError: 11111:0:(osd_io.c:2103:osd_ldiskfs_write_record()) lustre-MDT0002: error reading offset 278528 (block 68, size 24, offs 278496), credits 0/0: rc = -28 LustreError: 11111:0:(update_trans.c:985:top_trans_stop()) lustre-MDT0002-osd: write updates failed: rc = -28 LustreError: 11111:0:(osd_io.c:2110:osd_ldiskfs_write_record()) journal_get_write_access() returned error -30 LustreError: 11111:0:(tgt_lastrcvd.c:1253:tgt_add_reply_data()) lustre-MDT0002: can't update reply_data file: rc = -30 LustreError: 11111:0:(osd_handler.c:2075:osd_trans_stop()) lustre-MDT0002: failed in transaction hook: rc = -30 LDISKFS-fs error (device dm-2) in osd_trans_stop:2082: error 28 LustreError: 13847:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-OST0002-osc-MDT0002: fail to cancel 24 llog-records: rc = -30 LustreError: 14484:0:(osp_sync.c:1094:osp_sync_process_committed()) lustre-OST0003-osc-MDT0002: can't cancel 26 records: rc = -30 LustreError: 13844:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-OST0001-osc-MDT0002: fail to cancel 1 of 1 llog-records: rc = -30 LustreError: 13844:0:(osp_sync.c:1079:osp_sync_process_committed()) lustre-OST0001-osc-MDT0002: can't cancel record: rc = -30 LustreError: 11111:0:(osd_handler.c:2085:osd_trans_stop()) lustre-MDT0002: failed to stop transaction: rc = -28 LustreError: 11111:0:(update_trans.c:1011:top_trans_stop()) lustre-MDT0002-osd: stop trans failed: rc = -30 LustreError: 12248:0:(osd_handler.c:1779:osd_trans_commit_cb()) transaction @0xffff88024feb4948 commit error: 2 LustreError: 8536:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -30 LDISKFS-fs error (device dm-2) in osd_trans_stop:2082: IO failure LustreError: 13847:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) Skipped 7 previous similar messages LustreError: 13847:0:(osp_sync.c:1079:osp_sync_process_committed()) lustre-OST0002-osc-MDT0002: can't cancel record: rc = -30 LustreError: 13847:0:(osp_sync.c:1079:osp_sync_process_committed()) Skipped 7 previous similar messages LustreError: 13062:0:(osp_sync.c:1094:osp_sync_process_committed()) lustre-OST0000-osc-MDT0002: can't cancel 5 records: rc = -30 LustreError: 13062:0:(osp_sync.c:1094:osp_sync_process_committed()) Skipped 12 previous similar messages | Externally reported by onyx-68 boilpot email |
racer test 1: racer on clients: centos-70.localnet DURATION=2700 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa1016e52>] ll_migrate+0x9b2/0xec0 [lustre] PGD 80000002867d9067 PUD 292cdf067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: loop zfs(PO) zunicode(PO) zzstd(O) zlua(O) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) jbd2 mbcache lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) dm_flakey dm_mod libcfs(OE) crc_t10dif crct10dif_generic sb_edac edac_core iosf_mbi crc32_pclmul ghash_clmulni_intel aesni_intel lrw gf128mul glue_helper ablk_helper cryptd virtio_balloon virtio_console pcspkr i2c_piix4 ip_tables rpcsec_gss_krb5 ata_generic drm_kms_helper pata_acpi ttm crct10dif_pclmul drm crct10dif_common crc32c_intel ata_piix drm_panel_orientation_quirks serio_raw virtio_blk libata i2c_core floppy CPU: 1 PID: 32579 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011 task: ffff880321528010 ti: ffff880260dac000 task.ti: ffff880260dac000 RIP: 0010:[<ffffffffa1016e52>] [<ffffffffa1016e52>] ll_migrate+0x9b2/0xec0 [lustre] RSP: 0018:ffff880260dafbc8 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff8800764623f8 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff8803272ca038 RDI: ffff8803272ca018 RBP: ffff880260dafc48 R08: ffff880249700bb8 R09: ffff880260dafa60 R10: 0000000000000006 R11: 0000000000000007 R12: 0000000000000000 R13: ffff88027c335cc8 R14: ffff880268131cd8 R15: 0000000000000030 FS: 00007fabd0075740(0000) GS:ffff880331a40000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 CR2: 0000000000000008 CR3: 00000002852da000 CR4: 00000000001607e0 Call Trace: [<ffffffff81242213>] ? __check_object_size+0x1c3/0x220 [<ffffffffa1004561>] ll_dir_ioctl+0x5d01/0x6ed0 [lustre] [<ffffffff8103de0f>] ? save_stack_trace+0x1f/0x30 [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817d5822>] ? free_debug_processing+0x240/0x272 [<ffffffffa10326ad>] ? ll_finish_md_op_data+0xad/0x1f0 [lustre] [<ffffffff8121e474>] ? __slab_free+0xa4/0x300 [<ffffffff811becd5>] ? __free_pages+0x25/0x30 [<ffffffffa0ffaf3e>] ? ll_release_page+0x8e/0xa0 [lustre] [<ffffffffa04b3a39>] ? lprocfs_counter_add+0xf9/0x160 [obdclass] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e326e>] ? _raw_spin_unlock+0xe/0x20 [<ffffffff81277ec8>] ? __mark_inode_dirty+0x1a8/0x2c0 [<ffffffff8125b3fd>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff817df62e>] ? mutex_unlock+0xe/0x10 [<ffffffff8125b81a>] ? iterate_dir+0xaa/0x120 [<ffffffff8125b751>] SyS_ioctl+0xa1/0xc0 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 | Lustre: lustre-OST0000-osc-ffff88008dee12a8: disconnect after 21s idle LustreError: 11516:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0002-osp-MDT0000: fail to cancel 1 llog-records: rc = -2 LustreError: 11516:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0002-osp-MDT0000: fail to cancel 1 of 1 llog-records: rc = -2 Lustre: 17175:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 8: before 511 < left 8447, rollback = 9 Lustre: 17175:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 495/1980/0, destroy: 1/4/0 Lustre: 17175:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 5/5/0, xattr_set: 503/283/0 Lustre: 17175:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 2476/21295/0, punch: 0/0/0, quota 6/6/0 Lustre: 17175:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 497/8447/0, delete: 3/6/1 Lustre: 17175:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 1/1/0, ref_del: 1/1/0 Lustre: 17175:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 8: before 510 < left 8447, rollback = 9 Lustre: 17175:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 495/1980/0, destroy: 1/4/0 Lustre: 17175:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 5/5/1, xattr_set: 503/283/0 Lustre: 17175:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 2476/21295/0, punch: 0/0/0, quota 6/6/0 Lustre: 17175:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 497/8447/0, delete: 3/6/1 Lustre: 17175:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 1/1/0, ref_del: 1/1/0 Lustre: 17175:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 508 < left 21295, rollback = 7 Lustre: 17406:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 508 < left 610, rollback = 2 Lustre: 17406:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 3/12/2, destroy: 0/0/0 Lustre: 17406:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 17406:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 2/2/0, xattr_set: 8/610/0 Lustre: 17406:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 17406:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 5/136/0, punch: 0/0/0, quota 6/6/0 Lustre: 17406:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 17406:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 12/215/2, delete: 1/1/0 Lustre: 17406:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 17406:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 6/6/0, ref_del: 1/1/0 Lustre: 17406:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 13880:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 504 < left 671, rollback = 2 Lustre: 13880:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 5 previous similar messages Lustre: 13880:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 3/12/4, destroy: 0/0/0 Lustre: 13880:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 13880:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 7/671/0 Lustre: 13880:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 13880:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 5/136/0, punch: 0/0/0, quota 8/56/0 Lustre: 13880:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 13880:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 10/203/4, delete: 0/0/0 Lustre: 13880:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 13880:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 4/4/0, ref_del: 0/0/0 Lustre: 13880:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 17478:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 511 < left 3020, rollback = 7 Lustre: 16667:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 500 < left 610, rollback = 2 Lustre: 16667:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 7 previous similar messages Lustre: 16667:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 3/12/8, destroy: 0/0/0 Lustre: 16667:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 287 previous similar messages Lustre: 16667:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 2/2/0, xattr_set: 8/610/0 Lustre: 16667:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 287 previous similar messages Lustre: 16667:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 5/136/0, punch: 0/0/0, quota 6/6/0 Lustre: 16667:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 287 previous similar messages Lustre: 16667:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 12/215/4, delete: 1/1/0 Lustre: 16667:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 287 previous similar messages Lustre: 16667:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 6/6/0, ref_del: 1/1/0 Lustre: 16667:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 287 previous similar messages Lustre: 11186:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 499 < left 1606, rollback = 2 Lustre: 11186:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 15 previous similar messages Lustre: 11186:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 2/8/8, destroy: 1/4/0 Lustre: 11186:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 15 previous similar messages Lustre: 11186:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 2/2/0, xattr_set: 19/1606/0 Lustre: 11186:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 15 previous similar messages Lustre: 11186:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1/10/0, punch: 0/0/0, quota 8/104/0 Lustre: 11186:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 15 previous similar messages Lustre: 11186:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 12/194/4, delete: 3/6/1 Lustre: 11186:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 15 previous similar messages Lustre: 11186:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 7/7/0, ref_del: 3/3/0 Lustre: 11186:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 15 previous similar messages Lustre: 11175:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 504 < left 19317, rollback = 7 Lustre: 11175:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 279 previous similar messages Lustre: 19140:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 510 < left 21295, rollback = 7 Lustre: 19140:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 1800 previous similar messages LustreError: 17580:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0001: failed to get lu_attr of [0x240000404:0xc:0x0]: rc = -2 ------------[ cut here ]------------ WARNING: CPU: 7 PID: 19140 at /home/green/git/lustre-release/ldiskfs/namei.c:3331 ldiskfs_orphan_add+0x11e/0x2a0 [ldiskfs] Modules linked in: loop zfs(PO) zunicode(PO) zzstd(O) zlua(O) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) jbd2 mbcache lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) dm_flakey dm_mod libcfs(OE) crc_t10dif crct10dif_generic sb_edac edac_core iosf_mbi crc32_pclmul ghash_clmulni_intel aesni_intel lrw gf128mul glue_helper ablk_helper cryptd virtio_balloon virtio_console pcspkr i2c_piix4 ip_tables rpcsec_gss_krb5 ata_generic drm_kms_helper pata_acpi ttm crct10dif_pclmul drm crct10dif_common crc32c_intel ata_piix drm_panel_orientation_quirks serio_raw virtio_blk libata i2c_core floppy CPU: 7 PID: 19140 Comm: mdt_rdpg03_003 Tainted: P OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011 Call Trace: [<ffffffff817d9418>] dump_stack+0x19/0x1b [<ffffffff8108d538>] __warn+0xd8/0x100 [<ffffffff8108d67d>] warn_slowpath_null+0x1d/0x20 [<ffffffffa0bcd4ae>] ldiskfs_orphan_add+0x11e/0x2a0 [ldiskfs] [<ffffffffa0bbe9eb>] ldiskfs_xattr_inode_orphan_add.constprop.17+0xbb/0x110 [ldiskfs] [<ffffffffa0bc1417>] ldiskfs_xattr_delete_inode+0x77/0x3c0 [ldiskfs] [<ffffffffa0be1aa0>] ? ldiskfs_evict_inode+0x190/0x670 [ldiskfs] [<ffffffffa0be1ad9>] ldiskfs_evict_inode+0x1c9/0x670 [ldiskfs] [<ffffffff812649af>] evict+0xaf/0x180 [<ffffffff81264de5>] iput+0xf5/0x180 [<ffffffffa0c76d08>] osd_object_delete+0x1f8/0x370 [osd_ldiskfs] [<ffffffffa04d359d>] lu_object_free.isra.26+0xcd/0x1d0 [obdclass] [<ffffffffa04d7cf1>] lu_object_put+0xa1/0x480 [obdclass] [<ffffffffa0e0faf4>] mdt_mfd_close+0x1f4/0x870 [mdt] [<ffffffffa0e15e14>] mdt_close_internal+0xb4/0x240 [mdt] [<ffffffffa0e16224>] mdt_close+0x284/0x850 [mdt] [<ffffffffa07da2d4>] ? tgt_request_preprocess.isra.27+0xf4/0x850 [ptlrpc] [<ffffffffa07db26e>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<ffffffffa032945e>] ? libcfs_nid2str_r+0xfe/0x130 [lnet] [<ffffffffa0786990>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<ffffffffa0788529>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e32be>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffffa0787950>] ? ptlrpc_wait_event+0x620/0x620 [ptlrpc] [<ffffffff810ba0f4>] kthread+0xe4/0xf0 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 ---[ end trace 3e032eb585932b8d ]--- Lustre: 11168:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 509 < left 19317, rollback = 7 Lustre: 11168:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 1982 previous similar messages Lustre: dir [0x200000404:0x21:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: 20569:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 20569:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 3828 previous similar messages Lustre: 20569:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 1/89/0 Lustre: 20569:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 3828 previous similar messages Lustre: 20569:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 626/635/0, punch: 0/0/0, quota 3/3/0 Lustre: 20569:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 3828 previous similar messages Lustre: 20569:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 20569:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 3828 previous similar messages Lustre: 20569:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 20569:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 3828 previous similar messages LustreError: 11164:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0000: '4' migration was interrupted, run 'lfs migrate -m 2 -c 1 -H crush 4' to finish migration. Lustre: 16667:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 509 < left 1125, rollback = 7 Lustre: 16667:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 2277 previous similar messages LustreError: 11349:0:(out_handler.c:910:out_tx_end()) lustre-MDT0002-osd: undo for /home/green/git/lustre-release/lustre/ptlrpc/../../lustre/target/out_handler.c:445: rc = -524 LustreError: 11949:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0002-osp-MDT0001: fail to cancel 1 llog-records: rc = -2 LustreError: 11949:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0002-osp-MDT0001: fail to cancel 1 of 1 llog-records: rc = -2 LustreError: 18661:0:(mdt_xattr.c:429:mdt_dir_layout_update()) lustre-MDT0001: [0x240000404:0x1e:0x0] migrate mdt index mismatch 1 != 0 Lustre: 16688:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 504 < left 26928, rollback = 7 Lustre: 16688:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 6320 previous similar messages LustreError: 22103:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff88008dee12a8: inode [0x240000403:0xc1:0x0] mdc close failed: rc = -13 Lustre: 16718:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 8: before 509 < left 10658, rollback = 9 Lustre: 16718:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 43 previous similar messages Lustre: 16718:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 626/2504/0, destroy: 1/4/0 Lustre: 16718:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 11107 previous similar messages Lustre: 16718:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 3/3/1, xattr_set: 630/148/0 Lustre: 16718:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 11107 previous similar messages Lustre: 16718:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 3131/26928/0, punch: 0/0/0, quota 10/74/0 Lustre: 16718:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 11107 previous similar messages Lustre: 16718:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 627/10658/0, delete: 2/5/1 Lustre: 16718:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 11107 previous similar messages Lustre: 16718:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 1/1/0, ref_del: 2/2/1 Lustre: 16718:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 11107 previous similar messages Lustre: ll_ost07_006: service thread pid 20126 was inactive for 40.063 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 20302, comm: ll_ost07_009 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] ofd_destroy_by_fid+0x1d1/0x520 [ofd] [<0>] ofd_destroy_hdl+0x27f/0xa20 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: Skipped 1 previous similar message Pid: 20126, comm: ll_ost07_006 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] ofd_destroy_by_fid+0x1d1/0x520 [ofd] [<0>] ofd_destroy_hdl+0x27f/0xa20 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 20290, comm: ll_ost07_008 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] ofd_destroy_by_fid+0x1d1/0x520 [ofd] [<0>] ofd_destroy_hdl+0x27f/0xa20 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: ll_ost07_002: service thread pid 12722 was inactive for 40.506 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: 16806:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 507 < left 47955, rollback = 7 Lustre: 16806:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 3626 previous similar messages Lustre: 21569:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1630890481/real 1630890481] req@ffff8802848f57c0 x1710112383577664/t0(0) o101->lustre-MDT0002-mdc-ffff880299a4ca88@0@lo:12/10 lens 576/33552 e 0 to 1 dl 1630890527 ref 2 fl Rpc:XQr/2/ffffffff rc 0/-1 job:'ln.0' Lustre: lustre-MDT0002-mdc-ffff880299a4ca88: Connection to lustre-MDT0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: lustre-MDT0002: Client 989be46e-abfb-4210-a299-8eaa2ef57189 (at 0@lo) reconnecting Lustre: lustre-MDT0002-mdc-ffff880299a4ca88: Connection restored to 192.168.123.170@tcp (at 0@lo) Lustre: ll_ost03_001: service thread pid 12705 was inactive for 62.121 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 4 previous similar messages Lustre: ll_ost03_000: service thread pid 12704 was inactive for 62.012 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 1 previous similar message Lustre: 16812:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 501 < left 894, rollback = 2 Lustre: 16812:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 46 previous similar messages Lustre: ll_ost03_016: service thread pid 21607 was inactive for 62.068 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 11 previous similar messages Lustre: dir [0x240000404:0x56:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 6 previous similar messages sched: RT throttling activated Lustre: dir [0x240000403:0x3b:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: dir [0x200000404:0x105:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: 25120:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 473/1892/0, destroy: 1/4/0 Lustre: 25120:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 10812 previous similar messages Lustre: 25120:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 3/3/1, xattr_set: 477/148/0 Lustre: 25120:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 10812 previous similar messages Lustre: 25120:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 2366/20349/0, punch: 0/0/0, quota 10/90/0 Lustre: 25120:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 10812 previous similar messages Lustre: 25120:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 474/8057/0, delete: 2/5/1 Lustre: 25120:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 10812 previous similar messages Lustre: 17206:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 17206:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 10812 previous similar messages Lustre: mdt00_010: service thread pid 19511 was inactive for 62.001 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 9 previous similar messages Lustre: 16679:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 510 < left 1678, rollback = 7 Lustre: 16679:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 11426 previous similar messages LustreError: 11026:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0001_UUID lock: ffff880085b661c0/0x99ae7c71d2c4aef4 lrc: 3/0,0 mode: PW/PW res: [0x300000400:0x25:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->524287) gid 0 flags: 0x60000400030020 nid: 0@lo remote: 0x99ae7c71d2c4acd2 expref: 2005 pid: 18408 timeout: 437 lvb_type: 0 LustreError: 11-0: lustre-OST0001-osc-ffff88008dee12a8: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-OST0001-osc-ffff88008dee12a8: Connection to lustre-OST0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 11-0: lustre-OST0001-osc-ffff88008dee12a8: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: 12712:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88008397f0c0 x1710112388083264/t0(0) o104->lustre-OST0001@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 167-0: lustre-OST0001-osc-ffff88008dee12a8: This client was evicted by lustre-OST0001; in progress operations using this service will fail. LustreError: 17633:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802705a5180 x1710112388122944/t0(0) o104->lustre-OST0001@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 26071:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0001-osc-ffff88008dee12a8: namespace resource [0x100:0x0:0x0].0x0 (ffff880274faa0c0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 17633:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 1 previous similar message Lustre: 11170:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 8: before 509 < left 6391, rollback = 9 Lustre: 11170:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 43 previous similar messages Lustre: dir [0x200000403:0xb9:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: 4139:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff8802594b4500 x1710112388789568/t0(0) o10->lustre-OST0003-osc-ffff880299a4ca88@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630890657 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dd.0' LustreError: 11026:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0001_UUID lock: ffff88008344f480/0x99ae7c71d2c89f17 lrc: 3/0,0 mode: PR/PR res: [0x240000404:0x1e:0x0].0x0 bits 0x12/0x0 rrc: 7 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0x99ae7c71d2c89f09 expref: 124 pid: 11175 timeout: 459 lvb_type: 0 LustreError: 11-0: lustre-MDT0001-mdc-ffff880299a4ca88: operation mds_reint to node 0@lo failed: rc = -107 LustreError: Skipped 2 previous similar messages Lustre: lustre-MDT0001-mdc-ffff880299a4ca88: Connection to lustre-MDT0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-MDT0001-mdc-ffff880299a4ca88: This client was evicted by lustre-MDT0001; in progress operations using this service will fail. LustreError: 19511:0:(ldlm_lockd.c:1425:ldlm_handle_enqueue0()) ### lock on destroyed export ffff880298c95d28 ns: mdt-lustre-MDT0001_UUID lock: ffff880083f4cb40/0x99ae7c71d2c8a14e lrc: 3/0,0 mode: PR/PR res: [0x240000404:0x1e:0x0].0x0 bits 0x12/0x0 rrc: 4 type: IBT gid 0 flags: 0x50200000000000 nid: 0@lo remote: 0x99ae7c71d2c8a124 expref: 9 pid: 19511 timeout: 0 lvb_type: 0 LustreError: 22976:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -5 LustreError: 22526:0:(file.c:5042:ll_inode_revalidate_fini()) lustre: revalidate FID [0x240000404:0x1e:0x0] error: rc = -5 LustreError: 22976:0:(llite_lib.c:1836:ll_md_setattr()) Skipped 1 previous similar message LustreError: 22525:0:(mdc_request.c:1436:mdc_read_page()) lustre-MDT0001-mdc-ffff880299a4ca88: [0x240000400:0x6:0x0] lock enqueue fails: rc = -108 LustreError: 25916:0:(file.c:5042:ll_inode_revalidate_fini()) lustre: revalidate FID [0x240000404:0x1e:0x0] error: rc = -108 LustreError: 25916:0:(file.c:5042:ll_inode_revalidate_fini()) Skipped 35 previous similar messages LustreError: 23067:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -108 LustreError: 23067:0:(llite_lib.c:1836:ll_md_setattr()) Skipped 1 previous similar message LustreError: 22517:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff880299a4ca88: inode [0x240000404:0x1e:0x0] mdc close failed: rc = -108 LustreError: 24645:0:(file.c:5042:ll_inode_revalidate_fini()) lustre: revalidate FID [0x240000404:0x1e:0x0] error: rc = -108 LustreError: 23925:0:(mdc_request.c:1436:mdc_read_page()) lustre-MDT0001-mdc-ffff880299a4ca88: [0x240000400:0xa:0x0] lock enqueue fails: rc = -108 LustreError: 23925:0:(mdc_request.c:1436:mdc_read_page()) Skipped 6 previous similar messages LustreError: 26512:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff880299a4ca88: inode [0x240000404:0x85:0x0] mdc close failed: rc = -108 LustreError: 26512:0:(file.c:234:ll_close_inode_openhandle()) Skipped 2 previous similar messages Lustre: lustre-MDT0001-mdc-ffff880299a4ca88: Connection restored to 192.168.123.170@tcp (at 0@lo) LustreError: 11026:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0000_UUID lock: ffff880264eaa980/0x99ae7c71d2ca562d lrc: 3/0,0 mode: PW/PW res: [0x259:0x0:0x0].0x0 rrc: 4 type: EXT [0->18446744073709551615] (req 2097152->18446744073709551615) gid 0 flags: 0x60000000020020 nid: 0@lo remote: 0x99ae7c71d2ca55d9 expref: 2594 pid: 21804 timeout: 477 lvb_type: 0 LustreError: 21784:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff880271884b40 x1710112389583168/t0(0) o106->lustre-OST0000@0@lo:15/16 lens 328/280 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 11-0: lustre-OST0000-osc-ffff880299a4ca88: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-OST0000-osc-ffff880299a4ca88: Connection to lustre-OST0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: 4122:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff880281b55e00 x1710112389660928/t0(0) o10->lustre-OST0003-osc-ffff88008dee12a8@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630890646 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dir_create.sh.0' INFO: task mv:20961 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mv D ffff880087920040 11040 20961 16069 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff81251fe1>] lock_rename+0x31/0xe0 [<ffffffff8125862f>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed432>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817e8a76>] ? trace_do_page_fault+0x56/0x170 [<ffffffff8125975e>] SyS_renameat2+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task mv:22715 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mv D ffff8802935fc9d0 11168 22715 16174 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff81251fe1>] lock_rename+0x31/0xe0 [<ffffffff8125862f>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed432>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817e8a76>] ? trace_do_page_fault+0x56/0x170 [<ffffffff8125975e>] SyS_renameat2+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task mv:23043 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mv D ffff88027c4ac9d0 11344 23043 16557 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff81251fe1>] lock_rename+0x31/0xe0 [<ffffffff8125862f>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed432>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817e8a76>] ? trace_do_page_fault+0x56/0x170 [<ffffffff8125975e>] SyS_renameat2+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 Lustre: lustre-OST0001: haven't heard from client 099f8133-efe2-4519-9be0-b3a78691d601 (at 0@lo) in 47 seconds. I think it's dead, and I am evicting it. exp ffff8800ab1bdd28, cur 1630890633 expire 1630890603 last 1630890586 LustreError: 167-0: lustre-OST0000-osc-ffff880299a4ca88: This client was evicted by lustre-OST0000; in progress operations using this service will fail. Lustre: lustre-OST0000: Client 989be46e-abfb-4210-a299-8eaa2ef57189 (at 0@lo) reconnecting LustreError: 676:0:(ofd_obd.c:501:ofd_destroy_export()) lustre-OST0000: cli 989be46e-abfb-4210-a299-8eaa2ef57189/ffff88008b3f5d28 has 28672 pending on destroyed export Lustre: lustre-OST0001-osc-ffff88008dee12a8: Connection restored to 192.168.123.170@tcp (at 0@lo) LustreError: 676:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_destroy_export: tot_granted 108219584 != fo_tot_granted 108248256 LustreError: 676:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_destroy_export: tot_pending 0 != fo_tot_pending 28672 LustreError: 12726:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 108219584 != fo_tot_granted 108248256 LustreError: 12726:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 28672 Lustre: 4135:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.170@tcp:/lustre/fid: [0x280000404:0x116:0x0]// may get corrupted (rc -108) LustreError: 12736:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 108218688 != fo_tot_granted 108247360 LustreError: 12736:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 28672 Lustre: lustre-OST0001-osc-ffff88008dee12a8: disconnect after 56s idle Lustre: Skipped 7 previous similar messages LustreError: 11-0: lustre-OST0001-osc-ffff88008dee12a8: operation ost_disconnect to node 0@lo failed: rc = -107 LustreError: Skipped 55 previous similar messages LustreError: 27236:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0000-osc-ffff880299a4ca88: namespace resource [0x258:0x0:0x0].0x0 (ffff880265aea0c0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 27236:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 91 previous similar messages Lustre: lustre-OST0000-osc-ffff880299a4ca88: Connection restored to 192.168.123.170@tcp (at 0@lo) LustreError: 12736:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 108218304 != fo_tot_granted 108246976 LustreError: 12724:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 896 != fo_tot_pending 29568 LustreError: 12724:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 1 previous similar message LustreError: 12736:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 2 previous similar messages LustreError: 12735:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 116655424 != fo_tot_granted 116684096 LustreError: 12735:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 1 previous similar message LustreError: 12735:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 61440 != fo_tot_pending 90112 LustreError: 12735:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 2 previous similar messages Lustre: 4128:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff8802596e3240 x1710112390445824/t0(0) o10->lustre-OST0003-osc-ffff88008dee12a8@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630890712 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dd.0' LustreError: 12735:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 116595008 != fo_tot_granted 116623680 LustreError: 12735:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 5 previous similar messages LustreError: 12735:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 28672 LustreError: 12735:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 5 previous similar messages LustreError: 12736:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 125033408 != fo_tot_granted 125062080 LustreError: 12736:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 5 previous similar messages LustreError: 12736:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 28672 LustreError: 12736:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 5 previous similar messages Lustre: 11180:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 1/4/0, destroy: 0/0/0 Lustre: 11180:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 54542 previous similar messages Lustre: 11180:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 4/209/0 Lustre: 11180:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 54542 previous similar messages Lustre: 11180:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1377/1386/0, punch: 0/0/0, quota 8/80/0 Lustre: 11180:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 54542 previous similar messages Lustre: 11180:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 2/33/0, delete: 0/0/0 Lustre: 11180:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 54433 previous similar messages Lustre: 11180:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 11180:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 54542 previous similar messages Lustre: 4128:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff8802596e3240 x1710112392380992/t0(0) o10->lustre-OST0003-osc-ffff88008dee12a8@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630890763 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dd.0' Lustre: 25409:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 511 < left 6761, rollback = 7 Lustre: 25409:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 55169 previous similar messages LustreError: 12730:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 125032768 != fo_tot_granted 125061440 LustreError: 12730:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 13 previous similar messages LustreError: 12730:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 896 != fo_tot_pending 29568 LustreError: 12730:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 13 previous similar messages Lustre: 16808:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 8: before 509 < left 23408, rollback = 9 Lustre: 16808:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 78 previous similar messages LustreError: 19511:0:(mdd_object.c:401:mdd_xattr_get()) lustre-MDD0001: object [0x240000403:0x1b8:0x0] not found: rc = -2 LustreError: 23151:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 200271168 != fo_tot_granted 200299840 LustreError: 23151:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 33 previous similar messages LustreError: 23151:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 896 != fo_tot_pending 29568 LustreError: 23151:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 33 previous similar messages LustreError: 11026:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0001_UUID lock: ffff88007afee940/0x99ae7c71d2cfaf86 lrc: 3/0,0 mode: PW/PW res: [0x7fa:0x0:0x0].0x0 rrc: 4 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000400030020 nid: 0@lo remote: 0x99ae7c71d2cfaf7f expref: 2008 pid: 19847 timeout: 641 lvb_type: 0 LustreError: 14707:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff880298f957c0 x1710112395697792/t0(0) o106->lustre-OST0001@0@lo:15/16 lens 328/280 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 11-0: lustre-OST0001-osc-ffff88008dee12a8: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-OST0001-osc-ffff88008dee12a8: Connection to lustre-OST0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 4117:0:(osc_request.c:1035:osc_init_grant()) lustre-OST0001-osc-ffff88008dee12a8: granted 4419584 but already consumed 6287360 LustreError: 167-0: lustre-OST0001-osc-ffff88008dee12a8: This client was evicted by lustre-OST0001; in progress operations using this service will fail. LustreError: 14707:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 13 previous similar messages LustreError: 11344:0:(out_handler.c:910:out_tx_end()) lustre-MDT0002-osd: undo for /home/green/git/lustre-release/lustre/ptlrpc/../../lustre/target/out_handler.c:659: rc = -524 LustreError: 11344:0:(out_handler.c:910:out_tx_end()) Skipped 1 previous similar message LustreError: 11949:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0002-osp-MDT0001: fail to cancel 1 llog-records: rc = -116 LustreError: 11949:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0002-osp-MDT0001: fail to cancel 1 of 1 llog-records: rc = -116 Lustre: 4137:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.170@tcp:/lustre/fid: [0x200000403:0x14a:0x0]/ may get corrupted (rc -108) Lustre: 4137:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.170@tcp:/lustre/fid: [0x240000403:0x1fe:0x0]// may get corrupted (rc -108) Lustre: 4136:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.170@tcp:/lustre/fid: [0x200000403:0x11c:0x0]/ may get corrupted (rc -108) Lustre: 4136:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.170@tcp:/lustre/fid: [0x200000403:0x15d:0x0]/ may get corrupted (rc -108) Lustre: 4136:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.170@tcp:/lustre/fid: [0x280000403:0x20a:0x0]/ may get corrupted (rc -108) LustreError: 30045:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0001-osc-ffff88008dee12a8: namespace resource [0x300000401:0xc6f:0x0].0x0 (ffff88007bda34c0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 30045:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 1 previous similar message Lustre: lustre-OST0001-osc-ffff88008dee12a8: Connection restored to 192.168.123.170@tcp (at 0@lo) Lustre: 11338:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 6s req@ffff88028d2d25c0 x1710112396732672/t0(0) o1000->lustre-MDT0000-mdtlov_UUID@0@lo:0/0 lens 1528/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'osp_up1-0.0' Lustre: dir [0x280000404:0x127:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 10 previous similar messages LustreError: 11155:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0001: '5' migration was interrupted, run 'lfs migrate -m 0 -c 1 -H crush 5' to finish migration. LustreError: 11026:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0000_UUID lock: ffff8802615b07c0/0x99ae7c71d2d620e6 lrc: 3/0,0 mode: PW/PW res: [0x82f:0x0:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000400030020 nid: 0@lo remote: 0x99ae7c71d2d620ca expref: 3007 pid: 12703 timeout: 724 lvb_type: 0 LustreError: 18556:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802864e9300 x1710112398720576/t0(0) o105->lustre-OST0000@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 18556:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 10 previous similar messages LustreError: 21853:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630890872 with bad export cookie 11073925362060566975 LustreError: 11008:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630890873 with bad export cookie 11073925362060566975 LustreError: 11008:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) Skipped 1 previous similar message LustreError: 11-0: lustre-OST0000-osc-ffff880299a4ca88: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-OST0000-osc-ffff880299a4ca88: Connection to lustre-OST0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0000-osc-ffff880299a4ca88: This client was evicted by lustre-OST0000; in progress operations using this service will fail. LustreError: 11026:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0001_UUID lock: ffff88027be0c3c0/0x99ae7c71d2d1d3a8 lrc: 3/0,0 mode: PW/PW res: [0x7fa:0x0:0x0].0x0 rrc: 2 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000080020020 nid: 0@lo remote: 0x99ae7c71d2d1d3a1 expref: 3624 pid: 26915 timeout: 742 lvb_type: 0 LustreError: 10994:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630890890 with bad export cookie 11073925362059591728 Lustre: lustre-OST0001-osc-ffff880299a4ca88: Connection to lustre-OST0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0001-osc-ffff880299a4ca88: This client was evicted by lustre-OST0001; in progress operations using this service will fail. LustreError: 27098:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff880262a79f80 x1710112399324672/t0(0) o104->lustre-OST0001@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 27098:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 14 previous similar messages Lustre: 4130:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.170@tcp:/lustre/fid: [0x200000403:0x15d:0x0]/ may get corrupted (rc -108) Lustre: 4128:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.170@tcp:/lustre/fid: [0x280000404:0x209:0x0]/ may get corrupted (rc -108) Lustre: 17478:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 6s req@ffff8802734b9300 x1710112399235136/t0(0) o35->099f8133-efe2-4519-9be0-b3a78691d601@0@lo:0/0 lens 392/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'cat.0' Lustre: 21298:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 8s req@ffff88027b19f700 x1710112399317952/t0(0) o1000->lustre-MDT0001-mdtlov_UUID@0@lo:0/0 lens 1528/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'osp_up0-1.0' Lustre: 21298:0:(service.c:2157:ptlrpc_server_handle_req_in()) Skipped 2 previous similar messages Lustre: 4125:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.170@tcp:/lustre/fid: [0x240000405:0xbd:0x0]/ may get corrupted (rc -108) LustreError: 5:0:(ofd_obd.c:501:ofd_destroy_export()) lustre-OST0001: cli 989be46e-abfb-4210-a299-8eaa2ef57189/ffff88028ea5c138 has 3194880 pending on destroyed export LustreError: 31947:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0001-osc-ffff880299a4ca88: namespace resource [0x3d0:0x0:0x0].0x0 (ffff880261e256c0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 31947:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 195 previous similar messages Lustre: lustre-OST0001-osc-ffff880299a4ca88: Connection restored to 192.168.123.170@tcp (at 0@lo) LustreError: 12732:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 15350976 != fo_tot_granted 18545856 LustreError: 12732:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 49 previous similar messages LustreError: 12732:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 3194880 LustreError: 12732:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 49 previous similar messages LustreError: 31536:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0000-osc-ffff880299a4ca88: namespace resource [0x520:0x0:0x0].0x0 (ffff88025cdce840) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 31536:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 509 previous similar messages Lustre: 12727:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 6s req@ffff88029842de00 x1710112400440960/t0(0) o13->lustre-MDT0000-mdtlov_UUID@0@lo:0/0 lens 224/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'osp-pre-0-0.0' Lustre: 12727:0:(service.c:2157:ptlrpc_server_handle_req_in()) Skipped 7 previous similar messages Lustre: 18579:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 6s req@ffff880266d9a5c0 x1710112400811328/t0(0) o1000->lustre-MDT0000-mdtlov_UUID@0@lo:0/0 lens 1320/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'osp_up2-0.0' LustreError: 16716:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0001: '4' migration was interrupted, run 'lfs migrate -m 2 -c 1 -H crush 4' to finish migration. LustreError: 11949:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0000-osp-MDT0001: fail to cancel 1 llog-records: rc = -2 LustreError: 11949:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0000-osp-MDT0001: fail to cancel 1 of 1 llog-records: rc = -2 Lustre: 11165:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 1/4/0, destroy: 0/0/0 Lustre: 11165:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 106211 previous similar messages Lustre: 11165:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 4/209/0 Lustre: 11165:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 106211 previous similar messages Lustre: 11165:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 713/722/0, punch: 0/0/0, quota 8/56/0 Lustre: 11165:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 106211 previous similar messages Lustre: 11165:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 2/33/0, delete: 0/0/0 Lustre: 11165:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 106211 previous similar messages Lustre: 11165:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 11165:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 106211 previous similar messages Lustre: 11187:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 512 < left 1069, rollback = 7 Lustre: 11187:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 104223 previous similar messages INFO: task mv:30640 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mv D ffff8800823cc9d0 11216 30640 16676 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff81251fe1>] lock_rename+0x31/0xe0 [<ffffffff8125862f>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed432>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817e8a76>] ? trace_do_page_fault+0x56/0x170 [<ffffffff8125975e>] SyS_renameat2+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 LustreError: 11026:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0003_UUID lock: ffff8800790b34c0/0x99ae7c71d2dd801c lrc: 3/0,0 mode: PW/PW res: [0x3a7:0x0:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000080020020 nid: 0@lo remote: 0x99ae7c71d2dd7ff9 expref: 3184 pid: 19978 timeout: 869 lvb_type: 0 LustreError: 11-0: lustre-OST0003-osc-ffff88008dee12a8: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 6 previous similar messages Lustre: lustre-OST0003-osc-ffff88008dee12a8: Connection to lustre-OST0003 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0003-osc-ffff88008dee12a8: This client was evicted by lustre-OST0003; in progress operations using this service will fail. Lustre: 4122:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.170@tcp:/lustre/fid: [0x240000403:0x2a6:0x0]/ may get corrupted (rc -108) Lustre: 4123:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.170@tcp:/lustre/fid: [0x280000403:0x2da:0x0]/ may get corrupted (rc -108) Lustre: 4125:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.170@tcp:/lustre/fid: [0x200000403:0x23b:0x0]/ may get corrupted (rc -108) Lustre: 4121:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.170@tcp:/lustre/fid: [0x240000403:0x289:0x0]// may get corrupted (rc -5) Lustre: 11198:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 6s req@ffff880260ba8680 x1710112404577984/t0(0) o35->099f8133-efe2-4519-9be0-b3a78691d601@0@lo:0/0 lens 392/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'dir_create.sh.0' Lustre: 11164:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 8: before 512 < left 23034, rollback = 9 Lustre: 11164:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 237 previous similar messages ptlrpc_watchdog_fire: 32 callbacks suppressed Lustre: mdt01_008: service thread pid 17206 was inactive for 56.985 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Lustre: Skipped 1 previous similar message Pid: 17206, comm: mdt01_008 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_fini+0xa25/0xec0 [ptlrpc] [<0>] ldlm_cli_enqueue+0x436/0xa30 [ptlrpc] [<0>] osp_md_object_lock+0x160/0x300 [osp] [<0>] lod_object_lock+0xdb/0x7d0 [lod] [<0>] mdd_object_lock+0x2d/0xd0 [mdd] [<0>] mdt_remote_object_lock_try+0x1cc/0x510 [mdt] [<0>] mdt_remote_object_lock+0x2a/0x30 [mdt] [<0>] mdt_rename_lock+0xbe/0x4d0 [mdt] [<0>] mdt_reint_migrate+0x6e8/0x11d0 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x76c/0xb40 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: lustre-OST0003-osc-ffff88008dee12a8: Connection restored to 192.168.123.170@tcp (at 0@lo) Lustre: Skipped 1 previous similar message Lustre: mdt01_004: service thread pid 16766 was inactive for 56.218 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 16766, comm: mdt01_004 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] mdt_rename_lock+0x24b/0x4d0 [mdt] [<0>] mdt_reint_migrate+0x6e8/0x11d0 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x76c/0xb40 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 11172, comm: mdt04_000 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] __cond_resched+0x26/0x30 [<0>] kmem_cache_alloc_trace+0x58/0x310 [<0>] sub_updates_write+0x759/0xe32 [ptlrpc] [<0>] top_trans_stop+0x7c2/0xf60 [ptlrpc] [<0>] lod_trans_stop+0x25c/0x340 [lod] [<0>] mdd_trans_stop+0x28/0x16e [mdd] [<0>] mdd_migrate_object+0xf1f/0x1ba0 [mdd] [<0>] mdd_migrate+0x3d6/0x8e0 [mdd] [<0>] mdo_migrate+0x4c/0x4e [mdt] [<0>] mdt_reint_migrate+0x1023/0x11d0 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x76c/0xb40 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: mdt04_011: service thread pid 25409 was inactive for 72.240 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 2 previous similar messages ------------[ cut here ]------------ WARNING: CPU: 8 PID: 11172 at /home/green/git/lustre-release/ldiskfs/ext4_jbd2.c:266 __ldiskfs_handle_dirty_metadata+0x1d2/0x230 [ldiskfs] Modules linked in: loop zfs(PO) zunicode(PO) zzstd(O) zlua(O) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) jbd2 mbcache lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) dm_flakey dm_mod libcfs(OE) crc_t10dif crct10dif_generic sb_edac edac_core iosf_mbi crc32_pclmul ghash_clmulni_intel aesni_intel lrw gf128mul glue_helper ablk_helper cryptd virtio_balloon virtio_console pcspkr i2c_piix4 ip_tables rpcsec_gss_krb5 ata_generic drm_kms_helper pata_acpi ttm crct10dif_pclmul drm crct10dif_common crc32c_intel ata_piix drm_panel_orientation_quirks serio_raw virtio_blk libata i2c_core floppy CPU: 8 PID: 11172 Comm: mdt04_000 Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011 Call Trace: [<ffffffff817d9418>] dump_stack+0x19/0x1b [<ffffffff8108d538>] __warn+0xd8/0x100 [<ffffffff8108d67d>] warn_slowpath_null+0x1d/0x20 [<ffffffffa0bfc262>] __ldiskfs_handle_dirty_metadata+0x1d2/0x230 [ldiskfs] [<ffffffffa0bdc33d>] ldiskfs_getblk+0x13d/0x210 [ldiskfs] [<ffffffffa0bdc437>] ldiskfs_bread+0x27/0xe0 [ldiskfs] [<ffffffffa0ca0e97>] osd_write+0x4e7/0xcb0 [osd_ldiskfs] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffffa04d96f3>] dt_record_write+0x33/0x120 [obdclass] [<ffffffffa04986b2>] llog_osd_write_rec+0x1052/0x19a0 [obdclass] [<ffffffffa0485760>] llog_write_rec+0x290/0x590 [obdclass] [<ffffffffa048b4ee>] llog_cat_add_rec+0x23e/0x950 [obdclass] [<ffffffffa04825ff>] llog_add+0x17f/0x1f0 [obdclass] [<ffffffffa04b3a39>] ? lprocfs_counter_add+0xf9/0x160 [obdclass] [<ffffffffa081c60d>] sub_updates_write+0xb6b/0xe32 [ptlrpc] [<ffffffffa07ec892>] top_trans_stop+0x7c2/0xf60 [ptlrpc] [<ffffffffa0ebcf3c>] lod_trans_stop+0x25c/0x340 [lod] [<ffffffffa0d91408>] mdd_trans_stop+0x28/0x16e [mdd] [<ffffffffa0d72b2f>] mdd_migrate_object+0xf1f/0x1ba0 [mdd] [<ffffffffa04d8375>] ? lu_object_find_at+0x285/0xb10 [obdclass] [<ffffffffa0d73b86>] mdd_migrate+0x3d6/0x8e0 [mdd] [<ffffffffa0e450d8>] mdo_migrate+0x4c/0x4e [mdt] [<ffffffffa0e04e53>] mdt_reint_migrate+0x1023/0x11d0 [mdt] [<ffffffffa0e05087>] mdt_reint_rec+0x87/0x240 [mdt] [<ffffffffa0dda55c>] mdt_reint_internal+0x76c/0xb40 [mdt] [<ffffffffa0de7657>] mdt_reint+0x67/0x150 [mdt] [<ffffffffa07db26e>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<ffffffffa032945e>] ? libcfs_nid2str_r+0xfe/0x130 [lnet] [<ffffffffa0786990>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<ffffffffa0788529>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffffa0787950>] ? ptlrpc_wait_event+0x620/0x620 [ptlrpc] [<ffffffff810ba0f4>] kthread+0xe4/0xf0 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 ---[ end trace 3e032eb585932b8e ]--- LDISKFS-fs: ldiskfs_getblk:888: aborting transaction: error 28 in __ldiskfs_handle_dirty_metadata LDISKFS-fs error (device dm-0): ldiskfs_getblk:888: inode #193: block 52433: comm mdt04_000: journal_dirty_metadata failed: handle type 0 started at line 1972, credits 512/0, errcode -28 Aborting journal on device dm-0-8. LDISKFS-fs (dm-0): Remounting filesystem read-only LustreError: 11172:0:(osd_io.c:2103:osd_ldiskfs_write_record()) lustre-MDT0000: error reading offset 233472 (block 57, size 48, offs 233464), credits 0/0: rc = -28 LustreError: 11172:0:(update_trans.c:985:top_trans_stop()) lustre-MDT0000-osd: write updates failed: rc = -28 LustreError: 11172:0:(osd_io.c:2110:osd_ldiskfs_write_record()) journal_get_write_access() returned error -30 LustreError: 11172:0:(tgt_lastrcvd.c:1253:tgt_add_reply_data()) lustre-MDT0000: can't update reply_data file: rc = -30 LustreError: 11172:0:(osd_handler.c:2075:osd_trans_stop()) lustre-MDT0000: failed in transaction hook: rc = -30 LDISKFS-fs error (device dm-0) in osd_trans_stop:2082: error 28 LustreError: 14938:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-OST0003-osc-MDT0000: fail to cancel 1 llog-records: rc = -30 LustreError: 14938:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-OST0003-osc-MDT0000: fail to cancel 1 of 1 llog-records: rc = -30 LustreError: 14938:0:(osp_sync.c:1079:osp_sync_process_committed()) lustre-OST0003-osc-MDT0000: can't cancel record: rc = -30 LustreError: 11172:0:(osd_handler.c:2085:osd_trans_stop()) lustre-MDT0000: failed to stop transaction: rc = -28 LustreError: 11172:0:(update_trans.c:1011:top_trans_stop()) lustre-MDT0000-osd: stop trans failed: rc = -30 LustreError: 2510:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -5 LustreError: 10954:0:(osd_handler.c:1779:osd_trans_commit_cb()) transaction @0xffff8802c82744a8 commit error: 2 LustreError: 20203:0:(llog_cat.c:602:llog_cat_add_rec()) llog_write_rec -116: lh=ffff8802995589f8 LustreError: 20203:0:(update_trans.c:1062:top_trans_stop()) lustre-MDT0000-osp-MDT0001: write updates failed: rc = -116 LustreError: 11174:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0000: '10' migration was interrupted, run 'lfs migrate -m 2 -c 1 -H crush 10' to finish migration. LustreError: 2444:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -30 LustreError: 2444:0:(llite_lib.c:1836:ll_md_setattr()) Skipped 4 previous similar messages LustreError: 11174:0:(mdt_open.c:1237:mdt_cross_open()) lustre-MDT0002: [0x280000404:0x264:0x0] doesn't exist!: rc = -14 LustreError: 14938:0:(osp_sync.c:1079:osp_sync_process_committed()) lustre-OST0003-osc-MDT0000: can't cancel record: rc = -30 LustreError: 14938:0:(osp_sync.c:1079:osp_sync_process_committed()) Skipped 1 previous similar message LustreError: 14241:0:(osp_sync.c:1094:osp_sync_process_committed()) lustre-OST0002-osc-MDT0000: can't cancel 4 records: rc = -30 LustreError: 15996:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -30 LustreError: 15996:0:(llite_lib.c:1836:ll_md_setattr()) Skipped 5 previous similar messages LustreError: 16581:0:(mdt_open.c:1237:mdt_cross_open()) lustre-MDT0001: [0x240000405:0x178:0x0] doesn't exist!: rc = -14 LustreError: 12360:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0000-osp-MDT0002: fail to cancel 1 llog-records: rc = -2 LustreError: 12360:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) Skipped 11 previous similar messages LustreError: 12360:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0000-osp-MDT0002: fail to cancel 1 of 1 llog-records: rc = -2 LustreError: 12360:0:(llog_cat.c:787:llog_cat_cancel_records()) Skipped 10 previous similar messages LustreError: 21581:0:(mdt_open.c:1237:mdt_cross_open()) lustre-MDT0001: [0x240000405:0x178:0x0] doesn't exist!: rc = -14 LustreError: 17132:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0000: '4' migration was interrupted, run 'lfs migrate -m 2 -c 1 -H crush 4' to finish migration. | Externally reported by onyx-68 boilpot email |
racer test 1: racer on clients: centos-0.localnet DURATION=2700 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa178ae52>] ll_migrate+0x9b2/0xec0 [lustre] PGD 8000000091660067 PUD 918db067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC LustreError: 19549:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -30 Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) dm_flakey dm_mod loop zfs(PO) zunicode(PO) zzstd(O) zlua(O) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) jbd2 mbcache crc_t10dif crct10dif_generic sb_edac edac_core iosf_mbi crc32_pclmul ghash_clmulni_intel aesni_intel lrw gf128mul glue_helper ablk_helper cryptd virtio_console virtio_balloon pcspkr i2c_piix4 ip_tables rpcsec_gss_krb5 drm_kms_helper ata_generic ttm pata_acpi crct10dif_pclmul crct10dif_common drm ata_piix crc32c_intel drm_panel_orientation_quirks virtio_blk serio_raw libata i2c_core floppy [last unloaded: libcfs] CPU: 12 PID: 19320 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011 task: ffff8802a43649d0 ti: ffff8802b832c000 task.ti: ffff8802b832c000 RIP: 0010:[<ffffffffa178ae52>] [<ffffffffa178ae52>] ll_migrate+0x9b2/0xec0 [lustre] RSP: 0018:ffff8802b832fbc8 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff8802b980a588 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff880326e307b8 RDI: ffff880326e30798 RBP: ffff8802b832fc48 R08: ffff880094d6f138 R09: 0000000000000001 R10: 0000000000000000 R11: ffff8802b832f5e6 R12: 0000000000000000 R13: ffff8802b101e018 R14: ffff880272d5a618 R15: 0000000000000030 FS: 00007f5a95396740(0000) GS:ffff880331d00000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 CR2: 0000000000000008 CR3: 00000002b12fe000 CR4: 00000000001607e0 Call Trace: [<ffffffff81242213>] ? __check_object_size+0x1c3/0x220 [<ffffffffa1778561>] ll_dir_ioctl+0x5d01/0x6ed0 [lustre] [<ffffffff811b515b>] ? unlock_page+0x2b/0x30 [<ffffffff811e5e2e>] ? do_read_fault.isra.64+0x12e/0x170 [<ffffffff811ea1cf>] ? handle_pte_fault+0x2ef/0xca0 [<ffffffff810d64a6>] ? select_task_rq_fair+0x5c6/0x780 [<ffffffff811ecf6d>] ? __handle_mm_fault+0x31d/0x720 LustreError: 10400:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0001-osp-MDT0002: fail to cancel 1 llog-records: rc = -116 LustreError: 10400:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) Skipped 1 previous similar message LustreError: 10400:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0001-osp-MDT0002: fail to cancel 1 of 1 llog-records: rc = -116 LustreError: 10400:0:(llog_cat.c:787:llog_cat_cancel_records()) Skipped 1 previous similar message [<ffffffff8125b3fd>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff81245a38>] ? vfs_write+0x178/0x230 [<ffffffff8125b751>] SyS_ioctl+0xa1/0xc0 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 | LustreError: 12241:0:(out_handler.c:910:out_tx_end()) lustre-MDT0000-osd: undo for /home/green/git/lustre-release/lustre/ptlrpc/../../lustre/target/out_handler.c:445: rc = -524 LustreError: 10193:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0000-osp-MDT0001: fail to cancel 1 llog-records: rc = -116 LustreError: 10193:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0000-osp-MDT0001: fail to cancel 1 of 1 llog-records: rc = -116 LustreError: 10193:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0000-osp-MDT0001: fail to cancel 1 llog-records: rc = -116 LustreError: 10193:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0000-osp-MDT0001: fail to cancel 1 of 1 llog-records: rc = -116 Lustre: 11404:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 512 < left 1209, rollback = 7 Lustre: 11404:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 11404:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 1/89/0 Lustre: 11404:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1200/1209/0, punch: 0/0/0, quota 3/3/0 Lustre: 11404:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 11404:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 13738:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 8: before 510 < left 20399, rollback = 9 Lustre: 13784:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 3/12/1, destroy: 0/0/0 Lustre: 13784:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 1201 previous similar messages Lustre: 13784:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 2/2/0, xattr_set: 8/610/0 Lustre: 13784:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 1201 previous similar messages Lustre: 9930:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 512 < left 51567, rollback = 7 Lustre: 9930:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 1200 previous similar messages Lustre: 9930:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 5996/51567/0, punch: 0/0/0, quota 6/6/0 Lustre: 9930:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 1201 previous similar messages Lustre: 9930:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 1199/20383/0, delete: 2/5/0 Lustre: 9930:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 1201 previous similar messages Lustre: 9930:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 1/1/0 Lustre: 9930:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 1201 previous similar messages Lustre: 13784:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 508 < left 610, rollback = 2 Lustre: 13784:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 1 previous similar message ------------[ cut here ]------------ WARNING: CPU: 5 PID: 9930 at /home/green/git/lustre-release/ldiskfs/namei.c:3331 ldiskfs_orphan_add+0x11e/0x2a0 [ldiskfs] Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) dm_flakey dm_mod loop zfs(PO) zunicode(PO) zzstd(O) zlua(O) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) jbd2 mbcache crc_t10dif crct10dif_generic sb_edac edac_core iosf_mbi crc32_pclmul ghash_clmulni_intel aesni_intel lrw gf128mul glue_helper ablk_helper cryptd virtio_console virtio_balloon pcspkr i2c_piix4 ip_tables rpcsec_gss_krb5 drm_kms_helper ata_generic ttm pata_acpi crct10dif_pclmul crct10dif_common drm ata_piix crc32c_intel drm_panel_orientation_quirks virtio_blk serio_raw libata i2c_core floppy [last unloaded: libcfs] CPU: 5 PID: 9930 Comm: mdt_rdpg02_001 Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011 Call Trace: [<ffffffff817d9418>] dump_stack+0x19/0x1b [<ffffffff8108d538>] __warn+0xd8/0x100 [<ffffffff8108d67d>] warn_slowpath_null+0x1d/0x20 [<ffffffffa0d564ae>] ldiskfs_orphan_add+0x11e/0x2a0 [ldiskfs] [<ffffffffa0d479eb>] ldiskfs_xattr_inode_orphan_add.constprop.17+0xbb/0x110 [ldiskfs] [<ffffffffa0d4a417>] ldiskfs_xattr_delete_inode+0x77/0x3c0 [ldiskfs] [<ffffffffa0d6aaa0>] ? ldiskfs_evict_inode+0x190/0x670 [ldiskfs] [<ffffffffa0d6aad9>] ldiskfs_evict_inode+0x1c9/0x670 [ldiskfs] [<ffffffff812649af>] evict+0xaf/0x180 [<ffffffff81264de5>] iput+0xf5/0x180 [<ffffffffa0c86d08>] osd_object_delete+0x1f8/0x370 [osd_ldiskfs] [<ffffffffa04e459d>] lu_object_free.isra.26+0xcd/0x1d0 [obdclass] [<ffffffffa04e8cf1>] lu_object_put+0xa1/0x480 [obdclass] [<ffffffffa0ee6af4>] mdt_mfd_close+0x1f4/0x870 [mdt] [<ffffffffa0eece14>] mdt_close_internal+0xb4/0x240 [mdt] [<ffffffffa0eed224>] mdt_close+0x284/0x850 [mdt] [<ffffffffa07eb2d4>] ? tgt_request_preprocess.isra.27+0xf4/0x850 [ptlrpc] [<ffffffffa07ec26e>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<ffffffffa03b745e>] ? libcfs_nid2str_r+0xfe/0x130 [lnet] [<ffffffffa0797990>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<ffffffffa0799529>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffffa0798950>] ? ptlrpc_wait_event+0x620/0x620 [ptlrpc] [<ffffffff810ba0f4>] kthread+0xe4/0xf0 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 ---[ end trace 5192ad504da6a0a6 ]--- Lustre: 13784:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 505 < left 869, rollback = 7 Lustre: 13784:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 4798 previous similar messages Lustre: 13784:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 1/4/4, destroy: 0/0/0 Lustre: 13784:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 4804 previous similar messages Lustre: 13784:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 4/209/0 Lustre: 13784:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 908 previous similar messages Lustre: 13784:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 860/869/0, punch: 0/0/0, quota 8/80/2 Lustre: 13784:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 4804 previous similar messages Lustre: 13784:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 2/33/1, delete: 0/0/0 Lustre: 13784:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 4804 previous similar messages Lustre: 13784:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 13784:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 4804 previous similar messages Lustre: 9905:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 498 < left 760, rollback = 2 Lustre: 9905:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 4 previous similar messages Lustre: 9923:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 8: before 510 < left 14619, rollback = 9 Lustre: 9923:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 9 previous similar messages Lustre: 9923:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 859/3436/0, destroy: 1/4/0 Lustre: 9923:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 869 previous similar messages Lustre: 9923:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 3/3/0, xattr_set: 863/148/0 Lustre: 9923:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 869 previous similar messages Lustre: 9923:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 4296/36947/0, punch: 0/0/0, quota 10/90/0 Lustre: 9923:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 869 previous similar messages Lustre: 9923:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 860/14619/0, delete: 2/5/1 Lustre: 9923:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 869 previous similar messages Lustre: 9923:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 1/1/0, ref_del: 2/2/1 Lustre: 9923:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 869 previous similar messages Lustre: 9923:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 508 < left 36947, rollback = 7 Lustre: 9923:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 859 previous similar messages Lustre: 12406:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 508 < left 36947, rollback = 7 Lustre: 12406:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 860 previous similar messages Lustre: 12406:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 859/3436/0, destroy: 0/0/0 Lustre: 12406:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 861 previous similar messages Lustre: 12406:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/1, xattr_set: 861/15/0 Lustre: 12406:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 861 previous similar messages Lustre: 12406:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 4296/36947/0, punch: 0/0/0, quota 8/40/3 Lustre: 12406:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 861 previous similar messages Lustre: 12406:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 859/14603/0, delete: 0/0/0 Lustre: 12406:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 861 previous similar messages Lustre: 12406:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 12406:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 861 previous similar messages Lustre: mdt07_000: service thread pid 9922 was inactive for 40.051 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 9922, comm: mdt07_000 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] osp_precreate_reserve+0x49f/0x9b0 [osp] [<0>] osp_declare_create+0x1ac/0x650 [osp] [<0>] lod_sub_declare_create+0xe2/0x210 [lod] [<0>] lod_qos_declare_object_on+0xf3/0x420 [lod] [<0>] lod_ost_alloc_rr.constprop.23+0xa9d/0x1140 [lod] [<0>] lod_qos_prep_create+0x123d/0x1a20 [lod] [<0>] lod_declare_instantiate_components+0xa7/0x1e0 [lod] [<0>] lod_declare_update_plain+0x67b/0x970 [lod] [<0>] lod_declare_layout_change+0x69e/0xba0 [lod] [<0>] mdd_declare_layout_change+0x4b/0x100 [mdd] [<0>] mdd_layout_change+0xde9/0x1ad0 [mdd] [<0>] mdt_layout_change+0x2ff/0x490 [mdt] [<0>] mdt_intent_layout+0x940/0xe90 [mdt] [<0>] mdt_intent_opc+0x1dd/0xc10 [mdt] [<0>] mdt_intent_policy+0x1a1/0x360 [mdt] [<0>] ldlm_lock_enqueue+0x3c2/0xb40 [ptlrpc] [<0>] ldlm_handle_enqueue0+0xaa6/0x16d0 [ptlrpc] [<0>] tgt_enqueue+0x64/0x240 [ptlrpc] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: mdt00_003: service thread pid 10323 was inactive for 40.089 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 10323, comm: mdt00_003 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] mdt_object_local_lock+0x527/0xb90 [mdt] [<0>] mdt_object_lock_internal+0x70/0x390 [mdt] [<0>] mdt_object_lock_try+0x27/0xb0 [mdt] [<0>] mdt_getattr_name_lock+0x17ae/0x2590 [mdt] [<0>] mdt_intent_getattr+0x2d5/0x4b0 [mdt] [<0>] mdt_intent_opc+0x1dd/0xc10 [mdt] [<0>] mdt_intent_policy+0x1a1/0x360 [mdt] [<0>] ldlm_lock_enqueue+0x3c2/0xb40 [ptlrpc] [<0>] ldlm_handle_enqueue0+0xaa6/0x16d0 [ptlrpc] [<0>] tgt_enqueue+0x64/0x240 [ptlrpc] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: mdt04_005: service thread pid 13324 was inactive for 40.033 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 13324, comm: mdt04_005 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] call_rwsem_down_write_failed+0x17/0x30 [<0>] lod_qos_statfs_update+0x97/0x2e0 [lod] [<0>] lod_qos_prep_create+0x1b9/0x1a20 [lod] [<0>] lod_prepare_create+0x23b/0x320 [lod] [<0>] lod_declare_striped_create+0x1fe/0xa00 [lod] [<0>] lod_declare_create+0x1f5/0x560 [lod] [<0>] mdd_declare_create_object_internal+0x12a/0x340 [mdd] [<0>] mdd_declare_create_object.isra.34+0x51/0xc00 [mdd] [<0>] mdd_declare_create+0x62/0x6f0 [mdd] [<0>] mdd_create+0x9b9/0x1a40 [mdd] [<0>] mdt_reint_open+0x250b/0x3830 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x76c/0xb40 [mdt] [<0>] mdt_intent_open+0x93/0x480 [mdt] [<0>] mdt_intent_opc+0x1dd/0xc10 [mdt] [<0>] mdt_intent_policy+0x1a1/0x360 [mdt] [<0>] ldlm_lock_enqueue+0x3c2/0xb40 [ptlrpc] [<0>] ldlm_handle_enqueue0+0xaa6/0x16d0 [ptlrpc] Lustre: mdt03_004: service thread pid 12227 was inactive for 40.030 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. [<0>] tgt_enqueue+0x64/0x240 [ptlrpc] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: mdt03_002: service thread pid 9912 was inactive for 40.134 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 2 previous similar messages Lustre: 13815:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 512 < left 36947, rollback = 7 Lustre: 13815:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 10327 previous similar messages Lustre: 13815:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 859/3436/0, destroy: 0/0/0 Lustre: 13815:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 10327 previous similar messages Lustre: 13815:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 861/15/0 Lustre: 13815:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 10327 previous similar messages Lustre: 13815:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 4296/36947/0, punch: 0/0/0, quota 6/6/0 Lustre: 13815:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 10327 previous similar messages Lustre: 13815:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 859/14603/0, delete: 0/0/0 Lustre: 13815:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 10327 previous similar messages Lustre: 13815:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 13815:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 10327 previous similar messages Lustre: mdt07_002: service thread pid 9924 was inactive for 62.094 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: mdt06_005: service thread pid 12914 was inactive for 62.008 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: mdt04_004: service thread pid 13169 was inactive for 62.179 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 7 previous similar messages LustreError: 9894:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0002_UUID lock: ffff880090d3cb40/0x4b279abf16b2c051 lrc: 3/0,0 mode: PR/PR res: [0x280000404:0x1a:0x0].0x0 bits 0x12/0x0 rrc: 7 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0x4b279abf16b2c027 expref: 39 pid: 12227 timeout: 3750 lvb_type: 0 LustreError: 11-0: lustre-MDT0002-mdc-ffff880274528008: operation mds_reint to node 0@lo failed: rc = -107 Lustre: lustre-MDT0002-mdc-ffff880274528008: Connection to lustre-MDT0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-MDT0002-mdc-ffff880274528008: This client was evicted by lustre-MDT0002; in progress operations using this service will fail. LustreError: 13212:0:(file.c:5042:ll_inode_revalidate_fini()) lustre: revalidate FID [0x280000404:0x1a:0x0] error: rc = -5 LustreError: 12457:0:(vvp_io.c:1793:vvp_io_init()) lustre: refresh file layout [0x280000403:0x4:0x0] error -5. Lustre: lustre-MDT0002-mdc-ffff880274528008: Connection restored to 192.168.123.100@tcp (at 0@lo) LustreError: 13212:0:(file.c:5042:ll_inode_revalidate_fini()) Skipped 31 previous similar messages Lustre: lustre-MDT0002-mdc-ffff880223af37e8: Connection to lustre-MDT0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-MDT0002-mdc-ffff880223af37e8: This client was evicted by lustre-MDT0002; in progress operations using this service will fail. LustreError: 13465:0:(file.c:5042:ll_inode_revalidate_fini()) lustre: revalidate FID [0x280000404:0x1a:0x0] error: rc = -108 LustreError: 13465:0:(file.c:5042:ll_inode_revalidate_fini()) Skipped 5 previous similar messages LustreError: 12469:0:(mdc_request.c:1436:mdc_read_page()) lustre-MDT0002-mdc-ffff880223af37e8: [0x280000403:0x1:0x0] lock enqueue fails: rc = -108 LustreError: 14796:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff880223af37e8: inode [0x280000404:0x5:0x0] mdc close failed: rc = -108 LustreError: 14796:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-MDT0002-mdc-ffff880223af37e8: namespace resource [0x280000403:0x1:0x0].0x0 (ffff8802ab414a40) refcount nonzero (1) after lock cleanup; forcing cleanup. Lustre: lustre-MDT0002-mdc-ffff880223af37e8: Connection restored to 192.168.123.100@tcp (at 0@lo) Lustre: 13690:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 512 < left 1113, rollback = 7 Lustre: 13690:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 10338 previous similar messages Lustre: 13690:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 13690:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 10338 previous similar messages Lustre: 13690:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 1/89/0 Lustre: 13690:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 10338 previous similar messages Lustre: 13690:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1104/1113/0, punch: 0/0/0, quota 3/3/0 Lustre: 13690:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 10338 previous similar messages Lustre: 13690:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 13690:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 10338 previous similar messages Lustre: 13690:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 13690:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 10338 previous similar messages LustreError: 9894:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0002_UUID lock: ffff8802a89b6940/0x4b279abf16b55ec2 lrc: 3/0,0 mode: PR/PR res: [0x280000404:0x1a:0x0].0x0 bits 0x12/0x0 rrc: 4 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0x4b279abf16b55d09 expref: 30 pid: 13945 timeout: 3857 lvb_type: 0 LustreError: 9894:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 1 previous similar message LustreError: 11-0: lustre-MDT0002-mdc-ffff880223af37e8: operation mds_reint to node 0@lo failed: rc = -107 Lustre: lustre-MDT0002-mdc-ffff880223af37e8: Connection to lustre-MDT0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-MDT0002-mdc-ffff880223af37e8: This client was evicted by lustre-MDT0002; in progress operations using this service will fail. LustreError: 14839:0:(file.c:5042:ll_inode_revalidate_fini()) lustre: revalidate FID [0x280000404:0x2e:0x0] error: rc = -5 LustreError: 13789:0:(mdc_request.c:1436:mdc_read_page()) lustre-MDT0002-mdc-ffff880223af37e8: [0x280000400:0x5:0x0] lock enqueue fails: rc = -108 Lustre: dir [0x280000404:0x4e:0x0] stripe 0 readdir failed: -108, directory is partially accessed! LustreError: 15028:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff880223af37e8: inode [0x280000403:0xa:0x0] mdc close failed: rc = -108 LustreError: 15028:0:(file.c:234:ll_close_inode_openhandle()) Skipped 3 previous similar messages LustreError: 15028:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-MDT0002-mdc-ffff880223af37e8: namespace resource [0x280000404:0x4e:0x0].0x0 (ffff8802ab449bc0) refcount nonzero (1) after lock cleanup; forcing cleanup. Lustre: lustre-MDT0002-mdc-ffff880223af37e8: Connection restored to 192.168.123.100@tcp (at 0@lo) LustreError: 14839:0:(file.c:5042:ll_inode_revalidate_fini()) Skipped 32 previous similar messages LustreError: 10323:0:(ldlm_request.c:123:ldlm_expired_completion_wait()) ### lock timed out (enqueued at 1630872220, 300s ago); not entering recovery in server code, just going back to sleep ns: mdt-lustre-MDT0002_UUID lock: ffff880224bd9300/0x4b279abf16b2559f lrc: 3/1,0 mode: --/PR res: [0x280000403:0x4:0x0].0x0 bits 0x13/0x40 rrc: 8 type: IBT gid 0 flags: 0x40210000000000 nid: local remote: 0x0 expref: -99 pid: 10323 timeout: 0 lvb_type: 0 LustreError: dumping log to /tmp/lustre-log.1630872520.10323 LustreError: 9924:0:(ldlm_request.c:123:ldlm_expired_completion_wait()) ### lock timed out (enqueued at 1630872221, 300s ago); not entering recovery in server code, just going back to sleep ns: mdt-lustre-MDT0002_UUID lock: ffff88023ed07480/0x4b279abf16b25ae6 lrc: 3/1,0 mode: --/PR res: [0x280000403:0x4:0x0].0x0 bits 0x13/0x40 rrc: 8 type: IBT gid 0 flags: 0x40210000000000 nid: local remote: 0x0 expref: -99 pid: 9924 timeout: 0 lvb_type: 0 LustreError: 12227:0:(ldlm_request.c:123:ldlm_expired_completion_wait()) ### lock timed out (enqueued at 1630872230, 300s ago); not entering recovery in server code, just going back to sleep ns: mdt-lustre-MDT0002_UUID lock: ffff880090d396c0/0x4b279abf16b2c066 lrc: 3/1,0 mode: --/PR res: [0x280000400:0x4:0x0].0x0 bits 0x12/0x0 rrc: 29 type: IBT gid 0 flags: 0x40210000000000 nid: local remote: 0x0 expref: -99 pid: 12227 timeout: 0 lvb_type: 0 LustreError: 12227:0:(ldlm_request.c:123:ldlm_expired_completion_wait()) Skipped 1 previous similar message LustreError: 9894:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0002_UUID lock: ffff8802a8923c40/0x4b279abf16b5ca69 lrc: 3/0,0 mode: PR/PR res: [0x280000404:0x4e:0x0].0x0 bits 0x12/0x0 rrc: 3 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0x4b279abf16b5ca5b expref: 30 pid: 13793 timeout: 3958 lvb_type: 0 LustreError: 11-0: lustre-MDT0002-mdc-ffff880223af37e8: operation mds_reint to node 0@lo failed: rc = -107 Lustre: lustre-MDT0002-mdc-ffff880223af37e8: Connection to lustre-MDT0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-MDT0002-mdc-ffff880223af37e8: This client was evicted by lustre-MDT0002; in progress operations using this service will fail. LustreError: 12742:0:(file.c:5042:ll_inode_revalidate_fini()) lustre: revalidate FID [0x280000404:0x43:0x0] error: rc = -5 LustreError: 13728:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff880223af37e8: inode [0x280000406:0x6:0x0] mdc close failed: rc = -108 LustreError: 13728:0:(file.c:234:ll_close_inode_openhandle()) Skipped 1 previous similar message LustreError: 12742:0:(file.c:5042:ll_inode_revalidate_fini()) Skipped 8 previous similar messages Lustre: lustre-MDT0002-mdc-ffff880223af37e8: Connection restored to 192.168.123.100@tcp (at 0@lo) Lustre: lustre-OST0000-osc-ffff880274528008: disconnect after 23s idle ptlrpc_watchdog_fire: 13 callbacks suppressed Lustre: mdt06_000: service thread pid 9919 was inactive for 236.214 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 9919, comm: mdt06_000 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] mdt_object_local_lock+0x527/0xb90 [mdt] [<0>] mdt_object_lock_internal+0x70/0x390 [mdt] [<0>] mdt_getattr_name_lock+0xaca/0x2590 [mdt] [<0>] mdt_intent_getattr+0x2d5/0x4b0 [mdt] [<0>] mdt_intent_opc+0x1dd/0xc10 [mdt] [<0>] mdt_intent_policy+0x1a1/0x360 [mdt] [<0>] ldlm_lock_enqueue+0x3c2/0xb40 [ptlrpc] [<0>] ldlm_handle_enqueue0+0xaa6/0x16d0 [ptlrpc] [<0>] tgt_enqueue+0x64/0x240 [ptlrpc] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 12406, comm: mdt06_004 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] mdt_object_local_lock+0x527/0xb90 [mdt] [<0>] mdt_object_lock_internal+0x70/0x390 [mdt] [<0>] mdt_getattr_name_lock+0xaca/0x2590 [mdt] [<0>] mdt_intent_getattr+0x2d5/0x4b0 [mdt] [<0>] mdt_intent_opc+0x1dd/0xc10 [mdt] [<0>] mdt_intent_policy+0x1a1/0x360 [mdt] [<0>] ldlm_lock_enqueue+0x3c2/0xb40 [ptlrpc] [<0>] ldlm_handle_enqueue0+0xaa6/0x16d0 [ptlrpc] [<0>] tgt_enqueue+0x64/0x240 [ptlrpc] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 9920, comm: mdt06_001 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] mdt_object_local_lock+0x527/0xb90 [mdt] [<0>] mdt_object_lock_internal+0x70/0x390 [mdt] [<0>] mdt_getattr_name_lock+0xaca/0x2590 [mdt] [<0>] mdt_intent_getattr+0x2d5/0x4b0 [mdt] [<0>] mdt_intent_opc+0x1dd/0xc10 [mdt] [<0>] mdt_intent_policy+0x1a1/0x360 [mdt] [<0>] ldlm_lock_enqueue+0x3c2/0xb40 [ptlrpc] [<0>] ldlm_handle_enqueue0+0xaa6/0x16d0 [ptlrpc] [<0>] tgt_enqueue+0x64/0x240 [ptlrpc] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: mdt06_003: service thread pid 12383 was inactive for 236.752 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 1 previous similar message Lustre: mdt04_002: service thread pid 9915 was inactive for 262.015 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 1 previous similar message LustreError: 9915:0:(ldlm_request.c:123:ldlm_expired_completion_wait()) ### lock timed out (enqueued at 1630872333, 300s ago); not entering recovery in server code, just going back to sleep ns: mdt-lustre-MDT0002_UUID lock: ffff8802aa45f480/0x4b279abf16b5235e lrc: 3/1,0 mode: --/PR res: [0x280000403:0x4:0x0].0x0 bits 0x13/0x40 rrc: 8 type: IBT gid 0 flags: 0x40210000000000 nid: local remote: 0x0 expref: -99 pid: 9915 timeout: 0 lvb_type: 0 LustreError: 9915:0:(ldlm_request.c:123:ldlm_expired_completion_wait()) Skipped 6 previous similar messages LustreError: 9921:0:(ldlm_request.c:123:ldlm_expired_completion_wait()) ### lock timed out (enqueued at 1630872338, 300s ago); not entering recovery in server code, just going back to sleep ns: mdt-lustre-MDT0002_UUID lock: ffff8802a42eda40/0x4b279abf16b54b43 lrc: 3/1,0 mode: --/CR res: [0x280000404:0x2e:0x0].0x0 bits 0x2/0x0 rrc: 24 type: IBT gid 0 flags: 0x40210000000000 nid: local remote: 0x0 expref: -99 pid: 9921 timeout: 0 lvb_type: 0 LustreError: 9921:0:(ldlm_request.c:123:ldlm_expired_completion_wait()) Skipped 8 previous similar messages Lustre: mdt06_009: service thread pid 14818 was inactive for 286.621 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 10 previous similar messages LustreError: 9906:0:(ldlm_request.c:123:ldlm_expired_completion_wait()) ### lock timed out (enqueued at 1630872440, 300s ago); not entering recovery in server code, just going back to sleep ns: mdt-lustre-MDT0002_UUID lock: ffff8802a6d1a980/0x4b279abf16b5bff7 lrc: 3/1,0 mode: --/PR res: [0x280000400:0x4:0x0].0x0 bits 0x12/0x0 rrc: 29 type: IBT gid 0 flags: 0x40210000000000 nid: local remote: 0x0 expref: -99 pid: 9906 timeout: 0 lvb_type: 0 LustreError: 9906:0:(ldlm_request.c:123:ldlm_expired_completion_wait()) Skipped 6 previous similar messages LustreError: 14848:0:(ldlm_request.c:123:ldlm_expired_completion_wait()) ### lock timed out (enqueued at 1630872541, 300s ago); not entering recovery in server code, just going back to sleep ns: mdt-lustre-MDT0002_UUID lock: ffff8802a89ff480/0x4b279abf16b6041b lrc: 3/1,0 mode: --/PR res: [0x280000404:0x2e:0x0].0x0 bits 0x12/0x0 rrc: 24 type: IBT gid 0 flags: 0x40210000000000 nid: local remote: 0x0 expref: -99 pid: 14848 timeout: 0 lvb_type: 0 LustreError: 14848:0:(ldlm_request.c:123:ldlm_expired_completion_wait()) Skipped 7 previous similar messages Lustre: lustre-MDT0002-osp-MDT0001: Connection to lustre-MDT0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 15053:0:(ldlm_request.c:142:ldlm_expired_completion_wait()) ### lock timed out (enqueued at 1630872541, 300s ago), entering recovery for lustre-MDT0002_UUID@192.168.123.100@tcp ns: lustre-MDT0002-osp-MDT0001 lock: ffff8802a5e8d2c0/0x4b279abf16b607da lrc: 4/0,1 mode: --/EX res: [0x280000403:0x1:0x0].0x0 bits 0x2/0x0 rrc: 2 type: IBT gid 0 flags: 0x1000001000000 nid: local remote: 0x4b279abf16b607e1 expref: -99 pid: 15053 timeout: 0 lvb_type: 0 Lustre: lustre-MDT0002: Received new MDS connection from 0@lo, keep former export from same NID Lustre: lustre-MDT0002-osp-MDT0001: Connection restored to 192.168.123.100@tcp (at 0@lo) Lustre: mdt00_008: service thread pid 13214 was inactive for 312.185 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 7 previous similar messages Lustre: 13506:0:(service.c:1436:ptlrpc_at_send_early_reply()) @@@ Could not add any time (5/5), not sending early reply req@ffff8802ace0cb40 x1710093363309760/t0(0) o101->90d32ca6-eab3-4950-9694-28f5f48d5398@0@lo:453/0 lens 576/3576 e 18 to 0 dl 1630872933 ref 2 fl Interpret:/0/0 rc 0/0 job:'ln.0' Lustre: 13514:0:(service.c:1436:ptlrpc_at_send_early_reply()) @@@ Could not add any time (5/5), not sending early reply req@ffff8802aa0bac00 x1710093363325056/t0(0) o101->90d32ca6-eab3-4950-9694-28f5f48d5398@0@lo:453/0 lens 576/3576 e 18 to 0 dl 1630872933 ref 2 fl Interpret:/0/0 rc 0/0 job:'ls.0' Lustre: 13514:0:(service.c:1436:ptlrpc_at_send_early_reply()) Skipped 1 previous similar message Lustre: 13514:0:(service.c:1436:ptlrpc_at_send_early_reply()) @@@ Could not add any time (5/5), not sending early reply req@ffff8800942f6440 x1710093363336640/t0(0) o101->90d32ca6-eab3-4950-9694-28f5f48d5398@0@lo:454/0 lens 576/3576 e 18 to 0 dl 1630872934 ref 2 fl Interpret:/0/0 rc 0/0 job:'ls.0' Lustre: 13514:0:(service.c:1436:ptlrpc_at_send_early_reply()) Skipped 3 previous similar messages ptlrpc_watchdog_fire: 22 callbacks suppressed Lustre: mdt05_009: service thread pid 14955 was inactive for 390.193 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Lustre: Skipped 2 previous similar messages Pid: 14955, comm: mdt05_009 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x66d/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] mdt_object_local_lock+0x527/0xb90 [mdt] [<0>] mdt_object_lock_internal+0x70/0x390 [mdt] [<0>] mdt_getattr_name_lock+0xaca/0x2590 [mdt] [<0>] mdt_intent_getattr+0x2d5/0x4b0 [mdt] [<0>] mdt_intent_opc+0x1dd/0xc10 [mdt] [<0>] mdt_intent_policy+0x1a1/0x360 [mdt] [<0>] ldlm_lock_enqueue+0x3c2/0xb40 [ptlrpc] [<0>] ldlm_handle_enqueue0+0xaa6/0x16d0 [ptlrpc] [<0>] tgt_enqueue+0x64/0x240 [ptlrpc] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: 12646:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1630872333/real 1630872333] req@ffff8802ace0b880 x1710093363309760/t0(0) o101->lustre-MDT0002-mdc-ffff880274528008@0@lo:12/10 lens 576/1136 e 18 to 1 dl 1630872934 ref 2 fl Rpc:XQr/0/ffffffff rc 0/-1 job:'ln.0' Lustre: lustre-MDT0002-mdc-ffff880274528008: Connection to lustre-MDT0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: lustre-MDT0002: Client 90d32ca6-eab3-4950-9694-28f5f48d5398 (at 0@lo) reconnecting Lustre: lustre-MDT0002-mdc-ffff880274528008: Connection restored to 192.168.123.100@tcp (at 0@lo) Lustre: mdt02_007: service thread pid 14848 was inactive for 437.436 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 15199, comm: mdt02_010 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x66d/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] mdt_object_local_lock+0x527/0xb90 [mdt] [<0>] mdt_object_lock_internal+0x70/0x390 [mdt] [<0>] mdt_getattr_name_lock+0xaca/0x2590 [mdt] [<0>] mdt_intent_getattr+0x2d5/0x4b0 [mdt] [<0>] mdt_intent_opc+0x1dd/0xc10 [mdt] [<0>] mdt_intent_policy+0x1a1/0x360 [mdt] [<0>] ldlm_lock_enqueue+0x3c2/0xb40 [ptlrpc] [<0>] ldlm_handle_enqueue0+0xaa6/0x16d0 [ptlrpc] [<0>] tgt_enqueue+0x64/0x240 [ptlrpc] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: Skipped 1 previous similar message Pid: 14848, comm: mdt02_007 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x66d/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] mdt_object_local_lock+0x527/0xb90 [mdt] [<0>] mdt_object_lock_internal+0x70/0x390 [mdt] [<0>] mdt_getattr_name_lock+0xaca/0x2590 [mdt] [<0>] mdt_intent_getattr+0x2d5/0x4b0 [mdt] [<0>] mdt_intent_opc+0x1dd/0xc10 [mdt] [<0>] mdt_intent_policy+0x1a1/0x360 [mdt] [<0>] ldlm_lock_enqueue+0x3c2/0xb40 [ptlrpc] [<0>] ldlm_handle_enqueue0+0xaa6/0x16d0 [ptlrpc] [<0>] tgt_enqueue+0x64/0x240 [ptlrpc] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: mdt04_010: service thread pid 15228 was inactive for 441.468 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 3 previous similar messages Lustre: 9918:0:(service.c:1436:ptlrpc_at_send_early_reply()) @@@ Could not add any time (5/5), not sending early reply req@ffff8800943eb240 x1710093364182720/t0(0) o101->f86a0b4a-dd01-45d5-b8a0-59cf1c1be86c@0@lo:661/0 lens 576/3576 e 13 to 0 dl 1630873141 ref 2 fl Interpret:/0/0 rc 0/0 job:'getfattr.0' Lustre: 9918:0:(service.c:1436:ptlrpc_at_send_early_reply()) Skipped 2 previous similar messages Lustre: 12709:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1630872541/real 1630872541] req@ffff8802a8a00040 x1710093364180288/t0(0) o101->lustre-MDT0002-mdc-ffff880223af37e8@0@lo:12/10 lens 576/1136 e 2 to 1 dl 1630873142 ref 2 fl Rpc:XQr/0/ffffffff rc 0/-1 job:'ls.0' Lustre: lustre-MDT0002-mdc-ffff880223af37e8: Connection to lustre-MDT0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: lustre-MDT0002: Client f86a0b4a-dd01-45d5-b8a0-59cf1c1be86c (at 0@lo) reconnecting Lustre: lustre-MDT0002-mdc-ffff880223af37e8: Connection restored to 192.168.123.100@tcp (at 0@lo) Lustre: 13751:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1630872934/real 1630872934] req@ffff8802aaf7e440 x1710093363325568/t0(0) o36->lustre-MDT0002-mdc-ffff880274528008@0@lo:12/10 lens 544/440 e 18 to 1 dl 1630873535 ref 2 fl Rpc:XQr/2/ffffffff rc -11/-1 job:'setfattr.0' Lustre: 13284:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1630872934/real 1630872934] req@ffff8802aa1ed180 x1710093363320320/t0(0) o101->lustre-MDT0002-mdc-ffff880274528008@0@lo:12/10 lens 576/1136 e 18 to 1 dl 1630873535 ref 2 fl Rpc:XQr/2/ffffffff rc -11/-1 job:'ls.0' Lustre: 13284:0:(client.c:2285:ptlrpc_expire_one_request()) Skipped 5 previous similar messages Lustre: lustre-MDT0002-mdc-ffff880274528008: Connection to lustre-MDT0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 1 previous similar message Lustre: lustre-MDT0002: Client 90d32ca6-eab3-4950-9694-28f5f48d5398 (at 0@lo) reconnecting Lustre: Skipped 1 previous similar message Lustre: lustre-MDT0002-mdc-ffff880274528008: Connection restored to 192.168.123.100@tcp (at 0@lo) Lustre: Skipped 1 previous similar message Lustre: 13751:0:(client.c:2285:ptlrpc_expire_one_request()) Skipped 4 previous similar messages LustreError: 12914:0:(ldlm_lockd.c:1425:ldlm_handle_enqueue0()) ### lock on destroyed export ffff88026017efc8 ns: mdt-lustre-MDT0002_UUID lock: ffff8800930d1a80/0x4b279abf16b63904 lrc: 3/0,0 mode: CR/CR res: [0x280000404:0x47:0x0].0x0 bits 0x9/0x0 rrc: 2 type: IBT gid 0 flags: 0x50200000000000 nid: 0@lo remote: 0x4b279abf16b2a3b0 expref: 17 pid: 12914 timeout: 0 lvb_type: 0 Lustre: 12383:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (170/1092s); client may timeout req@ffff880225e00cc0 x1710093363489088/t0(0) o101->f86a0b4a-dd01-45d5-b8a0-59cf1c1be86c@0@lo:28/0 lens 576/7192 e 0 to 0 dl 1630872508 ref 1 fl Complete:/0/0 rc -107/-107 job:'cat.0' LustreError: 12914:0:(ldlm_lockd.c:1425:ldlm_handle_enqueue0()) Skipped 20 previous similar messages Lustre: 9904:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 500 < left 521, rollback = 2 Lustre: 9904:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 3/12/8, destroy: 0/0/0 Lustre: 9904:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 5523 previous similar messages Lustre: 9904:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 2/2/0, xattr_set: 7/521/0 Lustre: 9904:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 5523 previous similar messages Lustre: 9904:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 5/136/0, punch: 0/0/0, quota 6/6/0 Lustre: 9904:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 5523 previous similar messages Lustre: 9904:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 12/215/4, delete: 1/1/0 Lustre: 9904:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 5523 previous similar messages Lustre: 9904:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 6/6/0, ref_del: 1/1/0 Lustre: 9904:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 5523 previous similar messages Lustre: 9904:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 496 < left 521, rollback = 2 Lustre: 9904:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 2 previous similar messages Lustre: 12869:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 512 < left 1698, rollback = 7 Lustre: 12869:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 5523 previous similar messages Lustre: 12869:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 1/4/0, destroy: 0/0/0 Lustre: 12869:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 12869:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 4/209/0 Lustre: 12869:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 12869:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1689/1698/0, punch: 0/0/0, quota 8/80/0 Lustre: 12869:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 12869:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 2/33/0, delete: 0/0/0 Lustre: 12869:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 12869:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 12869:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 3 previous similar messages Lustre: 13002:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 504 < left 671, rollback = 2 Lustre: 9906:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 511 < left 610, rollback = 2 Lustre: 9906:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 15 previous similar messages Lustre: 9923:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 511 < left 1698, rollback = 7 Lustre: 9923:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 1688 previous similar messages Lustre: 9923:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 1/4/1, destroy: 0/0/0 Lustre: 9923:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 1710 previous similar messages Lustre: 9923:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 4/209/0 Lustre: 9923:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 1710 previous similar messages Lustre: 9923:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1689/1698/0, punch: 0/0/0, quota 8/80/0 Lustre: 9923:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 1710 previous similar messages Lustre: 9923:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 2/33/0, delete: 0/0/0 Lustre: 9923:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 1710 previous similar messages Lustre: 9923:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 9923:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 1710 previous similar messages Lustre: 12192:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 505 < left 849, rollback = 2 Lustre: 12192:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 5 previous similar messages LustreError: 14897:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0002: failed to get lu_attr of [0x280000404:0x4e:0x0]: rc = -2 Lustre: 13815:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 511 < left 11405, rollback = 7 Lustre: 13815:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 1688 previous similar messages Lustre: dir [0x200000403:0x93:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 2 previous similar messages Lustre: 12914:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 8: before 511 < left 4521, rollback = 9 Lustre: 12914:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 17 previous similar messages Lustre: 12383:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 238/952/0, destroy: 0/0/0 Lustre: 12383:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 2770 previous similar messages Lustre: 12383:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/1, xattr_set: 240/15/0 Lustre: 12383:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 2770 previous similar messages Lustre: 12383:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1191/10244/0, punch: 0/0/0, quota 10/106/7 Lustre: 12383:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 2770 previous similar messages Lustre: 12383:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 238/4046/0, delete: 0/0/0 Lustre: 12383:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 2770 previous similar messages Lustre: 12383:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 12383:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 2770 previous similar messages Lustre: dir [0x280000405:0x4:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: 6717:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff8802aa1cf0c0 x1710093367269824/t0(0) o10->lustre-OST0003-osc-ffff880274528008@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630873665 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'truncate.0' Lustre: 12364:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 510 < left 1498, rollback = 7 Lustre: 12364:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 4896 previous similar messages Lustre: 16738:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 499 < left 866, rollback = 2 Lustre: 16738:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 12 previous similar messages Lustre: 14848:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 1688/6752/0, destroy: 1/4/0 Lustre: 14848:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 21057 previous similar messages Lustre: 14848:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 3/3/0, xattr_set: 1692/148/0 Lustre: 14848:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 21057 previous similar messages Lustre: 14848:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 8441/72594/0, punch: 0/0/0, quota 9/65/0 Lustre: 14848:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 21057 previous similar messages Lustre: 14848:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 1689/28712/0, delete: 2/5/0 Lustre: 14848:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 21057 previous similar messages Lustre: 14848:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 1/1/0, ref_del: 2/2/1 Lustre: 14848:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 21057 previous similar messages Lustre: 9920:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 475 < left 72558, rollback = 7 Lustre: 9920:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 38038 previous similar messages ptlrpc_watchdog_fire: 3 callbacks suppressed Lustre: mdt03_000: service thread pid 9910 was inactive for 62.050 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 9910, comm: mdt03_000 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] mdt_object_local_lock+0x527/0xb90 [mdt] [<0>] mdt_object_lock_internal+0x70/0x390 [mdt] [<0>] mdt_getattr_name_lock+0x1130/0x2590 [mdt] [<0>] mdt_intent_getattr+0x2d5/0x4b0 [mdt] [<0>] mdt_intent_opc+0x1dd/0xc10 [mdt] [<0>] mdt_intent_policy+0x1a1/0x360 [mdt] [<0>] ldlm_lock_enqueue+0x3c2/0xb40 [ptlrpc] [<0>] ldlm_handle_enqueue0+0xaa6/0x16d0 [ptlrpc] [<0>] tgt_enqueue+0x64/0x240 [ptlrpc] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: mdt03_006: service thread pid 13719 was inactive for 62.194 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 13719, comm: mdt03_006 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] mdt_object_local_lock+0x527/0xb90 [mdt] [<0>] mdt_object_lock_internal+0x70/0x390 [mdt] [<0>] mdt_getattr_name_lock+0xaca/0x2590 [mdt] [<0>] mdt_intent_getattr+0x2d5/0x4b0 [mdt] [<0>] mdt_intent_opc+0x1dd/0xc10 [mdt] [<0>] mdt_intent_policy+0x1a1/0x360 [mdt] [<0>] ldlm_lock_enqueue+0x3c2/0xb40 [ptlrpc] [<0>] ldlm_handle_enqueue0+0xaa6/0x16d0 [ptlrpc] [<0>] tgt_enqueue+0x64/0x240 [ptlrpc] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: mdt03_008: service thread pid 13815 was inactive for 62.098 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 13815, comm: mdt03_008 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] mdt_object_local_lock+0x527/0xb90 [mdt] [<0>] mdt_object_lock_internal+0x70/0x390 [mdt] [<0>] mdt_getattr_name_lock+0xaca/0x2590 [mdt] [<0>] mdt_intent_getattr+0x2d5/0x4b0 [mdt] [<0>] mdt_intent_opc+0x1dd/0xc10 [mdt] [<0>] mdt_intent_policy+0x1a1/0x360 [mdt] [<0>] ldlm_lock_enqueue+0x3c2/0xb40 [ptlrpc] [<0>] ldlm_handle_enqueue0+0xaa6/0x16d0 [ptlrpc] [<0>] tgt_enqueue+0x64/0x240 [ptlrpc] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: mdt03_009: service thread pid 13945 was inactive for 62.063 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 1 previous similar message LustreError: 9894:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0002_UUID lock: ffff8802ae17c780/0x4b279abf16ba193d lrc: 3/0,0 mode: PR/PR res: [0x280000408:0x59:0x0].0x0 bits 0x12/0x0 rrc: 6 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0x4b279abf16ba1928 expref: 76 pid: 14818 timeout: 5179 lvb_type: 0 LustreError: 18033:0:(ldlm_lockd.c:1425:ldlm_handle_enqueue0()) ### lock on destroyed export ffff8802aab537e8 ns: mdt-lustre-MDT0002_UUID lock: ffff8802b284c000/0x4b279abf16ba3560 lrc: 3/0,0 mode: PR/PR res: [0x280000408:0x59:0x0].0x0 bits 0x12/0x0 rrc: 5 type: IBT gid 0 flags: 0x50200000000000 nid: 0@lo remote: 0x4b279abf16ba3528 expref: 60 pid: 18033 timeout: 0 lvb_type: 0 LustreError: 18033:0:(ldlm_lockd.c:1425:ldlm_handle_enqueue0()) Skipped 5 previous similar messages LustreError: 11-0: lustre-MDT0002-mdc-ffff880274528008: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-MDT0002-mdc-ffff880274528008: Connection to lustre-MDT0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-MDT0002-mdc-ffff880274528008: This client was evicted by lustre-MDT0002; in progress operations using this service will fail. LustreError: 18034:0:(file.c:5042:ll_inode_revalidate_fini()) lustre: revalidate FID [0x280000408:0x59:0x0] error: rc = -5 LustreError: Skipped 1 previous similar message LustreError: 18209:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff880274528008: inode [0x280000404:0x43:0x0] mdc close failed: rc = -108 LustreError: 18209:0:(file.c:234:ll_close_inode_openhandle()) Skipped 1 previous similar message Lustre: lustre-MDT0002-mdc-ffff880274528008: Connection restored to 192.168.123.100@tcp (at 0@lo) Lustre: 16657:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 511 < left 1498, rollback = 7 Lustre: 16657:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 22350 previous similar messages Lustre: 16657:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 1/4/0, destroy: 0/0/0 Lustre: 16657:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 43210 previous similar messages Lustre: 16657:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 4/209/0 Lustre: 16657:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 43210 previous similar messages Lustre: 16657:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1489/1498/0, punch: 0/0/0, quota 8/80/0 Lustre: 16657:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 43210 previous similar messages Lustre: 16657:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 2/33/1, delete: 0/0/0 Lustre: 16657:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 43210 previous similar messages Lustre: 16657:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 16657:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 43210 previous similar messages LustreError: 13514:0:(ldlm_lockd.c:1425:ldlm_handle_enqueue0()) ### lock on destroyed export ffff8802aab537e8 ns: mdt-lustre-MDT0002_UUID lock: ffff8802ae187840/0x4b279abf16ba1975 lrc: 3/0,0 mode: PR/PR res: [0x280000400:0xb:0x0].0x0 bits 0x12/0x0 rrc: 2 type: IBT gid 0 flags: 0x50200000000000 nid: 0@lo remote: 0x4b279abf16ba1967 expref: 13 pid: 13514 timeout: 0 lvb_type: 0 LustreError: 13514:0:(ldlm_lockd.c:1425:ldlm_handle_enqueue0()) Skipped 2 previous similar messages Lustre: 16812:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 8: before 509 < left 2328, rollback = 9 Lustre: 16812:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 32 previous similar messages Lustre: 9902:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000408:0x60:0x0] with magic=0xbd60bd0 Lustre: 13171:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000408:0x64:0x0] with magic=0xbd60bd0 Lustre: 13171:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 1 previous similar message Lustre: 14915:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 3/12/8, destroy: 0/0/0 Lustre: 14915:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 7072 previous similar messages Lustre: 14915:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 6/582/0 Lustre: 14915:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 7072 previous similar messages Lustre: 14915:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 5/136/0, punch: 0/0/0, quota 8/56/0 Lustre: 14915:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 7072 previous similar messages Lustre: 14915:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 10/203/3, delete: 0/0/0 Lustre: 14915:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 7072 previous similar messages Lustre: 14915:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 4/4/0, ref_del: 0/0/0 Lustre: 14915:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 7072 previous similar messages Lustre: 12482:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 509 < left 10244, rollback = 7 Lustre: 12482:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 7028 previous similar messages Lustre: lustre-OST0001-osc-ffff880223af37e8: disconnect after 22s idle LustreError: 9894:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 140s: evicting client at 0@lo ns: filter-lustre-OST0000_UUID lock: ffff8802b9458040/0x4b279abf16bd23b7 lrc: 3/0,0 mode: PW/PW res: [0x8b4:0x0:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 1310720->1376255) gid 0 flags: 0x60000400020020 nid: 0@lo remote: 0x4b279abf16bd2187 expref: 13 pid: 17494 timeout: 5391 lvb_type: 0 LustreError: 15227:0:(ldlm_lockd.c:1425:ldlm_handle_enqueue0()) ### lock on destroyed export ffff88006fd2b7e8 ns: mdt-lustre-MDT0000_UUID lock: ffff8802b949c780/0x4b279abf16bd200d lrc: 3/0,0 mode: PR/PR res: [0x200000404:0x8a:0x0].0x0 bits 0x12/0x0 rrc: 11 type: IBT gid 0 flags: 0x50200400000020 nid: 0@lo remote: 0x4b279abf16bd1f81 expref: 9 pid: 15227 timeout: 0 lvb_type: 0 LustreError: 15227:0:(ldlm_lockd.c:1425:ldlm_handle_enqueue0()) Skipped 3 previous similar messages LustreError: 11-0: lustre-MDT0000-mdc-ffff880274528008: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-MDT0000-mdc-ffff880274528008: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-MDT0000-mdc-ffff880274528008: This client was evicted by lustre-MDT0000; in progress operations using this service will fail. LustreError: 19384:0:(file.c:5042:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000403:0x8c:0x0] error: rc = -5 LustreError: 19384:0:(file.c:5042:ll_inode_revalidate_fini()) Skipped 27 previous similar messages LustreError: 19375:0:(mdc_request.c:1436:mdc_read_page()) lustre-MDT0000-mdc-ffff880274528008: [0x200000400:0x8:0x0] lock enqueue fails: rc = -108 LustreError: 19375:0:(mdc_request.c:1436:mdc_read_page()) Skipped 3 previous similar messages Lustre: dir [0x200000403:0x8c:0x0] stripe 0 readdir failed: -108, directory is partially accessed! Lustre: Skipped 16 previous similar messages LustreError: 19375:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff880274528008: inode [0x200000403:0x5e:0x0] mdc close failed: rc = -108 LustreError: 19375:0:(file.c:234:ll_close_inode_openhandle()) Skipped 3 previous similar messages LustreError: 18530:0:(vvp_io.c:1793:vvp_io_init()) lustre: refresh file layout [0x200000403:0xee:0x0] error -108. LustreError: 18530:0:(vvp_io.c:1793:vvp_io_init()) Skipped 1 previous similar message LustreError: Skipped 7 previous similar messages LustreError: 18595:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -108 Lustre: 6710:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.100@tcp:/lustre/fid: [0x200000404:0x101:0x0]/ may get corrupted (rc -5) Lustre: lustre-OST0000-osc-ffff880223af37e8: Connection restored to 192.168.123.100@tcp (at 0@lo) LustreError: 19754:0:(file.c:5042:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000007:0x1:0x0] error: rc = -108 LustreError: 19754:0:(file.c:5042:ll_inode_revalidate_fini()) Skipped 261 previous similar messages LustreError: 19747:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff880274528008: inode [0x200000403:0x76:0x0] mdc close failed: rc = -108 LustreError: 19747:0:(file.c:234:ll_close_inode_openhandle()) Skipped 14 previous similar messages LustreError: 19752:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-MDT0000-mdc-ffff880223af37e8: namespace resource [0x200000007:0x1:0x0].0x0 (ffff880094cded40) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 19689:0:(mdc_request.c:1436:mdc_read_page()) lustre-MDT0000-mdc-ffff880274528008: [0x200000403:0x1:0x0] lock enqueue fails: rc = -108 LustreError: 19689:0:(mdc_request.c:1436:mdc_read_page()) Skipped 12 previous similar messages Lustre: 15228:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 507 < left 1400, rollback = 2 Lustre: 15228:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 49 previous similar messages ------------[ cut here ]------------ WARNING: CPU: 3 PID: 15270 at /home/green/git/lustre-release/ldiskfs/ext4_jbd2.c:266 __ldiskfs_handle_dirty_metadata+0x1d2/0x230 [ldiskfs] Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) dm_flakey dm_mod loop zfs(PO) zunicode(PO) zzstd(O) zlua(O) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) jbd2 mbcache crc_t10dif crct10dif_generic sb_edac edac_core iosf_mbi crc32_pclmul ghash_clmulni_intel aesni_intel lrw gf128mul glue_helper ablk_helper cryptd virtio_console virtio_balloon pcspkr i2c_piix4 ip_tables rpcsec_gss_krb5 drm_kms_helper ata_generic ttm pata_acpi crct10dif_pclmul crct10dif_common drm ata_piix crc32c_intel drm_panel_orientation_quirks virtio_blk serio_raw libata i2c_core floppy [last unloaded: libcfs] CPU: 3 PID: 15270 Comm: mdt01_011 Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011 Call Trace: [<ffffffff817d9418>] dump_stack+0x19/0x1b [<ffffffff8108d538>] __warn+0xd8/0x100 [<ffffffff8108d67d>] warn_slowpath_null+0x1d/0x20 [<ffffffffa0d85262>] __ldiskfs_handle_dirty_metadata+0x1d2/0x230 [ldiskfs] [<ffffffffa0d66b10>] ldiskfs_do_update_inode+0x4a0/0x830 [ldiskfs] [<ffffffffa0b3c2d2>] ? jbd2_journal_get_write_access+0x32/0x40 [jbd2] [<ffffffffa0d68538>] ldiskfs_mark_iloc_dirty+0x58/0x80 [ldiskfs] [<ffffffffa0d6c410>] ? ldiskfs_dirty_inode+0x40/0x60 [ldiskfs] [<ffffffffa0d68686>] ldiskfs_mark_inode_dirty+0x86/0x2b0 [ldiskfs] [<ffffffffa0d6c3fa>] ? ldiskfs_dirty_inode+0x2a/0x60 [ldiskfs] [<ffffffffa0d6c410>] ldiskfs_dirty_inode+0x40/0x60 [ldiskfs] [<ffffffffa0cb10f7>] osd_write+0x747/0xcb0 [osd_ldiskfs] [<ffffffffa04ea6f3>] dt_record_write+0x33/0x120 [obdclass] [<ffffffffa04a96b2>] llog_osd_write_rec+0x1052/0x19a0 [obdclass] [<ffffffffa0c838bb>] ? osd_read_unlock+0x5b/0x90 [osd_ldiskfs] [<ffffffffa0496760>] llog_write_rec+0x290/0x590 [obdclass] [<ffffffffa049bb08>] llog_cat_new_log+0x6a8/0xe50 [obdclass] [<ffffffffa049c4ca>] llog_cat_add_rec+0x21a/0x950 [obdclass] [<ffffffffa04935ff>] llog_add+0x17f/0x1f0 [obdclass] [<ffffffffa04c4a39>] ? lprocfs_counter_add+0xf9/0x160 [obdclass] [<ffffffffa082d60d>] sub_updates_write+0xb6b/0xe32 [ptlrpc] [<ffffffffa07fd892>] top_trans_stop+0x7c2/0xf60 [ptlrpc] [<ffffffffa0f93f3c>] lod_trans_stop+0x25c/0x340 [lod] [<ffffffffa0e68408>] mdd_trans_stop+0x28/0x16e [mdd] [<ffffffffa0e49b2f>] mdd_migrate_object+0xf1f/0x1ba0 [mdd] [<ffffffffa04e9375>] ? lu_object_find_at+0x285/0xb10 [obdclass] [<ffffffffa0e4ab86>] mdd_migrate+0x3d6/0x8e0 [mdd] [<ffffffffa0f1c0d8>] mdo_migrate+0x4c/0x4e [mdt] [<ffffffffa0edbe53>] mdt_reint_migrate+0x1023/0x11d0 [mdt] [<ffffffffa0edc087>] mdt_reint_rec+0x87/0x240 [mdt] [<ffffffffa0eb155c>] mdt_reint_internal+0x76c/0xb40 [mdt] [<ffffffffa0ebe657>] mdt_reint+0x67/0x150 [mdt] [<ffffffffa07ec26e>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<ffffffffa03b745e>] ? libcfs_nid2str_r+0xfe/0x130 [lnet] [<ffffffffa0797990>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<ffffffffa0799529>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e32be>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffffa0798950>] ? ptlrpc_wait_event+0x620/0x620 [ptlrpc] [<ffffffff810ba0f4>] kthread+0xe4/0xf0 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 ---[ end trace 5192ad504da6a0a7 ]--- LDISKFS-fs: ldiskfs_do_update_inode:4844: aborting transaction: error 28 in __ldiskfs_handle_dirty_metadata LDISKFS: jbd2_journal_dirty_metadata failed: handle type 0 started at line 1972, credits 512/0, errcode -28 LDISKFS-fs error (device dm-1) in ldiskfs_do_update_inode:4862: error 28 Aborting journal on device dm-1-8. LDISKFS-fs (dm-1): Remounting filesystem read-only LDISKFS-fs error (device dm-1) in ldiskfs_dirty_inode:5467: error 28 LDISKFS-fs warning (device dm-1): ldiskfs_mb_new_blocks:5106: Updating bitmap error: [err -30] [pa ffff88029b1033b0] [phy 52096] [logic 52096] [len 16] [free 16] [error 1] [inode 0] LustreError: 15270:0:(osd_io.c:2103:osd_ldiskfs_write_record()) lustre-MDT0001: error reading offset 0 (block 0, size 32768, offs 0), credits 0/0: rc = -30 LustreError: 15270:0:(llog_cat.c:602:llog_cat_add_rec()) llog_write_rec -30: lh=ffff88026035bef8 LustreError: 15270:0:(update_trans.c:985:top_trans_stop()) lustre-MDT0001-osd: write updates failed: rc = -30 LustreError: 15270:0:(osd_io.c:2110:osd_ldiskfs_write_record()) journal_get_write_access() returned error -30 LustreError: 15270:0:(tgt_lastrcvd.c:1253:tgt_add_reply_data()) lustre-MDT0001: can't update reply_data file: rc = -30 LustreError: 15270:0:(osd_handler.c:2075:osd_trans_stop()) lustre-MDT0001: failed in transaction hook: rc = -30 LDISKFS-fs error (device dm-1) in osd_trans_stop:2082: error 28 LustreError: 10178:0:(osd_handler.c:1779:osd_trans_commit_cb()) transaction @0xffff8802a6a48948 commit error: 2 LustreError: 15270:0:(osd_handler.c:2085:osd_trans_stop()) lustre-MDT0001: failed to stop transaction: rc = -28 LustreError: 15270:0:(update_trans.c:1011:top_trans_stop()) lustre-MDT0001-osd: stop trans failed: rc = -30 | Externally reported by onyx-68 boilpot email |
racer test 1: racer on clients: centos-20.localnet DURATION=2700 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa1073e52>] ll_migrate+0x9b2/0xec0 [lustre] PGD 80000002528c9067 PUD 28b071067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: loop zfs(PO) zunicode(PO) zzstd(O) zlua(O) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) jbd2 mbcache lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) dm_flakey dm_mod libcfs(OE) crc_t10dif crct10dif_generic sb_edac edac_core iosf_mbi crc32_pclmul ghash_clmulni_intel aesni_intel lrw gf128mul glue_helper ablk_helper cryptd virtio_balloon pcspkr i2c_piix4 virtio_console ip_tables rpcsec_gss_krb5 ata_generic drm_kms_helper ttm pata_acpi crct10dif_pclmul drm crct10dif_common crc32c_intel drm_panel_orientation_quirks ata_piix virtio_blk serio_raw i2c_core libata floppy CPU: 4 PID: 12081 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011 task: ffff88025c4e24f0 ti: ffff880075d0c000 task.ti: ffff880075d0c000 RIP: 0010:[<ffffffffa1073e52>] [<ffffffffa1073e52>] ll_migrate+0x9b2/0xec0 [lustre] RSP: 0018:ffff880075d0fbc8 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff880285bdb528 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff880320fe80e8 RDI: ffff880320fe80c8 RBP: ffff880075d0fc48 R08: ffff8802508aaf58 R09: ffff880075d0fa60 R10: 0000000000000003 R11: ffff88008964ce78 R12: 0000000000000000 R13: ffff880272833508 R14: ffff8802443bb898 R15: 0000000000000030 FS: 00007fd3e83dc740(0000) GS:ffff880331b00000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 CR2: 0000000000000008 CR3: 0000000283ea0000 CR4: 00000000001607e0 Call Trace: [<ffffffff81242213>] ? __check_object_size+0x1c3/0x220 [<ffffffffa1061561>] ll_dir_ioctl+0x5d01/0x6ed0 [lustre] [<ffffffff811b515b>] ? unlock_page+0x2b/0x30 [<ffffffff811e5e2e>] ? do_read_fault.isra.64+0x12e/0x170 [<ffffffff811ea1cf>] ? handle_pte_fault+0x2ef/0xca0 [<ffffffff811ecf6d>] ? __handle_mm_fault+0x31d/0x720 [<ffffffff8125b3fd>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff81245a38>] ? vfs_write+0x178/0x230 [<ffffffff8125b751>] SyS_ioctl+0xa1/0xc0 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 | LustreError: 11588:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0002-osp-MDT0000: fail to cancel 1 llog-records: rc = -116 LustreError: 11588:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0002-osp-MDT0000: fail to cancel 1 of 1 llog-records: rc = -116 cp (17084) used greatest stack depth: 10080 bytes left Lustre: 11233:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 507 < left 849, rollback = 2 Lustre: 11233:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 3/12/2, destroy: 0/0/0 Lustre: 11233:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 9/849/0 Lustre: 11233:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 5/136/0, punch: 0/0/0, quota 6/6/0 Lustre: 11233:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 10/203/3, delete: 0/0/0 Lustre: 11233:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 4/4/0, ref_del: 0/0/0 LustreError: 12441:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0000-osp-MDT0002: fail to cancel 1 llog-records: rc = -2 LustreError: 12441:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0000-osp-MDT0002: fail to cancel 1 of 1 llog-records: rc = -2 Lustre: 17509:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 499 < left 866, rollback = 2 Lustre: 17509:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 17 previous similar messages Lustre: 17509:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 3/12/8, destroy: 1/4/0 Lustre: 17509:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 17 previous similar messages Lustre: 17509:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 9/866/0 Lustre: 17509:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 17 previous similar messages Lustre: 17509:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 5/136/0, punch: 0/0/0, quota 6/6/0 Lustre: 17509:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 17 previous similar messages Lustre: 17509:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 12/215/4, delete: 2/5/1 Lustre: 17509:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 17 previous similar messages Lustre: 17509:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 5/5/0, ref_del: 2/2/0 Lustre: 17509:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 17 previous similar messages 10[17529]: segfault at 8 ip 00007f95546c17e8 sp 00007ffdf72cf540 error 4 in ld-2.17.so[7f95546b6000+22000] Lustre: 17389:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 502 < left 671, rollback = 2 Lustre: 17389:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 8 previous similar messages Lustre: 17389:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 3/12/6, destroy: 0/0/0 Lustre: 17389:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 8 previous similar messages Lustre: 17389:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 7/671/0 Lustre: 17389:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 8 previous similar messages Lustre: 17389:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 5/136/0, punch: 0/0/0, quota 6/6/0 Lustre: 17389:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 8 previous similar messages Lustre: 17389:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 10/203/4, delete: 0/0/0 Lustre: 17389:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 8 previous similar messages Lustre: 17329:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 5/5/0, ref_del: 2/2/0 Lustre: 17329:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 16 previous similar messages LustreError: 17670:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0002: '7' migration was interrupted, run 'lfs migrate -m 0 -c 1 -H crush 7' to finish migration. Lustre: 11226:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 512 < left 1421, rollback = 7 Lustre: 11226:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 11226:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 22 previous similar messages Lustre: 11226:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 1/89/0 Lustre: 11226:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 22 previous similar messages Lustre: 11226:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1412/1421/0, punch: 0/0/0, quota 3/3/0 Lustre: 11226:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 22 previous similar messages Lustre: 11226:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 11226:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 22 previous similar messages Lustre: 11226:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 11226:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 14 previous similar messages Lustre: 11226:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 511 < left 1420, rollback = 7 Lustre: 11226:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 304 previous similar messages Lustre: 11226:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 511 < left 1420, rollback = 7 Lustre: 11226:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 890 previous similar messages Lustre: 11233:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 506 < left 535, rollback = 2 Lustre: 11233:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 22 previous similar messages Lustre: 11233:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 3/12/2, destroy: 0/0/0 Lustre: 11233:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 1411 previous similar messages Lustre: 11233:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 3/3/0, xattr_set: 8/535/0 Lustre: 11233:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 1411 previous similar messages Lustre: 11233:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 5/136/0, punch: 0/0/0, quota 8/56/0 Lustre: 11233:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 1411 previous similar messages Lustre: 11233:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 12/215/4, delete: 1/1/0 Lustre: 11233:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 1411 previous similar messages Lustre: 11233:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 6/6/0, ref_del: 1/1/0 Lustre: 11233:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 1411 previous similar messages Lustre: 11250:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 512 < left 617, rollback = 7 Lustre: 11250:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 215 previous similar messages Lustre: dir [0x200000404:0x56:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: 11250:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 505 < left 866, rollback = 2 Lustre: 11250:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 3 previous similar messages Lustre: 11250:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 3/12/3, destroy: 1/4/0 Lustre: 11250:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 1128 previous similar messages Lustre: 11250:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 9/866/0 Lustre: 11250:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 1128 previous similar messages Lustre: 11250:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 5/136/0, punch: 0/0/0, quota 6/6/0 Lustre: 11250:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 1128 previous similar messages Lustre: 11250:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 12/215/3, delete: 2/5/1 Lustre: 11250:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 1128 previous similar messages Lustre: 11250:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 5/5/0, ref_del: 2/2/0 Lustre: 11250:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 1128 previous similar messages Lustre: dir [0x200000403:0x33:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: 17441:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 508 < left 22198, rollback = 7 Lustre: 17441:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 1124 previous similar messages ------------[ cut here ]------------ WARNING: CPU: 11 PID: 11282 at /home/green/git/lustre-release/ldiskfs/namei.c:3331 ldiskfs_orphan_add+0x11e/0x2a0 [ldiskfs] Modules linked in: loop zfs(PO) zunicode(PO) zzstd(O) zlua(O) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) jbd2 mbcache lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) dm_flakey dm_mod libcfs(OE) crc_t10dif crct10dif_generic sb_edac edac_core iosf_mbi crc32_pclmul ghash_clmulni_intel aesni_intel lrw gf128mul glue_helper ablk_helper cryptd virtio_balloon pcspkr i2c_piix4 virtio_console ip_tables rpcsec_gss_krb5 ata_generic drm_kms_helper ttm pata_acpi crct10dif_pclmul drm crct10dif_common crc32c_intel drm_panel_orientation_quirks ata_piix virtio_blk serio_raw i2c_core libata floppy CPU: 11 PID: 11282 Comm: mdt_rdpg05_000 Tainted: P OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011 Call Trace: [<ffffffff817d9418>] dump_stack+0x19/0x1b [<ffffffff8108d538>] __warn+0xd8/0x100 [<ffffffff8108d67d>] warn_slowpath_null+0x1d/0x20 [<ffffffffa0c0b4ae>] ldiskfs_orphan_add+0x11e/0x2a0 [ldiskfs] [<ffffffffa0bfc9eb>] ldiskfs_xattr_inode_orphan_add.constprop.17+0xbb/0x110 [ldiskfs] [<ffffffffa0bff417>] ldiskfs_xattr_delete_inode+0x77/0x3c0 [ldiskfs] [<ffffffffa0c1faa0>] ? ldiskfs_evict_inode+0x190/0x670 [ldiskfs] [<ffffffffa0c1fad9>] ldiskfs_evict_inode+0x1c9/0x670 [ldiskfs] [<ffffffff812649af>] evict+0xaf/0x180 [<ffffffff81264de5>] iput+0xf5/0x180 [<ffffffffa0cb4d08>] osd_object_delete+0x1f8/0x370 [osd_ldiskfs] [<ffffffffa04e359d>] lu_object_free.isra.26+0xcd/0x1d0 [obdclass] [<ffffffffa04e7cf1>] lu_object_put+0xa1/0x480 [obdclass] [<ffffffffa0e6caf4>] mdt_mfd_close+0x1f4/0x870 [mdt] [<ffffffffa0e72e14>] mdt_close_internal+0xb4/0x240 [mdt] [<ffffffffa0e73224>] mdt_close+0x284/0x850 [mdt] [<ffffffffa07ea2d4>] ? tgt_request_preprocess.isra.27+0xf4/0x850 [ptlrpc] [<ffffffffa07eb26e>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<ffffffffa031f45e>] ? libcfs_nid2str_r+0xfe/0x130 [lnet] [<ffffffffa0796990>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<ffffffffa0798529>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffffa0797950>] ? ptlrpc_wait_event+0x620/0x620 [ptlrpc] [<ffffffff810ba0f4>] kthread+0xe4/0xf0 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 ---[ end trace 147a0d2842d7da1d ]--- LustreError: 17242:0:(lustre_lmv.h:438:lmv_is_sane()) insane LMV: magic=0xcd40cd0 count=2 index=1 hash=crush:0x82000003 version=0 migrate offset=1 migrate hash=crush:3. Lustre: 4122:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff88007c723240 x1710093003540800/t0(0) o10->lustre-OST0002-osc-ffff88029ab153d8@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630871958 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'truncate.0' sched: RT throttling activated Lustre: 11251:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 505 < left 671, rollback = 2 Lustre: 11251:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 36 previous similar messages Lustre: 11251:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 3/12/5, destroy: 0/0/0 Lustre: 11251:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 2391 previous similar messages Lustre: 11251:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 7/671/0 Lustre: 11251:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 2391 previous similar messages Lustre: 11251:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 5/136/0, punch: 0/0/0, quota 8/56/0 Lustre: 11251:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 2391 previous similar messages Lustre: 11251:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 10/203/2, delete: 0/0/0 Lustre: 11251:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 2391 previous similar messages Lustre: 11251:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 4/4/0, ref_del: 0/0/0 Lustre: 11251:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 2391 previous similar messages LustreError: 17726:0:(lustre_lmv.h:438:lmv_is_sane()) insane LMV: magic=0xcd40cd0 count=2 index=1 hash=crush:0x82000003 version=0 migrate offset=1 migrate hash=crush:3. Lustre: 17260:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 510 < left 26154, rollback = 7 Lustre: 17260:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 2354 previous similar messages LustreError: 16857:0:(lustre_lmv.h:438:lmv_is_sane()) insane LMV: magic=0xcd40cd0 count=2 index=1 hash=crush:0x82000003 version=0 migrate offset=1 migrate hash=crush:3. LustreError: 17300:0:(lustre_lmv.h:438:lmv_is_sane()) insane LMV: magic=0xcd40cd0 count=2 index=1 hash=crush:0x82000003 version=0 migrate offset=1 migrate hash=crush:3. Lustre: 4138:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff880266f5ac00 x1710093005069376/t0(0) o10->lustre-OST0003-osc-ffff880299a64138@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630872012 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'cp.0' LustreError: 17565:0:(lustre_lmv.h:438:lmv_is_sane()) insane LMV: magic=0xcd40cd0 count=2 index=1 hash=crush:0x82000003 version=0 migrate offset=1 migrate hash=crush:3. LustreError: 17144:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000404:0x33:0x0]: rc = -2 LustreError: 11251:0:(lustre_lmv.h:438:lmv_is_sane()) insane LMV: magic=0xcd40cd0 count=2 index=1 hash=crush:0x82000003 version=0 migrate offset=1 migrate hash=crush:3. LustreError: 11253:0:(lustre_lmv.h:438:lmv_is_sane()) insane LMV: magic=0xcd40cd0 count=2 index=1 hash=crush:0x82000003 version=0 migrate offset=1 migrate hash=crush:3. LustreError: 17485:0:(lustre_lmv.h:438:lmv_is_sane()) insane LMV: magic=0xcd40cd0 count=2 index=1 hash=crush:0x82000003 version=0 migrate offset=1 migrate hash=crush:3. Lustre: 17329:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 512 < left 1439, rollback = 7 Lustre: 17329:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 11810 previous similar messages LustreError: 17485:0:(lustre_lmv.h:438:lmv_is_sane()) insane LMV: magic=0xcd40cd0 count=2 index=1 hash=crush:0x82000003 version=0 migrate offset=1 migrate hash=crush:3. Lustre: 20598:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 778/3112/0, destroy: 0/0/0 Lustre: 20598:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 22656 previous similar messages Lustre: 20598:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/1, xattr_set: 780/15/0 Lustre: 20598:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 22656 previous similar messages Lustre: 20598:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 3891/33464/29, punch: 0/0/0, quota 10/106/3 Lustre: 20598:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 22656 previous similar messages Lustre: 20598:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 778/13226/0, delete: 0/0/0 Lustre: 20598:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 22656 previous similar messages Lustre: 22272:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 22272:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 23282 previous similar messages Lustre: 14736:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 8: before 511 < left 13258, rollback = 9 Lustre: 14736:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 56 previous similar messages Lustre: mdt05_003: service thread pid 16404 was inactive for 62.243 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 16404, comm: mdt05_003 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] mdt_object_local_lock+0x527/0xb90 [mdt] [<0>] mdt_object_lock_internal+0x70/0x390 [mdt] [<0>] mdt_reint_object_lock+0x2c/0x60 [mdt] [<0>] mdt_reint_striped_lock+0x89/0x5b0 [mdt] [<0>] mdt_attr_set+0x9c/0x800 [mdt] [<0>] mdt_reint_setattr+0x77a/0x1020 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x76c/0xb40 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: dir [0x280000404:0xa8:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: mdt02_010: service thread pid 20776 was inactive for 62.060 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 20776, comm: mdt02_010 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] mdt_object_local_lock+0x527/0xb90 [mdt] [<0>] mdt_object_lock_internal+0x70/0x390 [mdt] [<0>] mdt_getattr_name_lock+0x12c/0x2590 [mdt] [<0>] mdt_intent_getattr+0x2d5/0x4b0 [mdt] [<0>] mdt_intent_opc+0x1dd/0xc10 [mdt] [<0>] mdt_intent_policy+0x1a1/0x360 [mdt] [<0>] ldlm_lock_enqueue+0x3c2/0xb40 [ptlrpc] [<0>] ldlm_handle_enqueue0+0xaa6/0x16d0 [ptlrpc] [<0>] tgt_enqueue+0x64/0x240 [ptlrpc] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: 11231:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 504 < left 866, rollback = 2 Lustre: 11231:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 29 previous similar messages LustreError: 11081:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0002_UUID lock: ffff88027af89a80/0xab3c18598041e449 lrc: 3/0,0 mode: PW/PW res: [0x340000401:0x7:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000000020020 nid: 0@lo remote: 0xab3c18598041e385 expref: 3170 pid: 12844 timeout: 365 lvb_type: 0 LustreError: 11079:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff880251282c00 x1710093008192576/t0(0) o105->lustre-OST0002@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 11-0: lustre-OST0002-osc-ffff880299a64138: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-OST0002-osc-ffff880299a64138: Connection to lustre-OST0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0002-osc-ffff880299a64138: This client was evicted by lustre-OST0002; in progress operations using this service will fail. LustreError: 11072:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630872043 with bad export cookie 12338763851771075597 LustreError: 12837:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff880263dcc500 x1710093008209920/t0(0) o106->lustre-OST0002@0@lo:15/16 lens 328/280 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 12837:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 13 previous similar messages Lustre: 12872:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 7s req@ffff880267d88680 x1710093008028992/t0(0) o13->lustre-MDT0000-mdtlov_UUID@0@lo:0/0 lens 224/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'osp-pre-0-0.0' Lustre: ost_create: This server is not able to keep up with request traffic (cpu-bound). Lustre: 12874:0:(service.c:1612:ptlrpc_at_check_timed()) earlyQ=1 reqQ=0 recA=0, svcEst=1, delay=0ms Lustre: 12874:0:(service.c:1378:ptlrpc_at_send_early_reply()) @@@ Already past deadline (-1s), not sending early reply. Consider increasing at_early_margin (5)? req@ffff880267d88680 x1710093008028992/t0(0) o13->lustre-MDT0000-mdtlov_UUID@0@lo:320/0 lens 224/0 e 0 to 0 dl 1630872045 ref 2 fl New:/0/ffffffff rc 0/-1 job:'osp-pre-0-0.0' LustreError: 12874:0:(service.c:2290:ptlrpc_server_handle_request()) @@@ Dropping timed-out request from 12345-0@lo: deadline 6/1s ago req@ffff880267d88680 x1710093008028992/t0(0) o13->lustre-MDT0000-mdtlov_UUID@0@lo:320/0 lens 224/0 e 0 to 0 dl 1630872045 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'osp-pre-0-0.0' Lustre: 12874:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (6/1s); client may timeout req@ffff880267d88680 x1710093008028992/t0(0) o13->lustre-MDT0000-mdtlov_UUID@0@lo:320/0 lens 224/0 e 0 to 0 dl 1630872045 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'osp-pre-0-0.0' LustreError: 11081:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0003_UUID lock: ffff880080602980/0xab3c18598041dc69 lrc: 3/0,0 mode: PW/PW res: [0x380000401:0x5:0x0].0x0 rrc: 4 type: EXT [0->18446744073709551615] (req 0->2621439) gid 0 flags: 0x60000400020020 nid: 0@lo remote: 0xab3c18598041dc4d expref: 3157 pid: 16431 timeout: 369 lvb_type: 0 LustreError: 20678:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88026da8d180 x1710093008262144/t0(0) o104->lustre-OST0003@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 20678:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 3 previous similar messages Lustre: lustre-OST0003-osc-ffff880299a64138: Connection to lustre-OST0003 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0003-osc-ffff880299a64138: This client was evicted by lustre-OST0003; in progress operations using this service will fail. Lustre: 16183:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 508 < left 60726, rollback = 7 Lustre: 16183:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 12936 previous similar messages Lustre: 4122:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.120@tcp:/lustre/fid: [0x200000404:0x101:0x0]/ may get corrupted (rc -108) LustreError: 11081:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0000_UUID lock: ffff880082391e40/0xab3c185980428ea3 lrc: 3/0,0 mode: PR/PR res: [0x200000403:0x6d:0x0].0x0 bits 0x12/0x0 rrc: 12 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0xab3c185980428d76 expref: 99 pid: 17509 timeout: 372 lvb_type: 0 LustreError: 17820:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630872051 with bad export cookie 12338763851771075289 LustreError: 11-0: lustre-MDT0000-mdc-ffff88029ab153d8: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-MDT0000-mdc-ffff88029ab153d8: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-MDT0000-mdc-ffff88029ab153d8: This client was evicted by lustre-MDT0000; in progress operations using this service will fail. LustreError: 23405:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -5 LustreError: 20135:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -5 LustreError: 20135:0:(llite_lib.c:1836:ll_md_setattr()) Skipped 1 previous similar message LustreError: 19920:0:(file.c:5042:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000403:0x6d:0x0] error: rc = -108 LustreError: 19920:0:(file.c:5042:ll_inode_revalidate_fini()) Skipped 29 previous similar messages LustreError: 23926:0:(file.c:5042:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000007:0x1:0x0] error: rc = -108 LustreError: 23926:0:(file.c:5042:ll_inode_revalidate_fini()) Skipped 6 previous similar messages LustreError: 23856:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff88029ab153d8: inode [0x200000404:0xb:0x0] mdc close failed: rc = -108 Lustre: lustre-MDT0000-mdc-ffff88029ab153d8: Connection restored to 192.168.123.120@tcp (at 0@lo) LustreError: 23774:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0003-osc-ffff880299a64138: namespace resource [0x380000401:0x8f:0x0].0x0 (ffff880270b5e5c0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 23774:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0003-osc-ffff880299a64138: namespace resource [0x61:0x0:0x0].0x0 (ffff880254b06d40) refcount nonzero (1) after lock cleanup; forcing cleanup. Lustre: lustre-OST0003-osc-ffff880299a64138: Connection restored to 192.168.123.120@tcp (at 0@lo) LustreError: 23629:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0002-osc-ffff880299a64138: namespace resource [0x340000401:0xff:0x0].0x0 (ffff880252532fc0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 23629:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 555 previous similar messages Lustre: lustre-OST0002-osc-ffff880299a64138: Connection restored to 192.168.123.120@tcp (at 0@lo) Lustre: 21301:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 608/2432/0, destroy: 1/4/1 Lustre: 21301:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 2222 previous similar messages Lustre: 21301:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 608/0/0 Lustre: 21301:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 2222 previous similar messages Lustre: 21301:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 3041/26154/0, punch: 0/0/0, quota 10/74/0 Lustre: 21301:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 2222 previous similar messages Lustre: 21301:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 608/10336/0, delete: 2/5/1 Lustre: 21301:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 2222 previous similar messages Lustre: 11243:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 4/4/0, ref_del: 0/0/0 Lustre: 11243:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 1593 previous similar messages Lustre: 11280:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 6s req@ffff8800a9975d78 x1710093009295424/t0(0) o37->7eadd345-2683-4b20-82d7-c180a3661ff1@0@lo:0/0 lens 448/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'ls.0' Lustre: 4133:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1630872039/real 1630872039] req@ffff880267d8f0c0 x1710093008028992/t0(0) o13->lustre-OST0000-osc-MDT0000@0@lo:7/4 lens 224/368 e 0 to 1 dl 1630872068 ref 1 fl Rpc:XQr/0/ffffffff rc 0/-1 job:'osp-pre-0-0.0' Lustre: lustre-OST0000-osc-MDT0000: Connection to lustre-OST0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: lustre-OST0000: Client lustre-MDT0000-mdtlov_UUID (at 0@lo) reconnecting Lustre: lustre-OST0000-osc-MDT0000: Connection restored to 192.168.123.120@tcp (at 0@lo) LustreError: 11081:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0000_UUID lock: ffff880266588400/0xab3c1859804831de lrc: 3/0,0 mode: PW/PW res: [0x2c0000400:0x97:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000000020020 nid: 0@lo remote: 0xab3c18598048319f expref: 5924 pid: 12829 timeout: 422 lvb_type: 0 LustreError: 19778:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88024f0a3ec0 x1710093010110976/t0(0) o105->lustre-OST0000@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 19778:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 1 previous similar message LustreError: 11-0: lustre-OST0000-osc-ffff88029ab153d8: operation ost_getattr to node 0@lo failed: rc = -107 Lustre: lustre-OST0000-osc-ffff88029ab153d8: Connection to lustre-OST0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 11081:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0002_UUID lock: ffff8800824dbc40/0xab3c18598048c5e8 lrc: 3/0,0 mode: PW/PW res: [0x340000400:0x14e:0x0].0x0 rrc: 4 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000480000020 nid: 0@lo remote: 0xab3c18598048c5a9 expref: 6004 pid: 19545 timeout: 432 lvb_type: 0 LustreError: 11-0: lustre-OST0002-osc-ffff88029ab153d8: operation ost_getattr to node 0@lo failed: rc = -107 LustreError: Skipped 1 previous similar message Lustre: lustre-OST0002-osc-ffff88029ab153d8: Connection to lustre-OST0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: ll_ost05_017: service thread pid 22315 was inactive for 62.022 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Lustre: ll_ost05_006: service thread pid 20405 was inactive for 62.169 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Pid: 22315, comm: ll_ost05_017 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] tgt_extent_lock+0xea/0x2a0 [ptlrpc] [<0>] ofd_getattr_hdl+0x365/0x720 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: ll_ost05_023: service thread pid 23788 was inactive for 62.126 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 8 previous similar messages Lustre: ll_ost05_031: service thread pid 23823 was inactive for 62.079 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 3 previous similar messages Lustre: ll_ost05_046: service thread pid 24045 was inactive for 62.012 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 4 previous similar messages Lustre: ll_ost05_049: service thread pid 24236 was inactive for 62.064 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 2 previous similar messages Lustre: mdt07_012: service thread pid 21335 was inactive for 62.204 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 9 previous similar messages Lustre: 4133:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1630872049/real 1630872049] req@ffff880285bef700 x1710093008292032/t0(0) o1->lustre-OST0000-osc-ffff88029ab153d8@0@lo:28/4 lens 440/432 e 1 to 1 dl 1630872138 ref 1 fl Rpc:XQr/0/ffffffff rc 0/-1 job:'lfs.0' Lustre: 4133:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1630872050/real 1630872050] req@ffff88026a0d8040 x1710093008299200/t0(0) o1->lustre-OST0000-osc-ffff88029ab153d8@0@lo:28/4 lens 440/432 e 1 to 1 dl 1630872139 ref 1 fl Rpc:XQr/0/ffffffff rc 0/-1 job:'lfs.0' Lustre: 4133:0:(client.c:2285:ptlrpc_expire_one_request()) Skipped 3 previous similar messages Lustre: 4133:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1630872051/real 1630872051] req@ffff88026339e440 x1710093008337536/t0(0) o1->lustre-OST0000-osc-ffff88029ab153d8@0@lo:28/4 lens 440/432 e 1 to 1 dl 1630872141 ref 1 fl Rpc:XQr/0/ffffffff rc 0/-1 job:'lfs.0' Lustre: 4133:0:(client.c:2285:ptlrpc_expire_one_request()) Skipped 156 previous similar messages Lustre: 4133:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1630872052/real 1630872052] req@ffff880252461f80 x1710093008396480/t0(0) o1->lustre-OST0000-osc-ffff88029ab153d8@0@lo:28/4 lens 440/432 e 1 to 1 dl 1630872145 ref 1 fl Rpc:XQr/0/ffffffff rc 0/-1 job:'lfs.0' Lustre: 4133:0:(client.c:2285:ptlrpc_expire_one_request()) Skipped 431 previous similar messages LustreError: 11081:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0001_UUID lock: ffff880269882d40/0xab3c18598048b25b lrc: 3/0,0 mode: PW/PW res: [0x300000400:0x174:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000480000020 nid: 0@lo remote: 0xab3c18598048b246 expref: 5942 pid: 20827 timeout: 470 lvb_type: 0 LustreError: 23791:0:(service.c:2290:ptlrpc_server_handle_request()) @@@ Dropping timed-out request from 12345-0@lo: deadline 68/32s ago req@ffff88007add1f80 x1710093008291904/t0(0) o1->7eadd345-2683-4b20-82d7-c180a3661ff1@0@lo:392/0 lens 440/0 e 1 to 0 dl 1630872117 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'lfs.0' Lustre: 12844:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (68/32s); client may timeout req@ffff88007add0680 x1710093008292032/t0(0) o1->7eadd345-2683-4b20-82d7-c180a3661ff1@0@lo:392/0 lens 440/0 e 1 to 0 dl 1630872117 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'lfs.0' LustreError: 24408:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88025bfacb40 x1710093010150336/t0(0) o104->lustre-OST0001@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 24408:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 1 previous similar message LustreError: 23791:0:(service.c:2290:ptlrpc_server_handle_request()) Skipped 170 previous similar messages LustreError: 11-0: lustre-OST0003-osc-ffff88029ab153d8: operation ost_getattr to node 0@lo failed: rc = -107 LustreError: Skipped 19 previous similar messages Lustre: lustre-OST0003-osc-ffff88029ab153d8: Connection to lustre-OST0003 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 20414:0:(service.c:2290:ptlrpc_server_handle_request()) @@@ Dropping timed-out request from 12345-0@lo: deadline 43/30s ago req@ffff880262cc5180 x1710093009444864/t0(0) o101->7eadd345-2683-4b20-82d7-c180a3661ff1@0@lo:396/0 lens 328/0 e 0 to 0 dl 1630872121 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'ln.0' LustreError: 20414:0:(service.c:2290:ptlrpc_server_handle_request()) Skipped 507 previous similar messages Lustre: 20414:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (43/30s); client may timeout req@ffff880262cc5180 x1710093009444864/t0(0) o101->7eadd345-2683-4b20-82d7-c180a3661ff1@0@lo:396/0 lens 328/0 e 0 to 0 dl 1630872121 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'ln.0' Lustre: 20414:0:(service.c:2328:ptlrpc_server_handle_request()) Skipped 678 previous similar messages LustreError: 167-0: lustre-OST0003-osc-ffff88029ab153d8: This client was evicted by lustre-OST0003; in progress operations using this service will fail. Lustre: 22255:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 505 < left 1166, rollback = 7 Lustre: 22255:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 2436 previous similar messages Lustre: 16274:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1630872061/real 1630872061] req@ffff880267c01f80 x1710093008841280/t0(0) o101->lustre-OST0000-osc-ffff88029ab153d8@0@lo:28/4 lens 328/400 e 1 to 1 dl 1630872154 ref 2 fl Rpc:XQr/0/ffffffff rc 0/-1 job:'dir_create.sh.0' Lustre: 16274:0:(client.c:2285:ptlrpc_expire_one_request()) Skipped 68 previous similar messages LustreError: 24709:0:(import.c:354:ptlrpc_invalidate_import()) lustre-OST0002_UUID: timeout waiting for callback (300 != 0) LustreError: 24709:0:(import.c:382:ptlrpc_invalidate_import()) @@@ still on delayed list req@ffff88026339cb40 x1710093008337664/t0(0) o1->lustre-OST0002-osc-ffff88029ab153d8@0@lo:28/4 lens 440/432 e 1 to 1 dl 1630872142 ref 1 fl Rpc:EXQU/0/ffffffff rc -5/-1 job:'lfs.0' LustreError: 24709:0:(import.c:388:ptlrpc_invalidate_import()) lustre-OST0002_UUID: Unregistering RPCs found (0). Network is sluggish? Waiting for them to error out. Lustre: mdt01_000: service thread pid 11231 was inactive for 64.174 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 23 previous similar messages LustreError: 13734:0:(osp_precreate.c:676:osp_precreate_send()) lustre-OST0000-osc-MDT0000: can't precreate: rc = -107 LustreError: 24712:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0000-osc-ffff88029ab153d8: namespace resource [0x2c0000400:0x170:0x0].0x0 (ffff880263705bc0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 24712:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 659 previous similar messages LustreError: 167-0: lustre-MDT0001-mdc-ffff88029ab153d8: This client was evicted by lustre-MDT0001; in progress operations using this service will fail. LustreError: Skipped 3 previous similar messages LustreError: 23851:0:(file.c:5042:ll_inode_revalidate_fini()) lustre: revalidate FID [0x240000404:0x51:0x0] error: rc = -5 LustreError: 4127:0:(mdc_locks.c:1386:mdc_intent_getattr_async_interpret()) lustre-MDT0001-mdc-ffff88029ab153d8: ldlm_cli_enqueue_fini() failed: rc = -5 LustreError: 23851:0:(file.c:5042:ll_inode_revalidate_fini()) Skipped 29 previous similar messages INFO: task mv:19365 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mv D ffff880290105c40 10640 19365 15942 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 LustreError: 20888:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff88029ab153d8: inode [0x240000404:0x51:0x0] mdc close failed: rc = -108 LustreError: 20888:0:(file.c:234:ll_close_inode_openhandle()) Skipped 8 previous similar messages [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff81251fe1>] lock_rename+0x31/0xe0 [<ffffffff8125862f>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed432>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817e8a76>] ? trace_do_page_fault+0x56/0x170 [<ffffffff8125975e>] SyS_renameat2+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task mv:20404 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mv D ffff88028a7ac9d0 12128 20404 16266 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff81251fe1>] lock_rename+0x31/0xe0 [<ffffffff8125862f>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed432>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817e8a76>] ? trace_do_page_fault+0x56/0x170 [<ffffffff8125975e>] SyS_renameat2+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task lfs:20786 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. lfs D ffff880288159900 11632 20786 16423 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff812550a6>] do_last+0x296/0x1280 [<ffffffff812546de>] ? link_path_walk+0x27e/0x8c0 [<ffffffff812570ad>] path_openat+0xcd/0x5b0 [<ffffffff81258dfd>] do_filp_open+0x4d/0xb0 [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e326e>] ? _raw_spin_unlock+0xe/0x20 [<ffffffff81267893>] ? __alloc_fd+0xc3/0x170 [<ffffffff812447e4>] do_sys_open+0x124/0x220 [<ffffffff812448fe>] SyS_open+0x1e/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task lfs:20834 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. lfs D ffff880288a70010 11616 20834 16001 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff812550a6>] do_last+0x296/0x1280 [<ffffffff812546de>] ? link_path_walk+0x27e/0x8c0 [<ffffffff812570ad>] path_openat+0xcd/0x5b0 [<ffffffff81258dfd>] do_filp_open+0x4d/0xb0 [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e326e>] ? _raw_spin_unlock+0xe/0x20 [<ffffffff81267893>] ? __alloc_fd+0xc3/0x170 [<ffffffff812447e4>] do_sys_open+0x124/0x220 [<ffffffff812448fe>] SyS_open+0x1e/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task mv:21069 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mv D ffff88028d056340 11216 21069 16018 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff81251fe1>] lock_rename+0x31/0xe0 [<ffffffff8125862f>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed432>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817e8a76>] ? trace_do_page_fault+0x56/0x170 [<ffffffff8125975e>] SyS_renameat2+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task truncate:21212 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. truncate D ffff880294409900 12904 21212 15964 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff817d5bb2>] lookup_slow+0x33/0xa7 [<ffffffff81256d6e>] path_lookupat+0x89e/0x8d0 [<ffffffff8121f126>] ? kmem_cache_alloc+0x126/0x2e0 [<ffffffff81257b9f>] ? getname_flags+0x4f/0x1a0 [<ffffffff81256dcb>] filename_lookup+0x2b/0xc0 [<ffffffff81258cc7>] user_path_at_empty+0x67/0xc0 [<ffffffff817e8655>] ? __do_page_fault+0x1c5/0x500 [<ffffffff81258d31>] user_path_at+0x11/0x20 [<ffffffff81243914>] do_sys_truncate+0x44/0xb0 [<ffffffff81243b0e>] SyS_truncate+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 LustreError: 24706:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0001-osc-ffff88029ab153d8: namespace resource [0x300000400:0x170:0x0].0x0 (ffff88027e62cf40) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 24706:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 101 previous similar messages LustreError: 24722:0:(file.c:5042:ll_inode_revalidate_fini()) lustre: revalidate FID [0x240000404:0x51:0x0] error: rc = -5 LustreError: 24722:0:(file.c:5042:ll_inode_revalidate_fini()) Skipped 9 previous similar messages LustreError: 24738:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff88029ab153d8: inode [0x240000403:0xe2:0x0] mdc close failed: rc = -108 LustreError: 24738:0:(file.c:234:ll_close_inode_openhandle()) Skipped 2 previous similar messages Lustre: lustre-MDT0001-mdc-ffff88029ab153d8: Connection restored to 192.168.123.120@tcp (at 0@lo) LustreError: 11253:0:(ldlm_lockd.c:1425:ldlm_handle_enqueue0()) ### lock on destroyed export ffff88029ab14a88 ns: mdt-lustre-MDT0001_UUID lock: ffff8802608d96c0/0xab3c1859804fc4f3 lrc: 3/0,0 mode: PR/PR res: [0x240000404:0x51:0x0].0x0 bits 0x1b/0x0 rrc: 11 type: IBT gid 0 flags: 0x50200000000000 nid: 0@lo remote: 0xab3c1859804fc4e5 expref: 14 pid: 11253 timeout: 0 lvb_type: 0 Lustre: 11234:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 8: before 510 < left 24309, rollback = 9 Lustre: 11234:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 48 previous similar messages LustreError: 11255:0:(ldlm_lockd.c:1425:ldlm_handle_enqueue0()) ### lock on destroyed export ffff88029ab14a88 ns: mdt-lustre-MDT0001_UUID lock: ffff88026b89ed00/0xab3c1859804f7147 lrc: 3/0,0 mode: PR/PR res: [0x240000404:0x51:0x0].0x0 bits 0x1b/0x0 rrc: 7 type: IBT gid 0 flags: 0x50200400000020 nid: 0@lo remote: 0xab3c1859804f712b expref: 6 pid: 11255 timeout: 0 lvb_type: 0 LustreError: 11255:0:(ldlm_lockd.c:1425:ldlm_handle_enqueue0()) Skipped 3 previous similar messages Lustre: 11245:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (93/1s); client may timeout req@ffff880284e79300 x1710093009489984/t0(0) o101->7eadd345-2683-4b20-82d7-c180a3661ff1@0@lo:445/0 lens 576/48088 e 2 to 0 dl 1630872170 ref 1 fl Complete:/0/0 rc -107/-107 job:'rm.0' Lustre: 11245:0:(service.c:2328:ptlrpc_server_handle_request()) Skipped 8 previous similar messages Lustre: mdt02_004: service thread pid 17389 was inactive for 186.186 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 1 previous similar message INFO: task mdt05_002:11250 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mdt05_002 D ffff8803252e8010 10656 11250 2 0x00000080 Call Trace: [<ffffffff817e0ab9>] schedule+0x29/0x70 [<ffffffff817e2245>] rwsem_down_write_failed+0x195/0x3c0 [<ffffffff814063f7>] call_rwsem_down_write_failed+0x17/0x30 [<ffffffff817dfe80>] down_write+0x40/0x50 [<ffffffffa0f56b57>] lod_qos_statfs_update+0x97/0x2e0 [lod] [<ffffffffa0f5a389>] lod_qos_prep_create+0x1b9/0x1a20 [lod] [<ffffffff817e340e>] ? _raw_read_unlock+0xe/0x20 [<ffffffffa0b95ea5>] ? qsd_op_begin+0x185/0x520 [lquota] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffffa0cf80b3>] ? osd_declare_inode_qid+0x283/0x450 [osd_ldiskfs] [<ffffffffa0f5be2b>] lod_prepare_create+0x23b/0x320 [lod] [<ffffffffa0f49aee>] lod_declare_striped_create+0x1fe/0xa00 [lod] [<ffffffffa0f5c462>] ? lod_sub_declare_create+0xe2/0x210 [lod] [<ffffffffa0f4f745>] lod_declare_create+0x1f5/0x560 [lod] [<ffffffffa0dde07a>] mdd_declare_create_object_internal+0x12a/0x340 [mdd] [<ffffffffa04c3a39>] ? lprocfs_counter_add+0xf9/0x160 [obdclass] [<ffffffffa0dc60e1>] mdd_declare_create_object.isra.34+0x51/0xc00 [mdd] [<ffffffffa0dc95a2>] mdd_declare_create+0x62/0x6f0 [mdd] [<ffffffffa0dc6ffe>] ? mdd_linkea_prepare+0x33e/0x4b0 [mdd] [<ffffffffa0dccf39>] mdd_create+0x9b9/0x1a40 [mdd] [<ffffffffa0e70e2b>] mdt_reint_open+0x250b/0x3830 [mdt] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e326e>] ? _raw_spin_unlock+0xe/0x20 [<ffffffffa05045bd>] ? upcall_cache_get_entry+0x3ed/0x900 [obdclass] [<ffffffffa0e4aaf2>] ? ucred_set_audit_enabled.isra.13+0x22/0x60 [mdt] [<ffffffffa0e62087>] mdt_reint_rec+0x87/0x240 [mdt] [<ffffffffa0e3755c>] mdt_reint_internal+0x76c/0xb40 [mdt] [<ffffffffa0e46233>] mdt_intent_open+0x93/0x480 [mdt] [<ffffffffa0e3b28d>] mdt_intent_opc+0x1dd/0xc10 [mdt] [<ffffffffa0e461a0>] ? mdt_intent_fixup_resent+0x240/0x240 [mdt] [<ffffffffa0e442a1>] mdt_intent_policy+0x1a1/0x360 [mdt] [<ffffffffa0735962>] ldlm_lock_enqueue+0x3c2/0xb40 [ptlrpc] [<ffffffff817e340e>] ? _raw_read_unlock+0xe/0x20 [<ffffffffa01cbd85>] ? cfs_hash_rw_unlock+0x15/0x20 [libcfs] [<ffffffffa01cefb6>] ? cfs_hash_add+0xa6/0x180 [libcfs] [<ffffffffa075f7a6>] ldlm_handle_enqueue0+0xaa6/0x16d0 [ptlrpc] [<ffffffffa0784000>] ? lustre_msg_buf_v2+0x160/0x1e0 [ptlrpc] [<ffffffffa07e1494>] tgt_enqueue+0x64/0x240 [ptlrpc] [<ffffffffa07eb26e>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<ffffffffa031f45e>] ? libcfs_nid2str_r+0xfe/0x130 [lnet] [<ffffffffa0796990>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<ffffffffa0798529>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffffa0797950>] ? ptlrpc_wait_event+0x620/0x620 [ptlrpc] [<ffffffff810ba0f4>] kthread+0xe4/0xf0 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 INFO: task mdt03_003:14983 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mdt03_003 D ffff88028cfcba00 10624 14983 2 0x00000080 Call Trace: [<ffffffff817e0ab9>] schedule+0x29/0x70 [<ffffffff817e2245>] rwsem_down_write_failed+0x195/0x3c0 [<ffffffff817d5822>] ? free_debug_processing+0x240/0x272 [<ffffffff814063f7>] call_rwsem_down_write_failed+0x17/0x30 [<ffffffff817dfe80>] down_write+0x40/0x50 [<ffffffffa0f56b57>] lod_qos_statfs_update+0x97/0x2e0 [lod] [<ffffffffa0f5a389>] lod_qos_prep_create+0x1b9/0x1a20 [lod] [<ffffffff812204ff>] ? __kmalloc+0x1ef/0x370 [<ffffffffa0f2a27b>] ? lod_verify_striping+0xc7b/0x1260 [lod] [<ffffffffa0f256f3>] ? lod_alloc_comp_entries+0xd3/0x530 [lod] [<ffffffffa0f5be2b>] lod_prepare_create+0x23b/0x320 [lod] [<ffffffffa0f49aee>] lod_declare_striped_create+0x1fe/0xa00 [lod] [<ffffffff817e326e>] ? _raw_spin_unlock+0xe/0x20 [<ffffffffa0f52219>] lod_declare_xattr_set+0x139/0x14f0 [lod] [<ffffffffa04c3a39>] ? lprocfs_counter_add+0xf9/0x160 [obdclass] [<ffffffffa0cb742b>] ? osd_trans_create+0xbb/0x670 [osd_ldiskfs] [<ffffffffa0f1635c>] ? lod_trans_create+0x3c/0x50 [lod] [<ffffffffa0dcaaf2>] mdd_create_data+0x492/0x760 [mdd] [<ffffffffa0ddb600>] ? mdd_xattr_get+0x10/0x620 [mdd] [<ffffffffa0e6de3c>] mdt_mfd_open+0xccc/0xee0 [mdt] [<ffffffffa0e35f0f>] ? mdt_pack_acl2body+0x1cf/0x950 [mdt] [<ffffffffa0e6e66b>] mdt_finish_open+0x61b/0x8d0 [mdt] [<ffffffffa0e705cc>] mdt_reint_open+0x1cac/0x3830 [mdt] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e326e>] ? _raw_spin_unlock+0xe/0x20 [<ffffffffa05045bd>] ? upcall_cache_get_entry+0x3ed/0x900 [obdclass] [<ffffffffa050979e>] ? lu_ucred+0x1e/0x30 [obdclass] [<ffffffffa0e4aaf2>] ? ucred_set_audit_enabled.isra.13+0x22/0x60 [mdt] [<ffffffffa0e62087>] mdt_reint_rec+0x87/0x240 [mdt] [<ffffffffa0e3755c>] mdt_reint_internal+0x76c/0xb40 [mdt] [<ffffffffa0e46233>] mdt_intent_open+0x93/0x480 [mdt] [<ffffffffa0e3b28d>] mdt_intent_opc+0x1dd/0xc10 [mdt] [<ffffffffa0e461a0>] ? mdt_intent_fixup_resent+0x240/0x240 [mdt] [<ffffffffa0e442a1>] mdt_intent_policy+0x1a1/0x360 [mdt] [<ffffffffa0735962>] ldlm_lock_enqueue+0x3c2/0xb40 [ptlrpc] [<ffffffff817e340e>] ? _raw_read_unlock+0xe/0x20 [<ffffffffa01cbd85>] ? cfs_hash_rw_unlock+0x15/0x20 [libcfs] [<ffffffffa01cefb6>] ? cfs_hash_add+0xa6/0x180 [libcfs] [<ffffffffa075f7a6>] ldlm_handle_enqueue0+0xaa6/0x16d0 [ptlrpc] [<ffffffffa0784000>] ? lustre_msg_buf_v2+0x160/0x1e0 [ptlrpc] [<ffffffffa07e1494>] tgt_enqueue+0x64/0x240 [ptlrpc] [<ffffffffa07eb26e>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<ffffffffa031f45e>] ? libcfs_nid2str_r+0xfe/0x130 [lnet] [<ffffffffa0796990>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<ffffffffa0798529>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffffa0797950>] ? ptlrpc_wait_event+0x620/0x620 [ptlrpc] [<ffffffff810ba0f4>] kthread+0xe4/0xf0 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 INFO: task dir_create.sh:15928 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. dir_create.sh D ffff88008d29c9d0 11232 15928 15903 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff817d5bb2>] lookup_slow+0x33/0xa7 [<ffffffff81254c7f>] link_path_walk+0x81f/0x8c0 [<ffffffff8124818c>] ? get_empty_filp+0x5c/0x1f0 [<ffffffff8125708e>] path_openat+0xae/0x5b0 [<ffffffff817d589c>] ? __slab_alloc+0x48/0x5c [<ffffffff81258dfd>] do_filp_open+0x4d/0xb0 [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e326e>] ? _raw_spin_unlock+0xe/0x20 [<ffffffff81267893>] ? __alloc_fd+0xc3/0x170 [<ffffffff812447e4>] do_sys_open+0x124/0x220 [<ffffffff812448fe>] SyS_open+0x1e/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task dir_create.sh:15933 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. dir_create.sh D ffff88008d29dc40 11232 15933 15902 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff817d5bb2>] lookup_slow+0x33/0xa7 [<ffffffff81254c7f>] link_path_walk+0x81f/0x8c0 [<ffffffff8124818c>] ? get_empty_filp+0x5c/0x1f0 [<ffffffff8125708e>] path_openat+0xae/0x5b0 [<ffffffff81258dfd>] do_filp_open+0x4d/0xb0 [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e326e>] ? _raw_spin_unlock+0xe/0x20 [<ffffffff81267893>] ? __alloc_fd+0xc3/0x170 [<ffffffff812447e4>] do_sys_open+0x124/0x220 [<ffffffff812448fe>] SyS_open+0x1e/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 ptlrpc_watchdog_fire: 58 callbacks suppressed Lustre: mdt05_002: service thread pid 11250 was inactive for 236.519 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 11250, comm: mdt05_002 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] call_rwsem_down_write_failed+0x17/0x30 [<0>] lod_qos_statfs_update+0x97/0x2e0 [lod] [<0>] lod_qos_prep_create+0x1b9/0x1a20 [lod] [<0>] lod_prepare_create+0x23b/0x320 [lod] [<0>] lod_declare_striped_create+0x1fe/0xa00 [lod] [<0>] lod_declare_create+0x1f5/0x560 [lod] [<0>] mdd_declare_create_object_internal+0x12a/0x340 [mdd] [<0>] mdd_declare_create_object.isra.34+0x51/0xc00 [mdd] [<0>] mdd_declare_create+0x62/0x6f0 [mdd] [<0>] mdd_create+0x9b9/0x1a40 [mdd] [<0>] mdt_reint_open+0x250b/0x3830 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x76c/0xb40 [mdt] [<0>] mdt_intent_open+0x93/0x480 [mdt] [<0>] mdt_intent_opc+0x1dd/0xc10 [mdt] [<0>] mdt_intent_policy+0x1a1/0x360 [mdt] [<0>] ldlm_lock_enqueue+0x3c2/0xb40 [ptlrpc] [<0>] ldlm_handle_enqueue0+0xaa6/0x16d0 [ptlrpc] [<0>] tgt_enqueue+0x64/0x240 [ptlrpc] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: mdt02_006: service thread pid 17469 was inactive for 186.126 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 17469, comm: mdt02_006 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] mdt_object_local_lock+0x527/0xb90 [mdt] [<0>] mdt_object_lock_internal+0x70/0x390 [mdt] [<0>] mdt_object_find_lock+0x6a/0x1a0 [mdt] [<0>] mdt_reint_setxattr+0x1fa/0x1110 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x76c/0xb40 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe LustreError: 18724:0:(ldlm_request.c:123:ldlm_expired_completion_wait()) ### lock timed out (enqueued at 1630872066, 300s ago); not entering recovery in server code, just going back to sleep ns: mdt-lustre-MDT0000_UUID lock: ffff88026ab30b80/0xab3c1859804e4fe6 lrc: 3/1,0 mode: --/PR res: [0x200000404:0x88:0x0].0x0 bits 0x12/0x0 rrc: 9 type: IBT gid 0 flags: 0x40210400000020 nid: local remote: 0x0 expref: -99 pid: 18724 timeout: 0 lvb_type: 0 LustreError: dumping log to /tmp/lustre-log.1630872366.18724 LustreError: 21335:0:(ldlm_request.c:123:ldlm_expired_completion_wait()) ### lock timed out (enqueued at 1630872069, 300s ago); not entering recovery in server code, just going back to sleep ns: mdt-lustre-MDT0000_UUID lock: ffff88026b5c3100/0xab3c1859804e8c70 lrc: 3/1,0 mode: --/PR res: [0x200000404:0x88:0x0].0x0 bits 0x12/0x0 rrc: 9 type: IBT gid 0 flags: 0x40210400000020 nid: local remote: 0x0 expref: -99 pid: 21335 timeout: 0 lvb_type: 0 LustreError: 17389:0:(ldlm_request.c:123:ldlm_expired_completion_wait()) ### lock timed out (enqueued at 1630872075, 300s ago); not entering recovery in server code, just going back to sleep ns: mdt-lustre-MDT0000_UUID lock: ffff8802554cc3c0/0xab3c1859804eac13 lrc: 3/1,0 mode: --/PR res: [0x200000404:0x88:0x0].0x0 bits 0x12/0x0 rrc: 9 type: IBT gid 0 flags: 0x40210400000020 nid: local remote: 0x0 expref: -99 pid: 17389 timeout: 0 lvb_type: 0 LustreError: 17538:0:(ldlm_request.c:123:ldlm_expired_completion_wait()) ### lock timed out (enqueued at 1630872080, 300s ago); not entering recovery in server code, just going back to sleep ns: mdt-lustre-MDT0000_UUID lock: ffff880087b2b4c0/0xab3c1859804f081d lrc: 3/1,0 mode: --/PR res: [0x200000403:0x113:0x0].0x0 bits 0x12/0x0 rrc: 11 type: IBT gid 0 flags: 0x40210400000020 nid: local remote: 0x0 expref: -99 pid: 17538 timeout: 0 lvb_type: 0 Lustre: lustre-MDT0000-osp-MDT0001: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 2 previous similar messages LustreError: 11247:0:(ldlm_request.c:142:ldlm_expired_completion_wait()) ### lock timed out (enqueued at 1630872082, 300s ago), entering recovery for lustre-MDT0000_UUID@192.168.123.120@tcp ns: lustre-MDT0000-osp-MDT0001 lock: ffff8802648dc000/0xab3c1859804f2b08 lrc: 4/0,1 mode: --/PW res: [0x200000403:0x113:0x0].0x0 bits 0x2/0x0 rrc: 2 type: IBT gid 0 flags: 0x1000001000000 nid: local remote: 0xab3c1859804f2c12 expref: -99 pid: 11247 timeout: 0 lvb_type: 0 Lustre: lustre-MDT0000: Received new MDS connection from 0@lo, keep former export from same NID Lustre: lustre-MDT0000-osp-MDT0001: Connection restored to 192.168.123.120@tcp (at 0@lo) Lustre: Skipped 4 previous similar messages LustreError: 17509:0:(ldlm_request.c:123:ldlm_expired_completion_wait()) ### lock timed out (enqueued at 1630872085, 300s ago); not entering recovery in server code, just going back to sleep ns: mdt-lustre-MDT0000_UUID lock: ffff88026d90e1c0/0xab3c1859804f573f lrc: 3/0,1 mode: --/EX res: [0x200000004:0x1:0x0].0x0 bits 0x2/0x0 rrc: 4 type: IBT gid 0 flags: 0x40210400000020 nid: local remote: 0x0 expref: -99 pid: 17509 timeout: 0 lvb_type: 0 LustreError: 17509:0:(ldlm_request.c:123:ldlm_expired_completion_wait()) Skipped 5 previous similar messages LustreError: 11231:0:(ldlm_request.c:142:ldlm_expired_completion_wait()) ### lock timed out (enqueued at 1630872092, 300s ago), entering recovery for lustre-MDT0000_UUID@192.168.123.120@tcp ns: lustre-MDT0000-osp-MDT0001 lock: ffff88026df761c0/0xab3c1859804f6df1 lrc: 4/0,1 mode: --/EX res: [0x200000004:0x1:0x0].0x0 bits 0x2/0x0 rrc: 3 type: IBT gid 0 flags: 0x1000001000000 nid: local remote: 0xab3c1859804f6df8 expref: -99 pid: 11231 timeout: 0 lvb_type: 0 Lustre: 11247:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 501 < left 582, rollback = 2 Lustre: 11247:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 3/12/7, destroy: 0/0/0 Lustre: 11247:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 20416 previous similar messages Lustre: 11247:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 6/582/0 Lustre: 11247:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 20416 previous similar messages Lustre: 11247:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 5/136/0, punch: 0/0/0, quota 6/6/0 Lustre: 11247:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 20416 previous similar messages Lustre: 11247:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 10/203/4, delete: 0/0/0 Lustre: 11247:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 20410 previous similar messages Lustre: 11247:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 4/4/0, ref_del: 0/0/0 Lustre: 11247:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 20416 previous similar messages Lustre: 11247:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (93/242s); client may timeout req@ffff880267d1d7c0 x1710093009495680/t4294970110(0) o36->7eadd345-2683-4b20-82d7-c180a3661ff1@0@lo:445/0 lens 688/456 e 2 to 0 dl 1630872170 ref 1 fl Complete:/0/0 rc 0/0 job:'lfs.0' Lustre: 11280:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 509 < left 61457, rollback = 7 Lustre: 11280:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 17973 previous similar messages Lustre: 4131:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff88026ac63880 x1710093013366080/t0(0) o10->lustre-OST0002-osc-ffff880299a64138@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630872493 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dd.0' LustreError: 19700:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0002: '7' migration was interrupted, run 'lfs migrate -m 1 -c 1 -H crush 7' to finish migration. LustreError: 17512:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0000: '19' migration was interrupted, run 'lfs migrate -m 2 -c 1 -H crush 19' to finish migration. LustreError: 11081:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0001_UUID lock: ffff88026c57a200/0xab3c18598052d651 lrc: 3/0,0 mode: PW/PW res: [0xdb:0x0:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->786431) gid 0 flags: 0x60000400020020 nid: 0@lo remote: 0xab3c18598052d64a expref: 2254 pid: 23882 timeout: 900 lvb_type: 0 LustreError: 11081:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 9 previous similar messages LustreError: 11-0: lustre-OST0001-osc-ffff88029ab153d8: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 137 previous similar messages Lustre: lustre-OST0001-osc-ffff88029ab153d8: Connection to lustre-OST0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0001-osc-ffff88029ab153d8: This client was evicted by lustre-OST0001; in progress operations using this service will fail. Lustre: ll_ost07_011: service thread pid 20701 was inactive for 40.052 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 20701, comm: ll_ost07_011 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] tgt_extent_lock+0xea/0x2a0 [ptlrpc] [<0>] ofd_getattr_hdl+0x365/0x720 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: 4136:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.120@tcp:/lustre/fid: [0x200000404:0x1bb:0x0]// may get corrupted (rc -108) LustreError: 28880:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0001-osc-ffff88029ab153d8: namespace resource [0x140:0x0:0x0].0x0 (ffff880259c651c0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 28880:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 528 previous similar messages Lustre: lustre-OST0001-osc-ffff88029ab153d8: Connection restored to 192.168.123.120@tcp (at 0@lo) Lustre: 4135:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff88007c7ab880 x1710093020673280/t0(0) o10->lustre-OST0001-osc-ffff88029ab153d8@0@lo:6/4 lens 440/432 e 1 to 0 dl 1630872599 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'cp.0' LustreError: 19778:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88026db91300 x1710093021114688/t0(0) o105->lustre-OST0003@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 19778:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 65 previous similar messages LustreError: 11-0: lustre-OST0003-osc-ffff880299a64138: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 24 previous similar messages LustreError: 167-0: lustre-OST0003-osc-ffff880299a64138: This client was evicted by lustre-OST0003; in progress operations using this service will fail. Lustre: 4134:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.120@tcp:/lustre/fid: [0x280000404:0x188:0x0]// may get corrupted (rc -5) LustreError: 11269:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000404:0x17a:0x0]: rc = -2 LustreError: 29348:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0003-osc-ffff880299a64138: namespace resource [0x380000401:0x3cf:0x0].0x0 (ffff880087d356c0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 11081:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0000_UUID lock: ffff88025f837c00/0xab3c185980598855 lrc: 3/0,0 mode: PW/PW res: [0x2c0000401:0x431:0x0].0x0 rrc: 4 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000400020020 nid: 0@lo remote: 0xab3c18598059884e expref: 2113 pid: 23857 timeout: 965 lvb_type: 0 LustreError: 11081:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 1 previous similar message LustreError: 28909:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88026bed7700 x1710093022428480/t0(0) o105->lustre-OST0000@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 11-0: lustre-OST0000-osc-ffff880299a64138: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: 167-0: lustre-OST0000-osc-ffff880299a64138: This client was evicted by lustre-OST0000; in progress operations using this service will fail. Lustre: 4135:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.120@tcp:/lustre/fid: [0x200000404:0x1e0:0x0]// may get corrupted (rc -108) Lustre: 4136:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.120@tcp:/lustre/fid: [0x280000404:0x19d:0x0]/ may get corrupted (rc -108) Lustre: 4135:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.120@tcp:/lustre/fid: [0x200000404:0x1ee:0x0]// may get corrupted (rc -108) Lustre: 4136:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.120@tcp:/lustre/fid: [0x200000404:0x155:0x0]// may get corrupted (rc -108) Lustre: ll_ost06_012: service thread pid 25250 was inactive for 40.201 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 25250, comm: ll_ost06_012 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: Lustre: dir [0x240000404:0x189:0x0] stripe 1 readdir failed: -2, directory is partially accessed! [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] tgt_extent_lock+0xea/0x2a0 [ptlrpc] [<0>] ofd_getattr_hdl+0x365/0x720 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe LustreError: 30159:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0000-osc-ffff880299a64138: namespace resource [0x2c0000401:0x3ff:0x0].0x0 (ffff88026d7d4040) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 30159:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 131 previous similar messages Lustre: lustre-OST0000-osc-ffff880299a64138: Connection restored to 192.168.123.120@tcp (at 0@lo) Lustre: Skipped 1 previous similar message Lustre: 11239:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 504 < left 39226, rollback = 7 Lustre: 11239:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 101910 previous similar messages Lustre: 11239:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 912/3648/0, destroy: 0/0/0 Lustre: 11239:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 102129 previous similar messages Lustre: 11239:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/1, xattr_set: 914/15/0 Lustre: 11239:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 102128 previous similar messages Lustre: 11239:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 4561/39226/0, punch: 0/0/0, quota 10/106/7 Lustre: 11239:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 102128 previous similar messages Lustre: 11239:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 912/15504/0, delete: 0/0/0 Lustre: 11239:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 102129 previous similar messages Lustre: 11239:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 11239:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 102128 previous similar messages Lustre: 11250:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 504 < left 877, rollback = 2 Lustre: 11250:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 219 previous similar messages Lustre: 11249:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000403:0x1f8:0x0] with magic=0xbd60bd0 Lustre: 17485:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000403:0x1ed:0x0] with magic=0xbd60bd0 Lustre: 17485:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 1 previous similar message Lustre: 20776:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000403:0x1e9:0x0] with magic=0xbd60bd0 Lustre: 20776:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 1 previous similar message LustreError: 28891:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88028dda3880 x1710093024615168/t0(0) o106->lustre-OST0003@0@lo:15/16 lens 328/280 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 28891:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 9 previous similar messages Lustre: 26228:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000405:0x16d:0x0] with magic=0xbd60bd0 Lustre: 26228:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 1 previous similar message Lustre: dir [0x280000403:0x1ec:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 1 previous similar message LustreError: 11276:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0002: failed to get lu_attr of [0x280000403:0x1e2:0x0]: rc = -2 Lustre: 4123:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff8800863bcb40 x1710093026263616/t0(0) o10->lustre-OST0003-osc-ffff88029ab153d8@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630872794 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dir_create.sh.0' LustreError: 761:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff88029ab153d8: inode [0x200000405:0x104:0x0] mdc close failed: rc = -13 LustreError: 761:0:(file.c:234:ll_close_inode_openhandle()) Skipped 10 previous similar messages Lustre: 21335:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000405:0x196:0x0] with magic=0xbd60bd0 Lustre: 21335:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 1 previous similar message Lustre: ll_ost03_013: service thread pid 22169 was inactive for 40.048 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 12834, comm: ll_ost02_001 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] Lustre: ll_ost03_002: service thread pid 12839 was inactive for 40.302 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] Lustre: Skipped 1 previous similar message [<0>] ofd_destroy_by_fid+0x1d1/0x520 [ofd] [<0>] ofd_destroy_hdl+0x27f/0xa20 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: Skipped 1 previous similar message Pid: 22169, comm: ll_ost03_013 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] ofd_destroy_by_fid+0x1d1/0x520 [ofd] [<0>] ofd_destroy_hdl+0x27f/0xa20 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: 17441:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000405:0x180:0x0] with magic=0xbd60bd0 Lustre: 17441:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 7 previous similar messages Lustre: 4123:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff8800863bcb40 x1710093028023680/t0(0) o10->lustre-OST0003-osc-ffff88029ab153d8@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630872803 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dir_create.sh.0' Lustre: dir [0x280000404:0xd7:0x0] stripe 1 readdir failed: -2, directory is partially accessed! LustreError: 11081:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0001_UUID lock: ffff880077c99e40/0xab3c18598061df4e lrc: 3/0,0 mode: PW/PW res: [0x300000400:0x980:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000000020020 nid: 0@lo remote: 0xab3c18598061df40 expref: 1328 pid: 24646 timeout: 1130 lvb_type: 0 LustreError: 11081:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 1 previous similar message LustreError: 11-0: lustre-OST0001-osc-ffff880299a64138: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 18 previous similar messages LustreError: 17266:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88028bfca5c0 x1710093028888128/t0(0) o105->lustre-OST0001@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 17266:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 7 previous similar messages Lustre: lustre-OST0001-osc-ffff880299a64138: Connection to lustre-OST0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 3 previous similar messages LustreError: 167-0: lustre-OST0001-osc-ffff880299a64138: This client was evicted by lustre-OST0001; in progress operations using this service will fail. LustreError: Skipped 1 previous similar message Lustre: 4136:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.120@tcp:/lustre/fid: [0x200000405:0x78:0x0]/ may get corrupted (rc -108) Lustre: 4136:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.120@tcp:/lustre/fid: [0x280000404:0x2cb:0x0]/ may get corrupted (rc -108) Lustre: lustre-OST0001-osc-ffff880299a64138: Connection restored to 192.168.123.120@tcp (at 0@lo) Lustre: Skipped 1 previous similar message Lustre: 30566:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000404:0x25a:0x0] with magic=0xbd60bd0 Lustre: 30566:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 1 previous similar message 6[4398]: segfault at 0 ip 0000000000403e5f sp 00007ffde8bc6a30 error 6 in 6[400000+6000] LustreError: 17512:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0000: '7' migration was interrupted, run 'lfs migrate -m 1 -c 1 -H crush 7' to finish migration. Lustre: 4124:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.120@tcp:/lustre/fid: [0x280000403:0x3bc:0x0]/ may get corrupted (rc -108) LustreError: 6768:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0003-osc-ffff88029ab153d8: namespace resource [0x380000401:0x50f:0x0].0x0 (ffff880260f782c0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 6768:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 175 previous similar messages Lustre: 22800:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 6s req@ffff880263db9f80 x1710093032846720/t0(0) o103->e99e81a8-06e2-48aa-9753-9862b5dc40ec@0@lo:0/0 lens 328/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'ldlm_bl_07.0' Lustre: 22800:0:(service.c:2157:ptlrpc_server_handle_req_in()) Skipped 1 previous similar message 4[6838]: segfault at 8 ip 00007fe97623b7e8 sp 00007ffe1d2be9e0 error 4 in ld-2.17.so[7fe976230000+22000] Lustre: 4125:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff88029ad857c0 x1710093034246464/t0(0) o10->lustre-OST0003-osc-ffff88029ab153d8@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630872964 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dd.0' Lustre: 26228:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000405:0x2d9:0x0] with magic=0xbd60bd0 Lustre: 26228:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 7 previous similar messages LustreError: 21302:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0000: '5' migration was interrupted, run 'lfs migrate -m 1 -c 1 -H crush 5' to finish migration. LustreError: 11081:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 102s: evicting client at 0@lo ns: filter-lustre-OST0001_UUID lock: ffff880266e9f0c0/0xab3c1859806d4306 lrc: 3/0,0 mode: PW/PW res: [0x300000401:0x695:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000400030020 nid: 0@lo remote: 0xab3c1859806d42ff expref: 1475 pid: 20683 timeout: 1303 lvb_type: 0 LustreError: 11081:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 1 previous similar message LustreError: 11080:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88026c8b25c0 x1710093035663680/t0(0) o105->lustre-OST0001@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 11080:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 5 previous similar messages LustreError: 11-0: lustre-OST0001-osc-ffff880299a64138: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 1 previous similar message LustreError: 16986:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630872984 with bad export cookie 12338763851773797988 LustreError: 167-0: lustre-OST0001-osc-ffff880299a64138: This client was evicted by lustre-OST0001; in progress operations using this service will fail. LustreError: Skipped 1 previous similar message LustreError: 7300:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff88029ab153d8: inode [0x200000405:0x29f:0x0] mdc close failed: rc = -13 Lustre: 4134:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.120@tcp:/lustre/fid: [0x240000404:0x317:0x0]// may get corrupted (rc -108) Lustre: 4133:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.120@tcp:/lustre/fid: [0x200000404:0x3ba:0x0]// may get corrupted (rc -108) Lustre: 4134:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.120@tcp:/lustre/fid: [0x280000404:0x374:0x0]// may get corrupted (rc -108) Lustre: lustre-OST0001-osc-ffff880299a64138: Connection restored to 192.168.123.120@tcp (at 0@lo) Lustre: Skipped 1 previous similar message Lustre: dir [0x280000403:0x2e6:0x0] stripe 1 readdir failed: -2, directory is partially accessed! LustreError: 11282:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0002: failed to get lu_attr of [0x280000403:0x2cc:0x0]: rc = -2 Lustre: 4122:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff88026c72c500 x1710093038494720/t0(0) o10->lustre-OST0001-osc-ffff880299a64138@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630873071 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dd.0' ptlrpc_watchdog_fire: 22 callbacks suppressed Lustre: ll_ost05_004: service thread pid 20370 was inactive for 40.078 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 20370, comm: ll_ost05_004 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] ofd_destroy_by_fid+0x1d1/0x520 [ofd] [<0>] ofd_destroy_hdl+0x27f/0xa20 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 23873, comm: ll_ost05_039 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] ofd_destroy_by_fid+0x1d1/0x520 [ofd] [<0>] ofd_destroy_hdl+0x27f/0xa20 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 24324, comm: ll_ost05_058 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] ofd_destroy_by_fid+0x1d1/0x520 [ofd] [<0>] ofd_destroy_hdl+0x27f/0xa20 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: ll_ost05_003: service thread pid 16431 was inactive for 41.847 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 21 previous similar messages Lustre: lustre-OST0003-osc-ffff88029ab153d8: Connection to lustre-OST0003 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 2 previous similar messages LustreError: 12338:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0003-osc-ffff88029ab153d8: namespace resource [0x15b0:0x0:0x0].0x0 (ffff88008403b240) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 12338:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 1 previous similar message Lustre: 4123:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff88028560ac00 x1710093041433344/t0(0) o10->lustre-OST0002-osc-ffff880299a64138@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630873180 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dir_create.sh.0' Lustre: 4123:0:(client.c:1485:after_reply()) Skipped 1 previous similar message LustreError: 11282:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0002: failed to get lu_attr of [0x280000403:0x2e6:0x0]: rc = -2 Lustre: 22258:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 508 < left 84032, rollback = 7 Lustre: 22258:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 186765 previous similar messages Lustre: 22258:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 1954/7816/0, destroy: 1/4/1 Lustre: 22258:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 190709 previous similar messages Lustre: 22258:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 3/3/1, xattr_set: 1958/148/0 Lustre: 22258:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 190709 previous similar messages Lustre: 22258:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 9771/84032/16, punch: 0/0/0, quota 6/6/0 Lustre: 22258:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 190709 previous similar messages Lustre: 22258:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 1955/33234/0, delete: 2/5/1 Lustre: 22258:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 190708 previous similar messages Lustre: 22258:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 1/1/0, ref_del: 2/2/1 Lustre: 22258:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 190709 previous similar messages Lustre: 17232:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 504 < left 716, rollback = 2 Lustre: 17232:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 2869 previous similar messages LustreError: 17491:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630873210 with bad export cookie 12338763851772001816 Lustre: 4132:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.120@tcp:/lustre/fid: [0x240000404:0x3e8:0x0]// may get corrupted (rc -108) Lustre: 4124:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.120@tcp:/lustre/fid: [0x240000404:0x401:0x0]// may get corrupted (rc -108) Lustre: 4125:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.120@tcp:/lustre/fid: [0x240000404:0x3f3:0x0]// may get corrupted (rc -108) Lustre: 4124:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.120@tcp:/lustre/fid: [0x200000404:0x48f:0x0]/ may get corrupted (rc -108) Lustre: 4131:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff8800799e2c00 x1710093048476800/t0(0) o10->lustre-OST0002-osc-ffff88029ab153d8@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630873344 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dir_create.sh.0' Lustre: 4131:0:(client.c:1485:after_reply()) Skipped 4 previous similar messages ------------[ cut here ]------------ WARNING: CPU: 14 PID: 16209 at /home/green/git/lustre-release/ldiskfs/ext4_jbd2.c:266 __ldiskfs_handle_dirty_metadata+0x1d2/0x230 [ldiskfs] Modules linked in: loop zfs(PO) zunicode(PO) zzstd(O) zlua(O) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) jbd2 mbcache lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) dm_flakey dm_mod libcfs(OE) crc_t10dif crct10dif_generic sb_edac edac_core iosf_mbi crc32_pclmul ghash_clmulni_intel aesni_intel lrw gf128mul glue_helper ablk_helper cryptd virtio_balloon pcspkr i2c_piix4 virtio_console ip_tables rpcsec_gss_krb5 ata_generic drm_kms_helper ttm pata_acpi crct10dif_pclmul drm crct10dif_common crc32c_intel drm_panel_orientation_quirks ata_piix virtio_blk serio_raw i2c_core libata floppy CPU: 14 PID: 16209 Comm: mdt07_003 Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011 Call Trace: [<ffffffff817d9418>] dump_stack+0x19/0x1b [<ffffffff8108d538>] __warn+0xd8/0x100 [<ffffffff8108d67d>] warn_slowpath_null+0x1d/0x20 [<ffffffffa0c3a262>] __ldiskfs_handle_dirty_metadata+0x1d2/0x230 [ldiskfs] [<ffffffffa0c1a33d>] ldiskfs_getblk+0x13d/0x210 [ldiskfs] [<ffffffffa0c1a437>] ldiskfs_bread+0x27/0xe0 [ldiskfs] [<ffffffffa0cdee97>] osd_write+0x4e7/0xcb0 [osd_ldiskfs] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffffa04e96f3>] dt_record_write+0x33/0x120 [obdclass] [<ffffffffa04a86b2>] llog_osd_write_rec+0x1052/0x19a0 [obdclass] [<ffffffffa0495760>] llog_write_rec+0x290/0x590 [obdclass] [<ffffffffa049b4ee>] llog_cat_add_rec+0x23e/0x950 [obdclass] [<ffffffffa04925ff>] llog_add+0x17f/0x1f0 [obdclass] [<ffffffffa04c3a39>] ? lprocfs_counter_add+0xf9/0x160 [obdclass] [<ffffffffa082c60d>] sub_updates_write+0xb6b/0xe32 [ptlrpc] [<ffffffffa07fc892>] top_trans_stop+0x7c2/0xf60 [ptlrpc] [<ffffffffa0f19f3c>] lod_trans_stop+0x25c/0x340 [lod] [<ffffffffa0dee408>] mdd_trans_stop+0x28/0x16e [mdd] [<ffffffffa0dcfb2f>] mdd_migrate_object+0xf1f/0x1ba0 [mdd] [<ffffffffa04e8375>] ? lu_object_find_at+0x285/0xb10 [obdclass] [<ffffffffa0dd0b86>] mdd_migrate+0x3d6/0x8e0 [mdd] [<ffffffffa0ea20d8>] mdo_migrate+0x4c/0x4e [mdt] [<ffffffffa0e61e53>] mdt_reint_migrate+0x1023/0x11d0 [mdt] [<ffffffffa0e62087>] mdt_reint_rec+0x87/0x240 [mdt] [<ffffffffa0e3755c>] mdt_reint_internal+0x76c/0xb40 [mdt] [<ffffffffa0e44657>] mdt_reint+0x67/0x150 [mdt] [<ffffffffa07eb26e>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<ffffffffa031f45e>] ? libcfs_nid2str_r+0xfe/0x130 [lnet] [<ffffffffa0796990>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<ffffffffa0798529>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffffa0797950>] ? ptlrpc_wait_event+0x620/0x620 [ptlrpc] [<ffffffff810ba0f4>] kthread+0xe4/0xf0 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 Lustre: mdt07_003: service thread pid 16209 was inactive for 142.354 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 8 previous similar messages [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 ---[ end trace 147a0d2842d7da1e ]--- LDISKFS-fs: ldiskfs_getblk:888: aborting transaction: error 28 in __ldiskfs_handle_dirty_metadata LDISKFS-fs error (device dm-2): ldiskfs_getblk:888: inode #211: block 53099: comm mdt07_003: journal_dirty_metadata failed: handle type 0 started at line 1972, credits 512/0, errcode -28 Aborting journal on device dm-2-8. LDISKFS-fs (dm-2): Remounting filesystem read-only LustreError: 16209:0:(osd_io.c:2103:osd_ldiskfs_write_record()) lustre-MDT0002: error reading offset 434176 (block 106, size 40, offs 434160), credits 0/0: rc = -28 LustreError: 16209:0:(update_trans.c:985:top_trans_stop()) lustre-MDT0002-osd: write updates failed: rc = -28 LustreError: 16209:0:(osd_io.c:2110:osd_ldiskfs_write_record()) journal_get_write_access() returned error -30 LustreError: 16209:0:(tgt_lastrcvd.c:1253:tgt_add_reply_data()) lustre-MDT0002: can't update reply_data file: rc = -30 LustreError: 16209:0:(osd_handler.c:2075:osd_trans_stop()) lustre-MDT0002: failed in transaction hook: rc = -30 LDISKFS-fs error (device dm-2) in osd_trans_stop:2082: error 28 LustreError: 16209:0:(osd_handler.c:2085:osd_trans_stop()) lustre-MDT0002: failed to stop transaction: rc = -28 LustreError: 16209:0:(update_trans.c:1011:top_trans_stop()) lustre-MDT0002-osd: stop trans failed: rc = -30 LustreError: 13712:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-OST0000-osc-MDT0002: fail to cancel 76 llog-records: rc = -30 LustreError: 13712:0:(osp_sync.c:1094:osp_sync_process_committed()) lustre-OST0000-osc-MDT0002: can't cancel 76 records: rc = -30 LustreError: 14583:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-OST0002-osc-MDT0002: fail to cancel 1 of 1 llog-records: rc = -30 LustreError: 14583:0:(osp_sync.c:1079:osp_sync_process_committed()) lustre-OST0002-osc-MDT0002: can't cancel record: rc = -30 LustreError: 16267:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -30 LustreError: 12403:0:(osd_handler.c:1779:osd_trans_commit_cb()) transaction @0xffff88025fee39d8 commit error: 2 LustreError: 16943:0:(tgt_handler.c:1367:tgt_blocking_ast()) lustre-MDT0002: syncing [0x280000401:0x7:0x0] (19-0) on lock cancel: rc = -5 LustreError: 14587:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-OST0003-osc-MDT0002: fail to cancel 19 llog-records: rc = -30 LustreError: 14587:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) Skipped 12 previous similar messages LustreError: 14587:0:(osp_sync.c:1094:osp_sync_process_committed()) lustre-OST0003-osc-MDT0002: can't cancel 19 records: rc = -30 LustreError: 14587:0:(osp_sync.c:1094:osp_sync_process_committed()) Skipped 3 previous similar messages LustreError: 13717:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-OST0001-osc-MDT0002: fail to cancel 1 of 1 llog-records: rc = -30 LustreError: 13717:0:(llog_cat.c:787:llog_cat_cancel_records()) Skipped 8 previous similar messages LustreError: 13717:0:(osp_sync.c:1079:osp_sync_process_committed()) lustre-OST0001-osc-MDT0002: can't cancel record: rc = -30 LustreError: 13717:0:(osp_sync.c:1079:osp_sync_process_committed()) Skipped 7 previous similar messages | Externally reported by onyx-68 boilpot email |
racer test 1: racer on clients: centos-105.localnet DURATION=2700 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa1037e52>] ll_migrate+0x9b2/0xec0 [lustre] PGD 8000000259822067 PUD 259823067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: loop zfs(PO) zunicode(PO) zzstd(O) zlua(O) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) jbd2 mbcache lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) dm_flakey dm_mod libcfs(OE) crc_t10dif crct10dif_generic sb_edac edac_core iosf_mbi crc32_pclmul ghash_clmulni_intel aesni_intel lrw gf128mul glue_helper ablk_helper cryptd pcspkr virtio_balloon virtio_console i2c_piix4 ip_tables rpcsec_gss_krb5 ata_generic drm_kms_helper pata_acpi ttm crct10dif_pclmul crct10dif_common drm ata_piix crc32c_intel drm_panel_orientation_quirks serio_raw virtio_blk libata i2c_core floppy CPU: 0 PID: 32638 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011 task: ffff88027f4bdc40 ti: ffff8802762c8000 task.ti: ffff8802762c8000 RIP: 0010:[<ffffffffa1037e52>] [<ffffffffa1037e52>] ll_migrate+0x9b2/0xec0 [lustre] RSP: 0018:ffff8802762cbbc8 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff8802350f5a98 RCX: 0000000000000000 RDX: 0000000100006c53 RSI: ffff880326c86fb8 RDI: ffff880326c86f98 RBP: ffff8802762cbc48 R08: ffff880270fffa98 R09: ffff8802762cba60 R10: 000000000000000a R11: ffff880260bcb1b8 R12: 0000000000000000 R13: ffff8802529f31b8 R14: ffff8800739f9398 R15: 0000000000000030 FS: 00007f88063a5740(0000) GS:ffff880331a00000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 CR2: 0000000000000008 CR3: 0000000276768000 CR4: 00000000001607f0 Call Trace: [<ffffffff81242213>] ? __check_object_size+0x1c3/0x220 [<ffffffffa1025561>] ll_dir_ioctl+0x5d01/0x6ed0 [lustre] [<ffffffff811b515b>] ? unlock_page+0x2b/0x30 [<ffffffff811e5e2e>] ? do_read_fault.isra.64+0x12e/0x170 [<ffffffff811ea1cf>] ? handle_pte_fault+0x2ef/0xca0 [<ffffffff810d64a6>] ? select_task_rq_fair+0x5c6/0x780 [<ffffffff811ecf6d>] ? __handle_mm_fault+0x31d/0x720 [<ffffffff8125b3fd>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff81245a38>] ? vfs_write+0x178/0x230 [<ffffffff8125b751>] SyS_ioctl+0xa1/0xc0 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 | Lustre: lustre-OST0001-osc-ffff88008f3a1bf8: disconnect after 21s idle Lustre: 11169:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 508 < left 11061, rollback = 7 Lustre: 11169:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 257/1028/0, destroy: 0/0/0 Lustre: 11169:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 259/115/4 Lustre: 11169:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1286/11061/0, punch: 0/0/0, quota 6/6/0 Lustre: 11169:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 257/4369/0, delete: 0/0/0 Lustre: 11169:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 cp (16523) used greatest stack depth: 10080 bytes left Lustre: 11169:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 507 < left 11060, rollback = 7 Lustre: 11169:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 257/1028/0, destroy: 0/0/0 Lustre: 11169:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 259/115/4 Lustre: 11169:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1286/11061/1, punch: 0/0/0, quota 6/6/0 Lustre: 11169:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 257/4369/0, delete: 0/0/0 Lustre: 11169:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 11169:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 496 < left 11049, rollback = 7 Lustre: 11169:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 321 previous similar messages Lustre: 11169:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 257/1028/0, destroy: 0/0/0 Lustre: 11169:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 321 previous similar messages Lustre: 11169:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 259/115/4 Lustre: 11169:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 321 previous similar messages Lustre: 11169:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1286/11061/12, punch: 0/0/0, quota 6/6/0 Lustre: 11169:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 321 previous similar messages Lustre: 11169:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 257/4369/0, delete: 0/0/0 Lustre: 11169:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 321 previous similar messages Lustre: 11169:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 11169:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 321 previous similar messages Lustre: 11169:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 494 < left 11047, rollback = 7 Lustre: 11169:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 534 previous similar messages Lustre: 11169:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 257/1028/0, destroy: 0/0/0 Lustre: 11169:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 534 previous similar messages Lustre: 11169:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 259/115/4 Lustre: 11169:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 534 previous similar messages Lustre: 11169:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1286/11061/14, punch: 0/0/0, quota 6/6/0 Lustre: 11169:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 534 previous similar messages Lustre: 11169:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 257/4369/0, delete: 0/0/0 Lustre: 11169:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 534 previous similar messages Lustre: 11169:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 11169:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 534 previous similar messages Lustre: 16790:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 512 < left 1695, rollback = 7 Lustre: 16790:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 1632 previous similar messages Lustre: 16790:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 16790:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 1632 previous similar messages Lustre: 16790:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 1/89/0 Lustre: 16790:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 1632 previous similar messages Lustre: 16790:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1686/1695/0, punch: 0/0/0, quota 3/3/0 Lustre: 16790:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 1632 previous similar messages Lustre: 16790:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 16790:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 1632 previous similar messages Lustre: 16790:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 16790:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 1632 previous similar messages Lustre: 13403:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 505 < left 671, rollback = 2 Lustre: 13403:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 3/12/4, destroy: 0/0/0 Lustre: 13403:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 7398 previous similar messages Lustre: 13403:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 7/671/0 Lustre: 13403:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 7398 previous similar messages Lustre: 13403:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 5/136/0, punch: 0/0/0, quota 8/56/0 Lustre: 13403:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 7398 previous similar messages Lustre: 13403:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 10/203/3, delete: 0/0/0 Lustre: 13403:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 7398 previous similar messages Lustre: 13403:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 4/4/0, ref_del: 0/0/0 Lustre: 13403:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 7398 previous similar messages Lustre: 17935:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 501 < left 1072, rollback = 2 Lustre: 17935:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 7 previous similar messages Lustre: 11149:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 512 < left 4095, rollback = 7 Lustre: 11149:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 7398 previous similar messages Lustre: 11152:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 506 < left 1044, rollback = 2 Lustre: 11152:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 11 previous similar messages ------------[ cut here ]------------ WARNING: CPU: 5 PID: 16796 at /home/green/git/lustre-release/ldiskfs/namei.c:3331 ldiskfs_orphan_add+0x11e/0x2a0 [ldiskfs] Modules linked in: loop zfs(PO) zunicode(PO) zzstd(O) zlua(O) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) jbd2 mbcache lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) dm_flakey dm_mod libcfs(OE) crc_t10dif crct10dif_generic sb_edac edac_core iosf_mbi crc32_pclmul ghash_clmulni_intel aesni_intel lrw gf128mul glue_helper ablk_helper cryptd pcspkr virtio_balloon virtio_console i2c_piix4 ip_tables rpcsec_gss_krb5 ata_generic drm_kms_helper pata_acpi ttm crct10dif_pclmul crct10dif_common drm ata_piix crc32c_intel drm_panel_orientation_quirks serio_raw virtio_blk libata i2c_core floppy CPU: 5 PID: 16796 Comm: mdt02_006 Tainted: P OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011 Call Trace: [<ffffffff817d9418>] dump_stack+0x19/0x1b [<ffffffff8108d538>] __warn+0xd8/0x100 [<ffffffff8108d67d>] warn_slowpath_null+0x1d/0x20 [<ffffffffa0bcf4ae>] ldiskfs_orphan_add+0x11e/0x2a0 [ldiskfs] [<ffffffffa0bc09eb>] ldiskfs_xattr_inode_orphan_add.constprop.17+0xbb/0x110 [ldiskfs] [<ffffffffa0bc3417>] ldiskfs_xattr_delete_inode+0x77/0x3c0 [ldiskfs] [<ffffffffa0be3aa0>] ? ldiskfs_evict_inode+0x190/0x670 [ldiskfs] [<ffffffffa0be3ad9>] ldiskfs_evict_inode+0x1c9/0x670 [ldiskfs] [<ffffffff812649af>] evict+0xaf/0x180 [<ffffffff81264de5>] iput+0xf5/0x180 [<ffffffffa0c78d08>] osd_object_delete+0x1f8/0x370 [osd_ldiskfs] [<ffffffffa04ca59d>] lu_object_free.isra.26+0xcd/0x1d0 [obdclass] [<ffffffffa04cecf1>] lu_object_put+0xa1/0x480 [obdclass] [<ffffffffa0e17440>] mdt_object_put+0x30/0x110 [mdt] [<ffffffffa0e1ee18>] mdt_reint_unlink+0x918/0x19b0 [mdt] [<ffffffffa04f079e>] ? lu_ucred+0x1e/0x30 [obdclass] [<ffffffffa0e26087>] mdt_reint_rec+0x87/0x240 [mdt] [<ffffffffa0dfb55c>] mdt_reint_internal+0x76c/0xb40 [mdt] [<ffffffffa0e08657>] mdt_reint+0x67/0x150 [mdt] [<ffffffffa07d226e>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<ffffffffa032945e>] ? libcfs_nid2str_r+0xfe/0x130 [lnet] [<ffffffffa077d990>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<ffffffffa077f529>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e32be>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffffa077e950>] ? ptlrpc_wait_event+0x620/0x620 [ptlrpc] [<ffffffff810ba0f4>] kthread+0xe4/0xf0 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 ---[ end trace 6d269e43193aaa7d ]--- Lustre: 11155:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 65/260/0, destroy: 0/0/0 Lustre: 11155:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 6219 previous similar messages Lustre: 11155:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/1, xattr_set: 67/15/0 Lustre: 11155:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 6219 previous similar messages Lustre: 11155:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 326/2805/0, punch: 0/0/0, quota 10/106/7 Lustre: 11155:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 6219 previous similar messages Lustre: 11155:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 65/1105/0, delete: 0/0/0 Lustre: 11155:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 6219 previous similar messages Lustre: 11155:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 11155:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 6219 previous similar messages Lustre: 16583:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 496 < left 671, rollback = 2 Lustre: 16583:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 11 previous similar messages sched: RT throttling activated Lustre: 16602:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 507 < left 11061, rollback = 7 Lustre: 16602:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 6453 previous similar messages 1[19061]: segfault at 0 ip (null) sp 00007ffd72083d78 error 14 in 1[400000+6000] Lustre: ll_ost01_007: service thread pid 18089 was inactive for 40.111 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 18089, comm: ll_ost01_007 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] tgt_extent_lock+0xea/0x2a0 [ptlrpc] [<0>] ofd_getattr_hdl+0x365/0x720 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 12669, comm: ll_ost01_000 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] tgt_extent_lock+0xea/0x2a0 [ptlrpc] [<0>] ofd_getattr_hdl+0x365/0x720 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: ll_ost01_004: service thread pid 18004 was inactive for 40.757 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Lustre: Skipped 1 previous similar message Pid: 18004, comm: ll_ost01_004 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] tgt_extent_lock+0xea/0x2a0 [ptlrpc] [<0>] ofd_getattr_hdl+0x365/0x720 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: ll_ost01_006: service thread pid 18085 was inactive for 40.914 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: 11156:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 8: before 509 < left 16166, rollback = 9 Lustre: 11156:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 7 previous similar messages Lustre: 18307:0:(service.c:1378:ptlrpc_at_send_early_reply()) @@@ Already past deadline (-1s), not sending early reply. Consider increasing at_early_margin (5)? req@ffff880084bc3880 x1710022310609600/t0(0) o101->ca1369a5-6907-43ae-a0ae-a780fc0636a7@0@lo:54/0 lens 328/0 e 0 to 0 dl 1630804584 ref 2 fl New:/0/ffffffff rc 0/-1 job:'getfattr.0' Lustre: 18316:0:(service.c:1378:ptlrpc_at_send_early_reply()) @@@ Already past deadline (-1s), not sending early reply. Consider increasing at_early_margin (5)? req@ffff880275bf7700 x1710022310668544/t0(0) o101->ca1369a5-6907-43ae-a0ae-a780fc0636a7@0@lo:55/0 lens 328/0 e 0 to 0 dl 1630804585 ref 2 fl New:/0/ffffffff rc 0/-1 job:'getfattr.0' Lustre: 18316:0:(service.c:1378:ptlrpc_at_send_early_reply()) Skipped 74 previous similar messages Lustre: 18316:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 6s req@ffff880076b18680 x1710022312364416/t0(0) o101->1fc751ea-90e3-421e-8d25-6e7f755f5b87@0@lo:0/0 lens 328/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'ln.0' Lustre: 4135:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1630804541/real 1630804542] req@ffff88008eb170c0 x1710022310640640/t0(0) o101->lustre-OST0002-osc-ffff88008f3a1bf8@0@lo:28/4 lens 328/400 e 0 to 1 dl 1630804590 ref 1 fl Rpc:XQr/0/ffffffff rc 0/-1 job:'getfattr.0' Lustre: lustre-OST0002-osc-ffff88008f3a1bf8: Connection to lustre-OST0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: lustre-OST0000: Client ca1369a5-6907-43ae-a0ae-a780fc0636a7 (at 0@lo) reconnecting Lustre: lustre-OST0000-osc-ffff88008f3a1bf8: Connection restored to 192.168.123.205@tcp (at 0@lo) Lustre: 17023:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 506 < left 535, rollback = 2 Lustre: 14643:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 257/1028/0, destroy: 0/0/0 Lustre: 14643:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 1311 previous similar messages Lustre: 14643:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/1, xattr_set: 259/15/0 Lustre: 14643:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 1311 previous similar messages Lustre: 14643:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1286/11061/0, punch: 0/0/0, quota 14/174/3 Lustre: 14643:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 1311 previous similar messages Lustre: 14643:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 257/4369/0, delete: 0/0/0 Lustre: 14643:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 1311 previous similar messages Lustre: 14643:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 14643:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 1311 previous similar messages Lustre: ll_ost01_018: service thread pid 18154 was inactive for 62.205 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 14 previous similar messages LustreError: 16527:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0002: '1' migration was interrupted, run 'lfs migrate -m 0 -c 1 -H crush 1' to finish migration. Lustre: ll_ost01_034: service thread pid 18225 was inactive for 62.055 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 11 previous similar messages Lustre: ll_ost01_051: service thread pid 18344 was inactive for 62.028 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 15 previous similar messages Lustre: 11163:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 508 < left 72508, rollback = 7 Lustre: 11163:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 8429 previous similar messages Lustre: ll_ost01_059: service thread pid 18551 was inactive for 62.021 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 9 previous similar messages Lustre: 4136:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1630804541/real 1630804541] req@ffff88008490d7c0 x1710022310642432/t0(0) o101->lustre-OST0001-osc-ffff88008f3a1bf8@0@lo:28/4 lens 328/400 e 0 to 1 dl 1630804591 ref 1 fl Rpc:XQr/0/ffffffff rc 0/-1 job:'getfattr.0' Lustre: 4136:0:(client.c:2285:ptlrpc_expire_one_request()) Skipped 60 previous similar messages Lustre: lustre-OST0001-osc-ffff88008f3a1bf8: Connection to lustre-OST0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 1 previous similar message Lustre: lustre-OST0001: Client ca1369a5-6907-43ae-a0ae-a780fc0636a7 (at 0@lo) reconnecting Lustre: Skipped 1 previous similar message Lustre: lustre-OST0001-osc-ffff88008f3a1bf8: Connection restored to 192.168.123.205@tcp (at 0@lo) Lustre: Skipped 1 previous similar message LustreError: 11033:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0003_UUID lock: ffff880289114b40/0x9a131711b0f40568 lrc: 3/0,0 mode: PW/PW res: [0x380000400:0x3:0x0].0x0 rrc: 5 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000480020020 nid: 0@lo remote: 0x9a131711b0f40561 expref: 3189 pid: 12685 timeout: 385 lvb_type: 0 Lustre: lustre-OST0003-osc-ffff88008f3a1bf8: Connection to lustre-OST0003 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 1 previous similar message LustreError: 167-0: lustre-OST0003-osc-ffff88008f3a1bf8: This client was evicted by lustre-OST0003; in progress operations using this service will fail. Lustre: 18143:0:(service.c:1378:ptlrpc_at_send_early_reply()) @@@ Already past deadline (-1s), not sending early reply. Consider increasing at_early_margin (5)? req@ffff880075238680 x1710022313081856/t0(0) o101->1fc751ea-90e3-421e-8d25-6e7f755f5b87@0@lo:104/0 lens 328/0 e 0 to 0 dl 1630804634 ref 2 fl New:/0/ffffffff rc 0/-1 job:'chmod.0' Lustre: 18143:0:(service.c:1378:ptlrpc_at_send_early_reply()) Skipped 119 previous similar messages Lustre: 18247:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 507 < left 1222, rollback = 2 Lustre: 18247:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 14 previous similar messages LustreError: 18221:0:(service.c:2290:ptlrpc_server_handle_request()) @@@ Dropping timed-out request from 12345-0@lo: deadline 43/2s ago req@ffff88024f985180 x1710022313070336/t0(0) o101->1fc751ea-90e3-421e-8d25-6e7f755f5b87@0@lo:104/0 lens 328/0 e 0 to 0 dl 1630804634 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'chmod.0' Lustre: 18221:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (43/2s); client may timeout req@ffff88024f985180 x1710022313070336/t0(0) o101->1fc751ea-90e3-421e-8d25-6e7f755f5b87@0@lo:104/0 lens 328/0 e 0 to 0 dl 1630804634 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'chmod.0' LustreError: 11033:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0002_UUID lock: ffff88008bffad40/0x9a131711b0f411d2 lrc: 3/0,0 mode: PW/PW res: [0x340000400:0x6:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000480020020 nid: 0@lo remote: 0x9a131711b0f411c4 expref: 2330 pid: 12678 timeout: 397 lvb_type: 0 LustreError: 17922:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88024c9d8040 x1710022314476672/t0(0) o104->lustre-OST0000@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 11-0: lustre-OST0000-osc-ffff88008f3a1bf8: operation ost_setattr to node 0@lo failed: rc = -107 Lustre: lustre-OST0000-osc-ffff88008f3a1bf8: Connection to lustre-OST0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0000-osc-ffff88008f3a1bf8: This client was evicted by lustre-OST0000; in progress operations using this service will fail. LustreError: 18201:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88024de03240 x1710022314492736/t0(0) o106->lustre-OST0000@0@lo:15/16 lens 328/280 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 18201:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 28 previous similar messages LustreError: 19702:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630804640 with bad export cookie 11102242871149054265 LustreError: 19702:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) Skipped 4 previous similar messages LustreError: 11-0: lustre-OST0002-osc-ffff88008f3a1bf8: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: 12680:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff880084419940 x1710022314534592/t0(0) o106->lustre-OST0001@0@lo:15/16 lens 328/280 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 12680:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 57 previous similar messages Lustre: 4141:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.205@tcp:/lustre/fid: [0x200000404:0x28:0x0]/ may get corrupted (rc -108) Lustre: 4142:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.205@tcp:/lustre/fid: [0x280000403:0x3c:0x0]/ may get corrupted (rc -108) Lustre: 4142:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.205@tcp:/lustre/fid: [0x200000403:0x3d:0x0]// may get corrupted (rc -108) LustreError: 19952:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0003-osc-ffff88008f3a1bf8: namespace resource [0x10:0x0:0x0].0x0 (ffff8800855dca40) refcount nonzero (1) after lock cleanup; forcing cleanup. Lustre: lustre-OST0003-osc-ffff88008f3a1bf8: Connection restored to 192.168.123.205@tcp (at 0@lo) Lustre: Skipped 1 previous similar message LustreError: 19068:0:(vvp_io.c:1793:vvp_io_init()) lustre: refresh file layout [0x200000404:0x28:0x0] error -108. Lustre: 4139:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.205@tcp:/lustre/fid: [0x280000403:0x9e:0x0]// may get corrupted (rc -108) Lustre: 4140:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.205@tcp:/lustre/fid: [0x200000403:0x46:0x0]// may get corrupted (rc -108) Lustre: 4139:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.205@tcp:/lustre/fid: [0x280000403:0x90:0x0]/ may get corrupted (rc -108) Lustre: 4136:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1630804591/real 1630804591] req@ffff880262340680 x1710022313070272/t0(0) o101->lustre-OST0002-osc-ffff880298d3c138@0@lo:28/4 lens 328/400 e 0 to 1 dl 1630804646 ref 1 fl Rpc:XQr/0/ffffffff rc 0/-1 job:'chmod.0' Lustre: 4136:0:(client.c:2285:ptlrpc_expire_one_request()) Skipped 60 previous similar messages Lustre: lustre-OST0002-osc-ffff880298d3c138: Connection to lustre-OST0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 2 previous similar messages Lustre: lustre-OST0002: Client 1fc751ea-90e3-421e-8d25-6e7f755f5b87 (at 0@lo) reconnecting Lustre: Skipped 1 previous similar message Lustre: lustre-OST0002-osc-ffff880298d3c138: Connection restored to 192.168.123.205@tcp (at 0@lo) Lustre: 4136:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1630804591/real 1630804591] req@ffff880088991940 x1710022313070464/t0(0) o101->lustre-OST0000-osc-ffff880298d3c138@0@lo:28/4 lens 328/400 e 0 to 1 dl 1630804646 ref 1 fl Rpc:XQr/0/ffffffff rc 0/-1 job:'chmod.0' Lustre: lustre-OST0003: Client 1fc751ea-90e3-421e-8d25-6e7f755f5b87 (at 0@lo) reconnecting Lustre: Skipped 1 previous similar message Lustre: 11162:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 1587/6348/0, destroy: 1/4/0 Lustre: 11162:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 20724 previous similar messages Lustre: 11162:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 3/3/0, xattr_set: 1591/148/0 Lustre: 11162:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 20724 previous similar messages Lustre: 11162:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 7936/68251/0, punch: 0/0/0, quota 10/74/0 Lustre: 11162:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 20724 previous similar messages Lustre: 11162:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 1588/26995/0, delete: 2/5/0 Lustre: 11162:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 20724 previous similar messages Lustre: 11162:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 1/1/0, ref_del: 2/2/1 Lustre: 11162:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 20724 previous similar messages LustreError: 11033:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0001_UUID lock: ffff880263bee1c0/0x9a131711b0f7eafd lrc: 3/0,0 mode: PW/PW res: [0xf2:0x0:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000400000020 nid: 0@lo remote: 0x9a131711b0f7ea8d expref: 2016 pid: 12675 timeout: 428 lvb_type: 0 LustreError: 11033:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 2 previous similar messages LustreError: 11904:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff880088991940 x1710022315531904/t0(0) o105->lustre-OST0001@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 11904:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 20 previous similar messages LustreError: 11-0: lustre-OST0001-osc-ffff880298d3c138: operation ost_setattr to node 0@lo failed: rc = -107 LustreError: Skipped 6 previous similar messages Lustre: lustre-OST0001-osc-ffff880298d3c138: Connection to lustre-OST0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 3 previous similar messages LustreError: 167-0: lustre-OST0001-osc-ffff880298d3c138: This client was evicted by lustre-OST0001; in progress operations using this service will fail. LustreError: Skipped 2 previous similar messages Lustre: 11013:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 6s req@ffff88025016e440 x1710022315374208/t0(0) o103->1fc751ea-90e3-421e-8d25-6e7f755f5b87@0@lo:0/0 lens 328/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'ldlm_bl_03.0' Lustre: 11013:0:(service.c:2157:ptlrpc_server_handle_req_in()) Skipped 31 previous similar messages Lustre: 4133:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.205@tcp:/lustre/fid: [0x280000404:0x1:0x0]// may get corrupted (rc -108) Lustre: 4134:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.205@tcp:/lustre/fid: [0x200000404:0x5e:0x0]// may get corrupted (rc -108) Lustre: 11187:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 470 < left 72469, rollback = 7 Lustre: 11187:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 20851 previous similar messages Lustre: 17869:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 8: before 511 < left 865, rollback = 9 Lustre: 17869:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 18 previous similar messages Lustre: 12757:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 6s req@ffff8802751ef0c0 x1710022316029952/t0(0) o4->1fc751ea-90e3-421e-8d25-6e7f755f5b87@0@lo:0/0 lens 488/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'cp.0' Lustre: 12757:0:(service.c:2157:ptlrpc_server_handle_req_in()) Skipped 1 previous similar message Lustre: ost_create: This server is not able to keep up with request traffic (cpu-bound). Lustre: 12705:0:(service.c:1612:ptlrpc_at_check_timed()) earlyQ=1 reqQ=0 recA=0, svcEst=1, delay=0ms Lustre: 12705:0:(service.c:1378:ptlrpc_at_send_early_reply()) @@@ Already past deadline (-1s), not sending early reply. Consider increasing at_early_margin (5)? req@ffff88027711b240 x1710022316055168/t0(0) o13->lustre-MDT0002-mdtlov_UUID@0@lo:161/0 lens 224/0 e 0 to 0 dl 1630804691 ref 2 fl New:/0/ffffffff rc 0/-1 job:'osp-pre-0-2.0' Lustre: 12705:0:(service.c:1378:ptlrpc_at_send_early_reply()) Skipped 115 previous similar messages LustreError: 12705:0:(service.c:2290:ptlrpc_server_handle_request()) @@@ Dropping timed-out request from 12345-0@lo: deadline 6/1s ago req@ffff88027711b240 x1710022316055168/t0(0) o13->lustre-MDT0002-mdtlov_UUID@0@lo:161/0 lens 224/0 e 0 to 0 dl 1630804691 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'osp-pre-0-2.0' LustreError: 12705:0:(service.c:2290:ptlrpc_server_handle_request()) Skipped 115 previous similar messages Lustre: 12705:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (6/1s); client may timeout req@ffff88027711b240 x1710022316055168/t0(0) o13->lustre-MDT0002-mdtlov_UUID@0@lo:161/0 lens 224/0 e 0 to 0 dl 1630804691 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'osp-pre-0-2.0' Lustre: 12705:0:(service.c:2328:ptlrpc_server_handle_request()) Skipped 115 previous similar messages LustreError: 20339:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0000-osc-ffff88008f3a1bf8: namespace resource [0x130:0x0:0x0].0x0 (ffff880082d8d440) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 20339:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 1331 previous similar messages Lustre: lustre-OST0000-osc-ffff88008f3a1bf8: Connection restored to 192.168.123.205@tcp (at 0@lo) Lustre: Skipped 3 previous similar messages Lustre: 11015:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 7s req@ffff88024bc53ec0 x1710022316317056/t0(0) o103->ca1369a5-6907-43ae-a0ae-a780fc0636a7@0@lo:0/0 lens 328/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'ldlm_bl_05.0' Lustre: 11015:0:(service.c:2157:ptlrpc_server_handle_req_in()) Skipped 3 previous similar messages LustreError: 20393:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0001-osc-ffff88008f3a1bf8: namespace resource [0x300000400:0x50:0x0].0x0 (ffff880278c51940) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 20393:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 445 previous similar messages Lustre: lustre-OST0001-osc-ffff88008f3a1bf8: Connection restored to 192.168.123.205@tcp (at 0@lo) LustreError: 21149:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0001-osc-ffff880298d3c138: namespace resource [0x140:0x0:0x0].0x0 (ffff880076ca0a40) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 21149:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 316 previous similar messages Lustre: dir [0x200000404:0x90:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: 16521:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 499 < left 1756, rollback = 2 Lustre: 16521:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 6 previous similar messages Lustre: dir [0x200000404:0x26:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: 4141:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1630804685/real 1630804685] req@ffff88027759f700 x1710022316055168/t0(0) o13->lustre-OST0000-osc-MDT0002@0@lo:7/4 lens 224/368 e 0 to 1 dl 1630804729 ref 1 fl Rpc:XQr/0/ffffffff rc 0/-1 job:'osp-pre-0-2.0' Lustre: 4141:0:(client.c:2285:ptlrpc_expire_one_request()) Skipped 14 previous similar messages Lustre: lustre-OST0000-osc-MDT0002: Connection to lustre-OST0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: lustre-OST0000: Client lustre-MDT0002-mdtlov_UUID (at 0@lo) reconnecting Lustre: Skipped 1 previous similar message Lustre: lustre-OST0000-osc-MDT0002: Connection restored to 192.168.123.205@tcp (at 0@lo) Lustre: Skipped 2 previous similar messages Lustre: lustre-OST0000: deleting orphan objects from 0x2c0000400:1550 to 0x2c0000400:1569 Lustre: 11154:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 158/632/0, destroy: 0/0/0 Lustre: 11154:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 59725 previous similar messages Lustre: 11154:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/1, xattr_set: 160/15/0 Lustre: 11154:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 59725 previous similar messages Lustre: 11154:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 791/6804/0, punch: 0/0/0, quota 14/142/5 Lustre: 11154:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 59725 previous similar messages Lustre: 11154:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 158/2686/0, delete: 0/0/0 Lustre: 11154:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 59725 previous similar messages Lustre: 11154:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 11154:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 59725 previous similar messages LustreError: 11197:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0001: failed to get lu_attr of [0x240000404:0x105:0x0]: rc = -2 Lustre: 16593:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 511 < left 1863, rollback = 7 Lustre: 16593:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 53409 previous similar messages LustreError: 11184:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0001: failed to get lu_attr of [0x240000404:0x6d:0x0]: rc = -2 Lustre: 4136:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff88024ebc5e00 x1710022322154368/t0(0) o10->lustre-OST0003-osc-ffff880298d3c138@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630804900 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'cp.0' LustreError: 22921:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0001: '18' migration was interrupted, run 'lfs migrate -m 2 -c 1 -H crush 18' to finish migration. LustreError: 11190:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0002: failed to get lu_attr of [0x280000404:0x164:0x0]: rc = -2 Lustre: 11159:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 495 < left 966, rollback = 2 Lustre: 11159:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 104 previous similar messages LustreError: 11033:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0001_UUID lock: ffff880084282d40/0x9a131711b102ff93 lrc: 3/0,0 mode: PW/PW res: [0x3f5:0x0:0x0].0x0 rrc: 4 type: EXT [0->18446744073709551615] (req 0->524287) gid 0 flags: 0x60000400020020 nid: 0@lo remote: 0x9a131711b102ff8c expref: 3849 pid: 18271 timeout: 695 lvb_type: 0 LustreError: 11-0: lustre-OST0001-osc-ffff880298d3c138: operation ost_punch to node 0@lo failed: rc = -107 Lustre: lustre-OST0001-osc-ffff880298d3c138: Connection to lustre-OST0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0001-osc-ffff880298d3c138: This client was evicted by lustre-OST0001; in progress operations using this service will fail. LustreError: 11018:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630804938 with bad export cookie 11102242871149046278 LustreError: 20379:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff880276d57700 x1710022327693568/t0(0) o105->lustre-OST0003@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 20379:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 1 previous similar message Lustre: 4136:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.205@tcp:/lustre/fid: [0x240000404:0x175:0x0]// may get corrupted (rc -108) Lustre: 4142:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.205@tcp:/lustre/fid: [0x240000404:0x16a:0x0]/ may get corrupted (rc -108) Lustre: 4141:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.205@tcp:/lustre/fid: [0x200000403:0x158:0x0]/ may get corrupted (rc -108) Lustre: 4141:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.205@tcp:/lustre/fid: [0x280000404:0x20f:0x0]// may get corrupted (rc -108) Lustre: lustre-OST0000-osc-ffff880298d3c138: Connection restored to 192.168.123.205@tcp (at 0@lo) LustreError: 26510:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0003-osc-ffff880298d3c138: namespace resource [0x380000401:0x1d2:0x0].0x0 (ffff88008ac75940) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 26510:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 1802 previous similar messages Lustre: 4142:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.205@tcp:/lustre/fid: [0x240000403:0x131:0x0]/ may get corrupted (rc -108) Lustre: 14643:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000403:0x148:0x0] with magic=0xbd60bd0 Lustre: 11149:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000403:0x149:0x0] with magic=0xbd60bd0 Lustre: 11149:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 1 previous similar message Lustre: 17009:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000403:0x14a:0x0] with magic=0xbd60bd0 Lustre: 17009:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 1 previous similar message LustreError: 11296:0:(out_handler.c:910:out_tx_end()) lustre-MDT0001-osd: undo for /home/green/git/lustre-release/lustre/ptlrpc/../../lustre/target/out_handler.c:445: rc = -524 LustreError: 12312:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0001-osp-MDT0002: fail to cancel 1 llog-records: rc = -116 LustreError: 12312:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0001-osp-MDT0002: fail to cancel 1 of 1 llog-records: rc = -116 LustreError: 12312:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0001-osp-MDT0002: fail to cancel 1 llog-records: rc = -116 LustreError: 12312:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) Skipped 1 previous similar message LustreError: 12312:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0001-osp-MDT0002: fail to cancel 1 of 1 llog-records: rc = -116 LustreError: 12312:0:(llog_cat.c:787:llog_cat_cancel_records()) Skipped 1 previous similar message LustreError: 27458:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff88008f3a1bf8: inode [0x280000404:0x1cd:0x0] mdc close failed: rc = -13 LustreError: 11927:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0002-osp-MDT0001: fail to cancel 1 llog-records: rc = -2 LustreError: 11927:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) Skipped 2 previous similar messages LustreError: 11927:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0002-osp-MDT0001: fail to cancel 1 of 1 llog-records: rc = -2 LustreError: 11927:0:(llog_cat.c:787:llog_cat_cancel_records()) Skipped 2 previous similar messages LustreError: 11033:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0001_UUID lock: ffff8802650dad40/0x9a131711b107a5cb lrc: 3/0,0 mode: PW/PW res: [0x3f5:0x0:0x0].0x0 rrc: 2 type: EXT [0->18446744073709551615] (req 0->524287) gid 0 flags: 0x60000000030020 nid: 0@lo remote: 0x9a131711b107a4ba expref: 2944 pid: 18182 timeout: 796 lvb_type: 0 LustreError: 11033:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 2 previous similar messages LustreError: 21560:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630805038 with bad export cookie 11102242871149599495 LustreError: 12677:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff880076a1b880 x1710022332480192/t0(0) o104->lustre-OST0001@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 12677:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 7 previous similar messages LustreError: 11-0: lustre-OST0001-osc-ffff88008f3a1bf8: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 30 previous similar messages Lustre: lustre-OST0001-osc-ffff88008f3a1bf8: Connection to lustre-OST0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 2 previous similar messages LustreError: 167-0: lustre-OST0001-osc-ffff88008f3a1bf8: This client was evicted by lustre-OST0001; in progress operations using this service will fail. LustreError: Skipped 2 previous similar messages Lustre: 18646:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 1853/7412/0, destroy: 1/4/0 Lustre: 18646:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 108597 previous similar messages Lustre: 18646:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 1853/0/0 Lustre: 18646:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 108626 previous similar messages Lustre: 18646:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 9266/79689/25, punch: 0/0/0, quota 8/40/0 Lustre: 27789:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 1853/31501/0, delete: 2/5/1 Lustre: 27789:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 108626 previous similar messages Lustre: 27789:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 1/1/1 Lustre: 27789:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 108626 previous similar messages Lustre: 18646:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 108924 previous similar messages Lustre: lustre-OST0001-osc-ffff88008f3a1bf8: Connection restored to 192.168.123.205@tcp (at 0@lo) Lustre: Skipped 2 previous similar messages LustreError: 28601:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0000-osc-ffff88008f3a1bf8: namespace resource [0x2c0000400:0x5a0:0x0].0x0 (ffff880287ee5e40) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 28601:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 100 previous similar messages Lustre: 18128:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 510 < left 68208, rollback = 7 Lustre: 18128:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 131305 previous similar messages LustreError: 11033:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0003_UUID lock: ffff88007771da40/0x9a131711b1117a4b lrc: 3/0,0 mode: PW/PW res: [0x380000401:0x1ce:0x0].0x0 rrc: 4 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000400020020 nid: 0@lo remote: 0x9a131711b1117a44 expref: 2286 pid: 17641 timeout: 898 lvb_type: 0 LustreError: 11033:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 1 previous similar message LustreError: 11032:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff880282c3cb40 x1710022336512256/t0(0) o105->lustre-OST0003@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 11032:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 34 previous similar messages LustreError: 11-0: lustre-OST0003-osc-ffff880298d3c138: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 1 previous similar message LustreError: 167-0: lustre-OST0003-osc-ffff880298d3c138: This client was evicted by lustre-OST0003; in progress operations using this service will fail. LustreError: Skipped 1 previous similar message LustreError: 20125:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630805141 with bad export cookie 11102242871150783545 LustreError: 20125:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) Skipped 1 previous similar message LustreError: 17808:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0001: '18' migration was interrupted, run 'lfs migrate -m 0 -c 1 -H crush 18' to finish migration. LustreError: 30519:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff880298d3c138: inode [0x280000404:0x1cd:0x0] mdc close failed: rc = -13 Lustre: 4141:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.205@tcp:/lustre/fid: [0x240000404:0x18a:0x0]// may get corrupted (rc -108) LustreError: 30283:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0003-osc-ffff880298d3c138: namespace resource [0x380000401:0x1a9f:0x0].0x0 (ffff880076020540) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 30283:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 108 previous similar messages Lustre: 17060:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 504 < left 1934, rollback = 2 Lustre: 17060:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 194 previous similar messages LustreError: 23959:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0001: '18' migration was interrupted, run 'lfs migrate -m 0 -c 1 -H crush 18' to finish migration. Lustre: 25951:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 8s req@ffff880322329940 x1710022338557056/t0(0) o2->lustre-MDT0001-mdtlov_UUID@0@lo:0/0 lens 440/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'osp-syn-3-1.0' Lustre: 25951:0:(service.c:2157:ptlrpc_server_handle_req_in()) Skipped 3 previous similar messages INFO: task mv:27865 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mv D ffff880270ab24f0 11216 27865 16497 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff81251fe1>] lock_rename+0x31/0xe0 [<ffffffff8125862f>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed432>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817e8a76>] ? trace_do_page_fault+0x56/0x170 [<ffffffff8125975e>] SyS_renameat2+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task mv:28143 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mv D ffff88028c5be340 11344 28143 16519 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff81251fe1>] lock_rename+0x31/0xe0 [<ffffffff8125862f>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed432>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817e8a76>] ? trace_do_page_fault+0x56/0x170 [<ffffffff8125975e>] SyS_renameat2+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task mv:28361 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mv D ffff8803291924f0 11344 28361 16242 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff81251fe1>] lock_rename+0x31/0xe0 [<ffffffff8125862f>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed432>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817e8a76>] ? trace_do_page_fault+0x56/0x170 [<ffffffff8125975e>] SyS_renameat2+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task mv:28864 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mv D ffff880294e11280 11344 28864 15973 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff81251fe1>] lock_rename+0x31/0xe0 [<ffffffff8125862f>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed432>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817e8a76>] ? trace_do_page_fault+0x56/0x170 [<ffffffff8125975e>] SyS_renameat2+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task mv:29431 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mv D ffff880260bd24f0 12128 29431 16557 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff81251fe1>] lock_rename+0x31/0xe0 [<ffffffff8125862f>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed432>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817e8a76>] ? trace_do_page_fault+0x56/0x170 [<ffffffff8125975e>] SyS_renameat2+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 Lustre: 11017:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 6s req@ffff880259faf0c0 x1710022339024512/t0(0) o103->1fc751ea-90e3-421e-8d25-6e7f755f5b87@0@lo:0/0 lens 328/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'ldlm_bl_01.0' Lustre: 11017:0:(service.c:2157:ptlrpc_server_handle_req_in()) Skipped 3 previous similar messages LustreError: 11159:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0001: '4' migration was interrupted, run 'lfs migrate -m 2 -c 1 -H crush 4' to finish migration. LustreError: 11033:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 128s: evicting client at 0@lo ns: filter-lustre-OST0000_UUID lock: ffff88024b2796c0/0x9a131711b119b097 lrc: 3/0,0 mode: PW/PW res: [0x2c0000401:0x1e1:0x0].0x0 rrc: 4 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000000020020 nid: 0@lo remote: 0x9a131711b119b090 expref: 3148 pid: 17965 timeout: 1006 lvb_type: 0 LustreError: 20380:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff880255209f80 x1710022341940928/t0(0) o105->lustre-OST0000@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 11-0: lustre-OST0000-osc-ffff88008f3a1bf8: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-OST0000-osc-ffff88008f3a1bf8: Connection to lustre-OST0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 2 previous similar messages LustreError: 10998:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630805276 with bad export cookie 11102242871151193661 LustreError: 20380:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 71 previous similar messages LustreError: 167-0: lustre-OST0000-osc-ffff88008f3a1bf8: This client was evicted by lustre-OST0000; in progress operations using this service will fail. Lustre: 4134:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.205@tcp:/lustre/fid: [0x240000403:0x2c1:0x0]// may get corrupted (rc -5) Lustre: 4134:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.205@tcp:/lustre/fid: [0x200000404:0x216:0x0]// may get corrupted (rc -5) Lustre: 4134:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.205@tcp:/lustre/fid: [0x280000403:0x1fb:0x0]// may get corrupted (rc -5) Lustre: dir [0x200000403:0xbe:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: 4146:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.205@tcp:/lustre/fid: [0x240000403:0x1e7:0x0]// may get corrupted (rc -108) Lustre: 4145:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.205@tcp:/lustre/fid: [0x200000404:0x18d:0x0]/ may get corrupted (rc -108) Lustre: 4145:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.205@tcp:/lustre/fid: [0x200000404:0x209:0x0]// may get corrupted (rc -108) Lustre: 4145:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.205@tcp:/lustre/fid: [0x280000403:0x281:0x0]// may get corrupted (rc -108) Lustre: 4146:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.205@tcp:/lustre/fid: [0x200000403:0x243:0x0]/ may get corrupted (rc -108) Lustre: 4146:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.205@tcp:/lustre/fid: [0x200000404:0x221:0x0]// may get corrupted (rc -108) LustreError: 11179:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000403:0xbe:0x0]: rc = -2 LustreError: 762:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0003-osc-ffff88008f3a1bf8: namespace resource [0x380000401:0x1cf:0x0].0x0 (ffff8802642a2fc0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 762:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 78 previous similar messages Lustre: lustre-OST0003-osc-ffff88008f3a1bf8: Connection restored to 192.168.123.205@tcp (at 0@lo) Lustre: Skipped 2 previous similar messages Lustre: 17536:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000404:0x295:0x0] with magic=0xbd60bd0 Lustre: 17536:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 1 previous similar message INFO: task mv:30485 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mv D ffff88008cb0c240 11344 30485 17231 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff81251fe1>] lock_rename+0x31/0xe0 [<ffffffff8125862f>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed432>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817e8a76>] ? trace_do_page_fault+0x56/0x170 [<ffffffff8125975e>] SyS_renameat2+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 LustreError: 11033:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 101s: evicting client at 0@lo ns: filter-lustre-OST0001_UUID lock: ffff88027a1f4780/0x9a131711b11f7b1d lrc: 3/0,0 mode: PW/PW res: [0x300000400:0x955:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000480020020 nid: 0@lo remote: 0x9a131711b11f79aa expref: 3652 pid: 18571 timeout: 1130 lvb_type: 0 LustreError: 11033:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 1 previous similar message LustreError: 20332:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88007bcf1f80 x1710022345453120/t0(0) o105->lustre-OST0001@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 20332:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 43 previous similar messages LustreError: 11-0: lustre-OST0001-osc-ffff880298d3c138: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 1 previous similar message LustreError: 167-0: lustre-OST0001-osc-ffff880298d3c138: This client was evicted by lustre-OST0001; in progress operations using this service will fail. LustreError: Skipped 1 previous similar message Lustre: 4142:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.205@tcp:/lustre/fid: [0x240000404:0x312:0x0]// may get corrupted (rc -108) Lustre: 4141:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.205@tcp:/lustre/fid: [0x200000404:0x24a:0x0]// may get corrupted (rc -108) Lustre: 25733:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 6s req@ffff880075738680 x1710022345353728/t0(0) o6->lustre-MDT0001-mdtlov_UUID@0@lo:0/0 lens 544/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'osp-syn-2-1.0' LustreError: 16755:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630805372 with bad export cookie 11102242871150779947 LustreError: 2445:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0001-osc-ffff880298d3c138: namespace resource [0x300000401:0x1c9f:0x0].0x0 (ffff88027870b4c0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 2445:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 1575 previous similar messages Lustre: 4134:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff8802582ba5c0 x1710022345923200/t0(0) o10->lustre-OST0003-osc-ffff880298d3c138@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630805413 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dir_create.sh.0' INFO: task dir_create.sh:16412 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. dir_create.sh D ffff8802937c1280 11184 16412 15940 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff817d5bb2>] lookup_slow+0x33/0xa7 [<ffffffff81254c7f>] link_path_walk+0x81f/0x8c0 [<ffffffff8124818c>] ? get_empty_filp+0x5c/0x1f0 [<ffffffff8125708e>] path_openat+0xae/0x5b0 [<ffffffff81258dfd>] do_filp_open+0x4d/0xb0 [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e326e>] ? _raw_spin_unlock+0xe/0x20 [<ffffffff81267893>] ? __alloc_fd+0xc3/0x170 [<ffffffff812447e4>] do_sys_open+0x124/0x220 [<ffffffff812448fe>] SyS_open+0x1e/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task getfattr:32681 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. getfattr D ffff880085caa4f0 11296 32681 17271 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff817d5bb2>] lookup_slow+0x33/0xa7 [<ffffffff81256d6e>] path_lookupat+0x89e/0x8d0 [<ffffffff8121f126>] ? kmem_cache_alloc+0x126/0x2e0 [<ffffffff81257b9f>] ? getname_flags+0x4f/0x1a0 [<ffffffff81256dcb>] filename_lookup+0x2b/0xc0 [<ffffffff81258cc7>] user_path_at_empty+0x67/0xc0 [<ffffffff817e8655>] ? __do_page_fault+0x1c5/0x500 [<ffffffff81258d31>] user_path_at+0x11/0x20 [<ffffffff81270aa8>] SyS_listxattr+0x48/0xb0 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task mkdir:1060 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mkdir D ffff88008d54ba00 11824 1060 17207 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff81256ee5>] filename_create+0x85/0x180 [<ffffffff8121f126>] ? kmem_cache_alloc+0x126/0x2e0 [<ffffffff81257c14>] ? getname_flags+0xc4/0x1a0 [<ffffffff81257ec1>] user_path_create+0x41/0x60 [<ffffffff81259206>] SyS_mkdirat+0x46/0xe0 [<ffffffff812592b9>] SyS_mkdir+0x19/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task mcreate:1074 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mcreate D ffff88008a1cd2c0 12856 1074 16138 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff81256ee5>] filename_create+0x85/0x180 [<ffffffff81257c14>] ? getname_flags+0xc4/0x1a0 [<ffffffff81257ec1>] user_path_create+0x41/0x60 [<ffffffff8125905a>] SyS_mknodat+0xda/0x220 [<ffffffff812591bd>] SyS_mknod+0x1d/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 ptlrpc_watchdog_fire: 59 callbacks suppressed Lustre: mdt00_013: service thread pid 32302 was inactive for 64.234 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 32302, comm: mdt00_013 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_fini+0xa25/0xec0 [ptlrpc] [<0>] ldlm_cli_enqueue+0x436/0xa30 [ptlrpc] [<0>] osp_md_object_lock+0x160/0x300 [osp] [<0>] lod_object_lock+0xdb/0x7d0 [lod] [<0>] mdd_object_lock+0x2d/0xd0 [mdd] [<0>] mdt_remote_object_lock_try+0x1cc/0x510 [mdt] [<0>] mdt_remote_object_lock+0x2a/0x30 [mdt] [<0>] mdt_rename_lock+0xbe/0x4d0 [mdt] [<0>] mdt_reint_migrate+0x6e8/0x11d0 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x76c/0xb40 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe ------------[ cut here ]------------ WARNING: CPU: 2 PID: 16879 at /home/green/git/lustre-release/ldiskfs/ext4_jbd2.c:266 __ldiskfs_handle_dirty_metadata+0x1d2/0x230 [ldiskfs] Modules linked in: loop zfs(PO) zunicode(PO) zzstd(O) zlua(O) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) jbd2 mbcache lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) dm_flakey dm_mod libcfs(OE) crc_t10dif crct10dif_generic sb_edac edac_core iosf_mbi crc32_pclmul ghash_clmulni_intel aesni_intel lrw gf128mul glue_helper ablk_helper cryptd pcspkr virtio_balloon virtio_console i2c_piix4 ip_tables rpcsec_gss_krb5 ata_generic drm_kms_helper pata_acpi ttm crct10dif_pclmul crct10dif_common drm ata_piix crc32c_intel drm_panel_orientation_quirks serio_raw virtio_blk libata i2c_core floppy CPU: 2 PID: 16879 Comm: mdt01_005 Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011 Call Trace: [<ffffffff817d9418>] dump_stack+0x19/0x1b [<ffffffff8108d538>] __warn+0xd8/0x100 [<ffffffff8108d67d>] warn_slowpath_null+0x1d/0x20 [<ffffffffa0bfe262>] __ldiskfs_handle_dirty_metadata+0x1d2/0x230 [ldiskfs] [<ffffffffa0bde33d>] ldiskfs_getblk+0x13d/0x210 [ldiskfs] [<ffffffffa0bde437>] ldiskfs_bread+0x27/0xe0 [ldiskfs] [<ffffffffa0ca2e97>] osd_write+0x4e7/0xcb0 [osd_ldiskfs] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffffa04d06f3>] dt_record_write+0x33/0x120 [obdclass] [<ffffffffa048f6b2>] llog_osd_write_rec+0x1052/0x19a0 [obdclass] [<ffffffffa047c760>] llog_write_rec+0x290/0x590 [obdclass] [<ffffffffa04824ee>] llog_cat_add_rec+0x23e/0x950 [obdclass] [<ffffffffa04795ff>] llog_add+0x17f/0x1f0 [obdclass] [<ffffffffa04aaa39>] ? lprocfs_counter_add+0xf9/0x160 [obdclass] [<ffffffffa081360d>] sub_updates_write+0xb6b/0xe32 [ptlrpc] [<ffffffff811beccd>] ? __free_pages+0x1d/0x30 [<ffffffff8121efe4>] ? kfree+0x1e4/0x200 [<ffffffffa07e3892>] top_trans_stop+0x7c2/0xf60 [ptlrpc] [<ffffffffa0eddf3c>] lod_trans_stop+0x25c/0x340 [lod] [<ffffffffa0db2408>] mdd_trans_stop+0x28/0x16e [mdd] [<ffffffffa0d93b2f>] mdd_migrate_object+0xf1f/0x1ba0 [mdd] [<ffffffffa04cf375>] ? lu_object_find_at+0x285/0xb10 [obdclass] [<ffffffffa0d94b86>] mdd_migrate+0x3d6/0x8e0 [mdd] [<ffffffffa0e660d8>] mdo_migrate+0x4c/0x4e [mdt] [<ffffffffa0e25e53>] mdt_reint_migrate+0x1023/0x11d0 [mdt] [<ffffffffa0e26087>] mdt_reint_rec+0x87/0x240 [mdt] [<ffffffffa0dfb55c>] mdt_reint_internal+0x76c/0xb40 [mdt] [<ffffffffa0e08657>] mdt_reint+0x67/0x150 [mdt] [<ffffffffa07d226e>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<ffffffffa032945e>] ? libcfs_nid2str_r+0xfe/0x130 [lnet] [<ffffffffa077d990>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<ffffffffa077f529>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffffa077e950>] ? ptlrpc_wait_event+0x620/0x620 [ptlrpc] [<ffffffff810ba0f4>] kthread+0xe4/0xf0 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 ---[ end trace 6d269e43193aaa7e ]--- LDISKFS-fs: ldiskfs_getblk:888: aborting transaction: error 28 in __ldiskfs_handle_dirty_metadata LDISKFS-fs error (device dm-1): ldiskfs_getblk:888: inode #200: block 49288: comm mdt01_005: journal_dirty_metadata failed: handle type 0 started at line 1972, credits 512/0, errcode -28 Aborting journal on device dm-1-8. LDISKFS-fs (dm-1): Remounting filesystem read-only LustreError: 16879:0:(osd_io.c:2103:osd_ldiskfs_write_record()) lustre-MDT0001: error reading offset 180224 (block 44, size 24, offs 180192), credits 0/0: rc = -28 LustreError: 16879:0:(update_trans.c:985:top_trans_stop()) lustre-MDT0001-osd: write updates failed: rc = -28 LustreError: 16879:0:(osd_io.c:2110:osd_ldiskfs_write_record()) journal_get_write_access() returned error -30 LustreError: 16879:0:(tgt_lastrcvd.c:1253:tgt_add_reply_data()) lustre-MDT0001: can't update reply_data file: rc = -30 LustreError: 16879:0:(osd_handler.c:2075:osd_trans_stop()) lustre-MDT0001: failed in transaction hook: rc = -30 LDISKFS-fs error (device dm-1) in osd_trans_stop:2082: error 28 LustreError: 12978:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-OST0000-osc-MDT0001: fail to cancel 92 llog-records: rc = -30 LustreError: 14595:0:(osp_sync.c:1094:osp_sync_process_committed()) lustre-OST0003-osc-MDT0001: can't cancel 12 records: rc = -30 LustreError: 11890:0:(osd_handler.c:1779:osd_trans_commit_cb()) transaction @0xffff88008ac332e8 commit error: 2 LustreError: 16879:0:(osd_handler.c:2085:osd_trans_stop()) lustre-MDT0001: failed to stop transaction: rc = -28 LustreError: 16879:0:(update_trans.c:1011:top_trans_stop()) lustre-MDT0001-osd: stop trans failed: rc = -30 LustreError: 12978:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) Skipped 6 previous similar messages LustreError: 16879:0:(update_trans.c:1084:top_trans_stop()) lustre-OST0002-osc-MDT0001: stop trans failed: rc = -30 LustreError: 3485:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -30 LustreError: 11019:0:(tgt_handler.c:1367:tgt_blocking_ast()) lustre-MDT0001: syncing [0x240000401:0x4:0x0] (19-0) on lock cancel: rc = -5 LustreError: 14174:0:(osp_sync.c:1094:osp_sync_process_committed()) lustre-OST0002-osc-MDT0001: can't cancel 8 records: rc = -30 LustreError: 14174:0:(osp_sync.c:1094:osp_sync_process_committed()) Skipped 6 previous similar messages LustreError: 14174:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-OST0002-osc-MDT0001: fail to cancel 1 of 1 llog-records: rc = -30 LustreError: 14174:0:(osp_sync.c:1079:osp_sync_process_committed()) lustre-OST0002-osc-MDT0001: can't cancel record: rc = -30 LustreError: 3358:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -30 LustreError: 14174:0:(osp_sync.c:1094:osp_sync_process_committed()) lustre-OST0002-osc-MDT0001: can't cancel 13 records: rc = -30 LustreError: 14174:0:(osp_sync.c:1094:osp_sync_process_committed()) Skipped 6 previous similar messages LustreError: 3431:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -30 LustreError: 3431:0:(llite_lib.c:1836:ll_md_setattr()) Skipped 5 previous similar messages LustreError: 14174:0:(osp_sync.c:1079:osp_sync_process_committed()) lustre-OST0002-osc-MDT0001: can't cancel record: rc = -30 LustreError: 13414:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-OST0001-osc-MDT0001: fail to cancel 241 llog-records: rc = -30 LustreError: 13414:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) Skipped 29 previous similar messages LustreError: 13414:0:(osp_sync.c:1094:osp_sync_process_committed()) lustre-OST0001-osc-MDT0001: can't cancel 241 records: rc = -30 LustreError: 13414:0:(osp_sync.c:1094:osp_sync_process_committed()) Skipped 10 previous similar messages LustreError: 13414:0:(osp_sync.c:1079:osp_sync_process_committed()) lustre-OST0001-osc-MDT0001: can't cancel record: rc = -30 LustreError: 13414:0:(osp_sync.c:1079:osp_sync_process_committed()) Skipped 9 previous similar messages LustreError: 12312:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0001-osp-MDT0002: fail to cancel 1 of 1 llog-records: rc = -116 LustreError: 12312:0:(llog_cat.c:787:llog_cat_cancel_records()) Skipped 12 previous similar messages LustreError: 3628:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -30 LustreError: 3628:0:(llite_lib.c:1836:ll_md_setattr()) Skipped 4 previous similar messages LustreError: 14174:0:(osp_sync.c:1079:osp_sync_process_committed()) lustre-OST0002-osc-MDT0001: can't cancel record: rc = -30 LustreError: 14174:0:(osp_sync.c:1079:osp_sync_process_committed()) Skipped 10 previous similar messages | Externally reported by onyx-68 boilpot email |
racer test 1: racer on clients: centos-115.localnet DURATION=2700 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa1770e52>] ll_migrate+0x9b2/0xec0 [lustre] PGD 80000002bdf42067 PUD 249ac9067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) dm_flakey dm_mod loop zfs(PO) zunicode(PO) zzstd(O) zlua(O) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) jbd2 mbcache crc_t10dif crct10dif_generic sb_edac edac_core iosf_mbi crc32_pclmul ghash_clmulni_intel aesni_intel lrw gf128mul glue_helper ablk_helper cryptd virtio_balloon pcspkr virtio_console i2c_piix4 ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper crct10dif_pclmul ttm crct10dif_common drm crc32c_intel ata_piix drm_panel_orientation_quirks serio_raw virtio_blk libata i2c_core floppy [last unloaded: libcfs] CPU: 5 PID: 6286 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011 task: ffff8802d612a4f0 ti: ffff8800aceec000 task.ti: ffff8800aceec000 RIP: 0010:[<ffffffffa1770e52>] [<ffffffffa1770e52>] ll_migrate+0x9b2/0xec0 [lustre] RSP: 0018:ffff8800aceefbc8 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff8800ace97cf8 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffffea000bb0fd00 RDI: 0000000000000246 RBP: ffff8800aceefc48 R08: ffff8802ec3f6d08 R09: 0000000000000000 R10: ffff8802ec3f4d28 R11: ffff8802ec3f6d08 R12: 0000000000000000 R13: ffff8802e7b5e6b8 R14: ffff88026113f018 R15: 0000000000000030 FS: 00007fcab4626740(0000) GS:ffff880331b40000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 CR2: 00007f88180909e5 CR3: 0000000098232000 CR4: 00000000001607e0 Call Trace: [<ffffffff81242213>] ? __check_object_size+0x1c3/0x220 [<ffffffffa175e561>] ll_dir_ioctl+0x5d01/0x6ed0 [lustre] [<ffffffff811b515b>] ? unlock_page+0x2b/0x30 [<ffffffff811e5e2e>] ? do_read_fault.isra.64+0x12e/0x170 [<ffffffff811ea1cf>] ? handle_pte_fault+0x2ef/0xca0 [<ffffffff811ecf6d>] ? __handle_mm_fault+0x31d/0x720 [<ffffffff8125b3fd>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff81245a38>] ? vfs_write+0x178/0x230 [<ffffffff8125b751>] SyS_ioctl+0xa1/0xc0 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 | Lustre: lustre-OST0000-osc-ffff8802a433e678: disconnect after 20s idle Lustre: 14811:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 512 < left 1819, rollback = 7 Lustre: 14811:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 14811:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 1/89/0 Lustre: 14811:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1810/1819/0, punch: 0/0/0, quota 3/3/0 Lustre: 14811:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 14811:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 14811:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 511 < left 1818, rollback = 7 Lustre: 14811:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 1108 previous similar messages Lustre: 14811:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 14811:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 1149 previous similar messages Lustre: 14811:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 1/89/0 Lustre: 14811:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 1149 previous similar messages Lustre: 14811:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1810/1819/1, punch: 0/0/0, quota 3/3/0 Lustre: 14811:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 1149 previous similar messages Lustre: 14811:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 14811:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 1149 previous similar messages Lustre: 14811:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 14811:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 1149 previous similar messages cp (15331) used greatest stack depth: 9728 bytes left Lustre: 15230:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 505 < left 1701, rollback = 7 Lustre: 15230:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 700 previous similar messages Lustre: 15230:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 1/4/4, destroy: 0/0/0 Lustre: 15230:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 659 previous similar messages Lustre: 15230:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 4/209/0 Lustre: 15230:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 659 previous similar messages Lustre: 15230:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1692/1701/0, punch: 0/0/0, quota 6/6/2 Lustre: 15230:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 659 previous similar messages Lustre: 15230:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 2/33/1, delete: 0/0/0 Lustre: 15230:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 659 previous similar messages Lustre: 15230:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 15230:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 659 previous similar messages Lustre: 15544:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 512 < left 1642, rollback = 7 Lustre: 15544:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 1691 previous similar messages Lustre: 15544:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 15544:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 1691 previous similar messages Lustre: 15544:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 1/89/0 Lustre: 15544:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 1691 previous similar messages Lustre: 15544:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1633/1642/0, punch: 0/0/0, quota 3/3/0 Lustre: 15544:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 1691 previous similar messages Lustre: 15544:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 15544:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 1691 previous similar messages Lustre: 15544:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 15544:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 1691 previous similar messages Lustre: 12405:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 499 < left 671, rollback = 2 Lustre: 16956:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 7s req@ffff8802d77c1940 x1709942352329856/t0(0) o104->8263f8d3-ce0c-4db1-ac74-5bc989b69a05@0@lo:0/0 lens 328/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'' Lustre: ldlm_cbd: This server is not able to keep up with request traffic (cpu-bound). Lustre: 12359:0:(service.c:1612:ptlrpc_at_check_timed()) earlyQ=1 reqQ=0 recA=0, svcEst=1, delay=0ms Lustre: 12359:0:(service.c:1378:ptlrpc_at_send_early_reply()) @@@ Already past deadline (-2s), not sending early reply. Consider increasing at_early_margin (5)? req@ffff880231b44ad8 x1709942352326656/t0(0) o105->LOV_OSC_UUID@0@lo:48/0 lens 392/0 e 0 to 0 dl 1630728323 ref 2 fl New:/0/ffffffff rc 0/-1 job:'' LustreError: 12359:0:(service.c:2290:ptlrpc_server_handle_request()) @@@ Dropping timed-out request from 12345-0@lo: deadline 6/2s ago req@ffff880231b44ad8 x1709942352326656/t0(0) o105->LOV_OSC_UUID@0@lo:48/0 lens 392/0 e 0 to 0 dl 1630728323 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'' Lustre: 12359:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (6/2s); client may timeout req@ffff880231b44ad8 x1709942352326656/t0(0) o105->LOV_OSC_UUID@0@lo:48/0 lens 392/0 e 0 to 0 dl 1630728323 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'' Lustre: ldlm_canceld: This server is not able to keep up with request traffic (cpu-bound). Lustre: 12380:0:(service.c:1612:ptlrpc_at_check_timed()) earlyQ=1 reqQ=0 recA=1, svcEst=31, delay=9555ms Lustre: 12380:0:(service.c:1378:ptlrpc_at_send_early_reply()) @@@ Already past deadline (-5s), not sending early reply. Consider increasing at_early_margin (5)? req@ffff8802d75fea80 x1709942352326592/t0(0) o103->8263f8d3-ce0c-4db1-ac74-5bc989b69a05@0@lo:48/0 lens 328/224 e 0 to 0 dl 1630728323 ref 2 fl Interpret:H/0/0 rc 0/0 job:'ldlm_bl_05.0' Lustre: 15835:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1630728317/real 1630728317] req@ffff8802d75fcb40 x1709942352326656/t0(0) o105->lustre-OST0002@0@lo:15/16 lens 392/224 e 0 to 1 dl 1630728324 ref 1 fl Rpc:XQr/0/ffffffff rc 0/-1 job:'' Lustre: 15341:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (43/2s); client may timeout req@ffff8802d2591f80 x1709942352151680/t0(0) o101->8263f8d3-ce0c-4db1-ac74-5bc989b69a05@0@lo:51/0 lens 576/640 e 0 to 0 dl 1630728326 ref 1 fl Complete:/0/0 rc 0/0 job:'ls.0' Lustre: 12435:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 8s req@ffff8802d77f3240 x1709942352331712/t0(0) o35->7a8d0ba0-86fb-4c66-8dd9-af0523f59dad@0@lo:0/0 lens 392/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'ls.0' Lustre: 12435:0:(service.c:2157:ptlrpc_server_handle_req_in()) Skipped 2 previous similar messages Lustre: mdt_readpage: This server is not able to keep up with request traffic (cpu-bound). Lustre: Skipped 1 previous similar message Lustre: 16789:0:(service.c:1612:ptlrpc_at_check_timed()) earlyQ=1 reqQ=0 recA=0, svcEst=1, delay=0ms Lustre: 16789:0:(service.c:1612:ptlrpc_at_check_timed()) Skipped 1 previous similar message Lustre: 16789:0:(service.c:1378:ptlrpc_at_send_early_reply()) @@@ Already past deadline (-2s), not sending early reply. Consider increasing at_early_margin (5)? req@ffff8802d77f3240 x1709942352331712/t0(0) o35->7a8d0ba0-86fb-4c66-8dd9-af0523f59dad@0@lo:52/0 lens 392/0 e 0 to 0 dl 1630728327 ref 2 fl New:/0/ffffffff rc 0/-1 job:'ls.0' Lustre: 16789:0:(service.c:1378:ptlrpc_at_send_early_reply()) Skipped 2 previous similar messages LustreError: 16789:0:(service.c:2290:ptlrpc_server_handle_request()) @@@ Dropping timed-out request from 12345-0@lo: deadline 6/3s ago req@ffff8802d77f3240 x1709942352331712/t0(0) o35->7a8d0ba0-86fb-4c66-8dd9-af0523f59dad@0@lo:52/0 lens 392/0 e 0 to 0 dl 1630728327 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'ls.0' Lustre: 16789:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (6/3s); client may timeout req@ffff8802d77f3240 x1709942352331712/t0(0) o35->7a8d0ba0-86fb-4c66-8dd9-af0523f59dad@0@lo:52/0 lens 392/0 e 0 to 0 dl 1630728327 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'ls.0' Lustre: 16789:0:(service.c:2328:ptlrpc_server_handle_request()) Skipped 2 previous similar messages Lustre: ost_io: This server is not able to keep up with request traffic (cpu-bound). Lustre: 13150:0:(service.c:1612:ptlrpc_at_check_timed()) earlyQ=1 reqQ=0 recA=1, svcEst=1, delay=5645ms Lustre: 13150:0:(service.c:1378:ptlrpc_at_send_early_reply()) @@@ Already past deadline (-1s), not sending early reply. Consider increasing at_early_margin (5)? req@ffff8802d7b9e440 x1709942352338688/t0(0) o10->8263f8d3-ce0c-4db1-ac74-5bc989b69a05@0@lo:56/0 lens 440/432 e 0 to 0 dl 1630728331 ref 2 fl Interpret:/0/0 rc 0/0 job:'cp.0' Lustre: 15255:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (6/2s); client may timeout req@ffff8802d7b9e440 x1709942352338688/t4294967871(0) o10->8263f8d3-ce0c-4db1-ac74-5bc989b69a05@0@lo:56/0 lens 440/432 e 0 to 0 dl 1630728331 ref 1 fl Complete:/0/0 rc 0/0 job:'cp.0' Lustre: 12415:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 508 < left 20908, rollback = 7 Lustre: 12415:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 1632 previous similar messages Lustre: 12415:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 486/1944/0, destroy: 0/0/0 Lustre: 12415:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 1640 previous similar messages Lustre: 12415:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/1, xattr_set: 488/15/0 Lustre: 12415:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 1640 previous similar messages Lustre: 12415:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 2431/20908/0, punch: 0/0/0, quota 10/106/3 Lustre: 12415:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 1640 previous similar messages Lustre: 12415:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 486/8262/0, delete: 0/0/0 Lustre: 12415:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 1640 previous similar messages Lustre: 12415:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 12415:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 1640 previous similar messages Lustre: 17578:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 506 < left 20908, rollback = 7 Lustre: 17578:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 14316 previous similar messages Lustre: 17578:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 486/1944/0, destroy: 0/0/0 Lustre: 17578:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 14317 previous similar messages Lustre: 17578:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/1, xattr_set: 488/15/0 Lustre: 17578:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 14316 previous similar messages Lustre: 17578:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 2431/20908/0, punch: 0/0/0, quota 14/174/5 Lustre: 17578:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 14317 previous similar messages Lustre: 17578:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 486/8262/0, delete: 0/0/0 Lustre: 17578:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 14317 previous similar messages Lustre: 17578:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 17578:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 14317 previous similar messages Lustre: 12412:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 8: before 510 < left 8278, rollback = 9 Lustre: 12412:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 7 previous similar messages Lustre: 15358:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1630728321/real 1630728321] req@ffff8802d77f4500 x1709942352331712/t0(0) o35->lustre-MDT0000-mdc-ffff8802a433e678@0@lo:23/10 lens 392/4592 e 0 to 1 dl 1630728369 ref 2 fl Rpc:XQr/0/ffffffff rc 0/-1 job:'ls.0' Lustre: lustre-MDT0000-mdc-ffff8802a433e678: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: lustre-MDT0000: Client 7a8d0ba0-86fb-4c66-8dd9-af0523f59dad (at 0@lo) reconnecting Lustre: lustre-MDT0000-mdc-ffff8802a433e678: Connection restored to 192.168.123.215@tcp (at 0@lo) Lustre: 13997:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 512 < left 70229, rollback = 7 Lustre: 13997:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 1949 previous similar messages Lustre: 13997:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 1633/6532/0, destroy: 0/0/0 Lustre: 13997:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 1950 previous similar messages Lustre: 13997:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 1635/15/0 Lustre: 13997:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 1950 previous similar messages Lustre: 13997:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 8166/70229/0, punch: 0/0/0, quota 10/106/0 Lustre: 13997:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 1950 previous similar messages Lustre: 13997:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 1633/27761/0, delete: 0/0/0 Lustre: 13997:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 1950 previous similar messages Lustre: 13997:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 13997:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 1950 previous similar messages Lustre: 16202:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 501 < left 688, rollback = 2 Lustre: 16905:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 7s req@ffff8802cc6e9940 x1709942354685888/t0(0) o101->8263f8d3-ce0c-4db1-ac74-5bc989b69a05@0@lo:0/0 lens 328/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'rm.0' LustreError: 12392:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0001_UUID lock: ffff8802dd6d4780/0xfc2cc2c14c1e6029 lrc: 3/0,0 mode: PW/PW res: [0x7:0x0:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000000020020 nid: 0@lo remote: 0xfc2cc2c14c1e6014 expref: 5927 pid: 13603 timeout: 3749 lvb_type: 0 LustreError: 11-0: lustre-OST0001-osc-ffff8802a433e678: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-OST0001-osc-ffff8802a433e678: Connection to lustre-OST0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0001-osc-ffff8802a433e678: This client was evicted by lustre-OST0001; in progress operations using this service will fail. LustreError: 16899:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802d82d6a80 x1709942354976384/t0(0) o104->lustre-OST0000@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 17683:0:(ldlm_lockd.c:1425:ldlm_handle_enqueue0()) ### lock on destroyed export ffff8802a7218958 ns: filter-lustre-OST0000_UUID lock: ffff8802bfde8040/0xfc2cc2c14c231771 lrc: 3/0,0 mode: --/PW res: [0x2c0000401:0x209:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x50000000020000 nid: 0@lo remote: 0xfc2cc2c14c231763 expref: 3681 pid: 17683 timeout: 0 lvb_type: 0 LustreError: 17881:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802d6169940 x1709942354980928/t0(0) o104->lustre-OST0000@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 17881:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 2 previous similar messages LustreError: 16968:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802c96aea80 x1709942355007424/t0(0) o106->lustre-OST0001@0@lo:15/16 lens 328/280 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 16968:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 1 previous similar message ------------[ cut here ]------------ WARNING: CPU: 0 PID: 15963 at /home/green/git/lustre-release/ldiskfs/namei.c:3331 ldiskfs_orphan_add+0x11e/0x2a0 [ldiskfs] Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) dm_flakey dm_mod loop zfs(PO) zunicode(PO) zzstd(O) zlua(O) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) jbd2 mbcache crc_t10dif crct10dif_generic sb_edac edac_core iosf_mbi crc32_pclmul ghash_clmulni_intel aesni_intel lrw gf128mul glue_helper ablk_helper cryptd virtio_balloon pcspkr virtio_console i2c_piix4 ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper crct10dif_pclmul ttm crct10dif_common drm crc32c_intel ata_piix drm_panel_orientation_quirks serio_raw virtio_blk libata i2c_core floppy [last unloaded: libcfs] CPU: 0 PID: 15963 Comm: mdt_rdpg00_002 Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011 Call Trace: [<ffffffff817d9418>] dump_stack+0x19/0x1b [<ffffffff8108d538>] __warn+0xd8/0x100 [<ffffffff8108d67d>] warn_slowpath_null+0x1d/0x20 [<ffffffffa0ca94ae>] ldiskfs_orphan_add+0x11e/0x2a0 [ldiskfs] [<ffffffffa0c9a9eb>] ldiskfs_xattr_inode_orphan_add.constprop.17+0xbb/0x110 [ldiskfs] [<ffffffffa0c9d417>] ldiskfs_xattr_delete_inode+0x77/0x3c0 [ldiskfs] [<ffffffffa0cbdaa0>] ? ldiskfs_evict_inode+0x190/0x670 [ldiskfs] [<ffffffffa0cbdad9>] ldiskfs_evict_inode+0x1c9/0x670 [ldiskfs] [<ffffffff812649af>] evict+0xaf/0x180 [<ffffffff81264de5>] iput+0xf5/0x180 [<ffffffffa0d2ad08>] osd_object_delete+0x1f8/0x370 [osd_ldiskfs] [<ffffffffa050c59d>] lu_object_free.isra.26+0xcd/0x1d0 [obdclass] [<ffffffffa0510cf1>] lu_object_put+0xa1/0x480 [obdclass] [<ffffffffa0ec3af4>] mdt_mfd_close+0x1f4/0x870 [mdt] [<ffffffffa0ec9e14>] mdt_close_internal+0xb4/0x240 [mdt] [<ffffffffa0eca224>] mdt_close+0x284/0x850 [mdt] [<ffffffffa08132d4>] ? tgt_request_preprocess.isra.27+0xf4/0x850 [ptlrpc] [<ffffffffa081426e>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<ffffffffa034745e>] ? libcfs_nid2str_r+0xfe/0x130 [lnet] [<ffffffffa07bf990>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<ffffffffa07c1529>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffffa07c0950>] ? ptlrpc_wait_event+0x620/0x620 [ptlrpc] [<ffffffff810ba0f4>] kthread+0xe4/0xf0 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 ---[ end trace c7297906fceace18 ]--- Lustre: 13750:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 8s req@ffff8802d9b925c0 x1709942354959104/t0(0) o13->lustre-MDT0001-mdtlov_UUID@0@lo:0/0 lens 224/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'osp-pre-0-1.0' Lustre: 12438:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 509 < left 32776, rollback = 7 Lustre: 12438:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 10232 previous similar messages Lustre: 12438:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 762/3048/0, destroy: 1/4/1 Lustre: 12438:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 10238 previous similar messages Lustre: 12438:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 762/0/0 Lustre: 12438:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 10238 previous similar messages Lustre: 12438:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 3811/32776/0, punch: 0/0/0, quota 6/6/0 Lustre: 12438:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 10238 previous similar messages Lustre: 12438:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 762/12954/0, delete: 2/5/1 Lustre: 12438:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 10238 previous similar messages Lustre: 12438:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 1/1/1 Lustre: 12438:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 10238 previous similar messages Lustre: 16818:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 504 < left 1400, rollback = 2 Lustre: 16818:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 5 previous similar messages Lustre: lustre-OST0000-osc-ffff8802a4db5d28: Connection restored to 192.168.123.215@tcp (at 0@lo) Lustre: 15865:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 8s req@ffff8802942089a8 x1709942355171840/t0(0) o35->7a8d0ba0-86fb-4c66-8dd9-af0523f59dad@0@lo:0/0 lens 392/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'cat.0' Lustre: 15865:0:(service.c:2157:ptlrpc_server_handle_req_in()) Skipped 2 previous similar messages Lustre: 12450:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 509 < left 551, rollback = 2 Lustre: 12450:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 23 previous similar messages LustreError: 13922:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0001: '3' migration was interrupted, run 'lfs migrate -m 2 -c 1 -H crush 3' to finish migration. Lustre: 15399:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 8: before 510 < left 12749, rollback = 9 Lustre: 15399:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 5 previous similar messages Lustre: 15843:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 502 < left 582, rollback = 2 Lustre: 15843:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 3/12/6, destroy: 0/0/0 Lustre: 15843:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 37208 previous similar messages Lustre: 15843:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 6/582/0 Lustre: 15843:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 37208 previous similar messages Lustre: 15843:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 5/136/0, punch: 0/0/0, quota 6/6/0 Lustre: 15843:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 37208 previous similar messages Lustre: 15843:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 10/203/4, delete: 0/0/0 Lustre: 15843:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 37208 previous similar messages Lustre: 15843:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 4/4/0, ref_del: 0/0/0 Lustre: 15843:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 37208 previous similar messages Lustre: 12446:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 10s req@ffff8802b5014500 x1709942357531008/t0(0) o1000->lustre-MDT0002-mdtlov_UUID@0@lo:0/0 lens 1304/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'osp_up1-2.0' Lustre: 12414:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 501 < left 35872, rollback = 7 Lustre: 12414:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 37177 previous similar messages LustreError: 14889:0:(out_handler.c:910:out_tx_end()) lustre-MDT0002-osd: undo for /home/green/git/lustre-release/lustre/ptlrpc/../../lustre/target/out_handler.c:445: rc = -524 LustreError: 12499:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0002-osp-MDT0000: fail to cancel 1 llog-records: rc = -2 LustreError: 12499:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0002-osp-MDT0000: fail to cancel 1 of 1 llog-records: rc = -2 Lustre: 18202:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 496 < left 760, rollback = 2 Lustre: 18202:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 6 previous similar messages LustreError: 12392:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0000_UUID lock: ffff8802d8664000/0xfc2cc2c14c246907 lrc: 3/0,0 mode: PW/PW res: [0x2c0000401:0x325:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 2097152->18446744073709551615) gid 0 flags: 0x60000000020020 nid: 0@lo remote: 0xfc2cc2c14c246900 expref: 7130 pid: 18210 timeout: 3886 lvb_type: 0 LustreError: 12392:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 1 previous similar message LustreError: 13394:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802cd342c00 x1709942359054336/t0(0) o105->lustre-OST0000@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 13394:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 8 previous similar messages LustreError: 18802:0:(ldlm_lockd.c:1425:ldlm_handle_enqueue0()) ### lock on destroyed export ffff88023ffa8958 ns: filter-lustre-OST0000_UUID lock: ffff8802b23f4f00/0xfc2cc2c14c28a3e5 lrc: 3/0,0 mode: --/PW res: [0x2c0000401:0x33d:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x50000000020000 nid: 0@lo remote: 0xfc2cc2c14c28a3de expref: 6842 pid: 18802 timeout: 0 lvb_type: 0 LustreError: 18802:0:(ldlm_lockd.c:1425:ldlm_handle_enqueue0()) Skipped 1 previous similar message LustreError: 11-0: lustre-OST0000-osc-ffff8802a433e678: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 2 previous similar messages Lustre: lustre-OST0000-osc-ffff8802a433e678: Connection to lustre-OST0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 1 previous similar message LustreError: 167-0: lustre-OST0000-osc-ffff8802a433e678: This client was evicted by lustre-OST0000; in progress operations using this service will fail. LustreError: Skipped 1 previous similar message Lustre: 16973:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 7s req@ffff8802cef7a5c0 x1709942358864896/t0(0) o2->7a8d0ba0-86fb-4c66-8dd9-af0523f59dad@0@lo:0/0 lens 440/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'chown.0' LustreError: 12392:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 102s: evicting client at 0@lo ns: filter-lustre-OST0002_UUID lock: ffff8802db074f00/0xfc2cc2c14c2409df lrc: 3/0,0 mode: PW/PW res: [0x340000401:0x85:0x0].0x0 rrc: 4 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000400030020 nid: 0@lo remote: 0xfc2cc2c14c2407f5 expref: 5540 pid: 16900 timeout: 3889 lvb_type: 0 LustreError: 13093:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802d220c500 x1709942359107456/t0(0) o106->lustre-OST0002@0@lo:15/16 lens 328/280 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 18670:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630728551 with bad export cookie 18171112731994175395 Lustre: lustre-OST0002-osc-ffff8802a4db5d28: Connection to lustre-OST0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0002-osc-ffff8802a4db5d28: This client was evicted by lustre-OST0002; in progress operations using this service will fail. Lustre: 9200:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x280000404:0xa3:0x0]// may get corrupted (rc -5) Lustre: lustre-OST0002-osc-ffff8802a4db5d28: Connection restored to 192.168.123.215@tcp (at 0@lo) Lustre: Skipped 1 previous similar message Lustre: 18578:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1630728481/real 1630728481] req@ffff8802cb979940 x1709942357341568/t0(0) o101->lustre-MDT0001-mdc-ffff8802a433e678@0@lo:12/10 lens 576/22120 e 0 to 1 dl 1630728564 ref 2 fl Rpc:XQr/2/ffffffff rc 0/-1 job:'cat.0' Lustre: lustre-MDT0001-mdc-ffff8802a433e678: Connection to lustre-MDT0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: lustre-MDT0001: Client 7a8d0ba0-86fb-4c66-8dd9-af0523f59dad (at 0@lo) reconnecting LustreError: 19517:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0000-osc-ffff8802a433e678: namespace resource [0x2c0000400:0x180:0x0].0x0 (ffff8802db9bd1c0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 19517:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0000-osc-ffff8802a433e678: namespace resource [0x2c0000400:0x181:0x0].0x0 (ffff8802231e4040) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 19517:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 9 previous similar messages Lustre: 9209:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff880249d225c0 x1709942359476928/t0(0) o10->lustre-OST0002-osc-ffff8802a433e678@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630728629 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dir_create.sh.0' LustreError: 19517:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0000-osc-ffff8802a433e678: namespace resource [0x2c0000401:0x19f:0x0].0x0 (ffff88022537cf40) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 19517:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 151 previous similar messages Lustre: lustre-OST0000-osc-ffff8802a433e678: Connection restored to 192.168.123.215@tcp (at 0@lo) Lustre: Skipped 1 previous similar message Lustre: 13331:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 8: before 511 < left 27793, rollback = 9 Lustre: 13331:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 21 previous similar messages LustreError: 15325:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0000: '13' migration was interrupted, run 'lfs migrate -m 2 -c 1 -H crush 13' to finish migration. Lustre: 15399:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 15399:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 35883 previous similar messages Lustre: 15399:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 2/178/0 Lustre: 15399:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 35883 previous similar messages Lustre: 15399:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1686/1695/0, punch: 0/0/0, quota 3/3/0 Lustre: 15399:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 35883 previous similar messages Lustre: 15399:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 15399:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 35883 previous similar messages Lustre: 15399:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 15399:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 35883 previous similar messages Lustre: 17589:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 508 < left 32690, rollback = 7 Lustre: 17589:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 44786 previous similar messages Lustre: ll_ost00_009: service thread pid 16970 was inactive for 62.100 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 19442, comm: ll_ost00_018 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] tgt_extent_lock+0xea/0x2a0 [ptlrpc] [<0>] ofd_getattr_hdl+0x365/0x720 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: Skipped 1 previous similar message Pid: 16970, comm: ll_ost00_009 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] tgt_extent_lock+0xea/0x2a0 [ptlrpc] [<0>] ofd_getattr_hdl+0x365/0x720 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 18788, comm: ll_ost00_012 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] tgt_extent_lock+0xea/0x2a0 [ptlrpc] [<0>] ofd_getattr_hdl+0x365/0x720 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 Lustre: ll_ost00_004: service thread pid 16863 was inactive for 62.424 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. [<0>] 0xfffffffffffffffe Lustre: ll_ost00_034: service thread pid 20194 was inactive for 62.006 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 13 previous similar messages Lustre: ll_ost00_048: service thread pid 20218 was inactive for 62.059 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 2 previous similar messages Lustre: ll_ost00_028: service thread pid 20174 was inactive for 62.188 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 2 previous similar messages Lustre: dir [0x280000404:0x83:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: dir [0x280000404:0x83:0x0] stripe 2 readdir failed: -2, directory is partially accessed! LustreError: 12392:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 104s: evicting client at 0@lo ns: filter-lustre-OST0002_UUID lock: ffff8802c3e57840/0xfc2cc2c14c2a70b1 lrc: 3/0,0 mode: PW/PW res: [0x340000401:0x373:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000400020020 nid: 0@lo remote: 0xfc2cc2c14c2a706b expref: 1679 pid: 16951 timeout: 4043 lvb_type: 0 LustreError: 13394:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802a40be440 x1709942365235392/t0(0) o105->lustre-OST0002@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 13394:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 8 previous similar messages LustreError: 20850:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630728708 with bad export cookie 18171112731995319104 LustreError: 20850:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) Skipped 4 previous similar messages Lustre: lustre-OST0002-osc-ffff8802a4db5d28: Connection to lustre-OST0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0002-osc-ffff8802a4db5d28: This client was evicted by lustre-OST0002; in progress operations using this service will fail. Lustre: 9200:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x240000404:0xf:0x0]/ may get corrupted (rc -5) LustreError: 11-0: lustre-OST0000-osc-ffff8802a4db5d28: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: 9202:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x280000404:0x9:0x0]// may get corrupted (rc -5) Lustre: 9198:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x200000404:0x143:0x0]// may get corrupted (rc -5) Lustre: 9198:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x200000404:0x12d:0x0]/ may get corrupted (rc -108) LustreError: 22912:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0002-osc-ffff8802a4db5d28: namespace resource [0x340000401:0x39f:0x0].0x0 (ffff8802d71d79c0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 22912:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 88 previous similar messages Lustre: lustre-OST0002-osc-ffff8802a4db5d28: Connection restored to 192.168.123.215@tcp (at 0@lo) Lustre: 13997:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 507 < left 760, rollback = 2 Lustre: 13997:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 75 previous similar messages LustreError: 12694:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0000-osp-MDT0001: fail to cancel 1 llog-records: rc = -2 LustreError: 12694:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0000-osp-MDT0001: fail to cancel 1 of 1 llog-records: rc = -2 LustreError: 12405:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0000: '0' migration was interrupted, run 'lfs migrate -m 2 -c 1 -H crush 0' to finish migration. Lustre: 12413:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 342/1368/0, destroy: 0/0/0 Lustre: 12413:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 69652 previous similar messages Lustre: 12413:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/1, xattr_set: 344/15/0 Lustre: 12413:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 69652 previous similar messages Lustre: 12413:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1711/14716/0, punch: 0/0/0, quota 10/106/7 Lustre: 12413:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 69652 previous similar messages Lustre: 12413:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 342/5814/0, delete: 0/0/0 Lustre: 12413:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 69652 previous similar messages Lustre: 12413:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 12413:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 69651 previous similar messages Lustre: 15211:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 510 < left 14243, rollback = 7 Lustre: 15211:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 61995 previous similar messages LustreError: 12392:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 102s: evicting client at 0@lo ns: filter-lustre-OST0001_UUID lock: ffff8800acd49300/0xfc2cc2c14c34cabd lrc: 3/0,0 mode: PW/PW res: [0x463:0x0:0x0].0x0 rrc: 4 type: EXT [0->18446744073709551615] (req 0->134217727) gid 0 flags: 0x60000400020020 nid: 0@lo remote: 0xfc2cc2c14c34caaf expref: 3347 pid: 13092 timeout: 4280 lvb_type: 0 LustreError: 12392:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 16 previous similar messages LustreError: 20208:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802d5059300 x1709942374141312/t0(0) o105->lustre-OST0001@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 20208:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 184 previous similar messages LustreError: 11-0: lustre-OST0001-osc-ffff8802a433e678: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 1 previous similar message Lustre: lustre-OST0001-osc-ffff8802a433e678: Connection to lustre-OST0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 3 previous similar messages LustreError: 17733:0:(ldlm_lockd.c:1425:ldlm_handle_enqueue0()) ### lock on destroyed export ffff8802a4cc6678 ns: filter-lustre-OST0002_UUID lock: ffff8802d2168b80/0xfc2cc2c14c3a582d lrc: 3/0,0 mode: --/PW res: [0x340000400:0x457:0x0].0x0 rrc: 5 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x50000000020000 nid: 0@lo remote: 0xfc2cc2c14c3a5818 expref: 1773 pid: 17733 timeout: 0 lvb_type: 0 LustreError: 12377:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630728942 with bad export cookie 18171112731995715794 LustreError: 12377:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) Skipped 32 previous similar messages LustreError: 167-0: lustre-OST0002-osc-ffff8802a433e678: This client was evicted by lustre-OST0002; in progress operations using this service will fail. LustreError: Skipped 3 previous similar messages Lustre: 9199:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x240000404:0x101:0x0]/ may get corrupted (rc -108) Lustre: 9198:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x200000404:0x171:0x0]/ may get corrupted (rc -108) Lustre: 9200:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x280000404:0xdd:0x0]// may get corrupted (rc -108) Lustre: 9200:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x280000403:0x18b:0x0]/ may get corrupted (rc -108) Lustre: 9201:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x280000403:0x18d:0x0]// may get corrupted (rc -108) Lustre: 9201:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x280000404:0xfa:0x0]// may get corrupted (rc -108) Lustre: 9212:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x200000404:0x25e:0x0]// may get corrupted (rc -108) LustreError: 27648:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0001-osc-ffff8802a433e678: namespace resource [0x300000400:0x360:0x0].0x0 (ffff8802c2a9afc0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 27648:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 183 previous similar messages Lustre: lustre-OST0000-osc-ffff8802a433e678: Connection restored to 192.168.123.215@tcp (at 0@lo) Lustre: Skipped 3 previous similar messages LustreError: 27640:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0001-osc-ffff8802a4db5d28: namespace resource [0x6e0:0x0:0x0].0x0 (ffff8802cdfaaac0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 27640:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 117 previous similar messages Lustre: 15484:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 499 < left 1322, rollback = 2 Lustre: 15484:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 59 previous similar messages LustreError: 19228:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000403:0x10e:0x0]: rc = -2 LustreError: 18589:0:(mdd_object.c:401:mdd_xattr_get()) lustre-MDD0000: object [0x200000403:0x10e:0x0] not found: rc = -2 LustreError: 12894:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0000-osp-MDT0002: fail to cancel 1 llog-records: rc = -2 LustreError: 12894:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0000-osp-MDT0002: fail to cancel 1 of 1 llog-records: rc = -2 LustreError: 12392:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 101s: evicting client at 0@lo ns: mdt-lustre-MDT0002_UUID lock: ffff8802d65252c0/0xfc2cc2c14c3cdd30 lrc: 3/0,0 mode: PR/PR res: [0x280000404:0xed:0x0].0x0 bits 0x12/0x0 rrc: 14 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0xfc2cc2c14c3cdc6c expref: 135 pid: 15441 timeout: 4437 lvb_type: 0 LustreError: 12392:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 3 previous similar messages LustreError: 25593:0:(ldlm_lockd.c:1425:ldlm_handle_enqueue0()) ### lock on destroyed export ffff8802a4bea548 ns: mdt-lustre-MDT0002_UUID lock: ffff8802d888c000/0xfc2cc2c14c3e0505 lrc: 3/0,0 mode: PR/PR res: [0x280000404:0xed:0x0].0x0 bits 0x12/0x0 rrc: 12 type: IBT gid 0 flags: 0x50200000000000 nid: 0@lo remote: 0xfc2cc2c14c3ce112 expref: 21 pid: 25593 timeout: 0 lvb_type: 0 LustreError: 11-0: lustre-MDT0002-mdc-ffff8802a433e678: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 37 previous similar messages Lustre: lustre-MDT0002-mdc-ffff8802a433e678: Connection to lustre-MDT0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 3 previous similar messages LustreError: 167-0: lustre-MDT0002-mdc-ffff8802a433e678: This client was evicted by lustre-MDT0002; in progress operations using this service will fail. LustreError: Skipped 3 previous similar messages LustreError: 25313:0:(file.c:5042:ll_inode_revalidate_fini()) lustre: revalidate FID [0x280000404:0xed:0x0] error: rc = -5 LustreError: 25313:0:(file.c:5042:ll_inode_revalidate_fini()) Skipped 6 previous similar messages LustreError: 28008:0:(file.c:5042:ll_inode_revalidate_fini()) lustre: revalidate FID [0x240000403:0x155:0x0] error: rc = -108 LustreError: 28008:0:(file.c:5042:ll_inode_revalidate_fini()) Skipped 2 previous similar messages LustreError: 29364:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff8802a433e678: inode [0x280000404:0x5f:0x0] mdc close failed: rc = -108 LustreError: 29364:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-MDT0002-mdc-ffff8802a433e678: namespace resource [0x280000403:0x1:0x0].0x0 (ffff88007b216340) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 29364:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 81 previous similar messages Lustre: lustre-MDT0002-mdc-ffff8802a433e678: Connection restored to 192.168.123.215@tcp (at 0@lo) Lustre: Skipped 3 previous similar messages LustreError: 20896:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88009d3725c0 x1709942379063040/t0(0) o104->lustre-OST0002@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 20896:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 6 previous similar messages Lustre: 9198:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x280000405:0x3:0x0]// may get corrupted (rc -108) LustreError: 12392:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 96s: evicting client at 0@lo ns: filter-lustre-OST0002_UUID lock: ffff8802cda41e40/0xfc2cc2c14c3ef0b9 lrc: 3/0,0 mode: PW/PW res: [0x340000401:0x6ff:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000000020020 nid: 0@lo remote: 0xfc2cc2c14c3ef0ab expref: 4045 pid: 19657 timeout: 4475 lvb_type: 0 LustreError: 12392:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 1 previous similar message LustreError: 11-0: lustre-OST0002-osc-ffff8802a4db5d28: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 7 previous similar messages LustreError: 167-0: lustre-OST0002-osc-ffff8802a4db5d28: This client was evicted by lustre-OST0002; in progress operations using this service will fail. LustreError: Skipped 1 previous similar message LustreError: 24175:0:(ofd_obd.c:501:ofd_destroy_export()) lustre-OST0002: cli 8263f8d3-ce0c-4db1-ac74-5bc989b69a05/ffff8802498153d8 has 61440 pending on destroyed export LustreError: 24175:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_destroy_export: tot_granted 11002816 != fo_tot_granted 11064256 LustreError: 24175:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_destroy_export: tot_pending 0 != fo_tot_pending 61440 LustreError: 20911:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 11002816 != fo_tot_granted 11064256 LustreError: 20911:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 61440 Lustre: 18589:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 6s req@ffff8802c96c9940 x1709942379683520/t0(0) o101->7a8d0ba0-86fb-4c66-8dd9-af0523f59dad@0@lo:0/0 lens 576/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'ls.0' Lustre: 18589:0:(service.c:2157:ptlrpc_server_handle_req_in()) Skipped 1621 previous similar messages LustreError: 13112:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 11002816 != fo_tot_granted 11064256 LustreError: 13112:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 61440 LustreError: 13111:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 11002816 != fo_tot_granted 11064256 LustreError: 13111:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 61440 LustreError: 20911:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 11002816 != fo_tot_granted 11064256 LustreError: 20911:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 1 previous similar message LustreError: 20911:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 61440 LustreError: 20911:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 1 previous similar message Lustre: ost_io: This server is not able to keep up with request traffic (cpu-bound). LustreError: 16129:0:(service.c:2290:ptlrpc_server_handle_request()) @@@ Dropping timed-out request from 12345-0@lo: deadline 6/2s ago req@ffff8802c299b880 x1709942380122048/t0(0) o4->8263f8d3-ce0c-4db1-ac74-5bc989b69a05@0@lo:123/0 lens 488/0 e 0 to 0 dl 1630729153 ref 2 fl Interpret:/0/ffffffff rc 0/-1 job:'dd.0' Lustre: 16129:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (6/2s); client may timeout req@ffff8802c299b880 x1709942380122048/t0(0) o4->8263f8d3-ce0c-4db1-ac74-5bc989b69a05@0@lo:123/0 lens 488/0 e 0 to 0 dl 1630729153 ref 2 fl Interpret:/0/ffffffff rc 0/-1 job:'dd.0' Lustre: 13147:0:(service.c:1612:ptlrpc_at_check_timed()) earlyQ=1 reqQ=0 recA=1, svcEst=35, delay=2ms Lustre: 13147:0:(service.c:1378:ptlrpc_at_send_early_reply()) @@@ Already past deadline (-2s), not sending early reply. Consider increasing at_early_margin (5)? req@ffff8802c299b880 x1709942380122048/t0(0) o4->8263f8d3-ce0c-4db1-ac74-5bc989b69a05@0@lo:123/0 lens 488/0 e 0 to 0 dl 1630729153 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'dd.0' Lustre: 9207:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x280000404:0x272:0x0]// may get corrupted (rc -108) Lustre: 9206:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x200000404:0x2c5:0x0]// may get corrupted (rc -108) Lustre: 9207:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x200000404:0x2cc:0x0]// may get corrupted (rc -108) Lustre: 9206:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x280000403:0x18d:0x0]// may get corrupted (rc -108) Lustre: 9206:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x200000404:0x25d:0x0]// may get corrupted (rc -108) LustreError: 28267:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 18957248 != fo_tot_granted 19018688 LustreError: 28267:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 3 previous similar messages LustreError: 28267:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 61440 LustreError: 28267:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 3 previous similar messages LustreError: 29962:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0002-osc-ffff8802a4db5d28: namespace resource [0x340000400:0x680:0x0].0x0 (ffff8802b659de40) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 29962:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 32 previous similar messages Lustre: lustre-OST0002-osc-ffff8802a4db5d28: Connection restored to 192.168.123.215@tcp (at 0@lo) Lustre: Skipped 1 previous similar message LustreError: 13119:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 35746752 != fo_tot_granted 35808192 LustreError: 13119:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 5 previous similar messages LustreError: 13119:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 61440 LustreError: 13119:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 5 previous similar messages Lustre: 9210:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1630729147/real 1630729147] req@ffff8802c2998cc0 x1709942380122048/t0(0) o4->lustre-OST0003-osc-ffff8802a4db5d28@0@lo:6/4 lens 488/448 e 0 to 1 dl 1630729182 ref 2 fl Rpc:XQr/0/ffffffff rc 0/-1 job:'dd.0' Lustre: lustre-OST0003-osc-ffff8802a4db5d28: Connection to lustre-OST0003 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 2 previous similar messages Lustre: lustre-OST0003: Client 8263f8d3-ce0c-4db1-ac74-5bc989b69a05 (at 0@lo) reconnecting LustreError: 13750:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 82037696 != fo_tot_granted 82099136 LustreError: 13750:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 7 previous similar messages LustreError: 13750:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 1454080 != fo_tot_pending 1515520 LustreError: 13750:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 7 previous similar messages Lustre: 12432:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 6s req@ffff8802a85425c0 x1709942382986688/t0(0) o35->7a8d0ba0-86fb-4c66-8dd9-af0523f59dad@0@lo:0/0 lens 392/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'cat.0' Lustre: 12432:0:(service.c:2157:ptlrpc_server_handle_req_in()) Skipped 113 previous similar messages LustreError: 13110:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 132554944 != fo_tot_granted 132616384 LustreError: 13110:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 27 previous similar messages LustreError: 13110:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 61440 LustreError: 13110:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 27 previous similar messages LustreError: 12392:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0001_UUID lock: ffff8802dc638040/0xfc2cc2c14c44b183 lrc: 3/0,0 mode: PW/PW res: [0x7c2:0x0:0x0].0x0 rrc: 4 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000400020020 nid: 0@lo remote: 0xfc2cc2c14c44b17c expref: 2311 pid: 13109 timeout: 4707 lvb_type: 0 LustreError: 12383:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630729366 with bad export cookie 18171112731996485990 LustreError: 11-0: lustre-OST0001-osc-ffff8802a433e678: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-OST0001-osc-ffff8802a433e678: Connection to lustre-OST0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 17882:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802cdf7d180 x1709942388260160/t0(0) o104->lustre-OST0001@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 167-0: lustre-OST0001-osc-ffff8802a433e678: This client was evicted by lustre-OST0001; in progress operations using this service will fail. LustreError: 17882:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 19 previous similar messages LustreError: 1238:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0001-osc-ffff8802a433e678: namespace resource [0x7d0:0x0:0x0].0x0 (ffff8802d3b02ac0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 1238:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 431 previous similar messages Lustre: lustre-OST0001-osc-ffff8802a433e678: Connection restored to 192.168.123.215@tcp (at 0@lo) Lustre: Skipped 1 previous similar message Lustre: 18589:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 627/2508/0, destroy: 1/4/0 Lustre: 18589:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 199227 previous similar messages Lustre: 18589:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 3/3/1, xattr_set: 631/148/0 Lustre: 18589:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 199227 previous similar messages Lustre: 18589:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 3136/26971/0, punch: 0/0/0, quota 6/6/0 Lustre: 18589:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 199229 previous similar messages Lustre: 18589:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 628/10675/0, delete: 2/5/1 Lustre: 18589:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 199227 previous similar messages Lustre: 18589:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 1/1/0, ref_del: 2/2/1 Lustre: 18589:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 199229 previous similar messages Lustre: 14998:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 493 < left 14354, rollback = 7 Lustre: 14998:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 199942 previous similar messages LustreError: 15301:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0001: failed to get lu_attr of [0x240000404:0x29b:0x0]: rc = -2 LustreError: 13115:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 259936576 != fo_tot_granted 259998016 LustreError: 13115:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 63 previous similar messages LustreError: 13115:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 61440 LustreError: 13115:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 63 previous similar messages LustreError: 11-0: lustre-OST0002-osc-ffff8802a4db5d28: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 5 previous similar messages Lustre: 9207:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x240000404:0x324:0x0]// may get corrupted (rc -108) Lustre: 9206:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x280000404:0x388:0x0]/ may get corrupted (rc -108) Lustre: 9206:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x200000404:0x357:0x0]/ may get corrupted (rc -108) Lustre: 9207:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x280000404:0x3a6:0x0]// may get corrupted (rc -108) Lustre: 9207:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x200000404:0x354:0x0]/ may get corrupted (rc -108) Lustre: 13147:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 6s req@ffff8802b3cede00 x1709942391114304/t0(0) o10->7a8d0ba0-86fb-4c66-8dd9-af0523f59dad@0@lo:0/0 lens 440/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'dir_create.sh.0' Lustre: 9202:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff8802d008c500 x1709942393025408/t0(0) o10->lustre-OST0003-osc-ffff8802a433e678@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630729557 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'cp.0' Lustre: 9208:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff88027ed9cb40 x1709942393189248/t0(0) o10->lustre-OST0001-osc-ffff8802a4db5d28@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630729557 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dir_create.sh.0' LustreError: 12392:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0002_UUID lock: ffff8800a69c4b40/0xfc2cc2c14c4fa27f lrc: 3/0,0 mode: PW/PW res: [0x340000400:0xcfc:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000000030020 nid: 0@lo remote: 0xfc2cc2c14c4fa263 expref: 3745 pid: 20210 timeout: 4874 lvb_type: 0 LustreError: 12392:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 1 previous similar message LustreError: 16343:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff880097870040 x1709942394851136/t0(0) o105->lustre-OST0002@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 15876:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630729534 with bad export cookie 18171112731996896428 LustreError: 15876:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) Skipped 1 previous similar message LustreError: 16343:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 31 previous similar messages LustreError: 11-0: lustre-OST0002-osc-ffff8802a433e678: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 1 previous similar message LustreError: 167-0: lustre-OST0002-osc-ffff8802a433e678: This client was evicted by lustre-OST0002; in progress operations using this service will fail. LustreError: Skipped 1 previous similar message Lustre: 9211:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x200000403:0x32c:0x0]/ may get corrupted (rc -5) Lustre: 9211:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x280000404:0x2b4:0x0]/ may get corrupted (rc -5) Lustre: 9211:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x280000405:0x102:0x0]/ may get corrupted (rc -108) Lustre: 9211:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x200000404:0x338:0x0]/ may get corrupted (rc -108) Lustre: 17063:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 8: before 509 < left 33455, rollback = 9 Lustre: 17063:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 312 previous similar messages LustreError: 4405:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0002-osc-ffff8802a433e678: namespace resource [0x340000401:0x12a5:0x0].0x0 (ffff8802f4eb7740) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 4405:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 123 previous similar messages Lustre: lustre-OST0002-osc-ffff8802a433e678: Connection restored to 192.168.123.215@tcp (at 0@lo) Lustre: Skipped 1 previous similar message LustreError: 3399:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff8802a4db5d28: inode [0x280000404:0x3a9:0x0] mdc close failed: rc = -13 LustreError: 3399:0:(file.c:234:ll_close_inode_openhandle()) Skipped 10 previous similar messages Lustre: 9201:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff880223003880 x1709942397772736/t0(0) o10->lustre-OST0000-osc-ffff8802a433e678@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630729683 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dir_create.sh.0' ptlrpc_watchdog_fire: 24 callbacks suppressed Lustre: ll_ost01_047: service thread pid 19505 was inactive for 40.123 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Lustre: Skipped 1 previous similar message Pid: 19505, comm: ll_ost01_047 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] ofd_destroy_by_fid+0x1d1/0x520 [ofd] [<0>] ofd_destroy_hdl+0x27f/0xa20 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 19503, comm: ll_ost01_045 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] ofd_destroy_by_fid+0x1d1/0x520 [ofd] [<0>] ofd_destroy_hdl+0x27f/0xa20 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: ll_ost01_022: service thread pid 19143 was inactive for 43.150 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Lustre: Skipped 1 previous similar message Pid: 19143, comm: ll_ost01_022 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] ofd_destroy_by_fid+0x1d1/0x520 [ofd] Lustre: ll_ost01_007: service thread pid 16972 was inactive for 43.095 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 2 previous similar messages [<0>] ofd_destroy_hdl+0x27f/0xa20 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe LustreError: 13124:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 77677888 != fo_tot_granted 77739328 LustreError: 13124:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 106 previous similar messages LustreError: 13124:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 61440 LustreError: 13124:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 106 previous similar messages Lustre: 9201:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff880223003880 x1709942400012032/t0(0) o10->lustre-OST0000-osc-ffff8802a433e678@0@lo:6/4 lens 440/432 e 1 to 0 dl 1630729747 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dir_create.sh.0' LustreError: 12694:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0000-osp-MDT0001: fail to cancel 1 llog-records: rc = -116 LustreError: 12694:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0000-osp-MDT0001: fail to cancel 1 of 1 llog-records: rc = -116 Lustre: 9212:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff8802cc0c0040 x1709942400456640/t0(0) o10->lustre-OST0003-osc-ffff8802a4db5d28@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630729758 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dir_create.sh.0' LustreError: 12694:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0000-osp-MDT0001: fail to cancel 1 llog-records: rc = -2 LustreError: 12694:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0000-osp-MDT0001: fail to cancel 1 of 1 llog-records: rc = -2 Lustre: ll_ost06_013: service thread pid 17816 was inactive for 40.025 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 4 previous similar messages Lustre: dir [0x240000404:0x37c:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 10 previous similar messages Lustre: ll_ost06_047: service thread pid 6308 was inactive for 40.096 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 43 previous similar messages Lustre: ll_ost06_054: service thread pid 6327 was inactive for 40.059 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 7 previous similar messages Lustre: dir [0x200000404:0x350:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 2 previous similar messages LustreError: 12392:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 103s: evicting client at 0@lo ns: filter-lustre-OST0003_UUID lock: ffff8802ea2b0400/0xfc2cc2c14c522b56 lrc: 3/0,0 mode: PW/PW res: [0x380000401:0xcd8:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000400020020 nid: 0@lo remote: 0xfc2cc2c14c522b1e expref: 4591 pid: 19446 timeout: 5052 lvb_type: 0 LustreError: 12392:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 1 previous similar message LustreError: 18375:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630729715 with bad export cookie 18171112731995715668 LustreError: 18375:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) Skipped 1 previous similar message LustreError: 11-0: lustre-OST0003-osc-ffff8802a4db5d28: operation ost_getattr to node 0@lo failed: rc = -107 Lustre: lustre-OST0003-osc-ffff8802a4db5d28: Connection to lustre-OST0003 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 3 previous similar messages LustreError: 167-0: lustre-OST0003-osc-ffff8802a4db5d28: This client was evicted by lustre-OST0003; in progress operations using this service will fail. LustreError: Skipped 1 previous similar message Lustre: 12429:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 8s req@ffff8802d213d7c0 x1709942400830656/t0(0) o37->8263f8d3-ce0c-4db1-ac74-5bc989b69a05@0@lo:0/0 lens 448/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'lfs.0' Lustre: 12429:0:(service.c:2157:ptlrpc_server_handle_req_in()) Skipped 24 previous similar messages Lustre: 9199:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x280000405:0x131:0x0]// may get corrupted (rc -108) Lustre: 9198:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x200000404:0x338:0x0]/ may get corrupted (rc -108) Lustre: 9198:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x280000404:0x469:0x0]/ may get corrupted (rc -108) Lustre: 9199:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x240000403:0x372:0x0]/ may get corrupted (rc -108) Lustre: mdt_out: This server is not able to keep up with request traffic (cpu-bound). Lustre: 15848:0:(service.c:1612:ptlrpc_at_check_timed()) earlyQ=1 reqQ=0 recA=0, svcEst=5, delay=0ms Lustre: 15848:0:(service.c:1378:ptlrpc_at_send_early_reply()) @@@ Already past deadline (-2s), not sending early reply. Consider increasing at_early_margin (5)? req@ffff8802be260cc0 x1709942401045440/t0(0) o1000->lustre-MDT0000-mdtlov_UUID@0@lo:695/0 lens 304/0 e 0 to 0 dl 1630729725 ref 2 fl New:/0/ffffffff rc 0/-1 job:'mdt07_010.0' LustreError: 15848:0:(service.c:2290:ptlrpc_server_handle_request()) @@@ Dropping timed-out request from 12345-0@lo: deadline 8/2s ago req@ffff8802be260cc0 x1709942401045440/t0(0) o1000->lustre-MDT0000-mdtlov_UUID@0@lo:695/0 lens 304/0 e 0 to 0 dl 1630729725 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'mdt07_010.0' Lustre: 15848:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (8/2s); client may timeout req@ffff8802be260cc0 x1709942401045440/t0(0) o1000->lustre-MDT0000-mdtlov_UUID@0@lo:695/0 lens 304/0 e 0 to 0 dl 1630729725 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'mdt07_010.0' Lustre: 9199:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x280000404:0x187:0x0]/ may get corrupted (rc -108) Lustre: ldlm_canceld: This server is not able to keep up with request traffic (cpu-bound). Lustre: 12370:0:(service.c:1612:ptlrpc_at_check_timed()) earlyQ=2 reqQ=0 recA=2, svcEst=4, delay=5852ms Lustre: 12370:0:(service.c:1378:ptlrpc_at_send_early_reply()) @@@ Already past deadline (-2s), not sending early reply. Consider increasing at_early_margin (5)? req@ffff8802de265e00 x1709942401028224/t0(0) o103->lustre-MDT0001-mdtlov_UUID@0@lo:697/0 lens 328/224 e 0 to 0 dl 1630729727 ref 2 fl Interpret:H/0/0 rc 0/0 job:'ldlm_bl_07.0' Lustre: 12371:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (12/2s); client may timeout req@ffff8802de265e00 x1709942401028224/t0(0) o103->lustre-MDT0001-mdtlov_UUID@0@lo:697/0 lens 328/224 e 0 to 0 dl 1630729727 ref 1 fl Complete:H/0/0 rc 0/0 job:'ldlm_bl_07.0' LustreError: 18809:0:(service.c:2290:ptlrpc_server_handle_request()) @@@ Dropping timed-out request from 12345-0@lo: deadline 56/1s ago req@ffff8802ae207700 x1709942399552640/t0(0) o1->8263f8d3-ce0c-4db1-ac74-5bc989b69a05@0@lo:698/0 lens 440/0 e 0 to 0 dl 1630729728 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'lfs.0' Lustre: 9200:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x280000404:0x429:0x0]/ may get corrupted (rc -108) Lustre: 9201:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x200000404:0x3a9:0x0]// may get corrupted (rc -108) Lustre: 9200:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x280000404:0x445:0x0]// may get corrupted (rc -108) Lustre: 9200:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x200000404:0x39f:0x0]// may get corrupted (rc -108) Lustre: 9201:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x280000405:0x14c:0x0]/ may get corrupted (rc -108) Lustre: 19648:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (56/1s); client may timeout req@ffff8802cdbaea80 x1709942399535744/t0(0) o1->8263f8d3-ce0c-4db1-ac74-5bc989b69a05@0@lo:700/0 lens 440/0 e 0 to 0 dl 1630729730 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'lfs.0' Lustre: 19648:0:(service.c:2328:ptlrpc_server_handle_request()) Skipped 41 previous similar messages LustreError: 19645:0:(service.c:2290:ptlrpc_server_handle_request()) @@@ Dropping timed-out request from 12345-0@lo: deadline 56/2s ago req@ffff8802cc0c3240 x1709942399561280/t0(0) o1->8263f8d3-ce0c-4db1-ac74-5bc989b69a05@0@lo:700/0 lens 440/0 e 0 to 0 dl 1630729730 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'lfs.0' LustreError: 19645:0:(service.c:2290:ptlrpc_server_handle_request()) Skipped 42 previous similar messages Lustre: 18799:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (57/2s); client may timeout req@ffff88009da4b880 x1709942399558848/t0(0) o1->8263f8d3-ce0c-4db1-ac74-5bc989b69a05@0@lo:703/0 lens 440/0 e 1 to 0 dl 1630729733 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'lfs.0' Lustre: 18799:0:(service.c:2328:ptlrpc_server_handle_request()) Skipped 66 previous similar messages LustreError: 6331:0:(service.c:2290:ptlrpc_server_handle_request()) @@@ Dropping timed-out request from 12345-0@lo: deadline 57/3s ago req@ffff8802a4d9c500 x1709942399559488/t0(0) o1->8263f8d3-ce0c-4db1-ac74-5bc989b69a05@0@lo:703/0 lens 440/0 e 1 to 0 dl 1630729733 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'lfs.0' LustreError: 6331:0:(service.c:2290:ptlrpc_server_handle_request()) Skipped 71 previous similar messages LustreError: 6356:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0002: failed to get lu_attr of [0x280000404:0x9a:0x0]: rc = -2 Lustre: 18599:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1630729717/real 1630729717] req@ffff8802be266440 x1709942401045440/t0(0) o1000->lustre-MDT0001-osp-MDT0000@0@lo:24/4 lens 304/4320 e 0 to 1 dl 1630729737 ref 2 fl Rpc:XQr/0/ffffffff rc 0/-1 job:'mdt07_010.0' Lustre: lustre-MDT0001: Received new MDS connection from 0@lo, keep former export from same NID LustreError: 6306:0:(service.c:2290:ptlrpc_server_handle_request()) @@@ Dropping timed-out request from 12345-0@lo: deadline 56/6s ago req@ffff8802bae78cc0 x1709942399958464/t0(0) o2->8263f8d3-ce0c-4db1-ac74-5bc989b69a05@0@lo:711/0 lens 440/0 e 0 to 0 dl 1630729741 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'touch.0' LustreError: 6306:0:(service.c:2290:ptlrpc_server_handle_request()) Skipped 189 previous similar messages Lustre: 6306:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (56/6s); client may timeout req@ffff8802bae78cc0 x1709942399958464/t0(0) o2->8263f8d3-ce0c-4db1-ac74-5bc989b69a05@0@lo:711/0 lens 440/0 e 0 to 0 dl 1630729741 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'touch.0' Lustre: 6306:0:(service.c:2328:ptlrpc_server_handle_request()) Skipped 196 previous similar messages LustreError: 12427:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0001: failed to get lu_attr of [0x240000403:0x9b:0x0]: rc = -2 Lustre: 9212:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff8800acc1e440 x1709942402725696/t0(0) o10->lustre-OST0002-osc-ffff8802a433e678@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630729849 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dd.0' Lustre: mdt04_010: service thread pid 18651 was inactive for 84.105 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 9 previous similar messages LustreError: 12431:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000403:0x358:0x0]: rc = -2 LustreError: 20573:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630729826 with bad export cookie 18171112731996486011 LustreError: 20573:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) Skipped 4 previous similar messages LustreError: 12373:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802d77f1f80 x1709942404464192/t0(0) o105->lustre-OST0000@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 12373:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 237 previous similar messages Lustre: 9198:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x200000403:0x3d1:0x0]// may get corrupted (rc -108) Lustre: 9198:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x240000403:0x3c1:0x0]/ may get corrupted (rc -108) Lustre: 9199:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x280000405:0x182:0x0]// may get corrupted (rc -108) Lustre: lustre-OST0000-osc-ffff8802a433e678: Connection restored to 192.168.123.215@tcp (at 0@lo) Lustre: Skipped 4 previous similar messages LustreError: 19515:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630729938 with bad export cookie 18171112731998696058 Lustre: 9213:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x240000404:0x42c:0x0]/ may get corrupted (rc -5) Lustre: 9213:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x280000405:0x131:0x0]// may get corrupted (rc -5) Lustre: 9213:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x200000404:0x51c:0x0]/ may get corrupted (rc -5) Lustre: 9213:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x240000404:0x448:0x0]// may get corrupted (rc -5) LustreError: 8211:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff8802a433e678: inode [0x280000404:0x3a9:0x0] mdc close failed: rc = -13 Lustre: 9204:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x240000403:0x49b:0x0]// may get corrupted (rc -108) LustreError: 9294:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0003-osc-ffff8802a4db5d28: namespace resource [0x380000401:0xcdf:0x0].0x0 (ffff880251d1ed40) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 9294:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 4450 previous similar messages ------------[ cut here ]------------ WARNING: CPU: 2 PID: 12402 at /home/green/git/lustre-release/ldiskfs/ext4_jbd2.c:266 __ldiskfs_handle_dirty_metadata+0x1d2/0x230 [ldiskfs] Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) dm_flakey dm_mod loop zfs(PO) zunicode(PO) zzstd(O) zlua(O) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) jbd2 mbcache crc_t10dif crct10dif_generic sb_edac edac_core iosf_mbi crc32_pclmul ghash_clmulni_intel aesni_intel lrw gf128mul glue_helper ablk_helper cryptd virtio_balloon pcspkr virtio_console i2c_piix4 ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper crct10dif_pclmul ttm crct10dif_common drm crc32c_intel ata_piix drm_panel_orientation_quirks serio_raw virtio_blk libata i2c_core floppy [last unloaded: libcfs] CPU: 2 PID: 12402 Comm: mdt01_000 Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011 Call Trace: [<ffffffff817d9418>] dump_stack+0x19/0x1b [<ffffffff8108d538>] __warn+0xd8/0x100 [<ffffffff8108d67d>] warn_slowpath_null+0x1d/0x20 [<ffffffffa0cd8262>] __ldiskfs_handle_dirty_metadata+0x1d2/0x230 [ldiskfs] [<ffffffffa0cb833d>] ldiskfs_getblk+0x13d/0x210 [ldiskfs] [<ffffffffa0cb8437>] ldiskfs_bread+0x27/0xe0 [ldiskfs] [<ffffffffa0d54e97>] osd_write+0x4e7/0xcb0 [osd_ldiskfs] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffffa05126f3>] dt_record_write+0x33/0x120 [obdclass] [<ffffffffa04d16b2>] llog_osd_write_rec+0x1052/0x19a0 [obdclass] [<ffffffffa04be760>] llog_write_rec+0x290/0x590 [obdclass] [<ffffffffa04c44ee>] llog_cat_add_rec+0x23e/0x950 [obdclass] [<ffffffffa04bb5ff>] llog_add+0x17f/0x1f0 [obdclass] [<ffffffffa04eca39>] ? lprocfs_counter_add+0xf9/0x160 [obdclass] [<ffffffffa085560d>] sub_updates_write+0xb6b/0xe32 [ptlrpc] [<ffffffffa0825892>] top_trans_stop+0x7c2/0xf60 [ptlrpc] [<ffffffffa0f70f3c>] lod_trans_stop+0x25c/0x340 [lod] [<ffffffffa0e45408>] mdd_trans_stop+0x28/0x16e [mdd] [<ffffffffa0e26b2f>] mdd_migrate_object+0xf1f/0x1ba0 [mdd] [<ffffffffa0511375>] ? lu_object_find_at+0x285/0xb10 [obdclass] [<ffffffffa0e27b86>] mdd_migrate+0x3d6/0x8e0 [mdd] [<ffffffffa0ef90d8>] mdo_migrate+0x4c/0x4e [mdt] [<ffffffffa0eb8e53>] mdt_reint_migrate+0x1023/0x11d0 [mdt] [<ffffffffa0eb9087>] mdt_reint_rec+0x87/0x240 [mdt] [<ffffffffa0e8e55c>] mdt_reint_internal+0x76c/0xb40 [mdt] [<ffffffffa0e9b657>] mdt_reint+0x67/0x150 [mdt] [<ffffffffa081426e>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<ffffffffa034745e>] ? libcfs_nid2str_r+0xfe/0x130 [lnet] [<ffffffffa07bf990>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<ffffffffa07c1529>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffffa07c0950>] ? ptlrpc_wait_event+0x620/0x620 [ptlrpc] [<ffffffff810ba0f4>] kthread+0xe4/0xf0 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 ---[ end trace c7297906fceace19 ]--- LDISKFS-fs: ldiskfs_getblk:888: aborting transaction: error 28 in __ldiskfs_handle_dirty_metadata LDISKFS-fs error (device dm-2): ldiskfs_getblk:888: inode #159: block 51697: comm mdt01_000: journal_dirty_metadata failed: handle type 0 started at line 1972, credits 512/0, errcode -28 Aborting journal on device dm-2-8. LDISKFS-fs (dm-2): Remounting filesystem read-only LustreError: 12402:0:(osd_io.c:2103:osd_ldiskfs_write_record()) lustre-MDT0002: error reading offset 401408 (block 98, size 40, offs 401392), credits 0/0: rc = -28 LustreError: 12402:0:(update_trans.c:985:top_trans_stop()) lustre-MDT0002-osd: write updates failed: rc = -28 LustreError: 12402:0:(osd_io.c:2110:osd_ldiskfs_write_record()) journal_get_write_access() returned error -30 LustreError: 12402:0:(tgt_lastrcvd.c:1253:tgt_add_reply_data()) lustre-MDT0002: can't update reply_data file: rc = -30 LustreError: 12402:0:(osd_handler.c:2075:osd_trans_stop()) lustre-MDT0002: failed in transaction hook: rc = -30 LDISKFS-fs error (device dm-2) in osd_trans_stop:2082: error 28 LustreError: 13870:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-OST0003-osc-MDT0002: fail to cancel 1 llog-records: rc = -30 LustreError: 13870:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-OST0003-osc-MDT0002: fail to cancel 1 of 1 llog-records: rc = -30 LustreError: 9676:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -30 LustreError: 12876:0:(osd_handler.c:1779:osd_trans_commit_cb()) transaction @0xffff8802bfbb1e18 commit error: 2 LustreError: 12402:0:(osd_handler.c:2085:osd_trans_stop()) lustre-MDT0002: failed to stop transaction: rc = -28 LustreError: 12402:0:(update_trans.c:1011:top_trans_stop()) lustre-MDT0002-osd: stop trans failed: rc = -30 LustreError: 13870:0:(osp_sync.c:1079:osp_sync_process_committed()) lustre-OST0003-osc-MDT0002: can't cancel record: rc = -30 LustreError: 13588:0:(osp_sync.c:1079:osp_sync_process_committed()) lustre-OST0001-osc-MDT0002: can't cancel record: rc = -30 ptlrpc_watchdog_fire: 68 callbacks suppressed Lustre: ll_ost07_023: service thread pid 19382 was inactive for 68.058 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: ll_ost07_002: service thread pid 13109 was inactive for 68.070 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 13109, comm: ll_ost07_002 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] tgt_extent_lock+0xea/0x2a0 [ptlrpc] [<0>] ofd_getattr_hdl+0x365/0x720 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 19501, comm: ll_ost07_025 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] tgt_extent_lock+0xea/0x2a0 [ptlrpc] [<0>] ofd_getattr_hdl+0x365/0x720 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 6625, comm: ll_ost07_063 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] tgt_extent_lock+0xea/0x2a0 [ptlrpc] [<0>] ofd_getattr_hdl+0x365/0x720 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: ll_ost02_016: service thread pid 18214 was inactive for 86.264 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 81 previous similar messages LustreError: 12392:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 101s: evicting client at 0@lo ns: filter-lustre-OST0001_UUID lock: ffff8800a2bdad40/0xfc2cc2c14c652dbb lrc: 3/0,0 mode: PW/PW res: [0x300000400:0x2660:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000400020020 nid: 0@lo remote: 0xfc2cc2c14c652db4 expref: 2600 pid: 16951 timeout: 5374 lvb_type: 0 LustreError: 12392:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 3 previous similar messages LustreError: 11-0: lustre-OST0002-osc-ffff8802a433e678: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 31 previous similar messages LustreError: 167-0: lustre-OST0002-osc-ffff8802a433e678: This client was evicted by lustre-OST0002; in progress operations using this service will fail. LustreError: Skipped 3 previous similar messages Lustre: 9204:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x240000403:0x4bb:0x0]// may get corrupted (rc -108) LustreError: 9857:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -30 LustreError: 13868:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-OST0002-osc-MDT0002: fail to cancel 198 llog-records: rc = -30 LustreError: 13868:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) Skipped 2 previous similar messages LustreError: 13868:0:(osp_sync.c:1094:osp_sync_process_committed()) lustre-OST0002-osc-MDT0002: can't cancel 198 records: rc = -30 LustreError: 699:0:(ofd_obd.c:501:ofd_destroy_export()) lustre-OST0003: cli 7a8d0ba0-86fb-4c66-8dd9-af0523f59dad/ffff8802c4b253d8 has 61440 pending on destroyed export LustreError: 9864:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -30 LustreError: 13868:0:(osp_sync.c:1094:osp_sync_process_committed()) lustre-OST0002-osc-MDT0002: can't cancel 8 records: rc = -30 LustreError: 13868:0:(osp_sync.c:1094:osp_sync_process_committed()) lustre-OST0002-osc-MDT0002: can't cancel 46 records: rc = -30 Lustre: 14955:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 512 < left 1719, rollback = 7 Lustre: 14955:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 340915 previous similar messages Lustre: 14955:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 14955:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 345505 previous similar messages Lustre: 14955:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 1/89/0 Lustre: 14955:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 345505 previous similar messages Lustre: 14955:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1710/1719/0, punch: 0/0/0, quota 3/3/0 Lustre: 14955:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 345505 previous similar messages Lustre: 14955:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 14955:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 345504 previous similar messages Lustre: 14955:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 14955:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 345505 previous similar messages LustreError: 13868:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-OST0002-osc-MDT0002: fail to cancel 1 of 1 llog-records: rc = -30 LustreError: 13868:0:(llog_cat.c:787:llog_cat_cancel_records()) Skipped 2 previous similar messages LustreError: 13868:0:(osp_sync.c:1079:osp_sync_process_committed()) lustre-OST0002-osc-MDT0002: can't cancel record: rc = -30 LustreError: 8763:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -30 LustreError: 8763:0:(llite_lib.c:1836:ll_md_setattr()) Skipped 1 previous similar message LustreError: 12431:0:(llog_cat.c:602:llog_cat_add_rec()) llog_write_rec -116: lh=ffff880092c69098 LustreError: 12431:0:(update_trans.c:1062:top_trans_stop()) lustre-MDT0002-osp-MDT0000: write updates failed: rc = -116 | Externally reported by onyx-68 boilpot email |
racer test 1: racer on clients: centos-95.localnet DURATION=2700 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa17cae52>] ll_migrate+0x9b2/0xec0 [lustre] PGD 80000000a585c067 PUD bae45067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) dm_flakey dm_mod loop zfs(PO) zunicode(PO) zzstd(O) zlua(O) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) jbd2 mbcache crc_t10dif crct10dif_generic sb_edac edac_core iosf_mbi crc32_pclmul ghash_clmulni_intel aesni_intel lrw gf128mul glue_helper ablk_helper cryptd virtio_balloon pcspkr i2c_piix4 virtio_console ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm crct10dif_pclmul drm ata_piix crct10dif_common drm_panel_orientation_quirks crc32c_intel serio_raw virtio_blk i2c_core libata floppy [last unloaded: libcfs] CPU: 12 PID: 19679 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011 task: ffff8802e4945c40 ti: ffff8802d7a1c000 task.ti: ffff8802d7a1c000 RIP: 0010:[<ffffffffa17cae52>] [<ffffffffa17cae52>] ll_migrate+0x9b2/0xec0 [lustre] RSP: 0018:ffff8802d7a1fbc8 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff8802bf8b1778 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff880326b64228 RDI: ffff880326b64208 RBP: ffff8802d7a1fc48 R08: ffff8802d5372238 R09: ffff8802d7a1fa60 R10: 0000000000000001 R11: ffff8802ef0b57c0 R12: 0000000000000000 R13: ffff8802b448cf88 R14: ffff8802c5882618 R15: 0000000000000030 FS: 00007f6d176c4740(0000) GS:ffff880331d00000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 CR2: 0000000000000008 CR3: 00000000b5ab2000 CR4: 00000000001607e0 Call Trace: [<ffffffff81242213>] ? __check_object_size+0x1c3/0x220 [<ffffffffa17b8561>] ll_dir_ioctl+0x5d01/0x6ed0 [lustre] [<ffffffff811b515b>] ? unlock_page+0x2b/0x30 [<ffffffff811e5e2e>] ? do_read_fault.isra.64+0x12e/0x170 [<ffffffff811ea1cf>] ? handle_pte_fault+0x2ef/0xca0 [<ffffffff811ecf6d>] ? __handle_mm_fault+0x31d/0x720 [<ffffffff8125b3fd>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff81245a38>] ? vfs_write+0x178/0x230 [<ffffffff8125b751>] SyS_ioctl+0xa1/0xc0 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 | Lustre: lustre-OST0001-osc-ffff8802a308ca88: disconnect after 21s idle Lustre: 15197:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 512 < left 1754, rollback = 7 Lustre: 15197:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 15197:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 1/89/0 Lustre: 15197:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1745/1754/0, punch: 0/0/0, quota 3/3/0 Lustre: 15197:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 15197:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 15197:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 512 < left 1754, rollback = 7 Lustre: 15197:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 15197:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 1/89/0 Lustre: 15197:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1745/1754/0, punch: 0/0/0, quota 3/3/0 Lustre: 15197:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 15197:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 15197:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 512 < left 1754, rollback = 7 Lustre: 15197:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 1517 previous similar messages Lustre: 15197:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 15197:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 1517 previous similar messages Lustre: 15197:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 1/89/0 Lustre: 15197:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 1516 previous similar messages Lustre: 15197:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1745/1754/0, punch: 0/0/0, quota 3/3/0 Lustre: 15197:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 1517 previous similar messages Lustre: 15197:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 15197:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 1516 previous similar messages Lustre: 15197:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 15197:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 1517 previous similar messages Lustre: 18288:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 507 < left 26412, rollback = 7 Lustre: 18288:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 9734 previous similar messages Lustre: 18288:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 614/2456/0, destroy: 0/0/0 Lustre: 18288:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 9735 previous similar messages Lustre: 18288:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/1, xattr_set: 616/15/0 Lustre: 18288:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 9733 previous similar messages Lustre: 18288:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 3071/26412/0, punch: 0/0/0, quota 10/106/4 Lustre: 18288:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 9734 previous similar messages Lustre: 18288:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 614/10438/0, delete: 0/0/0 Lustre: 18288:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 9734 previous similar messages Lustre: 18288:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 18288:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 9734 previous similar messages Lustre: 18310:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 505 < left 1055, rollback = 2 Lustre: 18310:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 3/12/4, destroy: 0/0/0 Lustre: 18310:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 2460 previous similar messages Lustre: 18310:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 2/2/0, xattr_set: 13/1055/0 Lustre: 18310:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 2460 previous similar messages Lustre: 18310:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 5/136/0, punch: 0/0/0, quota 6/6/0 Lustre: 18310:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 2460 previous similar messages Lustre: 18310:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 12/215/3, delete: 1/1/0 Lustre: 18310:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 2460 previous similar messages Lustre: 18310:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 6/6/0, ref_del: 1/1/0 Lustre: 18310:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 2460 previous similar messages Lustre: 16980:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 505 < left 1767, rollback = 7 Lustre: 16980:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 2460 previous similar messages Lustre: ldlm_canceld: This server is not able to keep up with request traffic (cpu-bound). Lustre: 18819:0:(service.c:1612:ptlrpc_at_check_timed()) earlyQ=0 reqQ=0 recA=1, svcEst=31, delay=6756ms Lustre: 15241:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 7s req@ffff8800a5d55180 x1709915770739072/t0(0) o400->lustre-MDT0001-mdtlov_UUID@0@lo:0/0 lens 224/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'kworker/u32:0.0' Lustre: mdt_out: This server is not able to keep up with request traffic (cpu-bound). Lustre: 15242:0:(service.c:1612:ptlrpc_at_check_timed()) earlyQ=2 reqQ=0 recA=0, svcEst=1, delay=0ms Lustre: 15242:0:(service.c:1378:ptlrpc_at_send_early_reply()) @@@ Already past deadline (-1s), not sending early reply. Consider increasing at_early_margin (5)? req@ffff8802ebfb3ec0 x1709915770739392/t0(0) o400->lustre-MDT0001-mdtlov_UUID@0@lo:351/0 lens 224/0 e 0 to 0 dl 1630702956 ref 2 fl New:/0/ffffffff rc 0/-1 job:'kworker/u32:0.0' LustreError: 15242:0:(service.c:2290:ptlrpc_server_handle_request()) @@@ Dropping timed-out request from 12345-0@lo: deadline 6/1s ago req@ffff8800a5d53ec0 x1709915770739456/t0(0) o400->lustre-MDT0000-mdtlov_UUID@0@lo:351/0 lens 224/0 e 0 to 0 dl 1630702956 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'kworker/u32:0.0' Lustre: 15242:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (6/1s); client may timeout req@ffff8800a5d53ec0 x1709915770739456/t0(0) o400->lustre-MDT0000-mdtlov_UUID@0@lo:351/0 lens 224/0 e 0 to 0 dl 1630702956 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'kworker/u32:0.0' LustreError: 15692:0:(service.c:2290:ptlrpc_server_handle_request()) @@@ Dropping timed-out request from 12345-0@lo: deadline 6/7s ago req@ffff8802ebfb3ec0 x1709915770739392/t0(0) o400->lustre-MDT0001-mdtlov_UUID@0@lo:351/0 lens 224/0 e 0 to 0 dl 1630702956 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'kworker/u32:0.0' Lustre: 15692:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (6/7s); client may timeout req@ffff8802ebfb3ec0 x1709915770739392/t0(0) o400->lustre-MDT0001-mdtlov_UUID@0@lo:351/0 lens 224/0 e 0 to 0 dl 1630702956 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'kworker/u32:0.0' Lustre: mdt_out: This server is not able to keep up with request traffic (cpu-bound). Lustre: 15241:0:(service.c:1612:ptlrpc_at_check_timed()) earlyQ=1 reqQ=0 recA=0, svcEst=13, delay=0ms Lustre: 15241:0:(service.c:1378:ptlrpc_at_send_early_reply()) @@@ Already past deadline (-8s), not sending early reply. Consider increasing at_early_margin (5)? req@ffff8800a5d55180 x1709915770739072/t0(0) o400->lustre-MDT0001-mdtlov_UUID@0@lo:351/0 lens 224/0 e 0 to 0 dl 1630702956 ref 2 fl New:/0/ffffffff rc 0/-1 job:'kworker/u32:0.0' Lustre: 15241:0:(service.c:1378:ptlrpc_at_send_early_reply()) Skipped 1 previous similar message 15[19101]: segfault at 8 ip 00007f45e6fc17e8 sp 00007ffe5721e4e0 error 4 in ld-2.17.so[7f45e6fb6000+22000] Lustre: 18288:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 507 < left 26412, rollback = 7 Lustre: 18288:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 1757 previous similar messages Lustre: 18288:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 614/2456/0, destroy: 0/0/0 Lustre: 18288:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 1781 previous similar messages Lustre: 18288:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/1, xattr_set: 616/15/0 Lustre: 18288:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 1781 previous similar messages Lustre: 18288:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 3071/26412/0, punch: 0/0/0, quota 14/174/4 Lustre: 18288:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 1781 previous similar messages Lustre: 18288:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 614/10438/0, delete: 0/0/0 Lustre: 18288:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 1781 previous similar messages Lustre: 18288:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 18288:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 1781 previous similar messages LustreError: 15190:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0001_UUID lock: ffff8802b0c98400/0x4e6ec640de903b67 lrc: 3/0,0 mode: PW/PW res: [0x300000400:0xd:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000400000020 nid: 0@lo remote: 0x4e6ec640de903b52 expref: 4882 pid: 18349 timeout: 3685 lvb_type: 0 LustreError: 11-0: lustre-OST0001-osc-ffff8802a308ca88: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-OST0001-osc-ffff8802a308ca88: Connection to lustre-OST0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0001-osc-ffff8802a308ca88: This client was evicted by lustre-OST0001; in progress operations using this service will fail. LustreError: 19368:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802c54b9f80 x1709915771181696/t0(0) o104->lustre-OST0001@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 18634:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802e4ef2c00 x1709915771211392/t0(0) o104->lustre-OST0001@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' Lustre: 12001:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1630702950/real 1630702950] req@ffff8802df0f3240 x1709915770739072/t0(0) o400->lustre-MDT0000-osp-MDT0001@0@lo:24/4 lens 224/224 e 0 to 1 dl 1630702977 ref 1 fl Rpc:XNQr/0/ffffffff rc 0/-1 job:'kworker/u32:0.0' Lustre: lustre-MDT0000-osp-MDT0001: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 19328:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802e382de00 x1709915771224192/t0(0) o104->lustre-OST0001@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' Lustre: lustre-MDT0000: Received new MDS connection from 0@lo, keep former export from same NID Lustre: lustre-MDT0001-osp-MDT0000: Connection restored to 192.168.123.195@tcp (at 0@lo) LustreError: 19920:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0001-osc-ffff8802a308ca88: namespace resource [0x10:0x0:0x0].0x0 (ffff8800a13ad6c0) refcount nonzero (1) after lock cleanup; forcing cleanup. Lustre: lustre-OST0001-osc-ffff8802a308ca88: Connection restored to 192.168.123.195@tcp (at 0@lo) Lustre: Skipped 1 previous similar message Lustre: 15231:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 7s req@ffff8800843d9c48 x1709915771192960/t0(0) o37->f7d9d6fe-f8af-4096-948f-7758ffb89974@0@lo:0/0 lens 448/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'ls.0' Lustre: 15231:0:(service.c:2157:ptlrpc_server_handle_req_in()) Skipped 5 previous similar messages ------------[ cut here ]------------ WARNING: CPU: 0 PID: 18260 at /home/green/git/lustre-release/ldiskfs/namei.c:3331 ldiskfs_orphan_add+0x11e/0x2a0 [ldiskfs] Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) dm_flakey dm_mod loop zfs(PO) zunicode(PO) zzstd(O) zlua(O) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) jbd2 mbcache crc_t10dif crct10dif_generic sb_edac edac_core iosf_mbi crc32_pclmul ghash_clmulni_intel aesni_intel lrw gf128mul glue_helper ablk_helper cryptd virtio_balloon pcspkr i2c_piix4 virtio_console ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm crct10dif_pclmul drm ata_piix crct10dif_common drm_panel_orientation_quirks crc32c_intel serio_raw virtio_blk i2c_core libata floppy [last unloaded: libcfs] CPU: 0 PID: 18260 Comm: mdt00_006 Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011 Call Trace: [<ffffffff817d9418>] dump_stack+0x19/0x1b [<ffffffff8108d538>] __warn+0xd8/0x100 [<ffffffff8108d67d>] warn_slowpath_null+0x1d/0x20 [<ffffffffa0d964ae>] ldiskfs_orphan_add+0x11e/0x2a0 [ldiskfs] [<ffffffffa0d879eb>] ldiskfs_xattr_inode_orphan_add.constprop.17+0xbb/0x110 [ldiskfs] [<ffffffffa0d8a417>] ldiskfs_xattr_delete_inode+0x77/0x3c0 [ldiskfs] [<ffffffffa0daaaa0>] ? ldiskfs_evict_inode+0x190/0x670 [ldiskfs] [<ffffffffa0daaad9>] ldiskfs_evict_inode+0x1c9/0x670 [ldiskfs] [<ffffffff812649af>] evict+0xaf/0x180 [<ffffffff81264de5>] iput+0xf5/0x180 [<ffffffffa0cc6d08>] osd_object_delete+0x1f8/0x370 [osd_ldiskfs] [<ffffffffa04ea59d>] lu_object_free.isra.26+0xcd/0x1d0 [obdclass] [<ffffffffa04eecf1>] lu_object_put+0xa1/0x480 [obdclass] [<ffffffffa0eb3440>] mdt_object_put+0x30/0x110 [mdt] [<ffffffffa0ebae18>] mdt_reint_unlink+0x918/0x19b0 [mdt] [<ffffffffa051079e>] ? lu_ucred+0x1e/0x30 [obdclass] [<ffffffffa0eaaaf2>] ? ucred_set_audit_enabled.isra.13+0x22/0x60 [mdt] [<ffffffffa0ec2087>] mdt_reint_rec+0x87/0x240 [mdt] [<ffffffffa0e9755c>] mdt_reint_internal+0x76c/0xb40 [mdt] [<ffffffffa0ea4657>] mdt_reint+0x67/0x150 [mdt] [<ffffffffa07f226e>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<ffffffffa03d645e>] ? libcfs_nid2str_r+0xfe/0x130 [lnet] [<ffffffffa079d990>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<ffffffffa079f529>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e32be>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffffa079e950>] ? ptlrpc_wait_event+0x620/0x620 [ptlrpc] [<ffffffff810ba0f4>] kthread+0xe4/0xf0 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 ---[ end trace cfdfa726484435f1 ]--- Lustre: 18046:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 507 < left 1072, rollback = 2 Lustre: 18046:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 23 previous similar messages Lustre: 18046:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 2/8/2, destroy: 1/4/0 Lustre: 18046:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 11254 previous similar messages Lustre: 18046:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 2/2/0, xattr_set: 13/1072/0 Lustre: 18046:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 11254 previous similar messages Lustre: 18046:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1/10/0, punch: 0/0/0, quota 6/6/0 Lustre: 18046:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 11254 previous similar messages Lustre: 18046:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 12/194/2, delete: 3/6/1 Lustre: 18046:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 11254 previous similar messages Lustre: 18046:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 7/7/0, ref_del: 3/3/0 Lustre: 18046:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 11254 previous similar messages Lustre: 18138:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 501 < left 699, rollback = 2 Lustre: 18138:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 11 previous similar messages Lustre: 15208:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 511 < left 75002, rollback = 7 Lustre: 15208:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 11254 previous similar messages Lustre: 12000:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1630702950/real 1630702950] req@ffff8802ec24b880 x1709915770739392/t0(0) o400->lustre-MDT0002-osp-MDT0001@0@lo:24/4 lens 224/224 e 0 to 1 dl 1630702980 ref 1 fl Rpc:XNQr/0/ffffffff rc 0/-1 job:'kworker/u32:0.0' Lustre: 12000:0:(client.c:2285:ptlrpc_expire_one_request()) Skipped 1 previous similar message Lustre: lustre-MDT0002-osp-MDT0001: Connection to lustre-MDT0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 1 previous similar message Lustre: lustre-MDT0002: Received new MDS connection from 0@lo, keep former export from same NID Lustre: Skipped 1 previous similar message Lustre: lustre-MDT0002-osp-MDT0001: Connection restored to 192.168.123.195@tcp (at 0@lo) Lustre: 18116:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 8: before 509 < left 29885, rollback = 9 Lustre: 18116:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 7 previous similar messages Lustre: 15868:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 2/8/0, destroy: 0/0/0 Lustre: 15868:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 25632 previous similar messages Lustre: 15868:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 5/192/0 Lustre: 15868:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 25634 previous similar messages Lustre: 15868:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 812/897/0, punch: 0/0/0, quota 6/6/0 Lustre: 15868:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 25631 previous similar messages Lustre: 15868:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 2/34/0, delete: 0/0/0 Lustre: 15868:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 25631 previous similar messages Lustre: 15868:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 15868:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 25633 previous similar messages Lustre: 15868:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 481 < left 873, rollback = 7 Lustre: 15868:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 27616 previous similar messages Lustre: 20448:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 497 < left 671, rollback = 2 LustreError: 18138:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0002: '14' migration was interrupted, run 'lfs migrate -m 1 -c 1 -H crush 14' to finish migration. Lustre: 17831:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 8: before 509 < left 10199, rollback = 9 Lustre: 17831:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 7 previous similar messages Lustre: 20531:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 500 < left 730, rollback = 2 LustreError: 15190:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0002_UUID lock: ffff8802b14d1a80/0x4e6ec640de9a89bf lrc: 3/0,0 mode: PW/PW res: [0x340000401:0x6:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000000020020 nid: 0@lo remote: 0x4e6ec640de9a89b8 expref: 7344 pid: 15892 timeout: 3789 lvb_type: 0 LustreError: 19814:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630703080 with bad export cookie 5651672564310743290 LustreError: 11-0: lustre-OST0002-osc-ffff8802a37492a8: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-OST0002-osc-ffff8802a37492a8: Connection to lustre-OST0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: Skipped 1 previous similar message LustreError: 167-0: lustre-OST0002-osc-ffff8802a37492a8: This client was evicted by lustre-OST0002; in progress operations using this service will fail. LustreError: 19204:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff880096a66a80 x1709915775782272/t0(0) o104->lustre-OST0002@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 21315:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88022d6e70c0 x1709915775951104/t0(0) o104->lustre-OST0002@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 21315:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 24 previous similar messages Lustre: 12002:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.195@tcp:/lustre/fid: [0x280000404:0x59:0x0]/ may get corrupted (rc -108) Lustre: 12003:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.195@tcp:/lustre/fid: [0x240000404:0x55:0x0]/ may get corrupted (rc -108) Lustre: 12003:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.195@tcp:/lustre/fid: [0x240000404:0x97:0x0]/ may get corrupted (rc -108) Lustre: 12002:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.195@tcp:/lustre/fid: [0x240000404:0x31:0x0]// may get corrupted (rc -108) Lustre: 12002:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.195@tcp:/lustre/fid: [0x200000404:0x72:0x0]// may get corrupted (rc -108) Lustre: 21261:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 7s req@ffff8802e272b880 x1709915775806336/t0(0) o105->LOV_OSC_UUID@0@lo:0/0 lens 392/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'' Lustre: 15197:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 1438/5752/0, destroy: 1/4/0 Lustre: 15197:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 26053 previous similar messages Lustre: 15197:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 3/3/0, xattr_set: 1442/148/0 Lustre: 15197:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 26053 previous similar messages Lustre: 15197:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 7191/61844/26, punch: 0/0/0, quota 8/56/0 Lustre: 15197:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 26053 previous similar messages Lustre: 15197:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 1439/24462/0, delete: 2/5/0 Lustre: 15197:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 26053 previous similar messages Lustre: 15197:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 1/1/0, ref_del: 2/2/1 Lustre: 15197:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 26053 previous similar messages LustreError: 15207:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0001: '18' migration was interrupted, run 'lfs migrate -m 0 -c 1 -H crush 18' to finish migration. Lustre: lustre-OST0002-osc-ffff8802a37492a8: Connection restored to 192.168.123.195@tcp (at 0@lo) LustreError: 21430:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88009f5d57c0 x1709915776365568/t0(0) o106->lustre-OST0002@0@lo:15/16 lens 328/280 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 21430:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 11 previous similar messages Lustre: 20864:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 509 < left 81839, rollback = 7 Lustre: 20864:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 28765 previous similar messages Lustre: 20769:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 8: before 511 < left 10470, rollback = 9 Lustre: 20769:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 14 previous similar messages Lustre: dir [0x280000403:0x4d:0x0] stripe 1 readdir failed: -2, directory is partially accessed! LustreError: 22080:0:(dir.c:713:ll_dir_get_default_layout()) unknown magic: CD40CD0 Lustre: 11996:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff880248bd25c0 x1709915777935808/t0(0) o10->lustre-OST0000-osc-ffff8802a308ca88@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630703214 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dd.0' Lustre: 11996:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff880248bd25c0 x1709915780509696/t0(0) o10->lustre-OST0000-osc-ffff8802a308ca88@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630703263 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dd.0' LustreError: 15190:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0000_UUID lock: ffff8802c4524b40/0x4e6ec640dea1867e lrc: 3/0,0 mode: PW/PW res: [0x2c0000400:0x3:0x0].0x0 rrc: 5 type: EXT [0->18446744073709551615] (req 786432->18446744073709551615) gid 0 flags: 0x60000000030020 nid: 0@lo remote: 0x4e6ec640dea18669 expref: 5905 pid: 18731 timeout: 3903 lvb_type: 0 LustreError: 19888:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802e68a57c0 x1709915781136320/t0(0) o105->lustre-OST0000@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 11-0: lustre-OST0000-osc-ffff8802a308ca88: operation ost_setattr to node 0@lo failed: rc = -107 Lustre: lustre-OST0000-osc-ffff8802a308ca88: Connection to lustre-OST0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0000-osc-ffff8802a308ca88: This client was evicted by lustre-OST0000; in progress operations using this service will fail. LustreError: 15296:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0002-osp-MDT0000: fail to cancel 1 llog-records: rc = -116 LustreError: 15296:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0002-osp-MDT0000: fail to cancel 1 of 1 llog-records: rc = -116 LustreError: 15296:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0002-osp-MDT0000: fail to cancel 1 llog-records: rc = -2 LustreError: 15296:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) Skipped 1 previous similar message LustreError: 15296:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0002-osp-MDT0000: fail to cancel 1 of 1 llog-records: rc = -2 LustreError: 15296:0:(llog_cat.c:787:llog_cat_cancel_records()) Skipped 1 previous similar message Lustre: 15248:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 505 < left 817, rollback = 2 Lustre: 15248:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 64 previous similar messages LustreError: 25215:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0000-osc-ffff8802a308ca88: namespace resource [0xba0:0x0:0x0].0x0 (ffff8802e4a9f9c0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 25215:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 132 previous similar messages Lustre: lustre-OST0000-osc-ffff8802a308ca88: Connection restored to 192.168.123.195@tcp (at 0@lo) LustreError: 16980:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0001: '6' migration was interrupted, run 'lfs migrate -m 0 -c 1 -H crush 6' to finish migration. Lustre: 15201:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 1438/5752/0, destroy: 1/4/0 Lustre: 15201:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 47031 previous similar messages Lustre: 15201:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 3/3/0, xattr_set: 1442/148/0 Lustre: 15201:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 47030 previous similar messages Lustre: 15201:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 7191/61844/0, punch: 0/0/0, quota 10/90/0 Lustre: 15201:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 47031 previous similar messages Lustre: 15201:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 1439/24462/0, delete: 2/5/1 Lustre: 15201:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 47030 previous similar messages Lustre: 15201:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 1/1/0, ref_del: 2/2/1 Lustre: 15201:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 47030 previous similar messages Lustre: 20065:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 505 < left 42279, rollback = 7 Lustre: 20065:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 51582 previous similar messages LustreError: 18677:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0002: '19' migration was interrupted, run 'lfs migrate -m 0 -c 1 -H crush 19' to finish migration. LustreError: 15190:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 133s: evicting client at 0@lo ns: filter-lustre-OST0000_UUID lock: ffff8802c228da40/0x4e6ec640dea3a41e lrc: 3/0,0 mode: PW/PW res: [0x2c0000401:0x4:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000000020020 nid: 0@lo remote: 0x4e6ec640dea3a402 expref: 7440 pid: 21176 timeout: 3974 lvb_type: 0 LustreError: 15985:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802a37025c0 x1709915784208128/t0(0) o105->lustre-OST0000@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 15985:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 12 previous similar messages LustreError: 11-0: lustre-OST0000-osc-ffff8802a37492a8: operation ost_setattr to node 0@lo failed: rc = -107 LustreError: Skipped 1 previous similar message Lustre: lustre-OST0000-osc-ffff8802a37492a8: Connection to lustre-OST0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0000-osc-ffff8802a37492a8: This client was evicted by lustre-OST0000; in progress operations using this service will fail. Lustre: 12004:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.195@tcp:/lustre/fid: [0x240000404:0xc8:0x0]// may get corrupted (rc -108) Lustre: 12004:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.195@tcp:/lustre/fid: [0x200000404:0x17f:0x0]/ may get corrupted (rc -108) Lustre: 12004:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.195@tcp:/lustre/fid: [0x240000404:0x185:0x0]// may get corrupted (rc -108) LustreError: 26752:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0000-osc-ffff8802a37492a8: namespace resource [0x2c0000401:0x3ef:0x0].0x0 (ffff8800a7a374c0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 26752:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 192 previous similar messages Lustre: lustre-OST0000-osc-ffff8802a37492a8: Connection restored to 192.168.123.195@tcp (at 0@lo) Lustre: 12006:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff8802e6dcc500 x1709915785616256/t0(0) o10->lustre-OST0002-osc-ffff8802a308ca88@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630703377 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dir_create.sh.0' LustreError: 15190:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0000_UUID lock: ffff8802e78b7840/0x4e6ec640dea9cb8e lrc: 3/0,0 mode: PW/PW res: [0x2c0000401:0x3b2:0x0].0x0 rrc: 4 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000000020020 nid: 0@lo remote: 0x4e6ec640dea9cb87 expref: 2326 pid: 18699 timeout: 4035 lvb_type: 0 LustreError: 15174:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630703326 with bad export cookie 5651672564312319711 LustreError: 15174:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) Skipped 3 previous similar messages LustreError: 11-0: lustre-OST0000-osc-ffff8802a308ca88: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-OST0000-osc-ffff8802a308ca88: Connection to lustre-OST0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0000-osc-ffff8802a308ca88: This client was evicted by lustre-OST0000; in progress operations using this service will fail. LustreError: 23165:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802b7ebe440 x1709915786860480/t0(0) o106->lustre-OST0000@0@lo:15/16 lens 328/280 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 23165:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 117 previous similar messages LustreError: 18677:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0002: '17' migration was interrupted, run 'lfs migrate -m 1 -c 1 -H crush 17' to finish migration. LustreError: 27554:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0000-osc-ffff8802a308ca88: namespace resource [0xc10:0x0:0x0].0x0 (ffff8802b93a96c0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 27554:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 401 previous similar messages Lustre: lustre-OST0000-osc-ffff8802a308ca88: Connection restored to 192.168.123.195@tcp (at 0@lo) Lustre: 12006:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff8802e6dcc500 x1709915788052928/t0(0) o10->lustre-OST0002-osc-ffff8802a308ca88@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630703442 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dir_create.sh.0' Lustre: 23760:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 506 < left 671, rollback = 2 Lustre: 23760:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 66 previous similar messages LustreError: 28083:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff8802a308ca88: inode [0x240000403:0x191:0x0] mdc close failed: rc = -13 LustreError: 15227:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000404:0xc5:0x0]: rc = -2 LustreError: 15190:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 116s: evicting client at 0@lo ns: filter-lustre-OST0000_UUID lock: ffff8802e9f616c0/0x4e6ec640deb0ed53 lrc: 3/0,0 mode: PW/PW res: [0x2c0000401:0x3b1:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 458752->18446744073709551615) gid 0 flags: 0x60000000020020 nid: 0@lo remote: 0x4e6ec640deb0ed4c expref: 2974 pid: 21374 timeout: 4154 lvb_type: 0 LustreError: 19468:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802cce05e00 x1709915792485184/t0(0) o105->lustre-OST0000@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 19468:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 18 previous similar messages LustreError: 15177:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630703460 with bad export cookie 5651672564310742359 LustreError: 15177:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) Skipped 1 previous similar message LustreError: 11-0: lustre-OST0000-osc-ffff8802a37492a8: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-OST0000-osc-ffff8802a37492a8: Connection to lustre-OST0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0001-osc-ffff8802a37492a8: This client was evicted by lustre-OST0001; in progress operations using this service will fail. Lustre: 12006:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.195@tcp:/lustre/fid: [0x280000404:0x2b9:0x0]// may get corrupted (rc -108) Lustre: 11999:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.195@tcp:/lustre/fid: [0x280000404:0xe7:0x0]/ may get corrupted (rc -108) LustreError: 29317:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0001-osc-ffff8802a37492a8: namespace resource [0x300000401:0x3bf:0x0].0x0 (ffff8802de2ae0c0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 29317:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 11 previous similar messages Lustre: 18183:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 3/12/9, destroy: 0/0/0 Lustre: 18183:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 86627 previous similar messages Lustre: 18183:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 12/1116/0 Lustre: 18183:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 86626 previous similar messages Lustre: 18183:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 5/136/0, punch: 0/0/0, quota 8/56/0 Lustre: 18183:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 86623 previous similar messages Lustre: 18183:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 10/203/2, delete: 0/0/0 Lustre: 18183:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 86627 previous similar messages Lustre: 18183:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 4/4/0, ref_del: 0/0/0 Lustre: 18183:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 86626 previous similar messages LustreError: 29317:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0001-osc-ffff8802a37492a8: namespace resource [0x300000400:0x650:0x0].0x0 (ffff8802b4bdaac0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 29317:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 82 previous similar messages Lustre: lustre-OST0001-osc-ffff8802a37492a8: Connection restored to 192.168.123.195@tcp (at 0@lo) Lustre: 18288:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 510 < left 77711, rollback = 7 Lustre: 18288:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 77091 previous similar messages Lustre: ll_ost04_007: service thread pid 19190 was inactive for 40.095 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 19190, comm: ll_ost04_007 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] tgt_extent_lock+0xea/0x2a0 [ptlrpc] [<0>] ofd_getattr_hdl+0x365/0x720 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 19352, comm: ll_ost04_011 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] tgt_extent_lock+0xea/0x2a0 [ptlrpc] [<0>] ofd_getattr_hdl+0x365/0x720 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 27992, comm: ll_ost04_036 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] tgt_extent_lock+0xea/0x2a0 [ptlrpc] [<0>] ofd_getattr_hdl+0x365/0x720 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe LustreError: 15190:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 132s: evicting client at 0@lo ns: filter-lustre-OST0000_UUID lock: ffff8803222dd2c0/0x4e6ec640deb26ff0 lrc: 3/0,0 mode: PW/PW res: [0xc0a:0x0:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->524287) gid 0 flags: 0x60000400030020 nid: 0@lo remote: 0x4e6ec640deb26fb8 expref: 1582 pid: 19179 timeout: 4198 lvb_type: 0 LustreError: 15190:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 1 previous similar message LustreError: 15173:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630703521 with bad export cookie 5651672564312830403 LustreError: 167-0: lustre-OST0000-osc-ffff8802a308ca88: This client was evicted by lustre-OST0000; in progress operations using this service will fail. LustreError: Skipped 1 previous similar message Lustre: 11997:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.195@tcp:/lustre/fid: [0x280000403:0x22d:0x0]/ may get corrupted (rc -5) LustreError: 11960:0:(ofd_obd.c:501:ofd_destroy_export()) lustre-OST0002: cli ece76561-205c-4cfb-9cba-ea44907c54be/ffff8802a3922548 has 61440 pending on destroyed export LustreError: 11960:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_destroy_export: tot_granted 128431680 != fo_tot_granted 128493120 LustreError: 11960:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_destroy_export: tot_pending 151552 != fo_tot_pending 212992 LustreError: 11960:0:(ofd_obd.c:501:ofd_destroy_export()) lustre-OST0000: cli ece76561-205c-4cfb-9cba-ea44907c54be/ffff880272175d28 has 61440 pending on destroyed export LustreError: 11960:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_destroy_export: tot_granted 10389056 != fo_tot_granted 10450496 LustreError: 11960:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_destroy_export: tot_pending 0 != fo_tot_pending 61440 LustreError: 15912:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 10389056 != fo_tot_granted 10450496 LustreError: 15912:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 1 previous similar message LustreError: 15912:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 61440 LustreError: 15912:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 1 previous similar message LustreError: 15912:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 18826816 != fo_tot_granted 18888256 LustreError: 15912:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 61440 LustreError: 15910:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 145309504 != fo_tot_granted 145370944 LustreError: 15910:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 3 previous similar messages LustreError: 15910:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 152448 != fo_tot_pending 213888 LustreError: 15910:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 3 previous similar messages Lustre: 12008:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff8802a7f31940 x1709915795924032/t0(0) o10->lustre-OST0002-osc-ffff8802a308ca88@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630703548 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'cp.0' LustreError: 15913:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 18828864 != fo_tot_granted 18890304 LustreError: 15913:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 6 previous similar messages LustreError: 15913:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 61440 LustreError: 15913:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 6 previous similar messages LustreError: 15922:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 187497664 != fo_tot_granted 187559104 LustreError: 15922:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 13 previous similar messages LustreError: 15922:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 212992 != fo_tot_pending 274432 LustreError: 15922:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 13 previous similar messages Lustre: dir [0x280000403:0x21f:0x0] stripe 1 readdir failed: -2, directory is partially accessed! LustreError: 15917:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 117372352 != fo_tot_granted 117433792 LustreError: 15917:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 19 previous similar messages LustreError: 15917:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 876544 != fo_tot_pending 937984 LustreError: 15917:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 19 previous similar messages LustreError: 15190:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 101s: evicting client at 0@lo ns: filter-lustre-OST0002_UUID lock: ffff8802e004de00/0x4e6ec640deb1b9e2 lrc: 3/0,0 mode: PW/PW res: [0x340000400:0x400:0x0].0x0 rrc: 2 type: EXT [0->18446744073709551615] (req 0->1048575) gid 0 flags: 0x60000000020020 nid: 0@lo remote: 0x4e6ec640deb1b9b1 expref: 3658 pid: 18884 timeout: 4331 lvb_type: 0 LustreError: 15190:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 1 previous similar message LustreError: 11-0: lustre-OST0002-osc-ffff8802a37492a8: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 20 previous similar messages Lustre: lustre-OST0002-osc-ffff8802a37492a8: Connection to lustre-OST0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 3 previous similar messages LustreError: 167-0: lustre-OST0002-osc-ffff8802a37492a8: This client was evicted by lustre-OST0002; in progress operations using this service will fail. LustreError: Skipped 1 previous similar message LustreError: 11960:0:(ofd_obd.c:501:ofd_destroy_export()) lustre-OST0002: cli f7d9d6fe-f8af-4096-948f-7758ffb89974/ffff880097c15d28 has 151552 pending on destroyed export Lustre: 19290:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 6s req@ffff8802d798cb40 x1709915799926208/t0(0) o5->lustre-MDT0001-mdtlov_UUID@0@lo:0/0 lens 432/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'osp-pre-0-1.0' Lustre: 21337:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 502 < left 1044, rollback = 2 Lustre: 21337:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 178 previous similar messages Lustre: 18920:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 7s req@ffff880299d03838 x1709915800127360/t0(0) o104->LOV_OSC_UUID@0@lo:0/0 lens 328/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'' Lustre: 18920:0:(service.c:2157:ptlrpc_server_handle_req_in()) Skipped 1 previous similar message Lustre: ll_ost01_020: service thread pid 21319 was inactive for 84.176 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: ll_ost01_062: service thread pid 21428 was inactive for 84.174 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 11 previous similar messages Lustre: ll_ost01_029: service thread pid 21330 was inactive for 84.145 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 17 previous similar messages LustreError: 32216:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0002-osc-ffff8802a37492a8: namespace resource [0xc60:0x0:0x0].0x0 (ffff8800967c25c0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 32216:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 208 previous similar messages Lustre: lustre-OST0002-osc-ffff8802a37492a8: Connection restored to 192.168.123.195@tcp (at 0@lo) Lustre: Skipped 3 previous similar messages Lustre: 24040:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000403:0x2c8:0x0] with magic=0xbd60bd0 Lustre: 19163:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000403:0x268:0x0] with magic=0xbd60bd0 Lustre: 19163:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 1 previous similar message LustreError: 15184:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630703660 with bad export cookie 5651672564313455832 LustreError: 15184:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) Skipped 6 previous similar messages LustreError: 15912:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 8560832 != fo_tot_granted 8773824 LustreError: 15912:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 46 previous similar messages LustreError: 15912:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 212992 LustreError: 15912:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 46 previous similar messages Lustre: 12006:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.195@tcp:/lustre/fid: [0x200000403:0x230:0x0]// may get corrupted (rc -108) LustreError: 21290:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630703672 with bad export cookie 5651672564310733511 LustreError: 21290:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) Skipped 2 previous similar messages LustreError: 11-0: lustre-OST0003-osc-ffff8802a308ca88: operation ost_getattr to node 0@lo failed: rc = -107 Lustre: 30749:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1630703560/real 1630703560] req@ffff8800a14cd180 x1709915796906816/t0(0) o101->lustre-MDT0002-mdc-ffff8802a37492a8@0@lo:12/10 lens 576/8216 e 0 to 1 dl 1630703678 ref 2 fl Rpc:XQr/2/ffffffff rc 0/-1 job:'chmod.0' Lustre: lustre-MDT0002: Client f7d9d6fe-f8af-4096-948f-7758ffb89974 (at 0@lo) reconnecting Lustre: 27836:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 7s req@ffff8800a8419300 x1709915801718016/t0(0) o104->LOV_OSC_UUID@0@lo:0/0 lens 328/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'' LustreError: 32704:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0003-osc-ffff8802a308ca88: namespace resource [0x380000401:0x59f:0x0].0x0 (ffff8802def38040) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 32704:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 284 previous similar messages LustreError: 18534:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0001: failed to get lu_attr of [0x240000403:0x167:0x0]: rc = -2 Lustre: 23650:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000403:0x278:0x0] with magic=0xbd60bd0 Lustre: 23650:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 1 previous similar message Lustre: 18760:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 6s req@ffff8802e38025c0 x1709915805797568/t0(0) o37->f7d9d6fe-f8af-4096-948f-7758ffb89974@0@lo:0/0 lens 448/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'ls.0' Lustre: 18760:0:(service.c:2157:ptlrpc_server_handle_req_in()) Skipped 6 previous similar messages LustreError: 32286:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 248093120 != fo_tot_granted 248154560 LustreError: 32286:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 88 previous similar messages LustreError: 32286:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 61440 LustreError: 32286:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 89 previous similar messages Lustre: mdt_readpage: This server is not able to keep up with request traffic (cpu-bound). Lustre: 15225:0:(service.c:1612:ptlrpc_at_check_timed()) earlyQ=1 reqQ=0 recA=1, svcEst=8, delay=0ms Lustre: 15225:0:(service.c:1378:ptlrpc_at_send_early_reply()) @@@ Already past deadline (-6s), not sending early reply. Consider increasing at_early_margin (5)? req@ffff8802e38025c0 x1709915805797568/t0(0) o37->f7d9d6fe-f8af-4096-948f-7758ffb89974@0@lo:426/0 lens 448/0 e 0 to 0 dl 1630703786 ref 2 fl New:/0/ffffffff rc 0/-1 job:'ls.0' LustreError: 15225:0:(service.c:2290:ptlrpc_server_handle_request()) @@@ Dropping timed-out request from 12345-0@lo: deadline 8/6s ago req@ffff8802e38025c0 x1709915805797568/t0(0) o37->f7d9d6fe-f8af-4096-948f-7758ffb89974@0@lo:426/0 lens 448/0 e 0 to 0 dl 1630703786 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'ls.0' LustreError: 15225:0:(service.c:2290:ptlrpc_server_handle_request()) Skipped 1 previous similar message Lustre: 15225:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (8/6s); client may timeout req@ffff8802e38025c0 x1709915805797568/t0(0) o37->f7d9d6fe-f8af-4096-948f-7758ffb89974@0@lo:426/0 lens 448/0 e 0 to 0 dl 1630703786 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'ls.0' Lustre: 15225:0:(service.c:2328:ptlrpc_server_handle_request()) Skipped 1 previous similar message Lustre: 30843:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1630703778/real 1630703778] req@ffff8802e3804b40 x1709915805797568/t0(0) o37->lustre-MDT0002-mdc-ffff8802a37492a8@0@lo:23/10 lens 448/440 e 0 to 1 dl 1630703807 ref 2 fl Rpc:XQr/0/ffffffff rc 0/-1 job:'ls.0' Lustre: lustre-MDT0002-mdc-ffff8802a37492a8: Connection to lustre-MDT0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 3 previous similar messages Lustre: lustre-MDT0002: Client f7d9d6fe-f8af-4096-948f-7758ffb89974 (at 0@lo) reconnecting Lustre: lustre-MDT0002-mdc-ffff8802a37492a8: Connection restored to 192.168.123.195@tcp (at 0@lo) Lustre: Skipped 3 previous similar messages LustreError: 15190:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0003_UUID lock: ffff8802f744d680/0x4e6ec640dec38e36 lrc: 3/0,0 mode: PW/PW res: [0x380000400:0x8b1:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 2097152->18446744073709551615) gid 0 flags: 0x60000000020020 nid: 0@lo remote: 0x4e6ec640dec38db1 expref: 5291 pid: 29400 timeout: 4517 lvb_type: 0 LustreError: 15190:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 2 previous similar messages LustreError: 21189:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630703808 with bad export cookie 5651672564310740021 LustreError: 21189:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) Skipped 2 previous similar messages LustreError: 18678:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802eec1c500 x1709915806951360/t0(0) o104->lustre-OST0003@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 18678:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 224 previous similar messages LustreError: 11-0: lustre-OST0003-osc-ffff8802a37492a8: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 204 previous similar messages LustreError: 167-0: lustre-OST0003-osc-ffff8802a37492a8: This client was evicted by lustre-OST0003; in progress operations using this service will fail. LustreError: Skipped 2 previous similar messages LustreError: 1593:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0003-osc-ffff8802a37492a8: namespace resource [0x1000:0x0:0x0].0x0 (ffff8802c38a5e40) refcount nonzero (4) after lock cleanup; forcing cleanup. LustreError: 1593:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 883 previous similar messages Lustre: 15150:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 6s req@ffff8802e6478680 x1709915807375232/t0(0) o105->LOV_OSC_UUID@0@lo:0/0 lens 392/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'' Lustre: 27195:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (61/28s); client may timeout req@ffff8802d881bec0 x1709915804916224/t0(0) o1->f7d9d6fe-f8af-4096-948f-7758ffb89974@0@lo:462/0 lens 440/432 e 0 to 0 dl 1630703822 ref 1 fl Complete:/0/0 rc 0/0 job:'lfs.0' LustreError: 2041:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630703876 with bad export cookie 5651672564313870911 Lustre: 12000:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.195@tcp:/lustre/fid: [0x280000404:0x2f6:0x0]/ may get corrupted (rc -108) Lustre: ldlm_canceld: This server is not able to keep up with request traffic (cpu-bound). Lustre: 16959:0:(service.c:1612:ptlrpc_at_check_timed()) earlyQ=0 reqQ=0 recA=1, svcEst=4, delay=6594ms Lustre: 15151:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 6s req@ffff8802e5b8a5c0 x1709915809522624/t0(0) o104->LOV_OSC_UUID@0@lo:0/0 lens 328/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'' Lustre: 15151:0:(service.c:2157:ptlrpc_server_handle_req_in()) Skipped 3 previous similar messages LustreError: 23019:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0001: '14' migration was interrupted, run 'lfs migrate -m 0 -c 1 -H crush 14' to finish migration. Lustre: 24049:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000404:0x287:0x0] with magic=0xbd60bd0 Lustre: 24049:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 1 previous similar message Lustre: 23019:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000404:0x249:0x0] with magic=0xbd60bd0 Lustre: 23019:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 1 previous similar message Lustre: 15203:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000404:0x267:0x0] with magic=0xbd60bd0 Lustre: 15203:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 1 previous similar message Lustre: dir [0x200000404:0x231:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 2 previous similar messages Lustre: 17807:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000404:0x39a:0x0] with magic=0xbd60bd0 Lustre: 17807:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 3 previous similar messages Lustre: 15214:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 674/2696/0, destroy: 1/4/1 Lustre: 15214:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 157287 previous similar messages Lustre: 15214:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 2/2/0, xattr_set: 676/134/0 Lustre: 15214:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 157294 previous similar messages Lustre: 15214:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 3370/29075/18, punch: 0/0/0, quota 10/74/0 Lustre: 15214:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 157295 previous similar messages Lustre: 15214:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 676/11490/0, delete: 1/4/0 Lustre: 15214:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 157293 previous similar messages Lustre: 15214:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 1/1/0, ref_del: 2/2/1 Lustre: 15214:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 157448 previous similar messages LustreError: 15868:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0000: '15' migration was interrupted, run 'lfs migrate -m 1 -c 1 -H crush 15' to finish migration. 13[4661]: segfault at 8 ip 00007fd7dfad77e8 sp 00007fff9edfc990 error 4 in ld-2.17.so[7fd7dfacc000+22000] Lustre: 18760:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 510 < left 11018, rollback = 7 Lustre: 18760:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 159122 previous similar messages LustreError: 20444:0:(lustre_lmv.h:438:lmv_is_sane()) insane LMV: magic=0xcd40cd0 count=2 index=1 hash=crush:0x82000003 version=0 migrate offset=1 migrate hash=crush:3. LustreError: 26960:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000404:0x107:0x0]: rc = -2 Lustre: 20531:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0x2bb:0x0] with magic=0xbd60bd0 Lustre: 20531:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 15 previous similar messages LustreError: 15217:0:(mdt_xattr.c:429:mdt_dir_layout_update()) lustre-MDT0002: [0x280000404:0x451:0x0] migrate mdt index mismatch 2 != 0 LustreError: 15916:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 205163072 != fo_tot_granted 205376064 LustreError: 15916:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 181 previous similar messages LustreError: 15916:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 212992 LustreError: 15916:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 180 previous similar messages Lustre: 12003:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff8802af173880 x1709915815579264/t0(0) o10->lustre-OST0002-osc-ffff8802a37492a8@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630704072 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dd.0' Lustre: dir [0x240000404:0x3b0:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 2 previous similar messages Lustre: 12007:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff8802a6b12c00 x1709915816334720/t0(0) o10->lustre-OST0002-osc-ffff8802a308ca88@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630704127 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'cp.0' LustreError: 15201:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0001: '8' migration was interrupted, run 'lfs migrate -m 2 -c 1 -H crush 8' to finish migration. LustreError: 15201:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) Skipped 1 previous similar message Lustre: dir [0x240000403:0x3b5:0x0] stripe 1 readdir failed: -2, directory is partially accessed! LustreError: 6022:0:(lov_object.c:1299:lov_layout_change()) lustre-clilov-ffff8802a37492a8: cannot apply new layout on [0x240000403:0x417:0x0] : rc = -22 LustreError: 6022:0:(lcommon_cl.c:196:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000403:0x417:0x0]: rc = -22 LustreError: 7316:0:(lov_object.c:1299:lov_layout_change()) lustre-clilov-ffff8802a308ca88: cannot apply new layout on [0x240000403:0x417:0x0] : rc = -22 LustreError: 7316:0:(lcommon_cl.c:196:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000403:0x417:0x0]: rc = -22 LustreError: 7316:0:(llite_lib.c:2983:ll_prep_inode()) new_inode -fatal: rc -22 LustreError: 7702:0:(lov_object.c:1299:lov_layout_change()) lustre-clilov-ffff8802a37492a8: cannot apply new layout on [0x240000403:0x417:0x0] : rc = -22 LustreError: 7702:0:(lov_object.c:1299:lov_layout_change()) Skipped 3 previous similar messages LustreError: 7702:0:(lcommon_cl.c:196:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000403:0x417:0x0]: rc = -22 LustreError: 7702:0:(lcommon_cl.c:196:cl_file_inode_init()) Skipped 3 previous similar messages LustreError: 7702:0:(llite_lib.c:2983:ll_prep_inode()) new_inode -fatal: rc -22 LustreError: 7702:0:(llite_lib.c:2983:ll_prep_inode()) Skipped 3 previous similar messages LustreError: 7948:0:(lov_object.c:1299:lov_layout_change()) lustre-clilov-ffff8802a308ca88: cannot apply new layout on [0x240000403:0x417:0x0] : rc = -22 LustreError: 7948:0:(lov_object.c:1299:lov_layout_change()) Skipped 1 previous similar message LustreError: 7948:0:(lcommon_cl.c:196:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000403:0x417:0x0]: rc = -22 LustreError: 7948:0:(lcommon_cl.c:196:cl_file_inode_init()) Skipped 1 previous similar message LustreError: 7948:0:(llite_lib.c:2983:ll_prep_inode()) new_inode -fatal: rc -22 LustreError: 7948:0:(llite_lib.c:2983:ll_prep_inode()) Skipped 1 previous similar message LustreError: 7316:0:(lov_object.c:1299:lov_layout_change()) lustre-clilov-ffff8802a308ca88: cannot apply new layout on [0x240000403:0x417:0x0] : rc = -22 LustreError: 7316:0:(lov_object.c:1299:lov_layout_change()) Skipped 6 previous similar messages LustreError: 7316:0:(lcommon_cl.c:196:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000403:0x417:0x0]: rc = -22 LustreError: 7316:0:(lcommon_cl.c:196:cl_file_inode_init()) Skipped 6 previous similar messages LustreError: 7316:0:(llite_lib.c:2983:ll_prep_inode()) new_inode -fatal: rc -22 LustreError: 7316:0:(llite_lib.c:2983:ll_prep_inode()) Skipped 6 previous similar messages LustreError: 7893:0:(lov_object.c:1299:lov_layout_change()) lustre-clilov-ffff8802a37492a8: cannot apply new layout on [0x240000403:0x417:0x0] : rc = -22 LustreError: 7893:0:(lov_object.c:1299:lov_layout_change()) Skipped 1 previous similar message LustreError: 7893:0:(lcommon_cl.c:196:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000403:0x417:0x0]: rc = -22 LustreError: 7893:0:(lcommon_cl.c:196:cl_file_inode_init()) Skipped 1 previous similar message LustreError: 7893:0:(llite_lib.c:2983:ll_prep_inode()) new_inode -fatal: rc -22 LustreError: 7893:0:(llite_lib.c:2983:ll_prep_inode()) Skipped 1 previous similar message LustreError: 26960:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0001: failed to get lu_attr of [0x240000404:0x3b0:0x0]: rc = -2 LustreError: 7933:0:(llite_lib.c:2983:ll_prep_inode()) new_inode -fatal: rc -22 LustreError: 7933:0:(llite_lib.c:2983:ll_prep_inode()) Skipped 1 previous similar message LustreError: 8562:0:(lov_object.c:1299:lov_layout_change()) lustre-clilov-ffff8802a37492a8: cannot apply new layout on [0x240000403:0x417:0x0] : rc = -22 LustreError: 8562:0:(lov_object.c:1299:lov_layout_change()) Skipped 4 previous similar messages LustreError: 8562:0:(lcommon_cl.c:196:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000403:0x417:0x0]: rc = -22 LustreError: 8562:0:(lcommon_cl.c:196:cl_file_inode_init()) Skipped 4 previous similar messages LustreError: 8773:0:(llite_lib.c:2983:ll_prep_inode()) new_inode -fatal: rc -22 LustreError: 8773:0:(llite_lib.c:2983:ll_prep_inode()) Skipped 4 previous similar messages LustreError: 15190:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 175s: evicting client at 0@lo ns: filter-lustre-OST0001_UUID lock: ffff8800a9da5680/0x4e6ec640dece0966 lrc: 3/0,0 mode: PW/PW res: [0x1549:0x0:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->458751) gid 0 flags: 0x60000400000020 nid: 0@lo remote: 0x4e6ec640dece0927 expref: 2727 pid: 21421 timeout: 4851 lvb_type: 0 LustreError: 15190:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 2 previous similar messages LustreError: 11-0: lustre-OST0001-osc-ffff8802a37492a8: operation ost_setattr to node 0@lo failed: rc = -107 LustreError: Skipped 3 previous similar messages Lustre: lustre-OST0001-osc-ffff8802a37492a8: Connection to lustre-OST0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 3 previous similar messages LustreError: 167-0: lustre-OST0001-osc-ffff8802a37492a8: This client was evicted by lustre-OST0001; in progress operations using this service will fail. LustreError: Skipped 2 previous similar messages LustreError: 697:0:(ofd_obd.c:501:ofd_destroy_export()) lustre-OST0001: cli f7d9d6fe-f8af-4096-948f-7758ffb89974/ffff8802bb288958 has 2277376 pending on destroyed export LustreError: 15236:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0001: failed to get lu_attr of [0x240000404:0x3b0:0x0]: rc = -2 Lustre: 12008:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.195@tcp:/lustre/fid: [0x240000404:0x459:0x0]// may get corrupted (rc -108) Lustre: 12008:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.195@tcp:/lustre/fid: [0x200000404:0x388:0x0]/ may get corrupted (rc -108) Lustre: 12011:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.195@tcp:/lustre/fid: [0x200000403:0x389:0x0]// may get corrupted (rc -108) Lustre: 12010:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.195@tcp:/lustre/fid: [0x280000403:0x484:0x0]// may get corrupted (rc -108) LustreError: 8926:0:(lov_object.c:1299:lov_layout_change()) lustre-clilov-ffff8802a308ca88: cannot apply new layout on [0x240000403:0x417:0x0] : rc = -22 LustreError: 8926:0:(lov_object.c:1299:lov_layout_change()) Skipped 18 previous similar messages LustreError: 8926:0:(lcommon_cl.c:196:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000403:0x417:0x0]: rc = -22 LustreError: 8926:0:(lcommon_cl.c:196:cl_file_inode_init()) Skipped 18 previous similar messages LustreError: 9105:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0001-osc-ffff8802a37492a8: namespace resource [0x18a0:0x0:0x0].0x0 (ffff8802ce63fc40) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 9105:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 388 previous similar messages Lustre: lustre-OST0001-osc-ffff8802a37492a8: Connection restored to 192.168.123.195@tcp (at 0@lo) Lustre: Skipped 3 previous similar messages LustreError: 9558:0:(llite_lib.c:2983:ll_prep_inode()) new_inode -fatal: rc -22 LustreError: 9558:0:(llite_lib.c:2983:ll_prep_inode()) Skipped 19 previous similar messages Lustre: 1419:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 504 < left 610, rollback = 2 Lustre: 1419:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 225 previous similar messages Lustre: 11996:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff8800a6a0e440 x1709915821868032/t0(0) o10->lustre-OST0001-osc-ffff8802a308ca88@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630704311 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'truncate.0' LustreError: 10785:0:(lov_object.c:1299:lov_layout_change()) lustre-clilov-ffff8802a37492a8: cannot apply new layout on [0x240000403:0x417:0x0] : rc = -22 LustreError: 10785:0:(lov_object.c:1299:lov_layout_change()) Skipped 20 previous similar messages LustreError: 10785:0:(lcommon_cl.c:196:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000403:0x417:0x0]: rc = -22 LustreError: 10785:0:(lcommon_cl.c:196:cl_file_inode_init()) Skipped 20 previous similar messages LustreError: 10589:0:(llite_lib.c:2983:ll_prep_inode()) new_inode -fatal: rc -22 LustreError: 10589:0:(llite_lib.c:2983:ll_prep_inode()) Skipped 18 previous similar messages LustreError: 16980:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0002: '19' migration was interrupted, run 'lfs migrate -m 0 -c 1 -H crush 19' to finish migration. LustreError: 20176:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630704338 with bad export cookie 5651672564314653497 LustreError: 20176:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) Skipped 2 previous similar messages LustreError: 21401:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802e1393880 x1709915826958784/t0(0) o104->lustre-OST0003@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 21401:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 180 previous similar messages ptlrpc_watchdog_fire: 34 callbacks suppressed Lustre: ll_ost00_037: service thread pid 21426 was inactive for 40.096 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 3 previous similar messages Lustre: ll_ost00_006: service thread pid 19015 was inactive for 40.018 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 20901, comm: ll_ost00_015 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] ofd_destroy_by_fid+0x1d1/0x520 [ofd] [<0>] ofd_destroy_hdl+0x27f/0xa20 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: Skipped 3 previous similar messages Pid: 19015, comm: ll_ost00_006 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] ofd_destroy_by_fid+0x1d1/0x520 [ofd] [<0>] ofd_destroy_hdl+0x27f/0xa20 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: ll_ost00_035: service thread pid 21418 was inactive for 40.811 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 21418, comm: ll_ost00_035 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] ofd_destroy_by_fid+0x1d1/0x520 [ofd] [<0>] ofd_destroy_hdl+0x27f/0xa20 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: ll_ost00_008: service thread pid 19302 was inactive for 41.232 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 1 previous similar message Lustre: ll_ost01_023: service thread pid 21322 was inactive for 40.079 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 5 previous similar messages LustreError: 13547:0:(lov_object.c:1299:lov_layout_change()) lustre-clilov-ffff8802a37492a8: cannot apply new layout on [0x240000403:0x417:0x0] : rc = -22 LustreError: 13547:0:(lov_object.c:1299:lov_layout_change()) Skipped 53 previous similar messages LustreError: 13547:0:(lcommon_cl.c:196:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000403:0x417:0x0]: rc = -22 LustreError: 13547:0:(lcommon_cl.c:196:cl_file_inode_init()) Skipped 53 previous similar messages Lustre: ll_ost02_009: service thread pid 19304 was inactive for 46.556 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 10 previous similar messages Lustre: ll_ost02_007: service thread pid 18917 was inactive for 40.188 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 1 previous similar message LustreError: 13740:0:(llite_lib.c:2983:ll_prep_inode()) new_inode -fatal: rc -22 LustreError: 13740:0:(llite_lib.c:2983:ll_prep_inode()) Skipped 61 previous similar messages Lustre: 12007:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff8802e8b8a5c0 x1709915828932224/t0(0) o10->lustre-OST0001-osc-ffff8802a308ca88@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630704420 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'truncate.0' Lustre: 12007:0:(client.c:1485:after_reply()) Skipped 1 previous similar message 9[14575]: segfault at 8 ip 00007f03bff847e8 sp 00007ffc83ec0d80 error 4 in ld-2.17.so[7f03bff79000+22000] Lustre: mdt04_013: service thread pid 29457 was inactive for 46.149 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 7 previous similar messages Lustre: mdt01_008: service thread pid 21991 was inactive for 58.134 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 1 previous similar message LustreError: 15190:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 106s: evicting client at 0@lo ns: filter-lustre-OST0000_UUID lock: ffff8800a26507c0/0x4e6ec640dee2d197 lrc: 3/0,0 mode: PW/PW res: [0x2c0000400:0x189a:0x0].0x0 rrc: 4 type: EXT [0->18446744073709551615] (req 0->1048575) gid 0 flags: 0x60000400020020 nid: 0@lo remote: 0x4e6ec640dee2cff3 expref: 2318 pid: 23159 timeout: 5217 lvb_type: 0 LustreError: 15190:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 2 previous similar messages LustreError: 21838:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630704508 with bad export cookie 5651672564315497354 LustreError: 21838:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) Skipped 1 previous similar message LustreError: 167-0: lustre-OST0000-osc-ffff8802a308ca88: This client was evicted by lustre-OST0000; in progress operations using this service will fail. LustreError: Skipped 2 previous similar messages Lustre: 12008:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.195@tcp:/lustre/fid: [0x200000403:0x4e7:0x0]// may get corrupted (rc -108) LustreError: 18248:0:(vvp_io.c:1793:vvp_io_init()) lustre: refresh file layout [0x200000403:0x4e7:0x0] error -108. Lustre: 12008:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.195@tcp:/lustre/fid: [0x200000404:0x504:0x0]// may get corrupted (rc -108) Lustre: 12009:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.195@tcp:/lustre/fid: [0x200000403:0x527:0x0]// may get corrupted (rc -108) LustreError: 31761:0:(ofd_obd.c:501:ofd_destroy_export()) lustre-OST0000: cli ece76561-205c-4cfb-9cba-ea44907c54be/ffff88009b83c138 has 61440 pending on destroyed export LustreError: 18986:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0000-osc-ffff8802a308ca88: namespace resource [0x2c0000400:0x2090:0x0].0x0 (ffff8802b6afb9c0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 18986:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 184 previous similar messages Lustre: 12007:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff8802e8b8a5c0 x1709915835094848/t0(0) o10->lustre-OST0001-osc-ffff8802a308ca88@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630704596 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'truncate.0' Lustre: 12007:0:(client.c:1485:after_reply()) Skipped 481 previous similar messages LustreError: 11-0: lustre-MDT0001-mdc-ffff8802a308ca88: operation mds_close to node 0@lo failed: rc = -107 LustreError: Skipped 1 previous similar message LustreError: 9257:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff8802a308ca88: inode [0x240000403:0x430:0x0] mdc close failed: rc = -5 LustreError: 13990:0:(mdc_request.c:1436:mdc_read_page()) lustre-MDT0001-mdc-ffff8802a308ca88: [0x240000404:0x4c7:0x0] lock enqueue fails: rc = -108 LustreError: 18495:0:(mdc_request.c:1436:mdc_read_page()) lustre-MDT0001-mdc-ffff8802a308ca88: [0x240000404:0x4c7:0x0] lock enqueue fails: rc = -108 LustreError: 18495:0:(mdc_request.c:1436:mdc_read_page()) Skipped 24 previous similar messages LustreError: 19666:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff8802a308ca88: inode [0x240000403:0x3b5:0x0] mdc close failed: rc = -108 LustreError: 19666:0:(file.c:234:ll_close_inode_openhandle()) Skipped 7 previous similar messages LustreError: 18961:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0001: failed to get lu_attr of [0x240000403:0x44c:0x0]: rc = -2 Lustre: lustre-OST0000-osc-ffff8802a308ca88: disconnect after 43s idle Lustre: Skipped 3 previous similar messages LustreError: 15922:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 277431872 != fo_tot_granted 277644864 LustreError: 15922:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 585 previous similar messages LustreError: 15922:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 212992 LustreError: 15922:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 585 previous similar messages Lustre: 12002:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.195@tcp:/lustre/fid: [0x200000403:0x4e5:0x0]/ may get corrupted (rc -108) Lustre: 12003:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.195@tcp:/lustre/fid: [0x240000403:0x567:0x0]/ may get corrupted (rc -108) Lustre: 12003:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.195@tcp:/lustre/fid: [0x240000404:0x633:0x0]// may get corrupted (rc -108) LustreError: 20616:0:(vvp_io.c:1793:vvp_io_init()) lustre: refresh file layout [0x240000403:0x417:0x0] error -22. LustreError: 18760:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0002: failed to get lu_attr of [0x280000404:0x4de:0x0]: rc = -2 Lustre: 23714:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 3/12/12, destroy: 0/0/0 Lustre: 23714:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 258856 previous similar messages Lustre: 23714:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 2/2/0, xattr_set: 7/521/0 Lustre: 23714:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 258846 previous similar messages Lustre: 23714:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 5/136/0, punch: 0/0/0, quota 8/56/1 Lustre: 23714:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 258848 previous similar messages Lustre: 23714:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 12/215/4, delete: 1/1/0 Lustre: 23714:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 258848 previous similar messages Lustre: 23714:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 6/6/0, ref_del: 1/1/0 Lustre: 23714:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 258695 previous similar messages LustreError: 22361:0:(lov_object.c:1299:lov_layout_change()) lustre-clilov-ffff8802a37492a8: cannot apply new layout on [0x240000403:0x417:0x0] : rc = -22 LustreError: 22361:0:(lov_object.c:1299:lov_layout_change()) Skipped 106 previous similar messages LustreError: 22361:0:(lcommon_cl.c:196:cl_file_inode_init()) lustre: failed to initialize cl_object [0x240000403:0x417:0x0]: rc = -22 LustreError: 22361:0:(lcommon_cl.c:196:cl_file_inode_init()) Skipped 105 previous similar messages Lustre: 15219:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 510 < left 850, rollback = 7 Lustre: 15219:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 262363 previous similar messages LustreError: 23133:0:(ldlm_lockd.c:1425:ldlm_handle_enqueue0()) ### lock on destroyed export ffff8802774bca88 ns: filter-lustre-OST0000_UUID lock: ffff8802f6c196c0/0x4e6ec640def19631 lrc: 3/0,0 mode: --/PW res: [0x210c:0x0:0x0].0x0 rrc: 4 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x50000000020000 nid: 0@lo remote: 0x4e6ec640def19582 expref: 1615 pid: 23133 timeout: 0 lvb_type: 0 Lustre: lustre-OST0000-osc-ffff8802a308ca88: Connection to lustre-OST0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 5 previous similar messages ------------[ cut here ]------------ WARNING: CPU: 12 PID: 21114 at /home/green/git/lustre-release/ldiskfs/ext4_jbd2.c:266 __ldiskfs_handle_dirty_metadata+0x1d2/0x230 [ldiskfs] Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) dm_flakey dm_mod loop zfs(PO) zunicode(PO) zzstd(O) zlua(O) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) jbd2 mbcache crc_t10dif crct10dif_generic sb_edac edac_core iosf_mbi crc32_pclmul ghash_clmulni_intel aesni_intel lrw gf128mul glue_helper ablk_helper cryptd virtio_balloon pcspkr i2c_piix4 virtio_console ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm crct10dif_pclmul drm ata_piix crct10dif_common drm_panel_orientation_quirks crc32c_intel serio_raw virtio_blk i2c_core libata floppy [last unloaded: libcfs] CPU: 12 PID: 21114 Comm: mdt06_011 Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011 Call Trace: [<ffffffff817d9418>] dump_stack+0x19/0x1b [<ffffffff8108d538>] __warn+0xd8/0x100 [<ffffffff8108d67d>] warn_slowpath_null+0x1d/0x20 [<ffffffffa0dc5262>] __ldiskfs_handle_dirty_metadata+0x1d2/0x230 [ldiskfs] [<ffffffffa0da533d>] ldiskfs_getblk+0x13d/0x210 [ldiskfs] [<ffffffffa0da5437>] ldiskfs_bread+0x27/0xe0 [ldiskfs] [<ffffffffa0cf0e97>] osd_write+0x4e7/0xcb0 [osd_ldiskfs] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffffa04f06f3>] dt_record_write+0x33/0x120 [obdclass] [<ffffffffa04af6b2>] llog_osd_write_rec+0x1052/0x19a0 [obdclass] [<ffffffffa049c760>] llog_write_rec+0x290/0x590 [obdclass] [<ffffffffa04a24ee>] llog_cat_add_rec+0x23e/0x950 [obdclass] [<ffffffffa04995ff>] llog_add+0x17f/0x1f0 [obdclass] [<ffffffffa04caa39>] ? lprocfs_counter_add+0xf9/0x160 [obdclass] [<ffffffffa083360d>] sub_updates_write+0xb6b/0xe32 [ptlrpc] [<ffffffffa0803892>] top_trans_stop+0x7c2/0xf60 [ptlrpc] [<ffffffffa0f79f3c>] lod_trans_stop+0x25c/0x340 [lod] [<ffffffffa0e4e408>] mdd_trans_stop+0x28/0x16e [mdd] [<ffffffffa0e2fb2f>] mdd_migrate_object+0xf1f/0x1ba0 [mdd] [<ffffffffa04ef375>] ? lu_object_find_at+0x285/0xb10 [obdclass] [<ffffffffa0e30b86>] mdd_migrate+0x3d6/0x8e0 [mdd] [<ffffffffa0f020d8>] mdo_migrate+0x4c/0x4e [mdt] [<ffffffffa0ec1e53>] mdt_reint_migrate+0x1023/0x11d0 [mdt] [<ffffffffa0ec2087>] mdt_reint_rec+0x87/0x240 [mdt] [<ffffffffa0e9755c>] mdt_reint_internal+0x76c/0xb40 [mdt] [<ffffffffa0ea4657>] mdt_reint+0x67/0x150 [mdt] [<ffffffffa07f226e>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<ffffffffa03d645e>] ? libcfs_nid2str_r+0xfe/0x130 [lnet] [<ffffffffa079d990>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<ffffffffa079f529>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e32be>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffffa079e950>] ? ptlrpc_wait_event+0x620/0x620 [ptlrpc] [<ffffffff810ba0f4>] kthread+0xe4/0xf0 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 ---[ end trace cfdfa726484435f2 ]--- LDISKFS-fs: ldiskfs_getblk:888: aborting transaction: error 28 in __ldiskfs_handle_dirty_metadata LDISKFS-fs error (device dm-1): ldiskfs_getblk:888: inode #168: block 20810: comm mdt06_011: journal_dirty_metadata failed: handle type 0 started at line 1972, credits 512/0, errcode -28 Aborting journal on device dm-1-8. LDISKFS-fs (dm-1): Remounting filesystem read-only LustreError: 21114:0:(osd_io.c:2103:osd_ldiskfs_write_record()) lustre-MDT0001: error reading offset 401408 (block 98, size 40, offs 401392), credits 0/0: rc = -28 LustreError: 21114:0:(update_trans.c:985:top_trans_stop()) lustre-MDT0001-osd: write updates failed: rc = -28 LustreError: 21114:0:(osd_io.c:2110:osd_ldiskfs_write_record()) journal_get_write_access() returned error -30 LustreError: 21114:0:(tgt_lastrcvd.c:1253:tgt_add_reply_data()) lustre-MDT0001: can't update reply_data file: rc = -30 LustreError: 21114:0:(osd_handler.c:2075:osd_trans_stop()) lustre-MDT0001: failed in transaction hook: rc = -30 LDISKFS-fs error (device dm-1) in osd_trans_stop:2082: error 28 LustreError: 15473:0:(osd_handler.c:1779:osd_trans_commit_cb()) transaction @0xffff8802e2d93538 commit error: 2 LustreError: 21114:0:(osd_handler.c:2085:osd_trans_stop()) lustre-MDT0001: failed to stop transaction: rc = -28 LustreError: 21114:0:(update_trans.c:1011:top_trans_stop()) lustre-MDT0001-osd: stop trans failed: rc = -30 LustreError: 23395:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -5 LustreError: 15488:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0001-osd: fail to cancel 1 llog-records: rc = -30 LustreError: 15488:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0001-osd: fail to cancel 1 of 1 llog-records: rc = -30 | Externally reported by onyx-68 boilpot email |
racer test 1: racer on clients: centos-25.localnet DURATION=2700 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa17b9e52>] ll_migrate+0x9b2/0xec0 [lustre] PGD 800000028f7ab067 PUD 274721067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) dm_flakey dm_mod loop zfs(PO) zunicode(PO) zzstd(O) zlua(O) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) jbd2 mbcache crc_t10dif crct10dif_generic sb_edac edac_core iosf_mbi crc32_pclmul ghash_clmulni_intel aesni_intel lrw gf128mul glue_helper ablk_helper cryptd virtio_console pcspkr i2c_piix4 virtio_balloon ip_tables rpcsec_gss_krb5 drm_kms_helper ttm ata_generic pata_acpi drm crct10dif_pclmul crct10dif_common drm_panel_orientation_quirks ata_piix crc32c_intel serio_raw virtio_blk i2c_core libata floppy [last unloaded: libcfs] CPU: 3 PID: 25389 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011 task: ffff88029d7bb760 ti: ffff8802bc38c000 task.ti: ffff8802bc38c000 RIP: 0010:[<ffffffffa17b9e52>] [<ffffffffa17b9e52>] ll_migrate+0x9b2/0xec0 [lustre] RSP: 0018:ffff8802bc38fbc8 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff8802a130f9d8 RCX: 0000000000000000 RDX: 000000010001709b RSI: ffff88026f6180b8 RDI: ffff88026f618098 RBP: ffff8802bc38fc48 R08: ffff8802f4812d78 R09: ffff8802bc38fa60 R10: 0000000000000008 R11: ffff8802ca403a76 R12: 0000000000000000 R13: ffff88009f189dd8 R14: ffff8803242f5d98 R15: 0000000000000030 FS: 00007f7412ff6740(0000) GS:ffff880331ac0000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 CR2: 0000000000000008 CR3: 000000026e708000 CR4: 00000000001607e0 Call Trace: [<ffffffff81242213>] ? __check_object_size+0x1c3/0x220 [<ffffffffa17a7561>] ll_dir_ioctl+0x5d01/0x6ed0 [lustre] [<ffffffff811b515b>] ? unlock_page+0x2b/0x30 [<ffffffff811e5e2e>] ? do_read_fault.isra.64+0x12e/0x170 [<ffffffff811ea1cf>] ? handle_pte_fault+0x2ef/0xca0 [<ffffffff8103de0f>] ? save_stack_trace+0x1f/0x30 [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff811ecf6d>] ? __handle_mm_fault+0x31d/0x720 [<ffffffff8125b3fd>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff817e8655>] ? __do_page_fault+0x1c5/0x500 [<ffffffff8125b751>] SyS_ioctl+0xa1/0xc0 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 | Lustre: 16387:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 512 < left 1608, rollback = 7 Lustre: 16387:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 16387:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 1/89/0 Lustre: 16387:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1599/1608/0, punch: 0/0/0, quota 3/3/0 Lustre: 16387:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 16387:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 19127:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 8: before 511 < left 27198, rollback = 9 LustreError: 16410:0:(out_handler.c:910:out_tx_end()) lustre-MDT0002-osd: undo for /home/green/git/lustre-release/lustre/ptlrpc/../../lustre/target/out_handler.c:445: rc = -524 Lustre: 16378:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 512 < left 1290, rollback = 7 Lustre: 16378:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 1599 previous similar messages Lustre: 16378:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 2/8/0, destroy: 0/0/0 Lustre: 16378:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 1601 previous similar messages Lustre: 16378:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 5/192/0 Lustre: 16378:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 1601 previous similar messages Lustre: 16378:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1205/1290/0, punch: 0/0/0, quota 6/6/0 Lustre: 16378:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 1601 previous similar messages Lustre: 16378:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 2/34/0, delete: 0/0/0 Lustre: 16378:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 1601 previous similar messages Lustre: 16378:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 16378:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 1601 previous similar messages Lustre: 16378:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 512 < left 1290, rollback = 7 Lustre: 16378:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 686 previous similar messages Lustre: 16378:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 2/8/0, destroy: 0/0/0 Lustre: 16378:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 686 previous similar messages Lustre: 16378:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 5/192/0 Lustre: 16378:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 686 previous similar messages Lustre: 16378:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1205/1290/0, punch: 0/0/0, quota 6/6/0 Lustre: 16378:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 686 previous similar messages Lustre: 16378:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 2/34/0, delete: 0/0/0 Lustre: 16378:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 1329 previous similar messages Lustre: 16378:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 16378:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 1476 previous similar messages LustreError: 16862:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0000-osp-MDT0002: fail to cancel 1 llog-records: rc = -116 LustreError: 16862:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0000-osp-MDT0002: fail to cancel 1 of 1 llog-records: rc = -116 LustreError: 16661:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0002-osp-MDT0001: fail to cancel 1 llog-records: rc = -116 LustreError: 16661:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) Skipped 1 previous similar message LustreError: 16661:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0002-osp-MDT0001: fail to cancel 1 of 1 llog-records: rc = -116 LustreError: 16661:0:(llog_cat.c:787:llog_cat_cancel_records()) Skipped 1 previous similar message LustreError: 16661:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0002-osp-MDT0001: fail to cancel 1 llog-records: rc = -116 LustreError: 16661:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0002-osp-MDT0001: fail to cancel 1 of 1 llog-records: rc = -116 Lustre: 18773:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 512 < left 962, rollback = 7 Lustre: 18773:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 5299 previous similar messages Lustre: 18773:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 18773:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 5299 previous similar messages Lustre: 18773:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 1/89/0 Lustre: 18773:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 5299 previous similar messages Lustre: 18773:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 953/962/0, punch: 0/0/0, quota 3/3/0 Lustre: 18773:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 5299 previous similar messages Lustre: 18773:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 18773:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 4656 previous similar messages Lustre: 18773:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 18773:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 4509 previous similar messages Lustre: 19069:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 508 < left 671, rollback = 2 Lustre: 19069:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 1 previous similar message Lustre: 20694:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 498 < left 671, rollback = 2 Lustre: 20694:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 7 previous similar messages LustreError: 19405:0:(lustre_lmv.h:438:lmv_is_sane()) insane LMV: magic=0xcd40cd0 count=2 index=1 hash=crush:0x82000003 version=0 migrate offset=1 migrate hash=crush:3. Lustre: 16374:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 512 < left 1149, rollback = 7 Lustre: 16374:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 952 previous similar messages Lustre: 16374:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 2/8/0, destroy: 0/0/0 Lustre: 16374:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 968 previous similar messages Lustre: 16374:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 5/192/0 Lustre: 16374:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 968 previous similar messages Lustre: 16374:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1064/1149/0, punch: 0/0/0, quota 6/6/0 Lustre: 16374:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 968 previous similar messages Lustre: 16374:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 2/34/0, delete: 0/0/0 Lustre: 16374:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 968 previous similar messages Lustre: 16374:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 16374:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 968 previous similar messages Lustre: 19127:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 498 < left 45375, rollback = 7 Lustre: 19127:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 5275 previous similar messages Lustre: 19127:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 1055/4220/0, destroy: 0/0/0 Lustre: 19127:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 5275 previous similar messages Lustre: 19127:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 1058/219/14 Lustre: 19127:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 5275 previous similar messages Lustre: 19127:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 5276/45375/0, punch: 0/0/0, quota 6/6/0 Lustre: 19127:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 5275 previous similar messages Lustre: 19127:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 1055/17935/0, delete: 0/0/0 Lustre: 19127:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 5275 previous similar messages Lustre: 19127:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 19127:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 5275 previous similar messages Lustre: 19484:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 8: before 511 < left 17951, rollback = 9 Lustre: 19484:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 7 previous similar messages LustreError: 16854:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0002: '4' migration was interrupted, run 'lfs migrate -m 1 -c 1 -H crush 4' to finish migration. Lustre: 17789:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 459 < left 51393, rollback = 7 Lustre: 17789:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 8604 previous similar messages Lustre: 17789:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 1196/4784/0, destroy: 0/0/0 Lustre: 17789:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 8605 previous similar messages Lustre: 17789:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/1, xattr_set: 1198/15/0 Lustre: 17789:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 8605 previous similar messages Lustre: 17789:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 5981/51438/45, punch: 0/0/0, quota 10/106/7 Lustre: 17789:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 8605 previous similar messages Lustre: 17789:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 1196/20332/0, delete: 0/0/0 Lustre: 17789:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 8606 previous similar messages Lustre: 17789:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 17789:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 8606 previous similar messages Lustre: 17789:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 8: before 511 < left 20348, rollback = 9 LustreError: 22289:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0000: '10' migration was interrupted, run 'lfs migrate -m 1 -c 1 -H crush 10' to finish migration. LustreError: 16401:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0002: failed to get lu_attr of [0x280000403:0x21:0x0]: rc = -2 Lustre: 19484:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 508 < left 688, rollback = 2 Lustre: 19484:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 9 previous similar messages Lustre: 19405:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 506 < left 40946, rollback = 7 Lustre: 19405:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 1183 previous similar messages Lustre: 19405:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 952/3808/0, destroy: 0/0/0 Lustre: 19405:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 1199 previous similar messages Lustre: 19405:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/1, xattr_set: 954/15/0 Lustre: 19405:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 1199 previous similar messages Lustre: 19405:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 4761/40946/0, punch: 0/0/0, quota 10/106/5 Lustre: 19405:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 1199 previous similar messages Lustre: 19405:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 952/16184/0, delete: 0/0/0 Lustre: 19405:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 1198 previous similar messages Lustre: 19405:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 19405:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 1198 previous similar messages Lustre: 19128:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 500 < left 535, rollback = 2 Lustre: 19128:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 5 previous similar messages LustreError: 16362:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0000_UUID lock: ffff88024522c780/0xb62322e87040efeb lrc: 3/0,0 mode: PW/PW res: [0x2c0000401:0x3:0x0].0x0 rrc: 4 type: EXT [0->18446744073709551615] (req 0->262143) gid 0 flags: 0x60000400020020 nid: 0@lo remote: 0xb62322e87040efe4 expref: 1718 pid: 17061 timeout: 11519 lvb_type: 0 LustreError: 20807:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802e4f49940 x1709914362135040/t0(0) o105->lustre-OST0000@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 11-0: lustre-OST0000-osc-ffff88006dee1bf8: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: 16356:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630701653 with bad export cookie 13124372120798736516 Lustre: lustre-OST0000-osc-ffff88006dee1bf8: Connection to lustre-OST0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0000-osc-ffff88006dee1bf8: This client was evicted by lustre-OST0000; in progress operations using this service will fail. LustreError: 21109:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802ebda7700 x1709914362218944/t0(0) o104->lustre-OST0000@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' Lustre: 13178:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.125@tcp:/lustre/fid: [0x240000403:0x54:0x0]// may get corrupted (rc -108) LustreError: 23902:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0000-osc-ffff88006dee1bf8: namespace resource [0x2c0000401:0x25:0x0].0x0 (ffff88023db316c0) refcount nonzero (1) after lock cleanup; forcing cleanup. Lustre: lustre-OST0000-osc-ffff88006dee1bf8: Connection restored to 192.168.123.125@tcp (at 0@lo) LustreError: 16469:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0001-osp-MDT0000: fail to cancel 1 llog-records: rc = -116 LustreError: 16469:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) Skipped 1 previous similar message LustreError: 16469:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0001-osp-MDT0000: fail to cancel 1 of 1 llog-records: rc = -116 LustreError: 16469:0:(llog_cat.c:787:llog_cat_cancel_records()) Skipped 1 previous similar message ------------[ cut here ]------------ WARNING: CPU: 8 PID: 16401 at /home/green/git/lustre-release/ldiskfs/namei.c:3331 ldiskfs_orphan_add+0x11e/0x2a0 [ldiskfs] Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) dm_flakey dm_mod loop zfs(PO) zunicode(PO) zzstd(O) zlua(O) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) jbd2 mbcache crc_t10dif crct10dif_generic sb_edac edac_core iosf_mbi crc32_pclmul ghash_clmulni_intel aesni_intel lrw gf128mul glue_helper ablk_helper cryptd virtio_console pcspkr i2c_piix4 virtio_balloon ip_tables rpcsec_gss_krb5 drm_kms_helper ttm ata_generic pata_acpi drm crct10dif_pclmul crct10dif_common drm_panel_orientation_quirks ata_piix crc32c_intel serio_raw virtio_blk i2c_core libata floppy [last unloaded: libcfs] CPU: 8 PID: 16401 Comm: mdt_rdpg04_000 Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011 Call Trace: [<ffffffff817d9418>] dump_stack+0x19/0x1b [<ffffffff8108d538>] __warn+0xd8/0x100 [<ffffffff8108d67d>] warn_slowpath_null+0x1d/0x20 [<ffffffffa0d854ae>] ldiskfs_orphan_add+0x11e/0x2a0 [ldiskfs] [<ffffffffa0d769eb>] ldiskfs_xattr_inode_orphan_add.constprop.17+0xbb/0x110 [ldiskfs] [<ffffffffa0d79417>] ldiskfs_xattr_delete_inode+0x77/0x3c0 [ldiskfs] [<ffffffffa0d99aa0>] ? ldiskfs_evict_inode+0x190/0x670 [ldiskfs] [<ffffffffa0d99ad9>] ldiskfs_evict_inode+0x1c9/0x670 [ldiskfs] [<ffffffff812649af>] evict+0xaf/0x180 [<ffffffff81264de5>] iput+0xf5/0x180 [<ffffffffa0cb5d08>] osd_object_delete+0x1f8/0x370 [osd_ldiskfs] [<ffffffffa04d259d>] lu_object_free.isra.26+0xcd/0x1d0 [obdclass] [<ffffffffa04d6cf1>] lu_object_put+0xa1/0x480 [obdclass] [<ffffffffa0ebbaf4>] mdt_mfd_close+0x1f4/0x870 [mdt] [<ffffffffa0ec1e14>] mdt_close_internal+0xb4/0x240 [mdt] [<ffffffffa0ec2224>] mdt_close+0x284/0x850 [mdt] [<ffffffffa07d92d4>] ? tgt_request_preprocess.isra.27+0xf4/0x850 [ptlrpc] [<ffffffffa07da26e>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<ffffffffa035a45e>] ? libcfs_nid2str_r+0xfe/0x130 [lnet] [<ffffffffa0785990>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<ffffffffa0787529>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffffa0786950>] ? ptlrpc_wait_event+0x620/0x620 [ptlrpc] [<ffffffff810ba0f4>] kthread+0xe4/0xf0 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 ---[ end trace e5b03de8bbfe556d ]--- Lustre: 17789:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 512 < left 1160, rollback = 7 Lustre: 17789:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 29410 previous similar messages Lustre: 17789:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 17789:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 29476 previous similar messages Lustre: 17789:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 1/89/0 Lustre: 17789:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 29476 previous similar messages Lustre: 17789:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1151/1160/0, punch: 0/0/0, quota 3/3/0 Lustre: 17789:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 29476 previous similar messages Lustre: 17789:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 17789:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 29476 previous similar messages Lustre: 17789:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 17789:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 29476 previous similar messages Lustre: 13171:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff88029bb4c500 x1709914363842304/t0(0) o10->lustre-OST0001-osc-ffff88006dee1bf8@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630701760 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'truncate.0' Lustre: 20995:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 503 < left 849, rollback = 2 Lustre: 20995:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 65 previous similar messages LustreError: 16362:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0001_UUID lock: ffff88023ad64f00/0xb62322e87044c557 lrc: 3/0,0 mode: PW/PW res: [0x300000400:0x1ae:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000400020020 nid: 0@lo remote: 0xb62322e87044c550 expref: 1887 pid: 17058 timeout: 11571 lvb_type: 0 LustreError: 22003:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8800926c6a80 x1709914364096704/t0(0) o106->lustre-OST0001@0@lo:15/16 lens 328/280 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' Lustre: lustre-OST0001-osc-ffff88006dee4a88: Connection to lustre-OST0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0001-osc-ffff88006dee4a88: This client was evicted by lustre-OST0001; in progress operations using this service will fail. LustreError: 17059:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802b74b3880 x1709914364209088/t0(0) o106->lustre-OST0001@0@lo:15/16 lens 328/280 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 17059:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 60 previous similar messages Lustre: 13169:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.125@tcp:/lustre/fid: [0x200000404:0xa2:0x0]// may get corrupted (rc -108) LustreError: 25378:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0001-osc-ffff88006dee4a88: namespace resource [0x110:0x0:0x0].0x0 (ffff880223810cc0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 25378:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 3 previous similar messages LustreError: 25378:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0001-osc-ffff88006dee4a88: namespace resource [0x300000401:0xf:0x0].0x0 (ffff880271e8e0c0) refcount nonzero (1) after lock cleanup; forcing cleanup. Lustre: lustre-OST0001: haven't heard from client 88448a6d-0db1-45c7-85b6-dee4316839af (at 0@lo) in 47 seconds. I think it's dead, and I am evicting it. exp ffff88007cff5d28, cur 1630701754 expire 1630701724 last 1630701707 LustreError: 20472:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0001: '2' migration was interrupted, run 'lfs migrate -m 2 -c 1 -H crush 2' to finish migration. Lustre: 13171:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff88029bb4c500 x1709914366541824/t0(0) o10->lustre-OST0001-osc-ffff88006dee1bf8@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630701826 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'truncate.0' Lustre: 16371:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 507 < left 1044, rollback = 2 Lustre: 16371:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 49 previous similar messages Lustre: 13181:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff880276898cc0 x1709914367834752/t0(0) o10->lustre-OST0000-osc-ffff88006dee1bf8@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630701861 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dir_create.sh.0' LustreError: 16362:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0000_UUID lock: ffff8802ab28f840/0xb62322e87049e544 lrc: 3/0,0 mode: PW/PW res: [0x107:0x0:0x0].0x0 rrc: 5 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000000020020 nid: 0@lo remote: 0xb62322e87049e4e9 expref: 1151 pid: 23565 timeout: 11678 lvb_type: 0 Lustre: 13178:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff88009d2457c0 x1709914368227648/t0(0) o10->lustre-OST0002-osc-ffff88006dee1bf8@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630701871 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dir_create.sh.0' LustreError: 11-0: lustre-OST0000-osc-ffff88006dee1bf8: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-OST0000-osc-ffff88006dee1bf8: Connection to lustre-OST0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 21953:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88009efb70c0 x1709914368233280/t0(0) o104->lustre-OST0000@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 167-0: lustre-OST0000-osc-ffff88006dee1bf8: This client was evicted by lustre-OST0000; in progress operations using this service will fail. LustreError: 21953:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 14 previous similar messages LustreError: 25378:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0001-osc-ffff88006dee4a88: namespace resource [0x300000401:0x26b:0x0].0x0 (ffff880267cff740) refcount nonzero (3) after lock cleanup; forcing cleanup. LustreError: 25378:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 126 previous similar messages LustreError: 21390:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630701813 with bad export cookie 13124372120798736537 Lustre: lustre-OST0001-osc-ffff88006dee4a88: Connection restored to 192.168.123.125@tcp (at 0@lo) LustreError: 11-0: lustre-OST0001-osc-ffff88006dee4a88: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: 13171:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.125@tcp:/lustre/fid: [0x280000404:0x13b:0x0]/ may get corrupted (rc -108) Lustre: 13176:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.125@tcp:/lustre/fid: [0x240000403:0xe6:0x0]// may get corrupted (rc -108) Lustre: 13175:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.125@tcp:/lustre/fid: [0x200000403:0x125:0x0]/ may get corrupted (rc -108) Lustre: 13172:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.125@tcp:/lustre/fid: [0x280000404:0x127:0x0]// may get corrupted (rc -108) Lustre: 13173:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.125@tcp:/lustre/fid: [0x280000404:0x11e:0x0]/ may get corrupted (rc -108) LustreError: 27343:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0000-osc-ffff88006dee1bf8: namespace resource [0x2c0000400:0x1a0:0x0].0x0 (ffff880091a70040) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 27343:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 60 previous similar messages Lustre: lustre-OST0000-osc-ffff88006dee1bf8: Connection restored to 192.168.123.125@tcp (at 0@lo) Lustre: Skipped 1 previous similar message Lustre: lustre-OST0003-osc-ffff88006dee1bf8: Connection restored to 192.168.123.125@tcp (at 0@lo) Lustre: 17086:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 6s req@ffff88007a502c00 x1709914368371136/t0(0) o13->lustre-MDT0002-mdtlov_UUID@0@lo:0/0 lens 224/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'osp-pre-3-2.0' Lustre: 20391:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 510 < left 45375, rollback = 7 Lustre: 20391:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 23951 previous similar messages Lustre: 20391:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 1055/4220/0, destroy: 1/4/0 Lustre: 20391:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 24023 previous similar messages Lustre: 20391:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 1055/0/0 Lustre: 20391:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 24023 previous similar messages Lustre: 20391:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 5276/45375/0, punch: 0/0/0, quota 10/74/0 Lustre: 20391:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 24023 previous similar messages Lustre: 20391:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 1055/17935/0, delete: 2/5/1 Lustre: 20391:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 24023 previous similar messages Lustre: 20391:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 1/1/1 Lustre: 20391:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 24023 previous similar messages Lustre: 13181:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff880077aee440 x1709914368680448/t0(0) o10->lustre-OST0000-osc-ffff88006dee1bf8@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630701880 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'cp.0' Lustre: dir [0x240000404:0x91:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: 23429:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000404:0x168:0x0] with magic=0xbd60bd0 Lustre: 16371:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000404:0x145:0x0] with magic=0xbd60bd0 Lustre: 16371:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 1 previous similar message Lustre: 16379:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 504 < left 805, rollback = 2 Lustre: 16379:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 29 previous similar messages LustreError: 16362:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 128s: evicting client at 0@lo ns: filter-lustre-OST0002_UUID lock: ffff880278b32200/0xb62322e8704ec2c6 lrc: 3/0,0 mode: PW/PW res: [0x124:0x0:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000400030020 nid: 0@lo remote: 0xb62322e8704ec2b8 expref: 1945 pid: 25490 timeout: 11803 lvb_type: 0 LustreError: 16362:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 1 previous similar message LustreError: 20807:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802821a1300 x1709914373278976/t0(0) o105->lustre-OST0002@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 20807:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 15 previous similar messages LustreError: 23717:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630701948 with bad export cookie 13124372120798736768 LustreError: 23717:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) Skipped 1 previous similar message LustreError: 11-0: lustre-OST0002-osc-ffff88006dee4a88: operation ost_setattr to node 0@lo failed: rc = -107 LustreError: Skipped 2 previous similar messages Lustre: lustre-OST0002-osc-ffff88006dee4a88: Connection to lustre-OST0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 2 previous similar messages LustreError: 167-0: lustre-OST0002-osc-ffff88006dee4a88: This client was evicted by lustre-OST0002; in progress operations using this service will fail. LustreError: Skipped 2 previous similar messages LustreError: 17138:0:(ofd_obd.c:501:ofd_destroy_export()) lustre-OST0002: cli 88448a6d-0db1-45c7-85b6-dee4316839af/ffff880248796678 has 61440 pending on destroyed export LustreError: 17138:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_destroy_export: tot_granted 165441984 != fo_tot_granted 165503424 LustreError: 17138:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_destroy_export: tot_pending 896 != fo_tot_pending 62336 LustreError: 17091:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 165439296 != fo_tot_granted 165500736 LustreError: 17091:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 61440 Lustre: 13179:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.125@tcp:/lustre/fid: [0x200000404:0x1d4:0x0]/ may get corrupted (rc -108) Lustre: 13179:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.125@tcp:/lustre/fid: [0x240000404:0x1c0:0x0]// may get corrupted (rc -108) LustreError: 30416:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0003-osc-ffff88006dee4a88: namespace resource [0x130:0x0:0x0].0x0 (ffff88025cb00540) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 30416:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 1329 previous similar messages LustreError: 17080:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 165439296 != fo_tot_granted 165500736 LustreError: 17080:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 1 previous similar message LustreError: 17080:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 61440 LustreError: 17080:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 1 previous similar message Lustre: lustre-OST0003-osc-ffff88006dee4a88: Connection restored to 192.168.123.125@tcp (at 0@lo) LustreError: 30065:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 165439296 != fo_tot_granted 165500736 LustreError: 30065:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 61440 LustreError: 17081:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 173879488 != fo_tot_granted 173940928 LustreError: 17081:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 2 previous similar messages LustreError: 17081:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 61440 != fo_tot_pending 122880 LustreError: 17081:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 2 previous similar messages LustreError: 17089:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 182255808 != fo_tot_granted 182317248 LustreError: 17089:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 2 previous similar messages LustreError: 17089:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 61440 LustreError: 17089:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 2 previous similar messages LustreError: 16862:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0000-osp-MDT0002: fail to cancel 1 llog-records: rc = -116 LustreError: 16862:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) Skipped 1 previous similar message LustreError: 16862:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0000-osp-MDT0002: fail to cancel 1 of 1 llog-records: rc = -116 LustreError: 16862:0:(llog_cat.c:787:llog_cat_cancel_records()) Skipped 1 previous similar message LustreError: 17089:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 203013568 != fo_tot_granted 203075008 LustreError: 17089:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 5 previous similar messages LustreError: 17089:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 61440 LustreError: 17089:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 5 previous similar messages Lustre: 21721:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000404:0x154:0x0] with magic=0xbd60bd0 Lustre: 21721:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 1 previous similar message Lustre: 19128:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000404:0x18f:0x0] with magic=0xbd60bd0 Lustre: 19128:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 1 previous similar message Lustre: 19569:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000404:0x19a:0x0] with magic=0xbd60bd0 Lustre: 19569:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 3 previous similar messages Lustre: dir [0x240000403:0xc1:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: 13167:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff8802b4cb1940 x1709914376072768/t0(0) o10->lustre-OST0000-osc-ffff88006dee4a88@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630702080 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'cp.0' LustreError: 17080:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 244768320 != fo_tot_granted 244829760 LustreError: 17080:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 6 previous similar messages LustreError: 17080:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 61440 LustreError: 17080:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 6 previous similar messages Lustre: dir [0x200000404:0x1b0:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: 13177:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff880273f47700 x1709914377297984/t0(0) o10->lustre-OST0000-osc-ffff88006dee4a88@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630702102 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dir_create.sh.0' Lustre: 20632:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000403:0x13d:0x0] with magic=0xbd60bd0 Lustre: 20632:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 3 previous similar messages LustreError: 16362:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0001_UUID lock: ffff8802a50fd2c0/0xb62322e8705636d9 lrc: 3/0,0 mode: PW/PW res: [0x300000401:0x152:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000000030020 nid: 0@lo remote: 0xb62322e8705636d2 expref: 1896 pid: 20126 timeout: 11929 lvb_type: 0 LustreError: 16362:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 1 previous similar message LustreError: 21109:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff880068def0c0 x1709914378191680/t0(0) o106->lustre-OST0001@0@lo:15/16 lens 328/280 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 21109:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 10 previous similar messages LustreError: 11-0: lustre-OST0001-osc-ffff88006dee1bf8: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-OST0001-osc-ffff88006dee1bf8: Connection to lustre-OST0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 1 previous similar message LustreError: Skipped 2 previous similar messages LustreError: 167-0: lustre-OST0001-osc-ffff88006dee1bf8: This client was evicted by lustre-OST0001; in progress operations using this service will fail. LustreError: Skipped 1 previous similar message Lustre: lustre-OST0001-osc-ffff88006dee1bf8: Connection restored to 192.168.123.125@tcp (at 0@lo) Lustre: Skipped 1 previous similar message LustreError: 20581:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000403:0x220:0x0]: rc = -2 Lustre: 19428:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000403:0x1cd:0x0] with magic=0xbd60bd0 Lustre: 19428:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 1 previous similar message LustreError: 17789:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0002: '11' migration was interrupted, run 'lfs migrate -m 0 -c 1 -H crush 11' to finish migration. LustreError: 17789:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) Skipped 1 previous similar message Lustre: 19511:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 509 < left 53803, rollback = 7 Lustre: 19511:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 132118 previous similar messages Lustre: 19511:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 1251/5004/0, destroy: 1/4/0 Lustre: 19511:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 132256 previous similar messages Lustre: 19511:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 3/3/1, xattr_set: 1255/148/0 Lustre: 19511:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 132254 previous similar messages Lustre: 19511:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 6256/53803/0, punch: 0/0/0, quota 10/74/0 Lustre: 19511:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 132257 previous similar messages Lustre: 19511:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 1252/21283/0, delete: 2/5/1 Lustre: 19511:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 132258 previous similar messages Lustre: 19511:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 1/1/0, ref_del: 2/2/1 Lustre: 19511:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 132256 previous similar messages LustreError: 17085:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 274210880 != fo_tot_granted 274272320 LustreError: 17085:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 24 previous similar messages LustreError: 17085:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 90112 != fo_tot_pending 151552 LustreError: 17085:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 24 previous similar messages Lustre: 19127:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000403:0x15a:0x0] with magic=0xbd60bd0 Lustre: 19127:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 5 previous similar messages Lustre: 13177:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff8802a58df0c0 x1709914380863232/t0(0) o10->lustre-OST0000-osc-ffff88006dee1bf8@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630702180 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'cp.0' LustreError: 1446:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0002: failed to get lu_attr of [0x280000404:0x267:0x0]: rc = -2 Lustre: 19088:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 8: before 509 < left 19685, rollback = 9 Lustre: 19088:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 170 previous similar messages LustreError: 16661:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0000-osp-MDT0001: fail to cancel 1 llog-records: rc = -2 LustreError: 16661:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) Skipped 1 previous similar message LustreError: 16661:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0000-osp-MDT0001: fail to cancel 1 of 1 llog-records: rc = -2 LustreError: 16661:0:(llog_cat.c:787:llog_cat_cancel_records()) Skipped 1 previous similar message LustreError: 16362:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0002_UUID lock: ffff88025ad6cf00/0xb62322e8705bf7a3 lrc: 3/0,0 mode: PW/PW res: [0x340000401:0xae4:0x0].0x0 rrc: 4 type: EXT [0->18446744073709551615] (req 2097152->18446744073709551615) gid 0 flags: 0x60000400020020 nid: 0@lo remote: 0xb62322e8705bf78e expref: 1508 pid: 20943 timeout: 12032 lvb_type: 0 LustreError: 20807:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff880273e557c0 x1709914382858112/t0(0) o105->lustre-OST0002@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 25939:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630702166 with bad export cookie 13124372120800156774 LustreError: 25939:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) Skipped 2 previous similar messages LustreError: 20807:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 78 previous similar messages LustreError: 11-0: lustre-OST0002-osc-ffff88006dee4a88: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-OST0002-osc-ffff88006dee4a88: Connection to lustre-OST0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0002-osc-ffff88006dee4a88: This client was evicted by lustre-OST0002; in progress operations using this service will fail. LustreError: 3814:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0002-osc-ffff88006dee4a88: namespace resource [0x4d0:0x0:0x0].0x0 (ffff8800b029a340) refcount nonzero (2) after lock cleanup; forcing cleanup. LustreError: 3814:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 333 previous similar messages Lustre: lustre-OST0002-osc-ffff88006dee4a88: Connection restored to 192.168.123.125@tcp (at 0@lo) Lustre: 13177:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff8800913657c0 x1709914384540096/t0(0) o10->lustre-OST0000-osc-ffff88006dee4a88@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630702224 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'cp.0' Lustre: 13177:0:(client.c:1485:after_reply()) Skipped 2 previous similar messages LustreError: 16390:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0000: '15' migration was interrupted, run 'lfs migrate -m 1 -c 1 -H crush 15' to finish migration. LustreError: 17086:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 241919936 != fo_tot_granted 241981376 LustreError: 17086:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 54 previous similar messages LustreError: 17086:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 1792 != fo_tot_pending 63232 LustreError: 17086:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 54 previous similar messages LustreError: 16362:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0000_UUID lock: ffff8802c4b5e580/0xb62322e8705ee5ad lrc: 3/0,0 mode: PW/PW res: [0x2c0000400:0x516:0x0].0x0 rrc: 4 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000000030020 nid: 0@lo remote: 0xb62322e8705edc4c expref: 3911 pid: 23532 timeout: 12091 lvb_type: 0 LustreError: 11-0: lustre-OST0000-osc-ffff88006dee1bf8: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-OST0000-osc-ffff88006dee1bf8: Connection to lustre-OST0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: Skipped 1 previous similar message LustreError: 167-0: lustre-OST0000-osc-ffff88006dee1bf8: This client was evicted by lustre-OST0000; in progress operations using this service will fail. LustreError: 28438:0:(ofd_obd.c:501:ofd_destroy_export()) lustre-OST0000: cli 44f0f5e5-394f-4061-bca1-57e5378d4d07/ffff880275eeefc8 has 90112 pending on destroyed export Lustre: 13175:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.125@tcp:/lustre/fid: [0x280000403:0x32e:0x0]// may get corrupted (rc -108) Lustre: 13176:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.125@tcp:/lustre/fid: [0x240000403:0x2eb:0x0]// may get corrupted (rc -108) Lustre: 13176:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.125@tcp:/lustre/fid: [0x240000403:0x2ea:0x0]/ may get corrupted (rc -108) LustreError: 4776:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0000-osc-ffff88006dee1bf8: namespace resource [0x4a0:0x0:0x0].0x0 (ffff8802d5724540) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 4776:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 73 previous similar messages Lustre: lustre-OST0000: haven't heard from client 44f0f5e5-394f-4061-bca1-57e5378d4d07 (at 0@lo) in 47 seconds. I think it's dead, and I am evicting it. exp ffff880274268008, cur 1630702273 expire 1630702243 last 1630702226 Lustre: 20391:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 6s req@ffff880327b94ad8 x1709914387969280/t0(0) o35->88448a6d-0db1-45c7-85b6-dee4316839af@0@lo:0/0 lens 392/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'touch.0' Lustre: 20391:0:(service.c:2157:ptlrpc_server_handle_req_in()) Skipped 3 previous similar messages Lustre: 23452:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 6s req@ffff8802ec083880 x1709914388155776/t0(0) o10->88448a6d-0db1-45c7-85b6-dee4316839af@0@lo:0/0 lens 440/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'truncate.0' Lustre: 20643:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000404:0x180:0x0] with magic=0xbd60bd0 Lustre: 20643:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 1 previous similar message Lustre: lustre-OST0000-osc-ffff88006dee1bf8: Connection restored to 192.168.123.125@tcp (at 0@lo) LustreError: 11-0: lustre-OST0000-osc-ffff88006dee1bf8: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-OST0000-osc-ffff88006dee1bf8: Connection to lustre-OST0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0000-osc-ffff88006dee1bf8: This client was evicted by lustre-OST0000; in progress operations using this service will fail. LustreError: 21883:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0000: '10' migration was interrupted, run 'lfs migrate -m 1 -c 1 -H crush 10' to finish migration. Lustre: dir [0x240000403:0x2b9:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 7 previous similar messages LustreError: 16362:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 141s: evicting client at 0@lo ns: filter-lustre-OST0001_UUID lock: ffff8800a7e4c780/0xb62322e87066316b lrc: 3/0,0 mode: PW/PW res: [0x300000400:0x69f:0x0].0x0 rrc: 4 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000480020020 nid: 0@lo remote: 0xb62322e870663156 expref: 4756 pid: 22316 timeout: 12255 lvb_type: 0 LustreError: 6953:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802752cd7c0 x1709914391187264/t0(0) o104->lustre-OST0001@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 6953:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 12 previous similar messages LustreError: 11-0: lustre-OST0001-osc-ffff88006dee4a88: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-OST0001-osc-ffff88006dee4a88: Connection to lustre-OST0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0001-osc-ffff88006dee4a88: This client was evicted by lustre-OST0001; in progress operations using this service will fail. LustreError: 7266:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0001-osc-ffff88006dee4a88: namespace resource [0x580:0x0:0x0].0x0 (ffff88028e497240) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 7266:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 129 previous similar messages Lustre: 6950:0:(service.c:1378:ptlrpc_at_send_early_reply()) @@@ Already past deadline (-1s), not sending early reply. Consider increasing at_early_margin (5)? req@ffff8802dd4e5e00 x1709914390347072/t0(0) o2->44f0f5e5-394f-4061-bca1-57e5378d4d07@0@lo:545/0 lens 440/0 e 1 to 0 dl 1630702395 ref 2 fl New:/0/ffffffff rc 0/-1 job:'touch.0' Lustre: 22300:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (58/1s); client may timeout req@ffff8802d0b48040 x1709914390150656/t0(0) o101->44f0f5e5-394f-4061-bca1-57e5378d4d07@0@lo:545/0 lens 328/400 e 0 to 0 dl 1630702395 ref 1 fl Complete:/0/0 rc 0/0 job:'ln.0' LustreError: 22324:0:(service.c:2290:ptlrpc_server_handle_request()) @@@ Dropping timed-out request from 12345-0@lo: deadline 51/2s ago req@ffff8802dd4e5e00 x1709914390347072/t0(0) o2->44f0f5e5-394f-4061-bca1-57e5378d4d07@0@lo:545/0 lens 440/0 e 1 to 0 dl 1630702395 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'touch.0' Lustre: 21142:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (51/2s); client may timeout req@ffff8802dd4e1300 x1709914390361984/t0(0) o101->44f0f5e5-394f-4061-bca1-57e5378d4d07@0@lo:545/0 lens 328/0 e 1 to 0 dl 1630702395 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'rm.0' LustreError: 22324:0:(service.c:2290:ptlrpc_server_handle_request()) Skipped 27 previous similar messages Lustre: 6932:0:(service.c:1378:ptlrpc_at_send_early_reply()) @@@ Already past deadline (-1s), not sending early reply. Consider increasing at_early_margin (5)? req@ffff8802821f1300 x1709914390367552/t0(0) o2->44f0f5e5-394f-4061-bca1-57e5378d4d07@0@lo:546/0 lens 440/0 e 1 to 0 dl 1630702396 ref 2 fl New:/0/ffffffff rc 0/-1 job:'touch.0' Lustre: 6932:0:(service.c:1378:ptlrpc_at_send_early_reply()) Skipped 67 previous similar messages LustreError: 22343:0:(service.c:2290:ptlrpc_server_handle_request()) @@@ Dropping timed-out request from 12345-0@lo: deadline 51/1s ago req@ffff8802b2581300 x1709914390418432/t0(0) o2->44f0f5e5-394f-4061-bca1-57e5378d4d07@0@lo:547/0 lens 440/0 e 1 to 0 dl 1630702397 ref 2 fl Interpret:/0/ffffffff rc 0/-1 job:'touch.0' Lustre: 22339:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (51/1s); client may timeout req@ffff8802367f57c0 x1709914390409984/t4295001761(0) o2->44f0f5e5-394f-4061-bca1-57e5378d4d07@0@lo:547/0 lens 440/432 e 1 to 0 dl 1630702397 ref 2 fl Complete:/0/0 rc 0/0 job:'touch.0' Lustre: 22339:0:(service.c:2328:ptlrpc_server_handle_request()) Skipped 221 previous similar messages LustreError: 22343:0:(service.c:2290:ptlrpc_server_handle_request()) Skipped 190 previous similar messages LustreError: 28438:0:(ofd_obd.c:501:ofd_destroy_export()) lustre-OST0001: cli 88448a6d-0db1-45c7-85b6-dee4316839af/ffff880246771bf8 has 90112 pending on destroyed export Lustre: 22310:0:(service.c:1378:ptlrpc_at_send_early_reply()) @@@ Already past deadline (-2s), not sending early reply. Consider increasing at_early_margin (5)? req@ffff8802a81e2c00 x1709914390432192/t4295001449(0) o2->44f0f5e5-394f-4061-bca1-57e5378d4d07@0@lo:548/0 lens 440/432 e 1 to 0 dl 1630702398 ref 1 fl Complete:/0/0 rc 0/0 job:'touch.0' Lustre: 22310:0:(service.c:1378:ptlrpc_at_send_early_reply()) Skipped 567 previous similar messages Lustre: lustre-OST0001-osc-ffff88006dee4a88: Connection restored to 192.168.123.125@tcp (at 0@lo) Lustre: Skipped 1 previous similar message Lustre: 13169:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff8802d6840680 x1709914392196608/t0(0) o10->lustre-OST0001-osc-ffff88006dee1bf8@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630702441 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'truncate.0' Lustre: 13171:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1630702345/real 1630702345] req@ffff88029a2e9f80 x1709914390378368/t0(0) o2->lustre-OST0000-osc-ffff88006dee1bf8@0@lo:28/4 lens 440/432 e 1 to 1 dl 1630702409 ref 1 fl Rpc:XQr/0/ffffffff rc 0/-1 job:'touch.0' Lustre: lustre-OST0000: Client 44f0f5e5-394f-4061-bca1-57e5378d4d07 (at 0@lo) reconnecting Lustre: 13172:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1630702344/real 1630702344] req@ffff8802763770c0 x1709914390354560/t0(0) o101->lustre-OST0002-osc-ffff88006dee1bf8@0@lo:28/4 lens 328/400 e 1 to 1 dl 1630702414 ref 1 fl Rpc:XNQr/0/ffffffff rc 0/-1 job:'rm.0' Lustre: 13172:0:(client.c:2285:ptlrpc_expire_one_request()) Skipped 22 previous similar messages Lustre: lustre-OST0003: Client 44f0f5e5-394f-4061-bca1-57e5378d4d07 (at 0@lo) reconnecting Lustre: Skipped 2 previous similar messages LustreError: 19428:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0000: '16' migration was interrupted, run 'lfs migrate -m 1 -c 1 -H crush 16' to finish migration. Lustre: ll_ost01_034: service thread pid 4620 was inactive for 40.006 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 4676, comm: ll_ost03_045 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] ofd_destroy_by_fid+0x1d1/0x520 [ofd] [<0>] ofd_destroy_hdl+0x27f/0xa20 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 4608, comm: ll_ost01_027 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] ofd_destroy_by_fid+0x1d1/0x520 [ofd] [<0>] ofd_destroy_hdl+0x27f/0xa20 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: Skipped 2 previous similar messages Pid: 4620, comm: ll_ost01_034 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] ofd_destroy_by_fid+0x1d1/0x520 [ofd] [<0>] ofd_destroy_hdl+0x27f/0xa20 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: ll_ost01_016: service thread pid 25379 was inactive for 40.112 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: ll_ost01_020: service thread pid 25509 was inactive for 40.059 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 11 previous similar messages Lustre: ll_ost01_009: service thread pid 21146 was inactive for 41.514 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 3 previous similar messages LustreError: 16405:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000403:0x201:0x0]: rc = -2 Lustre: 18773:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0x356:0x0] with magic=0xbd60bd0 Lustre: 18773:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 5 previous similar messages Lustre: ll_ost05_002: service thread pid 17073 was inactive for 41.017 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 1 previous similar message LustreError: 17080:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 284206784 != fo_tot_granted 284296896 LustreError: 17080:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 231 previous similar messages LustreError: 17080:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 212992 != fo_tot_pending 303104 LustreError: 17080:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 231 previous similar messages Lustre: ll_ost07_023: service thread pid 4260 was inactive for 86.098 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 4 previous similar messages LustreError: 16362:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0000_UUID lock: ffff88029dc87840/0xb62322e8706ba9db lrc: 3/0,0 mode: PW/PW res: [0x45d:0x0:0x0].0x0 rrc: 4 type: EXT [0->18446744073709551615] (req 0->8388607) gid 0 flags: 0x60000000020020 nid: 0@lo remote: 0xb62322e8706ba9d4 expref: 2814 pid: 21133 timeout: 12355 lvb_type: 0 Lustre: 13182:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.125@tcp:/lustre/fid: [0x200000403:0x348:0x0]// may get corrupted (rc -108) Lustre: 13182:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.125@tcp:/lustre/fid: [0x280000403:0x3a3:0x0]// may get corrupted (rc -108) Lustre: 13170:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.125@tcp:/lustre/fid: [0x200000403:0x357:0x0]// may get corrupted (rc -108) LustreError: 7651:0:(vvp_io.c:1793:vvp_io_init()) lustre: refresh file layout [0x200000403:0x357:0x0] error -108. Lustre: 13181:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.125@tcp:/lustre/fid: [0x200000403:0x377:0x0]// may get corrupted (rc -108) Lustre: 13181:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.125@tcp:/lustre/fid: [0x280000403:0x3b2:0x0]// may get corrupted (rc -108) LustreError: 21311:0:(ofd_obd.c:501:ofd_destroy_export()) lustre-OST0000: cli 44f0f5e5-394f-4061-bca1-57e5378d4d07/ffff88026406dd28 has 61440 pending on destroyed export Lustre: 13174:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.125@tcp:/lustre/fid: [0x280000403:0x31a:0x0]// may get corrupted (rc -108) LustreError: 16356:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630702508 with bad export cookie 13124372120798736530 LustreError: 16356:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) Skipped 1 previous similar message LustreError: 9176:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0002-osc-ffff88006dee1bf8: namespace resource [0xfa2:0x0:0x0].0x0 (ffff880074252fc0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 9176:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 1176 previous similar messages Lustre: dir [0x200000403:0x37a:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 2 previous similar messages 1[8510]: segfault at 8 ip 00007f0fbaffa7e8 sp 00007fff7d8b3e50 error 4 in ld-2.17.so[7f0fbafef000+22000] LustreError: 8510:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff88006dee1bf8: inode [0x240000403:0x35d:0x0] mdc close failed: rc = -13 LustreError: 17075:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88025922bec0 x1709914398331712/t0(0) o104->lustre-OST0003@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 16357:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630702547 with bad export cookie 13124372120799707640 LustreError: 17075:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 57 previous similar messages LustreError: 11-0: lustre-OST0003-osc-ffff88006dee1bf8: operation ost_getattr to node 0@lo failed: rc = -107 Lustre: lustre-OST0003-osc-ffff88006dee1bf8: Connection to lustre-OST0003 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 6 previous similar messages LustreError: Skipped 2 previous similar messages LustreError: 167-0: lustre-OST0003-osc-ffff88006dee1bf8: This client was evicted by lustre-OST0003; in progress operations using this service will fail. LustreError: Skipped 2 previous similar messages Lustre: 13167:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.125@tcp:/lustre/fid: [0x200000403:0x3fc:0x0]/ may get corrupted (rc -108) LustreError: 21311:0:(ofd_obd.c:501:ofd_destroy_export()) lustre-OST0003: cli 44f0f5e5-394f-4061-bca1-57e5378d4d07/ffff88025a8812a8 has 61440 pending on destroyed export Lustre: lustre-OST0003-osc-ffff88006dee1bf8: Connection restored to 192.168.123.125@tcp (at 0@lo) 1[10053]: segfault at 8 ip 00007fbaa2eba7e8 sp 00007ffde6ddecb0 error 4 in ld-2.17.so[7fbaa2eaf000+22000] Lustre: Skipped 6 previous similar messages LustreError: 10053:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff88006dee4a88: inode [0x240000403:0x35d:0x0] mdc close failed: rc = -13 Lustre: 21665:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 479 < left 43328, rollback = 7 Lustre: 21665:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 221796 previous similar messages Lustre: 21665:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 1008/4032/0, destroy: 0/0/0 Lustre: 21665:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 222077 previous similar messages Lustre: 21665:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/1, xattr_set: 1010/15/0 Lustre: 21665:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 222077 previous similar messages Lustre: 21665:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 5041/43354/26, punch: 0/0/0, quota 10/106/6 Lustre: 21665:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 222076 previous similar messages Lustre: 21665:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 1008/17136/0, delete: 0/0/0 Lustre: 21665:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 222076 previous similar messages Lustre: 21665:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 21665:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 222077 previous similar messages 1[10601]: segfault at 8 ip 00007f95d48907e8 sp 00007ffc8d6e39f0 error 4 in ld-2.17.so[7f95d4885000+22000] LustreError: 9997:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff88006dee4a88: inode [0x200000404:0x2d6:0x0] mdc close failed: rc = -13 Lustre: dir [0x240000403:0x30c:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 4 previous similar messages LustreError: 16362:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0001_UUID lock: ffff8802d8b99a80/0xb62322e87077be5a lrc: 3/0,0 mode: PW/PW res: [0x300000400:0x928:0x0].0x0 rrc: 4 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000400020020 nid: 0@lo remote: 0xb62322e87077be3e expref: 2870 pid: 23533 timeout: 12538 lvb_type: 0 LustreError: 16362:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 2 previous similar messages Lustre: 19127:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 8: before 511 < left 2276, rollback = 9 Lustre: 19127:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 241 previous similar messages Lustre: 13168:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.125@tcp:/lustre/fid: [0x280000404:0x3a2:0x0]// may get corrupted (rc -108) Lustre: 13167:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.125@tcp:/lustre/fid: [0x280000404:0x413:0x0]/ may get corrupted (rc -108) 4[12354]: segfault at 8 ip 00007f1e4a6de7e8 sp 00007ffdf46ba3d0 error 4 in ld-2.17.so[7f1e4a6d3000+22000] 1[12422]: segfault at 8 ip 00007fe1e2b3b7e8 sp 00007ffec0144040 error 4 in ld-2.17.so[7fe1e2b30000+22000] Lustre: 13178:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.125@tcp:/lustre/fid: [0x240000403:0x384:0x0]// may get corrupted (rc -108) LustreError: 16406:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000404:0x1b0:0x0]: rc = -2 Lustre: lustre-OST0002-osc-ffff88006dee1bf8: disconnect after 20s idle LustreError: 20383:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000404:0x1b0:0x0]: rc = -2 LustreError: 21731:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0000: '11' migration was interrupted, run 'lfs migrate -m 2 -c 1 -H crush 11' to finish migration. Lustre: 19931:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 6s req@ffff8800a8a00040 x1709914409762624/t0(0) o36->88448a6d-0db1-45c7-85b6-dee4316839af@0@lo:0/0 lens 688/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'lfs.0' Lustre: 16369:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000404:0x2fd:0x0] with magic=0xbd60bd0 Lustre: 16369:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 13 previous similar messages LustreError: 11-0: lustre-OST0003-osc-ffff88006dee4a88: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 6 previous similar messages Lustre: lustre-OST0003-osc-ffff88006dee4a88: Connection to lustre-OST0003 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 21953:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8800a154f700 x1709914412151808/t0(0) o106->lustre-OST0003@0@lo:15/16 lens 328/280 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 21953:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 65 previous similar messages Lustre: Skipped 2 previous similar messages LustreError: 167-0: lustre-OST0003-osc-ffff88006dee4a88: This client was evicted by lustre-OST0003; in progress operations using this service will fail. LustreError: Skipped 2 previous similar messages LustreError: 26788:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630702881 with bad export cookie 13124372120798736754 LustreError: 26788:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) Skipped 2 previous similar messages Lustre: 13171:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.125@tcp:/lustre/fid: [0x240000404:0x3ef:0x0]// may get corrupted (rc -108) Lustre: 13172:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.125@tcp:/lustre/fid: [0x200000403:0x456:0x0]// may get corrupted (rc -108) LustreError: 30631:0:(ofd_obd.c:501:ofd_destroy_export()) lustre-OST0000: cli 88448a6d-0db1-45c7-85b6-dee4316839af/ffff88024451c138 has 151552 pending on destroyed export Lustre: 13172:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.125@tcp:/lustre/fid: [0x200000404:0x4a1:0x0]/ may get corrupted (rc -108) Lustre: 13171:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.125@tcp:/lustre/fid: [0x200000404:0x466:0x0]// may get corrupted (rc -108) LustreError: 14543:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0000-osc-ffff88006dee4a88: namespace resource [0x2340:0x0:0x0].0x0 (ffff8802708aeac0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 14543:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 1224 previous similar messages Lustre: lustre-OST0000-osc-ffff88006dee4a88: Connection restored to 192.168.123.125@tcp (at 0@lo) Lustre: Skipped 2 previous similar messages Lustre: 13173:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff8802bda19f80 x1709914415892864/t0(0) o10->lustre-OST0002-osc-ffff88006dee1bf8@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630702997 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dir_create.sh.0' Lustre: 13173:0:(client.c:1485:after_reply()) Skipped 1 previous similar message LustreError: 20702:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000404:0x419:0x0]: rc = -2 LustreError: 17089:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 276718784 != fo_tot_granted 276808896 LustreError: 17089:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 827 previous similar messages LustreError: 17089:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 90112 LustreError: 17089:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 827 previous similar messages LustreError: 16362:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0003_UUID lock: ffff8800a5544000/0xb62322e87090063e lrc: 3/0,0 mode: PW/PW res: [0x380000401:0x2b13:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 2097152->18446744073709551615) gid 0 flags: 0x60000400020020 nid: 0@lo remote: 0xb62322e870900637 expref: 4502 pid: 8985 timeout: 12900 lvb_type: 0 LustreError: 16362:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 4 previous similar messages LustreError: 16350:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630703035 with bad export cookie 13124372120800953381 LustreError: 16350:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) Skipped 3 previous similar messages Lustre: 16395:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 6s req@ffff8802207d1c48 x1709914419735232/t0(0) o37->88448a6d-0db1-45c7-85b6-dee4316839af@0@lo:0/0 lens 448/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'ls.0' Lustre: 13182:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.125@tcp:/lustre/fid: [0x200000404:0x4d1:0x0]// may get corrupted (rc -108) Lustre: 13181:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.125@tcp:/lustre/fid: [0x200000404:0x569:0x0]// may get corrupted (rc -108) LustreError: 16422:0:(out_handler.c:910:out_tx_end()) lustre-MDT0002-osd: undo for /home/green/git/lustre-release/lustre/ptlrpc/../../lustre/target/out_handler.c:445: rc = -524 LustreError: 16422:0:(out_handler.c:910:out_tx_end()) Skipped 1 previous similar message Lustre: 13177:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.125@tcp:/lustre/fid: [0x200000404:0x57f:0x0]/ may get corrupted (rc -108) LustreError: 19530:0:(ofd_obd.c:501:ofd_destroy_export()) lustre-OST0002: cli 88448a6d-0db1-45c7-85b6-dee4316839af/ffff8800a8359bf8 has 28672 pending on destroyed export LustreError: 16661:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0002-osp-MDT0001: fail to cancel 1 llog-records: rc = -2 LustreError: 16661:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0002-osp-MDT0001: fail to cancel 1 of 1 llog-records: rc = -2 LustreError: 8713:0:(ldlm_lockd.c:1425:ldlm_handle_enqueue0()) ### lock on destroyed export ffff8802e3bf0008 ns: filter-lustre-OST0000_UUID lock: ffff8802e2210b80/0xb62322e8709c67d9 lrc: 3/0,0 mode: PW/PW res: [0x2d3c:0x0:0x0].0x0 rrc: 2 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x50000000020000 nid: 0@lo remote: 0xb62322e8709c66f9 expref: 3653 pid: 8713 timeout: 0 lvb_type: 0 Lustre: 13178:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.125@tcp:/lustre/fid: [0x200000403:0x4b2:0x0]// may get corrupted (rc -5) Lustre: 13177:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.125@tcp:/lustre/fid: [0x240000403:0x490:0x0]// may get corrupted (rc -108) Lustre: 13182:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.125@tcp:/lustre/fid: [0x280000403:0x34f:0x0]// may get corrupted (rc -108) Lustre: 15232:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 6s req@ffff8802a0bf70c0 x1709914425951872/t0(0) o400->lustre-MDT0000-mdtlov_UUID@0@lo:0/0 lens 224/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'kworker/u32:8.0' Lustre: 15232:0:(service.c:2157:ptlrpc_server_handle_req_in()) Skipped 1 previous similar message Lustre: 16382:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 504 < left 30583, rollback = 7 Lustre: 16382:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 358539 previous similar messages Lustre: 16382:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 711/2844/0, destroy: 0/0/0 Lustre: 16382:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 358847 previous similar messages Lustre: 16382:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/1, xattr_set: 713/15/0 Lustre: 16382:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 358847 previous similar messages Lustre: 16382:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 3556/30583/0, punch: 0/0/0, quota 10/106/7 Lustre: 16382:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 358848 previous similar messages Lustre: 16382:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 711/12087/0, delete: 0/0/0 Lustre: 16382:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 358849 previous similar messages Lustre: 16382:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 16382:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 358847 previous similar messages ptlrpc_watchdog_fire: 27 callbacks suppressed Lustre: ll_ost07_020: service thread pid 2483 was inactive for 250.393 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 2483, comm: ll_ost07_020 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] tgt_extent_lock+0xea/0x2a0 [ptlrpc] [<0>] ofd_getattr_hdl+0x365/0x720 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: 19587:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 6s req@ffff8802f4d21f80 x1709914428569408/t0(0) o106->LOV_OSC_UUID@0@lo:0/0 lens 328/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'' Lustre: 19587:0:(service.c:2157:ptlrpc_server_handle_req_in()) Skipped 4 previous similar messages LustreError: 2483:0:(ldlm_request.c:123:ldlm_expired_completion_wait()) ### lock timed out (enqueued at 1630702977, 300s ago); not entering recovery in server code, just going back to sleep ns: filter-lustre-OST0002_UUID lock: ffff8802e2a84780/0xb62322e87093df38 lrc: 3/0,1 mode: --/PW res: [0x2222:0x0:0x0].0x0 rrc: 4 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x40010000000000 nid: local remote: 0x0 expref: -99 pid: 2483 timeout: 0 lvb_type: 0 LustreError: dumping log to /tmp/lustre-log.1630703284.2483 Lustre: 19649:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 502 < left 1116, rollback = 2 Lustre: 19649:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 333 previous similar messages Lustre: mdt: This server is not able to keep up with request traffic (cpu-bound). Lustre: 19716:0:(service.c:1612:ptlrpc_at_check_timed()) earlyQ=0 reqQ=0 recA=4, svcEst=92, delay=5987ms Lustre: mdt: This server is not able to keep up with request traffic (cpu-bound). Lustre: 17888:0:(service.c:1612:ptlrpc_at_check_timed()) earlyQ=0 reqQ=0 recA=2, svcEst=91, delay=6154ms Lustre: 22343:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 6s req@ffff8802f2648cc0 x1709914430890368/t0(0) o2->88448a6d-0db1-45c7-85b6-dee4316839af@0@lo:0/0 lens 440/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'chown.0' Lustre: 22343:0:(service.c:2157:ptlrpc_server_handle_req_in()) Skipped 1 previous similar message Lustre: 16392:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000404:0x4ed:0x0] with magic=0xbd60bd0 Lustre: 16392:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 43 previous similar messages Lustre: lustre-OST0001: haven't heard from client 44f0f5e5-394f-4061-bca1-57e5378d4d07 (at 0@lo) in 50 seconds. I think it's dead, and I am evicting it. exp ffff880239398008, cur 1630703369 expire 1630703339 last 1630703319 Lustre: lustre-OST0001-osc-ffff88006dee1bf8: disconnect after 55s idle LustreError: 17801:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630703377 with bad export cookie 13124372120804603930 LustreError: 17801:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) Skipped 13 previous similar messages LustreError: 10095:0:(ldlm_lockd.c:1425:ldlm_handle_enqueue0()) ### lock on destroyed export ffff88027c526fc8 ns: filter-lustre-OST0003_UUID lock: ffff88008d953c40/0xb62322e870aee74b lrc: 3/0,0 mode: --/PW res: [0x380000401:0x2b26:0x0].0x0 rrc: 3 type: EXT [2097152->18446744073709551615] (req 2097152->18446744073709551615) gid 0 flags: 0x50000000020000 nid: 0@lo remote: 0xb62322e870aee705 expref: 6998 pid: 10095 timeout: 0 lvb_type: 0 Lustre: dir [0x200000404:0x38d:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 3 previous similar messages Lustre: 13176:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.125@tcp:/lustre/fid: [0x200000404:0x4b7:0x0]/ may get corrupted (rc -5) Lustre: 13175:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.125@tcp:/lustre/fid: [0x240000404:0x504:0x0]// may get corrupted (rc -108) Lustre: 13176:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.125@tcp:/lustre/fid: [0x280000404:0x563:0x0]// may get corrupted (rc -108) Lustre: 13176:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.125@tcp:/lustre/fid: [0x280000403:0x34f:0x0]// may get corrupted (rc -108) Lustre: 13176:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.125@tcp:/lustre/fid: [0x200000404:0x69f:0x0]/ may get corrupted (rc -108) Lustre: 13175:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.125@tcp:/lustre/fid: [0x280000404:0x521:0x0]// may get corrupted (rc -108) Lustre: 24577:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 7s req@ffff880294df8680 x1709914433092608/t0(0) o101->44f0f5e5-394f-4061-bca1-57e5378d4d07@0@lo:0/0 lens 4616/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'lfs.0' Lustre: 24577:0:(service.c:2157:ptlrpc_server_handle_req_in()) Skipped 3 previous similar messages Lustre: mdt: This server is not able to keep up with request traffic (cpu-bound). Lustre: 16381:0:(service.c:1612:ptlrpc_at_check_timed()) earlyQ=0 reqQ=0 recA=2, svcEst=78, delay=5871ms LustreError: 20581:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000403:0x414:0x0]: rc = -2 Lustre: 13173:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff8802bda19f80 x1709914436146880/t0(0) o10->lustre-OST0002-osc-ffff88006dee1bf8@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630703558 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dir_create.sh.0' Lustre: 13173:0:(client.c:1485:after_reply()) Skipped 4 previous similar messages Lustre: 13982:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (61/1s); client may timeout req@ffff8802538d8040 x1709914434192896/t0(0) o1->88448a6d-0db1-45c7-85b6-dee4316839af@0@lo:135/0 lens 440/432 e 0 to 0 dl 1630703495 ref 2 fl Complete:/0/0 rc 0/0 job:'lfs.0' LustreError: 17072:0:(service.c:2290:ptlrpc_server_handle_request()) @@@ Dropping timed-out request from 12345-0@lo: deadline 61/1s ago req@ffff8800779d6a80 x1709914434230144/t0(0) o1->88448a6d-0db1-45c7-85b6-dee4316839af@0@lo:135/0 lens 440/0 e 0 to 0 dl 1630703495 ref 2 fl Interpret:/0/ffffffff rc 0/-1 job:'lfs.0' Lustre: 13982:0:(service.c:2328:ptlrpc_server_handle_request()) Skipped 110 previous similar messages Lustre: 13991:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (61/1s); client may timeout req@ffff880247db0040 x1709914434198656/t0(0) o1->88448a6d-0db1-45c7-85b6-dee4316839af@0@lo:135/0 lens 440/432 e 0 to 0 dl 1630703495 ref 2 fl Complete:/0/0 rc 0/0 job:'lfs.0' Lustre: 13991:0:(service.c:2328:ptlrpc_server_handle_request()) Skipped 6 previous similar messages Lustre: 23561:0:(service.c:1378:ptlrpc_at_send_early_reply()) @@@ Already past deadline (-1s), not sending early reply. Consider increasing at_early_margin (5)? req@ffff88025922cb40 x1709914434224448/t0(0) o1->88448a6d-0db1-45c7-85b6-dee4316839af@0@lo:135/0 lens 440/0 e 0 to 0 dl 1630703495 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'lfs.0' Lustre: 23561:0:(service.c:1378:ptlrpc_at_send_early_reply()) Skipped 549 previous similar messages Lustre: 13986:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (61/2s); client may timeout req@ffff880247db57c0 x1709914434198272/t0(0) o1->88448a6d-0db1-45c7-85b6-dee4316839af@0@lo:135/0 lens 440/432 e 0 to 0 dl 1630703495 ref 1 fl Complete:/0/0 rc 0/0 job:'lfs.0' Lustre: 13986:0:(service.c:2328:ptlrpc_server_handle_request()) Skipped 17 previous similar messages LustreError: 23502:0:(service.c:2290:ptlrpc_server_handle_request()) @@@ Dropping timed-out request from 12345-0@lo: deadline 61/2s ago req@ffff8802f1811300 x1709914434236928/t0(0) o1->88448a6d-0db1-45c7-85b6-dee4316839af@0@lo:136/0 lens 440/0 e 0 to 0 dl 1630703496 ref 2 fl Interpret:/0/ffffffff rc 0/-1 job:'lfs.0' LustreError: 23502:0:(service.c:2290:ptlrpc_server_handle_request()) Skipped 58 previous similar messages Lustre: 23559:0:(service.c:1378:ptlrpc_at_send_early_reply()) @@@ Already past deadline (-2s), not sending early reply. Consider increasing at_early_margin (5)? req@ffff8802eda84500 x1709914434238848/t0(0) o1->88448a6d-0db1-45c7-85b6-dee4316839af@0@lo:136/0 lens 440/0 e 0 to 0 dl 1630703496 ref 2 fl New:/0/ffffffff rc 0/-1 job:'lfs.0' Lustre: 23559:0:(service.c:1378:ptlrpc_at_send_early_reply()) Skipped 242 previous similar messages Lustre: 13980:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (61/4s); client may timeout req@ffff880071caf0c0 x1709914434197888/t0(0) o1->88448a6d-0db1-45c7-85b6-dee4316839af@0@lo:135/0 lens 440/432 e 0 to 0 dl 1630703495 ref 1 fl Complete:/0/0 rc 0/0 job:'lfs.0' Lustre: 13980:0:(service.c:2328:ptlrpc_server_handle_request()) Skipped 63 previous similar messages Lustre: 13177:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1630703435/real 1630703435] req@ffff88027085bec0 x1709914434235840/t0(0) o1->lustre-OST0003-osc-ffff88006dee4a88@0@lo:28/4 lens 440/432 e 0 to 1 dl 1630703500 ref 1 fl Rpc:XQr/0/ffffffff rc 0/-1 job:'lfs.0' Lustre: 13177:0:(client.c:2285:ptlrpc_expire_one_request()) Skipped 18 previous similar messages Lustre: lustre-OST0003-osc-ffff88006dee4a88: Connection to lustre-OST0003 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 11 previous similar messages Lustre: lustre-OST0003: Client 88448a6d-0db1-45c7-85b6-dee4316839af (at 0@lo) reconnecting Lustre: lustre-OST0003-osc-ffff88006dee4a88: Connection restored to 192.168.123.125@tcp (at 0@lo) Lustre: Skipped 11 previous similar messages Lustre: 21137:0:(service.c:1436:ptlrpc_at_send_early_reply()) @@@ Could not add any time (5/5), not sending early reply req@ffff8802e4e0c500 x1709914416356160/t0(0) o1->44f0f5e5-394f-4061-bca1-57e5378d4d07@0@lo:217/0 lens 440/432 e 18 to 0 dl 1630703577 ref 2 fl Interpret:/0/0 rc 0/0 job:'lfs.0' LustreError: 16362:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0002_UUID lock: ffff8802276516c0/0xb62322e870b3cf54 lrc: 3/0,0 mode: PW/PW res: [0x340000400:0x1d15:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000400000020 nid: 0@lo remote: 0xb62322e870b3cf38 expref: 6555 pid: 15311 timeout: 13445 lvb_type: 0 LustreError: 16362:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 8 previous similar messages LustreError: 4661:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88009591f700 x1709914440479872/t0(0) o104->lustre-OST0002@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 4661:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 638 previous similar messages Lustre: 2483:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (600/3s); client may timeout req@ffff8802e4e0c500 x1709914416356160/t0(0) o1->44f0f5e5-394f-4061-bca1-57e5378d4d07@0@lo:217/0 lens 440/432 e 18 to 0 dl 1630703577 ref 1 fl Complete:/0/0 rc 0/0 job:'lfs.0' LustreError: 12232:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630703580 with bad export cookie 13124372120803290079 Lustre: 2483:0:(service.c:2328:ptlrpc_server_handle_request()) Skipped 7 previous similar messages LustreError: 11-0: lustre-OST0002-osc-ffff88006dee1bf8: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 33 previous similar messages LustreError: 167-0: lustre-OST0002-osc-ffff88006dee1bf8: This client was evicted by lustre-OST0002; in progress operations using this service will fail. LustreError: Skipped 11 previous similar messages LustreError: 12227:0:(ofd_obd.c:501:ofd_destroy_export()) lustre-OST0002: cli 44f0f5e5-394f-4061-bca1-57e5378d4d07/ffff8802154b2548 has 61440 pending on destroyed export LustreError: 17091:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 276482496 != fo_tot_granted 276785600 LustreError: 17091:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 780 previous similar messages LustreError: 17091:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 28672 != fo_tot_pending 331776 LustreError: 17091:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 780 previous similar messages Lustre: 13174:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.125@tcp:/lustre/fid: [0x200000404:0x660:0x0]/ may get corrupted (rc -108) Lustre: 13173:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.125@tcp:/lustre/fid: [0x240000404:0x50d:0x0]// may get corrupted (rc -108) Lustre: 13178:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1630703434/real 1630703434] req@ffff8802ecaac500 x1709914434230144/t0(0) o1->lustre-OST0000-osc-ffff88006dee4a88@0@lo:28/4 lens 440/432 e 0 to 1 dl 1630703532 ref 1 fl Rpc:XQr/0/ffffffff rc 0/-1 job:'lfs.0' Lustre: 13178:0:(client.c:2285:ptlrpc_expire_one_request()) Skipped 11 previous similar messages Lustre: lustre-OST0000: Client 88448a6d-0db1-45c7-85b6-dee4316839af (at 0@lo) reconnecting LustreError: 19486:0:(mdt_open.c:1237:mdt_cross_open()) lustre-MDT0001: [0x240000404:0x2ec:0x0] doesn't exist!: rc = -14 LustreError: 23939:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0002-osc-ffff88006dee1bf8: namespace resource [0x340000401:0x2c3f:0x0].0x0 (ffff8802a9a287c0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 23939:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 3575 previous similar messages Lustre: 13174:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.125@tcp:/lustre/fid: [0x200000404:0x75d:0x0]// may get corrupted (rc -108) Lustre: 13173:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.125@tcp:/lustre/fid: [0x280000403:0x574:0x0]// may get corrupted (rc -108) Lustre: lustre-OST0003-osc-ffff88006dee4a88: disconnect after 30s idle ------------[ cut here ]------------ WARNING: CPU: 1 PID: 12901 at /home/green/git/lustre-release/ldiskfs/ext4_jbd2.c:266 __ldiskfs_handle_dirty_metadata+0x1d2/0x230 [ldiskfs] Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) dm_flakey dm_mod loop zfs(PO) zunicode(PO) zzstd(O) zlua(O) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) jbd2 mbcache crc_t10dif crct10dif_generic sb_edac edac_core iosf_mbi crc32_pclmul ghash_clmulni_intel aesni_intel lrw gf128mul glue_helper ablk_helper cryptd virtio_console pcspkr i2c_piix4 virtio_balloon ip_tables rpcsec_gss_krb5 drm_kms_helper ttm ata_generic pata_acpi drm crct10dif_pclmul crct10dif_common drm_panel_orientation_quirks ata_piix crc32c_intel serio_raw virtio_blk i2c_core libata floppy [last unloaded: libcfs] CPU: 1 PID: 12901 Comm: mdt00_014 Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011 Call Trace: [<ffffffff817d9418>] dump_stack+0x19/0x1b [<ffffffff8108d538>] __warn+0xd8/0x100 [<ffffffff8108d67d>] warn_slowpath_null+0x1d/0x20 [<ffffffffa0db4262>] __ldiskfs_handle_dirty_metadata+0x1d2/0x230 [ldiskfs] [<ffffffffa0d9433d>] ldiskfs_getblk+0x13d/0x210 [ldiskfs] [<ffffffffa0d94437>] ldiskfs_bread+0x27/0xe0 [ldiskfs] [<ffffffffa0cdfe97>] osd_write+0x4e7/0xcb0 [osd_ldiskfs] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffffa04d86f3>] dt_record_write+0x33/0x120 [obdclass] [<ffffffffa04976b2>] llog_osd_write_rec+0x1052/0x19a0 [obdclass] [<ffffffffa0484760>] llog_write_rec+0x290/0x590 [obdclass] [<ffffffffa048a4ee>] llog_cat_add_rec+0x23e/0x950 [obdclass] [<ffffffffa04815ff>] llog_add+0x17f/0x1f0 [obdclass] [<ffffffffa04b2a39>] ? lprocfs_counter_add+0xf9/0x160 [obdclass] [<ffffffffa081b60d>] sub_updates_write+0xb6b/0xe32 [ptlrpc] [<ffffffff811beccd>] ? __free_pages+0x1d/0x30 [<ffffffff8121efe4>] ? kfree+0x1e4/0x200 [<ffffffffa07eb892>] top_trans_stop+0x7c2/0xf60 [ptlrpc] [<ffffffffa0f68f3c>] lod_trans_stop+0x25c/0x340 [lod] [<ffffffffa0e3d408>] mdd_trans_stop+0x28/0x16e [mdd] [<ffffffffa0e1eb2f>] mdd_migrate_object+0xf1f/0x1ba0 [mdd] [<ffffffffa04d7375>] ? lu_object_find_at+0x285/0xb10 [obdclass] [<ffffffffa0e1fb86>] mdd_migrate+0x3d6/0x8e0 [mdd] [<ffffffffa0ef10d8>] mdo_migrate+0x4c/0x4e [mdt] [<ffffffffa0eb0e53>] mdt_reint_migrate+0x1023/0x11d0 [mdt] [<ffffffffa0eb1087>] mdt_reint_rec+0x87/0x240 [mdt] [<ffffffffa0e8655c>] mdt_reint_internal+0x76c/0xb40 [mdt] [<ffffffffa0e93657>] mdt_reint+0x67/0x150 [mdt] [<ffffffffa07da26e>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<ffffffffa035a45e>] ? libcfs_nid2str_r+0xfe/0x130 [lnet] [<ffffffffa0785990>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<ffffffffa0787529>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffffa0786950>] ? ptlrpc_wait_event+0x620/0x620 [ptlrpc] [<ffffffff810ba0f4>] kthread+0xe4/0xf0 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 ---[ end trace e5b03de8bbfe556e ]--- LDISKFS-fs: ldiskfs_getblk:888: aborting transaction: error 28 in __ldiskfs_handle_dirty_metadata LDISKFS-fs error (device dm-0): ldiskfs_getblk:888: inode #182: block 21574: comm mdt00_014: journal_dirty_metadata failed: handle type 0 started at line 1972, credits 512/0, errcode -28 Aborting journal on device dm-0-8. LDISKFS-fs (dm-0): Remounting filesystem read-only LustreError: 12901:0:(osd_io.c:2103:osd_ldiskfs_write_record()) lustre-MDT0000: error reading offset 475136 (block 116, size 24, offs 475104), credits 0/0: rc = -28 LustreError: 12901:0:(update_trans.c:985:top_trans_stop()) lustre-MDT0000-osd: write updates failed: rc = -28 LustreError: 12901:0:(osd_io.c:2110:osd_ldiskfs_write_record()) journal_get_write_access() returned error -30 LustreError: 12901:0:(tgt_lastrcvd.c:1253:tgt_add_reply_data()) lustre-MDT0000: can't update reply_data file: rc = -30 LustreError: 12901:0:(osd_handler.c:2075:osd_trans_stop()) lustre-MDT0000: failed in transaction hook: rc = -30 LDISKFS-fs error (device dm-0) in osd_trans_stop:2082: error 28 LustreError: 17474:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-OST0001-osc-MDT0000: fail to cancel 38 llog-records: rc = -30 LustreError: 17805:0:(osp_sync.c:1094:osp_sync_process_committed()) lustre-OST0002-osc-MDT0000: can't cancel 107 records: rc = -30 LustreError: 12901:0:(osd_handler.c:2085:osd_trans_stop()) lustre-MDT0000: failed to stop transaction: rc = -28 LustreError: 12901:0:(update_trans.c:1011:top_trans_stop()) lustre-MDT0000-osd: stop trans failed: rc = -30 LustreError: 16318:0:(osd_handler.c:1779:osd_trans_commit_cb()) transaction @0xffff88025b16e758 commit error: 2 LustreError: 17474:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) Skipped 5 previous similar messages LustreError: 12901:0:(update_trans.c:1084:top_trans_stop()) lustre-OST0003-osc-MDT0000: stop trans failed: rc = -30 LustreError: 17805:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-OST0002-osc-MDT0000: fail to cancel 1 llog-records: rc = -30 LustreError: 17805:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) Skipped 3 previous similar messages LustreError: 17805:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-OST0002-osc-MDT0000: fail to cancel 1 of 1 llog-records: rc = -30 LustreError: 17805:0:(osp_sync.c:1079:osp_sync_process_committed()) lustre-OST0002-osc-MDT0000: can't cancel record: rc = -30 LustreError: 17805:0:(osp_sync.c:1094:osp_sync_process_committed()) lustre-OST0002-osc-MDT0000: can't cancel 3 records: rc = -30 LustreError: 17805:0:(osp_sync.c:1094:osp_sync_process_committed()) Skipped 8 previous similar messages LustreError: 25185:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -30 LustreError: 25220:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -30 LustreError: 17807:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-OST0003-osc-MDT0000: fail to cancel 55 llog-records: rc = -30 LustreError: 17807:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) Skipped 3 previous similar messages LustreError: 17807:0:(osp_sync.c:1094:osp_sync_process_committed()) lustre-OST0003-osc-MDT0000: can't cancel 55 records: rc = -30 LustreError: 17807:0:(osp_sync.c:1094:osp_sync_process_committed()) Skipped 2 previous similar messages LustreError: 17474:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-OST0001-osc-MDT0000: fail to cancel 12 llog-records: rc = -30 LustreError: 17474:0:(osp_sync.c:1094:osp_sync_process_committed()) lustre-OST0001-osc-MDT0000: can't cancel 12 records: rc = -30 LustreError: 17474:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-OST0001-osc-MDT0000: fail to cancel 1 of 1 llog-records: rc = -30 LustreError: 17474:0:(osp_sync.c:1079:osp_sync_process_committed()) lustre-OST0001-osc-MDT0000: can't cancel record: rc = -30 LustreError: 25309:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -30 LustreError: 25309:0:(llite_lib.c:1836:ll_md_setattr()) Skipped 1 previous similar message LustreError: 17474:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-OST0001-osc-MDT0000: fail to cancel 11 llog-records: rc = -30 LustreError: 17474:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) Skipped 11 previous similar messages LustreError: 17474:0:(osp_sync.c:1094:osp_sync_process_committed()) lustre-OST0001-osc-MDT0000: can't cancel 11 records: rc = -30 LustreError: 17474:0:(osp_sync.c:1094:osp_sync_process_committed()) Skipped 6 previous similar messages LustreError: 16862:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0000-osp-MDT0002: fail to cancel 1 of 1 llog-records: rc = -2 LustreError: 16862:0:(llog_cat.c:787:llog_cat_cancel_records()) Skipped 4 previous similar messages LustreError: 24774:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -30 LustreError: 24774:0:(llite_lib.c:1836:ll_md_setattr()) Skipped 1 previous similar message LustreError: 17474:0:(osp_sync.c:1079:osp_sync_process_committed()) lustre-OST0001-osc-MDT0000: can't cancel record: rc = -30 LustreError: 17474:0:(osp_sync.c:1079:osp_sync_process_committed()) Skipped 4 previous similar messages LustreError: 25353:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -30 LustreError: 25353:0:(llite_lib.c:1836:ll_md_setattr()) Skipped 1 previous similar message LustreError: 16862:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0000-osp-MDT0002: fail to cancel 1 of 1 llog-records: rc = -116 LustreError: 16862:0:(llog_cat.c:787:llog_cat_cancel_records()) Skipped 4 previous similar messages LustreError: 17805:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-OST0002-osc-MDT0000: fail to cancel 67 llog-records: rc = -30 LustreError: 17805:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) Skipped 21 previous similar messages LustreError: 17805:0:(osp_sync.c:1094:osp_sync_process_committed()) lustre-OST0002-osc-MDT0000: can't cancel 67 records: rc = -30 LustreError: 17805:0:(osp_sync.c:1094:osp_sync_process_committed()) Skipped 15 previous similar messages LustreError: 24795:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -30 LustreError: 24795:0:(llite_lib.c:1836:ll_md_setattr()) Skipped 8 previous similar messages LustreError: 16862:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0000-osp-MDT0002: fail to cancel 1 of 1 llog-records: rc = -116 LustreError: 17474:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-OST0001-osc-MDT0000: fail to cancel 208 llog-records: rc = -30 LustreError: 17474:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) Skipped 14 previous similar messages LustreError: 17474:0:(osp_sync.c:1094:osp_sync_process_committed()) lustre-OST0001-osc-MDT0000: can't cancel 208 records: rc = -30 LustreError: 17474:0:(osp_sync.c:1094:osp_sync_process_committed()) Skipped 10 previous similar messages | Externally reported by onyx-68 boilpot email |
racer test 1: racer on clients: centos-10.localnet DURATION=2700 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa17a4e52>] ll_migrate+0x9b2/0xec0 [lustre] PGD 8000000275226067 PUD 2b77ac067 PMD 0 LustreError: 10518:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0000-osp-MDT0002: fail to cancel 1 llog-records: rc = -116 LustreError: 10518:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0000-osp-MDT0002: fail to cancel 1 of 1 llog-records: rc = -116 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) dm_flakey dm_mod loop zfs(PO) zunicode(PO) zzstd(O) zlua(O) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) jbd2 mbcache crc_t10dif crct10dif_generic sb_edac edac_core iosf_mbi crc32_pclmul ghash_clmulni_intel aesni_intel lrw gf128mul glue_helper ablk_helper cryptd virtio_balloon pcspkr i2c_piix4 virtio_console ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm crct10dif_pclmul ata_piix drm crct10dif_common crc32c_intel drm_panel_orientation_quirks serio_raw libata virtio_blk i2c_core floppy [last unloaded: libcfs] LustreError: 12890:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -30 LustreError: 12890:0:(llite_lib.c:1836:ll_md_setattr()) Skipped 1 previous similar message CPU: 1 PID: 9875 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011 task: ffff880294d68010 ti: ffff88027d360000 task.ti: ffff88027d360000 RIP: 0010:[<ffffffffa17a4e52>] [<ffffffffa17a4e52>] ll_migrate+0x9b2/0xec0 [lustre] RSP: 0018:ffff88027d363bc8 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff8800968760d8 RCX: 0000000000000000 RDX: 000000010000fdd3 RSI: ffff8802710f0058 RDI: ffff8802710f0038 RBP: ffff88027d363c48 R08: ffff8802e15a8bb8 R09: ffff88027d363a60 R10: ffff8803315bb240 R11: ffff8800a4bd8040 R12: 0000000000000000 R13: ffff880096935628 R14: ffff8802863fc1d8 R15: 0000000000000030 FS: 00007fa1e9c5e740(0000) GS:ffff880331a40000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 CR2: 0000000000000008 CR3: 00000002903e0000 CR4: 00000000001607e0 Call Trace: [<ffffffff81242213>] ? __check_object_size+0x1c3/0x220 [<ffffffffa1792561>] ll_dir_ioctl+0x5d01/0x6ed0 [lustre] [<ffffffff811b515b>] ? unlock_page+0x2b/0x30 [<ffffffff811e5e2e>] ? do_read_fault.isra.64+0x12e/0x170 [<ffffffff811ea1cf>] ? handle_pte_fault+0x2ef/0xca0 [<ffffffff811ecf6d>] ? __handle_mm_fault+0x31d/0x720 [<ffffffff810d5248>] ? __enqueue_entity+0x78/0x80 [<ffffffff810dbb8b>] ? put_prev_entity+0x8b/0x400 [<ffffffff8125b3fd>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e32be>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffff810c832d>] ? finish_task_switch+0x5d/0x1b0 [<ffffffff8125b751>] SyS_ioctl+0xa1/0xc0 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 | Lustre: lustre-OST0000-osc-ffff88009baa1bf8: disconnect after 21s idle Lustre: 10025:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 512 < left 1324, rollback = 7 Lustre: 10025:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 10025:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 2/178/0 Lustre: 10025:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1315/1324/0, punch: 0/0/0, quota 3/3/0 Lustre: 10025:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 10025:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 LustreError: 10518:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0001-osp-MDT0002: fail to cancel 1 llog-records: rc = -2 LustreError: 10518:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0001-osp-MDT0002: fail to cancel 1 of 1 llog-records: rc = -2 Lustre: 10025:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 512 < left 1324, rollback = 7 Lustre: 10025:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 10025:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 2/178/0 Lustre: 10025:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1315/1324/0, punch: 0/0/0, quota 3/3/0 Lustre: 10025:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 10025:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 10025:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 512 < left 1324, rollback = 7 Lustre: 10025:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 669 previous similar messages Lustre: 10025:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 10025:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 669 previous similar messages Lustre: 10025:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 2/178/0 Lustre: 10025:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 669 previous similar messages Lustre: 10025:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1315/1324/0, punch: 0/0/0, quota 3/3/0 Lustre: 10025:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 669 previous similar messages Lustre: 10025:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 10025:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 669 previous similar messages Lustre: 10025:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 10025:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 669 previous similar messages Lustre: 10025:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 512 < left 1324, rollback = 7 Lustre: 10025:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 120 previous similar messages Lustre: 10025:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 10025:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 121 previous similar messages Lustre: 10025:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 2/178/0 Lustre: 10025:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 121 previous similar messages Lustre: 10025:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1315/1324/0, punch: 0/0/0, quota 3/3/0 Lustre: 10025:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 121 previous similar messages Lustre: 10025:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 10025:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 121 previous similar messages Lustre: 10025:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 10025:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 121 previous similar messages Lustre: 10047:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 8: before 509 < left 4810, rollback = 9 Lustre: 10047:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 282/1128/0, destroy: 1/4/0 Lustre: 10047:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 521 previous similar messages Lustre: 10047:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 3/3/1, xattr_set: 286/148/0 Lustre: 10047:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 521 previous similar messages Lustre: 10047:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1411/12136/0, punch: 0/0/0, quota 6/6/0 Lustre: 10047:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 521 previous similar messages Lustre: 10047:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 283/4810/0, delete: 2/5/1 Lustre: 10047:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 521 previous similar messages Lustre: 10047:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 1/1/0, ref_del: 2/2/1 Lustre: 10047:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 521 previous similar messages Lustre: 10047:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 508 < left 12136, rollback = 7 Lustre: 10047:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 522 previous similar messages Lustre: 13883:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 501 < left 1400, rollback = 2 Lustre: 13854:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 512 < left 1119, rollback = 7 Lustre: 13854:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 2/8/0, destroy: 0/0/0 Lustre: 13854:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 19 previous similar messages Lustre: 13854:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 5/192/0 Lustre: 13854:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 19 previous similar messages Lustre: 13854:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1034/1119/0, punch: 0/0/0, quota 6/6/0 Lustre: 13854:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 19 previous similar messages Lustre: 13854:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 2/34/0, delete: 0/0/0 Lustre: 13854:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 19 previous similar messages Lustre: 13854:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 13854:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 19 previous similar messages Lustre: 10043:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 505 < left 699, rollback = 2 Lustre: 10043:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 17 previous similar messages Lustre: 15108:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 508 < left 40387, rollback = 7 Lustre: 15108:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 9553 previous similar messages Lustre: 15108:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 939/3756/0, destroy: 1/4/1 Lustre: 15108:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 9562 previous similar messages Lustre: 15108:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 3/3/1, xattr_set: 943/148/0 Lustre: 15108:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 9562 previous similar messages Lustre: 15108:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 4696/40387/1, punch: 0/0/0, quota 10/90/0 Lustre: 15108:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 9562 previous similar messages Lustre: 15108:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 940/15979/0, delete: 2/5/1 Lustre: 15108:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 9562 previous similar messages Lustre: 15108:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 1/1/0, ref_del: 2/2/1 Lustre: 15108:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 9562 previous similar messages ------------[ cut here ]------------ WARNING: CPU: 9 PID: 15108 at /home/green/git/lustre-release/ldiskfs/namei.c:3331 ldiskfs_orphan_add+0x11e/0x2a0 [ldiskfs] Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) dm_flakey dm_mod loop zfs(PO) zunicode(PO) zzstd(O) zlua(O) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) jbd2 mbcache crc_t10dif crct10dif_generic sb_edac edac_core iosf_mbi crc32_pclmul ghash_clmulni_intel aesni_intel lrw gf128mul glue_helper ablk_helper cryptd virtio_balloon pcspkr i2c_piix4 virtio_console ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm crct10dif_pclmul ata_piix drm crct10dif_common crc32c_intel drm_panel_orientation_quirks serio_raw libata virtio_blk i2c_core floppy [last unloaded: libcfs] CPU: 9 PID: 15108 Comm: mdt04_009 Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011 Call Trace: [<ffffffff817d9418>] dump_stack+0x19/0x1b [<ffffffff8108d538>] __warn+0xd8/0x100 [<ffffffff8108d67d>] warn_slowpath_null+0x1d/0x20 [<ffffffffa0d704ae>] ldiskfs_orphan_add+0x11e/0x2a0 [ldiskfs] [<ffffffffa0d619eb>] ldiskfs_xattr_inode_orphan_add.constprop.17+0xbb/0x110 [ldiskfs] [<ffffffffa0d64417>] ldiskfs_xattr_delete_inode+0x77/0x3c0 [ldiskfs] [<ffffffffa0d84aa0>] ? ldiskfs_evict_inode+0x190/0x670 [ldiskfs] [<ffffffffa0d84ad9>] ldiskfs_evict_inode+0x1c9/0x670 [ldiskfs] [<ffffffff812649af>] evict+0xaf/0x180 [<ffffffff81264de5>] iput+0xf5/0x180 [<ffffffffa0ca0d08>] osd_object_delete+0x1f8/0x370 [osd_ldiskfs] [<ffffffffa04da59d>] lu_object_free.isra.26+0xcd/0x1d0 [obdclass] [<ffffffffa04decf1>] lu_object_put+0xa1/0x480 [obdclass] [<ffffffffa0ee7440>] mdt_object_put+0x30/0x110 [mdt] [<ffffffffa0eeee18>] mdt_reint_unlink+0x918/0x19b0 [mdt] [<ffffffffa050079e>] ? lu_ucred+0x1e/0x30 [obdclass] [<ffffffffa0edeaf2>] ? ucred_set_audit_enabled.isra.13+0x22/0x60 [mdt] [<ffffffffa0ef6087>] mdt_reint_rec+0x87/0x240 [mdt] [<ffffffffa0ecb55c>] mdt_reint_internal+0x76c/0xb40 [mdt] [<ffffffffa0ed8657>] mdt_reint+0x67/0x150 [mdt] [<ffffffffa07e226e>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<ffffffffa03c645e>] ? libcfs_nid2str_r+0xfe/0x130 [lnet] [<ffffffffa078d990>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<ffffffffa078f529>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e32be>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffffa078e950>] ? ptlrpc_wait_event+0x620/0x620 [ptlrpc] [<ffffffff810ba0f4>] kthread+0xe4/0xf0 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 ---[ end trace b1a5eb572da447e9 ]--- Lustre: 13870:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 502 < left 671, rollback = 2 Lustre: 13870:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 7 previous similar messages Lustre: 13162:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 8: before 509 < left 15979, rollback = 9 Lustre: 13162:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 7 previous similar messages Lustre: ll_ost01_008: service thread pid 13964 was inactive for 62.111 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 13964, comm: ll_ost01_008 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] tgt_extent_lock+0xea/0x2a0 [ptlrpc] [<0>] ofd_getattr_hdl+0x365/0x720 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: ll_ost01_000: service thread pid 10715 was inactive for 63.137 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 10715, comm: ll_ost01_000 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] tgt_extent_lock+0xea/0x2a0 [ptlrpc] [<0>] ofd_getattr_hdl+0x365/0x720 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 13965, comm: ll_ost01_009 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] tgt_extent_lock+0xea/0x2a0 [ptlrpc] [<0>] ofd_getattr_hdl+0x365/0x720 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: 10572:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 510 < left 949, rollback = 7 Lustre: 10572:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 4700 previous similar messages Lustre: 10572:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 1/4/2, destroy: 0/0/0 Lustre: 10572:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 4708 previous similar messages Lustre: 10572:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 4/209/0 Lustre: 10572:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 4708 previous similar messages Lustre: 10572:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 940/949/0, punch: 0/0/0, quota 8/80/0 Lustre: 10572:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 4708 previous similar messages Lustre: 10572:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 2/33/0, delete: 0/0/0 Lustre: 10572:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 4708 previous similar messages Lustre: 10572:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 10572:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 4708 previous similar messages Lustre: dir [0x280000403:0x1f:0x0] stripe 1 readdir failed: -2, directory is partially accessed! LustreError: 10017:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0002_UUID lock: ffff8800947c6d00/0xdbfb804ebc8039ae lrc: 3/0,0 mode: PW/PW res: [0x340000401:0x3:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000400030020 nid: 0@lo remote: 0xdbfb804ebc8039a7 expref: 1424 pid: 10721 timeout: 10519 lvb_type: 0 LustreError: 15500:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802eb218040 x1709885388430400/t0(0) o105->lustre-OST0002@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 11-0: lustre-OST0000-osc-ffff8802a9804138: operation ost_setattr to node 0@lo failed: rc = -107 Lustre: lustre-OST0000-osc-ffff8802a9804138: Connection to lustre-OST0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0002-osc-ffff8802a9804138: This client was evicted by lustre-OST0002; in progress operations using this service will fail. LustreError: 14019:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802e553bec0 x1709885388437056/t0(0) o105->lustre-OST0000@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 14019:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 2 previous similar messages Lustre: 10028:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 8: before 509 < left 21878, rollback = 9 LustreError: 167-0: lustre-OST0000-osc-ffff8802a9804138: This client was evicted by lustre-OST0000; in progress operations using this service will fail. LustreError: 10753:0:(tgt_grant.c:758:tgt_grant_check()) lustre-OST0000: cli 379425da-f4d8-45a0-a674-9fb5206d5eaf claims 61440 GRANT, real grant 0 LustreError: 10753:0:(ldlm_lib.c:3481:target_bulk_io()) @@@ bulk WRITE failed: rc = -107 req@ffff88028754d180 x1709885388197312/t0(0) o4->379425da-f4d8-45a0-a674-9fb5206d5eaf@0@lo:93/0 lens 488/448 e 0 to 0 dl 1630674008 ref 1 fl Interpret:/0/0 rc 0/0 job:'cp.0' Lustre: lustre-OST0000: Bulk IO write error with 379425da-f4d8-45a0-a674-9fb5206d5eaf (at 0@lo), client will retry: rc = -107 Lustre: 6823:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.110@tcp:/lustre/fid: [0x280000404:0x48:0x0]// may get corrupted (rc -5) Lustre: 6823:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.110@tcp:/lustre/fid: [0x240000404:0x5d:0x0]/ may get corrupted (rc -108) Lustre: 6830:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.110@tcp:/lustre/fid: [0x280000404:0x63:0x0]/ may get corrupted (rc -108) LustreError: 10017:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0001_UUID lock: ffff8802cfe95680/0xdbfb804ebc80bedf lrc: 3/0,0 mode: PW/PW res: [0x300000400:0x4:0x0].0x0 rrc: 4 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000400020020 nid: 0@lo remote: 0xdbfb804ebc80be99 expref: 1598 pid: 13765 timeout: 10526 lvb_type: 0 LustreError: 10017:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 1 previous similar message LustreError: 11-0: lustre-OST0001-osc-ffff8802a9804138: operation ost_setattr to node 0@lo failed: rc = -107 LustreError: Skipped 35 previous similar messages Lustre: lustre-OST0001-osc-ffff8802a9804138: Connection to lustre-OST0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 1 previous similar message LustreError: 16356:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802c2272c00 x1709885388494912/t0(0) o104->lustre-OST0001@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 11-0: lustre-OST0001-osc-ffff8802a9804138: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 44 previous similar messages LustreError: 167-0: lustre-OST0001-osc-ffff8802a9804138: This client was evicted by lustre-OST0001; in progress operations using this service will fail. LustreError: 13964:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff880082c39300 x1709885388543168/t0(0) o104->lustre-OST0001@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 16374:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0002-osc-ffff8802a9804138: namespace resource [0x340000400:0x10:0x0].0x0 (ffff88009229b9c0) refcount nonzero (1) after lock cleanup; forcing cleanup. Lustre: 6834:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.110@tcp:/lustre/fid: [0x280000404:0x2:0x0]/ may get corrupted (rc -108) LustreError: 16374:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0002-osc-ffff8802a9804138: namespace resource [0x340000400:0x63:0x0].0x0 (ffff88009fda02c0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 16374:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 311 previous similar messages Lustre: lustre-OST0002-osc-ffff8802a9804138: Connection restored to 192.168.123.110@tcp (at 0@lo) Lustre: lustre-OST0000-osc-ffff8802a9804138: Connection restored to 192.168.123.110@tcp (at 0@lo) LustreError: 16499:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0001-osc-ffff8802a9804138: namespace resource [0x300000400:0x80:0x0].0x0 (ffff8802c10c2340) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 16499:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 7 previous similar messages Lustre: lustre-OST0001-osc-ffff8802a9804138: Connection restored to 192.168.123.110@tcp (at 0@lo) Lustre: 14898:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 8: before 509 < left 6374, rollback = 9 Lustre: 14898:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 21 previous similar messages Lustre: 14898:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 374/1496/0, destroy: 1/4/0 Lustre: 14898:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 28757 previous similar messages Lustre: 14898:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 3/3/1, xattr_set: 378/148/0 Lustre: 14898:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 28757 previous similar messages Lustre: 14898:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1871/16092/0, punch: 0/0/0, quota 10/74/0 Lustre: 14898:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 28757 previous similar messages Lustre: 14898:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 375/6374/0, delete: 2/5/1 Lustre: 14898:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 28757 previous similar messages Lustre: 14898:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 1/1/0, ref_del: 2/2/1 Lustre: 14898:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 28757 previous similar messages Lustre: 14898:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 508 < left 16092, rollback = 7 Lustre: 14898:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 28735 previous similar messages LustreError: 17274:0:(lustre_lmv.h:438:lmv_is_sane()) insane LMV: magic=0xcd40cd0 count=2 index=1 hash=crush:0x82000003 version=0 migrate offset=1 migrate hash=crush:3. LustreError: 10017:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0000_UUID lock: ffff8802e1b9b100/0xdbfb804ebc8538a8 lrc: 3/0,0 mode: PW/PW res: [0x2c0000400:0x2:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->655359) gid 0 flags: 0x60000480020020 nid: 0@lo remote: 0xdbfb804ebc8529c1 expref: 2152 pid: 15419 timeout: 10590 lvb_type: 0 LustreError: 10728:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88025b0e5180 x1709885390566720/t0(0) o106->lustre-OST0000@0@lo:15/16 lens 328/280 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 10728:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 1 previous similar message LustreError: 11-0: lustre-OST0000-osc-ffff88009baa1bf8: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 46 previous similar messages Lustre: lustre-OST0000-osc-ffff88009baa1bf8: Connection to lustre-OST0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0000-osc-ffff88009baa1bf8: This client was evicted by lustre-OST0000; in progress operations using this service will fail. Lustre: 6828:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff880255300680 x1709885390579968/t0(0) o10->lustre-OST0003-osc-ffff8802a9804138@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630674130 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'cp.0' Lustre: 6832:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.110@tcp:/lustre/fid: [0x240000403:0x114:0x0]// may get corrupted (rc -108) LustreError: 17987:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0000-osc-ffff88009baa1bf8: namespace resource [0x230:0x0:0x0].0x0 (ffff8802a63f79c0) refcount nonzero (3) after lock cleanup; forcing cleanup. LustreError: 17987:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 175 previous similar messages Lustre: 10028:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 8: before 512 < left 1104, rollback = 9 Lustre: 10028:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 42 previous similar messages Lustre: lustre-OST0000-osc-ffff88009baa1bf8: Connection restored to 192.168.123.110@tcp (at 0@lo) Lustre: ll_ost05_016: service thread pid 16367 was inactive for 62.204 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. LustreError: 13869:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0000: '6' migration was interrupted, run 'lfs migrate -m 1 -c 1 -H crush 6' to finish migration. LustreError: 10017:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0000_UUID lock: ffff8802e55ba5c0/0xdbfb804ebc898354 lrc: 3/0,0 mode: PW/PW res: [0x229:0x0:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->36863) gid 0 flags: 0x60000000020020 nid: 0@lo remote: 0xdbfb804ebc89834d expref: 1623 pid: 15531 timeout: 10659 lvb_type: 0 LustreError: 15658:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630674143 with bad export cookie 15851404389119394571 LustreError: 14047:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802f24f2c00 x1709885393174400/t0(0) o104->lustre-OST0000@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 14047:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 12 previous similar messages LustreError: 14301:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630674143 with bad export cookie 15851404389119394571 LustreError: 14301:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) Skipped 2 previous similar messages Lustre: lustre-OST0000-osc-ffff8802a9804138: Connection to lustre-OST0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0000-osc-ffff8802a9804138: This client was evicted by lustre-OST0000; in progress operations using this service will fail. Lustre: 6834:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.110@tcp:/lustre/fid: [0x280000404:0xa7:0x0]/ may get corrupted (rc -108) Lustre: 6835:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.110@tcp:/lustre/fid: [0x240000403:0x114:0x0]// may get corrupted (rc -108) LustreError: 19090:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0000-osc-ffff8802a9804138: namespace resource [0x13b:0x0:0x0].0x0 (ffff8802d8132840) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 19090:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 411 previous similar messages Lustre: lustre-OST0000-osc-ffff8802a9804138: Connection restored to 192.168.123.110@tcp (at 0@lo) Lustre: 13870:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 495 < left 760, rollback = 2 Lustre: 13870:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 9 previous similar messages LustreError: 10017:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0001_UUID lock: ffff8802dd9e7840/0xdbfb804ebc8a72e3 lrc: 3/0,0 mode: PW/PW res: [0x300000401:0x4f:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000080030020 nid: 0@lo remote: 0xdbfb804ebc8a725e expref: 1402 pid: 14131 timeout: 10684 lvb_type: 0 LustreError: 15442:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88020b292c00 x1709885393953728/t0(0) o104->lustre-OST0001@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 15442:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 34 previous similar messages LustreError: 10007:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630674168 with bad export cookie 15851404389118883403 Lustre: lustre-OST0003-osc-ffff8802a9804138: Connection to lustre-OST0003 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0003-osc-ffff8802a9804138: This client was evicted by lustre-OST0003; in progress operations using this service will fail. LustreError: 11-0: lustre-OST0001-osc-ffff8802a9804138: operation ost_setattr to node 0@lo failed: rc = -107 LustreError: Skipped 1 previous similar message Lustre: 6829:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.110@tcp:/lustre/fid: [0x280000404:0x115:0x0]// may get corrupted (rc -108) Lustre: 6837:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.110@tcp:/lustre/fid: [0x200000403:0xb9:0x0]// may get corrupted (rc -108) LustreError: 19616:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0003-osc-ffff8802a9804138: namespace resource [0x380000401:0xd23:0x0].0x0 (ffff880099812840) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 19616:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 1 previous similar message Lustre: lustre-OST0003-osc-ffff8802a9804138: Connection restored to 192.168.123.110@tcp (at 0@lo) Lustre: 13162:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 511 < left 1963, rollback = 7 Lustre: 13162:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 27948 previous similar messages Lustre: 13162:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 1/4/0, destroy: 0/0/0 Lustre: 13162:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 28013 previous similar messages Lustre: 13162:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 4/209/0 Lustre: 13162:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 28013 previous similar messages Lustre: 13162:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1954/1963/0, punch: 0/0/0, quota 8/80/0 Lustre: 13162:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 28013 previous similar messages Lustre: 13162:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 2/33/1, delete: 0/0/0 Lustre: 13162:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 28013 previous similar messages Lustre: 13162:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 13162:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 28013 previous similar messages Lustre: mdt03_008: service thread pid 13996 was inactive for 62.071 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. LustreError: 19968:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0002: failed to get lu_attr of [0x280000404:0x9:0x0]: rc = -2 Lustre: mdt06_001: service thread pid 10043 was inactive for 66.179 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. LustreError: 13996:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0002: '0' migration was interrupted, run 'lfs migrate -m 0 -c 1 -H crush 0' to finish migration. Lustre: 9988:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 6s req@ffff880289678040 x1709885396114432/t0(0) o104->42274881-cfc4-4ff5-935b-f0524dbb475a@0@lo:0/0 lens 328/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'' LustreError: 10017:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0001_UUID lock: ffff88025747bc40/0xdbfb804ebc8e9a73 lrc: 3/0,0 mode: PW/PW res: [0x300000400:0x552:0x0].0x0 rrc: 4 type: EXT [0->18446744073709551615] (req 0->3145727) gid 0 flags: 0x60000400020020 nid: 0@lo remote: 0xdbfb804ebc8e9a6c expref: 2667 pid: 10722 timeout: 10781 lvb_type: 0 LustreError: 10017:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 1 previous similar message LustreError: 14019:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88023f0c6a80 x1709885397161984/t0(0) o105->lustre-OST0001@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 14019:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 7 previous similar messages LustreError: 11-0: lustre-OST0001-osc-ffff88009baa1bf8: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-OST0001-osc-ffff88009baa1bf8: Connection to lustre-OST0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 1 previous similar message LustreError: 167-0: lustre-OST0001-osc-ffff88009baa1bf8: This client was evicted by lustre-OST0001; in progress operations using this service will fail. LustreError: Skipped 1 previous similar message Lustre: 6823:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.110@tcp:/lustre/fid: [0x240000403:0x149:0x0]// may get corrupted (rc -108) Lustre: 6822:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.110@tcp:/lustre/fid: [0x280000403:0xce:0x0]/ may get corrupted (rc -108) Lustre: 6822:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.110@tcp:/lustre/fid: [0x200000403:0xfd:0x0]// may get corrupted (rc -108) Lustre: 6837:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.110@tcp:/lustre/fid: [0x200000403:0x109:0x0]// may get corrupted (rc -5) Lustre: 6823:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.110@tcp:/lustre/fid: [0x240000403:0x1bd:0x0]/ may get corrupted (rc -108) LustreError: 21179:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0001-osc-ffff88009baa1bf8: namespace resource [0x300000400:0x560:0x0].0x0 (ffff880240fd02c0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 21179:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 91 previous similar messages Lustre: lustre-OST0001-osc-ffff88009baa1bf8: Connection restored to 192.168.123.110@tcp (at 0@lo) Lustre: Skipped 1 previous similar message Lustre: 10030:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 500 < left 699, rollback = 2 Lustre: 10030:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 43 previous similar messages Lustre: 6837:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff88022d8c5e00 x1709885398798528/t0(0) o10->lustre-OST0002-osc-ffff88009baa1bf8@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630674379 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dir_create.sh.0' LustreError: 10017:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0002_UUID lock: ffff8802d8fac780/0xdbfb804ebc91117d lrc: 3/0,0 mode: PW/PW res: [0x340000400:0x5b3:0x0].0x0 rrc: 4 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000400020020 nid: 0@lo remote: 0xdbfb804ebc91116f expref: 1483 pid: 15114 timeout: 10857 lvb_type: 0 LustreError: 10017:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 1 previous similar message LustreError: 11-0: lustre-OST0002-osc-ffff8802a9804138: operation ost_setattr to node 0@lo failed: rc = -107 LustreError: Skipped 2 previous similar messages Lustre: lustre-OST0002-osc-ffff8802a9804138: Connection to lustre-OST0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 1 previous similar message LustreError: 14114:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802e6d93240 x1709885399733504/t0(0) o106->lustre-OST0002@0@lo:15/16 lens 328/280 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 14114:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 6 previous similar messages LustreError: 167-0: lustre-OST0002-osc-ffff8802a9804138: This client was evicted by lustre-OST0002; in progress operations using this service will fail. LustreError: Skipped 1 previous similar message Lustre: 6829:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.110@tcp:/lustre/fid: [0x280000404:0x16d:0x0]/ may get corrupted (rc -5) Lustre: 6828:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.110@tcp:/lustre/fid: [0x280000404:0x182:0x0]// may get corrupted (rc -108) Lustre: 6829:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.110@tcp:/lustre/fid: [0x240000404:0xb:0x0]/ may get corrupted (rc -108) LustreError: 22707:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0002-osc-ffff8802a9804138: namespace resource [0x340000401:0xa9f:0x0].0x0 (ffff880265a01940) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 22707:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 278 previous similar messages Lustre: lustre-OST0002-osc-ffff8802a9804138: Connection restored to 192.168.123.110@tcp (at 0@lo) Lustre: Skipped 1 previous similar message LustreError: 14015:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000403:0xd3:0x0]: rc = -2 Lustre: 6836:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff88009a5a4500 x1709885402705472/t0(0) o10->lustre-OST0003-osc-ffff88009baa1bf8@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630674494 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dir_create.sh.0' Lustre: 14967:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 506 < left 13549, rollback = 7 Lustre: 14967:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 107863 previous similar messages Lustre: 14967:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 311/1244/0, destroy: 1/4/1 Lustre: 14967:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 108256 previous similar messages Lustre: 14967:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 311/28/4 Lustre: 14967:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 108255 previous similar messages Lustre: 14967:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1554/13549/0, punch: 0/0/0, quota 10/90/0 Lustre: 14967:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 108256 previous similar messages Lustre: 14967:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 313/5319/0, delete: 2/5/1 Lustre: 14967:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 108256 previous similar messages Lustre: 14967:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 1/1/0 Lustre: 14967:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 108256 previous similar messages Lustre: 12631:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0x194:0x0] with magic=0xbd60bd0 Lustre: 21362:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0x10d:0x0] with magic=0xbd60bd0 Lustre: 21362:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 1 previous similar message Lustre: 15668:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 6s req@ffff8800a4e59c48 x1709885404320512/t0(0) o104->LOV_OSC_UUID@0@lo:0/0 lens 328/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'' Lustre: 13869:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0x1a8:0x0] with magic=0xbd60bd0 Lustre: 13869:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 3 previous similar messages Lustre: 10014:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 6s req@ffff88029cf63838 x1709885404413632/t0(0) o103->42274881-cfc4-4ff5-935b-f0524dbb475a@0@lo:0/0 lens 328/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'ldlm_bl_04.0' Lustre: dir [0x240000403:0x1c2:0x0] stripe 1 readdir failed: -2, directory is partially accessed! LustreError: 18821:0:(lustre_lmv.h:438:lmv_is_sane()) insane LMV: magic=0xcd40cd0 count=2 index=1 hash=crush:0x82000003 version=0 migrate offset=1 migrate hash=crush:3. LustreError: 10017:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0000_UUID lock: ffff8802dc1aa5c0/0xdbfb804ebc95fe3a lrc: 3/0,0 mode: PW/PW res: [0x2c0000401:0x4b:0x0].0x0 rrc: 4 type: EXT [0->18446744073709551615] (req 0->917503) gid 0 flags: 0x60000480020020 nid: 0@lo remote: 0xdbfb804ebc95fe33 expref: 2023 pid: 15114 timeout: 11013 lvb_type: 0 LustreError: 11-0: lustre-OST0000-osc-ffff88009baa1bf8: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-OST0000-osc-ffff88009baa1bf8: Connection to lustre-OST0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 15900:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88022a163880 x1709885404847936/t0(0) o104->lustre-OST0000@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 15900:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 3 previous similar messages LustreError: 167-0: lustre-OST0000-osc-ffff88009baa1bf8: This client was evicted by lustre-OST0000; in progress operations using this service will fail. Lustre: 6831:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.110@tcp:/lustre/fid: [0x200000403:0x1e1:0x0]/ may get corrupted (rc -108) Lustre: 6830:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.110@tcp:/lustre/fid: [0x240000404:0x1f5:0x0]// may get corrupted (rc -108) Lustre: 6831:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.110@tcp:/lustre/fid: [0x280000403:0x145:0x0]// may get corrupted (rc -108) Lustre: 10041:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000404:0x17e:0x0] with magic=0xbd60bd0 Lustre: 10041:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 3 previous similar messages Lustre: lustre-OST0000-osc-ffff88009baa1bf8: Connection restored to 192.168.123.110@tcp (at 0@lo) LustreError: 9991:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630674551 with bad export cookie 15851404389119832141 Lustre: 6833:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.110@tcp:/lustre/fid: [0x200000404:0x208:0x0]// may get corrupted (rc -108) Lustre: 6832:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.110@tcp:/lustre/fid: [0x240000403:0x14f:0x0]// may get corrupted (rc -108) LustreError: 26232:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0000-osc-ffff8802a9804138: namespace resource [0xc35:0x0:0x0].0x0 (ffff8802dd9f0a40) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 26232:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 305 previous similar messages Lustre: 15415:0:(service.c:1378:ptlrpc_at_send_early_reply()) @@@ Already past deadline (-1s), not sending early reply. Consider increasing at_early_margin (5)? req@ffff88026b35f700 x1709885406535680/t4294995242(0) o2->42274881-cfc4-4ff5-935b-f0524dbb475a@0@lo:660/0 lens 440/432 e 0 to 0 dl 1630674575 ref 1 fl Complete:/0/0 rc 0/0 job:'chmod.0' Lustre: 12986:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 8: before 509 < left 29868, rollback = 9 Lustre: 12986:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 1269 previous similar messages LustreError: 26250:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff8802a9804138: inode [0x280000404:0x1f8:0x0] mdc close failed: rc = -13 10[26352]: segfault at 0 ip (null) sp 00007ffefeda8748 error 14 in 10[400000+6000] Lustre: 6836:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff88020ad75180 x1709885411053568/t0(0) o10->lustre-OST0000-osc-ffff88009baa1bf8@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630674735 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'cp.0' LustreError: 20095:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0001: failed to get lu_attr of [0x240000403:0x296:0x0]: rc = -2 Lustre: 15931:0:(service.c:1378:ptlrpc_at_send_early_reply()) @@@ Already past deadline (-1s), not sending early reply. Consider increasing at_early_margin (5)? req@ffff880096ee9300 x1709885410915136/t0(0) o1->379425da-f4d8-45a0-a674-9fb5206d5eaf@0@lo:29/0 lens 440/432 e 0 to 0 dl 1630674699 ref 1 fl Complete:/0/0 rc 0/0 job:'lfs.0' Lustre: 15931:0:(service.c:1378:ptlrpc_at_send_early_reply()) Skipped 34 previous similar messages Lustre: 16410:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (25/2s); client may timeout req@ffff8802db38f0c0 x1709885410913152/t0(0) o1->379425da-f4d8-45a0-a674-9fb5206d5eaf@0@lo:29/0 lens 440/432 e 0 to 0 dl 1630674699 ref 1 fl Complete:/0/0 rc 0/0 job:'lfs.0' Lustre: 27776:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (25/3s); client may timeout req@ffff8800833d57c0 x1709885410905216/t0(0) o1->379425da-f4d8-45a0-a674-9fb5206d5eaf@0@lo:29/0 lens 440/432 e 0 to 0 dl 1630674699 ref 1 fl Complete:/0/0 rc 0/0 job:'lfs.0' Lustre: 27776:0:(service.c:2328:ptlrpc_server_handle_request()) Skipped 1 previous similar message LustreError: 10017:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0003_UUID lock: ffff8802e7cf0b80/0xdbfb804ebc9ebe41 lrc: 3/0,0 mode: PW/PW res: [0x380000401:0x51:0x0].0x0 rrc: 4 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000480000020 nid: 0@lo remote: 0xdbfb804ebc9ebe33 expref: 3650 pid: 18869 timeout: 11219 lvb_type: 0 Lustre: 26319:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (25/4s); client may timeout req@ffff8802cc816440 x1709885410904704/t0(0) o1->379425da-f4d8-45a0-a674-9fb5206d5eaf@0@lo:29/0 lens 440/432 e 0 to 0 dl 1630674699 ref 1 fl Complete:/0/0 rc 0/0 job:'lfs.0' Lustre: 26319:0:(service.c:2328:ptlrpc_server_handle_request()) Skipped 8 previous similar messages LustreError: 10017:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 1 previous similar message LustreError: 14309:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630674703 with bad export cookie 15851404389118887603 LustreError: 11-0: lustre-OST0003-osc-ffff88009baa1bf8: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 2 previous similar messages Lustre: lustre-OST0003-osc-ffff88009baa1bf8: Connection to lustre-OST0003 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 1 previous similar message LustreError: 167-0: lustre-OST0003-osc-ffff88009baa1bf8: This client was evicted by lustre-OST0003; in progress operations using this service will fail. LustreError: Skipped 1 previous similar message Lustre: 6830:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.110@tcp:/lustre/fid: [0x200000403:0x223:0x0]/ may get corrupted (rc -108) Lustre: 6831:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.110@tcp:/lustre/fid: [0x200000404:0x244:0x0]// may get corrupted (rc -108) LustreError: 21510:0:(ofd_obd.c:501:ofd_destroy_export()) lustre-OST0003: cli 42274881-cfc4-4ff5-935b-f0524dbb475a/ffff8800a914e678 has 90112 pending on destroyed export LustreError: 21510:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_destroy_export: tot_granted 97121856 != fo_tot_granted 97211968 LustreError: 21510:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_destroy_export: tot_pending 122880 != fo_tot_pending 212992 Lustre: 6831:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff880269a1ea80 x1709885412621888/t0(0) o10->lustre-OST0003-osc-ffff8802a9804138@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630674782 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'cp.0' LustreError: 10747:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 97121856 != fo_tot_granted 97211968 LustreError: 10747:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 122880 != fo_tot_pending 212992 LustreError: 14331:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630674719 with bad export cookie 15851404389120369244 LustreError: 14331:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) Skipped 7 previous similar messages LustreError: 24954:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 105559616 != fo_tot_granted 105649728 LustreError: 24954:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 122880 != fo_tot_pending 212992 LustreError: 10746:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 105559616 != fo_tot_granted 105649728 LustreError: 10746:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 122880 != fo_tot_pending 212992 LustreError: 28268:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0003-osc-ffff88009baa1bf8: namespace resource [0x380000400:0x93e:0x0].0x0 (ffff8802d8e5b740) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 28268:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 185 previous similar messages Lustre: lustre-OST0003-osc-ffff88009baa1bf8: Connection restored to 192.168.123.110@tcp (at 0@lo) Lustre: Skipped 1 previous similar message Lustre: 6836:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.110@tcp:/lustre/fid: [0x280000404:0x266:0x0]/ may get corrupted (rc -108) LustreError: 10746:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 113997376 != fo_tot_granted 114087488 LustreError: 10746:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 1 previous similar message LustreError: 10746:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 122880 != fo_tot_pending 212992 LustreError: 10746:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 1 previous similar message LustreError: 10737:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 159901248 != fo_tot_granted 159991360 LustreError: 10737:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 2 previous similar messages LustreError: 10737:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 123776 != fo_tot_pending 213888 LustreError: 10737:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 2 previous similar messages LustreError: 10744:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 181528640 != fo_tot_granted 181618752 LustreError: 10744:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 9 previous similar messages LustreError: 10744:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 122880 != fo_tot_pending 212992 LustreError: 10744:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 9 previous similar messages Lustre: 6831:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff880269a1ea80 x1709885414195968/t0(0) o10->lustre-OST0003-osc-ffff8802a9804138@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630674829 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'cp.0' Lustre: 13131:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000404:0x27b:0x0] with magic=0xbd60bd0 Lustre: 13131:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 1 previous similar message LustreError: 12756:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0000: '0' migration was interrupted, run 'lfs migrate -m 2 -c 1 -H crush 0' to finish migration. LustreError: 10745:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 223508544 != fo_tot_granted 223598656 LustreError: 10745:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 17 previous similar messages LustreError: 10745:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 122880 != fo_tot_pending 212992 LustreError: 10745:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 17 previous similar messages Lustre: 22006:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000404:0x298:0x0] with magic=0xbd60bd0 Lustre: 22006:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 3 previous similar messages LustreError: 13834:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0000: '0' migration was interrupted, run 'lfs migrate -m 2 -c 1 -H crush 0' to finish migration. LustreError: 11807:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 260146496 != fo_tot_granted 260236608 LustreError: 11807:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 17 previous similar messages LustreError: 11807:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 122880 != fo_tot_pending 212992 LustreError: 11807:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 17 previous similar messages LustreError: 10017:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0001_UUID lock: ffff8802d8e05680/0xdbfb804ebca6c77d lrc: 3/0,0 mode: PW/PW res: [0x300000400:0x913:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000000020020 nid: 0@lo remote: 0xdbfb804ebca6c737 expref: 1824 pid: 27782 timeout: 11380 lvb_type: 0 LustreError: 10017:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 2 previous similar messages LustreError: 10016:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802732357c0 x1709885417436032/t0(0) o105->lustre-OST0001@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 14073:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630674864 with bad export cookie 15851404389120148415 LustreError: 11-0: lustre-OST0001-osc-ffff88009baa1bf8: operation ost_setattr to node 0@lo failed: rc = -107 LustreError: Skipped 5 previous similar messages Lustre: lustre-OST0001-osc-ffff88009baa1bf8: Connection to lustre-OST0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 2 previous similar messages LustreError: 10016:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 51 previous similar messages Lustre: 6832:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.110@tcp:/lustre/fid: [0x240000404:0x2e9:0x0]/ may get corrupted (rc -108) Lustre: 6834:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff8800a26bf700 x1709885418381312/t0(0) o10->lustre-OST0000-osc-ffff88009baa1bf8@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630674962 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dd.0' Lustre: 6834:0:(client.c:1485:after_reply()) Skipped 1 previous similar message LustreError: 13935:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630674922 with bad export cookie 15851404389119887791 LustreError: 13935:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) Skipped 3 previous similar messages LustreError: 20975:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0002: '18' migration was interrupted, run 'lfs migrate -m 0 -c 1 -H crush 18' to finish migration. LustreError: 18609:0:(ofd_obd.c:501:ofd_destroy_export()) lustre-OST0003: cli 379425da-f4d8-45a0-a674-9fb5206d5eaf/ffff88006b020008 has 122880 pending on destroyed export Lustre: 6829:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.110@tcp:/lustre/fid: [0x240000404:0x450:0x0]/ may get corrupted (rc -108) Lustre: 10760:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 6s req@ffff8800852cd7c0 x1709885420004032/t0(0) o10->42274881-cfc4-4ff5-935b-f0524dbb475a@0@lo:0/0 lens 440/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'truncate.0' Lustre: 10760:0:(service.c:2157:ptlrpc_server_handle_req_in()) Skipped 1 previous similar message Lustre: ost_io: This server is not able to keep up with request traffic (cpu-bound). Lustre: 17809:0:(service.c:1612:ptlrpc_at_check_timed()) earlyQ=1 reqQ=0 recA=0, svcEst=33, delay=0ms Lustre: 17809:0:(service.c:1378:ptlrpc_at_send_early_reply()) @@@ Already past deadline (-5s), not sending early reply. Consider increasing at_early_margin (5)? req@ffff8800852cd7c0 x1709885420004032/t0(0) o10->42274881-cfc4-4ff5-935b-f0524dbb475a@0@lo:274/0 lens 440/0 e 0 to 0 dl 1630674944 ref 2 fl New:/0/ffffffff rc 0/-1 job:'truncate.0' Lustre: 17809:0:(service.c:1378:ptlrpc_at_send_early_reply()) Skipped 213 previous similar messages LustreError: 17809:0:(service.c:2290:ptlrpc_server_handle_request()) @@@ Dropping timed-out request from 12345-0@lo: deadline 8/5s ago req@ffff8800852cd7c0 x1709885420004032/t0(0) o10->42274881-cfc4-4ff5-935b-f0524dbb475a@0@lo:274/0 lens 440/0 e 0 to 0 dl 1630674944 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'truncate.0' Lustre: 17809:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (8/5s); client may timeout req@ffff8800852cd7c0 x1709885420004032/t0(0) o10->42274881-cfc4-4ff5-935b-f0524dbb475a@0@lo:274/0 lens 440/0 e 0 to 0 dl 1630674944 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'truncate.0' Lustre: 17809:0:(service.c:2328:ptlrpc_server_handle_request()) Skipped 3 previous similar messages Lustre: 24736:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 3/12/1, destroy: 0/0/0 Lustre: 24736:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 206545 previous similar messages Lustre: 24736:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 13/1205/0 Lustre: 24736:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 206544 previous similar messages Lustre: 24736:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 5/136/0, punch: 0/0/0, quota 8/56/0 Lustre: 24736:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 206545 previous similar messages Lustre: 24736:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 10/203/1, delete: 0/0/0 Lustre: 24736:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 206543 previous similar messages Lustre: 24736:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 4/4/0, ref_del: 0/0/0 Lustre: 24736:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 206545 previous similar messages Lustre: 13465:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 509 < left 12824, rollback = 7 Lustre: 13465:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 204976 previous similar messages Lustre: 6823:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff8802578ec500 x1709885420976256/t0(0) o10->lustre-OST0001-osc-ffff8802a9804138@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630675031 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dir_create.sh.0' Lustre: 6823:0:(client.c:1485:after_reply()) Skipped 1 previous similar message Lustre: dir [0x200000403:0x248:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: lustre-OST0003: haven't heard from client 379425da-f4d8-45a0-a674-9fb5206d5eaf (at 0@lo) in 47 seconds. I think it's dead, and I am evicting it. exp ffff8802295aa548, cur 1630674969 expire 1630674939 last 1630674922 LustreError: 10742:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 44451264 != fo_tot_granted 44664256 LustreError: 10742:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 49 previous similar messages LustreError: 10742:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 212992 LustreError: 10742:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 49 previous similar messages Lustre: 6827:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1630674936/real 1630674936] req@ffff8802d56e9300 x1709885420004032/t0(0) o10->lustre-OST0001-osc-ffff88009baa1bf8@0@lo:6/4 lens 440/432 e 0 to 1 dl 1630674961 ref 1 fl Rpc:XQr/0/ffffffff rc 0/-1 job:'truncate.0' Lustre: lustre-OST0001: Client 42274881-cfc4-4ff5-935b-f0524dbb475a (at 0@lo) reconnecting Lustre: lustre-OST0001-osc-ffff88009baa1bf8: Connection restored to 192.168.123.110@tcp (at 0@lo) Lustre: Skipped 3 previous similar messages LustreError: 167-0: lustre-OST0003-osc-ffff8802a9804138: This client was evicted by lustre-OST0003; in progress operations using this service will fail. LustreError: Skipped 4 previous similar messages LustreError: 1352:0:(llite_lib.c:1674:ll_update_lsm_md()) lustre: [0x200000403:0x31a:0x0] dir layout mismatch: LustreError: 1352:0:(lustre_lmv.h:134:lsm_md_dump()) magic 0xcd20cd0 stripe count 1 master mdt 0 hash type crush:0x2000003 max inherit 0 version 1 migrate offset 0 migrate hash 0x0 pool LustreError: 1352:0:(lustre_lmv.h:141:lsm_md_dump()) stripe[0] [0x200000400:0x14:0x0] LustreError: 1352:0:(lustre_lmv.h:134:lsm_md_dump()) magic 0xcd20cd0 stripe count 2 master mdt 0 hash type crush:0x82000003 max inherit 0 version 0 migrate offset 1 migrate hash 0x3 pool LustreError: 1352:0:(lustre_lmv.h:141:lsm_md_dump()) stripe[0] [0x200000400:0x14:0x0] LustreError: 1352:0:(lustre_lmv.h:141:lsm_md_dump()) stripe[1] [0x280000403:0x19b:0x0] LustreError: 28156:0:(ofd_obd.c:501:ofd_destroy_export()) lustre-OST0002: cli 42274881-cfc4-4ff5-935b-f0524dbb475a/ffff8802f45b12a8 has 28672 pending on destroyed export Lustre: 6831:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.110@tcp:/lustre/fid: [0x200000404:0x45e:0x0]// may get corrupted (rc -108) Lustre: 6831:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.110@tcp:/lustre/fid: [0x280000403:0x33b:0x0]// may get corrupted (rc -108) Lustre: 6830:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.110@tcp:/lustre/fid: [0x200000404:0x468:0x0]// may get corrupted (rc -108) Lustre: 6830:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.110@tcp:/lustre/fid: [0x240000403:0x410:0x0]// may get corrupted (rc -108) LustreError: 3199:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0002-osc-ffff88009baa1bf8: namespace resource [0x1b40:0x0:0x0].0x0 (ffff8802558216c0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 3199:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 314 previous similar messages Lustre: lustre-OST0002: haven't heard from client 42274881-cfc4-4ff5-935b-f0524dbb475a (at 0@lo) in 47 seconds. I think it's dead, and I am evicting it. exp ffff8802339cdd28, cur 1630675092 expire 1630675062 last 1630675045 LustreError: 10039:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0000: '0' migration was interrupted, run 'lfs migrate -m 2 -c 1 -H crush 0' to finish migration. Lustre: 13794:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 507 < left 699, rollback = 2 Lustre: 13794:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 773 previous similar messages LustreError: 10056:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0002: failed to get lu_attr of [0x280000404:0x2e3:0x0]: rc = -2 Lustre: 11479:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 8s req@ffff8802af1eac00 x1709885426144768/t0(0) o101->lustre-MDT0001-mdtlov_UUID@0@lo:0/0 lens 328/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'mdt04_012.0' LustreError: 11-0: lustre-OST0002-osc-ffff88009baa1bf8: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 3 previous similar messages Lustre: lustre-OST0002-osc-ffff88009baa1bf8: Connection to lustre-OST0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 5 previous similar messages Lustre: 14366:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 6s req@ffff8802b10f5e00 x1709885426916992/t0(0) o104->LOV_OSC_UUID@0@lo:0/0 lens 328/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'' Lustre: 14366:0:(service.c:2157:ptlrpc_server_handle_req_in()) Skipped 1 previous similar message ptlrpc_watchdog_fire: 3 callbacks suppressed Lustre: mdt03_001: service thread pid 10034 was inactive for 70.075 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Lustre: Skipped 1 previous similar message Pid: 10034, comm: mdt03_001 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] mdt_object_local_lock+0x527/0xb90 [mdt] [<0>] mdt_object_lock_internal+0x70/0x390 [mdt] [<0>] mdt_reint_object_lock+0x2c/0x60 [mdt] [<0>] mdt_reint_striped_lock+0x89/0x5b0 [mdt] [<0>] mdt_migrate_object_lock+0x139/0x162 [mdt] [<0>] mdt_reint_migrate+0xb47/0x11d0 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x76c/0xb40 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: mdt03_010: service thread pid 17300 was inactive for 70.111 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 17300, comm: mdt03_010 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] mdt_rename_lock+0x24b/0x4d0 [mdt] [<0>] mdt_reint_migrate+0x6e8/0x11d0 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x76c/0xb40 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: mdt04_001: service thread pid 10037 was inactive for 73.915 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 10037, comm: mdt04_001 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_fini+0xa25/0xec0 [ptlrpc] [<0>] ldlm_cli_enqueue+0x436/0xa30 [ptlrpc] [<0>] osp_md_object_lock+0x160/0x300 [osp] [<0>] lod_object_lock+0x7bb/0x7d0 [lod] [<0>] mdd_object_lock+0x2d/0xd0 [mdd] [<0>] mdt_reint_striped_lock+0x378/0x5b0 [mdt] [<0>] mdt_migrate_parent_lock+0x3f/0xb0 [mdt] [<0>] mdt_reint_migrate+0x8b7/0x11d0 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x76c/0xb40 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: 23249:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 7s req@ffff8802a11b5180 x1709885427325824/t0(0) o4->379425da-f4d8-45a0-a674-9fb5206d5eaf@0@lo:0/0 lens 488/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'cp.0' Lustre: 23249:0:(service.c:2157:ptlrpc_server_handle_req_in()) Skipped 1 previous similar message Lustre: mdt02_010: service thread pid 18821 was inactive for 70.048 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: 6831:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff880096e0e440 x1709885427511744/t0(0) o10->lustre-OST0003-osc-ffff88009baa1bf8@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630675196 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dd.0' Lustre: dir [0x200000404:0x497:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 2 previous similar messages LustreError: 15895:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 273533376 != fo_tot_granted 273746368 LustreError: 15895:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 162 previous similar messages LustreError: 15895:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 212992 LustreError: 15895:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 162 previous similar messages Lustre: 15417:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 7s req@ffff8802a5404500 x1709885429524160/t0(0) o101->379425da-f4d8-45a0-a674-9fb5206d5eaf@0@lo:0/0 lens 328/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'mkdir.0' Lustre: 15417:0:(service.c:2157:ptlrpc_server_handle_req_in()) Skipped 19 previous similar messages Lustre: 14947:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0x566:0x0] with magic=0xbd60bd0 Lustre: 14947:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 5 previous similar messages LustreError: 10017:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0001_UUID lock: ffff8802d6339a80/0xdbfb804ebcbc462c lrc: 3/0,0 mode: PW/PW res: [0x300000400:0xf0b:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000000020020 nid: 0@lo remote: 0xdbfb804ebcbc4450 expref: 3081 pid: 10734 timeout: 11812 lvb_type: 0 LustreError: 10017:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 3 previous similar messages LustreError: 12897:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630675296 with bad export cookie 15851404389121327019 LustreError: 28156:0:(ofd_obd.c:501:ofd_destroy_export()) lustre-OST0001: cli 379425da-f4d8-45a0-a674-9fb5206d5eaf/ffff8802418a8008 has 28672 pending on destroyed export Lustre: 6828:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.110@tcp:/lustre/fid: [0x280000404:0x403:0x0]/ may get corrupted (rc -108) Lustre: 6832:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.110@tcp:/lustre/fid: [0x240000403:0x43b:0x0]/ may get corrupted (rc -5) LustreError: 12729:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0002: '12' migration was interrupted, run 'lfs migrate -m 1 -c 1 -H crush 12' to finish migration. Lustre: lustre-OST0003-osc-ffff88009baa1bf8: disconnect after 30s idle Lustre: Skipped 5 previous similar messages Lustre: 11806:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000404:0x5ab:0x0] with magic=0xbd60bd0 Lustre: 11806:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 3 previous similar messages Lustre: dir [0x200000403:0x506:0x0] stripe 1 readdir failed: -2, directory is partially accessed! LustreError: 8470:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff88009baa1bf8: inode [0x240000404:0x5b7:0x0] mdc close failed: rc = -5 LustreError: 7029:0:(file.c:5042:ll_inode_revalidate_fini()) lustre: revalidate FID [0x240000403:0x1:0x0] error: rc = -108 LustreError: 8470:0:(file.c:234:ll_close_inode_openhandle()) Skipped 3 previous similar messages Lustre: 13927:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000404:0x5be:0x0] with magic=0xbd60bd0 Lustre: 13927:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 1 previous similar message LustreError: 8887:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff88009baa1bf8: inode [0x200000403:0x579:0x0] mdc close failed: rc = -108 LustreError: 8100:0:(llite_lib.c:2983:ll_prep_inode()) new_inode -fatal: rc -108 LustreError: 8887:0:(file.c:234:ll_close_inode_openhandle()) Skipped 9 previous similar messages LustreError: 8704:0:(file.c:5042:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000007:0x1:0x0] error: rc = -108 LustreError: 8704:0:(file.c:5042:ll_inode_revalidate_fini()) Skipped 11 previous similar messages LustreError: 15500:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8800ad2de440 x1709885436518080/t0(0) o105->lustre-OST0001@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 15500:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 323 previous similar messages LustreError: 13639:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630675477 with bad export cookie 15851404389123096402 LustreError: 13639:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) Skipped 4 previous similar messages Lustre: 6834:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.110@tcp:/lustre/fid: [0x280000404:0x41a:0x0]/ may get corrupted (rc -108) LustreError: 1094:0:(ofd_obd.c:501:ofd_destroy_export()) lustre-OST0002: cli 42274881-cfc4-4ff5-935b-f0524dbb475a/ffff880284718958 has 679936 pending on destroyed export Lustre: 17057:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 7s req@ffff8802dfe9a5c0 x1709885436643136/t0(0) o35->379425da-f4d8-45a0-a674-9fb5206d5eaf@0@lo:0/0 lens 392/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'cat.0' Lustre: 17057:0:(service.c:2157:ptlrpc_server_handle_req_in()) Skipped 1 previous similar message LustreError: 10043:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0002: '12' migration was interrupted, run 'lfs migrate -m 1 -c 1 -H crush 12' to finish migration. Lustre: lustre-OST0002-osc-ffff88009baa1bf8: Connection restored to 192.168.123.110@tcp (at 0@lo) Lustre: Skipped 10 previous similar messages Lustre: 10030:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 801/3204/0, destroy: 0/0/0 Lustre: 10030:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 155683 previous similar messages Lustre: 10030:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/1, xattr_set: 803/15/0 Lustre: 10030:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 155685 previous similar messages Lustre: 10030:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 4006/34453/0, punch: 0/0/0, quota 14/174/8 Lustre: 10030:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 155684 previous similar messages Lustre: 10030:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 801/13617/0, delete: 0/0/0 Lustre: 10030:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 155684 previous similar messages Lustre: 10030:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 10030:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 155684 previous similar messages Lustre: ost_io: This server is not able to keep up with request traffic (cpu-bound). Lustre: 2890:0:(service.c:1612:ptlrpc_at_check_timed()) earlyQ=0 reqQ=0 recA=1, svcEst=34, delay=7030ms Lustre: 10042:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 512 < left 1658, rollback = 7 Lustre: 10042:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 158665 previous similar messages LustreError: 167-0: lustre-OST0001-osc-ffff8802a9804138: This client was evicted by lustre-OST0001; in progress operations using this service will fail. LustreError: Skipped 9 previous similar messages Lustre: 6833:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.110@tcp:/lustre/fid: [0x200000404:0x6e0:0x0]// may get corrupted (rc -108) LustreError: 11009:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0001-osc-ffff8802a9804138: namespace resource [0x2330:0x0:0x0].0x0 (ffff8802db2fa340) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 11009:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 989 previous similar messages Lustre: 9983:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 6s req@ffff8802b7256a80 x1709885440999872/t0(0) o105->LOV_OSC_UUID@0@lo:0/0 lens 392/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'' Lustre: 9983:0:(service.c:2157:ptlrpc_server_handle_req_in()) Skipped 7 previous similar messages Lustre: 10572:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 8: before 510 < left 13633, rollback = 9 Lustre: 10572:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 177 previous similar messages LustreError: 10750:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 233510720 != fo_tot_granted 233723712 LustreError: 10750:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 473 previous similar messages LustreError: 10750:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 212992 LustreError: 10750:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 473 previous similar messages LustreError: 11-0: lustre-OST0003-osc-ffff88009baa1bf8: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 38 previous similar messages Lustre: lustre-OST0003-osc-ffff88009baa1bf8: Connection to lustre-OST0003 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 7 previous similar messages Lustre: 6825:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.110@tcp:/lustre/fid: [0x200000405:0x6f:0x0]// may get corrupted (rc -108) Lustre: 6824:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.110@tcp:/lustre/fid: [0x280000403:0x471:0x0]// may get corrupted (rc -108) Lustre: 6824:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.110@tcp:/lustre/fid: [0x200000404:0x753:0x0]/ may get corrupted (rc -108) ------------[ cut here ]------------ WARNING: CPU: 1 PID: 10025 at /home/green/git/lustre-release/ldiskfs/ext4_jbd2.c:266 __ldiskfs_handle_dirty_metadata+0x1d2/0x230 [ldiskfs] Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) dm_flakey dm_mod loop zfs(PO) zunicode(PO) zzstd(O) zlua(O) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) jbd2 mbcache crc_t10dif crct10dif_generic sb_edac edac_core iosf_mbi crc32_pclmul ghash_clmulni_intel aesni_intel lrw gf128mul glue_helper ablk_helper cryptd virtio_balloon pcspkr i2c_piix4 virtio_console ip_tables rpcsec_gss_krb5 ata_generic pata_acpi drm_kms_helper ttm crct10dif_pclmul ata_piix drm crct10dif_common crc32c_intel drm_panel_orientation_quirks serio_raw libata virtio_blk i2c_core floppy [last unloaded: libcfs] CPU: 1 PID: 10025 Comm: mdt00_001 Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011 Call Trace: [<ffffffff817d9418>] dump_stack+0x19/0x1b [<ffffffff8108d538>] __warn+0xd8/0x100 [<ffffffff8108d67d>] warn_slowpath_null+0x1d/0x20 [<ffffffffa0d9f262>] __ldiskfs_handle_dirty_metadata+0x1d2/0x230 [ldiskfs] [<ffffffffa0d7f33d>] ldiskfs_getblk+0x13d/0x210 [ldiskfs] [<ffffffffa0d7f437>] ldiskfs_bread+0x27/0xe0 [ldiskfs] [<ffffffffa0ccae97>] osd_write+0x4e7/0xcb0 [osd_ldiskfs] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffffa04e06f3>] dt_record_write+0x33/0x120 [obdclass] [<ffffffffa049f6b2>] llog_osd_write_rec+0x1052/0x19a0 [obdclass] [<ffffffffa048c760>] llog_write_rec+0x290/0x590 [obdclass] [<ffffffffa04924ee>] llog_cat_add_rec+0x23e/0x950 [obdclass] [<ffffffffa04895ff>] llog_add+0x17f/0x1f0 [obdclass] [<ffffffffa04baa39>] ? lprocfs_counter_add+0xf9/0x160 [obdclass] [<ffffffffa082360d>] sub_updates_write+0xb6b/0xe32 [ptlrpc] [<ffffffffa07f3892>] top_trans_stop+0x7c2/0xf60 [ptlrpc] [<ffffffffa0fadf3c>] lod_trans_stop+0x25c/0x340 [lod] [<ffffffffa0e82408>] mdd_trans_stop+0x28/0x16e [mdd] [<ffffffffa0e63b2f>] mdd_migrate_object+0xf1f/0x1ba0 [mdd] [<ffffffffa04df375>] ? lu_object_find_at+0x285/0xb10 [obdclass] [<ffffffffa0e64b86>] mdd_migrate+0x3d6/0x8e0 [mdd] [<ffffffffa0f360d8>] mdo_migrate+0x4c/0x4e [mdt] [<ffffffffa0ef5e53>] mdt_reint_migrate+0x1023/0x11d0 [mdt] [<ffffffffa0ef6087>] mdt_reint_rec+0x87/0x240 [mdt] [<ffffffffa0ecb55c>] mdt_reint_internal+0x76c/0xb40 [mdt] [<ffffffffa0ed8657>] mdt_reint+0x67/0x150 [mdt] [<ffffffffa07e226e>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<ffffffffa03c645e>] ? libcfs_nid2str_r+0xfe/0x130 [lnet] [<ffffffffa078d990>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<ffffffffa078f529>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffffa078e950>] ? ptlrpc_wait_event+0x620/0x620 [ptlrpc] [<ffffffff810ba0f4>] kthread+0xe4/0xf0 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 ---[ end trace b1a5eb572da447ea ]--- LDISKFS-fs: ldiskfs_getblk:888: aborting transaction: error 28 in __ldiskfs_handle_dirty_metadata LDISKFS-fs error (device dm-0): ldiskfs_getblk:888: inode #230: block 53736: comm mdt00_001: journal_dirty_metadata failed: handle type 0 started at line 1972, credits 512/0, errcode -28 Aborting journal on device dm-0-8. LDISKFS-fs (dm-0): Remounting filesystem read-only LustreError: 10025:0:(osd_io.c:2103:osd_ldiskfs_write_record()) lustre-MDT0000: error reading offset 438272 (block 107, size 32, offs 438248), credits 0/0: rc = -28 LustreError: 10025:0:(update_trans.c:985:top_trans_stop()) lustre-MDT0000-osd: write updates failed: rc = -28 LustreError: 10025:0:(osd_io.c:2110:osd_ldiskfs_write_record()) journal_get_write_access() returned error -30 LustreError: 10025:0:(tgt_lastrcvd.c:1253:tgt_add_reply_data()) lustre-MDT0000: can't update reply_data file: rc = -30 LustreError: 10025:0:(osd_handler.c:2075:osd_trans_stop()) lustre-MDT0000: failed in transaction hook: rc = -30 LDISKFS-fs error (device dm-0) in osd_trans_stop:2082: error 28 LustreError: 9972:0:(osd_handler.c:1779:osd_trans_commit_cb()) transaction @0xffff8802f45da758 commit error: 2 LustreError: 12652:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -30 LustreError: 10025:0:(osd_handler.c:2085:osd_trans_stop()) lustre-MDT0000: failed to stop transaction: rc = -28 LustreError: 10025:0:(update_trans.c:1011:top_trans_stop()) lustre-MDT0000-osd: stop trans failed: rc = -30 | Externally reported by onyx-68 boilpot email |
racer test 1: racer on clients: centos-25.localnet DURATION=2700 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa178be52>] ll_migrate+0x9b2/0xec0 [lustre] PGD 800000009c015067 PUD 9c014067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) dm_flakey dm_mod loop zfs(PO) zunicode(PO) zzstd(O) zlua(O) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) jbd2 mbcache crc_t10dif crct10dif_generic sb_edac edac_core iosf_mbi crc32_pclmul ghash_clmulni_intel aesni_intel lrw gf128mul glue_helper ablk_helper cryptd virtio_balloon pcspkr virtio_console i2c_piix4 ip_tables rpcsec_gss_krb5 drm_kms_helper ata_generic ttm crct10dif_pclmul pata_acpi drm crct10dif_common ata_piix drm_panel_orientation_quirks crc32c_intel serio_raw virtio_blk libata i2c_core floppy [last unloaded: libcfs] CPU: 12 PID: 30942 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011 task: ffff8800972924f0 ti: ffff8800a04bc000 task.ti: ffff8800a04bc000 RIP: 0010:[<ffffffffa178be52>] [<ffffffffa178be52>] ll_migrate+0x9b2/0xec0 [lustre] RSP: 0018:ffff8800a04bfbc8 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff8800685cd2c8 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff880326407c98 RDI: ffff880326407c78 RBP: ffff8800a04bfc48 R08: ffff8802b7cb8d98 R09: ffff8800a04bfa60 R10: 0000000000000005 R11: 0000000000000400 R12: 0000000000000000 R13: ffff8802b1122128 R14: ffff880218ba8a58 R15: 0000000000000030 FS: 00007fb70c5d7740(0000) GS:ffff880331d00000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 CR2: 0000000000000008 CR3: 0000000099ea8000 CR4: 00000000001607e0 Call Trace: [<ffffffff81242213>] ? __check_object_size+0x1c3/0x220 [<ffffffffa1779561>] ll_dir_ioctl+0x5d01/0x6ed0 [lustre] [<ffffffff811d5aba>] ? kvfree+0x2a/0x40 [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff81415f91>] ? debug_object_activate+0x191/0x210 [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff8141676b>] ? debug_object_active_state+0x15b/0x190 [<ffffffff8125e410>] ? release_dentry_name_snapshot+0x30/0x30 [<ffffffff8125b3fd>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff81406d25>] ? lockref_put_or_lock+0x25/0x30 [<ffffffff8126a634>] ? mntput+0x24/0x40 [<ffffffff8125b751>] SyS_ioctl+0xa1/0xc0 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 | Lustre: 27085:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 512 < left 1705, rollback = 7 Lustre: 27085:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 27085:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 1/89/0 Lustre: 27085:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1696/1705/0, punch: 0/0/0, quota 3/3/0 Lustre: 27085:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 27085:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 27085:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 512 < left 1705, rollback = 7 Lustre: 27085:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 74 previous similar messages Lustre: 27085:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 27085:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 199 previous similar messages Lustre: 27085:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 1/89/0 Lustre: 27085:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 199 previous similar messages Lustre: 27085:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1696/1705/0, punch: 0/0/0, quota 3/3/0 Lustre: 27085:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 199 previous similar messages Lustre: 27085:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 27085:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 199 previous similar messages Lustre: 27085:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 27085:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 199 previous similar messages Lustre: 27084:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 505 < left 688, rollback = 2 Lustre: 29827:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 8: before 509 < left 3399, rollback = 9 Lustre: 29827:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 5 previous similar messages Lustre: 29827:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 199/796/0, destroy: 1/4/0 Lustre: 29827:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 1501 previous similar messages Lustre: 29827:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 3/3/1, xattr_set: 203/148/0 Lustre: 29827:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 1501 previous similar messages Lustre: 29827:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 996/8567/0, punch: 0/0/0, quota 6/6/0 Lustre: 29827:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 1501 previous similar messages Lustre: 29827:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 200/3399/0, delete: 2/5/1 Lustre: 29827:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 1501 previous similar messages Lustre: 29827:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 1/1/0, ref_del: 2/2/1 Lustre: 29827:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 1501 previous similar messages Lustre: 29827:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 508 < left 8567, rollback = 7 Lustre: 29827:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 1620 previous similar messages Lustre: dir [0x240000403:0x26:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: 29615:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 505 < left 511, rollback = 7 Lustre: 29615:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 29615:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 29615:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 1/89/6 Lustre: 29615:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 29615:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 503/512/1, punch: 0/0/0, quota 3/3/0 Lustre: 29615:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 29615:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 29615:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 29615:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 29615:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 1 previous similar message Lustre: 27068:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 509 < left 688, rollback = 2 Lustre: 30839:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 494 < left 610, rollback = 2 Lustre: 30839:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 5 previous similar messages Lustre: 30839:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 3/12/12, destroy: 0/0/0 Lustre: 30839:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 1046 previous similar messages Lustre: 30839:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 2/2/0, xattr_set: 8/610/0 Lustre: 30839:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 1046 previous similar messages Lustre: 30839:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 5/136/0, punch: 0/0/0, quota 8/56/2 Lustre: 30839:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 1046 previous similar messages Lustre: 30839:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 12/215/4, delete: 1/1/0 Lustre: 30839:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 1046 previous similar messages Lustre: 30839:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 6/6/0, ref_del: 1/1/0 Lustre: 30839:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 1046 previous similar messages Lustre: 27063:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 506 < left 535, rollback = 2 Lustre: 27063:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 19 previous similar messages Lustre: 27063:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 3/12/3, destroy: 0/0/0 Lustre: 27063:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 19 previous similar messages Lustre: 27063:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 3/3/0, xattr_set: 8/535/0 Lustre: 27063:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 19 previous similar messages Lustre: 27063:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 5/136/0, punch: 0/0/0, quota 8/56/0 Lustre: 27063:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 19 previous similar messages Lustre: 27063:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 12/215/3, delete: 1/1/0 Lustre: 27063:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 19 previous similar messages Lustre: 27063:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 6/6/0, ref_del: 1/1/0 Lustre: 27063:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 19 previous similar messages Lustre: 27074:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 509 < left 21765, rollback = 7 Lustre: 27074:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 1040 previous similar messages 13[31480]: segfault at 8 ip 00007f1f7d15c7e8 sp 00007fff9d919270 error 4 in ld-2.17.so[7f1f7d151000+22000] Lustre: 29732:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 507 < left 1578, rollback = 2 Lustre: 29732:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 4 previous similar messages Lustre: 32062:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 1040/4160/0, destroy: 1/4/0 Lustre: 32062:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 30 previous similar messages Lustre: 32062:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 3/3/1, xattr_set: 1044/148/0 Lustre: 32062:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 30 previous similar messages Lustre: 32062:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 5201/44730/0, punch: 0/0/0, quota 6/6/0 Lustre: 32062:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 30 previous similar messages Lustre: 32062:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 1041/17696/0, delete: 2/5/1 Lustre: 32062:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 30 previous similar messages Lustre: 32062:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 1/1/0, ref_del: 2/2/1 Lustre: 32062:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 30 previous similar messages Lustre: 32062:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 507 < left 44730, rollback = 7 Lustre: 32062:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 4 previous similar messages Lustre: ll_ost04_004: service thread pid 30117 was inactive for 40.124 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 30117, comm: ll_ost04_004 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] tgt_extent_lock+0xea/0x2a0 [ptlrpc] [<0>] ofd_getattr_hdl+0x365/0x720 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: 32698:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 506 < left 521, rollback = 2 Lustre: 32698:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 21 previous similar messages Lustre: ll_ost06_000: service thread pid 27767 was inactive for 40.008 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 27767, comm: ll_ost06_000 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] tgt_extent_lock+0xea/0x2a0 [ptlrpc] [<0>] ofd_getattr_hdl+0x365/0x720 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 32093, comm: ll_ost06_008 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] tgt_extent_lock+0xea/0x2a0 [ptlrpc] [<0>] ofd_getattr_hdl+0x365/0x720 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: ll_ost06_005: service thread pid 30862 was inactive for 40.442 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: 27079:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 512 < left 1615, rollback = 7 Lustre: 27079:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 5123 previous similar messages ------------[ cut here ]------------ WARNING: CPU: 3 PID: 31611 at /home/green/git/lustre-release/ldiskfs/namei.c:3331 ldiskfs_orphan_add+0x11e/0x2a0 [ldiskfs] Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) dm_flakey dm_mod loop zfs(PO) zunicode(PO) zzstd(O) zlua(O) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) jbd2 mbcache crc_t10dif crct10dif_generic sb_edac edac_core iosf_mbi crc32_pclmul ghash_clmulni_intel aesni_intel lrw gf128mul glue_helper ablk_helper cryptd virtio_balloon pcspkr virtio_console i2c_piix4 ip_tables rpcsec_gss_krb5 drm_kms_helper ata_generic ttm crct10dif_pclmul pata_acpi drm crct10dif_common ata_piix drm_panel_orientation_quirks crc32c_intel serio_raw virtio_blk libata i2c_core floppy [last unloaded: libcfs] CPU: 3 PID: 31611 Comm: mdt_rdpg01_003 Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011 Call Trace: [<ffffffff817d9418>] dump_stack+0x19/0x1b [<ffffffff8108d538>] __warn+0xd8/0x100 [<ffffffff8108d67d>] warn_slowpath_null+0x1d/0x20 [<ffffffffa0d574ae>] ldiskfs_orphan_add+0x11e/0x2a0 [ldiskfs] [<ffffffffa0d489eb>] ldiskfs_xattr_inode_orphan_add.constprop.17+0xbb/0x110 [ldiskfs] [<ffffffffa0d4b417>] ldiskfs_xattr_delete_inode+0x77/0x3c0 [ldiskfs] [<ffffffffa0d6baa0>] ? ldiskfs_evict_inode+0x190/0x670 [ldiskfs] [<ffffffffa0d6bad9>] ldiskfs_evict_inode+0x1c9/0x670 [ldiskfs] [<ffffffff812649af>] evict+0xaf/0x180 [<ffffffff81264de5>] iput+0xf5/0x180 [<ffffffffa0c87d08>] osd_object_delete+0x1f8/0x370 [osd_ldiskfs] [<ffffffffa04d859d>] lu_object_free.isra.26+0xcd/0x1d0 [obdclass] [<ffffffffa04dccf1>] lu_object_put+0xa1/0x480 [obdclass] [<ffffffffa0ee7af4>] mdt_mfd_close+0x1f4/0x870 [mdt] [<ffffffffa0eede14>] mdt_close_internal+0xb4/0x240 [mdt] [<ffffffffa0eee224>] mdt_close+0x284/0x850 [mdt] [<ffffffffa07df2d4>] ? tgt_request_preprocess.isra.27+0xf4/0x850 [ptlrpc] [<ffffffffa07e026e>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<ffffffffa033045e>] ? libcfs_nid2str_r+0xfe/0x130 [lnet] [<ffffffffa078b990>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<ffffffffa078d529>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffffa078c950>] ? ptlrpc_wait_event+0x620/0x620 [ptlrpc] [<ffffffff810ba0f4>] kthread+0xe4/0xf0 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 ---[ end trace be1905b362903f57 ]--- Lustre: ll_ost01_001: service thread pid 27753 was inactive for 62.060 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 1 previous similar message Lustre: 27084:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 506 < left 22456, rollback = 7 Lustre: 27084:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 2403 previous similar messages Lustre: 27084:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 522/2088/0, destroy: 0/0/0 Lustre: 27084:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 7537 previous similar messages Lustre: 27084:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/1, xattr_set: 524/15/0 Lustre: 27084:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 7537 previous similar messages Lustre: 27084:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 2611/22456/0, punch: 0/0/0, quota 10/106/5 Lustre: 27084:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 7537 previous similar messages Lustre: 27084:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 522/8874/0, delete: 0/0/0 Lustre: 27084:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 7537 previous similar messages Lustre: 27084:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 27084:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 7537 previous similar messages Lustre: dir [0x200000404:0x61:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: mdt04_004: service thread pid 29875 was inactive for 40.145 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. LustreError: 27055:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0003_UUID lock: ffff880092f68400/0x6d2fd72113cbf18e lrc: 3/0,0 mode: PW/PW res: [0x380000401:0x40:0x0].0x0 rrc: 4 type: EXT [0->18446744073709551615] (req 1048576->1114111) gid 0 flags: 0x60000400000020 nid: 0@lo remote: 0x6d2fd72113cbf187 expref: 4866 pid: 28436 timeout: 3836 lvb_type: 0 LustreError: 11-0: lustre-OST0003-osc-ffff8802157b0958: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-OST0003-osc-ffff8802157b0958: Connection to lustre-OST0003 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0003-osc-ffff8802157b0958: This client was evicted by lustre-OST0003; in progress operations using this service will fail. Lustre: dir [0x240000403:0x4b:0x0] stripe 1 readdir failed: -2, directory is partially accessed! LustreError: 32192:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802d8c59940 x1709865910043904/t0(0) o104->lustre-OST0003@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 2263:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0003-osc-ffff8802157b0958: namespace resource [0x3c0:0x0:0x0].0x0 (ffff8802d0a05440) refcount nonzero (2) after lock cleanup; forcing cleanup. LustreError: 31829:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802b03b5180 x1709865910079936/t0(0) o104->lustre-OST0003@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 31829:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 4 previous similar messages Lustre: 32698:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 8: before 510 < left 12817, rollback = 9 Lustre: 32698:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 8 previous similar messages LustreError: 31821:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802e2fa0680 x1709865910106624/t0(0) o104->lustre-OST0003@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 29881:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0001: '6' migration was interrupted, run 'lfs migrate -m 0 -c 1 -H crush 6' to finish migration. LustreError: 2263:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0003-osc-ffff8802157b0958: namespace resource [0x380000401:0x40:0x0].0x0 (ffff88008edb96c0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 2263:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 727 previous similar messages Lustre: lustre-OST0003-osc-ffff8802157b0958: Connection restored to 192.168.123.125@tcp (at 0@lo) LustreError: 30333:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0001: failed to get lu_attr of [0x240000403:0x4b:0x0]: rc = -2 LustreError: 27055:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0003_UUID lock: ffff880092e77c00/0x6d2fd72113cc23a6 lrc: 3/0,0 mode: PW/PW res: [0x2:0x0:0x0].0x0 rrc: 2 type: EXT [0->18446744073709551615] (req 0->3145727) gid 0 flags: 0x60000400020020 nid: 0@lo remote: 0x6d2fd72113cc2398 expref: 3239 pid: 27749 timeout: 3867 lvb_type: 0 LustreError: 28479:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802ba8d3ec0 x1709865911190976/t0(0) o105->lustre-OST0003@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 28479:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 7 previous similar messages LustreError: 11-0: lustre-OST0003-osc-ffff880246d712a8: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-OST0003-osc-ffff880246d712a8: Connection to lustre-OST0003 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0003-osc-ffff880246d712a8: This client was evicted by lustre-OST0003; in progress operations using this service will fail. LustreError: 1704:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802d8313880 x1709865911372800/t0(0) o104->lustre-OST0003@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 1704:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 20 previous similar messages Lustre: 23868:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.125@tcp:/lustre/fid: [0x280000404:0x44:0x0]// may get corrupted (rc -108) Lustre: lustre-OST0003-osc-ffff880246d712a8: Connection restored to 192.168.123.125@tcp (at 0@lo) Lustre: 31611:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 509 < left 44730, rollback = 7 Lustre: 31611:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 26785 previous similar messages Lustre: 31611:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 1040/4160/0, destroy: 1/4/1 Lustre: 31611:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 26858 previous similar messages Lustre: 31611:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 1040/0/0 Lustre: 31611:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 26856 previous similar messages Lustre: 31611:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 5201/44730/0, punch: 0/0/0, quota 6/6/0 Lustre: 31611:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 26855 previous similar messages Lustre: 31611:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 1040/17680/0, delete: 2/5/1 Lustre: 31611:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 26856 previous similar messages Lustre: 31611:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 1/1/1 Lustre: 31611:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 26856 previous similar messages Lustre: 23866:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff8800a0adea80 x1709865912189504/t0(0) o10->lustre-OST0000-osc-ffff8802157b0958@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630655507 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'cp.0' LustreError: 32310:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0001: '0' migration was interrupted, run 'lfs migrate -m 2 -c 1 -H crush 0' to finish migration. Lustre: 27066:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 8: before 510 < left 27318, rollback = 9 Lustre: 27066:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 70 previous similar messages Lustre: ll_ost05_006: service thread pid 31826 was inactive for 65.988 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. LustreError: 27055:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0003_UUID lock: ffff88026abf87c0/0x6d2fd72113d8d79d lrc: 3/0,0 mode: PW/PW res: [0x19f:0x0:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000400030020 nid: 0@lo remote: 0x6d2fd72113d8d796 expref: 2870 pid: 32384 timeout: 3992 lvb_type: 0 LustreError: 32175:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802b2037700 x1709865916913344/t0(0) o104->lustre-OST0003@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 32175:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 10 previous similar messages Lustre: lustre-OST0003-osc-ffff8802157b0958: Connection to lustre-OST0003 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: 23866:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1630655547/real 1630655547] req@ffff8802a2579940 x1709865915766144/t0(0) o1->lustre-OST0003-osc-ffff8802157b0958@0@lo:28/4 lens 440/432 e 0 to 1 dl 1630655605 ref 1 fl Rpc:XQr/0/ffffffff rc 0/-1 job:'lfs.0' Lustre: 23866:0:(client.c:2285:ptlrpc_expire_one_request()) Skipped 2 previous similar messages Lustre: 23865:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1630655549/real 1630655549] req@ffff88029d8f25c0 x1709865915777280/t0(0) o1->lustre-OST0003-osc-ffff8802157b0958@0@lo:28/4 lens 440/432 e 0 to 1 dl 1630655607 ref 1 fl Rpc:XQr/0/ffffffff rc 0/-1 job:'lfs.0' Lustre: 23865:0:(client.c:2285:ptlrpc_expire_one_request()) Skipped 10 previous similar messages Lustre: ll_ost02_003: service thread pid 28409 was inactive for 62.039 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 12 previous similar messages Lustre: 776:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1630655552/real 1630655552] req@ffff8802a532de00 x1709865916092544/t0(0) o101->lustre-OST0003-osc-ffff8802157b0958@0@lo:28/4 lens 328/400 e 0 to 1 dl 1630655610 ref 2 fl Rpc:XQr/0/ffffffff rc 0/-1 job:'dd.0' Lustre: 776:0:(client.c:2285:ptlrpc_expire_one_request()) Skipped 5 previous similar messages LustreError: 27055:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 138s: evicting client at 0@lo ns: filter-lustre-OST0001_UUID lock: ffff8802d61b8f40/0x6d2fd72113d8ebfc lrc: 3/0,0 mode: PW/PW res: [0x1bf:0x0:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000400030020 nid: 0@lo remote: 0x6d2fd72113d8ebf5 expref: 5410 pid: 32366 timeout: 4031 lvb_type: 0 LustreError: 4127:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802df501f80 x1709865917265408/t0(0) o104->lustre-OST0001@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 4127:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 7 previous similar messages Lustre: 29313:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1630655560/real 1630655560] req@ffff8802cd094b40 x1709865916378368/t0(0) o101->lustre-OST0003-osc-ffff8802157b0958@0@lo:28/4 lens 328/400 e 0 to 1 dl 1630655618 ref 2 fl Rpc:XQr/0/ffffffff rc 0/-1 job:'dir_create.sh.0' Lustre: mdt07_014: service thread pid 3172 was inactive for 86.194 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 59 previous similar messages LustreError: 27055:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 139s: evicting client at 0@lo ns: filter-lustre-OST0000_UUID lock: ffff8802ba4f25c0/0x6d2fd72113cf5422 lrc: 3/0,0 mode: PW/PW res: [0x3:0x0:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->917503) gid 0 flags: 0x60000400030020 nid: 0@lo remote: 0x6d2fd72113cf53ab expref: 5352 pid: 32100 timeout: 4035 lvb_type: 0 LustreError: 27055:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 7 previous similar messages LustreError: 11-0: lustre-OST0002-osc-ffff880246d712a8: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-OST0002-osc-ffff880246d712a8: Connection to lustre-OST0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 5399:0:(service.c:2290:ptlrpc_server_handle_request()) @@@ Dropping timed-out request from 12345-0@lo: deadline 43/29s ago req@ffff8802df4f8040 x1709865915753600/t0(0) o1->8f82a635-3fbb-4c4d-a2ce-769d96ea45a7@0@lo:550/0 lens 440/0 e 0 to 0 dl 1630655590 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'lfs.0' Lustre: 5399:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (43/29s); client may timeout req@ffff8802df4f8040 x1709865915753600/t0(0) o1->8f82a635-3fbb-4c4d-a2ce-769d96ea45a7@0@lo:550/0 lens 440/0 e 0 to 0 dl 1630655590 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'lfs.0' LustreError: 5399:0:(service.c:2290:ptlrpc_server_handle_request()) @@@ Dropping timed-out request from 12345-0@lo: deadline 43/10s ago req@ffff8802ba838680 x1709865916588864/t0(0) o101->8f82a635-3fbb-4c4d-a2ce-769d96ea45a7@0@lo:570/0 lens 328/0 e 0 to 0 dl 1630655610 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'ll_sa_5992.0' LustreError: 5399:0:(service.c:2290:ptlrpc_server_handle_request()) Skipped 47 previous similar messages Lustre: 5399:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (43/10s); client may timeout req@ffff8802ba838680 x1709865916588864/t0(0) o101->8f82a635-3fbb-4c4d-a2ce-769d96ea45a7@0@lo:570/0 lens 328/0 e 0 to 0 dl 1630655610 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'ll_sa_5992.0' Lustre: 5399:0:(service.c:2328:ptlrpc_server_handle_request()) Skipped 47 previous similar messages LustreError: 717:0:(ofd_obd.c:501:ofd_destroy_export()) lustre-OST0003: cli 8f82a635-3fbb-4c4d-a2ce-769d96ea45a7/ffff88009d21e678 has 1859584 pending on destroyed export LustreError: 717:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_destroy_export: tot_granted 57239872 != fo_tot_granted 59099456 LustreError: 2046:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_destroy_export: tot_pending 0 != fo_tot_pending 2056192 LustreError: 717:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 1 previous similar message LustreError: 11-0: lustre-OST0002-osc-ffff880246d712a8: operation ost_setattr to node 0@lo failed: rc = -107 LustreError: Skipped 588 previous similar messages LustreError: 167-0: lustre-OST0003-osc-ffff8802157b0958: This client was evicted by lustre-OST0003; in progress operations using this service will fail. Lustre: 23870:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.125@tcp:/lustre/fid: [0x200000404:0x111:0x0]// may get corrupted (rc -5) LustreError: 27781:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 59189568 != fo_tot_granted 61049152 LustreError: 27781:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 1859584 LustreError: 27781:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 1 previous similar message LustreError: 5421:0:(service.c:2290:ptlrpc_server_handle_request()) @@@ Dropping timed-out request from 12345-0@lo: deadline 20/26s ago req@ffff8802b9c425c0 x1709865917012864/t0(0) o8-><?>@<?>:0/0 lens 520/0 e 0 to 0 dl 1630655598 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'ptlrpcd_01_00.0' Lustre: 5421:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (20/26s); client may timeout req@ffff8802b9c425c0 x1709865917012864/t0(0) o8-><?>@<?>:0/0 lens 520/0 e 0 to 0 dl 1630655598 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'ptlrpcd_01_00.0' LustreError: 167-0: lustre-OST0000-osc-ffff8802157b0958: This client was evicted by lustre-OST0000; in progress operations using this service will fail. LustreError: Skipped 1 previous similar message LustreError: 27780:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 10588736 != fo_tot_granted 12644928 LustreError: 27780:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 5 previous similar messages LustreError: 27780:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 896 != fo_tot_pending 2057088 LustreError: 27780:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 5 previous similar messages Lustre: 23864:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.125@tcp:/lustre/fid: [0x280000404:0x12a:0x0]// may get corrupted (rc -5) Lustre: lustre-OST0002-osc-ffff880246d712a8: Connection restored to 192.168.123.125@tcp (at 0@lo) LustreError: 27781:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 10587840 != fo_tot_granted 12644032 LustreError: 27781:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 2056192 LustreError: 9873:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0003-osc-ffff8802157b0958: namespace resource [0x380000401:0x1bf:0x0].0x0 (ffff8802e07b6840) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 9873:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 1 previous similar message Lustre: lustre-OST0003-osc-ffff8802157b0958: Connection restored to 192.168.123.125@tcp (at 0@lo) Lustre: 30396:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 8: before 510 < left 12885, rollback = 9 Lustre: 30396:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 27 previous similar messages Lustre: 30396:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 757/3028/0, destroy: 1/4/0 Lustre: 30396:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 32016 previous similar messages Lustre: 30396:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 3/3/1, xattr_set: 761/148/0 Lustre: 30396:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 32016 previous similar messages Lustre: 30396:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 3786/32561/0, punch: 0/0/0, quota 6/6/0 Lustre: 30396:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 32016 previous similar messages Lustre: 30396:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 758/12885/0, delete: 2/5/1 Lustre: 30396:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 32016 previous similar messages Lustre: 30396:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 1/1/0, ref_del: 2/2/0 Lustre: 30396:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 32016 previous similar messages Lustre: 30396:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 509 < left 32561, rollback = 7 Lustre: 30396:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 31988 previous similar messages LustreError: 27780:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 10589376 != fo_tot_granted 12645568 LustreError: 27780:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 5 previous similar messages LustreError: 27780:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 896 != fo_tot_pending 2057088 LustreError: 27780:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 5 previous similar messages LustreError: 9883:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0002-osc-ffff8802157b0958: namespace resource [0x400:0x0:0x0].0x0 (ffff88021c239940) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 9883:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 306 previous similar messages Lustre: lustre-OST0002-osc-ffff8802157b0958: Connection restored to 192.168.123.125@tcp (at 0@lo) Lustre: Skipped 2 previous similar messages LustreError: 27774:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 10588736 != fo_tot_granted 12644928 LustreError: 27055:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0002_UUID lock: ffff8802bd882980/0x6d2fd72113dded96 lrc: 3/0,0 mode: PR/PR res: [0x280000404:0x3e:0x0].0x0 bits 0x12/0x0 rrc: 5 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0x6d2fd72113dded65 expref: 97 pid: 31524 timeout: 4059 lvb_type: 0 LustreError: 27055:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 1 previous similar message LustreError: 27774:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 8 previous similar messages LustreError: 11-0: lustre-MDT0002-mdc-ffff8802157b0958: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 417 previous similar messages Lustre: lustre-MDT0002-mdc-ffff8802157b0958: Connection to lustre-MDT0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 3 previous similar messages LustreError: 167-0: lustre-MDT0002-mdc-ffff8802157b0958: This client was evicted by lustre-MDT0002; in progress operations using this service will fail. LustreError: Skipped 2 previous similar messages LustreError: 27774:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 2056192 LustreError: 27774:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 8 previous similar messages LustreError: 3004:0:(llite_lib.c:2983:ll_prep_inode()) new_inode -fatal: rc -5 LustreError: 3477:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff8802157b0958: inode [0x280000404:0x3e:0x0] mdc close failed: rc = -108 Lustre: mdt07_001: service thread pid 27084 was inactive for 86.338 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 3 previous similar messages LustreError: 4992:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff880246d712a8: inode [0x200000404:0xde:0x0] mdc close failed: rc = -13 LustreError: 4992:0:(file.c:234:ll_close_inode_openhandle()) Skipped 12 previous similar messages LustreError: 27776:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 103341888 != fo_tot_granted 105398080 LustreError: 27776:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 14 previous similar messages LustreError: 10169:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 2211840 != fo_tot_pending 4268032 LustreError: 10169:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 15 previous similar messages Lustre: dir [0x200000403:0x41:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 2 previous similar messages LustreError: 31784:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0001: '9' migration was interrupted, run 'lfs migrate -m 2 -c 1 -H crush 9' to finish migration. LustreError: 29989:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000403:0x4d:0x0]: rc = -2 LustreError: 10649:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff880246d712a8: inode [0x240000404:0x125:0x0] mdc close failed: rc = -13 LustreError: 27773:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 126412736 != fo_tot_granted 128468928 LustreError: 27773:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 23 previous similar messages LustreError: 27774:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 61440 != fo_tot_pending 1921024 LustreError: 27774:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 23 previous similar messages Lustre: 23876:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff8800a12ed7c0 x1709865922506432/t0(0) o10->lustre-OST0002-osc-ffff880246d712a8@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630655787 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dd.0' Lustre: 23876:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff8802d5303880 x1709865922516928/t0(0) o10->lustre-OST0001-osc-ffff880246d712a8@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630655803 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dd.0' Lustre: 23876:0:(client.c:1485:after_reply()) Skipped 19 previous similar messages Lustre: 23876:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff8800a377ea80 x1709865922524800/t0(0) o10->lustre-OST0001-osc-ffff880246d712a8@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630655803 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dd.0' Lustre: 23876:0:(client.c:1485:after_reply()) Skipped 66 previous similar messages LustreError: 27055:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0002_UUID lock: ffff880095bf0b80/0x6d2fd72113e1138b lrc: 3/0,0 mode: PW/PW res: [0x340000401:0x192:0x0].0x0 rrc: 5 type: EXT [0->18446744073709551615] (req 0->1048575) gid 0 flags: 0x60000400020020 nid: 0@lo remote: 0x6d2fd72113e11384 expref: 3367 pid: 32144 timeout: 4157 lvb_type: 0 LustreError: 28421:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8800a377bec0 x1709865923125056/t0(0) o104->lustre-OST0002@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 28421:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 71 previous similar messages LustreError: 11-0: lustre-OST0002-osc-ffff880246d712a8: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 1 previous similar message Lustre: lustre-OST0002-osc-ffff880246d712a8: Connection to lustre-OST0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0002-osc-ffff880246d712a8: This client was evicted by lustre-OST0002; in progress operations using this service will fail. Lustre: 23876:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.125@tcp:/lustre/fid: [0x200000404:0xf6:0x0]// may get corrupted (rc -108) Lustre: 23875:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.125@tcp:/lustre/fid: [0x240000403:0x35f:0x0]// may get corrupted (rc -108) Lustre: 23876:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.125@tcp:/lustre/fid: [0x280000403:0x109:0x0]// may get corrupted (rc -108) LustreError: 27786:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 145364800 != fo_tot_granted 147420992 LustreError: 27786:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 44 previous similar messages Lustre: 29076:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000404:0x157:0x0] with magic=0xbd60bd0 LustreError: 27780:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 1859584 LustreError: 27780:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 46 previous similar messages LustreError: 12630:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0002-osc-ffff880246d712a8: namespace resource [0x340000400:0xc0:0x0].0x0 (ffff880095bba840) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 12630:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 558 previous similar messages LustreError: 11280:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630655760 with bad export cookie 7867743611106902789 LustreError: 11-0: lustre-OST0002-osc-ffff8802157b0958: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-OST0002-osc-ffff8802157b0958: Connection to lustre-OST0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0002-osc-ffff8802157b0958: This client was evicted by lustre-OST0002; in progress operations using this service will fail. Lustre: 23873:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.125@tcp:/lustre/fid: [0x280000404:0x12d:0x0]// may get corrupted (rc -5) Lustre: 23873:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.125@tcp:/lustre/fid: [0x240000403:0x373:0x0]/ may get corrupted (rc -5) Lustre: 23875:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff8800a12ecb40 x1709865924940352/t0(0) o10->lustre-OST0001-osc-ffff880246d712a8@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630655848 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dd.0' Lustre: lustre-OST0002-osc-ffff8802157b0958: Connection restored to 192.168.123.125@tcp (at 0@lo) Lustre: Skipped 1 previous similar message Lustre: lustre-OST0002: haven't heard from client efb7de0b-f08a-47d7-895a-a686c416123c (at 0@lo) in 48 seconds. I think it's dead, and I am evicting it. exp ffff8802cf6912a8, cur 1630655790 expire 1630655760 last 1630655742 LustreError: 13162:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff8802157b0958: inode [0x240000404:0x173:0x0] mdc close failed: rc = -13 LustreError: 12630:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0002-osc-ffff880246d712a8: namespace resource [0x340000400:0x925:0x0].0x0 (ffff8802bdd84540) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 12630:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 115 previous similar messages LustreError: 11-0: lustre-OST0002-osc-ffff880246d712a8: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 2 previous similar messages Lustre: lustre-OST0002-osc-ffff880246d712a8: Connection to lustre-OST0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0002-osc-ffff880246d712a8: This client was evicted by lustre-OST0002; in progress operations using this service will fail. Lustre: 23871:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff8802e30c8040 x1709865927161664/t0(0) o10->lustre-OST0000-osc-ffff8802157b0958@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630655842 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'cp.0' Lustre: 23871:0:(client.c:1485:after_reply()) Skipped 43 previous similar messages LustreError: 27090:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0001: failed to get lu_attr of [0x240000404:0x145:0x0]: rc = -2 Lustre: 23863:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff8802b4b51f80 x1709865930200960/t0(0) o10->lustre-OST0000-osc-ffff880246d712a8@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630655935 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dir_create.sh.0' LustreError: 27775:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 188851008 != fo_tot_granted 190907200 LustreError: 27775:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 132 previous similar messages LustreError: 27788:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 3506176 != fo_tot_pending 5562368 LustreError: 27788:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 131 previous similar messages Lustre: 31784:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 512 < left 525, rollback = 7 Lustre: 31784:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 190421 previous similar messages Lustre: 31784:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 31784:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 190618 previous similar messages Lustre: 31784:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 1/89/0 Lustre: 31784:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 190622 previous similar messages Lustre: 31784:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 516/525/0, punch: 0/0/0, quota 3/3/0 Lustre: 31784:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 190616 previous similar messages Lustre: 31784:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 31784:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 190618 previous similar messages Lustre: 31784:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 31784:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 190612 previous similar messages Lustre: 30396:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 8: before 510 < left 4776, rollback = 9 Lustre: 30396:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 198 previous similar messages Lustre: 23863:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff8802b4b51f80 x1709865933156736/t0(0) o10->lustre-OST0000-osc-ffff880246d712a8@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630655928 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dir_create.sh.0' LustreError: 27055:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0002_UUID lock: ffff880099e08f40/0x6d2fd72113ee1205 lrc: 3/0,0 mode: PW/PW res: [0x340000400:0x947:0x0].0x0 rrc: 4 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000000020020 nid: 0@lo remote: 0x6d2fd72113ee11c6 expref: 905 pid: 32331 timeout: 4330 lvb_type: 0 LustreError: 27055:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 1 previous similar message LustreError: 32508:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630655914 with bad export cookie 7867743611107543471 LustreError: 11-0: lustre-OST0002-osc-ffff8802157b0958: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-OST0002-osc-ffff8802157b0958: Connection to lustre-OST0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 23860:0:(osc_request.c:1035:osc_init_grant()) lustre-OST0002-osc-ffff8802157b0958: granted 8437760 but already consumed 11702272 LustreError: 28423:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802b9e01f80 x1709865933613184/t0(0) o104->lustre-OST0002@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 28423:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 31 previous similar messages LustreError: 167-0: lustre-OST0002-osc-ffff8802157b0958: This client was evicted by lustre-OST0002; in progress operations using this service will fail. ptlrpc_watchdog_fire: 82 callbacks suppressed Lustre: ll_ost03_040: service thread pid 12525 was inactive for 62.068 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: ll_ost03_001: service thread pid 27759 was inactive for 62.054 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Lustre: Skipped 1 previous similar message Pid: 27759, comm: ll_ost03_001 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] tgt_extent_lock+0xea/0x2a0 [ptlrpc] [<0>] ofd_getattr_hdl+0x365/0x720 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 12498, comm: ll_ost03_038 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] tgt_extent_lock+0xea/0x2a0 [ptlrpc] [<0>] ofd_getattr_hdl+0x365/0x720 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 3676, comm: ll_ost03_016 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] tgt_extent_lock+0xea/0x2a0 [ptlrpc] [<0>] ofd_getattr_hdl+0x365/0x720 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: 23862:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.125@tcp:/lustre/fid: [0x280000403:0x1c6:0x0]// may get corrupted (rc -108) Lustre: 23861:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.125@tcp:/lustre/fid: [0x280000405:0xca:0x0]// may get corrupted (rc -108) Lustre: 23862:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.125@tcp:/lustre/fid: [0x200000404:0x2e3:0x0]// may get corrupted (rc -108) LustreError: 18114:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0002-osc-ffff8802157b0958: namespace resource [0x340000401:0x6a8:0x0].0x0 (ffff8802b1bfa5c0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 18114:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 34 previous similar messages Lustre: lustre-OST0002-osc-ffff8802157b0958: Connection restored to 192.168.123.125@tcp (at 0@lo) Lustre: Skipped 2 previous similar messages LustreError: 27086:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000404:0x61:0x0]: rc = -2 Lustre: 23872:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.125@tcp:/lustre/fid: [0x280000403:0x143:0x0]/ may get corrupted (rc -108) Lustre: lustre-OST0003-osc-ffff880246d712a8: Connection restored to 192.168.123.125@tcp (at 0@lo) LustreError: 20106:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff8802157b0958: inode [0x280000403:0x143:0x0] mdc close failed: rc = -13 Lustre: 23873:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff8802bb5eb880 x1709865937054144/t0(0) o10->lustre-OST0002-osc-ffff880246d712a8@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630656034 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'truncate.0' Lustre: 23873:0:(client.c:1485:after_reply()) Skipped 1 previous similar message LustreError: 27097:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000403:0x313:0x0]: rc = -2 truncate (20967) used greatest stack depth: 10176 bytes left mv (20143) used greatest stack depth: 10000 bytes left LustreError: 27055:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0003_UUID lock: ffff8802a76070c0/0x6d2fd72113f89905 lrc: 3/0,0 mode: PW/PW res: [0x1b48:0x0:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000080020020 nid: 0@lo remote: 0x6d2fd72113f898f0 expref: 2867 pid: 32146 timeout: 4495 lvb_type: 0 LustreError: 27055:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 1 previous similar message LustreError: 4718:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630656079 with bad export cookie 7867743611108406039 LustreError: 4718:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) Skipped 2 previous similar messages LustreError: 11-0: lustre-OST0003-osc-ffff880246d712a8: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: 32654:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802b8b68040 x1709865942472192/t0(0) o105->lustre-OST0003@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 32654:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 21 previous similar messages LustreError: Skipped 1 previous similar message Lustre: lustre-OST0003-osc-ffff880246d712a8: Connection to lustre-OST0003 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 1 previous similar message LustreError: 167-0: lustre-OST0003-osc-ffff880246d712a8: This client was evicted by lustre-OST0003; in progress operations using this service will fail. LustreError: Skipped 1 previous similar message Lustre: dir [0x200000403:0x308:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 14 previous similar messages LustreError: 22254:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0003-osc-ffff880246d712a8: namespace resource [0x1bd0:0x0:0x0].0x0 (ffff8800a316f9c0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 22254:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 1 previous similar message Lustre: lustre-OST0003-osc-ffff880246d712a8: Connection restored to 192.168.123.125@tcp (at 0@lo) LustreError: 30515:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0001: failed to get lu_attr of [0x240000404:0x2a3:0x0]: rc = -2 LustreError: 30515:0:(mdd_object.c:3461:mdd_close()) Skipped 1 previous similar message Lustre: ll_ost07_025: service thread pid 5459 was inactive for 40.047 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 14 previous similar messages Lustre: 23867:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.125@tcp:/lustre/fid: [0x280000405:0x1c7:0x0]/ may get corrupted (rc -108) Lustre: 23868:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.125@tcp:/lustre/fid: [0x240000404:0x2d9:0x0]/ may get corrupted (rc -108) Lustre: 23867:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.125@tcp:/lustre/fid: [0x200000404:0x3ac:0x0]// may get corrupted (rc -108) LustreError: 27774:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 233602240 != fo_tot_granted 235658432 LustreError: 27774:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 241 previous similar messages LustreError: 27782:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 2056192 LustreError: 27782:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 244 previous similar messages Lustre: dir [0x240000404:0x2ba:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: dir [0x200000403:0x398:0x0] stripe 1 readdir failed: -2, directory is partially accessed! LustreError: 27055:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0003_UUID lock: ffff8800a33ecf00/0x6d2fd7211405bed1 lrc: 3/0,0 mode: PW/PW res: [0x380000400:0x1192:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000000030020 nid: 0@lo remote: 0x6d2fd7211405beb5 expref: 3983 pid: 19657 timeout: 4687 lvb_type: 0 LustreError: 27055:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 1 previous similar message Lustre: 23873:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.125@tcp:/lustre/fid: [0x200000404:0x3ac:0x0]// may get corrupted (rc -5) Lustre: 23873:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.125@tcp:/lustre/fid: [0x200000404:0x40c:0x0]// may get corrupted (rc -5) Lustre: 23873:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.125@tcp:/lustre/fid: [0x200000404:0x32f:0x0]// may get corrupted (rc -5) LustreError: 16336:0:(ofd_obd.c:501:ofd_destroy_export()) lustre-OST0001: cli efb7de0b-f08a-47d7-895a-a686c416123c/ffff88007a3392a8 has 28672 pending on destroyed export LustreError: 16336:0:(ofd_obd.c:501:ofd_destroy_export()) Skipped 1 previous similar message Lustre: 27026:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 6s req@ffff88029e124500 x1709865951924352/t0(0) o105->lustre-MDT0001-mdtlov_UUID@0@lo:0/0 lens 336/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'' LustreError: 25651:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0003-osc-ffff880246d712a8: namespace resource [0x380000400:0x1da0:0x0].0x0 (ffff8800942887c0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 25651:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 507 previous similar messages Lustre: 23875:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.125@tcp:/lustre/fid: [0x240000403:0x537:0x0]// may get corrupted (rc -108) Lustre: lustre-OST0003-osc-ffff880246d712a8: Connection restored to 192.168.123.125@tcp (at 0@lo) Lustre: Skipped 2 previous similar messages LustreError: 30613:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630656290 with bad export cookie 7867743611109200469 Lustre: 30984:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 6s req@ffff8802d80eb240 x1709865955384832/t0(0) o106->LOV_OSC_UUID@0@lo:0/0 lens 328/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'' 2[26428]: segfault at 8 ip 00007f367294a7e8 sp 00007ffee6ed5f40 error 4 in ld-2.17.so[7f367293f000+22000] Lustre: 882:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 511 < left 17855, rollback = 7 Lustre: 882:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 272302 previous similar messages Lustre: 882:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 415/1660/0, destroy: 1/4/1 Lustre: 882:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 273371 previous similar messages Lustre: 882:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 3/3/0, xattr_set: 419/148/0 Lustre: 882:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 273370 previous similar messages Lustre: 882:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 2076/17855/0, punch: 0/0/0, quota 10/90/0 Lustre: 882:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 273372 previous similar messages Lustre: 882:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 416/7071/0, delete: 2/5/0 Lustre: 882:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 273372 previous similar messages Lustre: 882:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 1/1/0, ref_del: 2/2/0 Lustre: 882:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 273369 previous similar messages Lustre: 29830:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 504 < left 610, rollback = 2 Lustre: 29830:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 1066 previous similar messages LustreError: 1695:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0002: '3' migration was interrupted, run 'lfs migrate -m 0 -c 1 -H crush 3' to finish migration. LustreError: 11280:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630656459 with bad export cookie 7867743611106902782 LustreError: 24684:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff880065beac00 x1709865960633600/t0(0) o106->lustre-OST0000@0@lo:15/16 lens 328/280 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 24684:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 249 previous similar messages LustreError: 11280:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) Skipped 1 previous similar message LustreError: 11-0: lustre-OST0000-osc-ffff8802157b0958: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 5 previous similar messages Lustre: lustre-OST0000-osc-ffff8802157b0958: Connection to lustre-OST0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 5 previous similar messages LustreError: 167-0: lustre-OST0000-osc-ffff8802157b0958: This client was evicted by lustre-OST0000; in progress operations using this service will fail. LustreError: Skipped 5 previous similar messages LustreError: 30142:0:(ofd_obd.c:501:ofd_destroy_export()) lustre-OST0000: cli 8f82a635-3fbb-4c4d-a2ce-769d96ea45a7/ffff880276ac53d8 has 2121728 pending on destroyed export ptlrpc_watchdog_fire: 9 callbacks suppressed Lustre: ll_ost06_018: service thread pid 32125 was inactive for 40.067 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Lustre: Skipped 2 previous similar messages Pid: 32125, comm: ll_ost06_018 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] tgt_extent_lock+0xea/0x2a0 [ptlrpc] [<0>] ofd_getattr_hdl+0x365/0x720 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 20048, comm: ll_ost06_046 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] tgt_extent_lock+0xea/0x2a0 [ptlrpc] [<0>] ofd_getattr_hdl+0x365/0x720 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 25272, comm: ll_ost06_059 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] tgt_extent_lock+0xea/0x2a0 [ptlrpc] [<0>] ofd_getattr_hdl+0x365/0x720 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe 18[29855]: segfault at 8 ip 00007f537538c7e8 sp 00007ffde2bf3480 error 4 in ld-2.17.so[7f5375381000+22000] LustreError: 27055:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0002_UUID lock: ffff8802e89cb4c0/0x6d2fd7211415da02 lrc: 3/0,0 mode: PW/PW res: [0x340000401:0x1396:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 1048576->18446744073709551615) gid 0 flags: 0x60000400020020 nid: 0@lo remote: 0x6d2fd7211415d9fb expref: 2506 pid: 10952 timeout: 4945 lvb_type: 0 LustreError: 27055:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 4 previous similar messages LustreError: 30520:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630656529 with bad export cookie 7867743611109815363 Lustre: 23871:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.125@tcp:/lustre/fid: [0x200000404:0x44f:0x0]/ may get corrupted (rc -108) Lustre: 23872:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.125@tcp:/lustre/fid: [0x280000405:0x2b5:0x0]// may get corrupted (rc -108) LustreError: 31352:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0003-osc-ffff880246d712a8: namespace resource [0x380000401:0x153f:0x0].0x0 (ffff8802d74b2840) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 31352:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 2167 previous similar messages Lustre: lustre-OST0002-osc-ffff880246d712a8: Connection restored to 192.168.123.125@tcp (at 0@lo) Lustre: Skipped 4 previous similar messages Lustre: dir [0x200000403:0x461:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 4 previous similar messages LustreError: 27089:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0001: failed to get lu_attr of [0x240000403:0x588:0x0]: rc = -2 LustreError: 27089:0:(mdd_object.c:3461:mdd_close()) Skipped 1 previous similar message Lustre: ll_ost04_049: service thread pid 4114 was inactive for 40.102 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 11 previous similar messages LustreError: 27777:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 124078272 != fo_tot_granted 125937856 LustreError: 27777:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 704 previous similar messages LustreError: 27777:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 28672 != fo_tot_pending 1888256 LustreError: 27777:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 708 previous similar messages Lustre: 29539:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1630656602/real 1630656602] req@ffff8802b74c70c0 x1709865968398080/t0(0) o101->lustre-OST0003-osc-ffff8802157b0958@0@lo:28/4 lens 328/400 e 1 to 1 dl 1630656665 ref 2 fl Rpc:XQr/0/ffffffff rc 0/-1 job:'cp.0' LustreError: 1521:0:(service.c:2290:ptlrpc_server_handle_request()) @@@ Dropping timed-out request from 12345-0@lo: deadline 45/31s ago req@ffff8802d7a9f700 x1709865968398080/t0(0) o101->8f82a635-3fbb-4c4d-a2ce-769d96ea45a7@0@lo:97/0 lens 328/0 e 1 to 0 dl 1630656647 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'cp.0' LustreError: 1521:0:(service.c:2290:ptlrpc_server_handle_request()) Skipped 1 previous similar message Lustre: 1521:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (45/31s); client may timeout req@ffff8802d7a9f700 x1709865968398080/t0(0) o101->8f82a635-3fbb-4c4d-a2ce-769d96ea45a7@0@lo:97/0 lens 328/0 e 1 to 0 dl 1630656647 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'cp.0' Lustre: 1521:0:(service.c:2328:ptlrpc_server_handle_request()) Skipped 1 previous similar message Lustre: 23863:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1630656609/real 1630656609] req@ffff8802e82d9940 x1709865968718400/t0(0) o2->lustre-OST0002-osc-ffff8802157b0958@0@lo:28/4 lens 440/432 e 0 to 1 dl 1630656694 ref 1 fl Rpc:XQr/0/ffffffff rc 0/-1 job:'chmod.0' Lustre: 23863:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1630656610/real 1630656610] req@ffff8802b2034500 x1709865968719936/t0(0) o2->lustre-OST0002-osc-ffff8802157b0958@0@lo:28/4 lens 440/432 e 0 to 1 dl 1630656695 ref 1 fl Rpc:XQr/0/ffffffff rc 0/-1 job:'chmod.0' Lustre: 23863:0:(client.c:2285:ptlrpc_expire_one_request()) Skipped 5 previous similar messages LustreError: 11995:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630656706 with bad export cookie 7867743611110585083 LustreError: 11995:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) Skipped 10 previous similar messages LustreError: 32335:0:(service.c:2290:ptlrpc_server_handle_request()) @@@ Dropping timed-out request from 12345-0@lo: deadline 86/16s ago req@ffff8802a2a88680 x1709865968470144/t0(0) o101->8f82a635-3fbb-4c4d-a2ce-769d96ea45a7@0@lo:140/0 lens 328/0 e 0 to 0 dl 1630656690 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'cp.0' Lustre: 32335:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (86/16s); client may timeout req@ffff8802a2a88680 x1709865968470144/t0(0) o101->8f82a635-3fbb-4c4d-a2ce-769d96ea45a7@0@lo:140/0 lens 328/0 e 0 to 0 dl 1630656690 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'cp.0' LustreError: 23570:0:(ofd_obd.c:501:ofd_destroy_export()) lustre-OST0001: cli 8f82a635-3fbb-4c4d-a2ce-769d96ea45a7/ffff8802698b9bf8 has 2273280 pending on destroyed export LustreError: 29278:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff8802157b0958: inode [0x200000404:0x467:0x0] mdc close failed: rc = -5 LustreError: 763:0:(file.c:5042:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000007:0x1:0x0] error: rc = -108 Lustre: lustre-OST0001-osc-ffff8802157b0958: disconnect after 31s idle LustreError: 3098:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0002: '3' migration was interrupted, run 'lfs migrate -m 0 -c 1 -H crush 3' to finish migration. LustreError: 29961:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630656776 with bad export cookie 7867743611110852392 Lustre: 23863:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.125@tcp:/lustre/fid: [0x280000405:0x2ed:0x0]// may get corrupted (rc -108) ------------[ cut here ]------------ WARNING: CPU: 15 PID: 29594 at /home/green/git/lustre-release/ldiskfs/ext4_jbd2.c:266 __ldiskfs_handle_dirty_metadata+0x1d2/0x230 [ldiskfs] Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) dm_flakey dm_mod loop zfs(PO) zunicode(PO) zzstd(O) zlua(O) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) jbd2 mbcache crc_t10dif crct10dif_generic sb_edac edac_core iosf_mbi crc32_pclmul ghash_clmulni_intel aesni_intel lrw gf128mul glue_helper ablk_helper cryptd virtio_balloon pcspkr virtio_console i2c_piix4 ip_tables rpcsec_gss_krb5 drm_kms_helper ata_generic ttm crct10dif_pclmul pata_acpi drm crct10dif_common ata_piix drm_panel_orientation_quirks crc32c_intel serio_raw virtio_blk libata i2c_core floppy [last unloaded: libcfs] CPU: 15 PID: 29594 Comm: mdt07_003 Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011 Call Trace: [<ffffffff817d9418>] dump_stack+0x19/0x1b [<ffffffff8108d538>] __warn+0xd8/0x100 [<ffffffff8108d67d>] warn_slowpath_null+0x1d/0x20 [<ffffffffa0d86262>] __ldiskfs_handle_dirty_metadata+0x1d2/0x230 [ldiskfs] [<ffffffffa0d6633d>] ldiskfs_getblk+0x13d/0x210 [ldiskfs] [<ffffffffa0d66437>] ldiskfs_bread+0x27/0xe0 [ldiskfs] [<ffffffffa0cb1e97>] osd_write+0x4e7/0xcb0 [osd_ldiskfs] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffffa04de6f3>] dt_record_write+0x33/0x120 [obdclass] [<ffffffffa049d6b2>] llog_osd_write_rec+0x1052/0x19a0 [obdclass] [<ffffffffa048a760>] llog_write_rec+0x290/0x590 [obdclass] [<ffffffffa04904ee>] llog_cat_add_rec+0x23e/0x950 [obdclass] [<ffffffffa04875ff>] llog_add+0x17f/0x1f0 [obdclass] [<ffffffffa04b8a39>] ? lprocfs_counter_add+0xf9/0x160 [obdclass] [<ffffffffa082160d>] sub_updates_write+0xb6b/0xe32 [ptlrpc] [<ffffffffa07f1892>] top_trans_stop+0x7c2/0xf60 [ptlrpc] [<ffffffffa0f94f3c>] lod_trans_stop+0x25c/0x340 [lod] [<ffffffffa0e69408>] mdd_trans_stop+0x28/0x16e [mdd] [<ffffffffa0e4ab2f>] mdd_migrate_object+0xf1f/0x1ba0 [mdd] [<ffffffffa04dd375>] ? lu_object_find_at+0x285/0xb10 [obdclass] [<ffffffffa0e4bb86>] mdd_migrate+0x3d6/0x8e0 [mdd] [<ffffffffa0f1d0d8>] mdo_migrate+0x4c/0x4e [mdt] [<ffffffffa0edce53>] mdt_reint_migrate+0x1023/0x11d0 [mdt] [<ffffffffa0edd087>] mdt_reint_rec+0x87/0x240 [mdt] [<ffffffffa0eb255c>] mdt_reint_internal+0x76c/0xb40 [mdt] [<ffffffffa0ebf657>] mdt_reint+0x67/0x150 [mdt] [<ffffffffa07e026e>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<ffffffffa033045e>] ? libcfs_nid2str_r+0xfe/0x130 [lnet] [<ffffffffa078b990>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<ffffffffa078d529>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffffa078c950>] ? ptlrpc_wait_event+0x620/0x620 [ptlrpc] [<ffffffff810ba0f4>] kthread+0xe4/0xf0 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 ---[ end trace be1905b362903f58 ]--- LDISKFS-fs: ldiskfs_getblk:888: aborting transaction: error 28 in __ldiskfs_handle_dirty_metadata LDISKFS-fs error (device dm-2): ldiskfs_getblk:888: inode #176: block 53737: comm mdt07_003: journal_dirty_metadata failed: handle type 0 started at line 1972, credits 512/0, errcode -28 Aborting journal on device dm-2-8. LDISKFS-fs (dm-2): Remounting filesystem read-only LustreError: 29594:0:(osd_io.c:2103:osd_ldiskfs_write_record()) lustre-MDT0002: error reading offset 430080 (block 105, size 48, offs 430072), credits 0/0: rc = -28 LustreError: 29594:0:(update_trans.c:985:top_trans_stop()) lustre-MDT0002-osd: write updates failed: rc = -28 LustreError: 29594:0:(osd_io.c:2110:osd_ldiskfs_write_record()) journal_get_write_access() returned error -30 LustreError: 29594:0:(tgt_lastrcvd.c:1253:tgt_add_reply_data()) lustre-MDT0002: can't update reply_data file: rc = -30 LustreError: 29594:0:(osd_handler.c:2075:osd_trans_stop()) lustre-MDT0002: failed in transaction hook: rc = -30 LDISKFS-fs error (device dm-2) in osd_trans_stop:2082: error 28 LustreError: 27992:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-OST0000-osc-MDT0002: fail to cancel 138 llog-records: rc = -30 LustreError: 29594:0:(osd_handler.c:2085:osd_trans_stop()) lustre-MDT0002: failed to stop transaction: rc = -28 LustreError: 28830:0:(osp_sync.c:1094:osp_sync_process_committed()) lustre-OST0003-osc-MDT0002: can't cancel 156 records: rc = -30 LustreError: 29594:0:(update_trans.c:1011:top_trans_stop()) lustre-MDT0002-osd: stop trans failed: rc = -30 LustreError: 27536:0:(osd_handler.c:1779:osd_trans_commit_cb()) transaction @0xffff8802d0109978 commit error: 2 LustreError: 27992:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) Skipped 4 previous similar messages LDISKFS-fs error (device dm-2) in osd_trans_stop:2082: IO failure LustreError: 27555:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0002-osd: fail to cancel 1 of 1 llog-records: rc = -30 LustreError: 2536:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -5 LustreError: 2256:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -30 Lustre: lustre-OST0002-osc-ffff880246d712a8: disconnect after 24s idle Lustre: Skipped 1 previous similar message LustreError: 1533:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -30 LustreError: 1533:0:(llite_lib.c:1836:ll_md_setattr()) Skipped 6 previous similar messages LustreError: 22989:0:(llog_cat.c:602:llog_cat_add_rec()) llog_write_rec -116: lh=ffff88029e772478 LustreError: 22989:0:(update_trans.c:1062:top_trans_stop()) lustre-MDT0002-osp-MDT0001: write updates failed: rc = -116 LustreError: 28830:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-OST0003-osc-MDT0002: fail to cancel 48 llog-records: rc = -30 LustreError: 28830:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) Skipped 5 previous similar messages LustreError: 28830:0:(osp_sync.c:1094:osp_sync_process_committed()) lustre-OST0003-osc-MDT0002: can't cancel 48 records: rc = -30 LustreError: 28830:0:(osp_sync.c:1094:osp_sync_process_committed()) Skipped 5 previous similar messages LustreError: 27162:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0002-osp-MDT0000: fail to cancel 1 of 1 llog-records: rc = -2 LustreError: 27162:0:(llog_cat.c:787:llog_cat_cancel_records()) Skipped 3 previous similar messages LustreError: 21179:0:(llog_cat.c:602:llog_cat_add_rec()) llog_write_rec -116: lh=ffff88029e772478 LustreError: 21179:0:(update_trans.c:1062:top_trans_stop()) lustre-MDT0002-osp-MDT0001: write updates failed: rc = -116 LustreError: 2391:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -116 LustreError: 2391:0:(llite_lib.c:1836:ll_md_setattr()) Skipped 5 previous similar messages LustreError: 27354:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0002-osp-MDT0001: fail to cancel 1 llog-records: rc = -116 LustreError: 27354:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) Skipped 8 previous similar messages LustreError: 27354:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0000-osp-MDT0001: fail to cancel 1 of 1 llog-records: rc = -116 LustreError: 27354:0:(llog_cat.c:787:llog_cat_cancel_records()) Skipped 4 previous similar messages | Externally reported by onyx-68 boilpot email |
racer test 1: racer on clients: centos-40.localnet DURATION=2700 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa106be52>] ll_migrate+0x9b2/0xec0 [lustre] PGD 8000000293866067 PUD 2705ca067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: loop zfs(PO) zunicode(PO) zzstd(O) zlua(O) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) jbd2 mbcache lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) dm_flakey dm_mod libcfs(OE) crc_t10dif crct10dif_generic sb_edac edac_core iosf_mbi crc32_pclmul ghash_clmulni_intel aesni_intel lrw gf128mul glue_helper ablk_helper cryptd pcspkr virtio_console i2c_piix4 virtio_balloon ip_tables rpcsec_gss_krb5 ata_generic drm_kms_helper pata_acpi ttm drm crct10dif_pclmul crct10dif_common drm_panel_orientation_quirks ata_piix crc32c_intel virtio_blk serio_raw i2c_core libata floppy CPU: 11 PID: 21031 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011 task: ffff88024c450010 ti: ffff8802532f8000 task.ti: ffff8802532f8000 RIP: 0010:[<ffffffffa106be52>] [<ffffffffa106be52>] ll_migrate+0x9b2/0xec0 [lustre] RSP: 0018:ffff8802532fbbc8 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff880257f22bc8 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff8800b5d7d738 RDI: ffff8800b5d7d718 RBP: ffff8802532fbc48 R08: ffff8802787536d8 R09: ffff8802532fba60 R10: 0000000000000001 R11: 0000000000000005 R12: 0000000000000000 R13: ffff88025f291a88 R14: ffff88025ba74b18 R15: 0000000000000030 FS: 00007f4bdb022740(0000) GS:ffff880331cc0000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 CR2: 0000000000000008 CR3: 000000024bfda000 CR4: 00000000001607e0 Call Trace: [<ffffffff81242213>] ? __check_object_size+0x1c3/0x220 [<ffffffffa1059561>] ll_dir_ioctl+0x5d01/0x6ed0 [lustre] [<ffffffff811b515b>] ? unlock_page+0x2b/0x30 [<ffffffff811e5e2e>] ? do_read_fault.isra.64+0x12e/0x170 [<ffffffff811ea1cf>] ? handle_pte_fault+0x2ef/0xca0 [<ffffffff811ecf6d>] ? __handle_mm_fault+0x31d/0x720 [<ffffffff8125b3fd>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff81245a38>] ? vfs_write+0x178/0x230 [<ffffffff8125b751>] SyS_ioctl+0xa1/0xc0 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 | Lustre: lustre-OST0000-osc-ffff88029a7ac138: disconnect after 21s idle Lustre: 11052:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 512 < left 1022, rollback = 7 Lustre: 11052:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 11052:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 1/89/0 Lustre: 11052:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1013/1022/0, punch: 0/0/0, quota 3/3/0 Lustre: 11052:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 11052:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 11052:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 512 < left 1022, rollback = 7 Lustre: 11052:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 11052:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 1/89/0 Lustre: 11052:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1013/1022/0, punch: 0/0/0, quota 3/3/0 Lustre: 11052:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 11052:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 11052:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 512 < left 1022, rollback = 7 Lustre: 11052:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 254 previous similar messages Lustre: 11052:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 11052:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 254 previous similar messages Lustre: 11052:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 1/89/0 Lustre: 11052:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 254 previous similar messages Lustre: 11052:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1013/1022/0, punch: 0/0/0, quota 3/3/0 Lustre: 11052:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 254 previous similar messages Lustre: 11052:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 11052:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 254 previous similar messages Lustre: 11052:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 11052:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 254 previous similar messages Lustre: 11052:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 512 < left 1022, rollback = 7 Lustre: 11052:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 11052:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 1/89/0 Lustre: 11052:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1013/1022/0, punch: 0/0/0, quota 3/3/0 Lustre: 11052:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 11062:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 11062:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 702 previous similar messages Lustre: 16613:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 494 < left 55093, rollback = 7 Lustre: 16613:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 14737 previous similar messages Lustre: 16613:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 1281/5124/0, destroy: 0/0/0 Lustre: 16613:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 14738 previous similar messages Lustre: 16613:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 1284/222/18 Lustre: 16613:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 14741 previous similar messages Lustre: 16613:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 6406/55093/0, punch: 0/0/0, quota 6/6/0 Lustre: 16613:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 14739 previous similar messages Lustre: 16613:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 1281/21777/0, delete: 0/0/0 Lustre: 16613:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 14740 previous similar messages Lustre: 16613:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 16613:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 13815 previous similar messages Lustre: 11059:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 498 < left 688, rollback = 2 Lustre: 11059:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 3/12/7, destroy: 2/8/0 Lustre: 11059:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 5134 previous similar messages Lustre: 11059:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 7/688/0 Lustre: 11059:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 5134 previous similar messages Lustre: 11059:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 5/136/0, punch: 0/0/0, quota 10/106/0 Lustre: 11059:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 5134 previous similar messages Lustre: 11059:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 13/251/4, delete: 7/13/2 Lustre: 11059:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 5134 previous similar messages Lustre: 11059:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 6/6/0, ref_del: 7/7/1 Lustre: 11059:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 5134 previous similar messages sched: RT throttling activated Lustre: 11062:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 509 < left 866, rollback = 2 Lustre: 11062:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 5 previous similar messages LustreError: 17711:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000404:0x22:0x0]: rc = -2 Lustre: 14412:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 503 < left 1044, rollback = 2 Lustre: 14412:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 8 previous similar messages Lustre: 11072:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 505 < left 688, rollback = 2 Lustre: 11072:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 15 previous similar messages LustreError: 11068:0:(lustre_lmv.h:438:lmv_is_sane()) insane LMV: magic=0xcd40cd0 count=2 index=1 hash=crush:0x82000003 version=0 migrate offset=1 migrate hash=crush:3. LustreError: 11067:0:(lustre_lmv.h:438:lmv_is_sane()) insane LMV: magic=0xcd40cd0 count=2 index=1 hash=crush:0x82000003 version=0 migrate offset=1 migrate hash=crush:3. Lustre: 11065:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 505 < left 829, rollback = 7 Lustre: 11065:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 5134 previous similar messages Lustre: 11065:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 1/4/4, destroy: 0/0/0 Lustre: 11065:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 36 previous similar messages Lustre: 11065:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 4/209/0 Lustre: 11065:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 36 previous similar messages Lustre: 11065:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 820/829/0, punch: 0/0/0, quota 8/56/2 Lustre: 11065:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 36 previous similar messages Lustre: 11065:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 2/33/1, delete: 0/0/0 Lustre: 11065:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 36 previous similar messages Lustre: 11065:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 11065:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 36 previous similar messages Lustre: 16603:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 8: before 509 < left 13939, rollback = 9 Lustre: 16603:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 5 previous similar messages Lustre: 11090:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 509 < left 35227, rollback = 7 Lustre: 11090:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 820 previous similar messages LustreError: 13284:0:(lustre_lmv.h:438:lmv_is_sane()) insane LMV: magic=0xcd40cd0 count=2 index=1 hash=crush:0x82000003 version=0 migrate offset=1 migrate hash=crush:3. Lustre: 11090:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 819/3276/0, destroy: 1/4/1 Lustre: 11090:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 1069 previous similar messages Lustre: 11090:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 819/0/0 Lustre: 11090:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 1178 previous similar messages Lustre: 11090:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 4096/35227/11, punch: 0/0/0, quota 6/6/0 Lustre: 11090:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 1249 previous similar messages Lustre: 11090:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 819/13923/0, delete: 2/5/1 Lustre: 11090:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 1250 previous similar messages Lustre: 11090:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 1/1/1 Lustre: 11090:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 1274 previous similar messages ------------[ cut here ]------------ WARNING: CPU: 4 PID: 11090 at /home/green/git/lustre-release/ldiskfs/namei.c:3331 ldiskfs_orphan_add+0x11e/0x2a0 [ldiskfs] Modules linked in: loop zfs(PO) zunicode(PO) zzstd(O) zlua(O) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) jbd2 mbcache lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) dm_flakey dm_mod libcfs(OE) crc_t10dif crct10dif_generic sb_edac edac_core iosf_mbi crc32_pclmul ghash_clmulni_intel aesni_intel lrw gf128mul glue_helper ablk_helper cryptd pcspkr virtio_console i2c_piix4 virtio_balloon ip_tables rpcsec_gss_krb5 ata_generic drm_kms_helper pata_acpi ttm drm crct10dif_pclmul crct10dif_common drm_panel_orientation_quirks ata_piix crc32c_intel virtio_blk serio_raw i2c_core libata floppy CPU: 4 PID: 11090 Comm: mdt_rdpg02_001 Tainted: P OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011 Call Trace: [<ffffffff817d9418>] dump_stack+0x19/0x1b [<ffffffff8108d538>] __warn+0xd8/0x100 [<ffffffff8108d67d>] warn_slowpath_null+0x1d/0x20 [<ffffffffa0c034ae>] ldiskfs_orphan_add+0x11e/0x2a0 [ldiskfs] [<ffffffffa0bf49eb>] ldiskfs_xattr_inode_orphan_add.constprop.17+0xbb/0x110 [ldiskfs] [<ffffffffa0bf7417>] ldiskfs_xattr_delete_inode+0x77/0x3c0 [ldiskfs] [<ffffffffa0c17aa0>] ? ldiskfs_evict_inode+0x190/0x670 [ldiskfs] [<ffffffffa0c17ad9>] ldiskfs_evict_inode+0x1c9/0x670 [ldiskfs] [<ffffffff812649af>] evict+0xaf/0x180 [<ffffffff81264de5>] iput+0xf5/0x180 [<ffffffffa0cacd08>] osd_object_delete+0x1f8/0x370 [osd_ldiskfs] [<ffffffffa04cf59d>] lu_object_free.isra.26+0xcd/0x1d0 [obdclass] [<ffffffffa04d3cf1>] lu_object_put+0xa1/0x480 [obdclass] [<ffffffffa0e64af4>] mdt_mfd_close+0x1f4/0x870 [mdt] [<ffffffffa0e6ae14>] mdt_close_internal+0xb4/0x240 [mdt] [<ffffffffa0e6b224>] mdt_close+0x284/0x850 [mdt] [<ffffffffa07d62d4>] ? tgt_request_preprocess.isra.27+0xf4/0x850 [ptlrpc] [<ffffffffa07d726e>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<ffffffffa036145e>] ? libcfs_nid2str_r+0xfe/0x130 [lnet] [<ffffffffa0782990>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<ffffffffa0784529>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffffa0783950>] ? ptlrpc_wait_event+0x620/0x620 [ptlrpc] [<ffffffff810ba0f4>] kthread+0xe4/0xf0 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 ---[ end trace b9bb1aeb4926296c ]--- Lustre: 11062:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 502 < left 610, rollback = 2 tsc: Refined TSC clocksource calibration: 2799.991 MHz Lustre: 16509:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 499 < left 1428, rollback = 2 Lustre: 16509:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 17 previous similar messages Lustre: ll_ost03_005: service thread pid 17532 was inactive for 62.079 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 12663, comm: ll_ost03_000 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] tgt_extent_lock+0xea/0x2a0 [ptlrpc] [<0>] ofd_getattr_hdl+0x365/0x720 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 17671, comm: ll_ost03_007 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] tgt_extent_lock+0xea/0x2a0 [ptlrpc] [<0>] ofd_getattr_hdl+0x365/0x720 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: ll_ost03_008: service thread pid 17675 was inactive for 62.181 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 2 previous similar messages Pid: 17532, comm: ll_ost03_005 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] tgt_extent_lock+0xea/0x2a0 [ptlrpc] [<0>] ofd_getattr_hdl+0x365/0x720 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: ll_ost03_019: service thread pid 19033 was inactive for 62.174 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 13 previous similar messages Lustre: ll_ost06_006: service thread pid 17622 was inactive for 64.025 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 4 previous similar messages LustreError: 10957:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0003_UUID lock: ffff88028a2a6580/0xde8200069090e8cd lrc: 3/0,0 mode: PW/PW res: [0x380000401:0x2:0x0].0x0 rrc: 5 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000480000020 nid: 0@lo remote: 0xde8200069090e8bf expref: 2124 pid: 17666 timeout: 446 lvb_type: 0 LustreError: 11-0: lustre-OST0003-osc-ffff8802987b2e98: operation ost_getattr to node 0@lo failed: rc = -107 Lustre: lustre-OST0003-osc-ffff8802987b2e98: Connection to lustre-OST0003 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0003-osc-ffff8802987b2e98: This client was evicted by lustre-OST0003; in progress operations using this service will fail. Lustre: ll_ost06_045: service thread pid 19177 was inactive for 67.240 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 39 previous similar messages Lustre: 4128:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.140@tcp:/lustre/fid: [0x240000404:0x83:0x0]// may get corrupted (rc -108) LustreError: 20693:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0003-osc-ffff8802987b2e98: namespace resource [0x20:0x0:0x0].0x0 (ffff880274dd9940) refcount nonzero (1) after lock cleanup; forcing cleanup. Lustre: 19182:0:(service.c:1378:ptlrpc_at_send_early_reply()) @@@ Already past deadline (-1s), not sending early reply. Consider increasing at_early_margin (5)? req@ffff88025e3bf700 x1709793828328896/t0(0) o1->096ce305-f0a6-4066-bcab-4fa0aee3b962@0@lo:449/0 lens 440/0 e 1 to 0 dl 1630586784 ref 2 fl New:/0/ffffffff rc 0/-1 job:'lfs.0' Lustre: 18915:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1630586743/real 1630586743] req@ffff88025d4fb880 x1709793829783424/t0(0) o101->lustre-MDT0002-mdc-ffff8802987b2e98@0@lo:12/10 lens 576/25184 e 0 to 1 dl 1630586797 ref 2 fl Rpc:XQr/2/ffffffff rc 0/-1 job:'cp.0' Lustre: lustre-MDT0002-mdc-ffff8802987b2e98: Connection to lustre-MDT0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: lustre-MDT0002: Client 096ce305-f0a6-4066-bcab-4fa0aee3b962 (at 0@lo) reconnecting Lustre: lustre-MDT0002-mdc-ffff8802987b2e98: Connection restored to 192.168.123.140@tcp (at 0@lo) Lustre: 4133:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1630586740/real 1630586741] req@ffff880297513ec0 x1709793829655936/t0(0) o101->lustre-OST0003-osc-ffff88029a7ac138@0@lo:28/4 lens 328/400 e 0 to 1 dl 1630586801 ref 1 fl Rpc:XQr/0/ffffffff rc 0/-1 job:'rm.0' Lustre: lustre-OST0003-osc-ffff88029a7ac138: Connection to lustre-OST0003 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: lustre-OST0003: Client aa2e35db-d012-4ae7-8ed1-fe69c47e17f6 (at 0@lo) reconnecting Lustre: lustre-OST0003-osc-ffff88029a7ac138: Connection restored to 192.168.123.140@tcp (at 0@lo) Lustre: 4133:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1630586741/real 1630586741] req@ffff8800846e6440 x1709793829690176/t0(0) o101->lustre-OST0001-osc-ffff88029a7ac138@0@lo:28/4 lens 328/400 e 0 to 1 dl 1630586802 ref 1 fl Rpc:XQr/0/ffffffff rc 0/-1 job:'rm.0' Lustre: lustre-OST0000: Client aa2e35db-d012-4ae7-8ed1-fe69c47e17f6 (at 0@lo) reconnecting Lustre: lustre-OST0000-osc-ffff88029a7ac138: Connection restored to 192.168.123.140@tcp (at 0@lo) Lustre: lustre-OST0002-osc-MDT0002: Connection to lustre-OST0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 6 previous similar messages LustreError: 10957:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0001_UUID lock: ffff88026e7afc00/0xde8200069092f2a8 lrc: 3/0,0 mode: PW/PW res: [0x300000401:0x65:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000480020020 nid: 0@lo remote: 0xde8200069092f28c expref: 2717 pid: 17842 timeout: 474 lvb_type: 0 LustreError: 19158:0:(service.c:2290:ptlrpc_server_handle_request()) @@@ Dropping timed-out request from 12345-0@lo: deadline 71/21s ago req@ffff8802869c0cc0 x1709793828335616/t0(0) o1->096ce305-f0a6-4066-bcab-4fa0aee3b962@0@lo:450/0 lens 440/0 e 1 to 0 dl 1630586785 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'lfs.0' Lustre: 17855:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (71/21s); client may timeout req@ffff88025e213880 x1709793828335872/t0(0) o1->096ce305-f0a6-4066-bcab-4fa0aee3b962@0@lo:450/0 lens 440/0 e 1 to 0 dl 1630586785 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'lfs.0' LustreError: 19158:0:(service.c:2290:ptlrpc_server_handle_request()) Skipped 28 previous similar messages LustreError: 19172:0:(service.c:2290:ptlrpc_server_handle_request()) @@@ Dropping timed-out request from 12345-0@lo: deadline 43/3s ago req@ffff8800776ebec0 x1709793830378752/t0(0) o101->096ce305-f0a6-4066-bcab-4fa0aee3b962@0@lo:468/0 lens 328/0 e 0 to 0 dl 1630586803 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'lfs.0' Lustre: 17826:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (43/3s); client may timeout req@ffff880250aff0c0 x1709793830379072/t0(0) o101->096ce305-f0a6-4066-bcab-4fa0aee3b962@0@lo:468/0 lens 328/0 e 0 to 0 dl 1630586803 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'lfs.0' Lustre: 17826:0:(service.c:2328:ptlrpc_server_handle_request()) Skipped 250 previous similar messages LustreError: 19172:0:(service.c:2290:ptlrpc_server_handle_request()) Skipped 140 previous similar messages Lustre: 19364:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 511 < left 1533, rollback = 7 Lustre: 19364:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 5770 previous similar messages Lustre: 19364:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 1/4/1, destroy: 0/0/0 Lustre: 19364:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 5558 previous similar messages Lustre: 19364:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 4/209/0 Lustre: 19364:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 5449 previous similar messages Lustre: 19364:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1524/1533/0, punch: 0/0/0, quota 8/80/0 Lustre: 19364:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 5378 previous similar messages Lustre: 19364:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 2/33/0, delete: 0/0/0 Lustre: 19364:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 5377 previous similar messages Lustre: 19364:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 19364:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 5353 previous similar messages LustreError: 11-0: lustre-OST0000-osc-ffff8802987b2e98: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 7 previous similar messages Lustre: lustre-OST0000-osc-ffff8802987b2e98: Connection to lustre-OST0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 2 previous similar messages LustreError: 167-0: lustre-OST0001-osc-ffff8802987b2e98: This client was evicted by lustre-OST0001; in progress operations using this service will fail. LustreError: 10957:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0002_UUID lock: ffff880288c6da40/0xde820006908fbe7b lrc: 3/0,0 mode: PW/PW res: [0x2:0x0:0x0].0x0 rrc: 2 type: EXT [0->18446744073709551615] (req 0->1048575) gid 0 flags: 0x60000400020020 nid: 0@lo remote: 0xde820006908fbe51 expref: 1945 pid: 17727 timeout: 478 lvb_type: 0 LustreError: 10957:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 9 previous similar messages LustreError: 10954:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff880275fbe440 x1709793831649664/t0(0) o105->lustre-OST0002@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 10922:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630586810 with bad export cookie 16033377651587489798 LustreError: 19102:0:(ldlm_lockd.c:1425:ldlm_handle_enqueue0()) ### lock on destroyed export ffff88028e08efc8 ns: filter-lustre-OST0002_UUID lock: ffff880076db5e00/0xde820006909876a9 lrc: 4/0,0 mode: PW/PW res: [0x340000401:0x23:0x0].0x0 rrc: 4 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x50000000020000 nid: 0@lo remote: 0xde82000690987655 expref: 1946 pid: 19102 timeout: 0 lvb_type: 0 LustreError: 20693:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0003-osc-ffff8802987b2e98: namespace resource [0x20:0x0:0x0].0x0 (ffff880274dd9940) refcount nonzero (1) after lock cleanup; forcing cleanup. Lustre: mdt01_001: service thread pid 11050 was inactive for 62.267 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 9 previous similar messages LustreError: 20693:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 497 previous similar messages Lustre: lustre-OST0003-osc-ffff8802987b2e98: Connection restored to 192.168.123.140@tcp (at 0@lo) Lustre: Skipped 6 previous similar messages INFO: task mv:16252 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mv D ffff88028c97c240 11664 16252 16037 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffffa106e0b1>] ll_layout_refresh+0x201/0x990 [lustre] [<ffffffffa10b9cfc>] vvp_io_fini+0x57c/0x710 [lustre] [<ffffffffa04e4455>] cl_io_fini+0x75/0x250 [obdclass] [<ffffffffa10b27a4>] cl_glimpse_size0+0xe4/0x280 [lustre] [<ffffffffa106cfdc>] ll_getattr_dentry+0x50c/0x9a0 [lustre] [<ffffffffa106d48e>] ll_getattr+0x1e/0x20 [lustre] [<ffffffff8124b596>] vfs_getattr+0x46/0x80 [<ffffffff8124b6c5>] vfs_fstatat+0x75/0xc0 [<ffffffff8124ba7e>] SYSC_newstat+0x2e/0x60 [<ffffffff817e8a76>] ? trace_do_page_fault+0x56/0x170 [<ffffffff8124bf4e>] SyS_newstat+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task mkdir:16588 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mkdir D ffff88029067b760 11296 16588 16235 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 Lustre: 4128:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.140@tcp:/lustre/fid: [0x200000404:0x52:0x0]/ may get corrupted (rc -108) [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffffa106e0b1>] ll_layout_refresh+0x201/0x990 [lustre] Lustre: 4130:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.140@tcp:/lustre/fid: [0x240000403:0xc0:0x0]// may get corrupted (rc -108) [<ffffffffa10b9cfc>] vvp_io_fini+0x57c/0x710 [lustre] Lustre: 4130:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.140@tcp:/lustre/fid: [0x240000404:0x32:0x0]/ may get corrupted (rc -108) [<ffffffffa04e4455>] cl_io_fini+0x75/0x250 [obdclass] [<ffffffffa10b27a4>] cl_glimpse_size0+0xe4/0x280 [lustre] [<ffffffffa106cfdc>] ll_getattr_dentry+0x50c/0x9a0 [lustre] [<ffffffffa106d48e>] ll_getattr+0x1e/0x20 [lustre] [<ffffffff8124b596>] vfs_getattr+0x46/0x80 [<ffffffff8124b6c5>] vfs_fstatat+0x75/0xc0 [<ffffffff8124ba7e>] SYSC_newstat+0x2e/0x60 [<ffffffff8126a634>] ? mntput+0x24/0x40 [<ffffffff81251532>] ? done_path_create+0x42/0x50 [<ffffffff81132fb6>] ? __audit_syscall_exit+0x1f6/0x2b0 [<ffffffff8124bf4e>] SyS_newstat+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task mv:16913 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mv D ffff8802902d4a80 11344 16913 16415 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffffa106e0b1>] ll_layout_refresh+0x201/0x990 [lustre] [<ffffffffa04b00c9>] ? class_handle2object+0xb9/0x200 [obdclass] [<ffffffffa10bd5c7>] vvp_io_init+0x357/0x490 [lustre] [<ffffffffa04e3e90>] cl_io_init0.isra.7+0x90/0x170 [obdclass] [<ffffffffa04e4033>] cl_io_init+0x43/0x80 [obdclass] [<ffffffffa10b275f>] cl_glimpse_size0+0x9f/0x280 [lustre] [<ffffffffa106cfdc>] ll_getattr_dentry+0x50c/0x9a0 [lustre] [<ffffffffa106d48e>] ll_getattr+0x1e/0x20 [lustre] [<ffffffff8124b596>] vfs_getattr+0x46/0x80 [<ffffffff8124b6c5>] vfs_fstatat+0x75/0xc0 [<ffffffff8124ba7e>] SYSC_newstat+0x2e/0x60 [<ffffffff817e8a76>] ? trace_do_page_fault+0x56/0x170 [<ffffffff8124bf4e>] SyS_newstat+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task ln:17095 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. ln D ffff88028bdb8010 11296 17095 16436 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffffa106e0b1>] ll_layout_refresh+0x201/0x990 [lustre] [<ffffffffa04b00c9>] ? class_handle2object+0xb9/0x200 [obdclass] [<ffffffffa10bd5c7>] vvp_io_init+0x357/0x490 [lustre] [<ffffffffa04e3e90>] cl_io_init0.isra.7+0x90/0x170 [obdclass] [<ffffffffa04e4033>] cl_io_init+0x43/0x80 [obdclass] [<ffffffffa10b275f>] cl_glimpse_size0+0x9f/0x280 [lustre] [<ffffffffa106cfdc>] ll_getattr_dentry+0x50c/0x9a0 [lustre] [<ffffffffa106d48e>] ll_getattr+0x1e/0x20 [lustre] [<ffffffff8124b596>] vfs_getattr+0x46/0x80 [<ffffffff8124b6c5>] vfs_fstatat+0x75/0xc0 [<ffffffff8124ba7e>] SYSC_newstat+0x2e/0x60 [<ffffffff817e8a76>] ? trace_do_page_fault+0x56/0x170 [<ffffffff8124bf4e>] SyS_newstat+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task chown:17628 blocked for more than 120 seconds. Lustre: 11065:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 501 < left 610, rollback = 2 Lustre: 11065:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 17 previous similar messages "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. chown D ffff880291648880 11296 17628 16096 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffffa106e0b1>] ll_layout_refresh+0x201/0x990 [lustre] [<ffffffff817e05ea>] ? __schedule+0x32a/0x7d0 [<ffffffffa10bd5c7>] vvp_io_init+0x357/0x490 [lustre] [<ffffffffa04e3e90>] cl_io_init0.isra.7+0x90/0x170 [obdclass] [<ffffffffa04e4033>] cl_io_init+0x43/0x80 [obdclass] [<ffffffffa10b275f>] cl_glimpse_size0+0x9f/0x280 [lustre] [<ffffffffa106cfdc>] ll_getattr_dentry+0x50c/0x9a0 [lustre] [<ffffffffa106d48e>] ll_getattr+0x1e/0x20 [lustre] [<ffffffff8124b596>] vfs_getattr+0x46/0x80 [<ffffffff8124b6c5>] vfs_fstatat+0x75/0xc0 [<ffffffff8124bb34>] SYSC_newfstatat+0x24/0x60 [<ffffffff817e8a76>] ? trace_do_page_fault+0x56/0x170 [<ffffffff8124bf6e>] SyS_newfstatat+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task truncate:17637 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. Lustre: 17810:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (93/11s); client may timeout req@ffff8802c542b838 x1709793828300608/t0(0) o1->096ce305-f0a6-4066-bcab-4fa0aee3b962@0@lo:471/0 lens 440/432 e 2 to 0 dl 1630586806 ref 1 fl Complete:/0/0 rc 0/0 job:'lfs.0' Lustre: 17810:0:(service.c:2328:ptlrpc_server_handle_request()) Skipped 61 previous similar messages truncate D ffff88028e39eb80 11872 17637 16160 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffffa106e0b1>] ll_layout_refresh+0x201/0x990 [lustre] [<ffffffffa10bd5c7>] vvp_io_init+0x357/0x490 [lustre] [<ffffffffa04e3e90>] cl_io_init0.isra.7+0x90/0x170 [obdclass] [<ffffffffa10c17f6>] ? pcc_io_init+0x56/0x160 [lustre] [<ffffffffa04e4033>] cl_io_init+0x43/0x80 [obdclass] [<ffffffffa10b2a92>] cl_setattr_ost+0x152/0x410 [lustre] [<ffffffffa1088915>] ll_setattr_raw+0x1125/0x1260 [lustre] [<ffffffffa1088ab3>] ll_setattr+0x63/0xc0 [lustre] [<ffffffff812661a9>] notify_change+0x309/0x4d0 [<ffffffff812436a5>] do_truncate+0x75/0xc0 [<ffffffff81243861>] vfs_truncate+0x171/0x1e0 [<ffffffff8124395c>] do_sys_truncate+0x8c/0xb0 [<ffffffff81243b0e>] SyS_truncate+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task cp:17790 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. cp D ffff880089a25c40 12720 17790 16699 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffffa106e0b1>] ll_layout_refresh+0x201/0x990 [lustre] [<ffffffffa04b00c9>] ? class_handle2object+0xb9/0x200 [obdclass] [<ffffffffa10bd5c7>] vvp_io_init+0x357/0x490 [lustre] [<ffffffffa04e3e90>] cl_io_init0.isra.7+0x90/0x170 [obdclass] [<ffffffffa04e4033>] cl_io_init+0x43/0x80 [obdclass] [<ffffffffa10b275f>] cl_glimpse_size0+0x9f/0x280 [lustre] [<ffffffffa106cfdc>] ll_getattr_dentry+0x50c/0x9a0 [lustre] [<ffffffffa106d48e>] ll_getattr+0x1e/0x20 [lustre] [<ffffffff8124b596>] vfs_getattr+0x46/0x80 [<ffffffff8124b6c5>] vfs_fstatat+0x75/0xc0 [<ffffffff8124ba7e>] SYSC_newstat+0x2e/0x60 [<ffffffff817e8a76>] ? trace_do_page_fault+0x56/0x170 [<ffffffff8124bf4e>] SyS_newstat+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 Lustre: mdt00_008: service thread pid 17037 was inactive for 66.185 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 2 previous similar messages Lustre: 17479:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 6s req@ffff88025ef3a5c0 x1709793831646784/t0(0) o10->aa2e35db-d012-4ae7-8ed1-fe69c47e17f6@0@lo:0/0 lens 440/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'cp.0' Lustre: 11219:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 7s req@ffff88027431de00 x1709793831714432/t0(0) o41->lustre-MDT0001-mdtlov_UUID@0@lo:0/0 lens 224/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'osp-pre-2-1.0' Lustre: mdt_out: This server is not able to keep up with request traffic (cpu-bound). Lustre: 18409:0:(service.c:1612:ptlrpc_at_check_timed()) earlyQ=1 reqQ=0 recA=0, svcEst=1, delay=0ms Lustre: 18409:0:(service.c:1378:ptlrpc_at_send_early_reply()) @@@ Already past deadline (-1s), not sending early reply. Consider increasing at_early_margin (5)? req@ffff88027431de00 x1709793831714432/t0(0) o41->lustre-MDT0001-mdtlov_UUID@0@lo:489/0 lens 224/0 e 0 to 0 dl 1630586824 ref 2 fl New:/0/ffffffff rc 0/-1 job:'osp-pre-2-1.0' Lustre: 18409:0:(service.c:1378:ptlrpc_at_send_early_reply()) Skipped 190 previous similar messages LustreError: 18409:0:(service.c:2290:ptlrpc_server_handle_request()) @@@ Dropping timed-out request from 12345-0@lo: deadline 6/1s ago req@ffff88027431de00 x1709793831714432/t0(0) o41->lustre-MDT0001-mdtlov_UUID@0@lo:489/0 lens 224/0 e 0 to 0 dl 1630586824 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'osp-pre-2-1.0' LustreError: 18409:0:(service.c:2290:ptlrpc_server_handle_request()) Skipped 31 previous similar messages Lustre: 18409:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (6/1s); client may timeout req@ffff88027431de00 x1709793831714432/t0(0) o41->lustre-MDT0001-mdtlov_UUID@0@lo:489/0 lens 224/0 e 0 to 0 dl 1630586824 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'osp-pre-2-1.0' Lustre: 4124:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1630586818/real 1630586818] req@ffff880274318040 x1709793831714432/t0(0) o41->lustre-MDT0002-osp-MDT0001@0@lo:24/4 lens 224/368 e 0 to 1 dl 1630586829 ref 1 fl Rpc:XQr/0/ffffffff rc 0/-1 job:'osp-pre-2-1.0' Lustre: 4124:0:(client.c:2285:ptlrpc_expire_one_request()) Skipped 62 previous similar messages Lustre: lustre-MDT0002-osp-MDT0001: Connection to lustre-MDT0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 1 previous similar message LustreError: 20960:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0001-osc-ffff8802987b2e98: namespace resource [0x40:0x0:0x0].0x0 (ffff880274dd9bc0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 20960:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 497 previous similar messages Lustre: lustre-OST0001-osc-ffff8802987b2e98: Connection restored to 192.168.123.140@tcp (at 0@lo) Lustre: lustre-MDT0002: Received new MDS connection from 0@lo, keep former export from same NID LustreError: 20962:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0002-osc-ffff8802987b2e98: namespace resource [0x20:0x0:0x0].0x0 (ffff880271da91c0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 20962:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 2061 previous similar messages Lustre: 19507:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 6s req@ffff88026cb15428 x1709793831830208/t0(0) o106->LOV_OSC_UUID@0@lo:0/0 lens 328/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'' Lustre: dir [0x240000403:0xad:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: mdt07_000: service thread pid 11072 was inactive for 62.013 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 11 previous similar messages LustreError: 10957:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0002_UUID lock: ffff88028f774000/0xde820006909518ec lrc: 4/0,0 mode: PR/PR res: [0x280000404:0x6:0x0].0x0 bits 0x1b/0x0 rrc: 6 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0xde8200069095189f expref: 78 pid: 11065 timeout: 517 lvb_type: 0 Lustre: lustre-MDT0002-mdc-ffff8802987b2e98: Connection to lustre-MDT0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-MDT0002-mdc-ffff8802987b2e98: This client was evicted by lustre-MDT0002; in progress operations using this service will fail. LustreError: Skipped 3 previous similar messages LustreError: 20393:0:(file.c:5042:ll_inode_revalidate_fini()) lustre: revalidate FID [0x280000404:0xa:0x0] error: rc = -5 LustreError: 20393:0:(file.c:5042:ll_inode_revalidate_fini()) Skipped 9 previous similar messages LustreError: 17458:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff8802987b2e98: inode [0x280000404:0x6:0x0] mdc close failed: rc = -108 LustreError: 19064:0:(vvp_io.c:1793:vvp_io_init()) lustre: refresh file layout [0x280000404:0x68:0x0] error -108. Lustre: 4134:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1630586803/real 1630586803] req@ffff880078ed6a80 x1709793831109632/t0(0) o101->lustre-OST0000-osc-ffff88029a7ac138@0@lo:28/4 lens 328/400 e 0 to 1 dl 1630586847 ref 1 fl Rpc:XQr/2/ffffffff rc -11/-1 job:'ln.0' Lustre: lustre-OST0000: Client aa2e35db-d012-4ae7-8ed1-fe69c47e17f6 (at 0@lo) reconnecting Lustre: Skipped 7 previous similar messages Lustre: lustre-OST0001-osc-ffff88029a7ac138: Connection restored to 192.168.123.140@tcp (at 0@lo) Lustre: Skipped 4 previous similar messages LustreError: 21036:0:(file.c:5042:ll_inode_revalidate_fini()) lustre: revalidate FID [0x280000404:0xa:0x0] error: rc = -108 LustreError: 21036:0:(file.c:5042:ll_inode_revalidate_fini()) Skipped 5 previous similar messages LustreError: 21183:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff8802987b2e98: inode [0x280000403:0x1:0x0] mdc close failed: rc = -108 LustreError: 21183:0:(file.c:234:ll_close_inode_openhandle()) Skipped 2 previous similar messages LustreError: 20506:0:(llite_lib.c:2983:ll_prep_inode()) new_inode -fatal: rc -108 Lustre: 4126:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff880274ec8cc0 x1709793832314432/t0(0) o10->lustre-OST0000-osc-ffff88029a7ac138@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630586863 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dd.0' Lustre: 4124:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff8800830c6440 x1709793832635520/t0(0) o10->lustre-OST0001-osc-ffff8802987b2e98@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630586881 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dd.0' Lustre: 21586:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 509 < left 65152, rollback = 7 Lustre: 21586:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 32355 previous similar messages Lustre: 11065:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 3: before 7 < left 5476, rollback = 4 Lustre: 11065:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 6 previous similar messages ------------[ cut here ]------------ WARNING: CPU: 11 PID: 11065 at /home/green/git/lustre-release/ldiskfs/ext4_jbd2.c:266 __ldiskfs_handle_dirty_metadata+0x1d2/0x230 [ldiskfs] Modules linked in: loop zfs(PO) zunicode(PO) zzstd(O) zlua(O) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) jbd2 mbcache lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) dm_flakey dm_mod libcfs(OE) crc_t10dif crct10dif_generic sb_edac edac_core iosf_mbi crc32_pclmul ghash_clmulni_intel aesni_intel lrw gf128mul glue_helper ablk_helper cryptd pcspkr virtio_console i2c_piix4 virtio_balloon ip_tables rpcsec_gss_krb5 ata_generic drm_kms_helper pata_acpi ttm drm crct10dif_pclmul crct10dif_common drm_panel_orientation_quirks ata_piix crc32c_intel virtio_blk serio_raw i2c_core libata floppy CPU: 11 PID: 11065 Comm: mdt05_001 Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011 Call Trace: [<ffffffff817d9418>] dump_stack+0x19/0x1b [<ffffffff8108d538>] __warn+0xd8/0x100 [<ffffffff8108d67d>] warn_slowpath_null+0x1d/0x20 [<ffffffffa0c32262>] __ldiskfs_handle_dirty_metadata+0x1d2/0x230 [ldiskfs] [<ffffffffa0c13b10>] ldiskfs_do_update_inode+0x4a0/0x830 [ldiskfs] [<ffffffffa0b532d2>] ? jbd2_journal_get_write_access+0x32/0x40 [jbd2] [<ffffffffa0c15538>] ldiskfs_mark_iloc_dirty+0x58/0x80 [ldiskfs] [<ffffffffa0c19410>] ? ldiskfs_dirty_inode+0x40/0x60 [ldiskfs] [<ffffffffa0c15686>] ldiskfs_mark_inode_dirty+0x86/0x2b0 [ldiskfs] [<ffffffffa0c193fa>] ? ldiskfs_dirty_inode+0x2a/0x60 [ldiskfs] [<ffffffffa0c19410>] ldiskfs_dirty_inode+0x40/0x60 [ldiskfs] [<ffffffffa0cd70f7>] osd_write+0x747/0xcb0 [osd_ldiskfs] [<ffffffffa04d56f3>] dt_record_write+0x33/0x120 [obdclass] [<ffffffffa04946b2>] llog_osd_write_rec+0x1052/0x19a0 [obdclass] [<ffffffffa0ca98bb>] ? osd_read_unlock+0x5b/0x90 [osd_ldiskfs] [<ffffffffa0481760>] llog_write_rec+0x290/0x590 [obdclass] [<ffffffffa0486b08>] llog_cat_new_log+0x6a8/0xe50 [obdclass] [<ffffffffa04874ca>] llog_cat_add_rec+0x21a/0x950 [obdclass] [<ffffffffa047e5ff>] llog_add+0x17f/0x1f0 [obdclass] [<ffffffffa04afa39>] ? lprocfs_counter_add+0xf9/0x160 [obdclass] [<ffffffffa081860d>] sub_updates_write+0xb6b/0xe32 [ptlrpc] [<ffffffff811beccd>] ? __free_pages+0x1d/0x30 [<ffffffff8121efe4>] ? kfree+0x1e4/0x200 Lustre: mdt05_010: service thread pid 19211 was inactive for 62.136 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. [<ffffffffa07e8892>] top_trans_stop+0x7c2/0xf60 [ptlrpc] [<ffffffffa0f11f3c>] lod_trans_stop+0x25c/0x340 [lod] [<ffffffffa0de6408>] mdd_trans_stop+0x28/0x16e [mdd] [<ffffffffa0dc7b2f>] mdd_migrate_object+0xf1f/0x1ba0 [mdd] [<ffffffffa04d4375>] ? lu_object_find_at+0x285/0xb10 [obdclass] [<ffffffffa0dc8b86>] mdd_migrate+0x3d6/0x8e0 [mdd] [<ffffffffa0e9a0d8>] mdo_migrate+0x4c/0x4e [mdt] [<ffffffffa0e59e53>] mdt_reint_migrate+0x1023/0x11d0 [mdt] [<ffffffffa0e5a087>] mdt_reint_rec+0x87/0x240 [mdt] [<ffffffffa0e2f55c>] mdt_reint_internal+0x76c/0xb40 [mdt] [<ffffffffa0e3c657>] mdt_reint+0x67/0x150 [mdt] [<ffffffffa07d726e>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<ffffffffa036145e>] ? libcfs_nid2str_r+0xfe/0x130 [lnet] [<ffffffffa0782990>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<ffffffffa0784529>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e32be>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffffa0783950>] ? ptlrpc_wait_event+0x620/0x620 [ptlrpc] [<ffffffff810ba0f4>] kthread+0xe4/0xf0 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 ---[ end trace b9bb1aeb4926296d ]--- LDISKFS-fs: ldiskfs_do_update_inode:4844: aborting transaction: error 28 in __ldiskfs_handle_dirty_metadata LDISKFS: jbd2_journal_dirty_metadata failed: handle type 0 started at line 1972, credits 512/0, errcode -28 LDISKFS-fs error (device dm-0) in ldiskfs_do_update_inode:4862: error 28 Aborting journal on device dm-0-8. LDISKFS-fs (dm-0): Remounting filesystem read-only LDISKFS-fs error (device dm-0) in ldiskfs_dirty_inode:5467: error 28 LDISKFS-fs warning (device dm-0): ldiskfs_mb_new_blocks:5106: Updating bitmap error: [err -30] [pa ffff8800b93dbad0] [phy 21976] [logic 21976] [len 296] [free 296] [error 1] [inode 0] LustreError: 11065:0:(osd_io.c:2103:osd_ldiskfs_write_record()) lustre-MDT0000: error reading offset 0 (block 0, size 32768, offs 0), credits 0/0: rc = -30 LustreError: 11065:0:(llog_cat.c:602:llog_cat_add_rec()) llog_write_rec -30: lh=ffff8802924c9738 LustreError: 11065:0:(update_trans.c:985:top_trans_stop()) lustre-MDT0000-osd: write updates failed: rc = -30 LustreError: 11065:0:(osd_io.c:2110:osd_ldiskfs_write_record()) journal_get_write_access() returned error -30 LustreError: 11065:0:(tgt_lastrcvd.c:1253:tgt_add_reply_data()) lustre-MDT0000: can't update reply_data file: rc = -30 LustreError: 11065:0:(osd_handler.c:2075:osd_trans_stop()) lustre-MDT0000: failed in transaction hook: rc = -30 LDISKFS-fs error (device dm-0) in osd_trans_stop:2082: error 28 LustreError: 13611:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-OST0001-osc-MDT0000: fail to cancel 58 llog-records: rc = -30 LustreError: 13611:0:(osp_sync.c:1094:osp_sync_process_committed()) lustre-OST0001-osc-MDT0000: can't cancel 58 records: rc = -30 LustreError: 10872:0:(osd_handler.c:1779:osd_trans_commit_cb()) transaction @0xffff8800832886f8 commit error: 2 LustreError: 11065:0:(osd_handler.c:2085:osd_trans_stop()) lustre-MDT0000: failed to stop transaction: rc = -28 LustreError: 11065:0:(update_trans.c:1011:top_trans_stop()) lustre-MDT0000-osd: stop trans failed: rc = -30 LustreError: 11065:0:(update_trans.c:1084:top_trans_stop()) lustre-OST0003-osc-MDT0000: stop trans failed: rc = -30 LustreError: 10872:0:(osd_handler.c:1779:osd_trans_commit_cb()) transaction @0xffff88024e888de8 commit error: 2 LustreError: 10947:0:(tgt_handler.c:1367:tgt_blocking_ast()) lustre-MDT0000: syncing [0x200000402:0x5:0x0] (19-0) on lock cancel: rc = -5 LustreError: 18339:0:(tgt_handler.c:1367:tgt_blocking_ast()) lustre-MDT0000: syncing [0x200000400:0x5:0x0] (19-0) on lock cancel: rc = -5 LustreError: 14193:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-OST0002-osc-MDT0000: fail to cancel 68 llog-records: rc = -30 LustreError: 14193:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) Skipped 5 previous similar messages LustreError: 14193:0:(osp_sync.c:1094:osp_sync_process_committed()) lustre-OST0002-osc-MDT0000: can't cancel 68 records: rc = -30 LustreError: 14193:0:(osp_sync.c:1094:osp_sync_process_committed()) Skipped 5 previous similar messages LustreError: 21200:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -30 LustreError: 14545:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-OST0003-osc-MDT0000: fail to cancel 1 llog-records: rc = -30 LustreError: 14545:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-OST0003-osc-MDT0000: fail to cancel 1 of 1 llog-records: rc = -30 LustreError: 14545:0:(osp_sync.c:1079:osp_sync_process_committed()) lustre-OST0003-osc-MDT0000: can't cancel record: rc = -30 LustreError: 16508:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0002: '16' migration was interrupted, run 'lfs migrate -m 0 -c 1 -H crush 16' to finish migration. LustreError: 21415:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -30 LustreError: 21531:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -30 LustreError: 21531:0:(llite_lib.c:1836:ll_md_setattr()) Skipped 1 previous similar message LustreError: 13611:0:(osp_sync.c:1094:osp_sync_process_committed()) lustre-OST0001-osc-MDT0000: can't cancel 110 records: rc = -30 LustreError: 13611:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-OST0001-osc-MDT0000: fail to cancel 39 llog-records: rc = -30 LustreError: 13611:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) Skipped 1 previous similar message LustreError: 14193:0:(osp_sync.c:1094:osp_sync_process_committed()) lustre-OST0002-osc-MDT0000: can't cancel 40 records: rc = -30 LustreError: 14193:0:(osp_sync.c:1094:osp_sync_process_committed()) Skipped 1 previous similar message | Externally reported by onyx-68 boilpot email |
racer test 1: racer on clients: centos-105.localnet DURATION=2700 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa105de52>] ll_migrate+0x9b2/0xec0 [lustre] PGD 80000000852c5067 PUD 7a188067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: loop zfs(PO) zunicode(PO) zzstd(O) zlua(O) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) jbd2 mbcache lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) dm_flakey dm_mod libcfs(OE) crc_t10dif crct10dif_generic sb_edac edac_core iosf_mbi crc32_pclmul ghash_clmulni_intel aesni_intel lrw gf128mul glue_helper ablk_helper cryptd virtio_balloon pcspkr i2c_piix4 virtio_console ip_tables rpcsec_gss_krb5 drm_kms_helper ata_generic ttm pata_acpi drm crct10dif_pclmul crct10dif_common drm_panel_orientation_quirks crc32c_intel ata_piix serio_raw virtio_blk i2c_core libata floppy CPU: 15 PID: 20221 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011 task: ffff8800800a9280 ti: ffff880293828000 task.ti: ffff880293828000 RIP: 0010:[<ffffffffa105de52>] [<ffffffffa105de52>] ll_migrate+0x9b2/0xec0 [lustre] RSP: 0018:ffff88029382bbc8 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff88026ab6cfa8 RCX: 0000000000000000 RDX: 0000000100000f62 RSI: ffff880325051838 RDI: ffff880325051818 RBP: ffff88029382bc48 R08: ffff88025d0bd338 R09: ffff88029382ba60 R10: 0000000000000003 R11: ffff8802629d1940 R12: 0000000000000000 R13: ffff880281e7c248 R14: ffff88007ad25d98 R15: 0000000000000030 FS: 00007fb5b7ec3740(0000) GS:ffff880331dc0000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 CR2: 0000000000000008 CR3: 0000000286622000 CR4: 00000000001607e0 Call Trace: [<ffffffff81242213>] ? __check_object_size+0x1c3/0x220 [<ffffffffa104b561>] ll_dir_ioctl+0x5d01/0x6ed0 [lustre] [<ffffffff811b515b>] ? unlock_page+0x2b/0x30 [<ffffffff811e5e2e>] ? do_read_fault.isra.64+0x12e/0x170 [<ffffffff811ea1cf>] ? handle_pte_fault+0x2ef/0xca0 [<ffffffff811ecf6d>] ? __handle_mm_fault+0x31d/0x720 [<ffffffff8125b3fd>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff81245a38>] ? vfs_write+0x178/0x230 [<ffffffff8125b751>] SyS_ioctl+0xa1/0xc0 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 | Lustre: lustre-OST0001-osc-ffff880298ec6678: disconnect after 21s idle Lustre: 11135:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 512 < left 922, rollback = 7 Lustre: 11135:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 11135:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 1/89/0 Lustre: 11135:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 913/922/0, punch: 0/0/0, quota 3/3/0 Lustre: 11135:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 11135:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 11135:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 511 < left 921, rollback = 7 Lustre: 11135:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 759 previous similar messages Lustre: 11135:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 11135:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 759 previous similar messages Lustre: 11135:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 1/89/0 Lustre: 11135:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 759 previous similar messages Lustre: 11135:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 913/922/1, punch: 0/0/0, quota 3/3/0 Lustre: 11135:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 759 previous similar messages Lustre: 11135:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 11135:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 759 previous similar messages Lustre: 11135:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 11135:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 759 previous similar messages Lustre: 13063:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 512 < left 1340, rollback = 7 Lustre: 13063:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 1567 previous similar messages Lustre: 13063:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 13063:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 1567 previous similar messages Lustre: 13063:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 1/89/0 Lustre: 13063:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 1567 previous similar messages Lustre: 13063:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1331/1340/0, punch: 0/0/0, quota 3/3/0 Lustre: 13063:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 1567 previous similar messages Lustre: 13063:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 13063:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 1567 previous similar messages Lustre: 13063:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 13063:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 1567 previous similar messages Lustre: 17040:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 506 < left 582, rollback = 2 Lustre: 17040:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 3/12/2, destroy: 0/0/0 Lustre: 17040:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 2260 previous similar messages Lustre: 17040:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 6/582/0 Lustre: 17040:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 2260 previous similar messages Lustre: 17040:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 5/136/0, punch: 0/0/0, quota 6/6/0 Lustre: 17040:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 2260 previous similar messages Lustre: 11150:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 512 < left 768, rollback = 7 Lustre: 11150:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 2260 previous similar messages Lustre: 11150:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 11150:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 2260 previous similar messages Lustre: 11150:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 11150:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 2260 previous similar messages Lustre: 17040:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 505 < left 582, rollback = 2 Lustre: 17091:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 8: before 511 < left 12935, rollback = 9 Lustre: 17091:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 4 previous similar messages Lustre: 17091:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 759/3036/0, destroy: 1/4/0 Lustre: 17091:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 764 previous similar messages Lustre: 17091:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 5/5/0, xattr_set: 767/283/0 Lustre: 17091:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 764 previous similar messages Lustre: 17091:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 3796/32647/0, punch: 0/0/0, quota 6/6/0 Lustre: 17091:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 393 previous similar messages Lustre: 17091:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 761/12935/0, delete: 3/6/1 Lustre: 17091:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 764 previous similar messages Lustre: 17091:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 1/1/0, ref_del: 1/1/0 Lustre: 17091:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 764 previous similar messages Lustre: 17091:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 507 < left 32647, rollback = 7 Lustre: 17091:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 758 previous similar messages ------------[ cut here ]------------ WARNING: CPU: 13 PID: 17091 at /home/green/git/lustre-release/ldiskfs/namei.c:3331 ldiskfs_orphan_add+0x11e/0x2a0 [ldiskfs] Modules linked in: loop zfs(PO) zunicode(PO) zzstd(O) zlua(O) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) jbd2 mbcache lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) dm_flakey dm_mod libcfs(OE) crc_t10dif crct10dif_generic sb_edac edac_core iosf_mbi crc32_pclmul ghash_clmulni_intel aesni_intel lrw gf128mul glue_helper ablk_helper cryptd virtio_balloon pcspkr i2c_piix4 virtio_console ip_tables rpcsec_gss_krb5 drm_kms_helper ata_generic ttm pata_acpi drm crct10dif_pclmul crct10dif_common drm_panel_orientation_quirks crc32c_intel ata_piix serio_raw virtio_blk i2c_core libata floppy CPU: 13 PID: 17091 Comm: mdt06_005 Tainted: P OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011 Call Trace: [<ffffffff817d9418>] dump_stack+0x19/0x1b [<ffffffff8108d538>] __warn+0xd8/0x100 [<ffffffff8108d67d>] warn_slowpath_null+0x1d/0x20 [<ffffffffa0bf54ae>] ldiskfs_orphan_add+0x11e/0x2a0 [ldiskfs] [<ffffffffa0be69eb>] ldiskfs_xattr_inode_orphan_add.constprop.17+0xbb/0x110 [ldiskfs] [<ffffffffa0be9417>] ldiskfs_xattr_delete_inode+0x77/0x3c0 [ldiskfs] [<ffffffffa0c09aa0>] ? ldiskfs_evict_inode+0x190/0x670 [ldiskfs] [<ffffffffa0c09ad9>] ldiskfs_evict_inode+0x1c9/0x670 [ldiskfs] [<ffffffff812649af>] evict+0xaf/0x180 [<ffffffff81264de5>] iput+0xf5/0x180 [<ffffffffa0c9ed08>] osd_object_delete+0x1f8/0x370 [osd_ldiskfs] [<ffffffffa04fe59d>] lu_object_free.isra.26+0xcd/0x1d0 [obdclass] [<ffffffffa0502cf1>] lu_object_put+0xa1/0x480 [obdclass] [<ffffffffa0e41144>] mdt_reint_rename+0x11f4/0x2ff0 [mdt] [<ffffffffa076cfb0>] ? ldlm_expired_completion_wait+0x2a0/0x2a0 [ptlrpc] [<ffffffffa0e34af2>] ? ucred_set_audit_enabled.isra.13+0x22/0x60 [mdt] [<ffffffffa0e4c087>] mdt_reint_rec+0x87/0x240 [mdt] [<ffffffffa0e2155c>] mdt_reint_internal+0x76c/0xb40 [mdt] [<ffffffffa0e2e657>] mdt_reint+0x67/0x150 [mdt] [<ffffffffa080626e>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<ffffffffa032745e>] ? libcfs_nid2str_r+0xfe/0x130 [lnet] [<ffffffffa07b1990>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<ffffffffa07b3529>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e32be>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffffa07b2950>] ? ptlrpc_wait_event+0x620/0x620 [ptlrpc] [<ffffffff810ba0f4>] kthread+0xe4/0xf0 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 ---[ end trace 9770dc902265b5ed ]--- Lustre: 18323:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 504 < left 1072, rollback = 2 Lustre: 18323:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 2/8/3, destroy: 1/4/0 Lustre: 18323:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 4541 previous similar messages Lustre: 18323:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 2/2/0, xattr_set: 13/1072/0 Lustre: 18323:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 4541 previous similar messages Lustre: 18323:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1/10/0, punch: 0/0/0, quota 8/104/0 Lustre: 18323:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 4541 previous similar messages Lustre: 18323:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 12/194/4, delete: 3/6/1 Lustre: 18323:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 4541 previous similar messages Lustre: 18323:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 7/7/0, ref_del: 3/3/0 Lustre: 18323:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 4541 previous similar messages sched: RT throttling activated Lustre: 11143:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 508 < left 60812, rollback = 7 Lustre: 11143:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 4540 previous similar messages Lustre: 17946:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 502 < left 671, rollback = 2 Lustre: 17946:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 11 previous similar messages Lustre: 17946:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 3/12/6, destroy: 0/0/0 Lustre: 17946:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 6506 previous similar messages Lustre: 17946:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 7/671/0 Lustre: 17946:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 6506 previous similar messages Lustre: 17946:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 5/136/0, punch: 0/0/0, quota 6/6/0 Lustre: 17946:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 6505 previous similar messages Lustre: 17946:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 10/203/4, delete: 0/0/0 Lustre: 17946:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 6506 previous similar messages Lustre: 17946:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 4/4/0, ref_del: 0/0/0 Lustre: 17946:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 6504 previous similar messages Lustre: 11127:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 509 < left 39269, rollback = 7 Lustre: 11127:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 6494 previous similar messages Lustre: ll_ost01_001: service thread pid 12697 was inactive for 40.151 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 12697, comm: ll_ost01_001 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] tgt_extent_lock+0xea/0x2a0 [ptlrpc] [<0>] ofd_getattr_hdl+0x365/0x720 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: 16866:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 8: before 511 < left 24054, rollback = 9 Lustre: 16866:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 7 previous similar messages Lustre: 16866:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 1414/5656/0, destroy: 1/4/0 Lustre: 16866:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 3669 previous similar messages Lustre: 16866:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 3/3/0, xattr_set: 1418/148/0 Lustre: 16866:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 3669 previous similar messages Lustre: 16866:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 7071/60812/0, punch: 0/0/0, quota 10/74/0 Lustre: 16866:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 3669 previous similar messages Lustre: 16866:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 1415/24054/0, delete: 2/5/0 Lustre: 16866:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 3669 previous similar messages Lustre: 16866:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 1/1/0, ref_del: 2/2/1 Lustre: 16866:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 3669 previous similar messages Lustre: 16983:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 505 < left 35313, rollback = 7 Lustre: 16983:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 3662 previous similar messages LustreError: 11016:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0000_UUID lock: ffff8802736e2200/0x4cdd18e9baca3cbe lrc: 3/0,0 mode: PW/PW res: [0x163:0x0:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000480000020 nid: 0@lo remote: 0x4cdd18e9baca3c86 expref: 3061 pid: 15503 timeout: 415 lvb_type: 0 LustreError: 18402:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88026625d180 x1709753732551936/t0(0) o105->lustre-OST0000@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 18343:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630548511 with bad export cookie 5538610508875765095 LustreError: 11-0: lustre-OST0000-osc-ffff880297dfdd28: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-OST0000-osc-ffff880297dfdd28: Connection to lustre-OST0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0000-osc-ffff880297dfdd28: This client was evicted by lustre-OST0000; in progress operations using this service will fail. LustreError: 19342:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88007b7eac00 x1709753732653696/t0(0) o104->lustre-OST0000@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' Lustre: 13392:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 8: before 511 < left 15825, rollback = 9 Lustre: 13392:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 1 previous similar message Lustre: 4139:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.205@tcp:/lustre/fid: [0x200000403:0x14:0x0]/ may get corrupted (rc -108) LustreError: 19540:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0000-osc-ffff880297dfdd28: namespace resource [0x50:0x0:0x0].0x0 (ffff880263b6de40) refcount nonzero (1) after lock cleanup; forcing cleanup. Lustre: lustre-OST0000-osc-ffff880297dfdd28: Connection restored to 192.168.123.205@tcp (at 0@lo) LustreError: 11016:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0001_UUID lock: ffff88026a7a70c0/0x4cdd18e9bacae6d9 lrc: 3/0,0 mode: PW/PW res: [0x300000401:0x1f:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000000020020 nid: 0@lo remote: 0x4cdd18e9bacae5b3 expref: 2778 pid: 12706 timeout: 424 lvb_type: 0 LustreError: 19163:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88028c7e5180 x1709753732849408/t0(0) o104->lustre-OST0001@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 11-0: lustre-OST0000-osc-ffff880298ec6678: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 9 previous similar messages Lustre: lustre-OST0000-osc-ffff880298ec6678: Connection to lustre-OST0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0000-osc-ffff880298ec6678: This client was evicted by lustre-OST0000; in progress operations using this service will fail. LustreError: 11-0: lustre-OST0001-osc-ffff880298ec6678: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-OST0001-osc-ffff880298ec6678: Connection to lustre-OST0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: 4138:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.205@tcp:/lustre/fid: [0x280000403:0x5e:0x0]// may get corrupted (rc -108) Lustre: 4139:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.205@tcp:/lustre/fid: [0x240000403:0x47:0x0]// may get corrupted (rc -108) LustreError: 19642:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0000-osc-ffff880298ec6678: namespace resource [0x10:0x0:0x0].0x0 (ffff88026e2ecf40) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 19642:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 245 previous similar messages LustreError: 19173:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88028ef870c0 x1709753732960576/t0(0) o104->lustre-OST0000@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 19173:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 2 previous similar messages LustreError: 167-0: lustre-OST0001-osc-ffff880298ec6678: This client was evicted by lustre-OST0001; in progress operations using this service will fail. Lustre: 4122:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.205@tcp:/lustre/fid: [0x200000403:0x56:0x0]/ may get corrupted (rc -108) Lustre: 4122:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.205@tcp:/lustre/fid: [0x240000403:0x3d:0x0]/ may get corrupted (rc -108) Lustre: 4123:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.205@tcp:/lustre/fid: [0x200000403:0x50:0x0]/ may get corrupted (rc -108) LustreError: 19117:0:(service.c:2290:ptlrpc_server_handle_request()) @@@ Dropping timed-out request from 12345-0@lo: deadline 45/4s ago req@ffff880263f9bec0 x1709753731842880/t0(0) o1->6d185bae-390e-4035-915f-f0e8ed765f60@0@lo:701/0 lens 440/0 e 0 to 0 dl 1630548531 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'lfs.0' Lustre: 19117:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (45/4s); client may timeout req@ffff880263f9bec0 x1709753731842880/t0(0) o1->6d185bae-390e-4035-915f-f0e8ed765f60@0@lo:701/0 lens 440/0 e 0 to 0 dl 1630548531 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'lfs.0' LustreError: 19159:0:(service.c:2290:ptlrpc_server_handle_request()) @@@ Dropping timed-out request from 12345-0@lo: deadline 45/4s ago req@ffff88026afe5180 x1709753731843648/t0(0) o1->6d185bae-390e-4035-915f-f0e8ed765f60@0@lo:701/0 lens 440/0 e 0 to 0 dl 1630548531 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'lfs.0' LustreError: 19159:0:(service.c:2290:ptlrpc_server_handle_request()) Skipped 2 previous similar messages Lustre: 19159:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (45/5s); client may timeout req@ffff88026afe5180 x1709753731843648/t0(0) o1->6d185bae-390e-4035-915f-f0e8ed765f60@0@lo:701/0 lens 440/0 e 0 to 0 dl 1630548531 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'lfs.0' Lustre: 19159:0:(service.c:2328:ptlrpc_server_handle_request()) Skipped 2 previous similar messages LustreError: 19718:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0001-osc-ffff880298ec6678: namespace resource [0x250:0x0:0x0].0x0 (ffff8802658a2340) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 19718:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 76 previous similar messages LustreError: 11016:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 93s: evicting client at 0@lo ns: filter-lustre-OST0002_UUID lock: ffff880265ab2200/0x4cdd18e9bacbc81b lrc: 3/0,0 mode: PW/PW res: [0x340000400:0x181:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->65535) gid 0 flags: 0x60000400000020 nid: 0@lo remote: 0x4cdd18e9bacbc7ff expref: 2442 pid: 16754 timeout: 440 lvb_type: 0 LustreError: 11016:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 1 previous similar message LustreError: 18198:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802641e8680 x1709753733300800/t0(0) o106->lustre-OST0002@0@lo:15/16 lens 328/280 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 17682:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630548536 with bad export cookie 5538610508875759005 LustreError: 17682:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) Skipped 3 previous similar messages Lustre: lustre-OST0002-osc-ffff880297dfdd28: Connection to lustre-OST0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0002-osc-ffff880297dfdd28: This client was evicted by lustre-OST0002; in progress operations using this service will fail. LustreError: 18198:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 12 previous similar messages LustreError: 19086:0:(service.c:2290:ptlrpc_server_handle_request()) @@@ Dropping timed-out request from 12345-0@lo: deadline 45/6s ago req@ffff88025c9c57c0 x1709753731847872/t0(0) o1->6d185bae-390e-4035-915f-f0e8ed765f60@0@lo:701/0 lens 440/0 e 0 to 0 dl 1630548531 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'lfs.0' LustreError: 19086:0:(service.c:2290:ptlrpc_server_handle_request()) Skipped 15 previous similar messages Lustre: 19086:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (45/6s); client may timeout req@ffff88025c9c57c0 x1709753731847872/t0(0) o1->6d185bae-390e-4035-915f-f0e8ed765f60@0@lo:701/0 lens 440/0 e 0 to 0 dl 1630548531 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'lfs.0' Lustre: 19086:0:(service.c:2328:ptlrpc_server_handle_request()) Skipped 15 previous similar messages LustreError: 19113:0:(service.c:2290:ptlrpc_server_handle_request()) @@@ Dropping timed-out request from 12345-0@lo: deadline 45/7s ago req@ffff8802646ce440 x1709753731870720/t0(0) o1->6d185bae-390e-4035-915f-f0e8ed765f60@0@lo:702/0 lens 440/0 e 0 to 0 dl 1630548532 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'lfs.0' LustreError: 19113:0:(service.c:2290:ptlrpc_server_handle_request()) Skipped 45 previous similar messages Lustre: 19113:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (45/7s); client may timeout req@ffff8802646ce440 x1709753731870720/t0(0) o1->6d185bae-390e-4035-915f-f0e8ed765f60@0@lo:702/0 lens 440/0 e 0 to 0 dl 1630548532 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'lfs.0' Lustre: 19113:0:(service.c:2328:ptlrpc_server_handle_request()) Skipped 45 previous similar messages Lustre: 4139:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.205@tcp:/lustre/fid: [0x240000404:0x46:0x0]/ may get corrupted (rc -108) Lustre: 4138:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.205@tcp:/lustre/fid: [0x200000404:0xd6:0x0]// may get corrupted (rc -108) LustreError: 19170:0:(service.c:2290:ptlrpc_server_handle_request()) @@@ Dropping timed-out request from 12345-0@lo: deadline 43/12s ago req@ffff880263909f80 x1709753731887808/t0(0) o1->6d185bae-390e-4035-915f-f0e8ed765f60@0@lo:701/0 lens 440/0 e 0 to 0 dl 1630548531 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'lfs.0' Lustre: 18160:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (43/12s); client may timeout req@ffff880260319300 x1709753731887936/t0(0) o1->6d185bae-390e-4035-915f-f0e8ed765f60@0@lo:701/0 lens 440/0 e 0 to 0 dl 1630548531 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'lfs.0' Lustre: 18160:0:(service.c:2328:ptlrpc_server_handle_request()) Skipped 26 previous similar messages LustreError: 19170:0:(service.c:2290:ptlrpc_server_handle_request()) Skipped 27 previous similar messages Lustre: ll_ost05_002: service thread pid 12711 was inactive for 62.205 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 19082, comm: ll_ost05_012 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] tgt_extent_lock+0xea/0x2a0 [ptlrpc] [<0>] ofd_getattr_hdl+0x365/0x720 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: ll_ost05_004: service thread pid 17775 was inactive for 62.201 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 1 previous similar message Pid: 12711, comm: ll_ost05_002 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] tgt_extent_lock+0xea/0x2a0 [ptlrpc] [<0>] ofd_getattr_hdl+0x365/0x720 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: ll_ost05_017: service thread pid 19089 was inactive for 62.246 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 3 previous similar messages Lustre: ll_ost05_010: service thread pid 18265 was inactive for 63.808 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 4 previous similar messages LustreError: 19718:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0001-osc-ffff880298ec6678: namespace resource [0x300000400:0x4:0x0].0x0 (ffff880265a92340) refcount nonzero (1) after lock cleanup; forcing cleanup. Lustre: lustre-OST0000-osc-ffff880298ec6678: Connection restored to 192.168.123.205@tcp (at 0@lo) LustreError: 19718:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 351 previous similar messages Lustre: lustre-OST0001-osc-ffff880298ec6678: Connection restored to 192.168.123.205@tcp (at 0@lo) Lustre: ll_ost05_041: service thread pid 19120 was inactive for 66.365 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 2 previous similar messages Lustre: 19598:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 507 < left 671, rollback = 2 Lustre: 19598:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 37 previous similar messages Lustre: 16866:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 373/1492/0, destroy: 0/0/0 Lustre: 16866:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 13069 previous similar messages Lustre: 16866:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/1, xattr_set: 375/15/0 Lustre: 19598:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 5/136/0, punch: 0/0/0, quota 6/6/0 Lustre: 19598:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 13069 previous similar messages Lustre: 16866:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 13074 previous similar messages Lustre: 16866:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 373/6341/0, delete: 0/0/0 Lustre: 16866:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 13074 previous similar messages Lustre: 16866:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 16866:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 13074 previous similar messages Lustre: 11171:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 509 < left 39269, rollback = 7 Lustre: 11171:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 24827 previous similar messages INFO: task mv:16599 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mv D ffff880290af1900 11344 16599 16185 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff81251fe1>] lock_rename+0x31/0xe0 [<ffffffff8125862f>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed432>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817e8a76>] ? trace_do_page_fault+0x56/0x170 [<ffffffff8125975e>] SyS_renameat2+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task mv:17719 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mv D ffff8802885e24f0 11344 17719 15954 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff81251fe1>] lock_rename+0x31/0xe0 [<ffffffff8125862f>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed432>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817e8a76>] ? trace_do_page_fault+0x56/0x170 [<ffffffff8125975e>] SyS_renameat2+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 Lustre: 4130:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff88028a810040 x1709753736107584/t0(0) o10->lustre-OST0001-osc-ffff880298ec6678@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630548684 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dd.0' Lustre: 17344:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 505 < left 582, rollback = 2 Lustre: 17344:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 21 previous similar messages LustreError: 11164:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000403:0x67:0x0]: rc = -2 Lustre: 13392:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 373/1492/0, destroy: 0/0/0 Lustre: 13392:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 48464 previous similar messages Lustre: 13392:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/1, xattr_set: 375/15/0 Lustre: 13392:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 48459 previous similar messages Lustre: 13392:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1866/16049/0, punch: 0/0/0, quota 14/174/11 Lustre: 13392:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 48464 previous similar messages Lustre: 13392:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 373/6341/0, delete: 0/0/0 Lustre: 13392:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 48459 previous similar messages Lustre: 13392:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 13392:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 48459 previous similar messages LustreError: 11016:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 99s: evicting client at 0@lo ns: filter-lustre-OST0002_UUID lock: ffff88007d254780/0x4cdd18e9bad41589 lrc: 3/0,0 mode: PW/PW res: [0x340000400:0x7:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->134217727) gid 0 flags: 0x60000400020020 nid: 0@lo remote: 0x4cdd18e9bad4156d expref: 3341 pid: 18894 timeout: 603 lvb_type: 0 LustreError: 18343:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630548701 with bad export cookie 5538610508876324976 LustreError: 13873:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff880253d14b40 x1709753739679680/t0(0) o104->lustre-OST0002@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 13873:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 41 previous similar messages INFO: task mv:20038 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mv D ffff8802957b4240 10608 20038 15948 0x00000080 Lustre: lustre-OST0002-osc-ffff880297dfdd28: Connection to lustre-OST0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Call Trace: LustreError: 167-0: lustre-OST0002-osc-ffff880297dfdd28: This client was evicted by lustre-OST0002; in progress operations using this service will fail. [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff81251fe1>] lock_rename+0x31/0xe0 [<ffffffff8125862f>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed432>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817e8a76>] ? trace_do_page_fault+0x56/0x170 [<ffffffff8125975e>] SyS_renameat2+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 LustreError: 11016:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0001_UUID lock: ffff88025dee4780/0x4cdd18e9bad63257 lrc: 3/0,0 mode: PW/PW res: [0x300000401:0x310:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000000020020 nid: 0@lo remote: 0x4cdd18e9bad6319a expref: 3443 pid: 18192 timeout: 611 lvb_type: 0 LustreError: 18282:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630548707 with bad export cookie 5538610508875758998 LustreError: 18282:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) Skipped 2 previous similar messages Lustre: 4139:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.205@tcp:/lustre/fid: [0x280000404:0xaa:0x0]// may get corrupted (rc -108) Lustre: 4138:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.205@tcp:/lustre/fid: [0x200000404:0x135:0x0]/ may get corrupted (rc -108) Lustre: 4139:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.205@tcp:/lustre/fid: [0x280000403:0xbb:0x0]/ may get corrupted (rc -108) Lustre: 4130:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff88007cd1b240 x1709753740163840/t0(0) o10->lustre-OST0003-osc-ffff880298ec6678@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630548791 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'cp.0' LustreError: 22454:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0001-osc-ffff880297dfdd28: namespace resource [0x10:0x0:0x0].0x0 (ffff880281758040) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 22454:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 166 previous similar messages Lustre: lustre-OST0001-osc-ffff880297dfdd28: Connection restored to 192.168.123.205@tcp (at 0@lo) Lustre: Skipped 1 previous similar message Lustre: 16928:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 503 < left 27616, rollback = 7 Lustre: 16928:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 40757 previous similar messages LustreError: 22136:0:(service.c:2290:ptlrpc_server_handle_request()) @@@ Dropping timed-out request from 12345-0@lo: deadline 47/5s ago req@ffff8800835cde00 x1709753738638720/t0(0) o101->6d185bae-390e-4035-915f-f0e8ed765f60@0@lo:142/0 lens 328/0 e 0 to 0 dl 1630548727 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'cp.0' LustreError: 22136:0:(service.c:2290:ptlrpc_server_handle_request()) Skipped 13 previous similar messages Lustre: 22136:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (47/5s); client may timeout req@ffff8800835cde00 x1709753738638720/t0(0) o101->6d185bae-390e-4035-915f-f0e8ed765f60@0@lo:142/0 lens 328/0 e 0 to 0 dl 1630548727 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'cp.0' Lustre: 22136:0:(service.c:2328:ptlrpc_server_handle_request()) Skipped 14 previous similar messages Lustre: ll_ost02_005: service thread pid 17823 was inactive for 62.024 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. LustreError: 11016:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0000_UUID lock: ffff8802659a6580/0x4cdd18e9bad7cb6e lrc: 3/0,0 mode: PW/PW res: [0x319:0x0:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000400020020 nid: 0@lo remote: 0x4cdd18e9bad7cb67 expref: 2676 pid: 18230 timeout: 641 lvb_type: 0 LustreError: 17149:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630548737 with bad export cookie 5538610508876289815 LustreError: 17149:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) Skipped 4 previous similar messages Lustre: lustre-OST0000-osc-ffff880298ec6678: Connection to lustre-OST0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 1 previous similar message LustreError: 18229:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff880086c2cb40 x1709753741110976/t0(0) o104->lustre-OST0000@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 167-0: lustre-OST0000-osc-ffff880298ec6678: This client was evicted by lustre-OST0000; in progress operations using this service will fail. LustreError: Skipped 1 previous similar message LustreError: 18229:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 37 previous similar messages LustreError: 22739:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0000-osc-ffff880298ec6678: namespace resource [0x10:0x0:0x0].0x0 (ffff88007d5bd940) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 22739:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 917 previous similar messages Lustre: lustre-OST0000-osc-ffff880298ec6678: Connection restored to 192.168.123.205@tcp (at 0@lo) Lustre: Skipped 1 previous similar message Lustre: 20336:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 8: before 509 < left 22779, rollback = 9 Lustre: 20336:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 26 previous similar messages Lustre: mdt06_005: service thread pid 17091 was inactive for 63.455 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. ptlrpc_watchdog_fire: 15 callbacks suppressed Lustre: mdt06_002: service thread pid 11151 was inactive for 70.037 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 11151, comm: mdt06_002 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_fini+0xa25/0xec0 [ptlrpc] [<0>] ldlm_cli_enqueue+0x436/0xa30 [ptlrpc] [<0>] osp_md_object_lock+0x160/0x300 [osp] [<0>] lod_object_lock+0xdb/0x7d0 [lod] [<0>] mdd_object_lock+0x2d/0xd0 [mdd] [<0>] mdt_remote_object_lock_try+0x1cc/0x510 [mdt] [<0>] mdt_remote_object_lock+0x2a/0x30 [mdt] [<0>] mdt_rename_lock+0xbe/0x4d0 [mdt] [<0>] mdt_reint_rename+0x31c/0x2ff0 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x76c/0xb40 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 17126, comm: mdt07_010 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_fini+0xa25/0xec0 [ptlrpc] [<0>] ldlm_cli_enqueue+0x436/0xa30 [ptlrpc] [<0>] osp_md_object_lock+0x160/0x300 [osp] [<0>] lod_object_lock+0xdb/0x7d0 [lod] [<0>] mdd_object_lock+0x2d/0xd0 [mdd] [<0>] mdt_remote_object_lock_try+0x1cc/0x510 [mdt] [<0>] mdt_remote_object_lock+0x2a/0x30 [mdt] [<0>] mdt_rename_lock+0xbe/0x4d0 [mdt] [<0>] mdt_reint_migrate+0x6e8/0x11d0 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x76c/0xb40 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: mdt01_006: service thread pid 17925 was inactive for 86.186 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Lustre: Skipped 1 previous similar message Pid: 17925, comm: mdt01_006 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] mdt_rename_lock+0x24b/0x4d0 [mdt] [<0>] mdt_reint_migrate+0x6e8/0x11d0 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x76c/0xb40 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe INFO: task mv:20038 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mv D ffff8802957b4240 10608 20038 15948 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff81251fe1>] lock_rename+0x31/0xe0 [<ffffffff8125862f>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed432>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817e8a76>] ? trace_do_page_fault+0x56/0x170 [<ffffffff8125975e>] SyS_renameat2+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task mv:20679 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mv D ffff88027d049280 11344 20679 16071 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff81251fe1>] lock_rename+0x31/0xe0 [<ffffffff8125862f>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed432>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817e8a76>] ? trace_do_page_fault+0x56/0x170 [<ffffffff8125975e>] SyS_renameat2+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task rm:21007 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. rm D ffff88008a2c9900 11760 21007 15991 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff812535bd>] vfs_unlink+0x4d/0x190 [<ffffffff812583be>] do_unlinkat+0x26e/0x2b0 [<ffffffff817df62e>] ? mutex_unlock+0xe/0x10 [<ffffffff81293556>] ? dnotify_flush+0x46/0x110 [<ffffffff812592fb>] SyS_unlinkat+0x1b/0x40 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task ls:21296 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. ls D ffff88031f5e6b80 11360 21296 16658 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff812550a6>] do_last+0x296/0x1280 [<ffffffff812546de>] ? link_path_walk+0x27e/0x8c0 [<ffffffff812570ad>] path_openat+0xcd/0x5b0 [<ffffffff817d5822>] ? free_debug_processing+0x240/0x272 [<ffffffff81257b2d>] ? putname+0x3d/0x60 [<ffffffff81258dfd>] do_filp_open+0x4d/0xb0 [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e326e>] ? _raw_spin_unlock+0xe/0x20 [<ffffffff81267893>] ? __alloc_fd+0xc3/0x170 [<ffffffff812447e4>] do_sys_open+0x124/0x220 [<ffffffff81244914>] SyS_openat+0x14/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task mv:21347 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mv D ffff88025d381280 11168 21347 16669 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff81251fe1>] lock_rename+0x31/0xe0 [<ffffffff8125862f>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed432>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817e8a76>] ? trace_do_page_fault+0x56/0x170 [<ffffffff8125975e>] SyS_renameat2+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task setfattr:21822 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. setfattr D ffff8802914073c0 12920 21822 17221 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff812519c2>] ? __inode_permission+0x52/0xd0 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff8126fdcf>] vfs_removexattr+0x5f/0x130 [<ffffffff8126fef5>] removexattr+0x55/0x80 [<ffffffff81257b2d>] ? putname+0x3d/0x60 [<ffffffff81258cd2>] ? user_path_at_empty+0x72/0xc0 [<ffffffff81248b85>] ? __sb_start_write+0x75/0x140 [<ffffffff817e8655>] ? __do_page_fault+0x1c5/0x500 [<ffffffff8126a8ac>] ? mnt_want_write+0x2c/0x50 [<ffffffff81270d14>] SyS_removexattr+0x94/0xd0 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task cp:21849 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. cp D ffff88029300a140 12760 21849 17331 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff817d5bb2>] lookup_slow+0x33/0xa7 [<ffffffff81256d6e>] path_lookupat+0x89e/0x8d0 [<ffffffff8121f126>] ? kmem_cache_alloc+0x126/0x2e0 [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff81257b9f>] ? getname_flags+0x4f/0x1a0 [<ffffffff81256dcb>] filename_lookup+0x2b/0xc0 [<ffffffff81258cc7>] user_path_at_empty+0x67/0xc0 [<ffffffff8124ba16>] ? cp_new_stat+0x176/0x1b0 [<ffffffff81258d31>] user_path_at+0x11/0x20 [<ffffffff8124b6b3>] vfs_fstatat+0x63/0xc0 [<ffffffff8124ba7e>] SYSC_newstat+0x2e/0x60 [<ffffffff81132fb6>] ? __audit_syscall_exit+0x1f6/0x2b0 [<ffffffff8124bf4e>] SyS_newstat+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 ------------[ cut here ]------------ Lustre: mdt01_007: service thread pid 17946 was inactive for 86.183 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. WARNING: CPU: 12 PID: 17091 at /home/green/git/lustre-release/ldiskfs/ext4_jbd2.c:266 __ldiskfs_handle_dirty_metadata+0x1d2/0x230 [ldiskfs] Modules linked in: loop zfs(PO) zunicode(PO) zzstd(O) zlua(O) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) jbd2 mbcache lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) dm_flakey dm_mod libcfs(OE) crc_t10dif crct10dif_generic sb_edac edac_core iosf_mbi crc32_pclmul ghash_clmulni_intel aesni_intel lrw gf128mul glue_helper ablk_helper cryptd virtio_balloon pcspkr i2c_piix4 virtio_console ip_tables rpcsec_gss_krb5 drm_kms_helper ata_generic ttm pata_acpi drm crct10dif_pclmul crct10dif_common drm_panel_orientation_quirks crc32c_intel ata_piix serio_raw virtio_blk i2c_core libata floppy CPU: 12 PID: 17091 Comm: mdt06_005 Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011 Call Trace: [<ffffffff817d9418>] dump_stack+0x19/0x1b [<ffffffff8108d538>] __warn+0xd8/0x100 [<ffffffff8108d67d>] warn_slowpath_null+0x1d/0x20 [<ffffffffa0c24262>] __ldiskfs_handle_dirty_metadata+0x1d2/0x230 [ldiskfs] [<ffffffffa0c0433d>] ldiskfs_getblk+0x13d/0x210 [ldiskfs] [<ffffffffa0c04437>] ldiskfs_bread+0x27/0xe0 [ldiskfs] [<ffffffffa0cc8e97>] osd_write+0x4e7/0xcb0 [osd_ldiskfs] [<ffffffffa05046f3>] dt_record_write+0x33/0x120 [obdclass] [<ffffffffa04c36b2>] llog_osd_write_rec+0x1052/0x19a0 [obdclass] [<ffffffffa04b0760>] llog_write_rec+0x290/0x590 [obdclass] [<ffffffffa04b64ee>] llog_cat_add_rec+0x23e/0x950 [obdclass] [<ffffffffa04ad5ff>] llog_add+0x17f/0x1f0 [obdclass] [<ffffffffa04dea39>] ? lprocfs_counter_add+0xf9/0x160 [obdclass] [<ffffffffa084760d>] sub_updates_write+0xb6b/0xe32 [ptlrpc] [<ffffffffa0817892>] top_trans_stop+0x7c2/0xf60 [ptlrpc] [<ffffffffa0f03f3c>] lod_trans_stop+0x25c/0x340 [lod] [<ffffffffa0dd8408>] mdd_trans_stop+0x28/0x16e [mdd] [<ffffffffa0db9b2f>] mdd_migrate_object+0xf1f/0x1ba0 [mdd] [<ffffffffa0503375>] ? lu_object_find_at+0x285/0xb10 [obdclass] [<ffffffffa0dbab86>] mdd_migrate+0x3d6/0x8e0 [mdd] [<ffffffffa0e8c0d8>] mdo_migrate+0x4c/0x4e [mdt] [<ffffffffa0e4be53>] mdt_reint_migrate+0x1023/0x11d0 [mdt] [<ffffffffa0e4c087>] mdt_reint_rec+0x87/0x240 [mdt] [<ffffffffa0e2155c>] mdt_reint_internal+0x76c/0xb40 [mdt] [<ffffffffa0e2e657>] mdt_reint+0x67/0x150 [mdt] [<ffffffffa080626e>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<ffffffffa032745e>] ? libcfs_nid2str_r+0xfe/0x130 [lnet] [<ffffffffa07b1990>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<ffffffffa07b3529>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e32be>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffffa07b2950>] ? ptlrpc_wait_event+0x620/0x620 [ptlrpc] [<ffffffff810ba0f4>] kthread+0xe4/0xf0 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 ---[ end trace 9770dc902265b5ee ]--- LDISKFS-fs: ldiskfs_getblk:888: aborting transaction: error 28 in __ldiskfs_handle_dirty_metadata LDISKFS-fs error (device dm-2): ldiskfs_getblk:888: inode #172: block 49122: comm mdt06_005: journal_dirty_metadata failed: handle type 0 started at line 1972, credits 512/0, errcode -28 Aborting journal on device dm-2-8. LDISKFS-fs (dm-2): Remounting filesystem read-only LustreError: 17091:0:(osd_io.c:2103:osd_ldiskfs_write_record()) lustre-MDT0002: error reading offset 32768 (block 8, size 56, offs 32768), credits 0/0: rc = -28 LustreError: 17091:0:(update_trans.c:985:top_trans_stop()) lustre-MDT0002-osd: write updates failed: rc = -28 LustreError: 17091:0:(osd_io.c:2110:osd_ldiskfs_write_record()) journal_get_write_access() returned error -30 LustreError: 17091:0:(tgt_lastrcvd.c:1253:tgt_add_reply_data()) lustre-MDT0002: can't update reply_data file: rc = -30 LustreError: 17091:0:(osd_handler.c:2075:osd_trans_stop()) lustre-MDT0002: failed in transaction hook: rc = -30 LDISKFS-fs error (device dm-2) in osd_trans_stop:2082: error 28 LustreError: 17091:0:(osd_handler.c:2085:osd_trans_stop()) lustre-MDT0002: failed to stop transaction: rc = -28 LustreError: 13875:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-OST0002-osc-MDT0002: fail to cancel 1 llog-records: rc = -30 LustreError: 13875:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-OST0002-osc-MDT0002: fail to cancel 1 of 1 llog-records: rc = -30 LustreError: 13875:0:(osp_sync.c:1079:osp_sync_process_committed()) lustre-OST0002-osc-MDT0002: can't cancel record: rc = -30 LustreError: 13875:0:(osp_sync.c:1094:osp_sync_process_committed()) lustre-OST0002-osc-MDT0002: can't cancel 8 records: rc = -30 LustreError: 12280:0:(osd_handler.c:1779:osd_trans_commit_cb()) transaction @0xffff880263234258 commit error: 2 LustreError: 17091:0:(update_trans.c:1011:top_trans_stop()) lustre-MDT0002-osd: stop trans failed: rc = -30 LDISKFS-fs error (device dm-2) in osd_trans_stop:2082: IO failure LustreError: 19203:0:(tgt_handler.c:1367:tgt_blocking_ast()) lustre-MDT0002: syncing [0x280000402:0x4:0x0] (19-0) on lock cancel: rc = -5 LustreError: 23687:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -5 | Externally reported by onyx-68 boilpot email |
racer test 1: racer on clients: centos-0.localnet DURATION=2700 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa1050e52>] ll_migrate+0x9b2/0xec0 [lustre] PGD 80000002876dd067 PUD 23add6067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: loop zfs(PO) zunicode(PO) zzstd(O) zlua(O) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) jbd2 mbcache lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) dm_flakey dm_mod libcfs(OE) crc_t10dif crct10dif_generic sb_edac edac_core iosf_mbi crc32_pclmul ghash_clmulni_intel aesni_intel lrw gf128mul glue_helper ablk_helper cryptd virtio_balloon pcspkr virtio_console i2c_piix4 ip_tables rpcsec_gss_krb5 ata_generic drm_kms_helper pata_acpi ttm crct10dif_pclmul drm crct10dif_common drm_panel_orientation_quirks ata_piix crc32c_intel serio_raw virtio_blk i2c_core libata floppy CPU: 5 PID: 9446 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011 task: ffff880079a649d0 ti: ffff880233610000 task.ti: ffff880233610000 RIP: 0010:[<ffffffffa1050e52>] [<ffffffffa1050e52>] ll_migrate+0x9b2/0xec0 [lustre] RSP: 0018:ffff880233613bc8 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff88027a52f528 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff880326a6ab98 RDI: ffff880326a6ab78 RBP: ffff880233613c48 R08: ffff88026be16418 R09: ffff880233613a60 R10: 0000000000000002 R11: 0000000000000005 R12: 0000000000000000 R13: ffff88025dcd7a98 R14: ffff88023c679cd8 R15: 0000000000000030 FS: 00007fc75fd8b740(0000) GS:ffff880331b40000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 CR2: 0000000000000008 CR3: 000000023077e000 CR4: 00000000001607e0 Call Trace: [<ffffffff81242213>] ? __check_object_size+0x1c3/0x220 [<ffffffffa103e561>] ll_dir_ioctl+0x5d01/0x6ed0 [lustre] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff8115260f>] ? delayacct_end+0x8f/0xb0 [<ffffffff81152744>] ? __delayacct_blkio_end+0x34/0x60 [<ffffffff817e0257>] ? io_schedule_timeout+0xe7/0x130 [<ffffffff811b62dd>] ? find_get_pages_tag+0x10d/0x260 [<ffffffff811c3691>] ? pagevec_lookup_tag+0x21/0x30 [<ffffffff811b400e>] ? __filemap_fdatawait_range+0xbe/0x1a0 [<ffffffff8125b3fd>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff81245a38>] ? vfs_write+0x178/0x230 [<ffffffff8125b751>] SyS_ioctl+0xa1/0xc0 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 | Lustre: lustre-OST0000-osc-ffff8802c6fd6fc8: disconnect after 21s idle cp (16323) used greatest stack depth: 10080 bytes left Lustre: 16611:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 512 < left 1834, rollback = 7 Lustre: 16611:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 16611:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 1/89/0 Lustre: 16611:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1825/1834/0, punch: 0/0/0, quota 3/3/0 Lustre: 16611:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 16611:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 16611:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 512 < left 1834, rollback = 7 Lustre: 16611:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 504 previous similar messages Lustre: 16611:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 16611:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 504 previous similar messages Lustre: 16611:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 1/89/0 Lustre: 16611:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 504 previous similar messages Lustre: 16611:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1825/1834/0, punch: 0/0/0, quota 3/3/0 Lustre: 16611:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 504 previous similar messages Lustre: 16611:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 16611:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 504 previous similar messages Lustre: 16611:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 16611:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 504 previous similar messages Lustre: 11190:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 505 < left 763, rollback = 7 Lustre: 11190:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 1319 previous similar messages Lustre: 11190:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 1/4/4, destroy: 0/0/0 Lustre: 11190:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 1319 previous similar messages Lustre: 11190:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 4/209/0 Lustre: 11190:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 1319 previous similar messages Lustre: 11190:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 754/763/0, punch: 0/0/0, quota 6/6/2 Lustre: 11190:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 1319 previous similar messages Lustre: 11190:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 2/33/1, delete: 0/0/0 Lustre: 11190:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 1319 previous similar messages Lustre: 11190:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 11190:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 1319 previous similar messages Lustre: 13068:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 507 < left 9857, rollback = 7 Lustre: 13068:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 753 previous similar messages Lustre: 13068:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 229/916/0, destroy: 0/0/0 Lustre: 13068:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 753 previous similar messages Lustre: 13068:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/1, xattr_set: 231/15/0 Lustre: 13068:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 753 previous similar messages Lustre: 13068:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1146/9857/0, punch: 0/0/0, quota 10/106/4 Lustre: 13068:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 753 previous similar messages Lustre: 13068:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 229/3893/0, delete: 0/0/0 Lustre: 13068:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 753 previous similar messages Lustre: 13068:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 13068:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 753 previous similar messages Lustre: 17236:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 8: before 510 < left 3909, rollback = 9 Lustre: 16766:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 511 < left 763, rollback = 7 Lustre: 16766:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 917 previous similar messages Lustre: 16766:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 1/4/0, destroy: 0/0/0 Lustre: 16766:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 918 previous similar messages Lustre: 16766:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 4/209/0 Lustre: 16766:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 918 previous similar messages Lustre: 16766:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 754/763/0, punch: 0/0/0, quota 6/6/0 Lustre: 16766:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 918 previous similar messages Lustre: 16766:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 2/33/1, delete: 0/0/0 Lustre: 16766:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 918 previous similar messages Lustre: 16766:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 16766:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 918 previous similar messages Lustre: 11198:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 507 < left 582, rollback = 2 Lustre: 11200:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 510 < left 763, rollback = 7 Lustre: 11200:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 753 previous similar messages Lustre: 11200:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 1/4/1, destroy: 0/0/0 Lustre: 11200:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 759 previous similar messages Lustre: 11200:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 4/209/0 Lustre: 11200:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 759 previous similar messages Lustre: 11200:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 754/763/0, punch: 0/0/0, quota 6/6/0 Lustre: 11200:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 759 previous similar messages Lustre: 11200:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 2/33/1, delete: 0/0/0 Lustre: 11200:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 759 previous similar messages Lustre: 11200:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 11200:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 759 previous similar messages Lustre: 18819:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 496 < left 788, rollback = 2 Lustre: 18819:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 5 previous similar messages Lustre: 18819:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 3/12/12, destroy: 0/0/0 Lustre: 18819:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 753 previous similar messages Lustre: 18819:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 2/2/0, xattr_set: 10/788/0 Lustre: 18819:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 753 previous similar messages Lustre: 18819:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 5/136/0, punch: 0/0/0, quota 8/56/0 Lustre: 18819:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 753 previous similar messages Lustre: 18819:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 12/215/4, delete: 1/1/0 Lustre: 18819:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 753 previous similar messages Lustre: 18819:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 6/6/0, ref_del: 1/1/0 Lustre: 18819:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 753 previous similar messages Lustre: 15879:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 511 < left 551, rollback = 2 Lustre: 15879:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 27 previous similar messages Lustre: 16342:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 8: before 512 < left 31041, rollback = 9 Lustre: 16342:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 11 previous similar messages Lustre: 16342:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 512 < left 78485, rollback = 7 Lustre: 16342:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 753 previous similar messages sched: RT throttling activated Lustre: 4122:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff8802800b9f80 x1709747209921216/t0(0) o10->lustre-OST0002-osc-ffff8802995d2e98@0@lo:6/4 lens 440/432 e 1 to 0 dl 1630542317 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dd.0' ------------[ cut here ]------------ WARNING: CPU: 15 PID: 11241 at /home/green/git/lustre-release/ldiskfs/namei.c:3331 ldiskfs_orphan_add+0x11e/0x2a0 [ldiskfs] Modules linked in: loop zfs(PO) zunicode(PO) zzstd(O) zlua(O) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) jbd2 mbcache lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) dm_flakey dm_mod libcfs(OE) crc_t10dif crct10dif_generic sb_edac edac_core iosf_mbi crc32_pclmul ghash_clmulni_intel aesni_intel lrw gf128mul glue_helper ablk_helper cryptd virtio_balloon pcspkr virtio_console i2c_piix4 ip_tables rpcsec_gss_krb5 ata_generic drm_kms_helper pata_acpi ttm crct10dif_pclmul drm crct10dif_common drm_panel_orientation_quirks ata_piix crc32c_intel serio_raw virtio_blk i2c_core libata floppy CPU: 15 PID: 11241 Comm: mdt_rdpg07_001 Tainted: P OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011 Call Trace: [<ffffffff817d9418>] dump_stack+0x19/0x1b [<ffffffff8108d538>] __warn+0xd8/0x100 [<ffffffff8108d67d>] warn_slowpath_null+0x1d/0x20 [<ffffffffa0be84ae>] ldiskfs_orphan_add+0x11e/0x2a0 [ldiskfs] [<ffffffffa0bd99eb>] ldiskfs_xattr_inode_orphan_add.constprop.17+0xbb/0x110 [ldiskfs] [<ffffffffa0bdc417>] ldiskfs_xattr_delete_inode+0x77/0x3c0 [ldiskfs] [<ffffffffa0bfcaa0>] ? ldiskfs_evict_inode+0x190/0x670 [ldiskfs] Lustre: 11180:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 506 < left 610, rollback = 2 Lustre: 11180:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 1 previous similar message [<ffffffffa0bfcad9>] ldiskfs_evict_inode+0x1c9/0x670 [ldiskfs] [<ffffffff812649af>] evict+0xaf/0x180 [<ffffffff81264de5>] iput+0xf5/0x180 [<ffffffffa0c91d08>] osd_object_delete+0x1f8/0x370 [osd_ldiskfs] [<ffffffffa04c759d>] lu_object_free.isra.26+0xcd/0x1d0 [obdclass] [<ffffffffa04cbcf1>] lu_object_put+0xa1/0x480 [obdclass] [<ffffffffa0e49af4>] mdt_mfd_close+0x1f4/0x870 [mdt] [<ffffffffa0e4fe14>] mdt_close_internal+0xb4/0x240 [mdt] [<ffffffffa0e50224>] mdt_close+0x284/0x850 [mdt] [<ffffffffa07ce2d4>] ? tgt_request_preprocess.isra.27+0xf4/0x850 [ptlrpc] [<ffffffffa07cf26e>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<ffffffffa02f245e>] ? libcfs_nid2str_r+0xfe/0x130 [lnet] [<ffffffffa077a990>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<ffffffffa077c529>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffffa077b950>] ? ptlrpc_wait_event+0x620/0x620 [ptlrpc] [<ffffffff810ba0f4>] kthread+0xe4/0xf0 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 ---[ end trace fc07225a3244f178 ]--- Lustre: 11175:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 11175:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 4306 previous similar messages Lustre: 11175:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 2/178/0 Lustre: 11175:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 4306 previous similar messages Lustre: 11175:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1382/1391/0, punch: 0/0/0, quota 3/3/0 Lustre: 11175:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 4306 previous similar messages Lustre: 11175:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 11175:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 4306 previous similar messages Lustre: 11175:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 11175:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 4306 previous similar messages Lustre: 17406:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 508 < left 9814, rollback = 7 Lustre: 17406:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 5640 previous similar messages Lustre: 4118:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff8802701f8040 x1709747210555264/t0(0) o10->lustre-OST0002-osc-ffff8802995d2e98@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630542342 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'cp.0' Lustre: 4118:0:(client.c:1485:after_reply()) Skipped 3 previous similar messages Lustre: 4120:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff88026c644b40 x1709747210950080/t0(0) o10->lustre-OST0000-osc-ffff8802c6fd6fc8@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630542395 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dir_create.sh.0' Lustre: 19354:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 496 < left 671, rollback = 2 Lustre: 19354:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 38 previous similar messages LustreError: 11235:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0002: failed to get lu_attr of [0x280000404:0x7e:0x0]: rc = -2 Lustre: 16838:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 507 < left 671, rollback = 2 Lustre: 16838:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 24 previous similar messages Lustre: 16838:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 3/12/3, destroy: 0/0/0 Lustre: 16838:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 16851 previous similar messages Lustre: 16838:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 7/671/0 Lustre: 16838:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 16850 previous similar messages Lustre: 16838:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 5/136/0, punch: 0/0/0, quota 8/56/0 Lustre: 16838:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 16851 previous similar messages Lustre: 16838:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 10/203/3, delete: 0/0/0 Lustre: 16838:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 16852 previous similar messages Lustre: 16838:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 4/4/0, ref_del: 0/0/0 Lustre: 16838:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 16852 previous similar messages Lustre: 13850:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 505 < left 1904, rollback = 7 Lustre: 13850:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 18425 previous similar messages LustreError: 11071:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0003_UUID lock: ffff88027506ad40/0x14bfa02baf9ba43b lrc: 3/0,0 mode: PW/PW res: [0x380000401:0xc9:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000480020020 nid: 0@lo remote: 0x14bfa02baf9ba42d expref: 1913 pid: 20402 timeout: 527 lvb_type: 0 LustreError: 16315:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88008c381f80 x1709747214186048/t0(0) o106->lustre-OST0003@0@lo:15/16 lens 328/280 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 11-0: lustre-OST0003-osc-ffff8802c6fd6fc8: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-OST0003-osc-ffff8802c6fd6fc8: Connection to lustre-OST0003 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0003-osc-ffff8802c6fd6fc8: This client was evicted by lustre-OST0003; in progress operations using this service will fail. LustreError: 12327:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88026dbb3880 x1709747214229184/t0(0) o105->lustre-OST0003@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 12327:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 3 previous similar messages Lustre: 4130:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.100@tcp:/lustre/fid: [0x240000403:0x16a:0x0]// may get corrupted (rc -108) Lustre: 4128:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.100@tcp:/lustre/fid: [0x280000403:0xc7:0x0]// may get corrupted (rc -108) LustreError: 24520:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0003-osc-ffff8802c6fd6fc8: namespace resource [0x380000400:0x24f:0x0].0x0 (ffff88026070dbc0) refcount nonzero (1) after lock cleanup; forcing cleanup. Lustre: lustre-OST0003-osc-ffff8802c6fd6fc8: Connection restored to 192.168.123.100@tcp (at 0@lo) LustreError: 11071:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0002_UUID lock: ffff8802713043c0/0x14bfa02baf9bd780 lrc: 3/0,0 mode: PW/PW res: [0x340000401:0x2:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->134217727) gid 0 flags: 0x60000400020020 nid: 0@lo remote: 0x14bfa02baf9bd73a expref: 2206 pid: 18282 timeout: 534 lvb_type: 0 LustreError: 11-0: lustre-OST0002-osc-ffff8802995d2e98: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-OST0002-osc-ffff8802995d2e98: Connection to lustre-OST0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0002-osc-ffff8802995d2e98: This client was evicted by lustre-OST0002; in progress operations using this service will fail. Lustre: lustre-OST0002-osc-ffff8802995d2e98: Connection restored to 192.168.123.100@tcp (at 0@lo) Lustre: 19055:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 8: before 509 < left 29358, rollback = 9 Lustre: 19055:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 64 previous similar messages Lustre: 4120:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff8800881d1f80 x1709747216543296/t0(0) o10->lustre-OST0001-osc-ffff8802995d2e98@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630542497 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dd.0' INFO: task mv:21751 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mv D ffff8802884c24f0 11344 21751 17016 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff81251fe1>] lock_rename+0x31/0xe0 [<ffffffff8125862f>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed432>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817e8a76>] ? trace_do_page_fault+0x56/0x170 [<ffffffff8125975e>] SyS_renameat2+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 Lustre: 19066:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 3/12/8, destroy: 0/0/0 Lustre: 19066:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 57929 previous similar messages Lustre: 19066:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 11/1027/0 Lustre: 19066:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 57928 previous similar messages Lustre: 19066:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 5/136/0, punch: 0/0/0, quota 8/56/0 Lustre: 19066:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 57928 previous similar messages Lustre: 19066:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 10/203/4, delete: 0/0/0 Lustre: 19066:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 57927 previous similar messages Lustre: 19066:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 4/4/0, ref_del: 0/0/0 Lustre: 19066:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 57925 previous similar messages Lustre: 19035:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 511 < left 1300, rollback = 7 Lustre: 19035:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 54839 previous similar messages Lustre: mdt04_002: service thread pid 11195 was inactive for 64.122 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 11195, comm: mdt04_002 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] mdt_object_local_lock+0x527/0xb90 [mdt] [<0>] mdt_object_lock_internal+0x70/0x390 [mdt] [<0>] mdt_reint_object_lock+0x2c/0x60 [mdt] [<0>] mdt_reint_striped_lock+0x89/0x5b0 [mdt] [<0>] mdt_migrate_parent_lock+0x3f/0xb0 [mdt] [<0>] mdt_reint_migrate+0x8b7/0x11d0 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x76c/0xb40 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: 4120:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff8800881d1f80 x1709747218728064/t0(0) o10->lustre-OST0001-osc-ffff8802995d2e98@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630542586 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dd.0' Lustre: mdt07_002: service thread pid 11209 was inactive for 66.084 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 11209, comm: mdt07_002 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_fini+0xa25/0xec0 [ptlrpc] [<0>] ldlm_cli_enqueue+0x436/0xa30 [ptlrpc] [<0>] osp_md_object_lock+0x160/0x300 [osp] [<0>] lod_object_lock+0xdb/0x7d0 [lod] [<0>] mdd_object_lock+0x2d/0xd0 [mdd] [<0>] mdt_remote_object_lock_try+0x1cc/0x510 [mdt] [<0>] mdt_remote_object_lock+0x2a/0x30 [mdt] [<0>] mdt_rename_lock+0xbe/0x4d0 [mdt] [<0>] mdt_reint_rename+0x31c/0x2ff0 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x76c/0xb40 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: mdt04_007: service thread pid 17510 was inactive for 64.255 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 17510, comm: mdt04_007 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] mdt_rename_lock+0x24b/0x4d0 [mdt] [<0>] mdt_reint_migrate+0x6e8/0x11d0 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x76c/0xb40 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe LustreError: 11071:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0003_UUID lock: ffff880271d22980/0x14bfa02bafa25447 lrc: 3/0,0 mode: PW/PW res: [0x380000400:0x26a:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000400000020 nid: 0@lo remote: 0x14bfa02bafa2542b expref: 2074 pid: 23426 timeout: 655 lvb_type: 0 LustreError: 27424:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88027c761940 x1709747219831680/t0(0) o106->lustre-OST0003@0@lo:15/16 lens 328/280 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 27424:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 1 previous similar message LustreError: 11062:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630542547 with bad export cookie 1495089710800122125 Lustre: lustre-OST0003-osc-ffff8802995d2e98: Connection to lustre-OST0003 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0003-osc-ffff8802995d2e98: This client was evicted by lustre-OST0003; in progress operations using this service will fail. Lustre: 4120:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.100@tcp:/lustre/fid: [0x240000404:0x1ce:0x0]// may get corrupted (rc -5) Lustre: 4123:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff880273a8d7c0 x1709747219876096/t0(0) o10->lustre-OST0000-osc-ffff8802995d2e98@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630542609 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'cp.0' Lustre: lustre-OST0003-osc-ffff8802995d2e98: Connection restored to 192.168.123.100@tcp (at 0@lo) Lustre: mdt04_006: service thread pid 16766 was inactive for 64.080 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: mdt02_011: service thread pid 22876 was inactive for 68.175 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. LustreError: 11071:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 93s: evicting client at 0@lo ns: filter-lustre-OST0002_UUID lock: ffff88026b187480/0x14bfa02bafa1ad3c lrc: 4/0,0 mode: PW/PW res: [0x340000401:0xb0a:0x0].0x0 rrc: 4 type: EXT [0->18446744073709551615] (req 0->2097151) gid 0 flags: 0x60000400020020 nid: 0@lo remote: 0x14bfa02bafa1ac71 expref: 2267 pid: 13516 timeout: 669 lvb_type: 0 LustreError: 11069:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff880087820040 x1709747220298240/t0(0) o105->lustre-OST0002@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 11069:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 45 previous similar messages LustreError: 11064:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630542562 with bad export cookie 1495089710800122839 LustreError: 11-0: lustre-OST0002-osc-ffff8802c6fd6fc8: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-OST0002-osc-ffff8802c6fd6fc8: Connection to lustre-OST0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0002-osc-ffff8802c6fd6fc8: This client was evicted by lustre-OST0002; in progress operations using this service will fail. Lustre: 4122:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.100@tcp:/lustre/fid: [0x240000403:0x1cd:0x0]/ may get corrupted (rc -108) Lustre: 4123:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.100@tcp:/lustre/fid: [0x200000403:0xe3:0x0]/ may get corrupted (rc -108) Lustre: 12744:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 6s req@ffff880080613880 x1709747220426176/t0(0) o13->lustre-MDT0001-mdtlov_UUID@0@lo:0/0 lens 224/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'osp-pre-3-1.0' Lustre: lustre-OST0002-osc-ffff8802c6fd6fc8: Connection restored to 192.168.123.100@tcp (at 0@lo) Lustre: mdt03_007: service thread pid 17886 was inactive for 62.069 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: lustre-OST0002-osc-ffff8802c6fd6fc8: disconnect after 24s idle Lustre: Skipped 7 previous similar messages Lustre: mdt02_013: service thread pid 23967 was inactive for 68.109 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. LustreError: 11071:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0000_UUID lock: ffff880085740400/0x14bfa02bafa54695 lrc: 3/0,0 mode: PR/PR res: [0x200000404:0xa8:0x0].0x0 bits 0x12/0x0 rrc: 14 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0x14bfa02bafa4e68d expref: 116 pid: 20331 timeout: 710 lvb_type: 0 LustreError: 11-0: lustre-MDT0000-mdc-ffff8802995d2e98: operation mds_close to node 0@lo failed: rc = -107 Lustre: lustre-MDT0000-mdc-ffff8802995d2e98: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-MDT0000-mdc-ffff8802995d2e98: This client was evicted by lustre-MDT0000; in progress operations using this service will fail. LustreError: 24440:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff8802995d2e98: inode [0x200000404:0x11c:0x0] mdc close failed: rc = -5 LustreError: 25537:0:(llite_lib.c:2983:ll_prep_inode()) new_inode -fatal: rc -108 LustreError: 25022:0:(mdc_request.c:1436:mdc_read_page()) lustre-MDT0000-mdc-ffff8802995d2e98: [0x200000404:0x10a:0x0] lock enqueue fails: rc = -108 Lustre: dir [0x280000404:0x1a1:0x0] stripe 1 readdir failed: -108, directory is partially accessed! LustreError: 28319:0:(file.c:5042:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000007:0x1:0x0] error: rc = -108 LustreError: 28319:0:(file.c:5042:ll_inode_revalidate_fini()) Skipped 22 previous similar messages LustreError: 27208:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -108 LustreError: 24993:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff8802995d2e98: inode [0x200000403:0x1:0x0] mdc close failed: rc = -108 LustreError: 25002:0:(file.c:5042:ll_inode_revalidate_fini()) lustre: revalidate FID [0x200000007:0x1:0x0] error: rc = -108 LustreError: 25002:0:(file.c:5042:ll_inode_revalidate_fini()) Skipped 187 previous similar messages LustreError: 24993:0:(file.c:234:ll_close_inode_openhandle()) Skipped 11 previous similar messages Lustre: lustre-MDT0000-mdc-ffff8802995d2e98: Connection restored to 192.168.123.100@tcp (at 0@lo) Lustre: 22876:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 501 < left 582, rollback = 2 Lustre: 22876:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 36 previous similar messages INFO: task touch:16658 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. touch D ffff880295388010 10912 16658 16089 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff817d5bb2>] lookup_slow+0x33/0xa7 [<ffffffff81254c7f>] link_path_walk+0x81f/0x8c0 [<ffffffff8124818c>] ? get_empty_filp+0x5c/0x1f0 [<ffffffff8125708e>] path_openat+0xae/0x5b0 [<ffffffff81258dfd>] do_filp_open+0x4d/0xb0 [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e326e>] ? _raw_spin_unlock+0xe/0x20 [<ffffffff81267893>] ? __alloc_fd+0xc3/0x170 [<ffffffff812447e4>] do_sys_open+0x124/0x220 [<ffffffff812448fe>] SyS_open+0x1e/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task lfs:22066 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. lfs D ffff880081cfb760 11200 22066 16960 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff817d5bb2>] lookup_slow+0x33/0xa7 [<ffffffff81256d6e>] path_lookupat+0x89e/0x8d0 [<ffffffff8121f126>] ? kmem_cache_alloc+0x126/0x2e0 [<ffffffff81257b9f>] ? getname_flags+0x4f/0x1a0 [<ffffffff81256dcb>] filename_lookup+0x2b/0xc0 [<ffffffff81258cc7>] user_path_at_empty+0x67/0xc0 [<ffffffff8126a634>] ? mntput+0x24/0x40 [<ffffffff81247f88>] ? __fput+0x1a8/0x240 [<ffffffff81258d31>] user_path_at+0x11/0x20 [<ffffffff8127056e>] SyS_setxattr+0x4e/0x100 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task mv:22152 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mv D ffff8802952f24f0 11040 22152 15990 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff81251fe1>] lock_rename+0x31/0xe0 [<ffffffff8125862f>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed432>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817e8a76>] ? trace_do_page_fault+0x56/0x170 [<ffffffff8125975e>] SyS_renameat2+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task rm:22165 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. rm D ffff880289740010 11296 22165 16135 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff8125828b>] do_unlinkat+0x13b/0x2b0 [<ffffffff817e8a76>] ? trace_do_page_fault+0x56/0x170 [<ffffffff812592fb>] SyS_unlinkat+0x1b/0x40 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task mv:23050 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mv D ffff88028c773760 11344 23050 16517 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff81251fe1>] lock_rename+0x31/0xe0 [<ffffffff8125862f>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed432>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817e8a76>] ? trace_do_page_fault+0x56/0x170 [<ffffffff8125975e>] SyS_renameat2+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task mv:23914 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mv D ffff880295b03760 11088 23914 16403 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff81251fe1>] lock_rename+0x31/0xe0 [<ffffffff8125862f>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed432>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817e8a76>] ? trace_do_page_fault+0x56/0x170 [<ffffffff8125975e>] SyS_renameat2+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task mv:24248 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mv D ffff880295be4240 11040 24248 16066 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff81251fe1>] lock_rename+0x31/0xe0 [<ffffffff8125862f>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed432>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817e8a76>] ? trace_do_page_fault+0x56/0x170 [<ffffffff8125975e>] SyS_renameat2+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task mv:24539 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mv D ffff88026b00c9d0 11216 24539 15991 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff81251fe1>] lock_rename+0x31/0xe0 [<ffffffff8125862f>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed432>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817e8a76>] ? trace_do_page_fault+0x56/0x170 [<ffffffff8125975e>] SyS_renameat2+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task mv:24775 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mv D ffff8802877c4240 11344 24775 17059 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff81251fe1>] lock_rename+0x31/0xe0 [<ffffffff8125862f>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed432>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817e8a76>] ? trace_do_page_fault+0x56/0x170 [<ffffffff8125975e>] SyS_renameat2+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 Lustre: 4126:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff880256856440 x1709747222281728/t0(0) o10->lustre-OST0003-osc-ffff8802995d2e98@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630542635 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dd.0' LustreError: 20617:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0000: '19' migration was interrupted, run 'lfs migrate -m 2 -c 1 -H crush 19' to finish migration. LustreError: 11239:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0002: failed to get lu_attr of [0x280000404:0x1a1:0x0]: rc = -2 LustreError: 11071:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0000_UUID lock: ffff880081727840/0x14bfa02bafa8c0d7 lrc: 3/0,0 mode: PW/PW res: [0x2c0000401:0xaed:0x0].0x0 rrc: 5 type: EXT [0->18446744073709551615] (req 2097152->18446744073709551615) gid 0 flags: 0x60000000020020 nid: 0@lo remote: 0x14bfa02bafa8c0d0 expref: 3090 pid: 20991 timeout: 774 lvb_type: 0 LustreError: 11-0: lustre-OST0000-osc-ffff8802995d2e98: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-OST0000-osc-ffff8802995d2e98: Connection to lustre-OST0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0000-osc-ffff8802995d2e98: This client was evicted by lustre-OST0000; in progress operations using this service will fail. LustreError: 11069:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802595d9f80 x1709747224637376/t0(0) o105->lustre-OST0000@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 11069:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 37 previous similar messages Lustre: 4126:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff880256856440 x1709747224774272/t0(0) o10->lustre-OST0003-osc-ffff8802995d2e98@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630542743 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dd.0' LustreError: 30153:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0000-osc-ffff8802995d2e98: namespace resource [0x2c0000400:0x491:0x0].0x0 (ffff88026264ad40) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 30153:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 1 previous similar message Lustre: lustre-OST0000-osc-ffff8802995d2e98: Connection restored to 192.168.123.100@tcp (at 0@lo) LustreError: 18307:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0002: '12' migration was interrupted, run 'lfs migrate -m 1 -c 1 -H crush 12' to finish migration. LustreError: 11226:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000403:0x5b:0x0]: rc = -2 LustreError: 20973:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0001: '15' migration was interrupted, run 'lfs migrate -m 0 -c 1 -H crush 15' to finish migration. 4[31606]: segfault at 8 ip 00007f8bb75507e8 sp 00007ffeaf8e3e00 error 4 in ld-2.17.so[7f8bb7545000+22000] LustreError: 11071:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0001_UUID lock: ffff88026b32a980/0x14bfa02bafab11ca lrc: 3/0,0 mode: PW/PW res: [0x300000400:0x425:0x0].0x0 rrc: 5 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000400020020 nid: 0@lo remote: 0x14bfa02bafab11b5 expref: 2441 pid: 26224 timeout: 853 lvb_type: 0 LustreError: 19884:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802594bb880 x1709747228575040/t0(0) o106->lustre-OST0001@0@lo:15/16 lens 328/280 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 11049:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630542747 with bad export cookie 1495089710800119892 Lustre: lustre-OST0001-osc-ffff8802c6fd6fc8: Connection to lustre-OST0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0001-osc-ffff8802c6fd6fc8: This client was evicted by lustre-OST0001; in progress operations using this service will fail. LustreError: 22876:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0002: '12' migration was interrupted, run 'lfs migrate -m 1 -c 1 -H crush 12' to finish migration. LustreError: 32458:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0001-osc-ffff8802c6fd6fc8: namespace resource [0x300000400:0x430:0x0].0x0 (ffff880281ec9bc0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 32458:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 25 previous similar messages LustreError: 11049:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630542757 with bad export cookie 1495089710800123154 LustreError: 18744:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0002: failed to get lu_attr of [0x280000403:0xc6:0x0]: rc = -2 LustreError: 377:0:(ofd_obd.c:501:ofd_destroy_export()) lustre-OST0000: cli 0b88cf86-b008-4fef-8d8e-b1fd14341648/ffff880293b2e678 has 61440 pending on destroyed export LustreError: 377:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_destroy_export: tot_granted 36774464 != fo_tot_granted 36835904 LustreError: 377:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_destroy_export: tot_pending 0 != fo_tot_pending 61440 LustreError: 32458:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0001-osc-ffff8802c6fd6fc8: namespace resource [0x300000400:0x610:0x0].0x0 (ffff8802705a4540) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 32458:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 38 previous similar messages Lustre: lustre-OST0001-osc-ffff8802c6fd6fc8: Connection restored to 192.168.123.100@tcp (at 0@lo) LustreError: 12732:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 45150784 != fo_tot_granted 45212224 LustreError: 12732:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 1 previous similar message LustreError: 12732:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 61440 != fo_tot_pending 122880 LustreError: 12732:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 1 previous similar message LustreError: 12732:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 45089344 != fo_tot_granted 45150784 LustreError: 12732:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 61440 Lustre: 16577:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 509 < left 2117, rollback = 7 Lustre: 16577:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 125367 previous similar messages Lustre: 16577:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 49/196/0, destroy: 1/4/0 Lustre: 16577:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 125560 previous similar messages Lustre: 16577:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 3/3/1, xattr_set: 53/148/0 Lustre: 16577:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 125558 previous similar messages Lustre: 16577:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 246/2117/0, punch: 0/0/0, quota 12/124/0 Lustre: 16577:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 125558 previous similar messages Lustre: 16577:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 50/849/0, delete: 2/5/1 Lustre: 16577:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 125552 previous similar messages Lustre: 16577:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 1/1/0, ref_del: 2/2/1 Lustre: 16577:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 125558 previous similar messages LustreError: 473:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0000-osc-ffff8802c6fd6fc8: namespace resource [0x2c0000400:0x683:0x0].0x0 (ffff880254e2e840) refcount nonzero (2) after lock cleanup; forcing cleanup. LustreError: 473:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 16 previous similar messages LustreError: 12733:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 45089344 != fo_tot_granted 45150784 LustreError: 12733:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 61440 LustreError: 12733:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 45088960 != fo_tot_granted 45150400 LustreError: 12733:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 1 previous similar message LustreError: 12733:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 61440 LustreError: 12733:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 1 previous similar message LustreError: 12738:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 51338560 != fo_tot_granted 51400000 LustreError: 12738:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 1 previous similar message LustreError: 12738:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 61440 LustreError: 12738:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 1 previous similar message LustreError: 12732:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 76650560 != fo_tot_granted 76712000 LustreError: 12732:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 3 previous similar messages LustreError: 12732:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 86016 != fo_tot_pending 147456 LustreError: 12732:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 3 previous similar messages LustreError: 12745:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 109091264 != fo_tot_granted 109152704 LustreError: 12745:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 16 previous similar messages LustreError: 12745:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 28672 != fo_tot_pending 90112 LustreError: 12745:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 16 previous similar messages LustreError: 16881:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0002: failed to get lu_attr of [0x280000403:0xc6:0x0]: rc = -2 LustreError: 19354:0:(mdt_xattr.c:429:mdt_dir_layout_update()) lustre-MDT0001: [0x240000403:0x25c:0x0] migrate mdt index mismatch 1 != 2 LustreError: 11071:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0001_UUID lock: ffff88026b154780/0x14bfa02bafadb806 lrc: 3/0,0 mode: PW/PW res: [0x300000400:0x425:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000000020020 nid: 0@lo remote: 0x14bfa02bafadb7f1 expref: 4332 pid: 26276 timeout: 956 lvb_type: 0 LustreError: 11071:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 1 previous similar message LustreError: 1020:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88025fbc0cc0 x1709747233141440/t0(0) o105->lustre-OST0001@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 11-0: lustre-OST0001-osc-ffff8802995d2e98: operation ost_setattr to node 0@lo failed: rc = -107 Lustre: lustre-OST0001-osc-ffff8802995d2e98: Connection to lustre-OST0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 1 previous similar message LustreError: 1020:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 29 previous similar messages LustreError: 167-0: lustre-OST0001-osc-ffff8802995d2e98: This client was evicted by lustre-OST0001; in progress operations using this service will fail. LustreError: Skipped 1 previous similar message LustreError: 1830:0:(ofd_obd.c:501:ofd_destroy_export()) lustre-OST0001: cli b26b540b-5129-4273-8036-cfcdc698a503/ffff88029264c138 has 61440 pending on destroyed export Lustre: 4123:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.100@tcp:/lustre/fid: [0x240000404:0x2cd:0x0]// may get corrupted (rc -108) Lustre: 4123:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.100@tcp:/lustre/fid: [0x240000404:0x27f:0x0]// may get corrupted (rc -108) LustreError: 2989:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0001-osc-ffff8802995d2e98: namespace resource [0x300000401:0x1413:0x0].0x0 (ffff88007d6856c0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 2989:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 13 previous similar messages Lustre: lustre-OST0001-osc-ffff8802995d2e98: Connection restored to 192.168.123.100@tcp (at 0@lo) Lustre: Skipped 1 previous similar message Lustre: 17854:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 497 < left 610, rollback = 2 Lustre: 17854:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 290 previous similar messages Lustre: 18025:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 6s req@ffff88024c41aee8 x1709747234792000/t0(0) o104->0b88cf86-b008-4fef-8d8e-b1fd14341648@0@lo:0/0 lens 328/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'' LustreError: 12733:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 209849280 != fo_tot_granted 209910720 LustreError: 12733:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 48 previous similar messages LustreError: 12733:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 62336 != fo_tot_pending 123776 LustreError: 12733:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 48 previous similar messages LustreError: 11071:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0003_UUID lock: ffff88026e3dc000/0x14bfa02bafbcbf0b lrc: 3/0,0 mode: PW/PW res: [0x5a5:0x0:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000400020020 nid: 0@lo remote: 0x14bfa02bafbcbeef expref: 5505 pid: 30689 timeout: 1061 lvb_type: 0 LustreError: 11042:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802560e0cc0 x1709747237545792/t0(0) o105->lustre-OST0003@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 11042:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 6 previous similar messages LustreError: 11-0: lustre-OST0003-osc-ffff8802c6fd6fc8: operation ost_punch to node 0@lo failed: rc = -107 Lustre: lustre-OST0003-osc-ffff8802c6fd6fc8: Connection to lustre-OST0003 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0003-osc-ffff8802c6fd6fc8: This client was evicted by lustre-OST0003; in progress operations using this service will fail. LustreError: 1830:0:(ofd_obd.c:501:ofd_destroy_export()) lustre-OST0003: cli 0b88cf86-b008-4fef-8d8e-b1fd14341648/ffff88026de2ae98 has 90112 pending on destroyed export Lustre: 12740:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 6s req@ffff880088fbb240 x1709747237761024/t0(0) o13->lustre-MDT0001-mdtlov_UUID@0@lo:0/0 lens 224/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'osp-pre-3-1.0' Lustre: 12740:0:(service.c:2157:ptlrpc_server_handle_req_in()) Skipped 1 previous similar message Lustre: 4118:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.100@tcp:/lustre/fid: [0x280000403:0x397:0x0]// may get corrupted (rc -108) LustreError: 5083:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0003-osc-ffff8802c6fd6fc8: namespace resource [0xbe0:0x0:0x0].0x0 (ffff880270128a40) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 5083:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 79 previous similar messages Lustre: lustre-OST0003-osc-ffff8802c6fd6fc8: Connection restored to 192.168.123.100@tcp (at 0@lo) LustreError: 4858:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff8802995d2e98: inode [0x280000403:0x315:0x0] mdc close failed: rc = -13 Lustre: dir [0x280000403:0x25e:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 9 previous similar messages Lustre: 4118:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff88024bbe70c0 x1709747240047616/t0(0) o10->lustre-OST0003-osc-ffff8802995d2e98@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630543080 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dd.0' LustreError: 19905:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 240219584 != fo_tot_granted 240281024 LustreError: 19905:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 147 previous similar messages LustreError: 19905:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 61440 LustreError: 19905:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 148 previous similar messages LustreError: 12923:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88025171b880 x1709747242565376/t0(0) o105->lustre-OST0000@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 11-0: lustre-OST0000-osc-ffff8802995d2e98: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: 12923:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 11 previous similar messages Lustre: 4131:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.100@tcp:/lustre/fid: [0x240000404:0x3a6:0x0]/ may get corrupted (rc -108) LustreError: 7494:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0000-osc-ffff8802995d2e98: namespace resource [0x13d0:0x0:0x0].0x0 (ffff88026c5947c0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 7494:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 409 previous similar messages LustreError: 11381:0:(out_handler.c:910:out_tx_end()) lustre-MDT0002-osd: undo for /home/green/git/lustre-release/lustre/ptlrpc/../../lustre/target/out_handler.c:445: rc = -524 LustreError: 11957:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0002-osp-MDT0001: fail to cancel 1 llog-records: rc = -2 LustreError: 11957:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0002-osp-MDT0001: fail to cancel 1 of 1 llog-records: rc = -2 LustreError: 11957:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0000-osp-MDT0001: fail to cancel 1 llog-records: rc = -116 LustreError: 11957:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0000-osp-MDT0001: fail to cancel 1 of 1 llog-records: rc = -116 Lustre: 4124:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff880277fc7700 x1709747247020608/t0(0) o10->lustre-OST0001-osc-ffff8802995d2e98@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630543227 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'cp.0' Lustre: 4124:0:(client.c:1485:after_reply()) Skipped 3 previous similar messages LustreError: 11071:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 101s: evicting client at 0@lo ns: filter-lustre-OST0002_UUID lock: ffff880078ff0b80/0x14bfa02bafc81587 lrc: 3/0,0 mode: PW/PW res: [0x584:0x0:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 262144->18446744073709551615) gid 0 flags: 0x60000080020020 nid: 0@lo remote: 0x14bfa02bafc81580 expref: 7311 pid: 12699 timeout: 1277 lvb_type: 0 LustreError: 11071:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 1 previous similar message LustreError: 11049:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630543171 with bad export cookie 1495089710801464375 LustreError: 11049:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) Skipped 2 previous similar messages LustreError: 11-0: lustre-OST0002-osc-ffff8802c6fd6fc8: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-OST0002-osc-ffff8802c6fd6fc8: Connection to lustre-OST0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 1 previous similar message LustreError: 167-0: lustre-OST0002-osc-ffff8802c6fd6fc8: This client was evicted by lustre-OST0002; in progress operations using this service will fail. LustreError: Skipped 1 previous similar message Lustre: 4133:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.100@tcp:/lustre/fid: [0x240000403:0x3f1:0x0]// may get corrupted (rc -108) LustreError: 417:0:(ofd_obd.c:501:ofd_destroy_export()) lustre-OST0002: cli 0b88cf86-b008-4fef-8d8e-b1fd14341648/ffff8800ab904138 has 28672 pending on destroyed export Lustre: lustre-OST0002-osc-ffff8802c6fd6fc8: Connection restored to 192.168.123.100@tcp (at 0@lo) Lustre: Skipped 1 previous similar message Lustre: lustre-OST0002-osc-ffff8802c6fd6fc8: disconnect after 20s idle ptlrpc_watchdog_fire: 4 callbacks suppressed Lustre: ll_ost06_004: service thread pid 17790 was inactive for 46.022 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 17790, comm: ll_ost06_004 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] tgt_extent_lock+0xea/0x2a0 [ptlrpc] [<0>] ofd_getattr_hdl+0x365/0x720 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 5246, comm: ll_ost06_050 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] tgt_extent_lock+0xea/0x2a0 [ptlrpc] [<0>] ofd_getattr_hdl+0x365/0x720 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: ll_ost06_051: service thread pid 5247 was inactive for 46.062 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Pid: 30750, comm: ll_ost06_037 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] tgt_extent_lock+0xea/0x2a0 [ptlrpc] [<0>] ofd_getattr_hdl+0x365/0x720 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: Skipped 3 previous similar messages Lustre: 16577:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000403:0x3fb:0x0] with magic=0xbd60bd0 Lustre: 16770:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000403:0x3e4:0x0] with magic=0xbd60bd0 Lustre: 16770:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 1 previous similar message Lustre: 16770:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000403:0x396:0x0] with magic=0xbd60bd0 Lustre: 16770:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 1 previous similar message Lustre: dir [0x240000403:0x37f:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 2 previous similar messages Lustre: 16577:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000403:0x395:0x0] with magic=0xbd60bd0 Lustre: 16577:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 1 previous similar message LustreError: 11231:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000403:0x2cd:0x0]: rc = -2 Lustre: ll_ost01_039: service thread pid 4625 was inactive for 82.199 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 4 previous similar messages LustreError: 19928:0:(ldlm_lockd.c:1425:ldlm_handle_enqueue0()) ### lock on destroyed export ffff880073e9ca88 ns: filter-lustre-OST0003_UUID lock: ffff880234db61c0/0x14bfa02bafd5e528 lrc: 3/0,0 mode: --/PW res: [0x122b:0x0:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x50000080020000 nid: 0@lo remote: 0x14bfa02bafd5e505 expref: 6283 pid: 19928 timeout: 0 lvb_type: 0 LustreError: 4726:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88029d2df700 x1709747251579200/t0(0) o104->lustre-OST0003@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 18312:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630543265 with bad export cookie 1495089710803436590 LustreError: 18312:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) Skipped 1 previous similar message LustreError: 4726:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 73 previous similar messages LustreError: 10748:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0000-osc-ffff8802995d2e98: namespace resource [0x13da:0x0:0x0].0x0 (ffff8802558a6d40) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 10748:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 551 previous similar messages LustreError: 11537:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0002-osp-MDT0000: fail to cancel 1 llog-records: rc = -2 LustreError: 11537:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) Skipped 1 previous similar message LustreError: 11537:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0002-osp-MDT0000: fail to cancel 1 of 1 llog-records: rc = -2 LustreError: 11537:0:(llog_cat.c:787:llog_cat_cancel_records()) Skipped 1 previous similar message LustreError: 647:0:(ofd_obd.c:501:ofd_destroy_export()) lustre-OST0001: cli b26b540b-5129-4273-8036-cfcdc698a503/ffff880275422548 has 61440 pending on destroyed export Lustre: 21871:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 526/2104/0, destroy: 1/4/0 Lustre: 21871:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 292185 previous similar messages Lustre: 21871:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 3/3/1, xattr_set: 530/148/0 Lustre: 21871:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 292187 previous similar messages Lustre: 21871:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 2631/22628/0, punch: 0/0/0, quota 8/56/0 Lustre: 21871:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 292184 previous similar messages Lustre: 21871:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 527/8958/0, delete: 2/5/1 Lustre: 21871:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 292180 previous similar messages Lustre: 21871:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 1/1/0, ref_del: 2/2/1 Lustre: 21871:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 292178 previous similar messages Lustre: 21871:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 507 < left 22628, rollback = 7 Lustre: 21871:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 291693 previous similar messages LustreError: 12746:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 56823232 != fo_tot_granted 56851904 LustreError: 12746:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 358 previous similar messages LustreError: 12746:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 896 != fo_tot_pending 29568 LustreError: 12746:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 357 previous similar messages LustreError: 11065:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630543375 with bad export cookie 1495089710802177185 LustreError: 11065:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) Skipped 11 previous similar messages LustreError: 12826:0:(ofd_obd.c:501:ofd_destroy_export()) lustre-OST0000: cli 0b88cf86-b008-4fef-8d8e-b1fd14341648/ffff8802843d4138 has 61440 pending on destroyed export Lustre: 4122:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.100@tcp:/lustre/fid: [0x200000403:0x475:0x0]/ may get corrupted (rc -108) Lustre: dir [0x240000403:0x40a:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: 17406:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 502 < left 699, rollback = 2 Lustre: 17406:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 380 previous similar messages LustreError: 11359:0:(out_handler.c:910:out_tx_end()) lustre-MDT0002-osd: undo for /home/green/git/lustre-release/lustre/ptlrpc/../../lustre/target/out_handler.c:445: rc = -524 LustreError: 11359:0:(out_handler.c:910:out_tx_end()) Skipped 1 previous similar message LustreError: 13889:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -5 LustreError: 11537:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0002-osp-MDT0000: fail to cancel 1 llog-records: rc = -2 LustreError: 11537:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0002-osp-MDT0000: fail to cancel 1 of 1 llog-records: rc = -2 LustreError: 11071:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 104s: evicting client at 0@lo ns: filter-lustre-OST0001_UUID lock: ffff8802417187c0/0x14bfa02bafdab242 lrc: 3/0,0 mode: PW/PW res: [0x300000401:0x18a7:0x0].0x0 rrc: 6 type: EXT [0->18446744073709551615] (req 0->134217727) gid 0 flags: 0x60000000020020 nid: 0@lo remote: 0x14bfa02bafdab234 expref: 3169 pid: 23426 timeout: 1601 lvb_type: 0 LustreError: 11071:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 6 previous similar messages LustreError: 11-0: lustre-OST0001-osc-ffff8802995d2e98: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 4 previous similar messages Lustre: lustre-OST0001-osc-ffff8802995d2e98: Connection to lustre-OST0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 5 previous similar messages LustreError: 167-0: lustre-OST0001-osc-ffff8802995d2e98: This client was evicted by lustre-OST0001; in progress operations using this service will fail. LustreError: Skipped 5 previous similar messages Lustre: 4118:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.100@tcp:/lustre/fid: [0x200000403:0x4a8:0x0]/ may get corrupted (rc -5) Lustre: 4119:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.100@tcp:/lustre/fid: [0x200000403:0x47c:0x0]// may get corrupted (rc -108) Lustre: 4118:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.100@tcp:/lustre/fid: [0x200000405:0x338:0x0]// may get corrupted (rc -108) Lustre: 4119:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.100@tcp:/lustre/fid: [0x280000404:0x5c2:0x0]/ may get corrupted (rc -108) LustreError: 676:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630543513 with bad export cookie 1495089710804090222 LustreError: 676:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) Skipped 2 previous similar messages LustreError: 15749:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0001-osc-ffff8802995d2e98: namespace resource [0x300000400:0x20c0:0x0].0x0 (ffff88028148b240) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 15749:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 564 previous similar messages LustreError: 5256:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88028ccf4b40 x1709747263029376/t0(0) o106->lustre-OST0000@0@lo:15/16 lens 328/280 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 5256:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 616 previous similar messages LustreError: 417:0:(ofd_obd.c:501:ofd_destroy_export()) lustre-OST0002: cli 0b88cf86-b008-4fef-8d8e-b1fd14341648/ffff8800ab9237e8 has 28672 pending on destroyed export Lustre: lustre-OST0001-osc-ffff8802995d2e98: Connection restored to 192.168.123.100@tcp (at 0@lo) Lustre: Skipped 5 previous similar messages Lustre: 4130:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.100@tcp:/lustre/fid: [0x200000403:0x4a8:0x0]/ may get corrupted (rc -108) Lustre: 4133:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.100@tcp:/lustre/fid: [0x280000403:0x4e4:0x0]// may get corrupted (rc -108) Lustre: 13486:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1630543539/real 1630543539] req@ffff880261210680 x1709747263642752/t0(0) o101->lustre-MDT0001-mdc-ffff8802c6fd6fc8@0@lo:12/10 lens 576/9536 e 0 to 1 dl 1630543642 ref 2 fl Rpc:XQr/2/ffffffff rc 0/-1 job:'dd.0' Lustre: lustre-MDT0001: Client 0b88cf86-b008-4fef-8d8e-b1fd14341648 (at 0@lo) reconnecting ptlrpc_watchdog_fire: 32 callbacks suppressed Lustre: mdt01_011: service thread pid 21349 was inactive for 56.171 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Lustre: Skipped 2 previous similar messages Pid: 21349, comm: mdt01_011 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_fini+0xa25/0xec0 [ptlrpc] [<0>] ldlm_cli_enqueue+0x436/0xa30 [ptlrpc] [<0>] osp_md_object_lock+0x160/0x300 [osp] [<0>] lod_object_lock+0xdb/0x7d0 [lod] [<0>] mdd_object_lock+0x2d/0xd0 [mdd] [<0>] mdt_remote_object_lock_try+0x1cc/0x510 [mdt] [<0>] mdt_remote_object_lock+0x2a/0x30 [mdt] [<0>] mdt_rename_lock+0xbe/0x4d0 [mdt] [<0>] mdt_reint_migrate+0x6e8/0x11d0 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x76c/0xb40 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe ------------[ cut here ]------------ WARNING: CPU: 11 PID: 19035 at /home/green/git/lustre-release/ldiskfs/ext4_jbd2.c:266 __ldiskfs_handle_dirty_metadata+0x1d2/0x230 [ldiskfs] Modules linked in: loop zfs(PO) zunicode(PO) zzstd(O) zlua(O) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) jbd2 mbcache lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) dm_flakey dm_mod libcfs(OE) crc_t10dif crct10dif_generic sb_edac edac_core iosf_mbi crc32_pclmul ghash_clmulni_intel aesni_intel lrw gf128mul glue_helper ablk_helper cryptd virtio_balloon pcspkr virtio_console i2c_piix4 ip_tables rpcsec_gss_krb5 ata_generic drm_kms_helper pata_acpi ttm crct10dif_pclmul drm crct10dif_common drm_panel_orientation_quirks ata_piix crc32c_intel serio_raw virtio_blk i2c_core libata floppy CPU: 11 PID: 19035 Comm: mdt05_009 Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011 Call Trace: [<ffffffff817d9418>] dump_stack+0x19/0x1b [<ffffffff8108d538>] __warn+0xd8/0x100 [<ffffffff8108d67d>] warn_slowpath_null+0x1d/0x20 [<ffffffffa0c17262>] __ldiskfs_handle_dirty_metadata+0x1d2/0x230 [ldiskfs] [<ffffffffa0bf733d>] ldiskfs_getblk+0x13d/0x210 [ldiskfs] [<ffffffffa0bf7437>] ldiskfs_bread+0x27/0xe0 [ldiskfs] [<ffffffffa0cbbe97>] osd_write+0x4e7/0xcb0 [osd_ldiskfs] [<ffffffffa04cd6f3>] dt_record_write+0x33/0x120 [obdclass] [<ffffffffa048c6b2>] llog_osd_write_rec+0x1052/0x19a0 [obdclass] [<ffffffffa0479760>] llog_write_rec+0x290/0x590 [obdclass] [<ffffffffa047f4ee>] llog_cat_add_rec+0x23e/0x950 [obdclass] [<ffffffffa04765ff>] llog_add+0x17f/0x1f0 [obdclass] [<ffffffffa04a7a39>] ? lprocfs_counter_add+0xf9/0x160 [obdclass] [<ffffffffa081060d>] sub_updates_write+0xb6b/0xe32 [ptlrpc] [<ffffffffa07e0892>] top_trans_stop+0x7c2/0xf60 [ptlrpc] [<ffffffffa0ef6f3c>] lod_trans_stop+0x25c/0x340 [lod] [<ffffffffa0dcb408>] mdd_trans_stop+0x28/0x16e [mdd] [<ffffffffa0dacb2f>] mdd_migrate_object+0xf1f/0x1ba0 [mdd] [<ffffffffa04cc375>] ? lu_object_find_at+0x285/0xb10 [obdclass] [<ffffffffa0dadb86>] mdd_migrate+0x3d6/0x8e0 [mdd] [<ffffffffa0e7f0d8>] mdo_migrate+0x4c/0x4e [mdt] [<ffffffffa0e3ee53>] mdt_reint_migrate+0x1023/0x11d0 [mdt] [<ffffffffa0e3f087>] mdt_reint_rec+0x87/0x240 [mdt] [<ffffffffa0e1455c>] mdt_reint_internal+0x76c/0xb40 [mdt] [<ffffffffa0e21657>] mdt_reint+0x67/0x150 [mdt] [<ffffffffa07cf26e>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<ffffffffa02f245e>] ? libcfs_nid2str_r+0xfe/0x130 [lnet] [<ffffffffa077a990>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<ffffffffa077c529>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffffa077b950>] ? ptlrpc_wait_event+0x620/0x620 [ptlrpc] [<ffffffff810ba0f4>] kthread+0xe4/0xf0 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 ---[ end trace fc07225a3244f179 ]--- LDISKFS-fs: ldiskfs_getblk:888: aborting transaction: error 28 in __ldiskfs_handle_dirty_metadata LDISKFS-fs error (device dm-2): ldiskfs_getblk:888: inode #162: block 20741: comm mdt05_009: journal_dirty_metadata failed: handle type 0 started at line 1972, credits 512/0, errcode -28 Aborting journal on device dm-2-8. LDISKFS-fs (dm-2): Remounting filesystem read-only LustreError: 19035:0:(osd_io.c:2103:osd_ldiskfs_write_record()) lustre-MDT0002: error reading offset 327680 (block 80, size 56, offs 327680), credits 0/0: rc = -28 LustreError: 19035:0:(update_trans.c:985:top_trans_stop()) lustre-MDT0002-osd: write updates failed: rc = -28 LustreError: 19035:0:(osd_io.c:2110:osd_ldiskfs_write_record()) journal_get_write_access() returned error -30 LustreError: 19035:0:(tgt_lastrcvd.c:1253:tgt_add_reply_data()) lustre-MDT0002: can't update reply_data file: rc = -30 LustreError: 19035:0:(osd_handler.c:2075:osd_trans_stop()) lustre-MDT0002: failed in transaction hook: rc = -30 LDISKFS-fs error (device dm-2) in osd_trans_stop:2082: error 28 LustreError: 14048:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-OST0002-osc-MDT0002: fail to cancel 73 llog-records: rc = -30 LustreError: 13492:0:(osp_sync.c:1094:osp_sync_process_committed()) lustre-OST0001-osc-MDT0002: can't cancel 45 records: rc = -30 LustreError: 14048:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) Skipped 4 previous similar messages LustreError: 18472:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -30 LustreError: 12310:0:(osd_handler.c:1779:osd_trans_commit_cb()) transaction @0xffff88007daf44a8 commit error: 2 LustreError: 19035:0:(osd_handler.c:2085:osd_trans_stop()) lustre-MDT0002: failed to stop transaction: rc = -28 LustreError: 19035:0:(update_trans.c:1011:top_trans_stop()) lustre-MDT0002-osd: stop trans failed: rc = -30 LustreError: 25718:0:(tgt_handler.c:1367:tgt_blocking_ast()) lustre-MDT0002: syncing [0x280000404:0x527:0x0] (2-0) on lock cancel: rc = -5 LustreError: 14048:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-OST0002-osc-MDT0002: fail to cancel 1 of 1 llog-records: rc = -30 LustreError: 14048:0:(llog_cat.c:787:llog_cat_cancel_records()) Skipped 1 previous similar message LustreError: 14048:0:(osp_sync.c:1079:osp_sync_process_committed()) lustre-OST0002-osc-MDT0002: can't cancel record: rc = -30 LustreError: 13492:0:(osp_sync.c:1094:osp_sync_process_committed()) lustre-OST0001-osc-MDT0002: can't cancel 6 records: rc = -30 LustreError: 13492:0:(osp_sync.c:1094:osp_sync_process_committed()) Skipped 7 previous similar messages LustreError: 18325:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -30 LustreError: 18325:0:(llite_lib.c:1836:ll_md_setattr()) Skipped 2 previous similar messages | Externally reported by onyx-68 boilpot email |
racer test 1: racer on clients: centos-40.localnet DURATION=2700 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa178de52>] ll_migrate+0x9b2/0xec0 [lustre] PGD 80000002f24c4067 PUD 27e9c8067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) dm_flakey dm_mod loop zfs(PO) zunicode(PO) zzstd(O) zlua(O) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) jbd2 mbcache crc_t10dif crct10dif_generic sb_edac edac_core iosf_mbi crc32_pclmul ghash_clmulni_intel aesni_intel lrw gf128mul glue_helper ablk_helper cryptd pcspkr virtio_console virtio_balloon i2c_piix4 ip_tables rpcsec_gss_krb5 ata_generic drm_kms_helper pata_acpi ttm drm crct10dif_pclmul crct10dif_common ata_piix drm_panel_orientation_quirks crc32c_intel serio_raw virtio_blk libata i2c_core floppy [last unloaded: libcfs] CPU: 12 PID: 14358 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011 task: ffff88027fea49d0 ti: ffff880275b18000 task.ti: ffff880275b18000 RIP: 0010:[<ffffffffa178de52>] [<ffffffffa178de52>] ll_migrate+0x9b2/0xec0 [lustre] RSP: 0018:ffff880275b1bbc8 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff880250343398 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff8800b8508268 RDI: ffff8800b8508248 RBP: ffff880275b1bc48 R08: ffff8800823b4618 R09: ffff880275b1ba60 R10: 0000000000000004 R11: ffff8800a7489200 R12: 0000000000000000 R13: ffff8802651309f8 R14: ffff880262d4d458 R15: 0000000000000030 FS: 00007f88e0a78740(0000) GS:ffff880331d00000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 CR2: 0000000000000008 CR3: 00000002403a0000 CR4: 00000000001607e0 Call Trace: [<ffffffff81242213>] ? __check_object_size+0x1c3/0x220 [<ffffffffa177b561>] ll_dir_ioctl+0x5d01/0x6ed0 [lustre] [<ffffffff811b515b>] ? unlock_page+0x2b/0x30 [<ffffffff811e5e2e>] ? do_read_fault.isra.64+0x12e/0x170 [<ffffffff811ea1cf>] ? handle_pte_fault+0x2ef/0xca0 [<ffffffff811ecf6d>] ? __handle_mm_fault+0x31d/0x720 [<ffffffff810d5248>] ? __enqueue_entity+0x78/0x80 [<ffffffff810d5248>] ? __enqueue_entity+0x78/0x80 [<ffffffff810dbb8b>] ? put_prev_entity+0x8b/0x400 [<ffffffff8125b3fd>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff81245a38>] ? vfs_write+0x178/0x230 [<ffffffff8125b751>] SyS_ioctl+0xa1/0xc0 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 | Lustre: 29438:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 512 < left 1800, rollback = 7 Lustre: 29438:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 29438:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 1/89/0 Lustre: 29438:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1791/1800/0, punch: 0/0/0, quota 3/3/0 Lustre: 29438:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 29438:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 29547:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 511 < left 1395, rollback = 7 Lustre: 29547:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 2095 previous similar messages Lustre: 29547:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 29547:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 2095 previous similar messages Lustre: 29547:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 1/89/0 Lustre: 29547:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 2095 previous similar messages Lustre: 29547:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1387/1396/1, punch: 0/0/0, quota 3/3/0 Lustre: 29547:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 2095 previous similar messages Lustre: 29547:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 29547:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 2095 previous similar messages Lustre: 29547:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 29547:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 2095 previous similar messages Lustre: 26750:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 512 < left 2005, rollback = 7 Lustre: 26750:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 1081 previous similar messages Lustre: 26750:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 26750:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 1081 previous similar messages Lustre: 26750:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 1/89/0 Lustre: 26750:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 1081 previous similar messages Lustre: 26750:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1996/2005/0, punch: 0/0/0, quota 3/3/0 Lustre: 26750:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 1081 previous similar messages Lustre: 26750:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 26750:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 1081 previous similar messages Lustre: 26750:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 26750:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 1081 previous similar messages Lustre: 26756:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 498 < left 582, rollback = 2 Lustre: 26756:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 3/12/10, destroy: 0/0/0 Lustre: 26756:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 3398 previous similar messages Lustre: 26756:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 6/582/0 Lustre: 26756:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 3398 previous similar messages Lustre: 26756:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 5/136/0, punch: 0/0/0, quota 8/56/0 Lustre: 26756:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 3398 previous similar messages Lustre: 26756:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 10/203/4, delete: 0/0/0 Lustre: 26756:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 3398 previous similar messages Lustre: 26756:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 4/4/0, ref_del: 0/0/0 Lustre: 26756:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 3398 previous similar messages Lustre: 29529:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 8: before 510 < left 23850, rollback = 9 Lustre: 29529:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 5 previous similar messages Lustre: 29529:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 1402/5608/0, destroy: 1/4/0 Lustre: 29529:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 29529:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 3/3/0, xattr_set: 1406/148/0 Lustre: 29529:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 29529:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 7011/60296/0, punch: 0/0/0, quota 9/65/0 Lustre: 29529:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 29529:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 1403/23850/0, delete: 2/5/1 Lustre: 29529:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 29529:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 1/1/0, ref_del: 2/2/1 Lustre: 29529:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 5 previous similar messages Lustre: 29529:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 508 < left 60296, rollback = 7 Lustre: 29529:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 3398 previous similar messages Lustre: 26760:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 497 < left 582, rollback = 2 Lustre: 31188:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 480 < left 76998, rollback = 7 Lustre: 31188:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 9107 previous similar messages sched: RT throttling activated Lustre: 26747:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 26747:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 13617 previous similar messages Lustre: 26747:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 1/89/0 Lustre: 26747:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 13617 previous similar messages Lustre: 26747:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 940/949/0, punch: 0/0/0, quota 3/3/0 Lustre: 26747:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 13617 previous similar messages Lustre: 26747:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 26747:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 13617 previous similar messages Lustre: 26747:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 26747:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 13617 previous similar messages Lustre: 29438:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 8: before 509 < left 21844, rollback = 9 Lustre: 29438:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 5 previous similar messages LustreError: 26755:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0000: '16' migration was interrupted, run 'lfs migrate -m 1 -c 1 -H crush 16' to finish migration. Lustre: 26765:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 511 < left 1412, rollback = 7 Lustre: 26765:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 5444 previous similar messages Lustre: 26751:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 8: before 512 < left 30463, rollback = 9 Lustre: 26751:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 5 previous similar messages Lustre: 26750:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 3/12/2, destroy: 1/4/0 Lustre: 26750:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 2352 previous similar messages Lustre: dir [0x240000404:0x48:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: 31556:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/1, xattr_set: 1998/15/0 Lustre: 31556:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 2367 previous similar messages Lustre: 31556:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 9981/85838/0, punch: 0/0/0, quota 14/174/1 Lustre: 31556:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 2367 previous similar messages Lustre: 31556:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 1996/33932/0, delete: 0/0/0 Lustre: 31556:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 2367 previous similar messages Lustre: 31556:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 31556:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 2367 previous similar messages Lustre: 26770:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 509 < left 77023, rollback = 7 Lustre: 26770:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 10810 previous similar messages ------------[ cut here ]------------ WARNING: CPU: 4 PID: 26770 at /home/green/git/lustre-release/ldiskfs/namei.c:3331 ldiskfs_orphan_add+0x11e/0x2a0 [ldiskfs] Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) dm_flakey dm_mod loop zfs(PO) zunicode(PO) zzstd(O) zlua(O) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) jbd2 mbcache crc_t10dif crct10dif_generic sb_edac edac_core iosf_mbi crc32_pclmul ghash_clmulni_intel aesni_intel lrw gf128mul glue_helper ablk_helper cryptd pcspkr virtio_console virtio_balloon i2c_piix4 ip_tables rpcsec_gss_krb5 ata_generic drm_kms_helper pata_acpi ttm drm crct10dif_pclmul crct10dif_common ata_piix drm_panel_orientation_quirks crc32c_intel serio_raw virtio_blk libata i2c_core floppy [last unloaded: libcfs] CPU: 4 PID: 26770 Comm: mdt_rdpg02_000 Kdump: loaded Tainted: P OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011 Call Trace: [<ffffffff817d9418>] dump_stack+0x19/0x1b [<ffffffff8108d538>] __warn+0xd8/0x100 [<ffffffff8108d67d>] warn_slowpath_null+0x1d/0x20 [<ffffffffa0d594ae>] ldiskfs_orphan_add+0x11e/0x2a0 [ldiskfs] [<ffffffffa0d4a9eb>] ldiskfs_xattr_inode_orphan_add.constprop.17+0xbb/0x110 [ldiskfs] [<ffffffffa0d4d417>] ldiskfs_xattr_delete_inode+0x77/0x3c0 [ldiskfs] [<ffffffffa0d6daa0>] ? ldiskfs_evict_inode+0x190/0x670 [ldiskfs] [<ffffffffa0d6dad9>] ldiskfs_evict_inode+0x1c9/0x670 [ldiskfs] [<ffffffff812649af>] evict+0xaf/0x180 [<ffffffff81264de5>] iput+0xf5/0x180 [<ffffffffa0c89d08>] osd_object_delete+0x1f8/0x370 [osd_ldiskfs] [<ffffffffa04d659d>] lu_object_free.isra.26+0xcd/0x1d0 [obdclass] [<ffffffffa04dacf1>] lu_object_put+0xa1/0x480 [obdclass] [<ffffffffa0ee9af4>] mdt_mfd_close+0x1f4/0x870 [mdt] [<ffffffffa0eefe14>] mdt_close_internal+0xb4/0x240 [mdt] [<ffffffffa0ef0224>] mdt_close+0x284/0x850 [mdt] [<ffffffffa07dd2d4>] ? tgt_request_preprocess.isra.27+0xf4/0x850 [ptlrpc] [<ffffffffa07de26e>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<ffffffffa03c245e>] ? libcfs_nid2str_r+0xfe/0x130 [lnet] [<ffffffffa0789990>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<ffffffffa078b529>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffffa078a950>] ? ptlrpc_wait_event+0x620/0x620 [ptlrpc] [<ffffffff810ba0f4>] kthread+0xe4/0xf0 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 ---[ end trace 9deceedc8d9bbe24 ]--- Lustre: 26759:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 8: before 511 < left 16012, rollback = 9 Lustre: 26759:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 15 previous similar messages Lustre: 26758:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 26758:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 23767 previous similar messages Lustre: 26744:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 8: before 509 < left 18393, rollback = 9 Lustre: 26744:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 3/3/1, xattr_set: 1085/148/0 Lustre: 26744:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 26548 previous similar messages Lustre: 26744:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 5406/46493/0, punch: 0/0/0, quota 6/6/0 Lustre: 26744:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 26549 previous similar messages Lustre: 26744:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 1082/18393/0, delete: 2/5/1 Lustre: 26744:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 26549 previous similar messages Lustre: 26744:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 1/1/0, ref_del: 2/2/1 Lustre: 26744:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 26549 previous similar messages Lustre: 26744:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 507 < left 46493, rollback = 7 Lustre: 26744:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 17141 previous similar messages Lustre: mdt06_007: service thread pid 31776 was inactive for 62.155 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 26750, comm: mdt02_002 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] mdt_object_local_lock+0x527/0xb90 [mdt] [<0>] mdt_object_lock_internal+0x70/0x390 [mdt] [<0>] mdt_reint_object_lock+0x2c/0x60 [mdt] [<0>] mdt_reint_striped_lock+0x89/0x5b0 [mdt] [<0>] mdt_attr_set+0x9c/0x800 [mdt] [<0>] mdt_reint_setattr+0x77a/0x1020 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x76c/0xb40 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: Skipped 1 previous similar message Pid: 31776, comm: mdt06_007 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] mdt_object_local_lock+0x527/0xb90 [mdt] [<0>] mdt_object_lock_internal+0x70/0x390 [mdt] [<0>] mdt_reint_object_lock+0x2c/0x60 [mdt] [<0>] mdt_reint_striped_lock+0x89/0x5b0 [mdt] [<0>] mdt_reint_unlink+0x8dd/0x19b0 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x76c/0xb40 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 32603, comm: mdt03_011 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] mdt_object_local_lock+0x527/0xb90 [mdt] [<0>] mdt_object_lock_internal+0x70/0x390 [mdt] [<0>] mdt_getattr_name_lock+0x1130/0x2590 [mdt] [<0>] mdt_intent_getattr+0x2d5/0x4b0 [mdt] [<0>] mdt_intent_opc+0x1dd/0xc10 [mdt] [<0>] mdt_intent_policy+0x1a1/0x360 [mdt] [<0>] ldlm_lock_enqueue+0x3c2/0xb40 [ptlrpc] [<0>] ldlm_handle_enqueue0+0xaa6/0x16d0 [ptlrpc] [<0>] tgt_enqueue+0x64/0x240 [ptlrpc] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: mdt03_010: service thread pid 32003 was inactive for 62.071 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: mdt07_002: service thread pid 26765 was inactive for 62.189 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 2 previous similar messages Lustre: mdt01_000: service thread pid 26745 was inactive for 62.097 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 1 previous similar message LustreError: 26735:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0002_UUID lock: ffff88026133ed00/0x368f4f940a97c194 lrc: 3/0,0 mode: PW/PW res: [0x340000400:0x12:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000400020020 nid: 0@lo remote: 0x368f4f940a97c163 expref: 2076 pid: 27442 timeout: 4204 lvb_type: 0 LustreError: 31021:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff880254274500 x1709743670718976/t0(0) o106->lustre-OST0002@0@lo:15/16 lens 328/280 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 11-0: lustre-OST0002-osc-ffff8802eba492a8: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-OST0002-osc-ffff8802eba492a8: Connection to lustre-OST0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0002-osc-ffff8802eba492a8: This client was evicted by lustre-OST0002; in progress operations using this service will fail. LustreError: 1143:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff880254a15e00 x1709743670744768/t0(0) o104->lustre-OST0002@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 1143:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 29 previous similar messages Lustre: mdt06_001: service thread pid 26761 was inactive for 62.065 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: 23560:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.140@tcp:/lustre/fid: [0x200000403:0x169:0x0]// may get corrupted (rc -108) Lustre: lustre-OST0002-osc-ffff8802eba492a8: Connection restored to 192.168.123.140@tcp (at 0@lo) Lustre: mdt05_007: service thread pid 32607 was inactive for 62.238 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 1 previous similar message LustreError: 26735:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0001_UUID lock: ffff880255ae0f40/0x368f4f940a9bf1f2 lrc: 3/0,0 mode: PW/PW res: [0x300000401:0x344:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 2097152->18446744073709551615) gid 0 flags: 0x60000080020020 nid: 0@lo remote: 0x368f4f940a9bf1e4 expref: 3597 pid: 31021 timeout: 4220 lvb_type: 0 LustreError: 32229:0:(ldlm_lockd.c:1425:ldlm_handle_enqueue0()) ### lock on destroyed export ffff8802ebba8008 ns: filter-lustre-OST0001_UUID lock: ffff8802460707c0/0x368f4f940aa53a4f lrc: 3/0,0 mode: --/PW res: [0x300000401:0xb0:0x0].0x0 rrc: 7 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x50000000020000 nid: 0@lo remote: 0x368f4f940aa53a33 expref: 3594 pid: 32229 timeout: 0 lvb_type: 0 LustreError: 29870:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630538832 with bad export cookie 3931448496969056262 Lustre: lustre-OST0001-osc-ffff8800a64a2e98: Connection to lustre-OST0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 11-0: lustre-OST0001-osc-ffff8800a64a2e98: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: 167-0: lustre-OST0001-osc-ffff8800a64a2e98: This client was evicted by lustre-OST0001; in progress operations using this service will fail. LustreError: 32027:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88027835a5c0 x1709743671999936/t0(0) o104->lustre-OST0001@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 32027:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 2 previous similar messages Lustre: lustre-OST0001-osc-ffff8800a64a2e98: Connection restored to 192.168.123.140@tcp (at 0@lo) Lustre: mdt05_002: service thread pid 26759 was inactive for 62.196 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: 26775:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 1081/4324/0, destroy: 1/4/1 Lustre: 26775:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 15518 previous similar messages Lustre: 23564:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff8802724cf700 x1709743672746112/t0(0) o10->lustre-OST0000-osc-ffff8800a64a2e98@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630538897 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dd.0' LustreError: 26735:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0001_UUID lock: ffff8802670ab4c0/0x368f4f940a9cab72 lrc: 4/0,0 mode: PR/PR res: [0x240000404:0x5:0x0].0x0 bits 0x1b/0x0 rrc: 6 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0x368f4f940a9cab56 expref: 57 pid: 26753 timeout: 4235 lvb_type: 0 Lustre: lustre-MDT0001-mdc-ffff8802eba492a8: Connection to lustre-MDT0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-MDT0001-mdc-ffff8802eba492a8: This client was evicted by lustre-MDT0001; in progress operations using this service will fail. LustreError: 9227:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff8802eba492a8: inode [0x240000403:0x1:0x0] mdc close failed: rc = -108 Lustre: lustre-MDT0001-mdc-ffff8802eba492a8: Connection restored to 192.168.123.140@tcp (at 0@lo) Lustre: 26765:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 503 < left 760, rollback = 2 Lustre: 26765:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 1 previous similar message Lustre: 531:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 6/582/0 Lustre: 531:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 22794 previous similar messages Lustre: 531:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 5/136/0, punch: 0/0/0, quota 6/6/0 Lustre: 531:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 22794 previous similar messages Lustre: 531:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 10/203/4, delete: 0/0/0 Lustre: 531:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 22794 previous similar messages Lustre: 531:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 4/4/0, ref_del: 0/0/0 Lustre: 531:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 22794 previous similar messages Lustre: 29819:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 509 < left 18371, rollback = 7 Lustre: 29819:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 22782 previous similar messages Lustre: 23556:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff880254036440 x1709743674585216/t0(0) o10->lustre-OST0001-osc-ffff8800a64a2e98@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630538913 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'cp.0' Lustre: 23558:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff88025f541940 x1709743676414592/t0(0) o10->lustre-OST0003-osc-ffff8802eba492a8@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630538941 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'truncate.0' LustreError: 26735:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0003_UUID lock: ffff88007a514000/0x368f4f940aa53ab1 lrc: 3/0,0 mode: PW/PW res: [0x380000400:0x6:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 2097152->18446744073709551615) gid 0 flags: 0x60000000020020 nid: 0@lo remote: 0x368f4f940aa53aa3 expref: 3257 pid: 31021 timeout: 4320 lvb_type: 0 Lustre: 9437:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 505 < left 1400, rollback = 2 Lustre: 9437:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 44 previous similar messages LustreError: 28247:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88029493bec0 x1709743679710400/t0(0) o104->lustre-OST0003@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 28247:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 32 previous similar messages LustreError: 11-0: lustre-OST0003-osc-ffff8802eba492a8: operation ost_setattr to node 0@lo failed: rc = -107 Lustre: lustre-OST0003-osc-ffff8802eba492a8: Connection to lustre-OST0003 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0003-osc-ffff8802eba492a8: This client was evicted by lustre-OST0003; in progress operations using this service will fail. LustreError: 12121:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0003-osc-ffff8802eba492a8: namespace resource [0x670:0x0:0x0].0x0 (ffff880268597240) refcount nonzero (1) after lock cleanup; forcing cleanup. Lustre: lustre-OST0003-osc-ffff8802eba492a8: Connection restored to 192.168.123.140@tcp (at 0@lo) Lustre: 23562:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff8802f089e440 x1709743681207104/t0(0) o10->lustre-OST0002-osc-ffff8802eba492a8@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630538981 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dir_create.sh.0' LustreError: 26735:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0000_UUID lock: ffff88008133c000/0x368f4f940aa72472 lrc: 3/0,0 mode: PW/PW res: [0x59d:0x0:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000000020020 nid: 0@lo remote: 0x368f4f940aa72464 expref: 4828 pid: 32140 timeout: 4346 lvb_type: 0 LustreError: 26727:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802438d1f80 x1709743681373504/t0(0) o105->lustre-OST0000@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 11-0: lustre-OST0000-osc-ffff8800a64a2e98: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-OST0000-osc-ffff8800a64a2e98: Connection to lustre-OST0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 26727:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 3 previous similar messages LustreError: 167-0: lustre-OST0000-osc-ffff8800a64a2e98: This client was evicted by lustre-OST0000; in progress operations using this service will fail. Lustre: dir [0x280000404:0x8b:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: 11906:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 3/12/8, destroy: 0/0/0 Lustre: 11906:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 72213 previous similar messages LustreError: 1480:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0002: '12' migration was interrupted, run 'lfs migrate -m 0 -c 1 -H crush 12' to finish migration. Lustre: 23565:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.140@tcp:/lustre/fid: [0x200000403:0x239:0x0]/ may get corrupted (rc -108) Lustre: 23566:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.140@tcp:/lustre/fid: [0x240000405:0x72:0x0]/ may get corrupted (rc -108) LustreError: 12985:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0000-osc-ffff8800a64a2e98: namespace resource [0x2c0000400:0x270:0x0].0x0 (ffff88026015efc0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 12985:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 275 previous similar messages LustreError: 12985:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0000-osc-ffff8800a64a2e98: namespace resource [0x2c0000400:0x4b0:0x0].0x0 (ffff880266fb5bc0) refcount nonzero (1) after lock cleanup; forcing cleanup. Lustre: lustre-OST0000-osc-ffff8800a64a2e98: Connection restored to 192.168.123.140@tcp (at 0@lo) Lustre: 32058:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 505 < left 82699, rollback = 7 Lustre: 32058:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 65946 previous similar messages Lustre: 32058:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/1, xattr_set: 1925/15/0 Lustre: 32058:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 66048 previous similar messages Lustre: 32058:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 9616/82699/0, punch: 0/0/0, quota 14/174/6 Lustre: 32058:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 66048 previous similar messages Lustre: 32058:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 1923/32691/0, delete: 0/0/0 Lustre: 32058:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 66048 previous similar messages Lustre: 32058:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 32058:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 66048 previous similar messages LustreError: 30155:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0002: '12' migration was interrupted, run 'lfs migrate -m 0 -c 1 -H crush 12' to finish migration. LustreError: 26772:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000404:0x291:0x0]: rc = -2 LustreError: 31826:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0001: '12' migration was interrupted, run 'lfs migrate -m 0 -c 1 -H crush 12' to finish migration. Lustre: 26750:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 502 < left 521, rollback = 2 Lustre: 26750:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 113 previous similar messages Lustre: ll_ost00_012: service thread pid 1363 was inactive for 62.082 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 3 previous similar messages LustreError: 26735:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 106s: evicting client at 0@lo ns: filter-lustre-OST0001_UUID lock: ffff8802784ba980/0x368f4f940aab05a0 lrc: 3/0,0 mode: PW/PW res: [0x5af:0x0:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->917503) gid 0 flags: 0x60000400020020 nid: 0@lo remote: 0x368f4f940aab0005 expref: 4260 pid: 32079 timeout: 4482 lvb_type: 0 LustreError: 14671:0:(ldlm_lockd.c:1425:ldlm_handle_enqueue0()) ### lock on destroyed export ffff8802ebf84138 ns: filter-lustre-OST0001_UUID lock: ffff880073b20f40/0x368f4f940abd788c lrc: 3/0,0 mode: --/PW res: [0x5fc:0x0:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x50000000020000 nid: 0@lo remote: 0x368f4f940abd787e expref: 4254 pid: 14671 timeout: 0 lvb_type: 0 LustreError: 14671:0:(ldlm_lockd.c:1425:ldlm_handle_enqueue0()) Skipped 1 previous similar message LustreError: 26721:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630539095 with bad export cookie 3931448496969056500 LustreError: 26721:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) Skipped 4 previous similar messages Lustre: lustre-OST0001-osc-ffff8802eba492a8: Connection to lustre-OST0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 28523:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88025710bec0 x1709743690722688/t0(0) o104->lustre-OST0001@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 28523:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 45 previous similar messages LustreError: 11-0: lustre-OST0001-osc-ffff8802eba492a8: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: 167-0: lustre-OST0001-osc-ffff8802eba492a8: This client was evicted by lustre-OST0001; in progress operations using this service will fail. Lustre: 23566:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.140@tcp:/lustre/fid: [0x200000403:0x227:0x0]/ may get corrupted (rc -108) Lustre: lustre-OST0001-osc-ffff8802eba492a8: Connection restored to 192.168.123.140@tcp (at 0@lo) LustreError: 32596:0:(ldlm_lockd.c:1425:ldlm_handle_enqueue0()) ### lock on destroyed export ffff8802eba4e678 ns: mdt-lustre-MDT0002_UUID lock: ffff8802755c0b80/0x368f4f940abf2602 lrc: 3/0,0 mode: PR/PR res: [0x280000403:0x2ad:0x0].0x0 bits 0x12/0x0 rrc: 2 type: IBT gid 0 flags: 0x50200000000000 nid: 0@lo remote: 0x368f4f940abf25ca expref: 4 pid: 32596 timeout: 0 lvb_type: 0 Lustre: dir [0x280000403:0x2b5:0x0] stripe 0 readdir failed: -108, directory is partially accessed! LustreError: 17366:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff8802eba492a8: inode [0x280000404:0x3c:0x0] mdc close failed: rc = -108 LustreError: 17366:0:(file.c:234:ll_close_inode_openhandle()) Skipped 4 previous similar messages LustreError: 17366:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-MDT0002-mdc-ffff8802eba492a8: namespace resource [0x280000403:0x1:0x0].0x0 (ffff88025b7c1bc0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 17366:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 713 previous similar messages LustreError: 26735:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 102s: evicting client at 0@lo ns: filter-lustre-OST0000_UUID lock: ffff88007c54d680/0x368f4f940ab6386d lrc: 3/0,0 mode: PW/PW res: [0x2c0000401:0x888:0x0].0x0 rrc: 4 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000400000020 nid: 0@lo remote: 0x368f4f940ab63866 expref: 2650 pid: 31723 timeout: 4513 lvb_type: 0 LustreError: 26735:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 1 previous similar message LustreError: 26712:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630539127 with bad export cookie 3931448496970914384 LustreError: 32071:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8800757d8040 x1709743693122176/t0(0) o104->lustre-OST0000@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 32071:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 16 previous similar messages LustreError: 11-0: lustre-OST0000-osc-ffff8800a64a2e98: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 1 previous similar message Lustre: lustre-OST0000-osc-ffff8800a64a2e98: Connection to lustre-OST0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 1 previous similar message LustreError: 167-0: lustre-OST0000-osc-ffff8800a64a2e98: This client was evicted by lustre-OST0000; in progress operations using this service will fail. LustreError: Skipped 1 previous similar message LustreError: 18011:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0000-osc-ffff8800a64a2e98: namespace resource [0x2c0000401:0x9ff:0x0].0x0 (ffff880255716340) refcount nonzero (1) after lock cleanup; forcing cleanup. Lustre: lustre-OST0000-osc-ffff8800a64a2e98: Connection restored to 192.168.123.140@tcp (at 0@lo) Lustre: Skipped 1 previous similar message ------------[ cut here ]------------ WARNING: CPU: 6 PID: 29652 at /home/green/git/lustre-release/ldiskfs/ext4_jbd2.c:266 __ldiskfs_handle_dirty_metadata+0x1d2/0x230 [ldiskfs] Modules linked in: lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) dm_flakey dm_mod loop zfs(PO) zunicode(PO) zzstd(O) zlua(O) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) jbd2 mbcache crc_t10dif crct10dif_generic sb_edac edac_core iosf_mbi crc32_pclmul ghash_clmulni_intel aesni_intel lrw gf128mul glue_helper ablk_helper cryptd pcspkr virtio_console virtio_balloon i2c_piix4 ip_tables rpcsec_gss_krb5 ata_generic drm_kms_helper pata_acpi ttm drm crct10dif_pclmul crct10dif_common ata_piix drm_panel_orientation_quirks crc32c_intel serio_raw virtio_blk libata i2c_core floppy [last unloaded: libcfs] CPU: 6 PID: 29652 Comm: mdt03_006 Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011 Call Trace: [<ffffffff817d9418>] dump_stack+0x19/0x1b [<ffffffff8108d538>] __warn+0xd8/0x100 [<ffffffff8108d67d>] warn_slowpath_null+0x1d/0x20 [<ffffffffa0d88262>] __ldiskfs_handle_dirty_metadata+0x1d2/0x230 [ldiskfs] [<ffffffffa0d6833d>] ldiskfs_getblk+0x13d/0x210 [ldiskfs] [<ffffffffa0d68437>] ldiskfs_bread+0x27/0xe0 [ldiskfs] [<ffffffffa0cb3e97>] osd_write+0x4e7/0xcb0 [osd_ldiskfs] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffffa04dc6f3>] dt_record_write+0x33/0x120 [obdclass] [<ffffffffa049b6b2>] llog_osd_write_rec+0x1052/0x19a0 [obdclass] [<ffffffffa0488760>] llog_write_rec+0x290/0x590 [obdclass] [<ffffffffa048e4ee>] llog_cat_add_rec+0x23e/0x950 [obdclass] [<ffffffffa04855ff>] llog_add+0x17f/0x1f0 [obdclass] [<ffffffffa04b6a39>] ? lprocfs_counter_add+0xf9/0x160 [obdclass] [<ffffffffa081f60d>] sub_updates_write+0xb6b/0xe32 [ptlrpc] [<ffffffffa07ef892>] top_trans_stop+0x7c2/0xf60 [ptlrpc] [<ffffffffa0f96f3c>] lod_trans_stop+0x25c/0x340 [lod] [<ffffffffa0e6b408>] mdd_trans_stop+0x28/0x16e [mdd] [<ffffffffa0e4cb2f>] mdd_migrate_object+0xf1f/0x1ba0 [mdd] [<ffffffffa04db375>] ? lu_object_find_at+0x285/0xb10 [obdclass] [<ffffffffa0e4db86>] mdd_migrate+0x3d6/0x8e0 [mdd] [<ffffffffa0f1f0d8>] mdo_migrate+0x4c/0x4e [mdt] [<ffffffffa0edee53>] mdt_reint_migrate+0x1023/0x11d0 [mdt] [<ffffffffa0edf087>] mdt_reint_rec+0x87/0x240 [mdt] [<ffffffffa0eb455c>] mdt_reint_internal+0x76c/0xb40 [mdt] [<ffffffffa0ec1657>] mdt_reint+0x67/0x150 [mdt] [<ffffffffa07de26e>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<ffffffffa03c245e>] ? libcfs_nid2str_r+0xfe/0x130 [lnet] [<ffffffffa0789990>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<ffffffffa078b529>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e32be>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffffa078a950>] ? ptlrpc_wait_event+0x620/0x620 [ptlrpc] [<ffffffff810ba0f4>] kthread+0xe4/0xf0 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 ---[ end trace 9deceedc8d9bbe25 ]--- LDISKFS-fs: ldiskfs_getblk:888: aborting transaction: error 28 in __ldiskfs_handle_dirty_metadata LDISKFS-fs error (device dm-1): ldiskfs_getblk:888: inode #187: block 99383: comm mdt03_006: journal_dirty_metadata failed: handle type 0 started at line 1972, credits 512/0, errcode -28 Aborting journal on device dm-1-8. LDISKFS-fs (dm-1): Remounting filesystem read-only LustreError: 29652:0:(osd_io.c:2103:osd_ldiskfs_write_record()) lustre-MDT0001: error reading offset 106496 (block 26, size 40, offs 106480), credits 0/0: rc = -28 LustreError: 29652:0:(update_trans.c:985:top_trans_stop()) lustre-MDT0001-osd: write updates failed: rc = -28 LustreError: 29652:0:(osd_io.c:2110:osd_ldiskfs_write_record()) journal_get_write_access() returned error -30 LustreError: 29652:0:(tgt_lastrcvd.c:1253:tgt_add_reply_data()) lustre-MDT0001: can't update reply_data file: rc = -30 LustreError: 29652:0:(osd_handler.c:2075:osd_trans_stop()) lustre-MDT0001: failed in transaction hook: rc = -30 LustreError: 28147:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-OST0003-osc-MDT0001: fail to cancel 16 llog-records: rc = -30 LustreError: 28142:0:(osp_sync.c:1094:osp_sync_process_committed()) lustre-OST0001-osc-MDT0001: can't cancel 43 records: rc = -30 LustreError: 17999:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -30 LDISKFS-fs error (device dm-1) in osd_trans_stop:2082: error 28 LustreError: 27019:0:(osd_handler.c:1779:osd_trans_commit_cb()) transaction @0xffff8802fa4469a8 commit error: 2 LustreError: 29652:0:(osd_handler.c:2085:osd_trans_stop()) lustre-MDT0001: failed to stop transaction: rc = -28 LustreError: 29652:0:(update_trans.c:1011:top_trans_stop()) lustre-MDT0001-osd: stop trans failed: rc = -30 LDISKFS-fs error (device dm-1) in osd_trans_stop:2082: IO failure LustreError: 27238:0:(tgt_handler.c:1367:tgt_blocking_ast()) lustre-MDT0001: syncing [0x240000402:0x4:0x0] (19-0) on lock cancel: rc = -5 LustreError: 28147:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) Skipped 6 previous similar messages LustreError: 28147:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-OST0003-osc-MDT0001: fail to cancel 37 llog-records: rc = -30 LustreError: 28147:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) Skipped 10 previous similar messages LustreError: 28147:0:(osp_sync.c:1094:osp_sync_process_committed()) lustre-OST0003-osc-MDT0001: can't cancel 37 records: rc = -30 LustreError: 28147:0:(osp_sync.c:1094:osp_sync_process_committed()) Skipped 16 previous similar messages LustreError: 28147:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-OST0003-osc-MDT0001: fail to cancel 1 of 1 llog-records: rc = -30 LustreError: 28147:0:(osp_sync.c:1079:osp_sync_process_committed()) lustre-OST0003-osc-MDT0001: can't cancel record: rc = -30 LustreError: 28966:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -30 LustreError: 28966:0:(llite_lib.c:1836:ll_md_setattr()) Skipped 3 previous similar messages LustreError: 28142:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-OST0001-osc-MDT0001: fail to cancel 84 llog-records: rc = -30 LustreError: 28142:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) Skipped 9 previous similar messages LustreError: 28142:0:(osp_sync.c:1094:osp_sync_process_committed()) lustre-OST0001-osc-MDT0001: can't cancel 84 records: rc = -30 LustreError: 28142:0:(osp_sync.c:1094:osp_sync_process_committed()) Skipped 6 previous similar messages LustreError: 28142:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-OST0001-osc-MDT0001: fail to cancel 1 of 1 llog-records: rc = -30 LustreError: 28142:0:(llog_cat.c:787:llog_cat_cancel_records()) Skipped 2 previous similar messages LustreError: 28142:0:(osp_sync.c:1079:osp_sync_process_committed()) lustre-OST0001-osc-MDT0001: can't cancel record: rc = -30 LustreError: 28142:0:(osp_sync.c:1079:osp_sync_process_committed()) Skipped 2 previous similar messages LustreError: 17951:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -30 LustreError: 17951:0:(llite_lib.c:1836:ll_md_setattr()) Skipped 3 previous similar messages LustreError: 28142:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-OST0001-osc-MDT0001: fail to cancel 1 of 1 llog-records: rc = -30 LustreError: 28142:0:(llog_cat.c:787:llog_cat_cancel_records()) Skipped 10 previous similar messages LustreError: 28142:0:(osp_sync.c:1079:osp_sync_process_committed()) lustre-OST0001-osc-MDT0001: can't cancel record: rc = -30 LustreError: 28142:0:(osp_sync.c:1079:osp_sync_process_committed()) Skipped 8 previous similar messages | Externally reported by onyx-68 boilpot email |
racer test 1: racer on clients: centos-115.localnet DURATION=2700 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa1032e52>] ll_migrate+0x9b2/0xec0 [lustre] PGD 800000008bb5f067 PUD 7b127067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: loop zfs(PO) zunicode(PO) zzstd(O) zlua(O) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) jbd2 mbcache lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) dm_flakey dm_mod libcfs(OE) crc_t10dif crct10dif_generic sb_edac edac_core iosf_mbi crc32_pclmul ghash_clmulni_intel aesni_intel lrw gf128mul glue_helper ablk_helper cryptd virtio_console virtio_balloon pcspkr i2c_piix4 ip_tables rpcsec_gss_krb5 drm_kms_helper ata_generic ttm crct10dif_pclmul pata_acpi crct10dif_common drm crc32c_intel ata_piix drm_panel_orientation_quirks serio_raw libata virtio_blk i2c_core floppy CPU: 2 PID: 4801 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011 task: ffff88025f59b760 ti: ffff8800706a0000 task.ti: ffff8800706a0000 RIP: 0010:[<ffffffffa1032e52>] [<ffffffffa1032e52>] ll_migrate+0x9b2/0xec0 [lustre] RSP: 0018:ffff8800706a3bc8 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff8802573c1c28 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffffea0001cbe900 RDI: 0000000000000246 RBP: ffff8800706a3c48 R08: ffff880072fa72a8 R09: 0000000000000000 R10: ffff880072fa43c8 R11: ffff880072fa72a8 R12: 0000000000000000 R13: ffff8802668489f8 R14: ffff880073d09398 R15: 0000000000000030 FS: 00007fab13b3c740(0000) GS:ffff880331a80000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 CR2: 00007f503307ec60 CR3: 000000006e4c0000 CR4: 00000000001607e0 Call Trace: [<ffffffff81242213>] ? __check_object_size+0x1c3/0x220 [<ffffffffa1020561>] ll_dir_ioctl+0x5d01/0x6ed0 [lustre] [<ffffffff8103de0f>] ? save_stack_trace+0x1f/0x30 [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817d5822>] ? free_debug_processing+0x240/0x272 [<ffffffffa104e6ad>] ? ll_finish_md_op_data+0xad/0x1f0 [lustre] [<ffffffff8121e474>] ? __slab_free+0xa4/0x300 [<ffffffff811becd5>] ? __free_pages+0x25/0x30 [<ffffffffa1016f3e>] ? ll_release_page+0x8e/0xa0 [lustre] [<ffffffffa04b1a39>] ? lprocfs_counter_add+0xf9/0x160 [obdclass] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e326e>] ? _raw_spin_unlock+0xe/0x20 [<ffffffff81277ec8>] ? __mark_inode_dirty+0x1a8/0x2c0 [<ffffffff8125b3fd>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff817df62e>] ? mutex_unlock+0xe/0x10 [<ffffffff8125b81a>] ? iterate_dir+0xaa/0x120 [<ffffffff8125b751>] SyS_ioctl+0xa1/0xc0 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 | Lustre: 11135:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 512 < left 958, rollback = 7 Lustre: 11135:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 11135:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 1/89/0 Lustre: 11135:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 949/958/0, punch: 0/0/0, quota 3/3/0 Lustre: 11135:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 11135:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 11135:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 511 < left 957, rollback = 7 Lustre: 11135:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 11135:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 1/89/0 Lustre: 11135:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 949/958/1, punch: 0/0/0, quota 3/3/0 Lustre: 11135:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 11135:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 dd (16913) used greatest stack depth: 10240 bytes left Lustre: 11129:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 511 < left 1451, rollback = 7 Lustre: 11129:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 1911 previous similar messages Lustre: 11129:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 11129:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 1911 previous similar messages Lustre: 11129:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 1/89/0 Lustre: 11129:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 1911 previous similar messages Lustre: 11129:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1443/1452/1, punch: 0/0/0, quota 3/3/0 Lustre: 11129:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 1911 previous similar messages Lustre: 11129:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 11129:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 1911 previous similar messages Lustre: 11129:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 11129:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 1911 previous similar messages Lustre: 17912:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 496 < left 877, rollback = 2 Lustre: 17912:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 3/12/12, destroy: 0/0/0 Lustre: 17912:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 4279 previous similar messages Lustre: 17912:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 2/2/0, xattr_set: 11/877/0 Lustre: 17912:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 4279 previous similar messages Lustre: 17912:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 5/136/0, punch: 0/0/0, quota 8/56/0 Lustre: 17912:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 4279 previous similar messages Lustre: 17912:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 12/215/4, delete: 1/1/0 Lustre: 17912:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 4279 previous similar messages Lustre: 17912:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 6/6/0, ref_del: 1/1/0 Lustre: 17912:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 4279 previous similar messages Lustre: 17912:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 494 < left 877, rollback = 2 Lustre: 17912:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 4 previous similar messages Lustre: 14291:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 507 < left 582, rollback = 2 Lustre: 14291:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 12 previous similar messages 0[18238]: segfault at 8 ip 00007f282634b7e8 sp 00007ffd41c36880 error 4 in ld-2.17.so[7f2826340000+22000] Lustre: 16900:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 506 < left 8051, rollback = 7 Lustre: 16900:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 4279 previous similar messages Lustre: 16900:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 187/748/0, destroy: 0/0/0 Lustre: 16900:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 23 previous similar messages Lustre: 16900:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 190/204/6 Lustre: 16900:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 23 previous similar messages Lustre: 16900:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 936/8051/0, punch: 0/0/0, quota 6/6/0 Lustre: 16900:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 23 previous similar messages Lustre: 16900:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 187/3179/0, delete: 0/0/0 Lustre: 16900:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 23 previous similar messages Lustre: 16900:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 16900:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 23 previous similar messages Lustre: 16934:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 507 < left 788, rollback = 2 Lustre: 16934:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 5 previous similar messages Lustre: 16895:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 511 < left 13598, rollback = 7 Lustre: 16895:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 4549 previous similar messages Lustre: 16895:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 316/1264/0, destroy: 1/4/0 Lustre: 16895:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 4559 previous similar messages Lustre: 16895:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 3/3/0, xattr_set: 320/148/0 Lustre: 16895:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 4559 previous similar messages Lustre: 16895:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1581/13598/0, punch: 0/0/0, quota 6/6/0 Lustre: 16895:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 4559 previous similar messages Lustre: 16895:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 317/5388/0, delete: 2/5/0 Lustre: 16895:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 4559 previous similar messages Lustre: 16895:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 1/1/0, ref_del: 2/2/1 Lustre: 16895:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 4559 previous similar messages ------------[ cut here ]------------ WARNING: CPU: 2 PID: 16895 at /home/green/git/lustre-release/ldiskfs/namei.c:3331 ldiskfs_orphan_add+0x11e/0x2a0 [ldiskfs] Modules linked in: loop zfs(PO) zunicode(PO) zzstd(O) zlua(O) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) jbd2 mbcache lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) dm_flakey dm_mod libcfs(OE) crc_t10dif crct10dif_generic sb_edac edac_core iosf_mbi crc32_pclmul ghash_clmulni_intel aesni_intel lrw gf128mul glue_helper ablk_helper cryptd virtio_console virtio_balloon pcspkr i2c_piix4 ip_tables rpcsec_gss_krb5 drm_kms_helper ata_generic ttm crct10dif_pclmul pata_acpi crct10dif_common drm crc32c_intel ata_piix drm_panel_orientation_quirks serio_raw libata virtio_blk i2c_core floppy CPU: 2 PID: 16895 Comm: mdt01_004 Tainted: P OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011 Call Trace: [<ffffffff817d9418>] dump_stack+0x19/0x1b [<ffffffff8108d538>] __warn+0xd8/0x100 [<ffffffff8108d67d>] warn_slowpath_null+0x1d/0x20 [<ffffffffa0bca4ae>] ldiskfs_orphan_add+0x11e/0x2a0 [ldiskfs] [<ffffffffa0bbb9eb>] ldiskfs_xattr_inode_orphan_add.constprop.17+0xbb/0x110 [ldiskfs] [<ffffffffa0bbe417>] ldiskfs_xattr_delete_inode+0x77/0x3c0 [ldiskfs] [<ffffffffa0bdeaa0>] ? ldiskfs_evict_inode+0x190/0x670 [ldiskfs] [<ffffffffa0bdead9>] ldiskfs_evict_inode+0x1c9/0x670 [ldiskfs] [<ffffffff812649af>] evict+0xaf/0x180 [<ffffffff81264de5>] iput+0xf5/0x180 [<ffffffffa0c73d08>] osd_object_delete+0x1f8/0x370 [osd_ldiskfs] [<ffffffffa04d159d>] lu_object_free.isra.26+0xcd/0x1d0 [obdclass] [<ffffffffa04d5cf1>] lu_object_put+0xa1/0x480 [obdclass] [<ffffffffa0e12440>] mdt_object_put+0x30/0x110 [mdt] [<ffffffffa0e19e18>] mdt_reint_unlink+0x918/0x19b0 [mdt] [<ffffffffa04f779e>] ? lu_ucred+0x1e/0x30 [obdclass] [<ffffffffa0e09af2>] ? ucred_set_audit_enabled.isra.13+0x22/0x60 [mdt] [<ffffffffa0e21087>] mdt_reint_rec+0x87/0x240 [mdt] [<ffffffffa0df655c>] mdt_reint_internal+0x76c/0xb40 [mdt] [<ffffffffa0e03657>] mdt_reint+0x67/0x150 [mdt] [<ffffffffa07d926e>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<ffffffffa03bd45e>] ? libcfs_nid2str_r+0xfe/0x130 [lnet] [<ffffffffa0784990>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<ffffffffa0786529>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffffa0785950>] ? ptlrpc_wait_event+0x620/0x620 [ptlrpc] [<ffffffff810ba0f4>] kthread+0xe4/0xf0 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 ---[ end trace af9871a41479a8b6 ]--- Lustre: 11131:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 502 < left 521, rollback = 2 Lustre: 11131:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 9 previous similar messages sched: RT throttling activated Lustre: 11115:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 512 < left 1015, rollback = 7 Lustre: 11115:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 5067 previous similar messages Lustre: 11115:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 11115:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 5071 previous similar messages Lustre: 11115:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 1/89/0 Lustre: 11115:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 5071 previous similar messages Lustre: 11115:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1006/1015/0, punch: 0/0/0, quota 3/3/0 Lustre: 11115:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 5071 previous similar messages Lustre: 11115:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 11115:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 5071 previous similar messages Lustre: 11115:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 11115:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 5071 previous similar messages Lustre: 14291:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 501 < left 521, rollback = 2 Lustre: 14291:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 3 previous similar messages LustreError: 16700:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0001: '15' migration was interrupted, run 'lfs migrate -m 0 -c 1 -H crush 15' to finish migration. Lustre: 17855:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 503 < left 62016, rollback = 7 Lustre: 17855:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 4809 previous similar messages Lustre: 20359:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 497 < left 788, rollback = 2 Lustre: 20359:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 14 previous similar messages LustreError: 16700:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0001: '15' migration was interrupted, run 'lfs migrate -m 0 -c 1 -H crush 15' to finish migration. Lustre: ll_ost02_000: service thread pid 12701 was inactive for 62.126 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 18388, comm: ll_ost02_009 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] tgt_extent_lock+0xea/0x2a0 [ptlrpc] [<0>] ofd_getattr_hdl+0x365/0x720 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 18165, comm: ll_ost02_005 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] tgt_extent_lock+0xea/0x2a0 [ptlrpc] [<0>] ofd_getattr_hdl+0x365/0x720 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: ll_ost02_010: service thread pid 18855 was inactive for 62.131 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 2 previous similar messages Pid: 12701, comm: ll_ost02_000 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] tgt_extent_lock+0xea/0x2a0 [ptlrpc] [<0>] ofd_getattr_hdl+0x365/0x720 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: ll_ost04_000: service thread pid 12707 was inactive for 62.148 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 21 previous similar messages Lustre: ll_ost02_063: service thread pid 20112 was inactive for 62.097 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 21 previous similar messages LustreError: 10989:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0001_UUID lock: ffff88027e6fb880/0x5011d9db3a736935 lrc: 3/0,0 mode: PW/PW res: [0x300000400:0x5:0x0].0x0 rrc: 5 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000400000020 nid: 0@lo remote: 0x5011d9db3a736426 expref: 3964 pid: 12707 timeout: 334 lvb_type: 0 LustreError: 18860:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff880257274500 x1709741049405568/t0(0) o104->lustre-OST0001@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' Lustre: lustre-OST0001-osc-ffff8802996aa548: Connection to lustre-OST0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0001-osc-ffff8802996aa548: This client was evicted by lustre-OST0001; in progress operations using this service will fail. LustreError: 21563:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0001-osc-ffff8802996aa548: namespace resource [0x300000400:0x10:0x0].0x0 (ffff8800889e1440) refcount nonzero (1) after lock cleanup; forcing cleanup. Lustre: mdt02_005: service thread pid 16915 was inactive for 62.196 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 15 previous similar messages Lustre: mdt06_005: service thread pid 16749 was inactive for 62.151 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 18 previous similar messages LustreError: 10989:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0003_UUID lock: ffff88028867f840/0x5011d9db3a738279 lrc: 3/0,0 mode: PW/PW res: [0x380000400:0x9:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000400000020 nid: 0@lo remote: 0x5011d9db3a738272 expref: 3788 pid: 12710 timeout: 370 lvb_type: 0 LustreError: 20021:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88008defac00 x1709741049434048/t0(0) o104->lustre-OST0000@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 20021:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 3 previous similar messages Lustre: lustre-OST0002-osc-ffff8802996aa548: Connection to lustre-OST0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0002-osc-ffff8802996aa548: This client was evicted by lustre-OST0002; in progress operations using this service will fail. LustreError: 19057:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630536397 with bad export cookie 5769632133239940938 LustreError: 19057:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) Skipped 1 previous similar message LustreError: 10989:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0003_UUID lock: ffff88007c519a80/0x5011d9db3a78f02a lrc: 3/0,0 mode: PW/PW res: [0x380000400:0x24f:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000400000020 nid: 0@lo remote: 0x5011d9db3a78f01c expref: 2225 pid: 12709 timeout: 377 lvb_type: 0 LustreError: 10989:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 16 previous similar messages LustreError: 10982:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630536403 with bad export cookie 5769632133239940700 LustreError: 21575:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff880283d46440 x1709741050053248/t0(0) o105->lustre-OST0003@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 21575:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 95 previous similar messages Lustre: lustre-OST0003-osc-ffff88029b341bf8: Connection to lustre-OST0003 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 2 previous similar messages LustreError: 167-0: lustre-OST0003-osc-ffff88029b341bf8: This client was evicted by lustre-OST0003; in progress operations using this service will fail. LustreError: Skipped 2 previous similar messages Lustre: 4134:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x240000403:0x94:0x0]/ may get corrupted (rc -108) LustreError: 20021:0:(service.c:2290:ptlrpc_server_handle_request()) @@@ Dropping timed-out request from 12345-0@lo: deadline 71/28s ago req@ffff880272be9940 x1709741047832576/t0(0) o2->f1de7ff5-7e63-4e7b-ac0f-06d879070ea9@0@lo:625/0 lens 440/0 e 1 to 0 dl 1630536375 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'chown.0' Lustre: 20038:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (71/28s); client may timeout req@ffff880272be9300 x1709741047832832/t0(0) o2->f1de7ff5-7e63-4e7b-ac0f-06d879070ea9@0@lo:625/0 lens 440/0 e 1 to 0 dl 1630536375 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'chown.0' LustreError: 20021:0:(service.c:2290:ptlrpc_server_handle_request()) Skipped 13 previous similar messages LustreError: 20089:0:(service.c:2290:ptlrpc_server_handle_request()) @@@ Dropping timed-out request from 12345-0@lo: deadline 71/24s ago req@ffff8802557d5180 x1709741048566080/t0(0) o101->f1de7ff5-7e63-4e7b-ac0f-06d879070ea9@0@lo:630/0 lens 328/0 e 1 to 0 dl 1630536380 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'mv.0' LustreError: 20089:0:(service.c:2290:ptlrpc_server_handle_request()) Skipped 490 previous similar messages Lustre: 20089:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (71/24s); client may timeout req@ffff8802557d5180 x1709741048566080/t0(0) o101->f1de7ff5-7e63-4e7b-ac0f-06d879070ea9@0@lo:630/0 lens 328/0 e 1 to 0 dl 1630536380 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'mv.0' Lustre: 20089:0:(service.c:2328:ptlrpc_server_handle_request()) Skipped 503 previous similar messages Lustre: 20671:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 505 < left 529, rollback = 7 Lustre: 20671:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 5784 previous similar messages Lustre: 20671:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 1/4/4, destroy: 0/0/0 Lustre: 20671:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 10628 previous similar messages Lustre: 20671:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 4/209/0 Lustre: 20671:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 10628 previous similar messages Lustre: 20671:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 520/529/0, punch: 0/0/0, quota 6/6/2 Lustre: 20671:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 10628 previous similar messages Lustre: 20671:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 2/33/1, delete: 0/0/0 Lustre: 20671:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 10628 previous similar messages Lustre: 20671:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 20671:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 10628 previous similar messages LustreError: 21563:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0001-osc-ffff8802996aa548: namespace resource [0x300000400:0x100:0x0].0x0 (ffff88025741c540) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 21563:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 807 previous similar messages Lustre: lustre-OST0001-osc-ffff8802996aa548: Connection restored to 192.168.123.215@tcp (at 0@lo) LustreError: 2149:0:(ofd_obd.c:501:ofd_destroy_export()) lustre-OST0001: cli 784550cb-a59c-40bc-a083-9d838a506239/ffff8802994fd3d8 has 28672 pending on destroyed export LustreError: 2149:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_destroy_export: tot_granted 8655168 != fo_tot_granted 8683840 LustreError: 2149:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_destroy_export: tot_pending 0 != fo_tot_pending 28672 LustreError: 12729:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 8655168 != fo_tot_granted 8683840 LustreError: 12729:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 28672 LustreError: 21558:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 8655168 != fo_tot_granted 8683840 LustreError: 21558:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 28672 Lustre: 17915:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 506 < left 610, rollback = 2 Lustre: 17915:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 18 previous similar messages LustreError: 21573:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 8655168 != fo_tot_granted 8683840 LustreError: 21573:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 1 previous similar message LustreError: 21573:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 28672 LustreError: 21573:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 1 previous similar message LustreError: 21578:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0000-osc-ffff8802996aa548: namespace resource [0x2c0000400:0x20:0x0].0x0 (ffff880274ff5e40) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 21578:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 355 previous similar messages Lustre: lustre-OST0000-osc-ffff8802996aa548: Connection restored to 192.168.123.215@tcp (at 0@lo) LustreError: 21576:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0002-osc-ffff8802996aa548: namespace resource [0x340000400:0x80:0x0].0x0 (ffff88026bd59940) refcount nonzero (2) after lock cleanup; forcing cleanup. LustreError: 21576:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 793 previous similar messages Lustre: lustre-OST0002-osc-ffff8802996aa548: Connection restored to 192.168.123.215@tcp (at 0@lo) Lustre: lustre-OST0003-osc-ffff8802996aa548: Connection restored to 192.168.123.215@tcp (at 0@lo) Lustre: Skipped 3 previous similar messages LustreError: 21573:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 8655168 != fo_tot_granted 8683840 LustreError: 21573:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 2 previous similar messages LustreError: 21573:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 28672 LustreError: 21573:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 2 previous similar messages LustreError: 12729:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 43888704 != fo_tot_granted 43917376 LustreError: 12729:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 4 previous similar messages LustreError: 12729:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 28672 LustreError: 12729:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 4 previous similar messages LustreError: 12738:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 56541760 != fo_tot_granted 56570432 LustreError: 12738:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 7 previous similar messages LustreError: 12738:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 1138688 != fo_tot_pending 1167360 LustreError: 12738:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 7 previous similar messages Lustre: 4133:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff8802715c9300 x1709741053320704/t0(0) o10->lustre-OST0002-osc-ffff88029b341bf8@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630536510 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'cp.0' Lustre: 16749:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 511 < left 22327, rollback = 7 Lustre: 16749:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 26025 previous similar messages Lustre: 16749:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 519/2076/0, destroy: 1/4/0 Lustre: 16749:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 26058 previous similar messages Lustre: 16749:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 3/3/0, xattr_set: 523/148/0 Lustre: 16749:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 26058 previous similar messages Lustre: 16749:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 2596/22327/0, punch: 0/0/0, quota 8/56/0 Lustre: 16749:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 26058 previous similar messages Lustre: 16749:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 520/8839/0, delete: 2/5/0 Lustre: 16749:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 26058 previous similar messages Lustre: 16749:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 1/1/0, ref_del: 2/2/1 Lustre: 16749:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 26058 previous similar messages Lustre: 17677:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000404:0x39:0x0] with magic=0xbd60bd0 LustreError: 16972:0:(mdt_xattr.c:429:mdt_dir_layout_update()) lustre-MDT0002: [0x280000404:0xb4:0x0] migrate mdt index mismatch 2 != 1 LustreError: 12737:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 149670720 != fo_tot_granted 149699392 LustreError: 12737:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 15 previous similar messages LustreError: 12737:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 28672 LustreError: 12737:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 15 previous similar messages Lustre: 22006:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 8: before 509 < left 5507, rollback = 9 Lustre: 22006:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 33 previous similar messages LustreError: 11165:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0002: failed to get lu_attr of [0x280000404:0x91:0x0]: rc = -2 LustreError: 22055:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0001: '9' migration was interrupted, run 'lfs migrate -m 0 -c 1 -H crush 9' to finish migration. Lustre: 22065:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000403:0xf8:0x0] with magic=0xbd60bd0 Lustre: 22065:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 3 previous similar messages LustreError: 12338:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0000-osp-MDT0002: fail to cancel 1 llog-records: rc = -2 LustreError: 12338:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0000-osp-MDT0002: fail to cancel 1 of 1 llog-records: rc = -2 LustreError: 11161:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0002: failed to get lu_attr of [0x280000404:0xc7:0x0]: rc = -2 Lustre: dir [0x200000403:0xd6:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: dir [0x280000404:0xf3:0x0] stripe 1 readdir failed: -2, directory is partially accessed! LustreError: 11131:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0000: '1' migration was interrupted, run 'lfs migrate -m 1 -c 1 -H crush 1' to finish migration. LustreError: 12731:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 246014784 != fo_tot_granted 246043456 LustreError: 12731:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 29 previous similar messages LustreError: 12731:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 28672 LustreError: 12731:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 29 previous similar messages LustreError: 10989:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0001_UUID lock: ffff88007cf3de00/0x5011d9db3a842631 lrc: 3/0,0 mode: PW/PW res: [0x16:0x0:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000000020020 nid: 0@lo remote: 0x5011d9db3a84260e expref: 2188 pid: 18729 timeout: 550 lvb_type: 0 LustreError: 10989:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 2 previous similar messages LustreError: 11-0: lustre-OST0001-osc-ffff88029b341bf8: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-OST0001-osc-ffff88029b341bf8: Connection to lustre-OST0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 2 previous similar messages LustreError: 167-0: lustre-OST0001-osc-ffff88029b341bf8: This client was evicted by lustre-OST0001; in progress operations using this service will fail. LustreError: Skipped 2 previous similar messages LustreError: 27808:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0001-osc-ffff88029b341bf8: namespace resource [0x1e0:0x0:0x0].0x0 (ffff8800895302c0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 27808:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 1520 previous similar messages LustreError: 10989:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 101s: evicting client at 0@lo ns: filter-lustre-OST0002_UUID lock: ffff880080b652c0/0x5011d9db3a848392 lrc: 3/0,0 mode: PW/PW res: [0x340000400:0xfe:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000000020020 nid: 0@lo remote: 0x5011d9db3a848345 expref: 2580 pid: 18721 timeout: 557 lvb_type: 0 LustreError: 10987:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8800729bf0c0 x1709741060177408/t0(0) o105->lustre-OST0002@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 10987:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 18 previous similar messages LustreError: 11-0: lustre-OST0002-osc-ffff8802996aa548: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-OST0002-osc-ffff8802996aa548: Connection to lustre-OST0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0002-osc-ffff8802996aa548: This client was evicted by lustre-OST0002; in progress operations using this service will fail. Lustre: lustre-OST0001-osc-ffff88029b341bf8: Connection restored to 192.168.123.215@tcp (at 0@lo) LustreError: 27956:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0003-osc-ffff88029b341bf8: namespace resource [0x2e0:0x0:0x0].0x0 (ffff8802818147c0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 27956:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 59 previous similar messages Lustre: 4135:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x280000404:0x111:0x0]// may get corrupted (rc -108) Lustre: 4135:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x240000404:0x1bf:0x0]// may get corrupted (rc -108) Lustre: 4134:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x280000404:0x125:0x0]// may get corrupted (rc -108) Lustre: 4135:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x200000404:0x124:0x0]/ may get corrupted (rc -108) LustreError: 10989:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0000_UUID lock: ffff8800897e34c0/0x5011d9db3a859977 lrc: 3/0,0 mode: PW/PW res: [0x2c0000400:0x564:0x0].0x0 rrc: 4 type: EXT [0->18446744073709551615] (req 0->36863) gid 0 flags: 0x60000000020020 nid: 0@lo remote: 0x5011d9db3a859970 expref: 2512 pid: 18945 timeout: 571 lvb_type: 0 LustreError: 10989:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 1 previous similar message LustreError: 18711:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88028212b880 x1709741060998336/t0(0) o104->lustre-OST0000@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 18711:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 2 previous similar messages LustreError: 10971:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630536597 with bad export cookie 5769632133240632615 LustreError: 10971:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) Skipped 2 previous similar messages LustreError: 11-0: lustre-OST0000-osc-ffff88029b341bf8: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 3 previous similar messages Lustre: lustre-OST0000-osc-ffff88029b341bf8: Connection to lustre-OST0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 1 previous similar message LustreError: 167-0: lustre-OST0000-osc-ffff88029b341bf8: This client was evicted by lustre-OST0000; in progress operations using this service will fail. LustreError: Skipped 1 previous similar message Lustre: 19255:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 512 < left 1955, rollback = 7 Lustre: 19255:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 29702 previous similar messages Lustre: 19255:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 1/4/0, destroy: 0/0/0 Lustre: 19255:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 29822 previous similar messages Lustre: 19255:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 4/209/0 Lustre: 19255:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 29822 previous similar messages Lustre: 19255:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1946/1955/0, punch: 0/0/0, quota 6/6/0 Lustre: 19255:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 29822 previous similar messages Lustre: 19255:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 2/33/0, delete: 0/0/0 Lustre: 19255:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 29822 previous similar messages Lustre: 19255:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 19255:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 29822 previous similar messages Lustre: 4143:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x240000403:0x174:0x0]// may get corrupted (rc -108) LustreError: 28146:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0000-osc-ffff88029b341bf8: namespace resource [0x30:0x0:0x0].0x0 (ffff8802791eb4c0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 28146:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 1508 previous similar messages Lustre: lustre-OST0000-osc-ffff88029b341bf8: Connection restored to 192.168.123.215@tcp (at 0@lo) Lustre: Skipped 1 previous similar message Lustre: lustre-OST0002: haven't heard from client f1de7ff5-7e63-4e7b-ac0f-06d879070ea9 (at 0@lo) in 47 seconds. I think it's dead, and I am evicting it. exp ffff8802716c6fc8, cur 1630536631 expire 1630536601 last 1630536584 Lustre: 4138:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff88024d566440 x1709741063172800/t0(0) o10->lustre-OST0001-osc-ffff88029b341bf8@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630536679 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'cp.0' Lustre: 16710:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 499 < left 1962, rollback = 2 Lustre: 16710:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 118 previous similar messages Lustre: ll_ost06_008: service thread pid 18671 was inactive for 62.067 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 3 previous similar messages Lustre: ll_ost05_003: service thread pid 15583 was inactive for 62.029 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 61 previous similar messages ptlrpc_watchdog_fire: 145 callbacks suppressed Lustre: ll_ost07_005: service thread pid 18695 was inactive for 62.150 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 18695, comm: ll_ost07_005 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] ofd_destroy_by_fid+0x1d1/0x520 [ofd] [<0>] ofd_destroy_hdl+0x27f/0xa20 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: ll_ost07_008: service thread pid 18753 was inactive for 62.091 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 18753, comm: ll_ost07_008 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] ofd_destroy_by_fid+0x1d1/0x520 [ofd] [<0>] ofd_destroy_hdl+0x27f/0xa20 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 18759, comm: ll_ost07_012 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] ofd_destroy_by_fid+0x1d1/0x520 [ofd] [<0>] ofd_destroy_hdl+0x27f/0xa20 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe LustreError: 12728:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 227999936 != fo_tot_granted 228028608 LustreError: 12728:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 69 previous similar messages LustreError: 12728:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 28672 LustreError: 12728:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 69 previous similar messages LustreError: 10989:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0001_UUID lock: ffff88026fe09300/0x5011d9db3a804e72 lrc: 3/0,0 mode: PW/PW res: [0x300000400:0x102:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000400020020 nid: 0@lo remote: 0x5011d9db3a804c9d expref: 2054 pid: 20053 timeout: 645 lvb_type: 0 LustreError: 18676:0:(service.c:2290:ptlrpc_server_handle_request()) @@@ Dropping timed-out request from 12345-0@lo: deadline 46/54s ago req@ffff88008193d180 x1709741059505024/t0(0) o1->f1de7ff5-7e63-4e7b-ac0f-06d879070ea9@0@lo:112/0 lens 440/0 e 0 to 0 dl 1630536617 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'lfs.0' Lustre: 27766:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (46/54s); client may timeout req@ffff88008193f700 x1709741059505280/t0(0) o1->f1de7ff5-7e63-4e7b-ac0f-06d879070ea9@0@lo:112/0 lens 440/0 e 0 to 0 dl 1630536617 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'lfs.0' Lustre: 27766:0:(service.c:2328:ptlrpc_server_handle_request()) Skipped 203 previous similar messages LustreError: 18676:0:(service.c:2290:ptlrpc_server_handle_request()) Skipped 222 previous similar messages LustreError: 11-0: lustre-OST0001-osc-ffff8802996aa548: operation ost_getattr to node 0@lo failed: rc = -107 Lustre: lustre-OST0001-osc-ffff8802996aa548: Connection to lustre-OST0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0000-osc-ffff8802996aa548: This client was evicted by lustre-OST0000; in progress operations using this service will fail. LustreError: 20666:0:(service.c:2290:ptlrpc_server_handle_request()) @@@ Dropping timed-out request from 12345-0@lo: deadline 43/50s ago req@ffff880080bfb240 x1709741060058880/t0(0) o2->784550cb-a59c-40bc-a083-9d838a506239@0@lo:118/0 lens 440/0 e 0 to 0 dl 1630536623 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'touch.0' Lustre: 27766:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (46/39s); client may timeout req@ffff88007dbe8680 x1709741060245824/t0(0) o101->784550cb-a59c-40bc-a083-9d838a506239@0@lo:129/0 lens 328/0 e 0 to 0 dl 1630536634 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'chmod.0' Lustre: 27766:0:(service.c:2328:ptlrpc_server_handle_request()) Skipped 378 previous similar messages LustreError: 20666:0:(service.c:2290:ptlrpc_server_handle_request()) Skipped 363 previous similar messages LustreError: 27941:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0002-osc-ffff8802996aa548: namespace resource [0x60:0x0:0x0].0x0 (ffff88007674b4c0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 27941:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 605 previous similar messages Lustre: lustre-OST0002-osc-ffff8802996aa548: Connection restored to 192.168.123.215@tcp (at 0@lo) LustreError: 11-0: lustre-OST0002-osc-ffff8802996aa548: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 43 previous similar messages LustreError: 11123:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0001: '7' migration was interrupted, run 'lfs migrate -m 0 -c 1 -H crush 7' to finish migration. Lustre: 4137:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff880289f6d180 x1709741064378496/t0(0) o10->lustre-OST0000-osc-ffff8802996aa548@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630536732 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'cp.0' Lustre: dir [0x280000404:0x104:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: 4137:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff88026bb0de00 x1709741066957952/t0(0) o10->lustre-OST0000-osc-ffff88029b341bf8@0@lo:6/4 lens 440/432 e 1 to 0 dl 1630536747 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dir_create.sh.0' 6[32753]: segfault at 8 ip 00007f85721c17e8 sp 00007fffbd9f68a0 error 4 in ld-2.17.so[7f85721b6000+22000] LustreError: 10989:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0002_UUID lock: ffff88026ac12980/0x5011d9db3a8b5c2b lrc: 3/0,0 mode: PW/PW res: [0x340000401:0x21c:0x0].0x0 rrc: 4 type: EXT [0->18446744073709551615] (req 0->4095) gid 0 flags: 0x60000400000020 nid: 0@lo remote: 0x5011d9db3a8b5bf3 expref: 1887 pid: 18817 timeout: 789 lvb_type: 0 LustreError: 10989:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 19 previous similar messages LustreError: 25480:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630536815 with bad export cookie 5769632133239940693 LustreError: 12332:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802720b1940 x1709741074631296/t0(0) o105->lustre-OST0002@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 12332:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 1 previous similar message LustreError: 11-0: lustre-OST0002-osc-ffff88029b341bf8: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-OST0002-osc-ffff88029b341bf8: Connection to lustre-OST0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 2 previous similar messages LustreError: 167-0: lustre-OST0002-osc-ffff88029b341bf8: This client was evicted by lustre-OST0002; in progress operations using this service will fail. LustreError: Skipped 2 previous similar messages LustreError: 25480:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) Skipped 3 previous similar messages LustreError: 478:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0002-osc-ffff88029b341bf8: namespace resource [0xd80:0x0:0x0].0x0 (ffff880272976ac0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 478:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 2647 previous similar messages Lustre: lustre-OST0002-osc-ffff88029b341bf8: Connection restored to 192.168.123.215@tcp (at 0@lo) Lustre: Skipped 3 previous similar messages Lustre: mdt05_000: service thread pid 11136 was inactive for 74.202 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 9 previous similar messages INFO: task dir_create.sh:15942 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. dir_create.sh D ffff88008cd5d2c0 10832 15942 15917 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff817d5bb2>] lookup_slow+0x33/0xa7 [<ffffffff81254c7f>] link_path_walk+0x81f/0x8c0 [<ffffffff8124818c>] ? get_empty_filp+0x5c/0x1f0 [<ffffffff8125708e>] path_openat+0xae/0x5b0 [<ffffffff81258dfd>] do_filp_open+0x4d/0xb0 [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e326e>] ? _raw_spin_unlock+0xe/0x20 [<ffffffff81267893>] ? __alloc_fd+0xc3/0x170 [<ffffffff812447e4>] do_sys_open+0x124/0x220 [<ffffffff812448fe>] SyS_open+0x1e/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task dir_create.sh:15973 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. dir_create.sh D ffff88008d569280 11184 15973 15913 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff817d5bb2>] lookup_slow+0x33/0xa7 [<ffffffff81254c7f>] link_path_walk+0x81f/0x8c0 [<ffffffff8124818c>] ? get_empty_filp+0x5c/0x1f0 [<ffffffff8125708e>] path_openat+0xae/0x5b0 [<ffffffff81258dfd>] do_filp_open+0x4d/0xb0 [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e326e>] ? _raw_spin_unlock+0xe/0x20 [<ffffffff81267893>] ? __alloc_fd+0xc3/0x170 [<ffffffff812447e4>] do_sys_open+0x124/0x220 [<ffffffff812448fe>] SyS_open+0x1e/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task dir_create.sh:16105 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. dir_create.sh D ffff8802977924f0 11232 16105 15915 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff817d5bb2>] lookup_slow+0x33/0xa7 [<ffffffff81254c7f>] link_path_walk+0x81f/0x8c0 [<ffffffff8124818c>] ? get_empty_filp+0x5c/0x1f0 [<ffffffff8125708e>] path_openat+0xae/0x5b0 [<ffffffff81258dfd>] do_filp_open+0x4d/0xb0 [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e326e>] ? _raw_spin_unlock+0xe/0x20 [<ffffffff81267893>] ? __alloc_fd+0xc3/0x170 [<ffffffff812447e4>] do_sys_open+0x124/0x220 [<ffffffff812448fe>] SyS_open+0x1e/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task ls:27439 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. ls D ffff88024d7fc9d0 11232 27439 16174 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff812550a6>] do_last+0x296/0x1280 [<ffffffff812546de>] ? link_path_walk+0x27e/0x8c0 [<ffffffff812570ad>] path_openat+0xcd/0x5b0 [<ffffffff817d5822>] ? free_debug_processing+0x240/0x272 [<ffffffff81258dfd>] do_filp_open+0x4d/0xb0 [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e326e>] ? _raw_spin_unlock+0xe/0x20 [<ffffffff81267893>] ? __alloc_fd+0xc3/0x170 [<ffffffff812447e4>] do_sys_open+0x124/0x220 [<ffffffff81244914>] SyS_openat+0x14/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task ls:27440 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. ls D ffff88024d7fa4f0 11232 27440 16174 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff812550a6>] do_last+0x296/0x1280 [<ffffffff812546de>] ? link_path_walk+0x27e/0x8c0 [<ffffffff812570ad>] path_openat+0xcd/0x5b0 [<ffffffff817d5822>] ? free_debug_processing+0x240/0x272 [<ffffffff81258dfd>] do_filp_open+0x4d/0xb0 [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e326e>] ? _raw_spin_unlock+0xe/0x20 [<ffffffff81267893>] ? __alloc_fd+0xc3/0x170 [<ffffffff812447e4>] do_sys_open+0x124/0x220 [<ffffffff81244914>] SyS_openat+0x14/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task ls:27441 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. ls D ffff88024d7fb760 11232 27441 16174 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff812550a6>] do_last+0x296/0x1280 [<ffffffff812546de>] ? link_path_walk+0x27e/0x8c0 [<ffffffff812570ad>] path_openat+0xcd/0x5b0 [<ffffffff81258dfd>] do_filp_open+0x4d/0xb0 [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e326e>] ? _raw_spin_unlock+0xe/0x20 [<ffffffff81267893>] ? __alloc_fd+0xc3/0x170 [<ffffffff812447e4>] do_sys_open+0x124/0x220 [<ffffffff81244914>] SyS_openat+0x14/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task ls:27443 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. ls D ffff880292ab73c0 11232 27443 16174 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff812550a6>] do_last+0x296/0x1280 [<ffffffff812546de>] ? link_path_walk+0x27e/0x8c0 [<ffffffff812570ad>] path_openat+0xcd/0x5b0 [<ffffffff817d5822>] ? free_debug_processing+0x240/0x272 [<ffffffff81257b2d>] ? putname+0x3d/0x60 [<ffffffff81258dfd>] do_filp_open+0x4d/0xb0 [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e326e>] ? _raw_spin_unlock+0xe/0x20 [<ffffffff81267893>] ? __alloc_fd+0xc3/0x170 [<ffffffff812447e4>] do_sys_open+0x124/0x220 [<ffffffff81244914>] SyS_openat+0x14/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task ls:27445 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. ls D ffff88008bfe90c0 11232 27445 16174 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff812550a6>] do_last+0x296/0x1280 [<ffffffff812546de>] ? link_path_walk+0x27e/0x8c0 [<ffffffff812570ad>] path_openat+0xcd/0x5b0 [<ffffffff81258dfd>] do_filp_open+0x4d/0xb0 [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e326e>] ? _raw_spin_unlock+0xe/0x20 [<ffffffff81267893>] ? __alloc_fd+0xc3/0x170 [<ffffffff812447e4>] do_sys_open+0x124/0x220 [<ffffffff81244914>] SyS_openat+0x14/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task ls:28394 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. ls D ffff8800889ad2c0 10640 28394 15965 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff812550a6>] do_last+0x296/0x1280 [<ffffffff812546de>] ? link_path_walk+0x27e/0x8c0 [<ffffffff812570ad>] path_openat+0xcd/0x5b0 [<ffffffff81258dfd>] do_filp_open+0x4d/0xb0 [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e326e>] ? _raw_spin_unlock+0xe/0x20 [<ffffffff81267893>] ? __alloc_fd+0xc3/0x170 [<ffffffff812447e4>] do_sys_open+0x124/0x220 [<ffffffff81244914>] SyS_openat+0x14/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task ls:28395 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. ls D ffff880276bb0010 11232 28395 15965 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff812550a6>] do_last+0x296/0x1280 [<ffffffff812546de>] ? link_path_walk+0x27e/0x8c0 [<ffffffff812570ad>] path_openat+0xcd/0x5b0 [<ffffffff81258dfd>] do_filp_open+0x4d/0xb0 [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e326e>] ? _raw_spin_unlock+0xe/0x20 [<ffffffff81267893>] ? __alloc_fd+0xc3/0x170 [<ffffffff812447e4>] do_sys_open+0x124/0x220 [<ffffffff81244914>] SyS_openat+0x14/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 Lustre: 22144:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 2/8/1, destroy: 1/4/0 Lustre: 22144:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 89642 previous similar messages Lustre: 22144:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 2/2/0, xattr_set: 9/716/0 Lustre: 22144:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 89642 previous similar messages Lustre: 22144:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1/10/0, punch: 0/0/0, quota 8/80/0 Lustre: 22144:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 89642 previous similar messages Lustre: 22144:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 12/194/2, delete: 3/6/1 Lustre: 22144:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 89642 previous similar messages Lustre: 22144:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 7/7/0, ref_del: 3/3/0 Lustre: 22144:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 89642 previous similar messages LustreError: 11-0: lustre-MDT0001-mdc-ffff88029b341bf8: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: 31817:0:(llite_lib.c:2983:ll_prep_inode()) new_inode -fatal: rc -5 LustreError: 524:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff88029b341bf8: inode [0x240000404:0xfd:0x0] mdc close failed: rc = -108 Lustre: 11146:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 508 < left 22327, rollback = 7 Lustre: 11146:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 89524 previous similar messages 6[531]: segfault at 8 ip 00007f8d656697e8 sp 00007ffddfa9d3d0 error 4 in ld-2.17.so[7f8d6565e000+22000] LustreError: 531:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff88029b341bf8: inode [0x240000404:0x2b1:0x0] mdc close failed: rc = -13 LustreError: 531:0:(file.c:234:ll_close_inode_openhandle()) Skipped 15 previous similar messages LustreError: 12338:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0001-osp-MDT0002: fail to cancel 1 llog-records: rc = -116 LustreError: 12338:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0001-osp-MDT0002: fail to cancel 1 of 1 llog-records: rc = -116 LustreError: 11171:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0002: failed to get lu_attr of [0x280000404:0x104:0x0]: rc = -2 LustreError: 12338:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0000-osp-MDT0002: fail to cancel 1 llog-records: rc = -116 LustreError: 12338:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0000-osp-MDT0002: fail to cancel 1 of 1 llog-records: rc = -116 Lustre: 19255:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 8: before 512 < left 1443, rollback = 9 Lustre: 19255:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 186 previous similar messages LustreError: 21558:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 273668928 != fo_tot_granted 273697600 LustreError: 21558:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 129 previous similar messages LustreError: 21558:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 28672 LustreError: 21558:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 129 previous similar messages Lustre: 11167:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 6s req@ffff880077fa6440 x1709741082021696/t0(0) o37->f1de7ff5-7e63-4e7b-ac0f-06d879070ea9@0@lo:0/0 lens 448/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'ls.0' Lustre: 16825:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000404:0x175:0x0] with magic=0xbd60bd0 Lustre: 16825:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 1 previous similar message Lustre: 16700:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000404:0x168:0x0] with magic=0xbd60bd0 Lustre: 16700:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 5 previous similar messages LustreError: 2126:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff8802996aa548: inode [0x200000403:0x254:0x0] mdc close failed: rc = -13 LustreError: 17542:0:(mdt_xattr.c:429:mdt_dir_layout_update()) lustre-MDT0002: [0x280000404:0x232:0x0] migrate mdt index mismatch 2 != 1 LustreError: 10989:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 108s: evicting client at 0@lo ns: filter-lustre-OST0003_UUID lock: ffff88026ba7b880/0x5011d9db3aad756a lrc: 3/0,0 mode: PW/PW res: [0xe52:0x0:0x0].0x0 rrc: 4 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000400000020 nid: 0@lo remote: 0x5011d9db3aad7547 expref: 2929 pid: 27847 timeout: 968 lvb_type: 0 LustreError: 10989:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 1 previous similar message LustreError: 18157:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88026a4fde00 x1709741085301760/t0(0) o106->lustre-OST0003@0@lo:15/16 lens 328/280 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 18157:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 1 previous similar message LustreError: 10958:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630537002 with bad export cookie 5769632133241635694 Lustre: lustre-OST0003-osc-ffff88029b341bf8: Connection to lustre-OST0003 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 1 previous similar message LustreError: 167-0: lustre-OST0003-osc-ffff88029b341bf8: This client was evicted by lustre-OST0003; in progress operations using this service will fail. LustreError: Skipped 1 previous similar message Lustre: lustre-OST0003-osc-ffff88029b341bf8: Connection restored to 192.168.123.215@tcp (at 0@lo) Lustre: Skipped 1 previous similar message LustreError: 3686:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff880086c5d7c0 x1709741086979136/t0(0) o104->lustre-OST0000@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 10967:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630537038 with bad export cookie 5769632133241989614 LustreError: 10967:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) Skipped 1 previous similar message LustreError: 3686:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 14 previous similar messages LustreError: 11-0: lustre-OST0000-osc-ffff8802996aa548: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: 22526:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000404:0x1c8:0x0]: rc = -2 Lustre: 20671:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000405:0x77:0x0] with magic=0xbd60bd0 Lustre: 20671:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 1 previous similar message LustreError: 3864:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff8802996aa548: inode [0x240000404:0x328:0x0] mdc close failed: rc = -13 LustreError: 4503:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff88029b341bf8: inode [0x200000403:0x254:0x0] mdc close failed: rc = -13 LustreError: 2755:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0001: failed to get lu_attr of [0x240000404:0xe2:0x0]: rc = -2 LustreError: 4787:0:(dir.c:713:ll_dir_get_default_layout()) unknown magic: CD40CD0 ------------[ cut here ]------------ WARNING: CPU: 0 PID: 20833 at /home/green/git/lustre-release/ldiskfs/ext4_jbd2.c:266 __ldiskfs_handle_dirty_metadata+0x1d2/0x230 [ldiskfs] Modules linked in: loop zfs(PO) zunicode(PO) zzstd(O) zlua(O) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) jbd2 mbcache lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) dm_flakey dm_mod libcfs(OE) crc_t10dif crct10dif_generic sb_edac edac_core iosf_mbi crc32_pclmul ghash_clmulni_intel aesni_intel lrw gf128mul glue_helper ablk_helper cryptd virtio_console virtio_balloon pcspkr i2c_piix4 ip_tables rpcsec_gss_krb5 drm_kms_helper ata_generic ttm crct10dif_pclmul pata_acpi crct10dif_common drm crc32c_intel ata_piix drm_panel_orientation_quirks serio_raw libata virtio_blk i2c_core floppy CPU: 0 PID: 20833 Comm: mdt00_011 Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011 Call Trace: [<ffffffff817d9418>] dump_stack+0x19/0x1b [<ffffffff8108d538>] __warn+0xd8/0x100 [<ffffffff8108d67d>] warn_slowpath_null+0x1d/0x20 [<ffffffffa0bf9262>] __ldiskfs_handle_dirty_metadata+0x1d2/0x230 [ldiskfs] [<ffffffffa0bd933d>] ldiskfs_getblk+0x13d/0x210 [ldiskfs] [<ffffffffa0bd9437>] ldiskfs_bread+0x27/0xe0 [ldiskfs] [<ffffffffa0c9de97>] osd_write+0x4e7/0xcb0 [osd_ldiskfs] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffffa04d76f3>] dt_record_write+0x33/0x120 [obdclass] [<ffffffffa04966b2>] llog_osd_write_rec+0x1052/0x19a0 [obdclass] [<ffffffffa0483760>] llog_write_rec+0x290/0x590 [obdclass] [<ffffffffa04894ee>] llog_cat_add_rec+0x23e/0x950 [obdclass] [<ffffffffa04805ff>] llog_add+0x17f/0x1f0 [obdclass] [<ffffffffa04b1a39>] ? lprocfs_counter_add+0xf9/0x160 [obdclass] [<ffffffffa081a60d>] sub_updates_write+0xb6b/0xe32 [ptlrpc] [<ffffffff811beccd>] ? __free_pages+0x1d/0x30 [<ffffffff8121efe4>] ? kfree+0x1e4/0x200 [<ffffffffa07ea892>] top_trans_stop+0x7c2/0xf60 [ptlrpc] [<ffffffffa0ed8f3c>] lod_trans_stop+0x25c/0x340 [lod] [<ffffffffa0dad408>] mdd_trans_stop+0x28/0x16e [mdd] [<ffffffffa0d8eb2f>] mdd_migrate_object+0xf1f/0x1ba0 [mdd] [<ffffffffa04d6375>] ? lu_object_find_at+0x285/0xb10 [obdclass] [<ffffffffa0d8fb86>] mdd_migrate+0x3d6/0x8e0 [mdd] [<ffffffffa0e610d8>] mdo_migrate+0x4c/0x4e [mdt] [<ffffffffa0e20e53>] mdt_reint_migrate+0x1023/0x11d0 [mdt] [<ffffffffa0e21087>] mdt_reint_rec+0x87/0x240 [mdt] [<ffffffffa0df655c>] mdt_reint_internal+0x76c/0xb40 [mdt] [<ffffffffa0e03657>] mdt_reint+0x67/0x150 [mdt] [<ffffffffa07d926e>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<ffffffffa03bd45e>] ? libcfs_nid2str_r+0xfe/0x130 [lnet] [<ffffffffa0784990>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<ffffffffa0786529>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffffa0785950>] ? ptlrpc_wait_event+0x620/0x620 [ptlrpc] [<ffffffff810ba0f4>] kthread+0xe4/0xf0 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 ---[ end trace af9871a41479a8b7 ]--- LDISKFS-fs: ldiskfs_getblk:888: aborting transaction: error 28 in __ldiskfs_handle_dirty_metadata LDISKFS-fs error (device dm-0): ldiskfs_getblk:888: inode #197: block 53692: comm mdt00_011: journal_dirty_metadata failed: handle type 0 started at line 1972, credits 512/0, errcode -28 Aborting journal on device dm-0-8. LDISKFS-fs (dm-0): Remounting filesystem read-only LustreError: 20833:0:(osd_io.c:2103:osd_ldiskfs_write_record()) lustre-MDT0000: error reading offset 253952 (block 62, size 8, offs 253904), credits 0/0: rc = -28 LustreError: 20833:0:(update_trans.c:985:top_trans_stop()) lustre-MDT0000-osd: write updates failed: rc = -28 LustreError: 20833:0:(osd_io.c:2110:osd_ldiskfs_write_record()) journal_get_write_access() returned error -30 LustreError: 20833:0:(tgt_lastrcvd.c:1253:tgt_add_reply_data()) lustre-MDT0000: can't update reply_data file: rc = -30 LustreError: 20833:0:(osd_handler.c:2075:osd_trans_stop()) lustre-MDT0000: failed in transaction hook: rc = -30 LDISKFS-fs error (device dm-0) in osd_trans_stop:2082: error 28 LustreError: 20833:0:(osd_handler.c:2085:osd_trans_stop()) lustre-MDT0000: failed to stop transaction: rc = -28 LustreError: 10916:0:(osd_handler.c:1779:osd_trans_commit_cb()) transaction @0xffff88027c178258 commit error: 2 LustreError: 20833:0:(update_trans.c:1011:top_trans_stop()) lustre-MDT0000-osd: stop trans failed: rc = -30 LustreError: 20833:0:(update_trans.c:1084:top_trans_stop()) lustre-OST0002-osc-MDT0000: stop trans failed: rc = -30 LustreError: 11480:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0000-osd: fail to cancel 1 llog-records: rc = -30 LustreError: 11480:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0000-osd: fail to cancel 1 of 1 llog-records: rc = -30 LustreError: 4311:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -30 LustreError: 10989:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0000_UUID lock: ffff880080824b40/0x5011d9db3a9d361a lrc: 3/0,0 mode: PW/PW res: [0x2c0000401:0x39b:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000400020020 nid: 0@lo remote: 0x5011d9db3a9d35e9 expref: 3026 pid: 18809 timeout: 1178 lvb_type: 0 LustreError: 10989:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 1 previous similar message LustreError: 27844:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff880267358040 x1709741093114240/t0(0) o104->lustre-OST0003@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 27844:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 60 previous similar messages LustreError: 11-0: lustre-OST0003-osc-ffff88029b341bf8: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-OST0003-osc-ffff88029b341bf8: Connection to lustre-OST0003 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 1 previous similar message LustreError: 167-0: lustre-OST0003-osc-ffff88029b341bf8: This client was evicted by lustre-OST0003; in progress operations using this service will fail. LustreError: Skipped 1 previous similar message LustreError: 19346:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630537206 with bad export cookie 5769632133243097623 LustreError: 19346:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) Skipped 11 previous similar messages LustreError: 5479:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -30 LustreError: 6172:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -30 LustreError: 6172:0:(llite_lib.c:1836:ll_md_setattr()) Skipped 2 previous similar messages LustreError: 17542:0:(llog_cat.c:602:llog_cat_add_rec()) llog_write_rec -116: lh=ffff880299451dd8 LustreError: 17542:0:(update_trans.c:1062:top_trans_stop()) lustre-MDT0000-osp-MDT0002: write updates failed: rc = -116 | Externally reported by onyx-68 boilpot email |
racer test 1: racer on clients: centos-115.localnet DURATION=2700 | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffa1033e52>] ll_migrate+0x9b2/0xec0 [lustre] PGD 8000000273f2c067 PUD 29123c067 PMD 0 Oops: 0000 [#1] SMP DEBUG_PAGEALLOC Modules linked in: loop zfs(PO) zunicode(PO) zzstd(O) zlua(O) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) jbd2 mbcache lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) dm_flakey dm_mod libcfs(OE) crc_t10dif crct10dif_generic sb_edac edac_core iosf_mbi crc32_pclmul ghash_clmulni_intel aesni_intel lrw gf128mul glue_helper ablk_helper cryptd virtio_console virtio_balloon pcspkr i2c_piix4 ip_tables rpcsec_gss_krb5 drm_kms_helper ata_generic ttm pata_acpi drm crct10dif_pclmul crct10dif_common drm_panel_orientation_quirks ata_piix crc32c_intel serio_raw virtio_blk i2c_core libata floppy CPU: 3 PID: 17684 Comm: lfs Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011 task: ffff88008ad449d0 ti: ffff8802298c0000 task.ti: ffff8802298c0000 RIP: 0010:[<ffffffffa1033e52>] [<ffffffffa1033e52>] ll_migrate+0x9b2/0xec0 [lustre] RSP: 0018:ffff8802298c3bc8 EFLAGS: 00010206 RAX: 0000000000000000 RBX: ffff88007414ad58 RCX: 0000000000000000 RDX: 000000010000edd5 RSI: ffff880326c18608 RDI: ffff880326c185e8 RBP: ffff8802298c3c48 R08: ffff880084945518 R09: ffff8802298c3a60 R10: 0000000000000003 R11: ffff8802802bd978 R12: 0000000000000000 R13: ffff880249945628 R14: ffff88029ba9d458 R15: 0000000000000030 FS: 00007f4f15c5c740(0000) GS:ffff880331ac0000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 CR2: 0000000000000008 CR3: 00000000890e8000 CR4: 00000000001607e0 Call Trace: [<ffffffff81242213>] ? __check_object_size+0x1c3/0x220 [<ffffffffa1021561>] ll_dir_ioctl+0x5d01/0x6ed0 [lustre] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff8115260f>] ? delayacct_end+0x8f/0xb0 [<ffffffff81152744>] ? __delayacct_blkio_end+0x34/0x60 [<ffffffff817e0257>] ? io_schedule_timeout+0xe7/0x130 [<ffffffff811b62dd>] ? find_get_pages_tag+0x10d/0x260 [<ffffffff811c3691>] ? pagevec_lookup_tag+0x21/0x30 [<ffffffff811b400e>] ? __filemap_fdatawait_range+0xbe/0x1a0 [<ffffffff8125b3fd>] do_vfs_ioctl+0x40d/0x6c0 [<ffffffff81264d2b>] ? iput+0x3b/0x180 [<ffffffff8125b751>] SyS_ioctl+0xa1/0xc0 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 | Lustre: lustre-OST0002-osc-ffff88029a49c138: disconnect after 21s idle LustreError: 11141:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0002-osp-MDT0001: fail to cancel 1 llog-records: rc = -116 LustreError: 11141:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0002-osp-MDT0001: fail to cancel 1 of 1 llog-records: rc = -116 LustreError: 11141:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0002-osp-MDT0001: fail to cancel 1 llog-records: rc = -116 LustreError: 11141:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0002-osp-MDT0001: fail to cancel 1 of 1 llog-records: rc = -116 LustreError: 10387:0:(out_handler.c:910:out_tx_end()) lustre-MDT0001-osd: undo for /home/green/git/lustre-release/lustre/ptlrpc/../../lustre/target/out_handler.c:445: rc = -524 LustreError: 10387:0:(out_handler.c:910:out_tx_end()) lustre-MDT0001-osd: undo for /home/green/git/lustre-release/lustre/ptlrpc/../../lustre/target/out_handler.c:445: rc = -524 LustreError: 10579:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0001-osp-MDT0000: fail to cancel 1 llog-records: rc = -116 LustreError: 10579:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0001-osp-MDT0000: fail to cancel 1 of 1 llog-records: rc = -116 Lustre: 16454:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 512 < left 1669, rollback = 7 Lustre: 16454:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 16454:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 1/89/0 Lustre: 16454:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1660/1669/0, punch: 0/0/0, quota 3/3/0 Lustre: 16454:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 16454:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 10210:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 512 < left 1216, rollback = 7 Lustre: 10210:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 1659 previous similar messages Lustre: 10210:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 10210:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 1659 previous similar messages Lustre: 10210:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 1/89/0 Lustre: 10210:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 1659 previous similar messages Lustre: 10210:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1207/1216/0, punch: 0/0/0, quota 3/3/0 Lustre: 10210:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 1659 previous similar messages Lustre: 10210:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 10210:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 1659 previous similar messages Lustre: 10210:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 10210:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 1659 previous similar messages ln (16600) used greatest stack depth: 10144 bytes left LustreError: 10579:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0001-osp-MDT0000: fail to cancel 1 llog-records: rc = -2 LustreError: 10579:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) Skipped 1 previous similar message LustreError: 10579:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0001-osp-MDT0000: fail to cancel 1 of 1 llog-records: rc = -2 LustreError: 10579:0:(llog_cat.c:787:llog_cat_cancel_records()) Skipped 1 previous similar message LustreError: 16859:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0000: '10' migration was interrupted, run 'lfs migrate -m 1 -c 1 -H crush 10' to finish migration. Lustre: 16306:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 506 < left 15017, rollback = 7 Lustre: 16306:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 3108 previous similar messages Lustre: 16306:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 349/1396/0, destroy: 0/0/0 Lustre: 16306:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 3108 previous similar messages Lustre: 16306:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/1, xattr_set: 351/15/0 Lustre: 16306:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 3108 previous similar messages Lustre: 16306:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1746/15017/0, punch: 0/0/0, quota 10/106/5 Lustre: 16306:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 3108 previous similar messages Lustre: 16306:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 349/5933/0, delete: 0/0/0 Lustre: 16306:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 3108 previous similar messages Lustre: 16306:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 16306:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 3108 previous similar messages Lustre: 16413:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 506 < left 71347, rollback = 7 Lustre: 16413:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 1398 previous similar messages Lustre: 16413:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 1659/6636/0, destroy: 0/0/0 Lustre: 16413:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 1398 previous similar messages Lustre: 16413:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/1, xattr_set: 1661/15/0 Lustre: 16413:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 1398 previous similar messages Lustre: 16413:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 8296/71347/0, punch: 0/0/0, quota 10/106/5 Lustre: 16413:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 1398 previous similar messages Lustre: 16413:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 1659/28203/0, delete: 0/0/0 Lustre: 16413:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 1398 previous similar messages Lustre: 16413:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 16413:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 1398 previous similar messages sched: RT throttling activated Lustre: 17812:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 512 < left 1433, rollback = 7 Lustre: 17812:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 6652 previous similar messages Lustre: 17812:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 0/0/0, destroy: 0/0/0 Lustre: 17812:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 6652 previous similar messages Lustre: 17812:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 0/0/0, xattr_set: 1/89/0 Lustre: 17812:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 6652 previous similar messages Lustre: 17812:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1424/1433/0, punch: 0/0/0, quota 3/3/0 Lustre: 17812:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 6652 previous similar messages Lustre: 17812:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 0/0/0, delete: 0/0/0 Lustre: 17812:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 6652 previous similar messages Lustre: 17812:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 17812:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 6652 previous similar messages Lustre: 13276:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 500 < left 716, rollback = 2 LustreError: 10234:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0002: failed to get lu_attr of [0x280000403:0x29:0x0]: rc = -2 Lustre: 10194:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 504 < left 521, rollback = 2 Lustre: 10194:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 5 previous similar messages Lustre: 10198:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 504 < left 71347, rollback = 7 Lustre: 10198:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 1423 previous similar messages Lustre: 10198:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 1659/6636/0, destroy: 0/0/0 Lustre: 10198:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 1433 previous similar messages Lustre: 10198:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/1, xattr_set: 1661/15/0 Lustre: 10198:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 1433 previous similar messages Lustre: 10198:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 8296/71347/0, punch: 0/0/0, quota 14/174/7 Lustre: 10198:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 1433 previous similar messages Lustre: 10198:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 1659/28203/0, delete: 0/0/0 Lustre: 10198:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 1433 previous similar messages Lustre: 10198:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 10198:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 1433 previous similar messages ------------[ cut here ]------------ WARNING: CPU: 6 PID: 16454 at /home/green/git/lustre-release/ldiskfs/namei.c:3331 ldiskfs_orphan_add+0x11e/0x2a0 [ldiskfs] Modules linked in: loop zfs(PO) zunicode(PO) zzstd(O) zlua(O) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) jbd2 mbcache lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) dm_flakey dm_mod libcfs(OE) crc_t10dif crct10dif_generic sb_edac edac_core iosf_mbi crc32_pclmul ghash_clmulni_intel aesni_intel lrw gf128mul glue_helper ablk_helper cryptd virtio_console virtio_balloon pcspkr i2c_piix4 ip_tables rpcsec_gss_krb5 drm_kms_helper ata_generic ttm pata_acpi drm crct10dif_pclmul crct10dif_common drm_panel_orientation_quirks ata_piix crc32c_intel serio_raw virtio_blk i2c_core libata floppy CPU: 6 PID: 16454 Comm: mdt03_007 Tainted: P OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011 Call Trace: [<ffffffff817d9418>] dump_stack+0x19/0x1b [<ffffffff8108d538>] __warn+0xd8/0x100 [<ffffffff8108d67d>] warn_slowpath_null+0x1d/0x20 [<ffffffffa0bcb4ae>] ldiskfs_orphan_add+0x11e/0x2a0 [ldiskfs] [<ffffffffa0bbc9eb>] ldiskfs_xattr_inode_orphan_add.constprop.17+0xbb/0x110 [ldiskfs] [<ffffffffa0bbf417>] ldiskfs_xattr_delete_inode+0x77/0x3c0 [ldiskfs] [<ffffffffa0bdfaa0>] ? ldiskfs_evict_inode+0x190/0x670 [ldiskfs] [<ffffffffa0bdfad9>] ldiskfs_evict_inode+0x1c9/0x670 [ldiskfs] [<ffffffff812649af>] evict+0xaf/0x180 [<ffffffff81264de5>] iput+0xf5/0x180 [<ffffffffa0c74d08>] osd_object_delete+0x1f8/0x370 [osd_ldiskfs] [<ffffffffa04cc59d>] lu_object_free.isra.26+0xcd/0x1d0 [obdclass] [<ffffffffa04d0cf1>] lu_object_put+0xa1/0x480 [obdclass] [<ffffffffa0e13440>] mdt_object_put+0x30/0x110 [mdt] [<ffffffffa0e1ae18>] mdt_reint_unlink+0x918/0x19b0 [mdt] [<ffffffffa04f279e>] ? lu_ucred+0x1e/0x30 [obdclass] [<ffffffffa0e0aaf2>] ? ucred_set_audit_enabled.isra.13+0x22/0x60 [mdt] [<ffffffffa0e22087>] mdt_reint_rec+0x87/0x240 [mdt] [<ffffffffa0df755c>] mdt_reint_internal+0x76c/0xb40 [mdt] [<ffffffffa0e04657>] mdt_reint+0x67/0x150 [mdt] [<ffffffffa07d426e>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<ffffffffa03e345e>] ? libcfs_nid2str_r+0xfe/0x130 [lnet] [<ffffffffa077f990>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<ffffffffa0781529>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e32be>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffffa0780950>] ? ptlrpc_wait_event+0x620/0x620 [ptlrpc] [<ffffffff810ba0f4>] kthread+0xe4/0xf0 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 ---[ end trace 00015c15de20e74d ]--- Lustre: 10212:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 506 < left 521, rollback = 2 Lustre: 10212:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 3 previous similar messages Lustre: 15892:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 8: before 511 < left 28219, rollback = 9 Lustre: 15892:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 5 previous similar messages Lustre: 17811:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 8: before 510 < left 5949, rollback = 9 Lustre: 17811:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 349/1396/0, destroy: 1/4/0 Lustre: 17811:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 20932 previous similar messages Lustre: 17811:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 3/3/1, xattr_set: 353/148/0 Lustre: 17811:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 20932 previous similar messages Lustre: 17811:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 1746/15017/0, punch: 0/0/0, quota 12/124/0 Lustre: 17811:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 20932 previous similar messages Lustre: 17811:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 350/5949/0, delete: 2/5/1 Lustre: 17811:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 20932 previous similar messages Lustre: 17811:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 1/1/0, ref_del: 2/2/0 Lustre: 17811:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 20932 previous similar messages Lustre: 17811:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 508 < left 15017, rollback = 7 Lustre: 17811:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 20925 previous similar messages Lustre: 15902:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 498 < left 849, rollback = 2 Lustre: 15902:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 6 previous similar messages LustreError: 10240:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0002: failed to get lu_attr of [0x280000404:0x52:0x0]: rc = -2 LustreError: 10090:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0002_UUID lock: ffff8800840a3100/0xb3325594733966fb lrc: 3/0,0 mode: PW/PW res: [0x340000401:0x2:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->262143) gid 0 flags: 0x60000000020020 nid: 0@lo remote: 0xb3325594733966f4 expref: 5296 pid: 12303 timeout: 589 lvb_type: 0 LustreError: 10068:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8800827d6a80 x1709739784832192/t0(0) o105->lustre-OST0002@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 11-0: lustre-OST0002-osc-ffff880299bd4138: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-OST0002-osc-ffff880299bd4138: Connection to lustre-OST0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0002-osc-ffff880299bd4138: This client was evicted by lustre-OST0002; in progress operations using this service will fail. Lustre: dir [0x240000404:0x46:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: 21649:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 509 < left 32988, rollback = 7 Lustre: 21649:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 4958 previous similar messages Lustre: 21649:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 765/3060/0, destroy: 1/4/0 Lustre: 21649:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 5008 previous similar messages Lustre: 21649:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 2/2/0, xattr_set: 767/134/2 Lustre: 21649:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 5008 previous similar messages Lustre: 21649:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 3825/32988/0, punch: 0/0/0, quota 6/6/0 Lustre: 21649:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 5008 previous similar messages Lustre: 21649:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 767/13037/0, delete: 1/4/0 Lustre: 21649:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 5008 previous similar messages Lustre: 21649:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 1/1/0, ref_del: 2/2/1 Lustre: 21649:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 5008 previous similar messages Lustre: 3910:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x200000404:0xca:0x0]// may get corrupted (rc -108) LustreError: 23535:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0002-osc-ffff880299bd4138: namespace resource [0x340000400:0x1c0:0x0].0x0 (ffff88025a8f4cc0) refcount nonzero (1) after lock cleanup; forcing cleanup. Lustre: 10192:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 2: before 508 < left 1222, rollback = 2 Lustre: 10192:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 42 previous similar messages LustreError: 10090:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 101s: evicting client at 0@lo ns: filter-lustre-OST0002_UUID lock: ffff8800806843c0/0xb3325594733a5b14 lrc: 3/0,0 mode: PW/PW res: [0x340000400:0x4:0x0].0x0 rrc: 2 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000400020020 nid: 0@lo remote: 0xb3325594733a598c expref: 4685 pid: 16418 timeout: 601 lvb_type: 0 LustreError: 11135:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff880268ae57c0 x1709739785699712/t0(0) o105->lustre-OST0002@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 11135:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 10 previous similar messages LustreError: 11-0: lustre-OST0002-osc-ffff88029a49c138: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-OST0002-osc-ffff88029a49c138: Connection to lustre-OST0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0002-osc-ffff88029a49c138: This client was evicted by lustre-OST0002; in progress operations using this service will fail. Lustre: ll_ost04_011: service thread pid 19858 was inactive for 62.021 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 19858, comm: ll_ost04_011 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] tgt_extent_lock+0xea/0x2a0 [ptlrpc] [<0>] ofd_getattr_hdl+0x365/0x720 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 12318, comm: ll_ost04_002 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] tgt_extent_lock+0xea/0x2a0 [ptlrpc] [<0>] ofd_getattr_hdl+0x365/0x720 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Pid: 17725, comm: ll_ost04_004 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] tgt_extent_lock+0xea/0x2a0 [ptlrpc] [<0>] ofd_getattr_hdl+0x365/0x720 [ofd] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: ll_ost04_006: service thread pid 17880 was inactive for 62.322 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: ll_ost04_020: service thread pid 19869 was inactive for 62.090 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 12 previous similar messages Lustre: ll_ost04_036: service thread pid 19894 was inactive for 62.087 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 9 previous similar messages LustreError: 10090:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0000_UUID lock: ffff88025ab86d00/0xb3325594733e23c2 lrc: 3/0,0 mode: PW/PW res: [0x2c0000400:0x5:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000000030020 nid: 0@lo remote: 0xb3325594733e239f expref: 4215 pid: 17701 timeout: 608 lvb_type: 0 LustreError: 11-0: lustre-OST0000-osc-ffff88029a49c138: operation ldlm_enqueue to node 0@lo failed: rc = -107 Lustre: lustre-OST0000-osc-ffff88029a49c138: Connection to lustre-OST0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: ll_ost04_051: service thread pid 19917 was inactive for 62.037 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 1 previous similar message LustreError: 10229:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0002: failed to get lu_attr of [0x280000404:0x97:0x0]: rc = -2 LustreError: 24786:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0002-osc-ffff88029a49c138: namespace resource [0x40:0x0:0x0].0x0 (ffff880272382d40) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 24786:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 420 previous similar messages LustreError: 24786:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0002-osc-ffff88029a49c138: namespace resource [0x50:0x0:0x0].0x0 (ffff88007b5487c0) refcount nonzero (1) after lock cleanup; forcing cleanup. Lustre: ll_ost04_063: service thread pid 19998 was inactive for 62.244 seconds. Watchdog stack traces are limited to 3 per 300 seconds, skipping this one. Lustre: Skipped 11 previous similar messages Lustre: 10212:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 497 < left 1411, rollback = 2 Lustre: 10212:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 61 previous similar messages Lustre: lustre-OST0002: haven't heard from client 25ea6364-a008-4682-b7ef-9cf7da42069f (at 0@lo) in 48 seconds. I think it's dead, and I am evicting it. exp ffff88008314efc8, cur 1630535354 expire 1630535324 last 1630535306 Lustre: 3907:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1630535294/real 1630535294] req@ffff880078891f80 x1709739783994560/t0(0) o101->lustre-OST0000-osc-ffff88029a49c138@0@lo:28/4 lens 328/400 e 0 to 1 dl 1630535356 ref 1 fl Rpc:XNQr/0/ffffffff rc 0/-1 job:'ln.0' Lustre: 3907:0:(client.c:2285:ptlrpc_expire_one_request()) Skipped 2 previous similar messages Lustre: 3907:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1630535291/real 1630535291] req@ffff88007c7fb240 x1709739783756544/t0(0) o101->lustre-OST0000-osc-ffff88029a49c138@0@lo:28/4 lens 328/400 e 0 to 1 dl 1630535353 ref 1 fl Rpc:XNQr/0/ffffffff rc 0/-1 job:'ln.0' Lustre: 3907:0:(client.c:2285:ptlrpc_expire_one_request()) Skipped 143 previous similar messages LustreError: 10187:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0002: '10' migration was interrupted, run 'lfs migrate -m 0 -c 1 -H crush 10' to finish migration. Lustre: 3908:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1630535294/real 1630535294] req@ffff88025ebca5c0 x1709739783991168/t0(0) o101->lustre-OST0000-osc-ffff88029a49c138@0@lo:28/4 lens 328/400 e 0 to 1 dl 1630535356 ref 1 fl Rpc:XNQr/0/ffffffff rc 0/-1 job:'ln.0' Lustre: 3908:0:(client.c:2285:ptlrpc_expire_one_request()) Skipped 287 previous similar messages LustreError: 10090:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0001_UUID lock: ffff8802799e6580/0xb33255947340ee01 lrc: 3/0,0 mode: PW/PW res: [0x300000400:0x1c1:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000400000020 nid: 0@lo remote: 0xb33255947340edfa expref: 4889 pid: 17977 timeout: 643 lvb_type: 0 LustreError: 19901:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8802638bbec0 x1709739788132160/t0(0) o104->lustre-OST0001@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 19901:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 1 previous similar message LustreError: 10067:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630535360 with bad export cookie 12912477177701234064 Lustre: lustre-OST0003-osc-ffff880299bd4138: Connection to lustre-OST0003 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 167-0: lustre-OST0003-osc-ffff880299bd4138: This client was evicted by lustre-OST0003; in progress operations using this service will fail. Lustre: 3899:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x280000404:0xdc:0x0]// may get corrupted (rc -108) LustreError: 15419:0:(vvp_io.c:1793:vvp_io_init()) lustre: refresh file layout [0x280000404:0xdc:0x0] error -108. Lustre: 3900:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x280000403:0x60:0x0]// may get corrupted (rc -108) Lustre: 3899:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x240000404:0x43:0x0]/ may get corrupted (rc -108) Lustre: 3900:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x280000404:0xe:0x0]/ may get corrupted (rc -108) Lustre: 3900:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x240000404:0x1a2:0x0]// may get corrupted (rc -108) Lustre: 3900:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x200000403:0x55:0x0]// may get corrupted (rc -108) Lustre: 3899:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x280000404:0x4:0x0]/ may get corrupted (rc -108) Lustre: 3899:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x240000404:0xa8:0x0]// may get corrupted (rc -108) Lustre: 3899:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x280000404:0x145:0x0]// may get corrupted (rc -108) Lustre: 19943:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (94/7s); client may timeout req@ffff8802c5c089a8 x1709739781102016/t0(0) o1->1815ede3-8f0f-490c-9f90-6f2bb8c847ec@0@lo:359/0 lens 440/432 e 2 to 0 dl 1630535354 ref 1 fl Complete:/0/0 rc 0/0 job:'lfs.0' LustreError: 19865:0:(service.c:2290:ptlrpc_server_handle_request()) @@@ Dropping timed-out request from 12345-0@lo: deadline 68/33s ago req@ffff88007cb9bec0 x1709739781118272/t0(0) o1->1815ede3-8f0f-490c-9f90-6f2bb8c847ec@0@lo:333/0 lens 440/0 e 1 to 0 dl 1630535328 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'lfs.0' Lustre: 19865:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (68/33s); client may timeout req@ffff88008c320cc0 x1709739781118400/t0(0) o1->1815ede3-8f0f-490c-9f90-6f2bb8c847ec@0@lo:333/0 lens 440/0 e 1 to 0 dl 1630535328 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'lfs.0' Lustre: 19865:0:(service.c:2328:ptlrpc_server_handle_request()) Skipped 14 previous similar messages LustreError: 19865:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88026eae25c0 x1709739788236928/t0(0) o104->lustre-OST0001@0@lo:15/16 lens 328/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 19865:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 61 previous similar messages LustreError: 19927:0:(service.c:2290:ptlrpc_server_handle_request()) @@@ Dropping timed-out request from 12345-0@lo: deadline 69/32s ago req@ffff88026ba87700 x1709739781161920/t0(0) o1->1815ede3-8f0f-490c-9f90-6f2bb8c847ec@0@lo:335/0 lens 440/0 e 1 to 0 dl 1630535330 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'lfs.0' LustreError: 19927:0:(service.c:2290:ptlrpc_server_handle_request()) Skipped 166 previous similar messages LustreError: 11-0: lustre-OST0001-osc-ffff880299bd4138: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: 167-0: lustre-OST0001-osc-ffff880299bd4138: This client was evicted by lustre-OST0001; in progress operations using this service will fail. LustreError: 10084:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630535362 with bad export cookie 12912477177701234036 Lustre: 17771:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (69/33s); client may timeout req@ffff8802c63a09a8 x1709739781179712/t0(0) o1->1815ede3-8f0f-490c-9f90-6f2bb8c847ec@0@lo:335/0 lens 440/0 e 1 to 0 dl 1630535330 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'lfs.0' Lustre: 17771:0:(service.c:2328:ptlrpc_server_handle_request()) Skipped 219 previous similar messages LustreError: 10084:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) Skipped 7 previous similar messages LustreError: 23535:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0002-osc-ffff880299bd4138: namespace resource [0xd0:0x0:0x0].0x0 (ffff88025a9c6840) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 23535:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 502 previous similar messages Lustre: lustre-OST0002-osc-ffff880299bd4138: Connection restored to 192.168.123.215@tcp (at 0@lo) LustreError: 19866:0:(service.c:2290:ptlrpc_server_handle_request()) @@@ Dropping timed-out request from 12345-0@lo: deadline 69/33s ago req@ffff880078a870c0 x1709739781182208/t0(0) o1->1815ede3-8f0f-490c-9f90-6f2bb8c847ec@0@lo:335/0 lens 440/0 e 1 to 0 dl 1630535330 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'lfs.0' LustreError: 19866:0:(service.c:2290:ptlrpc_server_handle_request()) Skipped 68 previous similar messages Lustre: 15500:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1630535276/real 1630535276] req@ffff880268508040 x1709739782592320/t0(0) o101->lustre-OST0000-osc-ffff88029a49c138@0@lo:28/4 lens 328/400 e 1 to 1 dl 1630535364 ref 2 fl Rpc:XQr/0/ffffffff rc 0/-1 job:'dir_create.sh.0' Lustre: 15500:0:(client.c:2285:ptlrpc_expire_one_request()) Skipped 127 previous similar messages Lustre: 3914:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x240000404:0xd9:0x0]// may get corrupted (rc -108) LustreError: 15512:0:(vvp_io.c:1793:vvp_io_init()) lustre: refresh file layout [0x240000404:0xd9:0x0] error -108. LustreError: 15512:0:(vvp_io.c:1793:vvp_io_init()) Skipped 1 previous similar message Lustre: lustre-OST0002-osc-ffff880299bd4138: Connection restored to 192.168.123.215@tcp (at 0@lo) Lustre: 19867:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (69/34s); client may timeout req@ffff88027d83a5c0 x1709739781207232/t0(0) o1->1815ede3-8f0f-490c-9f90-6f2bb8c847ec@0@lo:336/0 lens 440/0 e 1 to 0 dl 1630535331 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'lfs.0' Lustre: 3913:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x200000404:0x40:0x0]/ may get corrupted (rc -108) Lustre: 3913:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x200000404:0x17a:0x0]// may get corrupted (rc -108) Lustre: 3913:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x200000403:0xee:0x0]/ may get corrupted (rc -108) Lustre: 19867:0:(service.c:2328:ptlrpc_server_handle_request()) Skipped 138 previous similar messages Lustre: 3914:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x200000403:0xba:0x0]/ may get corrupted (rc -108) Lustre: 3914:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x280000404:0x1c6:0x0]/ may get corrupted (rc -108) Lustre: 3914:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x240000404:0x207:0x0]// may get corrupted (rc -108) LustreError: 19865:0:(service.c:2290:ptlrpc_server_handle_request()) @@@ Dropping timed-out request from 12345-0@lo: deadline 69/34s ago req@ffff8802711557c0 x1709739781214208/t0(0) o1->1815ede3-8f0f-490c-9f90-6f2bb8c847ec@0@lo:336/0 lens 440/0 e 1 to 0 dl 1630535331 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'lfs.0' LustreError: 19865:0:(service.c:2290:ptlrpc_server_handle_request()) Skipped 154 previous similar messages Lustre: 3913:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x240000404:0x54:0x0]/ may get corrupted (rc -108) Lustre: 3914:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x280000404:0xb7:0x0]/ may get corrupted (rc -108) Lustre: 3914:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x240000403:0xe:0x0]/ may get corrupted (rc -108) Lustre: 3913:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x280000404:0x1a1:0x0]/ may get corrupted (rc -108) Lustre: 19927:0:(service.c:2328:ptlrpc_server_handle_request()) @@@ Request took longer than estimated (43/33s); client may timeout req@ffff88027c0625c0 x1709739783968448/t0(0) o101->1815ede3-8f0f-490c-9f90-6f2bb8c847ec@0@lo:341/0 lens 328/0 e 0 to 0 dl 1630535336 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'ln.0' Lustre: 19927:0:(service.c:2328:ptlrpc_server_handle_request()) Skipped 817 previous similar messages LustreError: 19869:0:(service.c:2290:ptlrpc_server_handle_request()) @@@ Dropping timed-out request from 12345-0@lo: deadline 43/32s ago req@ffff88007bfa1940 x1709739783976896/t0(0) o101->1815ede3-8f0f-490c-9f90-6f2bb8c847ec@0@lo:342/0 lens 328/0 e 0 to 0 dl 1630535337 ref 1 fl Interpret:/0/ffffffff rc 0/-1 job:'ln.0' LustreError: 19869:0:(service.c:2290:ptlrpc_server_handle_request()) Skipped 819 previous similar messages LustreError: 888:0:(ofd_obd.c:501:ofd_destroy_export()) lustre-OST0000: cli 25ea6364-a008-4682-b7ef-9cf7da42069f/ffff88008ee89bf8 has 28672 pending on destroyed export LustreError: 888:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_destroy_export: tot_granted 5152192 != fo_tot_granted 5180864 LustreError: 888:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_destroy_export: tot_pending 0 != fo_tot_pending 28672 LustreError: 167-0: lustre-OST0000-osc-ffff88029a49c138: This client was evicted by lustre-OST0000; in progress operations using this service will fail. LustreError: Skipped 2 previous similar messages LustreError: 12341:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 123592128 != fo_tot_granted 123620800 LustreError: 12341:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 1 previous similar message LustreError: 12341:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 28672 LustreError: 12341:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 1 previous similar message LustreError: 28578:0:(import.c:354:ptlrpc_invalidate_import()) lustre-OST0000_UUID: timeout waiting for callback (222 != 0) LustreError: 28578:0:(import.c:377:ptlrpc_invalidate_import()) @@@ still on sending list req@ffff8802659d4500 x1709739781145024/t0(0) o1->lustre-OST0000-osc-ffff88029a49c138@0@lo:28/4 lens 440/432 e 2 to 0 dl 1630535372 ref 1 fl Rpc:EQr/0/ffffffff rc -5/-1 job:'lfs.0' LustreError: 28578:0:(import.c:382:ptlrpc_invalidate_import()) @@@ still on delayed list req@ffff8802683a0040 x1709739781132992/t0(0) o1->lustre-OST0000-osc-ffff88029a49c138@0@lo:28/4 lens 440/432 e 1 to 1 dl 1630535349 ref 1 fl Rpc:EXQU/0/ffffffff rc -5/-1 job:'lfs.0' LustreError: 28578:0:(import.c:388:ptlrpc_invalidate_import()) lustre-OST0000_UUID: Unregistering RPCs found (0). Network is sluggish? Waiting for them to error out. LustreError: 12349:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 123591232 != fo_tot_granted 123619904 LustreError: 12349:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 2 previous similar messages LustreError: 12349:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 28672 LustreError: 12349:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 2 previous similar messages LustreError: 12336:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 13537216 != fo_tot_granted 13565888 LustreError: 12336:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 28672 LustreError: 28304:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0001-osc-ffff880299bd4138: namespace resource [0xa0:0x0:0x0].0x0 (ffff88007c2174c0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 28304:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 1288 previous similar messages Lustre: lustre-OST0001-osc-ffff880299bd4138: Connection restored to 192.168.123.215@tcp (at 0@lo) Lustre: 15892:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 7: before 508 < left 30153, rollback = 7 Lustre: 15892:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 11474 previous similar messages Lustre: 15892:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 701/2804/0, destroy: 1/4/1 Lustre: 15892:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 11572 previous similar messages Lustre: 15892:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 3/3/1, xattr_set: 705/148/0 Lustre: 15892:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 11572 previous similar messages Lustre: 15892:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 3506/30153/0, punch: 0/0/0, quota 10/74/0 Lustre: 15892:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 11572 previous similar messages Lustre: 15892:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 702/11933/0, delete: 2/5/1 Lustre: 15892:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 11572 previous similar messages Lustre: 15892:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 1/1/0, ref_del: 2/2/1 Lustre: 15892:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 11572 previous similar messages LustreError: 12349:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 123588544 != fo_tot_granted 123617216 LustreError: 12349:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 2 previous similar messages LustreError: 12349:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 28672 LustreError: 11-0: lustre-OST0002-osc-ffff88029a49c138: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 17 previous similar messages Lustre: lustre-OST0002-osc-ffff88029a49c138: Connection to lustre-OST0002 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 3 previous similar messages LustreError: 12349:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 2 previous similar messages LustreError: 167-0: lustre-OST0002-osc-ffff88029a49c138: This client was evicted by lustre-OST0002; in progress operations using this service will fail. Lustre: lustre-OST0002-osc-ffff88029a49c138: Connection restored to 192.168.123.215@tcp (at 0@lo) Lustre: Skipped 1 previous similar message LustreError: 28396:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0000-osc-ffff880299bd4138: namespace resource [0xa0:0x0:0x0].0x0 (ffff88007b9f65c0) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 28396:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 214 previous similar messages Lustre: lustre-OST0003-osc-ffff880299bd4138: Connection restored to 192.168.123.215@tcp (at 0@lo) LustreError: 12349:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 13536320 != fo_tot_granted 13564992 LustreError: 12349:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 5 previous similar messages LustreError: 12349:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 28672 LustreError: 12349:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 5 previous similar messages LustreError: 12342:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 139909440 != fo_tot_granted 139938112 LustreError: 12342:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 13 previous similar messages LustreError: 12342:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 28672 LustreError: 12342:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 13 previous similar messages Lustre: 16197:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0000: opcode 2: before 509 < left 983, rollback = 2 Lustre: 16197:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 64 previous similar messages LustreError: 10233:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0001: failed to get lu_attr of [0x240000404:0x1be:0x0]: rc = -2 LustreError: 20132:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0001: '15' migration was interrupted, run 'lfs migrate -m 0 -c 1 -H crush 15' to finish migration. LustreError: 12332:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 87527232 != fo_tot_granted 87555904 LustreError: 12332:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 29 previous similar messages LustreError: 12332:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 151552 != fo_tot_pending 180224 LustreError: 12332:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 29 previous similar messages Lustre: 3905:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff88026b199f80 x1709739794710912/t0(0) o10->lustre-OST0003-osc-ffff88029a49c138@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630535514 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dd.0' LustreError: 32595:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0000: '0' migration was interrupted, run 'lfs migrate -m 2 -c 1 -H crush 0' to finish migration. LustreError: 12342:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 270712384 != fo_tot_granted 270741056 LustreError: 12342:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 71 previous similar messages LustreError: 12342:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 122880 != fo_tot_pending 151552 LustreError: 12342:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 71 previous similar messages 11[3471]: segfault at 8 ip 00007f59f76bb7e8 sp 00007ffdd8e83e30 error 4 in ld-2.17.so[7f59f76b0000+22000] Lustre: 16069:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 7: before 512 < left 960, rollback = 7 Lustre: 16069:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 56741 previous similar messages Lustre: 16069:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 1/4/0, destroy: 0/0/0 Lustre: 16069:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 56912 previous similar messages Lustre: 16069:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 1/1/0, xattr_set: 4/209/0 Lustre: 16069:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 56912 previous similar messages Lustre: 16069:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 951/960/0, punch: 0/0/0, quota 8/80/0 Lustre: 16069:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 56912 previous similar messages Lustre: 16069:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 2/33/0, delete: 0/0/0 Lustre: 16069:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 56912 previous similar messages Lustre: 16069:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 0/0/0, ref_del: 0/0/0 Lustre: 16069:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 56912 previous similar messages Lustre: 16881:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000404:0x2e6:0x0] with magic=0xbd60bd0 Lustre: 3901:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff88029cd56a80 x1709739800909824/t0(0) o10->lustre-OST0000-osc-ffff88029a49c138@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630535535 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dir_create.sh.0' Lustre: 3901:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff88026177cb40 x1709739800992384/t0(0) o10->lustre-OST0002-osc-ffff880299bd4138@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630535572 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'dir_create.sh.0' 9[4602]: segfault at 8 ip 00007f74275d17e8 sp 00007fff954c07e0 error 4 in ld-2.17.so[7f74275c6000+22000] LustreError: 16110:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0002: '15' migration was interrupted, run 'lfs migrate -m 0 -c 1 -H crush 15' to finish migration. LustreError: 10090:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0001_UUID lock: ffff88025b6087c0/0xb332559473541216 lrc: 3/0,0 mode: PW/PW res: [0x300000401:0x6c1:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000400000020 nid: 0@lo remote: 0xb3325594735411d0 expref: 1934 pid: 19930 timeout: 821 lvb_type: 0 LustreError: 10090:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 15 previous similar messages LustreError: 11-0: lustre-OST0001-osc-ffff88029a49c138: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 1 previous similar message Lustre: lustre-OST0001-osc-ffff88029a49c138: Connection to lustre-OST0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete LustreError: 12890:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff8800850fb240 x1709739802372544/t0(0) o105->lustre-OST0001@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 12890:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 21 previous similar messages LustreError: 167-0: lustre-OST0001-osc-ffff88029a49c138: This client was evicted by lustre-OST0001; in progress operations using this service will fail. LustreError: 887:0:(ofd_obd.c:501:ofd_destroy_export()) lustre-OST0001: cli 1815ede3-8f0f-490c-9f90-6f2bb8c847ec/ffff880299ac2548 has 61440 pending on destroyed export LustreError: 887:0:(ofd_obd.c:501:ofd_destroy_export()) Skipped 1 previous similar message LustreError: 3979:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630535539 with bad export cookie 12912477177702680446 LustreError: 3979:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) Skipped 2 previous similar messages Lustre: 32595:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000404:0x2f7:0x0] with magic=0xbd60bd0 Lustre: 32595:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 3 previous similar messages Lustre: 3910:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x240000403:0x27d:0x0]// may get corrupted (rc -108) Lustre: 3906:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x280000403:0x27e:0x0]/ may get corrupted (rc -108) LustreError: 5813:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0002-osc-ffff88029a49c138: namespace resource [0x340000400:0x850:0x0].0x0 (ffff88026fb1d940) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 5813:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 1312 previous similar messages Lustre: lustre-OST0002-osc-ffff88029a49c138: Connection restored to 192.168.123.215@tcp (at 0@lo) Lustre: Skipped 2 previous similar messages Lustre: 3906:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x240000404:0x353:0x0]/ may get corrupted (rc -108) LustreError: 10090:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0002_UUID lock: ffff880265fa1300/0xb332559473566bb4 lrc: 3/0,0 mode: PW/PW res: [0x472:0x0:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000000030020 nid: 0@lo remote: 0xb332559473566ba6 expref: 2847 pid: 29009 timeout: 835 lvb_type: 0 LustreError: 10090:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 1 previous similar message LustreError: 20136:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88026ef4ac00 x1709739803455104/t0(0) o106->lustre-OST0002@0@lo:15/16 lens 328/280 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 20136:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 11 previous similar messages Lustre: dir [0x200000403:0x218:0x0] stripe 1 readdir failed: -2, directory is partially accessed! LustreError: 17357:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0001: failed to get lu_attr of [0x240000404:0x2f8:0x0]: rc = -2 Lustre: 16200:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0001: opcode 8: before 512 < left 13616, rollback = 9 Lustre: 16200:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 208 previous similar messages Lustre: 3902:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x200000404:0x390:0x0]// may get corrupted (rc -108) Lustre: 3901:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x280000404:0x3ba:0x0]/ may get corrupted (rc -108) LustreError: 16029:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0000: '4' migration was interrupted, run 'lfs migrate -m 2 -c 1 -H crush 4' to finish migration. Lustre: 17820:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000404:0x1a4:0x0] with magic=0xbd60bd0 Lustre: 17820:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 1 previous similar message LustreError: 20718:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000403:0x218:0x0]: rc = -2 Lustre: dir [0x280000404:0x3af:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 2 previous similar messages Lustre: 16784:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000404:0x236:0x0] with magic=0xbd60bd0 Lustre: 16784:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 3 previous similar messages Lustre: 32595:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0000-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x200000404:0x24d:0x0] with magic=0xbd60bd0 Lustre: 32595:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 1 previous similar message Lustre: dir [0x200000404:0x193:0x0] stripe 1 readdir failed: -2, directory is partially accessed! LustreError: 22000:0:(mdd_object.c:3461:mdd_close()) lustre-MDD0000: failed to get lu_attr of [0x200000404:0x193:0x0]: rc = -2 LustreError: 12345:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 155633728 != fo_tot_granted 155695168 LustreError: 12345:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 181 previous similar messages LustreError: 12345:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 896 != fo_tot_pending 62336 LustreError: 12345:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 181 previous similar messages LustreError: 17811:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0000: '17' migration was interrupted, run 'lfs migrate -m 1 -c 1 -H crush 17' to finish migration. Lustre: 6019:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0002-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x280000403:0x363:0x0] with magic=0xbd60bd0 Lustre: 6019:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 9 previous similar messages LustreError: 10090:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: filter-lustre-OST0000_UUID lock: ffff8802581b4f00/0xb332559473623dd6 lrc: 3/0,0 mode: PW/PW res: [0x494:0x0:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->18446744073709551615) gid 0 flags: 0x60000000030020 nid: 0@lo remote: 0xb332559473623dc8 expref: 2993 pid: 5948 timeout: 943 lvb_type: 0 LustreError: 13759:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff880078593240 x1709739812491648/t0(0) o105->lustre-OST0000@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 13759:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 1 previous similar message LustreError: 10074:0:(ldlm_lockd.c:2496:ldlm_cancel_handler()) ldlm_cancel from 0@lo arrived at 1630535660 with bad export cookie 12912477177702592001 LustreError: 11-0: lustre-OST0003-osc-ffff88029a49c138: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 26 previous similar messages Lustre: lustre-OST0003-osc-ffff88029a49c138: Connection to lustre-OST0003 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 2 previous similar messages LustreError: 167-0: lustre-OST0003-osc-ffff88029a49c138: This client was evicted by lustre-OST0003; in progress operations using this service will fail. LustreError: Skipped 2 previous similar messages LustreError: 3898:0:(osc_request.c:1035:osc_init_grant()) lustre-OST0000-osc-ffff88029a49c138: granted 8437760 but already consumed 11804672 Lustre: 3901:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x240000403:0x306:0x0]/ may get corrupted (rc -108) Lustre: 3902:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x200000403:0x449:0x0]/ may get corrupted (rc -108) Lustre: 3902:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x200000403:0x4b9:0x0]/ may get corrupted (rc -108) LustreError: 669:0:(ofd_obd.c:501:ofd_destroy_export()) lustre-OST0003: cli 1815ede3-8f0f-490c-9f90-6f2bb8c847ec/ffff880299b2efc8 has 61440 pending on destroyed export Lustre: 3902:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x200000404:0x404:0x0]// may get corrupted (rc -108) Lustre: 3902:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x240000403:0x310:0x0]// may get corrupted (rc -108) Lustre: 3899:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x200000403:0x408:0x0]// may get corrupted (rc -108) Lustre: 3899:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x200000404:0x423:0x0]// may get corrupted (rc -108) Lustre: 3899:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x240000403:0x35e:0x0]/ may get corrupted (rc -108) Lustre: 3899:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x200000403:0x453:0x0]// may get corrupted (rc -108) Lustre: 3901:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x280000403:0x312:0x0]/ may get corrupted (rc -108) LustreError: 10762:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0003-osc-ffff88029a49c138: namespace resource [0x4a0:0x0:0x0].0x0 (ffff88025bd72d40) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 10762:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 216 previous similar messages Lustre: lustre-OST0003-osc-ffff88029a49c138: Connection restored to 192.168.123.215@tcp (at 0@lo) Lustre: Skipped 3 previous similar messages INFO: task mv:32690 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mv D ffff88027c800010 11344 32690 15509 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff81251fe1>] lock_rename+0x31/0xe0 [<ffffffff8125862f>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed432>] ? handle_mm_fault+0xc2/0x150 [<ffffffff8125975e>] SyS_renameat2+0xe/0x10 [<ffffffff81259771>] SyS_renameat+0x11/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 LustreError: 15908:0:(mdd_dir.c:4121:mdd_migrate_cmd_check()) lustre-MDD0002: '15' migration was interrupted, run 'lfs migrate -m 0 -c 1 -H crush 15' to finish migration. Lustre: dir [0x240000403:0x37f:0x0] stripe 1 readdir failed: -2, directory is partially accessed! Lustre: Skipped 10 previous similar messages Lustre: 10746:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1630535663/real 1630535663] req@ffff88026a838680 x1709739812636224/t0(0) o101->lustre-MDT0000-mdc-ffff88029a49c138@0@lo:12/10 lens 576/28064 e 0 to 1 dl 1630535721 ref 2 fl Rpc:XQr/2/ffffffff rc 0/-1 job:'rm.0' Lustre: lustre-MDT0000-mdc-ffff88029a49c138: Connection to lustre-MDT0000 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 1 previous similar message Lustre: lustre-MDT0000: Client 1815ede3-8f0f-490c-9f90-6f2bb8c847ec (at 0@lo) reconnecting Lustre: lustre-MDT0000-mdc-ffff88029a49c138: Connection restored to 192.168.123.215@tcp (at 0@lo) ptlrpc_watchdog_fire: 40 callbacks suppressed Lustre: mdt07_006: service thread pid 16876 was inactive for 64.147 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Lustre: Skipped 2 previous similar messages Pid: 16876, comm: mdt07_006 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] mdt_object_local_lock+0x527/0xb90 [mdt] [<0>] mdt_object_lock_internal+0x70/0x390 [mdt] [<0>] mdt_object_find_lock+0x6a/0x1a0 [mdt] [<0>] mdt_reint_setxattr+0x1fa/0x1110 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x76c/0xb40 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: mdt05_009: service thread pid 20132 was inactive for 62.045 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 20132, comm: mdt05_009 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] mdt_object_local_lock+0x527/0xb90 [mdt] [<0>] mdt_object_lock_internal+0x70/0x390 [mdt] [<0>] mdt_object_lock_try+0x27/0xb0 [mdt] [<0>] mdt_getattr_name_lock+0x17ae/0x2590 [mdt] [<0>] mdt_intent_getattr+0x2d5/0x4b0 [mdt] [<0>] mdt_intent_opc+0x1dd/0xc10 [mdt] [<0>] mdt_intent_policy+0x1a1/0x360 [mdt] [<0>] ldlm_lock_enqueue+0x3c2/0xb40 [ptlrpc] [<0>] ldlm_handle_enqueue0+0xaa6/0x16d0 [ptlrpc] [<0>] tgt_enqueue+0x64/0x240 [ptlrpc] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: 10635:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1630535657/real 1630535657] req@ffff880261754500 x1709739812251008/t0(0) o101->lustre-MDT0001-mdc-ffff88029a49c138@0@lo:12/10 lens 576/17904 e 0 to 1 dl 1630535745 ref 2 fl Rpc:XQr/2/ffffffff rc 0/-1 job:'setfattr.0' Lustre: lustre-MDT0001: Client 1815ede3-8f0f-490c-9f90-6f2bb8c847ec (at 0@lo) reconnecting Lustre: mdt04_003: service thread pid 15908 was inactive for 62.070 seconds. The thread might be hung, or it might only be slow and will resume later. Dumping the stack trace for debugging purposes: Pid: 15908, comm: mdt04_003 3.10.0-7.9-debug #2 SMP Wed Aug 4 02:24:01 EDT 2021 Call Trace: [<0>] ldlm_completion_ast+0x7a7/0xa00 [ptlrpc] [<0>] ldlm_cli_enqueue_local+0x259/0x880 [ptlrpc] [<0>] mdt_object_local_lock+0x527/0xb90 [mdt] [<0>] mdt_object_lock_internal+0x70/0x390 [mdt] [<0>] mdt_reint_object_lock+0x2c/0x60 [mdt] [<0>] mdt_reint_striped_lock+0x89/0x5b0 [mdt] [<0>] mdt_attr_set+0x9c/0x800 [mdt] [<0>] mdt_reint_setattr+0x77a/0x1020 [mdt] [<0>] mdt_reint_rec+0x87/0x240 [mdt] [<0>] mdt_reint_internal+0x76c/0xb40 [mdt] [<0>] mdt_reint+0x67/0x150 [mdt] [<0>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<0>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<0>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<0>] kthread+0xe4/0xf0 [<0>] ret_from_fork_nospec_begin+0x7/0x21 [<0>] 0xfffffffffffffffe Lustre: 18759:0:(lod_lov.c:1225:lod_parse_striping()) lustre-MDT0001-mdtlov: EXTENSION flags=40 set on component[2]=1 of non-SEL file [0x240000404:0x4d1:0x0] with magic=0xbd60bd0 Lustre: 18759:0:(lod_lov.c:1225:lod_parse_striping()) Skipped 1 previous similar message Lustre: 15870:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 7: before 510 < left 44601, rollback = 7 Lustre: 15870:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 282681 previous similar messages Lustre: 15870:0:(osd_handler.c:1867:osd_trans_dump_creds()) create: 1037/4148/0, destroy: 1/4/1 Lustre: 15870:0:(osd_handler.c:1867:osd_trans_dump_creds()) Skipped 282958 previous similar messages Lustre: 15870:0:(osd_handler.c:1874:osd_trans_dump_creds()) attr_set: 3/3/0, xattr_set: 1041/148/0 Lustre: 15870:0:(osd_handler.c:1874:osd_trans_dump_creds()) Skipped 282958 previous similar messages Lustre: 15870:0:(osd_handler.c:1884:osd_trans_dump_creds()) write: 5186/44601/0, punch: 0/0/0, quota 10/90/0 Lustre: 15870:0:(osd_handler.c:1884:osd_trans_dump_creds()) Skipped 282953 previous similar messages Lustre: 15870:0:(osd_handler.c:1891:osd_trans_dump_creds()) insert: 1038/17645/0, delete: 2/5/0 Lustre: 15870:0:(osd_handler.c:1891:osd_trans_dump_creds()) Skipped 282957 previous similar messages Lustre: 15870:0:(osd_handler.c:1898:osd_trans_dump_creds()) ref_add: 1/1/0, ref_del: 2/2/1 Lustre: 15870:0:(osd_handler.c:1898:osd_trans_dump_creds()) Skipped 282960 previous similar messages LustreError: 14276:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff88029a49c138: inode [0x240000403:0x272:0x0] mdc close failed: rc = -13 Lustre: 3912:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff88028ca2b240 x1709739824306432/t0(0) o10->lustre-OST0002-osc-ffff88029a49c138@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630535790 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'cp.0' Lustre: dir [0x200000403:0x5de:0x0] stripe 1 readdir failed: -2, directory is partially accessed! LustreError: 14745:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff880299bd4138: inode [0x240000403:0x272:0x0] mdc close failed: rc = -13 INFO: task mv:3638 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mv D ffff88008a6ea140 10944 3638 15570 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff81251fe1>] lock_rename+0x31/0xe0 [<ffffffff8125862f>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed432>] ? handle_mm_fault+0xc2/0x150 [<ffffffff8125975e>] SyS_renameat2+0xe/0x10 [<ffffffff81259771>] SyS_renameat+0x11/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task mv:5170 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mv D ffff88027c5fc9d0 11168 5170 15414 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff81251fe1>] lock_rename+0x31/0xe0 [<ffffffff8125862f>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed432>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817e8a76>] ? trace_do_page_fault+0x56/0x170 [<ffffffff8125975e>] SyS_renameat2+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task mv:5374 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mv D ffff880292bf0010 11040 5374 15601 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff81251fe1>] lock_rename+0x31/0xe0 [<ffffffff8125862f>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed432>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817e8a76>] ? trace_do_page_fault+0x56/0x170 [<ffffffff8125975e>] SyS_renameat2+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task mv:6664 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mv D ffff8800859db760 10608 6664 15446 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff81251fe1>] lock_rename+0x31/0xe0 [<ffffffff8125862f>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed432>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817e8a76>] ? trace_do_page_fault+0x56/0x170 [<ffffffff8125975e>] SyS_renameat2+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task mv:7549 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mv D ffff880264df3760 11216 7549 15437 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff81251fe1>] lock_rename+0x31/0xe0 [<ffffffff8125862f>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed432>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817e8a76>] ? trace_do_page_fault+0x56/0x170 [<ffffffff8125975e>] SyS_renameat2+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task mv:7725 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mv D ffff880295380040 11344 7725 15443 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff81251fe1>] lock_rename+0x31/0xe0 [<ffffffff8125862f>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed432>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817e8a76>] ? trace_do_page_fault+0x56/0x170 [<ffffffff8125975e>] SyS_renameat2+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task mv:8897 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mv D ffff880079e949d0 11216 8897 15409 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff81251fe1>] lock_rename+0x31/0xe0 [<ffffffff8125862f>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed432>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817e8a76>] ? trace_do_page_fault+0x56/0x170 [<ffffffff8125975e>] SyS_renameat2+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task mv:8993 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. mv D ffff88025a47c9d0 11216 8993 16020 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff81251fe1>] lock_rename+0x31/0xe0 [<ffffffff8125862f>] SYSC_renameat2+0x22f/0x570 [<ffffffff811ed432>] ? handle_mm_fault+0xc2/0x150 [<ffffffff817e8a76>] ? trace_do_page_fault+0x56/0x170 [<ffffffff8125975e>] SyS_renameat2+0xe/0x10 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 INFO: task ls:9120 blocked for more than 120 seconds. "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. ls D ffff880287830040 11232 9120 15764 0x00000080 Call Trace: [<ffffffff817e19f9>] schedule_preempt_disabled+0x39/0x90 [<ffffffff817df80a>] __mutex_lock_slowpath+0x13a/0x340 [<ffffffff817dfa3d>] mutex_lock+0x2d/0x40 [<ffffffff812550a6>] do_last+0x296/0x1280 [<ffffffff812546de>] ? link_path_walk+0x27e/0x8c0 [<ffffffff812570ad>] path_openat+0xcd/0x5b0 [<ffffffff81258dfd>] do_filp_open+0x4d/0xb0 [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e326e>] ? _raw_spin_unlock+0xe/0x20 [<ffffffff81267893>] ? __alloc_fd+0xc3/0x170 [<ffffffff812447e4>] do_sys_open+0x124/0x220 [<ffffffff81244914>] SyS_openat+0x14/0x20 [<ffffffff817ee00c>] system_call_fastpath+0x1f/0x24 LustreError: 10090:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 100s: evicting client at 0@lo ns: mdt-lustre-MDT0001_UUID lock: ffff88028bd56940/0xb332559473724c2d lrc: 3/0,0 mode: PR/PR res: [0x240000403:0x390:0x0].0x0 bits 0x12/0x0 rrc: 5 type: IBT gid 0 flags: 0x60200400000020 nid: 0@lo remote: 0xb332559473724c0a expref: 115 pid: 10215 timeout: 1083 lvb_type: 0 LustreError: 10090:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 1 previous similar message LustreError: 10193:0:(ldlm_lockd.c:1425:ldlm_handle_enqueue0()) ### lock on destroyed export ffff8802998dca88 ns: mdt-lustre-MDT0001_UUID lock: ffff8802617352c0/0xb3325594737cecfd lrc: 3/0,0 mode: PR/PR res: [0x240000403:0x48d:0x0].0x0 bits 0x13/0x0 rrc: 2 type: IBT gid 0 flags: 0x50200000000000 nid: 0@lo remote: 0xb3325594737cecb7 expref: 115 pid: 10193 timeout: 0 lvb_type: 0 LustreError: 11-0: lustre-MDT0001-mdc-ffff88029a49c138: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 1 previous similar message Lustre: lustre-MDT0001-mdc-ffff88029a49c138: Connection to lustre-MDT0001 (at 0@lo) was lost; in progress operations using this service will wait for recovery to complete Lustre: Skipped 1 previous similar message LustreError: 167-0: lustre-MDT0001-mdc-ffff88029a49c138: This client was evicted by lustre-MDT0001; in progress operations using this service will fail. LustreError: Skipped 1 previous similar message LustreError: 13968:0:(file.c:234:ll_close_inode_openhandle()) lustre-clilmv-ffff88029a49c138: inode [0x240000403:0x42b:0x0] mdc close failed: rc = -108 LustreError: 11980:0:(llite_lib.c:2983:ll_prep_inode()) new_inode -fatal: rc -5 Lustre: lustre-MDT0001-mdc-ffff88029a49c138: Connection restored to 192.168.123.215@tcp (at 0@lo) Lustre: Skipped 1 previous similar message Lustre: 3906:0:(client.c:1485:after_reply()) @@@ resending request on EINPROGRESS req@ffff880281fe0cc0 x1709739827579264/t0(0) o10->lustre-OST0003-osc-ffff880299bd4138@0@lo:6/4 lens 440/432 e 0 to 0 dl 1630535865 ref 1 fl Rpc:RQU/2/0 rc 0/-115 job:'cp.0' Lustre: 13398:0:(client.c:2285:ptlrpc_expire_one_request()) @@@ Request sent has timed out for slow reply: [sent 1630535751/real 1630535751] req@ffff880275f9c500 x1709739822156544/t0(0) o101->lustre-MDT0000-mdc-ffff88029a49c138@0@lo:12/10 lens 576/13016 e 0 to 1 dl 1630535828 ref 2 fl Rpc:XQr/2/ffffffff rc 0/-1 job:'rm.0' Lustre: lustre-MDT0000: Client 1815ede3-8f0f-490c-9f90-6f2bb8c847ec (at 0@lo) reconnecting LustreError: 10089:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff880272250cc0 x1709739829483712/t0(0) o105->lustre-OST0002@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 10089:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 66 previous similar messages LustreError: 16656:0:(ldlm_resource.c:1124:ldlm_resource_complain()) lustre-OST0002-osc-ffff880299bd4138: namespace resource [0x2170:0x0:0x0].0x0 (ffff880261711e40) refcount nonzero (1) after lock cleanup; forcing cleanup. LustreError: 16656:0:(ldlm_resource.c:1124:ldlm_resource_complain()) Skipped 138 previous similar messages Lustre: 16929:0:(osd_internal.h:1328:osd_trans_exec_op()) lustre-MDT0002: opcode 2: before 501 < left 1027, rollback = 2 Lustre: 16929:0:(osd_internal.h:1328:osd_trans_exec_op()) Skipped 283 previous similar messages LustreError: 11141:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0000-osp-MDT0001: fail to cancel 1 llog-records: rc = -116 LustreError: 11141:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0000-osp-MDT0001: fail to cancel 1 of 1 llog-records: rc = -116 LustreError: 11141:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-MDT0000-osp-MDT0001: fail to cancel 1 llog-records: rc = -116 LustreError: 11141:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-MDT0000-osp-MDT0001: fail to cancel 1 of 1 llog-records: rc = -116 Lustre: 7165:0:(service.c:2157:ptlrpc_server_handle_req_in()) @@@ Slow req_in handling 6s req@ffff880264cc2ee8 x1709739830932160/t0(0) o106->LOV_OSC_UUID@0@lo:0/0 lens 328/0 e 0 to 0 dl 0 ref 1 fl New:/0/ffffffff rc 0/-1 job:'' LustreError: 12343:0:(tgt_grant.c:248:tgt_grant_sanity_check()) ofd_statfs: tot_granted 283326144 != fo_tot_granted 283354816 LustreError: 12343:0:(tgt_grant.c:248:tgt_grant_sanity_check()) Skipped 344 previous similar messages LustreError: 12343:0:(tgt_grant.c:251:tgt_grant_sanity_check()) ofd_statfs: tot_pending 0 != fo_tot_pending 28672 LustreError: 12343:0:(tgt_grant.c:251:tgt_grant_sanity_check()) Skipped 344 previous similar messages LustreError: 10090:0:(ldlm_lockd.c:259:expired_lock_main()) ### lock callback timer expired after 101s: evicting client at 0@lo ns: filter-lustre-OST0000_UUID lock: ffff88025bd0a5c0/0xb332559473762e65 lrc: 3/0,0 mode: PW/PW res: [0x2116:0x0:0x0].0x0 rrc: 3 type: EXT [0->18446744073709551615] (req 0->36863) gid 0 flags: 0x60000400030020 nid: 0@lo remote: 0xb332559473762e5e expref: 3838 pid: 17920 timeout: 1191 lvb_type: 0 LustreError: 10090:0:(ldlm_lockd.c:259:expired_lock_main()) Skipped 2 previous similar messages LustreError: 12890:0:(client.c:1256:ptlrpc_import_delay_req()) @@@ IMP_CLOSED req@ffff88028a829300 x1709739834945728/t0(0) o105->lustre-OST0000@0@lo:15/16 lens 392/224 e 0 to 0 dl 0 ref 1 fl Rpc:QU/0/ffffffff rc 0/-1 job:'' LustreError: 12890:0:(client.c:1256:ptlrpc_import_delay_req()) Skipped 11 previous similar messages LustreError: 11-0: lustre-OST0000-osc-ffff88029a49c138: operation ldlm_enqueue to node 0@lo failed: rc = -107 LustreError: Skipped 2 previous similar messages Lustre: 3907:0:(llite_lib.c:3298:ll_dirty_page_discard_warn()) lustre: dirty page discard: 192.168.123.215@tcp:/lustre/fid: [0x240000404:0x54f:0x0]// may get corrupted (rc -108) ------------[ cut here ]------------ WARNING: CPU: 8 PID: 32595 at /home/green/git/lustre-release/ldiskfs/ext4_jbd2.c:266 __ldiskfs_handle_dirty_metadata+0x1d2/0x230 [ldiskfs] Modules linked in: loop zfs(PO) zunicode(PO) zzstd(O) zlua(O) zcommon(PO) znvpair(PO) zavl(PO) icp(PO) spl(O) lustre(OE) ofd(OE) osp(OE) lod(OE) ost(OE) mdt(OE) mdd(OE) mgs(OE) osd_ldiskfs(OE) ldiskfs(OE) jbd2 mbcache lquota(OE) lfsck(OE) obdecho(OE) mgc(OE) mdc(OE) lov(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) dm_flakey dm_mod libcfs(OE) crc_t10dif crct10dif_generic sb_edac edac_core iosf_mbi crc32_pclmul ghash_clmulni_intel aesni_intel lrw gf128mul glue_helper ablk_helper cryptd virtio_console virtio_balloon pcspkr i2c_piix4 ip_tables rpcsec_gss_krb5 drm_kms_helper ata_generic ttm pata_acpi drm crct10dif_pclmul crct10dif_common drm_panel_orientation_quirks ata_piix crc32c_intel serio_raw virtio_blk i2c_core libata floppy CPU: 8 PID: 32595 Comm: mdt04_011 Kdump: loaded Tainted: P W OE ------------ 3.10.0-7.9-debug #2 Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011 Call Trace: [<ffffffff817d9418>] dump_stack+0x19/0x1b [<ffffffff8108d538>] __warn+0xd8/0x100 [<ffffffff8108d67d>] warn_slowpath_null+0x1d/0x20 [<ffffffffa0bfa262>] __ldiskfs_handle_dirty_metadata+0x1d2/0x230 [ldiskfs] [<ffffffffa0bda33d>] ldiskfs_getblk+0x13d/0x210 [ldiskfs] [<ffffffffa0bda437>] ldiskfs_bread+0x27/0xe0 [ldiskfs] [<ffffffffa0c9ee97>] osd_write+0x4e7/0xcb0 [osd_ldiskfs] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffffa04d26f3>] dt_record_write+0x33/0x120 [obdclass] [<ffffffffa04916b2>] llog_osd_write_rec+0x1052/0x19a0 [obdclass] [<ffffffffa047e760>] llog_write_rec+0x290/0x590 [obdclass] [<ffffffffa04844ee>] llog_cat_add_rec+0x23e/0x950 [obdclass] [<ffffffffa047b5ff>] llog_add+0x17f/0x1f0 [obdclass] [<ffffffffa04aca39>] ? lprocfs_counter_add+0xf9/0x160 [obdclass] [<ffffffffa081560d>] sub_updates_write+0xb6b/0xe32 [ptlrpc] [<ffffffffa07e5892>] top_trans_stop+0x7c2/0xf60 [ptlrpc] [<ffffffffa0ed9f3c>] lod_trans_stop+0x25c/0x340 [lod] [<ffffffffa0dae408>] mdd_trans_stop+0x28/0x16e [mdd] [<ffffffffa0d8fb2f>] mdd_migrate_object+0xf1f/0x1ba0 [mdd] [<ffffffffa04d1375>] ? lu_object_find_at+0x285/0xb10 [obdclass] [<ffffffffa0d90b86>] mdd_migrate+0x3d6/0x8e0 [mdd] [<ffffffffa0e620d8>] mdo_migrate+0x4c/0x4e [mdt] [<ffffffffa0e21e53>] mdt_reint_migrate+0x1023/0x11d0 [mdt] [<ffffffffa0e22087>] mdt_reint_rec+0x87/0x240 [mdt] [<ffffffffa0df755c>] mdt_reint_internal+0x76c/0xb40 [mdt] [<ffffffffa0e04657>] mdt_reint+0x67/0x150 [mdt] [<ffffffffa07d426e>] tgt_request_handle+0x83e/0x1870 [ptlrpc] [<ffffffffa03e345e>] ? libcfs_nid2str_r+0xfe/0x130 [lnet] [<ffffffffa077f990>] ptlrpc_server_handle_request+0x250/0xb10 [ptlrpc] [<ffffffffa0781529>] ptlrpc_main+0xbd9/0x15f0 [ptlrpc] [<ffffffff81411979>] ? do_raw_spin_unlock+0x49/0x90 [<ffffffff817e32be>] ? _raw_spin_unlock_irq+0xe/0x30 [<ffffffffa0780950>] ? ptlrpc_wait_event+0x620/0x620 [ptlrpc] [<ffffffff810ba0f4>] kthread+0xe4/0xf0 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 [<ffffffff817ede5d>] ret_from_fork_nospec_begin+0x7/0x21 [<ffffffff810ba010>] ? kthread_create_on_node+0x140/0x140 ---[ end trace 00015c15de20e74e ]--- LDISKFS-fs: ldiskfs_getblk:888: aborting transaction: error 28 in __ldiskfs_handle_dirty_metadata LDISKFS-fs error (device dm-2): ldiskfs_getblk:888: inode #171: block 51515: comm mdt04_011: journal_dirty_metadata failed: handle type 0 started at line 1972, credits 512/0, errcode -28 Aborting journal on device dm-2-8. LDISKFS-fs (dm-2): Remounting filesystem read-only LustreError: 32595:0:(osd_io.c:2103:osd_ldiskfs_write_record()) lustre-MDT0002: error reading offset 344064 (block 84, size 24, offs 344032), credits 0/0: rc = -28 LustreError: 32595:0:(update_trans.c:985:top_trans_stop()) lustre-MDT0002-osd: write updates failed: rc = -28 LustreError: 32595:0:(osd_io.c:2110:osd_ldiskfs_write_record()) journal_get_write_access() returned error -30 LustreError: 32595:0:(tgt_lastrcvd.c:1253:tgt_add_reply_data()) lustre-MDT0002: can't update reply_data file: rc = -30 LustreError: 32595:0:(osd_handler.c:2075:osd_trans_stop()) lustre-MDT0002: failed in transaction hook: rc = -30 LDISKFS-fs error (device dm-2) in osd_trans_stop:2082: error 28 LustreError: 14047:0:(llog_cat.c:751:llog_cat_cancel_arr_rec()) lustre-OST0003-osc-MDT0002: fail to cancel 40 llog-records: rc = -30 LustreError: 14047:0:(osp_sync.c:1094:osp_sync_process_committed()) lustre-OST0003-osc-MDT0002: can't cancel 40 records: rc = -30 LustreError: 19311:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -30 LustreError: 11734:0:(osd_handler.c:1779:osd_trans_commit_cb()) transaction @0xffff880276141728 commit error: 2 LDISKFS-fs error (device dm-2) in osd_trans_stop:2082: IO failure LustreError: 32595:0:(osd_handler.c:2085:osd_trans_stop()) lustre-MDT0002: failed to stop transaction: rc = -28 LustreError: 32595:0:(update_trans.c:1011:top_trans_stop()) lustre-MDT0002-osd: stop trans failed: rc = -30 LustreError: 14047:0:(llog_cat.c:787:llog_cat_cancel_records()) lustre-OST0003-osc-MDT0002: fail to cancel 1 of 1 llog-records: rc = -30 LustreError: 14047:0:(osp_sync.c:1079:osp_sync_process_committed()) lustre-OST0003-osc-MDT0002: can't cancel record: rc = -30 LustreError: 19498:0:(llite_lib.c:1836:ll_md_setattr()) md_setattr fails: rc = -5 LustreError: 19498:0:(llite_lib.c:1836:ll_md_setattr()) Skipped 9 previous similar messages | Externally reported by onyx-68 boilpot email |
sanity test 230b: migrate directory | BUG: unable to handle kernel NULL pointer dereference at 0000000000000008 IP: [<ffffffffc0da6f8d>] ll_migrate+0x8fd/0xdb0 [lustre] PGD 8000000055f2e067 PUD 5dcaf067 PMD 0 Oops: 0000 [#1] SMP Modules linked in: brd lustre(OE) obdecho(OE) mgc(OE) lov(OE) mdc(OE) osc(OE) lmv(OE) fid(OE) fld(OE) ptlrpc_gss(OE) ptlrpc(OE) obdclass(OE) ksocklnd(OE) lnet(OE) libcfs(OE) loop rpcsec_gss_krb5 auth_rpcgss nfsv4 dns_resolver nfs lockd grace fscache rpcrdma ib_isert iscsi_target_mod ib_iser libiscsi scsi_transport_iscsi ib_srpt target_core_mod crc_t10dif crct10dif_generic ib_srp scsi_transport_srp scsi_tgt ib_ipoib rdma_ucm ib_ucm ib_uverbs ib_umad rdma_cm ib_cm iw_cm ib_core sunrpc iosf_mbi crc32_pclmul ghash_clmulni_intel ppdev aesni_intel lrw parport_pc gf128mul glue_helper ablk_helper pcspkr cryptd joydev parport virtio_balloon i2c_piix4 ip_tables ext4 mbcache jbd2 ata_generic pata_acpi virtio_blk ata_piix 8139too libata crct10dif_pclmul crct10dif_common virtio_pci crc32c_intel 8139cp virtio_ring serio_raw virtio mii floppy [last unloaded: libcfs] CPU: 1 PID: 22477 Comm: lfs Kdump: loaded Tainted: G OE ------------ 3.10.0-957.27.2.el7.x86_64 #1 Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011 task: ffff9e18a6afe180 ti: ffff9e18a74cc000 task.ti: ffff9e18a74cc000 RIP: 0010:[<ffffffffc0da6f8d>] [<ffffffffc0da6f8d>] ll_migrate+0x8fd/0xdb0 [lustre] RSP: 0018:ffff9e18a74cfbc0 EFLAGS: 00010206 RAX: 0000000000000000 RBX: 0000000000000000 RCX: 0000000000000000 RDX: 0000000000000000 RSI: ffff9e189f728ef0 RDI: ffff9e189bdb4070 RBP: ffff9e18a74cfc38 R08: ffff9e189c80ea90 R09: 0000000000000001 R10: 0000000000000000 R11: 0000000000000000 R12: ffff9e18a7d02910 R13: ffff9e189ff4d800 R14: ffff9e18a7132e40 R15: ffff9e189d72bb80 FS: 00007f977d69e740(0000) GS:ffff9e18bfd00000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 CR2: 0000000000000008 CR3: 000000000e61e000 CR4: 00000000000606e0 Call Trace: [<ffffffffb243f69a>] ? __check_object_size+0x1ca/0x250 [<ffffffffc0d957e4>] ll_dir_ioctl+0x61f4/0x71f0 [lustre] [<ffffffffc07be863>] ? fld_client_lookup+0x73/0x4a0 [fld] [<ffffffffc0913351>] ? lprocfs_counter_sub+0xc1/0x130 [obdclass] [<ffffffffc0913351>] ? lprocfs_counter_sub+0xc1/0x130 [obdclass] [<ffffffffc0b05d40>] ? ptlrpc_request_cache_free+0x90/0x1d0 [ptlrpc] [<ffffffffc0b07005>] ? __ptlrpc_req_finished+0x515/0x790 [ptlrpc] [<ffffffffc0b07290>] ? ptlrpc_req_finished+0x10/0x20 [ptlrpc] [<ffffffffc0d98f89>] ? ll_close_inode_openhandle+0x4a9/0xce0 [lustre] [<ffffffffc0913229>] ? lprocfs_counter_add+0xf9/0x160 [obdclass] [<ffffffffc0dc8428>] ? ll_stats_ops_tally+0x98/0x100 [lustre] [<ffffffffb235506d>] ? call_rcu_sched+0x1d/0x20 [<ffffffffb245a85f>] ? d_free+0x4f/0x70 [<ffffffffb245b1b8>] ? __dentry_kill+0x128/0x180 [<ffffffffb2457840>] do_vfs_ioctl+0x3a0/0x5a0 [<ffffffffb2465794>] ? mntput+0x24/0x40 [<ffffffffb2444ad6>] ? __fput+0x186/0x260 [<ffffffffb2457ae1>] SyS_ioctl+0xa1/0xc0 [<ffffffffb2976d15>] ? system_call_after_swapgs+0xa2/0x146 [<ffffffffb2976ddb>] system_call_fastpath+0x22/0x27 [<ffffffffb2976d21>] ? system_call_after_swapgs+0xae/0x146 | Link to test |