Status Job ID Links Posted Started Updated
Runtime
Duration
In Waiting
Machine Teuthology Branch OS Type OS Version Description Nodes
pass 5841019 2021-01-29 18:18:58 2021-01-29 18:23:21 2021-01-29 20:07:21 1:44:00 1:25:11 0:18:49 gibba master rhel 8.3 fs/mixed-clients/{begin clusters/1a3s-mds-2c-client conf/{client mds mon osd} kclient-overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped} objectstore-ec/bluestore-bitmap overrides/{frag_enable osd-asserts whitelist_health whitelist_wrongly_marked_down} tasks/kernel_cfuse_workunits_dbench_iozone} 2
fail 5841020 2021-01-29 18:19:00 2021-01-29 18:31:42 2021-01-29 19:29:42 0:58:00 0:39:54 0:18:06 gibba master rhel 8.3 fs/workload/{begin clusters/1a5s-mds-1c-client-3node conf/{client mds mon osd} distro/{rhel_8} mount/kclient/{mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} objectstore-ec/bluestore-comp-ec-root omap_limit/10000 overrides/{frag_enable osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/3 scrub/yes tasks/{0-check-counter workunit/fs/misc}} 3
Failure Reason:

"2021-01-29T19:00:04.671843+0000 mds.b (mds.0) 14 : cluster [WRN] Scrub error on inode 0x10000000262 (/client.0/tmp/testdir/dir1) see mds.b log and `damage ls` output for details" in cluster log

pass 5841021 2021-01-29 18:19:01 2021-01-29 18:34:54 2021-01-29 19:28:54 0:54:00 0:41:03 0:12:57 gibba master rhel 8.3 fs/thrash/workloads/{begin clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{rhel_8} mount/kclient/{mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} msgr-failures/osd-mds-delay objectstore-ec/bluestore-comp-ec-root overrides/{frag_enable session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} ranks/3 tasks/{1-thrash/mon 2-workunit/suites/ffsb}} 2
pass 5841022 2021-01-29 18:19:03 2021-01-29 18:36:29 2021-01-29 20:10:30 1:34:01 0:50:31 0:43:30 gibba master rhel 8.3 fs/workload/{begin clusters/1a5s-mds-1c-client-3node conf/{client mds mon osd} distro/{centos_8} mount/kclient/{mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} objectstore-ec/bluestore-comp-ec-root omap_limit/10000 overrides/{frag_enable osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/5 scrub/yes tasks/{0-check-counter workunit/suites/ffsb}} 3
pass 5841023 2021-01-29 18:19:04 2021-01-29 18:38:05 2021-01-29 19:14:04 0:35:59 0:13:58 0:22:01 gibba master rhel 8.3 fs/thrash/workloads/{begin clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{ubuntu_latest} mount/kclient/{mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} msgr-failures/osd-mds-delay objectstore-ec/bluestore-comp-ec-root overrides/{frag_enable session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} ranks/5 tasks/{1-thrash/osd 2-workunit/suites/pjd}} 2
pass 5841024 2021-01-29 18:19:06 2021-01-29 18:50:31 2021-01-29 19:52:30 1:01:59 0:11:52 0:50:07 gibba master rhel 8.3 fs/workload/{begin clusters/1a5s-mds-1c-client-3node conf/{client mds mon osd} distro/{ubuntu_latest} mount/kclient/{mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} objectstore-ec/bluestore-comp-ec-root omap_limit/10000 overrides/{frag_enable osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/1 scrub/yes tasks/{0-check-counter workunit/suites/fsync-tester}} 3
pass 5841025 2021-01-29 18:19:07 2021-01-29 18:59:13 2021-01-29 19:51:12 0:51:59 0:27:19 0:24:40 gibba master rhel 8.3 fs/functional/{begin clusters/1a3s-mds-4c-client conf/{client mds mon osd} distro/{rhel_8} mount/kclient/{mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} objectstore/bluestore-ec-root overrides/{frag_enable no_client_pidfile whitelist_health whitelist_wrongly_marked_down} tasks/admin} 2
pass 5841026 2021-01-29 18:19:08 2021-01-29 19:06:52 2021-01-29 21:28:53 2:22:01 0:13:16 2:08:45 gibba master rhel 8.3 fs/workload/{begin clusters/1a5s-mds-1c-client-3node conf/{client mds mon osd} distro/{centos_8} mount/kclient/{mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} objectstore-ec/bluestore-comp-ec-root omap_limit/10000 overrides/{frag_enable osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/3 scrub/yes tasks/{0-check-counter workunit/suites/pjd}} 3
pass 5841027 2021-01-29 18:19:09 2021-01-29 19:14:08 2021-01-29 19:50:08 0:36:00 0:11:08 0:24:52 gibba master rhel 8.3 fs/thrash/workloads/{begin clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{rhel_8} mount/kclient/{mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} msgr-failures/osd-mds-delay objectstore-ec/bluestore-comp-ec-root overrides/{frag_enable session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} ranks/1 tasks/{1-thrash/osd 2-workunit/fs/trivial_sync}} 2
fail 5841028 2021-01-29 18:19:11 2021-01-29 19:17:45 2021-01-30 01:07:51 5:50:06 1:49:13 4:00:53 gibba master rhel 8.3 fs/workload/{begin clusters/1a5s-mds-1c-client-3node conf/{client mds mon osd} distro/{centos_8} mount/kclient/{mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} objectstore-ec/bluestore-comp-ec-root omap_limit/10000 overrides/{frag_enable osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/5 scrub/yes tasks/{0-check-counter workunit/suites/blogbench}} 3
Failure Reason:

"2021-01-29T23:28:45.313641+0000 mds.f (mds.0) 14 : cluster [WRN] Scrub error on inode 0x100000001fa (/client.0/tmp/blogbench-1.0/man) see mds.f log and `damage ls` output for details" in cluster log

fail 5841029 2021-01-29 18:19:12 2021-01-29 19:29:21 2021-01-29 20:43:21 1:14:00 0:42:40 0:31:20 gibba master rhel 8.3 fs/thrash/workloads/{begin clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{ubuntu_latest} mount/kclient/{mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} msgr-failures/osd-mds-delay objectstore-ec/bluestore-comp-ec-root overrides/{frag_enable session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} ranks/3 tasks/{1-thrash/osd 2-workunit/suites/ffsb}} 2
Failure Reason:

SELinux denials found on ubuntu@gibba019.front.sepia.ceph.com: ['type=AVC msg=audit(1611950992.325:9102): avc: denied { create } for pid=15030 comm="cachefilesd" name="1611950992.15030.core" scontext=system_u:system_r:cachefilesd_t:s0 tcontext=system_u:object_r:user_home_t:s0 tclass=file permissive=1', 'type=AVC msg=audit(1611950992.325:9101): avc: denied { add_name } for pid=15030 comm="cachefilesd" name="1611950992.15030.core" scontext=system_u:system_r:cachefilesd_t:s0 tcontext=unconfined_u:object_r:user_home_t:s0 tclass=dir permissive=1', 'type=AVC msg=audit(1611950992.325:9098): avc: denied { search } for pid=15030 comm="cachefilesd" name="ubuntu" dev="sda1" ino=2387 scontext=system_u:system_r:cachefilesd_t:s0 tcontext=unconfined_u:object_r:user_home_dir_t:s0 tclass=dir permissive=1', 'type=AVC msg=audit(1611950992.325:9100): avc: denied { write } for pid=15030 comm="cachefilesd" name="coredump" dev="sda1" ino=524299 scontext=system_u:system_r:cachefilesd_t:s0 tcontext=unconfined_u:object_r:user_home_t:s0 tclass=dir permissive=1', 'type=AVC msg=audit(1611950992.325:9103): avc: denied { read write open } for pid=15030 comm="cachefilesd" path="/home/ubuntu/cephtest/archive/coredump/1611950992.15030.core" dev="sda1" ino=524499 scontext=system_u:system_r:cachefilesd_t:s0 tcontext=system_u:object_r:user_home_t:s0 tclass=file permissive=1']

pass 5841030 2021-01-29 18:19:14 2021-01-29 19:29:46 2021-01-29 20:19:45 0:49:59 0:15:40 0:34:19 gibba master rhel 8.3 fs/workload/{begin clusters/1a5s-mds-1c-client-3node conf/{client mds mon osd} distro/{rhel_8} mount/kclient/{mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} objectstore-ec/bluestore-comp-ec-root omap_limit/10000 overrides/{frag_enable osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/1 scrub/yes tasks/{0-check-counter workunit/suites/fsstress}} 3
pass 5841031 2021-01-29 18:19:16 2021-01-29 19:36:23 2021-01-29 21:28:23 1:52:00 1:07:10 0:44:50 gibba master rhel 8.3 fs/workload/{begin clusters/1a5s-mds-1c-client-3node conf/{client mds mon osd} distro/{ubuntu_latest} mount/kclient/{mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} objectstore-ec/bluestore-comp-ec-root omap_limit/10000 overrides/{frag_enable osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/3 scrub/yes tasks/{0-check-counter workunit/suites/iogen}} 3
pass 5841032 2021-01-29 18:19:17 2021-01-29 19:50:28 2021-01-29 20:14:28 0:24:00 0:13:04 0:10:56 gibba master rhel 8.3 fs/thrash/workloads/{begin clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{centos_8} mount/kclient/{mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} msgr-failures/osd-mds-delay objectstore-ec/bluestore-comp-ec-root overrides/{frag_enable session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} ranks/5 tasks/{1-thrash/mds 2-workunit/suites/pjd}} 2
pass 5841033 2021-01-29 18:19:19 2021-01-29 19:51:13 2021-01-29 20:37:13 0:46:00 0:19:38 0:26:22 gibba master rhel 8.3 fs/thrash/multifs/{begin clusters/1a3s-mds-2c-client conf/{client mds mon osd} distro/{centos_8} mount/kclient/{mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} msgr-failures/osd-mds-delay objectstore/bluestore-bitmap overrides/{frag_enable multifs session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} tasks/{1-thrash/mon 2-workunit/cfuse_workunit_suites_fsstress}} 2
fail 5841034 2021-01-29 18:19:22 2021-01-29 19:52:50 2021-01-29 21:58:52 2:06:02 0:32:03 1:33:59 gibba master rhel 8.3 fs/workload/{begin clusters/1a5s-mds-1c-client-3node conf/{client mds mon osd} distro/{centos_8} mount/kclient/{mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} objectstore-ec/bluestore-bitmap omap_limit/10000 overrides/{frag_enable osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/5 scrub/yes tasks/{0-check-counter workunit/fs/misc}} 3
Failure Reason:

"2021-01-29T21:37:16.977858+0000 mds.f (mds.4) 6 : cluster [WRN] Scrub error on inode 0x10000000263 (/client.0/tmp/testdir/dir1/dir2) see mds.f log and `damage ls` output for details" in cluster log

pass 5841035 2021-01-29 18:19:24 2021-01-29 20:07:25 2021-01-29 21:27:25 1:20:00 0:32:00 0:48:00 gibba master rhel 8.3 fs/workload/{begin clusters/1a5s-mds-1c-client-3node conf/{client mds mon osd} distro/{ubuntu_latest} mount/kclient/{mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} objectstore-ec/bluestore-bitmap omap_limit/10000 overrides/{frag_enable osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/1 scrub/yes tasks/{0-check-counter workunit/suites/ffsb}} 3
pass 5841036 2021-01-29 18:19:26 2021-01-29 20:10:59 2021-01-29 20:36:59 0:26:00 0:10:30 0:15:30 gibba master rhel 8.3 fs/thrash/workloads/{begin clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{centos_8} mount/kclient/{mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} msgr-failures/osd-mds-delay objectstore-ec/bluestore-comp-ec-root overrides/{frag_enable session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} ranks/1 tasks/{1-thrash/mds 2-workunit/fs/trivial_sync}} 2
pass 5841037 2021-01-29 18:19:28 2021-01-29 20:14:57 2021-01-29 20:42:56 0:27:59 0:12:07 0:15:52 gibba master rhel 8.3 fs/workload/{begin clusters/1a5s-mds-1c-client-3node conf/{client mds mon osd} distro/{centos_8} mount/kclient/{mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} objectstore-ec/bluestore-bitmap omap_limit/10000 overrides/{frag_enable osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/3 scrub/yes tasks/{0-check-counter workunit/suites/fsync-tester}} 3
pass 5841038 2021-01-29 18:19:29 2021-01-29 20:20:15 2021-01-29 21:24:14 1:03:59 0:37:52 0:26:07 gibba master rhel 8.3 fs/thrash/workloads/{begin clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{ubuntu_latest} mount/kclient/{mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} msgr-failures/osd-mds-delay objectstore-ec/bluestore-comp-ec-root overrides/{frag_enable session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} ranks/3 tasks/{1-thrash/mds 2-workunit/suites/ffsb}} 2