Status Job ID Links Posted Started Updated
Runtime
Duration
In Waiting
Machine Teuthology Branch OS Type OS Version Description Nodes
pass 6488898 2021-11-06 20:25:09 2021-11-07 02:15:21 2021-11-07 02:41:06 0:25:45 0:14:35 0:11:10 smithi master centos 8.3 rados/cephadm/smoke-roleless/{0-distro/centos_8.3_container_tools_3.0 0-nvme-loop 1-start 2-services/rgw 3-final} 2
pass 6488899 2021-11-06 20:25:10 2021-11-07 02:15:31 2021-11-07 02:53:09 0:37:38 0:27:13 0:10:25 smithi master centos 8.2 rados/cephadm/mgr-nfs-upgrade/{0-centos_8.2_container_tools_3.0 1-bootstrap/octopus 1-start 2-nfs 3-upgrade-with-workload 4-final} 2
pass 6488900 2021-11-06 20:25:11 2021-11-07 02:15:42 2021-11-07 02:54:55 0:39:13 0:28:48 0:10:25 smithi master centos 8.3 rados/basic/{ceph clusters/{fixed-2 openstack} mon_election/connectivity msgr-failures/many msgr/async-v1only objectstore/bluestore-comp-zlib rados supported-random-distro$/{centos_8} tasks/rados_api_tests} 2
dead 6488901 2021-11-06 20:25:12 2021-11-07 02:15:42 2021-11-07 14:26:28 12:10:46 smithi master centos 8.3 rados/cephadm/mds_upgrade_sequence/{bluestore-bitmap centos_8.3_container_tools_3.0 conf/{client mds mon osd} overrides/{pg-warn whitelist_health whitelist_wrongly_marked_down} roles tasks/{0-v16.2.4 1-volume/{0-create 1-ranks/1 2-allow_standby_replay/yes 3-verify} 2-client 3-upgrade-with-workload 4-verify}} 2
Failure Reason:

hit max job timeout

pass 6488902 2021-11-06 20:25:13 2021-11-07 02:15:42 2021-11-07 03:03:24 0:47:42 0:37:40 0:10:02 smithi master centos 8.3 rados/cephadm/with-work/{0-distro/centos_8.3_container_tools_3.0 fixed-2 mode/root mon_election/connectivity msgr/async start tasks/rados_python} 2
pass 6488903 2021-11-06 20:25:14 2021-11-07 02:16:13 2021-11-07 02:42:34 0:26:21 0:14:30 0:11:51 smithi master centos 8.3 rados/cephadm/smoke-roleless/{0-distro/centos_8.3_container_tools_3.0 0-nvme-loop 1-start 2-services/nfs-ingress-rgw 3-final} 2
fail 6488904 2021-11-06 20:25:15 2021-11-07 02:17:23 2021-11-07 02:52:48 0:35:25 0:25:59 0:09:26 smithi master centos 8.2 rados/dashboard/{centos_8.2_container_tools_3.0 debug/mgr mon_election/classic random-objectstore$/{bluestore-hybrid} tasks/dashboard} 2
Failure Reason:

Test failure: test_ganesha (unittest.loader._FailedTest)

dead 6488905 2021-11-06 20:25:16 2021-11-07 02:17:34 2021-11-07 14:33:04 12:15:30 smithi master centos 8.3 rados/cephadm/mds_upgrade_sequence/{bluestore-bitmap centos_8.3_container_tools_3.0 conf/{client mds mon osd} overrides/{pg-warn whitelist_health whitelist_wrongly_marked_down} roles tasks/{0-v16.2.4 1-volume/{0-create 1-ranks/1 2-allow_standby_replay/no 3-verify} 2-client 3-upgrade-with-workload 4-verify}} 2
Failure Reason:

hit max job timeout

pass 6488906 2021-11-06 20:25:17 2021-11-07 02:23:15 2021-11-07 03:03:18 0:40:03 0:14:57 0:25:06 smithi master centos 8.2 rados/cephadm/smoke/{0-nvme-loop agent/off distro/centos_8.2_container_tools_3.0 fixed-2 mon_election/classic start} 2
pass 6488907 2021-11-06 20:25:18 2021-11-07 02:38:07 2021-11-07 03:06:34 0:28:27 0:14:56 0:13:31 smithi master centos 8.3 rados/cephadm/smoke/{0-nvme-loop agent/on distro/centos_8.3_container_tools_3.0 fixed-2 mon_election/connectivity start} 2
pass 6488908 2021-11-06 20:25:19 2021-11-07 02:41:08 2021-11-07 03:22:04 0:40:56 0:33:31 0:07:25 smithi master rhel 8.4 rados/thrash/{0-size-min-size-overrides/2-size-2-min-size 1-pg-log-overrides/normal_pg_log 2-recovery-overrides/{more-partial-recovery} 3-scrub-overrides/{max-simultaneous-scrubs-3} backoff/normal ceph clusters/{fixed-2 openstack} crc-failures/bad_map_crc_failure d-balancer/crush-compat mon_election/classic msgr-failures/osd-dispatch-delay msgr/async-v1only objectstore/bluestore-bitmap rados supported-random-distro$/{rhel_8} thrashers/careful thrashosds-health workloads/small-objects} 2
pass 6488909 2021-11-06 20:25:20 2021-11-07 02:42:38 2021-11-07 03:27:05 0:44:27 0:31:23 0:13:04 smithi master ubuntu 20.04 rados/thrash-erasure-code-big/{ceph cluster/{12-osds openstack} mon_election/connectivity msgr-failures/few objectstore/bluestore-comp-lz4 rados recovery-overrides/{more-async-partial-recovery} supported-random-distro$/{ubuntu_latest} thrashers/pggrow thrashosds-health workloads/ec-rados-plugin=jerasure-k=4-m=2} 3
dead 6488910 2021-11-06 20:25:20 2021-11-07 02:43:08 2021-11-07 14:55:20 12:12:12 smithi master centos 8.3 rados/cephadm/mds_upgrade_sequence/{bluestore-bitmap centos_8.3_container_tools_3.0 conf/{client mds mon osd} overrides/{pg-warn whitelist_health whitelist_wrongly_marked_down} roles tasks/{0-v16.2.4 1-volume/{0-create 1-ranks/2 2-allow_standby_replay/yes 3-verify} 2-client 3-upgrade-with-workload 4-verify}} 2
Failure Reason:

hit max job timeout

fail 6488911 2021-11-06 20:25:21 2021-11-07 02:44:29 2021-11-07 03:13:41 0:29:12 0:19:31 0:09:41 smithi master centos 8.2 rados/dashboard/{centos_8.2_container_tools_3.0 debug/mgr mon_election/connectivity random-objectstore$/{bluestore-comp-zstd} tasks/e2e} 2
Failure Reason:

Command failed (workunit test cephadm/test_dashboard_e2e.sh) on smithi078 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=8449fa6794694cc653cc4ca07b755225039f4b6d TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 CEPH_MNT=/home/ubuntu/cephtest/mnt.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/cephadm/test_dashboard_e2e.sh'

dead 6488912 2021-11-06 20:25:22 2021-11-07 02:44:29 2021-11-07 15:00:23 12:15:54 smithi master centos 8.3 rados/cephadm/mds_upgrade_sequence/{bluestore-bitmap centos_8.3_container_tools_3.0 conf/{client mds mon osd} overrides/{pg-warn whitelist_health whitelist_wrongly_marked_down} roles tasks/{0-v16.2.4 1-volume/{0-create 1-ranks/2 2-allow_standby_replay/no 3-verify} 2-client 3-upgrade-with-workload 4-verify}} 2
Failure Reason:

hit max job timeout

pass 6488913 2021-11-06 20:25:23 2021-11-07 02:49:10 2021-11-07 04:29:59 1:40:49 1:27:24 0:13:25 smithi master centos 8.3 rados/verify/{centos_latest ceph clusters/{fixed-2 openstack} d-thrash/default/{default thrashosds-health} mon_election/connectivity msgr-failures/few msgr/async-v1only objectstore/bluestore-comp-snappy rados tasks/mon_recovery validater/valgrind} 2
pass 6488914 2021-11-06 20:25:24 2021-11-07 02:52:51 2021-11-07 03:32:10 0:39:19 0:28:48 0:10:31 smithi master centos 8.2 rados/cephadm/mgr-nfs-upgrade/{0-centos_8.2_container_tools_3.0 1-bootstrap/octopus 1-start 2-nfs 3-upgrade-with-workload 4-final} 2
dead 6488915 2021-11-06 20:25:25 2021-11-07 02:52:51 2021-11-07 15:04:22 12:11:31 smithi master centos 8.3 rados/cephadm/mds_upgrade_sequence/{bluestore-bitmap centos_8.3_container_tools_3.0 conf/{client mds mon osd} overrides/{pg-warn whitelist_health whitelist_wrongly_marked_down} roles tasks/{0-v16.2.4 1-volume/{0-create 1-ranks/1 2-allow_standby_replay/yes 3-verify} 2-client 3-upgrade-with-workload 4-verify}} 2
Failure Reason:

hit max job timeout

pass 6488916 2021-11-06 20:25:26 2021-11-07 02:53:12 2021-11-07 03:18:17 0:25:05 0:13:02 0:12:03 smithi master centos 8.3 rados/cephadm/osds/{0-distro/centos_8.3_container_tools_3.0 0-nvme-loop 1-start 2-ops/rm-zap-add} 2
fail 6488917 2021-11-06 20:25:27 2021-11-07 02:55:02 2021-11-07 03:35:07 0:40:05 0:25:08 0:14:57 smithi master centos 8.2 rados/dashboard/{centos_8.2_container_tools_3.0 debug/mgr mon_election/connectivity random-objectstore$/{bluestore-comp-zlib} tasks/dashboard} 2
Failure Reason:

Test failure: test_ganesha (unittest.loader._FailedTest)

dead 6488918 2021-11-06 20:25:28 2021-11-07 03:00:03 2021-11-07 15:10:21 12:10:18 smithi master centos 8.3 rados/cephadm/mds_upgrade_sequence/{bluestore-bitmap centos_8.3_container_tools_3.0 conf/{client mds mon osd} overrides/{pg-warn whitelist_health whitelist_wrongly_marked_down} roles tasks/{0-v16.2.4 1-volume/{0-create 1-ranks/1 2-allow_standby_replay/no 3-verify} 2-client 3-upgrade-with-workload 4-verify}} 2
Failure Reason:

hit max job timeout

dead 6488919 2021-11-06 20:25:29 2021-11-07 03:00:04 2021-11-07 15:11:57 12:11:53 smithi master centos 8.3 rados/cephadm/mds_upgrade_sequence/{bluestore-bitmap centos_8.3_container_tools_3.0 conf/{client mds mon osd} overrides/{pg-warn whitelist_health whitelist_wrongly_marked_down} roles tasks/{0-v16.2.4 1-volume/{0-create 1-ranks/2 2-allow_standby_replay/yes 3-verify} 2-client 3-upgrade-with-workload 4-verify}} 2
Failure Reason:

hit max job timeout

pass 6488920 2021-11-06 20:25:30 2021-11-07 03:00:04 2021-11-07 04:58:57 1:58:53 1:44:05 0:14:48 smithi master centos 8.3 rados/verify/{centos_latest ceph clusters/{fixed-2 openstack} d-thrash/default/{default thrashosds-health} mon_election/connectivity msgr-failures/few msgr/async-v2only objectstore/bluestore-low-osd-mem-target rados tasks/rados_api_tests validater/valgrind} 2
pass 6488921 2021-11-06 20:25:31 2021-11-07 03:03:25 2021-11-07 03:53:17 0:49:52 0:32:55 0:16:57 smithi master ubuntu 20.04 rados/thrash-erasure-code-big/{ceph cluster/{12-osds openstack} mon_election/connectivity msgr-failures/osd-dispatch-delay objectstore/bluestore-stupid rados recovery-overrides/{more-partial-recovery} supported-random-distro$/{ubuntu_latest} thrashers/pggrow thrashosds-health workloads/ec-rados-plugin=jerasure-k=4-m=2} 3
fail 6488922 2021-11-06 20:25:32 2021-11-07 03:06:35 2021-11-07 03:40:48 0:34:13 0:19:06 0:15:07 smithi master centos 8.2 rados/dashboard/{centos_8.2_container_tools_3.0 debug/mgr mon_election/classic random-objectstore$/{bluestore-hybrid} tasks/e2e} 2
Failure Reason:

Command failed (workunit test cephadm/test_dashboard_e2e.sh) on smithi129 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=8449fa6794694cc653cc4ca07b755225039f4b6d TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 CEPH_MNT=/home/ubuntu/cephtest/mnt.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/cephadm/test_dashboard_e2e.sh'

dead 6488923 2021-11-06 20:25:33 2021-11-07 03:11:36 2021-11-07 15:23:39 12:12:03 smithi master centos 8.3 rados/cephadm/mds_upgrade_sequence/{bluestore-bitmap centos_8.3_container_tools_3.0 conf/{client mds mon osd} overrides/{pg-warn whitelist_health whitelist_wrongly_marked_down} roles tasks/{0-v16.2.4 1-volume/{0-create 1-ranks/2 2-allow_standby_replay/no 3-verify} 2-client 3-upgrade-with-workload 4-verify}} 2
Failure Reason:

hit max job timeout