Status Job ID Links Posted Started Updated
Runtime
Duration
In Waiting
Machine Teuthology Branch OS Type OS Version Description Nodes
fail 6248610 2021-07-02 03:55:08 2021-07-02 03:58:32 2021-07-02 04:22:00 0:23:28 0:12:38 0:10:50 smithi wip-focal centos 8.2 rados/cephadm/thrash/{0-distro/centos_8.2_kubic_stable 1-start 2-thrash 3-tasks/small-objects fixed-2 msgr/async-v1only root} 2
Failure Reason:

Command failed on smithi111 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 530a3d24-daec-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi111:vg_nvme/lv_4'

fail 6248611 2021-07-02 03:55:10 2021-07-02 03:58:33 2021-07-02 04:22:04 0:23:31 0:12:05 0:11:26 smithi wip-focal centos 8.3 rados/thrash-old-clients/{0-size-min-size-overrides/2-size-2-min-size 1-install/pacific backoff/normal ceph clusters/{openstack three-plus-one} d-balancer/on distro$/{centos_latest} mon_election/connectivity msgr-failures/osd-delay rados thrashers/default thrashosds-health workloads/test_rbd_api} 3
Failure Reason:

Command failed on smithi043 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 36b5ec72-daec-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi043:vg_nvme/lv_4'

fail 6248612 2021-07-02 03:55:11 2021-07-02 03:58:33 2021-07-02 04:18:51 0:20:18 0:08:01 0:12:17 smithi wip-focal ubuntu 20.04 rados/cephadm/smoke/{distro/ubuntu_20.04 fixed-2 mon_election/connectivity start} 2
Failure Reason:

Command failed on smithi102 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid d3e4bb8c-daeb-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi102:vg_nvme/lv_4'

fail 6248613 2021-07-02 03:55:12 2021-07-02 03:59:14 2021-07-02 04:20:44 0:21:30 0:14:41 0:06:49 smithi wip-focal rhel 8.3 rados/cephadm/smoke-roleless/{0-distro/rhel_8.3_kubic_stable 1-start 2-services/nfs 3-final} 2
Failure Reason:

Command failed on smithi153 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 296d3052-daec-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi153:vg_nvme/lv_4'

fail 6248614 2021-07-02 03:55:13 2021-07-02 03:59:34 2021-07-02 04:32:38 0:33:04 0:23:33 0:09:31 smithi wip-focal centos 8.stream rados/mgr/{clusters/{2-node-mgr} debug/mgr mon_election/connectivity objectstore/bluestore-comp-zstd supported-random-distro$/{centos_8.stream} tasks/progress} 2
Failure Reason:

"2021-07-02T04:19:16.849237+0000 mon.a (mon.0) 233 : cluster [WRN] Health check failed: norecover flag(s) set (OSDMAP_FLAGS)" in cluster log

fail 6248615 2021-07-02 03:55:14 2021-07-02 03:59:34 2021-07-02 04:20:15 0:20:41 0:11:24 0:09:17 smithi wip-focal centos 8.2 rados/cephadm/workunits/{0-distro/centos_8.2_kubic_stable mon_election/classic task/test_orch_cli} 1
Failure Reason:

Command failed on smithi005 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 3c808c3e-daec-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi005:vg_nvme/lv_4'

fail 6248616 2021-07-02 03:55:15 2021-07-02 03:59:35 2021-07-02 04:23:43 0:24:08 0:12:31 0:11:37 smithi wip-focal centos 8.2 rados/cephadm/with-work/{0-distro/centos_8.2_kubic_stable fixed-2 mode/root mon_election/classic msgr/async-v1only start tasks/rados_python} 2
Failure Reason:

Command failed on smithi033 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 845d44de-daec-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi033:vg_nvme/lv_4'

fail 6248617 2021-07-02 03:55:17 2021-07-02 04:00:45 2021-07-02 04:20:01 0:19:16 0:08:49 0:10:27 smithi wip-focal ubuntu 20.04 rados/cephadm/smoke-roleless/{0-distro/ubuntu_20.04 1-start 2-services/nfs2 3-final} 2
Failure Reason:

Command failed on smithi028 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid f663c6da-daeb-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi028:vg_nvme/lv_4'

fail 6248618 2021-07-02 03:55:18 2021-07-02 04:00:55 2021-07-02 04:23:15 0:22:20 0:10:05 0:12:15 smithi wip-focal ubuntu 20.04 rados/upgrade/parallel/{0-start 1-tasks distro1$/{ubuntu_20.04} mon_election/connectivity upgrade-sequence workload/{ec-rados-default rados_api rados_loadgenbig rbd_import_export test_rbd_api test_rbd_python}} 2
Failure Reason:

Command failed on smithi124 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image docker.io/ceph/daemon-base:latest-pacific shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 6db5be32-daec-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi124:vg_nvme/lv_4'

fail 6248619 2021-07-02 03:55:19 2021-07-02 04:01:46 2021-07-02 04:24:59 0:23:13 0:12:16 0:10:57 smithi wip-focal centos 8.2 rados/cephadm/thrash/{0-distro/centos_8.2_kubic_stable 1-start 2-thrash 3-tasks/snaps-few-objects fixed-2 msgr/async-v2only root} 2
Failure Reason:

Command failed on smithi085 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid b4e91e2a-daec-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi085:vg_nvme/lv_4'

fail 6248620 2021-07-02 03:55:20 2021-07-02 04:01:46 2021-07-02 04:23:05 0:21:19 0:10:16 0:11:03 smithi wip-focal centos 8.2 rados/cephadm/smoke-roleless/{0-distro/centos_8.2_kubic_stable 1-start 2-services/rgw-ingress 3-final} 2
Failure Reason:

Command failed on smithi073 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 665ac4de-daec-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi073:vg_nvme/lv_4'

fail 6248621 2021-07-02 03:55:21 2021-07-02 04:01:57 2021-07-02 04:21:31 0:19:34 0:09:48 0:09:46 smithi wip-focal centos 8.2 rados/cephadm/smoke/{distro/centos_8.2_kubic_stable fixed-2 mon_election/classic start} 2
Failure Reason:

Command failed on smithi086 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 7106d5c6-daec-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi086:vg_nvme/lv_4'

fail 6248622 2021-07-02 03:55:22 2021-07-02 04:02:37 2021-07-02 04:20:16 0:17:39 0:07:39 0:10:00 smithi wip-focal ubuntu 20.04 rados/cephadm/smoke-singlehost/{0-distro$/{ubuntu_20.04} 1-start 2-services/basic 3-final} 1
Failure Reason:

Command failed on smithi122 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 15b60cf0-daec-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi122:vg_nvme/lv_4'

pass 6248623 2021-07-02 03:55:23 2021-07-02 04:02:57 2021-07-02 06:11:35 2:08:38 2:00:01 0:08:37 smithi wip-focal centos 8.3 rados/verify/{centos_latest ceph clusters/{fixed-2 openstack} d-thrash/none mon_election/connectivity msgr-failures/few msgr/async-v1only objectstore/bluestore-bitmap rados tasks/rados_cls_all validater/valgrind} 2
fail 6248624 2021-07-02 03:55:24 2021-07-02 04:02:57 2021-07-02 04:23:05 0:20:08 0:07:38 0:12:30 smithi wip-focal ubuntu 20.04 rados/cephadm/upgrade/{1-start-distro/1-start-ubuntu_20.04 2-repo_digest/repo_digest 3-start-upgrade 4-wait mon_election/classic} 2
Failure Reason:

Command failed on smithi148 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image docker.io/ceph/ceph:v15.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 6affc5ca-daec-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi148:vg_nvme/lv_4'

dead 6248625 2021-07-02 03:55:25 2021-07-02 04:03:28 2021-07-02 04:18:51 0:15:23 smithi wip-focal rhel 8.3 rados/cephadm/smoke-roleless/{0-distro/rhel_8.3_kubic_stable 1-start 2-services/rgw 3-final} 2
Failure Reason:

Error reimaging machines: reached maximum tries (60) after waiting for 900 seconds

fail 6248626 2021-07-02 03:55:26 2021-07-02 04:03:48 2021-07-02 04:28:13 0:24:25 0:12:45 0:11:40 smithi wip-focal centos 8.3 rados/thrash-old-clients/{0-size-min-size-overrides/3-size-2-min-size 1-install/nautilus-v1only backoff/peering ceph clusters/{openstack three-plus-one} d-balancer/crush-compat distro$/{centos_latest} mon_election/classic msgr-failures/fastclose rados thrashers/mapgap thrashosds-health workloads/cache-snaps} 3
Failure Reason:

Command failed on smithi042 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 173b39b4-daed-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi042:vg_nvme/lv_4'

pass 6248627 2021-07-02 03:55:27 2021-07-02 04:05:09 2021-07-02 04:31:28 0:26:19 0:16:49 0:09:30 smithi wip-focal centos 8.2 rados/cephadm/workunits/{0-distro/centos_8.2_kubic_stable mon_election/connectivity task/test_cephadm} 1
fail 6248628 2021-07-02 03:55:28 2021-07-02 04:05:09 2021-07-02 04:24:39 0:19:30 0:08:51 0:10:39 smithi wip-focal ubuntu 20.04 rados/cephadm/smoke-roleless/{0-distro/ubuntu_20.04 1-start 2-services/basic 3-final} 2
Failure Reason:

Command failed on smithi082 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 9a1e812a-daec-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi082:vg_nvme/lv_4'

fail 6248629 2021-07-02 03:55:29 2021-07-02 04:05:19 2021-07-02 04:29:01 0:23:42 0:12:57 0:10:45 smithi wip-focal centos 8.2 rados/cephadm/thrash/{0-distro/centos_8.2_kubic_stable 1-start 2-thrash 3-tasks/rados_api_tests fixed-2 msgr/async root} 2
Failure Reason:

Command failed on smithi134 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 5335216e-daed-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi134:vg_nvme/lv_4'

fail 6248630 2021-07-02 03:55:30 2021-07-02 04:05:40 2021-07-02 04:27:39 0:21:59 0:14:19 0:07:40 smithi wip-focal rhel 8.3 rados/cephadm/smoke/{distro/rhel_8.3_kubic_stable fixed-2 mon_election/connectivity start} 2
Failure Reason:

Command failed on smithi132 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 1eeb33da-daed-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi132:vg_nvme/lv_4'

fail 6248631 2021-07-02 03:55:31 2021-07-02 04:06:40 2021-07-02 04:36:03 0:29:23 0:22:50 0:06:33 smithi wip-focal rhel 8.3 rados/cephadm/with-work/{0-distro/rhel_8.3_kubic_stable fixed-2 mode/packaged mon_election/connectivity msgr/async-v2only start tasks/rados_api_tests} 2
Failure Reason:

Command failed on smithi120 with status 22: 'sudo cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 59064e96-daee-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi120:vg_nvme/lv_4'

fail 6248632 2021-07-02 03:55:32 2021-07-02 04:07:11 2021-07-02 04:28:20 0:21:09 0:10:20 0:10:49 smithi wip-focal centos 8.2 rados/cephadm/smoke-roleless/{0-distro/centos_8.2_kubic_stable 1-start 2-services/client-keyring 3-final} 2
Failure Reason:

Command failed on smithi071 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 225afa50-daed-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi071:vg_nvme/lv_4'

fail 6248633 2021-07-02 03:55:33 2021-07-02 04:07:11 2021-07-02 04:26:54 0:19:43 0:08:50 0:10:53 smithi wip-focal centos 8.3 rados/cephadm/upgrade/{1-start-distro/1-start-centos_8.3-octopus 2-repo_digest/defaut 3-start-upgrade 4-wait mon_election/connectivity} 2
Failure Reason:

Command failed on smithi041 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:octopus shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 02a64d4a-daed-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi041:vg_nvme/lv_4'

pass 6248634 2021-07-02 03:55:34 2021-07-02 04:07:51 2021-07-02 04:38:22 0:30:31 0:22:36 0:07:55 smithi wip-focal centos 8.3 rados/valgrind-leaks/{1-start 2-inject-leak/osd centos_latest} 1
fail 6248635 2021-07-02 03:55:35 2021-07-02 04:07:52 2021-07-02 04:29:51 0:21:59 0:15:00 0:06:59 smithi wip-focal rhel 8.3 rados/cephadm/smoke-roleless/{0-distro/rhel_8.3_kubic_stable 1-start 2-services/iscsi 3-final} 2
Failure Reason:

Command failed on smithi006 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 780f22be-daed-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi006:vg_nvme/lv_4'

fail 6248636 2021-07-02 03:55:36 2021-07-02 04:08:42 2021-07-02 04:32:38 0:23:56 0:12:56 0:11:00 smithi wip-focal centos 8.2 rados/cephadm/thrash/{0-distro/centos_8.2_kubic_stable 1-start 2-thrash 3-tasks/radosbench fixed-2 msgr/async-v1only root} 2
Failure Reason:

Command failed on smithi023 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ea611746-daed-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi023:vg_nvme/lv_4'

fail 6248637 2021-07-02 03:55:37 2021-07-02 04:09:02 2021-07-02 04:28:47 0:19:45 0:08:46 0:10:59 smithi wip-focal ubuntu 20.04 rados/cephadm/smoke-roleless/{0-distro/ubuntu_20.04 1-start 2-services/mirror 3-final} 2
Failure Reason:

Command failed on smithi080 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 2b83fcf8-daed-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi080:vg_nvme/lv_4'

fail 6248638 2021-07-02 03:55:39 2021-07-02 04:09:23 2021-07-02 04:28:53 0:19:30 0:07:54 0:11:36 smithi wip-focal ubuntu 20.04 rados/cephadm/smoke/{distro/ubuntu_20.04 fixed-2 mon_election/classic start} 2
Failure Reason:

Command failed on smithi003 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 373b722e-daed-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi003:vg_nvme/lv_4'

fail 6248639 2021-07-02 03:55:40 2021-07-02 04:09:23 2021-07-02 04:28:55 0:19:32 0:11:04 0:08:28 smithi wip-focal centos 8.2 rados/cephadm/workunits/{0-distro/centos_8.2_kubic_stable mon_election/connectivity task/test_orch_cli} 1
Failure Reason:

Command failed on smithi179 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a3243b74-daed-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi179:vg_nvme/lv_4'

fail 6248640 2021-07-02 03:55:41 2021-07-02 04:09:43 2021-07-02 04:32:35 0:22:52 0:11:48 0:11:04 smithi wip-focal centos 8.3 rados/thrash-old-clients/{0-size-min-size-overrides/2-size-2-min-size 1-install/nautilus-v2only backoff/peering_and_degraded ceph clusters/{openstack three-plus-one} d-balancer/on distro$/{centos_latest} mon_election/connectivity msgr-failures/few rados thrashers/morepggrow thrashosds-health workloads/radosbench} 3
Failure Reason:

Command failed on smithi099 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a3a96998-daed-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi099:vg_nvme/lv_4'

fail 6248641 2021-07-02 03:55:42 2021-07-02 04:09:53 2021-07-02 04:31:14 0:21:21 0:10:35 0:10:46 smithi wip-focal centos 8.2 rados/cephadm/smoke-roleless/{0-distro/centos_8.2_kubic_stable 1-start 2-services/nfs-ingress 3-final} 2
Failure Reason:

Command failed on smithi159 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 83d0b9a0-daed-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi159:vg_nvme/lv_4'

fail 6248642 2021-07-02 03:55:43 2021-07-02 04:10:14 2021-07-02 04:31:49 0:21:35 0:10:30 0:11:05 smithi wip-focal ubuntu 20.04 rados/cephadm/with-work/{0-distro/ubuntu_20.04 fixed-2 mode/root mon_election/classic msgr/async start tasks/rados_python} 2
Failure Reason:

Command failed on smithi131 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 9bac0426-daed-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi131:vg_nvme/lv_4'

pass 6248643 2021-07-02 03:55:44 2021-07-02 04:10:14 2021-07-02 04:50:59 0:40:45 0:28:56 0:11:49 smithi wip-focal ubuntu 20.04 rados/thrash-erasure-code-isa/{arch/x86_64 ceph clusters/{fixed-2 openstack} mon_election/connectivity msgr-failures/osd-dispatch-delay objectstore/bluestore-comp-zstd rados recovery-overrides/{more-active-recovery} supported-random-distro$/{ubuntu_latest} thrashers/default thrashosds-health workloads/ec-rados-plugin=isa-k=2-m=1} 2
fail 6248644 2021-07-02 03:55:45 2021-07-02 04:11:35 2021-07-02 04:28:48 0:17:13 0:04:08 0:13:05 smithi wip-focal ubuntu 20.04 rados/dashboard/{centos_8.2_kubic_stable debug/mgr mon_election/classic random-objectstore$/{bluestore-comp-snappy} supported-random-distro$/{ubuntu_latest} tasks/e2e} 2
Failure Reason:

Command failed on smithi135 with status 1: 'TESTDIR=/home/ubuntu/cephtest bash -s'

fail 6248645 2021-07-02 03:55:46 2021-07-02 04:13:26 2021-07-02 04:34:45 0:21:19 0:14:31 0:06:48 smithi wip-focal rhel 8.3 rados/cephadm/smoke-roleless/{0-distro/rhel_8.3_kubic_stable 1-start 2-services/nfs-ingress2 3-final} 2
Failure Reason:

Command failed on smithi145 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 18782f3e-daee-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi145:vg_nvme/lv_4'

fail 6248646 2021-07-02 03:55:47 2021-07-02 04:13:36 2021-07-02 04:36:11 0:22:35 0:12:53 0:09:42 smithi wip-focal centos 8.2 rados/cephadm/thrash/{0-distro/centos_8.2_kubic_stable 1-start 2-thrash 3-tasks/small-objects fixed-2 msgr/async-v2only root} 2
Failure Reason:

Command failed on smithi026 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4b95b65c-daee-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi026:vg_nvme/lv_4'

fail 6248647 2021-07-02 03:55:48 2021-07-02 04:13:36 2021-07-02 04:34:16 0:20:40 0:10:24 0:10:16 smithi wip-focal centos 8.2 rados/cephadm/smoke/{distro/centos_8.2_kubic_stable fixed-2 mon_election/connectivity start} 2
Failure Reason:

Command failed on smithi007 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid fa9f375a-daed-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi007:vg_nvme/lv_4'

fail 6248648 2021-07-02 03:55:49 2021-07-02 04:13:37 2021-07-02 04:35:25 0:21:48 0:13:39 0:08:09 smithi wip-focal rhel 8.3 rados/cephadm/smoke-singlehost/{0-distro$/{rhel_8.3_kubic_stable} 1-start 2-services/rgw 3-final} 1
Failure Reason:

Command failed on smithi161 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 2c74fbe8-daee-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi161:vg_nvme/lv_4'

fail 6248649 2021-07-02 03:55:50 2021-07-02 04:14:07 2021-07-02 04:35:48 0:21:41 0:07:38 0:14:03 smithi wip-focal ubuntu 20.04 rados/cephadm/upgrade/{1-start-distro/1-start-ubuntu_20.04-15.2.9 2-repo_digest/repo_digest 3-start-upgrade 4-wait mon_election/classic} 2
Failure Reason:

Command failed on smithi100 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image docker.io/ceph/ceph:v15.2.9 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 2e7ff38e-daee-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi100:vg_nvme/lv_4'

fail 6248650 2021-07-02 03:55:51 2021-07-02 04:16:08 2021-07-02 04:36:12 0:20:04 0:08:25 0:11:39 smithi wip-focal ubuntu 20.04 rados/cephadm/smoke-roleless/{0-distro/ubuntu_20.04 1-start 2-services/nfs 3-final} 2
Failure Reason:

Command failed on smithi185 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 336fa574-daee-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi185:vg_nvme/lv_4'

fail 6248651 2021-07-02 03:55:52 2021-07-02 04:16:39 2021-07-02 04:38:59 0:22:20 0:10:15 0:12:05 smithi wip-focal centos 8.2 rados/cephadm/smoke-roleless/{0-distro/centos_8.2_kubic_stable 1-start 2-services/nfs2 3-final} 2
Failure Reason:

Command failed on smithi050 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 95867d46-daee-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi050:vg_nvme/lv_4'

pass 6248652 2021-07-02 03:55:53 2021-07-02 04:17:59 2021-07-02 04:43:02 0:25:03 0:16:17 0:08:46 smithi wip-focal centos 8.2 rados/cephadm/workunits/{0-distro/centos_8.2_kubic_stable mon_election/classic task/test_cephadm} 1
fail 6248653 2021-07-02 03:55:54 2021-07-02 04:17:59 2021-07-02 04:41:09 0:23:10 0:12:15 0:10:55 smithi wip-focal centos 8.2 rados/cephadm/thrash/{0-distro/centos_8.2_kubic_stable 1-start 2-thrash 3-tasks/snaps-few-objects fixed-2 msgr/async root} 2
Failure Reason:

Command failed on smithi176 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid fa7a9e12-daee-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi176:vg_nvme/lv_4'

fail 6248654 2021-07-02 03:55:55 2021-07-02 04:18:11 2021-07-02 04:39:50 0:21:39 0:14:50 0:06:49 smithi wip-focal rhel 8.3 rados/cephadm/smoke-roleless/{0-distro/rhel_8.3_kubic_stable 1-start 2-services/rgw-ingress 3-final} 2
Failure Reason:

Command failed on smithi067 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid dbd9327a-daee-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi067:vg_nvme/lv_4'

fail 6248655 2021-07-02 03:55:56 2021-07-02 04:18:41 2021-07-02 04:41:48 0:23:07 0:11:46 0:11:21 smithi wip-focal centos 8.3 rados/thrash-old-clients/{0-size-min-size-overrides/3-size-2-min-size 1-install/nautilus backoff/normal ceph clusters/{openstack three-plus-one} d-balancer/crush-compat distro$/{centos_latest} mon_election/classic msgr-failures/osd-delay rados thrashers/none thrashosds-health workloads/rbd_cls} 3
Failure Reason:

Command failed on smithi081 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ec90c376-daee-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi081:vg_nvme/lv_4'

dead 6248656 2021-07-02 03:55:57 2021-07-02 04:18:51 2021-07-02 04:34:02 0:15:11 smithi wip-focal rhel 8.3 rados/cephadm/smoke/{distro/rhel_8.3_kubic_stable fixed-2 mon_election/classic start} 2
Failure Reason:

Error reimaging machines: reached maximum tries (60) after waiting for 900 seconds

fail 6248657 2021-07-02 03:55:58 2021-07-02 04:19:02 2021-07-02 04:40:37 0:21:35 0:09:58 0:11:37 smithi wip-focal ubuntu 20.04 rados/cephadm/with-work/{0-distro/ubuntu_20.04 fixed-2 mode/packaged mon_election/connectivity msgr/async-v1only start tasks/rados_api_tests} 2
Failure Reason:

Command failed on smithi102 with status 22: 'sudo cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid d590cbd0-daee-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi102:vg_nvme/lv_4'

fail 6248658 2021-07-02 03:55:59 2021-07-02 04:19:02 2021-07-02 04:39:39 0:20:37 0:08:39 0:11:58 smithi wip-focal ubuntu 20.04 rados/cephadm/smoke-roleless/{0-distro/ubuntu_20.04 1-start 2-services/rgw 3-final} 2
Failure Reason:

Command failed on smithi118 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid b7f78744-daee-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi118:vg_nvme/lv_4'

fail 6248659 2021-07-02 03:56:00 2021-07-02 04:19:43 2021-07-02 04:39:05 0:19:22 0:07:40 0:11:42 smithi wip-focal ubuntu 20.04 rados/cephadm/upgrade/{1-start-distro/1-start-ubuntu_20.04 2-repo_digest/defaut 3-start-upgrade 4-wait mon_election/connectivity} 2
Failure Reason:

Command failed on smithi032 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image docker.io/ceph/ceph:v15.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid adcf0c38-daee-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi032:vg_nvme/lv_4'

fail 6248660 2021-07-02 03:56:02 2021-07-02 04:19:43 2021-07-02 04:41:12 0:21:29 0:11:13 0:10:16 smithi wip-focal centos 8.2 rados/cephadm/smoke-roleless/{0-distro/centos_8.2_kubic_stable 1-start 2-services/basic 3-final} 2
Failure Reason:

Command failed on smithi029 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 034002bc-daef-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi029:vg_nvme/lv_4'

fail 6248661 2021-07-02 03:56:03 2021-07-02 04:19:54 2021-07-02 04:42:27 0:22:33 0:12:50 0:09:43 smithi wip-focal centos 8.2 rados/cephadm/thrash/{0-distro/centos_8.2_kubic_stable 1-start 2-thrash 3-tasks/rados_api_tests fixed-2 msgr/async-v1only root} 2
Failure Reason:

Command failed on smithi028 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 26d6325a-daef-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi028:vg_nvme/lv_4'

pass 6248662 2021-07-02 03:56:04 2021-07-02 04:20:04 2021-07-02 04:57:40 0:37:36 0:26:09 0:11:27 smithi wip-focal ubuntu 20.04 rados/thrash/{0-size-min-size-overrides/3-size-2-min-size 1-pg-log-overrides/short_pg_log 2-recovery-overrides/{more-partial-recovery} 3-scrub-overrides/{default} backoff/peering_and_degraded ceph clusters/{fixed-2 openstack} crc-failures/bad_map_crc_failure d-balancer/crush-compat mon_election/connectivity msgr-failures/osd-dispatch-delay msgr/async objectstore/bluestore-comp-zlib rados supported-random-distro$/{ubuntu_latest} thrashers/default thrashosds-health workloads/radosbench} 2
fail 6248663 2021-07-02 03:56:05 2021-07-02 04:20:25 2021-07-02 04:39:51 0:19:26 0:08:07 0:11:19 smithi wip-focal ubuntu 20.04 rados/cephadm/smoke/{distro/ubuntu_20.04 fixed-2 mon_election/connectivity start} 2
Failure Reason:

Command failed on smithi112 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid b7342024-daee-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi112:vg_nvme/lv_4'

fail 6248664 2021-07-02 03:56:06 2021-07-02 04:20:35 2021-07-02 04:41:53 0:21:18 0:14:29 0:06:49 smithi wip-focal rhel 8.3 rados/cephadm/smoke-roleless/{0-distro/rhel_8.3_kubic_stable 1-start 2-services/client-keyring 3-final} 2
Failure Reason:

Command failed on smithi153 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 14ad65d0-daef-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi153:vg_nvme/lv_4'

fail 6248665 2021-07-02 03:56:07 2021-07-02 04:20:45 2021-07-02 04:42:41 0:21:56 0:11:24 0:10:32 smithi wip-focal centos 8.2 rados/cephadm/workunits/{0-distro/centos_8.2_kubic_stable mon_election/classic task/test_orch_cli} 1
Failure Reason:

Command failed on smithi158 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 678c9532-daef-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi158:vg_nvme/lv_4'

fail 6248666 2021-07-02 03:56:08 2021-07-02 04:20:46 2021-07-02 04:52:50 0:32:04 0:16:10 0:15:54 smithi wip-focal centos 8.2 rados/cephadm/with-work/{0-distro/centos_8.2_kubic_stable fixed-2 mode/root mon_election/classic msgr/async-v2only start tasks/rados_python} 2
Failure Reason:

Command failed on smithi019 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a87bec4a-daf0-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi019:vg_nvme/lv_4'

fail 6248667 2021-07-02 03:56:09 2021-07-02 04:20:56 2021-07-02 04:40:47 0:19:51 0:09:13 0:10:38 smithi wip-focal ubuntu 20.04 rados/cephadm/smoke-roleless/{0-distro/ubuntu_20.04 1-start 2-services/iscsi 3-final} 2
Failure Reason:

Command failed on smithi089 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid e54c5c2e-daee-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi089:vg_nvme/lv_4'

fail 6248668 2021-07-02 03:56:10 2021-07-02 04:21:36 2021-07-02 04:37:35 0:15:59 0:04:13 0:11:46 smithi wip-focal ubuntu 20.04 rados/dashboard/{centos_8.2_kubic_stable debug/mgr mon_election/classic random-objectstore$/{bluestore-comp-zstd} supported-random-distro$/{ubuntu_latest} tasks/dashboard} 2
Failure Reason:

Command failed on smithi160 with status 1: 'TESTDIR=/home/ubuntu/cephtest bash -s'

fail 6248669 2021-07-02 03:56:11 2021-07-02 04:21:57 2021-07-02 04:49:28 0:27:31 0:19:35 0:07:56 smithi wip-focal rhel 8.3 rados/upgrade/parallel/{0-start 1-tasks distro1$/{rhel_8.3_kubic_stable} mon_election/classic upgrade-sequence workload/{ec-rados-default rados_api rados_loadgenbig rbd_import_export test_rbd_api test_rbd_python}} 2
Failure Reason:

Command failed on smithi076 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image docker.io/ceph/daemon-base:latest-pacific shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 2744d2c2-daf0-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi076:vg_nvme/lv_4'

fail 6248670 2021-07-02 03:56:12 2021-07-02 04:22:07 2021-07-02 04:44:50 0:22:43 0:12:45 0:09:58 smithi wip-focal centos 8.2 rados/cephadm/thrash/{0-distro/centos_8.2_kubic_stable 1-start 2-thrash 3-tasks/radosbench fixed-2 msgr/async-v2only root} 2
Failure Reason:

Command failed on smithi043 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 7ce3a876-daef-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi043:vg_nvme/lv_4'

fail 6248671 2021-07-02 03:56:13 2021-07-02 04:22:07 2021-07-02 04:46:06 0:23:59 0:12:31 0:11:28 smithi wip-focal centos 8.3 rados/thrash-old-clients/{0-size-min-size-overrides/2-size-2-min-size 1-install/octopus backoff/peering ceph clusters/{openstack three-plus-one} d-balancer/on distro$/{centos_latest} mon_election/connectivity msgr-failures/fastclose rados thrashers/pggrow thrashosds-health workloads/snaps-few-objects} 3
Failure Reason:

Command failed on smithi073 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 97454b0c-daef-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi073:vg_nvme/lv_4'

pass 6248672 2021-07-02 03:56:14 2021-07-02 04:23:08 2021-07-02 04:46:28 0:23:20 0:13:27 0:09:53 smithi wip-focal centos 8.3 rados/thrash/{0-size-min-size-overrides/3-size-2-min-size 1-pg-log-overrides/short_pg_log 2-recovery-overrides/{more-async-recovery} 3-scrub-overrides/{default} backoff/peering ceph clusters/{fixed-2 openstack} crc-failures/bad_map_crc_failure d-balancer/crush-compat mon_election/connectivity msgr-failures/few msgr/async-v2only objectstore/bluestore-hybrid rados supported-random-distro$/{centos_8} thrashers/morepggrow thrashosds-health workloads/redirect_promote_tests} 2
fail 6248673 2021-07-02 03:56:15 2021-07-02 04:23:08 2021-07-02 04:44:19 0:21:11 0:10:20 0:10:51 smithi wip-focal centos 8.2 rados/cephadm/smoke-roleless/{0-distro/centos_8.2_kubic_stable 1-start 2-services/mirror 3-final} 2
Failure Reason:

Command failed on smithi124 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4fb3507c-daef-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi124:vg_nvme/lv_4'

fail 6248674 2021-07-02 03:56:16 2021-07-02 04:23:19 2021-07-02 04:44:35 0:21:16 0:10:26 0:10:50 smithi wip-focal centos 8.2 rados/cephadm/smoke/{distro/centos_8.2_kubic_stable fixed-2 mon_election/classic start} 2
Failure Reason:

Command failed on smithi033 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 6346c696-daef-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi033:vg_nvme/lv_4'

fail 6248675 2021-07-02 03:56:17 2021-07-02 04:23:49 2021-07-02 04:42:51 0:19:02 0:09:33 0:09:29 smithi wip-focal centos 8.2 rados/cephadm/smoke-singlehost/{0-distro$/{centos_8.2_kubic_stable} 1-start 2-services/basic 3-final} 1
Failure Reason:

Command failed on smithi172 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 5e2a9656-daef-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi172:vg_nvme/lv_4'

fail 6248676 2021-07-02 03:56:19 2021-07-02 04:23:49 2021-07-02 04:43:49 0:20:00 0:08:56 0:11:04 smithi wip-focal centos 8.3 rados/cephadm/upgrade/{1-start-distro/1-start-centos_8.3-octopus 2-repo_digest/defaut 3-start-upgrade 4-wait mon_election/classic} 2
Failure Reason:

Command failed on smithi152 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:octopus shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 5d025e80-daef-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi152:vg_nvme/lv_4'

fail 6248677 2021-07-02 03:56:20 2021-07-02 04:24:40 2021-07-02 04:45:26 0:20:46 0:14:44 0:06:02 smithi wip-focal rhel 8.3 rados/cephadm/smoke-roleless/{0-distro/rhel_8.3_kubic_stable 1-start 2-services/nfs-ingress 3-final} 2
Failure Reason:

Command failed on smithi082 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 986f3ed4-daef-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi082:vg_nvme/lv_4'

pass 6248678 2021-07-02 03:56:21 2021-07-02 04:24:40 2021-07-02 04:49:43 0:25:03 0:16:33 0:08:30 smithi wip-focal centos 8.2 rados/cephadm/workunits/{0-distro/centos_8.2_kubic_stable mon_election/connectivity task/test_cephadm} 1
fail 6248679 2021-07-02 03:56:22 2021-07-02 04:24:41 2021-07-02 04:44:15 0:19:34 0:08:43 0:10:51 smithi wip-focal ubuntu 20.04 rados/cephadm/smoke-roleless/{0-distro/ubuntu_20.04 1-start 2-services/nfs-ingress2 3-final} 2
Failure Reason:

Command failed on smithi085 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 57b895f2-daef-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi085:vg_nvme/lv_4'

fail 6248680 2021-07-02 03:56:23 2021-07-02 04:25:02 2021-07-02 04:49:05 0:24:03 0:12:13 0:11:50 smithi wip-focal centos 8.2 rados/cephadm/thrash/{0-distro/centos_8.2_kubic_stable 1-start 2-thrash 3-tasks/small-objects fixed-2 msgr/async root} 2
Failure Reason:

Command failed on smithi139 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 1d5b02c2-daf0-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi139:vg_nvme/lv_4'

pass 6248681 2021-07-02 03:56:24 2021-07-02 04:25:42 2021-07-02 05:09:44 0:44:02 0:33:00 0:11:02 smithi wip-focal centos 8.3 rados/monthrash/{ceph clusters/3-mons mon_election/connectivity msgr-failures/few msgr/async objectstore/bluestore-bitmap rados supported-random-distro$/{centos_8} thrashers/one workloads/rados_api_tests} 2
fail 6248682 2021-07-02 03:56:25 2021-07-02 04:26:13 2021-07-02 04:47:22 0:21:09 0:14:43 0:06:26 smithi wip-focal rhel 8.3 rados/cephadm/smoke/{distro/rhel_8.3_kubic_stable fixed-2 mon_election/connectivity start} 2
Failure Reason:

Command failed on smithi031 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid eb5a0804-daef-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi031:vg_nvme/lv_4'

pass 6248683 2021-07-02 03:56:26 2021-07-02 04:26:13 2021-07-02 04:52:40 0:26:27 0:13:03 0:13:24 smithi wip-focal centos 8.3 rados/thrash-erasure-code-shec/{ceph clusters/{fixed-4 openstack} mon_election/connectivity msgr-failures/few objectstore/bluestore-stupid rados recovery-overrides/{more-async-partial-recovery} supported-random-distro$/{centos_8} thrashers/default thrashosds-health workloads/ec-rados-plugin=shec-k=4-m=3-c=2} 4
fail 6248684 2021-07-02 03:56:27 2021-07-02 04:27:43 2021-07-02 04:57:09 0:29:26 0:22:22 0:07:04 smithi wip-focal rhel 8.3 rados/cephadm/with-work/{0-distro/rhel_8.3_kubic_stable fixed-2 mode/packaged mon_election/connectivity msgr/async start tasks/rados_api_tests} 2
Failure Reason:

Command failed on smithi078 with status 22: 'sudo cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 41f2d8d4-daf1-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi078:vg_nvme/lv_4'

fail 6248685 2021-07-02 03:56:28 2021-07-02 04:28:14 2021-07-02 05:01:24 0:33:10 0:22:43 0:10:27 smithi wip-focal centos 8.3 rados/mgr/{clusters/{2-node-mgr} debug/mgr mon_election/classic objectstore/bluestore-comp-snappy supported-random-distro$/{centos_8} tasks/progress} 2
Failure Reason:

"2021-07-02T04:47:41.400660+0000 mon.a (mon.0) 226 : cluster [WRN] Health check failed: norecover flag(s) set (OSDMAP_FLAGS)" in cluster log

fail 6248686 2021-07-02 03:56:29 2021-07-02 04:28:24 2021-07-02 04:49:52 0:21:28 0:10:28 0:11:00 smithi wip-focal centos 8.2 rados/cephadm/smoke-roleless/{0-distro/centos_8.2_kubic_stable 1-start 2-services/nfs 3-final} 2
Failure Reason:

Command failed on smithi003 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 2490146a-daf0-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi003:vg_nvme/lv_4'

fail 6248687 2021-07-02 03:56:30 2021-07-02 04:28:54 2021-07-02 04:52:43 0:23:49 0:12:51 0:10:58 smithi wip-focal centos 8.3 rados/thrash-old-clients/{0-size-min-size-overrides/3-size-2-min-size 1-install/pacific backoff/peering_and_degraded ceph clusters/{openstack three-plus-one} d-balancer/crush-compat distro$/{centos_latest} mon_election/classic msgr-failures/few rados thrashers/careful thrashosds-health workloads/test_rbd_api} 3
Failure Reason:

Command failed on smithi068 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 91b708e6-daf0-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi068:vg_nvme/lv_4'

fail 6248688 2021-07-02 03:56:31 2021-07-02 04:28:55 2021-07-02 04:46:03 0:17:08 0:07:31 0:09:37 smithi wip-focal ubuntu 20.04 rados/cephadm/upgrade/{1-start-distro/1-start-ubuntu_20.04-15.2.9 2-repo_digest/repo_digest 3-start-upgrade 4-wait mon_election/connectivity} 2
Failure Reason:

Command failed on smithi042 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image docker.io/ceph/ceph:v15.2.9 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid dcb34aae-daef-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi042:vg_nvme/lv_4'

dead 6248689 2021-07-02 03:56:32 2021-07-02 04:28:56 2021-07-02 16:38:10 12:09:14 smithi wip-focal centos 8.3 rados/thrash/{0-size-min-size-overrides/2-size-2-min-size 1-pg-log-overrides/normal_pg_log 2-recovery-overrides/{more-partial-recovery} 3-scrub-overrides/{default} backoff/normal ceph clusters/{fixed-2 openstack} crc-failures/default d-balancer/on mon_election/classic msgr-failures/osd-delay msgr/async-v1only objectstore/bluestore-comp-lz4 rados supported-random-distro$/{centos_8} thrashers/mapgap thrashosds-health workloads/small-objects} 2
Failure Reason:

hit max job timeout

fail 6248690 2021-07-02 03:56:33 2021-07-02 04:29:07 2021-07-02 04:50:52 0:21:45 0:14:59 0:06:46 smithi wip-focal rhel 8.3 rados/cephadm/smoke-roleless/{0-distro/rhel_8.3_kubic_stable 1-start 2-services/nfs2 3-final} 2
Failure Reason:

Command failed on smithi006 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 6a0e7932-daf0-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi006:vg_nvme/lv_4'

fail 6248691 2021-07-02 03:56:35 2021-07-02 04:29:57 2021-07-02 04:54:28 0:24:31 0:12:21 0:12:10 smithi wip-focal centos 8.2 rados/cephadm/thrash/{0-distro/centos_8.2_kubic_stable 1-start 2-thrash 3-tasks/snaps-few-objects fixed-2 msgr/async-v1only root} 2
Failure Reason:

Command failed on smithi159 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid d700142e-daf0-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi159:vg_nvme/lv_4'

fail 6248692 2021-07-02 03:56:36 2021-07-02 04:31:18 2021-07-02 04:50:58 0:19:40 0:08:56 0:10:44 smithi wip-focal ubuntu 20.04 rados/cephadm/smoke-roleless/{0-distro/ubuntu_20.04 1-start 2-services/rgw-ingress 3-final} 2
Failure Reason:

Command failed on smithi115 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 51da361c-daf0-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi115:vg_nvme/lv_4'

fail 6248693 2021-07-02 03:56:37 2021-07-02 04:31:28 2021-07-02 04:50:50 0:19:22 0:08:02 0:11:20 smithi wip-focal ubuntu 20.04 rados/cephadm/smoke/{distro/ubuntu_20.04 fixed-2 mon_election/classic start} 2
Failure Reason:

Command failed on smithi109 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 49b1502e-daf0-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi109:vg_nvme/lv_4'

fail 6248694 2021-07-02 03:56:38 2021-07-02 04:31:29 2021-07-02 04:53:09 0:21:40 0:11:01 0:10:39 smithi wip-focal centos 8.2 rados/cephadm/workunits/{0-distro/centos_8.2_kubic_stable mon_election/connectivity task/test_orch_cli} 1
Failure Reason:

Command failed on smithi167 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid dd2d7666-daf0-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi167:vg_nvme/lv_4'

fail 6248695 2021-07-02 03:56:39 2021-07-02 04:31:59 2021-07-02 04:53:33 0:21:34 0:10:19 0:11:15 smithi wip-focal centos 8.2 rados/cephadm/smoke-roleless/{0-distro/centos_8.2_kubic_stable 1-start 2-services/rgw 3-final} 2
Failure Reason:

Command failed on smithi143 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a6e83f64-daf0-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi143:vg_nvme/lv_4'

fail 6248696 2021-07-02 03:56:40 2021-07-02 04:32:39 2021-07-02 05:01:36 0:28:57 0:22:27 0:06:30 smithi wip-focal rhel 8.3 rados/cephadm/with-work/{0-distro/rhel_8.3_kubic_stable fixed-2 mode/root mon_election/classic msgr/async-v1only start tasks/rados_python} 2
Failure Reason:

Command failed on smithi099 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid e29aefd8-daf1-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi099:vg_nvme/lv_4'

pass 6248697 2021-07-02 03:56:41 2021-07-02 04:32:40 2021-07-02 05:10:12 0:37:32 0:28:05 0:09:27 smithi wip-focal centos 8.3 rados/thrash-erasure-code-isa/{arch/x86_64 ceph clusters/{fixed-2 openstack} mon_election/classic msgr-failures/fastclose objectstore/bluestore-bitmap rados recovery-overrides/{more-async-recovery} supported-random-distro$/{centos_8} thrashers/pggrow thrashosds-health workloads/ec-rados-plugin=isa-k=2-m=1} 2
fail 6248698 2021-07-02 03:56:42 2021-07-02 04:32:41 2021-07-02 04:55:20 0:22:39 0:12:18 0:10:21 smithi wip-focal centos 8.stream rados/dashboard/{centos_8.2_kubic_stable debug/mgr mon_election/connectivity random-objectstore$/{bluestore-bitmap} supported-random-distro$/{centos_8.stream} tasks/e2e} 2
Failure Reason:

Command failed on smithi036 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 048411fc-daf1-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi036:vg_nvme/lv_4'

dead 6248699 2021-07-02 03:56:43 2021-07-02 04:32:41 2021-07-02 04:49:12 0:16:31 smithi wip-focal rhel 8.3 rados/cephadm/smoke-roleless/{0-distro/rhel_8.3_kubic_stable 1-start 2-services/basic 3-final} 2
Failure Reason:

Error reimaging machines: reached maximum tries (60) after waiting for 900 seconds

fail 6248700 2021-07-02 03:56:44 2021-07-02 04:34:12 2021-07-02 04:57:47 0:23:35 0:13:04 0:10:31 smithi wip-focal centos 8.2 rados/cephadm/thrash/{0-distro/centos_8.2_kubic_stable 1-start 2-thrash 3-tasks/rados_api_tests fixed-2 msgr/async-v2only root} 2
Failure Reason:

Command failed on smithi039 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 6460e104-daf1-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi039:vg_nvme/lv_4'

fail 6248701 2021-07-02 03:56:45 2021-07-02 04:34:12 2021-07-02 04:54:51 0:20:39 0:10:41 0:09:58 smithi wip-focal centos 8.2 rados/cephadm/smoke/{distro/centos_8.2_kubic_stable fixed-2 mon_election/connectivity start} 2
Failure Reason:

Command failed on smithi007 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 0f3025dc-daf1-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi007:vg_nvme/lv_4'

fail 6248702 2021-07-02 03:56:46 2021-07-02 04:34:22 2021-07-02 04:53:08 0:18:46 0:09:21 0:09:25 smithi wip-focal centos 8.2 rados/cephadm/smoke-singlehost/{0-distro$/{centos_8.2_kubic_stable} 1-start 2-services/rgw 3-final} 1
Failure Reason:

Command failed on smithi038 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid c4056310-daf0-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi038:vg_nvme/lv_4'

fail 6248703 2021-07-02 03:56:49 2021-07-02 04:34:23 2021-07-02 04:58:41 0:24:18 0:12:04 0:12:14 smithi wip-focal centos 8.3 rados/thrash-old-clients/{0-size-min-size-overrides/2-size-2-min-size 1-install/nautilus-v1only backoff/normal ceph clusters/{openstack three-plus-one} d-balancer/on distro$/{centos_latest} mon_election/connectivity msgr-failures/osd-delay rados thrashers/default thrashosds-health workloads/cache-snaps} 3
Failure Reason:

Command failed on smithi145 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 6440f506-daf1-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi145:vg_nvme/lv_4'

fail 6248704 2021-07-02 03:56:50 2021-07-02 04:35:33 2021-07-02 04:55:25 0:19:52 0:07:46 0:12:06 smithi wip-focal ubuntu 20.04 rados/cephadm/upgrade/{1-start-distro/1-start-ubuntu_20.04 2-repo_digest/defaut 3-start-upgrade 4-wait mon_election/classic} 2
Failure Reason:

Command failed on smithi171 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image docker.io/ceph/ceph:v15.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid f43a124c-daf0-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi171:vg_nvme/lv_4'

fail 6248705 2021-07-02 03:56:51 2021-07-02 04:35:44 2021-07-02 04:55:29 0:19:45 0:08:52 0:10:53 smithi wip-focal ubuntu 20.04 rados/cephadm/smoke-roleless/{0-distro/ubuntu_20.04 1-start 2-services/client-keyring 3-final} 2
Failure Reason:

Command failed on smithi100 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid f26f9766-daf0-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi100:vg_nvme/lv_4'

fail 6248706 2021-07-02 03:56:52 2021-07-02 04:35:54 2021-07-02 04:56:57 0:21:03 0:10:44 0:10:19 smithi wip-focal centos 8.2 rados/cephadm/smoke-roleless/{0-distro/centos_8.2_kubic_stable 1-start 2-services/iscsi 3-final} 2
Failure Reason:

Command failed on smithi052 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 26cc0a9e-daf1-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi052:vg_nvme/lv_4'

pass 6248707 2021-07-02 03:56:53 2021-07-02 04:35:54 2021-07-02 05:03:57 0:28:03 0:17:02 0:11:01 smithi wip-focal ubuntu 20.04 rados/thrash/{0-size-min-size-overrides/3-size-2-min-size 1-pg-log-overrides/short_pg_log 2-recovery-overrides/{more-active-recovery} 3-scrub-overrides/{default} backoff/peering_and_degraded ceph clusters/{fixed-2 openstack} crc-failures/bad_map_crc_failure d-balancer/crush-compat mon_election/connectivity msgr-failures/osd-dispatch-delay msgr/async objectstore/bluestore-low-osd-mem-target rados supported-random-distro$/{ubuntu_latest} thrashers/default thrashosds-health workloads/admin_socket_objecter_requests} 2
pass 6248708 2021-07-02 03:56:54 2021-07-02 04:36:05 2021-07-02 05:01:11 0:25:06 0:16:19 0:08:47 smithi wip-focal centos 8.2 rados/cephadm/workunits/{0-distro/centos_8.2_kubic_stable mon_election/classic task/test_cephadm} 1
fail 6248709 2021-07-02 03:56:55 2021-07-02 04:36:15 2021-07-02 04:59:10 0:22:55 0:12:52 0:10:03 smithi wip-focal centos 8.2 rados/cephadm/thrash/{0-distro/centos_8.2_kubic_stable 1-start 2-thrash 3-tasks/radosbench fixed-2 msgr/async root} 2
Failure Reason:

Command failed on smithi026 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 809f02b0-daf1-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi026:vg_nvme/lv_4'

fail 6248710 2021-07-02 03:56:56 2021-07-02 04:36:15 2021-07-02 04:58:43 0:22:28 0:15:03 0:07:25 smithi wip-focal rhel 8.3 rados/cephadm/smoke-roleless/{0-distro/rhel_8.3_kubic_stable 1-start 2-services/mirror 3-final} 2
Failure Reason:

Command failed on smithi025 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 86428746-daf1-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi025:vg_nvme/lv_4'

pass 6248711 2021-07-02 03:56:57 2021-07-02 04:37:26 2021-07-02 05:06:45 0:29:19 0:23:18 0:06:01 smithi wip-focal rhel 8.4 rados/singleton/{all/mon-config-keys mon_election/connectivity msgr-failures/none msgr/async-v2only objectstore/bluestore-comp-zlib rados supported-random-distro$/{rhel_8}} 1
fail 6248712 2021-07-02 03:56:58 2021-07-02 04:37:26 2021-07-02 04:59:14 0:21:48 0:14:25 0:07:23 smithi wip-focal rhel 8.3 rados/cephadm/smoke/{distro/rhel_8.3_kubic_stable fixed-2 mon_election/classic start} 2
Failure Reason:

Command failed on smithi160 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 8e1cd944-daf1-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi160:vg_nvme/lv_4'

fail 6248713 2021-07-02 03:56:59 2021-07-02 04:37:36 2021-07-02 04:59:18 0:21:42 0:10:34 0:11:08 smithi wip-focal ubuntu 20.04 rados/cephadm/with-work/{0-distro/ubuntu_20.04 fixed-2 mode/packaged mon_election/connectivity msgr/async-v2only start tasks/rados_api_tests} 2
Failure Reason:

Command failed on smithi001 with status 22: 'sudo cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 7cc6b1ba-daf1-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi001:vg_nvme/lv_4'

fail 6248714 2021-07-02 03:57:00 2021-07-02 04:38:27 2021-07-02 04:58:18 0:19:51 0:08:51 0:11:00 smithi wip-focal ubuntu 20.04 rados/cephadm/smoke-roleless/{0-distro/ubuntu_20.04 1-start 2-services/nfs-ingress 3-final} 2
Failure Reason:

Command failed on smithi032 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:76d7281a58c1ad90714792f32480e2d1ae311698 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 597af6ee-daf1-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi032:vg_nvme/lv_4'

fail 6248715 2021-07-02 03:57:01 2021-07-02 04:39:07 2021-07-02 04:57:50 0:18:43 0:09:03 0:09:40 smithi wip-focal centos 8.3 rados/cephadm/upgrade/{1-start-distro/1-start-centos_8.3-octopus 2-repo_digest/repo_digest 3-start-upgrade 4-wait mon_election/connectivity} 2
Failure Reason:

Command failed on smithi050 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:octopus shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 7d841e1c-daf1-11eb-8c1c-001a4aab830c -- ceph orch daemon add osd smithi050:vg_nvme/lv_4'