Status Job ID Links Posted Started Updated
Runtime
Duration
In Waiting
Machine Teuthology Branch OS Type OS Version Description Nodes
fail 7058160 2022-10-08 00:10:27 2022-10-08 00:11:09 2022-10-08 00:34:09 0:23:00 0:14:50 0:08:10 smithi main centos 8.stream orch:cephadm/osds/{0-distro/centos_8.stream_container_tools 0-nvme-loop 1-start 2-ops/rm-zap-add} 2
Failure Reason:

Command failed on smithi085 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:32a5ea00c4bee0c51c66aa8902716d7900f94c76 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid b96a661e-469f-11ed-8436-001a4aab830c -- bash -c \'set -e\nset -x\nceph orch ps\nceph orch device ls\nDEVID=$(ceph device ls | grep osd.1 | awk \'"\'"\'{print $1}\'"\'"\')\nHOST=$(ceph orch device ls | grep $DEVID | awk \'"\'"\'{print $1}\'"\'"\')\nDEV=$(ceph orch device ls | grep $DEVID | awk \'"\'"\'{print $2}\'"\'"\')\necho "host $HOST, dev $DEV, devid $DEVID"\nceph orch osd rm 1\nwhile ceph orch osd rm status | grep ^1 ; do sleep 5 ; done\nceph orch device zap $HOST $DEV --force\nceph orch daemon add osd $HOST:$DEV\nwhile ! ceph osd dump | grep osd.1 | grep up ; do sleep 5 ; done\n\''

fail 7058161 2022-10-08 00:10:28 2022-10-08 00:11:10 2022-10-08 00:21:57 0:10:47 0:04:16 0:06:31 smithi main centos 8.stream orch:cephadm/workunits/{0-distro/rhel_8.6_container_tools_3.0 agent/on mon_election/connectivity task/test_iscsi_pids_limit/{centos_8.stream_container_tools test_iscsi_pids_limit}} 1
Failure Reason:

Command failed on smithi176 with status 1: 'TESTDIR=/home/ubuntu/cephtest bash -s'

fail 7058162 2022-10-08 00:10:29 2022-10-08 00:11:10 2022-10-10 12:45:20 2 days, 12:34:10 3:13:13 2 days, 9:20:57 smithi main rhel 8.6 orch:cephadm/with-work/{0-distro/rhel_8.6_container_tools_3.0 fixed-2 mode/packaged mon_election/classic msgr/async-v1only start tasks/rados_python} 2
Failure Reason:

{'smithi096.front.sepia.ceph.com': {'_ansible_no_log': False, 'changed': False, 'invocation': {'module_args': {'allow_downgrade': False, 'autoremove': False, 'bugfix': False, 'conf_file': None, 'disable_excludes': None, 'disable_gpg_check': False, 'disable_plugin': [], 'disablerepo': [], 'download_dir': None, 'download_only': False, 'enable_plugin': [], 'enablerepo': [], 'exclude': [], 'install_repoquery': True, 'install_weak_deps': True, 'installroot': '/', 'list': None, 'lock_timeout': 30, 'name': ['krb5-workstation'], 'releasever': None, 'security': False, 'skip_broken': False, 'state': 'present', 'update_cache': False, 'update_only': False, 'validate_certs': True}}, 'msg': "Loading repository 'rhel-8-for-x86_64-baseos-rpms' has failed", 'rc': 1, 'results': []}, 'smithi163.front.sepia.ceph.com': {'_ansible_no_log': False, 'changed': False, 'invocation': {'module_args': {'allow_downgrade': False, 'autoremove': False, 'bugfix': False, 'conf_file': None, 'disable_excludes': None, 'disable_gpg_check': False, 'disable_plugin': [], 'disablerepo': [], 'download_dir': None, 'download_only': False, 'enable_plugin': [], 'enablerepo': [], 'exclude': [], 'install_repoquery': True, 'install_weak_deps': True, 'installroot': '/', 'list': None, 'lock_timeout': 30, 'name': ['krb5-workstation'], 'releasever': None, 'security': False, 'skip_broken': False, 'state': 'present', 'update_cache': False, 'update_only': False, 'validate_certs': True}}, 'msg': "Loading repository 'rhel-8-for-x86_64-baseos-rpms' has failed", 'rc': 1, 'results': []}}

pass 7058163 2022-10-08 00:10:30 2022-10-08 00:11:10 2022-10-08 00:36:47 0:25:37 0:15:51 0:09:46 smithi main centos 8.stream orch:cephadm/smoke-roleless/{0-distro/centos_8.stream_container_tools_crun 0-nvme-loop 1-start 2-services/jaeger 3-final} 2
pass 7058164 2022-10-08 00:10:31 2022-10-08 00:11:11 2022-10-08 00:54:09 0:42:58 0:35:40 0:07:18 smithi main centos 8.stream orch:cephadm/mds_upgrade_sequence/{bluestore-bitmap centos_8.stream_container_tools conf/{client mds mon osd} fail_fs/no overrides/{ignorelist_health ignorelist_wrongly_marked_down pg-warn syntax} roles tasks/{0-from/v16.2.4 1-volume/{0-create 1-ranks/1 2-allow_standby_replay/no 3-inline/no 4-verify} 2-client 3-upgrade-mgr-staggered 4-config-upgrade/{fail_fs} 5-upgrade-with-workload 6-verify}} 2
pass 7058165 2022-10-08 00:10:33 2022-10-08 00:11:11 2022-10-08 01:03:10 0:51:59 0:40:37 0:11:22 smithi main centos 8.stream orch:cephadm/mgr-nfs-upgrade/{0-centos_8.stream_container_tools 1-bootstrap/16.2.0 1-start 2-nfs 3-upgrade-with-workload 4-final} 2
pass 7058166 2022-10-08 00:10:34 2022-10-08 00:11:11 2022-10-08 00:35:26 0:24:15 0:16:51 0:07:24 smithi main centos 8.stream orch:cephadm/orchestrator_cli/{0-random-distro$/{centos_8.stream_container_tools} 2-node-mgr agent/off orchestrator_cli} 2
fail 7058167 2022-10-08 00:10:35 2022-10-08 00:11:12 2022-10-08 00:33:36 0:22:24 0:11:17 0:11:07 smithi main centos 8.stream orch:cephadm/rbd_iscsi/{0-single-container-host base/install cluster/{fixed-3 openstack} workloads/cephadm_iscsi} 3
Failure Reason:

Command failed on smithi040 with status 1: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:32a5ea00c4bee0c51c66aa8902716d7900f94c76 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 251f70fc-46a0-11ed-8436-001a4aab830c -- ceph-volume lvm zap /dev/vg_nvme/lv_4'

fail 7058168 2022-10-08 00:10:36 2022-10-08 00:11:12 2022-10-08 00:27:17 0:16:05 0:07:40 0:08:25 smithi main centos 8.stream orch:cephadm/smoke-singlehost/{0-random-distro$/{centos_8.stream_container_tools} 1-start 2-services/basic 3-final} 1
Failure Reason:

Command failed on smithi111 with status 1: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:32a5ea00c4bee0c51c66aa8902716d7900f94c76 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 558d5084-469f-11ed-8436-001a4aab830c -- ceph-volume lvm zap /dev/vg_nvme/lv_4'

fail 7058169 2022-10-08 00:10:37 2022-10-08 00:11:12 2022-10-08 01:05:56 0:54:44 0:43:11 0:11:33 smithi main centos 8.stream orch:cephadm/upgrade/{1-start-distro/1-start-centos_8.stream_container-tools 2-repo_digest/defaut 3-upgrade/staggered 4-wait 5-upgrade-ls agent/off mon_election/classic} 2
Failure Reason:

Command failed on smithi007 with status 22: "sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v16.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 16f39562-46a0-11ed-8436-001a4aab830c -e sha1=32a5ea00c4bee0c51c66aa8902716d7900f94c76 -- bash -c 'ceph orch upgrade start --image quay.ceph.io/ceph-ci/ceph:$sha1 --services rgw.r'"

fail 7058170 2022-10-08 00:10:38 2022-10-08 00:11:13 2022-10-08 00:40:18 0:29:05 0:23:11 0:05:54 smithi main rhel 8.6 orch:cephadm/with-work/{0-distro/rhel_8.6_container_tools_rhel8 fixed-2 mode/root mon_election/connectivity msgr/async-v2only start tasks/rotate-keys} 2
Failure Reason:

Command failed on smithi063 with status 1: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:32a5ea00c4bee0c51c66aa8902716d7900f94c76 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid dd0e69f2-46a0-11ed-8436-001a4aab830c -- ceph-volume lvm zap /dev/vg_nvme/lv_4'

fail 7058171 2022-10-08 00:10:39 2022-10-08 00:11:13 2022-10-08 00:41:44 0:30:31 0:19:33 0:10:58 smithi main rhel 8.6 orch:cephadm/workunits/{0-distro/rhel_8.6_container_tools_rhel8 agent/off mon_election/classic task/test_nfs} 1
Failure Reason:

Command failed on smithi027 with status 1: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:32a5ea00c4bee0c51c66aa8902716d7900f94c76 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 680666d6-46a1-11ed-8436-001a4aab830c -- ceph-volume lvm zap /dev/vg_nvme/lv_4'

pass 7058172 2022-10-08 00:10:40 2022-10-08 00:11:14 2022-10-08 00:40:09 0:28:55 0:22:36 0:06:19 smithi main rhel 8.6 orch:cephadm/smoke-roleless/{0-distro/rhel_8.6_container_tools_3.0 0-nvme-loop 1-start 2-services/mirror 3-final} 2
fail 7058173 2022-10-08 00:10:41 2022-10-08 00:11:14 2022-10-08 00:37:41 0:26:27 0:19:08 0:07:19 smithi main rhel 8.6 orch:cephadm/smoke/{0-distro/rhel_8.6_container_tools_rhel8 0-nvme-loop agent/off fixed-2 mon_election/classic start} 2
Failure Reason:

Command failed on smithi037 with status 1: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:32a5ea00c4bee0c51c66aa8902716d7900f94c76 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 7d008504-46a0-11ed-8436-001a4aab830c -- ceph-volume lvm zap /dev/nvme4n1'

fail 7058174 2022-10-08 00:10:43 2022-10-08 00:11:14 2022-10-08 00:40:07 0:28:53 0:23:10 0:05:43 smithi main rhel 8.6 orch:cephadm/thrash/{0-distro/rhel_8.6_container_tools_rhel8 1-start 2-thrash 3-tasks/rados_api_tests fixed-2 msgr/async root} 2
Failure Reason:

Command failed on smithi018 with status 1: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:32a5ea00c4bee0c51c66aa8902716d7900f94c76 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ce2c6902-46a0-11ed-8436-001a4aab830c -- ceph-volume lvm zap /dev/vg_nvme/lv_4'

fail 7058175 2022-10-08 00:10:44 2022-10-08 00:11:15 2022-10-08 00:42:33 0:31:18 0:17:59 0:13:19 smithi main ubuntu 20.04 orch:cephadm/with-work/{0-distro/ubuntu_20.04 fixed-2 mode/packaged mon_election/classic msgr/async start tasks/rados_api_tests} 2
Failure Reason:

Command failed on smithi049 with status 1: 'sudo cephadm --image quay.ceph.io/ceph-ci/ceph:32a5ea00c4bee0c51c66aa8902716d7900f94c76 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 99d320d8-46a0-11ed-8436-001a4aab830c -- ceph-volume lvm zap /dev/vg_nvme/lv_4'

fail 7058176 2022-10-08 00:10:45 2022-10-08 00:11:15 2022-10-08 00:32:02 0:20:47 0:12:35 0:08:12 smithi main ubuntu 20.04 orch:cephadm/workunits/{0-distro/ubuntu_20.04 agent/on mon_election/connectivity task/test_orch_cli} 1
Failure Reason:

Command failed on smithi035 with status 1: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:32a5ea00c4bee0c51c66aa8902716d7900f94c76 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 047d0a62-46a0-11ed-8436-001a4aab830c -- ceph-volume lvm zap /dev/vg_nvme/lv_4'

pass 7058177 2022-10-08 00:10:46 2022-10-08 00:11:16 2022-10-08 00:45:48 0:34:32 0:26:19 0:08:13 smithi main rhel 8.6 orch:cephadm/smoke-roleless/{0-distro/rhel_8.6_container_tools_rhel8 0-nvme-loop 1-start 2-services/nfs-ingress-rgw-bucket 3-final} 2
fail 7058178 2022-10-08 00:10:47 2022-10-08 00:11:16 2022-10-08 00:39:14 0:27:58 0:16:49 0:11:09 smithi main centos 8.stream orch:cephadm/with-work/{0-distro/centos_8.stream_container_tools fixed-2 mode/root mon_election/connectivity msgr/async-v1only start tasks/rados_python} 2
Failure Reason:

Command failed on smithi012 with status 1: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:32a5ea00c4bee0c51c66aa8902716d7900f94c76 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 97db7500-46a0-11ed-8436-001a4aab830c -- ceph-volume lvm zap /dev/vg_nvme/lv_4'

pass 7058179 2022-10-08 00:10:48 2022-10-08 00:11:16 2022-10-08 00:33:01 0:21:45 0:13:58 0:07:47 smithi main centos 8.stream orch:cephadm/osds/{0-distro/centos_8.stream_container_tools_crun 0-nvme-loop 1-start 2-ops/rm-zap-flag} 2
fail 7058180 2022-10-08 00:10:49 2022-10-08 00:11:17 2022-10-08 00:39:10 0:27:53 0:16:30 0:11:23 smithi main centos 8.stream orch:cephadm/mds_upgrade_sequence/{bluestore-bitmap centos_8.stream_container_tools conf/{client mds mon osd} fail_fs/yes overrides/{ignorelist_health ignorelist_wrongly_marked_down pg-warn syntax} roles tasks/{0-from/pacific 1-volume/{0-create 1-ranks/2 2-allow_standby_replay/yes 3-inline/yes 4-verify} 2-client 3-upgrade-mgr-staggered 4-config-upgrade/{fail_fs} 5-upgrade-with-workload 6-verify}} 2
Failure Reason:

Command failed on smithi172 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:pacific shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 85add710-46a0-11ed-8436-001a4aab830c -- ceph orch daemon add osd smithi172:vg_nvme/lv_4'

fail 7058181 2022-10-08 00:10:51 2022-10-08 00:11:17 2022-10-08 00:42:37 0:31:20 0:20:51 0:10:29 smithi main centos 8.stream orch:cephadm/workunits/{0-distro/centos_8.stream_container_tools agent/off mon_election/classic task/test_orch_cli_mon} 5
Failure Reason:

Command failed on smithi033 with status 1: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:32a5ea00c4bee0c51c66aa8902716d7900f94c76 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid c317dc72-46a0-11ed-8436-001a4aab830c -- ceph-volume lvm zap /dev/vg_nvme/lv_4'

fail 7058182 2022-10-08 00:10:52 2022-10-08 00:11:18 2022-10-08 00:29:53 0:18:35 0:12:58 0:05:37 smithi main centos 8.stream orch:cephadm/with-work/{0-distro/centos_8.stream_container_tools_crun fixed-2 mode/packaged mon_election/classic msgr/async-v2only start tasks/rotate-keys} 2
Failure Reason:

Command failed on smithi005 with status 1: 'sudo cephadm --image quay.ceph.io/ceph-ci/ceph:32a5ea00c4bee0c51c66aa8902716d7900f94c76 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid b9d44912-469f-11ed-8436-001a4aab830c -- ceph-volume lvm zap /dev/vg_nvme/lv_4'

pass 7058183 2022-10-08 00:10:53 2022-10-08 00:11:18 2022-10-08 00:45:06 0:33:48 0:20:14 0:13:34 smithi main ubuntu 20.04 orch:cephadm/smoke-roleless/{0-distro/ubuntu_20.04 0-nvme-loop 1-start 2-services/nfs-ingress-rgw-user 3-final} 2
fail 7058184 2022-10-08 00:10:54 2022-10-08 00:11:18 2022-10-08 00:33:13 0:21:55 0:11:18 0:10:37 smithi main ubuntu 20.04 orch:cephadm/smoke/{0-distro/ubuntu_20.04 0-nvme-loop agent/on fixed-2 mon_election/connectivity start} 2
Failure Reason:

Command failed on smithi044 with status 1: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:32a5ea00c4bee0c51c66aa8902716d7900f94c76 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid dd749f66-469f-11ed-8436-001a4aab830c -- ceph-volume lvm zap /dev/nvme4n1'

fail 7058185 2022-10-08 00:10:55 2022-10-08 00:11:19 2022-10-08 00:41:49 0:30:30 0:16:26 0:14:04 smithi main ubuntu 20.04 orch:cephadm/thrash/{0-distro/ubuntu_20.04 1-start 2-thrash 3-tasks/radosbench fixed-2 msgr/async-v1only root} 2
Failure Reason:

Command failed on smithi109 with status 1: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:32a5ea00c4bee0c51c66aa8902716d7900f94c76 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid bee08d70-46a0-11ed-8436-001a4aab830c -- ceph-volume lvm zap /dev/vg_nvme/lv_4'

fail 7058186 2022-10-08 00:10:56 2022-10-08 00:11:19 2022-10-08 00:43:03 0:31:44 0:23:44 0:08:00 smithi main rhel 8.6 orch:cephadm/with-work/{0-distro/rhel_8.6_container_tools_3.0 fixed-2 mode/root mon_election/connectivity msgr/async start tasks/rados_api_tests} 2
Failure Reason:

Command failed on smithi186 with status 1: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:32a5ea00c4bee0c51c66aa8902716d7900f94c76 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid f7a01270-46a0-11ed-8436-001a4aab830c -- ceph-volume lvm zap /dev/vg_nvme/lv_4'

pass 7058187 2022-10-08 00:10:57 2022-10-08 00:11:20 2022-10-08 00:32:10 0:20:50 0:12:53 0:07:57 smithi main centos 8.stream orch:cephadm/workunits/{0-distro/centos_8.stream_container_tools_crun agent/on mon_election/connectivity task/test_adoption} 1
pass 7058188 2022-10-08 00:10:59 2022-10-08 00:11:20 2022-10-08 00:40:14 0:28:54 0:17:53 0:11:01 smithi main centos 8.stream orch:cephadm/smoke-roleless/{0-distro/centos_8.stream_container_tools 0-nvme-loop 1-start 2-services/nfs-ingress 3-final} 2
fail 7058189 2022-10-08 00:11:00 2022-10-08 00:11:20 2022-10-08 00:42:10 0:30:50 0:22:57 0:07:53 smithi main rhel 8.6 orch:cephadm/with-work/{0-distro/rhel_8.6_container_tools_rhel8 fixed-2 mode/packaged mon_election/classic msgr/async-v1only start tasks/rados_python} 2
Failure Reason:

Command failed on smithi134 with status 1: 'sudo cephadm --image quay.ceph.io/ceph-ci/ceph:32a5ea00c4bee0c51c66aa8902716d7900f94c76 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 2f714098-46a1-11ed-8436-001a4aab830c -- ceph-volume lvm zap /dev/vg_nvme/lv_4'

pass 7058190 2022-10-08 00:11:01 2022-10-08 00:11:21 2022-10-08 00:54:13 0:42:52 0:35:15 0:07:37 smithi main centos 8.stream orch:cephadm/mds_upgrade_sequence/{bluestore-bitmap centos_8.stream_container_tools conf/{client mds mon osd} fail_fs/no overrides/{ignorelist_health ignorelist_wrongly_marked_down pg-warn syntax} roles tasks/{0-from/v16.2.4 1-volume/{0-create 1-ranks/1 2-allow_standby_replay/yes 3-inline/no 4-verify} 2-client 3-upgrade-mgr-staggered 4-config-upgrade/{fail_fs} 5-upgrade-with-workload 6-verify}} 2
pass 7058191 2022-10-08 00:11:02 2022-10-08 00:11:21 2022-10-08 00:53:48 0:42:27 0:35:14 0:07:13 smithi main ubuntu 20.04 orch:cephadm/upgrade/{1-start-distro/1-start-ubuntu_20.04 2-repo_digest/repo_digest 3-upgrade/simple 4-wait 5-upgrade-ls agent/on mon_election/connectivity} 2
pass 7058192 2022-10-08 00:11:03 2022-10-08 00:11:22 2022-10-08 00:45:26 0:34:04 0:27:15 0:06:49 smithi main rhel 8.6 orch:cephadm/workunits/{0-distro/rhel_8.6_container_tools_3.0 agent/off mon_election/classic task/test_cephadm} 1
fail 7058193 2022-10-08 00:11:04 2022-10-08 00:11:32 2022-10-08 00:41:24 0:29:52 0:22:39 0:07:13 smithi main rhel 8.6 orch:cephadm/osds/{0-distro/rhel_8.6_container_tools_3.0 0-nvme-loop 1-start 2-ops/rm-zap-wait} 2
Failure Reason:

Command failed on smithi080 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:32a5ea00c4bee0c51c66aa8902716d7900f94c76 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 5e0ce052-46a0-11ed-8436-001a4aab830c -- bash -c \'set -e\nset -x\nceph orch ps\nceph orch device ls\nDEVID=$(ceph device ls | grep osd.1 | awk \'"\'"\'{print $1}\'"\'"\')\nHOST=$(ceph orch device ls | grep $DEVID | awk \'"\'"\'{print $1}\'"\'"\')\nDEV=$(ceph orch device ls | grep $DEVID | awk \'"\'"\'{print $2}\'"\'"\')\necho "host $HOST, dev $DEV, devid $DEVID"\nceph orch osd rm 1\nwhile ceph orch osd rm status | grep ^1 ; do sleep 5 ; done\nceph orch device zap $HOST $DEV --force\nwhile ! ceph osd dump | grep osd.1 | grep up ; do sleep 5 ; done\n\''

fail 7058194 2022-10-08 00:11:05 2022-10-08 00:11:32 2022-10-08 00:43:50 0:32:18 0:17:03 0:15:15 smithi main ubuntu 20.04 orch:cephadm/with-work/{0-distro/ubuntu_20.04 fixed-2 mode/root mon_election/connectivity msgr/async-v2only start tasks/rotate-keys} 2
Failure Reason:

Command failed on smithi150 with status 1: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:32a5ea00c4bee0c51c66aa8902716d7900f94c76 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid bf5f5b0a-46a0-11ed-8436-001a4aab830c -- ceph-volume lvm zap /dev/vg_nvme/lv_4'

pass 7058195 2022-10-08 00:11:06 2022-10-08 00:11:33 2022-10-08 00:37:55 0:26:22 0:18:16 0:08:06 smithi main centos 8.stream orch:cephadm/smoke-roleless/{0-distro/centos_8.stream_container_tools_crun 0-nvme-loop 1-start 2-services/nfs-ingress2 3-final} 2
pass 7058196 2022-10-08 00:11:08 2022-10-08 00:11:33 2022-10-08 00:35:36 0:24:03 0:12:40 0:11:23 smithi main rhel 8.6 orch:cephadm/workunits/{0-distro/rhel_8.6_container_tools_rhel8 agent/on mon_election/connectivity task/test_cephadm_repos} 1
pass 7058197 2022-10-08 00:11:09 2022-10-08 00:11:33 2022-10-08 00:59:41 0:48:08 0:41:09 0:06:59 smithi main centos 8.stream orch:cephadm/mgr-nfs-upgrade/{0-centos_8.stream_container_tools 1-bootstrap/16.2.4 1-start 2-nfs 3-upgrade-with-workload 4-final} 2
fail 7058198 2022-10-08 00:11:10 2022-10-08 00:11:34 2022-10-08 00:29:40 0:18:06 0:09:34 0:08:32 smithi main centos 8.stream orch:cephadm/smoke/{0-distro/centos_8.stream_container_tools 0-nvme-loop agent/on fixed-2 mon_election/classic start} 2
Failure Reason:

Command failed on smithi059 with status 1: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:32a5ea00c4bee0c51c66aa8902716d7900f94c76 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 9c6168d8-469f-11ed-8436-001a4aab830c -- ceph-volume lvm zap /dev/nvme4n1'

fail 7058199 2022-10-08 00:11:11 2022-10-08 00:11:34 2022-10-08 00:39:42 0:28:08 0:16:12 0:11:56 smithi main centos 8.stream orch:cephadm/thrash/{0-distro/centos_8.stream_container_tools 1-start 2-thrash 3-tasks/small-objects fixed-2 msgr/async-v2only root} 2
Failure Reason:

Command failed on smithi081 with status 1: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:32a5ea00c4bee0c51c66aa8902716d7900f94c76 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a400cf10-46a0-11ed-8436-001a4aab830c -- ceph-volume lvm zap /dev/vg_nvme/lv_4'

fail 7058200 2022-10-08 00:11:12 2022-10-08 00:11:34 2022-10-08 00:36:28 0:24:54 0:15:58 0:08:56 smithi main centos 8.stream orch:cephadm/with-work/{0-distro/centos_8.stream_container_tools fixed-2 mode/packaged mon_election/classic msgr/async-v2only start tasks/rados_api_tests} 2
Failure Reason:

Command failed on smithi046 with status 1: 'sudo cephadm --image quay.ceph.io/ceph-ci/ceph:32a5ea00c4bee0c51c66aa8902716d7900f94c76 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 52370fd2-46a0-11ed-8436-001a4aab830c -- ceph-volume lvm zap /dev/vg_nvme/lv_4'

pass 7058201 2022-10-08 00:11:14 2022-10-08 00:44:53 1465 smithi main rhel 8.6 orch:cephadm/smoke-roleless/{0-distro/rhel_8.6_container_tools_3.0 0-nvme-loop 1-start 2-services/nfs 3-final} 2
fail 7058202 2022-10-08 00:11:15 2022-10-08 00:30:45 706 smithi main centos 8.stream orch:cephadm/workunits/{0-distro/ubuntu_20.04 agent/off mon_election/classic task/test_iscsi_pids_limit/{centos_8.stream_container_tools test_iscsi_pids_limit}} 1
Failure Reason:

Command failed on smithi138 with status 1: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:32a5ea00c4bee0c51c66aa8902716d7900f94c76 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid c5031fe8-469f-11ed-8436-001a4aab830c -- ceph-volume lvm zap /dev/vg_nvme/lv_4'

pass 7058203 2022-10-08 00:11:16 2022-10-08 00:54:05 2107 smithi main centos 8.stream orch:cephadm/mds_upgrade_sequence/{bluestore-bitmap centos_8.stream_container_tools conf/{client mds mon osd} fail_fs/yes overrides/{ignorelist_health ignorelist_wrongly_marked_down pg-warn syntax} roles tasks/{0-from/pacific 1-volume/{0-create 1-ranks/2 2-allow_standby_replay/no 3-inline/yes 4-verify} 2-client 3-upgrade-mgr-staggered 4-config-upgrade/{fail_fs} 5-upgrade-with-workload 6-verify}} 2
fail 7058204 2022-10-08 00:11:18 2022-10-08 00:36:16 970 smithi main centos 8.stream orch:cephadm/with-work/{0-distro/centos_8.stream_container_tools_crun fixed-2 mode/root mon_election/connectivity msgr/async start tasks/rados_python} 2
Failure Reason:

Command failed on smithi171 with status 1: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:32a5ea00c4bee0c51c66aa8902716d7900f94c76 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4bd10c88-46a0-11ed-8436-001a4aab830c -- ceph-volume lvm zap /dev/vg_nvme/lv_4'

pass 7058205 2022-10-08 00:11:19 2022-10-08 00:11:36 2022-10-08 00:43:40 0:32:04 0:24:24 0:07:40 smithi main rhel 8.6 orch:cephadm/smoke-roleless/{0-distro/rhel_8.6_container_tools_rhel8 0-nvme-loop 1-start 2-services/nfs2 3-final} 2
fail 7058206 2022-10-08 00:11:20 2022-10-08 00:11:37 2022-10-08 00:44:04 0:32:27 0:23:39 0:08:48 smithi main rhel 8.6 orch:cephadm/with-work/{0-distro/rhel_8.6_container_tools_3.0 fixed-2 mode/packaged mon_election/classic msgr/async-v1only start tasks/rotate-keys} 2
Failure Reason:

Command failed on smithi120 with status 1: 'sudo cephadm --image quay.ceph.io/ceph-ci/ceph:32a5ea00c4bee0c51c66aa8902716d7900f94c76 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 2c024358-46a1-11ed-8436-001a4aab830c -- ceph-volume lvm zap /dev/vg_nvme/lv_4'

fail 7058207 2022-10-08 00:11:21 2022-10-08 00:11:37 2022-10-08 00:30:59 0:19:22 0:11:33 0:07:49 smithi main centos 8.stream orch:cephadm/workunits/{0-distro/centos_8.stream_container_tools agent/on mon_election/connectivity task/test_nfs} 1
Failure Reason:

Command failed on smithi096 with status 1: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:32a5ea00c4bee0c51c66aa8902716d7900f94c76 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid cf431454-469f-11ed-8436-001a4aab830c -- ceph-volume lvm zap /dev/vg_nvme/lv_4'

pass 7058208 2022-10-08 00:11:23 2022-10-08 00:11:37 2022-10-08 00:45:06 0:33:29 0:24:05 0:09:24 smithi main rhel 8.6 orch:cephadm/osds/{0-distro/rhel_8.6_container_tools_rhel8 0-nvme-loop 1-start 2-ops/rmdir-reactivate} 2
fail 7058209 2022-10-08 00:11:24 2022-10-08 00:11:38 2022-10-08 00:29:59 0:18:21 0:10:46 0:07:35 smithi main centos 8.stream orch:cephadm/smoke/{0-distro/centos_8.stream_container_tools_crun 0-nvme-loop agent/off fixed-2 mon_election/connectivity start} 2
Failure Reason:

Command failed on smithi032 with status 1: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:32a5ea00c4bee0c51c66aa8902716d7900f94c76 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid b8081a50-469f-11ed-8436-001a4aab830c -- ceph-volume lvm zap /dev/nvme4n1'

fail 7058210 2022-10-08 00:11:25 2022-10-08 00:11:38 2022-10-08 00:36:16 0:24:38 0:15:58 0:08:40 smithi main centos 8.stream orch:cephadm/thrash/{0-distro/centos_8.stream_container_tools_crun 1-start 2-thrash 3-tasks/snaps-few-objects fixed-2 msgr/async root} 2
Failure Reason:

Command failed on smithi099 with status 1: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:32a5ea00c4bee0c51c66aa8902716d7900f94c76 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 4ae18b86-46a0-11ed-8436-001a4aab830c -- ceph-volume lvm zap /dev/vg_nvme/lv_4'

fail 7058211 2022-10-08 00:11:26 2022-10-08 00:11:38 2022-10-08 00:43:09 0:31:31 0:20:55 0:10:36 smithi main rhel 8.6 orch:cephadm/with-work/{0-distro/rhel_8.6_container_tools_rhel8 fixed-2 mode/root mon_election/connectivity msgr/async-v2only start tasks/rados_api_tests} 2
Failure Reason:

Command failed on smithi097 with status 1: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:32a5ea00c4bee0c51c66aa8902716d7900f94c76 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 6d149580-46a1-11ed-8436-001a4aab830c -- ceph-volume lvm zap /dev/vg_nvme/lv_4'

pass 7058212 2022-10-08 00:11:27 2022-10-08 00:11:39 2022-10-08 00:51:56 0:40:17 0:26:09 0:14:08 smithi main ubuntu 20.04 orch:cephadm/smoke-roleless/{0-distro/ubuntu_20.04 0-nvme-loop 1-start 2-services/rgw-ingress 3-final} 2
fail 7058213 2022-10-08 00:11:28 2022-10-08 00:11:39 2022-10-08 00:37:14 0:25:35 0:14:37 0:10:58 smithi main centos 8.stream orch:cephadm/workunits/{0-distro/centos_8.stream_container_tools_crun agent/off mon_election/classic task/test_orch_cli} 1
Failure Reason:

Command failed on smithi057 with status 1: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:32a5ea00c4bee0c51c66aa8902716d7900f94c76 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 70032b36-46a0-11ed-8436-001a4aab830c -- ceph-volume lvm zap /dev/vg_nvme/lv_4'

pass 7058214 2022-10-08 00:11:30 2022-10-08 00:11:39 2022-10-08 00:54:39 0:43:00 0:34:11 0:08:49 smithi main centos 8.stream orch:cephadm/mds_upgrade_sequence/{bluestore-bitmap centos_8.stream_container_tools conf/{client mds mon osd} fail_fs/no overrides/{ignorelist_health ignorelist_wrongly_marked_down pg-warn syntax} roles tasks/{0-from/v16.2.4 1-volume/{0-create 1-ranks/2 2-allow_standby_replay/no 3-inline/no 4-verify} 2-client 3-upgrade-mgr-staggered 4-config-upgrade/{fail_fs} 5-upgrade-with-workload 6-verify}} 2
pass 7058215 2022-10-08 00:11:31 2022-10-08 00:11:40 2022-10-08 00:44:36 0:32:56 0:14:35 0:18:21 smithi main centos 8.stream orch:cephadm/orchestrator_cli/{0-random-distro$/{centos_8.stream_container_tools} 2-node-mgr agent/on orchestrator_cli} 2
fail 7058216 2022-10-08 00:11:32 2022-10-08 00:30:00 2022-10-08 00:43:00 0:13:00 0:05:52 0:07:08 smithi main centos 8.stream orch:cephadm/smoke-singlehost/{0-random-distro$/{centos_8.stream_container_tools_crun} 1-start 2-services/rgw 3-final} 1
Failure Reason:

Command failed on smithi153 with status 1: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:32a5ea00c4bee0c51c66aa8902716d7900f94c76 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 95e0cfe2-46a1-11ed-8436-001a4aab830c -- ceph-volume lvm zap /dev/vg_nvme/lv_4'

fail 7058217 2022-10-08 00:11:33 2022-10-08 01:17:21 2385 smithi main centos 8.stream orch:cephadm/upgrade/{1-start-distro/1-start-centos_8.stream_container-tools 2-repo_digest/defaut 3-upgrade/staggered 4-wait 5-upgrade-ls agent/on mon_election/classic} 2
Failure Reason:

Command failed on smithi032 with status 22: "sudo /home/ubuntu/cephtest/cephadm --image quay.io/ceph/ceph:v16.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a51f9006-46a1-11ed-8436-001a4aab830c -e sha1=32a5ea00c4bee0c51c66aa8902716d7900f94c76 -- bash -c 'ceph orch upgrade start --image quay.ceph.io/ceph-ci/ceph:$sha1 --services rgw.r'"

fail 7058218 2022-10-08 00:11:34 2022-10-08 00:52:26 676 smithi main ubuntu 20.04 orch:cephadm/with-work/{0-distro/ubuntu_20.04 fixed-2 mode/packaged mon_election/classic msgr/async start tasks/rados_python} 2
Failure Reason:

Command failed on smithi096 with status 1: 'sudo cephadm --image quay.ceph.io/ceph-ci/ceph:32a5ea00c4bee0c51c66aa8902716d7900f94c76 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a9dd2c6a-46a2-11ed-8436-001a4aab830c -- ceph-volume lvm zap /dev/vg_nvme/lv_4'

pass 7058219 2022-10-08 00:11:36 2022-10-08 00:31:01 2022-10-08 00:51:20 0:20:19 0:12:52 0:07:27 smithi main centos 8.stream orch:cephadm/smoke-roleless/{0-distro/centos_8.stream_container_tools 0-nvme-loop 1-start 2-services/rgw 3-final} 2
fail 7058220 2022-10-08 00:11:37 2022-10-08 00:32:12 2022-10-08 00:59:12 0:27:00 0:19:02 0:07:58 smithi main rhel 8.6 orch:cephadm/workunits/{0-distro/rhel_8.6_container_tools_3.0 agent/on mon_election/connectivity task/test_orch_cli_mon} 5
Failure Reason:

Command failed on smithi040 with status 1: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:32a5ea00c4bee0c51c66aa8902716d7900f94c76 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 98dd0bc8-46a3-11ed-8436-001a4aab830c -- ceph-volume lvm zap /dev/vg_nvme/lv_4'

fail 7058221 2022-10-08 00:11:38 2022-10-08 00:33:42 2022-10-08 00:50:36 0:16:54 0:10:29 0:06:25 smithi main centos 8.stream orch:cephadm/with-work/{0-distro/centos_8.stream_container_tools fixed-2 mode/root mon_election/connectivity msgr/async-v1only start tasks/rotate-keys} 2
Failure Reason:

Command failed on smithi084 with status 1: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:32a5ea00c4bee0c51c66aa8902716d7900f94c76 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid aed08514-46a2-11ed-8436-001a4aab830c -- ceph-volume lvm zap /dev/vg_nvme/lv_4'

pass 7058222 2022-10-08 00:11:39 2022-10-08 00:33:43 2022-10-08 00:55:09 0:21:26 0:15:35 0:05:51 smithi main rhel 8.6 orch:cephadm/osds/{0-distro/rhel_8.6_container_tools_rhel8 0-nvme-loop 1-start 2-ops/repave-all} 2
fail 7058223 2022-10-08 00:11:41 2022-10-08 00:55:07 721 smithi main rhel 8.6 orch:cephadm/smoke/{0-distro/rhel_8.6_container_tools_3.0 0-nvme-loop agent/on fixed-2 mon_election/classic start} 2
Failure Reason:

Command failed on smithi103 with status 1: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:32a5ea00c4bee0c51c66aa8902716d7900f94c76 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 23fe4e34-46a3-11ed-8436-001a4aab830c -- ceph-volume lvm zap /dev/nvme4n1'

pass 7058224 2022-10-08 00:11:42 2022-10-08 00:35:34 2022-10-08 00:53:39 0:18:05 0:11:25 0:06:40 smithi main centos 8.stream orch:cephadm/smoke-roleless/{0-distro/centos_8.stream_container_tools_crun 0-nvme-loop 1-start 2-services/basic 3-final} 2
fail 7058225 2022-10-08 00:11:43 2022-10-08 00:36:24 2022-10-08 00:59:34 0:23:10 0:17:22 0:05:48 smithi main rhel 8.6 orch:cephadm/thrash/{0-distro/rhel_8.6_container_tools_3.0 1-start 2-thrash 3-tasks/rados_api_tests fixed-2 msgr/async-v1only root} 2
Failure Reason:

Command failed on smithi171 with status 1: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:32a5ea00c4bee0c51c66aa8902716d7900f94c76 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ef9929f6-46a3-11ed-8436-001a4aab830c -- ceph-volume lvm zap /dev/vg_nvme/lv_4'

fail 7058226 2022-10-08 00:11:44 2022-10-08 00:36:24 2022-10-08 00:53:22 0:16:58 0:11:09 0:05:49 smithi main centos 8.stream orch:cephadm/with-work/{0-distro/centos_8.stream_container_tools_crun fixed-2 mode/packaged mon_election/classic msgr/async-v2only start tasks/rados_api_tests} 2
Failure Reason:

Command failed on smithi046 with status 1: 'sudo cephadm --image quay.ceph.io/ceph-ci/ceph:32a5ea00c4bee0c51c66aa8902716d7900f94c76 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 2597dd1e-46a3-11ed-8436-001a4aab830c -- ceph-volume lvm zap /dev/vg_nvme/lv_4'

pass 7058227 2022-10-08 00:11:45 2022-10-08 00:36:35 2022-10-08 00:59:33 0:22:58 0:16:28 0:06:30 smithi main rhel 8.6 orch:cephadm/workunits/{0-distro/rhel_8.6_container_tools_rhel8 agent/off mon_election/classic task/test_adoption} 1
pass 7058228 2022-10-08 00:11:46 2022-10-08 00:36:35 2022-10-08 01:16:06 0:39:31 0:32:00 0:07:31 smithi main centos 8.stream orch:cephadm/mds_upgrade_sequence/{bluestore-bitmap centos_8.stream_container_tools conf/{client mds mon osd} fail_fs/yes overrides/{ignorelist_health ignorelist_wrongly_marked_down pg-warn syntax} roles tasks/{0-from/pacific 1-volume/{0-create 1-ranks/1 2-allow_standby_replay/yes 3-inline/yes 4-verify} 2-client 3-upgrade-mgr-staggered 4-config-upgrade/{fail_fs} 5-upgrade-with-workload 6-verify}} 2
fail 7058229 2022-10-08 00:11:47 2022-10-08 00:36:56 2022-10-08 01:01:06 0:24:10 0:17:17 0:06:53 smithi main rhel 8.6 orch:cephadm/with-work/{0-distro/rhel_8.6_container_tools_3.0 fixed-2 mode/root mon_election/connectivity msgr/async start tasks/rados_python} 2
Failure Reason:

Command failed on smithi037 with status 1: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:32a5ea00c4bee0c51c66aa8902716d7900f94c76 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 2296ad74-46a4-11ed-8436-001a4aab830c -- ceph-volume lvm zap /dev/vg_nvme/lv_4'

pass 7058230 2022-10-08 00:11:49 2022-10-08 00:37:46 2022-10-08 01:08:50 0:31:04 0:21:01 0:10:03 smithi main ubuntu 20.04 orch:cephadm/workunits/{0-distro/ubuntu_20.04 agent/on mon_election/connectivity task/test_cephadm} 1
fail 7058231 2022-10-08 00:11:50 2022-10-08 00:37:46 2022-10-08 00:59:30 0:21:44 0:14:11 0:07:33 smithi main rhel 8.6 orch:cephadm/smoke-roleless/{0-distro/rhel_8.6_container_tools_3.0 0-nvme-loop 1-start 2-services/client-keyring 3-final} 2
Failure Reason:

Command failed on smithi036 with status 127: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:32a5ea00c4bee0c51c66aa8902716d7900f94c76 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 80cf5752-46a3-11ed-8436-001a4aab830c -- ceph osd stat -f json'

pass 7058232 2022-10-08 00:11:51 2022-10-08 00:37:57 2022-10-08 01:15:36 0:37:39 0:31:12 0:06:27 smithi main centos 8.stream orch:cephadm/mgr-nfs-upgrade/{0-centos_8.stream_container_tools 1-bootstrap/16.2.5 1-start 2-nfs 3-upgrade-with-workload 4-final} 2
fail 7058233 2022-10-08 00:11:52 2022-10-08 00:38:37 2022-10-08 01:02:51 0:24:14 0:16:50 0:07:24 smithi main rhel 8.6 orch:cephadm/with-work/{0-distro/rhel_8.6_container_tools_rhel8 fixed-2 mode/packaged mon_election/classic msgr/async-v1only start tasks/rotate-keys} 2
Failure Reason:

Command failed on smithi012 with status 1: 'sudo cephadm --image quay.ceph.io/ceph-ci/ceph:32a5ea00c4bee0c51c66aa8902716d7900f94c76 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 66ef3428-46a4-11ed-8436-001a4aab830c -- ceph-volume lvm zap /dev/vg_nvme/lv_4'

pass 7058234 2022-10-08 00:11:53 2022-10-08 00:39:18 2022-10-08 00:50:14 0:10:56 0:04:53 0:06:03 smithi main centos 8.stream orch:cephadm/workunits/{0-distro/centos_8.stream_container_tools agent/off mon_election/classic task/test_cephadm_repos} 1
pass 7058235 2022-10-08 00:11:54 2022-10-08 00:39:18 2022-10-08 01:03:18 0:24:00 0:16:26 0:07:34 smithi main rhel 8.6 orch:cephadm/smoke-roleless/{0-distro/rhel_8.6_container_tools_rhel8 0-nvme-loop 1-start 2-services/iscsi 3-final} 2
fail 7058236 2022-10-08 00:11:55 2022-10-08 00:39:48 2022-10-08 00:59:01 0:19:13 0:12:23 0:06:50 smithi main rhel 8.6 orch:cephadm/smoke/{0-distro/rhel_8.6_container_tools_rhel8 0-nvme-loop agent/off fixed-2 mon_election/connectivity start} 2
Failure Reason:

Command failed on smithi018 with status 1: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:32a5ea00c4bee0c51c66aa8902716d7900f94c76 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid c077dea6-46a3-11ed-8436-001a4aab830c -- ceph-volume lvm zap /dev/nvme4n1'

fail 7058237 2022-10-08 00:11:56 2022-10-08 00:40:09 2022-10-08 01:03:16 0:23:07 0:16:56 0:06:11 smithi main rhel 8.6 orch:cephadm/thrash/{0-distro/rhel_8.6_container_tools_rhel8 1-start 2-thrash 3-tasks/radosbench fixed-2 msgr/async-v2only root} 2
Failure Reason:

Command failed on smithi063 with status 1: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:32a5ea00c4bee0c51c66aa8902716d7900f94c76 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 78811300-46a4-11ed-8436-001a4aab830c -- ceph-volume lvm zap /dev/vg_nvme/lv_4'

fail 7058238 2022-10-08 00:11:58 2022-10-08 00:40:19 2022-10-08 01:04:32 0:24:13 0:12:12 0:12:01 smithi main ubuntu 20.04 orch:cephadm/with-work/{0-distro/ubuntu_20.04 fixed-2 mode/root mon_election/connectivity msgr/async-v2only start tasks/rados_api_tests} 2
Failure Reason:

Command failed on smithi090 with status 1: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:32a5ea00c4bee0c51c66aa8902716d7900f94c76 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 39120238-46a4-11ed-8436-001a4aab830c -- ceph-volume lvm zap /dev/vg_nvme/lv_4'

pass 7058239 2022-10-08 00:11:59 2022-10-08 00:40:20 2022-10-08 01:15:15 0:34:55 0:29:00 0:05:55 smithi main centos 8.stream orch:cephadm/mds_upgrade_sequence/{bluestore-bitmap centos_8.stream_container_tools conf/{client mds mon osd} fail_fs/no overrides/{ignorelist_health ignorelist_wrongly_marked_down pg-warn syntax} roles tasks/{0-from/v16.2.4 1-volume/{0-create 1-ranks/2 2-allow_standby_replay/yes 3-inline/no 4-verify} 2-client 3-upgrade-mgr-staggered 4-config-upgrade/{fail_fs} 5-upgrade-with-workload 6-verify}} 2
pass 7058240 2022-10-08 00:12:00 2022-10-08 00:40:20 2022-10-08 01:16:48 0:36:28 0:28:58 0:07:30 smithi main ubuntu 20.04 orch:cephadm/upgrade/{1-start-distro/1-start-ubuntu_20.04 2-repo_digest/repo_digest 3-upgrade/simple 4-wait 5-upgrade-ls agent/off mon_election/connectivity} 2
fail 7058241 2022-10-08 00:12:01 2022-10-08 00:41:31 2022-10-08 01:07:08 0:25:37 0:14:51 0:10:46 smithi main ubuntu 20.04 orch:cephadm/osds/{0-distro/ubuntu_20.04 0-nvme-loop 1-start 2-ops/rm-zap-add} 2
Failure Reason:

Command failed on smithi109 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:32a5ea00c4bee0c51c66aa8902716d7900f94c76 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 935d17e2-46a3-11ed-8436-001a4aab830c -- bash -c \'set -e\nset -x\nceph orch ps\nceph orch device ls\nDEVID=$(ceph device ls | grep osd.1 | awk \'"\'"\'{print $1}\'"\'"\')\nHOST=$(ceph orch device ls | grep $DEVID | awk \'"\'"\'{print $1}\'"\'"\')\nDEV=$(ceph orch device ls | grep $DEVID | awk \'"\'"\'{print $2}\'"\'"\')\necho "host $HOST, dev $DEV, devid $DEVID"\nceph orch osd rm 1\nwhile ceph orch osd rm status | grep ^1 ; do sleep 5 ; done\nceph orch device zap $HOST $DEV --force\nceph orch daemon add osd $HOST:$DEV\nwhile ! ceph osd dump | grep osd.1 | grep up ; do sleep 5 ; done\n\''

fail 7058242 2022-10-08 00:12:02 2022-10-08 00:41:51 2022-10-08 00:59:26 0:17:35 0:10:33 0:07:02 smithi main centos 8.stream orch:cephadm/workunits/{0-distro/centos_8.stream_container_tools_crun agent/on mon_election/connectivity task/test_iscsi_pids_limit/{centos_8.stream_container_tools test_iscsi_pids_limit}} 1
Failure Reason:

Command failed on smithi027 with status 1: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:32a5ea00c4bee0c51c66aa8902716d7900f94c76 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid f9acf6f2-46a3-11ed-8436-001a4aab830c -- ceph-volume lvm zap /dev/vg_nvme/lv_4'

fail 7058243 2022-10-08 00:12:03 2022-10-08 00:41:51 2022-10-08 01:01:19 0:19:28 0:11:39 0:07:49 smithi main centos 8.stream orch:cephadm/with-work/{0-distro/centos_8.stream_container_tools fixed-2 mode/packaged mon_election/classic msgr/async start tasks/rados_python} 2
Failure Reason:

Command failed on smithi134 with status 1: 'sudo cephadm --image quay.ceph.io/ceph-ci/ceph:32a5ea00c4bee0c51c66aa8902716d7900f94c76 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid f3c400b4-46a3-11ed-8436-001a4aab830c -- ceph-volume lvm zap /dev/vg_nvme/lv_4'

pass 7058244 2022-10-08 00:12:04 2022-10-08 00:42:12 2022-10-08 01:09:51 0:27:39 0:18:42 0:08:57 smithi main ubuntu 20.04 orch:cephadm/smoke-roleless/{0-distro/ubuntu_20.04 0-nvme-loop 1-start 2-services/jaeger 3-final} 2
fail 7058245 2022-10-08 00:12:05 2022-10-08 00:42:42 2022-10-08 01:01:38 0:18:56 0:11:38 0:07:18 smithi main centos 8.stream orch:cephadm/with-work/{0-distro/centos_8.stream_container_tools_crun fixed-2 mode/root mon_election/connectivity msgr/async-v1only start tasks/rotate-keys} 2
Failure Reason:

Command failed on smithi043 with status 1: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:32a5ea00c4bee0c51c66aa8902716d7900f94c76 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 0cccdf04-46a4-11ed-8436-001a4aab830c -- ceph-volume lvm zap /dev/vg_nvme/lv_4'

fail 7058246 2022-10-08 00:12:07 2022-10-08 00:42:43 2022-10-08 01:05:07 0:22:24 0:16:26 0:05:58 smithi main rhel 8.6 orch:cephadm/workunits/{0-distro/rhel_8.6_container_tools_3.0 agent/off mon_election/classic task/test_nfs} 1
Failure Reason:

Command failed on smithi033 with status 1: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:32a5ea00c4bee0c51c66aa8902716d7900f94c76 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid b14cacc6-46a4-11ed-8436-001a4aab830c -- ceph-volume lvm zap /dev/vg_nvme/lv_4'

fail 7058247 2022-10-08 00:12:08 2022-10-08 00:42:43 2022-10-08 00:58:22 0:15:39 0:08:54 0:06:45 smithi main centos 8.stream orch:cephadm/smoke-roleless/{0-distro/centos_8.stream_container_tools 0-nvme-loop 1-start 2-services/mirror 3-final} 2
Failure Reason:

Command failed on smithi049 with status 127: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:32a5ea00c4bee0c51c66aa8902716d7900f94c76 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 801bb2c4-46a3-11ed-8436-001a4aab830c -- ceph osd stat -f json'

fail 7058248 2022-10-08 00:12:09 2022-10-08 00:42:44 2022-10-08 01:00:25 0:17:41 0:06:52 0:10:49 smithi main ubuntu 20.04 orch:cephadm/smoke/{0-distro/ubuntu_20.04 0-nvme-loop agent/on fixed-2 mon_election/classic start} 2
Failure Reason:

Command failed on smithi186 with status 1: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:32a5ea00c4bee0c51c66aa8902716d7900f94c76 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid b9ff248a-46a3-11ed-8436-001a4aab830c -- ceph-volume lvm zap /dev/nvme4n1'

fail 7058249 2022-10-08 00:12:10 2022-10-08 00:43:04 2022-10-08 01:04:41 0:21:37 0:11:53 0:09:44 smithi main ubuntu 20.04 orch:cephadm/thrash/{0-distro/ubuntu_20.04 1-start 2-thrash 3-tasks/small-objects fixed-2 msgr/async root} 2
Failure Reason:

Command failed on smithi153 with status 1: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:32a5ea00c4bee0c51c66aa8902716d7900f94c76 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 7d51abba-46a4-11ed-8436-001a4aab830c -- ceph-volume lvm zap /dev/vg_nvme/lv_4'

fail 7058250 2022-10-08 00:12:11 2022-10-08 00:43:04 2022-10-08 01:06:09 0:23:05 0:16:38 0:06:27 smithi main rhel 8.6 orch:cephadm/with-work/{0-distro/rhel_8.6_container_tools_3.0 fixed-2 mode/packaged mon_election/classic msgr/async-v2only start tasks/rados_api_tests} 2
Failure Reason:

Command failed on smithi097 with status 1: 'sudo cephadm --image quay.ceph.io/ceph-ci/ceph:32a5ea00c4bee0c51c66aa8902716d7900f94c76 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ce6ee44a-46a4-11ed-8436-001a4aab830c -- ceph-volume lvm zap /dev/vg_nvme/lv_4'

pass 7058251 2022-10-08 00:12:12 2022-10-08 00:43:15 2022-10-08 01:22:35 0:39:20 0:32:57 0:06:23 smithi main centos 8.stream orch:cephadm/mds_upgrade_sequence/{bluestore-bitmap centos_8.stream_container_tools conf/{client mds mon osd} fail_fs/yes overrides/{ignorelist_health ignorelist_wrongly_marked_down pg-warn syntax} roles tasks/{0-from/pacific 1-volume/{0-create 1-ranks/1 2-allow_standby_replay/no 3-inline/yes 4-verify} 2-client 3-upgrade-mgr-staggered 4-config-upgrade/{fail_fs} 5-upgrade-with-workload 6-verify}} 2
fail 7058252 2022-10-08 00:12:13 2022-10-08 00:43:35 2022-10-08 01:06:45 0:23:10 0:16:12 0:06:58 smithi main rhel 8.6 orch:cephadm/workunits/{0-distro/rhel_8.6_container_tools_rhel8 agent/on mon_election/connectivity task/test_orch_cli} 1
Failure Reason:

Command failed on smithi137 with status 1: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:32a5ea00c4bee0c51c66aa8902716d7900f94c76 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid f33d11a2-46a4-11ed-8436-001a4aab830c -- ceph-volume lvm zap /dev/vg_nvme/lv_4'

pass 7058253 2022-10-08 00:12:14 2022-10-08 00:43:46 2022-10-08 01:05:06 0:21:20 0:13:37 0:07:43 smithi main centos 8.stream orch:cephadm/smoke-roleless/{0-distro/centos_8.stream_container_tools_crun 0-nvme-loop 1-start 2-services/nfs-ingress-rgw-bucket 3-final} 2
fail 7058254 2022-10-08 00:12:15 2022-10-08 00:43:56 2022-10-08 01:07:12 0:23:16 0:16:33 0:06:43 smithi main rhel 8.6 orch:cephadm/with-work/{0-distro/rhel_8.6_container_tools_rhel8 fixed-2 mode/root mon_election/connectivity msgr/async start tasks/rados_python} 2
Failure Reason:

Command failed on smithi120 with status 1: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:32a5ea00c4bee0c51c66aa8902716d7900f94c76 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid fcb4fa9c-46a4-11ed-8436-001a4aab830c -- ceph-volume lvm zap /dev/vg_nvme/lv_4'

pass 7058255 2022-10-08 00:12:16 2022-10-08 00:44:06 2022-10-08 01:03:33 0:19:27 0:11:40 0:07:47 smithi main centos 8.stream orch:cephadm/osds/{0-distro/centos_8.stream_container_tools 0-nvme-loop 1-start 2-ops/rm-zap-flag} 2
fail 7058256 2022-10-08 00:12:18 2022-10-08 00:44:37 2022-10-08 01:10:52 0:26:15 0:15:10 0:11:05 smithi main ubuntu 20.04 orch:cephadm/workunits/{0-distro/ubuntu_20.04 agent/off mon_election/classic task/test_orch_cli_mon} 5
Failure Reason:

Command failed on smithi045 with status 1: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:32a5ea00c4bee0c51c66aa8902716d7900f94c76 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid d19df28c-46a4-11ed-8436-001a4aab830c -- ceph-volume lvm zap /dev/vg_nvme/lv_4'