Status Job ID Links Posted Started Updated
Runtime
Duration
In Waiting
Machine Teuthology Branch OS Type OS Version Description Nodes
pass 7079681 2022-10-24 15:00:44 2022-10-24 18:54:57 2022-10-24 19:09:19 0:14:22 0:07:53 0:06:29 smithi main centos 8.stream rados/cephadm/workunits/{0-distro/centos_8.stream_container_tools mon_election/connectivity task/test_cephadm_repos} 1
fail 7079682 2022-10-24 15:00:46 2022-10-24 18:54:57 2022-10-24 19:08:24 0:13:27 0:06:06 0:07:21 smithi main ubuntu 18.04 rados/cephadm/thrash-old-clients/{0-size-min-size-overrides/3-size-2-min-size 1-install/mimic backoff/peering_and_degraded ceph clusters/{openstack three-plus-one} d-balancer/on distro$/{ubuntu_18.04} mon_election/connectivity msgr-failures/osd-delay rados thrashers/careful thrashosds-health workloads/radosbench} 3
Failure Reason:

Command failed on smithi037 with status 5: 'sudo systemctl stop ceph-f7e6af42-53ce-11ed-8438-001a4aab830c@mon.a'

fail 7079683 2022-10-24 15:00:47 2022-10-24 18:55:28 2022-10-24 19:12:50 0:17:22 0:07:48 0:09:34 smithi main rhel 8.4 rados/cephadm/smoke/{0-nvme-loop distro/rhel_8.4_container_tools_rhel8 fixed-2 mon_election/connectivity start} 2
Failure Reason:

Command failed on smithi055 with status 5: 'sudo systemctl stop ceph-8c100ba0-53cf-11ed-8438-001a4aab830c@mon.a'

fail 7079684 2022-10-24 15:00:48 2022-10-24 18:58:08 2022-10-24 19:13:50 0:15:42 0:08:02 0:07:40 smithi main rhel 8.4 rados/cephadm/smoke-roleless/{0-distro/rhel_8.4_container_tools_3.0 0-nvme-loop 1-start 2-services/iscsi 3-final} 2
Failure Reason:

Command failed on smithi063 with status 5: 'sudo systemctl stop ceph-b71424d0-53cf-11ed-8438-001a4aab830c@mon.smithi063'

fail 7079685 2022-10-24 15:00:49 2022-10-24 18:58:49 2022-10-24 19:26:07 0:27:18 0:21:01 0:06:17 smithi main centos 8.stream rados/cephadm/mds_upgrade_sequence/{bluestore-bitmap centos_8.stream_container_tools conf/{client mds mon osd} overrides/{ignorelist_health ignorelist_wrongly_marked_down pg-warn} roles tasks/{0-from/v16.2.4 1-volume/{0-create 1-ranks/2 2-allow_standby_replay/yes 3-inline/no 4-verify} 2-client 3-upgrade-with-workload 4-verify}} 2
Failure Reason:

Command failed on smithi110 with status 1: 'sudo /home/ubuntu/cephtest/cephadm --image docker.io/ceph/ceph:v16.2.4 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid fff9e220-53cf-11ed-8438-001a4aab830c -e sha1=d88f759b6ee30a8b0d9b03e2bbb29f39d48f9f2a -- bash -c \'ceph versions | jq -e \'"\'"\'.overall | keys\'"\'"\' | grep $sha1\''

fail 7079686 2022-10-24 15:00:51 2022-10-24 18:58:59 2022-10-24 19:30:03 0:31:04 0:19:37 0:11:27 smithi main ubuntu 20.04 rados/cephadm/upgrade/{1-start-distro/1-start-ubuntu_20.04 2-repo_digest/defaut 3-upgrade/simple 4-wait 5-upgrade-ls mon_election/connectivity} 2
Failure Reason:

Command failed on smithi134 with status 1: 'sudo /home/ubuntu/cephtest/cephadm --image docker.io/ceph/ceph:v15.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 170b4dc8-53d0-11ed-8438-001a4aab830c -e sha1=d88f759b6ee30a8b0d9b03e2bbb29f39d48f9f2a -- bash -c \'ceph versions | jq -e \'"\'"\'.overall | keys\'"\'"\' | grep $sha1\''

pass 7079687 2022-10-24 15:00:52 2022-10-24 19:00:20 2022-10-24 19:51:30 0:51:10 0:45:49 0:05:21 smithi main rhel 8.4 rados/monthrash/{ceph clusters/3-mons mon_election/classic msgr-failures/mon-delay msgr/async-v1only objectstore/bluestore-stupid rados supported-random-distro$/{rhel_8} thrashers/sync workloads/rados_mon_osdmap_prune} 2
fail 7079688 2022-10-24 15:00:53 2022-10-24 19:00:30 2022-10-24 19:17:38 0:17:08 0:09:29 0:07:39 smithi main centos 8.stream rados/cephadm/workunits/{0-distro/centos_8.stream_container_tools mon_election/classic task/test_nfs} 1
Failure Reason:

Command failed on smithi167 with status 5: 'sudo systemctl stop ceph-397efa80-53d0-11ed-8438-001a4aab830c@mon.a'

fail 7079689 2022-10-24 15:00:54 2022-10-24 19:00:31 2022-10-24 19:16:06 0:15:35 0:05:19 0:10:16 smithi main ubuntu 20.04 rados/cephadm/osds/{0-distro/ubuntu_20.04 0-nvme-loop 1-start 2-ops/rm-zap-add} 2
Failure Reason:

Command failed on smithi005 with status 5: 'sudo systemctl stop ceph-1ce8c09a-53d0-11ed-8438-001a4aab830c@mon.smithi005'

fail 7079690 2022-10-24 15:00:56 2022-10-24 19:00:41 2022-10-24 19:14:50 0:14:09 0:07:10 0:06:59 smithi main ubuntu 18.04 rados/cephadm/thrash-old-clients/{0-size-min-size-overrides/2-size-2-min-size 1-install/nautilus-v1only backoff/normal ceph clusters/{openstack three-plus-one} d-balancer/crush-compat distro$/{ubuntu_18.04} mon_election/classic msgr-failures/fastclose rados thrashers/default thrashosds-health workloads/rbd_cls} 3
Failure Reason:

Command failed on smithi003 with status 5: 'sudo systemctl stop ceph-03ebe3f6-53d0-11ed-8438-001a4aab830c@mon.a'

fail 7079691 2022-10-24 15:00:57 2022-10-24 19:00:52 2022-10-24 19:12:34 0:11:42 0:05:11 0:06:31 smithi main ubuntu 18.04 rados/cephadm/smoke/{0-nvme-loop distro/ubuntu_18.04 fixed-2 mon_election/classic start} 2
Failure Reason:

Command failed on smithi106 with status 5: 'sudo systemctl stop ceph-b2d22b24-53cf-11ed-8438-001a4aab830c@mon.a'

fail 7079692 2022-10-24 15:00:58 2022-10-24 19:00:52 2022-10-24 19:18:13 0:17:21 0:06:49 0:10:32 smithi main ubuntu 20.04 rados/rook/smoke/{0-distro/ubuntu_20.04 0-kubeadm 1-rook 2-workload/radosbench 3-final cluster/1-node k8s/1.21 net/calico rook/master} 1
Failure Reason:

[Errno 2] Cannot find file on the remote 'ubuntu@smithi179.front.sepia.ceph.com': 'rook/cluster/examples/kubernetes/ceph/operator.yaml'

fail 7079693 2022-10-24 15:00:59 2022-10-24 19:00:53 2022-10-24 19:16:19 0:15:26 0:07:39 0:07:47 smithi main rhel 8.4 rados/cephadm/smoke-roleless/{0-distro/rhel_8.4_container_tools_rhel8 0-nvme-loop 1-start 2-services/mirror 3-final} 2
Failure Reason:

Command failed on smithi120 with status 5: 'sudo systemctl stop ceph-035e9064-53d0-11ed-8438-001a4aab830c@mon.smithi120'

fail 7079694 2022-10-24 15:01:01 2022-10-24 19:01:03 2022-10-24 19:18:06 0:17:03 0:09:43 0:07:20 smithi main centos 8.stream rados/cephadm/thrash/{0-distro/centos_8.stream_container_tools 1-start 2-thrash 3-tasks/snaps-few-objects fixed-2 msgr/async root} 2
Failure Reason:

Command failed on smithi071 with status 5: 'sudo systemctl stop ceph-523d162e-53d0-11ed-8438-001a4aab830c@mon.a'

fail 7079695 2022-10-24 15:01:02 2022-10-24 19:01:03 2022-10-24 19:20:28 0:19:25 0:11:59 0:07:26 smithi main rhel 8.4 rados/cephadm/with-work/{0-distro/rhel_8.4_container_tools_3.0 fixed-2 mode/root mon_election/connectivity msgr/async start tasks/rados_python} 2
Failure Reason:

Command failed on smithi043 with status 5: 'sudo systemctl stop ceph-a2face3a-53d0-11ed-8438-001a4aab830c@mon.a'

fail 7079696 2022-10-24 15:01:03 2022-10-24 19:01:34 2022-10-24 19:20:06 0:18:32 0:09:30 0:09:02 smithi main centos 8.stream rados/cephadm/workunits/{0-distro/centos_8.stream_container_tools mon_election/connectivity task/test_orch_cli} 1
Failure Reason:

Command failed on smithi182 with status 5: 'sudo systemctl stop ceph-921fa13a-53d0-11ed-8438-001a4aab830c@mon.a'

fail 7079697 2022-10-24 15:01:04 2022-10-24 19:03:14 2022-10-24 19:14:39 0:11:25 0:05:16 0:06:09 smithi main ubuntu 18.04 rados/cephadm/smoke-roleless/{0-distro/ubuntu_18.04 0-nvme-loop 1-start 2-services/nfs-ingress-rgw-bucket 3-final} 2
Failure Reason:

Command failed on smithi035 with status 5: 'sudo systemctl stop ceph-00db5e4e-53d0-11ed-8438-001a4aab830c@mon.smithi035'

fail 7079698 2022-10-24 15:01:06 2022-10-24 19:03:25 2022-10-24 19:27:52 0:24:27 0:15:47 0:08:40 smithi main centos 8.stream rados/cephadm/mds_upgrade_sequence/{bluestore-bitmap centos_8.stream_container_tools conf/{client mds mon osd} overrides/{ignorelist_health ignorelist_wrongly_marked_down pg-warn} roles tasks/{0-from/v16.2.4 1-volume/{0-create 1-ranks/1 2-allow_standby_replay/no 3-inline/yes 4-verify} 2-client 3-upgrade-with-workload 4-verify}} 2
Failure Reason:

Command failed on smithi066 with status 1: "sudo /home/ubuntu/cephtest/cephadm --image docker.io/ceph/ceph:v16.2.4 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid d335f61a-53d0-11ed-8438-001a4aab830c -- bash -c 'ceph fs set cephfs max_mds 1'"

fail 7079699 2022-10-24 15:01:07 2022-10-24 19:03:35 2022-10-24 19:17:07 0:13:32 0:06:47 0:06:45 smithi main ubuntu 18.04 rados/cephadm/thrash-old-clients/{0-size-min-size-overrides/3-size-2-min-size 1-install/nautilus-v2only backoff/peering ceph clusters/{openstack three-plus-one} d-balancer/on distro$/{ubuntu_18.04} mon_election/connectivity msgr-failures/few rados thrashers/mapgap thrashosds-health workloads/snaps-few-objects} 3
Failure Reason:

Command failed on smithi099 with status 5: 'sudo systemctl stop ceph-486b59e4-53d0-11ed-8438-001a4aab830c@mon.a'

fail 7079700 2022-10-24 15:01:08 2022-10-24 19:03:46 2022-10-24 19:20:46 0:17:00 0:04:54 0:12:06 smithi main ubuntu 20.04 rados/cephadm/smoke/{0-nvme-loop distro/ubuntu_20.04 fixed-2 mon_election/connectivity start} 2
Failure Reason:

Command failed on smithi044 with status 5: 'sudo systemctl stop ceph-d2aabbf4-53d0-11ed-8438-001a4aab830c@mon.a'

fail 7079701 2022-10-24 15:01:09 2022-10-24 19:04:36 2022-10-24 19:24:38 0:20:02 0:10:03 0:09:59 smithi main centos 8.stream rados/cephadm/workunits/{0-distro/centos_8.stream_container_tools mon_election/classic task/test_orch_cli_mon} 5
Failure Reason:

Command failed on smithi018 with status 5: 'sudo systemctl stop ceph-45663f24-53d1-11ed-8438-001a4aab830c@mon.a'

fail 7079702 2022-10-24 15:01:11 2022-10-24 19:07:17 2022-10-24 19:20:47 0:13:30 0:06:38 0:06:52 smithi main centos 8.stream rados/cephadm/osds/{0-distro/centos_8.stream_container_tools 0-nvme-loop 1-start 2-ops/rm-zap-flag} 2
Failure Reason:

Command failed on smithi078 with status 5: 'sudo systemctl stop ceph-d1a41ffc-53d0-11ed-8438-001a4aab830c@mon.smithi078'

fail 7079703 2022-10-24 15:01:12 2022-10-24 19:07:38 2022-10-24 19:24:03 0:16:25 0:05:04 0:11:21 smithi main ubuntu 20.04 rados/cephadm/smoke-roleless/{0-distro/ubuntu_20.04 0-nvme-loop 1-start 2-services/nfs-ingress-rgw-user 3-final} 2
Failure Reason:

Command failed on smithi157 with status 5: 'sudo systemctl stop ceph-2ec908a0-53d1-11ed-8438-001a4aab830c@mon.smithi157'

fail 7079704 2022-10-24 15:01:13 2022-10-24 19:07:48 2022-10-24 19:21:17 0:13:29 0:04:44 0:08:45 smithi main ubuntu 20.04 rados/dashboard/{centos_8.stream_container_tools clusters/{2-node-mgr} debug/mgr mon_election/classic random-objectstore$/{bluestore-hybrid} supported-random-distro$/{ubuntu_latest} tasks/dashboard} 2
Failure Reason:

Command failed on smithi045 with status 1: 'TESTDIR=/home/ubuntu/cephtest bash -s'

fail 7079705 2022-10-24 15:01:14 2022-10-24 19:08:29 2022-10-24 19:25:25 0:16:56 0:09:38 0:07:18 smithi main centos 8.stream rados/cephadm/dashboard/{0-distro/centos_8.stream_container_tools task/test_e2e} 2
Failure Reason:

Command failed on smithi006 with status 5: 'sudo systemctl stop ceph-55632c5c-53d1-11ed-8438-001a4aab830c@mon.a'

fail 7079706 2022-10-24 15:01:16 2022-10-24 19:08:49 2022-10-24 19:38:36 0:29:47 0:22:01 0:07:46 smithi main centos 8.stream rados/cephadm/mds_upgrade_sequence/{bluestore-bitmap centos_8.stream_container_tools conf/{client mds mon osd} overrides/{ignorelist_health ignorelist_wrongly_marked_down pg-warn} roles tasks/{0-from/v16.2.4 1-volume/{0-create 1-ranks/1 2-allow_standby_replay/no 3-inline/no 4-verify} 2-client 3-upgrade-with-workload 4-verify}} 2
Failure Reason:

Command failed on smithi053 with status 1: 'sudo /home/ubuntu/cephtest/cephadm --image docker.io/ceph/ceph:v16.2.4 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 7b7933c8-53d1-11ed-8438-001a4aab830c -e sha1=d88f759b6ee30a8b0d9b03e2bbb29f39d48f9f2a -- bash -c \'ceph versions | jq -e \'"\'"\'.overall | keys\'"\'"\' | grep $sha1\''

fail 7079707 2022-10-24 15:01:17 2022-10-24 19:09:20 2022-10-24 19:37:29 0:28:09 0:21:02 0:07:07 smithi main centos 8.stream rados/cephadm/mgr-nfs-upgrade/{0-distro/centos_8.stream_container_tools 1-bootstrap/16.2.4 1-start 2-nfs 3-upgrade-with-workload 4-final} 2
Failure Reason:

Command failed on smithi143 with status 1: 'sudo /home/ubuntu/cephtest/cephadm --image docker.io/ceph/ceph:v16.2.4 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 32a247ac-53d1-11ed-8438-001a4aab830c -e sha1=d88f759b6ee30a8b0d9b03e2bbb29f39d48f9f2a -- bash -c \'ceph versions | jq -e \'"\'"\'.overall | keys\'"\'"\' | grep $sha1\''

fail 7079708 2022-10-24 15:01:18 2022-10-24 19:10:30 2022-10-24 19:25:36 0:15:06 0:07:04 0:08:02 smithi main centos 8.stream rados/cephadm/smoke/{0-nvme-loop distro/centos_8.stream_container_tools fixed-2 mon_election/classic start} 2
Failure Reason:

Command failed on smithi008 with status 5: 'sudo systemctl stop ceph-8cf666a2-53d1-11ed-8438-001a4aab830c@mon.a'

fail 7079709 2022-10-24 15:01:19 2022-10-24 19:12:41 2022-10-24 19:25:58 0:13:17 0:07:02 0:06:15 smithi main centos 8.stream rados/cephadm/smoke-roleless/{0-distro/centos_8.stream_container_tools 0-nvme-loop 1-start 2-services/nfs-ingress 3-final} 2
Failure Reason:

Command failed on smithi106 with status 5: 'sudo systemctl stop ceph-9861453e-53d1-11ed-8438-001a4aab830c@mon.smithi106'

fail 7079710 2022-10-24 15:01:20 2022-10-24 19:12:41 2022-10-24 19:27:00 0:14:19 0:07:41 0:06:38 smithi main rhel 8.4 rados/cephadm/smoke-singlehost/{0-distro$/{rhel_8.4_container_tools_rhel8} 1-start 2-services/basic 3-final} 1
Failure Reason:

Command failed on smithi059 with status 5: 'sudo systemctl stop ceph-8456b560-53d1-11ed-8438-001a4aab830c@mon.smithi059'

fail 7079711 2022-10-24 15:01:22 2022-10-24 19:12:52 2022-10-24 19:30:36 0:17:44 0:09:57 0:07:47 smithi main centos 8.stream rados/cephadm/thrash/{0-distro/centos_8.stream_container_tools 1-start 2-thrash 3-tasks/rados_api_tests fixed-2 msgr/async-v1only root} 2
Failure Reason:

Command failed on smithi055 with status 5: 'sudo systemctl stop ceph-1c5b91e6-53d2-11ed-8438-001a4aab830c@mon.a'

fail 7079712 2022-10-24 15:01:23 2022-10-24 19:13:12 2022-10-24 19:39:03 0:25:51 0:17:19 0:08:32 smithi main centos 8.stream rados/cephadm/upgrade/{1-start-distro/1-start-centos_8.stream_container-tools 2-repo_digest/defaut 3-upgrade/simple 4-wait 5-upgrade-ls mon_election/classic} 2
Failure Reason:

Command failed on smithi107 with status 1: 'sudo /home/ubuntu/cephtest/cephadm --image docker.io/ceph/ceph:v15.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid c02415a6-53d1-11ed-8438-001a4aab830c -e sha1=d88f759b6ee30a8b0d9b03e2bbb29f39d48f9f2a -- bash -c \'ceph versions | jq -e \'"\'"\'.overall | keys\'"\'"\' | grep $sha1\''

fail 7079713 2022-10-24 15:01:24 2022-10-24 19:13:43 2022-10-24 19:30:47 0:17:04 0:11:13 0:05:51 smithi main rhel 8.4 rados/cephadm/with-work/{0-distro/rhel_8.4_container_tools_rhel8 fixed-2 mode/packaged mon_election/classic msgr/async-v1only start tasks/rados_api_tests} 2
Failure Reason:

Command failed on smithi063 with status 5: 'sudo systemctl stop ceph-4455944e-53d2-11ed-8438-001a4aab830c@mon.a'

fail 7079714 2022-10-24 15:01:25 2022-10-24 19:13:53 2022-10-24 19:28:49 0:14:56 0:07:07 0:07:49 smithi main ubuntu 18.04 rados/cephadm/thrash-old-clients/{0-size-min-size-overrides/2-size-2-min-size 1-install/nautilus backoff/peering_and_degraded ceph clusters/{openstack three-plus-one} d-balancer/crush-compat distro$/{ubuntu_18.04} mon_election/classic msgr-failures/osd-delay rados thrashers/morepggrow thrashosds-health workloads/test_rbd_api} 3
Failure Reason:

Command failed on smithi003 with status 5: 'sudo systemctl stop ceph-f62440b8-53d1-11ed-8438-001a4aab830c@mon.a'

fail 7079715 2022-10-24 15:01:27 2022-10-24 19:14:54 2022-10-24 19:29:42 0:14:48 0:08:24 0:06:24 smithi main rhel 8.4 rados/cephadm/smoke-roleless/{0-distro/rhel_8.4_container_tools_3.0 0-nvme-loop 1-start 2-services/nfs-ingress2 3-final} 2
Failure Reason:

Command failed on smithi035 with status 5: 'sudo systemctl stop ceph-fc6d9bae-53d1-11ed-8438-001a4aab830c@mon.smithi035'

fail 7079716 2022-10-24 15:01:28 2022-10-24 19:14:54 2022-10-24 19:30:14 0:15:20 0:08:22 0:06:58 smithi main rhel 8.4 rados/cephadm/osds/{0-distro/rhel_8.4_container_tools_3.0 0-nvme-loop 1-start 2-ops/rm-zap-wait} 2
Failure Reason:

Command failed on smithi137 with status 5: 'sudo systemctl stop ceph-0d634b5c-53d2-11ed-8438-001a4aab830c@mon.smithi137'

fail 7079717 2022-10-24 15:01:29 2022-10-24 19:15:25 2022-10-24 19:31:50 0:16:25 0:10:36 0:05:49 smithi main centos 8.stream rados/cephadm/workunits/{0-distro/centos_8.stream_container_tools mon_election/connectivity task/test_cephadm} 1
Failure Reason:

Command failed (workunit test cephadm/test_cephadm.sh) on smithi040 with status 125: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=d88f759b6ee30a8b0d9b03e2bbb29f39d48f9f2a TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 CEPH_MNT=/home/ubuntu/cephtest/mnt.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/cephadm/test_cephadm.sh'

fail 7079718 2022-10-24 15:01:30 2022-10-24 19:15:25 2022-10-24 19:31:12 0:15:47 0:08:24 0:07:23 smithi main rhel 8.4 rados/cephadm/smoke/{0-nvme-loop distro/rhel_8.4_container_tools_3.0 fixed-2 mon_election/connectivity start} 2
Failure Reason:

Command failed on smithi186 with status 5: 'sudo systemctl stop ceph-314dd88e-53d2-11ed-8438-001a4aab830c@mon.a'

fail 7079719 2022-10-24 15:01:31 2022-10-24 19:15:55 2022-10-24 19:29:46 0:13:51 0:07:30 0:06:21 smithi main ubuntu 18.04 rados/cephadm/thrash-old-clients/{0-size-min-size-overrides/3-size-2-min-size 1-install/octopus backoff/normal ceph clusters/{openstack three-plus-one} d-balancer/on distro$/{ubuntu_18.04} mon_election/connectivity msgr-failures/fastclose rados thrashers/none thrashosds-health workloads/cache-snaps} 3
Failure Reason:

Command failed on smithi093 with status 5: 'sudo systemctl stop ceph-25adf8ce-53d2-11ed-8438-001a4aab830c@mon.a'

fail 7079720 2022-10-24 15:01:33 2022-10-24 19:16:26 2022-10-24 19:46:52 0:30:26 0:21:46 0:08:40 smithi main centos 8.stream rados/cephadm/mds_upgrade_sequence/{bluestore-bitmap centos_8.stream_container_tools conf/{client mds mon osd} overrides/{ignorelist_health ignorelist_wrongly_marked_down pg-warn} roles tasks/{0-from/v16.2.4 1-volume/{0-create 1-ranks/2 2-allow_standby_replay/yes 3-inline/yes 4-verify} 2-client 3-upgrade-with-workload 4-verify}} 2
Failure Reason:

Command failed on smithi049 with status 1: 'sudo /home/ubuntu/cephtest/cephadm --image docker.io/ceph/ceph:v16.2.4 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid b015672c-53d2-11ed-8438-001a4aab830c -e sha1=d88f759b6ee30a8b0d9b03e2bbb29f39d48f9f2a -- bash -c \'ceph versions | jq -e \'"\'"\'.overall | keys\'"\'"\' | grep $sha1\''

fail 7079721 2022-10-24 15:01:34 2022-10-24 19:16:46 2022-10-24 19:32:04 0:15:18 0:07:55 0:07:23 smithi main rhel 8.4 rados/cephadm/smoke-roleless/{0-distro/rhel_8.4_container_tools_rhel8 0-nvme-loop 1-start 2-services/nfs 3-final} 2
Failure Reason:

Command failed on smithi080 with status 5: 'sudo systemctl stop ceph-409cd196-53d2-11ed-8438-001a4aab830c@mon.smithi080'

fail 7079722 2022-10-24 15:01:35 2022-10-24 19:16:57 2022-10-24 19:34:02 0:17:05 0:10:13 0:06:52 smithi main centos 8.stream rados/cephadm/thrash/{0-distro/centos_8.stream_container_tools 1-start 2-thrash 3-tasks/radosbench fixed-2 msgr/async-v2only root} 2
Failure Reason:

Command failed on smithi032 with status 5: 'sudo systemctl stop ceph-9e202764-53d2-11ed-8438-001a4aab830c@mon.a'

fail 7079723 2022-10-24 15:01:36 2022-10-24 19:16:57 2022-10-24 19:30:28 0:13:31 0:07:21 0:06:10 smithi main ubuntu 18.04 rados/cephadm/with-work/{0-distro/ubuntu_18.04 fixed-2 mode/root mon_election/connectivity msgr/async-v2only start tasks/rados_python} 2
Failure Reason:

Command failed on smithi121 with status 5: 'sudo systemctl stop ceph-3b0e1b90-53d2-11ed-8438-001a4aab830c@mon.a'

pass 7079724 2022-10-24 15:01:37 2022-10-24 19:17:18 2022-10-24 19:31:58 0:14:40 0:08:09 0:06:31 smithi main centos 8.stream rados/cephadm/workunits/{0-distro/centos_8.stream_container_tools mon_election/classic task/test_cephadm_repos} 1
fail 7079725 2022-10-24 15:01:39 2022-10-24 19:17:18 2022-10-24 19:32:52 0:15:34 0:07:56 0:07:38 smithi main rhel 8.4 rados/cephadm/smoke/{0-nvme-loop distro/rhel_8.4_container_tools_rhel8 fixed-2 mon_election/classic start} 2
Failure Reason:

Command failed on smithi005 with status 5: 'sudo systemctl stop ceph-5dd2953e-53d2-11ed-8438-001a4aab830c@mon.a'

fail 7079726 2022-10-24 15:01:40 2022-10-24 19:17:48 2022-10-24 19:29:26 0:11:38 0:05:19 0:06:19 smithi main ubuntu 18.04 rados/cephadm/smoke-roleless/{0-distro/ubuntu_18.04 0-nvme-loop 1-start 2-services/nfs2 3-final} 2
Failure Reason:

Command failed on smithi071 with status 5: 'sudo systemctl stop ceph-14eb1eb8-53d2-11ed-8438-001a4aab830c@mon.smithi071'

fail 7079727 2022-10-24 15:01:41 2022-10-24 19:18:09 2022-10-24 19:33:42 0:15:33 0:06:12 0:09:21 smithi main ubuntu 18.04 rados/cephadm/thrash-old-clients/{0-size-min-size-overrides/2-size-2-min-size 1-install/luminous-v1only backoff/peering ceph clusters/{openstack three-plus-one} d-balancer/crush-compat distro$/{ubuntu_18.04} mon_election/classic msgr-failures/few rados thrashers/careful thrashosds-health workloads/radosbench} 3
Failure Reason:

Command failed on smithi043 with status 5: 'sudo systemctl stop ceph-849dc670-53d2-11ed-8438-001a4aab830c@mon.a'

fail 7079728 2022-10-24 15:01:42 2022-10-24 19:20:30 2022-10-24 19:33:48 0:13:18 0:07:29 0:05:49 smithi main rhel 8.4 rados/cephadm/osds/{0-distro/rhel_8.4_container_tools_rhel8 0-nvme-loop 1-start 2-ops/rmdir-reactivate} 2
Failure Reason:

Command failed on smithi078 with status 5: 'sudo systemctl stop ceph-b7623c44-53d2-11ed-8438-001a4aab830c@mon.smithi078'

fail 7079729 2022-10-24 15:01:44 2022-10-24 19:20:50 2022-10-24 19:50:27 0:29:37 0:21:28 0:08:09 smithi main centos 8.stream rados/cephadm/mds_upgrade_sequence/{bluestore-bitmap centos_8.stream_container_tools conf/{client mds mon osd} overrides/{ignorelist_health ignorelist_wrongly_marked_down pg-warn} roles tasks/{0-from/v16.2.4 1-volume/{0-create 1-ranks/1 2-allow_standby_replay/yes 3-inline/no 4-verify} 2-client 3-upgrade-with-workload 4-verify}} 2
Failure Reason:

Command failed on smithi044 with status 1: 'sudo /home/ubuntu/cephtest/cephadm --image docker.io/ceph/ceph:v16.2.4 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 2237bdfa-53d3-11ed-8438-001a4aab830c -e sha1=d88f759b6ee30a8b0d9b03e2bbb29f39d48f9f2a -- bash -c \'ceph versions | jq -e \'"\'"\'.overall | keys\'"\'"\' | grep $sha1\''

fail 7079730 2022-10-24 15:01:45 2022-10-24 19:20:51 2022-10-24 19:54:37 0:33:46 0:23:17 0:10:29 smithi main ubuntu 20.04 rados/cephadm/upgrade/{1-start-distro/1-start-ubuntu_20.04-15.2.9 2-repo_digest/repo_digest 3-upgrade/staggered 4-wait 5-upgrade-ls mon_election/connectivity} 2
Failure Reason:

Command failed on smithi036 with status 1: 'sudo /home/ubuntu/cephtest/cephadm --image docker.io/ceph/ceph:v15.2.9 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid f022f9ce-53d2-11ed-8438-001a4aab830c -e sha1=d88f759b6ee30a8b0d9b03e2bbb29f39d48f9f2a -- bash -c \'ceph versions | jq -e \'"\'"\'.mgr | length == 2\'"\'"\'\''

fail 7079731 2022-10-24 15:01:46 2022-10-24 19:21:01 2022-10-24 19:35:43 0:14:42 0:09:15 0:05:27 smithi main centos 8.stream rados/cephadm/workunits/{0-distro/centos_8.stream_container_tools mon_election/connectivity task/test_nfs} 1
Failure Reason:

Command failed on smithi179 with status 5: 'sudo systemctl stop ceph-ffe45e5c-53d2-11ed-8438-001a4aab830c@mon.a'

fail 7079732 2022-10-24 15:01:47 2022-10-24 19:21:01 2022-10-24 19:36:20 0:15:19 0:05:04 0:10:15 smithi main ubuntu 20.04 rados/cephadm/smoke-roleless/{0-distro/ubuntu_20.04 0-nvme-loop 1-start 2-services/rgw-ingress 3-final} 2
Failure Reason:

Command failed on smithi045 with status 5: 'sudo systemctl stop ceph-e3a7074e-53d2-11ed-8438-001a4aab830c@mon.smithi045'

fail 7079733 2022-10-24 15:01:49 2022-10-24 19:21:22 2022-10-24 19:32:53 0:11:31 0:04:51 0:06:40 smithi main ubuntu 18.04 rados/cephadm/smoke/{0-nvme-loop distro/ubuntu_18.04 fixed-2 mon_election/connectivity start} 2
Failure Reason:

Command failed on smithi085 with status 5: 'sudo systemctl stop ceph-7e307332-53d2-11ed-8438-001a4aab830c@mon.a'

fail 7079734 2022-10-24 15:01:50 2022-10-24 19:21:32 2022-10-24 19:37:39 0:16:07 0:06:07 0:10:00 smithi main ubuntu 18.04 rados/cephadm/thrash-old-clients/{0-size-min-size-overrides/3-size-2-min-size 1-install/luminous backoff/peering_and_degraded ceph clusters/{openstack three-plus-one} d-balancer/on distro$/{ubuntu_18.04} mon_election/connectivity msgr-failures/osd-delay rados thrashers/default thrashosds-health workloads/rbd_cls} 3
Failure Reason:

Command failed on smithi046 with status 5: 'sudo systemctl stop ceph-0db7f96c-53d3-11ed-8438-001a4aab830c@mon.a'

fail 7079735 2022-10-24 15:01:51 2022-10-24 19:24:13 2022-10-24 19:49:57 0:25:44 0:17:17 0:08:27 smithi main centos 8.stream rados/cephadm/mgr-nfs-upgrade/{0-distro/centos_8.stream_container_tools 1-bootstrap/16.2.5 1-start 2-nfs 3-upgrade-with-workload 4-final} 2
Failure Reason:

Command failed on smithi157 with status 1: 'sudo /home/ubuntu/cephtest/cephadm --image docker.io/ceph/ceph:v16.2.5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 3f18a4ac-53d3-11ed-8438-001a4aab830c -e sha1=d88f759b6ee30a8b0d9b03e2bbb29f39d48f9f2a -- bash -c \'ceph versions | jq -e \'"\'"\'.overall | keys\'"\'"\' | grep $sha1\''

fail 7079736 2022-10-24 15:01:52 2022-10-24 19:24:14 2022-10-24 19:37:54 0:13:40 0:06:51 0:06:49 smithi main centos 8.stream rados/cephadm/smoke-roleless/{0-distro/centos_8.stream_container_tools 0-nvme-loop 1-start 2-services/rgw 3-final} 2
Failure Reason:

Command failed on smithi087 with status 5: 'sudo systemctl stop ceph-3d1f39fe-53d3-11ed-8438-001a4aab830c@mon.smithi087'

fail 7079737 2022-10-24 15:01:53 2022-10-24 19:24:44 2022-10-24 19:41:42 0:16:58 0:09:44 0:07:14 smithi main centos 8.stream rados/cephadm/thrash/{0-distro/centos_8.stream_container_tools 1-start 2-thrash 3-tasks/small-objects fixed-2 msgr/async root} 2
Failure Reason:

Command failed on smithi062 with status 5: 'sudo systemctl stop ceph-9e9a9200-53d3-11ed-8438-001a4aab830c@mon.a'

fail 7079738 2022-10-24 15:01:55 2022-10-24 19:24:45 2022-10-24 19:42:31 0:17:46 0:07:25 0:10:21 smithi main ubuntu 20.04 rados/cephadm/with-work/{0-distro/ubuntu_20.04 fixed-2 mode/packaged mon_election/classic msgr/async start tasks/rados_api_tests} 2
Failure Reason:

Command failed on smithi018 with status 5: 'sudo systemctl stop ceph-da4493be-53d3-11ed-8438-001a4aab830c@mon.a'

fail 7079739 2022-10-24 15:01:56 2022-10-24 19:24:45 2022-10-24 19:41:47 0:17:02 0:09:36 0:07:26 smithi main centos 8.stream rados/cephadm/workunits/{0-distro/centos_8.stream_container_tools mon_election/classic task/test_orch_cli} 1
Failure Reason:

Command failed on smithi037 with status 5: 'sudo systemctl stop ceph-a237ed36-53d3-11ed-8438-001a4aab830c@mon.a'

fail 7079740 2022-10-24 15:01:57 2022-10-24 19:25:36 2022-10-24 19:55:03 0:29:27 0:21:20 0:08:07 smithi main centos 8.stream rados/cephadm/mds_upgrade_sequence/{bluestore-bitmap centos_8.stream_container_tools conf/{client mds mon osd} overrides/{ignorelist_health ignorelist_wrongly_marked_down pg-warn} roles tasks/{0-from/v16.2.4 1-volume/{0-create 1-ranks/2 2-allow_standby_replay/no 3-inline/yes 4-verify} 2-client 3-upgrade-with-workload 4-verify}} 2
Failure Reason:

Command failed on smithi008 with status 1: 'sudo /home/ubuntu/cephtest/cephadm --image docker.io/ceph/ceph:v16.2.4 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid dc34cae0-53d3-11ed-8438-001a4aab830c -e sha1=d88f759b6ee30a8b0d9b03e2bbb29f39d48f9f2a -- bash -c \'ceph versions | jq -e \'"\'"\'.overall | keys\'"\'"\' | grep $sha1\''

fail 7079741 2022-10-24 15:01:58 2022-10-24 19:25:47 2022-10-24 19:40:26 0:14:39 0:07:46 0:06:53 smithi main rhel 8.4 rados/cephadm/osds/{0-distro/rhel_8.4_container_tools_rhel8 0-nvme-loop 1-start 2-ops/repave-all} 2
Failure Reason:

Command failed on smithi006 with status 5: 'sudo systemctl stop ceph-688345c2-53d3-11ed-8438-001a4aab830c@mon.smithi006'

fail 7079742 2022-10-24 15:01:59 2022-10-24 19:25:47 2022-10-24 19:43:50 0:18:03 0:06:01 0:12:02 smithi main ubuntu 20.04 rados/cephadm/smoke/{0-nvme-loop distro/ubuntu_20.04 fixed-2 mon_election/classic start} 2
Failure Reason:

Command failed on smithi106 with status 5: 'sudo systemctl stop ceph-e6179f42-53d3-11ed-8438-001a4aab830c@mon.a'

fail 7079743 2022-10-24 15:02:01 2022-10-24 19:26:08 2022-10-24 19:41:03 0:14:55 0:08:12 0:06:43 smithi main rhel 8.4 rados/cephadm/smoke-roleless/{0-distro/rhel_8.4_container_tools_3.0 0-nvme-loop 1-start 2-services/basic 3-final} 2
Failure Reason:

Command failed on smithi110 with status 5: 'sudo systemctl stop ceph-89fe6b28-53d3-11ed-8438-001a4aab830c@mon.smithi110'

fail 7079744 2022-10-24 15:02:02 2022-10-24 19:26:08 2022-10-24 19:48:04 0:21:56 0:11:31 0:10:25 smithi main centos 8.stream rados/cephadm/workunits/{0-distro/centos_8.stream_container_tools mon_election/connectivity task/test_orch_cli_mon} 5
Failure Reason:

Command failed on smithi047 with status 5: 'sudo systemctl stop ceph-78ed68e2-53d4-11ed-8438-001a4aab830c@mon.a'

fail 7079745 2022-10-24 15:02:03 2022-10-24 19:27:59 2022-10-24 19:41:26 0:13:27 0:06:56 0:06:31 smithi main ubuntu 18.04 rados/rook/smoke/{0-distro/ubuntu_18.04 0-kubeadm 1-rook 2-workload/radosbench 3-final cluster/1-node k8s/1.21 net/calico rook/1.6.2} 1
Failure Reason:

Command failed on smithi074 with status 1: 'kubectl create -f rook/cluster/examples/kubernetes/ceph/crds.yaml -f rook/cluster/examples/kubernetes/ceph/common.yaml -f operator.yaml'

fail 7079746 2022-10-24 15:02:04 2022-10-24 19:28:09 2022-10-24 19:42:54 0:14:45 0:07:14 0:07:31 smithi main ubuntu 18.04 rados/cephadm/thrash-old-clients/{0-size-min-size-overrides/2-size-2-min-size 1-install/mimic-v1only backoff/normal ceph clusters/{openstack three-plus-one} d-balancer/crush-compat distro$/{ubuntu_18.04} mon_election/classic msgr-failures/fastclose rados thrashers/mapgap thrashosds-health workloads/snaps-few-objects} 3
Failure Reason:

Command failed on smithi003 with status 5: 'sudo systemctl stop ceph-f1820c32-53d3-11ed-8438-001a4aab830c@mon.a'

fail 7079747 2022-10-24 15:02:06 2022-10-24 19:29:00 2022-10-24 19:44:28 0:15:28 0:08:40 0:06:48 smithi main rhel 8.4 rados/cephadm/smoke-roleless/{0-distro/rhel_8.4_container_tools_rhel8 0-nvme-loop 1-start 2-services/client-keyring 3-final} 2
Failure Reason:

Command failed on smithi071 with status 5: 'sudo systemctl stop ceph-174616ca-53d4-11ed-8438-001a4aab830c@mon.smithi071'

fail 7079748 2022-10-24 15:02:07 2022-10-24 19:29:30 2022-10-24 19:45:35 0:16:05 0:06:04 0:10:01 smithi main rados/cephadm/dashboard/{0-distro/ignorelist_health task/test_e2e} 2
Failure Reason:

Failed to fetch package version from https://shaman.ceph.com/api/search/?status=ready&project=ceph&flavor=default&distros=ubuntu%2F22.04%2Fx86_64&sha1=d88f759b6ee30a8b0d9b03e2bbb29f39d48f9f2a

fail 7079749 2022-10-24 15:02:08 2022-10-24 19:29:50 2022-10-24 19:56:50 0:27:00 0:21:17 0:05:43 smithi main centos 8.stream rados/cephadm/mds_upgrade_sequence/{bluestore-bitmap centos_8.stream_container_tools conf/{client mds mon osd} overrides/{ignorelist_health ignorelist_wrongly_marked_down pg-warn} roles tasks/{0-from/v16.2.4 1-volume/{0-create 1-ranks/2 2-allow_standby_replay/no 3-inline/no 4-verify} 2-client 3-upgrade-with-workload 4-verify}} 2
Failure Reason:

Command failed on smithi093 with status 1: 'sudo /home/ubuntu/cephtest/cephadm --image docker.io/ceph/ceph:v16.2.4 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 6490f0f8-53d4-11ed-8438-001a4aab830c -e sha1=d88f759b6ee30a8b0d9b03e2bbb29f39d48f9f2a -- bash -c \'ceph versions | jq -e \'"\'"\'.overall | keys\'"\'"\' | grep $sha1\''

fail 7079750 2022-10-24 15:02:10 2022-10-24 19:29:51 2022-10-24 19:45:49 0:15:58 0:07:48 0:08:10 smithi main centos 8.stream rados/cephadm/smoke/{0-nvme-loop distro/centos_8.stream_container_tools fixed-2 mon_election/connectivity start} 2
Failure Reason:

Command failed on smithi027 with status 5: 'sudo systemctl stop ceph-33f3501c-53d4-11ed-8438-001a4aab830c@mon.a'

fail 7079751 2022-10-24 15:02:11 2022-10-24 19:29:51 2022-10-24 19:45:18 0:15:27 0:07:53 0:07:34 smithi main rhel 8.4 rados/cephadm/smoke-singlehost/{0-distro$/{rhel_8.4_container_tools_rhel8} 1-start 2-services/rgw 3-final} 1
Failure Reason:

Command failed on smithi150 with status 5: 'sudo systemctl stop ceph-199d233c-53d4-11ed-8438-001a4aab830c@mon.smithi150'

fail 7079752 2022-10-24 15:02:12 2022-10-24 19:30:12 2022-10-24 19:47:12 0:17:00 0:10:21 0:06:39 smithi main centos 8.stream rados/cephadm/thrash/{0-distro/centos_8.stream_container_tools 1-start 2-thrash 3-tasks/snaps-few-objects fixed-2 msgr/async-v1only root} 2
Failure Reason:

Command failed on smithi137 with status 5: 'sudo systemctl stop ceph-7894d290-53d4-11ed-8438-001a4aab830c@mon.a'

fail 7079753 2022-10-24 15:02:13 2022-10-24 19:30:22 2022-10-24 19:59:46 0:29:24 0:20:29 0:08:55 smithi main ubuntu 20.04 rados/cephadm/upgrade/{1-start-distro/1-start-ubuntu_20.04 2-repo_digest/defaut 3-upgrade/simple 4-wait 5-upgrade-ls mon_election/classic} 2
Failure Reason:

Command failed on smithi121 with status 1: 'sudo /home/ubuntu/cephtest/cephadm --image docker.io/ceph/ceph:v15.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 501dff4e-53d4-11ed-8438-001a4aab830c -e sha1=d88f759b6ee30a8b0d9b03e2bbb29f39d48f9f2a -- bash -c \'ceph versions | jq -e \'"\'"\'.overall | keys\'"\'"\' | grep $sha1\''

fail 7079754 2022-10-24 15:02:15 2022-10-24 19:30:32 2022-10-24 19:48:01 0:17:29 0:10:28 0:07:01 smithi main centos 8.stream rados/cephadm/with-work/{0-distro/centos_8.stream_container_tools fixed-2 mode/root mon_election/connectivity msgr/async-v1only start tasks/rados_python} 2
Failure Reason:

Command failed on smithi055 with status 5: 'sudo systemctl stop ceph-9b47d0bc-53d4-11ed-8438-001a4aab830c@mon.a'

fail 7079755 2022-10-24 15:02:16 2022-10-24 19:30:43 2022-10-24 19:44:07 0:13:24 0:05:57 0:07:27 smithi main ubuntu 18.04 rados/cephadm/smoke-roleless/{0-distro/ubuntu_18.04 0-nvme-loop 1-start 2-services/iscsi 3-final} 2
Failure Reason:

Command failed on smithi063 with status 5: 'sudo systemctl stop ceph-ef43f5c0-53d3-11ed-8438-001a4aab830c@mon.smithi063'

fail 7079756 2022-10-24 15:02:17 2022-10-24 19:30:53 2022-10-24 19:44:37 0:13:44 0:07:27 0:06:17 smithi main ubuntu 18.04 rados/cephadm/thrash-old-clients/{0-size-min-size-overrides/3-size-2-min-size 1-install/mimic backoff/peering ceph clusters/{openstack three-plus-one} d-balancer/on distro$/{ubuntu_18.04} mon_election/connectivity msgr-failures/few rados thrashers/morepggrow thrashosds-health workloads/test_rbd_api} 3
Failure Reason:

Command failed on smithi134 with status 5: 'sudo systemctl stop ceph-37b25f36-53d4-11ed-8438-001a4aab830c@mon.a'

fail 7079757 2022-10-24 15:02:19 2022-10-24 19:31:14 2022-10-24 19:45:18 0:14:04 0:05:59 0:08:05 smithi main ubuntu 18.04 rados/cephadm/osds/{0-distro/ubuntu_18.04 0-nvme-loop 1-start 2-ops/rm-zap-add} 2
Failure Reason:

Command failed on smithi040 with status 5: 'sudo systemctl stop ceph-1bd5a7e6-53d4-11ed-8438-001a4aab830c@mon.smithi040'

fail 7079758 2022-10-24 15:02:20 2022-10-24 19:32:04 2022-10-24 19:48:58 0:16:54 0:10:35 0:06:19 smithi main centos 8.stream rados/cephadm/workunits/{0-distro/centos_8.stream_container_tools mon_election/classic task/test_cephadm} 1
Failure Reason:

Command failed (workunit test cephadm/test_cephadm.sh) on smithi172 with status 125: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=d88f759b6ee30a8b0d9b03e2bbb29f39d48f9f2a TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 CEPH_MNT=/home/ubuntu/cephtest/mnt.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/cephadm/test_cephadm.sh'

fail 7079759 2022-10-24 15:02:21 2022-10-24 19:32:05 2022-10-24 19:47:53 0:15:48 0:08:41 0:07:07 smithi main rhel 8.4 rados/cephadm/smoke/{0-nvme-loop distro/rhel_8.4_container_tools_3.0 fixed-2 mon_election/classic start} 2
Failure Reason:

Command failed on smithi085 with status 5: 'sudo systemctl stop ceph-8fe14514-53d4-11ed-8438-001a4aab830c@mon.a'

fail 7079760 2022-10-24 15:02:22 2022-10-24 19:32:55 2022-10-24 19:48:29 0:15:34 0:05:17 0:10:17 smithi main ubuntu 20.04 rados/cephadm/smoke-roleless/{0-distro/ubuntu_20.04 0-nvme-loop 1-start 2-services/mirror 3-final} 2
Failure Reason:

Command failed on smithi005 with status 5: 'sudo systemctl stop ceph-9da2fd28-53d4-11ed-8438-001a4aab830c@mon.smithi005'

fail 7079761 2022-10-24 15:02:24 2022-10-24 19:32:56 2022-10-24 20:02:39 0:29:43 0:22:01 0:07:42 smithi main centos 8.stream rados/cephadm/mds_upgrade_sequence/{bluestore-bitmap centos_8.stream_container_tools conf/{client mds mon osd} overrides/{ignorelist_health ignorelist_wrongly_marked_down pg-warn} roles tasks/{0-from/v16.2.4 1-volume/{0-create 1-ranks/1 2-allow_standby_replay/yes 3-inline/yes 4-verify} 2-client 3-upgrade-with-workload 4-verify}} 2
Failure Reason:

Command failed on smithi175 with status 1: 'sudo /home/ubuntu/cephtest/cephadm --image docker.io/ceph/ceph:v16.2.4 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid e9a06d5a-53d4-11ed-8438-001a4aab830c -e sha1=d88f759b6ee30a8b0d9b03e2bbb29f39d48f9f2a -- bash -c \'ceph versions | jq -e \'"\'"\'.overall | keys\'"\'"\' | grep $sha1\''

fail 7079762 2022-10-24 15:02:25 2022-10-24 19:33:46 2022-10-24 19:47:15 0:13:29 0:07:30 0:05:59 smithi main ubuntu 18.04 rados/cephadm/thrash-old-clients/{0-size-min-size-overrides/2-size-2-min-size 1-install/nautilus-v1only backoff/peering_and_degraded ceph clusters/{openstack three-plus-one} d-balancer/crush-compat distro$/{ubuntu_18.04} mon_election/classic msgr-failures/osd-delay rados thrashers/none thrashosds-health workloads/cache-snaps} 3
Failure Reason:

Command failed on smithi043 with status 5: 'sudo systemctl stop ceph-9872628a-53d4-11ed-8438-001a4aab830c@mon.a'

fail 7079763 2022-10-24 15:02:26 2022-10-24 19:33:57 2022-10-24 20:02:56 0:28:59 0:21:21 0:07:38 smithi main centos 8.stream rados/cephadm/mgr-nfs-upgrade/{0-distro/centos_8.stream_container_tools 1-bootstrap/octopus 1-start 2-nfs 3-upgrade-with-workload 4-final} 2
Failure Reason:

Command failed on smithi032 with status 1: 'sudo /home/ubuntu/cephtest/cephadm --image docker.io/ceph/ceph:v15 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 91a4a648-53d4-11ed-8438-001a4aab830c -e sha1=d88f759b6ee30a8b0d9b03e2bbb29f39d48f9f2a -- bash -c \'ceph versions | jq -e \'"\'"\'.overall | keys\'"\'"\' | grep $sha1\''

fail 7079764 2022-10-24 15:02:28 2022-10-24 19:34:07 2022-10-24 19:48:41 0:14:34 0:06:59 0:07:35 smithi main centos 8.stream rados/cephadm/smoke-roleless/{0-distro/centos_8.stream_container_tools 0-nvme-loop 1-start 2-services/nfs-ingress-rgw-bucket 3-final} 2
Failure Reason:

Command failed on smithi080 with status 5: 'sudo systemctl stop ceph-c634c3fc-53d4-11ed-8438-001a4aab830c@mon.smithi080'

fail 7079765 2022-10-24 15:02:29 2022-10-24 19:35:48 2022-10-24 19:53:03 0:17:15 0:10:00 0:07:15 smithi main centos 8.stream rados/cephadm/thrash/{0-distro/centos_8.stream_container_tools 1-start 2-thrash 3-tasks/rados_api_tests fixed-2 msgr/async-v2only root} 2
Failure Reason:

Command failed on smithi045 with status 5: 'sudo systemctl stop ceph-3f31578e-53d5-11ed-8438-001a4aab830c@mon.a'

fail 7079766 2022-10-24 15:02:30 2022-10-24 19:36:28 2022-10-24 19:54:24 0:17:56 0:11:24 0:06:32 smithi main rhel 8.4 rados/cephadm/with-work/{0-distro/rhel_8.4_container_tools_3.0 fixed-2 mode/packaged mon_election/classic msgr/async-v2only start tasks/rados_api_tests} 2
Failure Reason:

Command failed on smithi143 with status 5: 'sudo systemctl stop ceph-952cfcb0-53d5-11ed-8438-001a4aab830c@mon.a'

pass 7079767 2022-10-24 15:02:31 2022-10-24 19:37:39 2022-10-24 19:52:33 0:14:54 0:07:45 0:07:09 smithi main centos 8.stream rados/cephadm/workunits/{0-distro/centos_8.stream_container_tools mon_election/connectivity task/test_cephadm_repos} 1
fail 7079768 2022-10-24 15:02:33 2022-10-24 19:37:49 2022-10-24 19:52:43 0:14:54 0:07:36 0:07:18 smithi main rhel 8.4 rados/cephadm/smoke/{0-nvme-loop distro/rhel_8.4_container_tools_rhel8 fixed-2 mon_election/connectivity start} 2
Failure Reason:

Command failed on smithi046 with status 5: 'sudo systemctl stop ceph-1965e542-53d5-11ed-8438-001a4aab830c@mon.a'

fail 7079769 2022-10-24 15:02:34 2022-10-24 19:37:50 2022-10-24 19:53:27 0:15:37 0:05:04 0:10:33 smithi main ubuntu 20.04 rados/cephadm/osds/{0-distro/ubuntu_20.04 0-nvme-loop 1-start 2-ops/rm-zap-flag} 2
Failure Reason:

Command failed on smithi087 with status 5: 'sudo systemctl stop ceph-4b53896a-53d5-11ed-8438-001a4aab830c@mon.smithi087'

fail 7079770 2022-10-24 15:02:35 2022-10-24 19:38:01 2022-10-24 19:53:45 0:15:44 0:08:05 0:07:39 smithi main rhel 8.4 rados/cephadm/smoke-roleless/{0-distro/rhel_8.4_container_tools_3.0 0-nvme-loop 1-start 2-services/nfs-ingress-rgw-user 3-final} 2
Failure Reason:

Command failed on smithi053 with status 5: 'sudo systemctl stop ceph-4c2d05d2-53d5-11ed-8438-001a4aab830c@mon.smithi053'

fail 7079772 2022-10-24 15:02:36 2022-10-24 19:38:43 2022-10-24 20:06:23 0:27:40 0:20:42 0:06:58 smithi main centos 8.stream rados/cephadm/mds_upgrade_sequence/{bluestore-bitmap centos_8.stream_container_tools conf/{client mds mon osd} overrides/{ignorelist_health ignorelist_wrongly_marked_down pg-warn} roles tasks/{0-from/v16.2.4 1-volume/{0-create 1-ranks/2 2-allow_standby_replay/yes 3-inline/no 4-verify} 2-client 3-upgrade-with-workload 4-verify}} 2
Failure Reason:

Command failed on smithi107 with status 1: 'sudo /home/ubuntu/cephtest/cephadm --image docker.io/ceph/ceph:v16.2.4 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 9faf55a2-53d5-11ed-8438-001a4aab830c -e sha1=d88f759b6ee30a8b0d9b03e2bbb29f39d48f9f2a -- bash -c \'ceph versions | jq -e \'"\'"\'.overall | keys\'"\'"\' | grep $sha1\''

fail 7079774 2022-10-24 15:02:38 2022-10-24 19:40:34 2022-10-24 20:03:49 0:23:15 0:16:09 0:07:06 smithi main centos 8.stream rados/cephadm/upgrade/{1-start-distro/1-start-centos_8.stream_container-tools 2-repo_digest/repo_digest 3-upgrade/staggered 4-wait 5-upgrade-ls mon_election/connectivity} 2
Failure Reason:

Command failed on smithi079 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image docker.io/ceph/ceph:v15.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 6bd831f4-53d5-11ed-8438-001a4aab830c -e sha1=d88f759b6ee30a8b0d9b03e2bbb29f39d48f9f2a -- bash -c \'ceph orch daemon redeploy "mgr.$(ceph mgr dump -f json | jq .standbys | jq .[] | jq -r .name)"\''

fail 7079776 2022-10-24 15:02:39 2022-10-24 19:41:05 2022-10-24 19:57:34 0:16:29 0:09:29 0:07:00 smithi main centos 8.stream rados/cephadm/workunits/{0-distro/centos_8.stream_container_tools mon_election/classic task/test_nfs} 1
Failure Reason:

Command failed on smithi006 with status 5: 'sudo systemctl stop ceph-ce42b008-53d5-11ed-8438-001a4aab830c@mon.a'

fail 7079778 2022-10-24 15:02:40 2022-10-24 19:41:47 2022-10-24 19:56:06 0:14:19 0:06:26 0:07:53 smithi main ubuntu 18.04 rados/cephadm/thrash-old-clients/{0-size-min-size-overrides/3-size-2-min-size 1-install/nautilus-v2only backoff/normal ceph clusters/{openstack three-plus-one} d-balancer/on distro$/{ubuntu_18.04} mon_election/connectivity msgr-failures/fastclose rados thrashers/pggrow thrashosds-health workloads/radosbench} 3
Failure Reason:

Command failed on smithi018 with status 5: 'sudo systemctl stop ceph-ad21355c-53d5-11ed-8438-001a4aab830c@mon.a'

fail 7079780 2022-10-24 15:02:41 2022-10-24 19:42:58 2022-10-24 19:55:36 0:12:38 0:05:02 0:07:36 smithi main ubuntu 18.04 rados/cephadm/smoke/{0-nvme-loop distro/ubuntu_18.04 fixed-2 mon_election/classic start} 2
Failure Reason:

Command failed on smithi106 with status 5: 'sudo systemctl stop ceph-b0ece2bc-53d5-11ed-8438-001a4aab830c@mon.a'

fail 7079782 2022-10-24 15:02:43 2022-10-24 19:44:09 2022-10-24 19:59:18 0:15:09 0:07:43 0:07:26 smithi main rhel 8.4 rados/cephadm/smoke-roleless/{0-distro/rhel_8.4_container_tools_rhel8 0-nvme-loop 1-start 2-services/nfs-ingress 3-final} 2
Failure Reason:

Command failed on smithi071 with status 5: 'sudo systemctl stop ceph-06cf78a2-53d6-11ed-8438-001a4aab830c@mon.smithi071'

fail 7079784 2022-10-24 15:02:44 2022-10-24 19:45:20 2022-10-24 20:02:04 0:16:44 0:10:01 0:06:43 smithi main centos 8.stream rados/cephadm/thrash/{0-distro/centos_8.stream_container_tools 1-start 2-thrash 3-tasks/radosbench fixed-2 msgr/async root} 2
Failure Reason:

Command failed on smithi074 with status 5: 'sudo systemctl stop ceph-80ca8944-53d6-11ed-8438-001a4aab830c@mon.a'

fail 7079786 2022-10-24 15:02:46 2022-10-24 19:45:41 2022-10-24 20:03:41 0:18:00 0:11:29 0:06:31 smithi main rhel 8.4 rados/cephadm/with-work/{0-distro/rhel_8.4_container_tools_rhel8 fixed-2 mode/root mon_election/connectivity msgr/async start tasks/rados_python} 2
Failure Reason:

Command failed on smithi027 with status 5: 'sudo systemctl stop ceph-e47808fe-53d6-11ed-8438-001a4aab830c@mon.a'

fail 7079788 2022-10-24 15:02:47 2022-10-24 19:47:02 2022-10-24 20:04:12 0:17:10 0:09:36 0:07:34 smithi main centos 8.stream rados/cephadm/workunits/{0-distro/centos_8.stream_container_tools mon_election/connectivity task/test_orch_cli} 1
Failure Reason:

Command failed on smithi138 with status 5: 'sudo systemctl stop ceph-bfe32046-53d6-11ed-8438-001a4aab830c@mon.a'

fail 7079790 2022-10-24 15:02:48 2022-10-24 19:47:23 2022-10-24 19:58:38 0:11:15 0:05:26 0:05:49 smithi main ubuntu 18.04 rados/cephadm/smoke-roleless/{0-distro/ubuntu_18.04 0-nvme-loop 1-start 2-services/nfs-ingress2 3-final} 2
Failure Reason:

Command failed on smithi137 with status 5: 'sudo systemctl stop ceph-2c11807e-53d6-11ed-8438-001a4aab830c@mon.smithi137'

fail 7079792 2022-10-24 15:02:49 2022-10-24 19:48:04 2022-10-24 20:16:54 0:28:50 0:21:37 0:07:13 smithi main centos 8.stream rados/cephadm/mds_upgrade_sequence/{bluestore-bitmap centos_8.stream_container_tools conf/{client mds mon osd} overrides/{ignorelist_health ignorelist_wrongly_marked_down pg-warn} roles tasks/{0-from/v16.2.4 1-volume/{0-create 1-ranks/1 2-allow_standby_replay/no 3-inline/yes 4-verify} 2-client 3-upgrade-with-workload 4-verify}} 2
Failure Reason:

Command failed on smithi085 with status 1: 'sudo /home/ubuntu/cephtest/cephadm --image docker.io/ceph/ceph:v16.2.4 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid da90fbde-53d6-11ed-8438-001a4aab830c -e sha1=d88f759b6ee30a8b0d9b03e2bbb29f39d48f9f2a -- bash -c \'ceph versions | jq -e \'"\'"\'.overall | keys\'"\'"\' | grep $sha1\''

fail 7079794 2022-10-24 15:02:51 2022-10-24 19:48:15 2022-10-24 20:03:23 0:15:08 0:07:21 0:07:47 smithi main centos 8.stream rados/cephadm/osds/{0-distro/centos_8.stream_container_tools 0-nvme-loop 1-start 2-ops/rm-zap-wait} 2
Failure Reason:

Command failed on smithi059 with status 5: 'sudo systemctl stop ceph-988e9d22-53d6-11ed-8438-001a4aab830c@mon.smithi059'

fail 7079796 2022-10-24 15:02:52 2022-10-24 19:48:35 2022-10-24 20:01:55 0:13:20 0:07:07 0:06:13 smithi main ubuntu 18.04 rados/cephadm/thrash-old-clients/{0-size-min-size-overrides/2-size-2-min-size 1-install/nautilus backoff/peering ceph clusters/{openstack three-plus-one} d-balancer/crush-compat distro$/{ubuntu_18.04} mon_election/classic msgr-failures/few rados thrashers/careful thrashosds-health workloads/rbd_cls} 3
Failure Reason:

Command failed on smithi047 with status 5: 'sudo systemctl stop ceph-966cb592-53d6-11ed-8438-001a4aab830c@mon.a'

fail 7079798 2022-10-24 15:02:53 2022-10-24 19:50:07 2022-10-24 20:06:39 0:16:32 0:05:23 0:11:09 smithi main ubuntu 20.04 rados/cephadm/smoke/{0-nvme-loop distro/ubuntu_20.04 fixed-2 mon_election/connectivity start} 2
Failure Reason:

Command failed on smithi044 with status 5: 'sudo systemctl stop ceph-4b1c842c-53d7-11ed-8438-001a4aab830c@mon.a'

fail 7079800 2022-10-24 15:02:54 2022-10-24 19:51:38 2022-10-24 20:09:43 0:18:05 0:10:21 0:07:44 smithi main centos 8.stream rados/cephadm/workunits/{0-distro/centos_8.stream_container_tools mon_election/classic task/test_orch_cli_mon} 5
Failure Reason:

Command failed on smithi046 with status 5: 'sudo systemctl stop ceph-9ca68c16-53d7-11ed-8438-001a4aab830c@mon.a'

fail 7079802 2022-10-24 15:02:56 2022-10-24 19:53:09 2022-10-24 20:08:54 0:15:45 0:05:18 0:10:27 smithi main ubuntu 20.04 rados/cephadm/smoke-roleless/{0-distro/ubuntu_20.04 0-nvme-loop 1-start 2-services/nfs 3-final} 2
Failure Reason:

Command failed on smithi087 with status 5: 'sudo systemctl stop ceph-7d51b246-53d7-11ed-8438-001a4aab830c@mon.smithi087'

fail 7079804 2022-10-24 15:02:57 2022-10-24 19:53:50 2022-10-24 20:07:45 0:13:55 0:06:55 0:07:00 smithi main ubuntu 18.04 rados/cephadm/thrash-old-clients/{0-size-min-size-overrides/3-size-2-min-size 1-install/octopus backoff/peering_and_degraded ceph clusters/{openstack three-plus-one} d-balancer/on distro$/{ubuntu_18.04} mon_election/connectivity msgr-failures/osd-delay rados thrashers/default thrashosds-health workloads/snaps-few-objects} 3
Failure Reason:

Command failed on smithi053 with status 5: 'sudo systemctl stop ceph-5f1cbe7e-53d7-11ed-8438-001a4aab830c@mon.a'

fail 7079806 2022-10-24 15:02:58 2022-10-24 19:54:40 2022-10-24 20:11:43 0:17:03 0:10:01 0:07:02 smithi main centos 8.stream rados/cephadm/dashboard/{0-distro/centos_8.stream_container_tools task/test_e2e} 2
Failure Reason:

Command failed on smithi008 with status 5: 'sudo systemctl stop ceph-da82f934-53d7-11ed-8438-001a4aab830c@mon.a'

fail 7079808 2022-10-24 15:03:00 2022-10-24 19:55:41 2022-10-24 20:25:26 0:29:45 0:21:44 0:08:01 smithi main centos 8.stream rados/cephadm/mds_upgrade_sequence/{bluestore-bitmap centos_8.stream_container_tools conf/{client mds mon osd} overrides/{ignorelist_health ignorelist_wrongly_marked_down pg-warn} roles tasks/{0-from/v16.2.4 1-volume/{0-create 1-ranks/1 2-allow_standby_replay/no 3-inline/no 4-verify} 2-client 3-upgrade-with-workload 4-verify}} 2
Failure Reason:

Command failed on smithi037 with status 1: 'sudo /home/ubuntu/cephtest/cephadm --image docker.io/ceph/ceph:v16.2.4 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 13bc480e-53d8-11ed-8438-001a4aab830c -e sha1=d88f759b6ee30a8b0d9b03e2bbb29f39d48f9f2a -- bash -c \'ceph versions | jq -e \'"\'"\'.overall | keys\'"\'"\' | grep $sha1\''

fail 7079810 2022-10-24 15:03:01 2022-10-24 19:56:13 2022-10-24 20:25:46 0:29:33 0:21:25 0:08:08 smithi main centos 8.stream rados/cephadm/mgr-nfs-upgrade/{0-distro/centos_8.stream_container_tools 1-bootstrap/16.2.4 1-start 2-nfs 3-upgrade-with-workload 4-final} 2
Failure Reason:

Command failed on smithi093 with status 1: 'sudo /home/ubuntu/cephtest/cephadm --image docker.io/ceph/ceph:v16.2.4 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid b8a76e76-53d7-11ed-8438-001a4aab830c -e sha1=d88f759b6ee30a8b0d9b03e2bbb29f39d48f9f2a -- bash -c \'ceph versions | jq -e \'"\'"\'.overall | keys\'"\'"\' | grep $sha1\''

fail 7079812 2022-10-24 15:03:02 2022-10-24 19:57:04 2022-10-24 20:09:56 0:12:52 0:07:13 0:05:39 smithi main centos 8.stream rados/cephadm/smoke/{0-nvme-loop distro/centos_8.stream_container_tools fixed-2 mon_election/classic start} 2
Failure Reason:

Command failed on smithi003 with status 5: 'sudo systemctl stop ceph-c48a1054-53d7-11ed-8438-001a4aab830c@mon.a'

fail 7079814 2022-10-24 15:03:04 2022-10-24 19:58:45 2022-10-24 20:12:18 0:13:33 0:07:11 0:06:22 smithi main centos 8.stream rados/cephadm/smoke-roleless/{0-distro/centos_8.stream_container_tools 0-nvme-loop 1-start 2-services/nfs2 3-final} 2
Failure Reason:

Command failed on smithi071 with status 5: 'sudo systemctl stop ceph-17dbf006-53d8-11ed-8438-001a4aab830c@mon.smithi071'

fail 7079816 2022-10-24 15:03:05 2022-10-24 19:59:36 2022-10-24 20:12:47 0:13:11 0:06:57 0:06:14 smithi main centos 8.stream rados/cephadm/smoke-singlehost/{0-distro$/{centos_8.stream_container_tools} 1-start 2-services/basic 3-final} 1
Failure Reason:

Command failed on smithi176 with status 5: 'sudo systemctl stop ceph-224c649e-53d8-11ed-8438-001a4aab830c@mon.smithi176'

fail 7079818 2022-10-24 15:03:06 2022-10-24 19:59:57 2022-10-24 20:18:52 0:18:55 0:10:25 0:08:30 smithi main centos 8.stream rados/cephadm/thrash/{0-distro/centos_8.stream_container_tools 1-start 2-thrash 3-tasks/small-objects fixed-2 msgr/async-v1only root} 2
Failure Reason:

Command failed on smithi049 with status 5: 'sudo systemctl stop ceph-e8508ad0-53d8-11ed-8438-001a4aab830c@mon.a'

fail 7079820 2022-10-24 15:03:08 2022-10-24 20:31:09 1191 smithi main ubuntu 20.04 rados/cephadm/upgrade/{1-start-distro/1-start-ubuntu_20.04-15.2.9 2-repo_digest/defaut 3-upgrade/simple 4-wait 5-upgrade-ls mon_election/classic} 2
Failure Reason:

Command failed on smithi047 with status 1: 'sudo /home/ubuntu/cephtest/cephadm --image docker.io/ceph/ceph:v15.2.9 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid a184814c-53d8-11ed-8438-001a4aab830c -e sha1=d88f759b6ee30a8b0d9b03e2bbb29f39d48f9f2a -- bash -c \'ceph versions | jq -e \'"\'"\'.overall | keys\'"\'"\' | grep $sha1\''

fail 7079822 2022-10-24 15:03:09 2022-10-24 20:02:09 2022-10-24 20:20:11 0:18:02 0:08:19 0:09:43 smithi main ubuntu 20.04 rados/rook/smoke/{0-distro/ubuntu_20.04 0-kubeadm 1-rook 2-workload/none 3-final cluster/3-node k8s/1.21 net/calico rook/master} 3
Failure Reason:

[Errno 2] Cannot find file on the remote 'ubuntu@smithi150.front.sepia.ceph.com': 'rook/cluster/examples/kubernetes/ceph/operator.yaml'

fail 7079824 2022-10-24 15:03:10 2022-10-24 20:02:40 2022-10-24 20:16:06 0:13:26 0:07:15 0:06:11 smithi main ubuntu 18.04 rados/cephadm/with-work/{0-distro/ubuntu_18.04 fixed-2 mode/packaged mon_election/classic msgr/async-v1only start tasks/rados_api_tests} 2
Failure Reason:

Command failed on smithi175 with status 5: 'sudo systemctl stop ceph-98783ed6-53d8-11ed-8438-001a4aab830c@mon.a'

fail 7079826 2022-10-24 15:03:11 2022-10-24 20:03:01 2022-10-24 20:18:30 0:15:29 0:09:08 0:06:21 smithi main rhel 8.4 rados/cephadm/osds/{0-distro/rhel_8.4_container_tools_3.0 0-nvme-loop 1-start 2-ops/rmdir-reactivate} 2
Failure Reason:

Command failed on smithi059 with status 5: 'sudo systemctl stop ceph-e7fa7c62-53d8-11ed-8438-001a4aab830c@mon.smithi059'

fail 7079828 2022-10-24 15:03:13 2022-10-24 20:03:52 2022-10-24 20:19:42 0:15:50 0:08:35 0:07:15 smithi main rhel 8.4 rados/cephadm/smoke-roleless/{0-distro/rhel_8.4_container_tools_3.0 0-nvme-loop 1-start 2-services/rgw-ingress 3-final} 2
Failure Reason:

Command failed on smithi027 with status 5: 'sudo systemctl stop ceph-00433c46-53d9-11ed-8438-001a4aab830c@mon.smithi027'

fail 7079830 2022-10-24 15:03:14 2022-10-24 20:04:22 2022-10-24 20:21:15 0:16:53 0:10:34 0:06:19 smithi main centos 8.stream rados/cephadm/workunits/{0-distro/centos_8.stream_container_tools mon_election/connectivity task/test_cephadm} 1
Failure Reason:

Command failed (workunit test cephadm/test_cephadm.sh) on smithi138 with status 125: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=d88f759b6ee30a8b0d9b03e2bbb29f39d48f9f2a TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 CEPH_MNT=/home/ubuntu/cephtest/mnt.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/cephadm/test_cephadm.sh'

fail 7079832 2022-10-24 15:03:16 2022-10-24 20:05:44 2022-10-24 20:21:42 0:15:58 0:08:11 0:07:47 smithi main rhel 8.4 rados/cephadm/smoke/{0-nvme-loop distro/rhel_8.4_container_tools_3.0 fixed-2 mon_election/connectivity start} 2
Failure Reason:

Command failed on smithi107 with status 5: 'sudo systemctl stop ceph-38c8002e-53d9-11ed-8438-001a4aab830c@mon.a'

fail 7079834 2022-10-24 15:03:17 2022-10-24 20:06:45 2022-10-24 20:21:11 0:14:26 0:06:27 0:07:59 smithi main ubuntu 18.04 rados/cephadm/thrash-old-clients/{0-size-min-size-overrides/2-size-2-min-size 1-install/luminous-v1only backoff/normal ceph clusters/{openstack three-plus-one} d-balancer/crush-compat distro$/{ubuntu_18.04} mon_election/classic msgr-failures/fastclose rados thrashers/mapgap thrashosds-health workloads/test_rbd_api} 3
Failure Reason:

Command failed on smithi053 with status 5: 'sudo systemctl stop ceph-2f1acff2-53d9-11ed-8438-001a4aab830c@mon.a'

fail 7079837 2022-10-24 15:03:18 2022-10-24 20:07:57 2022-10-24 20:37:59 0:30:02 0:21:53 0:08:09 smithi main centos 8.stream rados/cephadm/mds_upgrade_sequence/{bluestore-bitmap centos_8.stream_container_tools conf/{client mds mon osd} overrides/{ignorelist_health ignorelist_wrongly_marked_down pg-warn} roles tasks/{0-from/v16.2.4 1-volume/{0-create 1-ranks/2 2-allow_standby_replay/yes 3-inline/yes 4-verify} 2-client 3-upgrade-with-workload 4-verify}} 2
Failure Reason:

Command failed on smithi159 with status 1: 'sudo /home/ubuntu/cephtest/cephadm --image docker.io/ceph/ceph:v16.2.4 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid c2c358aa-53d9-11ed-8438-001a4aab830c -e sha1=d88f759b6ee30a8b0d9b03e2bbb29f39d48f9f2a -- bash -c \'ceph versions | jq -e \'"\'"\'.overall | keys\'"\'"\' | grep $sha1\''

fail 7079840 2022-10-24 15:03:20 2022-10-24 20:09:49 2022-10-24 20:24:42 0:14:53 0:07:35 0:07:18 smithi main rhel 8.4 rados/cephadm/smoke-roleless/{0-distro/rhel_8.4_container_tools_rhel8 0-nvme-loop 1-start 2-services/rgw 3-final} 2
Failure Reason:

Command failed on smithi046 with status 5: 'sudo systemctl stop ceph-93cecae8-53d9-11ed-8438-001a4aab830c@mon.smithi046'

fail 7079843 2022-10-24 15:03:21 2022-10-24 20:10:51 2022-10-24 20:28:41 0:17:50 0:09:32 0:08:18 smithi main centos 8.stream rados/cephadm/thrash/{0-distro/centos_8.stream_container_tools 1-start 2-thrash 3-tasks/snaps-few-objects fixed-2 msgr/async-v2only root} 2
Failure Reason:

Command failed on smithi036 with status 5: 'sudo systemctl stop ceph-280056c8-53da-11ed-8438-001a4aab830c@mon.a'

fail 7079846 2022-10-24 15:03:22 2022-10-24 20:12:03 2022-10-24 20:29:45 0:17:42 0:06:55 0:10:47 smithi main ubuntu 20.04 rados/cephadm/with-work/{0-distro/ubuntu_20.04 fixed-2 mode/root mon_election/connectivity msgr/async-v2only start tasks/rados_python} 2
Failure Reason:

Command failed on smithi071 with status 5: 'sudo systemctl stop ceph-5bc0ead6-53da-11ed-8438-001a4aab830c@mon.a'

pass 7079849 2022-10-24 15:03:23 2022-10-24 20:13:45 2022-10-24 20:29:27 0:15:42 0:07:51 0:07:51 smithi main centos 8.stream rados/cephadm/workunits/{0-distro/centos_8.stream_container_tools mon_election/classic task/test_cephadm_repos} 1
fail 7079852 2022-10-24 15:03:25 2022-10-24 20:15:26 2022-10-24 20:30:15 0:14:49 0:08:09 0:06:40 smithi main rhel 8.4 rados/cephadm/osds/{0-distro/rhel_8.4_container_tools_3.0 0-nvme-loop 1-start 2-ops/repave-all} 2
Failure Reason:

Command failed on smithi063 with status 5: 'sudo systemctl stop ceph-67d3c0f0-53da-11ed-8438-001a4aab830c@mon.smithi063'

fail 7079855 2022-10-24 15:03:26 2022-10-24 20:16:27 2022-10-24 20:32:36 0:16:09 0:07:40 0:08:29 smithi main rhel 8.4 rados/cephadm/smoke/{0-nvme-loop distro/rhel_8.4_container_tools_rhel8 fixed-2 mon_election/classic start} 2
Failure Reason:

Command failed on smithi120 with status 5: 'sudo systemctl stop ceph-aaded9fc-53da-11ed-8438-001a4aab830c@mon.a'

fail 7079858 2022-10-24 15:03:27 2022-10-24 20:17:29 2022-10-24 20:28:43 0:11:14 0:05:05 0:06:09 smithi main ubuntu 18.04 rados/cephadm/smoke-roleless/{0-distro/ubuntu_18.04 0-nvme-loop 1-start 2-services/basic 3-final} 2
Failure Reason:

Command failed on smithi040 with status 5: 'sudo systemctl stop ceph-52f86992-53da-11ed-8438-001a4aab830c@mon.smithi040'

fail 7079861 2022-10-24 15:03:29 2022-10-24 20:18:40 2022-10-24 20:32:05 0:13:25 0:06:29 0:06:56 smithi main ubuntu 18.04 rados/cephadm/thrash-old-clients/{0-size-min-size-overrides/3-size-2-min-size 1-install/luminous backoff/peering ceph clusters/{openstack three-plus-one} d-balancer/on distro$/{ubuntu_18.04} mon_election/connectivity msgr-failures/few rados thrashers/morepggrow thrashosds-health workloads/cache-snaps} 3
Failure Reason:

Command failed on smithi079 with status 5: 'sudo systemctl stop ceph-b53e8294-53da-11ed-8438-001a4aab830c@mon.a'

fail 7079864 2022-10-24 15:03:30 2022-10-24 20:19:51 2022-10-24 20:49:16 0:29:25 0:22:47 0:06:38 smithi main centos 8.stream rados/cephadm/mds_upgrade_sequence/{bluestore-bitmap centos_8.stream_container_tools conf/{client mds mon osd} overrides/{ignorelist_health ignorelist_wrongly_marked_down pg-warn} roles tasks/{0-from/v16.2.4 1-volume/{0-create 1-ranks/1 2-allow_standby_replay/yes 3-inline/no 4-verify} 2-client 3-upgrade-with-workload 4-verify}} 2
Failure Reason:

Command failed on smithi186 with status 1: 'sudo /home/ubuntu/cephtest/cephadm --image docker.io/ceph/ceph:v16.2.4 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 6ae044c0-53db-11ed-8438-001a4aab830c -e sha1=d88f759b6ee30a8b0d9b03e2bbb29f39d48f9f2a -- bash -c \'ceph versions | jq -e \'"\'"\'.overall | keys\'"\'"\' | grep $sha1\''

fail 7079867 2022-10-24 15:03:31 2022-10-24 20:20:53 2022-10-24 20:50:35 0:29:42 0:19:57 0:09:45 smithi main ubuntu 20.04 rados/cephadm/upgrade/{1-start-distro/1-start-ubuntu_20.04 2-repo_digest/repo_digest 3-upgrade/staggered 4-wait 5-upgrade-ls mon_election/connectivity} 2
Failure Reason:

Command failed on smithi143 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image docker.io/ceph/ceph:v15.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 581c089c-53db-11ed-8438-001a4aab830c -e sha1=d88f759b6ee30a8b0d9b03e2bbb29f39d48f9f2a -- bash -c \'ceph orch daemon redeploy "mgr.$(ceph mgr dump -f json | jq .standbys | jq .[] | jq -r .name)"\''

fail 7079870 2022-10-24 15:03:33 2022-10-24 20:21:24 2022-10-24 20:38:53 0:17:29 0:09:55 0:07:34 smithi main centos 8.stream rados/cephadm/workunits/{0-distro/centos_8.stream_container_tools mon_election/connectivity task/test_nfs} 1
Failure Reason:

Command failed on smithi155 with status 5: 'sudo systemctl stop ceph-a41cf7f6-53db-11ed-8438-001a4aab830c@mon.a'

fail 7079873 2022-10-24 15:03:34 2022-10-24 20:22:46 2022-10-24 20:39:38 0:16:52 0:05:23 0:11:29 smithi main ubuntu 20.04 rados/cephadm/smoke-roleless/{0-distro/ubuntu_20.04 0-nvme-loop 1-start 2-services/client-keyring 3-final} 2
Failure Reason:

Command failed on smithi164 with status 5: 'sudo systemctl stop ceph-cb0b529a-53db-11ed-8438-001a4aab830c@mon.smithi164'

fail 7079876 2022-10-24 15:03:35 2022-10-24 20:24:47 2022-10-24 20:37:56 0:13:09 0:05:45 0:07:24 smithi main ubuntu 18.04 rados/cephadm/smoke/{0-nvme-loop distro/ubuntu_18.04 fixed-2 mon_election/connectivity start} 2
Failure Reason:

Command failed on smithi110 with status 5: 'sudo systemctl stop ceph-6d1c0742-53db-11ed-8438-001a4aab830c@mon.a'

fail 7079879 2022-10-24 15:03:37 2022-10-24 20:25:28 2022-10-24 20:38:43 0:13:15 0:07:01 0:06:14 smithi main ubuntu 18.04 rados/cephadm/thrash-old-clients/{0-size-min-size-overrides/2-size-2-min-size 1-install/mimic-v1only backoff/peering_and_degraded ceph clusters/{openstack three-plus-one} d-balancer/crush-compat distro$/{ubuntu_18.04} mon_election/classic msgr-failures/osd-delay rados thrashers/none thrashosds-health workloads/radosbench} 3
Failure Reason:

Command failed on smithi037 with status 5: 'sudo systemctl stop ceph-b8374f7a-53db-11ed-8438-001a4aab830c@mon.a'

fail 7079882 2022-10-24 15:03:38 2022-10-24 20:28:50 2022-10-24 20:53:45 0:24:55 0:17:49 0:07:06 smithi main centos 8.stream rados/cephadm/mgr-nfs-upgrade/{0-distro/centos_8.stream_container_tools 1-bootstrap/16.2.5 1-start 2-nfs 3-upgrade-with-workload 4-final} 2
Failure Reason:

Command failed on smithi036 with status 1: 'sudo /home/ubuntu/cephtest/cephadm --image docker.io/ceph/ceph:v16.2.5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 2d163e96-53dc-11ed-8438-001a4aab830c -e sha1=d88f759b6ee30a8b0d9b03e2bbb29f39d48f9f2a -- bash -c \'ceph versions | jq -e \'"\'"\'.overall | keys\'"\'"\' | grep $sha1\''

fail 7079885 2022-10-24 15:03:39 2022-10-24 20:29:52 2022-10-24 20:43:06 0:13:14 0:06:50 0:06:24 smithi main centos 8.stream rados/cephadm/smoke-roleless/{0-distro/centos_8.stream_container_tools 0-nvme-loop 1-start 2-services/iscsi 3-final} 2
Failure Reason:

Command failed on smithi171 with status 5: 'sudo systemctl stop ceph-5a060cf6-53dc-11ed-8438-001a4aab830c@mon.smithi171'

fail 7079888 2022-10-24 15:03:40 2022-10-24 20:31:14 2022-10-24 20:48:49 0:17:35 0:09:38 0:07:57 smithi main centos 8.stream rados/cephadm/thrash/{0-distro/centos_8.stream_container_tools 1-start 2-thrash 3-tasks/rados_api_tests fixed-2 msgr/async root} 2
Failure Reason:

Command failed on smithi032 with status 5: 'sudo systemctl stop ceph-fbc066ea-53dc-11ed-8438-001a4aab830c@mon.a'

fail 7079891 2022-10-24 15:03:42 2022-10-24 20:32:16 2022-10-24 20:50:27 0:18:11 0:09:40 0:08:31 smithi main centos 8.stream rados/cephadm/with-work/{0-distro/centos_8.stream_container_tools fixed-2 mode/packaged mon_election/classic msgr/async start tasks/rados_api_tests} 2
Failure Reason:

Command failed on smithi120 with status 5: 'sudo systemctl stop ceph-36f4599c-53dd-11ed-8438-001a4aab830c@mon.a'

fail 7079894 2022-10-24 15:03:43 2022-10-24 20:35:18 2022-10-24 20:52:11 0:16:53 0:09:34 0:07:19 smithi main centos 8.stream rados/cephadm/workunits/{0-distro/centos_8.stream_container_tools mon_election/classic task/test_orch_cli} 1
Failure Reason:

Command failed on smithi079 with status 5: 'sudo systemctl stop ceph-718c5e60-53dd-11ed-8438-001a4aab830c@mon.a'

fail 7079897 2022-10-24 15:03:44 2022-10-24 20:36:19 2022-10-24 20:51:33 0:15:14 0:07:52 0:07:22 smithi main rhel 8.4 rados/cephadm/osds/{0-distro/rhel_8.4_container_tools_rhel8 0-nvme-loop 1-start 2-ops/rm-zap-add} 2
Failure Reason:

Command failed on smithi192 with status 5: 'sudo systemctl stop ceph-590c74d8-53dd-11ed-8438-001a4aab830c@mon.smithi192'

fail 7079900 2022-10-24 15:03:45 2022-10-24 20:38:01 2022-10-24 21:07:09 0:29:08 0:22:49 0:06:19 smithi main centos 8.stream rados/cephadm/mds_upgrade_sequence/{bluestore-bitmap centos_8.stream_container_tools conf/{client mds mon osd} overrides/{ignorelist_health ignorelist_wrongly_marked_down pg-warn} roles tasks/{0-from/v16.2.4 1-volume/{0-create 1-ranks/2 2-allow_standby_replay/no 3-inline/yes 4-verify} 2-client 3-upgrade-with-workload 4-verify}} 2
Failure Reason:

Command failed on smithi080 with status 1: 'sudo /home/ubuntu/cephtest/cephadm --image docker.io/ceph/ceph:v16.2.4 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid e4cc2978-53dd-11ed-8438-001a4aab830c -e sha1=d88f759b6ee30a8b0d9b03e2bbb29f39d48f9f2a -- bash -c \'ceph versions | jq -e \'"\'"\'.overall | keys\'"\'"\' | grep $sha1\''

fail 7079903 2022-10-24 15:03:47 2022-10-24 20:38:52 2022-10-24 20:53:57 0:15:05 0:05:00 0:10:05 smithi main ubuntu 20.04 rados/cephadm/smoke/{0-nvme-loop distro/ubuntu_20.04 fixed-2 mon_election/classic start} 2
Failure Reason:

Command failed on smithi037 with status 5: 'sudo systemctl stop ceph-bb0bf91a-53dd-11ed-8438-001a4aab830c@mon.a'

fail 7079906 2022-10-24 15:03:48 2022-10-24 20:39:53 2022-10-24 20:56:34 0:16:41 0:07:55 0:08:46 smithi main rhel 8.4 rados/cephadm/smoke-roleless/{0-distro/rhel_8.4_container_tools_3.0 0-nvme-loop 1-start 2-services/mirror 3-final} 2
Failure Reason:

Command failed on smithi133 with status 5: 'sudo systemctl stop ceph-0ca39486-53de-11ed-8438-001a4aab830c@mon.smithi133'

fail 7079909 2022-10-24 15:03:49 2022-10-24 20:42:25 2022-10-24 21:00:23 0:17:58 0:10:00 0:07:58 smithi main centos 8.stream rados/cephadm/workunits/{0-distro/centos_8.stream_container_tools mon_election/connectivity task/test_orch_cli_mon} 5
Failure Reason:

Command failed on smithi005 with status 5: 'sudo systemctl stop ceph-a44da90c-53de-11ed-8438-001a4aab830c@mon.a'

fail 7079912 2022-10-24 15:03:51 2022-10-24 20:44:07 2022-10-24 20:59:10 0:15:03 0:05:58 0:09:05 smithi main ubuntu 18.04 rados/cephadm/thrash-old-clients/{0-size-min-size-overrides/3-size-2-min-size 1-install/mimic backoff/normal ceph clusters/{openstack three-plus-one} d-balancer/on distro$/{ubuntu_18.04} mon_election/connectivity msgr-failures/fastclose rados thrashers/pggrow thrashosds-health workloads/rbd_cls} 3
Failure Reason:

Command failed on smithi040 with status 5: 'sudo systemctl stop ceph-6c31c77e-53de-11ed-8438-001a4aab830c@mon.a'

fail 7079915 2022-10-24 15:03:52 2022-10-24 20:47:19 2022-10-24 21:02:33 0:15:14 0:07:31 0:07:43 smithi main rhel 8.4 rados/cephadm/smoke-roleless/{0-distro/rhel_8.4_container_tools_rhel8 0-nvme-loop 1-start 2-services/nfs-ingress-rgw-bucket 3-final} 2
Failure Reason:

Command failed on smithi053 with status 5: 'sudo systemctl stop ceph-d90c06b6-53de-11ed-8438-001a4aab830c@mon.smithi053'

fail 7079918 2022-10-24 15:03:53 2022-10-24 20:49:00 2022-10-24 21:05:04 0:16:04 0:06:06 0:09:58 smithi main rados/cephadm/dashboard/{0-distro/ignorelist_health task/test_e2e} 2
Failure Reason:

Failed to fetch package version from https://shaman.ceph.com/api/search/?status=ready&project=ceph&flavor=default&distros=ubuntu%2F22.04%2Fx86_64&sha1=d88f759b6ee30a8b0d9b03e2bbb29f39d48f9f2a

fail 7079921 2022-10-24 15:03:54 2022-10-24 20:50:12 2022-10-24 21:18:09 0:27:57 0:20:59 0:06:58 smithi main centos 8.stream rados/cephadm/mds_upgrade_sequence/{bluestore-bitmap centos_8.stream_container_tools conf/{client mds mon osd} overrides/{ignorelist_health ignorelist_wrongly_marked_down pg-warn} roles tasks/{0-from/v16.2.4 1-volume/{0-create 1-ranks/2 2-allow_standby_replay/no 3-inline/no 4-verify} 2-client 3-upgrade-with-workload 4-verify}} 2
Failure Reason:

Command failed on smithi120 with status 1: 'sudo /home/ubuntu/cephtest/cephadm --image docker.io/ceph/ceph:v16.2.4 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ba257646-53df-11ed-8438-001a4aab830c -e sha1=d88f759b6ee30a8b0d9b03e2bbb29f39d48f9f2a -- bash -c \'ceph versions | jq -e \'"\'"\'.overall | keys\'"\'"\' | grep $sha1\''

fail 7079924 2022-10-24 15:03:56 2022-10-24 20:51:33 2022-10-24 21:06:26 0:14:53 0:07:26 0:07:27 smithi main centos 8.stream rados/cephadm/smoke/{0-nvme-loop distro/centos_8.stream_container_tools fixed-2 mon_election/connectivity start} 2
Failure Reason:

Command failed on smithi192 with status 5: 'sudo systemctl stop ceph-6771109a-53df-11ed-8438-001a4aab830c@mon.a'

fail 7079927 2022-10-24 15:03:57 2022-10-24 20:52:45 2022-10-24 21:04:05 0:11:20 0:05:22 0:05:58 smithi main ubuntu 18.04 rados/cephadm/smoke-singlehost/{0-distro$/{ubuntu_18.04} 1-start 2-services/rgw 3-final} 1
Failure Reason:

Command failed on smithi061 with status 5: 'sudo systemctl stop ceph-50d6375c-53df-11ed-8438-001a4aab830c@mon.smithi061'

fail 7079930 2022-10-24 15:03:58 2022-10-24 20:53:46 2022-10-24 21:10:40 0:16:54 0:10:34 0:06:20 smithi main centos 8.stream rados/cephadm/thrash/{0-distro/centos_8.stream_container_tools 1-start 2-thrash 3-tasks/radosbench fixed-2 msgr/async-v1only root} 2
Failure Reason:

Command failed on smithi003 with status 5: 'sudo systemctl stop ceph-2b0e0558-53e0-11ed-8438-001a4aab830c@mon.a'

fail 7079933 2022-10-24 15:04:00 2022-10-24 20:54:27 2022-10-24 21:20:01 0:25:34 0:17:23 0:08:11 smithi main centos 8.stream rados/cephadm/upgrade/{1-start-distro/1-start-centos_8.stream_container-tools 2-repo_digest/defaut 3-upgrade/staggered 4-wait 5-upgrade-ls mon_election/classic} 2
Failure Reason:

Command failed on smithi106 with status 22: 'sudo /home/ubuntu/cephtest/cephadm --image docker.io/ceph/ceph:v15.2.0 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid f1d08a7c-53df-11ed-8438-001a4aab830c -e sha1=d88f759b6ee30a8b0d9b03e2bbb29f39d48f9f2a -- bash -c \'ceph orch daemon redeploy "mgr.$(ceph mgr dump -f json | jq .standbys | jq .[] | jq -r .name)"\''

fail 7079936 2022-10-24 15:04:01 2022-10-24 20:55:39 2022-10-24 21:14:53 0:19:14 0:12:19 0:06:55 smithi main rhel 8.4 rados/cephadm/with-work/{0-distro/rhel_8.4_container_tools_3.0 fixed-2 mode/root mon_election/connectivity msgr/async-v1only start tasks/rados_python} 2
Failure Reason:

Command failed on smithi164 with status 5: 'sudo systemctl stop ceph-abda8422-53e0-11ed-8438-001a4aab830c@mon.a'

fail 7079939 2022-10-24 15:04:02 2022-10-24 20:56:40 2022-10-24 21:09:49 0:13:09 0:06:02 0:07:07 smithi main ubuntu 18.04 rados/cephadm/osds/{0-distro/ubuntu_18.04 0-nvme-loop 1-start 2-ops/rm-zap-flag} 2
Failure Reason:

Command failed on smithi006 with status 5: 'sudo systemctl stop ceph-edc680c6-53df-11ed-8438-001a4aab830c@mon.smithi006'

fail 7079942 2022-10-24 15:04:03 2022-10-24 20:57:01 2022-10-24 21:11:32 0:14:31 0:05:16 0:09:15 smithi main ubuntu 18.04 rados/cephadm/smoke-roleless/{0-distro/ubuntu_18.04 0-nvme-loop 1-start 2-services/nfs-ingress-rgw-user 3-final} 2
Failure Reason:

Command failed on smithi066 with status 5: 'sudo systemctl stop ceph-558cb7d4-53e0-11ed-8438-001a4aab830c@mon.smithi066'

fail 7079945 2022-10-24 15:04:05 2022-10-24 21:00:33 2022-10-24 21:14:01 0:13:28 0:06:40 0:06:48 smithi main ubuntu 18.04 rados/cephadm/thrash-old-clients/{0-size-min-size-overrides/2-size-2-min-size 1-install/nautilus-v1only backoff/peering ceph clusters/{openstack three-plus-one} d-balancer/crush-compat distro$/{ubuntu_18.04} mon_election/classic msgr-failures/few rados thrashers/careful thrashosds-health workloads/snaps-few-objects} 3
Failure Reason:

Command failed on smithi005 with status 5: 'sudo systemctl stop ceph-99a0b3a8-53e0-11ed-8438-001a4aab830c@mon.a'

fail 7079948 2022-10-24 15:04:06 2022-10-24 21:05:06 2022-10-24 21:21:39 0:16:33 0:10:16 0:06:17 smithi main centos 8.stream rados/cephadm/workunits/{0-distro/centos_8.stream_container_tools mon_election/classic task/test_cephadm} 1
Failure Reason:

Command failed (workunit test cephadm/test_cephadm.sh) on smithi017 with status 125: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=d88f759b6ee30a8b0d9b03e2bbb29f39d48f9f2a TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 CEPH_MNT=/home/ubuntu/cephtest/mnt.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/cephadm/test_cephadm.sh'

fail 7079951 2022-10-24 15:04:07 2022-10-24 21:05:48 2022-10-24 21:21:22 0:15:34 0:07:55 0:07:39 smithi main rhel 8.4 rados/cephadm/smoke/{0-nvme-loop distro/rhel_8.4_container_tools_3.0 fixed-2 mon_election/classic start} 2
Failure Reason:

Command failed on smithi192 with status 5: 'sudo systemctl stop ceph-840f0a16-53e1-11ed-8438-001a4aab830c@mon.a'

fail 7079954 2022-10-24 15:04:08 2022-10-24 21:07:19 2022-10-24 21:22:59 0:15:40 0:05:35 0:10:05 smithi main ubuntu 20.04 rados/cephadm/smoke-roleless/{0-distro/ubuntu_20.04 0-nvme-loop 1-start 2-services/nfs-ingress 3-final} 2
Failure Reason:

Command failed on smithi143 with status 5: 'sudo systemctl stop ceph-de49979e-53e1-11ed-8438-001a4aab830c@mon.smithi143'

fail 7079957 2022-10-24 15:04:10 2022-10-24 21:09:21 2022-10-24 21:39:14 0:29:53 0:21:31 0:08:22 smithi main centos 8.stream rados/cephadm/mds_upgrade_sequence/{bluestore-bitmap centos_8.stream_container_tools conf/{client mds mon osd} overrides/{ignorelist_health ignorelist_wrongly_marked_down pg-warn} roles tasks/{0-from/v16.2.4 1-volume/{0-create 1-ranks/1 2-allow_standby_replay/yes 3-inline/yes 4-verify} 2-client 3-upgrade-with-workload 4-verify}} 2
Failure Reason:

Command failed on smithi044 with status 1: 'sudo /home/ubuntu/cephtest/cephadm --image docker.io/ceph/ceph:v16.2.4 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 61d64bf2-53e2-11ed-8438-001a4aab830c -e sha1=d88f759b6ee30a8b0d9b03e2bbb29f39d48f9f2a -- bash -c \'ceph versions | jq -e \'"\'"\'.overall | keys\'"\'"\' | grep $sha1\''

fail 7079960 2022-10-24 15:04:11 2022-10-24 21:10:22 2022-10-24 21:24:10 0:13:48 0:07:03 0:06:45 smithi main ubuntu 18.04 rados/cephadm/thrash-old-clients/{0-size-min-size-overrides/3-size-2-min-size 1-install/nautilus-v2only backoff/peering_and_degraded ceph clusters/{openstack three-plus-one} d-balancer/on distro$/{ubuntu_18.04} mon_election/connectivity msgr-failures/osd-delay rados thrashers/default thrashosds-health workloads/test_rbd_api} 3
Failure Reason:

Command failed on smithi078 with status 5: 'sudo systemctl stop ceph-0fecc622-53e2-11ed-8438-001a4aab830c@mon.a'

fail 7079963 2022-10-24 15:04:12 2022-10-24 21:10:43 2022-10-24 21:38:20 0:27:37 0:20:51 0:06:46 smithi main centos 8.stream rados/cephadm/mgr-nfs-upgrade/{0-distro/centos_8.stream_container_tools 1-bootstrap/octopus 1-start 2-nfs 3-upgrade-with-workload 4-final} 2
Failure Reason:

Command failed on smithi175 with status 1: 'sudo /home/ubuntu/cephtest/cephadm --image docker.io/ceph/ceph:v15 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 23813754-53e2-11ed-8438-001a4aab830c -e sha1=d88f759b6ee30a8b0d9b03e2bbb29f39d48f9f2a -- bash -c \'ceph versions | jq -e \'"\'"\'.overall | keys\'"\'"\' | grep $sha1\''

fail 7079966 2022-10-24 15:04:13 2022-10-24 21:11:34 2022-10-24 21:24:52 0:13:18 0:07:06 0:06:12 smithi main centos 8.stream rados/cephadm/smoke-roleless/{0-distro/centos_8.stream_container_tools 0-nvme-loop 1-start 2-services/nfs-ingress2 3-final} 2
Failure Reason:

Command failed on smithi071 with status 5: 'sudo systemctl stop ceph-387d1d44-53e2-11ed-8438-001a4aab830c@mon.smithi071'

fail 7079969 2022-10-24 15:04:14 2022-10-24 21:14:06 2022-10-24 21:31:29 0:17:23 0:09:36 0:07:47 smithi main centos 8.stream rados/cephadm/thrash/{0-distro/centos_8.stream_container_tools 1-start 2-thrash 3-tasks/small-objects fixed-2 msgr/async-v2only root} 2
Failure Reason:

Command failed on smithi040 with status 5: 'sudo systemctl stop ceph-f138bf14-53e2-11ed-8438-001a4aab830c@mon.a'

fail 7079972 2022-10-24 15:04:15 2022-10-24 21:17:08 2022-10-24 21:32:38 0:15:30 0:07:08 0:08:22 smithi main ubuntu 18.04 rados/rook/smoke/{0-distro/ubuntu_18.04 0-kubeadm 1-rook 2-workload/radosbench 3-final cluster/3-node k8s/1.21 net/calico rook/1.6.2} 3
Failure Reason:

Command failed on smithi063 with status 1: 'kubectl create -f rook/cluster/examples/kubernetes/ceph/crds.yaml -f rook/cluster/examples/kubernetes/ceph/common.yaml -f operator.yaml'

fail 7079975 2022-10-24 15:04:17 2022-10-24 21:19:10 2022-10-24 21:37:26 0:18:16 0:10:53 0:07:23 smithi main rhel 8.4 rados/cephadm/with-work/{0-distro/rhel_8.4_container_tools_rhel8 fixed-2 mode/packaged mon_election/classic msgr/async-v2only start tasks/rados_api_tests} 2
Failure Reason:

Command failed on smithi106 with status 5: 'sudo systemctl stop ceph-e755f6c8-53e3-11ed-8438-001a4aab830c@mon.a'