Name Machine Type Up Locked Locked Since Locked By OS Type OS Version Arch Description
smithi111.front.sepia.ceph.com smithi True False ubuntu 20.04 x86_64 None
Status Job ID Links Posted Started Updated
Runtime
Duration
In Waiting
Machine Teuthology Branch OS Type OS Version Description Nodes
fail 6290014 2021-07-24 05:57:52 2021-07-24 08:49:02 2021-07-24 09:10:53 0:21:51 0:04:38 0:17:13 smithi master ubuntu 20.04 upgrade/octopus-x/stress-split-no-cephadm/{0-cluster/{openstack start} 1-ceph-install/octopus 1.1-pg-log-overrides/short_pg_log 2-partial-upgrade/firsthalf 3-thrash/default 4-workload/{radosbench rbd-cls rbd-import-export rbd_api readwrite rgw_ragweed_prepare snaps-few-objects} 5-finish-upgrade 6-pacific 7-msgr2 8-final-workload/{rbd-python snaps-many-objects} mon_election/classic objectstore/filestore-xfs thrashosds-health ubuntu_20.04} 5
Failure Reason:

Command failed on smithi111 with status 100: 'sudo DEBIAN_FRONTEND=noninteractive apt-get -y --force-yes -o Dpkg::Options::="--force-confdef" -o Dpkg::Options::="--force-confold" install ceph=15.2.13-209-g02dd0874-1focal ceph-mds=15.2.13-209-g02dd0874-1focal ceph-mgr=15.2.13-209-g02dd0874-1focal ceph-common=15.2.13-209-g02dd0874-1focal ceph-fuse=15.2.13-209-g02dd0874-1focal ceph-test=15.2.13-209-g02dd0874-1focal radosgw=15.2.13-209-g02dd0874-1focal python3-rados=15.2.13-209-g02dd0874-1focal python3-rgw=15.2.13-209-g02dd0874-1focal python3-cephfs=15.2.13-209-g02dd0874-1focal python3-rbd=15.2.13-209-g02dd0874-1focal libcephfs2=15.2.13-209-g02dd0874-1focal librados2=15.2.13-209-g02dd0874-1focal librbd1=15.2.13-209-g02dd0874-1focal rbd-fuse=15.2.13-209-g02dd0874-1focal'

fail 6289992 2021-07-24 05:57:30 2021-07-24 08:29:43 2021-07-24 08:51:08 0:21:25 0:04:34 0:16:51 smithi master ubuntu 20.04 upgrade/octopus-x/stress-split-erasure-code-no-cephadm/{0-cluster/{openstack start} 1-nautilus-install/octopus 1.1-pg-log-overrides/short_pg_log 2-partial-upgrade/firsthalf 3-thrash/default 3.1-objectstore/filestore-xfs 4-ec-workload/{rados-ec-workload rbd-ec-workload} 5-finish-upgrade 6-pacific 7-final-workload mon_election/connectivity thrashosds-health ubuntu_20.04} 5
Failure Reason:

Command failed on smithi134 with status 100: 'sudo DEBIAN_FRONTEND=noninteractive apt-get -y --force-yes -o Dpkg::Options::="--force-confdef" -o Dpkg::Options::="--force-confold" install ceph=15.2.13-209-g02dd0874-1focal ceph-mds=15.2.13-209-g02dd0874-1focal ceph-mgr=15.2.13-209-g02dd0874-1focal ceph-common=15.2.13-209-g02dd0874-1focal ceph-fuse=15.2.13-209-g02dd0874-1focal ceph-test=15.2.13-209-g02dd0874-1focal radosgw=15.2.13-209-g02dd0874-1focal python3-rados=15.2.13-209-g02dd0874-1focal python3-rgw=15.2.13-209-g02dd0874-1focal python3-cephfs=15.2.13-209-g02dd0874-1focal python3-rbd=15.2.13-209-g02dd0874-1focal libcephfs2=15.2.13-209-g02dd0874-1focal librados2=15.2.13-209-g02dd0874-1focal librbd1=15.2.13-209-g02dd0874-1focal rbd-fuse=15.2.13-209-g02dd0874-1focal'

fail 6289971 2021-07-24 05:57:09 2021-07-24 08:15:14 2021-07-24 08:31:13 0:15:59 0:04:08 0:11:51 smithi master upgrade/octopus-x/rgw-multisite/{clusters frontend overrides realm tasks upgrade/secondary} 2
Failure Reason:

Command failed on smithi152 with status 100: 'sudo DEBIAN_FRONTEND=noninteractive apt-get -y --force-yes -o Dpkg::Options::="--force-confdef" -o Dpkg::Options::="--force-confold" install ceph=15.2.13-209-g02dd0874-1focal cephadm=15.2.13-209-g02dd0874-1focal ceph-mds=15.2.13-209-g02dd0874-1focal ceph-mgr=15.2.13-209-g02dd0874-1focal ceph-common=15.2.13-209-g02dd0874-1focal ceph-fuse=15.2.13-209-g02dd0874-1focal ceph-test=15.2.13-209-g02dd0874-1focal radosgw=15.2.13-209-g02dd0874-1focal python3-rados=15.2.13-209-g02dd0874-1focal python3-rgw=15.2.13-209-g02dd0874-1focal python3-cephfs=15.2.13-209-g02dd0874-1focal python3-rbd=15.2.13-209-g02dd0874-1focal libcephfs2=15.2.13-209-g02dd0874-1focal libcephfs-dev=15.2.13-209-g02dd0874-1focal librados2=15.2.13-209-g02dd0874-1focal librbd1=15.2.13-209-g02dd0874-1focal rbd-fuse=15.2.13-209-g02dd0874-1focal'

pass 6289905 2021-07-24 05:28:36 2021-07-24 07:35:54 2021-07-24 08:15:23 0:39:29 0:28:08 0:11:21 smithi master centos 8.2 rados/cephadm/mgr-nfs-upgrade/{0-centos_8.2_kubic_stable 1-bootstrap/octopus 1-start 2-nfs 3-upgrade-with-workload 4-final} 2
pass 6289856 2021-07-24 05:27:53 2021-07-24 07:10:27 2021-07-24 07:35:24 0:24:57 0:16:14 0:08:43 smithi master rhel 8.4 rados/multimon/{clusters/6 mon_election/connectivity msgr-failures/few msgr/async no_pools objectstore/bluestore-comp-snappy rados supported-random-distro$/{rhel_8} tasks/mon_clock_no_skews} 2
pass 6289818 2021-07-24 05:27:20 2021-07-24 06:51:32 2021-07-24 07:11:42 0:20:10 0:09:57 0:10:13 smithi master ubuntu 20.04 rados/perf/{ceph mon_election/classic objectstore/bluestore-bitmap openstack scheduler/dmclock_default_shards settings/optimized ubuntu_latest workloads/sample_radosbench} 1
fail 6289736 2021-07-24 05:26:10 2021-07-24 06:14:00 2021-07-24 06:52:00 0:38:00 0:28:15 0:09:45 smithi master ubuntu 20.04 rados/rook/smoke/{0-distro/ubuntu_20.04 0-kubeadm 1-rook 2-workload/radosbench 3-final cluster/1-node k8s/1.21 net/calico rook/1.6.2} 1
Failure Reason:

Command failed on smithi111 with status 22: 'kubectl -n rook-ceph exec rook-ceph-tools-7467d8bf8-l2c7z -- ceph orch device ls'

fail 6289646 2021-07-24 05:24:42 2021-07-24 05:25:03 2021-07-24 06:14:13 0:49:10 0:36:46 0:12:24 smithi master ubuntu 20.04 rados/rook/smoke/{0-distro/ubuntu_20.04 0-kubeadm 1-rook 2-workload/radosbench 3-final cluster/1-node k8s/1.21 net/calico rook/master} 1
Failure Reason:

Command failed on smithi111 with status 22: 'kubectl -n rook-ceph exec rook-ceph-tools-78cdfd976c-47h7b -- ceph orch device ls'

dead 6289294 2021-07-24 03:40:01 2021-07-24 03:40:01 2021-07-24 03:58:35 0:18:34 smithi master rhel 8.4 rados/singleton-nomsgr/{all/health-warnings mon_election/classic rados supported-random-distro$/{rhel_8}} 1
pass 6289168 2021-07-23 17:41:37 2021-07-23 19:35:13 2021-07-23 20:12:59 0:37:46 0:26:17 0:11:29 smithi master centos 8.3 rados/thrash/{0-size-min-size-overrides/2-size-2-min-size 1-pg-log-overrides/normal_pg_log 2-recovery-overrides/{default} 3-scrub-overrides/{default} backoff/peering ceph clusters/{fixed-2 openstack} crc-failures/bad_map_crc_failure d-balancer/off msgr-failures/osd-dispatch-delay msgr/async-v2only objectstore/bluestore-stupid rados supported-random-distro$/{centos_8} thrashers/morepggrow thrashosds-health workloads/cache-snaps} 2
pass 6289081 2021-07-23 17:40:08 2021-07-23 18:55:29 2021-07-23 19:35:24 0:39:55 0:29:02 0:10:53 smithi master centos 8.3 rados/thrash/{0-size-min-size-overrides/3-size-2-min-size 1-pg-log-overrides/short_pg_log 2-recovery-overrides/{more-active-recovery} 3-scrub-overrides/{default} backoff/peering ceph clusters/{fixed-2 openstack} crc-failures/default d-balancer/off msgr-failures/osd-delay msgr/async-v2only objectstore/bluestore-stupid rados supported-random-distro$/{centos_8} thrashers/careful thrashosds-health workloads/snaps-few-objects-localized} 2
pass 6289011 2021-07-23 17:38:55 2021-07-23 18:27:18 2021-07-23 18:55:42 0:28:24 0:17:22 0:11:02 smithi master ubuntu 20.04 rados/thrash/{0-size-min-size-overrides/2-size-2-min-size 1-pg-log-overrides/normal_pg_log 2-recovery-overrides/{more-partial-recovery} 3-scrub-overrides/{default} backoff/normal ceph clusters/{fixed-2 openstack} crc-failures/bad_map_crc_failure d-balancer/crush-compat msgr-failures/osd-dispatch-delay msgr/async-v1only objectstore/bluestore-bitmap rados supported-random-distro$/{ubuntu_latest} thrashers/pggrow thrashosds-health workloads/redirect_set_object} 2
pass 6288972 2021-07-23 17:38:14 2021-07-23 18:06:36 2021-07-23 18:27:37 0:21:01 0:09:32 0:11:29 smithi master ubuntu 18.04 rados/perf/{ceph objectstore/bluestore-low-osd-mem-target openstack settings/optimized ubuntu_18.04 workloads/radosbench_4M_write} 1
pass 6288818 2021-07-23 15:31:55 2021-07-23 16:43:59 2021-07-23 18:07:39 1:23:40 1:16:05 0:07:35 smithi master rhel 8.4 fs:workload/{begin/{0-install 1-cephadm 2-logrotate} clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} mount/fuse objectstore-ec/bluestore-ec-root omap_limit/10000 overrides/{frag_enable osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/5 rhel_8 scrub/yes standby-replay tasks/{0-check-counter workunit/suites/ffsb} wsync/{yes}} 3
fail 6288798 2021-07-23 15:31:35 2021-07-23 16:22:39 2021-07-23 16:44:10 0:21:31 0:14:58 0:06:33 smithi master rhel 8.4 fs:workload/{begin/{0-install 1-cephadm 2-logrotate} clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} mount/kclient/{mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} objectstore-ec/bluestore-comp omap_limit/10000 overrides/{frag_enable osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/1 rhel_8 scrub/yes standby-replay tasks/{0-check-counter workunit/suites/pjd} wsync/{no}} 3
Failure Reason:

Command failed on smithi111 with status 1: 'sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:dc6e6aee83d05e539da09c20751974c378bcb8fa shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 7f7a43c4-ebd4-11eb-8c23-001a4aab830c -- ceph mon dump -f json'

pass 6288752 2021-07-23 15:30:45 2021-07-23 15:31:19 2021-07-23 16:22:41 0:51:22 0:40:44 0:10:38 smithi master rhel 8.4 fs:workload/{begin/{0-install 1-cephadm 2-logrotate} clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} mount/fuse objectstore-ec/bluestore-comp-ec-root omap_limit/10000 overrides/{frag_enable osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/5 rhel_8 scrub/yes standby-replay tasks/{0-check-counter workunit/suites/fsync-tester} wsync/{no}} 3
pass 6288716 2021-07-23 14:33:49 2021-07-23 14:33:49 2021-07-23 14:57:51 0:24:02 0:13:25 0:10:37 smithi master centos 8.stream rados:thrash/{0-size-min-size-overrides/3-size-2-min-size 1-pg-log-overrides/normal_pg_log 2-recovery-overrides/{more-async-partial-recovery} 3-scrub-overrides/{max-simultaneous-scrubs-2} backoff/normal ceph clusters/{fixed-2 openstack} crc-failures/default d-balancer/crush-compat mon_election/connectivity msgr-failures/fastclose msgr/async-v2only objectstore/bluestore-stupid rados supported-random-distro$/{centos_8.stream} thrashers/none thrashosds-health workloads/redirect_set_object} 2
pass 6288488 2021-07-23 06:46:10 2021-07-23 06:54:24 2021-07-23 07:24:32 0:30:08 0:19:42 0:10:26 smithi master ubuntu 20.04 rados:cephadm/smoke-roleless/{0-distro/ubuntu_20.04 1-start 2-services/nfs-ingress2 3-final} 2
pass 6288452 2021-07-23 05:20:10 2021-07-23 07:55:58 2021-07-23 08:26:37 0:30:39 0:12:23 0:18:16 smithi master krbd/rbd/{bluestore-bitmap clusters/fixed-3 conf ms_mode/crc msgr-failures/few tasks/rbd_workunit_suites_fsstress_ext4} 3
pass 6288400 2021-07-23 05:19:31 2021-07-23 07:24:00 2021-07-23 07:56:34 0:32:34 0:19:17 0:13:17 smithi master krbd/rbd/{bluestore-bitmap clusters/fixed-3 conf ms_mode/crc msgr-failures/many tasks/rbd_workunit_kernel_untar_build} 3