Status Job ID Links Posted Started Updated
Runtime
Duration
In Waiting
Machine Teuthology Branch OS Type OS Version Description Nodes
pass 5388605 2020-08-29 19:23:32 2020-08-29 23:28:52 2020-08-30 00:04:52 0:36:00 0:25:15 0:10:45 smithi master ubuntu 18.04 rados/thrash-erasure-code-big/{ceph cluster/{12-osds openstack} msgr-failures/osd-delay objectstore/bluestore-comp-lz4 rados recovery-overrides/{more-async-recovery} supported-random-distro$/{ubuntu_latest} thrashers/careful thrashosds-health workloads/ec-rados-plugin=jerasure-k=4-m=2} 3
pass 5388606 2020-08-29 19:23:33 2020-08-29 23:28:53 2020-08-29 23:48:52 0:19:59 0:13:24 0:06:35 smithi master rhel 8.0 rados/cephadm/smoke-roleless/{distro/rhel_8.0 start} 2
pass 5388607 2020-08-29 19:23:34 2020-08-29 23:28:53 2020-08-30 00:02:53 0:34:00 0:26:43 0:07:17 smithi master ubuntu 18.04 rados/thrash-erasure-code-isa/{arch/x86_64 ceph clusters/{fixed-2 openstack} msgr-failures/osd-delay objectstore/bluestore-stupid rados recovery-overrides/{more-async-recovery} supported-random-distro$/{ubuntu_latest} thrashers/default thrashosds-health workloads/ec-rados-plugin=isa-k=2-m=1} 2
fail 5388608 2020-08-29 19:23:35 2020-08-29 23:28:53 2020-08-30 00:10:53 0:42:00 0:24:18 0:17:42 smithi master centos 7.6 rados/thrash-old-clients/{0-size-min-size-overrides/3-size-2-min-size 1-install/luminous-v1only backoff/normal ceph clusters/{openstack three-plus-one} d-balancer/on distro$/{centos_7.6} msgr-failures/few rados thrashers/mapgap thrashosds-health workloads/snaps-few-objects} 3
Failure Reason:

reached maximum tries (180) after waiting for 180 seconds

fail 5388609 2020-08-29 19:23:36 2020-08-29 23:29:06 2020-08-30 00:07:06 0:38:00 0:24:34 0:13:26 smithi master centos 7.6 rados/thrash-old-clients/{0-size-min-size-overrides/2-size-2-min-size 1-install/luminous backoff/peering ceph clusters/{openstack three-plus-one} d-balancer/crush-compat distro$/{centos_7.6} msgr-failures/osd-delay rados thrashers/morepggrow thrashosds-health workloads/test_rbd_api} 3
Failure Reason:

reached maximum tries (180) after waiting for 180 seconds

pass 5388610 2020-08-29 19:23:37 2020-08-29 23:30:32 2020-08-30 00:04:30 0:33:58 0:26:54 0:07:04 smithi master centos 8.1 rados/thrash-erasure-code-isa/{arch/x86_64 ceph clusters/{fixed-2 openstack} msgr-failures/few objectstore/bluestore-bitmap rados recovery-overrides/{more-async-partial-recovery} supported-random-distro$/{centos_8} thrashers/morepggrow thrashosds-health workloads/ec-rados-plugin=isa-k=2-m=1} 2
pass 5388611 2020-08-29 19:23:38 2020-08-29 23:30:32 2020-08-29 23:58:30 0:27:58 0:21:24 0:06:34 smithi master ubuntu 18.04 rados/singleton/{all/thrash_cache_writeback_proxy_none msgr-failures/many msgr/async-v1only objectstore/bluestore-low-osd-mem-target rados supported-random-distro$/{ubuntu_latest}} 2
pass 5388612 2020-08-29 19:23:39 2020-08-29 23:32:21 2020-08-30 01:42:24 2:10:03 1:38:10 0:31:53 smithi master centos 7.6 rados/thrash-old-clients/{0-size-min-size-overrides/2-size-2-min-size 1-install/mimic backoff/normal ceph clusters/{openstack three-plus-one} d-balancer/crush-compat distro$/{centos_7.6} msgr-failures/few rados thrashers/pggrow thrashosds-health workloads/radosbench} 3
pass 5388613 2020-08-29 19:23:40 2020-08-29 23:32:24 2020-08-30 00:10:24 0:38:00 0:25:28 0:12:32 smithi master centos 8.1 rados/thrash-erasure-code-big/{ceph cluster/{12-osds openstack} msgr-failures/osd-delay objectstore/bluestore-stupid rados recovery-overrides/{more-async-partial-recovery} supported-random-distro$/{centos_8} thrashers/careful thrashosds-health workloads/ec-rados-plugin=jerasure-k=4-m=2} 3
pass 5388614 2020-08-29 19:23:40 2020-08-29 23:32:26 2020-08-29 23:52:26 0:20:00 0:14:06 0:05:54 smithi master rhel 8.0 rados/cephadm/smoke-roleless/{distro/rhel_8.0 start} 2
fail 5388615 2020-08-29 19:23:41 2020-08-29 23:32:26 2020-08-30 02:56:31 3:24:05 3:16:58 0:07:07 smithi master ubuntu 18.04 rados/standalone/{supported-random-distro$/{ubuntu_latest} workloads/osd} 1
Failure Reason:

Command failed (workunit test osd/osd-rep-recov-eio.sh) on smithi029 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=08fab7a8a9103f87935c685c0a66d28e361bc9f5 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/standalone/osd/osd-rep-recov-eio.sh'

pass 5388616 2020-08-29 19:23:42 2020-08-29 23:32:40 2020-08-29 23:54:39 0:21:59 0:16:46 0:05:13 smithi master rhel 8.1 rados/singleton-nomsgr/{all/admin_socket_output rados supported-random-distro$/{rhel_8}} 1
pass 5388617 2020-08-29 19:23:43 2020-08-29 23:32:40 2020-08-30 00:14:40 0:42:00 0:26:47 0:15:13 smithi master centos 8.1 rados/thrash-erasure-code-big/{ceph cluster/{12-osds openstack} msgr-failures/fastclose objectstore/bluestore-comp-snappy rados recovery-overrides/{more-active-recovery} supported-random-distro$/{centos_8} thrashers/morepggrow thrashosds-health workloads/ec-rados-plugin=jerasure-k=4-m=2} 3