Status Job ID Links Posted Started Updated
Runtime
Duration
In Waiting
Machine Teuthology Branch OS Type OS Version Description Nodes
pass 4513866 2019-11-17 01:29:41 2019-11-17 01:29:48 2019-11-17 02:19:48 0:50:00 0:27:09 0:22:51 mira master centos 7.6 rados/singleton/{all/random-eio.yaml msgr-failures/few.yaml msgr/async-v1only.yaml objectstore/bluestore-avl.yaml rados.yaml supported-random-distro$/{centos_7.yaml}} 2
pass 4513867 2019-11-17 01:29:42 2019-11-17 01:29:49 2019-11-17 01:57:48 0:27:59 0:19:18 0:08:41 mira master rhel 7.7 rados/mgr/{clusters/{2-node-mgr.yaml} debug/mgr.yaml objectstore/bluestore-comp.yaml supported-random-distro$/{rhel_7.yaml} tasks/ssh_orchestrator.yaml} 2
pass 4513868 2019-11-17 01:29:43 2019-11-17 01:29:49 2019-11-17 02:27:48 0:57:59 0:41:46 0:16:13 mira master centos 7.6 rados/thrash-old-clients/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-install/nautilus.yaml backoff/peering_and_degraded.yaml ceph.yaml clusters/{openstack.yaml three-plus-one.yaml} d-balancer/crush-compat.yaml distro$/{centos_latest.yaml} msgr-failures/fastclose.yaml rados.yaml thrashers/pggrow.yaml thrashosds-health.yaml workloads/snaps-few-objects.yaml} 4
pass 4513869 2019-11-17 01:29:44 2019-11-17 01:29:49 2019-11-17 02:03:48 0:33:59 0:23:49 0:10:10 mira master ubuntu 18.04 rados/basic/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/few.yaml msgr/async-v2only.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} tasks/rados_api_tests.yaml} 2
fail 4513870 2019-11-17 01:29:45 2019-11-17 01:29:49 2019-11-17 02:09:49 0:40:00 0:30:39 0:09:21 mira master ubuntu 18.04 rados/singleton-bluestore/{all/cephtool.yaml msgr-failures/few.yaml msgr/async-v2only.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}} 1
Failure Reason:

Command failed (workunit test cephtool/test.sh) on mira082 with status 13: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=70968846f95926b5f8ae4f251788ce3818969018 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/cephtool/test.sh'

pass 4513871 2019-11-17 01:29:45 2019-11-17 01:57:49 2019-11-17 02:25:49 0:28:00 0:21:37 0:06:23 mira master rhel 7.7 rados/singleton/{all/rebuild-mondb.yaml msgr-failures/many.yaml msgr/async-v2only.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}} 1
pass 4513872 2019-11-17 01:29:46 2019-11-17 02:03:56 2019-11-17 02:29:55 0:25:59 0:15:15 0:10:44 mira master ubuntu 18.04 rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml 2-recovery-overrides/{more-active-recovery.yaml} backoff/peering.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/upmap.yaml msgr-failures/osd-delay.yaml msgr/async-v1only.yaml objectstore/filestore-xfs.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} thrashers/none.yaml thrashosds-health.yaml workloads/write_fadvise_dontneed.yaml} 2
pass 4513873 2019-11-17 01:29:47 2019-11-17 02:09:50 2019-11-17 03:35:51 1:26:01 1:15:30 0:10:31 mira master ubuntu 18.04 rados/perf/{ceph.yaml objectstore/bluestore-low-osd-mem-target.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{ubuntu_latest.yaml} workloads/radosbench_omap_write.yaml} 1
pass 4513874 2019-11-17 01:29:48 2019-11-17 02:19:49 2019-11-17 03:05:49 0:46:00 0:35:37 0:10:23 mira master ubuntu 18.04 rados/objectstore/{backends/objectcacher-stress.yaml supported-random-distro$/{ubuntu_latest.yaml}} 1
pass 4513875 2019-11-17 01:29:49 2019-11-17 02:26:06 2019-11-17 03:10:05 0:43:59 0:22:50 0:21:09 mira master centos 7.6 rados/singleton-nomsgr/{all/osd_stale_reads.yaml rados.yaml supported-random-distro$/{centos_7.yaml}} 1
pass 4513876 2019-11-17 01:29:50 2019-11-17 02:27:50 2019-11-17 03:05:50 0:38:00 0:26:56 0:11:04 mira master ubuntu 18.04 rados/thrash-erasure-code-overwrites/{bluestore-bitmap.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} fast/fast.yaml msgr-failures/fastclose.yaml rados.yaml recovery-overrides/{more-partial-recovery.yaml} supported-random-distro$/{ubuntu_latest.yaml} thrashers/pggrow.yaml thrashosds-health.yaml workloads/ec-small-objects-overwrites.yaml} 2
pass 4513877 2019-11-17 01:29:51 2019-11-17 02:30:12 2019-11-17 05:16:14 2:46:02 2:27:21 0:18:41 mira master rhel 7.7 rados/thrash-erasure-code-shec/{ceph.yaml clusters/{fixed-4.yaml openstack.yaml} msgr-failures/fastclose.yaml objectstore/filestore-xfs.yaml rados.yaml recovery-overrides/{more-async-partial-recovery.yaml} supported-random-distro$/{rhel_7.yaml} thrashers/careful.yaml thrashosds-health.yaml workloads/ec-rados-plugin=shec-k=4-m=3-c=2.yaml} 4
pass 4513878 2019-11-17 01:29:52 2019-11-17 03:05:55 2019-11-17 04:07:55 1:02:00 0:50:28 0:11:32 mira master centos 7.6 rados/singleton/{all/recovery-preemption.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{centos_7.yaml}} 1
dead 4513879 2019-11-17 01:29:52 2019-11-17 03:05:55 2019-11-17 15:08:19 12:02:24 mira master ubuntu 18.04 rados/thrash-erasure-code/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} fast/fast.yaml msgr-failures/osd-delay.yaml objectstore/bluestore-stupid.yaml rados.yaml recovery-overrides/{more-partial-recovery.yaml} supported-random-distro$/{ubuntu_latest.yaml} thrashers/default.yaml thrashosds-health.yaml workloads/ec-rados-plugin=clay-k=4-m=2.yaml} 2
pass 4513880 2019-11-17 01:29:53 2019-11-17 03:10:18 2019-11-17 03:36:17 0:25:59 0:14:45 0:11:14 mira master ubuntu 18.04 rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml 2-recovery-overrides/{default.yaml} backoff/peering_and_degraded.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/crush-compat.yaml msgr-failures/fastclose.yaml msgr/async-v2only.yaml objectstore/bluestore-avl.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} thrashers/pggrow.yaml thrashosds-health.yaml workloads/admin_socket_objecter_requests.yaml} 2
pass 4513881 2019-11-17 01:29:54 2019-11-17 03:35:52 2019-11-17 06:21:54 2:46:02 2:26:58 0:19:04 mira master rhel 7.7 rados/singleton/{all/resolve_stuck_peering.yaml msgr-failures/many.yaml msgr/async-v1only.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}} 2
pass 4513882 2019-11-17 01:29:55 2019-11-17 03:36:18 2019-11-17 04:02:18 0:26:00 0:14:36 0:11:24 mira master centos 7.6 rados/perf/{ceph.yaml objectstore/bluestore-stupid.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{centos_7.yaml} workloads/sample_fio.yaml} 1
pass 4513883 2019-11-17 01:29:56 2019-11-17 04:02:31 2019-11-17 04:24:30 0:21:59 0:11:48 0:10:11 mira master ubuntu 18.04 rados/multimon/{clusters/6.yaml msgr-failures/few.yaml msgr/async-v2only.yaml no_pools.yaml objectstore/bluestore-stupid.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} tasks/mon_recovery.yaml} 2
pass 4513884 2019-11-17 01:29:57 2019-11-17 04:08:13 2019-11-17 04:44:13 0:36:00 0:27:54 0:08:06 mira master rhel 7.7 rados/monthrash/{ceph.yaml clusters/3-mons.yaml msgr-failures/mon-delay.yaml msgr/async-v2only.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml supported-random-distro$/{rhel_7.yaml} thrashers/sync.yaml workloads/pool-create-delete.yaml} 2
pass 4513885 2019-11-17 01:29:58 2019-11-17 04:24:33 2019-11-17 05:18:32 0:53:59 0:19:55 0:34:04 mira master centos 7.6 rados/thrash-old-clients/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-install/hammer.yaml backoff/normal.yaml ceph.yaml clusters/{openstack.yaml three-plus-one.yaml} d-balancer/off.yaml distro$/{centos_latest.yaml} msgr-failures/few.yaml rados.yaml thrashers/pggrow.yaml thrashosds-health.yaml workloads/test_rbd_api.yaml} 4
pass 4513886 2019-11-17 01:29:58 2019-11-17 04:44:15 2019-11-17 05:08:14 0:23:59 0:13:23 0:10:36 mira master ubuntu 18.04 rados/basic/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/many.yaml msgr/async.yaml objectstore/bluestore-stupid.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} tasks/rados_cls_all.yaml} 2
pass 4513887 2019-11-17 01:29:59 2019-11-17 05:08:17 2019-11-17 06:34:17 1:26:00 1:08:23 0:17:37 mira master centos 7.6 rados/dashboard/{clusters/{2-node-mgr.yaml} debug/mgr.yaml objectstore/filestore-xfs.yaml supported-random-distro$/{centos_7.yaml} tasks/dashboard.yaml} 2
pass 4513888 2019-11-17 01:30:00 2019-11-17 05:16:32 2019-11-17 05:42:31 0:25:59 0:19:48 0:06:11 mira master rhel 7.7 rados/singleton-nomsgr/{all/pool-access.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}} 1
pass 4513889 2019-11-17 01:30:01 2019-11-17 05:18:34 2019-11-17 05:38:33 0:19:59 0:09:08 0:10:51 mira master ubuntu 18.04 rados/mgr/{clusters/{2-node-mgr.yaml} debug/mgr.yaml objectstore/bluestore-low-osd-mem-target.yaml supported-random-distro$/{ubuntu_latest.yaml} tasks/workunits.yaml} 2
pass 4513890 2019-11-17 01:30:02 2019-11-17 05:38:37 2019-11-17 06:36:37 0:58:00 0:39:34 0:18:26 mira master centos 7.6 rados/thrash-erasure-code-isa/{arch/x86_64.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/few.yaml objectstore/bluestore-comp.yaml rados.yaml recovery-overrides/{default.yaml} supported-random-distro$/{centos_7.yaml} thrashers/mapgap.yaml thrashosds-health.yaml workloads/ec-rados-plugin=isa-k=2-m=1.yaml} 2
pass 4513891 2019-11-17 01:30:03 2019-11-17 05:42:33 2019-11-17 06:06:32 0:23:59 0:12:05 0:11:54 mira master ubuntu 18.04 rados/thrash-erasure-code-big/{ceph.yaml cluster/{12-osds.yaml openstack.yaml} msgr-failures/fastclose.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml recovery-overrides/{more-partial-recovery.yaml} supported-random-distro$/{ubuntu_latest.yaml} thrashers/morepggrow.yaml thrashosds-health.yaml workloads/ec-rados-plugin=lrc-k=4-m=2-l=3.yaml} 3
pass 4513892 2019-11-17 01:30:03 2019-11-17 06:06:35 2019-11-17 06:30:34 0:23:59 0:11:57 0:12:02 mira master rados/verify/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-thrash/none.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/bluestore-stupid.yaml rados.yaml tasks/mon_recovery.yaml validater/lockdep.yaml} 2
pass 4513893 2019-11-17 01:30:04 2019-11-17 06:22:12 2019-11-17 06:52:11 0:29:59 0:21:18 0:08:41 mira master rhel 7.7 rados/singleton/{all/test-crash.yaml msgr-failures/few.yaml msgr/async-v2only.yaml objectstore/bluestore-stupid.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}} 1
pass 4513894 2019-11-17 01:30:05 2019-11-17 06:30:46 2019-11-17 07:22:46 0:52:00 0:41:13 0:10:47 mira master ubuntu 18.04 rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml 2-recovery-overrides/{more-async-recovery.yaml} backoff/normal.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/off.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} thrashers/careful.yaml thrashosds-health.yaml workloads/cache-agent-big.yaml} 2
pass 4513895 2019-11-17 01:30:06 2019-11-17 06:34:57 2019-11-17 07:58:58 1:24:01 1:13:11 0:10:50 mira master ubuntu 18.04 rados/standalone/{supported-random-distro$/{ubuntu_latest.yaml} workloads/scrub.yaml} 1
pass 4513896 2019-11-17 01:30:07 2019-11-17 06:36:39 2019-11-17 07:06:38 0:29:59 0:21:35 0:08:24 mira master rhel 7.7 rados/perf/{ceph.yaml objectstore/bluestore-basic-min-osd-mem-target.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{rhel_7.yaml} workloads/sample_radosbench.yaml} 1
pass 4513897 2019-11-17 01:30:08 2019-11-17 06:52:29 2019-11-17 09:46:32 2:54:03 2:35:14 0:18:49 mira master rhel 7.7 rados/singleton/{all/test_envlibrados_for_rocksdb.yaml msgr-failures/many.yaml msgr/async.yaml objectstore/filestore-xfs.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}} 1
fail 4513898 2019-11-17 01:30:08 2019-11-17 07:06:40 2019-11-17 09:08:41 2:02:01 1:38:04 0:23:57 mira master ubuntu 18.04 rados/objectstore/{backends/objectstore.yaml supported-random-distro$/{ubuntu_latest.yaml}} 1
Failure Reason:

Command crashed: 'sudo TESTDIR=/home/ubuntu/cephtest bash -c \'mkdir $TESTDIR/archive/ostest && cd $TESTDIR/archive/ostest && ulimit -Sn 16384 && CEPH_ARGS="--no-log-to-stderr --log-file $TESTDIR/archive/ceph_test_objectstore.log --debug-filestore 20 --debug-bluestore 20" ceph_test_objectstore --gtest_filter=-*/3 --gtest_catch_exceptions=0\''

pass 4513899 2019-11-17 01:30:09 2019-11-17 07:22:50 2019-11-17 08:06:50 0:44:00 0:36:45 0:07:15 mira master rhel 7.7 rados/singleton-nomsgr/{all/recovery-unfound-found.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}} 1
pass 4513900 2019-11-17 01:30:10 2019-11-17 07:59:01 2019-11-17 08:39:00 0:39:59 0:30:10 0:09:49 mira master rhel 7.7 rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml 2-recovery-overrides/{more-active-recovery.yaml} backoff/peering.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/upmap.yaml msgr-failures/osd-delay.yaml msgr/async-v1only.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{rhel_7.yaml} thrashers/default.yaml thrashosds-health.yaml workloads/cache-agent-small.yaml} 2
pass 4513901 2019-11-17 01:30:11 2019-11-17 08:06:53 2019-11-17 09:04:53 0:58:00 0:47:37 0:10:23 mira master ubuntu 18.04 rados/singleton/{all/thrash-backfill-full.yaml msgr-failures/few.yaml msgr/async-v1only.yaml objectstore/bluestore-avl.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}} 2
pass 4513902 2019-11-17 01:30:12 2019-11-17 08:39:03 2019-11-17 09:13:03 0:34:00 0:20:16 0:13:44 mira master centos 7.6 rados/basic/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/few.yaml msgr/async-v1only.yaml objectstore/filestore-xfs.yaml rados.yaml supported-random-distro$/{centos_7.yaml} tasks/rados_python.yaml} 2
fail 4513903 2019-11-17 01:30:13 2019-11-17 09:05:02 2019-11-17 10:07:02 1:02:00 0:36:32 0:25:28 mira master centos 7.6 rados/thrash-old-clients/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-install/jewel-v1only.yaml backoff/peering.yaml ceph.yaml clusters/{openstack.yaml three-plus-one.yaml} d-balancer/crush-compat.yaml distro$/{centos_latest.yaml} msgr-failures/osd-delay.yaml rados.yaml thrashers/careful.yaml thrashosds-health.yaml workloads/cache-snaps.yaml} 4
Failure Reason:

"2019-11-17T10:02:29.359546+0000 osd.2 (osd.2) 38 : cluster [ERR] osd.2 found snap mapper error on pg 2.1 oid 2:8ac095d2:::mira0721068-471 oooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooooo:8d snaps missing in mapper, should be: 8a was r -2...repaired" in cluster log

pass 4513904 2019-11-17 01:30:13 2019-11-17 09:09:00 2019-11-17 12:11:02 3:02:02 2:42:49 0:19:13 mira master rhel 7.7 rados/thrash-erasure-code/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} fast/normal.yaml msgr-failures/fastclose.yaml objectstore/filestore-xfs.yaml rados.yaml recovery-overrides/{more-async-partial-recovery.yaml} supported-random-distro$/{rhel_7.yaml} thrashers/fastread.yaml thrashosds-health.yaml workloads/ec-rados-plugin=jerasure-k=2-m=1.yaml} 2
pass 4513905 2019-11-17 01:30:14 2019-11-17 09:13:05 2019-11-17 09:43:04 0:29:59 0:22:50 0:07:09 mira master rhel 7.7 rados/mgr/{clusters/{2-node-mgr.yaml} debug/mgr.yaml objectstore/bluestore-avl.yaml supported-random-distro$/{rhel_7.yaml} tasks/crash.yaml} 2
pass 4513906 2019-11-17 01:30:15 2019-11-17 09:44:20 2019-11-17 10:10:02 0:25:42 0:15:52 0:09:50 mira master ubuntu 18.04 rados/perf/{ceph.yaml objectstore/bluestore-basic-min-osd-mem-target.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{ubuntu_latest.yaml} workloads/cosbench_64K_read_write.yaml} 1
pass 4513907 2019-11-17 01:30:16 2019-11-17 09:46:50 2019-11-17 12:56:53 3:10:03 2:49:27 0:20:36 mira master rhel 7.7 rados/thrash-erasure-code-overwrites/{bluestore-bitmap.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} fast/fast.yaml msgr-failures/fastclose.yaml rados.yaml recovery-overrides/{more-partial-recovery.yaml} supported-random-distro$/{rhel_7.yaml} thrashers/careful.yaml thrashosds-health.yaml workloads/ec-pool-snaps-few-objects-overwrites.yaml} 2
pass 4513908 2019-11-17 01:30:17 2019-11-17 10:07:21 2019-11-17 10:41:21 0:34:00 0:25:42 0:08:18 mira master rhel 7.7 rados/thrash-erasure-code-shec/{ceph.yaml clusters/{fixed-4.yaml openstack.yaml} msgr-failures/fastclose.yaml objectstore/bluestore-avl.yaml rados.yaml recovery-overrides/{more-async-partial-recovery.yaml} supported-random-distro$/{rhel_7.yaml} thrashers/careful.yaml thrashosds-health.yaml workloads/ec-rados-plugin=shec-k=4-m=3-c=2.yaml} 4
pass 4513909 2019-11-17 01:30:18 2019-11-17 10:10:04 2019-11-17 11:00:04 0:50:00 0:35:03 0:14:57 mira master centos 7.6 rados/singleton/{all/thrash-eio.yaml msgr-failures/many.yaml msgr/async-v2only.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{centos_7.yaml}} 2
pass 4513910 2019-11-17 01:30:18 2019-11-17 10:41:23 2019-11-17 11:07:23 0:26:00 0:19:51 0:06:09 mira master rhel 7.7 rados/singleton-nomsgr/{all/version-number-sanity.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}} 1
pass 4513911 2019-11-17 01:30:19 2019-11-17 11:00:07 2019-11-17 11:54:07 0:54:00 0:36:56 0:17:04 mira master centos 7.6 rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml 2-recovery-overrides/{more-async-recovery.yaml} backoff/peering_and_degraded.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/crush-compat.yaml msgr-failures/fastclose.yaml msgr/async-v2only.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml supported-random-distro$/{centos_7.yaml} thrashers/mapgap.yaml thrashosds-health.yaml workloads/cache-pool-snaps-readproxy.yaml} 2
pass 4513912 2019-11-17 01:30:20 2019-11-17 11:07:25 2019-11-17 11:51:25 0:44:00 0:30:01 0:13:59 mira master centos 7.6 rados/singleton/{all/thrash-rados/{thrash-rados.yaml thrashosds-health.yaml} msgr-failures/few.yaml msgr/async.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{centos_7.yaml}} 2
pass 4513913 2019-11-17 01:30:21 2019-11-17 11:51:42 2019-11-17 12:19:42 0:28:00 0:17:57 0:10:03 mira master rhel 7.7 rados/multimon/{clusters/9.yaml msgr-failures/many.yaml msgr/async.yaml no_pools.yaml objectstore/filestore-xfs.yaml rados.yaml supported-random-distro$/{rhel_7.yaml} tasks/mon_clock_no_skews.yaml} 3
pass 4513914 2019-11-17 01:30:22 2019-11-17 11:54:09 2019-11-17 12:24:08 0:29:59 0:16:04 0:13:55 mira master centos 7.6 rados/monthrash/{ceph.yaml clusters/9-mons.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/bluestore-stupid.yaml rados.yaml supported-random-distro$/{centos_7.yaml} thrashers/force-sync-many.yaml workloads/rados_5925.yaml} 2
pass 4513915 2019-11-17 01:30:23 2019-11-17 12:11:22 2019-11-17 15:15:24 3:04:02 2:44:39 0:19:23 mira master rhel 7.7 rados/thrash-erasure-code-isa/{arch/x86_64.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/osd-delay.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml recovery-overrides/{more-async-partial-recovery.yaml} supported-random-distro$/{rhel_7.yaml} thrashers/morepggrow.yaml thrashosds-health.yaml workloads/ec-rados-plugin=isa-k=2-m=1.yaml} 2
pass 4513916 2019-11-17 01:30:23 2019-11-17 12:19:56 2019-11-17 13:03:56 0:44:00 0:32:57 0:11:03 mira master ubuntu 18.04 rados/thrash-erasure-code-big/{ceph.yaml cluster/{12-osds.yaml openstack.yaml} msgr-failures/few.yaml objectstore/bluestore-stupid.yaml rados.yaml recovery-overrides/{default.yaml} supported-random-distro$/{ubuntu_latest.yaml} thrashers/pggrow.yaml thrashosds-health.yaml workloads/ec-rados-plugin=jerasure-k=4-m=2.yaml} 3
fail 4513917 2019-11-17 01:30:24 2019-11-17 12:24:12 2019-11-17 19:10:18 6:46:06 6:32:40 0:13:26 mira master centos rados/verify/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-thrash/default/{default.yaml thrashosds-health.yaml} msgr-failures/few.yaml msgr/async-v1only.yaml objectstore/filestore-xfs.yaml rados.yaml tasks/rados_api_tests.yaml validater/valgrind.yaml} 2
Failure Reason:

Command failed (workunit test rados/test.sh) on mira063 with status 124: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=70968846f95926b5f8ae4f251788ce3818969018 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 6h /home/ubuntu/cephtest/clone.client.0/qa/workunits/rados/test.sh'

pass 4513918 2019-11-17 01:30:25 2019-11-17 12:57:11 2019-11-17 13:45:11 0:48:00 0:24:44 0:23:16 mira master centos 7.6 rados/basic/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/many.yaml msgr/async-v2only.yaml objectstore/bluestore-avl.yaml rados.yaml supported-random-distro$/{centos_7.yaml} tasks/rados_stress_watch.yaml} 2
pass 4513919 2019-11-17 01:30:26 2019-11-17 13:04:03 2019-11-17 13:30:02 0:25:59 0:15:43 0:10:16 mira master ubuntu 18.04 rados/perf/{ceph.yaml objectstore/bluestore-bitmap.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{ubuntu_latest.yaml} workloads/cosbench_64K_write.yaml} 1
pass 4513920 2019-11-17 01:30:27 2019-11-17 13:30:04 2019-11-17 14:16:04 0:46:00 0:32:20 0:13:40 mira master centos 7.6 rados/singleton/{all/thrash_cache_writeback_proxy_none.yaml msgr-failures/many.yaml msgr/async-v1only.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml supported-random-distro$/{centos_7.yaml}} 2
pass 4513921 2019-11-17 01:30:27 2019-11-17 13:45:14 2019-11-17 14:39:14 0:54:00 0:40:29 0:13:31 mira master centos 7.6 rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml 2-recovery-overrides/{more-active-recovery.yaml} backoff/normal.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/off.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/bluestore-stupid.yaml rados.yaml supported-random-distro$/{centos_7.yaml} thrashers/morepggrow.yaml thrashosds-health.yaml workloads/cache-pool-snaps.yaml} 2
pass 4513922 2019-11-17 01:30:28 2019-11-17 14:16:26 2019-11-17 14:54:25 0:37:59 0:21:24 0:16:35 mira master centos 7.6 rados/thrash-old-clients/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-install/jewel.yaml backoff/peering_and_degraded.yaml ceph.yaml clusters/{openstack.yaml three-plus-one.yaml} d-balancer/off.yaml distro$/{centos_latest.yaml} msgr-failures/fastclose.yaml rados.yaml thrashers/default.yaml thrashosds-health.yaml workloads/radosbench.yaml} 4
pass 4513923 2019-11-17 01:30:29 2019-11-17 14:39:16 2019-11-17 16:13:18 1:34:02 1:08:03 0:25:59 mira master centos 7.6 rados/dashboard/{clusters/{2-node-mgr.yaml} debug/mgr.yaml objectstore/bluestore-avl.yaml supported-random-distro$/{centos_7.yaml} tasks/dashboard.yaml} 2
pass 4513924 2019-11-17 01:30:30 2019-11-17 14:54:33 2019-11-17 15:20:33 0:26:00 0:20:20 0:05:40 mira master rhel 7.7 rados/objectstore/{backends/alloc-hint.yaml supported-random-distro$/{rhel_7.yaml}} 1
pass 4513925 2019-11-17 01:30:31 2019-11-17 15:08:24 2019-11-17 15:36:23 0:27:59 0:20:47 0:07:12 mira master rhel 7.7 rados/rest/{mgr-restful.yaml supported-random-distro$/{rhel_7.yaml}} 1
fail 4513926 2019-11-17 01:30:31 2019-11-17 15:15:27 2019-11-17 18:21:29 3:06:02 2:46:35 0:19:27 mira master rhel 7.7 rados/singleton-bluestore/{all/cephtool.yaml msgr-failures/many.yaml msgr/async.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}} 1
Failure Reason:

Command failed (workunit test cephtool/test.sh) on mira061 with status 13: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=70968846f95926b5f8ae4f251788ce3818969018 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/cephtool/test.sh'

pass 4513927 2019-11-17 01:30:32 2019-11-17 15:20:51 2019-11-17 15:52:50 0:31:59 0:21:54 0:10:05 mira master centos rados/singleton-flat/valgrind-leaks.yaml 1
pass 4513928 2019-11-17 01:30:33 2019-11-17 15:36:43 2019-11-17 16:06:42 0:29:59 0:24:00 0:05:59 mira master rhel 7.7 rados/singleton-nomsgr/{all/admin_socket_output.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}} 1
pass 4513929 2019-11-17 01:30:34 2019-11-17 15:52:52 2019-11-17 16:22:52 0:30:00 0:22:23 0:07:37 mira master rhel 7.7 rados/standalone/{supported-random-distro$/{rhel_7.yaml} workloads/crush.yaml} 1
fail 4513930 2019-11-17 01:30:35 2019-11-17 16:07:00 2019-11-17 16:44:59 0:37:59 0:27:20 0:10:39 mira master rhel 7.7 rados/upgrade/mimic-x-singleton/{0-cluster/{openstack.yaml start.yaml} 1-install/mimic.yaml 2-partial-upgrade/firsthalf.yaml 3-thrash/default.yaml 4-workload/{rbd-cls.yaml rbd-import-export.yaml readwrite.yaml snaps-few-objects.yaml} 5-workload/{radosbench.yaml rbd_api.yaml} 6-finish-upgrade.yaml 7-nautilus.yaml 8-workload/{rbd-python.yaml rgw-swift.yaml snaps-many-objects.yaml} bluestore-bitmap.yaml supported-random-distro$/{rhel_7.yaml} thrashosds-health.yaml} 4
Failure Reason:

Command failed (workunit test cls/test_cls_rbd.sh) on mira083 with status 139: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=mimic TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/cls/test_cls_rbd.sh'

pass 4513931 2019-11-17 01:30:36 2019-11-17 16:13:37 2019-11-17 16:39:37 0:26:00 0:19:05 0:06:55 mira master rhel 7.7 rados/singleton/{all/watch-notify-same-primary.yaml msgr-failures/few.yaml msgr/async-v2only.yaml objectstore/bluestore-stupid.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}} 1
pass 4513932 2019-11-17 01:30:36 2019-11-17 16:23:08 2019-11-17 19:07:10 2:44:02 2:25:46 0:18:16 mira master rhel 7.7 rados/mgr/{clusters/{2-node-mgr.yaml} debug/mgr.yaml objectstore/bluestore-bitmap.yaml supported-random-distro$/{rhel_7.yaml} tasks/failover.yaml} 2
pass 4513933 2019-11-17 01:30:37 2019-11-17 16:39:39 2019-11-17 17:05:38 0:25:59 0:14:49 0:11:10 mira master centos 7.6 rados/perf/{ceph.yaml objectstore/bluestore-comp.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{centos_7.yaml} workloads/fio_4K_rand_read.yaml} 1
pass 4513934 2019-11-17 01:30:38 2019-11-17 16:45:17 2019-11-17 17:15:17 0:30:00 0:20:10 0:09:50 mira master ubuntu 18.04 rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml 2-recovery-overrides/{more-active-recovery.yaml} backoff/peering.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/upmap.yaml msgr-failures/osd-delay.yaml msgr/async-v1only.yaml objectstore/filestore-xfs.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} thrashers/none.yaml thrashosds-health.yaml workloads/cache-snaps.yaml} 2
pass 4513935 2019-11-17 01:30:39 2019-11-17 17:05:40 2019-11-17 17:21:39 0:15:59 0:06:51 0:09:08 mira master ubuntu 18.04 rados/singleton/{all/admin-socket.yaml msgr-failures/many.yaml msgr/async-v2only.yaml objectstore/bluestore-avl.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}} 1
pass 4513936 2019-11-17 01:30:39 2019-11-17 17:15:36 2019-11-17 17:57:35 0:41:59 0:30:07 0:11:52 mira master ubuntu 18.04 rados/thrash-erasure-code/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} fast/fast.yaml msgr-failures/few.yaml objectstore/bluestore-avl.yaml rados.yaml recovery-overrides/{default.yaml} supported-random-distro$/{ubuntu_latest.yaml} thrashers/minsize_recovery.yaml thrashosds-health.yaml workloads/ec-rados-plugin=jerasure-k=3-m=1.yaml} 2
pass 4513937 2019-11-17 01:30:40 2019-11-17 17:21:41 2019-11-17 17:41:40 0:19:59 0:08:56 0:11:03 mira master ubuntu 18.04 rados/basic/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/few.yaml msgr/async.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} tasks/rados_striper.yaml} 2
pass 4513938 2019-11-17 01:30:41 2019-11-17 17:41:58 2019-11-17 20:24:00 2:42:02 2:24:26 0:17:36 mira master rhel 7.7 rados/singleton-nomsgr/{all/balancer.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}} 1
pass 4513939 2019-11-17 01:30:42 2019-11-17 17:57:37 2019-11-17 18:33:36 0:35:59 0:24:20 0:11:39 mira master ubuntu 18.04 rados/thrash-erasure-code-overwrites/{bluestore-bitmap.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} fast/normal.yaml msgr-failures/few.yaml rados.yaml recovery-overrides/{more-async-recovery.yaml} supported-random-distro$/{ubuntu_latest.yaml} thrashers/default.yaml thrashosds-health.yaml workloads/ec-small-objects-fast-read-overwrites.yaml} 2
pass 4513940 2019-11-17 01:30:43 2019-11-17 18:21:50 2019-11-17 21:07:52 2:46:02 2:27:50 0:18:12 mira master rhel 7.7 rados/thrash-erasure-code-shec/{ceph.yaml clusters/{fixed-4.yaml openstack.yaml} msgr-failures/few.yaml objectstore/bluestore-bitmap.yaml rados.yaml recovery-overrides/{more-async-partial-recovery.yaml} supported-random-distro$/{rhel_7.yaml} thrashers/default.yaml thrashosds-health.yaml workloads/ec-rados-plugin=shec-k=4-m=3-c=2.yaml} 4
pass 4513941 2019-11-17 01:30:43 2019-11-17 18:33:56 2019-11-17 18:51:55 0:17:59 0:08:36 0:09:23 mira master ubuntu 18.04 rados/singleton/{all/deduptool.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}} 1
fail 4513942 2019-11-17 01:30:44 2019-11-17 18:51:57 2019-11-17 19:31:57 0:40:00 0:21:50 0:18:10 mira master centos 7.6 rados/thrash-old-clients/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-install/luminous-v1only.yaml backoff/normal.yaml ceph.yaml clusters/{openstack.yaml three-plus-one.yaml} d-balancer/crush-compat.yaml distro$/{centos_latest.yaml} msgr-failures/few.yaml rados.yaml thrashers/mapgap.yaml thrashosds-health.yaml workloads/rbd_cls.yaml} 4
Failure Reason:

Command failed on mira105 with status 1: "sudo TESTDIR=/home/ubuntu/cephtest bash -c 'ceph_test_cls_rbd --gtest_filter=-TestClsRbd.get_features:TestClsRbd.parents'"

fail 4513943 2019-11-17 01:30:45 2019-11-17 19:07:28 2019-11-17 19:31:27 0:23:59 0:13:37 0:10:22 mira master ubuntu 18.04 rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml 2-recovery-overrides/{more-active-recovery.yaml} backoff/peering_and_degraded.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/crush-compat.yaml msgr-failures/fastclose.yaml msgr/async-v2only.yaml objectstore/bluestore-avl.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} thrashers/pggrow.yaml thrashosds-health.yaml workloads/cache.yaml} 2
Failure Reason:

Command crashed: 'CEPH_CLIENT_ID=0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph_test_rados --max-ops 4000 --objects 500 --max-in-flight 16 --size 4000000 --min-stride-size 400000 --max-stride-size 800000 --max-seconds 0 --op read 100 --op copy_from 50 --op write 50 --op write_excl 50 --op cache_try_flush 50 --op cache_flush 50 --op cache_evict 50 --op delete 50 --pool base'

pass 4513944 2019-11-17 01:30:46 2019-11-17 19:10:38 2019-11-17 19:30:37 0:19:59 0:09:47 0:10:12 mira master ubuntu 18.04 rados/perf/{ceph.yaml objectstore/bluestore-low-osd-mem-target.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{ubuntu_latest.yaml} workloads/fio_4K_rand_rw.yaml} 1
pass 4513945 2019-11-17 01:30:47 2019-11-17 19:30:57 2019-11-17 19:54:56 0:23:59 0:14:04 0:09:55 mira master ubuntu 18.04 rados/singleton/{all/divergent_priors.yaml msgr-failures/many.yaml msgr/async-v1only.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}} 1
pass 4513946 2019-11-17 01:30:47 2019-11-17 19:31:28 2019-11-17 19:49:27 0:17:59 0:07:31 0:10:28 mira master ubuntu 18.04 rados/multimon/{clusters/9.yaml msgr-failures/few.yaml msgr/async-v1only.yaml no_pools.yaml objectstore/bluestore-avl.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} tasks/mon_clock_with_skews.yaml} 3
pass 4513947 2019-11-17 01:30:48 2019-11-17 19:32:16 2019-11-17 20:20:16 0:48:00 0:39:17 0:08:43 mira master rhel 7.7 rados/monthrash/{ceph.yaml clusters/3-mons.yaml msgr-failures/mon-delay.yaml msgr/async-v1only.yaml objectstore/filestore-xfs.yaml rados.yaml supported-random-distro$/{rhel_7.yaml} thrashers/many.yaml workloads/rados_api_tests.yaml} 2
pass 4513948 2019-11-17 01:30:49 2019-11-17 19:49:29 2019-11-17 20:11:29 0:22:00 0:13:03 0:08:57 mira master ubuntu 18.04 rados/objectstore/{backends/ceph_objectstore_tool.yaml supported-random-distro$/{ubuntu_latest.yaml}} 1
pass 4513949 2019-11-17 01:30:50 2019-11-17 19:55:15 2019-11-17 20:23:14 0:27:59 0:13:46 0:14:13 mira master centos 7.6 rados/singleton-nomsgr/{all/cache-fs-trunc.yaml rados.yaml supported-random-distro$/{centos_7.yaml}} 1
pass 4513950 2019-11-17 01:30:50 2019-11-17 20:11:31 2019-11-17 20:41:30 0:29:59 0:23:34 0:06:25 mira master rhel 7.7 rados/mgr/{clusters/{2-node-mgr.yaml} debug/mgr.yaml objectstore/bluestore-comp.yaml supported-random-distro$/{rhel_7.yaml} tasks/insights.yaml} 2
pass 4513951 2019-11-17 01:30:51 2019-11-17 20:20:35 2019-11-17 21:08:35 0:48:00 0:41:12 0:06:48 mira master rhel 7.7 rados/thrash-erasure-code-isa/{arch/x86_64.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/fastclose.yaml objectstore/bluestore-stupid.yaml rados.yaml recovery-overrides/{default.yaml} supported-random-distro$/{rhel_7.yaml} thrashers/none.yaml thrashosds-health.yaml workloads/ec-rados-plugin=isa-k=2-m=1.yaml} 2
pass 4513952 2019-11-17 01:30:52 2019-11-17 20:23:16 2019-11-17 20:47:15 0:23:59 0:11:51 0:12:08 mira master ubuntu 18.04 rados/thrash-erasure-code-big/{ceph.yaml cluster/{12-osds.yaml openstack.yaml} msgr-failures/osd-delay.yaml objectstore/bluestore-stupid.yaml rados.yaml recovery-overrides/{more-active-recovery.yaml} supported-random-distro$/{ubuntu_latest.yaml} thrashers/careful.yaml thrashosds-health.yaml workloads/ec-rados-plugin=lrc-k=4-m=2-l=3.yaml} 3
pass 4513953 2019-11-17 01:30:53 2019-11-17 20:24:01 2019-11-17 20:56:01 0:32:00 0:20:35 0:11:25 mira master rados/verify/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-thrash/none.yaml msgr-failures/few.yaml msgr/async-v2only.yaml objectstore/bluestore-avl.yaml rados.yaml tasks/rados_api_tests.yaml validater/lockdep.yaml} 2
pass 4513954 2019-11-17 01:30:54 2019-11-17 20:41:47 2019-11-17 20:59:47 0:18:00 0:09:10 0:08:50 mira master ubuntu 18.04 rados/singleton/{all/divergent_priors2.yaml msgr-failures/few.yaml msgr/async-v2only.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}} 1
pass 4513955 2019-11-17 01:30:54 2019-11-17 20:47:35 2019-11-17 21:13:34 0:25:59 0:14:03 0:11:56 mira master ubuntu 18.04 rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml 2-recovery-overrides/{more-active-recovery.yaml} backoff/normal.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/off.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} thrashers/careful.yaml thrashosds-health.yaml workloads/dedup_tier.yaml} 2
pass 4513956 2019-11-17 01:30:55 2019-11-17 20:56:16 2019-11-18 00:00:19 3:04:03 2:45:28 0:18:35 mira master rhel 7.7 rados/basic/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/many.yaml msgr/async-v1only.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{rhel_7.yaml} tasks/rados_workunit_loadgen_big.yaml} 2
pass 4513957 2019-11-17 01:30:56 2019-11-17 20:59:48 2019-11-17 21:27:48 0:28:00 0:20:53 0:07:07 mira master rhel 7.7 rados/perf/{ceph.yaml objectstore/bluestore-stupid.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{rhel_7.yaml} workloads/fio_4M_rand_read.yaml} 1
pass 4513958 2019-11-17 01:30:57 2019-11-17 21:07:54 2019-11-17 21:51:53 0:43:59 0:35:22 0:08:37 mira master ubuntu 18.04 rados/standalone/{supported-random-distro$/{ubuntu_latest.yaml} workloads/erasure-code.yaml} 1
pass 4513959 2019-11-17 01:30:58 2019-11-17 21:08:37 2019-11-17 21:26:36 0:17:59 0:08:59 0:09:00 mira master ubuntu 18.04 rados/singleton/{all/dump-stuck.yaml msgr-failures/many.yaml msgr/async.yaml objectstore/bluestore-stupid.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}} 1
pass 4513960 2019-11-17 01:30:59 2019-11-17 21:13:52 2019-11-17 22:21:52 1:08:00 0:40:52 0:27:08 mira master centos 7.6 rados/thrash-old-clients/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-install/luminous.yaml backoff/peering.yaml ceph.yaml clusters/{openstack.yaml three-plus-one.yaml} d-balancer/off.yaml distro$/{centos_latest.yaml} msgr-failures/osd-delay.yaml rados.yaml thrashers/morepggrow.yaml thrashosds-health.yaml workloads/snaps-few-objects.yaml} 4
pass 4513961 2019-11-17 01:30:59 2019-11-17 21:26:55 2019-11-17 22:24:55 0:58:00 0:47:20 0:10:40 mira master ubuntu 18.04 rados/dashboard/{clusters/{2-node-mgr.yaml} debug/mgr.yaml objectstore/bluestore-bitmap.yaml supported-random-distro$/{ubuntu_latest.yaml} tasks/dashboard.yaml} 2
fail 4513962 2019-11-17 01:31:00 2019-11-17 21:27:49 2019-11-17 21:47:48 0:19:59 0:09:27 0:10:32 mira master ubuntu 18.04 rados/singleton-nomsgr/{all/ceph-daemon.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}} 1
Failure Reason:

Command failed (workunit test test_ceph_daemon.sh) on mira105 with status 2: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=70968846f95926b5f8ae4f251788ce3818969018 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/standalone/test_ceph_daemon.sh'

pass 4513963 2019-11-17 01:31:01 2019-11-17 21:47:57 2019-11-17 22:47:57 1:00:00 0:49:54 0:10:06 mira master ubuntu 18.04 rados/thrash-erasure-code/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} fast/normal.yaml msgr-failures/osd-delay.yaml objectstore/bluestore-bitmap.yaml rados.yaml recovery-overrides/{more-async-partial-recovery.yaml} supported-random-distro$/{ubuntu_latest.yaml} thrashers/morepggrow.yaml thrashosds-health.yaml workloads/ec-radosbench.yaml} 2
pass 4513964 2019-11-17 01:31:02 2019-11-17 21:51:55 2019-11-18 00:57:57 3:06:02 2:46:54 0:19:08 mira master rhel 7.7 rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml 2-recovery-overrides/{more-async-recovery.yaml} backoff/peering.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/upmap.yaml msgr-failures/osd-delay.yaml msgr/async-v1only.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{rhel_7.yaml} thrashers/default.yaml thrashosds-health.yaml workloads/pool-snaps-few-objects.yaml} 2
pass 4513965 2019-11-17 01:31:03 2019-11-17 22:22:10 2019-11-17 23:58:10 1:36:00 1:27:39 0:08:21 mira master rhel 7.7 rados/singleton/{all/ec-lost-unfound.yaml msgr-failures/few.yaml msgr/async-v1only.yaml objectstore/filestore-xfs.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}} 1
pass 4513966 2019-11-17 01:31:04 2019-11-17 22:25:13 2019-11-17 22:45:12 0:19:59 0:09:21 0:10:38 mira master ubuntu 18.04 rados/perf/{ceph.yaml objectstore/bluestore-basic-min-osd-mem-target.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{ubuntu_latest.yaml} workloads/fio_4M_rand_rw.yaml} 1
fail 4513967 2019-11-17 01:31:04 2019-11-17 22:45:31 2019-11-17 23:21:30 0:35:59 0:25:18 0:10:41 mira master ubuntu 18.04 rados/mgr/{clusters/{2-node-mgr.yaml} debug/mgr.yaml objectstore/bluestore-low-osd-mem-target.yaml supported-random-distro$/{ubuntu_latest.yaml} tasks/module_selftest.yaml} 2
Failure Reason:

"2019-11-17T23:06:15.291787+0000 mds.a (mds.0) 1 : cluster [WRN] evicting unresponsive client mira063:x (4415), after 300.764 seconds" in cluster log

pass 4513968 2019-11-17 01:31:05 2019-11-17 22:48:14 2019-11-17 23:34:13 0:45:59 0:32:01 0:13:58 mira master centos 7.6 rados/thrash-erasure-code-overwrites/{bluestore-bitmap.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} fast/fast.yaml msgr-failures/osd-delay.yaml rados.yaml recovery-overrides/{more-async-partial-recovery.yaml} supported-random-distro$/{centos_7.yaml} thrashers/fastread.yaml thrashosds-health.yaml workloads/ec-small-objects-overwrites.yaml} 2
pass 4513969 2019-11-17 01:31:06 2019-11-17 23:21:48 2019-11-18 00:09:48 0:48:00 0:21:44 0:26:16 mira master centos 7.6 rados/thrash-erasure-code-shec/{ceph.yaml clusters/{fixed-4.yaml openstack.yaml} msgr-failures/osd-delay.yaml objectstore/bluestore-comp.yaml rados.yaml recovery-overrides/{more-async-partial-recovery.yaml} supported-random-distro$/{centos_7.yaml} thrashers/careful.yaml thrashosds-health.yaml workloads/ec-rados-plugin=shec-k=4-m=3-c=2.yaml} 4
pass 4513970 2019-11-17 01:31:07 2019-11-17 23:34:15 2019-11-17 23:58:14 0:23:59 0:11:44 0:12:15 mira master centos 7.6 rados/singleton/{all/erasure-code-nonregression.yaml msgr-failures/many.yaml msgr/async-v2only.yaml objectstore/bluestore-avl.yaml rados.yaml supported-random-distro$/{centos_7.yaml}} 1
pass 4513971 2019-11-17 01:31:08 2019-11-17 23:58:14 2019-11-18 00:30:13 0:31:59 0:21:32 0:10:27 mira master ubuntu 18.04 rados/basic/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/few.yaml msgr/async-v2only.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} tasks/rados_workunit_loadgen_mix.yaml} 2
pass 4513972 2019-11-17 01:31:09 2019-11-17 23:58:16 2019-11-18 00:40:15 0:41:59 0:28:35 0:13:24 mira master centos 7.6 rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml 2-recovery-overrides/{more-async-recovery.yaml} backoff/peering_and_degraded.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/crush-compat.yaml msgr-failures/fastclose.yaml msgr/async-v2only.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml supported-random-distro$/{centos_7.yaml} thrashers/mapgap.yaml thrashosds-health.yaml workloads/rados_api_tests.yaml} 2
pass 4513973 2019-11-17 01:31:09 2019-11-18 00:00:37 2019-11-18 00:18:36 0:17:59 0:07:28 0:10:31 mira master ubuntu 18.04 rados/objectstore/{backends/filejournal.yaml supported-random-distro$/{ubuntu_latest.yaml}} 1
pass 4513974 2019-11-17 01:31:10 2019-11-18 00:10:12 2019-11-18 00:28:11 0:17:59 0:08:21 0:09:38 mira master ubuntu 18.04 rados/singleton-nomsgr/{all/ceph-kvstore-tool.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}} 1
pass 4513975 2019-11-17 01:31:11 2019-11-18 00:18:38 2019-11-18 01:48:39 1:30:01 1:23:45 0:06:16 mira master rhel 7.7 rados/singleton/{all/lost-unfound-delete.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}} 1
pass 4513976 2019-11-17 01:31:12 2019-11-18 00:28:13 2019-11-18 01:16:12 0:47:59 0:31:21 0:16:38 mira master centos 7.6 rados/multimon/{clusters/21.yaml msgr-failures/many.yaml msgr/async-v2only.yaml no_pools.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{centos_7.yaml} tasks/mon_recovery.yaml} 3
pass 4513977 2019-11-17 01:31:13 2019-11-18 00:30:15 2019-11-18 02:16:16 1:46:01 1:31:33 0:14:28 mira master centos 7.6 rados/monthrash/{ceph.yaml clusters/9-mons.yaml msgr-failures/few.yaml msgr/async-v2only.yaml objectstore/bluestore-avl.yaml rados.yaml supported-random-distro$/{centos_7.yaml} thrashers/one.yaml workloads/rados_mon_osdmap_prune.yaml} 2
pass 4513978 2019-11-17 01:31:14 2019-11-18 00:40:32 2019-11-18 01:28:32 0:48:00 0:18:53 0:29:07 mira master centos 7.6 rados/thrash-old-clients/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-install/mimic-v1only.yaml backoff/peering_and_degraded.yaml ceph.yaml clusters/{openstack.yaml three-plus-one.yaml} d-balancer/crush-compat.yaml distro$/{centos_latest.yaml} msgr-failures/fastclose.yaml rados.yaml thrashers/none.yaml thrashosds-health.yaml workloads/test_rbd_api.yaml} 4
pass 4513979 2019-11-17 01:31:14 2019-11-18 00:58:15 2019-11-18 01:18:15 0:20:00 0:09:35 0:10:25 mira master ubuntu 18.04 rados/perf/{ceph.yaml objectstore/bluestore-bitmap.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{ubuntu_latest.yaml} workloads/fio_4M_rand_write.yaml} 1
fail 4513980 2019-11-17 01:31:15 2019-11-18 01:16:16 2019-11-18 01:56:16 0:40:00 0:30:58 0:09:02 mira master ubuntu 18.04 rados/singleton-bluestore/{all/cephtool.yaml msgr-failures/few.yaml msgr/async-v1only.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}} 1
Failure Reason:

Command failed (workunit test cephtool/test.sh) on mira057 with status 13: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=70968846f95926b5f8ae4f251788ce3818969018 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/cephtool/test.sh'

pass 4513981 2019-11-17 01:31:16 2019-11-18 01:18:30 2019-11-18 04:24:33 3:06:03 2:42:40 0:23:23 mira master rhel 7.7 rados/thrash-erasure-code-isa/{arch/x86_64.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/few.yaml objectstore/filestore-xfs.yaml rados.yaml recovery-overrides/{more-active-recovery.yaml} supported-random-distro$/{rhel_7.yaml} thrashers/pggrow.yaml thrashosds-health.yaml workloads/ec-rados-plugin=isa-k=2-m=1.yaml} 2
pass 4513982 2019-11-17 01:31:17 2019-11-18 01:28:43 2019-11-18 02:23:00 0:54:17 0:37:58 0:16:19 mira master centos 7.6 rados/thrash-erasure-code-big/{ceph.yaml cluster/{12-osds.yaml openstack.yaml} msgr-failures/fastclose.yaml objectstore/filestore-xfs.yaml rados.yaml recovery-overrides/{default.yaml} supported-random-distro$/{centos_7.yaml} thrashers/default.yaml thrashosds-health.yaml workloads/ec-rados-plugin=jerasure-k=4-m=2.yaml} 3
dead 4513983 2019-11-17 01:31:18 2019-11-18 01:48:57 2019-11-18 13:51:19 12:02:22 mira master centos rados/verify/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-thrash/default/{default.yaml thrashosds-health.yaml} msgr-failures/few.yaml msgr/async.yaml objectstore/bluestore-bitmap.yaml rados.yaml tasks/rados_cls_all.yaml validater/valgrind.yaml} 2
pass 4513984 2019-11-17 01:31:19 2019-11-18 01:56:17 2019-11-18 03:28:18 1:32:01 1:21:31 0:10:30 mira master centos 7.6 rados/singleton/{all/lost-unfound.yaml msgr-failures/many.yaml msgr/async-v1only.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{centos_7.yaml}} 1
pass 4513985 2019-11-17 01:31:19 2019-11-18 02:16:35 2019-11-18 03:02:34 0:45:59 0:35:41 0:10:18 mira master ubuntu 18.04 rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml 2-recovery-overrides/{more-partial-recovery.yaml} backoff/normal.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/off.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/bluestore-stupid.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} thrashers/morepggrow.yaml thrashosds-health.yaml workloads/radosbench.yaml} 2
pass 4513986 2019-11-17 01:31:20 2019-11-18 02:23:17 2019-11-18 02:39:16 0:15:59 0:06:49 0:09:10 mira master ubuntu 18.04 rados/singleton-nomsgr/{all/ceph-post-file.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}} 1
pass 4513987 2019-11-17 01:31:21 2019-11-18 02:39:18 2019-11-18 03:29:18 0:50:00 0:36:30 0:13:30 mira master centos 7.6 rados/basic/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/many.yaml msgr/async.yaml objectstore/bluestore-stupid.yaml rados.yaml supported-random-distro$/{centos_7.yaml} tasks/rados_workunit_loadgen_mostlyread.yaml} 2
pass 4513988 2019-11-17 01:31:22 2019-11-18 03:02:37 2019-11-18 03:20:36 0:17:59 0:08:10 0:09:49 mira master ubuntu 18.04 rados/standalone/{supported-random-distro$/{ubuntu_latest.yaml} workloads/mgr.yaml} 1
pass 4513989 2019-11-17 01:31:23 2019-11-18 03:20:46 2019-11-18 03:38:45 0:17:59 0:07:38 0:10:21 mira master ubuntu 18.04 rados/singleton/{all/max-pg-per-osd.from-mon.yaml msgr-failures/few.yaml msgr/async-v2only.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}} 1
pass 4513990 2019-11-17 01:31:23 2019-11-18 03:28:37 2019-11-18 04:12:37 0:44:00 0:30:07 0:13:53 mira master centos 7.6 rados/thrash-erasure-code/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} fast/fast.yaml msgr-failures/fastclose.yaml objectstore/bluestore-comp.yaml rados.yaml recovery-overrides/{more-partial-recovery.yaml} supported-random-distro$/{centos_7.yaml} thrashers/pggrow.yaml thrashosds-health.yaml workloads/ec-small-objects-fast-read.yaml} 2
fail 4513991 2019-11-17 01:31:24 2019-11-18 03:29:19 2019-11-18 03:49:19 0:20:00 0:09:33 0:10:27 mira master ubuntu 18.04 rados/mgr/{clusters/{2-node-mgr.yaml} debug/mgr.yaml objectstore/bluestore-stupid.yaml supported-random-distro$/{ubuntu_latest.yaml} tasks/orchestrator_cli.yaml} 2
Failure Reason:

Test failure: test_mds_add (tasks.mgr.test_orchestrator_cli.TestOrchestratorCli)

pass 4513992 2019-11-17 01:31:25 2019-11-18 03:39:06 2019-11-18 04:11:05 0:31:59 0:23:28 0:08:31 mira master rhel 7.7 rados/perf/{ceph.yaml objectstore/bluestore-comp.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{rhel_7.yaml} workloads/radosbench_4K_rand_read.yaml} 1
fail 4513993 2019-11-17 01:31:26 2019-11-18 03:49:37 2019-11-18 03:59:36 0:09:59 0:03:04 0:06:55 mira master rhel 7.7 rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml 2-recovery-overrides/{more-active-recovery.yaml} backoff/peering.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/upmap.yaml msgr-failures/osd-delay.yaml msgr/async-v1only.yaml objectstore/filestore-xfs.yaml rados.yaml supported-random-distro$/{rhel_7.yaml} thrashers/none.yaml thrashosds-health.yaml workloads/redirect.yaml} 2
Failure Reason:

d': True, u'stderr': u'http://dl.fedoraproject.org/pub/epel/7/x86_64/repodata/repomd.xml: [Errno 14] curl#18 - "transfer closed with 1 bytes remaining to read"\nTrying other mirror.\n\n\n One of the configured repositories failed (Extra Packages for Enterprise Linux),\n and yum doesn\'t have enough cached data to continue. At this point the only\n safe thing yum can do is fail. There are a few ways to work "fix" this:\n\n 1. Contact the upstream for the repository and get them to fix the problem.\n\n 2. Reconfigure the baseurl/etc. for the repository, to point to a working\n upstream. This is most often useful if you are using a newer\n distribution release than is supported by the repository (and the\n packages for the previous distribution release still work).\n\n 3. Run the command with the repository temporarily disabled\n yum --disablerepo=epel ...\n\n 4. Disable the repository permanently, so yum won\'t use it by default. Yum\n will then just ignore the repository until you permanently enable it\n again or use --enablerepo for temporary usage:\n\n yum-config-manager --disable epel\n or\n subscription-manager repos --disable=epel\n\n 5. Configure the failing repository to be skipped, if it is unavailable.\n Note that yum will try to contact the repo. when it runs most commands,\n so will have to try and fail each time (and thus. yum will be be much\n slower). If it is a very temporary problem though, this is often a nice\n compromise:\n\n yum-config-manager --save --setopt=epel.skip_if_unavailable=true\n\nfailure: repodata/repomd.xml from epel: [Errno 256] No more mirrors to try.\nhttp://dl.fedoraproject.org/pub/epel/7/x86_64/repodata/repomd.xml: [Errno 14] curl#18 - "transfer closed with 1 bytes remaining to read"', u'stdout': u'Loaded plugins: fastestmirror, langpacks, priorities, product-id, search-\n : disabled-repos, subscription-manager\nCleaning repos: epel lab-extras rhel-7-fcgi-ceph rhel-7-server-extras-rpms\n : rhel-7-server-optional-rpms rhel-7-server-rpms\n : rhel-ha-for-rhel-7-server-rpms\nLoaded plugins: fastestmirror, langpacks, priorities, product-id, search-\n : disabled-repos, subscription-manager\nDetermining fastest mirrors', u'msg': u'non-zero return code', u'delta': u'0:00:02.690009', 'stdout_lines': [u'Loaded plugins: fastestmirror, langpacks, priorities, product-id, search-', u' : disabled-repos, subscription-manager', u'Cleaning repos: epel lab-extras rhel-7-fcgi-ceph rhel-7-server-extras-rpms', u' : rhel-7-server-optional-rpms rhel-7-server-rpms', u' : rhel-ha-for-rhel-7-server-rpms', u'Loaded plugins: fastestmirror, langpacks, priorities, product-id, search-', u' : disabled-repos, subscription-manager', u'Determining fastest mirrors'], u'end': u'2019-11-18 03:58:12.159671', '_ansible_no_log': False, u'cmd': u'rm -rf /var/cache/yum/*; yum clean all; yum makecache', u'start': u'2019-11-18 03:58:09.469662', u'warnings': [u"Consider using the file module with state=absent rather than running 'rm'. If you need to use command because file is insufficient you can add 'warn: false' to this command task or set 'command_warnings=False' in ansible.cfg to get rid of this message."], u'rc': 1, u'invocation': {u'module_args': {u'warn': True, u'executable': None, u'_uses_shell': True, u'strip_empty_ends': True, u'_raw_params': u'rm -rf /var/cache/yum/*; yum clean all; yum makecache', u'removes': None, u'argv': None, u'creates': None, u'chdir': None, u'stdin_add_newline': True, u'stdin': None}}}}Traceback (most recent call last): File "/home/teuthworker/src/git.ceph.com_git_ceph-cm-ansible_master/callback_plugins/failure_log.py", line 44, in log_failure log.error(yaml.safe_dump(failure)) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/__init__.py", line 309, in safe_dump return dump_all([data], stream, Dumper=SafeDumper, **kwds) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/__init__.py", line 281, in dump_all dumper.represent(data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 29, in represent node = self.represent_data(data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 58, in represent_data node = self.yaml_representers[data_types[0]](self, data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 227, in represent_dict return self.represent_mapping(u'tag:yaml.org,2002:map', data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 125, in represent_mapping node_value = self.represent_data(item_value) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 58, in represent_data node = self.yaml_representers[data_types[0]](self, data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 227, in represent_dict return self.represent_mapping(u'tag:yaml.org,2002:map', data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 125, in represent_mapping node_value = self.represent_data(item_value) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 68, in represent_data node = self.yaml_representers[None](self, data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 251, in represent_undefined raise RepresenterError("cannot represent an object", data)RepresenterError: ('cannot represent an object', u'rm -rf /var/cache/yum/*; yum clean all; yum makecache')Failure object was: {'mira066.front.sepia.ceph.com': {'stderr_lines': [u'http://dl.fedoraproject.org/pub/epel/7/x86_64/repodata/35487e8c13fe4d62fd897b5e8610b8de5436779d0800441eb38a7919f96ded32-filelists.sqlite.bz2: [Errno 14] curl#18 - "transfer closed with 1758586 bytes remaining to read"', u'Trying other mirror.', u'http://dl.fedoraproject.org/pub/epel/7/x86_64/repodata/78364192654f81f335a53ad2e150be332036a37b75ee154c010f36692077f4bc-other.sqlite.bz2: [Errno 14] HTTP Error 404 - Not Found', u'Trying other mirror.', u'To address this issue please refer to the below knowledge base article ', u'', u'https://access.redhat.com/articles/1320623', u'', u"If above article doesn't help to resolve this issue please open a ticket with Red Hat Support.", u'', u'http://dl.fedoraproject.org/pub/epel/7/x86_64/repodata/7ef61dc594674fe902020c298a6e8a943b6e15ced7768d0f8bd82047964fb22e-updateinfo.xml.zck: [Errno 14] HTTP Error 404 - Not Found', u'Trying other mirror.', u'http://dl.fedoraproject.org/pub/epel/7/x86_64/repodata/35487e8c13fe4d62fd897b5e8610b8de5436779d0800441eb38a7919f96ded32-filelists.sqlite.bz2: [Errno 14] HTTP Error 404 - Not Found', u'Trying other mirror.', u'http://dl.fedoraproject.org/pub/epel/7/x86_64/repodata/78364192654f81f335a53ad2e150be332036a37b75ee154c010f36692077f4bc-other.sqlite.bz2: [Errno 14] HTTP Error 404 - Not Found', u'Trying other mirror.', u'http://dl.fedoraproject.org/pub/epel/7/x86_64/repodata/35487e8c13fe4d62fd897b5e8610b8de5436779d0800441eb38a7919f96ded32-filelists.sqlite.bz2: [Errno 14] HTTP Error 404 - Not Found', u'Trying other mirror.', u'', u'', u' One of the configured repositories failed (Extra Packages for Enterprise Linux),', u" and yum doesn't have enough cached data to continue. At this point the only", u' safe thing yum can do is fail. There are a few ways to work "fix" this:', u'', u' 1. Contact the upstream for the repository and get them to fix the problem.', u'', u' 2. Reconfigure the baseurl/etc. for the repository, to point to a working', u' upstream. This is most often useful if you are using a newer', u' distribution release than is supported by the repository (and the', u' packages for the previous distribution release still work).', u'', u' 3. Run the command with the repository temporarily disabled', u' yum --disablerepo=epel ...', u'', u" 4. Disable the repository permanently, so yum won't use it by default. Yum", u' will then just ignore the repository until you permanently enable it', u' again or use --enablerepo for temporary usage:', u'', u' yum-config-manager --disable epel', u' or', u' subscription-manager repos --disable=epel', u'', u' 5. Configure the failing repository to be skipped, if it is unavailable.', u' Note that yum will try to contact the repo. when it runs most commands,', u' so will have to try and fail each time (and thus. yum will be be much', u' slower). If it is a very temporary problem though, this is often a nice', u' compromise:', u'', u' yum-config-manager --save --setopt=epel.skip_if_unavailable=true', u'', u'failure: repodata/35487e8c13fe4d62fd897b5e8610b8de5436779d0800441eb38a7919f96ded32-filelists.sqlite.bz2 from epel: [Errno 256] No more mirrors to try.', u'http://dl.fedoraproject.org/pub/epel/7/x86_64/repodata/35487e8c13fe4d62fd897b5e8610b8de5436779d0800441eb38a7919f96ded32-filelists.sqlite.bz2: [Errno 14] HTTP Error 404 - Not Found'], u'changed': True, u'stderr': u'http://dl.fedoraproject.org/pub/epel/7/x86_64/repodata/35487e8c13fe4d62fd897b5e8610b8de5436779d0800441eb38a7919f96ded32-filelists.sqlite.bz2: [Errno 14] curl#18 - "transfer closed with 1758586 bytes remaining to read"\nTrying other mirror.\nhttp://dl.fedoraproject.org/pub/epel/7/x86_64/repodata/78364192654f81f335a53ad2e150be332036a37b75ee154c010f36692077f4bc-other.sqlite.bz2: [Errno 14] HTTP Error 404 - Not Found\nTrying other mirror.\nTo address this issue please refer to the below knowledge base article \n\nhttps://access.redhat.com/articles/1320623\n\nIf above article doesn\'t help to resolve this issue please open a ticket with Red Hat Support.\n\nhttp://dl.fedoraproject.org/pub/epel/7/x86_64/repodata/7ef61dc594674fe902020c298a6e8a943b6e15ced7768d0f8bd82047964fb22e-updateinfo.xml.zck: [Errno 14] HTTP Error 404 - Not Found\nTrying other mirror.\nhttp://dl.fedoraproject.org/pub/epel/7/x86_64/repodata/35487e8c13fe4d62fd897b5e8610b8de5436779d0800441eb38a7919f96ded32-filelists.sqlite.bz2: [Errno 14] HTTP Error 404 - Not Found\nTrying other mirror.\nhttp://dl.fedoraproject.org/pub/epel/7/x86_64/repodata/78364192654f81f335a53ad2e150be332036a37b75ee154c010f36692077f4bc-other.sqlite.bz2: [Errno 14] HTTP Error 404 - Not Found\nTrying other mirror.\nhttp://dl.fedoraproject.org/pub/epel/7/x86_64/repodata/35487e8c13fe4d62fd897b5e8610b8de5436779d0800441eb38a7919f96ded32-filelists.sqlite.bz2: [Errno 14] HTTP Error 404 - Not Found\nTrying other mirror.\n\n\n One of the configured repositories failed (Extra Packages for Enterprise Linux),\n and yum doesn\'t have enough cached data to continue. At this point the only\n safe thing yum can do is fail. There are a few ways to work "fix" this:\n\n 1. Contact the upstream for the repository and get them to fix the problem.\n\n 2. Reconfigure the baseurl/etc. for the repository, to point to a working\n upstream. This is most often useful if you are using a newer\n distribution release than is supported by the repository (and the\n packages for the previous distribution release still work).\n\n 3. Run the command with the repository temporarily disabled\n yum --disablerepo=epel ...\n\n 4. Disable the repository permanently, so yum won\'t use it by default. Yum\n will then just ignore the repository until you permanently enable it\n again or use --enablerepo for temporary usage:\n\n yum-config-manager --disable epel\n or\n subscription-manager repos --disable=epel\n\n 5. Configure the failing repository to be skipped, if it is unavailable.\n Note that yum will try to contact the repo. when it runs most commands,\n so will have to try and fail each time (and thus. yum will be be much\n slower). If it is a very temporary problem though, this is often a nice\n compromise:\n\n yum-config-manager --save --setopt=epel.skip_if_unavailable=true\n\nfailure: repodata/35487e8c13fe4d62fd897b5e8610b8de5436779d0800441eb38a7919f96ded32-filelists.sqlite.bz2 from epel: [Errno 256] No more mirrors to try.\nhttp://dl.fedoraproject.org/pub/epel/7/x86_64/repodata/35487e8c13fe4d62fd897b5e8610b8de5436779d0800441eb38a7919f96ded32-filelists.sqlite.bz2: [Errno 14] HTTP Error 404 - Not Found', u'stdout': u'Loaded plugins: fastestmirror, langpacks, priorities, product-id, search-\n : disabled-repos, subscription-manager\nCleaning repos: epel lab-extras rhel-7-fcgi-ceph rhel-7-server-extras-rpms\n : rhel-7-server-optional-rpms rhel-7-server-rpms\n : rhel-ha-for-rhel-7-server-rpms\nLoaded plugins: fastestmirror, langpacks, priorities, product-id, search-\n : disabled-repos, subscription-manager\nDetermining fastest mirrors', u'msg': u'non-zero return code', u'delta': u'0:00:19.995013', 'stdout_lines': [u'Loaded plugins: fastestmirror, langpacks, priorities, product-id, search-', u' : disabled-repos, subscription-manager', u'Cleaning repos: epel lab-extras rhel-7-fcgi-ceph rhel-7-server-extras-rpms', u' : rhel-7-server-optional-rpms rhel-7-server-rpms', u' : rhel-ha-for-rhel-7-server-rpms', u'Loaded plugins: fastestmirror, langpacks, priorities, product-id, search-', u' : disabled-repos, subscription-manager', u'Determining fastest mirrors'], u'end': u'2019-11-18 03:58:24.378963', '_ansible_no_log': False, u'cmd': u'rm -rf /var/cache/yum/*; yum clean all; yum makecache', u'start': u'2019-11-18 03:58:04.383950', u'warnings': [u"Consider using the file module with state=absent rather than running 'rm'. If you need to use command because file is insufficient you can add 'warn: false' to this command task or set 'command_warnings=False' in ansible.cfg to get rid of this message."], u'rc': 1, u'invocation': {u'module_args': {u'warn': True, u'executable': None, u'_uses_shell': True, u'strip_empty_ends': True, u'_raw_params': u'rm -rf /var/cache/yum/*; yum clean all; yum makecache', u'removes': None, u'argv': None, u'creates': None, u'chdir': None, u'stdin_add_newline': True, u'stdin': None}}}}Traceback (most recent call last): File "/home/teuthworker/src/git.ceph.com_git_ceph-cm-ansible_master/callback_plugins/failure_log.py", line 44, in log_failure log.error(yaml.safe_dump(failure)) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/__init__.py", line 309, in safe_dump return dump_all([data], stream, Dumper=SafeDumper, **kwds) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/__init__.py", line 281, in dump_all dumper.represent(data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 29, in represent node = self.represent_data(data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 58, in represent_data node = self.yaml_representers[data_types[0]](self, data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 227, in represent_dict return self.represent_mapping(u'tag:yaml.org,2002:map', data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 125, in represent_mapping node_value = self.represent_data(item_value) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 58, in represent_data node = self.yaml_representers[data_types[0]](self, data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 227, in represent_dict return self.represent_mapping(u'tag:yaml.org,2002:map', data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 125, in represent_mapping node_value = self.represent_data(item_value) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 68, in represent_data node = self.yaml_representers[None](self, data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 251, in represent_undefined raise RepresenterError("cannot represent an object", data)RepresenterError: ('cannot represent an object', u'rm -rf /var/cache/yum/*; yum clean all; yum makecache')

pass 4513994 2019-11-17 01:31:27 2019-11-18 03:59:39 2019-11-18 04:29:38 0:29:59 0:23:26 0:06:33 mira master rhel 7.7 rados/singleton/{all/max-pg-per-osd.from-primary.yaml msgr-failures/many.yaml msgr/async.yaml objectstore/bluestore-stupid.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}} 1
pass 4513995 2019-11-17 01:31:27 2019-11-18 04:11:08 2019-11-18 05:09:08 0:58:00 0:32:49 0:25:11 mira master centos 7.6 rados/thrash-old-clients/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-install/mimic.yaml backoff/normal.yaml ceph.yaml clusters/{openstack.yaml three-plus-one.yaml} d-balancer/off.yaml distro$/{centos_latest.yaml} msgr-failures/few.yaml rados.yaml thrashers/pggrow.yaml thrashosds-health.yaml workloads/cache-snaps.yaml} 4
pass 4513996 2019-11-17 01:31:28 2019-11-18 04:12:39 2019-11-18 05:24:39 1:12:00 1:02:57 0:09:03 mira master rhel 7.7 rados/dashboard/{clusters/{2-node-mgr.yaml} debug/mgr.yaml objectstore/bluestore-comp.yaml supported-random-distro$/{rhel_7.yaml} tasks/dashboard.yaml} 2
pass 4513997 2019-11-17 01:31:29 2019-11-18 04:24:52 2019-11-18 05:02:52 0:38:00 0:24:03 0:13:57 mira master centos 7.6 rados/objectstore/{backends/filestore-idempotent-aio-journal.yaml supported-random-distro$/{centos_7.yaml}} 1
pass 4513998 2019-11-17 01:31:30 2019-11-18 04:29:40 2019-11-18 04:47:39 0:17:59 0:07:23 0:10:36 mira master ubuntu 18.04 rados/singleton-nomsgr/{all/export-after-evict.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}} 1
pass 4513999 2019-11-17 01:31:31 2019-11-18 04:47:42 2019-11-18 07:51:46 3:04:04 2:45:11 0:18:53 mira master rhel 7.7 rados/thrash-erasure-code-overwrites/{bluestore-bitmap.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} fast/normal.yaml msgr-failures/fastclose.yaml rados.yaml recovery-overrides/{more-async-partial-recovery.yaml} supported-random-distro$/{rhel_7.yaml} thrashers/minsize_recovery.yaml thrashosds-health.yaml workloads/ec-snaps-few-objects-overwrites.yaml} 2
pass 4514000 2019-11-17 01:31:31 2019-11-18 05:03:11 2019-11-18 05:51:11 0:48:00 0:20:18 0:27:42 mira master centos 7.6 rados/thrash-erasure-code-shec/{ceph.yaml clusters/{fixed-4.yaml openstack.yaml} msgr-failures/fastclose.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml recovery-overrides/{more-active-recovery.yaml} supported-random-distro$/{centos_7.yaml} thrashers/default.yaml thrashosds-health.yaml workloads/ec-rados-plugin=shec-k=4-m=3-c=2.yaml} 4
pass 4514001 2019-11-17 01:31:32 2019-11-18 05:09:14 2019-11-18 05:41:13 0:31:59 0:19:44 0:12:15 mira master centos 7.6 rados/singleton/{all/max-pg-per-osd.from-replica.yaml msgr-failures/few.yaml msgr/async-v1only.yaml objectstore/filestore-xfs.yaml rados.yaml supported-random-distro$/{centos_7.yaml}} 1
pass 4514002 2019-11-17 01:31:33 2019-11-18 05:24:45 2019-11-18 05:52:45 0:28:00 0:18:09 0:09:51 mira master ubuntu 18.04 rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml 2-recovery-overrides/{more-active-recovery.yaml} backoff/peering_and_degraded.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/crush-compat.yaml msgr-failures/fastclose.yaml msgr/async-v2only.yaml objectstore/bluestore-avl.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} thrashers/pggrow.yaml thrashosds-health.yaml workloads/redirect_promote_tests.yaml} 2
pass 4514003 2019-11-17 01:31:34 2019-11-18 05:41:15 2019-11-18 06:11:15 0:30:00 0:22:52 0:07:08 mira master rhel 7.7 rados/perf/{ceph.yaml objectstore/bluestore-low-osd-mem-target.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{rhel_7.yaml} workloads/radosbench_4K_seq_read.yaml} 1
pass 4514004 2019-11-17 01:31:35 2019-11-18 05:51:15 2019-11-18 06:25:14 0:33:59 0:21:14 0:12:45 mira master centos 7.6 rados/basic/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/few.yaml msgr/async-v1only.yaml objectstore/filestore-xfs.yaml rados.yaml supported-random-distro$/{centos_7.yaml} tasks/readwrite.yaml} 2
pass 4514005 2019-11-17 01:31:35 2019-11-18 05:52:47 2019-11-18 06:12:46 0:19:59 0:09:52 0:10:07 mira master ubuntu 18.04 rados/singleton/{all/mon-auth-caps.yaml msgr-failures/many.yaml msgr/async-v2only.yaml objectstore/bluestore-avl.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}} 1
fail 4514006 2019-11-17 01:31:36 2019-11-18 06:11:18 2019-11-18 08:43:20 2:32:02 2:13:24 0:18:38 mira master rhel 7.7 rados/multimon/{clusters/3.yaml msgr-failures/few.yaml msgr/async.yaml no_pools.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{rhel_7.yaml} tasks/mon_clock_no_skews.yaml} 2
Failure Reason:

'strip_empty_ends': True, u'_raw_params': u'sgdisk --zap-all /dev/sdd || sgdisk --zap-all /dev/sdd', u'removes': None, u'argv': None, u'creates': None, u'chdir': None, u'stdin_add_newline': True, u'stdin': None}}, u'start': u'2019-11-18 08:41:41.880138'}, {'stderr_lines': [], u'changed': True, u'stdout': u'Creating new GPT entries.\nGPT data structures destroyed! You may now partition the disk using fdisk or\nother utilities.', u'delta': u'0:00:01.011105', 'stdout_lines': [u'Creating new GPT entries.', u'GPT data structures destroyed! You may now partition the disk using fdisk or', u'other utilities.'], '_ansible_item_label': {'value': {u'sectorsize': u'512', u'vendor': u'Hitachi', u'links': {u'masters': [], u'labels': [], u'ids': [u'scsi-2001b4d2011a85600'], u'uuids': [u'c7985657-1ca5-421c-8e0b-6fb67d08df1c']}, u'sas_device_handle': None, u'host': u'RAID bus controller: Areca Technology Corp. ARC-1680 series PCIe to SAS/SATA 3Gb RAID Controller', u'support_discard': u'0', u'serial': u'JPW9K0N211XZVE', u'holders': [], u'size': u'931.51 GB', u'scheduler_mode': u'deadline', u'rotational': u'1', u'sectors': u'1953525168', u'sas_address': None, u'virtual': 1, u'removable': u'0', u'model': u'HUA722010CLA330', u'partitions': {}}, 'key': u'sde'}, 'ansible_loop_var': u'item', u'end': u'2019-11-18 08:41:44.684185', '_ansible_no_log': False, 'item': {'value': {u'sectorsize': u'512', u'vendor': u'Hitachi', u'links': {u'masters': [], u'labels': [], u'ids': [u'scsi-2001b4d2011a85600'], u'uuids': [u'c7985657-1ca5-421c-8e0b-6fb67d08df1c']}, u'sas_device_handle': None, u'host': u'RAID bus controller: Areca Technology Corp. ARC-1680 series PCIe to SAS/SATA 3Gb RAID Controller', u'support_discard': u'0', u'serial': u'JPW9K0N211XZVE', u'holders': [], u'size': u'931.51 GB', u'scheduler_mode': u'deadline', u'rotational': u'1', u'sectors': u'1953525168', u'sas_address': None, u'virtual': 1, u'removable': u'0', u'model': u'HUA722010CLA330', u'partitions': {}}, 'key': u'sde'}, u'cmd': u'sgdisk --zap-all /dev/sde || sgdisk --zap-all /dev/sde', 'failed': False, u'stderr': u'', u'rc': 0, u'invocation': {u'module_args': {u'warn': True, u'executable': None, u'_uses_shell': True, u'strip_empty_ends': True, u'_raw_params': u'sgdisk --zap-all /dev/sde || sgdisk --zap-all /dev/sde', u'removes': None, u'argv': None, u'creates': None, u'chdir': None, u'stdin_add_newline': True, u'stdin': None}}, u'start': u'2019-11-18 08:41:43.673080'}, {'stderr_lines': [], u'changed': True, u'stdout': u'Creating new GPT entries.\nGPT data structures destroyed! You may now partition the disk using fdisk or\nother utilities.', u'delta': u'0:00:01.011287', 'stdout_lines': [u'Creating new GPT entries.', u'GPT data structures destroyed! You may now partition the disk using fdisk or', u'other utilities.'], '_ansible_item_label': {'value': {u'sectorsize': u'512', u'vendor': u'Hitachi', u'links': {u'masters': [], u'labels': [], u'ids': [u'scsi-2001b4d2040775100'], u'uuids': [u'585c91be-4a56-4b5f-b892-3c9595219f7c']}, u'sas_device_handle': None, u'host': u'RAID bus controller: Areca Technology Corp. ARC-1680 series PCIe to SAS/SATA 3Gb RAID Controller', u'support_discard': u'0', u'serial': u'JPW9K0N204WG1E', u'holders': [], u'size': u'931.51 GB', u'scheduler_mode': u'deadline', u'rotational': u'1', u'sectors': u'1953525168', u'sas_address': None, u'virtual': 1, u'removable': u'0', u'model': u'HUA722010CLA330', u'partitions': {}}, 'key': u'sdf'}, 'ansible_loop_var': u'item', u'end': u'2019-11-18 08:41:45.946306', '_ansible_no_log': False, 'item': {'value': {u'sectorsize': u'512', u'vendor': u'Hitachi', u'links': {u'masters': [], u'labels': [], u'ids': [u'scsi-2001b4d2040775100'], u'uuids': [u'585c91be-4a56-4b5f-b892-3c9595219f7c']}, u'sas_device_handle': None, u'host': u'RAID bus controller: Areca Technology Corp. ARC-1680 series PCIe to SAS/SATA 3Gb RAID Controller', u'support_discard': u'0', u'serial': u'JPW9K0N204WG1E', u'holders': [], u'size': u'931.51 GB', u'scheduler_mode': u'deadline', u'rotational': u'1', u'sectors': u'1953525168', u'sas_address': None, u'virtual': 1, u'removable': u'0', u'model': u'HUA722010CLA330', u'partitions': {}}, 'key': u'sdf'}, u'cmd': u'sgdisk --zap-all /dev/sdf || sgdisk --zap-all /dev/sdf', 'failed': False, u'stderr': u'', u'rc': 0, u'invocation': {u'module_args': {u'warn': True, u'executable': None, u'_uses_shell': True, u'strip_empty_ends': True, u'_raw_params': u'sgdisk --zap-all /dev/sdf || sgdisk --zap-all /dev/sdf', u'removes': None, u'argv': None, u'creates': None, u'chdir': None, u'stdin_add_newline': True, u'stdin': None}}, u'start': u'2019-11-18 08:41:44.935019'}, {'ansible_loop_var': u'item', '_ansible_no_log': False, 'skip_reason': u'Conditional result was False', 'item': {'value': {u'sectorsize': u'512', u'vendor': u'Seagate', u'links': {u'masters': [], u'labels': [], u'ids': [u'scsi-2001b4d2000000000'], u'uuids': []}, u'sas_device_handle': None, u'host': u'RAID bus controller: Areca Technology Corp. ARC-1680 series PCIe to SAS/SATA 3Gb RAID Controller', u'support_discard': u'0', u'serial': u'5VP66QW9', u'holders': [], u'size': u'931.51 GB', u'scheduler_mode': u'deadline', u'rotational': u'1', u'sectors': u'1953525168', u'sas_address': None, u'virtual': 1, u'removable': u'0', u'model': u'ST31000524AS', u'partitions': {u'sda1': {u'start': u'2048', u'sectorsize': 512, u'uuid': u'f1bca609-9cd3-46ed-8f34-af0538e7246e', u'sectors': u'1953522688', u'holders': [], u'links': {u'masters': [], u'labels': [], u'ids': [u'scsi-2001b4d2000000000-part1'], u'uuids': [u'f1bca609-9cd3-46ed-8f34-af0538e7246e']}, u'size': u'931.51 GB'}}}, 'key': u'sda'}, 'skipped': True, 'changed': False, '_ansible_item_label': {'value': {u'sectorsize': u'512', u'vendor': u'Seagate', u'links': {u'masters': [], u'labels': [], u'ids': [u'scsi-2001b4d2000000000'], u'uuids': []}, u'sas_device_handle': None, u'host': u'RAID bus controller: Areca Technology Corp. ARC-1680 series PCIe to SAS/SATA 3Gb RAID Controller', u'support_discard': u'0', u'serial': u'5VP66QW9', u'holders': [], u'size': u'931.51 GB', u'scheduler_mode': u'deadline', u'rotational': u'1', u'sectors': u'1953525168', u'sas_address': None, u'virtual': 1, u'removable': u'0', u'model': u'ST31000524AS', u'partitions': {u'sda1': {u'start': u'2048', u'sectorsize': 512, u'uuid': u'f1bca609-9cd3-46ed-8f34-af0538e7246e', u'sectors': u'1953522688', u'holders': [], u'links': {u'masters': [], u'labels': [], u'ids': [u'scsi-2001b4d2000000000-part1'], u'uuids': [u'f1bca609-9cd3-46ed-8f34-af0538e7246e']}, u'size': u'931.51 GB'}}}, 'key': u'sda'}}, {'stderr_lines': [], u'changed': True, u'stdout': u'Creating new GPT entries.\nGPT data structures destroyed! You may now partition the disk using fdisk or\nother utilities.', u'delta': u'0:00:01.026288', 'stdout_lines': [u'Creating new GPT entries.', u'GPT data structures destroyed! You may now partition the disk using fdisk or', u'other utilities.'], '_ansible_item_label': {'value': {u'sectorsize': u'512', u'vendor': u'Seagate', u'links': {u'masters': [u'dm-0'], u'labels': [], u'ids': [], u'uuids': []}, u'sas_device_handle': None, u'host': u'RAID bus controller: Areca Technology Corp. ARC-1680 series PCIe to SAS/SATA 3Gb RAID Controller', u'support_discard': u'0', u'serial': u'5VP8NWLD', u'holders': [u'mpatha'], u'size': u'931.51 GB', u'scheduler_mode': u'deadline', u'rotational': u'1', u'sectors': u'1953525168', u'sas_address': None, u'virtual': 1, u'removable': u'0', u'model': u'ST31000528AS', u'partitions': {}}, 'key': u'sdb'}, 'ansible_loop_var': u'item', u'end': u'2019-11-18 08:41:47.226976', '_ansible_no_log': False, 'item': {'value': {u'sectorsize': u'512', u'vendor': u'Seagate', u'links': {u'masters': [u'dm-0'], u'labels': [], u'ids': [], u'uuids': []}, u'sas_device_handle': None, u'host': u'RAID bus controller: Areca Technology Corp. ARC-1680 series PCIe to SAS/SATA 3Gb RAID Controller', u'support_discard': u'0', u'serial': u'5VP8NWLD', u'holders': [u'mpatha'], u'size': u'931.51 GB', u'scheduler_mode': u'deadline', u'rotational': u'1', u'sectors': u'1953525168', u'sas_address': None, u'virtual': 1, u'removable': u'0', u'model': u'ST31000528AS', u'partitions': {}}, 'key': u'sdb'}, u'cmd': u'sgdisk --zap-all /dev/sdb || sgdisk --zap-all /dev/sdb', 'failed': False, u'stderr': u'', u'rc': 0, u'invocation': {u'module_args': {u'warn': True, u'executable': None, u'_uses_shell': True, u'strip_empty_ends': True, u'_raw_params': u'sgdisk --zap-all /dev/sdb || sgdisk --zap-all /dev/sdb', u'removes': None, u'argv': None, u'creates': None, u'chdir': None, u'stdin_add_newline': True, u'stdin': None}}, u'start': u'2019-11-18 08:41:46.200688'}, {'stderr_lines': [], u'changed': True, u'stdout': u'Creating new GPT entries.\nGPT data structures destroyed! You may now partition the disk using fdisk or\nother utilities.', u'delta': u'0:00:01.023082', 'stdout_lines': [u'Creating new GPT entries.', u'GPT data structures destroyed! You may now partition the disk using fdisk or', u'other utilities.'], '_ansible_item_label': {'value': {u'sectorsize': u'512', u'vendor': u'Seagate', u'links': {u'masters': [u'dm-0'], u'labels': [], u'ids': [], u'uuids': []}, u'sas_device_handle': None, u'host': u'RAID bus controller: Areca Technology Corp. ARC-1680 series PCIe to SAS/SATA 3Gb RAID Controller', u'support_discard': u'0', u'serial': u'6VPBDH90', u'holders': [u'mpatha'], u'size': u'931.51 GB', u'scheduler_mode': u'deadline', u'rotational': u'1', u'sectors': u'1953525168', u'sas_address': None, u'virtual': 1, u'removable': u'0', u'model': u'ST31000528AS', u'partitions': {}}, 'key': u'sdc'}, 'ansible_loop_var': u'item', u'end': u'2019-11-18 08:41:48.498868', '_ansible_no_log': False, 'item': {'value': {u'sectorsize': u'512', u'vendor': u'Seagate', u'links': {u'masters': [u'dm-0'], u'labels': [], u'ids': [], u'uuids': []}, u'sas_device_handle': None, u'host': u'RAID bus controller: Areca Technology Corp. ARC-1680 series PCIe to SAS/SATA 3Gb RAID Controller', u'support_discard': u'0', u'serial': u'6VPBDH90', u'holders': [u'mpatha'], u'size': u'931.51 GB', u'scheduler_mode': u'deadline', u'rotational': u'1', u'sectors': u'1953525168', u'sas_address': None, u'virtual': 1, u'removable': u'0', u'model': u'ST31000528AS', u'partitions': {}}, 'key': u'sdc'}, u'cmd': u'sgdisk --zap-all /dev/sdc || sgdisk --zap-all /dev/sdc', 'failed': False, u'stderr': u'', u'rc': 0, u'invocation': {u'module_args': {u'warn': True, u'executable': None, u'_uses_shell': True, u'strip_empty_ends': True, u'_raw_params': u'sgdisk --zap-all /dev/sdc || sgdisk --zap-all /dev/sdc', u'removes': None, u'argv': None, u'creates': None, u'chdir': None, u'stdin_add_newline': True, u'stdin': None}}, u'start': u'2019-11-18 08:41:47.475786'}, {'stderr_lines': [u'Problem opening /dev/dm-0 for reading! Error is 2.', u'The specified file does not exist!', u"Problem opening '' for writing! Program will now terminate.", u'Warning! MBR not overwritten! Error is 2!', u'Problem opening /dev/dm-0 for reading! Error is 2.', u'The specified file does not exist!', u"Problem opening '' for writing! Program will now terminate.", u'Warning! MBR not overwritten! Error is 2!'], u'changed': True, u'stdout': u'', u'invocation': {u'module_args': {u'warn': True, u'executable': None, u'_uses_shell': True, u'strip_empty_ends': True, u'_raw_params': u'sgdisk --zap-all /dev/dm-0 || sgdisk --zap-all /dev/dm-0', u'removes': None, u'argv': None, u'creates': None, u'chdir': None, u'stdin_add_newline': True, u'stdin': None}}, u'delta': u'0:00:00.008550', 'stdout_lines': [], '_ansible_item_label': {'value': {u'sectorsize': u'512', u'vendor': None, u'links': {u'masters': [], u'labels': [], u'ids': [u'dm-name-mpatha', u'dm-uuid-mpath-2001b4d2000000000'], u'uuids': []}, u'sas_device_handle': None, u'host': u'', u'support_discard': u'0', u'serial': u'5VP8NWLD', u'holders': [], u'size': u'931.51 GB', u'scheduler_mode': u'deadline', u'rotational': u'1', u'sectors': u'1953525168', u'sas_address': None, u'virtual': 1, u'removable': u'0', u'model': None, u'partitions': {}}, 'key': u'dm-0'}, 'ansible_loop_var': u'item', u'end': u'2019-11-18 08:41:48.766527', '_ansible_no_log': False, u'start': u'2019-11-18 08:41:48.757977', u'failed': True, u'cmd': u'sgdisk --zap-all /dev/dm-0 || sgdisk --zap-all /dev/dm-0', 'item': {'value': {u'sectorsize': u'512', u'vendor': None, u'links': {u'masters': [], u'labels': [], u'ids': [u'dm-name-mpatha', u'dm-uuid-mpath-2001b4d2000000000'], u'uuids': []}, u'sas_device_handle': None, u'host': u'', u'support_discard': u'0', u'serial': u'5VP8NWLD', u'holders': [], u'size': u'931.51 GB', u'scheduler_mode': u'deadline', u'rotational': u'1', u'sectors': u'1953525168', u'sas_address': None, u'virtual': 1, u'removable': u'0', u'model': None, u'partitions': {}}, 'key': u'dm-0'}, u'stderr': u"Problem opening /dev/dm-0 for reading! Error is 2.\nThe specified file does not exist!\nProblem opening '' for writing! Program will now terminate.\nWarning! MBR not overwritten! Error is 2!\nProblem opening /dev/dm-0 for reading! Error is 2.\nThe specified file does not exist!\nProblem opening '' for writing! Program will now terminate.\nWarning! MBR not overwritten! Error is 2!", u'rc': 2, u'msg': u'non-zero return code'}]}}Traceback (most recent call last): File "/home/teuthworker/src/git.ceph.com_git_ceph-cm-ansible_master/callback_plugins/failure_log.py", line 44, in log_failure log.error(yaml.safe_dump(failure)) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/__init__.py", line 309, in safe_dump return dump_all([data], stream, Dumper=SafeDumper, **kwds) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/__init__.py", line 281, in dump_all dumper.represent(data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 29, in represent node = self.represent_data(data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 58, in represent_data node = self.yaml_representers[data_types[0]](self, data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 227, in represent_dict return self.represent_mapping(u'tag:yaml.org,2002:map', data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 125, in represent_mapping node_value = self.represent_data(item_value) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 58, in represent_data node = self.yaml_representers[data_types[0]](self, data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 227, in represent_dict return self.represent_mapping(u'tag:yaml.org,2002:map', data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 125, in represent_mapping node_value = self.represent_data(item_value) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 58, in represent_data node = self.yaml_representers[data_types[0]](self, data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 219, in represent_list return self.represent_sequence(u'tag:yaml.org,2002:seq', data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 102, in represent_sequence node_item = self.represent_data(item) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 58, in represent_data node = self.yaml_representers[data_types[0]](self, data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 227, in represent_dict return self.represent_mapping(u'tag:yaml.org,2002:map', data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 125, in represent_mapping node_value = self.represent_data(item_value) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 58, in represent_data node = self.yaml_representers[data_types[0]](self, data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 227, in represent_dict return self.represent_mapping(u'tag:yaml.org,2002:map', data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 125, in represent_mapping node_value = self.represent_data(item_value) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 68, in represent_data node = self.yaml_representers[None](self, data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 251, in represent_undefined raise RepresenterError("cannot represent an object", data)RepresenterError: ('cannot represent an object', u'sdd')

pass 4514007 2019-11-17 01:31:37 2019-11-18 06:13:07 2019-11-18 09:15:09 3:02:02 2:43:18 0:18:44 mira master rhel 7.7 rados/monthrash/{ceph.yaml clusters/3-mons.yaml msgr-failures/mon-delay.yaml msgr/async.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{rhel_7.yaml} thrashers/sync-many.yaml workloads/rados_mon_workunits.yaml} 2
pass 4514008 2019-11-17 01:31:38 2019-11-18 06:25:30 2019-11-18 06:59:28 0:33:58 0:13:02 0:20:56 mira master centos 7.6 rados/singleton-nomsgr/{all/full-tiering.yaml rados.yaml supported-random-distro$/{centos_7.yaml}} 1
pass 4514009 2019-11-17 01:31:39 2019-11-18 06:59:32 2019-11-18 07:41:31 0:41:59 0:19:11 0:22:48 mira master centos 7.6 rados/mgr/{clusters/{2-node-mgr.yaml} debug/mgr.yaml objectstore/filestore-xfs.yaml supported-random-distro$/{centos_7.yaml} tasks/progress.yaml} 2
pass 4514010 2019-11-17 01:31:40 2019-11-18 07:41:35 2019-11-18 08:21:35 0:40:00 0:29:56 0:10:04 mira master ubuntu 18.04 rados/thrash-erasure-code-isa/{arch/x86_64.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/osd-delay.yaml objectstore/filestore-xfs.yaml rados.yaml recovery-overrides/{more-active-recovery.yaml} supported-random-distro$/{ubuntu_latest.yaml} thrashers/careful.yaml thrashosds-health.yaml workloads/ec-rados-plugin=isa-k=2-m=1.yaml} 2
pass 4514011 2019-11-17 01:31:40 2019-11-18 07:51:49 2019-11-18 08:25:48 0:33:59 0:23:50 0:10:09 mira master rhel 7.7 rados/thrash-erasure-code-big/{ceph.yaml cluster/{12-osds.yaml openstack.yaml} msgr-failures/few.yaml objectstore/bluestore-avl.yaml rados.yaml recovery-overrides/{more-partial-recovery.yaml} supported-random-distro$/{rhel_7.yaml} thrashers/fastread.yaml thrashosds-health.yaml workloads/ec-rados-plugin=lrc-k=4-m=2-l=3.yaml} 3
pass 4514012 2019-11-17 01:31:41 2019-11-18 08:21:54 2019-11-18 08:43:53 0:21:59 0:11:10 0:10:49 mira master rados/verify/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-thrash/none.yaml msgr-failures/few.yaml msgr/async-v1only.yaml objectstore/bluestore-comp.yaml rados.yaml tasks/mon_recovery.yaml validater/lockdep.yaml} 2
pass 4514013 2019-11-17 01:31:42 2019-11-18 08:25:51 2019-11-18 11:15:53 2:50:02 2:30:02 0:20:00 mira master rhel 7.7 rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml 2-recovery-overrides/{more-async-partial-recovery.yaml} backoff/normal.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/off.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{rhel_7.yaml} thrashers/careful.yaml thrashosds-health.yaml workloads/redirect_set_object.yaml} 2
pass 4514014 2019-11-17 01:31:43 2019-11-18 08:43:29 2019-11-18 09:13:27 0:29:58 0:17:47 0:12:11 mira master centos 7.6 rados/singleton/{all/mon-config-key-caps.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{centos_7.yaml}} 1
pass 4514015 2019-11-17 01:31:44 2019-11-18 08:43:54 2019-11-18 09:13:54 0:30:00 0:16:28 0:13:32 mira master centos 7.6 rados/perf/{ceph.yaml objectstore/bluestore-stupid.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{centos_7.yaml} workloads/radosbench_4M_rand_read.yaml} 1
pass 4514016 2019-11-17 01:31:45 2019-11-18 09:13:52 2019-11-18 09:53:51 0:39:59 0:22:05 0:17:54 mira master centos 7.6 rados/thrash-old-clients/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-install/nautilus-v1only.yaml backoff/peering.yaml ceph.yaml clusters/{openstack.yaml three-plus-one.yaml} d-balancer/crush-compat.yaml distro$/{centos_latest.yaml} msgr-failures/osd-delay.yaml rados.yaml thrashers/careful.yaml thrashosds-health.yaml workloads/radosbench.yaml} 4
pass 4514017 2019-11-17 01:31:45 2019-11-18 09:13:56 2019-11-18 09:53:55 0:39:59 0:33:00 0:06:59 mira master rhel 7.7 rados/thrash-erasure-code/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} fast/normal.yaml msgr-failures/few.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml recovery-overrides/{default.yaml} supported-random-distro$/{rhel_7.yaml} thrashers/careful.yaml thrashosds-health.yaml workloads/ec-small-objects-many-deletes.yaml} 2
pass 4514018 2019-11-17 01:31:46 2019-11-18 09:15:16 2019-11-18 09:39:15 0:23:59 0:13:59 0:10:00 mira master ubuntu 18.04 rados/singleton/{all/mon-config-keys.yaml msgr-failures/many.yaml msgr/async-v1only.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}} 1
fail 4514019 2019-11-17 01:31:47 2019-11-18 09:39:23 2019-11-18 10:09:23 0:30:00 0:21:54 0:08:06 mira master rhel 7.7 rados/standalone/{supported-random-distro$/{rhel_7.yaml} workloads/misc.yaml} 1
Failure Reason:

Command failed (workunit test misc/network-ping.sh) on mira082 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=70968846f95926b5f8ae4f251788ce3818969018 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/standalone/misc/network-ping.sh'

pass 4514020 2019-11-17 01:31:48 2019-11-18 09:54:10 2019-11-18 10:22:09 0:27:59 0:17:39 0:10:20 mira master ubuntu 18.04 rados/basic/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/many.yaml msgr/async-v2only.yaml objectstore/bluestore-avl.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} tasks/repair_test.yaml} 2
pass 4514021 2019-11-17 01:31:49 2019-11-18 09:54:10 2019-11-18 10:30:09 0:35:59 0:21:27 0:14:32 mira master centos 7.6 rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml 2-recovery-overrides/{more-active-recovery.yaml} backoff/peering.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/upmap.yaml msgr-failures/osd-delay.yaml msgr/async-v1only.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{centos_7.yaml} thrashers/default.yaml thrashosds-health.yaml workloads/set-chunk-promote-flush.yaml} 2
pass 4514022 2019-11-17 01:31:50 2019-11-18 10:09:54 2019-11-18 10:47:53 0:37:59 0:30:38 0:07:21 mira master rhel 7.7 rados/objectstore/{backends/filestore-idempotent.yaml supported-random-distro$/{rhel_7.yaml}} 1
pass 4514023 2019-11-17 01:31:51 2019-11-18 10:22:12 2019-11-18 13:06:14 2:44:02 2:25:26 0:18:36 mira master rhel 7.7 rados/singleton-nomsgr/{all/health-warnings.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}} 1
pass 4514024 2019-11-17 01:31:51 2019-11-18 10:30:17 2019-11-18 10:58:16 0:27:59 0:15:03 0:12:56 mira master centos 7.6 rados/singleton/{all/mon-config.yaml msgr-failures/few.yaml msgr/async-v2only.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml supported-random-distro$/{centos_7.yaml}} 1
pass 4514025 2019-11-17 01:31:52 2019-11-18 10:48:14 2019-11-18 11:08:13 0:19:59 0:10:30 0:09:29 mira master ubuntu 18.04 rados/perf/{ceph.yaml objectstore/bluestore-basic-min-osd-mem-target.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{ubuntu_latest.yaml} workloads/radosbench_4M_seq_read.yaml} 1
fail 4514026 2019-11-17 01:31:53 2019-11-18 10:58:22 2019-11-18 11:28:20 0:29:58 0:16:14 0:13:44 mira master centos 7.6 rados/mgr/{clusters/{2-node-mgr.yaml} debug/mgr.yaml objectstore/bluestore-avl.yaml supported-random-distro$/{centos_7.yaml} tasks/prometheus.yaml} 2
Failure Reason:

Test failure: test_standby (tasks.mgr.test_prometheus.TestPrometheus)

pass 4514027 2019-11-17 01:31:54 2019-11-18 11:08:17 2019-11-18 11:52:16 0:43:59 0:32:46 0:11:13 mira master ubuntu 18.04 rados/thrash-erasure-code-overwrites/{bluestore-bitmap.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} fast/fast.yaml msgr-failures/few.yaml rados.yaml recovery-overrides/{default.yaml} supported-random-distro$/{ubuntu_latest.yaml} thrashers/morepggrow.yaml thrashosds-health.yaml workloads/ec-pool-snaps-few-objects-overwrites.yaml} 2
pass 4514028 2019-11-17 01:31:55 2019-11-18 11:16:05 2019-11-18 11:42:04 0:25:59 0:13:05 0:12:54 mira master ubuntu 18.04 rados/thrash-erasure-code-shec/{ceph.yaml clusters/{fixed-4.yaml openstack.yaml} msgr-failures/few.yaml objectstore/bluestore-stupid.yaml rados.yaml recovery-overrides/{more-active-recovery.yaml} supported-random-distro$/{ubuntu_latest.yaml} thrashers/careful.yaml thrashosds-health.yaml workloads/ec-rados-plugin=shec-k=4-m=3-c=2.yaml} 4
pass 4514029 2019-11-17 01:31:56 2019-11-18 11:28:35 2019-11-18 12:48:35 1:20:00 1:09:44 0:10:16 mira master centos 7.6 rados/singleton/{all/osd-backfill.yaml msgr-failures/many.yaml msgr/async.yaml objectstore/bluestore-stupid.yaml rados.yaml supported-random-distro$/{centos_7.yaml}} 1
pass 4514030 2019-11-17 01:31:57 2019-11-18 11:42:24 2019-11-18 12:30:23 0:47:59 0:26:07 0:21:52 mira master centos 7.6 rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml 2-recovery-overrides/{more-partial-recovery.yaml} backoff/peering_and_degraded.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/crush-compat.yaml msgr-failures/fastclose.yaml msgr/async-v2only.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml supported-random-distro$/{centos_7.yaml} thrashers/mapgap.yaml thrashosds-health.yaml workloads/set-chunks-read.yaml} 2
pass 4514031 2019-11-17 01:31:57 2019-11-18 11:52:21 2019-11-18 12:40:20 0:47:59 0:20:50 0:27:09 mira master centos 7.6 rados/thrash-old-clients/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-install/nautilus-v2only.yaml backoff/peering_and_degraded.yaml ceph.yaml clusters/{openstack.yaml three-plus-one.yaml} d-balancer/off.yaml distro$/{centos_latest.yaml} msgr-failures/fastclose.yaml rados.yaml thrashers/default.yaml thrashosds-health.yaml workloads/rbd_cls.yaml} 4
pass 4514032 2019-11-17 01:31:58 2019-11-18 12:30:45 2019-11-18 13:52:46 1:22:01 1:08:15 0:13:46 mira master centos 7.6 rados/dashboard/{clusters/{2-node-mgr.yaml} debug/mgr.yaml objectstore/bluestore-low-osd-mem-target.yaml supported-random-distro$/{centos_7.yaml} tasks/dashboard.yaml} 2
fail 4514033 2019-11-17 01:31:59 2019-11-18 12:40:44 2019-11-18 13:44:43 1:03:59 0:43:46 0:20:13 mira master centos 7.6 rados/singleton-bluestore/{all/cephtool.yaml msgr-failures/many.yaml msgr/async-v2only.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{centos_7.yaml}} 1
Failure Reason:

Command failed (workunit test cephtool/test.sh) on mira072 with status 13: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=70968846f95926b5f8ae4f251788ce3818969018 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/cephtool/test.sh'

pass 4514034 2019-11-17 01:32:00 2019-11-18 12:48:51 2019-11-18 13:12:50 0:23:59 0:14:03 0:09:56 mira master centos 7.6 rados/singleton-nomsgr/{all/large-omap-object-warnings.yaml rados.yaml supported-random-distro$/{centos_7.yaml}} 1
fail 4514035 2019-11-17 01:32:01 2019-11-18 13:06:34 2019-11-18 15:16:36 2:10:02 1:57:00 0:13:02 mira master ubuntu 18.04 rados/upgrade/nautilus-x-singleton/{0-cluster/{openstack.yaml start.yaml} 1-install/nautilus.yaml 2-partial-upgrade/firsthalf.yaml 3-thrash/default.yaml 4-workload/{rbd-cls.yaml rbd-import-export.yaml readwrite.yaml snaps-few-objects.yaml} 5-workload/{radosbench.yaml rbd_api.yaml} 6-finish-upgrade.yaml 7-octopus.yaml 8-workload/{rbd-python.yaml rgw-swift.yaml snaps-many-objects.yaml} bluestore-bitmap.yaml supported-random-distro$/{ubuntu_latest.yaml} thrashosds-health.yaml} 4
Failure Reason:

"2019-11-18T13:32:15.744619+0000 osd.10 (osd.10) 70 : cluster [ERR] 5.c missing primary copy of 5:303ec511:::mira10515400-86:head, unfound" in cluster log

pass 4514036 2019-11-17 01:32:02 2019-11-18 13:12:52 2019-11-18 14:34:52 1:22:00 1:16:16 0:05:44 mira master rhel 7.7 rados/singleton/{all/osd-recovery-incomplete.yaml msgr-failures/few.yaml msgr/async-v1only.yaml objectstore/filestore-xfs.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}} 1
pass 4514037 2019-11-17 01:32:03 2019-11-18 13:44:55 2019-11-18 14:12:54 0:27:59 0:14:53 0:13:06 mira master centos 7.6 rados/perf/{ceph.yaml objectstore/bluestore-bitmap.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{centos_7.yaml} workloads/radosbench_4M_write.yaml} 1
pass 4514038 2019-11-17 01:32:03 2019-11-18 13:51:21 2019-11-18 14:27:21 0:36:00 0:23:46 0:12:14 mira master centos 7.6 rados/basic/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/few.yaml msgr/async.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{centos_7.yaml} tasks/rgw_snaps.yaml} 2
pass 4514039 2019-11-17 01:32:04 2019-11-18 13:53:03 2019-11-18 14:29:02 0:35:59 0:12:21 0:23:38 mira master centos 7.6 rados/multimon/{clusters/6.yaml msgr-failures/many.yaml msgr/async-v1only.yaml no_pools.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml supported-random-distro$/{centos_7.yaml} tasks/mon_clock_with_skews.yaml} 2
pass 4514040 2019-11-17 01:32:05 2019-11-18 14:13:17 2019-11-18 15:03:16 0:49:59 0:37:34 0:12:25 mira master centos 7.6 rados/monthrash/{ceph.yaml clusters/9-mons.yaml msgr-failures/few.yaml msgr/async-v1only.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{centos_7.yaml} thrashers/sync.yaml workloads/snaps-few-objects.yaml} 2
pass 4514041 2019-11-17 01:32:06 2019-11-18 14:27:22 2019-11-18 15:13:22 0:46:00 0:38:43 0:07:17 mira master rhel 7.7 rados/thrash-erasure-code-isa/{arch/x86_64.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/fastclose.yaml objectstore/bluestore-avl.yaml rados.yaml recovery-overrides/{more-async-partial-recovery.yaml} supported-random-distro$/{rhel_7.yaml} thrashers/default.yaml thrashosds-health.yaml workloads/ec-rados-plugin=isa-k=2-m=1.yaml} 2
pass 4514042 2019-11-17 01:32:07 2019-11-18 14:29:14 2019-11-18 15:49:15 1:20:01 0:56:57 0:23:04 mira master centos 7.6 rados/thrash-erasure-code-big/{ceph.yaml cluster/{12-osds.yaml openstack.yaml} msgr-failures/osd-delay.yaml objectstore/bluestore-bitmap.yaml rados.yaml recovery-overrides/{more-async-recovery.yaml} supported-random-distro$/{centos_7.yaml} thrashers/mapgap.yaml thrashosds-health.yaml workloads/ec-rados-plugin=jerasure-k=4-m=2.yaml} 3
pass 4514043 2019-11-17 01:32:08 2019-11-18 14:35:09 2019-11-18 16:05:10 1:30:01 1:17:15 0:12:46 mira master centos rados/verify/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-thrash/default/{default.yaml thrashosds-health.yaml} msgr-failures/few.yaml msgr/async-v2only.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml tasks/rados_api_tests.yaml validater/valgrind.yaml} 2
fail 4514044 2019-11-17 01:32:09 2019-11-18 15:03:29 2019-11-18 15:33:29 0:30:00 mira master centos 7.6 rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml 2-recovery-overrides/{more-async-partial-recovery.yaml} backoff/normal.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/off.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/bluestore-stupid.yaml rados.yaml supported-random-distro$/{centos_7.yaml} thrashers/morepggrow.yaml thrashosds-health.yaml workloads/small-objects.yaml} 2
Failure Reason:

[Errno 113] No route to host

pass 4514045 2019-11-17 01:32:09 2019-11-18 15:13:25 2019-11-18 15:45:24 0:31:59 0:21:59 0:10:00 mira master centos 7.6 rados/singleton/{all/osd-recovery.yaml msgr-failures/many.yaml msgr/async-v2only.yaml objectstore/bluestore-avl.yaml rados.yaml supported-random-distro$/{centos_7.yaml}} 1
pass 4514046 2019-11-17 01:32:10 2019-11-18 15:16:55 2019-11-18 18:16:57 3:00:02 2:40:26 0:19:36 mira master rhel 7.7 rados/thrash-erasure-code/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} fast/fast.yaml msgr-failures/osd-delay.yaml objectstore/bluestore-stupid.yaml rados.yaml recovery-overrides/{more-async-partial-recovery.yaml} supported-random-distro$/{rhel_7.yaml} thrashers/default.yaml thrashosds-health.yaml workloads/ec-small-objects.yaml} 2
pass 4514047 2019-11-17 01:32:11 2019-11-18 16:27:35 2019-11-18 16:51:35 0:24:00 0:18:27 0:05:33 mira master rhel 7.7 rados/objectstore/{backends/fusestore.yaml supported-random-distro$/{rhel_7.yaml}} 1
pass 4514048 2019-11-17 01:32:12 2019-11-18 16:27:44 2019-11-18 16:55:44 0:28:00 0:21:00 0:07:00 mira master rhel 7.7 rados/singleton-nomsgr/{all/lazy_omap_stats_output.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}} 1
pass 4514049 2019-11-17 01:32:13 2019-11-18 16:51:38 2019-11-18 17:09:37 0:17:59 0:08:20 0:09:39 mira master ubuntu 18.04 rados/singleton/{all/peer.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}} 1
pass 4514050 2019-11-17 01:32:14 2019-11-18 16:55:45 2019-11-18 17:25:45 0:30:00 0:20:11 0:09:49 mira master ubuntu 18.04 rados/perf/{ceph.yaml objectstore/bluestore-comp.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{ubuntu_latest.yaml} workloads/radosbench_omap_write.yaml} 1
pass 4514051 2019-11-17 01:32:15 2019-11-18 17:07:19 2019-11-18 18:21:19 1:14:00 1:05:07 0:08:53 mira master rhel 7.7 rados/standalone/{supported-random-distro$/{rhel_7.yaml} workloads/mon.yaml} 1
pass 4514052 2019-11-17 01:32:15 2019-11-18 17:09:43 2019-11-18 17:35:42 0:25:59 0:14:00 0:11:59 mira master centos 7.6 rados/mgr/{clusters/{2-node-mgr.yaml} debug/mgr.yaml objectstore/bluestore-bitmap.yaml supported-random-distro$/{centos_7.yaml} tasks/ssh_orchestrator.yaml} 2
pass 4514053 2019-11-17 01:32:16 2019-11-18 17:13:26 2019-11-18 17:57:26 0:44:00 0:32:16 0:11:44 mira master centos 7.6 rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml 2-recovery-overrides/{more-active-recovery.yaml} backoff/peering.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/upmap.yaml msgr-failures/osd-delay.yaml msgr/async-v1only.yaml objectstore/filestore-xfs.yaml rados.yaml supported-random-distro$/{centos_7.yaml} thrashers/none.yaml thrashosds-health.yaml workloads/snaps-few-objects.yaml} 2
pass 4514054 2019-11-17 01:32:17 2019-11-18 17:25:49 2019-11-18 18:29:48 1:03:59 0:43:04 0:20:55 mira master centos 7.6 rados/thrash-old-clients/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-install/nautilus.yaml backoff/normal.yaml ceph.yaml clusters/{openstack.yaml three-plus-one.yaml} d-balancer/crush-compat.yaml distro$/{centos_latest.yaml} msgr-failures/few.yaml rados.yaml thrashers/mapgap.yaml thrashosds-health.yaml workloads/snaps-few-objects.yaml} 4
pass 4514055 2019-11-17 01:32:18 2019-11-18 17:35:44 2019-11-18 20:19:46 2:44:02 2:25:49 0:18:13 mira master rhel 7.7 rados/basic/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/many.yaml msgr/async-v1only.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{rhel_7.yaml} tasks/scrub_test.yaml} 2
pass 4514056 2019-11-17 01:32:19 2019-11-18 17:57:46 2019-11-18 18:27:45 0:29:59 0:17:06 0:12:53 mira master centos 7.6 rados/singleton/{all/pg-autoscaler.yaml msgr-failures/many.yaml msgr/async-v1only.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{centos_7.yaml}} 2
pass 4514057 2019-11-17 01:32:20 2019-11-18 18:17:17 2019-11-18 18:37:16 0:19:59 0:10:09 0:09:50 mira master ubuntu 18.04 rados/singleton-nomsgr/{all/librados_hello_world.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}} 1
pass 4514058 2019-11-17 01:32:20 2019-11-18 18:21:36 2019-11-18 21:03:38 2:42:02 2:23:57 0:18:05 mira master rhel 7.7 rados/perf/{ceph.yaml objectstore/bluestore-low-osd-mem-target.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{rhel_7.yaml} workloads/sample_fio.yaml} 1
pass 4514059 2019-11-17 01:32:21 2019-11-18 18:27:48 2019-11-18 19:11:48 0:44:00 0:30:53 0:13:07 mira master centos 7.6 rados/thrash-erasure-code-overwrites/{bluestore-bitmap.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} fast/normal.yaml msgr-failures/osd-delay.yaml rados.yaml recovery-overrides/{default.yaml} supported-random-distro$/{centos_7.yaml} thrashers/pggrow.yaml thrashosds-health.yaml workloads/ec-small-objects-fast-read-overwrites.yaml} 2
pass 4514060 2019-11-17 01:32:22 2019-11-18 18:29:50 2019-11-18 19:07:50 0:38:00 0:20:23 0:17:37 mira master centos 7.6 rados/thrash-erasure-code-shec/{ceph.yaml clusters/{fixed-4.yaml openstack.yaml} msgr-failures/osd-delay.yaml objectstore/filestore-xfs.yaml rados.yaml recovery-overrides/{more-partial-recovery.yaml} supported-random-distro$/{centos_7.yaml} thrashers/default.yaml thrashosds-health.yaml workloads/ec-rados-plugin=shec-k=4-m=3-c=2.yaml} 4
pass 4514061 2019-11-17 01:32:23 2019-11-18 18:37:38 2019-11-18 19:01:38 0:24:00 0:13:09 0:10:51 mira master centos 7.6 rados/singleton/{all/pg-removal-interruption.yaml msgr-failures/few.yaml msgr/async-v2only.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml supported-random-distro$/{centos_7.yaml}} 1
pass 4514062 2019-11-17 01:32:24 2019-11-18 19:01:41 2019-11-18 19:39:40 0:37:59 0:25:04 0:12:55 mira master centos 7.6 rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml 2-recovery-overrides/{more-async-recovery.yaml} backoff/peering_and_degraded.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/crush-compat.yaml msgr-failures/fastclose.yaml msgr/async-v2only.yaml objectstore/bluestore-avl.yaml rados.yaml supported-random-distro$/{centos_7.yaml} thrashers/pggrow.yaml thrashosds-health.yaml workloads/write_fadvise_dontneed.yaml} 2
pass 4514063 2019-11-17 01:32:25 2019-11-18 19:07:53 2019-11-18 19:39:52 0:31:59 0:18:59 0:13:00 mira master centos 7.6 rados/singleton/{all/radostool.yaml msgr-failures/many.yaml msgr/async.yaml objectstore/bluestore-stupid.yaml rados.yaml supported-random-distro$/{centos_7.yaml}} 1
pass 4514064 2019-11-17 01:32:25 2019-11-18 19:11:51 2019-11-18 19:47:51 0:36:00 0:24:37 0:11:23 mira master ubuntu 18.04 rados/basic/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/few.yaml msgr/async-v2only.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} tasks/rados_api_tests.yaml} 2
pass 4514065 2019-11-17 01:32:26 2019-11-18 19:39:55 2019-11-18 20:37:55 0:58:00 0:47:23 0:10:37 mira master ubuntu 18.04 rados/dashboard/{clusters/{2-node-mgr.yaml} debug/mgr.yaml objectstore/bluestore-stupid.yaml supported-random-distro$/{ubuntu_latest.yaml} tasks/dashboard.yaml} 2
pass 4514066 2019-11-17 01:32:27 2019-11-18 19:39:55 2019-11-18 19:59:54 0:19:59 0:10:09 0:09:50 mira master ubuntu 18.04 rados/mgr/{clusters/{2-node-mgr.yaml} debug/mgr.yaml objectstore/bluestore-comp.yaml supported-random-distro$/{ubuntu_latest.yaml} tasks/workunits.yaml} 2
pass 4514067 2019-11-17 01:32:28 2019-11-18 19:48:06 2019-11-18 20:22:06 0:34:00 0:20:25 0:13:35 mira master centos 7.6 rados/monthrash/{ceph.yaml clusters/3-mons.yaml msgr-failures/mon-delay.yaml msgr/async-v2only.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{centos_7.yaml} thrashers/force-sync-many.yaml workloads/pool-create-delete.yaml} 2
pass 4514068 2019-11-17 01:32:29 2019-11-18 20:00:05 2019-11-18 20:26:04 0:25:59 0:14:36 0:11:23 mira master ubuntu 18.04 rados/multimon/{clusters/9.yaml msgr-failures/few.yaml msgr/async-v2only.yaml no_pools.yaml objectstore/bluestore-stupid.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} tasks/mon_recovery.yaml} 3
pass 4514069 2019-11-17 01:32:30 2019-11-18 20:19:56 2019-11-18 20:41:55 0:21:59 0:16:42 0:05:17 mira master rhel 7.7 rados/objectstore/{backends/keyvaluedb.yaml supported-random-distro$/{rhel_7.yaml}} 1
pass 4514070 2019-11-17 01:32:30 2019-11-18 20:22:13 2019-11-18 20:56:12 0:33:59 0:22:57 0:11:02 mira master centos 7.6 rados/singleton-nomsgr/{all/msgr.yaml rados.yaml supported-random-distro$/{centos_7.yaml}} 1
pass 4514071 2019-11-17 01:32:31 2019-11-18 20:26:25 2019-11-18 23:16:27 2:50:02 2:31:29 0:18:33 mira master rhel 7.7 rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml 2-recovery-overrides/{more-partial-recovery.yaml} backoff/normal.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/off.yaml msgr-failures/fastclose.yaml msgr/async-v2only.yaml objectstore/bluestore-avl.yaml rados.yaml supported-random-distro$/{rhel_7.yaml} thrashers/careful.yaml thrashosds-health.yaml workloads/admin_socket_objecter_requests.yaml} 2
pass 4514072 2019-11-17 01:32:32 2019-11-18 20:37:58 2019-11-18 21:19:57 0:41:59 0:30:58 0:11:01 mira master ubuntu 18.04 rados/thrash-erasure-code/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} fast/normal.yaml msgr-failures/fastclose.yaml objectstore/filestore-xfs.yaml rados.yaml recovery-overrides/{more-active-recovery.yaml} supported-random-distro$/{ubuntu_latest.yaml} thrashers/fastread.yaml thrashosds-health.yaml workloads/ec-rados-plugin=clay-k=4-m=2.yaml} 2
pass 4514073 2019-11-17 01:32:33 2019-11-18 20:42:14 2019-11-18 21:28:13 0:45:59 0:22:28 0:23:31 mira master centos 7.6 rados/thrash-erasure-code-big/{ceph.yaml cluster/{12-osds.yaml openstack.yaml} msgr-failures/fastclose.yaml objectstore/bluestore-comp.yaml rados.yaml recovery-overrides/{more-async-partial-recovery.yaml} supported-random-distro$/{centos_7.yaml} thrashers/morepggrow.yaml thrashosds-health.yaml workloads/ec-rados-plugin=lrc-k=4-m=2-l=3.yaml} 3
pass 4514074 2019-11-17 01:32:34 2019-11-18 20:56:25 2019-11-18 21:38:24 0:41:59 0:31:21 0:10:38 mira master ubuntu 18.04 rados/thrash-erasure-code-isa/{arch/x86_64.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/few.yaml objectstore/bluestore-bitmap.yaml rados.yaml recovery-overrides/{more-active-recovery.yaml} supported-random-distro$/{ubuntu_latest.yaml} thrashers/mapgap.yaml thrashosds-health.yaml workloads/ec-rados-plugin=isa-k=2-m=1.yaml} 2
dead 4514075 2019-11-17 01:32:35 2019-11-18 21:03:40 2019-11-18 21:27:39 0:23:59 mira master centos 7.6 rados/thrash-old-clients/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-install/hammer.yaml backoff/peering.yaml ceph.yaml clusters/{openstack.yaml three-plus-one.yaml} d-balancer/off.yaml distro$/{centos_latest.yaml} msgr-failures/osd-delay.yaml rados.yaml thrashers/morepggrow.yaml thrashosds-health.yaml workloads/test_rbd_api.yaml}
Failure Reason:

reached maximum tries (100) after waiting for 600 seconds

pass 4514076 2019-11-17 01:32:36 2019-11-18 21:20:16 2019-11-18 21:44:15 0:23:59 0:12:52 0:11:07 mira master rados/verify/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-thrash/none.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/bluestore-stupid.yaml rados.yaml tasks/rados_cls_all.yaml validater/lockdep.yaml} 2
pass 4514077 2019-11-17 01:32:37 2019-11-18 21:27:57 2019-11-18 21:55:56 0:27:59 0:21:35 0:06:24 mira master rhel 7.7 rados/perf/{ceph.yaml objectstore/bluestore-stupid.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{rhel_7.yaml} workloads/sample_radosbench.yaml} 1
pass 4514078 2019-11-17 01:32:37 2019-11-18 21:28:15 2019-11-18 22:14:15 0:46:00 0:23:54 0:22:06 mira master centos 7.6 rados/singleton/{all/random-eio.yaml msgr-failures/few.yaml msgr/async-v1only.yaml objectstore/filestore-xfs.yaml rados.yaml supported-random-distro$/{centos_7.yaml}} 2
pass 4514079 2019-11-17 01:32:38 2019-11-18 21:38:27 2019-11-18 22:06:26 0:27:59 0:17:04 0:10:55 mira master centos 7.6 rados/singleton/{all/rebuild-mondb.yaml msgr-failures/many.yaml msgr/async-v2only.yaml objectstore/bluestore-avl.yaml rados.yaml supported-random-distro$/{centos_7.yaml}} 1
pass 4514080 2019-11-17 01:32:39 2019-11-18 21:44:18 2019-11-18 22:38:17 0:53:59 0:46:49 0:07:10 mira master rhel 7.7 rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml 2-recovery-overrides/{more-async-recovery.yaml} backoff/peering.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/upmap.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{rhel_7.yaml} thrashers/default.yaml thrashosds-health.yaml workloads/cache-agent-big.yaml} 2
pass 4514081 2019-11-17 01:32:40 2019-11-18 21:56:16 2019-11-19 00:52:18 2:56:02 2:46:32 0:09:30 mira master centos 7.6 rados/standalone/{supported-random-distro$/{centos_7.yaml} workloads/osd.yaml} 1
pass 4514082 2019-11-17 01:32:41 2019-11-18 22:06:29 2019-11-18 22:44:28 0:37:59 0:27:51 0:10:08 mira master ubuntu 18.04 rados/singleton-nomsgr/{all/multi-backfill-reject.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}} 2
pass 4514083 2019-11-17 01:32:42 2019-11-18 22:14:26 2019-11-19 01:04:28 2:50:02 2:32:02 0:18:00 mira master rhel 7.7 rados/perf/{ceph.yaml objectstore/bluestore-stupid.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{rhel_7.yaml} workloads/cosbench_64K_read_write.yaml} 1
pass 4514084 2019-11-17 01:32:42 2019-11-18 22:38:37 2019-11-18 23:42:37 1:04:00 0:57:01 0:06:59 mira master rhel 7.7 rados/singleton/{all/recovery-preemption.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}} 1