Status Job ID Links Posted Started Updated
Runtime
Duration
In Waiting
Machine Teuthology Branch OS Type OS Version Description Nodes
pass 4554018 2019-11-30 02:34:44 2019-11-30 02:34:52 2019-11-30 05:44:54 3:10:02 2:47:01 0:23:01 mira master rhel 7.7 rados/thrash-erasure-code-big/{ceph.yaml cluster/{12-osds.yaml openstack.yaml} msgr-failures/fastclose.yaml objectstore/bluestore-comp.yaml rados.yaml recovery-overrides/{more-active-recovery.yaml} supported-random-distro$/{rhel_7.yaml} thrashers/default.yaml thrashosds-health.yaml workloads/ec-rados-plugin=jerasure-k=4-m=2.yaml} 3
pass 4554019 2019-11-30 02:34:45 2019-11-30 02:34:53 2019-11-30 02:58:52 0:23:59 0:13:09 0:10:50 mira master rados/verify/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-thrash/default/{default.yaml thrashosds-health.yaml} msgr-failures/few.yaml msgr/async-v2only.yaml objectstore/bluestore-bitmap.yaml rados.yaml tasks/rados_cls_all.yaml validater/lockdep.yaml} 2
pass 4554020 2019-11-30 02:34:46 2019-11-30 02:35:09 2019-11-30 03:19:08 0:43:59 0:28:22 0:15:37 mira master ubuntu 18.04 rados/thrash-erasure-code-isa/{arch/x86_64.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/few.yaml objectstore/bluestore-comp.yaml rados.yaml recovery-overrides/{more-partial-recovery.yaml} supported-random-distro$/{ubuntu_latest.yaml} thrashers/pggrow.yaml thrashosds-health.yaml workloads/ec-rados-plugin=isa-k=2-m=1.yaml} 2
pass 4554021 2019-11-30 02:34:47 2019-11-30 02:35:09 2019-11-30 03:23:08 0:47:59 0:37:52 0:10:07 mira master ubuntu 18.04 rados/singleton-bluestore/{all/cephtool.yaml msgr-failures/few.yaml msgr/async-v2only.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}} 1
pass 4554022 2019-11-30 02:34:48 2019-11-30 02:35:09 2019-11-30 05:19:10 2:44:01 2:24:56 0:19:05 mira master rhel 7.7 rados/perf/{ceph.yaml objectstore/bluestore-basic-min-osd-mem-target.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{rhel_7.yaml} workloads/radosbench_4K_seq_read.yaml} 1
pass 4554023 2019-11-30 02:34:49 2019-11-30 02:59:18 2019-11-30 03:49:18 0:50:00 0:34:11 0:15:49 mira master centos 7.6 rados/singleton/{all/thrash-eio.yaml msgr-failures/many.yaml msgr/async-v1only.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{centos_7.yaml}} 2
pass 4554024 2019-11-30 02:34:50 2019-11-30 03:19:29 2019-11-30 04:31:29 1:12:00 1:00:44 0:11:16 mira master ubuntu 18.04 rados/monthrash/{ceph.yaml clusters/9-mons.yaml msgr-failures/few.yaml msgr/async-v2only.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} thrashers/force-sync-many.yaml workloads/rados_mon_osdmap_prune.yaml} 2
pass 4554025 2019-11-30 02:34:51 2019-11-30 03:23:10 2019-11-30 04:29:10 1:06:00 0:41:52 0:24:08 mira master centos 7.6 rados/thrash-old-clients/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-install/mimic-v1only.yaml backoff/peering_and_degraded.yaml ceph.yaml clusters/{openstack.yaml three-plus-one.yaml} d-balancer/crush-compat.yaml distro$/{centos_latest.yaml} msgr-failures/fastclose.yaml rados.yaml thrashers/pggrow.yaml thrashosds-health.yaml workloads/snaps-few-objects.yaml} 4
pass 4554026 2019-11-30 02:34:52 2019-11-30 03:49:37 2019-11-30 04:33:36 0:43:59 0:33:51 0:10:08 mira master ubuntu 18.04 rados/objectstore/{backends/objectcacher-stress.yaml supported-random-distro$/{ubuntu_latest.yaml}} 1
pass 4554027 2019-11-30 02:34:52 2019-11-30 04:29:28 2019-11-30 05:07:27 0:37:59 0:26:16 0:11:43 mira master centos 7.6 rados/singleton-nomsgr/{all/osd_stale_reads.yaml rados.yaml supported-random-distro$/{centos_7.yaml}} 1
pass 4554028 2019-11-30 02:34:53 2019-11-30 04:31:46 2019-11-30 05:01:45 0:29:59 0:19:56 0:10:03 mira master ubuntu 18.04 rados/singleton/{all/thrash-rados/{thrash-rados.yaml thrashosds-health.yaml} msgr-failures/few.yaml msgr/async-v2only.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}} 2
pass 4554029 2019-11-30 02:34:54 2019-11-30 04:33:56 2019-11-30 04:51:55 0:17:59 0:06:37 0:11:22 mira master ubuntu 18.04 rados/multimon/{clusters/9.yaml msgr-failures/many.yaml msgr/async-v1only.yaml no_pools.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} tasks/mon_clock_with_skews.yaml} 3
pass 4554030 2019-11-30 02:34:55 2019-11-30 04:52:14 2019-11-30 05:40:13 0:47:59 0:37:52 0:10:07 mira master rhel 7.7 rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml 2-recovery-overrides/{more-async-partial-recovery.yaml} backoff/peering_and_degraded.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/crush-compat.yaml msgr-failures/osd-delay.yaml msgr/async-v1only.yaml objectstore/filestore-xfs.yaml rados.yaml supported-random-distro$/{rhel_7.yaml} thrashers/pggrow.yaml thrashosds-health.yaml workloads/rados_api_tests.yaml} 2
pass 4554031 2019-11-30 02:34:56 2019-11-30 05:02:06 2019-11-30 06:32:07 1:30:01 1:14:13 0:15:48 mira master ubuntu 18.04 rados/thrash-erasure-code/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} fast/fast.yaml msgr-failures/osd-delay.yaml objectstore/bluestore-stupid.yaml rados.yaml recovery-overrides/{default.yaml} supported-random-distro$/{ubuntu_latest.yaml} thrashers/default.yaml thrashosds-health.yaml workloads/ec-radosbench.yaml} 2
fail 4554032 2019-11-30 02:34:57 2019-11-30 05:07:30 2019-11-30 05:25:30 0:18:00 0:08:36 0:09:24 mira master ubuntu 18.04 rados/mgr/{clusters/{2-node-mgr.yaml} debug/mgr.yaml objectstore/bluestore-avl.yaml supported-random-distro$/{ubuntu_latest.yaml} tasks/orchestrator_cli.yaml} 2
Failure Reason:

Test failure: test_host_ls (tasks.mgr.test_orchestrator_cli.TestOrchestratorCli)

pass 4554033 2019-11-30 02:34:58 2019-11-30 05:19:13 2019-11-30 05:49:12 0:29:59 0:14:57 0:15:02 mira master centos 7.6 rados/basic/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/many.yaml msgr/async-v2only.yaml objectstore/bluestore-stupid.yaml rados.yaml supported-random-distro$/{centos_7.yaml} tasks/rados_striper.yaml} 2
pass 4554034 2019-11-30 02:34:58 2019-11-30 05:25:47 2019-11-30 05:45:46 0:19:59 0:10:19 0:09:40 mira master ubuntu 18.04 rados/perf/{ceph.yaml objectstore/bluestore-bitmap.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{ubuntu_latest.yaml} workloads/radosbench_4M_rand_read.yaml} 1
pass 4554035 2019-11-30 02:34:59 2019-11-30 05:40:31 2019-11-30 06:24:31 0:44:00 0:37:13 0:06:47 mira master rhel 7.7 rados/singleton/{all/thrash_cache_writeback_proxy_none.yaml msgr-failures/many.yaml msgr/async.yaml objectstore/bluestore-stupid.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}} 2
pass 4554036 2019-11-30 02:35:00 2019-11-30 05:44:56 2019-11-30 06:32:56 0:48:00 0:37:47 0:10:13 mira master ubuntu 18.04 rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml 2-recovery-overrides/{more-active-recovery.yaml} backoff/normal.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/off.yaml msgr-failures/fastclose.yaml msgr/async-v2only.yaml objectstore/bluestore-avl.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} thrashers/careful.yaml thrashosds-health.yaml workloads/radosbench.yaml} 2
pass 4554037 2019-11-30 02:35:01 2019-11-30 05:46:04 2019-11-30 06:58:04 1:12:00 1:02:27 0:09:33 mira master rhel 7.7 rados/dashboard/{clusters/{2-node-mgr.yaml} debug/mgr.yaml objectstore/filestore-xfs.yaml supported-random-distro$/{rhel_7.yaml} tasks/dashboard.yaml} 2
pass 4554038 2019-11-30 02:35:02 2019-11-30 05:49:15 2019-11-30 06:17:14 0:27:59 0:13:44 0:14:15 mira master centos 7.6 rados/singleton-nomsgr/{all/pool-access.yaml rados.yaml supported-random-distro$/{centos_7.yaml}} 1
fail 4554039 2019-11-30 02:35:03 2019-11-30 06:17:22 2019-11-30 06:35:21 0:17:59 0:07:45 0:10:14 mira master ubuntu 18.04 rados/ssh/{fixed-2.yaml msgr/async.yaml start.yaml supported-random-distro$/{ubuntu_latest.yaml} tasks/rados_python.yaml} 2
Failure Reason:

Command failed on mira038 with status 22: "sudo /home/ubuntu/cephtest/ceph-daemon --image quay.io/ceph-ci/ceph:378993f279ac979885d08442ca085f5699aa9c1b shell -c /home/ubuntu/cephtest/ceph.conf -k /home/ubuntu/cephtest/ceph.keyring --fsid fb3054d2-133a-11ea-8261-001a4aab830c -- ceph orchestrator mon update 2 'mira038:[v2:172.21.5.138:3301,v1:172.21.5.138:6790]=c'"

pass 4554040 2019-11-30 02:35:04 2019-11-30 06:24:51 2019-11-30 06:48:50 0:23:59 0:12:31 0:11:28 mira master centos 7.6 rados/singleton/{all/watch-notify-same-primary.yaml msgr-failures/few.yaml msgr/async-v1only.yaml objectstore/filestore-xfs.yaml rados.yaml supported-random-distro$/{centos_7.yaml}} 1
pass 4554041 2019-11-30 02:35:05 2019-11-30 06:32:29 2019-11-30 07:18:29 0:46:00 0:31:04 0:14:56 mira master ubuntu 18.04 rados/thrash-erasure-code-overwrites/{bluestore-bitmap.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} fast/normal.yaml msgr-failures/fastclose.yaml rados.yaml recovery-overrides/{more-async-partial-recovery.yaml} supported-random-distro$/{ubuntu_latest.yaml} thrashers/minsize_recovery.yaml thrashosds-health.yaml workloads/ec-pool-snaps-few-objects-overwrites.yaml} 2
pass 4554042 2019-11-30 02:35:06 2019-11-30 06:32:58 2019-11-30 07:12:57 0:39:59 0:20:42 0:19:17 mira master centos 7.6 rados/thrash-erasure-code-shec/{ceph.yaml clusters/{fixed-4.yaml openstack.yaml} msgr-failures/fastclose.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml recovery-overrides/{default.yaml} supported-random-distro$/{centos_7.yaml} thrashers/careful.yaml thrashosds-health.yaml workloads/ec-rados-plugin=shec-k=4-m=3-c=2.yaml} 4
pass 4554043 2019-11-30 02:35:07 2019-11-30 06:35:29 2019-11-30 07:13:29 0:38:00 0:20:22 0:17:38 mira master centos 7.6 rados/thrash-old-clients/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-install/mimic.yaml backoff/normal.yaml ceph.yaml clusters/{openstack.yaml three-plus-one.yaml} d-balancer/off.yaml distro$/{centos_latest.yaml} msgr-failures/few.yaml rados.yaml thrashers/careful.yaml thrashosds-health.yaml workloads/test_rbd_api.yaml} 4
pass 4554044 2019-11-30 02:35:07 2019-11-30 06:48:53 2019-11-30 08:08:54 1:20:01 1:11:38 0:08:23 mira master ubuntu 18.04 rados/standalone/{supported-random-distro$/{ubuntu_latest.yaml} workloads/scrub.yaml} 1
pass 4554045 2019-11-30 02:35:08 2019-11-30 06:58:22 2019-11-30 07:26:22 0:28:00 0:22:14 0:05:46 mira master rhel 7.7 rados/perf/{ceph.yaml objectstore/bluestore-comp.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{rhel_7.yaml} workloads/radosbench_4M_seq_read.yaml} 1
pass 4554046 2019-11-30 02:35:09 2019-11-30 07:13:18 2019-11-30 07:39:18 0:26:00 0:18:30 0:07:30 mira master rhel 7.7 rados/singleton/{all/admin-socket.yaml msgr-failures/many.yaml msgr/async-v2only.yaml objectstore/filestore-xfs.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}} 1
pass 4554047 2019-11-30 02:35:10 2019-11-30 07:13:31 2019-11-30 07:43:30 0:29:59 0:19:44 0:10:15 mira master ubuntu 18.04 rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml 2-recovery-overrides/{more-async-recovery.yaml} backoff/peering.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/upmap.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} thrashers/default.yaml thrashosds-health.yaml workloads/redirect.yaml} 2
pass 4554048 2019-11-30 02:35:11 2019-11-30 07:18:31 2019-11-30 08:08:31 0:50:00 0:42:58 0:07:02 mira master rhel 7.7 rados/basic/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/few.yaml msgr/async.yaml objectstore/filestore-xfs.yaml rados.yaml supported-random-distro$/{rhel_7.yaml} tasks/rados_workunit_loadgen_big.yaml} 2
pass 4554049 2019-11-30 02:35:12 2019-11-30 07:26:24 2019-11-30 07:50:23 0:23:59 0:12:05 0:11:54 mira master ubuntu 18.04 rados/thrash-erasure-code-big/{ceph.yaml cluster/{12-osds.yaml openstack.yaml} msgr-failures/few.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml recovery-overrides/{more-active-recovery.yaml} supported-random-distro$/{ubuntu_latest.yaml} thrashers/fastread.yaml thrashosds-health.yaml workloads/ec-rados-plugin=lrc-k=4-m=2-l=3.yaml} 3
pass 4554050 2019-11-30 02:35:13 2019-11-30 07:39:23 2019-11-30 08:29:23 0:50:00 0:34:38 0:15:22 mira master centos rados/verify/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-thrash/none.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/bluestore-comp.yaml rados.yaml tasks/mon_recovery.yaml validater/valgrind.yaml} 2
pass 4554051 2019-11-30 02:35:14 2019-11-30 07:43:48 2019-11-30 08:29:48 0:46:00 0:32:21 0:13:39 mira master centos 7.6 rados/thrash-erasure-code-isa/{arch/x86_64.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/osd-delay.yaml objectstore/bluestore-comp.yaml rados.yaml recovery-overrides/{more-active-recovery.yaml} supported-random-distro$/{centos_7.yaml} thrashers/careful.yaml thrashosds-health.yaml workloads/ec-rados-plugin=isa-k=2-m=1.yaml} 2
pass 4554052 2019-11-30 02:35:15 2019-11-30 07:50:42 2019-11-30 08:22:41 0:31:59 0:25:24 0:06:35 mira master rhel 7.7 rados/mgr/{clusters/{2-node-mgr.yaml} debug/mgr.yaml objectstore/bluestore-bitmap.yaml supported-random-distro$/{rhel_7.yaml} tasks/progress.yaml} 2
fail 4554053 2019-11-30 02:35:15 2019-11-30 08:08:49 2019-11-30 10:10:50 2:02:01 1:38:34 0:23:27 mira master ubuntu 18.04 rados/objectstore/{backends/objectstore.yaml supported-random-distro$/{ubuntu_latest.yaml}} 1
Failure Reason:

Command crashed: 'sudo TESTDIR=/home/ubuntu/cephtest bash -c \'mkdir $TESTDIR/archive/ostest && cd $TESTDIR/archive/ostest && ulimit -Sn 16384 && CEPH_ARGS="--no-log-to-stderr --log-file $TESTDIR/archive/ceph_test_objectstore.log --debug-filestore 20 --debug-bluestore 20" ceph_test_objectstore --gtest_filter=-*/3 --gtest_catch_exceptions=0\''

pass 4554054 2019-11-30 02:35:16 2019-11-30 08:08:55 2019-11-30 08:52:55 0:44:00 0:33:13 0:10:47 mira master centos 7.6 rados/singleton-nomsgr/{all/recovery-unfound-found.yaml rados.yaml supported-random-distro$/{centos_7.yaml}} 1
pass 4554055 2019-11-30 02:35:17 2019-11-30 08:23:00 2019-11-30 08:48:59 0:25:59 0:20:11 0:05:48 mira master rhel 7.7 rados/singleton/{all/deduptool.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/bluestore-avl.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}} 1
pass 4554056 2019-11-30 02:35:18 2019-11-30 08:29:45 2019-11-30 11:31:47 3:02:02 2:43:30 0:18:32 mira master rhel 7.7 rados/monthrash/{ceph.yaml clusters/3-mons.yaml msgr-failures/mon-delay.yaml msgr/async.yaml objectstore/bluestore-stupid.yaml rados.yaml supported-random-distro$/{rhel_7.yaml} thrashers/many.yaml workloads/rados_mon_workunits.yaml} 2
pass 4554057 2019-11-30 02:35:19 2019-11-30 08:29:49 2019-11-30 11:13:51 2:44:02 2:24:40 0:19:22 mira master rhel 7.7 rados/perf/{ceph.yaml objectstore/bluestore-low-osd-mem-target.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{rhel_7.yaml} workloads/radosbench_4M_write.yaml} 1
pass 4554058 2019-11-30 02:35:20 2019-11-30 08:49:18 2019-11-30 09:25:17 0:35:59 0:25:09 0:10:50 mira master ubuntu 18.04 rados/thrash-erasure-code/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} fast/normal.yaml msgr-failures/fastclose.yaml objectstore/filestore-xfs.yaml rados.yaml recovery-overrides/{more-async-partial-recovery.yaml} supported-random-distro$/{ubuntu_latest.yaml} thrashers/fastread.yaml thrashosds-health.yaml workloads/ec-small-objects-fast-read.yaml} 2
pass 4554059 2019-11-30 02:35:21 2019-11-30 08:52:57 2019-11-30 09:12:56 0:19:59 0:11:04 0:08:55 mira master ubuntu 18.04 rados/singleton/{all/divergent_priors.yaml msgr-failures/many.yaml msgr/async-v1only.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}} 1
pass 4554060 2019-11-30 02:35:22 2019-11-30 09:12:59 2019-11-30 09:52:59 0:40:00 0:32:33 0:07:27 mira master rhel 7.7 rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml 2-recovery-overrides/{more-async-recovery.yaml} backoff/peering_and_degraded.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/crush-compat.yaml msgr-failures/osd-delay.yaml msgr/async-v1only.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{rhel_7.yaml} thrashers/mapgap.yaml thrashosds-health.yaml workloads/redirect_promote_tests.yaml} 2
pass 4554061 2019-11-30 02:35:23 2019-11-30 09:25:35 2019-11-30 10:07:35 0:42:00 0:35:06 0:06:54 mira master rhel 7.7 rados/multimon/{clusters/21.yaml msgr-failures/few.yaml msgr/async-v2only.yaml no_pools.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{rhel_7.yaml} tasks/mon_recovery.yaml} 3
pass 4554062 2019-11-30 02:35:24 2019-11-30 09:53:19 2019-11-30 10:45:19 0:52:00 0:33:49 0:18:11 mira master centos 7.6 rados/thrash-old-clients/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-install/nautilus-v1only.yaml backoff/peering.yaml ceph.yaml clusters/{openstack.yaml three-plus-one.yaml} d-balancer/crush-compat.yaml distro$/{centos_latest.yaml} msgr-failures/osd-delay.yaml rados.yaml thrashers/default.yaml thrashosds-health.yaml workloads/cache-snaps.yaml} 4
pass 4554063 2019-11-30 02:35:25 2019-11-30 10:07:56 2019-11-30 10:35:56 0:28:00 0:20:59 0:07:01 mira master rhel 7.7 rados/singleton-nomsgr/{all/version-number-sanity.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}} 1
pass 4554064 2019-11-30 02:35:26 2019-11-30 10:10:52 2019-11-30 10:28:51 0:17:59 0:08:19 0:09:40 mira master ubuntu 18.04 rados/singleton/{all/divergent_priors2.yaml msgr-failures/few.yaml msgr/async-v2only.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}} 1
pass 4554065 2019-11-30 02:35:26 2019-11-30 10:29:11 2019-11-30 11:15:10 0:45:59 0:27:50 0:18:09 mira master centos 7.6 rados/basic/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/many.yaml msgr/async-v1only.yaml objectstore/bluestore-avl.yaml rados.yaml supported-random-distro$/{centos_7.yaml} tasks/rados_workunit_loadgen_mix.yaml} 2
pass 4554066 2019-11-30 02:35:27 2019-11-30 10:36:17 2019-11-30 11:14:16 0:37:59 0:24:25 0:13:34 mira master centos 7.6 rados/perf/{ceph.yaml objectstore/bluestore-stupid.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{centos_7.yaml} workloads/radosbench_omap_write.yaml} 1
pass 4554067 2019-11-30 02:35:28 2019-11-30 10:45:39 2019-11-30 11:27:39 0:42:00 0:24:33 0:17:27 mira master centos 7.6 rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml 2-recovery-overrides/{more-async-recovery.yaml} backoff/normal.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/off.yaml msgr-failures/fastclose.yaml msgr/async-v2only.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml supported-random-distro$/{centos_7.yaml} thrashers/morepggrow.yaml thrashosds-health.yaml workloads/redirect_set_object.yaml} 2
pass 4554068 2019-11-30 02:35:29 2019-11-30 11:14:10 2019-11-30 13:56:12 2:42:02 2:23:26 0:18:36 mira master rhel 7.7 rados/singleton/{all/dump-stuck.yaml msgr-failures/many.yaml msgr/async.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}} 1
pass 4554069 2019-11-30 02:35:30 2019-11-30 11:14:18 2019-11-30 11:42:17 0:27:59 0:21:58 0:06:01 mira master rhel 7.7 rados/mgr/{clusters/{2-node-mgr.yaml} debug/mgr.yaml objectstore/bluestore-comp.yaml supported-random-distro$/{rhel_7.yaml} tasks/prometheus.yaml} 2
pass 4554070 2019-11-30 02:35:31 2019-11-30 11:15:30 2019-11-30 11:51:29 0:35:59 0:25:03 0:10:56 mira master ubuntu 18.04 rados/thrash-erasure-code-overwrites/{bluestore-bitmap.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} fast/fast.yaml msgr-failures/few.yaml rados.yaml recovery-overrides/{default.yaml} supported-random-distro$/{ubuntu_latest.yaml} thrashers/morepggrow.yaml thrashosds-health.yaml workloads/ec-small-objects-fast-read-overwrites.yaml} 2
pass 4554071 2019-11-30 02:35:32 2019-11-30 11:27:58 2019-11-30 11:53:57 0:25:59 0:12:58 0:13:01 mira master ubuntu 18.04 rados/thrash-erasure-code-shec/{ceph.yaml clusters/{fixed-4.yaml openstack.yaml} msgr-failures/few.yaml objectstore/bluestore-stupid.yaml rados.yaml recovery-overrides/{default.yaml} supported-random-distro$/{ubuntu_latest.yaml} thrashers/default.yaml thrashosds-health.yaml workloads/ec-rados-plugin=shec-k=4-m=3-c=2.yaml} 4
pass 4554072 2019-11-30 02:35:33 2019-11-30 11:32:06 2019-11-30 13:02:06 1:30:00 1:00:33 0:29:27 mira master centos 7.6 rados/dashboard/{clusters/{2-node-mgr.yaml} debug/mgr.yaml objectstore/bluestore-avl.yaml supported-random-distro$/{centos_7.yaml} tasks/dashboard.yaml} 2
pass 4554073 2019-11-30 02:35:34 2019-11-30 11:42:19 2019-11-30 12:10:18 0:27:59 0:14:22 0:13:37 mira master centos 7.6 rados/objectstore/{backends/alloc-hint.yaml supported-random-distro$/{centos_7.yaml}} 1
pass 4554074 2019-11-30 02:35:34 2019-11-30 11:51:51 2019-11-30 12:15:50 0:23:59 0:14:56 0:09:03 mira master centos 7.6 rados/rest/{mgr-restful.yaml supported-random-distro$/{centos_7.yaml}} 1
pass 4554075 2019-11-30 02:35:35 2019-11-30 11:53:59 2019-11-30 12:59:59 1:06:00 0:54:46 0:11:14 mira master centos 7.6 rados/singleton-bluestore/{all/cephtool.yaml msgr-failures/many.yaml msgr/async.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{centos_7.yaml}} 1
pass 4554076 2019-11-30 02:35:36 2019-11-30 12:10:20 2019-11-30 12:44:20 0:34:00 0:21:41 0:12:19 mira master centos rados/singleton-flat/valgrind-leaks.yaml 1
pass 4554077 2019-11-30 02:35:37 2019-11-30 12:16:11 2019-11-30 12:46:11 0:30:00 0:17:59 0:12:01 mira master centos 7.6 rados/singleton-nomsgr/{all/admin_socket_output.yaml rados.yaml supported-random-distro$/{centos_7.yaml}} 1
fail 4554078 2019-11-30 02:35:38 2019-11-30 12:44:38 2019-11-30 13:12:37 0:27:59 0:20:20 0:07:39 mira master rhel 7.7 rados/ssh/{fixed-2.yaml msgr/async-v1only.yaml start.yaml supported-random-distro$/{rhel_7.yaml} tasks/rados_api_tests.yaml} 2
Failure Reason:

Command failed on mira105 with status 22: "sudo /home/ubuntu/cephtest/ceph-daemon --image quay.io/ceph-ci/ceph:378993f279ac979885d08442ca085f5699aa9c1b shell -c /home/ubuntu/cephtest/ceph.conf -k /home/ubuntu/cephtest/ceph.keyring --fsid 684e021c-1372-11ea-8261-001a4aab830c -- ceph orchestrator mon update 2 'mira105:[v1:172.21.8.110:6789]=b'"

pass 4554079 2019-11-30 02:35:39 2019-11-30 12:46:32 2019-11-30 13:14:30 0:27:58 0:22:21 0:05:37 mira master rhel 7.7 rados/standalone/{supported-random-distro$/{rhel_7.yaml} workloads/crush.yaml} 1
dead 4554080 2019-11-30 02:35:40 2019-11-30 13:00:19 2019-12-01 01:02:43 12:02:24 mira master rhel 7.7 rados/upgrade/mimic-x-singleton/{0-cluster/{openstack.yaml start.yaml} 1-install/mimic.yaml 2-partial-upgrade/firsthalf.yaml 3-thrash/default.yaml 4-workload/{rbd-cls.yaml rbd-import-export.yaml readwrite.yaml snaps-few-objects.yaml} 5-workload/{radosbench.yaml rbd_api.yaml} 6-finish-upgrade.yaml 7-nautilus.yaml 8-workload/{rbd-python.yaml rgw-swift.yaml snaps-many-objects.yaml} bluestore-bitmap.yaml supported-random-distro$/{rhel_7.yaml} thrashosds-health.yaml} 4
fail 4554081 2019-11-30 02:35:40 2019-11-30 13:02:27 2019-11-30 15:34:29 2:32:02 2:13:56 0:18:06 mira master rhel 7.7 rados/singleton/{all/ec-lost-unfound.yaml msgr-failures/few.yaml msgr/async-v1only.yaml objectstore/bluestore-stupid.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}} 1
Failure Reason:

'strip_empty_ends': True, u'_raw_params': u'sgdisk --zap-all /dev/sdd || sgdisk --zap-all /dev/sdd', u'removes': None, u'argv': None, u'creates': None, u'chdir': None, u'stdin_add_newline': True, u'stdin': None}}, u'start': u'2019-11-30 15:33:18.316452'}, {'stderr_lines': [], u'changed': True, u'stdout': u'Creating new GPT entries.\nGPT data structures destroyed! You may now partition the disk using fdisk or\nother utilities.', u'delta': u'0:00:01.021669', 'stdout_lines': [u'Creating new GPT entries.', u'GPT data structures destroyed! You may now partition the disk using fdisk or', u'other utilities.'], '_ansible_item_label': {'value': {u'sectorsize': u'512', u'vendor': u'Hitachi', u'links': {u'masters': [], u'labels': [], u'ids': [u'scsi-2001b4d2011a85600'], u'uuids': [u'd42d97e4-7a28-488a-9b24-8e4e9388d6d8']}, u'sas_device_handle': None, u'host': u'RAID bus controller: Areca Technology Corp. ARC-1680 series PCIe to SAS/SATA 3Gb RAID Controller', u'support_discard': u'0', u'serial': u'JPW9K0N211XZVE', u'holders': [], u'size': u'931.51 GB', u'scheduler_mode': u'deadline', u'rotational': u'1', u'sectors': u'1953525168', u'sas_address': None, u'virtual': 1, u'removable': u'0', u'model': u'HUA722010CLA330', u'partitions': {}}, 'key': u'sde'}, 'ansible_loop_var': u'item', u'end': u'2019-11-30 15:33:21.060383', '_ansible_no_log': False, 'item': {'value': {u'sectorsize': u'512', u'vendor': u'Hitachi', u'links': {u'masters': [], u'labels': [], u'ids': [u'scsi-2001b4d2011a85600'], u'uuids': [u'd42d97e4-7a28-488a-9b24-8e4e9388d6d8']}, u'sas_device_handle': None, u'host': u'RAID bus controller: Areca Technology Corp. ARC-1680 series PCIe to SAS/SATA 3Gb RAID Controller', u'support_discard': u'0', u'serial': u'JPW9K0N211XZVE', u'holders': [], u'size': u'931.51 GB', u'scheduler_mode': u'deadline', u'rotational': u'1', u'sectors': u'1953525168', u'sas_address': None, u'virtual': 1, u'removable': u'0', u'model': u'HUA722010CLA330', u'partitions': {}}, 'key': u'sde'}, u'cmd': u'sgdisk --zap-all /dev/sde || sgdisk --zap-all /dev/sde', 'failed': False, u'stderr': u'', u'rc': 0, u'invocation': {u'module_args': {u'warn': True, u'executable': None, u'_uses_shell': True, u'strip_empty_ends': True, u'_raw_params': u'sgdisk --zap-all /dev/sde || sgdisk --zap-all /dev/sde', u'removes': None, u'argv': None, u'creates': None, u'chdir': None, u'stdin_add_newline': True, u'stdin': None}}, u'start': u'2019-11-30 15:33:20.038714'}, {'stderr_lines': [], u'changed': True, u'stdout': u'Creating new GPT entries.\nGPT data structures destroyed! You may now partition the disk using fdisk or\nother utilities.', u'delta': u'0:00:01.011304', 'stdout_lines': [u'Creating new GPT entries.', u'GPT data structures destroyed! You may now partition the disk using fdisk or', u'other utilities.'], '_ansible_item_label': {'value': {u'sectorsize': u'512', u'vendor': u'Hitachi', u'links': {u'masters': [], u'labels': [], u'ids': [u'scsi-2001b4d2040775100'], u'uuids': [u'4bdaa43c-5f44-4d3d-8a8d-d2396dfea7b9']}, u'sas_device_handle': None, u'host': u'RAID bus controller: Areca Technology Corp. ARC-1680 series PCIe to SAS/SATA 3Gb RAID Controller', u'support_discard': u'0', u'serial': u'JPW9K0N204WG1E', u'holders': [], u'size': u'931.51 GB', u'scheduler_mode': u'deadline', u'rotational': u'1', u'sectors': u'1953525168', u'sas_address': None, u'virtual': 1, u'removable': u'0', u'model': u'HUA722010CLA330', u'partitions': {}}, 'key': u'sdf'}, 'ansible_loop_var': u'item', u'end': u'2019-11-30 15:33:22.318840', '_ansible_no_log': False, 'item': {'value': {u'sectorsize': u'512', u'vendor': u'Hitachi', u'links': {u'masters': [], u'labels': [], u'ids': [u'scsi-2001b4d2040775100'], u'uuids': [u'4bdaa43c-5f44-4d3d-8a8d-d2396dfea7b9']}, u'sas_device_handle': None, u'host': u'RAID bus controller: Areca Technology Corp. ARC-1680 series PCIe to SAS/SATA 3Gb RAID Controller', u'support_discard': u'0', u'serial': u'JPW9K0N204WG1E', u'holders': [], u'size': u'931.51 GB', u'scheduler_mode': u'deadline', u'rotational': u'1', u'sectors': u'1953525168', u'sas_address': None, u'virtual': 1, u'removable': u'0', u'model': u'HUA722010CLA330', u'partitions': {}}, 'key': u'sdf'}, u'cmd': u'sgdisk --zap-all /dev/sdf || sgdisk --zap-all /dev/sdf', 'failed': False, u'stderr': u'', u'rc': 0, u'invocation': {u'module_args': {u'warn': True, u'executable': None, u'_uses_shell': True, u'strip_empty_ends': True, u'_raw_params': u'sgdisk --zap-all /dev/sdf || sgdisk --zap-all /dev/sdf', u'removes': None, u'argv': None, u'creates': None, u'chdir': None, u'stdin_add_newline': True, u'stdin': None}}, u'start': u'2019-11-30 15:33:21.307536'}, {'ansible_loop_var': u'item', '_ansible_no_log': False, 'skip_reason': u'Conditional result was False', 'item': {'value': {u'sectorsize': u'512', u'vendor': u'Seagate', u'links': {u'masters': [], u'labels': [], u'ids': [u'scsi-2001b4d2000000000'], u'uuids': []}, u'sas_device_handle': None, u'host': u'RAID bus controller: Areca Technology Corp. ARC-1680 series PCIe to SAS/SATA 3Gb RAID Controller', u'support_discard': u'0', u'serial': u'5VP66QW9', u'holders': [], u'size': u'931.51 GB', u'scheduler_mode': u'deadline', u'rotational': u'1', u'sectors': u'1953525168', u'sas_address': None, u'virtual': 1, u'removable': u'0', u'model': u'ST31000524AS', u'partitions': {u'sda1': {u'start': u'2048', u'sectorsize': 512, u'uuid': u'f1bca609-9cd3-46ed-8f34-af0538e7246e', u'sectors': u'1953522688', u'holders': [], u'links': {u'masters': [], u'labels': [], u'ids': [u'scsi-2001b4d2000000000-part1'], u'uuids': [u'f1bca609-9cd3-46ed-8f34-af0538e7246e']}, u'size': u'931.51 GB'}}}, 'key': u'sda'}, 'skipped': True, 'changed': False, '_ansible_item_label': {'value': {u'sectorsize': u'512', u'vendor': u'Seagate', u'links': {u'masters': [], u'labels': [], u'ids': [u'scsi-2001b4d2000000000'], u'uuids': []}, u'sas_device_handle': None, u'host': u'RAID bus controller: Areca Technology Corp. ARC-1680 series PCIe to SAS/SATA 3Gb RAID Controller', u'support_discard': u'0', u'serial': u'5VP66QW9', u'holders': [], u'size': u'931.51 GB', u'scheduler_mode': u'deadline', u'rotational': u'1', u'sectors': u'1953525168', u'sas_address': None, u'virtual': 1, u'removable': u'0', u'model': u'ST31000524AS', u'partitions': {u'sda1': {u'start': u'2048', u'sectorsize': 512, u'uuid': u'f1bca609-9cd3-46ed-8f34-af0538e7246e', u'sectors': u'1953522688', u'holders': [], u'links': {u'masters': [], u'labels': [], u'ids': [u'scsi-2001b4d2000000000-part1'], u'uuids': [u'f1bca609-9cd3-46ed-8f34-af0538e7246e']}, u'size': u'931.51 GB'}}}, 'key': u'sda'}}, {'stderr_lines': [], u'changed': True, u'stdout': u'Creating new GPT entries.\nGPT data structures destroyed! You may now partition the disk using fdisk or\nother utilities.', u'delta': u'0:00:01.011196', 'stdout_lines': [u'Creating new GPT entries.', u'GPT data structures destroyed! You may now partition the disk using fdisk or', u'other utilities.'], '_ansible_item_label': {'value': {u'sectorsize': u'512', u'vendor': u'Seagate', u'links': {u'masters': [u'dm-0'], u'labels': [], u'ids': [], u'uuids': []}, u'sas_device_handle': None, u'host': u'RAID bus controller: Areca Technology Corp. ARC-1680 series PCIe to SAS/SATA 3Gb RAID Controller', u'support_discard': u'0', u'serial': u'5VP8NWLD', u'holders': [u'mpatha'], u'size': u'931.51 GB', u'scheduler_mode': u'deadline', u'rotational': u'1', u'sectors': u'1953525168', u'sas_address': None, u'virtual': 1, u'removable': u'0', u'model': u'ST31000528AS', u'partitions': {}}, 'key': u'sdb'}, 'ansible_loop_var': u'item', u'end': u'2019-11-30 15:33:23.599868', '_ansible_no_log': False, 'item': {'value': {u'sectorsize': u'512', u'vendor': u'Seagate', u'links': {u'masters': [u'dm-0'], u'labels': [], u'ids': [], u'uuids': []}, u'sas_device_handle': None, u'host': u'RAID bus controller: Areca Technology Corp. ARC-1680 series PCIe to SAS/SATA 3Gb RAID Controller', u'support_discard': u'0', u'serial': u'5VP8NWLD', u'holders': [u'mpatha'], u'size': u'931.51 GB', u'scheduler_mode': u'deadline', u'rotational': u'1', u'sectors': u'1953525168', u'sas_address': None, u'virtual': 1, u'removable': u'0', u'model': u'ST31000528AS', u'partitions': {}}, 'key': u'sdb'}, u'cmd': u'sgdisk --zap-all /dev/sdb || sgdisk --zap-all /dev/sdb', 'failed': False, u'stderr': u'', u'rc': 0, u'invocation': {u'module_args': {u'warn': True, u'executable': None, u'_uses_shell': True, u'strip_empty_ends': True, u'_raw_params': u'sgdisk --zap-all /dev/sdb || sgdisk --zap-all /dev/sdb', u'removes': None, u'argv': None, u'creates': None, u'chdir': None, u'stdin_add_newline': True, u'stdin': None}}, u'start': u'2019-11-30 15:33:22.588672'}, {'stderr_lines': [], u'changed': True, u'stdout': u'Creating new GPT entries.\nGPT data structures destroyed! You may now partition the disk using fdisk or\nother utilities.', u'delta': u'0:00:01.025515', 'stdout_lines': [u'Creating new GPT entries.', u'GPT data structures destroyed! You may now partition the disk using fdisk or', u'other utilities.'], '_ansible_item_label': {'value': {u'sectorsize': u'512', u'vendor': u'Seagate', u'links': {u'masters': [u'dm-0'], u'labels': [], u'ids': [], u'uuids': []}, u'sas_device_handle': None, u'host': u'RAID bus controller: Areca Technology Corp. ARC-1680 series PCIe to SAS/SATA 3Gb RAID Controller', u'support_discard': u'0', u'serial': u'6VPBDH90', u'holders': [u'mpatha'], u'size': u'931.51 GB', u'scheduler_mode': u'deadline', u'rotational': u'1', u'sectors': u'1953525168', u'sas_address': None, u'virtual': 1, u'removable': u'0', u'model': u'ST31000528AS', u'partitions': {}}, 'key': u'sdc'}, 'ansible_loop_var': u'item', u'end': u'2019-11-30 15:33:24.871910', '_ansible_no_log': False, 'item': {'value': {u'sectorsize': u'512', u'vendor': u'Seagate', u'links': {u'masters': [u'dm-0'], u'labels': [], u'ids': [], u'uuids': []}, u'sas_device_handle': None, u'host': u'RAID bus controller: Areca Technology Corp. ARC-1680 series PCIe to SAS/SATA 3Gb RAID Controller', u'support_discard': u'0', u'serial': u'6VPBDH90', u'holders': [u'mpatha'], u'size': u'931.51 GB', u'scheduler_mode': u'deadline', u'rotational': u'1', u'sectors': u'1953525168', u'sas_address': None, u'virtual': 1, u'removable': u'0', u'model': u'ST31000528AS', u'partitions': {}}, 'key': u'sdc'}, u'cmd': u'sgdisk --zap-all /dev/sdc || sgdisk --zap-all /dev/sdc', 'failed': False, u'stderr': u'', u'rc': 0, u'invocation': {u'module_args': {u'warn': True, u'executable': None, u'_uses_shell': True, u'strip_empty_ends': True, u'_raw_params': u'sgdisk --zap-all /dev/sdc || sgdisk --zap-all /dev/sdc', u'removes': None, u'argv': None, u'creates': None, u'chdir': None, u'stdin_add_newline': True, u'stdin': None}}, u'start': u'2019-11-30 15:33:23.846395'}, {'stderr_lines': [u'Problem opening /dev/dm-0 for reading! Error is 2.', u'The specified file does not exist!', u"Problem opening '' for writing! Program will now terminate.", u'Warning! MBR not overwritten! Error is 2!', u'Problem opening /dev/dm-0 for reading! Error is 2.', u'The specified file does not exist!', u"Problem opening '' for writing! Program will now terminate.", u'Warning! MBR not overwritten! Error is 2!'], u'changed': True, u'stdout': u'', u'invocation': {u'module_args': {u'warn': True, u'executable': None, u'_uses_shell': True, u'strip_empty_ends': True, u'_raw_params': u'sgdisk --zap-all /dev/dm-0 || sgdisk --zap-all /dev/dm-0', u'removes': None, u'argv': None, u'creates': None, u'chdir': None, u'stdin_add_newline': True, u'stdin': None}}, u'delta': u'0:00:00.008489', 'stdout_lines': [], '_ansible_item_label': {'value': {u'sectorsize': u'512', u'vendor': None, u'links': {u'masters': [], u'labels': [], u'ids': [u'dm-name-mpatha', u'dm-uuid-mpath-2001b4d2000000000'], u'uuids': []}, u'sas_device_handle': None, u'host': u'', u'support_discard': u'0', u'serial': u'5VP8NWLD', u'holders': [], u'size': u'931.51 GB', u'scheduler_mode': u'deadline', u'rotational': u'1', u'sectors': u'1953525168', u'sas_address': None, u'virtual': 1, u'removable': u'0', u'model': None, u'partitions': {}}, 'key': u'dm-0'}, 'ansible_loop_var': u'item', u'end': u'2019-11-30 15:33:25.142683', '_ansible_no_log': False, u'start': u'2019-11-30 15:33:25.134194', u'failed': True, u'cmd': u'sgdisk --zap-all /dev/dm-0 || sgdisk --zap-all /dev/dm-0', 'item': {'value': {u'sectorsize': u'512', u'vendor': None, u'links': {u'masters': [], u'labels': [], u'ids': [u'dm-name-mpatha', u'dm-uuid-mpath-2001b4d2000000000'], u'uuids': []}, u'sas_device_handle': None, u'host': u'', u'support_discard': u'0', u'serial': u'5VP8NWLD', u'holders': [], u'size': u'931.51 GB', u'scheduler_mode': u'deadline', u'rotational': u'1', u'sectors': u'1953525168', u'sas_address': None, u'virtual': 1, u'removable': u'0', u'model': None, u'partitions': {}}, 'key': u'dm-0'}, u'stderr': u"Problem opening /dev/dm-0 for reading! Error is 2.\nThe specified file does not exist!\nProblem opening '' for writing! Program will now terminate.\nWarning! MBR not overwritten! Error is 2!\nProblem opening /dev/dm-0 for reading! Error is 2.\nThe specified file does not exist!\nProblem opening '' for writing! Program will now terminate.\nWarning! MBR not overwritten! Error is 2!", u'rc': 2, u'msg': u'non-zero return code'}]}}Traceback (most recent call last): File "/home/teuthworker/src/git.ceph.com_git_ceph-cm-ansible_master/callback_plugins/failure_log.py", line 44, in log_failure log.error(yaml.safe_dump(failure)) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/__init__.py", line 309, in safe_dump return dump_all([data], stream, Dumper=SafeDumper, **kwds) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/__init__.py", line 281, in dump_all dumper.represent(data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 29, in represent node = self.represent_data(data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 58, in represent_data node = self.yaml_representers[data_types[0]](self, data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 227, in represent_dict return self.represent_mapping(u'tag:yaml.org,2002:map', data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 125, in represent_mapping node_value = self.represent_data(item_value) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 58, in represent_data node = self.yaml_representers[data_types[0]](self, data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 227, in represent_dict return self.represent_mapping(u'tag:yaml.org,2002:map', data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 125, in represent_mapping node_value = self.represent_data(item_value) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 58, in represent_data node = self.yaml_representers[data_types[0]](self, data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 219, in represent_list return self.represent_sequence(u'tag:yaml.org,2002:seq', data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 102, in represent_sequence node_item = self.represent_data(item) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 58, in represent_data node = self.yaml_representers[data_types[0]](self, data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 227, in represent_dict return self.represent_mapping(u'tag:yaml.org,2002:map', data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 125, in represent_mapping node_value = self.represent_data(item_value) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 58, in represent_data node = self.yaml_representers[data_types[0]](self, data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 227, in represent_dict return self.represent_mapping(u'tag:yaml.org,2002:map', data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 125, in represent_mapping node_value = self.represent_data(item_value) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 68, in represent_data node = self.yaml_representers[None](self, data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 251, in represent_undefined raise RepresenterError("cannot represent an object", data)RepresenterError: ('cannot represent an object', u'sdd')

pass 4554082 2019-11-30 02:35:41 2019-11-30 13:12:47 2019-11-30 15:58:49 2:46:02 2:26:31 0:19:31 mira master rhel 7.7 rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml 2-recovery-overrides/{more-async-partial-recovery.yaml} backoff/peering.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/upmap.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/bluestore-stupid.yaml rados.yaml supported-random-distro$/{rhel_7.yaml} thrashers/none.yaml thrashosds-health.yaml workloads/set-chunk-promote-flush.yaml} 2
pass 4554083 2019-11-30 02:35:42 2019-11-30 13:14:33 2019-11-30 13:42:32 0:27:59 0:21:12 0:06:47 mira master rhel 7.7 rados/perf/{ceph.yaml objectstore/bluestore-basic-min-osd-mem-target.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{rhel_7.yaml} workloads/sample_fio.yaml} 1
pass 4554084 2019-11-30 02:35:43 2019-11-30 13:42:52 2019-11-30 14:38:52 0:56:00 0:39:57 0:16:03 mira master centos 7.6 rados/thrash-erasure-code-big/{ceph.yaml cluster/{12-osds.yaml openstack.yaml} msgr-failures/osd-delay.yaml objectstore/bluestore-stupid.yaml rados.yaml recovery-overrides/{more-async-partial-recovery.yaml} supported-random-distro$/{centos_7.yaml} thrashers/mapgap.yaml thrashosds-health.yaml workloads/ec-rados-plugin=jerasure-k=4-m=2.yaml} 3
pass 4554085 2019-11-30 02:35:44 2019-11-30 13:56:34 2019-11-30 14:28:33 0:31:59 0:22:57 0:09:02 mira master rados/verify/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-thrash/default/{default.yaml thrashosds-health.yaml} msgr-failures/few.yaml msgr/async-v1only.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml tasks/rados_api_tests.yaml validater/lockdep.yaml} 2
pass 4554086 2019-11-30 02:35:45 2019-11-30 14:28:41 2019-11-30 15:20:41 0:52:00 0:25:19 0:26:41 mira master centos 7.6 rados/thrash-old-clients/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-install/nautilus-v2only.yaml backoff/peering_and_degraded.yaml ceph.yaml clusters/{openstack.yaml three-plus-one.yaml} d-balancer/off.yaml distro$/{centos_latest.yaml} msgr-failures/fastclose.yaml rados.yaml thrashers/mapgap.yaml thrashosds-health.yaml workloads/radosbench.yaml} 4
pass 4554087 2019-11-30 02:35:46 2019-11-30 14:38:55 2019-11-30 15:14:54 0:35:59 0:25:49 0:10:10 mira master ubuntu 18.04 rados/thrash-erasure-code-isa/{arch/x86_64.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/fastclose.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml recovery-overrides/{more-async-recovery.yaml} supported-random-distro$/{ubuntu_latest.yaml} thrashers/default.yaml thrashosds-health.yaml workloads/ec-rados-plugin=isa-k=2-m=1.yaml} 2
pass 4554088 2019-11-30 02:35:47 2019-11-30 15:14:57 2019-11-30 15:46:56 0:31:59 0:19:47 0:12:12 mira master centos 7.6 rados/singleton/{all/erasure-code-nonregression.yaml msgr-failures/many.yaml msgr/async-v2only.yaml objectstore/filestore-xfs.yaml rados.yaml supported-random-distro$/{centos_7.yaml}} 1
pass 4554089 2019-11-30 02:35:47 2019-11-30 15:21:00 2019-11-30 16:00:59 0:39:59 0:29:57 0:10:02 mira master ubuntu 18.04 rados/basic/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/few.yaml msgr/async-v2only.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} tasks/rados_workunit_loadgen_mostlyread.yaml} 2
pass 4554090 2019-11-30 02:35:48 2019-11-30 15:34:47 2019-11-30 18:32:50 2:58:03 2:36:12 0:21:51 mira master rhel 7.7 rados/thrash-erasure-code/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} fast/fast.yaml msgr-failures/few.yaml objectstore/bluestore-avl.yaml rados.yaml recovery-overrides/{more-async-recovery.yaml} supported-random-distro$/{rhel_7.yaml} thrashers/minsize_recovery.yaml thrashosds-health.yaml workloads/ec-small-objects-many-deletes.yaml} 2
pass 4554091 2019-11-30 02:35:49 2019-11-30 15:47:10 2019-11-30 16:29:10 0:42:00 0:31:22 0:10:38 mira master ubuntu 18.04 rados/monthrash/{ceph.yaml clusters/9-mons.yaml msgr-failures/few.yaml msgr/async-v1only.yaml objectstore/filestore-xfs.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} thrashers/one.yaml workloads/snaps-few-objects.yaml} 2
pass 4554092 2019-11-30 02:35:50 2019-11-30 15:59:15 2019-11-30 16:19:14 0:19:59 0:09:19 0:10:40 mira master ubuntu 18.04 rados/singleton-nomsgr/{all/balancer.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}} 1
pass 4554093 2019-11-30 02:35:51 2019-11-30 16:01:18 2019-11-30 17:37:19 1:36:01 1:28:28 0:07:33 mira master rhel 7.7 rados/singleton/{all/lost-unfound-delete.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/bluestore-avl.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}} 1
pass 4554094 2019-11-30 02:35:52 2019-11-30 16:19:16 2019-11-30 16:49:16 0:30:00 0:19:37 0:10:23 mira master ubuntu 18.04 rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml 2-recovery-overrides/{more-async-partial-recovery.yaml} backoff/peering_and_degraded.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/crush-compat.yaml msgr-failures/osd-delay.yaml msgr/async-v1only.yaml objectstore/filestore-xfs.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} thrashers/pggrow.yaml thrashosds-health.yaml workloads/set-chunks-read.yaml} 2
pass 4554095 2019-11-30 02:35:53 2019-11-30 16:29:12 2019-11-30 16:55:12 0:26:00 0:12:49 0:13:11 mira master centos 7.6 rados/multimon/{clusters/3.yaml msgr-failures/many.yaml msgr/async.yaml no_pools.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml supported-random-distro$/{centos_7.yaml} tasks/mon_clock_no_skews.yaml} 2
pass 4554096 2019-11-30 02:35:54 2019-11-30 16:49:23 2019-11-30 19:33:25 2:44:02 2:25:32 0:18:30 mira master rhel 7.7 rados/perf/{ceph.yaml objectstore/bluestore-bitmap.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{rhel_7.yaml} workloads/sample_radosbench.yaml} 1
pass 4554097 2019-11-30 02:35:55 2019-11-30 16:55:14 2019-11-30 17:13:14 0:18:00 0:07:37 0:10:23 mira master ubuntu 18.04 rados/mgr/{clusters/{2-node-mgr.yaml} debug/mgr.yaml objectstore/bluestore-low-osd-mem-target.yaml supported-random-distro$/{ubuntu_latest.yaml} tasks/ssh_orchestrator.yaml} 2
pass 4554098 2019-11-30 02:35:56 2019-11-30 17:13:35 2019-11-30 18:27:35 1:14:00 1:05:32 0:08:28 mira master ubuntu 18.04 rados/singleton/{all/lost-unfound.yaml msgr-failures/many.yaml msgr/async-v1only.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}} 1
pass 4554099 2019-11-30 02:35:57 2019-11-30 17:37:21 2019-11-30 18:09:20 0:31:59 0:21:12 0:10:47 mira master ubuntu 18.04 rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml 2-recovery-overrides/{more-partial-recovery.yaml} backoff/normal.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/off.yaml msgr-failures/fastclose.yaml msgr/async-v2only.yaml objectstore/bluestore-avl.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} thrashers/careful.yaml thrashosds-health.yaml workloads/small-objects.yaml} 2
pass 4554100 2019-11-30 02:35:58 2019-11-30 18:09:33 2019-11-30 18:47:32 0:37:59 0:20:35 0:17:24 mira master centos 7.6 rados/thrash-old-clients/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-install/nautilus.yaml backoff/normal.yaml ceph.yaml clusters/{openstack.yaml three-plus-one.yaml} d-balancer/crush-compat.yaml distro$/{centos_latest.yaml} msgr-failures/few.yaml rados.yaml thrashers/morepggrow.yaml thrashosds-health.yaml workloads/rbd_cls.yaml} 4
pass 4554101 2019-11-30 02:35:59 2019-11-30 18:27:53 2019-11-30 18:51:52 0:23:59 0:14:40 0:09:19 mira master ubuntu 18.04 rados/basic/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/many.yaml msgr/async.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} tasks/readwrite.yaml} 2
pass 4554102 2019-11-30 02:35:59 2019-11-30 18:33:08 2019-11-30 19:11:08 0:38:00 0:24:25 0:13:35 mira master centos 7.6 rados/objectstore/{backends/ceph_objectstore_tool.yaml supported-random-distro$/{centos_7.yaml}} 1
pass 4554103 2019-11-30 02:36:00 2019-11-30 18:47:50 2019-11-30 19:15:49 0:27:59 0:20:11 0:07:48 mira master rhel 7.7 rados/singleton-nomsgr/{all/cache-fs-trunc.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}} 1
pass 4554104 2019-11-30 02:36:01 2019-11-30 18:51:59 2019-11-30 19:27:58 0:35:59 0:12:49 0:23:10 mira master centos 7.6 rados/singleton/{all/max-pg-per-osd.from-mon.yaml msgr-failures/few.yaml msgr/async-v2only.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{centos_7.yaml}} 1
pass 4554105 2019-11-30 02:36:02 2019-11-30 19:11:26 2019-11-30 19:59:26 0:48:00 0:32:56 0:15:04 mira master centos 7.6 rados/thrash-erasure-code-overwrites/{bluestore-bitmap.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} fast/normal.yaml msgr-failures/osd-delay.yaml rados.yaml recovery-overrides/{more-async-partial-recovery.yaml} supported-random-distro$/{centos_7.yaml} thrashers/pggrow.yaml thrashosds-health.yaml workloads/ec-small-objects-overwrites.yaml} 2
pass 4554106 2019-11-30 02:36:03 2019-11-30 19:15:53 2019-11-30 19:39:53 0:24:00 0:12:55 0:11:05 mira master ubuntu 18.04 rados/thrash-erasure-code-shec/{ceph.yaml clusters/{fixed-4.yaml openstack.yaml} msgr-failures/osd-delay.yaml objectstore/filestore-xfs.yaml rados.yaml recovery-overrides/{more-async-recovery.yaml} supported-random-distro$/{ubuntu_latest.yaml} thrashers/careful.yaml thrashosds-health.yaml workloads/ec-rados-plugin=shec-k=4-m=3-c=2.yaml} 4
pass 4554107 2019-11-30 02:36:04 2019-11-30 19:28:01 2019-11-30 20:06:01 0:38:00 0:29:15 0:08:45 mira master rhel 7.7 rados/perf/{ceph.yaml objectstore/bluestore-bitmap.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{rhel_7.yaml} workloads/cosbench_64K_read_write.yaml} 1
pass 4554108 2019-11-30 02:36:05 2019-11-30 19:33:44 2019-11-30 19:55:43 0:21:59 0:11:32 0:10:27 mira master ubuntu 18.04 rados/singleton/{all/max-pg-per-osd.from-primary.yaml msgr-failures/many.yaml msgr/async.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}} 1
pass 4554109 2019-11-30 02:36:06 2019-11-30 19:39:57 2019-11-30 20:33:56 0:53:59 0:48:07 0:05:52 mira master rhel 7.7 rados/standalone/{supported-random-distro$/{rhel_7.yaml} workloads/erasure-code.yaml} 1
pass 4554110 2019-11-30 02:36:06 2019-11-30 19:56:01 2019-11-30 20:32:01 0:36:00 0:24:59 0:11:01 mira master ubuntu 18.04 rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml 2-recovery-overrides/{more-async-recovery.yaml} backoff/peering.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/upmap.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} thrashers/default.yaml thrashosds-health.yaml workloads/snaps-few-objects.yaml} 2
pass 4554111 2019-11-30 02:36:07 2019-11-30 19:59:43 2019-11-30 20:33:43 0:34:00 0:24:55 0:09:05 mira master rhel 7.7 rados/thrash-erasure-code-big/{ceph.yaml cluster/{12-osds.yaml openstack.yaml} msgr-failures/fastclose.yaml objectstore/filestore-xfs.yaml rados.yaml recovery-overrides/{more-partial-recovery.yaml} supported-random-distro$/{rhel_7.yaml} thrashers/morepggrow.yaml thrashosds-health.yaml workloads/ec-rados-plugin=lrc-k=4-m=2-l=3.yaml} 3
pass 4554112 2019-11-30 02:36:08 2019-11-30 20:06:15 2019-11-30 21:02:15 0:56:00 0:43:03 0:12:57 mira master centos rados/verify/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-thrash/none.yaml msgr-failures/few.yaml msgr/async-v2only.yaml objectstore/bluestore-stupid.yaml rados.yaml tasks/rados_cls_all.yaml validater/valgrind.yaml} 2
pass 4554113 2019-11-30 02:36:09 2019-11-30 20:32:04 2019-11-30 21:08:03 0:35:59 0:25:46 0:10:13 mira master ubuntu 18.04 rados/thrash-erasure-code/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} fast/normal.yaml msgr-failures/osd-delay.yaml objectstore/bluestore-bitmap.yaml rados.yaml recovery-overrides/{more-partial-recovery.yaml} supported-random-distro$/{ubuntu_latest.yaml} thrashers/morepggrow.yaml thrashosds-health.yaml workloads/ec-small-objects.yaml} 2
pass 4554114 2019-11-30 02:36:10 2019-11-30 20:34:01 2019-11-30 21:16:00 0:41:59 0:31:09 0:10:50 mira master ubuntu 18.04 rados/thrash-erasure-code-isa/{arch/x86_64.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/few.yaml objectstore/bluestore-stupid.yaml rados.yaml recovery-overrides/{more-async-partial-recovery.yaml} supported-random-distro$/{ubuntu_latest.yaml} thrashers/mapgap.yaml thrashosds-health.yaml workloads/ec-rados-plugin=isa-k=2-m=1.yaml} 2
pass 4554115 2019-11-30 02:36:11 2019-11-30 20:34:00 2019-11-30 20:53:59 0:19:59 0:09:01 0:10:58 mira master ubuntu 18.04 rados/mgr/{clusters/{2-node-mgr.yaml} debug/mgr.yaml objectstore/bluestore-stupid.yaml supported-random-distro$/{ubuntu_latest.yaml} tasks/workunits.yaml} 2
fail 4554116 2019-11-30 02:36:12 2019-11-30 20:54:08 2019-11-30 23:26:10 2:32:02 2:13:57 0:18:05 mira master rhel 7.7 rados/dashboard/{clusters/{2-node-mgr.yaml} debug/mgr.yaml objectstore/bluestore-bitmap.yaml supported-random-distro$/{rhel_7.yaml} tasks/dashboard.yaml} 2
Failure Reason:

k --zap-all /dev/sdd', u'removes': None, u'argv': None, u'creates': None, u'chdir': None, u'stdin_add_newline': True, u'stdin': None}}, u'start': u'2019-11-30 23:25:39.339835'}, {'stderr_lines': [], u'changed': True, u'stdout': u'Creating new GPT entries.\nGPT data structures destroyed! You may now partition the disk using fdisk or\nother utilities.', u'delta': u'0:00:01.011542', 'stdout_lines': [u'Creating new GPT entries.', u'GPT data structures destroyed! You may now partition the disk using fdisk or', u'other utilities.'], '_ansible_item_label': {'value': {u'sectorsize': u'512', u'vendor': u'Hitachi', u'links': {u'masters': [], u'labels': [], u'ids': [u'scsi-2001b4d2011a85600'], u'uuids': []}, u'sas_device_handle': None, u'host': u'RAID bus controller: Areca Technology Corp. ARC-1680 series PCIe to SAS/SATA 3Gb RAID Controller', u'support_discard': u'0', u'serial': u'JPW9K0N211XZVE', u'holders': [], u'size': u'931.51 GB', u'scheduler_mode': u'deadline', u'rotational': u'1', u'sectors': u'1953525168', u'sas_address': None, u'virtual': 1, u'removable': u'0', u'model': u'HUA722010CLA330', u'partitions': {}}, 'key': u'sde'}, 'ansible_loop_var': u'item', u'end': u'2019-11-30 23:25:41.646228', '_ansible_no_log': False, 'item': {'value': {u'sectorsize': u'512', u'vendor': u'Hitachi', u'links': {u'masters': [], u'labels': [], u'ids': [u'scsi-2001b4d2011a85600'], u'uuids': []}, u'sas_device_handle': None, u'host': u'RAID bus controller: Areca Technology Corp. ARC-1680 series PCIe to SAS/SATA 3Gb RAID Controller', u'support_discard': u'0', u'serial': u'JPW9K0N211XZVE', u'holders': [], u'size': u'931.51 GB', u'scheduler_mode': u'deadline', u'rotational': u'1', u'sectors': u'1953525168', u'sas_address': None, u'virtual': 1, u'removable': u'0', u'model': u'HUA722010CLA330', u'partitions': {}}, 'key': u'sde'}, u'cmd': u'sgdisk --zap-all /dev/sde || sgdisk --zap-all /dev/sde', 'failed': False, u'stderr': u'', u'rc': 0, u'invocation': {u'module_args': {u'warn': True, u'executable': None, u'_uses_shell': True, u'strip_empty_ends': True, u'_raw_params': u'sgdisk --zap-all /dev/sde || sgdisk --zap-all /dev/sde', u'removes': None, u'argv': None, u'creates': None, u'chdir': None, u'stdin_add_newline': True, u'stdin': None}}, u'start': u'2019-11-30 23:25:40.634686'}, {'stderr_lines': [], u'changed': True, u'stdout': u'Creating new GPT entries.\nGPT data structures destroyed! You may now partition the disk using fdisk or\nother utilities.', u'delta': u'0:00:01.011324', 'stdout_lines': [u'Creating new GPT entries.', u'GPT data structures destroyed! You may now partition the disk using fdisk or', u'other utilities.'], '_ansible_item_label': {'value': {u'sectorsize': u'512', u'vendor': u'Hitachi', u'links': {u'masters': [], u'labels': [], u'ids': [u'scsi-2001b4d2040775100'], u'uuids': [u'014617a6-3f01-4701-b086-35b12b1828e7']}, u'sas_device_handle': None, u'host': u'RAID bus controller: Areca Technology Corp. ARC-1680 series PCIe to SAS/SATA 3Gb RAID Controller', u'support_discard': u'0', u'serial': u'JPW9K0N204WG1E', u'holders': [], u'size': u'931.51 GB', u'scheduler_mode': u'deadline', u'rotational': u'1', u'sectors': u'1953525168', u'sas_address': None, u'virtual': 1, u'removable': u'0', u'model': u'HUA722010CLA330', u'partitions': {}}, 'key': u'sdf'}, 'ansible_loop_var': u'item', u'end': u'2019-11-30 23:25:42.920294', '_ansible_no_log': False, 'item': {'value': {u'sectorsize': u'512', u'vendor': u'Hitachi', u'links': {u'masters': [], u'labels': [], u'ids': [u'scsi-2001b4d2040775100'], u'uuids': [u'014617a6-3f01-4701-b086-35b12b1828e7']}, u'sas_device_handle': None, u'host': u'RAID bus controller: Areca Technology Corp. ARC-1680 series PCIe to SAS/SATA 3Gb RAID Controller', u'support_discard': u'0', u'serial': u'JPW9K0N204WG1E', u'holders': [], u'size': u'931.51 GB', u'scheduler_mode': u'deadline', u'rotational': u'1', u'sectors': u'1953525168', u'sas_address': None, u'virtual': 1, u'removable': u'0', u'model': u'HUA722010CLA330', u'partitions': {}}, 'key': u'sdf'}, u'cmd': u'sgdisk --zap-all /dev/sdf || sgdisk --zap-all /dev/sdf', 'failed': False, u'stderr': u'', u'rc': 0, u'invocation': {u'module_args': {u'warn': True, u'executable': None, u'_uses_shell': True, u'strip_empty_ends': True, u'_raw_params': u'sgdisk --zap-all /dev/sdf || sgdisk --zap-all /dev/sdf', u'removes': None, u'argv': None, u'creates': None, u'chdir': None, u'stdin_add_newline': True, u'stdin': None}}, u'start': u'2019-11-30 23:25:41.908970'}, {'ansible_loop_var': u'item', '_ansible_no_log': False, 'skip_reason': u'Conditional result was False', 'item': {'value': {u'sectorsize': u'512', u'vendor': u'Seagate', u'links': {u'masters': [], u'labels': [], u'ids': [u'scsi-2001b4d2000000000'], u'uuids': []}, u'sas_device_handle': None, u'host': u'RAID bus controller: Areca Technology Corp. ARC-1680 series PCIe to SAS/SATA 3Gb RAID Controller', u'support_discard': u'0', u'serial': u'5VP66QW9', u'holders': [], u'size': u'931.51 GB', u'scheduler_mode': u'deadline', u'rotational': u'1', u'sectors': u'1953525168', u'sas_address': None, u'virtual': 1, u'removable': u'0', u'model': u'ST31000524AS', u'partitions': {u'sda1': {u'start': u'2048', u'sectorsize': 512, u'uuid': u'f1bca609-9cd3-46ed-8f34-af0538e7246e', u'sectors': u'1953522688', u'holders': [], u'links': {u'masters': [], u'labels': [], u'ids': [u'scsi-2001b4d2000000000-part1'], u'uuids': [u'f1bca609-9cd3-46ed-8f34-af0538e7246e']}, u'size': u'931.51 GB'}}}, 'key': u'sda'}, 'skipped': True, 'changed': False, '_ansible_item_label': {'value': {u'sectorsize': u'512', u'vendor': u'Seagate', u'links': {u'masters': [], u'labels': [], u'ids': [u'scsi-2001b4d2000000000'], u'uuids': []}, u'sas_device_handle': None, u'host': u'RAID bus controller: Areca Technology Corp. ARC-1680 series PCIe to SAS/SATA 3Gb RAID Controller', u'support_discard': u'0', u'serial': u'5VP66QW9', u'holders': [], u'size': u'931.51 GB', u'scheduler_mode': u'deadline', u'rotational': u'1', u'sectors': u'1953525168', u'sas_address': None, u'virtual': 1, u'removable': u'0', u'model': u'ST31000524AS', u'partitions': {u'sda1': {u'start': u'2048', u'sectorsize': 512, u'uuid': u'f1bca609-9cd3-46ed-8f34-af0538e7246e', u'sectors': u'1953522688', u'holders': [], u'links': {u'masters': [], u'labels': [], u'ids': [u'scsi-2001b4d2000000000-part1'], u'uuids': [u'f1bca609-9cd3-46ed-8f34-af0538e7246e']}, u'size': u'931.51 GB'}}}, 'key': u'sda'}}, {'stderr_lines': [], u'changed': True, u'stdout': u'Creating new GPT entries.\nGPT data structures destroyed! You may now partition the disk using fdisk or\nother utilities.', u'delta': u'0:00:01.030875', 'stdout_lines': [u'Creating new GPT entries.', u'GPT data structures destroyed! You may now partition the disk using fdisk or', u'other utilities.'], '_ansible_item_label': {'value': {u'sectorsize': u'512', u'vendor': u'Seagate', u'links': {u'masters': [u'dm-0'], u'labels': [], u'ids': [], u'uuids': []}, u'sas_device_handle': None, u'host': u'RAID bus controller: Areca Technology Corp. ARC-1680 series PCIe to SAS/SATA 3Gb RAID Controller', u'support_discard': u'0', u'serial': u'5VP8NWLD', u'holders': [u'mpatha'], u'size': u'931.51 GB', u'scheduler_mode': u'deadline', u'rotational': u'1', u'sectors': u'1953525168', u'sas_address': None, u'virtual': 1, u'removable': u'0', u'model': u'ST31000528AS', u'partitions': {}}, 'key': u'sdb'}, 'ansible_loop_var': u'item', u'end': u'2019-11-30 23:25:44.226465', '_ansible_no_log': False, 'item': {'value': {u'sectorsize': u'512', u'vendor': u'Seagate', u'links': {u'masters': [u'dm-0'], u'labels': [], u'ids': [], u'uuids': []}, u'sas_device_handle': None, u'host': u'RAID bus controller: Areca Technology Corp. ARC-1680 series PCIe to SAS/SATA 3Gb RAID Controller', u'support_discard': u'0', u'serial': u'5VP8NWLD', u'holders': [u'mpatha'], u'size': u'931.51 GB', u'scheduler_mode': u'deadline', u'rotational': u'1', u'sectors': u'1953525168', u'sas_address': None, u'virtual': 1, u'removable': u'0', u'model': u'ST31000528AS', u'partitions': {}}, 'key': u'sdb'}, u'cmd': u'sgdisk --zap-all /dev/sdb || sgdisk --zap-all /dev/sdb', 'failed': False, u'stderr': u'', u'rc': 0, u'invocation': {u'module_args': {u'warn': True, u'executable': None, u'_uses_shell': True, u'strip_empty_ends': True, u'_raw_params': u'sgdisk --zap-all /dev/sdb || sgdisk --zap-all /dev/sdb', u'removes': None, u'argv': None, u'creates': None, u'chdir': None, u'stdin_add_newline': True, u'stdin': None}}, u'start': u'2019-11-30 23:25:43.195590'}, {'stderr_lines': [], u'changed': True, u'stdout': u'Creating new GPT entries.\nGPT data structures destroyed! You may now partition the disk using fdisk or\nother utilities.', u'delta': u'0:00:01.029144', 'stdout_lines': [u'Creating new GPT entries.', u'GPT data structures destroyed! You may now partition the disk using fdisk or', u'other utilities.'], '_ansible_item_label': {'value': {u'sectorsize': u'512', u'vendor': u'Seagate', u'links': {u'masters': [u'dm-0'], u'labels': [], u'ids': [], u'uuids': []}, u'sas_device_handle': None, u'host': u'RAID bus controller: Areca Technology Corp. ARC-1680 series PCIe to SAS/SATA 3Gb RAID Controller', u'support_discard': u'0', u'serial': u'6VPBDH90', u'holders': [u'mpatha'], u'size': u'931.51 GB', u'scheduler_mode': u'deadline', u'rotational': u'1', u'sectors': u'1953525168', u'sas_address': None, u'virtual': 1, u'removable': u'0', u'model': u'ST31000528AS', u'partitions': {}}, 'key': u'sdc'}, 'ansible_loop_var': u'item', u'end': u'2019-11-30 23:25:45.506753', '_ansible_no_log': False, 'item': {'value': {u'sectorsize': u'512', u'vendor': u'Seagate', u'links': {u'masters': [u'dm-0'], u'labels': [], u'ids': [], u'uuids': []}, u'sas_device_handle': None, u'host': u'RAID bus controller: Areca Technology Corp. ARC-1680 series PCIe to SAS/SATA 3Gb RAID Controller', u'support_discard': u'0', u'serial': u'6VPBDH90', u'holders': [u'mpatha'], u'size': u'931.51 GB', u'scheduler_mode': u'deadline', u'rotational': u'1', u'sectors': u'1953525168', u'sas_address': None, u'virtual': 1, u'removable': u'0', u'model': u'ST31000528AS', u'partitions': {}}, 'key': u'sdc'}, u'cmd': u'sgdisk --zap-all /dev/sdc || sgdisk --zap-all /dev/sdc', 'failed': False, u'stderr': u'', u'rc': 0, u'invocation': {u'module_args': {u'warn': True, u'executable': None, u'_uses_shell': True, u'strip_empty_ends': True, u'_raw_params': u'sgdisk --zap-all /dev/sdc || sgdisk --zap-all /dev/sdc', u'removes': None, u'argv': None, u'creates': None, u'chdir': None, u'stdin_add_newline': True, u'stdin': None}}, u'start': u'2019-11-30 23:25:44.477609'}, {'stderr_lines': [u'Problem opening /dev/dm-0 for reading! Error is 2.', u'The specified file does not exist!', u"Problem opening '' for writing! Program will now terminate.", u'Warning! MBR not overwritten! Error is 2!', u'Problem opening /dev/dm-0 for reading! Error is 2.', u'The specified file does not exist!', u"Problem opening '' for writing! Program will now terminate.", u'Warning! MBR not overwritten! Error is 2!'], u'changed': True, u'stdout': u'', u'invocation': {u'module_args': {u'warn': True, u'executable': None, u'_uses_shell': True, u'strip_empty_ends': True, u'_raw_params': u'sgdisk --zap-all /dev/dm-0 || sgdisk --zap-all /dev/dm-0', u'removes': None, u'argv': None, u'creates': None, u'chdir': None, u'stdin_add_newline': True, u'stdin': None}}, u'delta': u'0:00:00.008523', 'stdout_lines': [], '_ansible_item_label': {'value': {u'sectorsize': u'512', u'vendor': None, u'links': {u'masters': [], u'labels': [], u'ids': [u'dm-name-mpatha', u'dm-uuid-mpath-2001b4d2000000000'], u'uuids': []}, u'sas_device_handle': None, u'host': u'', u'support_discard': u'0', u'serial': u'5VP8NWLD', u'holders': [], u'size': u'931.51 GB', u'scheduler_mode': u'deadline', u'rotational': u'1', u'sectors': u'1953525168', u'sas_address': None, u'virtual': 1, u'removable': u'0', u'model': None, u'partitions': {}}, 'key': u'dm-0'}, 'ansible_loop_var': u'item', u'end': u'2019-11-30 23:25:45.768627', '_ansible_no_log': False, u'start': u'2019-11-30 23:25:45.760104', u'failed': True, u'cmd': u'sgdisk --zap-all /dev/dm-0 || sgdisk --zap-all /dev/dm-0', 'item': {'value': {u'sectorsize': u'512', u'vendor': None, u'links': {u'masters': [], u'labels': [], u'ids': [u'dm-name-mpatha', u'dm-uuid-mpath-2001b4d2000000000'], u'uuids': []}, u'sas_device_handle': None, u'host': u'', u'support_discard': u'0', u'serial': u'5VP8NWLD', u'holders': [], u'size': u'931.51 GB', u'scheduler_mode': u'deadline', u'rotational': u'1', u'sectors': u'1953525168', u'sas_address': None, u'virtual': 1, u'removable': u'0', u'model': None, u'partitions': {}}, 'key': u'dm-0'}, u'stderr': u"Problem opening /dev/dm-0 for reading! Error is 2.\nThe specified file does not exist!\nProblem opening '' for writing! Program will now terminate.\nWarning! MBR not overwritten! Error is 2!\nProblem opening /dev/dm-0 for reading! Error is 2.\nThe specified file does not exist!\nProblem opening '' for writing! Program will now terminate.\nWarning! MBR not overwritten! Error is 2!", u'rc': 2, u'msg': u'non-zero return code'}]}}Traceback (most recent call last): File "/home/teuthworker/src/git.ceph.com_git_ceph-cm-ansible_master/callback_plugins/failure_log.py", line 44, in log_failure log.error(yaml.safe_dump(failure)) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/__init__.py", line 309, in safe_dump return dump_all([data], stream, Dumper=SafeDumper, **kwds) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/__init__.py", line 281, in dump_all dumper.represent(data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 29, in represent node = self.represent_data(data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 58, in represent_data node = self.yaml_representers[data_types[0]](self, data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 227, in represent_dict return self.represent_mapping(u'tag:yaml.org,2002:map', data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 125, in represent_mapping node_value = self.represent_data(item_value) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 58, in represent_data node = self.yaml_representers[data_types[0]](self, data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 227, in represent_dict return self.represent_mapping(u'tag:yaml.org,2002:map', data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 125, in represent_mapping node_value = self.represent_data(item_value) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 58, in represent_data node = self.yaml_representers[data_types[0]](self, data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 219, in represent_list return self.represent_sequence(u'tag:yaml.org,2002:seq', data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 102, in represent_sequence node_item = self.represent_data(item) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 58, in represent_data node = self.yaml_representers[data_types[0]](self, data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 227, in represent_dict return self.represent_mapping(u'tag:yaml.org,2002:map', data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 125, in represent_mapping node_value = self.represent_data(item_value) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 58, in represent_data node = self.yaml_representers[data_types[0]](self, data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 227, in represent_dict return self.represent_mapping(u'tag:yaml.org,2002:map', data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 125, in represent_mapping node_value = self.represent_data(item_value) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 68, in represent_data node = self.yaml_representers[None](self, data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 251, in represent_undefined raise RepresenterError("cannot represent an object", data)RepresenterError: ('cannot represent an object', u'sdd')

pass 4554117 2019-11-30 02:36:13 2019-11-30 21:02:18 2019-11-30 23:56:20 2:54:02 2:34:24 0:19:38 mira master rhel 7.7 rados/singleton-nomsgr/{all/ceph-daemon.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}} 1
fail 4554118 2019-11-30 02:36:14 2019-11-30 21:08:21 2019-11-30 21:26:20 0:17:59 0:07:28 0:10:31 mira master ubuntu 18.04 rados/ssh/{fixed-2.yaml msgr/async-v2only.yaml start.yaml supported-random-distro$/{ubuntu_latest.yaml} tasks/rados_python.yaml} 2
Failure Reason:

Command failed on mira046 with status 22: "sudo /home/ubuntu/cephtest/ceph-daemon --image quay.io/ceph-ci/ceph:378993f279ac979885d08442ca085f5699aa9c1b shell -c /home/ubuntu/cephtest/ceph.conf -k /home/ubuntu/cephtest/ceph.keyring --fsid 6f97fd30-13b7-11ea-8261-001a4aab830c -- ceph orchestrator mon update 2 'mira046:[v2:172.21.5.120:3301,v1:172.21.5.120:6790]=c'"

pass 4554119 2019-11-30 02:36:14 2019-11-30 21:16:18 2019-11-30 21:38:17 0:21:59 0:12:05 0:09:54 mira master ubuntu 18.04 rados/singleton/{all/max-pg-per-osd.from-replica.yaml msgr-failures/few.yaml msgr/async-v1only.yaml objectstore/bluestore-stupid.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}} 1
pass 4554120 2019-11-30 02:36:15 2019-11-30 21:26:23 2019-12-01 00:18:25 2:52:02 2:33:13 0:18:49 mira master rhel 7.7 rados/perf/{ceph.yaml objectstore/bluestore-comp.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{rhel_7.yaml} workloads/cosbench_64K_write.yaml} 1
pass 4554121 2019-11-30 02:36:16 2019-11-30 21:38:27 2019-11-30 22:18:26 0:39:59 0:26:19 0:13:40 mira master centos 7.6 rados/basic/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/few.yaml msgr/async-v1only.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml supported-random-distro$/{centos_7.yaml} tasks/repair_test.yaml} 2
pass 4554122 2019-11-30 02:36:17 2019-11-30 22:18:45 2019-11-30 22:54:44 0:35:59 0:26:45 0:09:14 mira master rhel 7.7 rados/monthrash/{ceph.yaml clusters/3-mons.yaml msgr-failures/mon-delay.yaml msgr/async-v2only.yaml objectstore/filestore-xfs.yaml rados.yaml supported-random-distro$/{rhel_7.yaml} thrashers/sync-many.yaml workloads/pool-create-delete.yaml} 2
pass 4554123 2019-11-30 02:36:18 2019-11-30 22:54:47 2019-12-01 00:06:47 1:12:00 0:45:51 0:26:09 mira master centos 7.6 rados/thrash-old-clients/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-install/hammer.yaml backoff/peering.yaml ceph.yaml clusters/{openstack.yaml three-plus-one.yaml} d-balancer/off.yaml distro$/{centos_latest.yaml} msgr-failures/osd-delay.yaml rados.yaml thrashers/none.yaml thrashosds-health.yaml workloads/snaps-few-objects.yaml} 4
pass 4554124 2019-11-30 02:36:19 2019-11-30 23:26:28 2019-12-01 00:00:27 0:33:59 0:22:41 0:11:18 mira master ubuntu 18.04 rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml 2-recovery-overrides/{default.yaml} backoff/peering_and_degraded.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/crush-compat.yaml msgr-failures/osd-delay.yaml msgr/async-v1only.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} thrashers/mapgap.yaml thrashosds-health.yaml workloads/write_fadvise_dontneed.yaml} 2
pass 4554125 2019-11-30 02:36:20 2019-11-30 23:56:38 2019-12-01 00:14:37 0:17:59 0:09:09 0:08:50 mira master ubuntu 18.04 rados/singleton/{all/mon-auth-caps.yaml msgr-failures/many.yaml msgr/async-v2only.yaml objectstore/filestore-xfs.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}} 1
pass 4554126 2019-11-30 02:36:21 2019-12-01 00:00:29 2019-12-01 00:26:28 0:25:59 0:11:49 0:14:10 mira master centos 7.6 rados/multimon/{clusters/6.yaml msgr-failures/few.yaml msgr/async-v1only.yaml no_pools.yaml objectstore/bluestore-stupid.yaml rados.yaml supported-random-distro$/{centos_7.yaml} tasks/mon_clock_with_skews.yaml} 2
pass 4554127 2019-11-30 02:36:21 2019-12-01 00:06:57 2019-12-01 00:22:56 0:15:59 0:06:57 0:09:02 mira master ubuntu 18.04 rados/objectstore/{backends/filejournal.yaml supported-random-distro$/{ubuntu_latest.yaml}} 1
pass 4554128 2019-11-30 02:36:22 2019-12-01 00:14:57 2019-12-01 00:32:56 0:17:59 0:08:17 0:09:42 mira master ubuntu 18.04 rados/singleton-nomsgr/{all/ceph-kvstore-tool.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}} 1
pass 4554129 2019-11-30 02:36:23 2019-12-01 00:18:42 2019-12-01 00:50:41 0:31:59 0:24:05 0:07:54 mira master rhel 7.7 rados/singleton/{all/mon-config-key-caps.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/bluestore-avl.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}} 1
dead 4554130 2019-11-30 02:36:24 2019-12-01 00:22:58 2019-12-01 12:25:22 12:02:24 mira master centos 7.6 rados/perf/{ceph.yaml objectstore/bluestore-low-osd-mem-target.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{centos_7.yaml} workloads/fio_4K_rand_read.yaml} 1
pass 4554131 2019-11-30 02:36:25 2019-12-01 00:26:39 2019-12-01 00:52:38 0:25:59 0:14:58 0:11:01 mira master ubuntu 18.04 rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml 2-recovery-overrides/{more-active-recovery.yaml} backoff/normal.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/off.yaml msgr-failures/fastclose.yaml msgr/async-v2only.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} thrashers/morepggrow.yaml thrashosds-health.yaml workloads/admin_socket_objecter_requests.yaml} 2
pass 4554132 2019-11-30 02:36:26 2019-12-01 00:33:01 2019-12-01 01:21:01 0:48:00 0:37:55 0:10:05 mira master ubuntu 18.04 rados/singleton-bluestore/{all/cephtool.yaml msgr-failures/few.yaml msgr/async-v1only.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}} 1
pass 4554133 2019-11-30 02:36:27 2019-12-01 00:51:01 2019-12-01 01:13:00 0:21:59 0:13:14 0:08:45 mira master ubuntu 18.04 rados/singleton/{all/mon-config-keys.yaml msgr-failures/many.yaml msgr/async-v1only.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}} 1
pass 4554134 2019-11-30 02:36:27 2019-12-01 00:52:59 2019-12-01 03:37:01 2:44:02 2:24:57 0:19:05 mira master rhel 7.7 rados/mgr/{clusters/{2-node-mgr.yaml} debug/mgr.yaml objectstore/bluestore-stupid.yaml supported-random-distro$/{rhel_7.yaml} tasks/crash.yaml} 2
pass 4554135 2019-11-30 02:36:28 2019-12-01 01:03:01 2019-12-01 01:55:01 0:52:00 0:39:16 0:12:44 mira master centos 7.6 rados/thrash-erasure-code-overwrites/{bluestore-bitmap.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} fast/fast.yaml msgr-failures/osd-delay.yaml rados.yaml recovery-overrides/{more-async-partial-recovery.yaml} supported-random-distro$/{centos_7.yaml} thrashers/careful.yaml thrashosds-health.yaml workloads/ec-pool-snaps-few-objects-overwrites.yaml} 2
pass 4554136 2019-11-30 02:36:29 2019-12-01 01:13:19 2019-12-01 01:59:19 0:46:00 0:20:32 0:25:28 mira master centos 7.6 rados/thrash-erasure-code-shec/{ceph.yaml clusters/{fixed-4.yaml openstack.yaml} msgr-failures/osd-delay.yaml objectstore/bluestore-avl.yaml rados.yaml recovery-overrides/{default.yaml} supported-random-distro$/{centos_7.yaml} thrashers/default.yaml thrashosds-health.yaml workloads/ec-rados-plugin=shec-k=4-m=3-c=2.yaml} 4
fail 4554137 2019-11-30 02:36:30 2019-12-01 01:21:19 2019-12-01 01:41:18 0:19:59 0:10:24 0:09:35 mira master ubuntu 18.04 rados/basic/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/many.yaml msgr/async-v2only.yaml objectstore/bluestore-stupid.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} tasks/rgw_snaps.yaml} 2
Failure Reason:

Command failed on mira038 with status 127: 'cd /home/ubuntu/cephtest/s3-tests && ./bootstrap'

pass 4554138 2019-11-30 02:36:31 2019-12-01 01:41:32 2019-12-01 02:33:32 0:52:00 0:26:41 0:25:19 mira master centos 7.6 rados/thrash-old-clients/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-install/jewel-v1only.yaml backoff/peering_and_degraded.yaml ceph.yaml clusters/{openstack.yaml three-plus-one.yaml} d-balancer/crush-compat.yaml distro$/{centos_latest.yaml} msgr-failures/fastclose.yaml rados.yaml thrashers/pggrow.yaml thrashosds-health.yaml workloads/test_rbd_api.yaml} 4
pass 4554139 2019-11-30 02:36:32 2019-12-01 01:55:22 2019-12-01 02:11:21 0:15:59 0:06:53 0:09:06 mira master ubuntu 18.04 rados/singleton-nomsgr/{all/ceph-post-file.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}} 1
pass 4554140 2019-11-30 02:36:33 2019-12-01 01:59:20 2019-12-01 02:33:20 0:34:00 0:22:07 0:11:53 mira master centos 7.6 rados/singleton/{all/mon-config.yaml msgr-failures/few.yaml msgr/async-v2only.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{centos_7.yaml}} 1
pass 4554141 2019-11-30 02:36:34 2019-12-01 02:11:28 2019-12-01 02:45:27 0:33:59 0:22:58 0:11:01 mira master ubuntu 18.04 rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml 2-recovery-overrides/{more-async-recovery.yaml} backoff/peering.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/upmap.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/bluestore-stupid.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} thrashers/none.yaml thrashosds-health.yaml workloads/cache-agent-big.yaml} 2
pass 4554142 2019-11-30 02:36:34 2019-12-01 02:33:22 2019-12-01 02:51:21 0:17:59 0:08:51 0:09:08 mira master ubuntu 18.04 rados/perf/{ceph.yaml objectstore/bluestore-stupid.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{ubuntu_latest.yaml} workloads/fio_4K_rand_rw.yaml} 1
pass 4554143 2019-11-30 02:36:35 2019-12-01 02:33:34 2019-12-01 03:13:33 0:39:59 0:16:51 0:23:08 mira master centos 7.6 rados/standalone/{supported-random-distro$/{centos_7.yaml} workloads/mgr.yaml} 1
pass 4554144 2019-11-30 02:36:36 2019-12-01 02:45:43 2019-12-01 03:33:42 0:47:59 0:41:33 0:06:26 mira master rhel 7.7 rados/thrash-erasure-code/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} fast/fast.yaml msgr-failures/fastclose.yaml objectstore/bluestore-comp.yaml rados.yaml recovery-overrides/{more-async-recovery.yaml} supported-random-distro$/{rhel_7.yaml} thrashers/pggrow.yaml thrashosds-health.yaml workloads/ec-rados-plugin=clay-k=4-m=2.yaml} 2
pass 4554145 2019-11-30 02:36:37 2019-12-01 02:51:34 2019-12-01 05:55:36 3:04:02 2:45:26 0:18:36 mira master rhel 7.7 rados/thrash-erasure-code-big/{ceph.yaml cluster/{12-osds.yaml openstack.yaml} msgr-failures/few.yaml objectstore/bluestore-avl.yaml rados.yaml recovery-overrides/{more-active-recovery.yaml} supported-random-distro$/{rhel_7.yaml} thrashers/pggrow.yaml thrashosds-health.yaml workloads/ec-rados-plugin=jerasure-k=4-m=2.yaml} 3
pass 4554146 2019-11-30 02:36:38 2019-12-01 03:13:43 2019-12-01 03:35:42 0:21:59 0:12:21 0:09:38 mira master rados/verify/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-thrash/default/{default.yaml thrashosds-health.yaml} msgr-failures/few.yaml msgr/async.yaml objectstore/filestore-xfs.yaml rados.yaml tasks/mon_recovery.yaml validater/lockdep.yaml} 2
pass 4554147 2019-11-30 02:36:39 2019-12-01 03:33:49 2019-12-01 04:11:49 0:38:00 0:28:20 0:09:40 mira master ubuntu 18.04 rados/thrash-erasure-code-isa/{arch/x86_64.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/osd-delay.yaml objectstore/filestore-xfs.yaml rados.yaml recovery-overrides/{more-async-partial-recovery.yaml} supported-random-distro$/{ubuntu_latest.yaml} thrashers/morepggrow.yaml thrashosds-health.yaml workloads/ec-rados-plugin=isa-k=2-m=1.yaml} 2
pass 4554148 2019-11-30 02:36:40 2019-12-01 03:36:00 2019-12-01 06:58:02 3:22:02 3:02:49 0:19:13 mira master rhel 7.7 rados/singleton/{all/osd-backfill.yaml msgr-failures/many.yaml msgr/async.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}} 1
pass 4554149 2019-11-30 02:36:41 2019-12-01 03:37:18 2019-12-01 04:05:17 0:27:59 0:16:23 0:11:36 mira master ubuntu 18.04 rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml 2-recovery-overrides/{default.yaml} backoff/peering_and_degraded.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/crush-compat.yaml msgr-failures/osd-delay.yaml msgr/async-v1only.yaml objectstore/filestore-xfs.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} thrashers/pggrow.yaml thrashosds-health.yaml workloads/cache-agent-small.yaml} 2
pass 4554150 2019-11-30 02:36:42 2019-12-01 04:05:34 2019-12-01 04:35:33 0:29:59 0:21:37 0:08:22 mira master rhel 7.7 rados/monthrash/{ceph.yaml clusters/9-mons.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/bluestore-avl.yaml rados.yaml supported-random-distro$/{rhel_7.yaml} thrashers/sync.yaml workloads/rados_5925.yaml} 2
pass 4554151 2019-11-30 02:36:42 2019-12-01 04:12:06 2019-12-01 05:22:06 1:10:00 1:02:50 0:07:10 mira master rhel 7.7 rados/dashboard/{clusters/{2-node-mgr.yaml} debug/mgr.yaml objectstore/bluestore-comp.yaml supported-random-distro$/{rhel_7.yaml} tasks/dashboard.yaml} 2
pass 4554152 2019-11-30 02:36:43 2019-12-01 04:35:53 2019-12-01 05:33:53 0:58:00 0:35:56 0:22:04 mira master centos 7.6 rados/objectstore/{backends/filestore-idempotent-aio-journal.yaml supported-random-distro$/{centos_7.yaml}} 1
pass 4554153 2019-11-30 02:36:44 2019-12-01 05:22:08 2019-12-01 05:46:08 0:24:00 0:12:36 0:11:24 mira master centos 7.6 rados/singleton-nomsgr/{all/export-after-evict.yaml rados.yaml supported-random-distro$/{centos_7.yaml}} 1
fail 4554154 2019-11-30 02:36:45 2019-12-01 05:34:03 2019-12-01 06:02:03 0:28:00 0:13:42 0:14:18 mira master centos 7.6 rados/ssh/{fixed-2.yaml msgr/async.yaml start.yaml supported-random-distro$/{centos_7.yaml} tasks/rados_api_tests.yaml} 2
Failure Reason:

Command failed on mira072 with status 22: "sudo /home/ubuntu/cephtest/ceph-daemon --image quay.io/ceph-ci/ceph:378993f279ac979885d08442ca085f5699aa9c1b shell -c /home/ubuntu/cephtest/ceph.conf -k /home/ubuntu/cephtest/ceph.keyring --fsid 71678a66-13ff-11ea-8261-001a4aab830c -- ceph orchestrator mon update 2 'mira072:[v2:172.21.6.104:3301,v1:172.21.6.104:6790]=c'"

pass 4554155 2019-11-30 02:36:46 2019-12-01 05:46:14 2019-12-01 07:14:14 1:28:00 1:18:00 0:10:00 mira master centos 7.6 rados/singleton/{all/osd-recovery-incomplete.yaml msgr-failures/few.yaml msgr/async-v1only.yaml objectstore/bluestore-stupid.yaml rados.yaml supported-random-distro$/{centos_7.yaml}} 1
pass 4554156 2019-11-30 02:36:47 2019-12-01 05:55:54 2019-12-01 06:13:53 0:17:59 0:09:04 0:08:55 mira master ubuntu 18.04 rados/perf/{ceph.yaml objectstore/bluestore-basic-min-osd-mem-target.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{ubuntu_latest.yaml} workloads/fio_4M_rand_read.yaml} 1
pass 4554157 2019-11-30 02:36:48 2019-12-01 06:02:20 2019-12-01 08:48:22 2:46:02 2:26:43 0:19:19 mira master rhel 7.7 rados/basic/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/few.yaml msgr/async.yaml objectstore/filestore-xfs.yaml rados.yaml supported-random-distro$/{rhel_7.yaml} tasks/scrub_test.yaml} 2
pass 4554158 2019-11-30 02:36:49 2019-12-01 06:14:09 2019-12-01 06:52:09 0:38:00 0:22:19 0:15:41 mira master centos 7.6 rados/multimon/{clusters/9.yaml msgr-failures/many.yaml msgr/async-v2only.yaml no_pools.yaml objectstore/filestore-xfs.yaml rados.yaml supported-random-distro$/{centos_7.yaml} tasks/mon_recovery.yaml} 3
pass 4554159 2019-11-30 02:36:50 2019-12-01 06:52:11 2019-12-01 07:14:11 0:22:00 0:11:49 0:10:11 mira master ubuntu 18.04 rados/mgr/{clusters/{2-node-mgr.yaml} debug/mgr.yaml objectstore/filestore-xfs.yaml supported-random-distro$/{ubuntu_latest.yaml} tasks/failover.yaml} 2
pass 4554160 2019-11-30 02:36:51 2019-12-01 06:58:12 2019-12-01 08:00:12 1:02:00 0:34:05 0:27:55 mira master centos 7.6 rados/thrash-old-clients/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-install/jewel.yaml backoff/normal.yaml ceph.yaml clusters/{openstack.yaml three-plus-one.yaml} d-balancer/off.yaml distro$/{centos_latest.yaml} msgr-failures/few.yaml rados.yaml thrashers/careful.yaml thrashosds-health.yaml workloads/cache-snaps.yaml} 4
pass 4554161 2019-11-30 02:36:51 2019-12-01 07:14:18 2019-12-01 07:48:17 0:33:59 0:24:19 0:09:40 mira master rhel 7.7 rados/singleton/{all/osd-recovery.yaml msgr-failures/many.yaml msgr/async-v2only.yaml objectstore/filestore-xfs.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}} 1
pass 4554162 2019-11-30 02:36:52 2019-12-01 07:14:18 2019-12-01 07:58:18 0:44:00 0:31:08 0:12:52 mira master centos 7.6 rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml 2-recovery-overrides/{default.yaml} backoff/normal.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/off.yaml msgr-failures/fastclose.yaml msgr/async-v2only.yaml objectstore/bluestore-avl.yaml rados.yaml supported-random-distro$/{centos_7.yaml} thrashers/careful.yaml thrashosds-health.yaml workloads/cache-pool-snaps-readproxy.yaml} 2
pass 4554163 2019-11-30 02:36:53 2019-12-01 07:48:35 2019-12-01 08:24:35 0:36:00 0:14:01 0:21:59 mira master centos 7.6 rados/singleton-nomsgr/{all/full-tiering.yaml rados.yaml supported-random-distro$/{centos_7.yaml}} 1
pass 4554164 2019-11-30 02:36:54 2019-12-01 07:58:19 2019-12-01 08:16:19 0:18:00 0:08:21 0:09:39 mira master ubuntu 18.04 rados/singleton/{all/peer.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/bluestore-avl.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}} 1
pass 4554165 2019-11-30 02:36:55 2019-12-01 08:00:30 2019-12-01 08:20:29 0:19:59 0:10:50 0:09:09 mira master ubuntu 18.04 rados/perf/{ceph.yaml objectstore/bluestore-bitmap.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{ubuntu_latest.yaml} workloads/fio_4M_rand_rw.yaml} 1
pass 4554166 2019-11-30 02:36:56 2019-12-01 08:16:29 2019-12-01 08:56:28 0:39:59 0:33:21 0:06:38 mira master rhel 7.7 rados/thrash-erasure-code-overwrites/{bluestore-bitmap.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} fast/normal.yaml msgr-failures/fastclose.yaml rados.yaml recovery-overrides/{more-async-partial-recovery.yaml} supported-random-distro$/{rhel_7.yaml} thrashers/default.yaml thrashosds-health.yaml workloads/ec-small-objects-fast-read-overwrites.yaml} 2
pass 4554167 2019-11-30 02:36:57 2019-12-01 08:20:47 2019-12-01 09:00:47 0:40:00 0:21:22 0:18:38 mira master centos 7.6 rados/thrash-erasure-code-shec/{ceph.yaml clusters/{fixed-4.yaml openstack.yaml} msgr-failures/fastclose.yaml objectstore/bluestore-bitmap.yaml rados.yaml recovery-overrides/{more-async-partial-recovery.yaml} supported-random-distro$/{centos_7.yaml} thrashers/careful.yaml thrashosds-health.yaml workloads/ec-rados-plugin=shec-k=4-m=3-c=2.yaml} 4
pass 4554168 2019-11-30 02:36:58 2019-12-01 08:24:53 2019-12-01 11:26:55 3:02:02 2:43:30 0:18:32 mira master rhel 7.7 rados/thrash-erasure-code/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} fast/normal.yaml msgr-failures/few.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml recovery-overrides/{more-active-recovery.yaml} supported-random-distro$/{rhel_7.yaml} thrashers/careful.yaml thrashosds-health.yaml workloads/ec-rados-plugin=jerasure-k=2-m=1.yaml} 2
pass 4554169 2019-11-30 02:36:59 2019-12-01 08:48:34 2019-12-01 09:36:34 0:48:00 0:29:30 0:18:30 mira master centos 7.6 rados/basic/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/many.yaml msgr/async-v1only.yaml objectstore/filestore-xfs.yaml rados.yaml supported-random-distro$/{centos_7.yaml} tasks/rados_api_tests.yaml} 2
pass 4554170 2019-11-30 02:37:00 2019-12-01 08:56:31 2019-12-01 09:28:30 0:31:59 0:24:16 0:07:43 mira master rhel 7.7 rados/singleton/{all/pg-autoscaler.yaml msgr-failures/many.yaml msgr/async-v1only.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}} 2
pass 4554171 2019-11-30 02:37:01 2019-12-01 09:00:48 2019-12-01 09:52:48 0:52:00 0:43:42 0:08:18 mira master rhel 7.7 rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml 2-recovery-overrides/{more-partial-recovery.yaml} backoff/peering.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/upmap.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{rhel_7.yaml} thrashers/default.yaml thrashosds-health.yaml workloads/cache-pool-snaps.yaml} 2
pass 4554172 2019-11-30 02:37:02 2019-12-01 09:28:32 2019-12-01 10:04:32 0:36:00 0:26:43 0:09:17 mira master ubuntu 18.04 rados/standalone/{supported-random-distro$/{ubuntu_latest.yaml} workloads/misc.yaml} 1
pass 4554173 2019-11-30 02:37:03 2019-12-01 09:36:44 2019-12-01 10:10:44 0:34:00 0:18:06 0:15:54 mira master centos 7.6 rados/thrash-erasure-code-big/{ceph.yaml cluster/{12-osds.yaml openstack.yaml} msgr-failures/few.yaml objectstore/bluestore-avl.yaml rados.yaml recovery-overrides/{more-active-recovery.yaml} supported-random-distro$/{centos_7.yaml} thrashers/careful.yaml thrashosds-health.yaml workloads/ec-rados-plugin=lrc-k=4-m=2-l=3.yaml} 3
pass 4554174 2019-11-30 02:37:04 2019-12-01 09:52:51 2019-12-01 10:40:51 0:48:00 0:34:45 0:13:15 mira master centos rados/verify/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-thrash/none.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/bluestore-avl.yaml rados.yaml tasks/mon_recovery.yaml validater/valgrind.yaml} 2
pass 4554175 2019-11-30 02:37:04 2019-12-01 10:04:34 2019-12-01 10:52:34 0:48:00 0:40:29 0:07:31 mira master rhel 7.7 rados/thrash-erasure-code-isa/{arch/x86_64.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/fastclose.yaml objectstore/bluestore-avl.yaml rados.yaml recovery-overrides/{more-async-partial-recovery.yaml} supported-random-distro$/{rhel_7.yaml} thrashers/none.yaml thrashosds-health.yaml workloads/ec-rados-plugin=isa-k=2-m=1.yaml} 2
pass 4554176 2019-11-30 02:37:05 2019-12-01 10:10:46 2019-12-01 10:42:45 0:31:59 0:17:22 0:14:37 mira master centos 7.6 rados/mgr/{clusters/{2-node-mgr.yaml} debug/mgr.yaml objectstore/bluestore-avl.yaml supported-random-distro$/{centos_7.yaml} tasks/insights.yaml} 2
pass 4554177 2019-11-30 02:37:06 2019-12-01 10:40:56 2019-12-01 11:28:56 0:48:00 0:22:12 0:25:48 mira master centos 7.6 rados/thrash-old-clients/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-install/luminous-v1only.yaml backoff/peering.yaml ceph.yaml clusters/{openstack.yaml three-plus-one.yaml} d-balancer/crush-compat.yaml distro$/{centos_latest.yaml} msgr-failures/osd-delay.yaml rados.yaml thrashers/default.yaml thrashosds-health.yaml workloads/radosbench.yaml} 4
pass 4554178 2019-11-30 02:37:07 2019-12-01 10:42:47 2019-12-01 11:32:47 0:50:00 0:36:31 0:13:29 mira master centos 7.6 rados/objectstore/{backends/filestore-idempotent.yaml supported-random-distro$/{centos_7.yaml}} 1
pass 4554179 2019-11-30 02:37:08 2019-12-01 10:52:51 2019-12-01 11:20:50 0:27:59 0:16:57 0:11:02 mira master centos 7.6 rados/singleton-nomsgr/{all/health-warnings.yaml rados.yaml supported-random-distro$/{centos_7.yaml}} 1
pass 4554180 2019-11-30 02:37:09 2019-12-01 11:20:52 2019-12-01 11:36:51 0:15:59 0:06:51 0:09:08 mira master ubuntu 18.04 rados/singleton/{all/pg-removal-interruption.yaml msgr-failures/few.yaml msgr/async-v2only.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}} 1
pass 4554181 2019-11-30 02:37:10 2019-12-01 11:27:13 2019-12-01 12:03:12 0:35:59 0:14:58 0:21:01 mira master centos 7.6 rados/perf/{ceph.yaml objectstore/bluestore-comp.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{centos_7.yaml} workloads/fio_4M_rand_write.yaml} 1
pass 4554182 2019-11-30 02:37:11 2019-12-01 11:28:57 2019-12-01 12:18:57 0:50:00 0:27:07 0:22:53 mira master centos 7.6 rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml 2-recovery-overrides/{more-partial-recovery.yaml} backoff/peering_and_degraded.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/crush-compat.yaml msgr-failures/osd-delay.yaml msgr/async-v1only.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{centos_7.yaml} thrashers/mapgap.yaml thrashosds-health.yaml workloads/cache-snaps.yaml} 2
pass 4554183 2019-11-30 02:37:12 2019-12-01 11:33:04 2019-12-01 12:17:04 0:44:00 0:25:09 0:18:51 mira master ubuntu 18.04 rados/monthrash/{ceph.yaml clusters/3-mons.yaml msgr-failures/mon-delay.yaml msgr/async-v1only.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} thrashers/force-sync-many.yaml workloads/rados_api_tests.yaml} 2
pass 4554184 2019-11-30 02:37:13 2019-12-01 11:36:57 2019-12-01 12:06:57 0:30:00 0:18:49 0:11:11 mira master centos 7.6 rados/singleton/{all/radostool.yaml msgr-failures/many.yaml msgr/async.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml supported-random-distro$/{centos_7.yaml}} 1
pass 4554185 2019-11-30 02:37:14 2019-12-01 12:03:15 2019-12-01 12:23:14 0:19:59 0:08:34 0:11:25 mira master ubuntu 18.04 rados/multimon/{clusters/9.yaml msgr-failures/few.yaml msgr/async.yaml no_pools.yaml objectstore/bluestore-avl.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} tasks/mon_clock_no_skews.yaml} 3
pass 4554186 2019-11-30 02:37:15 2019-12-01 12:06:59 2019-12-01 12:40:58 0:33:59 0:18:32 0:15:27 mira master centos 7.6 rados/basic/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/few.yaml msgr/async-v2only.yaml objectstore/bluestore-avl.yaml rados.yaml supported-random-distro$/{centos_7.yaml} tasks/rados_cls_all.yaml} 2
pass 4554187 2019-11-30 02:37:16 2019-12-01 12:17:06 2019-12-01 12:47:06 0:30:00 0:16:42 0:13:18 mira master centos 7.6 rados/perf/{ceph.yaml objectstore/bluestore-low-osd-mem-target.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{centos_7.yaml} workloads/radosbench_4K_rand_read.yaml} 1
pass 4554188 2019-11-30 02:37:16 2019-12-01 12:18:59 2019-12-01 13:30:59 1:12:00 0:57:22 0:14:38 mira master centos 7.6 rados/dashboard/{clusters/{2-node-mgr.yaml} debug/mgr.yaml objectstore/bluestore-low-osd-mem-target.yaml supported-random-distro$/{centos_7.yaml} tasks/dashboard.yaml} 2
pass 4554189 2019-11-30 02:37:17 2019-12-01 12:23:16 2019-12-01 13:11:16 0:48:00 0:38:52 0:09:08 mira master ubuntu 18.04 rados/singleton-bluestore/{all/cephtool.yaml msgr-failures/many.yaml msgr/async-v2only.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}} 1
pass 4554190 2019-11-30 02:37:18 2019-12-01 12:25:40 2019-12-01 12:53:39 0:27:59 0:20:30 0:07:29 mira master rhel 7.7 rados/singleton-nomsgr/{all/large-omap-object-warnings.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}} 1
fail 4554191 2019-11-30 02:37:19 2019-12-01 12:41:17 2019-12-01 13:07:16 0:25:59 0:13:38 0:12:21 mira master centos 7.6 rados/ssh/{fixed-2.yaml msgr/async-v1only.yaml start.yaml supported-random-distro$/{centos_7.yaml} tasks/rados_python.yaml} 2
Failure Reason:

Command failed on mira066 with status 22: "sudo /home/ubuntu/cephtest/ceph-daemon --image quay.io/ceph-ci/ceph:378993f279ac979885d08442ca085f5699aa9c1b shell -c /home/ubuntu/cephtest/ceph.conf -k /home/ubuntu/cephtest/ceph.keyring --fsid f2408efe-143a-11ea-8261-001a4aab830c -- ceph orchestrator mon update 2 'mira066:[v1:172.21.6.116:6790]=c'"

pass 4554192 2019-11-30 02:37:20 2019-12-01 12:47:24 2019-12-01 15:05:25 2:18:01 2:00:07 0:17:54 mira master centos 7.6 rados/upgrade/nautilus-x-singleton/{0-cluster/{openstack.yaml start.yaml} 1-install/nautilus.yaml 2-partial-upgrade/firsthalf.yaml 3-thrash/default.yaml 4-workload/{rbd-cls.yaml rbd-import-export.yaml readwrite.yaml snaps-few-objects.yaml} 5-workload/{radosbench.yaml rbd_api.yaml} 6-finish-upgrade.yaml 7-octopus.yaml 8-workload/{rbd-python.yaml rgw-swift.yaml snaps-many-objects.yaml} bluestore-bitmap.yaml supported-random-distro$/{centos_7.yaml} thrashosds-health.yaml} 4
pass 4554193 2019-11-30 02:37:21 2019-12-01 12:53:41 2019-12-01 13:29:41 0:36:00 0:22:46 0:13:14 mira master centos 7.6 rados/singleton/{all/random-eio.yaml msgr-failures/few.yaml msgr/async-v1only.yaml objectstore/bluestore-stupid.yaml rados.yaml supported-random-distro$/{centos_7.yaml}} 2
pass 4554194 2019-11-30 02:37:22 2019-12-01 13:07:18 2019-12-01 13:43:18 0:36:00 0:28:14 0:07:46 mira master rhel 7.7 rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml 2-recovery-overrides/{more-active-recovery.yaml} backoff/normal.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/off.yaml msgr-failures/fastclose.yaml msgr/async-v2only.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml supported-random-distro$/{rhel_7.yaml} thrashers/morepggrow.yaml thrashosds-health.yaml workloads/cache.yaml} 2
pass 4554195 2019-11-30 02:37:23 2019-12-01 13:11:18 2019-12-01 15:55:20 2:44:02 2:25:19 0:18:43 mira master rhel 7.7 rados/singleton/{all/rebuild-mondb.yaml msgr-failures/many.yaml msgr/async-v2only.yaml objectstore/filestore-xfs.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}} 1
pass 4554196 2019-11-30 02:37:24 2019-12-01 13:29:59 2019-12-01 14:23:59 0:54:00 0:29:27 0:24:33 mira master centos 7.6 rados/thrash-old-clients/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-install/luminous.yaml backoff/peering_and_degraded.yaml ceph.yaml clusters/{openstack.yaml three-plus-one.yaml} d-balancer/off.yaml distro$/{centos_latest.yaml} msgr-failures/fastclose.yaml rados.yaml thrashers/mapgap.yaml thrashosds-health.yaml workloads/rbd_cls.yaml} 4
pass 4554197 2019-11-30 02:37:25 2019-12-01 13:31:17 2019-12-01 16:33:20 3:02:03 2:43:20 0:18:43 mira master rhel 7.7 rados/thrash-erasure-code/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} fast/fast.yaml msgr-failures/osd-delay.yaml objectstore/bluestore-stupid.yaml rados.yaml recovery-overrides/{more-active-recovery.yaml} supported-random-distro$/{rhel_7.yaml} thrashers/default.yaml thrashosds-health.yaml workloads/ec-rados-plugin=jerasure-k=3-m=1.yaml} 2
fail 4554198 2019-11-30 02:37:26 2019-12-01 13:43:20 2019-12-01 14:19:20 0:36:00 0:23:20 0:12:40 mira master centos 7.6 rados/mgr/{clusters/{2-node-mgr.yaml} debug/mgr.yaml objectstore/bluestore-bitmap.yaml supported-random-distro$/{centos_7.yaml} tasks/module_selftest.yaml} 2
Failure Reason:

Test failure: test_diskprediction_local (tasks.mgr.test_module_selftest.TestModuleSelftest)

pass 4554199 2019-11-30 02:37:27 2019-12-01 14:19:22 2019-12-01 14:59:22 0:40:00 0:32:42 0:07:18 mira master rhel 7.7 rados/thrash-erasure-code-overwrites/{bluestore-bitmap.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} fast/fast.yaml msgr-failures/few.yaml rados.yaml recovery-overrides/{default.yaml} supported-random-distro$/{rhel_7.yaml} thrashers/fastread.yaml thrashosds-health.yaml workloads/ec-small-objects-overwrites.yaml} 2
pass 4554200 2019-11-30 02:37:27 2019-12-01 14:24:02 2019-12-01 15:00:01 0:35:59 0:25:37 0:10:22 mira master rhel 7.7 rados/thrash-erasure-code-shec/{ceph.yaml clusters/{fixed-4.yaml openstack.yaml} msgr-failures/few.yaml objectstore/bluestore-comp.yaml rados.yaml recovery-overrides/{more-async-partial-recovery.yaml} supported-random-distro$/{rhel_7.yaml} thrashers/default.yaml thrashosds-health.yaml workloads/ec-rados-plugin=shec-k=4-m=3-c=2.yaml} 4
pass 4554201 2019-11-30 02:37:28 2019-12-01 14:59:30 2019-12-01 15:29:29 0:29:59 0:22:22 0:07:37 mira master rhel 7.7 rados/perf/{ceph.yaml objectstore/bluestore-stupid.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{rhel_7.yaml} workloads/radosbench_4K_seq_read.yaml} 1
pass 4554202 2019-11-30 02:37:29 2019-12-01 15:00:03 2019-12-01 15:22:02 0:21:59 0:11:14 0:10:45 mira master ubuntu 18.04 rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml 2-recovery-overrides/{more-partial-recovery.yaml} backoff/peering.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/upmap.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/bluestore-stupid.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} thrashers/none.yaml thrashosds-health.yaml workloads/dedup_tier.yaml} 2
pass 4554203 2019-11-30 02:37:30 2019-12-01 15:05:31 2019-12-01 17:45:32 2:40:01 2:21:36 0:18:25 mira master rhel 7.7 rados/objectstore/{backends/fusestore.yaml supported-random-distro$/{rhel_7.yaml}} 1
pass 4554204 2019-11-30 02:37:31 2019-12-01 15:22:05 2019-12-01 15:46:04 0:23:59 0:14:22 0:09:37 mira master centos 7.6 rados/singleton-nomsgr/{all/lazy_omap_stats_output.yaml rados.yaml supported-random-distro$/{centos_7.yaml}} 1
pass 4554205 2019-11-30 02:37:32 2019-12-01 15:29:32 2019-12-01 16:31:31 1:01:59 0:49:54 0:12:05 mira master centos 7.6 rados/singleton/{all/recovery-preemption.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/bluestore-avl.yaml rados.yaml supported-random-distro$/{centos_7.yaml}} 1
pass 4554206 2019-11-30 02:37:33 2019-12-01 15:46:07 2019-12-01 16:20:06 0:33:59 0:19:55 0:14:04 mira master centos 7.6 rados/basic/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/many.yaml msgr/async.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{centos_7.yaml} tasks/rados_python.yaml} 2
pass 4554207 2019-11-30 02:37:34 2019-12-01 15:55:27 2019-12-01 16:53:27 0:58:00 0:48:46 0:09:14 mira master ubuntu 18.04 rados/standalone/{supported-random-distro$/{ubuntu_latest.yaml} workloads/mon.yaml} 1
pass 4554208 2019-11-30 02:37:34 2019-12-01 16:20:09 2019-12-01 17:10:09 0:50:00 0:35:29 0:14:31 mira master centos 7.6 rados/thrash-erasure-code-big/{ceph.yaml cluster/{12-osds.yaml openstack.yaml} msgr-failures/osd-delay.yaml objectstore/bluestore-bitmap.yaml rados.yaml recovery-overrides/{more-async-partial-recovery.yaml} supported-random-distro$/{centos_7.yaml} thrashers/default.yaml thrashosds-health.yaml workloads/ec-rados-plugin=jerasure-k=4-m=2.yaml} 3
fail 4554209 2019-11-30 02:37:35 2019-12-01 16:31:35 2019-12-01 23:05:42 6:34:07 6:24:19 0:09:48 mira master rados/verify/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-thrash/default/{default.yaml thrashosds-health.yaml} msgr-failures/few.yaml msgr/async-v1only.yaml objectstore/bluestore-bitmap.yaml rados.yaml tasks/rados_api_tests.yaml validater/lockdep.yaml} 2
Failure Reason:

Command failed (workunit test rados/test.sh) on mira110 with status 124: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=378993f279ac979885d08442ca085f5699aa9c1b TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 6h /home/ubuntu/cephtest/clone.client.0/qa/workunits/rados/test.sh'

pass 4554210 2019-11-30 02:37:36 2019-12-01 16:33:41 2019-12-01 17:21:41 0:48:00 0:39:56 0:08:04 mira master rhel 7.7 rados/thrash-erasure-code-isa/{arch/x86_64.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/few.yaml objectstore/bluestore-bitmap.yaml rados.yaml recovery-overrides/{more-partial-recovery.yaml} supported-random-distro$/{rhel_7.yaml} thrashers/pggrow.yaml thrashosds-health.yaml workloads/ec-rados-plugin=isa-k=2-m=1.yaml} 2
pass 4554211 2019-11-30 02:37:37 2019-12-01 16:53:47 2019-12-01 17:33:46 0:39:59 0:17:25 0:22:34 mira master centos 7.6 rados/singleton/{all/resolve_stuck_peering.yaml msgr-failures/many.yaml msgr/async-v1only.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{centos_7.yaml}} 2
pass 4554212 2019-11-30 02:37:38 2019-12-01 17:10:11 2019-12-01 17:48:11 0:38:00 0:27:42 0:10:18 mira master ubuntu 18.04 rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml 2-recovery-overrides/{default.yaml} backoff/peering_and_degraded.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/crush-compat.yaml msgr-failures/osd-delay.yaml msgr/async-v1only.yaml objectstore/filestore-xfs.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} thrashers/pggrow.yaml thrashosds-health.yaml workloads/pool-snaps-few-objects.yaml} 2
pass 4554213 2019-11-30 02:37:39 2019-12-01 17:21:56 2019-12-01 17:57:55 0:35:59 0:15:55 0:20:04 mira master centos 7.6 rados/perf/{ceph.yaml objectstore/bluestore-basic-min-osd-mem-target.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{centos_7.yaml} workloads/radosbench_4M_rand_read.yaml} 1
pass 4554214 2019-11-30 02:37:39 2019-12-01 17:34:05 2019-12-01 18:00:04 0:25:59 0:15:24 0:10:35 mira master centos 7.6 rados/singleton-nomsgr/{all/librados_hello_world.yaml rados.yaml supported-random-distro$/{centos_7.yaml}} 1
pass 4554215 2019-11-30 02:37:40 2019-12-01 17:45:50 2019-12-01 18:05:49 0:19:59 0:09:41 0:10:18 mira master ubuntu 18.04 rados/singleton/{all/test-crash.yaml msgr-failures/few.yaml msgr/async-v2only.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}} 1
pass 4554216 2019-11-30 02:37:41 2019-12-01 17:48:13 2019-12-01 19:32:14 1:44:01 1:30:21 0:13:40 mira master centos 7.6 rados/monthrash/{ceph.yaml clusters/9-mons.yaml msgr-failures/few.yaml msgr/async-v2only.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{centos_7.yaml} thrashers/many.yaml workloads/rados_mon_osdmap_prune.yaml} 2
pass 4554217 2019-11-30 02:37:42 2019-12-01 17:57:58 2019-12-01 18:57:58 1:00:00 0:42:23 0:17:37 mira master centos 7.6 rados/thrash-old-clients/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-install/mimic-v1only.yaml backoff/normal.yaml ceph.yaml clusters/{openstack.yaml three-plus-one.yaml} d-balancer/crush-compat.yaml distro$/{centos_latest.yaml} msgr-failures/few.yaml rados.yaml thrashers/morepggrow.yaml thrashosds-health.yaml workloads/snaps-few-objects.yaml} 4
fail 4554218 2019-11-30 02:37:43 2019-12-01 18:00:10 2019-12-01 18:18:09 0:17:59 0:08:32 0:09:27 mira master ubuntu 18.04 rados/mgr/{clusters/{2-node-mgr.yaml} debug/mgr.yaml objectstore/bluestore-comp.yaml supported-random-distro$/{ubuntu_latest.yaml} tasks/orchestrator_cli.yaml} 2
Failure Reason:

Test failure: test_host_ls (tasks.mgr.test_orchestrator_cli.TestOrchestratorCli)

pass 4554219 2019-11-30 02:37:44 2019-12-01 18:05:52 2019-12-01 18:23:51 0:17:59 0:06:52 0:11:07 mira master ubuntu 18.04 rados/multimon/{clusters/21.yaml msgr-failures/many.yaml msgr/async-v1only.yaml no_pools.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} tasks/mon_clock_with_skews.yaml} 3
pass 4554220 2019-11-30 02:37:45 2019-12-01 18:18:12 2019-12-01 18:48:11 0:29:59 0:19:42 0:10:17 mira master ubuntu 18.04 rados/singleton/{all/test_envlibrados_for_rocksdb.yaml msgr-failures/many.yaml msgr/async.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}} 1
pass 4554221 2019-11-30 02:37:46 2019-12-01 18:24:00 2019-12-01 19:00:00 0:36:00 0:24:36 0:11:24 mira master ubuntu 18.04 rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml 2-recovery-overrides/{more-active-recovery.yaml} backoff/normal.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/off.yaml msgr-failures/fastclose.yaml msgr/async-v2only.yaml objectstore/bluestore-avl.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} thrashers/careful.yaml thrashosds-health.yaml workloads/rados_api_tests.yaml} 2
pass 4554222 2019-11-30 02:37:47 2019-12-01 18:48:37 2019-12-01 21:38:39 2:50:02 2:31:44 0:18:18 mira master rhel 7.7 rados/basic/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/few.yaml msgr/async-v1only.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{rhel_7.yaml} tasks/rados_stress_watch.yaml} 2
pass 4554223 2019-11-30 02:37:48 2019-12-01 18:58:21 2019-12-01 19:28:20 0:29:59 0:15:48 0:14:11 mira master centos 7.6 rados/perf/{ceph.yaml objectstore/bluestore-bitmap.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{centos_7.yaml} workloads/radosbench_4M_seq_read.yaml} 1
pass 4554224 2019-11-30 02:37:49 2019-12-01 19:00:13 2019-12-01 20:32:13 1:32:00 1:23:41 0:08:19 mira master rhel 7.7 rados/thrash-erasure-code/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} fast/normal.yaml msgr-failures/fastclose.yaml objectstore/filestore-xfs.yaml rados.yaml recovery-overrides/{more-partial-recovery.yaml} supported-random-distro$/{rhel_7.yaml} thrashers/fastread.yaml thrashosds-health.yaml workloads/ec-radosbench.yaml} 2
pass 4554225 2019-11-30 02:37:50 2019-12-01 19:28:46 2019-12-01 20:40:45 1:11:59 0:48:50 0:23:09 mira master centos 7.6 rados/singleton/{all/thrash-backfill-full.yaml msgr-failures/few.yaml msgr/async-v1only.yaml objectstore/bluestore-stupid.yaml rados.yaml supported-random-distro$/{centos_7.yaml}} 2
pass 4554226 2019-11-30 02:37:51 2019-12-01 19:32:40 2019-12-01 20:34:39 1:01:59 0:46:24 0:15:35 mira master ubuntu 18.04 rados/dashboard/{clusters/{2-node-mgr.yaml} debug/mgr.yaml objectstore/bluestore-stupid.yaml supported-random-distro$/{ubuntu_latest.yaml} tasks/dashboard.yaml} 2
pass 4554227 2019-11-30 02:37:51 2019-12-01 20:32:15 2019-12-01 20:58:15 0:26:00 0:18:05 0:07:55 mira master rhel 7.7 rados/objectstore/{backends/keyvaluedb.yaml supported-random-distro$/{rhel_7.yaml}} 1
dead 4554228 2019-11-30 02:37:52 2019-12-01 20:34:41 2019-12-01 20:50:40 0:15:59 mira master centos 7.6 rados/singleton-nomsgr/{all/msgr.yaml rados.yaml supported-random-distro$/{centos_7.yaml}}
Failure Reason:

reached maximum tries (60) after waiting for 900 seconds

fail 4554229 2019-11-30 02:37:53 2019-12-01 20:40:52 2019-12-01 23:22:54 2:42:02 2:22:53 0:19:09 mira master rhel 7.7 rados/ssh/{fixed-2.yaml msgr/async-v2only.yaml start.yaml supported-random-distro$/{rhel_7.yaml} tasks/rados_api_tests.yaml} 2
Failure Reason:

Command failed on mira038 with status 22: "sudo /home/ubuntu/cephtest/ceph-daemon --image quay.io/ceph-ci/ceph:378993f279ac979885d08442ca085f5699aa9c1b shell -c /home/ubuntu/cephtest/ceph.conf -k /home/ubuntu/cephtest/ceph.keyring --fsid c57dd9a2-1490-11ea-8261-001a4aab830c -- ceph orchestrator mon update 2 'mira038:[v2:172.21.5.138:3301,v1:172.21.5.138:6790]=c'"

pass 4554230 2019-11-30 02:37:54 2019-12-01 20:50:51 2019-12-01 21:38:51 0:48:00 0:41:05 0:06:55 mira master rhel 7.7 rados/thrash-erasure-code-overwrites/{bluestore-bitmap.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} fast/normal.yaml msgr-failures/osd-delay.yaml rados.yaml recovery-overrides/{more-partial-recovery.yaml} supported-random-distro$/{rhel_7.yaml} thrashers/minsize_recovery.yaml thrashosds-health.yaml workloads/ec-snaps-few-objects-overwrites.yaml} 2
pass 4554231 2019-11-30 02:37:55 2019-12-01 20:58:32 2019-12-01 21:42:31 0:43:59 0:26:11 0:17:48 mira master centos 7.6 rados/thrash-erasure-code-shec/{ceph.yaml clusters/{fixed-4.yaml openstack.yaml} msgr-failures/osd-delay.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml recovery-overrides/{more-async-recovery.yaml} supported-random-distro$/{centos_7.yaml} thrashers/careful.yaml thrashosds-health.yaml workloads/ec-rados-plugin=shec-k=4-m=3-c=2.yaml} 4
pass 4554232 2019-11-30 02:37:56 2019-12-01 21:38:58 2019-12-01 22:24:57 0:45:59 0:35:57 0:10:02 mira master ubuntu 18.04 rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml 2-recovery-overrides/{default.yaml} backoff/peering.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/upmap.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} thrashers/default.yaml thrashosds-health.yaml workloads/radosbench.yaml} 2
pass 4554233 2019-11-30 02:37:57 2019-12-01 21:38:58 2019-12-01 22:24:58 0:46:00 0:37:18 0:08:42 mira master rhel 7.7 rados/singleton/{all/thrash-eio.yaml msgr-failures/many.yaml msgr/async-v2only.yaml objectstore/filestore-xfs.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}} 2
pass 4554234 2019-11-30 02:37:58 2019-12-01 21:42:52 2019-12-01 22:26:51 0:43:59 0:18:48 0:25:11 mira master centos 7.6 rados/thrash-old-clients/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-install/mimic.yaml backoff/peering.yaml ceph.yaml clusters/{openstack.yaml three-plus-one.yaml} d-balancer/off.yaml distro$/{centos_latest.yaml} msgr-failures/osd-delay.yaml rados.yaml thrashers/none.yaml thrashosds-health.yaml workloads/test_rbd_api.yaml} 4
pass 4554235 2019-11-30 02:37:59 2019-12-01 22:25:00 2019-12-01 22:42:59 0:17:59 0:09:16 0:08:43 mira master ubuntu 18.04 rados/perf/{ceph.yaml objectstore/bluestore-comp.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{ubuntu_latest.yaml} workloads/radosbench_4M_write.yaml} 1
pass 4554236 2019-11-30 02:38:00 2019-12-01 22:25:00 2019-12-02 03:23:05 4:58:05 4:37:58 0:20:07 mira master rhel 7.7 rados/standalone/{supported-random-distro$/{rhel_7.yaml} workloads/osd.yaml} 1
pass 4554237 2019-11-30 02:38:01 2019-12-01 22:27:10 2019-12-01 22:49:09 0:21:59 0:11:35 0:10:24 mira master ubuntu 18.04 rados/thrash-erasure-code-big/{ceph.yaml cluster/{12-osds.yaml openstack.yaml} msgr-failures/fastclose.yaml objectstore/bluestore-comp.yaml rados.yaml recovery-overrides/{more-active-recovery.yaml} supported-random-distro$/{ubuntu_latest.yaml} thrashers/fastread.yaml thrashosds-health.yaml workloads/ec-rados-plugin=lrc-k=4-m=2-l=3.yaml} 3
fail 4554238 2019-11-30 02:38:01 2019-12-01 22:43:12 2019-12-01 23:53:12 1:10:00 0:47:11 0:22:49 mira master centos rados/verify/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-thrash/none.yaml msgr-failures/few.yaml msgr/async-v2only.yaml objectstore/bluestore-comp.yaml rados.yaml tasks/rados_cls_all.yaml validater/valgrind.yaml} 2
Failure Reason:

reached maximum tries (50) after waiting for 300 seconds

pass 4554239 2019-11-30 02:38:02 2019-12-01 22:49:30 2019-12-01 23:37:30 0:48:00 0:40:11 0:07:49 mira master rhel 7.7 rados/thrash-erasure-code-isa/{arch/x86_64.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/osd-delay.yaml objectstore/bluestore-bitmap.yaml rados.yaml recovery-overrides/{more-async-recovery.yaml} supported-random-distro$/{rhel_7.yaml} thrashers/careful.yaml thrashosds-health.yaml workloads/ec-rados-plugin=isa-k=2-m=1.yaml} 2
pass 4554240 2019-11-30 02:38:03 2019-12-01 23:05:45 2019-12-01 23:37:44 0:31:59 0:19:48 0:12:11 mira master centos 7.6 rados/mgr/{clusters/{2-node-mgr.yaml} debug/mgr.yaml objectstore/bluestore-low-osd-mem-target.yaml supported-random-distro$/{centos_7.yaml} tasks/progress.yaml} 2
pass 4554241 2019-11-30 02:38:04 2019-12-01 23:23:16 2019-12-02 00:15:15 0:51:59 0:29:28 0:22:31 mira master centos 7.6 rados/singleton/{all/thrash-rados/{thrash-rados.yaml thrashosds-health.yaml} msgr-failures/few.yaml msgr/async.yaml objectstore/bluestore-avl.yaml rados.yaml supported-random-distro$/{centos_7.yaml}} 2
pass 4554242 2019-11-30 02:38:05 2019-12-01 23:37:32 2019-12-02 00:05:32 0:28:00 0:14:58 0:13:02 mira master centos 7.6 rados/basic/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/many.yaml msgr/async-v2only.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml supported-random-distro$/{centos_7.yaml} tasks/rados_striper.yaml} 2