User | Scheduled | Started | Updated | Runtime | Suite | Branch | Machine Type | Revision | Pass | Fail | Dead |
---|---|---|---|---|---|---|---|---|---|---|---|
kchai | 2019-11-05 12:08:55 | 2019-11-05 12:09:15 | 2019-11-06 14:28:11 | 1 day, 2:18:56 | rados | wip-kefu-testing-2019-11-03-1250 | mira | 4bad7f7 | 201 | 13 | 4 |
Status | Job ID | Links | Posted | Started | Updated | Runtime |
Duration |
In Waiting |
Machine | Teuthology Branch | OS Type | OS Version | Description | Nodes |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
pass | 4474302 | 2019-11-05 12:09:11 | 2019-11-05 12:09:13 | 2019-11-05 12:43:13 | 0:34:00 | 0:20:52 | 0:13:08 | mira | master | centos | 7.6 | rados/basic/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/few.yaml msgr/async.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml supported-random-distro$/{centos_7.yaml} tasks/rados_python.yaml} | 2 | |
pass | 4474303 | 2019-11-05 12:09:12 | 2019-11-05 12:09:13 | 2019-11-05 12:51:13 | 0:42:00 | 0:31:29 | 0:10:31 | mira | master | ubuntu | 18.04 | rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml 2-recovery-overrides/{more-active-recovery.yaml} backoff/peering_and_degraded.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/crush-compat.yaml msgr-failures/fastclose.yaml msgr/async-v2only.yaml objectstore/bluestore-stupid.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} thrashers/mapgap.yaml thrashosds-health.yaml workloads/pool-snaps-few-objects.yaml} | 2 | |
pass | 4474304 | 2019-11-05 12:09:13 | 2019-11-05 12:09:14 | 2019-11-05 12:47:14 | 0:38:00 | 0:29:10 | 0:08:50 | mira | master | ubuntu | 18.04 | rados/singleton-bluestore/{all/cephtool.yaml msgr-failures/few.yaml msgr/async-v2only.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}} | 1 | |
pass | 4474305 | 2019-11-05 12:09:13 | 2019-11-05 12:09:15 | 2019-11-05 12:29:14 | 0:19:59 | 0:10:44 | 0:09:15 | mira | master | ubuntu | 18.04 | rados/perf/{ceph.yaml objectstore/bluestore-basic-min-osd-mem-target.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{ubuntu_latest.yaml} workloads/radosbench_4K_seq_read.yaml} | 1 | |
pass | 4474306 | 2019-11-05 12:09:14 | 2019-11-05 12:09:16 | 2019-11-05 12:43:15 | 0:33:59 | 0:24:21 | 0:09:38 | mira | master | rhel | 7.7 | rados/mgr/{clusters/{2-node-mgr.yaml} debug/mgr.yaml objectstore/bluestore-low-osd-mem-target.yaml supported-random-distro$/{rhel_7.yaml} tasks/crash.yaml} | 2 | |
pass | 4474307 | 2019-11-05 12:09:15 | 2019-11-05 12:09:17 | 2019-11-05 12:47:16 | 0:37:59 | 0:27:02 | 0:10:57 | mira | master | ubuntu | 18.04 | rados/thrash-erasure-code-overwrites/{bluestore-bitmap.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} fast/normal.yaml msgr-failures/fastclose.yaml rados.yaml recovery-overrides/{more-active-recovery.yaml} supported-random-distro$/{ubuntu_latest.yaml} thrashers/careful.yaml thrashosds-health.yaml workloads/ec-small-objects-overwrites.yaml} | 2 | |
pass | 4474308 | 2019-11-05 12:09:16 | 2019-11-05 12:09:18 | 2019-11-05 12:35:17 | 0:25:59 | 0:13:04 | 0:12:55 | mira | master | ubuntu | 18.04 | rados/thrash-erasure-code-shec/{ceph.yaml clusters/{fixed-4.yaml openstack.yaml} msgr-failures/osd-delay.yaml objectstore/bluestore-avl.yaml rados.yaml recovery-overrides/{more-async-recovery.yaml} supported-random-distro$/{ubuntu_latest.yaml} thrashers/careful.yaml thrashosds-health.yaml workloads/ec-rados-plugin=shec-k=4-m=3-c=2.yaml} | 4 | |
pass | 4474309 | 2019-11-05 12:09:17 | 2019-11-05 12:09:19 | 2019-11-05 13:47:19 | 1:38:00 | 1:26:49 | 0:11:11 | mira | master | centos | 7.6 | rados/singleton/{all/ec-lost-unfound.yaml msgr-failures/few.yaml msgr/async-v2only.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{centos_7.yaml}} | 1 | |
pass | 4474310 | 2019-11-05 12:09:18 | 2019-11-05 12:09:19 | 2019-11-05 12:55:19 | 0:46:00 | 0:35:25 | 0:10:35 | mira | master | ubuntu | 18.04 | rados/objectstore/{backends/objectcacher-stress.yaml supported-random-distro$/{ubuntu_latest.yaml}} | 1 | |
pass | 4474311 | 2019-11-05 12:09:19 | 2019-11-05 12:09:21 | 2019-11-05 12:37:20 | 0:27:59 | 0:17:21 | 0:10:38 | mira | master | ubuntu | 18.04 | rados/singleton-nomsgr/{all/osd_stale_reads.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}} | 1 | |
pass | 4474312 | 2019-11-05 12:09:20 | 2019-11-05 12:29:23 | 2019-11-05 15:33:25 | 3:04:02 | 2:38:42 | 0:25:20 | mira | master | rhel | 7.7 | rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml 2-recovery-overrides/{more-async-partial-recovery.yaml} backoff/normal.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/off.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/filestore-xfs.yaml rados.yaml supported-random-distro$/{rhel_7.yaml} thrashers/morepggrow.yaml thrashosds-health.yaml workloads/rados_api_tests.yaml} | 2 | |
pass | 4474313 | 2019-11-05 12:09:21 | 2019-11-05 12:35:19 | 2019-11-05 13:01:18 | 0:25:59 | 0:13:09 | 0:12:50 | mira | master | centos | 7.6 | rados/multimon/{clusters/3.yaml msgr-failures/many.yaml msgr/async-v2only.yaml no_pools.yaml objectstore/bluestore-avl.yaml rados.yaml supported-random-distro$/{centos_7.yaml} tasks/mon_clock_with_skews.yaml} | 2 | |
pass | 4474314 | 2019-11-05 12:09:21 | 2019-11-05 12:37:37 | 2019-11-05 12:55:36 | 0:17:59 | 0:08:16 | 0:09:43 | mira | master | ubuntu | 18.04 | rados/singleton/{all/erasure-code-nonregression.yaml msgr-failures/many.yaml msgr/async.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}} | 1 | |
pass | 4474315 | 2019-11-05 12:09:22 | 2019-11-05 12:43:26 | 2019-11-05 13:03:24 | 0:19:58 | 0:10:50 | 0:09:08 | mira | master | ubuntu | 18.04 | rados/perf/{ceph.yaml objectstore/bluestore-bitmap.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{ubuntu_latest.yaml} workloads/radosbench_4M_rand_read.yaml} | 1 | |
pass | 4474316 | 2019-11-05 12:09:23 | 2019-11-05 12:43:26 | 2019-11-05 13:13:25 | 0:29:59 | 0:20:10 | 0:09:49 | mira | master | ubuntu | 18.04 | rados/thrash-erasure-code/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} fast/fast.yaml msgr-failures/osd-delay.yaml objectstore/bluestore-avl.yaml rados.yaml recovery-overrides/{default.yaml} supported-random-distro$/{ubuntu_latest.yaml} thrashers/default.yaml thrashosds-health.yaml workloads/ec-small-objects-many-deletes.yaml} | 2 | |
pass | 4474317 | 2019-11-05 12:09:24 | 2019-11-05 12:47:27 | 2019-11-05 13:41:26 | 0:53:59 | 0:40:57 | 0:13:02 | mira | master | centos | 7.6 | rados/thrash-erasure-code-isa/{arch/x86_64.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/few.yaml objectstore/bluestore-comp.yaml rados.yaml recovery-overrides/{more-async-partial-recovery.yaml} supported-random-distro$/{centos_7.yaml} thrashers/morepggrow.yaml thrashosds-health.yaml workloads/ec-rados-plugin=isa-k=2-m=1.yaml} | 2 | |
pass | 4474318 | 2019-11-05 12:09:25 | 2019-11-05 12:47:27 | 2019-11-05 13:15:26 | 0:27:59 | 0:17:56 | 0:10:03 | mira | master | ubuntu | 18.04 | rados/basic/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/many.yaml msgr/async-v1only.yaml objectstore/bluestore-stupid.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} tasks/rados_stress_watch.yaml} | 2 | |
pass | 4474319 | 2019-11-05 12:09:26 | 2019-11-05 12:51:21 | 2019-11-05 14:33:22 | 1:42:01 | 1:35:51 | 0:06:10 | mira | master | rhel | 7.7 | rados/singleton/{all/lost-unfound-delete.yaml msgr-failures/few.yaml msgr/async-v1only.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}} | 1 | |
pass | 4474320 | 2019-11-05 12:09:27 | 2019-11-05 12:55:22 | 2019-11-05 13:35:22 | 0:40:00 | 0:22:52 | 0:17:08 | mira | master | centos | 7.6 | rados/thrash-old-clients/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-install/luminous-v1only.yaml backoff/peering.yaml ceph.yaml clusters/{openstack.yaml three-plus-one.yaml} d-balancer/crush-compat.yaml distro$/{centos_latest.yaml} msgr-failures/fastclose.yaml rados.yaml thrashers/careful.yaml thrashosds-health.yaml workloads/test_rbd_api.yaml} | 4 | |
pass | 4474321 | 2019-11-05 12:09:28 | 2019-11-05 12:55:38 | 2019-11-05 13:45:37 | 0:49:59 | 0:21:11 | 0:28:48 | mira | master | centos | 7.6 | rados/thrash-erasure-code-big/{ceph.yaml cluster/{12-osds.yaml openstack.yaml} msgr-failures/fastclose.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml recovery-overrides/{more-async-recovery.yaml} supported-random-distro$/{centos_7.yaml} thrashers/pggrow.yaml thrashosds-health.yaml workloads/ec-rados-plugin=lrc-k=4-m=2-l=3.yaml} | 3 | |
pass | 4474322 | 2019-11-05 12:09:28 | 2019-11-05 13:01:20 | 2019-11-05 13:23:19 | 0:21:59 | 0:11:25 | 0:10:34 | mira | master | rados/verify/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-thrash/none.yaml msgr-failures/few.yaml msgr/async-v2only.yaml objectstore/filestore-xfs.yaml rados.yaml tasks/mon_recovery.yaml validater/lockdep.yaml} | 2 | |||
pass | 4474323 | 2019-11-05 12:09:29 | 2019-11-05 13:03:45 | 2019-11-05 13:53:45 | 0:50:00 | 0:38:34 | 0:11:26 | mira | master | ubuntu | 18.04 | rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml 2-recovery-overrides/{more-async-partial-recovery.yaml} backoff/peering.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/upmap.yaml msgr-failures/osd-delay.yaml msgr/async-v1only.yaml objectstore/bluestore-avl.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} thrashers/none.yaml thrashosds-health.yaml workloads/radosbench.yaml} | 2 | |
pass | 4474324 | 2019-11-05 12:09:30 | 2019-11-05 13:13:44 | 2019-11-05 14:35:45 | 1:22:01 | 1:06:36 | 0:15:25 | mira | master | centos | 7.6 | rados/dashboard/{clusters/{2-node-mgr.yaml} debug/mgr.yaml objectstore/filestore-xfs.yaml supported-random-distro$/{centos_7.yaml} tasks/dashboard.yaml} | 2 | |
pass | 4474325 | 2019-11-05 12:09:31 | 2019-11-05 13:15:28 | 2019-11-05 13:41:27 | 0:25:59 | 0:14:59 | 0:11:00 | mira | master | centos | 7.6 | rados/singleton-nomsgr/{all/pool-access.yaml rados.yaml supported-random-distro$/{centos_7.yaml}} | 1 | |
pass | 4474326 | 2019-11-05 12:09:32 | 2019-11-05 13:23:22 | 2019-11-05 14:45:22 | 1:22:00 | 1:12:33 | 0:09:27 | mira | master | ubuntu | 18.04 | rados/standalone/{supported-random-distro$/{ubuntu_latest.yaml} workloads/scrub.yaml} | 1 | |
pass | 4474327 | 2019-11-05 12:09:33 | 2019-11-05 13:35:23 | 2019-11-05 14:51:24 | 1:16:01 | 1:06:31 | 0:09:30 | mira | master | ubuntu | 18.04 | rados/singleton/{all/lost-unfound.yaml msgr-failures/many.yaml msgr/async-v2only.yaml objectstore/bluestore-stupid.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}} | 1 | |
pass | 4474328 | 2019-11-05 12:09:34 | 2019-11-05 13:41:37 | 2019-11-05 16:37:39 | 2:56:02 | 2:36:24 | 0:19:38 | mira | master | rhel | 7.7 | rados/monthrash/{ceph.yaml clusters/3-mons.yaml msgr-failures/mon-delay.yaml msgr/async-v2only.yaml objectstore/bluestore-stupid.yaml rados.yaml supported-random-distro$/{rhel_7.yaml} thrashers/force-sync-many.yaml workloads/rados_api_tests.yaml} | 2 | |
pass | 4474329 | 2019-11-05 12:09:34 | 2019-11-05 13:41:37 | 2019-11-05 14:13:36 | 0:31:59 | 0:25:16 | 0:06:43 | mira | master | rhel | 7.7 | rados/mgr/{clusters/{2-node-mgr.yaml} debug/mgr.yaml objectstore/bluestore-stupid.yaml supported-random-distro$/{rhel_7.yaml} tasks/failover.yaml} | 2 | |
pass | 4474330 | 2019-11-05 12:09:35 | 2019-11-05 13:45:41 | 2019-11-05 14:05:41 | 0:20:00 | 0:10:26 | 0:09:34 | mira | master | ubuntu | 18.04 | rados/perf/{ceph.yaml objectstore/bluestore-comp.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{ubuntu_latest.yaml} workloads/radosbench_4M_seq_read.yaml} | 1 | |
pass | 4474331 | 2019-11-05 12:09:36 | 2019-11-05 13:47:34 | 2019-11-05 14:05:33 | 0:17:59 | 0:07:24 | 0:10:35 | mira | master | ubuntu | 18.04 | rados/singleton/{all/max-pg-per-osd.from-mon.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/filestore-xfs.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}} | 1 | |
pass | 4474332 | 2019-11-05 12:09:37 | 2019-11-05 13:53:47 | 2019-11-05 14:33:46 | 0:39:59 | 0:30:21 | 0:09:38 | mira | master | rhel | 7.7 | rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml 2-recovery-overrides/{more-async-partial-recovery.yaml} backoff/peering_and_degraded.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/crush-compat.yaml msgr-failures/fastclose.yaml msgr/async-v2only.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{rhel_7.yaml} thrashers/pggrow.yaml thrashosds-health.yaml workloads/redirect.yaml} | 2 | |
pass | 4474333 | 2019-11-05 12:09:38 | 2019-11-05 14:05:48 | 2019-11-05 14:25:47 | 0:19:59 | 0:09:23 | 0:10:36 | mira | master | ubuntu | 18.04 | rados/basic/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/few.yaml msgr/async-v2only.yaml objectstore/filestore-xfs.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} tasks/rados_striper.yaml} | 2 | |
dead | 4474334 | 2019-11-05 12:09:39 | 2019-11-05 14:05:48 | 2019-11-06 02:08:13 | 12:02:25 | mira | master | rhel | 7.7 | rados/objectstore/{backends/objectstore.yaml supported-random-distro$/{rhel_7.yaml}} | 1 | |||
pass | 4474335 | 2019-11-05 12:09:40 | 2019-11-05 14:13:39 | 2019-11-05 14:59:38 | 0:45:59 | 0:32:10 | 0:13:49 | mira | master | centos | 7.6 | rados/singleton-nomsgr/{all/recovery-unfound-found.yaml rados.yaml supported-random-distro$/{centos_7.yaml}} | 1 | |
pass | 4474336 | 2019-11-05 12:09:40 | 2019-11-05 14:25:50 | 2019-11-05 15:07:49 | 0:41:59 | 0:30:17 | 0:11:42 | mira | master | ubuntu | 18.04 | rados/thrash-erasure-code-overwrites/{bluestore-bitmap.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} fast/fast.yaml msgr-failures/few.yaml rados.yaml recovery-overrides/{more-async-partial-recovery.yaml} supported-random-distro$/{ubuntu_latest.yaml} thrashers/default.yaml thrashosds-health.yaml workloads/ec-snaps-few-objects-overwrites.yaml} | 2 | |
pass | 4474337 | 2019-11-05 12:09:41 | 2019-11-05 14:33:38 | 2019-11-05 15:11:38 | 0:38:00 | 0:28:16 | 0:09:44 | mira | master | rhel | 7.7 | rados/thrash-erasure-code-shec/{ceph.yaml clusters/{fixed-4.yaml openstack.yaml} msgr-failures/fastclose.yaml objectstore/bluestore-bitmap.yaml rados.yaml recovery-overrides/{more-active-recovery.yaml} supported-random-distro$/{rhel_7.yaml} thrashers/default.yaml thrashosds-health.yaml workloads/ec-rados-plugin=shec-k=4-m=3-c=2.yaml} | 4 | |
pass | 4474338 | 2019-11-05 12:09:42 | 2019-11-05 14:33:48 | 2019-11-05 17:21:50 | 2:48:02 | 2:29:43 | 0:18:19 | mira | master | rhel | 7.7 | rados/singleton/{all/max-pg-per-osd.from-primary.yaml msgr-failures/many.yaml msgr/async-v1only.yaml objectstore/bluestore-avl.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}} | 1 | |
fail | 4474339 | 2019-11-05 12:09:43 | 2019-11-05 14:35:46 | 2019-11-05 15:31:46 | 0:56:00 | 0:30:54 | 0:25:06 | mira | master | centos | 7.6 | rados/thrash-old-clients/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-install/luminous.yaml backoff/peering_and_degraded.yaml ceph.yaml clusters/{openstack.yaml three-plus-one.yaml} d-balancer/off.yaml distro$/{centos_latest.yaml} msgr-failures/few.yaml rados.yaml thrashers/default.yaml thrashosds-health.yaml workloads/cache-snaps.yaml} | 4 | |
Failure Reason:
Command crashed: 'CEPH_CLIENT_ID=2 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph_test_rados --max-ops 4000 --objects 500 --max-in-flight 16 --size 4000000 --min-stride-size 400000 --max-stride-size 800000 --max-seconds 0 --op snap_remove 50 --op snap_create 50 --op rollback 50 --op read 100 --op copy_from 50 --op write 50 --op write_excl 50 --op cache_try_flush 50 --op cache_flush 50 --op cache_evict 50 --op delete 50 --pool base' |
||||||||||||||
pass | 4474340 | 2019-11-05 12:09:44 | 2019-11-05 14:45:41 | 2019-11-05 15:15:41 | 0:30:00 | 0:24:09 | 0:05:51 | mira | master | rhel | 7.7 | rados/perf/{ceph.yaml objectstore/bluestore-low-osd-mem-target.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{rhel_7.yaml} workloads/radosbench_4M_write.yaml} | 1 | |
pass | 4474341 | 2019-11-05 12:09:45 | 2019-11-05 14:51:31 | 2019-11-05 15:21:30 | 0:29:59 | 0:19:50 | 0:10:09 | mira | master | ubuntu | 18.04 | rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml 2-recovery-overrides/{more-async-recovery.yaml} backoff/normal.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/off.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} thrashers/careful.yaml thrashosds-health.yaml workloads/redirect_promote_tests.yaml} | 2 | |
pass | 4474342 | 2019-11-05 12:09:46 | 2019-11-05 14:59:56 | 2019-11-05 15:31:56 | 0:32:00 | 0:26:03 | 0:05:57 | mira | master | rhel | 7.7 | rados/singleton/{all/max-pg-per-osd.from-replica.yaml msgr-failures/few.yaml msgr/async-v2only.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}} | 1 | |
pass | 4474343 | 2019-11-05 12:09:46 | 2019-11-05 15:07:51 | 2019-11-05 15:39:51 | 0:32:00 | 0:25:20 | 0:06:40 | mira | master | rhel | 7.7 | rados/multimon/{clusters/6.yaml msgr-failures/few.yaml msgr/async.yaml no_pools.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{rhel_7.yaml} tasks/mon_recovery.yaml} | 2 | |
pass | 4474344 | 2019-11-05 12:09:47 | 2019-11-05 15:11:40 | 2019-11-05 16:03:40 | 0:52:00 | 0:28:58 | 0:23:02 | mira | master | centos | 7.6 | rados/thrash-erasure-code/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} fast/normal.yaml msgr-failures/fastclose.yaml objectstore/bluestore-bitmap.yaml rados.yaml recovery-overrides/{more-async-partial-recovery.yaml} supported-random-distro$/{centos_7.yaml} thrashers/fastread.yaml thrashosds-health.yaml workloads/ec-small-objects.yaml} | 2 | |
pass | 4474345 | 2019-11-05 12:09:48 | 2019-11-05 15:16:00 | 2019-11-05 15:45:59 | 0:29:59 | 0:22:36 | 0:07:23 | mira | master | rhel | 7.7 | rados/singleton-nomsgr/{all/version-number-sanity.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}} | 1 | |
pass | 4474346 | 2019-11-05 12:09:49 | 2019-11-05 15:21:49 | 2019-11-05 15:43:48 | 0:21:59 | 0:10:39 | 0:11:20 | mira | master | ubuntu | 18.04 | rados/mgr/{clusters/{2-node-mgr.yaml} debug/mgr.yaml objectstore/filestore-xfs.yaml supported-random-distro$/{ubuntu_latest.yaml} tasks/insights.yaml} | 2 | |
pass | 4474347 | 2019-11-05 12:09:50 | 2019-11-05 15:31:48 | 2019-11-05 16:21:48 | 0:50:00 | 0:36:11 | 0:13:49 | mira | master | centos | 7.6 | rados/thrash-erasure-code-isa/{arch/x86_64.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/osd-delay.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml recovery-overrides/{more-active-recovery.yaml} supported-random-distro$/{centos_7.yaml} thrashers/none.yaml thrashosds-health.yaml workloads/ec-rados-plugin=isa-k=2-m=1.yaml} | 2 | |
pass | 4474348 | 2019-11-05 12:09:51 | 2019-11-05 15:31:57 | 2019-11-05 16:01:56 | 0:29:59 | 0:22:38 | 0:07:21 | mira | master | rhel | 7.7 | rados/singleton/{all/mon-auth-caps.yaml msgr-failures/many.yaml msgr/async.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}} | 1 | |
pass | 4474349 | 2019-11-05 12:09:51 | 2019-11-05 15:33:44 | 2019-11-05 16:25:44 | 0:52:00 | 0:42:23 | 0:09:37 | mira | master | rhel | 7.7 | rados/thrash-erasure-code-big/{ceph.yaml cluster/{12-osds.yaml openstack.yaml} msgr-failures/few.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml recovery-overrides/{more-active-recovery.yaml} supported-random-distro$/{rhel_7.yaml} thrashers/careful.yaml thrashosds-health.yaml workloads/ec-rados-plugin=jerasure-k=4-m=2.yaml} | 3 | |
pass | 4474350 | 2019-11-05 12:09:52 | 2019-11-05 15:39:57 | 2019-11-05 16:31:57 | 0:52:00 | 0:38:55 | 0:13:05 | mira | master | centos | rados/verify/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-thrash/default/{default.yaml thrashosds-health.yaml} msgr-failures/few.yaml msgr/async-v2only.yaml objectstore/bluestore-avl.yaml rados.yaml tasks/mon_recovery.yaml validater/valgrind.yaml} | 2 | ||
pass | 4474351 | 2019-11-05 12:09:53 | 2019-11-05 15:43:53 | 2019-11-05 16:35:52 | 0:51:59 | 0:44:24 | 0:07:35 | mira | master | rhel | 7.7 | rados/basic/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/many.yaml msgr/async.yaml objectstore/bluestore-avl.yaml rados.yaml supported-random-distro$/{rhel_7.yaml} tasks/rados_workunit_loadgen_big.yaml} | 2 | |
pass | 4474352 | 2019-11-05 12:09:54 | 2019-11-05 15:46:02 | 2019-11-05 16:22:01 | 0:35:59 | 0:24:04 | 0:11:55 | mira | master | ubuntu | 18.04 | rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml 2-recovery-overrides/{default.yaml} backoff/peering.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/upmap.yaml msgr-failures/osd-delay.yaml msgr/async-v1only.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} thrashers/default.yaml thrashosds-health.yaml workloads/redirect_set_object.yaml} | 2 | |
pass | 4474353 | 2019-11-05 12:09:55 | 2019-11-05 16:01:59 | 2019-11-05 16:37:58 | 0:35:59 | 0:25:18 | 0:10:41 | mira | master | centos | 7.6 | rados/perf/{ceph.yaml objectstore/bluestore-stupid.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{centos_7.yaml} workloads/radosbench_omap_write.yaml} | 1 | |
pass | 4474354 | 2019-11-05 12:09:56 | 2019-11-05 16:03:52 | 2019-11-05 16:23:51 | 0:19:59 | 0:11:21 | 0:08:38 | mira | master | ubuntu | 18.04 | rados/singleton/{all/mon-config-key-caps.yaml msgr-failures/few.yaml msgr/async-v1only.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}} | 1 | |
pass | 4474355 | 2019-11-05 12:09:57 | 2019-11-05 16:22:06 | 2019-11-05 18:08:06 | 1:46:00 | 1:33:49 | 0:12:11 | mira | master | centos | 7.6 | rados/monthrash/{ceph.yaml clusters/9-mons.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/filestore-xfs.yaml rados.yaml supported-random-distro$/{centos_7.yaml} thrashers/many.yaml workloads/rados_mon_osdmap_prune.yaml} | 2 | |
pass | 4474356 | 2019-11-05 12:09:58 | 2019-11-05 16:22:06 | 2019-11-05 18:00:06 | 1:38:00 | 1:05:56 | 0:32:04 | mira | master | centos | 7.6 | rados/thrash-old-clients/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-install/mimic-v1only.yaml backoff/normal.yaml ceph.yaml clusters/{openstack.yaml three-plus-one.yaml} d-balancer/crush-compat.yaml distro$/{centos_latest.yaml} msgr-failures/osd-delay.yaml rados.yaml thrashers/mapgap.yaml thrashosds-health.yaml workloads/radosbench.yaml} | 4 | |
pass | 4474357 | 2019-11-05 12:09:59 | 2019-11-05 16:24:09 | 2019-11-05 17:44:10 | 1:20:01 | 1:06:35 | 0:13:26 | mira | master | centos | 7.6 | rados/dashboard/{clusters/{2-node-mgr.yaml} debug/mgr.yaml objectstore/bluestore-avl.yaml supported-random-distro$/{centos_7.yaml} tasks/dashboard.yaml} | 2 | |
pass | 4474358 | 2019-11-05 12:09:59 | 2019-11-05 16:26:02 | 2019-11-05 16:52:01 | 0:25:59 | 0:15:09 | 0:10:50 | mira | master | centos | 7.6 | rados/objectstore/{backends/alloc-hint.yaml supported-random-distro$/{centos_7.yaml}} | 1 | |
pass | 4474359 | 2019-11-05 12:10:00 | 2019-11-05 16:32:16 | 2019-11-05 16:58:15 | 0:25:59 | 0:16:23 | 0:09:36 | mira | master | centos | 7.6 | rados/rest/{mgr-restful.yaml supported-random-distro$/{centos_7.yaml}} | 1 | |
pass | 4474360 | 2019-11-05 12:10:01 | 2019-11-05 16:36:10 | 2019-11-05 17:28:09 | 0:51:59 | 0:45:36 | 0:06:23 | mira | master | rhel | 7.7 | rados/singleton-bluestore/{all/cephtool.yaml msgr-failures/many.yaml msgr/async.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}} | 1 | |
fail | 4474361 | 2019-11-05 12:10:02 | 2019-11-05 16:37:57 | 2019-11-05 17:03:56 | 0:25:59 | mira | master | centos | rados/singleton-flat/valgrind-leaks.yaml | 1 | ||||
Failure Reason:
[Errno 113] No route to host |
||||||||||||||
fail | 4474362 | 2019-11-05 12:10:03 | 2019-11-05 16:38:00 | 2019-11-05 17:15:59 | 0:37:59 | 0:29:50 | 0:08:09 | mira | master | rhel | 7.7 | rados/singleton-nomsgr/{all/admin_socket_output.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}} | 1 | |
Failure Reason:
Command crashed: "sudo TESTDIR=/home/ubuntu/cephtest bash -c 'ceph_test_admin_socket_output --all'" |
||||||||||||||
pass | 4474363 | 2019-11-05 12:10:04 | 2019-11-05 16:52:22 | 2019-11-05 17:22:21 | 0:29:59 | 0:23:54 | 0:06:05 | mira | master | rhel | 7.7 | rados/standalone/{supported-random-distro$/{rhel_7.yaml} workloads/crush.yaml} | 1 | |
fail | 4474364 | 2019-11-05 12:10:04 | 2019-11-05 16:58:17 | 2019-11-05 17:38:17 | 0:40:00 | 0:30:03 | 0:09:57 | mira | master | rhel | 7.7 | rados/upgrade/mimic-x-singleton/{0-cluster/{openstack.yaml start.yaml} 1-install/mimic.yaml 2-partial-upgrade/firsthalf.yaml 3-thrash/default.yaml 4-workload/{rbd-cls.yaml rbd-import-export.yaml readwrite.yaml snaps-few-objects.yaml} 5-workload/{radosbench.yaml rbd_api.yaml} 6-finish-upgrade.yaml 7-nautilus.yaml 8-workload/{rbd-python.yaml rgw-swift.yaml snaps-many-objects.yaml} bluestore-bitmap.yaml supported-random-distro$/{rhel_7.yaml} thrashosds-health.yaml} | 4 | |
Failure Reason:
Command failed (workunit test cls/test_cls_rbd.sh) on mira056 with status 139: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=mimic TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/cls/test_cls_rbd.sh' |
||||||||||||||
pass | 4474365 | 2019-11-05 12:10:05 | 2019-11-05 17:03:59 | 2019-11-05 19:52:00 | 2:48:01 | 2:29:01 | 0:19:00 | mira | master | rhel | 7.7 | rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml 2-recovery-overrides/{more-async-partial-recovery.yaml} backoff/peering_and_degraded.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/crush-compat.yaml msgr-failures/fastclose.yaml msgr/async-v2only.yaml objectstore/bluestore-stupid.yaml rados.yaml supported-random-distro$/{rhel_7.yaml} thrashers/mapgap.yaml thrashosds-health.yaml workloads/set-chunk-promote-flush.yaml} | 2 | |
pass | 4474366 | 2019-11-05 12:10:06 | 2019-11-05 17:16:01 | 2019-11-05 17:50:01 | 0:34:00 | 0:27:00 | 0:07:00 | mira | master | rhel | 7.7 | rados/singleton/{all/mon-config-keys.yaml msgr-failures/many.yaml msgr/async-v2only.yaml objectstore/bluestore-stupid.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}} | 1 | |
pass | 4474367 | 2019-11-05 12:10:07 | 2019-11-05 17:22:08 | 2019-11-05 17:52:07 | 0:29:59 | 0:22:39 | 0:07:20 | mira | master | rhel | 7.7 | rados/perf/{ceph.yaml objectstore/bluestore-basic-min-osd-mem-target.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{rhel_7.yaml} workloads/sample_fio.yaml} | 1 | |
fail | 4474368 | 2019-11-05 12:10:08 | 2019-11-05 17:22:23 | 2019-11-05 18:02:22 | 0:39:59 | 0:16:14 | 0:23:45 | mira | master | centos | 7.6 | rados/mgr/{clusters/{2-node-mgr.yaml} debug/mgr.yaml objectstore/bluestore-avl.yaml supported-random-distro$/{centos_7.yaml} tasks/module_selftest.yaml} | 2 | |
Failure Reason:
Test failure: test_devicehealth (tasks.mgr.test_module_selftest.TestModuleSelftest) |
||||||||||||||
pass | 4474369 | 2019-11-05 12:10:09 | 2019-11-05 17:28:11 | 2019-11-05 18:10:11 | 0:42:00 | 0:31:25 | 0:10:35 | mira | master | ubuntu | 18.04 | rados/thrash-erasure-code-overwrites/{bluestore-bitmap.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} fast/normal.yaml msgr-failures/osd-delay.yaml rados.yaml recovery-overrides/{more-active-recovery.yaml} supported-random-distro$/{ubuntu_latest.yaml} thrashers/fastread.yaml thrashosds-health.yaml workloads/ec-pool-snaps-few-objects-overwrites.yaml} | 2 | |
pass | 4474370 | 2019-11-05 12:10:10 | 2019-11-05 17:38:20 | 2019-11-05 18:20:19 | 0:41:59 | 0:23:48 | 0:18:11 | mira | master | centos | 7.6 | rados/thrash-erasure-code-shec/{ceph.yaml clusters/{fixed-4.yaml openstack.yaml} msgr-failures/few.yaml objectstore/bluestore-comp.yaml rados.yaml recovery-overrides/{more-async-recovery.yaml} supported-random-distro$/{centos_7.yaml} thrashers/careful.yaml thrashosds-health.yaml workloads/ec-rados-plugin=shec-k=4-m=3-c=2.yaml} | 4 | |
pass | 4474371 | 2019-11-05 12:10:11 | 2019-11-05 17:44:11 | 2019-11-05 18:04:10 | 0:19:59 | 0:09:29 | 0:10:30 | mira | master | ubuntu | 18.04 | rados/singleton/{all/mon-config.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/filestore-xfs.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}} | 1 | |
pass | 4474372 | 2019-11-05 12:10:11 | 2019-11-05 17:50:19 | 2019-11-05 18:22:18 | 0:31:59 | 0:21:34 | 0:10:25 | mira | master | ubuntu | 18.04 | rados/basic/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/few.yaml msgr/async-v1only.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} tasks/rados_workunit_loadgen_mix.yaml} | 2 | |
pass | 4474373 | 2019-11-05 12:10:12 | 2019-11-05 17:52:27 | 2019-11-05 18:20:26 | 0:27:59 | 0:21:51 | 0:06:08 | mira | master | rhel | 7.7 | rados/singleton-nomsgr/{all/balancer.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}} | 1 | |
pass | 4474374 | 2019-11-05 12:10:13 | 2019-11-05 18:00:19 | 2019-11-05 18:32:18 | 0:31:59 | 0:22:22 | 0:09:37 | mira | master | ubuntu | 18.04 | rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml 2-recovery-overrides/{more-async-recovery.yaml} backoff/normal.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/off.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/filestore-xfs.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} thrashers/morepggrow.yaml thrashosds-health.yaml workloads/set-chunks-read.yaml} | 2 | |
dead | 4474375 | 2019-11-05 12:10:14 | 2019-11-05 18:02:24 | 2019-11-06 06:04:48 | 12:02:24 | mira | master | rhel | 7.7 | rados/thrash-erasure-code/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} fast/fast.yaml msgr-failures/few.yaml objectstore/bluestore-comp.yaml rados.yaml recovery-overrides/{more-async-partial-recovery.yaml} supported-random-distro$/{rhel_7.yaml} thrashers/minsize_recovery.yaml thrashosds-health.yaml workloads/ec-rados-plugin=clay-k=4-m=2.yaml} | 2 | |||
pass | 4474376 | 2019-11-05 12:10:15 | 2019-11-05 18:04:23 | 2019-11-05 19:12:23 | 1:08:00 | 0:56:20 | 0:11:40 | mira | master | ubuntu | 18.04 | rados/singleton/{all/osd-backfill.yaml msgr-failures/many.yaml msgr/async-v1only.yaml objectstore/bluestore-avl.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}} | 1 | |
pass | 4474377 | 2019-11-05 12:10:15 | 2019-11-05 18:08:20 | 2019-11-05 18:40:19 | 0:31:59 | 0:15:58 | 0:16:01 | mira | master | centos | 7.6 | rados/multimon/{clusters/9.yaml msgr-failures/many.yaml msgr/async-v1only.yaml no_pools.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{centos_7.yaml} tasks/mon_clock_no_skews.yaml} | 3 | |
fail | 4474378 | 2019-11-05 12:10:16 | 2019-11-05 18:10:22 | 2019-11-05 19:00:21 | 0:49:59 | 0:24:41 | 0:25:18 | mira | master | centos | 7.6 | rados/thrash-old-clients/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-install/mimic.yaml backoff/peering.yaml ceph.yaml clusters/{openstack.yaml three-plus-one.yaml} d-balancer/off.yaml distro$/{centos_latest.yaml} msgr-failures/fastclose.yaml rados.yaml thrashers/morepggrow.yaml thrashosds-health.yaml workloads/rbd_cls.yaml} | 4 | |
Failure Reason:
Command failed on mira018 with status 1: "sudo TESTDIR=/home/ubuntu/cephtest bash -c 'ceph_test_cls_rbd --gtest_filter=-TestClsRbd.get_features:TestClsRbd.parents'" |
||||||||||||||
pass | 4474379 | 2019-11-05 12:10:17 | 2019-11-05 18:20:21 | 2019-11-05 18:40:21 | 0:20:00 | 0:10:09 | 0:09:51 | mira | master | ubuntu | 18.04 | rados/perf/{ceph.yaml objectstore/bluestore-bitmap.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{ubuntu_latest.yaml} workloads/sample_radosbench.yaml} | 1 | |
pass | 4474380 | 2019-11-05 12:10:18 | 2019-11-05 18:20:28 | 2019-11-05 19:00:28 | 0:40:00 | 0:29:36 | 0:10:24 | mira | master | ubuntu | 18.04 | rados/thrash-erasure-code-isa/{arch/x86_64.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/fastclose.yaml objectstore/bluestore-stupid.yaml rados.yaml recovery-overrides/{more-partial-recovery.yaml} supported-random-distro$/{ubuntu_latest.yaml} thrashers/pggrow.yaml thrashosds-health.yaml workloads/ec-rados-plugin=isa-k=2-m=1.yaml} | 2 | |
pass | 4474381 | 2019-11-05 12:10:19 | 2019-11-05 18:22:23 | 2019-11-05 19:32:23 | 1:10:00 | 0:59:36 | 0:10:24 | mira | master | ubuntu | 18.04 | rados/singleton/{all/osd-recovery-incomplete.yaml msgr-failures/few.yaml msgr/async-v2only.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}} | 1 | |
pass | 4474382 | 2019-11-05 12:10:20 | 2019-11-05 18:32:40 | 2019-11-05 19:08:39 | 0:35:59 | 0:23:57 | 0:12:02 | mira | master | ubuntu | 18.04 | rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml 2-recovery-overrides/{more-async-partial-recovery.yaml} backoff/peering.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/upmap.yaml msgr-failures/osd-delay.yaml msgr/async-v1only.yaml objectstore/bluestore-avl.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} thrashers/none.yaml thrashosds-health.yaml workloads/small-objects.yaml} | 2 | |
pass | 4474383 | 2019-11-05 12:10:21 | 2019-11-05 18:40:22 | 2019-11-05 19:12:21 | 0:31:59 | 0:25:25 | 0:06:34 | mira | master | rhel | 7.7 | rados/thrash-erasure-code-big/{ceph.yaml cluster/{12-osds.yaml openstack.yaml} msgr-failures/osd-delay.yaml objectstore/bluestore-stupid.yaml rados.yaml recovery-overrides/{default.yaml} supported-random-distro$/{rhel_7.yaml} thrashers/default.yaml thrashosds-health.yaml workloads/ec-rados-plugin=lrc-k=4-m=2-l=3.yaml} | 3 | |
pass | 4474384 | 2019-11-05 12:10:22 | 2019-11-05 18:40:22 | 2019-11-05 19:12:22 | 0:32:00 | 0:20:21 | 0:11:39 | mira | master | rados/verify/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-thrash/none.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/bluestore-bitmap.yaml rados.yaml tasks/rados_api_tests.yaml validater/lockdep.yaml} | 2 | |||
pass | 4474385 | 2019-11-05 12:10:22 | 2019-11-05 19:00:24 | 2019-11-05 19:38:23 | 0:37:59 | 0:26:38 | 0:11:21 | mira | master | centos | 7.6 | rados/objectstore/{backends/ceph_objectstore_tool.yaml supported-random-distro$/{centos_7.yaml}} | 1 | |
pass | 4474386 | 2019-11-05 12:10:23 | 2019-11-05 19:00:29 | 2019-11-05 19:28:28 | 0:27:59 | 0:21:05 | 0:06:54 | mira | master | rhel | 7.7 | rados/singleton-nomsgr/{all/cache-fs-trunc.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}} | 1 | |
pass | 4474387 | 2019-11-05 12:10:24 | 2019-11-05 19:08:41 | 2019-11-05 20:10:41 | 1:02:00 | 0:39:49 | 0:22:11 | mira | master | centos | 7.6 | rados/basic/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/many.yaml msgr/async-v2only.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{centos_7.yaml} tasks/rados_workunit_loadgen_mostlyread.yaml} | 2 | |
pass | 4474388 | 2019-11-05 12:10:25 | 2019-11-05 19:12:37 | 2019-11-05 19:34:37 | 0:22:00 | 0:12:07 | 0:09:53 | mira | master | ubuntu | 18.04 | rados/singleton/{all/osd-recovery.yaml msgr-failures/many.yaml msgr/async.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}} | 1 | |
pass | 4474389 | 2019-11-05 12:10:26 | 2019-11-05 19:12:38 | 2019-11-05 19:36:37 | 0:23:59 | 0:15:33 | 0:08:26 | mira | master | ubuntu | 18.04 | rados/perf/{ceph.yaml objectstore/bluestore-bitmap.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{ubuntu_latest.yaml} workloads/cosbench_64K_read_write.yaml} | 1 | |
fail | 4474390 | 2019-11-05 12:10:27 | 2019-11-05 19:12:37 | 2019-11-05 19:48:37 | 0:36:00 | 0:29:03 | 0:06:57 | mira | master | rhel | 7.7 | rados/monthrash/{ceph.yaml clusters/3-mons.yaml msgr-failures/mon-delay.yaml msgr/async-v1only.yaml objectstore/bluestore-avl.yaml rados.yaml supported-random-distro$/{rhel_7.yaml} thrashers/one.yaml workloads/rados_mon_workunits.yaml} | 2 | |
Failure Reason:
Command failed (workunit test mon/crush_ops.sh) on mira109 with status 16: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=4bad7f756c2b0609ac17bae7e818ce6a17d3b451 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/mon/crush_ops.sh' |
||||||||||||||
pass | 4474391 | 2019-11-05 12:10:28 | 2019-11-05 19:28:30 | 2019-11-05 19:50:29 | 0:21:59 | 0:11:30 | 0:10:29 | mira | master | ubuntu | 18.04 | rados/mgr/{clusters/{2-node-mgr.yaml} debug/mgr.yaml objectstore/bluestore-bitmap.yaml supported-random-distro$/{ubuntu_latest.yaml} tasks/orchestrator_cli.yaml} | 2 | |
fail | 4474392 | 2019-11-05 12:10:29 | 2019-11-05 19:32:25 | 2019-11-05 22:06:27 | 2:34:02 | 2:15:35 | 0:18:27 | mira | master | rhel | 7.7 | rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml 2-recovery-overrides/{more-async-recovery.yaml} backoff/peering_and_degraded.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/crush-compat.yaml msgr-failures/fastclose.yaml msgr/async-v2only.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{rhel_7.yaml} thrashers/pggrow.yaml thrashosds-health.yaml workloads/snaps-few-objects.yaml} | 2 | |
Failure Reason:
'strip_empty_ends': True, u'_raw_params': u'sgdisk --zap-all /dev/sdd || sgdisk --zap-all /dev/sdd', u'removes': None, u'argv': None, u'creates': None, u'chdir': None, u'stdin_add_newline': True, u'stdin': None}}, u'start': u'2019-11-05 22:06:02.522604'}, {'stderr_lines': [], u'changed': True, u'stdout': u'Creating new GPT entries.\nGPT data structures destroyed! You may now partition the disk using fdisk or\nother utilities.', u'delta': u'0:00:01.012016', 'stdout_lines': [u'Creating new GPT entries.', u'GPT data structures destroyed! You may now partition the disk using fdisk or', u'other utilities.'], '_ansible_item_label': {'value': {u'sectorsize': u'512', u'vendor': u'Hitachi', u'links': {u'masters': [], u'labels': [], u'ids': [u'scsi-2001b4d20119e5500'], u'uuids': []}, u'sas_device_handle': None, u'host': u'RAID bus controller: Areca Technology Corp. ARC-1680 series PCIe to SAS/SATA 3Gb RAID Controller', u'support_discard': u'0', u'serial': u'JPW9K0N211N9UE', u'holders': [], u'size': u'931.51 GB', u'scheduler_mode': u'deadline', u'rotational': u'1', u'sectors': u'1953525168', u'sas_address': None, u'virtual': 1, u'removable': u'0', u'model': u'HUA722010CLA330', u'partitions': {}}, 'key': u'sde'}, 'ansible_loop_var': u'item', u'end': u'2019-11-05 22:06:04.814290', '_ansible_no_log': False, 'item': {'value': {u'sectorsize': u'512', u'vendor': u'Hitachi', u'links': {u'masters': [], u'labels': [], u'ids': [u'scsi-2001b4d20119e5500'], u'uuids': []}, u'sas_device_handle': None, u'host': u'RAID bus controller: Areca Technology Corp. ARC-1680 series PCIe to SAS/SATA 3Gb RAID Controller', u'support_discard': u'0', u'serial': u'JPW9K0N211N9UE', u'holders': [], u'size': u'931.51 GB', u'scheduler_mode': u'deadline', u'rotational': u'1', u'sectors': u'1953525168', u'sas_address': None, u'virtual': 1, u'removable': u'0', u'model': u'HUA722010CLA330', u'partitions': {}}, 'key': u'sde'}, u'cmd': u'sgdisk --zap-all /dev/sde || sgdisk --zap-all /dev/sde', 'failed': False, u'stderr': u'', u'rc': 0, u'invocation': {u'module_args': {u'warn': True, u'executable': None, u'_uses_shell': True, u'strip_empty_ends': True, u'_raw_params': u'sgdisk --zap-all /dev/sde || sgdisk --zap-all /dev/sde', u'removes': None, u'argv': None, u'creates': None, u'chdir': None, u'stdin_add_newline': True, u'stdin': None}}, u'start': u'2019-11-05 22:06:03.802274'}, {'stderr_lines': [], u'changed': True, u'stdout': u'Creating new GPT entries.\nGPT data structures destroyed! You may now partition the disk using fdisk or\nother utilities.', u'delta': u'0:00:01.011442', 'stdout_lines': [u'Creating new GPT entries.', u'GPT data structures destroyed! You may now partition the disk using fdisk or', u'other utilities.'], '_ansible_item_label': {'value': {u'sectorsize': u'512', u'vendor': u'Seagate', u'links': {u'masters': [u'dm-0'], u'labels': [], u'ids': [u'scsi-2001b4d2000000000'], u'uuids': []}, u'sas_device_handle': None, u'host': u'RAID bus controller: Areca Technology Corp. ARC-1680 series PCIe to SAS/SATA 3Gb RAID Controller', u'support_discard': u'0', u'serial': u'5VP7GG81', u'holders': [u'mpatha'], u'size': u'931.51 GB', u'scheduler_mode': u'deadline', u'rotational': u'1', u'sectors': u'1953525168', u'sas_address': None, u'virtual': 1, u'removable': u'0', u'model': u'ST31000528AS', u'partitions': {}}, 'key': u'sdf'}, 'ansible_loop_var': u'item', u'end': u'2019-11-05 22:06:06.078571', '_ansible_no_log': False, 'item': {'value': {u'sectorsize': u'512', u'vendor': u'Seagate', u'links': {u'masters': [u'dm-0'], u'labels': [], u'ids': [u'scsi-2001b4d2000000000'], u'uuids': []}, u'sas_device_handle': None, u'host': u'RAID bus controller: Areca Technology Corp. ARC-1680 series PCIe to SAS/SATA 3Gb RAID Controller', u'support_discard': u'0', u'serial': u'5VP7GG81', u'holders': [u'mpatha'], u'size': u'931.51 GB', u'scheduler_mode': u'deadline', u'rotational': u'1', u'sectors': u'1953525168', u'sas_address': None, u'virtual': 1, u'removable': u'0', u'model': u'ST31000528AS', u'partitions': {}}, 'key': u'sdf'}, u'cmd': u'sgdisk --zap-all /dev/sdf || sgdisk --zap-all /dev/sdf', 'failed': False, u'stderr': u'', u'rc': 0, u'invocation': {u'module_args': {u'warn': True, u'executable': None, u'_uses_shell': True, u'strip_empty_ends': True, u'_raw_params': u'sgdisk --zap-all /dev/sdf || sgdisk --zap-all /dev/sdf', u'removes': None, u'argv': None, u'creates': None, u'chdir': None, u'stdin_add_newline': True, u'stdin': None}}, u'start': u'2019-11-05 22:06:05.067129'}, {'stderr_lines': [], u'changed': True, u'stdout': u'Creating new GPT entries.\nGPT data structures destroyed! You may now partition the disk using fdisk or\nother utilities.', u'delta': u'0:00:01.012013', 'stdout_lines': [u'Creating new GPT entries.', u'GPT data structures destroyed! You may now partition the disk using fdisk or', u'other utilities.'], '_ansible_item_label': {'value': {u'sectorsize': u'512', u'vendor': u'Hitachi', u'links': {u'masters': [u'dm-0'], u'labels': [], u'ids': [], u'uuids': [u'1dc5a03b-3af9-428f-8b65-90446f05166a']}, u'sas_device_handle': None, u'host': u'RAID bus controller: Areca Technology Corp. ARC-1680 series PCIe to SAS/SATA 3Gb RAID Controller', u'support_discard': u'0', u'serial': u'N020YX8L', u'holders': [u'mpatha'], u'size': u'931.51 GB', u'scheduler_mode': u'deadline', u'rotational': u'1', u'sectors': u'1953525168', u'sas_address': None, u'virtual': 1, u'removable': u'0', u'model': u'HUA722010CLA331', u'partitions': {}}, 'key': u'sdg'}, 'ansible_loop_var': u'item', u'end': u'2019-11-05 22:06:07.358263', '_ansible_no_log': False, 'item': {'value': {u'sectorsize': u'512', u'vendor': u'Hitachi', u'links': {u'masters': [u'dm-0'], u'labels': [], u'ids': [], u'uuids': [u'1dc5a03b-3af9-428f-8b65-90446f05166a']}, u'sas_device_handle': None, u'host': u'RAID bus controller: Areca Technology Corp. ARC-1680 series PCIe to SAS/SATA 3Gb RAID Controller', u'support_discard': u'0', u'serial': u'N020YX8L', u'holders': [u'mpatha'], u'size': u'931.51 GB', u'scheduler_mode': u'deadline', u'rotational': u'1', u'sectors': u'1953525168', u'sas_address': None, u'virtual': 1, u'removable': u'0', u'model': u'HUA722010CLA331', u'partitions': {}}, 'key': u'sdg'}, u'cmd': u'sgdisk --zap-all /dev/sdg || sgdisk --zap-all /dev/sdg', 'failed': False, u'stderr': u'', u'rc': 0, u'invocation': {u'module_args': {u'warn': True, u'executable': None, u'_uses_shell': True, u'strip_empty_ends': True, u'_raw_params': u'sgdisk --zap-all /dev/sdg || sgdisk --zap-all /dev/sdg', u'removes': None, u'argv': None, u'creates': None, u'chdir': None, u'stdin_add_newline': True, u'stdin': None}}, u'start': u'2019-11-05 22:06:06.346250'}, {'ansible_loop_var': u'item', '_ansible_no_log': False, 'skip_reason': u'Conditional result was False', 'item': {'value': {u'sectorsize': u'512', u'vendor': u'Seagate', u'links': {u'masters': [], u'labels': [], u'ids': [], u'uuids': []}, u'sas_device_handle': None, u'host': u'RAID bus controller: Areca Technology Corp. ARC-1680 series PCIe to SAS/SATA 3Gb RAID Controller', u'support_discard': u'0', u'serial': u'5VP4BY3N', u'holders': [], u'size': u'931.51 GB', u'scheduler_mode': u'deadline', u'rotational': u'1', u'sectors': u'1953525168', u'sas_address': None, u'virtual': 1, u'removable': u'0', u'model': u'ST31000528AS', u'partitions': {u'sda1': {u'start': u'2048', u'sectorsize': 512, u'uuid': u'f1bca609-9cd3-46ed-8f34-af0538e7246e', u'sectors': u'1953522688', u'holders': [], u'links': {u'masters': [], u'labels': [], u'ids': [u'scsi-2001b4d2000000000-part1'], u'uuids': [u'f1bca609-9cd3-46ed-8f34-af0538e7246e']}, u'size': u'931.51 GB'}}}, 'key': u'sda'}, 'skipped': True, 'changed': False, '_ansible_item_label': {'value': {u'sectorsize': u'512', u'vendor': u'Seagate', u'links': {u'masters': [], u'labels': [], u'ids': [], u'uuids': []}, u'sas_device_handle': None, u'host': u'RAID bus controller: Areca Technology Corp. ARC-1680 series PCIe to SAS/SATA 3Gb RAID Controller', u'support_discard': u'0', u'serial': u'5VP4BY3N', u'holders': [], u'size': u'931.51 GB', u'scheduler_mode': u'deadline', u'rotational': u'1', u'sectors': u'1953525168', u'sas_address': None, u'virtual': 1, u'removable': u'0', u'model': u'ST31000528AS', u'partitions': {u'sda1': {u'start': u'2048', u'sectorsize': 512, u'uuid': u'f1bca609-9cd3-46ed-8f34-af0538e7246e', u'sectors': u'1953522688', u'holders': [], u'links': {u'masters': [], u'labels': [], u'ids': [u'scsi-2001b4d2000000000-part1'], u'uuids': [u'f1bca609-9cd3-46ed-8f34-af0538e7246e']}, u'size': u'931.51 GB'}}}, 'key': u'sda'}}, {'stderr_lines': [], u'changed': True, u'stdout': u'Creating new GPT entries.\nGPT data structures destroyed! You may now partition the disk using fdisk or\nother utilities.', u'delta': u'0:00:01.029201', 'stdout_lines': [u'Creating new GPT entries.', u'GPT data structures destroyed! You may now partition the disk using fdisk or', u'other utilities.'], '_ansible_item_label': {'value': {u'sectorsize': u'512', u'vendor': u'Hitachi', u'links': {u'masters': [], u'labels': [], u'ids': [u'scsi-2001b4d20220ec500'], u'uuids': []}, u'sas_device_handle': None, u'host': u'RAID bus controller: Areca Technology Corp. ARC-1680 series PCIe to SAS/SATA 3Gb RAID Controller', u'support_discard': u'0', u'serial': u'JPS930N122NPEL', u'holders': [], u'size': u'931.51 GB', u'scheduler_mode': u'deadline', u'rotational': u'1', u'sectors': u'1953525168', u'sas_address': None, u'virtual': 1, u'removable': u'0', u'model': u'HDS721010CLA330', u'partitions': {}}, 'key': u'sdb'}, 'ansible_loop_var': u'item', u'end': u'2019-11-05 22:06:08.644042', '_ansible_no_log': False, 'item': {'value': {u'sectorsize': u'512', u'vendor': u'Hitachi', u'links': {u'masters': [], u'labels': [], u'ids': [u'scsi-2001b4d20220ec500'], u'uuids': []}, u'sas_device_handle': None, u'host': u'RAID bus controller: Areca Technology Corp. ARC-1680 series PCIe to SAS/SATA 3Gb RAID Controller', u'support_discard': u'0', u'serial': u'JPS930N122NPEL', u'holders': [], u'size': u'931.51 GB', u'scheduler_mode': u'deadline', u'rotational': u'1', u'sectors': u'1953525168', u'sas_address': None, u'virtual': 1, u'removable': u'0', u'model': u'HDS721010CLA330', u'partitions': {}}, 'key': u'sdb'}, u'cmd': u'sgdisk --zap-all /dev/sdb || sgdisk --zap-all /dev/sdb', 'failed': False, u'stderr': u'', u'rc': 0, u'invocation': {u'module_args': {u'warn': True, u'executable': None, u'_uses_shell': True, u'strip_empty_ends': True, u'_raw_params': u'sgdisk --zap-all /dev/sdb || sgdisk --zap-all /dev/sdb', u'removes': None, u'argv': None, u'creates': None, u'chdir': None, u'stdin_add_newline': True, u'stdin': None}}, u'start': u'2019-11-05 22:06:07.614841'}, {'stderr_lines': [], u'changed': True, u'stdout': u'Creating new GPT entries.\nGPT data structures destroyed! You may now partition the disk using fdisk or\nother utilities.', u'delta': u'0:00:01.012084', 'stdout_lines': [u'Creating new GPT entries.', u'GPT data structures destroyed! You may now partition the disk using fdisk or', u'other utilities.'], '_ansible_item_label': {'value': {u'sectorsize': u'512', u'vendor': u'Hitachi', u'links': {u'masters': [u'dm-0'], u'labels': [], u'ids': [], u'uuids': []}, u'sas_device_handle': None, u'host': u'RAID bus controller: Areca Technology Corp. ARC-1680 series PCIe to SAS/SATA 3Gb RAID Controller', u'support_discard': u'0', u'serial': u'N01XYGUL', u'holders': [u'mpatha'], u'size': u'931.51 GB', u'scheduler_mode': u'deadline', u'rotational': u'1', u'sectors': u'1953525168', u'sas_address': None, u'virtual': 1, u'removable': u'0', u'model': u'HUA722010CLA331', u'partitions': {}}, 'key': u'sdc'}, 'ansible_loop_var': u'item', u'end': u'2019-11-05 22:06:09.921627', '_ansible_no_log': False, 'item': {'value': {u'sectorsize': u'512', u'vendor': u'Hitachi', u'links': {u'masters': [u'dm-0'], u'labels': [], u'ids': [], u'uuids': []}, u'sas_device_handle': None, u'host': u'RAID bus controller: Areca Technology Corp. ARC-1680 series PCIe to SAS/SATA 3Gb RAID Controller', u'support_discard': u'0', u'serial': u'N01XYGUL', u'holders': [u'mpatha'], u'size': u'931.51 GB', u'scheduler_mode': u'deadline', u'rotational': u'1', u'sectors': u'1953525168', u'sas_address': None, u'virtual': 1, u'removable': u'0', u'model': u'HUA722010CLA331', u'partitions': {}}, 'key': u'sdc'}, u'cmd': u'sgdisk --zap-all /dev/sdc || sgdisk --zap-all /dev/sdc', 'failed': False, u'stderr': u'', u'rc': 0, u'invocation': {u'module_args': {u'warn': True, u'executable': None, u'_uses_shell': True, u'strip_empty_ends': True, u'_raw_params': u'sgdisk --zap-all /dev/sdc || sgdisk --zap-all /dev/sdc', u'removes': None, u'argv': None, u'creates': None, u'chdir': None, u'stdin_add_newline': True, u'stdin': None}}, u'start': u'2019-11-05 22:06:08.909543'}, {'stderr_lines': [], u'changed': True, u'stdout': u'Creating new GPT entries.\nGPT data structures destroyed! You may now partition the disk using fdisk or\nother utilities.', u'delta': u'0:00:01.011334', 'stdout_lines': [u'Creating new GPT entries.', u'GPT data structures destroyed! You may now partition the disk using fdisk or', u'other utilities.'], '_ansible_item_label': {'value': {u'sectorsize': u'512', u'vendor': u'Hitachi', u'links': {u'masters': [], u'labels': [], u'ids': [u'scsi-2001b4d202274c700'], u'uuids': [u'1cf94ad8-e480-41ca-81d7-8a2b419ab6f7']}, u'sas_device_handle': None, u'host': u'RAID bus controller: Areca Technology Corp. ARC-1680 series PCIe to SAS/SATA 3Gb RAID Controller', u'support_discard': u'0', u'serial': u'JPS930N122T7GL', u'holders': [], u'size': u'931.51 GB', u'scheduler_mode': u'deadline', u'rotational': u'1', u'sectors': u'1953525168', u'sas_address': None, u'virtual': 1, u'removable': u'0', u'model': u'HDS721010CLA330', u'partitions': {}}, 'key': u'sdh'}, 'ansible_loop_var': u'item', u'end': u'2019-11-05 22:06:11.178870', '_ansible_no_log': False, 'item': {'value': {u'sectorsize': u'512', u'vendor': u'Hitachi', u'links': {u'masters': [], u'labels': [], u'ids': [u'scsi-2001b4d202274c700'], u'uuids': [u'1cf94ad8-e480-41ca-81d7-8a2b419ab6f7']}, u'sas_device_handle': None, u'host': u'RAID bus controller: Areca Technology Corp. ARC-1680 series PCIe to SAS/SATA 3Gb RAID Controller', u'support_discard': u'0', u'serial': u'JPS930N122T7GL', u'holders': [], u'size': u'931.51 GB', u'scheduler_mode': u'deadline', u'rotational': u'1', u'sectors': u'1953525168', u'sas_address': None, u'virtual': 1, u'removable': u'0', u'model': u'HDS721010CLA330', u'partitions': {}}, 'key': u'sdh'}, u'cmd': u'sgdisk --zap-all /dev/sdh || sgdisk --zap-all /dev/sdh', 'failed': False, u'stderr': u'', u'rc': 0, u'invocation': {u'module_args': {u'warn': True, u'executable': None, u'_uses_shell': True, u'strip_empty_ends': True, u'_raw_params': u'sgdisk --zap-all /dev/sdh || sgdisk --zap-all /dev/sdh', u'removes': None, u'argv': None, u'creates': None, u'chdir': None, u'stdin_add_newline': True, u'stdin': None}}, u'start': u'2019-11-05 22:06:10.167536'}, {'stderr_lines': [u'Problem opening /dev/dm-0 for reading! Error is 2.', u'The specified file does not exist!', u"Problem opening '' for writing! Program will now terminate.", u'Warning! MBR not overwritten! Error is 2!', u'Problem opening /dev/dm-0 for reading! Error is 2.', u'The specified file does not exist!', u"Problem opening '' for writing! Program will now terminate.", u'Warning! MBR not overwritten! Error is 2!'], u'changed': True, u'stdout': u'', u'invocation': {u'module_args': {u'warn': True, u'executable': None, u'_uses_shell': True, u'strip_empty_ends': True, u'_raw_params': u'sgdisk --zap-all /dev/dm-0 || sgdisk --zap-all /dev/dm-0', u'removes': None, u'argv': None, u'creates': None, u'chdir': None, u'stdin_add_newline': True, u'stdin': None}}, u'delta': u'0:00:00.008504', 'stdout_lines': [], '_ansible_item_label': {'value': {u'sectorsize': u'512', u'vendor': None, u'links': {u'masters': [], u'labels': [], u'ids': [u'dm-name-mpatha', u'dm-uuid-mpath-2001b4d2000000000'], u'uuids': [u'de90d16e-b910-4057-90a2-0e711f786253']}, u'sas_device_handle': None, u'host': u'', u'support_discard': u'0', u'serial': u'5VP7GG81', u'holders': [], u'size': u'931.51 GB', u'scheduler_mode': u'deadline', u'rotational': u'1', u'sectors': u'1953525168', u'sas_address': None, u'virtual': 1, u'removable': u'0', u'model': None, u'partitions': {}}, 'key': u'dm-0'}, 'ansible_loop_var': u'item', u'end': u'2019-11-05 22:06:11.435082', '_ansible_no_log': False, u'start': u'2019-11-05 22:06:11.426578', u'failed': True, u'cmd': u'sgdisk --zap-all /dev/dm-0 || sgdisk --zap-all /dev/dm-0', 'item': {'value': {u'sectorsize': u'512', u'vendor': None, u'links': {u'masters': [], u'labels': [], u'ids': [u'dm-name-mpatha', u'dm-uuid-mpath-2001b4d2000000000'], u'uuids': [u'de90d16e-b910-4057-90a2-0e711f786253']}, u'sas_device_handle': None, u'host': u'', u'support_discard': u'0', u'serial': u'5VP7GG81', u'holders': [], u'size': u'931.51 GB', u'scheduler_mode': u'deadline', u'rotational': u'1', u'sectors': u'1953525168', u'sas_address': None, u'virtual': 1, u'removable': u'0', u'model': None, u'partitions': {}}, 'key': u'dm-0'}, u'stderr': u"Problem opening /dev/dm-0 for reading! Error is 2.\nThe specified file does not exist!\nProblem opening '' for writing! Program will now terminate.\nWarning! MBR not overwritten! Error is 2!\nProblem opening /dev/dm-0 for reading! Error is 2.\nThe specified file does not exist!\nProblem opening '' for writing! Program will now terminate.\nWarning! MBR not overwritten! Error is 2!", u'rc': 2, u'msg': u'non-zero return code'}]}}Traceback (most recent call last): File "/home/teuthworker/src/git.ceph.com_git_ceph-cm-ansible_master/callback_plugins/failure_log.py", line 44, in log_failure log.error(yaml.safe_dump(failure)) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/__init__.py", line 309, in safe_dump return dump_all([data], stream, Dumper=SafeDumper, **kwds) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/__init__.py", line 281, in dump_all dumper.represent(data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 29, in represent node = self.represent_data(data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 58, in represent_data node = self.yaml_representers[data_types[0]](self, data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 227, in represent_dict return self.represent_mapping(u'tag:yaml.org,2002:map', data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 125, in represent_mapping node_value = self.represent_data(item_value) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 58, in represent_data node = self.yaml_representers[data_types[0]](self, data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 227, in represent_dict return self.represent_mapping(u'tag:yaml.org,2002:map', data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 125, in represent_mapping node_value = self.represent_data(item_value) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 58, in represent_data node = self.yaml_representers[data_types[0]](self, data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 219, in represent_list return self.represent_sequence(u'tag:yaml.org,2002:seq', data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 102, in represent_sequence node_item = self.represent_data(item) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 58, in represent_data node = self.yaml_representers[data_types[0]](self, data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 227, in represent_dict return self.represent_mapping(u'tag:yaml.org,2002:map', data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 125, in represent_mapping node_value = self.represent_data(item_value) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 58, in represent_data node = self.yaml_representers[data_types[0]](self, data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 227, in represent_dict return self.represent_mapping(u'tag:yaml.org,2002:map', data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 125, in represent_mapping node_value = self.represent_data(item_value) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 68, in represent_data node = self.yaml_representers[None](self, data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 251, in represent_undefined raise RepresenterError("cannot represent an object", data)RepresenterError: ('cannot represent an object', u'sdd') |
||||||||||||||
pass | 4474393 | 2019-11-05 12:10:30 | 2019-11-05 19:34:38 | 2019-11-05 19:52:37 | 0:17:59 | 0:07:54 | 0:10:05 | mira | master | ubuntu | 18.04 | rados/singleton/{all/peer.yaml msgr-failures/few.yaml msgr/async-v1only.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}} | 1 | |
pass | 4474394 | 2019-11-05 12:10:31 | 2019-11-05 19:36:39 | 2019-11-05 20:44:39 | 1:08:00 | 0:38:18 | 0:29:42 | mira | master | centos | 7.6 | rados/thrash-old-clients/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-install/nautilus-v1only.yaml backoff/peering_and_degraded.yaml ceph.yaml clusters/{openstack.yaml three-plus-one.yaml} d-balancer/crush-compat.yaml distro$/{centos_latest.yaml} msgr-failures/few.yaml rados.yaml thrashers/none.yaml thrashosds-health.yaml workloads/snaps-few-objects.yaml} | 4 | |
pass | 4474395 | 2019-11-05 12:10:31 | 2019-11-05 19:38:29 | 2019-11-05 20:52:31 | 1:14:02 | 1:04:58 | 0:09:04 | mira | master | rhel | 7.7 | rados/dashboard/{clusters/{2-node-mgr.yaml} debug/mgr.yaml objectstore/bluestore-bitmap.yaml supported-random-distro$/{rhel_7.yaml} tasks/dashboard.yaml} | 2 | |
pass | 4474396 | 2019-11-05 12:10:32 | 2019-11-05 19:48:39 | 2019-11-05 20:10:39 | 0:22:00 | 0:11:57 | 0:10:03 | mira | master | ubuntu | 18.04 | rados/singleton-nomsgr/{all/ceph-daemon.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}} | 1 | |
pass | 4474397 | 2019-11-05 12:10:33 | 2019-11-05 19:50:31 | 2019-11-05 20:36:32 | 0:46:01 | 0:36:01 | 0:10:00 | mira | master | ubuntu | 18.04 | rados/standalone/{supported-random-distro$/{ubuntu_latest.yaml} workloads/erasure-code.yaml} | 1 | |
pass | 4474398 | 2019-11-05 12:10:34 | 2019-11-05 19:52:19 | 2019-11-05 20:32:20 | 0:40:01 | 0:28:40 | 0:11:21 | mira | master | ubuntu | 18.04 | rados/thrash-erasure-code-overwrites/{bluestore-bitmap.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} fast/fast.yaml msgr-failures/fastclose.yaml rados.yaml recovery-overrides/{more-async-partial-recovery.yaml} supported-random-distro$/{ubuntu_latest.yaml} thrashers/minsize_recovery.yaml thrashosds-health.yaml workloads/ec-small-objects-fast-read-overwrites.yaml} | 2 | |
pass | 4474399 | 2019-11-05 12:10:35 | 2019-11-05 19:52:39 | 2019-11-05 20:30:39 | 0:38:00 | 0:28:50 | 0:09:10 | mira | master | rhel | 7.7 | rados/thrash-erasure-code-shec/{ceph.yaml clusters/{fixed-4.yaml openstack.yaml} msgr-failures/osd-delay.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml recovery-overrides/{more-async-recovery.yaml} supported-random-distro$/{rhel_7.yaml} thrashers/default.yaml thrashosds-health.yaml workloads/ec-rados-plugin=shec-k=4-m=3-c=2.yaml} | 4 | |
pass | 4474400 | 2019-11-05 12:10:36 | 2019-11-05 20:11:09 | 2019-11-05 23:05:02 | 2:53:53 | 2:30:50 | 0:23:03 | mira | master | rhel | 7.7 | rados/singleton/{all/pg-autoscaler.yaml msgr-failures/many.yaml msgr/async-v2only.yaml objectstore/bluestore-stupid.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}} | 2 | |
pass | 4474401 | 2019-11-05 12:10:37 | 2019-11-05 20:11:08 | 2019-11-05 20:37:00 | 0:25:52 | 0:16:08 | 0:09:44 | mira | master | ubuntu | 18.04 | rados/perf/{ceph.yaml objectstore/bluestore-comp.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{ubuntu_latest.yaml} workloads/cosbench_64K_write.yaml} | 1 | |
pass | 4474402 | 2019-11-05 12:10:37 | 2019-11-05 20:30:53 | 2019-11-05 23:42:55 | 3:12:02 | 2:53:04 | 0:18:58 | mira | master | rhel | 7.7 | rados/thrash-erasure-code/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} fast/normal.yaml msgr-failures/osd-delay.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml recovery-overrides/{more-partial-recovery.yaml} supported-random-distro$/{rhel_7.yaml} thrashers/morepggrow.yaml thrashosds-health.yaml workloads/ec-rados-plugin=jerasure-k=2-m=1.yaml} | 2 | |
pass | 4474403 | 2019-11-05 12:10:38 | 2019-11-05 20:32:45 | 2019-11-05 21:12:43 | 0:39:58 | 0:33:54 | 0:06:04 | mira | master | rhel | 7.7 | rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml 2-recovery-overrides/{default.yaml} backoff/normal.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/off.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{rhel_7.yaml} thrashers/careful.yaml thrashosds-health.yaml workloads/write_fadvise_dontneed.yaml} | 2 | |
pass | 4474404 | 2019-11-05 12:10:39 | 2019-11-05 20:36:38 | 2019-11-05 21:12:37 | 0:35:59 | 0:22:26 | 0:13:33 | mira | master | centos | 7.6 | rados/basic/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/few.yaml msgr/async.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml supported-random-distro$/{centos_7.yaml} tasks/readwrite.yaml} | 2 | |
pass | 4474405 | 2019-11-05 12:10:40 | 2019-11-05 20:37:04 | 2019-11-05 21:03:02 | 0:25:58 | 0:14:36 | 0:11:22 | mira | master | centos | 7.6 | rados/singleton/{all/pg-removal-interruption.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/filestore-xfs.yaml rados.yaml supported-random-distro$/{centos_7.yaml}} | 1 | |
pass | 4474406 | 2019-11-05 12:10:41 | 2019-11-05 20:44:44 | 2019-11-05 21:06:43 | 0:21:59 | 0:07:25 | 0:14:34 | mira | master | ubuntu | 18.04 | rados/multimon/{clusters/21.yaml msgr-failures/few.yaml msgr/async-v2only.yaml no_pools.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} tasks/mon_clock_with_skews.yaml} | 3 | |
pass | 4474407 | 2019-11-05 12:10:42 | 2019-11-05 20:52:34 | 2019-11-05 21:10:34 | 0:18:00 | 0:07:32 | 0:10:28 | mira | master | ubuntu | 18.04 | rados/objectstore/{backends/filejournal.yaml supported-random-distro$/{ubuntu_latest.yaml}} | 1 | |
pass | 4474408 | 2019-11-05 12:10:43 | 2019-11-05 21:03:04 | 2019-11-05 21:31:03 | 0:27:59 | 0:15:52 | 0:12:07 | mira | master | centos | 7.6 | rados/singleton-nomsgr/{all/ceph-kvstore-tool.yaml rados.yaml supported-random-distro$/{centos_7.yaml}} | 1 | |
pass | 4474409 | 2019-11-05 12:10:44 | 2019-11-05 21:06:45 | 2019-11-05 21:30:44 | 0:23:59 | 0:13:55 | 0:10:04 | mira | master | ubuntu | 18.04 | rados/mgr/{clusters/{2-node-mgr.yaml} debug/mgr.yaml objectstore/bluestore-comp.yaml supported-random-distro$/{ubuntu_latest.yaml} tasks/progress.yaml} | 2 | |
fail | 4474410 | 2019-11-05 12:10:45 | 2019-11-05 21:10:58 | 2019-11-05 23:48:58 | 2:38:00 | 2:15:56 | 0:22:04 | mira | master | rhel | 7.7 | rados/thrash-erasure-code-isa/{arch/x86_64.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/few.yaml objectstore/bluestore-stupid.yaml rados.yaml recovery-overrides/{more-partial-recovery.yaml} supported-random-distro$/{rhel_7.yaml} thrashers/careful.yaml thrashosds-health.yaml workloads/ec-rados-plugin=isa-k=2-m=1.yaml} | 2 | |
Failure Reason:
k --zap-all /dev/sdd', u'removes': None, u'argv': None, u'creates': None, u'chdir': None, u'stdin_add_newline': True, u'stdin': None}}, u'start': u'2019-11-05 23:47:17.485992'}, {'stderr_lines': [], u'changed': True, u'stdout': u'Creating new GPT entries.\nGPT data structures destroyed! You may now partition the disk using fdisk or\nother utilities.', u'delta': u'0:00:01.011233', 'stdout_lines': [u'Creating new GPT entries.', u'GPT data structures destroyed! You may now partition the disk using fdisk or', u'other utilities.'], '_ansible_item_label': {'value': {u'sectorsize': u'512', u'vendor': u'Hitachi', u'links': {u'masters': [], u'labels': [], u'ids': [u'scsi-2001b4d2001655500'], u'uuids': []}, u'sas_device_handle': None, u'host': u'RAID bus controller: Areca Technology Corp. ARC-1680 series PCIe to SAS/SATA 3Gb RAID Controller', u'support_discard': u'0', u'serial': u'JPW9K0N210EV5E', u'holders': [], u'size': u'931.51 GB', u'scheduler_mode': u'deadline', u'rotational': u'1', u'sectors': u'1953525168', u'sas_address': None, u'virtual': 1, u'removable': u'0', u'model': u'HUA722010CLA330', u'partitions': {}}, 'key': u'sde'}, 'ansible_loop_var': u'item', u'end': u'2019-11-05 23:47:19.802604', '_ansible_no_log': False, 'item': {'value': {u'sectorsize': u'512', u'vendor': u'Hitachi', u'links': {u'masters': [], u'labels': [], u'ids': [u'scsi-2001b4d2001655500'], u'uuids': []}, u'sas_device_handle': None, u'host': u'RAID bus controller: Areca Technology Corp. ARC-1680 series PCIe to SAS/SATA 3Gb RAID Controller', u'support_discard': u'0', u'serial': u'JPW9K0N210EV5E', u'holders': [], u'size': u'931.51 GB', u'scheduler_mode': u'deadline', u'rotational': u'1', u'sectors': u'1953525168', u'sas_address': None, u'virtual': 1, u'removable': u'0', u'model': u'HUA722010CLA330', u'partitions': {}}, 'key': u'sde'}, u'cmd': u'sgdisk --zap-all /dev/sde || sgdisk --zap-all /dev/sde', 'failed': False, u'stderr': u'', u'rc': 0, u'invocation': {u'module_args': {u'warn': True, u'executable': None, u'_uses_shell': True, u'strip_empty_ends': True, u'_raw_params': u'sgdisk --zap-all /dev/sde || sgdisk --zap-all /dev/sde', u'removes': None, u'argv': None, u'creates': None, u'chdir': None, u'stdin_add_newline': True, u'stdin': None}}, u'start': u'2019-11-05 23:47:18.791371'}, {'stderr_lines': [], u'changed': True, u'stdout': u'Creating new GPT entries.\nGPT data structures destroyed! You may now partition the disk using fdisk or\nother utilities.', u'delta': u'0:00:01.011521', 'stdout_lines': [u'Creating new GPT entries.', u'GPT data structures destroyed! You may now partition the disk using fdisk or', u'other utilities.'], '_ansible_item_label': {'value': {u'sectorsize': u'512', u'vendor': u'Hitachi', u'links': {u'masters': [], u'labels': [], u'ids': [u'scsi-2001b4d2012776300'], u'uuids': []}, u'sas_device_handle': None, u'host': u'RAID bus controller: Areca Technology Corp. ARC-1680 series PCIe to SAS/SATA 3Gb RAID Controller', u'support_discard': u'0', u'serial': u'JPS930N121G73V', u'holders': [], u'size': u'931.51 GB', u'scheduler_mode': u'deadline', u'rotational': u'1', u'sectors': u'1953525168', u'sas_address': None, u'virtual': 1, u'removable': u'0', u'model': u'HDS721010CLA330', u'partitions': {}}, 'key': u'sdf'}, 'ansible_loop_var': u'item', u'end': u'2019-11-05 23:47:21.066666', '_ansible_no_log': False, 'item': {'value': {u'sectorsize': u'512', u'vendor': u'Hitachi', u'links': {u'masters': [], u'labels': [], u'ids': [u'scsi-2001b4d2012776300'], u'uuids': []}, u'sas_device_handle': None, u'host': u'RAID bus controller: Areca Technology Corp. ARC-1680 series PCIe to SAS/SATA 3Gb RAID Controller', u'support_discard': u'0', u'serial': u'JPS930N121G73V', u'holders': [], u'size': u'931.51 GB', u'scheduler_mode': u'deadline', u'rotational': u'1', u'sectors': u'1953525168', u'sas_address': None, u'virtual': 1, u'removable': u'0', u'model': u'HDS721010CLA330', u'partitions': {}}, 'key': u'sdf'}, u'cmd': u'sgdisk --zap-all /dev/sdf || sgdisk --zap-all /dev/sdf', 'failed': False, u'stderr': u'', u'rc': 0, u'invocation': {u'module_args': {u'warn': True, u'executable': None, u'_uses_shell': True, u'strip_empty_ends': True, u'_raw_params': u'sgdisk --zap-all /dev/sdf || sgdisk --zap-all /dev/sdf', u'removes': None, u'argv': None, u'creates': None, u'chdir': None, u'stdin_add_newline': True, u'stdin': None}}, u'start': u'2019-11-05 23:47:20.055145'}, {'stderr_lines': [], u'changed': True, u'stdout': u'Creating new GPT entries.\nGPT data structures destroyed! You may now partition the disk using fdisk or\nother utilities.', u'delta': u'0:00:01.023926', 'stdout_lines': [u'Creating new GPT entries.', u'GPT data structures destroyed! You may now partition the disk using fdisk or', u'other utilities.'], '_ansible_item_label': {'value': {u'sectorsize': u'512', u'vendor': u'NA', u'links': {u'masters': [u'dm-0'], u'labels': [], u'ids': [u'scsi-2001b4d2000000000'], u'uuids': []}, u'sas_device_handle': None, u'host': u'RAID bus controller: Areca Technology Corp. ARC-1680 series PCIe to SAS/SATA 3Gb RAID Controller', u'support_discard': u'0', u'serial': u'PAJ55T7E', u'holders': [u'mpatha'], u'size': u'931.51 GB', u'scheduler_mode': u'deadline', u'rotational': u'1', u'sectors': u'1953525168', u'sas_address': None, u'virtual': 1, u'removable': u'0', u'model': u'HUA721010KLA330', u'partitions': {}}, 'key': u'sdg'}, 'ansible_loop_var': u'item', u'end': u'2019-11-05 23:47:22.337839', '_ansible_no_log': False, 'item': {'value': {u'sectorsize': u'512', u'vendor': u'NA', u'links': {u'masters': [u'dm-0'], u'labels': [], u'ids': [u'scsi-2001b4d2000000000'], u'uuids': []}, u'sas_device_handle': None, u'host': u'RAID bus controller: Areca Technology Corp. ARC-1680 series PCIe to SAS/SATA 3Gb RAID Controller', u'support_discard': u'0', u'serial': u'PAJ55T7E', u'holders': [u'mpatha'], u'size': u'931.51 GB', u'scheduler_mode': u'deadline', u'rotational': u'1', u'sectors': u'1953525168', u'sas_address': None, u'virtual': 1, u'removable': u'0', u'model': u'HUA721010KLA330', u'partitions': {}}, 'key': u'sdg'}, u'cmd': u'sgdisk --zap-all /dev/sdg || sgdisk --zap-all /dev/sdg', 'failed': False, u'stderr': u'', u'rc': 0, u'invocation': {u'module_args': {u'warn': True, u'executable': None, u'_uses_shell': True, u'strip_empty_ends': True, u'_raw_params': u'sgdisk --zap-all /dev/sdg || sgdisk --zap-all /dev/sdg', u'removes': None, u'argv': None, u'creates': None, u'chdir': None, u'stdin_add_newline': True, u'stdin': None}}, u'start': u'2019-11-05 23:47:21.313913'}, {'ansible_loop_var': u'item', '_ansible_no_log': False, 'skip_reason': u'Conditional result was False', 'item': {'value': {u'sectorsize': u'512', u'vendor': u'Seagate', u'links': {u'masters': [], u'labels': [], u'ids': [], u'uuids': []}, u'sas_device_handle': None, u'host': u'RAID bus controller: Areca Technology Corp. ARC-1680 series PCIe to SAS/SATA 3Gb RAID Controller', u'support_discard': u'0', u'serial': u'5VP52BEJ', u'holders': [], u'size': u'931.51 GB', u'scheduler_mode': u'deadline', u'rotational': u'1', u'sectors': u'1953525168', u'sas_address': None, u'virtual': 1, u'removable': u'0', u'model': u'ST31000528AS', u'partitions': {u'sda1': {u'start': u'2048', u'sectorsize': 512, u'uuid': u'f1bca609-9cd3-46ed-8f34-af0538e7246e', u'sectors': u'1953522688', u'holders': [], u'links': {u'masters': [], u'labels': [], u'ids': [u'scsi-2001b4d2000000000-part1'], u'uuids': [u'f1bca609-9cd3-46ed-8f34-af0538e7246e']}, u'size': u'931.51 GB'}}}, 'key': u'sda'}, 'skipped': True, 'changed': False, '_ansible_item_label': {'value': {u'sectorsize': u'512', u'vendor': u'Seagate', u'links': {u'masters': [], u'labels': [], u'ids': [], u'uuids': []}, u'sas_device_handle': None, u'host': u'RAID bus controller: Areca Technology Corp. ARC-1680 series PCIe to SAS/SATA 3Gb RAID Controller', u'support_discard': u'0', u'serial': u'5VP52BEJ', u'holders': [], u'size': u'931.51 GB', u'scheduler_mode': u'deadline', u'rotational': u'1', u'sectors': u'1953525168', u'sas_address': None, u'virtual': 1, u'removable': u'0', u'model': u'ST31000528AS', u'partitions': {u'sda1': {u'start': u'2048', u'sectorsize': 512, u'uuid': u'f1bca609-9cd3-46ed-8f34-af0538e7246e', u'sectors': u'1953522688', u'holders': [], u'links': {u'masters': [], u'labels': [], u'ids': [u'scsi-2001b4d2000000000-part1'], u'uuids': [u'f1bca609-9cd3-46ed-8f34-af0538e7246e']}, u'size': u'931.51 GB'}}}, 'key': u'sda'}}, {'stderr_lines': [], u'changed': True, u'stdout': u'Creating new GPT entries.\nGPT data structures destroyed! You may now partition the disk using fdisk or\nother utilities.', u'delta': u'0:00:01.012288', 'stdout_lines': [u'Creating new GPT entries.', u'GPT data structures destroyed! You may now partition the disk using fdisk or', u'other utilities.'], '_ansible_item_label': {'value': {u'sectorsize': u'512', u'vendor': u'Seagate', u'links': {u'masters': [u'dm-0'], u'labels': [], u'ids': [], u'uuids': []}, u'sas_device_handle': None, u'host': u'RAID bus controller: Areca Technology Corp. ARC-1680 series PCIe to SAS/SATA 3Gb RAID Controller', u'support_discard': u'0', u'serial': u'5VP53FPZ', u'holders': [u'mpatha'], u'size': u'931.51 GB', u'scheduler_mode': u'deadline', u'rotational': u'1', u'sectors': u'1953525168', u'sas_address': None, u'virtual': 1, u'removable': u'0', u'model': u'ST31000528AS', u'partitions': {}}, 'key': u'sdb'}, 'ansible_loop_var': u'item', u'end': u'2019-11-05 23:47:23.624342', '_ansible_no_log': False, 'item': {'value': {u'sectorsize': u'512', u'vendor': u'Seagate', u'links': {u'masters': [u'dm-0'], u'labels': [], u'ids': [], u'uuids': []}, u'sas_device_handle': None, u'host': u'RAID bus controller: Areca Technology Corp. ARC-1680 series PCIe to SAS/SATA 3Gb RAID Controller', u'support_discard': u'0', u'serial': u'5VP53FPZ', u'holders': [u'mpatha'], u'size': u'931.51 GB', u'scheduler_mode': u'deadline', u'rotational': u'1', u'sectors': u'1953525168', u'sas_address': None, u'virtual': 1, u'removable': u'0', u'model': u'ST31000528AS', u'partitions': {}}, 'key': u'sdb'}, u'cmd': u'sgdisk --zap-all /dev/sdb || sgdisk --zap-all /dev/sdb', 'failed': False, u'stderr': u'', u'rc': 0, u'invocation': {u'module_args': {u'warn': True, u'executable': None, u'_uses_shell': True, u'strip_empty_ends': True, u'_raw_params': u'sgdisk --zap-all /dev/sdb || sgdisk --zap-all /dev/sdb', u'removes': None, u'argv': None, u'creates': None, u'chdir': None, u'stdin_add_newline': True, u'stdin': None}}, u'start': u'2019-11-05 23:47:22.612054'}, {'stderr_lines': [], u'changed': True, u'stdout': u'Creating new GPT entries.\nGPT data structures destroyed! You may now partition the disk using fdisk or\nother utilities.', u'delta': u'0:00:01.011938', 'stdout_lines': [u'Creating new GPT entries.', u'GPT data structures destroyed! You may now partition the disk using fdisk or', u'other utilities.'], '_ansible_item_label': {'value': {u'sectorsize': u'512', u'vendor': u'Hitachi', u'links': {u'masters': [], u'labels': [], u'ids': [u'scsi-2001b4d208263c000'], u'uuids': []}, u'sas_device_handle': None, u'host': u'RAID bus controller: Areca Technology Corp. ARC-1680 series PCIe to SAS/SATA 3Gb RAID Controller', u'support_discard': u'0', u'serial': u'JPW9K0HD2H3VPL', u'holders': [], u'size': u'931.51 GB', u'scheduler_mode': u'deadline', u'rotational': u'1', u'sectors': u'1953525168', u'sas_address': None, u'virtual': 1, u'removable': u'0', u'model': u'HUA722010CLA330', u'partitions': {}}, 'key': u'sdc'}, 'ansible_loop_var': u'item', u'end': u'2019-11-05 23:47:24.883290', '_ansible_no_log': False, 'item': {'value': {u'sectorsize': u'512', u'vendor': u'Hitachi', u'links': {u'masters': [], u'labels': [], u'ids': [u'scsi-2001b4d208263c000'], u'uuids': []}, u'sas_device_handle': None, u'host': u'RAID bus controller: Areca Technology Corp. ARC-1680 series PCIe to SAS/SATA 3Gb RAID Controller', u'support_discard': u'0', u'serial': u'JPW9K0HD2H3VPL', u'holders': [], u'size': u'931.51 GB', u'scheduler_mode': u'deadline', u'rotational': u'1', u'sectors': u'1953525168', u'sas_address': None, u'virtual': 1, u'removable': u'0', u'model': u'HUA722010CLA330', u'partitions': {}}, 'key': u'sdc'}, u'cmd': u'sgdisk --zap-all /dev/sdc || sgdisk --zap-all /dev/sdc', 'failed': False, u'stderr': u'', u'rc': 0, u'invocation': {u'module_args': {u'warn': True, u'executable': None, u'_uses_shell': True, u'strip_empty_ends': True, u'_raw_params': u'sgdisk --zap-all /dev/sdc || sgdisk --zap-all /dev/sdc', u'removes': None, u'argv': None, u'creates': None, u'chdir': None, u'stdin_add_newline': True, u'stdin': None}}, u'start': u'2019-11-05 23:47:23.871352'}, {'stderr_lines': [u'Problem opening /dev/dm-0 for reading! Error is 2.', u'The specified file does not exist!', u"Problem opening '' for writing! Program will now terminate.", u'Warning! MBR not overwritten! Error is 2!', u'Problem opening /dev/dm-0 for reading! Error is 2.', u'The specified file does not exist!', u"Problem opening '' for writing! Program will now terminate.", u'Warning! MBR not overwritten! Error is 2!'], u'changed': True, u'stdout': u'', u'invocation': {u'module_args': {u'warn': True, u'executable': None, u'_uses_shell': True, u'strip_empty_ends': True, u'_raw_params': u'sgdisk --zap-all /dev/dm-0 || sgdisk --zap-all /dev/dm-0', u'removes': None, u'argv': None, u'creates': None, u'chdir': None, u'stdin_add_newline': True, u'stdin': None}}, u'delta': u'0:00:00.008650', 'stdout_lines': [], '_ansible_item_label': {'value': {u'sectorsize': u'512', u'vendor': None, u'links': {u'masters': [], u'labels': [], u'ids': [u'dm-name-mpatha', u'dm-uuid-mpath-2001b4d2000000000'], u'uuids': []}, u'sas_device_handle': None, u'host': u'', u'support_discard': u'0', u'serial': u'PAJ55T7E', u'holders': [], u'size': u'931.51 GB', u'scheduler_mode': u'deadline', u'rotational': u'1', u'sectors': u'1953525168', u'sas_address': None, u'virtual': 1, u'removable': u'0', u'model': None, u'partitions': {}}, 'key': u'dm-0'}, 'ansible_loop_var': u'item', u'end': u'2019-11-05 23:47:25.148862', '_ansible_no_log': False, u'start': u'2019-11-05 23:47:25.140212', u'failed': True, u'cmd': u'sgdisk --zap-all /dev/dm-0 || sgdisk --zap-all /dev/dm-0', 'item': {'value': {u'sectorsize': u'512', u'vendor': None, u'links': {u'masters': [], u'labels': [], u'ids': [u'dm-name-mpatha', u'dm-uuid-mpath-2001b4d2000000000'], u'uuids': []}, u'sas_device_handle': None, u'host': u'', u'support_discard': u'0', u'serial': u'PAJ55T7E', u'holders': [], u'size': u'931.51 GB', u'scheduler_mode': u'deadline', u'rotational': u'1', u'sectors': u'1953525168', u'sas_address': None, u'virtual': 1, u'removable': u'0', u'model': None, u'partitions': {}}, 'key': u'dm-0'}, u'stderr': u"Problem opening /dev/dm-0 for reading! Error is 2.\nThe specified file does not exist!\nProblem opening '' for writing! Program will now terminate.\nWarning! MBR not overwritten! Error is 2!\nProblem opening /dev/dm-0 for reading! Error is 2.\nThe specified file does not exist!\nProblem opening '' for writing! Program will now terminate.\nWarning! MBR not overwritten! Error is 2!", u'rc': 2, u'msg': u'non-zero return code'}]}}Traceback (most recent call last): File "/home/teuthworker/src/git.ceph.com_git_ceph-cm-ansible_master/callback_plugins/failure_log.py", line 44, in log_failure log.error(yaml.safe_dump(failure)) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/__init__.py", line 309, in safe_dump return dump_all([data], stream, Dumper=SafeDumper, **kwds) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/__init__.py", line 281, in dump_all dumper.represent(data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 29, in represent node = self.represent_data(data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 58, in represent_data node = self.yaml_representers[data_types[0]](self, data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 227, in represent_dict return self.represent_mapping(u'tag:yaml.org,2002:map', data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 125, in represent_mapping node_value = self.represent_data(item_value) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 58, in represent_data node = self.yaml_representers[data_types[0]](self, data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 227, in represent_dict return self.represent_mapping(u'tag:yaml.org,2002:map', data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 125, in represent_mapping node_value = self.represent_data(item_value) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 58, in represent_data node = self.yaml_representers[data_types[0]](self, data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 219, in represent_list return self.represent_sequence(u'tag:yaml.org,2002:seq', data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 102, in represent_sequence node_item = self.represent_data(item) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 58, in represent_data node = self.yaml_representers[data_types[0]](self, data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 227, in represent_dict return self.represent_mapping(u'tag:yaml.org,2002:map', data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 125, in represent_mapping node_value = self.represent_data(item_value) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 58, in represent_data node = self.yaml_representers[data_types[0]](self, data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 227, in represent_dict return self.represent_mapping(u'tag:yaml.org,2002:map', data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 125, in represent_mapping node_value = self.represent_data(item_value) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 68, in represent_data node = self.yaml_representers[None](self, data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/local/lib/python2.7/site-packages/yaml/representer.py", line 251, in represent_undefined raise RepresenterError("cannot represent an object", data)RepresenterError: ('cannot represent an object', u'sdd') |
||||||||||||||
pass | 4474411 | 2019-11-05 12:10:45 | 2019-11-05 21:12:59 | 2019-11-05 21:50:58 | 0:37:59 | 0:27:50 | 0:10:09 | mira | master | centos | 7.6 | rados/singleton/{all/radostool.yaml msgr-failures/many.yaml msgr/async-v1only.yaml objectstore/bluestore-avl.yaml rados.yaml supported-random-distro$/{centos_7.yaml}} | 1 | |
pass | 4474412 | 2019-11-05 12:10:46 | 2019-11-05 21:12:59 | 2019-11-05 21:38:58 | 0:25:59 | 0:16:21 | 0:09:38 | mira | master | centos | 7.6 | rados/perf/{ceph.yaml objectstore/bluestore-low-osd-mem-target.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{centos_7.yaml} workloads/fio_4K_rand_read.yaml} | 1 | |
pass | 4474413 | 2019-11-05 12:10:47 | 2019-11-05 21:30:47 | 2019-11-05 21:54:46 | 0:23:59 | 0:14:14 | 0:09:45 | mira | master | ubuntu | 18.04 | rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml 2-recovery-overrides/{more-partial-recovery.yaml} backoff/peering.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/upmap.yaml msgr-failures/osd-delay.yaml msgr/async-v1only.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} thrashers/default.yaml thrashosds-health.yaml workloads/admin_socket_objecter_requests.yaml} | 2 | |
pass | 4474414 | 2019-11-05 12:10:48 | 2019-11-05 21:31:05 | 2019-11-05 22:11:05 | 0:40:00 | 0:22:05 | 0:17:55 | mira | master | centos | 7.6 | rados/thrash-old-clients/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-install/nautilus-v2only.yaml backoff/normal.yaml ceph.yaml clusters/{openstack.yaml three-plus-one.yaml} d-balancer/off.yaml distro$/{centos_latest.yaml} msgr-failures/osd-delay.yaml rados.yaml thrashers/pggrow.yaml thrashosds-health.yaml workloads/test_rbd_api.yaml} | 4 | |
pass | 4474415 | 2019-11-05 12:10:49 | 2019-11-05 21:39:00 | 2019-11-05 22:33:00 | 0:54:00 | 0:44:15 | 0:09:45 | mira | master | rhel | 7.7 | rados/thrash-erasure-code-big/{ceph.yaml cluster/{12-osds.yaml openstack.yaml} msgr-failures/fastclose.yaml objectstore/filestore-xfs.yaml rados.yaml recovery-overrides/{more-active-recovery.yaml} supported-random-distro$/{rhel_7.yaml} thrashers/fastread.yaml thrashosds-health.yaml workloads/ec-rados-plugin=jerasure-k=4-m=2.yaml} | 3 | |
dead | 4474416 | 2019-11-05 12:10:50 | 2019-11-05 21:51:17 | 2019-11-06 09:53:44 | 12:02:27 | mira | master | centos | rados/verify/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-thrash/default/{default.yaml thrashosds-health.yaml} msgr-failures/few.yaml msgr/async-v1only.yaml objectstore/bluestore-comp.yaml rados.yaml tasks/rados_cls_all.yaml validater/valgrind.yaml} | 2 | ||||
pass | 4474417 | 2019-11-05 12:10:51 | 2019-11-05 21:55:07 | 2019-11-05 22:51:06 | 0:55:59 | 0:44:56 | 0:11:03 | mira | master | centos | 7.6 | rados/singleton-bluestore/{all/cephtool.yaml msgr-failures/few.yaml msgr/async-v1only.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{centos_7.yaml}} | 1 | |
pass | 4474418 | 2019-11-05 12:10:52 | 2019-11-05 22:06:48 | 2019-11-05 22:46:47 | 0:39:59 | 0:25:01 | 0:14:58 | mira | master | centos | 7.6 | rados/singleton/{all/random-eio.yaml msgr-failures/few.yaml msgr/async-v2only.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{centos_7.yaml}} | 2 | |
pass | 4474419 | 2019-11-05 12:10:52 | 2019-11-05 22:11:06 | 2019-11-05 23:15:06 | 1:04:00 | 0:40:53 | 0:23:07 | mira | master | centos | 7.6 | rados/monthrash/{ceph.yaml clusters/9-mons.yaml msgr-failures/few.yaml msgr/async-v2only.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{centos_7.yaml} thrashers/sync-many.yaml workloads/snaps-few-objects.yaml} | 2 | |
pass | 4474420 | 2019-11-05 12:10:53 | 2019-11-05 22:33:02 | 2019-11-05 23:03:01 | 0:29:59 | 0:18:52 | 0:11:07 | mira | master | ubuntu | 18.04 | rados/basic/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/many.yaml msgr/async-v1only.yaml objectstore/bluestore-stupid.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} tasks/repair_test.yaml} | 2 | |
pass | 4474421 | 2019-11-05 12:10:54 | 2019-11-05 22:47:04 | 2019-11-05 23:03:03 | 0:15:59 | 0:06:57 | 0:09:02 | mira | master | ubuntu | 18.04 | rados/singleton-nomsgr/{all/ceph-post-file.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}} | 1 | |
pass | 4474422 | 2019-11-05 12:10:55 | 2019-11-05 22:51:21 | 2019-11-05 23:47:21 | 0:56:00 | 0:43:29 | 0:12:31 | mira | master | centos | 7.6 | rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml 2-recovery-overrides/{more-async-recovery.yaml} backoff/peering_and_degraded.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/crush-compat.yaml msgr-failures/fastclose.yaml msgr/async-v2only.yaml objectstore/bluestore-stupid.yaml rados.yaml supported-random-distro$/{centos_7.yaml} thrashers/mapgap.yaml thrashosds-health.yaml workloads/cache-agent-big.yaml} | 2 | |
pass | 4474423 | 2019-11-05 12:10:56 | 2019-11-05 23:03:03 | 2019-11-05 23:23:02 | 0:19:59 | 0:09:27 | 0:10:32 | mira | master | ubuntu | 18.04 | rados/perf/{ceph.yaml objectstore/bluestore-stupid.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{ubuntu_latest.yaml} workloads/fio_4K_rand_rw.yaml} | 1 | |
fail | 4474424 | 2019-11-05 12:10:57 | 2019-11-05 23:03:05 | 2019-11-05 23:23:04 | 0:19:59 | 0:10:16 | 0:09:43 | mira | master | ubuntu | 18.04 | rados/singleton/{all/rebuild-mondb.yaml msgr-failures/many.yaml msgr/async.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}} | 1 | |
Failure Reason:
Command failed on mira063 with status 13: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json' |
||||||||||||||
pass | 4474425 | 2019-11-05 12:10:58 | 2019-11-05 23:05:21 | 2019-11-05 23:59:20 | 0:53:59 | 0:40:08 | 0:13:51 | mira | master | centos | 7.6 | rados/thrash-erasure-code/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} fast/fast.yaml msgr-failures/fastclose.yaml objectstore/bluestore-stupid.yaml rados.yaml recovery-overrides/{more-partial-recovery.yaml} supported-random-distro$/{centos_7.yaml} thrashers/pggrow.yaml thrashosds-health.yaml workloads/ec-rados-plugin=jerasure-k=3-m=1.yaml} | 2 | |
pass | 4474426 | 2019-11-05 12:10:59 | 2019-11-05 23:15:08 | 2019-11-05 23:37:07 | 0:21:59 | 0:10:08 | 0:11:51 | mira | master | ubuntu | 18.04 | rados/mgr/{clusters/{2-node-mgr.yaml} debug/mgr.yaml objectstore/bluestore-low-osd-mem-target.yaml supported-random-distro$/{ubuntu_latest.yaml} tasks/prometheus.yaml} | 2 | |
pass | 4474427 | 2019-11-05 12:10:59 | 2019-11-05 23:23:20 | 2019-11-06 00:15:19 | 0:51:59 | 0:37:08 | 0:14:51 | mira | master | centos | 7.6 | rados/thrash-erasure-code-overwrites/{bluestore-bitmap.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} fast/normal.yaml msgr-failures/few.yaml rados.yaml recovery-overrides/{more-async-partial-recovery.yaml} supported-random-distro$/{centos_7.yaml} thrashers/morepggrow.yaml thrashosds-health.yaml workloads/ec-small-objects-overwrites.yaml} | 2 | |
pass | 4474428 | 2019-11-05 12:11:00 | 2019-11-05 23:23:20 | 2019-11-06 00:01:19 | 0:37:59 | 0:29:01 | 0:08:58 | mira | master | rhel | 7.7 | rados/thrash-erasure-code-shec/{ceph.yaml clusters/{fixed-4.yaml openstack.yaml} msgr-failures/fastclose.yaml objectstore/bluestore-stupid.yaml rados.yaml recovery-overrides/{more-async-partial-recovery.yaml} supported-random-distro$/{rhel_7.yaml} thrashers/careful.yaml thrashosds-health.yaml workloads/ec-rados-plugin=shec-k=4-m=3-c=2.yaml} | 4 | |
pass | 4474429 | 2019-11-05 12:11:01 | 2019-11-05 23:37:19 | 2019-11-06 00:27:18 | 0:49:59 | 0:39:18 | 0:10:41 | mira | master | ubuntu | 18.04 | rados/singleton/{all/recovery-preemption.yaml msgr-failures/few.yaml msgr/async-v1only.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}} | 1 | |
pass | 4474430 | 2019-11-05 12:11:02 | 2019-11-05 23:43:13 | 2019-11-06 00:45:13 | 1:02:00 | 0:36:31 | 0:25:29 | mira | master | centos | 7.6 | rados/thrash-old-clients/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-install/nautilus.yaml backoff/peering.yaml ceph.yaml clusters/{openstack.yaml three-plus-one.yaml} d-balancer/crush-compat.yaml distro$/{centos_latest.yaml} msgr-failures/fastclose.yaml rados.yaml thrashers/careful.yaml thrashosds-health.yaml workloads/cache-snaps.yaml} | 4 | |
pass | 4474431 | 2019-11-05 12:11:03 | 2019-11-05 23:47:22 | 2019-11-06 00:17:22 | 0:30:00 | 0:19:49 | 0:10:11 | mira | master | ubuntu | 18.04 | rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml 2-recovery-overrides/{default.yaml} backoff/normal.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/off.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/filestore-xfs.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} thrashers/morepggrow.yaml thrashosds-health.yaml workloads/cache-agent-small.yaml} | 2 | |
pass | 4474432 | 2019-11-05 12:11:04 | 2019-11-05 23:49:16 | 2019-11-06 01:05:16 | 1:16:00 | 1:04:29 | 0:11:31 | mira | master | rhel | 7.7 | rados/dashboard/{clusters/{2-node-mgr.yaml} debug/mgr.yaml objectstore/bluestore-comp.yaml supported-random-distro$/{rhel_7.yaml} tasks/dashboard.yaml} | 2 | |
pass | 4474433 | 2019-11-05 12:11:05 | 2019-11-05 23:59:30 | 2019-11-06 02:55:32 | 2:56:02 | 2:38:09 | 0:17:53 | mira | master | rhel | 7.7 | rados/objectstore/{backends/filestore-idempotent-aio-journal.yaml supported-random-distro$/{rhel_7.yaml}} | 1 | |
pass | 4474434 | 2019-11-05 12:11:06 | 2019-11-06 00:01:21 | 2019-11-06 00:27:20 | 0:25:59 | 0:14:03 | 0:11:56 | mira | master | centos | 7.6 | rados/singleton-nomsgr/{all/export-after-evict.yaml rados.yaml supported-random-distro$/{centos_7.yaml}} | 1 | |
pass | 4474435 | 2019-11-05 12:11:06 | 2019-11-06 00:15:38 | 2019-11-06 00:59:37 | 0:43:59 | 0:34:08 | 0:09:51 | mira | master | centos | 7.6 | rados/standalone/{supported-random-distro$/{centos_7.yaml} workloads/misc.yaml} | 1 | |
pass | 4474436 | 2019-11-05 12:11:07 | 2019-11-06 00:17:38 | 2019-11-06 00:47:38 | 0:30:00 | 0:16:05 | 0:13:55 | mira | master | centos | 7.6 | rados/perf/{ceph.yaml objectstore/bluestore-basic-min-osd-mem-target.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{centos_7.yaml} workloads/fio_4M_rand_read.yaml} | 1 | |
pass | 4474437 | 2019-11-05 12:11:08 | 2019-11-06 00:27:20 | 2019-11-06 00:55:19 | 0:27:59 | 0:15:34 | 0:12:25 | mira | master | centos | 7.6 | rados/singleton/{all/resolve_stuck_peering.yaml msgr-failures/many.yaml msgr/async-v2only.yaml objectstore/bluestore-stupid.yaml rados.yaml supported-random-distro$/{centos_7.yaml}} | 2 | |
pass | 4474438 | 2019-11-05 12:11:09 | 2019-11-06 00:27:21 | 2019-11-06 00:55:21 | 0:28:00 | 0:16:42 | 0:11:18 | mira | master | ubuntu | 18.04 | rados/basic/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/few.yaml msgr/async-v2only.yaml objectstore/filestore-xfs.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} tasks/rgw_snaps.yaml} | 2 | |
pass | 4474439 | 2019-11-05 12:11:10 | 2019-11-06 00:45:14 | 2019-11-06 03:31:16 | 2:46:02 | 2:27:50 | 0:18:12 | mira | master | rhel | 7.7 | rados/multimon/{clusters/3.yaml msgr-failures/many.yaml msgr/async.yaml no_pools.yaml objectstore/bluestore-stupid.yaml rados.yaml supported-random-distro$/{rhel_7.yaml} tasks/mon_recovery.yaml} | 2 | |
pass | 4474440 | 2019-11-05 12:11:11 | 2019-11-06 00:47:39 | 2019-11-06 01:27:39 | 0:40:00 | 0:28:59 | 0:11:01 | mira | master | ubuntu | 18.04 | rados/thrash-erasure-code-isa/{arch/x86_64.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/osd-delay.yaml objectstore/filestore-xfs.yaml rados.yaml recovery-overrides/{more-active-recovery.yaml} supported-random-distro$/{ubuntu_latest.yaml} thrashers/default.yaml thrashosds-health.yaml workloads/ec-rados-plugin=isa-k=2-m=1.yaml} | 2 | |
pass | 4474441 | 2019-11-05 12:11:12 | 2019-11-06 00:55:21 | 2019-11-06 01:27:20 | 0:31:59 | 0:24:25 | 0:07:34 | mira | master | rhel | 7.7 | rados/singleton/{all/test-crash.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/filestore-xfs.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}} | 1 | |
pass | 4474442 | 2019-11-05 12:11:13 | 2019-11-06 00:55:22 | 2019-11-06 01:35:22 | 0:40:00 | 0:28:00 | 0:12:00 | mira | master | centos | 7.6 | rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml 2-recovery-overrides/{more-async-partial-recovery.yaml} backoff/peering.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/upmap.yaml msgr-failures/osd-delay.yaml msgr/async-v1only.yaml objectstore/bluestore-avl.yaml rados.yaml supported-random-distro$/{centos_7.yaml} thrashers/none.yaml thrashosds-health.yaml workloads/cache-pool-snaps-readproxy.yaml} | 2 | |
pass | 4474443 | 2019-11-05 12:11:14 | 2019-11-06 00:59:39 | 2019-11-06 01:35:38 | 0:35:59 | 0:20:31 | 0:15:28 | mira | master | centos | 7.6 | rados/thrash-erasure-code-big/{ceph.yaml cluster/{12-osds.yaml openstack.yaml} msgr-failures/few.yaml objectstore/bluestore-avl.yaml rados.yaml recovery-overrides/{default.yaml} supported-random-distro$/{centos_7.yaml} thrashers/mapgap.yaml thrashosds-health.yaml workloads/ec-rados-plugin=lrc-k=4-m=2-l=3.yaml} | 3 | |
pass | 4474444 | 2019-11-05 12:11:15 | 2019-11-06 01:05:21 | 2019-11-06 01:27:21 | 0:22:00 | 0:11:30 | 0:10:30 | mira | master | rados/verify/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-thrash/none.yaml msgr-failures/few.yaml msgr/async-v2only.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml tasks/mon_recovery.yaml validater/lockdep.yaml} | 2 | |||
pass | 4474445 | 2019-11-05 12:11:16 | 2019-11-06 01:27:30 | 2019-11-06 01:45:29 | 0:17:59 | 0:08:03 | 0:09:56 | mira | master | ubuntu | 18.04 | rados/singleton-nomsgr/{all/full-tiering.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}} | 1 | |
pass | 4474446 | 2019-11-05 12:11:17 | 2019-11-06 01:27:32 | 2019-11-06 01:47:31 | 0:19:59 | 0:09:37 | 0:10:22 | mira | master | ubuntu | 18.04 | rados/perf/{ceph.yaml objectstore/bluestore-bitmap.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{ubuntu_latest.yaml} workloads/fio_4M_rand_rw.yaml} | 1 | |
pass | 4474447 | 2019-11-05 12:11:18 | 2019-11-06 01:27:40 | 2019-11-06 02:11:40 | 0:44:00 | 0:35:32 | 0:08:28 | mira | master | rhel | 7.7 | rados/singleton/{all/test_envlibrados_for_rocksdb.yaml msgr-failures/many.yaml msgr/async-v1only.yaml objectstore/bluestore-avl.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}} | 1 | |
pass | 4474448 | 2019-11-05 12:11:18 | 2019-11-06 01:35:39 | 2019-11-06 02:05:38 | 0:29:59 | 0:23:28 | 0:06:31 | mira | master | rhel | 7.7 | rados/mgr/{clusters/{2-node-mgr.yaml} debug/mgr.yaml objectstore/bluestore-stupid.yaml supported-random-distro$/{rhel_7.yaml} tasks/ssh_orchestrator.yaml} | 2 | |
pass | 4474449 | 2019-11-05 12:11:19 | 2019-11-06 01:35:40 | 2019-11-06 02:13:39 | 0:37:59 | 0:30:20 | 0:07:39 | mira | master | rhel | 7.7 | rados/monthrash/{ceph.yaml clusters/3-mons.yaml msgr-failures/mon-delay.yaml msgr/async.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{rhel_7.yaml} thrashers/sync.yaml workloads/pool-create-delete.yaml} | 2 | |
pass | 4474450 | 2019-11-05 12:11:20 | 2019-11-06 01:45:47 | 2019-11-06 02:25:46 | 0:39:59 | 0:22:22 | 0:17:37 | mira | master | centos | 7.6 | rados/thrash-old-clients/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-install/hammer.yaml backoff/peering_and_degraded.yaml ceph.yaml clusters/{openstack.yaml three-plus-one.yaml} d-balancer/off.yaml distro$/{centos_latest.yaml} msgr-failures/few.yaml rados.yaml thrashers/default.yaml thrashosds-health.yaml workloads/radosbench.yaml} | 4 | |
pass | 4474451 | 2019-11-05 12:11:21 | 2019-11-06 01:47:50 | 2019-11-06 04:57:52 | 3:10:02 | 2:48:54 | 0:21:08 | mira | master | rhel | 7.7 | rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml 2-recovery-overrides/{more-active-recovery.yaml} backoff/peering_and_degraded.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/crush-compat.yaml msgr-failures/fastclose.yaml msgr/async-v2only.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{rhel_7.yaml} thrashers/pggrow.yaml thrashosds-health.yaml workloads/cache-pool-snaps.yaml} | 2 | |
pass | 4474452 | 2019-11-05 12:11:22 | 2019-11-06 02:05:40 | 2019-11-06 02:39:40 | 0:34:00 | 0:19:42 | 0:14:18 | mira | master | centos | 7.6 | rados/basic/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/many.yaml msgr/async.yaml objectstore/bluestore-avl.yaml rados.yaml supported-random-distro$/{centos_7.yaml} tasks/scrub_test.yaml} | 2 | |
pass | 4474453 | 2019-11-05 12:11:23 | 2019-11-06 02:08:15 | 2019-11-06 03:18:15 | 1:10:00 | 1:00:30 | 0:09:30 | mira | master | rhel | 7.7 | rados/singleton/{all/thrash-backfill-full.yaml msgr-failures/few.yaml msgr/async-v2only.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}} | 2 | |
pass | 4474454 | 2019-11-05 12:11:24 | 2019-11-06 02:11:42 | 2019-11-06 05:57:44 | 3:46:02 | 3:26:32 | 0:19:30 | mira | master | rhel | 7.7 | rados/thrash-erasure-code/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} fast/normal.yaml msgr-failures/few.yaml objectstore/filestore-xfs.yaml rados.yaml recovery-overrides/{more-partial-recovery.yaml} supported-random-distro$/{rhel_7.yaml} thrashers/careful.yaml thrashosds-health.yaml workloads/ec-radosbench.yaml} | 2 | |
pass | 4474455 | 2019-11-05 12:11:25 | 2019-11-06 02:13:41 | 2019-11-06 03:01:41 | 0:48:00 | 0:38:03 | 0:09:57 | mira | master | ubuntu | 18.04 | rados/objectstore/{backends/filestore-idempotent.yaml supported-random-distro$/{ubuntu_latest.yaml}} | 1 | |
dead | 4474456 | 2019-11-05 12:11:26 | 2019-11-06 02:25:48 | 2019-11-06 14:28:11 | 12:02:23 | mira | master | centos | 7.6 | rados/singleton-nomsgr/{all/health-warnings.yaml rados.yaml supported-random-distro$/{centos_7.yaml}} | 1 | |||
pass | 4474457 | 2019-11-05 12:11:27 | 2019-11-06 02:39:57 | 2019-11-06 03:07:56 | 0:27:59 | 0:17:25 | 0:10:34 | mira | master | centos | 7.6 | rados/perf/{ceph.yaml objectstore/bluestore-comp.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{centos_7.yaml} workloads/fio_4M_rand_write.yaml} | 1 | |
pass | 4474458 | 2019-11-05 12:11:28 | 2019-11-06 02:55:50 | 2019-11-06 05:57:52 | 3:02:02 | 2:41:47 | 0:20:15 | mira | master | rhel | 7.7 | rados/singleton/{all/thrash-eio.yaml msgr-failures/many.yaml msgr/async.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}} | 2 | |
pass | 4474459 | 2019-11-05 12:11:29 | 2019-11-06 03:01:59 | 2019-11-06 03:51:59 | 0:50:00 | 0:36:55 | 0:13:05 | mira | master | centos | 7.6 | rados/thrash-erasure-code-overwrites/{bluestore-bitmap.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} fast/fast.yaml msgr-failures/osd-delay.yaml rados.yaml recovery-overrides/{more-async-partial-recovery.yaml} supported-random-distro$/{centos_7.yaml} thrashers/pggrow.yaml thrashosds-health.yaml workloads/ec-snaps-few-objects-overwrites.yaml} | 2 | |
pass | 4474460 | 2019-11-05 12:11:30 | 2019-11-06 03:07:58 | 2019-11-06 03:45:58 | 0:38:00 | 0:21:49 | 0:16:11 | mira | master | centos | 7.6 | rados/thrash-erasure-code-shec/{ceph.yaml clusters/{fixed-4.yaml openstack.yaml} msgr-failures/few.yaml objectstore/filestore-xfs.yaml rados.yaml recovery-overrides/{more-partial-recovery.yaml} supported-random-distro$/{centos_7.yaml} thrashers/default.yaml thrashosds-health.yaml workloads/ec-rados-plugin=shec-k=4-m=3-c=2.yaml} | 4 | |
pass | 4474461 | 2019-11-05 12:11:31 | 2019-11-06 03:18:33 | 2019-11-06 03:54:32 | 0:35:59 | 0:25:00 | 0:10:59 | mira | master | ubuntu | 18.04 | rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml 2-recovery-overrides/{default.yaml} backoff/normal.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/off.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} thrashers/careful.yaml thrashosds-health.yaml workloads/cache-snaps.yaml} | 2 | |
pass | 4474462 | 2019-11-05 12:11:32 | 2019-11-06 03:31:35 | 2019-11-06 04:29:35 | 0:58:00 | 0:27:32 | 0:30:28 | mira | master | centos | 7.6 | rados/singleton/{all/thrash-rados/{thrash-rados.yaml thrashosds-health.yaml} msgr-failures/few.yaml msgr/async-v1only.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml supported-random-distro$/{centos_7.yaml}} | 2 | |
fail | 4474463 | 2019-11-05 12:11:33 | 2019-11-06 03:46:14 | 2019-11-06 04:42:14 | 0:56:00 | 0:26:00 | 0:30:00 | mira | master | centos | 7.6 | rados/thrash-old-clients/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-install/jewel-v1only.yaml backoff/normal.yaml ceph.yaml clusters/{openstack.yaml three-plus-one.yaml} d-balancer/crush-compat.yaml distro$/{centos_latest.yaml} msgr-failures/osd-delay.yaml rados.yaml thrashers/mapgap.yaml thrashosds-health.yaml workloads/rbd_cls.yaml} | 4 | |
Failure Reason:
Command failed on mira107 with status 1: "sudo TESTDIR=/home/ubuntu/cephtest bash -c 'ceph_test_cls_rbd --gtest_filter=-TestClsRbd.get_features:TestClsRbd.parents'" |
||||||||||||||
pass | 4474464 | 2019-11-05 12:11:34 | 2019-11-06 03:52:17 | 2019-11-06 04:26:16 | 0:33:59 | 0:13:30 | 0:20:29 | mira | master | centos | 7.6 | rados/multimon/{clusters/6.yaml msgr-failures/few.yaml msgr/async-v1only.yaml no_pools.yaml objectstore/filestore-xfs.yaml rados.yaml supported-random-distro$/{centos_7.yaml} tasks/mon_clock_no_skews.yaml} | 2 | |
pass | 4474465 | 2019-11-05 12:11:35 | 2019-11-06 03:54:34 | 2019-11-06 04:30:33 | 0:35:59 | 0:24:15 | 0:11:44 | mira | master | ubuntu | 18.04 | rados/basic/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/few.yaml msgr/async.yaml objectstore/bluestore-avl.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} tasks/rados_api_tests.yaml} | 2 | |
pass | 4474466 | 2019-11-05 12:11:36 | 2019-11-06 04:26:18 | 2019-11-06 05:46:18 | 1:20:00 | 1:07:00 | 0:13:00 | mira | master | centos | 7.6 | rados/dashboard/{clusters/{2-node-mgr.yaml} debug/mgr.yaml objectstore/bluestore-low-osd-mem-target.yaml supported-random-distro$/{centos_7.yaml} tasks/dashboard.yaml} | 2 | |
pass | 4474467 | 2019-11-05 12:11:37 | 2019-11-06 04:29:52 | 2019-11-06 05:09:51 | 0:39:59 | 0:29:40 | 0:10:19 | mira | master | ubuntu | 18.04 | rados/singleton-bluestore/{all/cephtool.yaml msgr-failures/many.yaml msgr/async-v2only.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}} | 1 | |
pass | 4474468 | 2019-11-05 12:11:38 | 2019-11-06 04:30:35 | 2019-11-06 04:56:34 | 0:25:59 | 0:14:56 | 0:11:03 | mira | master | centos | 7.6 | rados/singleton-nomsgr/{all/large-omap-object-warnings.yaml rados.yaml supported-random-distro$/{centos_7.yaml}} | 1 | |
pass | 4474469 | 2019-11-05 12:11:38 | 2019-11-06 04:42:16 | 2019-11-06 05:58:16 | 1:16:00 | 1:06:04 | 0:09:56 | mira | master | centos | 7.6 | rados/standalone/{supported-random-distro$/{centos_7.yaml} workloads/mon.yaml} | 1 | |
fail | 4474470 | 2019-11-05 12:11:39 | 2019-11-06 04:56:36 | 2019-11-06 05:38:35 | 0:41:59 | 0:22:35 | 0:19:24 | mira | master | centos | 7.6 | rados/upgrade/nautilus-x-singleton/{0-cluster/{openstack.yaml start.yaml} 1-install/nautilus.yaml 2-partial-upgrade/firsthalf.yaml 3-thrash/default.yaml 4-workload/{rbd-cls.yaml rbd-import-export.yaml readwrite.yaml snaps-few-objects.yaml} 5-workload/{radosbench.yaml rbd_api.yaml} 6-finish-upgrade.yaml 7-octopus.yaml 8-workload/{rbd-python.yaml rgw-swift.yaml snaps-many-objects.yaml} bluestore-bitmap.yaml supported-random-distro$/{centos_7.yaml} thrashosds-health.yaml} | 4 | |
Failure Reason:
Command failed (workunit test cls/test_cls_rbd.sh) on mira100 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=nautilus TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/cls/test_cls_rbd.sh' |
||||||||||||||
pass | 4474471 | 2019-11-05 12:11:40 | 2019-11-06 04:58:10 | 2019-11-06 05:20:09 | 0:21:59 | 0:12:41 | 0:09:18 | mira | master | ubuntu | 18.04 | rados/perf/{ceph.yaml objectstore/bluestore-low-osd-mem-target.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{ubuntu_latest.yaml} workloads/radosbench_4K_rand_read.yaml} | 1 | |
pass | 4474472 | 2019-11-05 12:11:41 | 2019-11-06 05:09:53 | 2019-11-06 05:43:52 | 0:33:59 | 0:24:26 | 0:09:33 | mira | master | rhel | 7.7 | rados/mgr/{clusters/{2-node-mgr.yaml} debug/mgr.yaml objectstore/filestore-xfs.yaml supported-random-distro$/{rhel_7.yaml} tasks/workunits.yaml} | 2 | |
pass | 4474473 | 2019-11-05 12:11:42 | 2019-11-06 05:20:10 | 2019-11-06 06:26:10 | 1:06:00 | 0:38:42 | 0:27:18 | mira | master | centos | 7.6 | rados/thrash-erasure-code-isa/{arch/x86_64.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/fastclose.yaml objectstore/bluestore-avl.yaml rados.yaml recovery-overrides/{default.yaml} supported-random-distro$/{centos_7.yaml} thrashers/mapgap.yaml thrashosds-health.yaml workloads/ec-rados-plugin=isa-k=2-m=1.yaml} | 2 | |
pass | 4474474 | 2019-11-05 12:11:43 | 2019-11-06 05:38:52 | 2019-11-06 06:16:52 | 0:38:00 | 0:23:55 | 0:14:05 | mira | master | centos | 7.6 | rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml 2-recovery-overrides/{default.yaml} backoff/peering.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/upmap.yaml msgr-failures/osd-delay.yaml msgr/async-v1only.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml supported-random-distro$/{centos_7.yaml} thrashers/default.yaml thrashosds-health.yaml workloads/cache.yaml} | 2 | |
pass | 4474475 | 2019-11-05 12:11:44 | 2019-11-06 05:44:10 | 2019-11-06 06:32:10 | 0:48:00 | 0:40:46 | 0:07:14 | mira | master | rhel | 7.7 | rados/singleton/{all/thrash_cache_writeback_proxy_none.yaml msgr-failures/many.yaml msgr/async-v2only.yaml objectstore/bluestore-stupid.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}} | 2 | |
pass | 4474476 | 2019-11-05 12:11:45 | 2019-11-06 05:46:31 | 2019-11-06 06:32:31 | 0:46:00 | 0:34:58 | 0:11:02 | mira | master | ubuntu | 18.04 | rados/thrash-erasure-code-big/{ceph.yaml cluster/{12-osds.yaml openstack.yaml} msgr-failures/osd-delay.yaml objectstore/bluestore-bitmap.yaml rados.yaml recovery-overrides/{more-async-partial-recovery.yaml} supported-random-distro$/{ubuntu_latest.yaml} thrashers/morepggrow.yaml thrashosds-health.yaml workloads/ec-rados-plugin=jerasure-k=4-m=2.yaml} | 3 | |
fail | 4474477 | 2019-11-05 12:11:46 | 2019-11-06 05:58:01 | 2019-11-06 08:00:04 | 2:02:03 | 1:39:44 | 0:22:19 | mira | master | centos | rados/verify/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-thrash/default/{default.yaml thrashosds-health.yaml} msgr-failures/few.yaml msgr/async.yaml objectstore/bluestore-stupid.yaml rados.yaml tasks/rados_api_tests.yaml validater/valgrind.yaml} | 2 | ||
Failure Reason:
saw valgrind issues |
||||||||||||||
pass | 4474478 | 2019-11-05 12:11:47 | 2019-11-06 05:58:01 | 2019-11-06 06:22:01 | 0:24:00 | 0:14:18 | 0:09:42 | mira | master | centos | 7.6 | rados/singleton/{all/watch-notify-same-primary.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/filestore-xfs.yaml rados.yaml supported-random-distro$/{centos_7.yaml}} | 1 | |
pass | 4474479 | 2019-11-05 12:11:48 | 2019-11-06 05:58:17 | 2019-11-06 06:32:17 | 0:34:00 | 0:25:57 | 0:08:03 | mira | master | rhel | 7.7 | rados/monthrash/{ceph.yaml clusters/9-mons.yaml msgr-failures/few.yaml msgr/async-v1only.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{rhel_7.yaml} thrashers/force-sync-many.yaml workloads/rados_5925.yaml} | 2 | |
pass | 4474480 | 2019-11-05 12:11:49 | 2019-11-06 06:04:53 | 2019-11-06 06:48:53 | 0:44:00 | 0:21:41 | 0:22:19 | mira | master | centos | 7.6 | rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml 2-recovery-overrides/{default.yaml} backoff/peering_and_degraded.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/crush-compat.yaml msgr-failures/fastclose.yaml msgr/async-v2only.yaml objectstore/bluestore-stupid.yaml rados.yaml supported-random-distro$/{centos_7.yaml} thrashers/mapgap.yaml thrashosds-health.yaml workloads/dedup_tier.yaml} | 2 | |
pass | 4474481 | 2019-11-05 12:11:50 | 2019-11-06 06:17:09 | 2019-11-06 06:51:08 | 0:33:59 | 0:26:07 | 0:07:52 | mira | master | rhel | 7.7 | rados/perf/{ceph.yaml objectstore/bluestore-stupid.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{rhel_7.yaml} workloads/radosbench_4K_seq_read.yaml} | 1 | |
pass | 4474482 | 2019-11-05 12:11:51 | 2019-11-06 06:22:02 | 2019-11-06 06:50:02 | 0:28:00 | 0:21:51 | 0:06:09 | mira | master | rhel | 7.7 | rados/objectstore/{backends/fusestore.yaml supported-random-distro$/{rhel_7.yaml}} | 1 | |
pass | 4474483 | 2019-11-05 12:11:52 | 2019-11-06 06:26:12 | 2019-11-06 06:56:12 | 0:30:00 | 0:23:48 | 0:06:12 | mira | master | rhel | 7.7 | rados/singleton-nomsgr/{all/lazy_omap_stats_output.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}} | 1 | |
pass | 4474484 | 2019-11-05 12:11:53 | 2019-11-06 06:32:28 | 2019-11-06 07:22:28 | 0:50:00 | 0:26:18 | 0:23:42 | mira | master | ubuntu | 18.04 | rados/thrash-erasure-code/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} fast/fast.yaml msgr-failures/osd-delay.yaml objectstore/bluestore-avl.yaml rados.yaml recovery-overrides/{more-async-partial-recovery.yaml} supported-random-distro$/{ubuntu_latest.yaml} thrashers/default.yaml thrashosds-health.yaml workloads/ec-small-objects-fast-read.yaml} | 2 | |
pass | 4474485 | 2019-11-05 12:11:54 | 2019-11-06 06:32:28 | 2019-11-06 07:18:28 | 0:46:00 | 0:26:47 | 0:19:13 | mira | master | ubuntu | 18.04 | rados/singleton/{all/admin-socket.yaml msgr-failures/many.yaml msgr/async-v1only.yaml objectstore/filestore-xfs.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}} | 1 | |
pass | 4474486 | 2019-11-05 12:11:55 | 2019-11-06 06:32:32 | 2019-11-06 07:08:32 | 0:36:00 | 0:27:37 | 0:08:23 | mira | master | rhel | 7.7 | rados/basic/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/many.yaml msgr/async-v1only.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{rhel_7.yaml} tasks/rados_cls_all.yaml} | 2 | |
pass | 4474487 | 2019-11-05 12:11:56 | 2019-11-06 06:49:12 | 2019-11-06 07:51:11 | 1:01:59 | 0:45:01 | 0:16:58 | mira | master | centos | 7.6 | rados/thrash-old-clients/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-install/jewel.yaml backoff/peering.yaml ceph.yaml clusters/{openstack.yaml three-plus-one.yaml} d-balancer/off.yaml distro$/{centos_latest.yaml} msgr-failures/fastclose.yaml rados.yaml thrashers/morepggrow.yaml thrashosds-health.yaml workloads/snaps-few-objects.yaml} | 4 | |
pass | 4474488 | 2019-11-05 12:11:57 | 2019-11-06 06:50:04 | 2019-11-06 07:18:03 | 0:27:59 | 0:15:38 | 0:12:21 | mira | master | centos | 7.6 | rados/singleton/{all/deduptool.yaml msgr-failures/few.yaml msgr/async-v2only.yaml objectstore/bluestore-avl.yaml rados.yaml supported-random-distro$/{centos_7.yaml}} | 1 | |
pass | 4474489 | 2019-11-05 12:11:58 | 2019-11-06 06:51:27 | 2019-11-06 07:23:26 | 0:31:59 | 0:18:22 | 0:13:37 | mira | master | centos | 7.6 | rados/mgr/{clusters/{2-node-mgr.yaml} debug/mgr.yaml objectstore/filestore-xfs.yaml supported-random-distro$/{centos_7.yaml} tasks/crash.yaml} | 2 | |
pass | 4474490 | 2019-11-05 12:11:59 | 2019-11-06 06:56:30 | 2019-11-06 07:54:29 | 0:57:59 | 0:46:33 | 0:11:26 | mira | master | rhel | 7.7 | rados/thrash-erasure-code-overwrites/{bluestore-bitmap.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} fast/normal.yaml msgr-failures/fastclose.yaml rados.yaml recovery-overrides/{more-async-recovery.yaml} supported-random-distro$/{rhel_7.yaml} thrashers/careful.yaml thrashosds-health.yaml workloads/ec-snaps-few-objects-overwrites.yaml} | 2 | |
pass | 4474491 | 2019-11-05 12:12:00 | 2019-11-06 07:08:49 | 2019-11-06 07:54:49 | 0:46:00 | 0:27:58 | 0:18:02 | mira | master | ubuntu | 18.04 | rados/thrash-erasure-code-shec/{ceph.yaml clusters/{fixed-4.yaml openstack.yaml} msgr-failures/few.yaml objectstore/bluestore-avl.yaml rados.yaml recovery-overrides/{more-active-recovery.yaml} supported-random-distro$/{ubuntu_latest.yaml} thrashers/careful.yaml thrashosds-health.yaml workloads/ec-rados-plugin=shec-k=4-m=3-c=2.yaml} | 4 | |
pass | 4474492 | 2019-11-05 12:12:01 | 2019-11-06 07:18:05 | 2019-11-06 08:12:05 | 0:54:00 | 0:40:03 | 0:13:57 | mira | master | centos | 7.6 | rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml 2-recovery-overrides/{more-async-partial-recovery.yaml} backoff/normal.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/off.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/filestore-xfs.yaml rados.yaml supported-random-distro$/{centos_7.yaml} thrashers/morepggrow.yaml thrashosds-health.yaml workloads/pool-snaps-few-objects.yaml} | 2 | |
pass | 4474493 | 2019-11-05 12:12:02 | 2019-11-06 07:18:29 | 2019-11-06 07:46:28 | 0:27:59 | 0:17:27 | 0:10:32 | mira | master | centos | 7.6 | rados/perf/{ceph.yaml objectstore/bluestore-basic-min-osd-mem-target.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{centos_7.yaml} workloads/radosbench_4M_rand_read.yaml} | 1 | |
pass | 4474494 | 2019-11-05 12:12:03 | 2019-11-06 07:22:45 | 2019-11-06 08:10:45 | 0:48:00 | 0:21:18 | 0:26:42 | mira | master | ubuntu | 18.04 | rados/singleton-nomsgr/{all/librados_hello_world.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}} | 1 | |
pass | 4474495 | 2019-11-05 12:12:04 | 2019-11-06 07:24:25 | 2019-11-06 08:00:25 | 0:36:00 | 0:15:55 | 0:20:05 | mira | master | centos | 7.6 | rados/singleton/{all/divergent_priors.yaml msgr-failures/many.yaml msgr/async.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{centos_7.yaml}} | 1 | |
pass | 4474496 | 2019-11-05 12:12:05 | 2019-11-06 07:46:30 | 2019-11-06 08:16:29 | 0:29:59 | 0:14:26 | 0:15:33 | mira | master | centos | 7.6 | rados/multimon/{clusters/21.yaml msgr-failures/few.yaml msgr/async-v1only.yaml no_pools.yaml objectstore/bluestore-avl.yaml rados.yaml supported-random-distro$/{centos_7.yaml} tasks/mon_clock_no_skews.yaml} | 3 | |
pass | 4474497 | 2019-11-05 12:12:06 | 2019-11-06 07:51:28 | 2019-11-06 08:33:28 | 0:42:00 | 0:28:23 | 0:13:37 | mira | master | centos | 7.6 | rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml 2-recovery-overrides/{more-async-recovery.yaml} backoff/peering.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/upmap.yaml msgr-failures/osd-delay.yaml msgr/async-v1only.yaml objectstore/bluestore-avl.yaml rados.yaml supported-random-distro$/{centos_7.yaml} thrashers/none.yaml thrashosds-health.yaml workloads/rados_api_tests.yaml} | 2 | |
pass | 4474498 | 2019-11-05 12:12:06 | 2019-11-06 07:54:46 | 2019-11-06 08:56:46 | 1:02:00 | 0:44:42 | 0:17:18 | mira | master | ubuntu | 18.04 | rados/thrash-erasure-code-isa/{arch/x86_64.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/few.yaml objectstore/bluestore-bitmap.yaml rados.yaml recovery-overrides/{default.yaml} supported-random-distro$/{ubuntu_latest.yaml} thrashers/morepggrow.yaml thrashosds-health.yaml workloads/ec-rados-plugin=isa-k=2-m=1.yaml} | 2 | |
pass | 4474499 | 2019-11-05 12:12:07 | 2019-11-06 07:54:53 | 2019-11-06 08:30:53 | 0:36:00 | 0:14:10 | 0:21:50 | mira | master | ubuntu | 18.04 | rados/singleton/{all/divergent_priors2.yaml msgr-failures/few.yaml msgr/async-v1only.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}} | 1 | |
pass | 4474500 | 2019-11-05 12:12:08 | 2019-11-06 08:00:21 | 2019-11-06 08:36:20 | 0:35:59 | 0:16:58 | 0:19:01 | mira | master | ubuntu | 18.04 | rados/basic/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/few.yaml msgr/async-v2only.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} tasks/rados_python.yaml} | 2 | |
pass | 4474501 | 2019-11-05 12:12:09 | 2019-11-06 08:00:26 | 2019-11-06 08:48:26 | 0:48:00 | 0:27:49 | 0:20:11 | mira | master | ubuntu | 18.04 | rados/thrash-erasure-code-big/{ceph.yaml cluster/{12-osds.yaml openstack.yaml} msgr-failures/fastclose.yaml objectstore/bluestore-comp.yaml rados.yaml recovery-overrides/{more-active-recovery.yaml} supported-random-distro$/{ubuntu_latest.yaml} thrashers/pggrow.yaml thrashosds-health.yaml workloads/ec-rados-plugin=lrc-k=4-m=2-l=3.yaml} | 3 | |
pass | 4474502 | 2019-11-05 12:12:10 | 2019-11-06 08:11:02 | 2019-11-06 09:09:02 | 0:58:00 | 0:21:06 | 0:36:54 | mira | master | rados/verify/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-thrash/none.yaml msgr-failures/few.yaml msgr/async-v1only.yaml objectstore/filestore-xfs.yaml rados.yaml tasks/rados_cls_all.yaml validater/lockdep.yaml} | 2 | |||
pass | 4474503 | 2019-11-05 12:12:11 | 2019-11-06 08:12:37 | 2019-11-06 08:44:36 | 0:31:59 | 0:25:25 | 0:06:34 | mira | master | rhel | 7.7 | rados/perf/{ceph.yaml objectstore/bluestore-bitmap.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{rhel_7.yaml} workloads/radosbench_4M_seq_read.yaml} | 1 | |
pass | 4474504 | 2019-11-05 12:12:12 | 2019-11-06 08:16:49 | 2019-11-06 08:58:49 | 0:42:00 | 0:20:00 | 0:22:00 | mira | master | centos | 7.6 | rados/thrash-old-clients/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-install/luminous-v1only.yaml backoff/peering_and_degraded.yaml ceph.yaml clusters/{openstack.yaml three-plus-one.yaml} d-balancer/crush-compat.yaml distro$/{centos_latest.yaml} msgr-failures/few.yaml rados.yaml thrashers/none.yaml thrashosds-health.yaml workloads/test_rbd_api.yaml} | 4 | |
pass | 4474505 | 2019-11-05 12:12:13 | 2019-11-06 08:30:55 | 2019-11-06 09:34:55 | 1:04:00 | 0:46:24 | 0:17:36 | mira | master | ubuntu | 18.04 | rados/dashboard/{clusters/{2-node-mgr.yaml} debug/mgr.yaml objectstore/bluestore-stupid.yaml supported-random-distro$/{ubuntu_latest.yaml} tasks/dashboard.yaml} | 2 | |
pass | 4474506 | 2019-11-05 12:12:13 | 2019-11-06 08:33:46 | 2019-11-06 09:01:45 | 0:27:59 | 0:20:33 | 0:07:26 | mira | master | rhel | 7.7 | rados/objectstore/{backends/keyvaluedb.yaml supported-random-distro$/{rhel_7.yaml}} | 1 | |
pass | 4474507 | 2019-11-05 12:12:14 | 2019-11-06 08:36:37 | 2019-11-06 09:14:37 | 0:38:00 | 0:23:26 | 0:14:34 | mira | master | centos | 7.6 | rados/singleton-nomsgr/{all/msgr.yaml rados.yaml supported-random-distro$/{centos_7.yaml}} | 1 | |
pass | 4474508 | 2019-11-05 12:12:15 | 2019-11-06 08:44:53 | 2019-11-06 11:38:55 | 2:54:02 | 2:35:46 | 0:18:16 | mira | master | ubuntu | 18.04 | rados/standalone/{supported-random-distro$/{ubuntu_latest.yaml} workloads/osd.yaml} | 1 | |
pass | 4474509 | 2019-11-05 12:12:16 | 2019-11-06 08:48:44 | 2019-11-06 11:34:46 | 2:46:02 | 2:26:48 | 0:19:14 | mira | master | rhel | 7.7 | rados/singleton/{all/dump-stuck.yaml msgr-failures/many.yaml msgr/async-v2only.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}} | 1 | |
pass | 4474510 | 2019-11-05 12:12:17 | 2019-11-06 08:56:54 | 2019-11-06 09:42:54 | 0:46:00 | 0:37:15 | 0:08:45 | mira | master | rhel | 7.7 | rados/thrash-erasure-code/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} fast/normal.yaml msgr-failures/fastclose.yaml objectstore/bluestore-bitmap.yaml rados.yaml recovery-overrides/{default.yaml} supported-random-distro$/{rhel_7.yaml} thrashers/fastread.yaml thrashosds-health.yaml workloads/ec-small-objects-many-deletes.yaml} | 2 | |
pass | 4474511 | 2019-11-05 12:12:18 | 2019-11-06 08:59:06 | 2019-11-06 09:35:06 | 0:36:00 | 0:14:03 | 0:21:57 | mira | master | ubuntu | 18.04 | rados/mgr/{clusters/{2-node-mgr.yaml} debug/mgr.yaml objectstore/bluestore-avl.yaml supported-random-distro$/{ubuntu_latest.yaml} tasks/failover.yaml} | 2 | |
pass | 4474512 | 2019-11-05 12:12:19 | 2019-11-06 09:01:47 | 2019-11-06 10:07:47 | 1:06:00 | 0:57:47 | 0:08:13 | mira | master | rhel | 7.7 | rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml 2-recovery-overrides/{more-async-partial-recovery.yaml} backoff/peering_and_degraded.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/crush-compat.yaml msgr-failures/fastclose.yaml msgr/async-v2only.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{rhel_7.yaml} thrashers/pggrow.yaml thrashosds-health.yaml workloads/radosbench.yaml} | 2 | |
pass | 4474513 | 2019-11-05 12:12:20 | 2019-11-06 09:09:19 | 2019-11-06 09:57:19 | 0:48:00 | 0:34:32 | 0:13:28 | mira | master | centos | 7.6 | rados/monthrash/{ceph.yaml clusters/3-mons.yaml msgr-failures/mon-delay.yaml msgr/async-v2only.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml supported-random-distro$/{centos_7.yaml} thrashers/many.yaml workloads/rados_api_tests.yaml} | 2 | |
pass | 4474514 | 2019-11-05 12:12:20 | 2019-11-06 09:14:39 | 2019-11-06 10:52:39 | 1:38:00 | 1:27:54 | 0:10:06 | mira | master | centos | 7.6 | rados/singleton/{all/ec-lost-unfound.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/bluestore-stupid.yaml rados.yaml supported-random-distro$/{centos_7.yaml}} | 1 | |
pass | 4474515 | 2019-11-05 12:12:21 | 2019-11-06 09:34:57 | 2019-11-06 09:58:57 | 0:24:00 | 0:12:44 | 0:11:16 | mira | master | ubuntu | 18.04 | rados/perf/{ceph.yaml objectstore/bluestore-comp.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{ubuntu_latest.yaml} workloads/radosbench_4M_write.yaml} | 1 | |
pass | 4474516 | 2019-11-05 12:12:22 | 2019-11-06 09:35:07 | 2019-11-06 10:27:08 | 0:52:01 | 0:43:59 | 0:08:02 | mira | master | rhel | 7.7 | rados/singleton-nomsgr/{all/multi-backfill-reject.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}} | 2 | |
pass | 4474517 | 2019-11-05 12:12:23 | 2019-11-06 09:43:11 | 2019-11-06 10:23:11 | 0:40:00 | 0:26:31 | 0:13:29 | mira | master | centos | 7.6 | rados/basic/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/many.yaml msgr/async.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml supported-random-distro$/{centos_7.yaml} tasks/rados_stress_watch.yaml} | 2 | |
pass | 4474518 | 2019-11-05 12:12:24 | 2019-11-06 09:53:46 | 2019-11-06 10:21:45 | 0:27:59 | 0:10:08 | 0:17:51 | mira | master | ubuntu | 18.04 | rados/singleton/{all/erasure-code-nonregression.yaml msgr-failures/many.yaml msgr/async-v1only.yaml objectstore/filestore-xfs.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}} | 1 | |
pass | 4474519 | 2019-11-05 12:12:25 | 2019-11-06 09:57:21 | 2019-11-06 10:55:21 | 0:58:00 | 0:50:03 | 0:07:57 | mira | master | rhel | 7.7 | rados/thrash-erasure-code-overwrites/{bluestore-bitmap.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} fast/fast.yaml msgr-failures/few.yaml rados.yaml recovery-overrides/{more-async-recovery.yaml} supported-random-distro$/{rhel_7.yaml} thrashers/default.yaml thrashosds-health.yaml workloads/ec-pool-snaps-few-objects-overwrites.yaml} | 2 |