ID
Status
Ceph Branch
Suite Branch
Teuthology Branch
Machine
OS
Nodes
Description
Failure Reason
wip-sage-testing-2019-02-17-0921
wip-sage-testing-2019-02-17-0921
master
smithi
rhel 7.5
rados/perf/{ceph.yaml objectstore/bluestore-stupid.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{rhel_latest.yaml} workloads/radosbench_4K_rand_read.yaml}
wip-sage-testing-2019-02-17-0921
wip-sage-testing-2019-02-17-0921
master
smithi
centos 7.5
rados/basic/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} mon_kv_backend/leveldb.yaml msgr-failures/few.yaml msgr/async-v2only.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{centos_latest.yaml} tasks/rados_python.yaml}
"2019-02-17 17:52:01.028094 mon.b (mon.0) 37 : cluster [WRN] Health check failed: 1/3 mons down, quorum b,a (MON_DOWN)" in cluster log
wip-sage-testing-2019-02-17-0921
wip-sage-testing-2019-02-17-0921
master
smithi
ubuntu 16.04
rados/objectstore/{backends/objectstore.yaml supported-random-distro$/{ubuntu_16.04.yaml}}
wip-sage-testing-2019-02-17-0921
wip-sage-testing-2019-02-17-0921
master
smithi
ubuntu 16.04
rados/thrash-erasure-code-overwrites/{bluestore-bitmap.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} fast/fast.yaml leveldb.yaml msgr-failures/fastclose.yaml rados.yaml recovery-overrides/{default.yaml} supported-random-distro$/{ubuntu_16.04.yaml} thrashers/default.yaml thrashosds-health.yaml workloads/ec-pool-snaps-few-objects-overwrites.yaml}
Command crashed: 'sudo ceph --cluster ceph osd crush tunables default'
wip-sage-testing-2019-02-17-0921
wip-sage-testing-2019-02-17-0921
master
smithi
centos 7.5
rados/thrash-erasure-code-shec/{ceph.yaml clusters/{fixed-4.yaml openstack.yaml} leveldb.yaml msgr-failures/fastclose.yaml objectstore/bluestore-bitmap.yaml rados.yaml recovery-overrides/{more-active-recovery.yaml} supported-random-distro$/{centos_latest.yaml} thrashers/default.yaml thrashosds-health.yaml workloads/ec-rados-plugin=shec-k=4-m=3-c=2.yaml}
"2019-02-17 17:57:02.199578 mon.c (mon.0) 19 : cluster [WRN] Health check failed: 1/3 mons down, quorum c,a (MON_DOWN)" in cluster log
wip-sage-testing-2019-02-17-0921
wip-sage-testing-2019-02-17-0921
master
smithi
 
rados/verify/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-thrash/none.yaml mon_kv_backend/rocksdb.yaml msgr-failures/few.yaml msgr/async-v2only.yaml objectstore/bluestore-bitmap.yaml rados.yaml tasks/mon_recovery.yaml validater/lockdep.yaml}
"2019-02-17 17:47:12.804473 mon.a (mon.1) 199 : cluster [WRN] Health check failed: 4 osds down (OSD_DOWN)" in cluster log
wip-sage-testing-2019-02-17-0921
wip-sage-testing-2019-02-17-0921
master
smithi
ubuntu 18.04
rados/singleton/{all/thrash_cache_writeback_proxy_none.yaml msgr-failures/few.yaml msgr/async-v1only.yaml objectstore/bluestore-stupid.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}}
wip-sage-testing-2019-02-17-0921
wip-sage-testing-2019-02-17-0921
master
smithi
ubuntu 18.04
rados/thrash-erasure-code/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} fast/fast.yaml leveldb.yaml msgr-failures/fastclose.yaml objectstore/filestore-xfs.yaml rados.yaml recovery-overrides/{default.yaml} supported-random-distro$/{ubuntu_latest.yaml} thrashers/pggrow.yaml thrashosds-health.yaml workloads/ec-rados-plugin=jerasure-k=2-m=1.yaml}
"2019-02-17 17:49:45.047399 mon.b (mon.0) 18 : cluster [WRN] Health check failed: 1/3 mons down, quorum b,c (MON_DOWN)" in cluster log
wip-sage-testing-2019-02-17-0921
wip-sage-testing-2019-02-17-0921
master
smithi
rhel 7.5
rados/mgr/{clusters/{2-node-mgr.yaml openstack.yaml} debug/mgr.yaml objectstore/filestore-xfs.yaml supported-random-distro$/{rhel_latest.yaml} tasks/orchestrator_cli.yaml}
"2019-02-17 17:52:16.339665 mon.a (mon.0) 66 : cluster [WRN] Health check failed: 1/3 mons down, quorum a,c (MON_DOWN)" in cluster log
wip-sage-testing-2019-02-17-0921
wip-sage-testing-2019-02-17-0921
master
smithi
centos 7.5
rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml 2-recovery-overrides/{more-active-recovery.yaml} backoff/peering.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/upmap.yaml msgr-failures/fastclose.yaml msgr/random.yaml objectstore/filestore-xfs.yaml rados.yaml rocksdb.yaml supported-random-distro$/{centos_latest.yaml} thrashers/mapgap.yaml thrashosds-health.yaml workloads/pool-snaps-few-objects.yaml}
wip-sage-testing-2019-02-17-0921
wip-sage-testing-2019-02-17-0921
master
smithi
centos 7.5
rados/singleton/{all/watch-notify-same-primary.yaml msgr-failures/many.yaml msgr/async-v2only.yaml objectstore/filestore-xfs.yaml rados.yaml supported-random-distro$/{centos_latest.yaml}}
"2019-02-17 17:53:47.495780 mon.b (mon.1) 10 : cluster [WRN] Health check failed: 1/3 mons down, quorum b,c (MON_DOWN)" in cluster log
wip-sage-testing-2019-02-17-0921
wip-sage-testing-2019-02-17-0921
master
smithi
ubuntu 18.04
rados/singleton-nomsgr/{all/recovery-unfound-found.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}}
Command failed on smithi178 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
wip-sage-testing-2019-02-17-0921
wip-sage-testing-2019-02-17-0921
master
smithi
ubuntu 18.04
rados/perf/{ceph.yaml objectstore/filestore-xfs.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{ubuntu_latest.yaml} workloads/radosbench_4K_seq_read.yaml}
wip-sage-testing-2019-02-17-0921
wip-sage-testing-2019-02-17-0921
master
smithi
rhel 7.5
rados/singleton/{all/admin-socket.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{rhel_latest.yaml}}
Command failed on smithi058 with status 1: 'adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph --cluster ceph osd dump --format=json'
wip-sage-testing-2019-02-17-0921
wip-sage-testing-2019-02-17-0921
master
smithi
ubuntu 16.04
rados/singleton-bluestore/{all/cephtool.yaml msgr-failures/few.yaml msgr/random.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{ubuntu_16.04.yaml}}
wip-sage-testing-2019-02-17-0921
wip-sage-testing-2019-02-17-0921
master
smithi
centos 7.5
rados/basic/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} mon_kv_backend/rocksdb.yaml msgr-failures/many.yaml msgr/async.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{centos_latest.yaml} tasks/rados_stress_watch.yaml}
wip-sage-testing-2019-02-17-0921
wip-sage-testing-2019-02-17-0921
master
smithi
rhel 7.5
rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml 2-recovery-overrides/{more-active-recovery.yaml} backoff/peering_and_degraded.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/crush-compat.yaml msgr-failures/few.yaml msgr/simple.yaml objectstore/bluestore-bitmap.yaml rados.yaml rocksdb.yaml supported-random-distro$/{rhel_latest.yaml} thrashers/morepggrow.yaml thrashosds-health.yaml workloads/rados_api_tests.yaml}
wip-sage-testing-2019-02-17-0921
wip-sage-testing-2019-02-17-0921
master
smithi
ubuntu 18.04
rados/singleton/{all/deduptool.yaml msgr-failures/many.yaml msgr/random.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}}
wip-sage-testing-2019-02-17-0921
wip-sage-testing-2019-02-17-0921
master
smithi
rhel 7.5
rados/perf/{ceph.yaml objectstore/bluestore-bitmap.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{rhel_latest.yaml} workloads/radosbench_4M_rand_read.yaml}
wip-sage-testing-2019-02-17-0921
wip-sage-testing-2019-02-17-0921
master
smithi
ubuntu 16.04
rados/thrash-old-clients/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-install/luminous.yaml backoff/peering.yaml ceph.yaml clusters/{openstack.yaml three-plus-one.yaml} d-balancer/off.yaml distro$/{ubuntu_16.04.yaml} msgr-failures/few.yaml msgr/async-v1only.yaml rados.yaml rocksdb.yaml thrashers/morepggrow.yaml thrashosds-health.yaml workloads/radosbench.yaml}
wip-sage-testing-2019-02-17-0921
wip-sage-testing-2019-02-17-0921
master
smithi
ubuntu 16.04
rados/mgr/{clusters/{2-node-mgr.yaml openstack.yaml} debug/mgr.yaml objectstore/bluestore-bitmap.yaml supported-random-distro$/{ubuntu_16.04.yaml} tasks/progress.yaml}
"2019-02-17 17:56:05.202021 mon.b (mon.0) 89 : cluster [WRN] Health check failed: 1/3 mons down, quorum b,a (MON_DOWN)" in cluster log
wip-sage-testing-2019-02-17-0921
wip-sage-testing-2019-02-17-0921
master
smithi
ubuntu 18.04
rados/objectstore/{backends/alloc-hint.yaml supported-random-distro$/{ubuntu_latest.yaml}}
Command failed on smithi049 with status 1: 'adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph --cluster ceph osd dump --format=json'
wip-sage-testing-2019-02-17-0921
wip-sage-testing-2019-02-17-0921
master
smithi
rhel 7.5
rados/rest/{mgr-restful.yaml supported-random-distro$/{rhel_latest.yaml}}
Command failed on smithi187 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph osd dump --format=json'