ID
Status
Ceph Branch
Suite Branch
Teuthology Branch
Machine
OS
Nodes
Description
Failure Reason
mimic-dev1
master
master
ovh
 
rados/thrash/{0-size-min-size-overrides/2-size-1-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml backoff/peering_and_degraded.yaml clusters/{fixed-2.yaml openstack.yaml} fs/xfs.yaml msgr-failures/osd-delay.yaml msgr/simple.yaml objectstore/bluestore.yaml rados.yaml rocksdb.yaml thrashers/morepggrow.yaml workloads/pool-snaps-few-objects.yaml}
Command failed on ovh057 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
 
rados/basic/{clusters/{fixed-2.yaml openstack.yaml} fs/btrfs.yaml mon_kv_backend/rocksdb.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/bluestore.yaml rados.yaml tasks/rados_stress_watch.yaml}
"2017-12-20 04:00:28.990746 mon.b mon.0 158.69.87.83:6789/0 73 : cluster [ERR] Health check failed: no active mgr (MGR_DOWN)" in cluster log
mimic-dev1
master
master
ovh
 
rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml backoff/normal.yaml clusters/{fixed-2.yaml openstack.yaml} fs/btrfs.yaml msgr-failures/fastclose.yaml msgr/async.yaml objectstore/filestore.yaml rados.yaml rocksdb.yaml thrashers/none.yaml workloads/rados_api_tests.yaml}
Command failed (workunit test rados/test.sh) on ovh007 with status 124: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=master TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/rados/test.sh'
mimic-dev1
master
master
ovh
 
rados/singleton/{all/reg11184.yaml fs/xfs.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/filestore.yaml rados.yaml}
Command failed on ovh054 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
 
rados/objectstore/filejournal.yaml
"2017-12-20 03:54:28.894700 mon.0 mon.0 158.69.87.40:6789/0 35 : cluster [ERR] Health check failed: no active mgr (MGR_DOWN)" in cluster log
mimic-dev1
master
master
ovh
 
rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml backoff/peering.yaml clusters/{fixed-2.yaml openstack.yaml} fs/xfs.yaml msgr-failures/few.yaml msgr/random.yaml objectstore/bluestore.yaml rados.yaml rocksdb.yaml thrashers/pggrow.yaml workloads/radosbench.yaml}
Command failed on ovh092 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
 
rados/thrash/{0-size-min-size-overrides/2-size-1-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml backoff/peering_and_degraded.yaml clusters/{fixed-2.yaml openstack.yaml} fs/btrfs.yaml msgr-failures/osd-delay.yaml msgr/simple.yaml objectstore/filestore.yaml rados.yaml rocksdb.yaml thrashers/default.yaml workloads/small-objects.yaml}
Command failed on ovh013 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
 
rados/monthrash/{ceph/ceph.yaml clusters/9-mons.yaml fs/xfs.yaml mon_kv_backend/rocksdb.yaml msgr-failures/mon-delay.yaml msgr/random.yaml objectstore/bluestore.yaml rados.yaml thrashers/one.yaml workloads/rados_mon_workunits.yaml}
Command failed on ovh095 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
 
rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml backoff/normal.yaml clusters/{fixed-2.yaml openstack.yaml} fs/xfs.yaml msgr-failures/fastclose.yaml msgr/async.yaml objectstore/bluestore.yaml rados.yaml rocksdb.yaml thrashers/mapgap.yaml workloads/snaps-few-objects.yaml}
Command failed on ovh062 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
 
rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml backoff/peering.yaml clusters/{fixed-2.yaml openstack.yaml} fs/btrfs.yaml msgr-failures/few.yaml msgr/random.yaml objectstore/filestore.yaml rados.yaml rocksdb.yaml thrashers/morepggrow.yaml workloads/write_fadvise_dontneed.yaml}
Command failed on ovh054 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
 
rados/singleton/{all/resolve_stuck_peering.yaml fs/xfs.yaml msgr-failures/many.yaml msgr/random.yaml objectstore/bluestore.yaml rados.yaml}
Command failed on ovh073 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
 
rados/thrash/{0-size-min-size-overrides/2-size-1-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml backoff/peering_and_degraded.yaml clusters/{fixed-2.yaml openstack.yaml} fs/xfs.yaml msgr-failures/osd-delay.yaml msgr/simple.yaml objectstore/bluestore.yaml rados.yaml rocksdb.yaml thrashers/none.yaml workloads/admin_socket_objecter_requests.yaml}
"2017-12-20 03:46:57.459123 mon.b mon.0 158.69.87.20:6789/0 68 : cluster [ERR] Health check failed: no active mgr (MGR_DOWN)" in cluster log
mimic-dev1
master
master
ovh
 
rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml backoff/normal.yaml clusters/{fixed-2.yaml openstack.yaml} fs/btrfs.yaml msgr-failures/fastclose.yaml msgr/async.yaml objectstore/filestore.yaml rados.yaml rocksdb.yaml thrashers/pggrow.yaml workloads/cache-agent-big.yaml}
Command failed on ovh020 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
 
rados/thrash-erasure-code-big/{cluster/{12-osds.yaml openstack.yaml} fs/xfs.yaml leveldb.yaml msgr-failures/fastclose.yaml objectstore/bluestore.yaml rados.yaml thrashers/mapgap.yaml workloads/ec-rados-plugin=lrc-k=4-m=2-l=3.yaml}
Command failed on ovh039 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
 
rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml backoff/peering.yaml clusters/{fixed-2.yaml openstack.yaml} fs/xfs.yaml msgr-failures/few.yaml msgr/random.yaml objectstore/bluestore.yaml rados.yaml rocksdb.yaml thrashers/default.yaml workloads/cache-agent-small.yaml}
Command failed on ovh017 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
 
rados/thrash/{0-size-min-size-overrides/2-size-1-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml backoff/peering_and_degraded.yaml clusters/{fixed-2.yaml openstack.yaml} fs/btrfs.yaml msgr-failures/osd-delay.yaml msgr/simple.yaml objectstore/filestore.yaml rados.yaml rocksdb.yaml thrashers/mapgap.yaml workloads/cache-pool-snaps-readproxy.yaml}
Command failed on ovh083 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
 
rados/basic/{clusters/{fixed-2.yaml openstack.yaml} fs/xfs.yaml mon_kv_backend/leveldb.yaml msgr-failures/many.yaml msgr/random.yaml objectstore/filestore.yaml rados.yaml tasks/rados_striper.yaml}
Command crashed: 'sudo TESTDIR=/home/ubuntu/cephtest bash -c ceph_test_rados_striper_api_aio'
mimic-dev1
master
master
ovh
 
rados/multimon/{clusters/6.yaml fs/xfs.yaml mon_kv_backend/rocksdb.yaml msgr-failures/many.yaml msgr/async.yaml objectstore/bluestore.yaml rados.yaml tasks/mon_clock_no_skews.yaml}
'timechecks'
mimic-dev1
master
master
ovh
 
rados/thrash-erasure-code/{clusters/{fixed-2.yaml openstack.yaml} fast/fast.yaml fs/btrfs.yaml leveldb.yaml msgr-failures/few.yaml objectstore/filestore.yaml rados.yaml thrashers/default.yaml workloads/ec-small-objects-overwrites.yaml}
Command failed on ovh053 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
centos 
rados/verify/{1thrash/default.yaml clusters/{fixed-2.yaml openstack.yaml} fs/btrfs.yaml mon_kv_backend/rocksdb.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/filestore.yaml rados.yaml tasks/rados_cls_all.yaml validater/valgrind.yaml}
Command failed on ovh052 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
 
rados/singleton-nomsgr/{all/export-after-evict.yaml rados.yaml}
"2017-12-20 03:52:44.168961 mon.a mon.0 158.69.87.37:6789/0 62 : cluster [ERR] Health check failed: no active mgr (MGR_DOWN)" in cluster log
mimic-dev1
master
master
ovh
 
rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml backoff/normal.yaml clusters/{fixed-2.yaml openstack.yaml} fs/xfs.yaml msgr-failures/fastclose.yaml msgr/async.yaml objectstore/bluestore.yaml rados.yaml rocksdb.yaml thrashers/morepggrow.yaml workloads/cache-pool-snaps.yaml}
Command failed on ovh053 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
 
rados/singleton/{all/rest-api.yaml fs/xfs.yaml msgr-failures/few.yaml msgr/simple.yaml objectstore/filestore.yaml rados.yaml}
Command failed (workunit test rest/test.py) on ovh024 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=master TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/rest/test.py'
mimic-dev1
master
master
ovh
 
rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml backoff/peering.yaml clusters/{fixed-2.yaml openstack.yaml} fs/btrfs.yaml msgr-failures/few.yaml msgr/random.yaml objectstore/filestore.yaml rados.yaml rocksdb.yaml thrashers/none.yaml workloads/cache-snaps.yaml}
"2017-12-20 03:48:52.372546 mon.a mon.0 158.69.87.192:6789/0 103 : cluster [ERR] Health check failed: no active mgr (MGR_DOWN)" in cluster log
mimic-dev1
master
master
ovh
 
rados/thrash/{0-size-min-size-overrides/2-size-1-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml backoff/peering_and_degraded.yaml clusters/{fixed-2.yaml openstack.yaml} fs/xfs.yaml msgr-failures/osd-delay.yaml msgr/simple.yaml objectstore/bluestore.yaml rados.yaml rocksdb.yaml thrashers/pggrow.yaml workloads/cache.yaml}
Command failed on ovh098 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
 
rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml backoff/normal.yaml clusters/{fixed-2.yaml openstack.yaml} fs/btrfs.yaml msgr-failures/fastclose.yaml msgr/async.yaml objectstore/filestore.yaml rados.yaml rocksdb.yaml thrashers/default.yaml workloads/pool-snaps-few-objects.yaml}
Command failed on ovh036 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
 
rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml backoff/peering.yaml clusters/{fixed-2.yaml openstack.yaml} fs/xfs.yaml msgr-failures/few.yaml msgr/random.yaml objectstore/bluestore.yaml rados.yaml rocksdb.yaml thrashers/mapgap.yaml workloads/rados_api_tests.yaml}
Command failed on ovh026 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
 
rados/singleton/{all/test_envlibrados_for_rocksdb.yaml fs/xfs.yaml msgr-failures/many.yaml msgr/async.yaml objectstore/bluestore.yaml rados.yaml}
"2017-12-20 03:52:41.910486 mon.a mon.0 158.69.87.38:6789/0 43 : cluster [ERR] Health check failed: no active mgr (MGR_DOWN)" in cluster log
mimic-dev1
master
master
ovh
 
rados/thrash/{0-size-min-size-overrides/2-size-1-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml backoff/peering_and_degraded.yaml clusters/{fixed-2.yaml openstack.yaml} fs/btrfs.yaml msgr-failures/osd-delay.yaml msgr/simple.yaml objectstore/filestore.yaml rados.yaml rocksdb.yaml thrashers/morepggrow.yaml workloads/radosbench.yaml}
Command failed on ovh040 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
 
rados/objectstore/filestore-idempotent-aio-journal.yaml
"2017-12-20 03:43:19.660719 mon.0 mon.0 158.69.87.174:6789/0 35 : cluster [ERR] Health check failed: no active mgr (MGR_DOWN)" in cluster log
mimic-dev1
master
master
ovh
 
rados/thrash-erasure-code-shec/{clusters/{fixed-4.yaml openstack.yaml} fs/xfs.yaml leveldb.yaml msgr-failures/osd-delay.yaml objectstore/bluestore.yaml rados.yaml thrashers/default.yaml workloads/ec-rados-plugin=shec-k=4-m=3-c=2.yaml}
Command failed on ovh003 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
 
rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml backoff/normal.yaml clusters/{fixed-2.yaml openstack.yaml} fs/xfs.yaml msgr-failures/fastclose.yaml msgr/async.yaml objectstore/bluestore.yaml rados.yaml rocksdb.yaml thrashers/none.yaml workloads/small-objects.yaml}
Command failed on ovh018 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
 
rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml backoff/peering.yaml clusters/{fixed-2.yaml openstack.yaml} fs/btrfs.yaml msgr-failures/few.yaml msgr/random.yaml objectstore/filestore.yaml rados.yaml rocksdb.yaml thrashers/pggrow.yaml workloads/snaps-few-objects.yaml}
Command failed on ovh015 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
 
rados/thrash/{0-size-min-size-overrides/2-size-1-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml backoff/peering_and_degraded.yaml clusters/{fixed-2.yaml openstack.yaml} fs/xfs.yaml msgr-failures/osd-delay.yaml msgr/simple.yaml objectstore/bluestore.yaml rados.yaml rocksdb.yaml thrashers/default.yaml workloads/write_fadvise_dontneed.yaml}
Command failed on ovh010 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
 
rados/basic/{clusters/{fixed-2.yaml openstack.yaml} fs/btrfs.yaml mon_kv_backend/rocksdb.yaml msgr-failures/few.yaml msgr/simple.yaml objectstore/bluestore.yaml rados.yaml tasks/rados_workunit_loadgen_big.yaml}
Command failed (workunit test rados/load-gen-big.sh) on ovh022 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/mnt.1/client.1/tmp && cd -- /home/ubuntu/cephtest/mnt.1/client.1/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=master TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="1" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.1 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.1/qa/workunits/rados/load-gen-big.sh'
mimic-dev1
master
master
ovh
 
rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml backoff/normal.yaml clusters/{fixed-2.yaml openstack.yaml} fs/btrfs.yaml msgr-failures/fastclose.yaml msgr/async.yaml objectstore/filestore.yaml rados.yaml rocksdb.yaml thrashers/mapgap.yaml workloads/admin_socket_objecter_requests.yaml}
Command failed on ovh095 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
 
rados/thrash-erasure-code/{clusters/{fixed-2.yaml openstack.yaml} fast/normal.yaml fs/xfs.yaml leveldb.yaml msgr-failures/osd-delay.yaml objectstore/bluestore.yaml rados.yaml thrashers/fastread.yaml workloads/ec-small-objects.yaml}
Command failed on ovh080 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
ubuntu 16.04
rados/thrash-erasure-code-isa/{arch/x86_64.yaml clusters/{fixed-2.yaml openstack.yaml} fs/xfs.yaml leveldb.yaml msgr-failures/osd-delay.yaml objectstore/bluestore.yaml rados.yaml supported/ubuntu_latest.yaml thrashers/none.yaml workloads/ec-rados-plugin=isa-k=2-m=1.yaml}
Command failed on ovh057 with status 22: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph osd erasure-code-profile set isaprofile name=isaprofile plugin=isa k=2 technique=reed_sol_van m=1 ruleset-failure-domain=osd'
mimic-dev1
master
master
ovh
 
rados/singleton/{all/thrash-rados.yaml fs/xfs.yaml msgr-failures/few.yaml msgr/random.yaml objectstore/filestore.yaml rados.yaml}
Command failed on ovh046 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
 
rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml backoff/peering.yaml clusters/{fixed-2.yaml openstack.yaml} fs/xfs.yaml msgr-failures/few.yaml msgr/random.yaml objectstore/bluestore.yaml rados.yaml rocksdb.yaml thrashers/morepggrow.yaml workloads/cache-agent-big.yaml}
Command failed on ovh040 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
 
rados/monthrash/{ceph/ceph.yaml clusters/3-mons.yaml fs/xfs.yaml mon_kv_backend/leveldb.yaml msgr-failures/few.yaml msgr/simple.yaml objectstore/filestore.yaml rados.yaml thrashers/sync-many.yaml workloads/snaps-few-objects.yaml}
Command failed on ovh017 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
 
rados/thrash/{0-size-min-size-overrides/2-size-1-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml backoff/peering_and_degraded.yaml clusters/{fixed-2.yaml openstack.yaml} fs/btrfs.yaml msgr-failures/osd-delay.yaml msgr/simple.yaml objectstore/filestore.yaml rados.yaml rocksdb.yaml thrashers/none.yaml workloads/cache-agent-small.yaml}
"2017-12-20 04:14:34.846276 mon.a mon.0 158.69.88.206:6789/0 103 : cluster [ERR] Health check failed: no active mgr (MGR_DOWN)" in cluster log
mimic-dev1
master
master
ovh
 
rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml backoff/normal.yaml clusters/{fixed-2.yaml openstack.yaml} fs/xfs.yaml msgr-failures/fastclose.yaml msgr/async.yaml objectstore/bluestore.yaml rados.yaml rocksdb.yaml thrashers/pggrow.yaml workloads/cache-pool-snaps-readproxy.yaml}
Command failed on ovh083 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
 
rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml backoff/peering.yaml clusters/{fixed-2.yaml openstack.yaml} fs/btrfs.yaml msgr-failures/few.yaml msgr/random.yaml objectstore/filestore.yaml rados.yaml rocksdb.yaml thrashers/default.yaml workloads/cache-pool-snaps.yaml}
Command failed on ovh015 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
 
rados/singleton/{all/thrash_cache_writeback_proxy_none.yaml fs/xfs.yaml msgr-failures/many.yaml msgr/simple.yaml objectstore/bluestore.yaml rados.yaml}
"2017-12-20 04:15:09.826777 mon.a mon.0 158.69.88.24:6789/0 102 : cluster [ERR] Health check failed: no active mgr (MGR_DOWN)" in cluster log
mimic-dev1
master
master
ovh
 
rados/thrash/{0-size-min-size-overrides/2-size-1-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml backoff/peering_and_degraded.yaml clusters/{fixed-2.yaml openstack.yaml} fs/xfs.yaml msgr-failures/osd-delay.yaml msgr/simple.yaml objectstore/bluestore.yaml rados.yaml rocksdb.yaml thrashers/mapgap.yaml workloads/cache-snaps.yaml}
Command failed on ovh006 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
 
rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml backoff/normal.yaml clusters/{fixed-2.yaml openstack.yaml} fs/btrfs.yaml msgr-failures/fastclose.yaml msgr/async.yaml objectstore/filestore.yaml rados.yaml rocksdb.yaml thrashers/morepggrow.yaml workloads/cache.yaml}
Command failed on ovh090 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
 
rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml backoff/peering.yaml clusters/{fixed-2.yaml openstack.yaml} fs/xfs.yaml msgr-failures/few.yaml msgr/random.yaml objectstore/bluestore.yaml rados.yaml rocksdb.yaml thrashers/none.yaml workloads/pool-snaps-few-objects.yaml}
Command failed on ovh003 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
 
rados/thrash/{0-size-min-size-overrides/2-size-1-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml backoff/peering_and_degraded.yaml clusters/{fixed-2.yaml openstack.yaml} fs/btrfs.yaml msgr-failures/osd-delay.yaml msgr/simple.yaml objectstore/filestore.yaml rados.yaml rocksdb.yaml thrashers/pggrow.yaml workloads/rados_api_tests.yaml}
Command failed on ovh064 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
 
rados/basic/{clusters/{fixed-2.yaml openstack.yaml} fs/xfs.yaml mon_kv_backend/leveldb.yaml msgr-failures/many.yaml msgr/async.yaml objectstore/filestore.yaml rados.yaml tasks/rados_workunit_loadgen_mix.yaml}
Command failed (workunit test rados/load-gen-mix.sh) on ovh085 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/mnt.1/client.1/tmp && cd -- /home/ubuntu/cephtest/mnt.1/client.1/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=master TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="1" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.1 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.1/qa/workunits/rados/load-gen-mix.sh'
mimic-dev1
master
master
ovh
 
rados/singleton-nomsgr/{all/full-tiering.yaml rados.yaml}
"2017-12-20 04:14:33.054799 mon.a mon.0 158.69.88.25:6789/0 74 : cluster [ERR] Health check failed: no active mgr (MGR_DOWN)" in cluster log
mimic-dev1
master
master
ovh
 
rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml backoff/normal.yaml clusters/{fixed-2.yaml openstack.yaml} fs/xfs.yaml msgr-failures/fastclose.yaml msgr/async.yaml objectstore/bluestore.yaml rados.yaml rocksdb.yaml thrashers/default.yaml workloads/radosbench.yaml}
Command failed on ovh078 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
 
rados/thrash-erasure-code-big/{cluster/{12-osds.yaml openstack.yaml} fs/btrfs.yaml leveldb.yaml msgr-failures/few.yaml objectstore/filestore.yaml rados.yaml thrashers/morepggrow.yaml workloads/ec-rados-plugin=lrc-k=4-m=2-l=3.yaml}
Command failed on ovh008 with status 1: 'sudo MALLOC_CHECK_=3 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-osd --cluster ceph --mkfs --mkkey -i 0 --monmap /home/ubuntu/cephtest/ceph.monmap'
mimic-dev1
master
master
ovh
 
rados/singleton/{all/watch-notify-same-primary.yaml fs/xfs.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/filestore.yaml rados.yaml}
"2017-12-20 04:21:18.892234 mon.0 mon.0 158.69.88.51:6789/0 70 : cluster [ERR] Health check failed: no active mgr (MGR_DOWN)" in cluster log
mimic-dev1
master
master
ovh
 
rados/thrash-erasure-code/{clusters/{fixed-2.yaml openstack.yaml} fast/fast.yaml fs/btrfs.yaml leveldb.yaml msgr-failures/fastclose.yaml objectstore/filestore.yaml rados.yaml thrashers/mapgap.yaml workloads/ec-snaps-few-objects-overwrites.yaml}
Command failed on ovh013 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
 
rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml backoff/peering.yaml clusters/{fixed-2.yaml openstack.yaml} fs/btrfs.yaml msgr-failures/few.yaml msgr/random.yaml objectstore/filestore.yaml rados.yaml rocksdb.yaml thrashers/mapgap.yaml workloads/small-objects.yaml}
Command failed on ovh092 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
 
rados/objectstore/filestore-idempotent.yaml
"2017-12-20 04:23:29.695720 mon.0 mon.0 158.69.88.63:6789/0 35 : cluster [ERR] Health check failed: no active mgr (MGR_DOWN)" in cluster log
mimic-dev1
master
master
ovh
 
rados/thrash/{0-size-min-size-overrides/2-size-1-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml backoff/peering_and_degraded.yaml clusters/{fixed-2.yaml openstack.yaml} fs/xfs.yaml msgr-failures/osd-delay.yaml msgr/simple.yaml objectstore/bluestore.yaml rados.yaml rocksdb.yaml thrashers/morepggrow.yaml workloads/snaps-few-objects.yaml}
Command failed on ovh010 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
 
rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml backoff/normal.yaml clusters/{fixed-2.yaml openstack.yaml} fs/btrfs.yaml msgr-failures/fastclose.yaml msgr/async.yaml objectstore/filestore.yaml rados.yaml rocksdb.yaml thrashers/none.yaml workloads/write_fadvise_dontneed.yaml}
Command failed on ovh058 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
 
rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml backoff/peering.yaml clusters/{fixed-2.yaml openstack.yaml} fs/xfs.yaml msgr-failures/few.yaml msgr/random.yaml objectstore/bluestore.yaml rados.yaml rocksdb.yaml thrashers/pggrow.yaml workloads/admin_socket_objecter_requests.yaml}
Command failed on ovh020 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
 
rados/singleton/{all/admin-socket.yaml fs/xfs.yaml msgr-failures/many.yaml msgr/random.yaml objectstore/bluestore.yaml rados.yaml}
"2017-12-20 04:23:17.353667 mon.a mon.0 158.69.88.7:6789/0 40 : cluster [ERR] Health check failed: no active mgr (MGR_DOWN)" in cluster log
mimic-dev1
master
master
ovh
 
rados/thrash/{0-size-min-size-overrides/2-size-1-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml backoff/peering_and_degraded.yaml clusters/{fixed-2.yaml openstack.yaml} fs/btrfs.yaml msgr-failures/osd-delay.yaml msgr/simple.yaml objectstore/filestore.yaml rados.yaml rocksdb.yaml thrashers/default.yaml workloads/cache-agent-big.yaml}
Command failed on ovh054 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
 
rados/multimon/{clusters/9.yaml fs/xfs.yaml mon_kv_backend/leveldb.yaml msgr-failures/few.yaml msgr/random.yaml objectstore/filestore.yaml rados.yaml tasks/mon_clock_with_skews.yaml}
'timechecks'
mimic-dev1
master
master
ovh
 
rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml backoff/normal.yaml clusters/{fixed-2.yaml openstack.yaml} fs/xfs.yaml msgr-failures/fastclose.yaml msgr/async.yaml objectstore/bluestore.yaml rados.yaml rocksdb.yaml thrashers/mapgap.yaml workloads/cache-agent-small.yaml}
Command failed on ovh031 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
 
rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml backoff/peering.yaml clusters/{fixed-2.yaml openstack.yaml} fs/btrfs.yaml msgr-failures/few.yaml msgr/random.yaml objectstore/filestore.yaml rados.yaml rocksdb.yaml thrashers/morepggrow.yaml workloads/cache-pool-snaps-readproxy.yaml}
Command failed on ovh057 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
 
rados/thrash/{0-size-min-size-overrides/2-size-1-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml backoff/peering_and_degraded.yaml clusters/{fixed-2.yaml openstack.yaml} fs/xfs.yaml msgr-failures/osd-delay.yaml msgr/simple.yaml objectstore/bluestore.yaml rados.yaml rocksdb.yaml thrashers/none.yaml workloads/cache-pool-snaps.yaml}
"2017-12-20 04:30:05.838918 mon.b mon.0 158.69.89.13:6789/0 115 : cluster [ERR] Health check failed: no active mgr (MGR_DOWN)" in cluster log
mimic-dev1
master
master
ovh
 
rados/basic/{clusters/{fixed-2.yaml openstack.yaml} fs/btrfs.yaml mon_kv_backend/rocksdb.yaml msgr-failures/few.yaml msgr/random.yaml objectstore/bluestore.yaml rados.yaml tasks/rados_workunit_loadgen_mostlyread.yaml}
Command failed (workunit test rados/load-gen-mostlyread.sh) on ovh030 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/mnt.1/client.1/tmp && cd -- /home/ubuntu/cephtest/mnt.1/client.1/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=master TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="1" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.1 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.1/qa/workunits/rados/load-gen-mostlyread.sh'
mimic-dev1
master
master
ovh
 
rados/thrash-erasure-code-shec/{clusters/{fixed-4.yaml openstack.yaml} fs/xfs.yaml leveldb.yaml msgr-failures/fastclose.yaml objectstore/filestore.yaml rados.yaml thrashers/default.yaml workloads/ec-rados-plugin=shec-k=4-m=3-c=2.yaml}
Command failed on ovh046 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
 
rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml backoff/normal.yaml clusters/{fixed-2.yaml openstack.yaml} fs/btrfs.yaml msgr-failures/fastclose.yaml msgr/async.yaml objectstore/filestore.yaml rados.yaml rocksdb.yaml thrashers/pggrow.yaml workloads/cache-snaps.yaml}
Command failed on ovh040 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
 
rados/singleton/{all/cephtool.yaml fs/xfs.yaml msgr-failures/few.yaml msgr/simple.yaml objectstore/filestore.yaml rados.yaml}
Command failed (workunit test cephtool/test.sh) on ovh015 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=master TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/cephtool/test.sh'
mimic-dev1
master
master
ovh
 
rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml backoff/peering.yaml clusters/{fixed-2.yaml openstack.yaml} fs/xfs.yaml msgr-failures/few.yaml msgr/random.yaml objectstore/bluestore.yaml rados.yaml rocksdb.yaml thrashers/default.yaml workloads/cache.yaml}
Command failed on ovh017 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
 
rados/monthrash/{ceph/ceph.yaml clusters/9-mons.yaml fs/xfs.yaml mon_kv_backend/rocksdb.yaml msgr-failures/mon-delay.yaml msgr/async.yaml objectstore/bluestore.yaml rados.yaml thrashers/sync-many.yaml workloads/pool-create-delete.yaml}
Command failed on ovh083 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
 
rados/thrash-erasure-code/{clusters/{fixed-2.yaml openstack.yaml} fast/normal.yaml fs/xfs.yaml leveldb.yaml msgr-failures/few.yaml objectstore/bluestore.yaml rados.yaml thrashers/morepggrow.yaml workloads/ec-pool-snaps-few-objects-overwrites.yaml}
Command failed on ovh077 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
centos 7.3
rados/thrash-erasure-code-isa/{arch/x86_64.yaml clusters/{fixed-2.yaml openstack.yaml} fs/btrfs.yaml leveldb.yaml msgr-failures/fastclose.yaml objectstore/filestore.yaml rados.yaml supported/centos_latest.yaml thrashers/pggrow.yaml workloads/ec-rados-plugin=isa-k=2-m=1.yaml}
Command failed on ovh090 with status 1: 'sudo MALLOC_CHECK_=3 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-osd --cluster ceph --mkfs --mkkey -i 0 --monmap /home/ubuntu/cephtest/ceph.monmap'
mimic-dev1
master
master
ovh
 
rados/verify/{1thrash/none.yaml clusters/{fixed-2.yaml openstack.yaml} fs/btrfs.yaml mon_kv_backend/leveldb.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/filestore.yaml rados.yaml tasks/mon_recovery.yaml validater/lockdep.yaml}
"2017-12-20 04:38:32.411278 mon.a mon.0 158.69.89.191:6789/0 3 : cluster [WRN] Health check failed: 6 osds exist in the crush map but not in the osdmap (OSD_ORPHAN)" in cluster log
mimic-dev1
master
master
ovh
 
rados/thrash/{0-size-min-size-overrides/2-size-1-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml backoff/peering_and_degraded.yaml clusters/{fixed-2.yaml openstack.yaml} fs/btrfs.yaml msgr-failures/osd-delay.yaml msgr/simple.yaml objectstore/filestore.yaml rados.yaml rocksdb.yaml thrashers/mapgap.yaml workloads/pool-snaps-few-objects.yaml}
Command failed on ovh062 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
 
rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml backoff/normal.yaml clusters/{fixed-2.yaml openstack.yaml} fs/xfs.yaml msgr-failures/fastclose.yaml msgr/async.yaml objectstore/bluestore.yaml rados.yaml rocksdb.yaml thrashers/morepggrow.yaml workloads/rados_api_tests.yaml}
Command failed on ovh018 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
 
rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml backoff/peering.yaml clusters/{fixed-2.yaml openstack.yaml} fs/btrfs.yaml msgr-failures/few.yaml msgr/random.yaml objectstore/filestore.yaml rados.yaml rocksdb.yaml thrashers/none.yaml workloads/radosbench.yaml}
"2017-12-20 04:41:09.511404 mon.b mon.0 158.69.89.248:6789/0 67 : cluster [ERR] Health check failed: no active mgr (MGR_DOWN)" in cluster log
mimic-dev1
master
master
ovh
 
rados/singleton/{all/divergent_priors.yaml fs/xfs.yaml msgr-failures/many.yaml msgr/async.yaml objectstore/bluestore.yaml rados.yaml}
Command failed on ovh033 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
 
rados/thrash/{0-size-min-size-overrides/2-size-1-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml backoff/peering_and_degraded.yaml clusters/{fixed-2.yaml openstack.yaml} fs/xfs.yaml msgr-failures/osd-delay.yaml msgr/simple.yaml objectstore/bluestore.yaml rados.yaml rocksdb.yaml thrashers/pggrow.yaml workloads/small-objects.yaml}
Command failed on ovh073 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
 
rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml backoff/normal.yaml clusters/{fixed-2.yaml openstack.yaml} fs/btrfs.yaml msgr-failures/fastclose.yaml msgr/async.yaml objectstore/filestore.yaml rados.yaml rocksdb.yaml thrashers/default.yaml workloads/snaps-few-objects.yaml}
Command failed on ovh036 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
 
rados/objectstore/fusestore.yaml
mimic-dev1
master
master
ovh
 
rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml backoff/peering.yaml clusters/{fixed-2.yaml openstack.yaml} fs/xfs.yaml msgr-failures/few.yaml msgr/random.yaml objectstore/bluestore.yaml rados.yaml rocksdb.yaml thrashers/mapgap.yaml workloads/write_fadvise_dontneed.yaml}
Command failed on ovh066 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
 
rados/thrash/{0-size-min-size-overrides/2-size-1-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml backoff/peering_and_degraded.yaml clusters/{fixed-2.yaml openstack.yaml} fs/btrfs.yaml msgr-failures/osd-delay.yaml msgr/simple.yaml objectstore/filestore.yaml rados.yaml rocksdb.yaml thrashers/morepggrow.yaml workloads/admin_socket_objecter_requests.yaml}
Command failed on ovh054 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
 
rados/basic/{clusters/{fixed-2.yaml openstack.yaml} fs/xfs.yaml mon_kv_backend/leveldb.yaml msgr-failures/many.yaml msgr/simple.yaml objectstore/filestore.yaml rados.yaml tasks/readwrite.yaml}
Command failed on ovh092 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
 
rados/singleton-nomsgr/{all/msgr.yaml rados.yaml}
mimic-dev1
master
master
ovh
 
rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml backoff/normal.yaml clusters/{fixed-2.yaml openstack.yaml} fs/xfs.yaml msgr-failures/fastclose.yaml msgr/async.yaml objectstore/bluestore.yaml rados.yaml rocksdb.yaml thrashers/none.yaml workloads/cache-agent-big.yaml}
Command failed on ovh093 with status 22: "sudo TESTDIR=/home/ubuntu/cephtest bash -c 'sudo ceph osd erasure-code-profile set teuthologyprofile ruleset-failure-domain=osd m=1 k=2'"
mimic-dev1
master
master
ovh
 
rados/singleton/{all/divergent_priors2.yaml fs/xfs.yaml msgr-failures/few.yaml msgr/random.yaml objectstore/filestore.yaml rados.yaml}
Command failed on ovh013 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
 
rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml backoff/peering.yaml clusters/{fixed-2.yaml openstack.yaml} fs/btrfs.yaml msgr-failures/few.yaml msgr/random.yaml objectstore/filestore.yaml rados.yaml rocksdb.yaml thrashers/pggrow.yaml workloads/cache-agent-small.yaml}
Command failed on ovh020 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
 
rados/thrash/{0-size-min-size-overrides/2-size-1-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml backoff/peering_and_degraded.yaml clusters/{fixed-2.yaml openstack.yaml} fs/xfs.yaml msgr-failures/osd-delay.yaml msgr/simple.yaml objectstore/bluestore.yaml rados.yaml rocksdb.yaml thrashers/default.yaml workloads/cache-pool-snaps-readproxy.yaml}
Command failed on ovh010 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
 
rados/thrash-erasure-code/{clusters/{fixed-2.yaml openstack.yaml} fast/fast.yaml fs/btrfs.yaml leveldb.yaml msgr-failures/osd-delay.yaml objectstore/filestore.yaml rados.yaml thrashers/pggrow.yaml workloads/ec-rados-plugin=jerasure-k=2-m=1.yaml}
Command failed on ovh057 with status 1: 'sudo MALLOC_CHECK_=3 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-osd --cluster ceph --mkfs --mkkey -i 0 --monmap /home/ubuntu/cephtest/ceph.monmap'
mimic-dev1
master
master
ovh
 
rados/thrash-erasure-code-big/{cluster/{12-osds.yaml openstack.yaml} fs/xfs.yaml leveldb.yaml msgr-failures/osd-delay.yaml objectstore/bluestore.yaml rados.yaml thrashers/pggrow.yaml workloads/ec-rados-plugin=lrc-k=4-m=2-l=3.yaml}
Command failed on ovh009 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
 
rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml backoff/normal.yaml clusters/{fixed-2.yaml openstack.yaml} fs/btrfs.yaml msgr-failures/fastclose.yaml msgr/async.yaml objectstore/filestore.yaml rados.yaml rocksdb.yaml thrashers/mapgap.yaml workloads/cache-pool-snaps.yaml}
Command failed on ovh053 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
 
rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml backoff/peering.yaml clusters/{fixed-2.yaml openstack.yaml} fs/xfs.yaml msgr-failures/few.yaml msgr/random.yaml objectstore/bluestore.yaml rados.yaml rocksdb.yaml thrashers/morepggrow.yaml workloads/cache-snaps.yaml}
Command failed on ovh100 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
 
rados/singleton/{all/dump-stuck.yaml fs/xfs.yaml msgr-failures/many.yaml msgr/simple.yaml objectstore/bluestore.yaml rados.yaml}
Command failed on ovh017 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
 
rados/thrash/{0-size-min-size-overrides/2-size-1-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml backoff/peering_and_degraded.yaml clusters/{fixed-2.yaml openstack.yaml} fs/btrfs.yaml msgr-failures/osd-delay.yaml msgr/simple.yaml objectstore/filestore.yaml rados.yaml rocksdb.yaml thrashers/none.yaml workloads/cache.yaml}
"2017-12-20 04:58:48.368121 mon.b mon.0 158.69.90.54:6789/0 102 : cluster [ERR] Health check failed: no active mgr (MGR_DOWN)" in cluster log
mimic-dev1
master
master
ovh
 
rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml backoff/normal.yaml clusters/{fixed-2.yaml openstack.yaml} fs/xfs.yaml msgr-failures/fastclose.yaml msgr/async.yaml objectstore/bluestore.yaml rados.yaml rocksdb.yaml thrashers/pggrow.yaml workloads/pool-snaps-few-objects.yaml}
Command failed on ovh040 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
 
rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml backoff/peering.yaml clusters/{fixed-2.yaml openstack.yaml} fs/btrfs.yaml msgr-failures/few.yaml msgr/random.yaml objectstore/filestore.yaml rados.yaml rocksdb.yaml thrashers/default.yaml workloads/rados_api_tests.yaml}
Command failed on ovh090 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
 
rados/thrash/{0-size-min-size-overrides/2-size-1-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml backoff/peering_and_degraded.yaml clusters/{fixed-2.yaml openstack.yaml} fs/xfs.yaml msgr-failures/osd-delay.yaml msgr/simple.yaml objectstore/bluestore.yaml rados.yaml rocksdb.yaml thrashers/mapgap.yaml workloads/radosbench.yaml}
Command failed on ovh083 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
 
rados/basic/{clusters/{fixed-2.yaml openstack.yaml} fs/btrfs.yaml mon_kv_backend/rocksdb.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/bluestore.yaml rados.yaml tasks/repair_test.yaml}
Command failed on ovh003 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
 
rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml backoff/normal.yaml clusters/{fixed-2.yaml openstack.yaml} fs/btrfs.yaml msgr-failures/fastclose.yaml msgr/async.yaml objectstore/filestore.yaml rados.yaml rocksdb.yaml thrashers/morepggrow.yaml workloads/small-objects.yaml}
Command failed on ovh006 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
 
rados/singleton/{all/ec-lost-unfound.yaml fs/xfs.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/filestore.yaml rados.yaml}
Command failed on ovh082 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
 
rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml backoff/peering.yaml clusters/{fixed-2.yaml openstack.yaml} fs/xfs.yaml msgr-failures/few.yaml msgr/random.yaml objectstore/bluestore.yaml rados.yaml rocksdb.yaml thrashers/none.yaml workloads/snaps-few-objects.yaml}
Command failed on ovh077 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
 
rados/monthrash/{ceph/ceph.yaml clusters/3-mons.yaml fs/xfs.yaml mon_kv_backend/leveldb.yaml msgr-failures/few.yaml msgr/random.yaml objectstore/filestore.yaml rados.yaml thrashers/sync.yaml workloads/rados_5925.yaml}
Command failed on ovh018 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
 
rados/thrash/{0-size-min-size-overrides/2-size-1-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml backoff/peering_and_degraded.yaml clusters/{fixed-2.yaml openstack.yaml} fs/btrfs.yaml msgr-failures/osd-delay.yaml msgr/simple.yaml objectstore/filestore.yaml rados.yaml rocksdb.yaml thrashers/pggrow.yaml workloads/write_fadvise_dontneed.yaml}
Command failed on ovh002 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
 
rados/multimon/{clusters/21.yaml fs/xfs.yaml mon_kv_backend/rocksdb.yaml msgr-failures/many.yaml msgr/simple.yaml objectstore/bluestore.yaml rados.yaml tasks/mon_recovery.yaml}
"2017-12-20 05:12:27.919665 mon.c mon.0 158.69.91.106:6789/0 108 : cluster [ERR] Health check failed: no active mgr (MGR_DOWN)" in cluster log
mimic-dev1
master
master
ovh
 
rados/objectstore/keyvaluedb.yaml
mimic-dev1
master
master
ovh
 
rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml backoff/normal.yaml clusters/{fixed-2.yaml openstack.yaml} fs/xfs.yaml msgr-failures/fastclose.yaml msgr/simple.yaml objectstore/bluestore.yaml rados.yaml rocksdb.yaml thrashers/default.yaml workloads/admin_socket_objecter_requests.yaml}
Command failed on ovh070 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'
mimic-dev1
master
master
ovh
 
rados/thrash-erasure-code/{clusters/{fixed-2.yaml openstack.yaml} fast/normal.yaml fs/xfs.yaml leveldb.yaml msgr-failures/fastclose.yaml objectstore/bluestore.yaml rados.yaml thrashers/default.yaml workloads/ec-rados-plugin=jerasure-k=3-m=1.yaml}
Command failed on ovh062 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json'