Status Job ID Links Posted Started Updated
Runtime
Duration
In Waiting
Machine Teuthology Branch OS Type OS Version Description Nodes
fail 1363914 2017-07-05 19:01:21 2017-07-05 19:09:07 2017-07-05 19:33:07 0:24:00 0:12:03 0:11:57 smithi master rados/thrash-luminous/{0-size-min-size-overrides/2-size-1-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml backoff/normal.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr/async.yaml objectstore/bluestore-comp.yaml rados.yaml rocksdb.yaml thrashers/mapgap.yaml thrashosds-health.yaml workloads/redirect.yaml} 2
Failure Reason:

"2017-07-05 19:20:31.957089 mon.a mon.0 172.21.15.55:6789/0 117 : cluster [WRN] Health check failed: noscrub flag(s) set (OSDMAP_FLAGS)" in cluster log

fail 1363916 2017-07-05 19:01:23 2017-07-05 19:09:07 2017-07-05 19:21:07 0:12:00 0:07:43 0:04:17 smithi master rados/singleton/{all/divergent_priors.yaml msgr-failures/few.yaml msgr/random.yaml objectstore/bluestore-comp.yaml rados.yaml} 1
Failure Reason:

"2017-07-05 19:15:25.202660 mon.a mon.0 172.21.15.10:6789/0 74 : cluster [WRN] Health check failed: noout flag(s) set (OSDMAP_FLAGS)" in cluster log

fail 1363918 2017-07-05 19:01:25 2017-07-05 19:11:04 2017-07-05 19:29:04 0:18:00 0:08:52 0:09:08 smithi master rados/monthrash/{ceph.yaml clusters/3-mons.yaml d-require-luminous/at-end.yaml mon_kv_backend/leveldb.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/bluestore-comp.yaml rados.yaml thrashers/force-sync-many.yaml workloads/rados_5925.yaml} 2
Failure Reason:

"2017-07-05 19:20:06.821597 mon.a mon.0 172.21.15.45:6789/0 112 : cluster [WRN] Health check failed: 1/3 mons down, quorum a,c (MON_DOWN)" in cluster log

fail 1363920 2017-07-05 19:01:27 2017-07-05 19:11:04 2017-07-05 19:29:04 0:18:00 0:09:46 0:08:14 smithi master rados/multimon/{clusters/3.yaml mon_kv_backend/rocksdb.yaml msgr-failures/many.yaml msgr/simple.yaml objectstore/filestore-xfs.yaml rados.yaml tasks/mon_recovery.yaml} 2
Failure Reason:

"2017-07-05 19:22:09.819610 mon.a mon.1 172.21.15.156:6789/0 17 : cluster [WRN] Health check failed: 1/3 mons down, quorum a,c (MON_DOWN)" in cluster log

fail 1363921 2017-07-05 19:01:29 2017-07-05 19:11:05 2017-07-05 19:57:05 0:46:00 0:35:37 0:10:23 smithi master rados/thrash-erasure-code/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-require-luminous/at-end.yaml fast/fast.yaml leveldb.yaml msgr-failures/fastclose.yaml objectstore/bluestore-comp.yaml rados.yaml thrashers/default.yaml thrashosds-health.yaml workloads/ec-rados-plugin=jerasure-k=3-m=1.yaml} 2
Failure Reason:

"2017-07-05 19:21:09.601661 mon.b mon.0 172.21.15.96:6789/0 91 : cluster [WRN] Health check failed: noscrub flag(s) set (OSDMAP_FLAGS)" in cluster log

fail 1363923 2017-07-05 19:01:31 2017-07-05 19:13:12 2017-07-05 19:41:12 0:28:00 0:19:58 0:08:02 smithi master rados/verify/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-require-luminous/at-end.yaml d-thrash/default/{default.yaml thrashosds-health.yaml} mon_kv_backend/leveldb.yaml msgr-failures/few.yaml msgr/random.yaml objectstore/bluestore-comp.yaml rados.yaml tasks/rados_api_tests.yaml validater/lockdep.yaml} 2
Failure Reason:

"2017-07-05 19:20:32.479130 mon.a mon.0 172.21.15.54:6789/0 144 : cluster [WRN] Health check failed: noscrub flag(s) set (OSDMAP_FLAGS)" in cluster log

fail 1363926 2017-07-05 19:01:33 2017-07-05 19:15:00 2017-07-05 19:53:00 0:38:00 0:30:20 0:07:40 smithi master centos 7.3 rados/thrash-erasure-code-isa/{arch/x86_64.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-require-luminous/at-end.yaml leveldb.yaml msgr-failures/fastclose.yaml objectstore/bluestore-comp.yaml rados.yaml supported/centos_latest.yaml thrashers/mapgap.yaml thrashosds-health.yaml workloads/ec-rados-plugin=isa-k=2-m=1.yaml} 2
Failure Reason:

"2017-07-05 19:25:22.146660 mon.a mon.0 172.21.15.138:6789/0 103 : cluster [WRN] Health check failed: noscrub flag(s) set (OSDMAP_FLAGS)" in cluster log

pass 1363928 2017-07-05 19:01:34 2017-07-05 19:15:00 2017-07-05 19:26:59 0:11:59 0:09:38 0:02:21 smithi master rados/basic/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-require-luminous/at-end.yaml mon_kv_backend/leveldb.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/bluestore-comp.yaml rados.yaml tasks/rados_cls_all.yaml} 2
fail 1363930 2017-07-05 19:01:36 2017-07-05 19:17:01 2017-07-05 19:35:01 0:18:00 0:11:16 0:06:44 smithi master rados/thrash-erasure-code-big/{ceph.yaml cluster/{12-osds.yaml openstack.yaml} d-require-luminous/at-end.yaml leveldb.yaml msgr-failures/fastclose.yaml objectstore/bluestore-comp.yaml rados.yaml thrashers/fastread.yaml thrashosds-health.yaml workloads/ec-rados-plugin=lrc-k=4-m=2-l=3.yaml} 4
Failure Reason:

"2017-07-05 19:26:31.164826 mon.a mon.0 172.21.15.1:6789/0 239 : cluster [WRN] Health check failed: noscrub flag(s) set (OSDMAP_FLAGS)" in cluster log

fail 1363932 2017-07-05 19:01:39 2017-07-05 19:17:01 2017-07-05 19:47:01 0:30:00 0:23:33 0:06:27 smithi master rados/thrash-erasure-code-overwrites/{bluestore.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-require-luminous/at-end.yaml fast/fast.yaml leveldb.yaml msgr-failures/fastclose.yaml rados.yaml thrashers/fastread.yaml thrashosds-health.yaml workloads/ec-small-objects-overwrites.yaml} 2
Failure Reason:

"2017-07-05 19:23:39.753315 mon.b mon.0 172.21.15.143:6789/0 109 : cluster [WRN] Health check failed: noscrub flag(s) set (OSDMAP_FLAGS)" in cluster log

fail 1363934 2017-07-05 19:01:42 2017-07-05 19:17:02 2017-07-05 19:31:01 0:13:59 0:09:24 0:04:35 smithi master rados/basic-luminous/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} objectstore/bluestore-comp.yaml rados.yaml scrub_test.yaml} 2
Failure Reason:

"2017-07-05 19:24:43.294559 mon.a mon.0 172.21.15.23:6789/0 122 : cluster [ERR] Health check failed: 2 scrub errors (OSD_SCRUB_ERRORS)" in cluster log

fail 1363936 2017-07-05 19:01:46 2017-07-05 19:17:01 2017-07-05 19:31:01 0:14:00 0:08:22 0:05:38 smithi master rados/mgr/{clusters/2-node-mgr.yaml debug/mgr.yaml objectstore/bluestore-comp.yaml tasks/failover.yaml} 2
Failure Reason:

"2017-07-05 19:22:12.656952 mon.b mon.0 172.21.15.58:6789/0 108 : cluster [WRN] Health check failed: no active mgr (MGR_DOWN)" in cluster log

pass 1363937 2017-07-05 19:01:50 2017-07-05 19:17:02 2017-07-05 19:23:01 0:05:59 0:04:41 0:01:18 smithi master rados/objectstore/alloc-hint.yaml 1
pass 1363939 2017-07-05 19:01:53 2017-07-05 19:17:02 2017-07-05 19:25:01 0:07:59 0:05:53 0:02:06 smithi master rados/rest/mgr-restful.yaml 1
fail 1363941 2017-07-05 19:01:57 2017-07-05 19:17:02 2017-07-05 19:35:01 0:17:59 0:12:05 0:05:54 smithi master rados/singleton-bluestore/{all/cephtool.yaml msgr-failures/few.yaml msgr/random.yaml objectstore/bluestore-comp.yaml rados.yaml} 1
Failure Reason:

Command failed (workunit test cephtool/test.sh) on smithi038 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=wip-health TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/cephtool/test.sh'

fail 1363943 2017-07-05 19:02:01 2017-07-05 19:17:14 2017-07-05 19:31:13 0:13:59 0:08:47 0:05:12 smithi master rados/singleton-nomsgr/{all/admin_socket_output.yaml rados.yaml} 1
Failure Reason:

"2017-07-05 19:24:19.580569 mon.a mon.0 172.21.15.141:6789/0 157 : cluster [ERR] Health check failed: 1 full osd(s) (OSD_FULL)" in cluster log

fail 1363945 2017-07-05 19:02:05 2017-07-05 19:19:02 2017-07-05 21:09:03 1:50:01 1:43:18 0:06:43 smithi master rados/upgrade/jewel-x-singleton/{0-cluster/{openstack.yaml start.yaml} 1-jewel-install/jewel.yaml 2-partial-upgrade/firsthalf.yaml 3-thrash/default.yaml 4-workload/{rbd-cls.yaml rbd-import-export.yaml readwrite.yaml snaps-few-objects.yaml} 5-workload/{radosbench.yaml rbd_api.yaml} 6-finish-upgrade.yaml 7-luminous.yaml 8-workload/{rbd-python.yaml rgw-swift.yaml snaps-many-objects.yaml} thrashosds-health.yaml} 3
Failure Reason:

Command failed on smithi085 with status 1: "SWIFT_TEST_CONFIG_FILE=/home/ubuntu/cephtest/archive/testswift.client.0.conf /home/ubuntu/cephtest/swift/virtualenv/bin/nosetests -w /home/ubuntu/cephtest/swift/test/functional -v -a '!fails_on_rgw'"

fail 1363947 2017-07-05 19:02:10 2017-07-05 19:19:02 2017-07-05 19:41:01 0:21:59 0:17:58 0:04:01 smithi master rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml backoff/peering.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-require-luminous/at-mkfs.yaml msgr-failures/few.yaml msgr/random.yaml objectstore/bluestore.yaml rados.yaml rocksdb.yaml thrashers/pggrow.yaml thrashosds-health.yaml workloads/cache.yaml} 2
Failure Reason:

"2017-07-05 19:25:02.556256 mon.a mon.0 172.21.15.136:6789/0 123 : cluster [WRN] Health check failed: noscrub flag(s) set (OSDMAP_FLAGS)" in cluster log

fail 1363949 2017-07-05 19:02:15 2017-07-05 19:19:02 2017-07-05 19:31:01 0:11:59 0:09:07 0:02:52 smithi master rados/singleton/{all/divergent_priors2.yaml msgr-failures/many.yaml msgr/simple.yaml objectstore/bluestore.yaml rados.yaml} 1
Failure Reason:

"2017-07-05 19:24:15.918016 mon.a mon.0 172.21.15.197:6789/0 72 : cluster [WRN] Health check failed: noout flag(s) set (OSDMAP_FLAGS)" in cluster log

fail 1363951 2017-07-05 19:02:19 2017-07-05 19:19:02 2017-07-05 19:31:01 0:11:59 0:07:24 0:04:35 smithi master rados/singleton/{all/dump-stuck.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/filestore-xfs.yaml rados.yaml} 1
Failure Reason:

"2017-07-05 19:25:03.266567 mon.a mon.0 172.21.15.36:6789/0 83 : cluster [WRN] Health check failed: Reduced data availability: 8 pgs peering (PG_AVAILABILITY)" in cluster log

fail 1363953 2017-07-05 19:02:23 2017-07-05 19:19:02 2017-07-05 20:01:02 0:42:00 0:36:56 0:05:04 smithi master rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml backoff/peering_and_degraded.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-require-luminous/at-end.yaml msgr-failures/osd-delay.yaml msgr/simple.yaml objectstore/filestore-xfs.yaml rados.yaml rocksdb.yaml thrashers/default.yaml thrashosds-health.yaml workloads/pool-snaps-few-objects.yaml} 2
Failure Reason:

"2017-07-05 19:24:37.514983 mon.b mon.0 172.21.15.140:6789/0 91 : cluster [WRN] Health check failed: noscrub flag(s) set (OSDMAP_FLAGS)" in cluster log

fail 1363955 2017-07-05 19:02:27 2017-07-05 19:19:02 2017-07-05 19:35:01 0:15:59 0:11:40 0:04:19 smithi master rados/basic/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-require-luminous/at-mkfs.yaml mon_kv_backend/rocksdb.yaml msgr-failures/many.yaml msgr/random.yaml objectstore/bluestore.yaml rados.yaml tasks/rados_python.yaml} 2
Failure Reason:

"2017-07-05 19:25:41.318363 mon.a mon.0 172.21.15.3:6789/0 87 : cluster [WRN] Health check failed: noup flag(s) set (OSDMAP_FLAGS)" in cluster log

fail 1363957 2017-07-05 19:02:31 2017-07-05 19:19:14 2017-07-05 19:41:14 0:22:00 0:17:18 0:04:42 smithi master rados/singleton/{all/ec-lost-unfound.yaml msgr-failures/many.yaml msgr/random.yaml objectstore/bluestore-comp.yaml rados.yaml} 1
Failure Reason:

"2017-07-05 19:23:16.427088 mon.a mon.0 172.21.15.179:6789/0 128 : cluster [WRN] Health check failed: 1 osds down (OSD_DOWN)" in cluster log

fail 1363959 2017-07-05 19:02:35 2017-07-05 19:21:02 2017-07-05 19:31:01 0:09:59 0:06:35 0:03:24 smithi master rados/singleton-nomsgr/{all/cache-fs-trunc.yaml rados.yaml} 1
Failure Reason:

"2017-07-05 19:26:10.559545 mon.a mon.0 172.21.15.137:6789/0 123 : cluster [WRN] Health check failed: 1 cache pools are missing hit_sets (CACHE_POOL_NO_HIT_SET)" in cluster log

fail 1363961 2017-07-05 19:02:38 2017-07-05 19:21:02 2017-07-05 19:45:02 0:24:00 0:20:29 0:03:31 smithi master rados/thrash/{0-size-min-size-overrides/2-size-1-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml backoff/normal.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-require-luminous/at-mkfs.yaml msgr-failures/fastclose.yaml msgr/async.yaml objectstore/bluestore-comp.yaml rados.yaml rocksdb.yaml thrashers/mapgap.yaml thrashosds-health.yaml workloads/rados_api_tests.yaml} 2
Failure Reason:

"2017-07-05 19:26:54.084243 mon.b mon.0 172.21.15.14:6789/0 63 : cluster [WRN] Health check failed: noscrub flag(s) set (OSDMAP_FLAGS)" in cluster log

fail 1363963 2017-07-05 19:02:42 2017-07-05 19:21:02 2017-07-05 19:37:01 0:15:59 0:10:30 0:05:29 smithi master rados/thrash-erasure-code-shec/{ceph.yaml clusters/{fixed-4.yaml openstack.yaml} d-require-luminous/at-mkfs.yaml leveldb.yaml msgr-failures/few.yaml objectstore/filestore-xfs.yaml rados.yaml thrashers/default.yaml thrashosds-health.yaml workloads/ec-rados-plugin=shec-k=4-m=3-c=2.yaml} 4
Failure Reason:

"2017-07-05 19:28:32.281083 mon.c mon.0 172.21.15.51:6789/0 133 : cluster [WRN] Health check failed: noscrub flag(s) set (OSDMAP_FLAGS)" in cluster log

fail 1363965 2017-07-05 19:02:47 2017-07-05 19:21:02 2017-07-05 19:43:02 0:22:00 0:19:15 0:02:45 smithi master rados/objectstore/ceph_objectstore_tool.yaml 1
Failure Reason:

"2017-07-05 19:24:56.613083 mon.a mon.0 172.21.15.193:6789/0 113 : cluster [WRN] Health check failed: noout flag(s) set (OSDMAP_FLAGS)" in cluster log

fail 1363967 2017-07-05 19:02:50 2017-07-05 19:21:02 2017-07-05 19:53:02 0:32:00 0:25:24 0:06:36 smithi master rados/singleton/{all/lost-unfound-delete.yaml msgr-failures/few.yaml msgr/simple.yaml objectstore/bluestore.yaml rados.yaml} 1
Failure Reason:

"2017-07-05 19:29:57.351384 mon.a mon.0 172.21.15.10:6789/0 112 : cluster [WRN] Health check failed: 1 osds down (OSD_DOWN)" in cluster log

fail 1363969 2017-07-05 19:02:53 2017-07-05 19:21:02 2017-07-05 19:41:01 0:19:59 0:14:15 0:05:44 smithi master rados/thrash-luminous/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml backoff/peering.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr/random.yaml objectstore/bluestore.yaml rados.yaml rocksdb.yaml thrashers/morepggrow.yaml thrashosds-health.yaml workloads/redirect_set_object.yaml} 2
Failure Reason:

"2017-07-05 19:26:55.653499 mon.b mon.0 172.21.15.74:6789/0 113 : cluster [WRN] Health check failed: noscrub flag(s) set (OSDMAP_FLAGS)" in cluster log

fail 1363971 2017-07-05 19:02:56 2017-07-05 19:21:08 2017-07-05 22:37:12 3:16:04 3:09:39 0:06:25 smithi master rados/monthrash/{ceph.yaml clusters/9-mons.yaml d-require-luminous/at-mkfs.yaml mon_kv_backend/rocksdb.yaml msgr-failures/mon-delay.yaml msgr/random.yaml objectstore/bluestore.yaml rados.yaml thrashers/many.yaml workloads/rados_api_tests.yaml} 2
Failure Reason:

Command failed (workunit test rados/test.sh) on smithi022 with status 124: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=wip-health TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/rados/test.sh'

fail 1363973 2017-07-05 19:02:59 2017-07-05 19:23:02 2017-07-05 19:45:02 0:22:00 0:18:22 0:03:38 smithi master rados/singleton/{all/lost-unfound.yaml msgr-failures/many.yaml msgr/async.yaml objectstore/filestore-xfs.yaml rados.yaml} 1
Failure Reason:

"2017-07-05 19:26:42.324157 mon.a mon.0 172.21.15.59:6789/0 84 : cluster [WRN] Health check failed: 1 osds down (OSD_DOWN)" in cluster log

fail 1363975 2017-07-05 19:03:03 2017-07-05 19:23:03 2017-07-05 20:05:03 0:42:00 0:37:32 0:04:28 smithi master rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml backoff/peering.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-require-luminous/at-end.yaml msgr-failures/few.yaml msgr/random.yaml objectstore/bluestore.yaml rados.yaml rocksdb.yaml thrashers/morepggrow.yaml thrashosds-health.yaml workloads/radosbench.yaml} 2
Failure Reason:

Command failed on smithi180 with status 22: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph osd pool create unique_pool_5 16'

fail 1363977 2017-07-05 19:03:06 2017-07-05 19:23:03 2017-07-05 20:21:03 0:58:00 0:52:11 0:05:49 smithi master rados/thrash-erasure-code/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-require-luminous/at-mkfs.yaml fast/normal.yaml leveldb.yaml msgr-failures/few.yaml objectstore/bluestore.yaml rados.yaml thrashers/fastread.yaml thrashosds-health.yaml workloads/ec-radosbench.yaml} 2
Failure Reason:

"2017-07-05 19:40:18.868390 mon.a mon.0 172.21.15.9:6789/0 2171 : cluster [ERR] Health check failed: full ratio(s) out of order (OSD_OUT_OF_ORDER_FULL)" in cluster log

fail 1363979 2017-07-05 19:03:11 2017-07-05 19:23:03 2017-07-05 19:57:03 0:34:00 0:28:45 0:05:15 smithi master ubuntu 14.04 rados/thrash-erasure-code-isa/{arch/x86_64.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-require-luminous/at-mkfs.yaml leveldb.yaml msgr-failures/few.yaml objectstore/bluestore.yaml rados.yaml supported/ubuntu_14.04.yaml thrashers/morepggrow.yaml thrashosds-health.yaml workloads/ec-rados-plugin=isa-k=2-m=1.yaml} 2
Failure Reason:

"2017-07-05 19:28:50.427549 mon.a mon.0 172.21.15.60:6789/0 191 : cluster [ERR] Health check failed: full ratio(s) out of order (OSD_OUT_OF_ORDER_FULL)" in cluster log

fail 1363981 2017-07-05 19:03:14 2017-07-05 19:23:03 2017-07-05 19:43:02 0:19:59 0:15:55 0:04:04 smithi master rados/basic/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-require-luminous/at-end.yaml mon_kv_backend/leveldb.yaml msgr-failures/few.yaml msgr/simple.yaml objectstore/filestore-xfs.yaml rados.yaml tasks/rados_stress_watch.yaml} 2
Failure Reason:

"2017-07-05 19:29:57.461041 mon.a mon.0 172.21.15.151:6789/0 199 : cluster [WRN] Health check failed: 1 cache pools are missing hit_sets (CACHE_POOL_NO_HIT_SET)" in cluster log

fail 1363982 2017-07-05 19:03:19 2017-07-05 19:23:03 2017-07-05 19:39:02 0:15:59 0:10:07 0:05:52 smithi master rados/thrash-erasure-code-big/{ceph.yaml cluster/{12-osds.yaml openstack.yaml} d-require-luminous/at-mkfs.yaml leveldb.yaml msgr-failures/few.yaml objectstore/bluestore.yaml rados.yaml thrashers/mapgap.yaml thrashosds-health.yaml workloads/ec-rados-plugin=lrc-k=4-m=2-l=3.yaml} 4
Failure Reason:

"2017-07-05 19:29:44.263610 mon.b mon.0 172.21.15.24:6789/0 112 : cluster [WRN] Health check failed: noscrub flag(s) set (OSDMAP_FLAGS)" in cluster log

fail 1363984 2017-07-05 19:03:24 2017-07-05 19:23:03 2017-07-05 19:57:03 0:34:00 0:28:46 0:05:14 smithi master rados/thrash-erasure-code-overwrites/{bluestore.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-require-luminous/at-mkfs.yaml fast/normal.yaml leveldb.yaml msgr-failures/few.yaml rados.yaml thrashers/mapgap.yaml thrashosds-health.yaml workloads/ec-snaps-few-objects-overwrites.yaml} 2
Failure Reason:

"2017-07-05 19:29:59.205997 mon.a mon.0 172.21.15.21:6789/0 212 : cluster [ERR] Health check failed: full ratio(s) out of order (OSD_OUT_OF_ORDER_FULL)" in cluster log

fail 1363986 2017-07-05 19:03:28 2017-07-05 19:23:03 2017-07-05 19:35:02 0:11:59 0:05:47 0:06:12 smithi master rados/multimon/{clusters/6.yaml mon_kv_backend/leveldb.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/bluestore-comp.yaml rados.yaml tasks/mon_clock_no_skews.yaml} 2
Failure Reason:

global name 'self' is not defined

fail 1363989 2017-07-05 19:03:34 2017-07-05 19:23:03 2017-07-05 19:55:03 0:32:00 0:27:48 0:04:12 smithi master rados/singleton/{all/mon-auth-caps.yaml msgr-failures/few.yaml msgr/random.yaml objectstore/bluestore-comp.yaml rados.yaml} 1
Failure Reason:

"2017-07-05 19:28:14.300893 mon.a mon.0 172.21.15.20:6789/0 39 : cluster [WRN] Health check failed: Degraded data redundancy: 8 pgs degraded (PG_DEGRADED)" in cluster log

pass 1363990 2017-07-05 19:03:37 2017-07-05 19:23:14 2017-07-05 19:29:13 0:05:59 0:04:22 0:01:37 smithi master rados/singleton-nomsgr/{all/ceph-post-file.yaml rados.yaml} 1
pass 1363992 2017-07-05 19:03:41 2017-07-05 19:25:04 2017-07-05 19:47:04 0:22:00 0:17:43 0:04:17 smithi master rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml backoff/peering_and_degraded.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-require-luminous/at-mkfs.yaml msgr-failures/osd-delay.yaml msgr/simple.yaml objectstore/filestore-xfs.yaml rados.yaml rocksdb.yaml thrashers/none.yaml thrashosds-health.yaml workloads/small-objects.yaml} 2
pass 1363994 2017-07-05 19:03:46 2017-07-05 19:25:04 2017-07-05 19:31:04 0:06:00 0:04:37 0:01:23 smithi master rados/objectstore/filejournal.yaml 1
pass 1363997 2017-07-05 19:03:50 2017-07-05 19:25:04 2017-07-05 19:39:04 0:14:00 0:11:24 0:02:36 smithi master rados/singleton/{all/mon-config-keys.yaml msgr-failures/many.yaml msgr/simple.yaml objectstore/bluestore.yaml rados.yaml} 1
fail 1363999 2017-07-05 19:03:53 2017-07-05 19:25:04 2017-07-05 19:49:04 0:24:00 0:20:10 0:03:50 smithi master centos rados/verify/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-require-luminous/at-mkfs.yaml d-thrash/none.yaml mon_kv_backend/rocksdb.yaml msgr-failures/few.yaml msgr/simple.yaml objectstore/bluestore.yaml rados.yaml tasks/rados_cls_all.yaml validater/valgrind.yaml} 2
Failure Reason:

saw valgrind issues

pass 1364001 2017-07-05 19:03:58 2017-07-05 19:25:14 2017-07-05 19:35:14 0:10:00 0:07:45 0:02:15 smithi master rados/basic/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-require-luminous/at-mkfs.yaml mon_kv_backend/rocksdb.yaml msgr-failures/many.yaml msgr/async.yaml objectstore/bluestore-comp.yaml rados.yaml tasks/rados_striper.yaml} 2
fail 1364003 2017-07-05 19:04:02 2017-07-05 19:27:01 2017-07-05 19:43:00 0:15:59 0:13:08 0:02:51 smithi master rados/singleton-bluestore/{all/cephtool.yaml msgr-failures/many.yaml msgr/simple.yaml objectstore/bluestore.yaml rados.yaml} 1
Failure Reason:

Command failed (workunit test cephtool/test.sh) on smithi004 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=wip-health TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/cephtool/test.sh'

pass 1364006 2017-07-05 19:04:06 2017-07-05 19:27:01 2017-07-05 19:33:00 0:05:59 0:05:42 0:00:17 smithi master rados/singleton/{all/mon-seesaw.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/filestore-xfs.yaml rados.yaml} 1
fail 1364008 2017-07-05 19:04:09 2017-07-05 19:27:01 2017-07-05 20:03:01 0:36:00 0:31:40 0:04:20 smithi master rados/thrash/{0-size-min-size-overrides/2-size-1-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml backoff/normal.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-require-luminous/at-end.yaml msgr-failures/fastclose.yaml msgr/async.yaml objectstore/bluestore-comp.yaml rados.yaml rocksdb.yaml thrashers/pggrow.yaml thrashosds-health.yaml workloads/snaps-few-objects.yaml} 2
Failure Reason:

"2017-07-05 19:32:59.429262 mon.a mon.0 172.21.15.152:6789/0 100 : cluster [WRN] Health check failed: noscrub flag(s) set (OSDMAP_FLAGS)" in cluster log

fail 1364009 2017-07-05 19:04:13 2017-07-05 19:27:01 2017-07-05 19:47:01 0:20:00 0:16:44 0:03:16 smithi master rados/singleton/{all/mon-thrasher.yaml msgr-failures/many.yaml msgr/random.yaml objectstore/bluestore-comp.yaml rados.yaml} 1
Failure Reason:

"2017-07-05 19:32:18.580773 mon.a mon.0 172.21.15.149:6789/0 58 : cluster [WRN] Health check failed: 1/3 mons down, quorum a,c (MON_DOWN)" in cluster log

pass 1364012 2017-07-05 19:04:17 2017-07-05 19:27:01 2017-07-05 19:39:00 0:11:59 0:10:22 0:01:37 smithi master rados/thrash-luminous/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml backoff/peering_and_degraded.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr/simple.yaml objectstore/filestore-xfs.yaml rados.yaml rocksdb.yaml thrashers/none.yaml thrashosds-health.yaml workloads/redirect.yaml} 2
fail 1364014 2017-07-05 19:04:21 2017-07-05 19:27:01 2017-07-05 19:37:00 0:09:59 0:04:41 0:05:18 smithi master rados/singleton-nomsgr/{all/export-after-evict.yaml rados.yaml} 1
Failure Reason:

"2017-07-05 19:30:10.115892 mon.a mon.0 172.21.15.46:6789/0 78 : cluster [WRN] Health check failed: 1 cache pools are missing hit_sets (CACHE_POOL_NO_HIT_SET)" in cluster log

fail 1364016 2017-07-05 19:04:25 2017-07-05 19:27:06 2017-07-05 19:43:05 0:15:59 0:10:29 0:05:30 smithi master rados/monthrash/{ceph.yaml clusters/3-mons.yaml d-require-luminous/at-end.yaml mon_kv_backend/leveldb.yaml msgr-failures/few.yaml msgr/simple.yaml objectstore/filestore-xfs.yaml rados.yaml thrashers/one.yaml workloads/rados_mon_workunits.yaml} 2
Failure Reason:

"2017-07-05 19:32:16.442876 mon.a mon.0 172.21.15.67:6789/0 97 : cluster [WRN] Health check failed: 1/3 mons down, quorum a,c (MON_DOWN)" in cluster log

fail 1364018 2017-07-05 19:04:28 2017-07-05 19:27:14 2017-07-05 19:45:14 0:18:00 0:13:45 0:04:15 smithi master rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml backoff/peering.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-require-luminous/at-mkfs.yaml msgr-failures/few.yaml msgr/random.yaml objectstore/bluestore.yaml rados.yaml rocksdb.yaml thrashers/default.yaml thrashosds-health.yaml workloads/write_fadvise_dontneed.yaml} 2
Failure Reason:

"2017-07-05 19:32:01.340568 mon.a mon.0 172.21.15.75:6789/0 120 : cluster [WRN] Health check failed: noscrub flag(s) set (OSDMAP_FLAGS)" in cluster log

fail 1364020 2017-07-05 19:04:32 2017-07-05 19:29:02 2017-07-05 19:55:02 0:26:00 0:20:59 0:05:01 smithi master rados/thrash-erasure-code/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-require-luminous/at-end.yaml fast/fast.yaml leveldb.yaml msgr-failures/osd-delay.yaml objectstore/filestore-xfs.yaml rados.yaml thrashers/mapgap.yaml thrashosds-health.yaml workloads/ec-small-objects-fast-read.yaml} 2
Failure Reason:

"2017-07-05 19:33:25.424845 mon.a mon.0 172.21.15.173:6789/0 121 : cluster [WRN] Health check failed: noscrub flag(s) set (OSDMAP_FLAGS)" in cluster log

fail 1364022 2017-07-05 19:04:35 2017-07-05 19:29:02 2017-07-05 19:47:02 0:18:00 0:13:29 0:04:31 smithi master rados/singleton/{all/osd-backfill.yaml msgr-failures/few.yaml msgr/simple.yaml objectstore/bluestore.yaml rados.yaml} 1
Failure Reason:

"2017-07-05 19:33:34.524194 mon.a mon.0 172.21.15.45:6789/0 89 : cluster [WRN] Health check failed: Degraded data redundancy: 18333/50106 objects degraded (36.588%), 6 pgs degraded (PG_DEGRADED)" in cluster log

pass 1364024 2017-07-05 19:04:39 2017-07-05 19:29:03 2017-07-05 19:55:02 0:25:59 0:24:44 0:01:15 smithi master ubuntu 16.04 rados/thrash-erasure-code-isa/{arch/x86_64.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-require-luminous/at-end.yaml leveldb.yaml msgr-failures/osd-delay.yaml objectstore/filestore-xfs.yaml rados.yaml supported/ubuntu_latest.yaml thrashers/none.yaml thrashosds-health.yaml workloads/ec-rados-plugin=isa-k=2-m=1.yaml} 2
pass 1364026 2017-07-05 19:04:43 2017-07-05 19:29:02 2017-07-05 19:59:02 0:30:00 0:28:36 0:01:24 smithi master rados/basic/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-require-luminous/at-end.yaml mon_kv_backend/leveldb.yaml msgr-failures/few.yaml msgr/random.yaml objectstore/bluestore.yaml rados.yaml tasks/rados_workunit_loadgen_big.yaml} 2
fail 1364028 2017-07-05 19:04:47 2017-07-05 19:29:02 2017-07-05 19:45:02 0:16:00 0:11:28 0:04:32 smithi master rados/thrash-erasure-code-big/{ceph.yaml cluster/{12-osds.yaml openstack.yaml} d-require-luminous/at-end.yaml leveldb.yaml msgr-failures/osd-delay.yaml objectstore/filestore-xfs.yaml rados.yaml thrashers/morepggrow.yaml thrashosds-health.yaml workloads/ec-rados-plugin=lrc-k=4-m=2-l=3.yaml} 4
Failure Reason:

"2017-07-05 19:35:56.609303 mon.a mon.0 172.21.15.11:6789/0 217 : cluster [WRN] Health check failed: noscrub flag(s) set (OSDMAP_FLAGS)" in cluster log

fail 1364030 2017-07-05 19:04:51 2017-07-05 19:29:03 2017-07-05 20:29:04 1:00:01 0:52:57 0:07:04 smithi master rados/thrash-erasure-code-overwrites/{bluestore.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-require-luminous/at-end.yaml fast/fast.yaml leveldb.yaml msgr-failures/osd-delay.yaml rados.yaml thrashers/morepggrow.yaml thrashosds-health.yaml workloads/ec-pool-snaps-few-objects-overwrites.yaml} 2
Failure Reason:

"2017-07-05 19:39:20.507134 mon.b mon.0 172.21.15.16:6789/0 151 : cluster [WRN] Health check failed: noscrub flag(s) set (OSDMAP_FLAGS)" in cluster log

fail 1364033 2017-07-05 19:04:53 2017-07-05 19:29:05 2017-07-05 19:47:05 0:18:00 0:12:41 0:05:19 smithi master rados/thrash-erasure-code-shec/{ceph.yaml clusters/{fixed-4.yaml openstack.yaml} d-require-luminous/at-end.yaml leveldb.yaml msgr-failures/few.yaml objectstore/bluestore-comp.yaml rados.yaml thrashers/default.yaml thrashosds-health.yaml workloads/ec-rados-plugin=shec-k=4-m=3-c=2.yaml} 4
Failure Reason:

"2017-07-05 19:37:43.702690 mon.a mon.0 172.21.15.154:6789/0 186 : cluster [WRN] Health check failed: noscrub flag(s) set (OSDMAP_FLAGS)" in cluster log

fail 1364035 2017-07-05 19:04:56 2017-07-05 19:29:05 2017-07-05 19:43:05 0:14:00 0:08:33 0:05:27 smithi master rados/basic-luminous/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} objectstore/bluestore.yaml rados.yaml scrub_test.yaml} 2
fail 1364036 2017-07-05 19:04:58 2017-07-05 19:29:14 2017-07-05 19:45:14 0:16:00 0:10:38 0:05:22 smithi master rados/mgr/{clusters/2-node-mgr.yaml debug/mgr.yaml objectstore/bluestore.yaml tasks/failover.yaml} 2
Failure Reason:

"2017-07-05 19:35:52.619162 mon.b mon.0 172.21.15.31:6789/0 92 : cluster [WRN] Health check failed: no active mgr (MGR_DOWN)" in cluster log

pass 1364038 2017-07-05 19:05:01 2017-07-05 19:31:02 2017-07-06 00:11:07 4:40:05 4:38:27 0:01:38 smithi master rados/objectstore/filestore-idempotent-aio-journal.yaml 1
fail 1364040 2017-07-05 19:05:03 2017-07-05 19:31:02 2017-07-05 19:49:02 0:18:00 0:13:13 0:04:47 smithi master rados/singleton/{all/osd-recovery-incomplete.yaml msgr-failures/many.yaml msgr/async.yaml objectstore/filestore-xfs.yaml rados.yaml} 1
Failure Reason:

"2017-07-05 19:35:10.681400 mon.a mon.0 172.21.15.170:6789/0 82 : cluster [WRN] Health check failed: 25882/45982 objects misplaced (56.287%) (OBJECT_MISPLACED)" in cluster log

fail 1364042 2017-07-05 19:05:06 2017-07-05 19:31:02 2017-07-05 19:49:02 0:18:00 0:11:32 0:06:28 smithi master rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml backoff/peering_and_degraded.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-require-luminous/at-end.yaml msgr-failures/osd-delay.yaml msgr/simple.yaml objectstore/filestore-xfs.yaml rados.yaml rocksdb.yaml thrashers/mapgap.yaml thrashosds-health.yaml workloads/admin_socket_objecter_requests.yaml} 2
Failure Reason:

"2017-07-05 19:40:22.044478 mon.b mon.0 172.21.15.134:6789/0 130 : cluster [WRN] Health check failed: noscrub flag(s) set (OSDMAP_FLAGS)" in cluster log

fail 1364044 2017-07-05 19:05:09 2017-07-05 19:31:02 2017-07-05 19:55:02 0:24:00 0:17:42 0:06:18 smithi master rados/multimon/{clusters/9.yaml mon_kv_backend/rocksdb.yaml msgr-failures/many.yaml msgr/random.yaml objectstore/bluestore.yaml rados.yaml tasks/mon_clock_with_skews.yaml} 3
Failure Reason:

global name 'self' is not defined

fail 1364046 2017-07-05 19:05:10 2017-07-05 19:31:02 2017-07-05 19:45:02 0:14:00 0:10:33 0:03:27 smithi master rados/singleton/{all/osd-recovery.yaml msgr-failures/few.yaml msgr/random.yaml objectstore/bluestore-comp.yaml rados.yaml} 1
Failure Reason:

"2017-07-05 19:36:30.631641 mon.a mon.0 172.21.15.139:6789/0 50 : cluster [WRN] Health check failed: noin flag(s) set (OSDMAP_FLAGS)" in cluster log

fail 1364048 2017-07-05 19:05:12 2017-07-05 19:31:02 2017-07-05 19:43:02 0:12:00 0:07:07 0:04:53 smithi master rados/singleton-nomsgr/{all/full-tiering.yaml rados.yaml} 1
Failure Reason:

"2017-07-05 19:36:17.935039 mon.a mon.0 172.21.15.197:6789/0 86 : cluster [WRN] Health check failed: 1 cache pools are missing hit_sets (CACHE_POOL_NO_HIT_SET)" in cluster log

pass 1364050 2017-07-05 19:05:14 2017-07-05 19:31:02 2017-07-05 19:49:02 0:18:00 0:17:27 0:00:33 smithi master rados/basic/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-require-luminous/at-mkfs.yaml mon_kv_backend/rocksdb.yaml msgr-failures/many.yaml msgr/simple.yaml objectstore/filestore-xfs.yaml rados.yaml tasks/rados_workunit_loadgen_mix.yaml} 2
fail 1364052 2017-07-05 19:05:16 2017-07-05 19:31:03 2017-07-05 19:59:03 0:28:00 0:22:37 0:05:23 smithi master rados/thrash/{0-size-min-size-overrides/2-size-1-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml backoff/normal.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-require-luminous/at-mkfs.yaml msgr-failures/fastclose.yaml msgr/async.yaml objectstore/bluestore-comp.yaml rados.yaml rocksdb.yaml thrashers/morepggrow.yaml thrashosds-health.yaml workloads/cache-agent-big.yaml} 2
Failure Reason:

"2017-07-05 19:35:52.409077 mon.b mon.0 172.21.15.40:6789/0 117 : cluster [WRN] Health check failed: 1 osds down (OSD_DOWN)" in cluster log

fail 1364054 2017-07-05 19:05:18 2017-07-05 19:31:04 2017-07-05 19:43:04 0:12:00 0:07:08 0:04:52 smithi master rados/singleton/{all/peer.yaml msgr-failures/many.yaml msgr/simple.yaml objectstore/bluestore.yaml rados.yaml} 1
Failure Reason:

"2017-07-05 19:36:45.243335 mon.a mon.0 172.21.15.23:6789/0 65 : cluster [WRN] Health check failed: 1 osds down (OSD_DOWN)" in cluster log

pass 1364056 2017-07-05 19:05:19 2017-07-05 19:31:04 2017-07-05 23:41:09 4:10:05 4:08:16 0:01:49 smithi master rados/objectstore/filestore-idempotent.yaml 1
fail 1364058 2017-07-05 19:05:21 2017-07-05 19:31:14 2017-07-05 19:55:14 0:24:00 0:16:38 0:07:22 smithi master rados/thrash-luminous/{0-size-min-size-overrides/2-size-1-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml backoff/normal.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr/async.yaml objectstore/bluestore-comp.yaml rados.yaml rocksdb.yaml thrashers/pggrow.yaml thrashosds-health.yaml workloads/redirect_set_object.yaml} 2
Failure Reason:

"2017-07-05 19:39:18.292515 mon.b mon.0 172.21.15.61:6789/0 72 : cluster [ERR] Health check failed: full ratio(s) out of order (OSD_OUT_OF_ORDER_FULL)" in cluster log

fail 1364060 2017-07-05 19:05:23 2017-07-05 19:31:14 2017-07-05 19:43:14 0:12:00 0:06:41 0:05:19 smithi master rados/singleton/{all/pg-removal-interruption.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/filestore-xfs.yaml rados.yaml} 1
Failure Reason:

"2017-07-05 19:36:01.462828 mon.a mon.0 172.21.15.57:6789/0 88 : cluster [WRN] Health check failed: 1 osds down (OSD_DOWN)" in cluster log

dead 1364062 2017-07-05 19:05:26 2017-07-05 19:33:07 2017-07-06 07:39:44 12:06:37 smithi master rados/monthrash/{ceph.yaml clusters/9-mons.yaml d-require-luminous/at-mkfs.yaml mon_kv_backend/rocksdb.yaml msgr-failures/mon-delay.yaml msgr/async.yaml objectstore/bluestore-comp.yaml rados.yaml thrashers/sync-many.yaml workloads/snaps-few-objects.yaml} 2
fail 1364063 2017-07-05 19:05:29 2017-07-05 19:33:07 2017-07-05 19:55:07 0:22:00 0:13:00 0:09:00 smithi master rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml backoff/peering.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-require-luminous/at-end.yaml msgr-failures/few.yaml msgr/random.yaml objectstore/bluestore.yaml rados.yaml rocksdb.yaml thrashers/none.yaml thrashosds-health.yaml workloads/cache-agent-small.yaml} 2
Failure Reason:

"2017-07-05 19:43:38.318772 mon.b mon.0 172.21.15.38:6789/0 217 : cluster [WRN] Health check failed: 1 cache pools are missing hit_sets (CACHE_POOL_NO_HIT_SET)" in cluster log

fail 1364065 2017-07-05 19:05:32 2017-07-05 19:33:08 2017-07-05 19:59:08 0:26:00 0:21:19 0:04:41 smithi master rados/thrash-erasure-code/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-require-luminous/at-mkfs.yaml fast/normal.yaml leveldb.yaml msgr-failures/fastclose.yaml objectstore/bluestore-comp.yaml rados.yaml thrashers/morepggrow.yaml thrashosds-health.yaml workloads/ec-small-objects.yaml} 2
Failure Reason:

"2017-07-05 19:40:22.981226 mon.b mon.0 172.21.15.5:6789/0 116 : cluster [WRN] Health check failed: 1 osds down (OSD_DOWN)" in cluster log

fail 1364067 2017-07-05 19:05:34 2017-07-05 19:35:01 2017-07-05 19:53:01 0:18:00 0:13:03 0:04:57 smithi master rados/verify/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-require-luminous/at-end.yaml d-thrash/default/{default.yaml thrashosds-health.yaml} mon_kv_backend/leveldb.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/bluestore.yaml rados.yaml tasks/mon_recovery.yaml validater/lockdep.yaml} 2
Failure Reason:

"2017-07-05 19:43:11.557693 mon.a mon.0 172.21.15.3:6789/0 127 : cluster [WRN] Health check failed: noscrub flag(s) set (OSDMAP_FLAGS)" in cluster log

fail 1364069 2017-07-05 19:05:36 2017-07-05 19:35:01 2017-07-05 20:09:01 0:34:00 0:30:09 0:03:51 smithi master centos 7.3 rados/thrash-erasure-code-isa/{arch/x86_64.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-require-luminous/at-mkfs.yaml leveldb.yaml msgr-failures/fastclose.yaml objectstore/bluestore-comp.yaml rados.yaml supported/centos_latest.yaml thrashers/pggrow.yaml thrashosds-health.yaml workloads/ec-rados-plugin=isa-k=2-m=1.yaml} 2
Failure Reason:

"2017-07-05 19:50:12.445257 mon.a mon.0 172.21.15.2:6789/0 2051 : cluster [ERR] Health check failed: full ratio(s) out of order (OSD_OUT_OF_ORDER_FULL)" in cluster log

pass 1364071 2017-07-05 19:05:38 2017-07-05 19:35:01 2017-07-05 20:01:01 0:26:00 0:24:38 0:01:22 smithi master rados/basic/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-require-luminous/at-end.yaml mon_kv_backend/leveldb.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/bluestore-comp.yaml rados.yaml tasks/rados_workunit_loadgen_mostlyread.yaml} 2
fail 1364073 2017-07-05 19:05:40 2017-07-05 19:35:01 2017-07-05 19:55:01 0:20:00 0:11:26 0:08:34 smithi master rados/thrash-erasure-code-big/{ceph.yaml cluster/{12-osds.yaml openstack.yaml} d-require-luminous/at-mkfs.yaml leveldb.yaml msgr-failures/fastclose.yaml objectstore/bluestore-comp.yaml rados.yaml thrashers/pggrow.yaml thrashosds-health.yaml workloads/ec-rados-plugin=lrc-k=4-m=2-l=3.yaml} 4
Failure Reason:

"2017-07-05 19:46:02.590804 mon.a mon.0 172.21.15.70:6789/0 106 : cluster [WRN] Health check failed: noscrub flag(s) set (OSDMAP_FLAGS)" in cluster log

fail 1364075 2017-07-05 19:05:43 2017-07-05 19:35:01 2017-07-05 20:03:01 0:28:00 0:22:30 0:05:30 smithi master rados/thrash-erasure-code-overwrites/{bluestore.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-require-luminous/at-mkfs.yaml fast/normal.yaml leveldb.yaml msgr-failures/fastclose.yaml rados.yaml thrashers/pggrow.yaml thrashosds-health.yaml workloads/ec-small-objects-fast-read-overwrites.yaml} 2
Failure Reason:

"2017-07-05 19:41:30.184447 mon.b mon.0 172.21.15.155:6789/0 114 : cluster [WRN] Health check failed: noscrub flag(s) set (OSDMAP_FLAGS)" in cluster log

fail 1364078 2017-07-05 19:05:44 2017-07-05 19:35:02 2017-07-05 19:53:01 0:17:59 0:12:07 0:05:52 smithi master rados/singleton-bluestore/{all/cephtool.yaml msgr-failures/many.yaml msgr/async.yaml objectstore/bluestore-comp.yaml rados.yaml} 1
Failure Reason:

Command failed (workunit test cephtool/test.sh) on smithi062 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=wip-health TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/cephtool/test.sh'

fail 1364079 2017-07-05 19:05:46 2017-07-05 19:35:02 2017-07-05 19:47:01 0:11:59 0:06:54 0:05:05 smithi master rados/singleton-nomsgr/{all/health-warnings.yaml rados.yaml} 1
Failure Reason:

"2017-07-05 19:40:30.337183 mon.a mon.0 172.21.15.110:6789/0 245 : cluster [WRN] Health check failed: noup flag(s) set (OSDMAP_FLAGS)" in cluster log

fail 1364081 2017-07-05 19:05:47 2017-07-05 19:35:03 2017-07-05 20:05:02 0:29:59 0:25:05 0:04:54 smithi master rados/singleton/{all/radostool.yaml msgr-failures/many.yaml msgr/random.yaml objectstore/bluestore-comp.yaml rados.yaml} 1
Failure Reason:

"2017-07-05 19:54:42.926803 mon.a mon.0 172.21.15.46:6789/0 173 : cluster [WRN] Health check failed: 1 pool(s) full (POOL_FULL)" in cluster log

fail 1364083 2017-07-05 19:05:49 2017-07-05 19:35:03 2017-07-05 20:11:03 0:36:00 0:30:18 0:05:42 smithi master rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml backoff/peering_and_degraded.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-require-luminous/at-mkfs.yaml msgr-failures/osd-delay.yaml msgr/simple.yaml objectstore/filestore-xfs.yaml rados.yaml rocksdb.yaml thrashers/pggrow.yaml thrashosds-health.yaml workloads/cache-pool-snaps-readproxy.yaml} 2
Failure Reason:

"2017-07-05 19:51:35.792328 mon.a mon.0 172.21.15.113:6789/0 2507 : cluster [ERR] Health check failed: full ratio(s) out of order (OSD_OUT_OF_ORDER_FULL)" in cluster log

pass 1364085 2017-07-05 19:05:50 2017-07-05 19:35:53 2017-07-05 19:57:52 0:21:59 0:21:13 0:00:46 smithi master rados/singleton/{all/random-eio.yaml msgr-failures/few.yaml msgr/simple.yaml objectstore/bluestore.yaml rados.yaml} 2
fail 1364087 2017-07-05 19:05:51 2017-07-05 19:37:03 2017-07-05 19:55:03 0:18:00 0:10:56 0:07:04 smithi master rados/thrash-erasure-code-shec/{ceph.yaml clusters/{fixed-4.yaml openstack.yaml} d-require-luminous/at-mkfs.yaml leveldb.yaml msgr-failures/osd-delay.yaml objectstore/bluestore.yaml rados.yaml thrashers/default.yaml thrashosds-health.yaml workloads/ec-rados-plugin=shec-k=4-m=3-c=2.yaml} 4
Failure Reason:

"2017-07-05 19:47:09.944193 mon.c mon.0 172.21.15.51:6789/0 317 : cluster [WRN] Health check failed: noscrub flag(s) set (OSDMAP_FLAGS)" in cluster log

pass 1364090 2017-07-05 19:05:53 2017-07-05 19:37:03 2017-07-05 19:47:03 0:10:00 0:05:26 0:04:34 smithi master rados/objectstore/fusestore.yaml 1
fail 1364092 2017-07-05 19:05:54 2017-07-05 19:37:03 2017-07-05 19:49:02 0:11:59 0:07:18 0:04:41 smithi master rados/singleton/{all/rebuild-mondb.yaml msgr-failures/many.yaml msgr/async.yaml objectstore/filestore-xfs.yaml rados.yaml} 1
Failure Reason:

"2017-07-05 19:42:08.756760 mon.a mon.0 172.21.15.202:6789/0 13 : cluster [WRN] Health check failed: 1 osds down (OSD_DOWN)" in cluster log

pass 1364094 2017-07-05 19:05:55 2017-07-05 19:37:03 2017-07-05 19:51:03 0:14:00 0:11:46 0:02:14 smithi master rados/basic/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-require-luminous/at-mkfs.yaml mon_kv_backend/rocksdb.yaml msgr-failures/many.yaml msgr/random.yaml objectstore/bluestore.yaml rados.yaml tasks/readwrite.yaml} 2
dead 1364096 2017-07-05 19:05:56 2017-07-05 19:39:19 2017-07-06 07:46:02 12:06:43 smithi master rados/multimon/{clusters/21.yaml mon_kv_backend/leveldb.yaml msgr-failures/few.yaml msgr/simple.yaml objectstore/filestore-xfs.yaml rados.yaml tasks/mon_recovery.yaml} 3
fail 1364098 2017-07-05 19:05:57 2017-07-05 19:39:19 2017-07-05 20:19:18 0:39:59 0:34:57 0:05:02 smithi master rados/thrash/{0-size-min-size-overrides/2-size-1-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml backoff/normal.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-require-luminous/at-end.yaml msgr-failures/fastclose.yaml msgr/async.yaml objectstore/bluestore-comp.yaml rados.yaml rocksdb.yaml thrashers/default.yaml thrashosds-health.yaml workloads/cache-pool-snaps.yaml} 2
Failure Reason:

"2017-07-05 19:44:20.650993 mon.a mon.0 172.21.15.163:6789/0 120 : cluster [WRN] Health check failed: noscrub flag(s) set (OSDMAP_FLAGS)" in cluster log

pass 1364100 2017-07-05 19:05:58 2017-07-05 19:39:19 2017-07-05 20:03:18 0:23:59 0:22:30 0:01:29 smithi master rados/singleton-nomsgr/{all/msgr.yaml rados.yaml} 1
fail 1364102 2017-07-05 19:05:59 2017-07-05 19:39:19 2017-07-05 19:53:18 0:13:59 0:09:58 0:04:01 smithi master rados/singleton/{all/reg11184.yaml msgr-failures/few.yaml msgr/random.yaml objectstore/bluestore-comp.yaml rados.yaml} 1
fail 1364104 2017-07-05 19:06:00 2017-07-05 19:39:19 2017-07-05 19:59:18 0:19:59 0:15:47 0:04:12 smithi master rados/thrash-luminous/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml backoff/peering.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr/random.yaml objectstore/bluestore.yaml rados.yaml rocksdb.yaml thrashers/default.yaml thrashosds-health.yaml workloads/redirect.yaml} 2
Failure Reason:

"2017-07-05 19:44:33.160283 mon.b mon.0 172.21.15.94:6789/0 110 : cluster [WRN] Health check failed: noscrub flag(s) set (OSDMAP_FLAGS)" in cluster log

fail 1364106 2017-07-05 19:06:00 2017-07-05 19:39:19 2017-07-05 19:59:18 0:19:59 0:15:43 0:04:16 smithi master rados/monthrash/{ceph.yaml clusters/3-mons.yaml d-require-luminous/at-end.yaml mon_kv_backend/leveldb.yaml msgr-failures/few.yaml msgr/random.yaml objectstore/bluestore.yaml rados.yaml thrashers/sync-many.yaml workloads/pool-create-delete.yaml} 2
Failure Reason:

"2017-07-05 19:45:28.080875 mon.a mon.1 172.21.15.41:6789/0 30 : cluster [WRN] Health check failed: 1/3 mons down, quorum a,c (MON_DOWN)" in cluster log

fail 1364108 2017-07-05 19:06:01 2017-07-05 19:41:06 2017-07-05 20:17:06 0:36:00 0:31:13 0:04:47 smithi master rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml backoff/peering.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-require-luminous/at-mkfs.yaml msgr-failures/few.yaml msgr/random.yaml objectstore/bluestore.yaml rados.yaml rocksdb.yaml thrashers/mapgap.yaml thrashosds-health.yaml workloads/cache-snaps.yaml} 2
Failure Reason:

"2017-07-05 19:45:40.264718 mon.b mon.0 172.21.15.76:6789/0 74 : cluster [WRN] Health check failed: noscrub flag(s) set (OSDMAP_FLAGS)" in cluster log

fail 1364110 2017-07-05 19:06:02 2017-07-05 19:41:07 2017-07-05 19:55:06 0:13:59 0:09:01 0:04:58 smithi master rados/singleton/{all/resolve_stuck_peering.yaml msgr-failures/many.yaml msgr/simple.yaml objectstore/bluestore.yaml rados.yaml} 2
Failure Reason:

"2017-07-05 19:46:38.850921 mon.a mon.0 172.21.15.78:6789/0 122 : cluster [WRN] Health check failed: 2 osds down (OSD_DOWN)" in cluster log

fail 1364112 2017-07-05 19:06:03 2017-07-05 19:41:07 2017-07-05 20:23:07 0:42:00 0:36:52 0:05:08 smithi master rados/thrash-erasure-code/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-require-luminous/at-end.yaml fast/fast.yaml leveldb.yaml msgr-failures/few.yaml objectstore/bluestore.yaml rados.yaml thrashers/morepggrow.yaml thrashosds-health.yaml workloads/ec-rados-plugin=jerasure-k=2-m=1.yaml} 2
Failure Reason:

"2017-07-05 19:45:48.052934 mon.b mon.0 172.21.15.54:6789/0 87 : cluster [WRN] Health check failed: noscrub flag(s) set (OSDMAP_FLAGS)" in cluster log

fail 1364113 2017-07-05 19:06:04 2017-07-05 19:41:07 2017-07-05 20:19:07 0:38:00 0:32:32 0:05:28 smithi master ubuntu 14.04 rados/thrash-erasure-code-isa/{arch/x86_64.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-require-luminous/at-end.yaml leveldb.yaml msgr-failures/few.yaml objectstore/bluestore.yaml rados.yaml supported/ubuntu_14.04.yaml thrashers/default.yaml thrashosds-health.yaml workloads/ec-rados-plugin=isa-k=2-m=1.yaml} 2
Failure Reason:

"2017-07-05 19:46:42.466940 mon.a mon.0 172.21.15.47:6789/0 165 : cluster [WRN] Health check failed: noscrub flag(s) set (OSDMAP_FLAGS)" in cluster log

fail 1364115 2017-07-05 19:06:05 2017-07-05 19:41:13 2017-07-05 20:01:13 0:20:00 0:14:52 0:05:08 smithi master rados/basic/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-require-luminous/at-end.yaml mon_kv_backend/leveldb.yaml msgr-failures/few.yaml msgr/simple.yaml objectstore/filestore-xfs.yaml rados.yaml tasks/repair_test.yaml} 2
Failure Reason:

"2017-07-05 19:46:44.404789 mon.a mon.0 172.21.15.84:6789/0 156 : cluster [ERR] Health check failed: 1 scrub errors (OSD_SCRUB_ERRORS)" in cluster log

fail 1364117 2017-07-05 19:06:06 2017-07-05 19:41:15 2017-07-05 19:57:14 0:15:59 0:11:18 0:04:41 smithi master rados/thrash-erasure-code-big/{ceph.yaml cluster/{12-osds.yaml openstack.yaml} d-require-luminous/at-end.yaml leveldb.yaml msgr-failures/few.yaml objectstore/bluestore.yaml rados.yaml thrashers/default.yaml thrashosds-health.yaml workloads/ec-rados-plugin=lrc-k=4-m=2-l=3.yaml} 4
Failure Reason:

"2017-07-05 19:48:34.558374 mon.b mon.0 172.21.15.68:6789/0 289 : cluster [WRN] Health check failed: noscrub flag(s) set (OSDMAP_FLAGS)" in cluster log

fail 1364119 2017-07-05 19:06:06 2017-07-05 19:41:15 2017-07-05 20:07:15 0:26:00 0:21:32 0:04:28 smithi master rados/thrash-erasure-code-overwrites/{bluestore.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-require-luminous/at-end.yaml fast/fast.yaml leveldb.yaml msgr-failures/few.yaml rados.yaml thrashers/default.yaml thrashosds-health.yaml workloads/ec-small-objects-overwrites.yaml} 2
Failure Reason:

"2017-07-05 19:47:41.712504 mon.a mon.0 172.21.15.37:6789/0 102 : cluster [WRN] Health check failed: noscrub flag(s) set (OSDMAP_FLAGS)" in cluster log

fail 1364122 2017-07-05 19:06:07 2017-07-05 19:43:03 2017-07-05 19:57:02 0:13:59 0:08:36 0:05:23 smithi master rados/basic-luminous/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} objectstore/filestore-xfs.yaml rados.yaml scrub_test.yaml} 2
Failure Reason:

"2017-07-05 19:47:57.601464 mon.b mon.0 172.21.15.57:6789/0 61 : cluster [ERR] Health check failed: 2 scrub errors (OSD_SCRUB_ERRORS)" in cluster log

fail 1364123 2017-07-05 19:06:08 2017-07-05 19:43:03 2017-07-05 20:01:02 0:17:59 0:10:11 0:07:48 smithi master rados/mgr/{clusters/2-node-mgr.yaml debug/mgr.yaml objectstore/filestore-xfs.yaml tasks/failover.yaml} 2
Failure Reason:

"2017-07-05 19:52:45.778794 mon.b mon.0 172.21.15.4:6789/0 81 : cluster [WRN] Health check failed: no active mgr (MGR_DOWN)" in cluster log

pass 1364125 2017-07-05 19:06:09 2017-07-05 19:43:03 2017-07-05 19:49:02 0:05:59 0:05:47 0:00:12 smithi master rados/objectstore/keyvaluedb.yaml 1
fail 1364127 2017-07-05 19:06:09 2017-07-05 19:43:03 2017-07-05 19:55:02 0:11:59 0:07:01 0:04:58 smithi master rados/singleton/{all/rest-api.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/filestore-xfs.yaml rados.yaml} 1
Failure Reason:

"2017-07-05 19:47:39.797811 mon.a mon.0 172.21.15.160:6789/0 174 : cluster [WRN] Health check failed: noup flag(s) set (OSDMAP_FLAGS)" in cluster log

fail 1364129 2017-07-05 19:06:10 2017-07-05 19:43:03 2017-07-05 20:03:03 0:20:00 0:15:55 0:04:05 smithi master rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml backoff/peering_and_degraded.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-require-luminous/at-end.yaml msgr-failures/osd-delay.yaml msgr/simple.yaml objectstore/filestore-xfs.yaml rados.yaml rocksdb.yaml thrashers/morepggrow.yaml thrashosds-health.yaml workloads/cache.yaml} 2
Failure Reason:

"2017-07-05 19:49:13.957019 mon.b mon.0 172.21.15.35:6789/0 136 : cluster [WRN] Health check failed: noscrub flag(s) set (OSDMAP_FLAGS)" in cluster log

fail 1364131 2017-07-05 19:06:11 2017-07-05 19:43:03 2017-07-05 20:01:02 0:17:59 0:13:44 0:04:15 smithi master rados/singleton-nomsgr/{all/multi-backfill-reject.yaml rados.yaml} 2
Failure Reason:

"2017-07-05 19:50:05.462880 mon.a mon.0 172.21.15.31:6789/0 165 : cluster [WRN] Health check failed: 9185/540577 objects misplaced (1.699%) (OBJECT_MISPLACED)" in cluster log

pass 1364134 2017-07-05 19:06:12 2017-07-05 19:43:03 2017-07-05 20:05:03 0:22:00 0:16:46 0:05:14 smithi master rados/singleton/{all/test_envlibrados_for_rocksdb.yaml msgr-failures/many.yaml msgr/random.yaml objectstore/bluestore-comp.yaml rados.yaml} 1
pass 1364136 2017-07-05 19:06:13 2017-07-05 19:43:04 2017-07-05 20:11:04 0:28:00 0:25:33 0:02:27 smithi master rados/thrash/{0-size-min-size-overrides/2-size-1-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml backoff/normal.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-require-luminous/at-mkfs.yaml msgr-failures/fastclose.yaml msgr/async.yaml objectstore/bluestore-comp.yaml rados.yaml rocksdb.yaml thrashers/none.yaml thrashosds-health.yaml workloads/pool-snaps-few-objects.yaml} 2
fail 1364137 2017-07-05 19:06:13 2017-07-05 19:43:05 2017-07-05 20:09:05 0:26:00 0:21:03 0:04:57 smithi master centos rados/verify/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-require-luminous/at-mkfs.yaml d-thrash/none.yaml mon_kv_backend/rocksdb.yaml msgr-failures/few.yaml msgr/random.yaml objectstore/filestore-xfs.yaml rados.yaml tasks/rados_api_tests.yaml validater/valgrind.yaml} 2
Failure Reason:

saw valgrind issues

pass 1364139 2017-07-05 19:06:14 2017-07-05 19:43:06 2017-07-05 20:03:05 0:19:59 0:19:32 0:00:27 smithi master rados/basic/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-require-luminous/at-mkfs.yaml mon_kv_backend/rocksdb.yaml msgr-failures/many.yaml msgr/async.yaml objectstore/bluestore-comp.yaml rados.yaml tasks/rgw_snaps.yaml} 2
fail 1364141 2017-07-05 19:06:15 2017-07-05 19:43:06 2017-07-05 20:19:06 0:36:00 0:31:37 0:04:23 smithi master rados/singleton/{all/thrash-eio.yaml msgr-failures/few.yaml msgr/simple.yaml objectstore/bluestore.yaml rados.yaml} 2
Failure Reason:

"2017-07-05 19:58:53.704248 mon.a mon.0 172.21.15.190:6789/0 2762 : cluster [ERR] Health check failed: full ratio(s) out of order (OSD_OUT_OF_ORDER_FULL)" in cluster log

fail 1364143 2017-07-05 19:06:15 2017-07-05 19:43:15 2017-07-05 19:59:14 0:15:59 0:10:59 0:05:00 smithi master rados/singleton-bluestore/{all/cephtool.yaml msgr-failures/few.yaml msgr/random.yaml objectstore/bluestore.yaml rados.yaml} 1
Failure Reason:

Command failed (workunit test cephtool/test.sh) on smithi056 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=wip-health TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/cephtool/test.sh'

fail 1364145 2017-07-05 19:06:16 2017-07-05 19:43:15 2017-07-05 20:05:14 0:21:59 0:12:07 0:09:52 smithi master rados/thrash-erasure-code-shec/{ceph.yaml clusters/{fixed-4.yaml openstack.yaml} d-require-luminous/at-end.yaml leveldb.yaml msgr-failures/fastclose.yaml objectstore/filestore-xfs.yaml rados.yaml thrashers/default.yaml thrashosds-health.yaml workloads/ec-rados-plugin=shec-k=4-m=3-c=2.yaml} 4
Failure Reason:

"2017-07-05 19:56:12.572271 mon.a mon.0 172.21.15.11:6789/0 333 : cluster [WRN] Health check failed: noscrub flag(s) set (OSDMAP_FLAGS)" in cluster log

pass 1364147 2017-07-05 19:06:17 2017-07-05 19:43:15 2017-07-05 20:03:14 0:19:59 0:15:33 0:04:26 smithi master rados/objectstore/objectcacher-stress.yaml 1
fail 1364149 2017-07-05 19:06:18 2017-07-05 19:45:10 2017-07-05 19:55:09 0:09:59 0:06:02 0:03:57 smithi master rados/multimon/{clusters/3.yaml mon_kv_backend/rocksdb.yaml msgr-failures/many.yaml msgr/async.yaml objectstore/bluestore-comp.yaml rados.yaml tasks/mon_clock_no_skews.yaml} 2
Failure Reason:

global name 'self' is not defined

fail 1364151 2017-07-05 19:06:19 2017-07-05 19:45:10 2017-07-05 20:09:10 0:24:00 0:19:13 0:04:47 smithi master rados/singleton/{all/thrash-rados/{thrash-rados.yaml thrashosds-health.yaml} msgr-failures/many.yaml msgr/async.yaml objectstore/filestore-xfs.yaml rados.yaml} 2
Failure Reason:

"2017-07-05 19:51:12.486542 mon.a mon.0 172.21.15.114:6789/0 138 : cluster [WRN] Health check failed: noscrub flag(s) set (OSDMAP_FLAGS)" in cluster log

fail 1364153 2017-07-05 19:06:20 2017-07-05 19:45:10 2017-07-05 20:03:09 0:17:59 0:13:06 0:04:53 smithi master rados/thrash-luminous/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml backoff/peering_and_degraded.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr/simple.yaml objectstore/filestore-xfs.yaml rados.yaml rocksdb.yaml thrashers/mapgap.yaml thrashosds-health.yaml workloads/redirect_set_object.yaml} 2
Failure Reason:

"2017-07-05 19:51:18.446086 mon.a mon.0 172.21.15.29:6789/0 72 : cluster [WRN] Health check failed: noscrub flag(s) set (OSDMAP_FLAGS)" in cluster log

fail 1364155 2017-07-05 19:06:21 2017-07-05 19:45:10 2017-07-05 20:13:10 0:28:00 0:22:52 0:05:08 smithi master rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml backoff/peering.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-require-luminous/at-end.yaml msgr-failures/few.yaml msgr/random.yaml objectstore/bluestore.yaml rados.yaml rocksdb.yaml thrashers/pggrow.yaml thrashosds-health.yaml workloads/rados_api_tests.yaml} 2
Failure Reason:

"2017-07-05 19:52:29.894416 mon.b mon.0 172.21.15.14:6789/0 191 : cluster [WRN] Health check failed: noscrub flag(s) set (OSDMAP_FLAGS)" in cluster log

fail 1364157 2017-07-05 19:06:21 2017-07-05 19:45:10 2017-07-05 19:57:09 0:11:59 0:08:33 0:03:26 smithi master rados/monthrash/{ceph.yaml clusters/9-mons.yaml d-require-luminous/at-mkfs.yaml mon_kv_backend/rocksdb.yaml msgr-failures/mon-delay.yaml msgr/simple.yaml objectstore/filestore-xfs.yaml rados.yaml thrashers/sync.yaml workloads/rados_5925.yaml} 2
Failure Reason:

"2017-07-05 19:50:21.361325 mon.f mon.1 172.21.15.118:6789/0 48 : cluster [WRN] Health check failed: 4/9 mons down, quorum f,g,d,i,e (MON_DOWN)" in cluster log

pass 1364159 2017-07-05 19:06:22 2017-07-05 19:45:10 2017-07-05 19:51:09 0:05:59 0:05:39 0:00:20 smithi master rados/singleton-nomsgr/{all/pool-access.yaml rados.yaml} 1
fail 1364162 2017-07-05 19:06:23 2017-07-05 19:45:15 2017-07-05 20:17:15 0:32:00 0:23:40 0:08:20 smithi master rados/singleton/{all/thrash_cache_writeback_proxy_none.yaml msgr-failures/few.yaml msgr/random.yaml objectstore/bluestore-comp.yaml rados.yaml} 2
Failure Reason:

"2017-07-05 19:54:54.238703 mon.a mon.0 172.21.15.110:6789/0 145 : cluster [WRN] Health check failed: 1 cache pools are missing hit_sets (CACHE_POOL_NO_HIT_SET)" in cluster log

fail 1364164 2017-07-05 19:06:24 2017-07-05 19:45:15 2017-07-05 20:25:15 0:40:00 0:35:20 0:04:40 smithi master rados/thrash-erasure-code/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-require-luminous/at-mkfs.yaml fast/normal.yaml leveldb.yaml msgr-failures/osd-delay.yaml objectstore/filestore-xfs.yaml rados.yaml thrashers/pggrow.yaml thrashosds-health.yaml workloads/ec-rados-plugin=jerasure-k=3-m=1.yaml} 2
Failure Reason:

"2017-07-05 19:50:04.090864 mon.a mon.0 172.21.15.188:6789/0 73 : cluster [WRN] Health check failed: noscrub flag(s) set (OSDMAP_FLAGS)" in cluster log

fail 1364166 2017-07-05 19:06:25 2017-07-05 19:47:01 2017-07-05 20:17:01 0:30:00 0:25:28 0:04:32 smithi master ubuntu 16.04 rados/thrash-erasure-code-isa/{arch/x86_64.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-require-luminous/at-mkfs.yaml leveldb.yaml msgr-failures/osd-delay.yaml objectstore/filestore-xfs.yaml rados.yaml supported/ubuntu_latest.yaml thrashers/mapgap.yaml thrashosds-health.yaml workloads/ec-rados-plugin=isa-k=2-m=1.yaml} 2
Failure Reason:

"2017-07-05 19:52:33.366330 mon.a mon.0 172.21.15.117:6789/0 100 : cluster [WRN] Health check failed: noscrub flag(s) set (OSDMAP_FLAGS)" in cluster log

fail 1364168 2017-07-05 19:06:25 2017-07-05 19:47:02 2017-07-05 20:13:02 0:26:00 0:21:21 0:04:39 smithi master rados/basic/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-require-luminous/at-end.yaml mon_kv_backend/leveldb.yaml msgr-failures/few.yaml msgr/random.yaml objectstore/bluestore.yaml rados.yaml tasks/rados_api_tests.yaml} 2
Failure Reason:

"2017-07-05 19:53:17.054106 mon.a mon.0 172.21.15.24:6789/0 190 : cluster [WRN] Health check failed: 1 cache pools are missing hit_sets (CACHE_POOL_NO_HIT_SET)" in cluster log

fail 1364170 2017-07-05 19:06:26 2017-07-05 19:47:02 2017-07-05 20:07:02 0:20:00 0:10:34 0:09:26 smithi master rados/thrash-erasure-code-big/{ceph.yaml cluster/{12-osds.yaml openstack.yaml} d-require-luminous/at-mkfs.yaml leveldb.yaml msgr-failures/osd-delay.yaml objectstore/filestore-xfs.yaml rados.yaml thrashers/fastread.yaml thrashosds-health.yaml workloads/ec-rados-plugin=lrc-k=4-m=2-l=3.yaml} 4
Failure Reason:

"2017-07-05 19:58:08.344707 mon.a mon.0 172.21.15.45:6789/0 105 : cluster [WRN] Health check failed: noscrub flag(s) set (OSDMAP_FLAGS)" in cluster log

fail 1364172 2017-07-05 19:06:27 2017-07-05 19:47:03 2017-07-05 20:31:03 0:44:00 0:36:57 0:07:03 smithi master rados/thrash-erasure-code-overwrites/{bluestore.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-require-luminous/at-mkfs.yaml fast/normal.yaml leveldb.yaml msgr-failures/osd-delay.yaml rados.yaml thrashers/fastread.yaml thrashosds-health.yaml workloads/ec-snaps-few-objects-overwrites.yaml} 2
Failure Reason:

"2017-07-05 19:56:31.341616 mon.a mon.0 172.21.15.136:6789/0 74 : cluster [WRN] Health check failed: noscrub flag(s) set (OSDMAP_FLAGS)" in cluster log

fail 1364174 2017-07-05 19:06:28 2017-07-05 19:47:04 2017-07-05 20:15:03 0:27:59 0:23:12 0:04:47 smithi master rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml backoff/peering_and_degraded.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-require-luminous/at-mkfs.yaml msgr-failures/osd-delay.yaml msgr/simple.yaml objectstore/filestore-xfs.yaml rados.yaml rocksdb.yaml thrashers/default.yaml thrashosds-health.yaml workloads/radosbench.yaml} 2
Failure Reason:

"2017-07-05 19:52:54.217482 mon.a mon.0 172.21.15.65:6789/0 69 : cluster [WRN] Health check failed: noscrub flag(s) set (OSDMAP_FLAGS)" in cluster log

fail 1364176 2017-07-05 19:06:28 2017-07-05 19:47:05 2017-07-05 19:59:04 0:11:59 0:05:52 0:06:07 smithi master rados/singleton/{all/watch-notify-same-primary.yaml msgr-failures/many.yaml msgr/simple.yaml objectstore/bluestore.yaml rados.yaml} 1
Failure Reason:

"2017-07-05 19:51:48.989540 mon.a mon.0 172.21.15.170:6789/0 46 : cluster [WRN] Health check failed: noout flag(s) set (OSDMAP_FLAGS)" in cluster log

pass 1364178 2017-07-05 19:06:29 2017-07-05 19:47:06 2017-07-05 19:57:05 0:09:59 0:06:41 0:03:18 smithi master rados/singleton/{all/admin-socket.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/bluestore.yaml rados.yaml} 1
pass 1364180 2017-07-05 19:06:30 2017-07-05 19:47:14 2017-07-06 03:37:24 7:50:10 7:48:53 0:01:17 smithi master rados/objectstore/objectstore.yaml 1
fail 1364182 2017-07-05 19:06:31 2017-07-05 19:49:02 2017-07-05 20:19:02 0:30:00 0:22:19 0:07:41 smithi master rados/thrash/{0-size-min-size-overrides/2-size-1-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml backoff/normal.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-require-luminous/at-end.yaml msgr-failures/fastclose.yaml msgr/async.yaml objectstore/bluestore-comp.yaml rados.yaml rocksdb.yaml thrashers/mapgap.yaml thrashosds-health.yaml workloads/small-objects.yaml} 2
Failure Reason:

"2017-07-05 19:58:34.046054 mon.a mon.0 172.21.15.26:6789/0 98 : cluster [WRN] Health check failed: noscrub flag(s) set (OSDMAP_FLAGS)" in cluster log

fail 1364184 2017-07-05 19:06:32 2017-07-05 19:49:02 2017-07-05 20:03:01 0:13:59 0:10:52 0:03:07 smithi master centos rados/singleton-nomsgr/{all/valgrind-leaks.yaml rados.yaml} 1
Failure Reason:

Command failed on smithi134 with status 1: 'find /home/ubuntu/cephtest -ls ; rmdir -- /home/ubuntu/cephtest'

pass 1364186 2017-07-05 19:06:32 2017-07-05 19:49:03 2017-07-05 19:59:02 0:09:59 0:09:18 0:00:41 smithi master rados/basic/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-require-luminous/at-mkfs.yaml mon_kv_backend/rocksdb.yaml msgr-failures/many.yaml msgr/simple.yaml objectstore/filestore-xfs.yaml rados.yaml tasks/rados_cls_all.yaml} 2
fail 1364188 2017-07-05 19:06:33 2017-07-05 19:49:03 2017-07-05 20:03:02 0:13:59 0:09:54 0:04:05 smithi master rados/singleton/{all/divergent_priors.yaml msgr-failures/many.yaml msgr/random.yaml objectstore/filestore-xfs.yaml rados.yaml} 1
Failure Reason:

"2017-07-05 19:54:45.709959 mon.a mon.0 172.21.15.141:6789/0 73 : cluster [WRN] Health check failed: noout flag(s) set (OSDMAP_FLAGS)" in cluster log

fail 1364190 2017-07-05 19:06:34 2017-07-05 19:49:03 2017-07-05 20:09:03 0:20:00 0:15:12 0:04:48 smithi master rados/thrash-luminous/{0-size-min-size-overrides/2-size-1-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml backoff/normal.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr/async.yaml objectstore/bluestore-comp.yaml rados.yaml rocksdb.yaml thrashers/morepggrow.yaml thrashosds-health.yaml workloads/redirect.yaml} 2
Failure Reason:

"2017-07-05 19:58:14.992973 mon.a mon.0 172.21.15.64:6789/0 1106 : cluster [ERR] Health check failed: full ratio(s) out of order (OSD_OUT_OF_ORDER_FULL)" in cluster log

fail 1364192 2017-07-05 19:06:35 2017-07-05 19:49:03 2017-07-05 20:27:03 0:38:00 0:31:08 0:06:52 smithi master rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml backoff/peering.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-require-luminous/at-mkfs.yaml msgr-failures/few.yaml msgr/random.yaml objectstore/bluestore.yaml rados.yaml rocksdb.yaml thrashers/morepggrow.yaml thrashosds-health.yaml workloads/snaps-few-objects.yaml} 2
Failure Reason:

"2017-07-05 19:56:17.634705 mon.a mon.0 172.21.15.48:6789/0 114 : cluster [WRN] Health check failed: noscrub flag(s) set (OSDMAP_FLAGS)" in cluster log