Status Job ID Links Posted Started Updated
Runtime
Duration
In Waiting
Machine Teuthology Branch OS Type OS Version Description Nodes
dead 1447945 2017-07-26 18:20:32 2017-07-26 18:30:39 2017-07-27 01:58:49 7:28:10 mira master rados/thrash/{0-size-min-size-overrides/2-size-1-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml backoff/peering.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-require-luminous/at-end.yaml msgr-failures/fastclose.yaml msgr/simple.yaml objectstore/bluestore.yaml rados.yaml rocksdb.yaml thrashers/none.yaml thrashosds-health.yaml workloads/cache-agent-big.yaml} 2
pass 1447946 2017-07-26 18:20:33 2017-07-26 18:38:50 2017-07-26 19:16:50 0:38:00 0:36:12 0:01:48 mira master rados/singleton/{all/osd-recovery-incomplete.yaml msgr-failures/many.yaml msgr/async.yaml objectstore/bluestore-comp.yaml rados.yaml} 1
fail 1447947 2017-07-26 18:20:34 2017-07-26 18:42:12 2017-07-26 19:06:12 0:24:00 0:04:49 0:19:11 mira master rados/multimon/{clusters/9.yaml mon_kv_backend/leveldb.yaml msgr-failures/many.yaml msgr/async.yaml objectstore/bluestore-comp.yaml rados.yaml tasks/mon_recovery.yaml} 3
Failure Reason:

HTTPSConnectionPool(host='shaman.ceph.com', port=443): Max retries exceeded with url: /api/search?status=ready&project=ceph&flavor=default&distros=ubuntu%2F14.04%2Fx86_64&sha1=a75af1371197696160f417902be9f584f2b77b90 (Caused by NewConnectionError('<requests.packages.urllib3.connection.VerifiedHTTPSConnection object at 0x7f266a5a83d0>: Failed to establish a new connection: [Errno 113] No route to host',))

pass 1447948 2017-07-26 18:20:34 2017-07-26 18:57:55 2017-07-26 19:27:55 0:30:00 0:21:48 0:08:12 mira master rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml backoff/peering_and_degraded.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-require-luminous/at-mkfs.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/filestore-xfs.yaml rados.yaml rocksdb.yaml thrashers/pggrow.yaml thrashosds-health.yaml workloads/cache-agent-small.yaml} 2
pass 1447949 2017-07-26 18:20:35 2017-07-26 19:00:07 2017-07-26 20:16:08 1:16:01 0:48:17 0:27:44 mira master rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml backoff/normal.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-require-luminous/at-end.yaml msgr-failures/osd-delay.yaml msgr/random.yaml objectstore/bluestore-comp.yaml rados.yaml rocksdb.yaml thrashers/default.yaml thrashosds-health.yaml workloads/cache-pool-snaps-readproxy.yaml} 2
pass 1447950 2017-07-26 18:20:36 2017-07-26 19:03:30 2017-07-26 20:11:30 1:08:00 0:32:30 0:35:30 mira master centos 7.3 rados/thrash-erasure-code-isa/{arch/x86_64.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-require-luminous/at-mkfs.yaml leveldb.yaml msgr-failures/few.yaml objectstore/filestore-xfs.yaml rados.yaml supported/centos_latest.yaml thrashers/default.yaml thrashosds-health.yaml workloads/ec-rados-plugin=isa-k=2-m=1.yaml} 2
pass 1447951 2017-07-26 18:20:36 2017-07-26 19:06:22 2017-07-26 19:38:21 0:31:59 0:30:57 0:01:02 mira master rados/standalone/scrub.yaml 1
pass 1447952 2017-07-26 18:20:37 2017-07-26 19:17:09 2017-07-26 20:21:09 1:04:00 0:33:19 0:30:41 mira master rados/thrash-erasure-code-overwrites/{bluestore.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-require-luminous/at-mkfs.yaml fast/fast.yaml leveldb.yaml msgr-failures/osd-delay.yaml rados.yaml thrashers/pggrow.yaml thrashosds-health.yaml workloads/ec-snaps-few-objects-overwrites.yaml} 2
fail 1447953 2017-07-26 18:20:38 2017-07-26 19:22:04 2017-07-26 20:26:04 1:04:00 0:28:54 0:35:06 mira master rados/thrash/{0-size-min-size-overrides/2-size-1-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml backoff/peering.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-require-luminous/at-mkfs.yaml msgr-failures/fastclose.yaml msgr/simple.yaml objectstore/bluestore.yaml rados.yaml rocksdb.yaml thrashers/mapgap.yaml thrashosds-health.yaml workloads/cache-pool-snaps.yaml} 2
Failure Reason:

ceph-objectstore-tool: exp list-pgs failure with status 1