Status Job ID Links Posted Started Updated
Runtime
Duration
In Waiting
Machine Teuthology Branch OS Type OS Version Description Nodes
fail 2176315 2018-02-09 19:23:58 2018-02-09 19:26:04 2018-02-09 22:48:07 3:22:03 3:05:51 0:16:12 smithi master rados/standalone/crush.yaml 1
Failure Reason:

Command failed (workunit test crush/crush-choose-args.sh) on smithi001 with status 124: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=wip-sage3-testing-2018-02-08-1259 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/standalone/crush/crush-choose-args.sh'

fail 2176316 2018-02-09 19:23:59 2018-02-09 19:26:06 2018-02-09 19:52:06 0:26:00 0:07:55 0:18:05 smithi master rados/upgrade/luminous-x-singleton/{0-cluster/{openstack.yaml start.yaml} 1-install/luminous.yaml 2-partial-upgrade/firsthalf.yaml 3-thrash/default.yaml 4-workload/{rbd-cls.yaml rbd-import-export.yaml readwrite.yaml snaps-few-objects.yaml} 5-workload/{radosbench.yaml rbd_api.yaml} 6-finish-upgrade.yaml 7-mimic.yaml 8-workload/{rbd-python.yaml rgw-swift.yaml snaps-many-objects.yaml} thrashosds-health.yaml} 3
Failure Reason:

Command failed on smithi071 with status 1: "sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph -- tell 'mon.*' injectargs --mon_health_to_clog=false"

fail 2176317 2018-02-09 19:23:59 2018-02-09 19:26:06 2018-02-09 22:42:10 3:16:04 3:05:34 0:10:30 smithi master rados/standalone/erasure-code.yaml 1
Failure Reason:

Command failed (workunit test erasure-code/test-erasure-code-plugins.sh) on smithi060 with status 124: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=wip-sage3-testing-2018-02-08-1259 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/standalone/erasure-code/test-erasure-code-plugins.sh'

fail 2176318 2018-02-09 19:24:00 2018-02-09 19:26:06 2018-02-09 22:42:10 3:16:04 3:05:29 0:10:35 smithi master rados/standalone/misc.yaml 1
Failure Reason:

Command failed (workunit test misc/rados-striper.sh) on smithi146 with status 124: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=wip-sage3-testing-2018-02-08-1259 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/standalone/misc/rados-striper.sh'

fail 2176319 2018-02-09 19:24:01 2018-02-09 19:26:14 2018-02-09 19:44:14 0:18:00 0:07:35 0:10:25 smithi master rados/standalone/mon.yaml 1
Failure Reason:

Command failed (workunit test mon/misc.sh) on smithi024 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=wip-sage3-testing-2018-02-08-1259 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/standalone/mon/misc.sh'

fail 2176320 2018-02-09 19:24:01 2018-02-09 19:26:26 2018-02-09 20:04:26 0:38:00 0:27:02 0:10:58 smithi master rados/monthrash/{ceph.yaml clusters/9-mons.yaml mon_kv_backend/leveldb.yaml msgr-failures/mon-delay.yaml msgr/random.yaml objectstore/bluestore.yaml rados.yaml thrashers/many.yaml workloads/rados_api_tests.yaml} 2
Failure Reason:

Command failed (workunit test rados/test.sh) on smithi045 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=wip-sage3-testing-2018-02-08-1259 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/rados/test.sh'

fail 2176321 2018-02-09 19:24:02 2018-02-09 19:27:55 2018-02-09 22:43:59 3:16:04 3:05:52 0:10:12 smithi master rados/standalone/osd.yaml 1
Failure Reason:

Command failed (workunit test osd/osd-backfill-stats.sh) on smithi029 with status 124: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=wip-sage3-testing-2018-02-08-1259 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/standalone/osd/osd-backfill-stats.sh'

fail 2176322 2018-02-09 19:24:03 2018-02-09 19:27:56 2018-02-09 22:41:59 3:14:03 3:05:30 0:08:33 smithi master rados/standalone/scrub.yaml 1
Failure Reason:

Command failed (workunit test scrub/osd-recovery-scrub.sh) on smithi157 with status 124: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=wip-sage3-testing-2018-02-08-1259 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/standalone/scrub/osd-recovery-scrub.sh'

pass 2176323 2018-02-09 19:24:04 2018-02-09 19:28:09 2018-02-09 19:56:09 0:28:00 0:15:10 0:12:50 smithi master rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml backoff/normal.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/crush-compat.yaml msgr-failures/few.yaml msgr/simple.yaml objectstore/bluestore-bitmap.yaml rados.yaml rocksdb.yaml thrashers/mapgap.yaml thrashosds-health.yaml workloads/set-chunks-read.yaml} 2
fail 2176324 2018-02-09 19:24:04 2018-02-09 19:28:37 2018-02-09 20:10:37 0:42:00 0:28:29 0:13:31 smithi master centos rados/singleton-nomsgr/{all/valgrind-leaks.yaml rados.yaml} 1
Failure Reason:

"2018-02-09 19:54:51.850547 mon.a mon.0 172.21.15.198:6789/0 49 : cluster [WRN] Manager daemon x is unresponsive. No standby daemons available." in cluster log

fail 2176325 2018-02-09 19:24:05 2018-02-09 19:29:34 2018-02-09 20:11:34 0:42:00 0:31:37 0:10:23 smithi master centos rados/verify/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-thrash/none.yaml mon_kv_backend/rocksdb.yaml msgr-failures/few.yaml msgr/random.yaml objectstore/bluestore-bitmap.yaml rados.yaml tasks/rados_cls_all.yaml validater/valgrind.yaml} 2
Failure Reason:

"2018-02-09 19:56:57.120922 mon.b mon.0 172.21.15.50:6789/0 106 : cluster [WRN] Manager daemon y is unresponsive. No standby daemons available." in cluster log

fail 2176326 2018-02-09 19:24:06 2018-02-09 19:29:56 2018-02-09 20:11:56 0:42:00 0:31:30 0:10:30 smithi master centos rados/verify/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-thrash/none.yaml mon_kv_backend/rocksdb.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/bluestore.yaml rados.yaml tasks/rados_api_tests.yaml validater/valgrind.yaml} 2
Failure Reason:

"2018-02-09 19:55:45.924196 mon.b mon.0 172.21.15.105:6789/0 109 : cluster [WRN] Manager daemon x is unresponsive. No standby daemons available." in cluster log

fail 2176327 2018-02-09 19:24:07 2018-02-09 19:30:12 2018-02-09 19:50:12 0:20:00 0:08:46 0:11:14 smithi master rados/singleton/{all/rebuild-mondb.yaml msgr-failures/many.yaml msgr/simple.yaml objectstore/filestore-xfs.yaml rados.yaml} 1
Failure Reason:

Command failed on smithi098 with status 234: 'sudo -u ceph CEPH_ARGS=--no-monstore-tool ceph-monstore-tool /var/lib/ceph/mon/ceph-a rebuild -- --keyring /etc/ceph/ceph.keyring'