ID
Status
Ceph Branch
Suite Branch
Teuthology Branch
Machine
OS
Nodes
Description
Failure Reason
luminous
wip-23208A
master
ovh
ubuntu 16.04
smoke/1node/{clusters/{fixed-1.yaml openstack.yaml} distros/ubuntu_latest.yaml objectstore/filestore-xfs.yaml tasks/ceph-deploy.yaml}
failed during ceph-deploy cmd: disk zap ovh077:/dev/sdb , ec=1
luminous
wip-23208A
master
ovh
ubuntu 16.04
smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} distros/{ubuntu_latest.yaml} objectstore/bluestore.yaml tasks/cfuse_workunit_suites_blogbench.yaml}
luminous
wip-23208A
master
ovh
centos 7.4
smoke/systemd/{clusters/{fixed-4.yaml openstack.yaml} distros/{centos_latest.yaml} objectstore/filestore-xfs.yaml tasks/systemd.yaml}
luminous
wip-23208A
master
ovh
ubuntu 16.04
smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} distros/{ubuntu_latest.yaml} objectstore/bluestore.yaml tasks/cfuse_workunit_suites_fsstress.yaml}
luminous
wip-23208A
master
ovh
ubuntu 16.04
smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} distros/{ubuntu_latest.yaml} objectstore/bluestore.yaml tasks/cfuse_workunit_suites_iozone.yaml}
luminous
wip-23208A
master
ovh
ubuntu 16.04
smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} distros/{ubuntu_latest.yaml} objectstore/bluestore.yaml tasks/cfuse_workunit_suites_pjd.yaml}
luminous
wip-23208A
master
ovh
ubuntu 16.04
smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} distros/{ubuntu_latest.yaml} objectstore/bluestore.yaml tasks/kclient_workunit_direct_io.yaml}
Command failed on ovh044 with status 22: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage /sbin/mount.ceph 158.69.68.216:6789,158.69.68.48:6790,158.69.68.48:6789:/ /home/ubuntu/cephtest/mnt.0 -v -o name=0,secretfile=/home/ubuntu/cephtest/ceph.data/client.0.secret,norequire_active_mds'
luminous
wip-23208A
master
ovh
centos 7.4
smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} distros/{centos_latest.yaml} objectstore/bluestore.yaml tasks/kclient_workunit_suites_dbench.yaml}
while scanning a plain scalar in "/tmp/teuth_ansible_failures_GWaHEA", line 1, column 2431 found unexpected ':' in "/tmp/teuth_ansible_failures_GWaHEA", line 1, column 2434 Please check http://pyyaml.org/wiki/YAMLColonInFlowContext for details.
luminous
wip-23208A
master
ovh
centos 7.4
smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} distros/{centos_latest.yaml} objectstore/bluestore.yaml tasks/kclient_workunit_suites_fsstress.yaml}
Command failed on ovh037 with status 22: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage /sbin/mount.ceph 158.69.68.20:6789,158.69.68.181:6790,158.69.68.181:6789:/ /home/ubuntu/cephtest/mnt.0 -v -o name=0,secretfile=/home/ubuntu/cephtest/ceph.data/client.0.secret,norequire_active_mds'
luminous
wip-23208A
master
ovh
ubuntu 16.04
smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} distros/{ubuntu_latest.yaml} objectstore/bluestore.yaml tasks/kclient_workunit_suites_pjd.yaml}
Command failed on ovh076 with status 22: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage /sbin/mount.ceph 158.69.69.198:6789,158.69.69.148:6790,158.69.69.148:6789:/ /home/ubuntu/cephtest/mnt.0 -v -o name=0,secretfile=/home/ubuntu/cephtest/ceph.data/client.0.secret,norequire_active_mds'
luminous
wip-23208A
master
ovh
centos 7.4
smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} distros/{centos_latest.yaml} objectstore/bluestore.yaml tasks/libcephfs_interface_tests.yaml}
luminous
wip-23208A
master
ovh
ubuntu 16.04
smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} distros/{ubuntu_latest.yaml} objectstore/bluestore.yaml tasks/mon_thrash.yaml}
Command failed (workunit test rados/test.sh) on ovh077 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=wip-23208A TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/rados/test.sh'
luminous
wip-23208A
master
ovh
centos 7.4
smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} distros/{centos_latest.yaml} objectstore/bluestore.yaml tasks/rados_api_tests.yaml}
Command failed (workunit test rados/test.sh) on ovh066 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=wip-23208A TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/rados/test.sh'
luminous
wip-23208A
master
ovh
ubuntu 16.04
smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} distros/{ubuntu_latest.yaml} objectstore/bluestore.yaml tasks/rados_bench.yaml}
luminous
wip-23208A
master
ovh
ubuntu 16.04
smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} distros/{ubuntu_latest.yaml} objectstore/bluestore.yaml tasks/rados_cache_snaps.yaml}
luminous
wip-23208A
master
ovh
ubuntu 16.04
smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} distros/{ubuntu_latest.yaml} objectstore/bluestore.yaml tasks/rados_cls_all.yaml}
luminous
wip-23208A
master
ovh
ubuntu 16.04
smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} distros/{ubuntu_latest.yaml} objectstore/bluestore.yaml tasks/rados_ec_snaps.yaml}
luminous
wip-23208A
master
ovh
ubuntu 16.04
smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} distros/{ubuntu_latest.yaml} objectstore/bluestore.yaml tasks/rados_python.yaml}
Command failed (workunit test rados/test_python.sh) on ovh094 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=wip-23208A TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/rados/test_python.sh'
luminous
wip-23208A
master
ovh
centos 7.4
smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} distros/{centos_latest.yaml} objectstore/bluestore.yaml tasks/rados_workunit_loadgen_mix.yaml}
luminous
wip-23208A
master
ovh
ubuntu 16.04
smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} distros/{ubuntu_latest.yaml} objectstore/bluestore.yaml tasks/rbd_api_tests.yaml}
luminous
wip-23208A
master
ovh
ubuntu 16.04
smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} distros/{ubuntu_latest.yaml} objectstore/bluestore.yaml tasks/rbd_cli_import_export.yaml}
Command failed (workunit test rbd/import_export.sh) on ovh067 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=wip-23208A TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 RBD_CREATE_ARGS=--new-format adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/rbd/import_export.sh'
luminous
wip-23208A
master
ovh
centos 7.4
smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} distros/{centos_latest.yaml} objectstore/bluestore.yaml tasks/rbd_fsx.yaml}
luminous
wip-23208A
master
ovh
centos 7.4
smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} distros/{centos_latest.yaml} objectstore/bluestore.yaml tasks/rbd_python_api_tests.yaml}
Command failed (workunit test rbd/test_librbd_python.sh) on ovh029 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=wip-23208A TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 RBD_FEATURES=1 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/rbd/test_librbd_python.sh'
luminous
wip-23208A
master
ovh
ubuntu 16.04
smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} distros/{ubuntu_latest.yaml} objectstore/bluestore.yaml tasks/rbd_workunit_suites_iozone.yaml}
Command failed on ovh049 with status 110: "sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage rbd --user 0 -p rbd map testimage.client.0 && while test '!' -e /dev/rbd/rbd/testimage.client.0 ; do sleep 1 ; done"
luminous
wip-23208A
master
ovh
ubuntu 16.04
smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} distros/{ubuntu_latest.yaml} objectstore/bluestore.yaml tasks/rgw_ec_s3tests.yaml}
Command failed on ovh046 with status 128: 'git clone -b ceph-wip-23208A git://git.ceph.com/git/s3-tests.git /home/ubuntu/cephtest/s3-tests'
luminous
wip-23208A
master
ovh
ubuntu 16.04
smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} distros/{ubuntu_latest.yaml} objectstore/bluestore.yaml tasks/rgw_s3tests.yaml}
Command failed on ovh044 with status 128: 'git clone -b ceph-wip-23208A git://git.ceph.com/git/s3-tests.git /home/ubuntu/cephtest/s3-tests'
luminous
wip-23208A
master
ovh
centos 7.4
smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} distros/{centos_latest.yaml} objectstore/bluestore.yaml tasks/rgw_swift.yaml}