ID
Status
Ceph Branch
Suite Branch
Teuthology Branch
Machine
OS
Nodes
Description
Failure Reason
next
master
vps
rhel 7.0
upgrade:hammer-x/stress-split-erasure-code-x86_64/{0-cluster/start.yaml 1-hammer-install/hammer.yaml 2-partial-upgrade/firsthalf.yaml 3-thrash/default.yaml 4-mon/mona.yaml 5-workload/ec-rados-default.yaml 6-next-mon/monb.yaml 8-next-mon/monc.yaml 9-workload/ec-rados-plugin=isa-k=2-m=1.yaml distros/rhel_7.0.yaml}
next
master
vps
centos 6.5
upgrade:hammer-x/point-to-point-x/{point-to-point.yaml distros/centos_6.5.yaml}
next
master
vps
centos 6.5
upgrade:hammer-x/stress-split/{0-cluster/start.yaml 1-hammer-install/hammer.yaml 2-partial-upgrade/firsthalf.yaml 3-thrash/default.yaml 4-mon/mona.yaml 5-workload/{rbd-cls.yaml rbd-import-export.yaml readwrite.yaml snaps-few-objects.yaml} 6-next-mon/monb.yaml 7-workload/{radosbench.yaml rbd_api.yaml} 8-next-mon/monc.yaml 9-workload/{rbd-python.yaml rgw-swift.yaml snaps-many-objects.yaml} distros/centos_6.5.yaml}
'wait_until_healthy' reached maximum tries (150) after waiting for 900 seconds
next
master
vps
centos 6.5
upgrade:hammer-x/stress-split-erasure-code/{0-cluster/start.yaml 1-hammer-install/hammer.yaml 2-partial-upgrade/firsthalf.yaml 3-thrash/default.yaml 4-mon/mona.yaml 5-workload/ec-rados-default.yaml 6-next-mon/monb.yaml 8-next-mon/monc.yaml 9-workload/ec-rados-plugin=jerasure-k=3-m=1.yaml distros/centos_6.5.yaml}
'wait_until_healthy' reached maximum tries (150) after waiting for 900 seconds
next
master
vps
centos 6.5
upgrade:hammer-x/parallel/{0-cluster/start.yaml 1-hammer-install/hammer.yaml 2-workload/{ec-rados-default.yaml rados_api.yaml rados_loadgenbig.yaml test_rbd_api.yaml test_rbd_python.yaml} 3-upgrade-sequence/upgrade-all.yaml 4-final-workload/{rados-snaps-few-objects.yaml rados_loadgenmix.yaml rados_mon_thrash.yaml rbd_cls.yaml rbd_import_export.yaml rgw_swift.yaml} distros/centos_6.5.yaml}
Command failed (workunit test rbd/test_librbd.sh) on vpm124 with status 124: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=hammer TESTDIR="/home/ubuntu/cephtest" CEPH_ID="0" PATH=$PATH:/usr/sbin adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/workunit.client.0/rbd/test_librbd.sh'
next
master
vps
debian 7.0
upgrade:hammer-x/parallel/{0-cluster/start.yaml 1-hammer-install/hammer.yaml 2-workload/{ec-rados-default.yaml rados_api.yaml rados_loadgenbig.yaml test_rbd_api.yaml test_rbd_python.yaml} 3-upgrade-sequence/upgrade-mon-osd-mds.yaml 4-final-workload/{rados-snaps-few-objects.yaml rados_loadgenmix.yaml rados_mon_thrash.yaml rbd_cls.yaml rbd_import_export.yaml rgw_swift.yaml} distros/debian_7.0.yaml}
'wait_until_healthy' reached maximum tries (150) after waiting for 900 seconds
next
master
vps
debian 7.0
upgrade:hammer-x/point-to-point-x/{point-to-point.yaml distros/debian_7.0.yaml}
'wait_until_healthy' reached maximum tries (150) after waiting for 900 seconds
next
master
vps
debian 7.0
upgrade:hammer-x/stress-split/{0-cluster/start.yaml 1-hammer-install/hammer.yaml 2-partial-upgrade/firsthalf.yaml 3-thrash/default.yaml 4-mon/mona.yaml 5-workload/{rbd-cls.yaml rbd-import-export.yaml readwrite.yaml snaps-few-objects.yaml} 6-next-mon/monb.yaml 7-workload/{radosbench.yaml rbd_api.yaml} 8-next-mon/monc.yaml 9-workload/{rbd-python.yaml rgw-swift.yaml snaps-many-objects.yaml} distros/debian_7.0.yaml}
'wait_until_healthy' reached maximum tries (150) after waiting for 900 seconds
next
master
vps
debian 7.0
upgrade:hammer-x/stress-split-erasure-code/{0-cluster/start.yaml 1-hammer-install/hammer.yaml 2-partial-upgrade/firsthalf.yaml 3-thrash/default.yaml 4-mon/mona.yaml 5-workload/ec-rados-default.yaml 6-next-mon/monb.yaml 8-next-mon/monc.yaml 9-workload/ec-rados-plugin=jerasure-k=3-m=1.yaml distros/debian_7.0.yaml}
'wait_until_healthy' reached maximum tries (150) after waiting for 900 seconds
next
master
vps
ubuntu 12.04
upgrade:hammer-x/parallel/{0-cluster/start.yaml 1-hammer-install/hammer.yaml 2-workload/{ec-rados-default.yaml rados_api.yaml rados_loadgenbig.yaml test_rbd_api.yaml test_rbd_python.yaml} 3-upgrade-sequence/upgrade-all.yaml 4-final-workload/{rados-snaps-few-objects.yaml rados_loadgenmix.yaml rados_mon_thrash.yaml rbd_cls.yaml rbd_import_export.yaml rgw_swift.yaml} distros/ubuntu_12.04.yaml}
next
master
vps
ubuntu 14.04
upgrade:hammer-x/parallel/{0-cluster/start.yaml 1-hammer-install/hammer.yaml 2-workload/{ec-rados-default.yaml rados_api.yaml rados_loadgenbig.yaml test_rbd_api.yaml test_rbd_python.yaml} 3-upgrade-sequence/upgrade-mon-osd-mds.yaml 4-final-workload/{rados-snaps-few-objects.yaml rados_loadgenmix.yaml rados_mon_thrash.yaml rbd_cls.yaml rbd_import_export.yaml rgw_swift.yaml} distros/ubuntu_14.04.yaml}
'wait_until_healthy' reached maximum tries (150) after waiting for 900 seconds
next
master
vps
ubuntu 14.04
upgrade:hammer-x/stress-split-erasure-code-x86_64/{0-cluster/start.yaml 1-hammer-install/hammer.yaml 2-partial-upgrade/firsthalf.yaml 3-thrash/default.yaml 4-mon/mona.yaml 5-workload/ec-rados-default.yaml 6-next-mon/monb.yaml 8-next-mon/monc.yaml 9-workload/ec-rados-plugin=isa-k=2-m=1.yaml distros/ubuntu_14.04.yaml}
SSH connection to vpm129 was lost: u'sudo DEBIAN_FRONTEND=noninteractive apt-get -y --force-yes -o Dpkg::Options::="--force-confdef" -o Dpkg::Options::="--force-confold" install ceph=0.94.2-156-g8355bda-1trusty ceph-dbg=0.94.2-156-g8355bda-1trusty ceph-mds=0.94.2-156-g8355bda-1trusty ceph-mds-dbg=0.94.2-156-g8355bda-1trusty ceph-common=0.94.2-156-g8355bda-1trusty ceph-common-dbg=0.94.2-156-g8355bda-1trusty ceph-fuse=0.94.2-156-g8355bda-1trusty ceph-fuse-dbg=0.94.2-156-g8355bda-1trusty ceph-test=0.94.2-156-g8355bda-1trusty ceph-test-dbg=0.94.2-156-g8355bda-1trusty radosgw=0.94.2-156-g8355bda-1trusty radosgw-dbg=0.94.2-156-g8355bda-1trusty python-ceph=0.94.2-156-g8355bda-1trusty libcephfs1=0.94.2-156-g8355bda-1trusty libcephfs1-dbg=0.94.2-156-g8355bda-1trusty libcephfs-java=0.94.2-156-g8355bda-1trusty libcephfs-jni=0.94.2-156-g8355bda-1trusty librados2=0.94.2-156-g8355bda-1trusty librados2-dbg=0.94.2-156-g8355bda-1trusty librbd1=0.94.2-156-g8355bda-1trusty librbd1-dbg=0.94.2-156-g8355bda-1trusty rbd-fuse=0.94.2-156-g8355bda-1trusty librados2=0.94.2-156-g8355bda-1trusty librados2-dbg=0.94.2-156-g8355bda-1trusty librbd1=0.94.2-156-g8355bda-1trusty librbd1-dbg=0.94.2-156-g8355bda-1trusty'
next
master
vps
ubuntu 12.04
upgrade:hammer-x/point-to-point-x/{point-to-point.yaml distros/ubuntu_12.04.yaml}
Command failed on vpm016 with status 100: u'sudo DEBIAN_FRONTEND=noninteractive apt-get -y --force-yes -o Dpkg::Options::="--force-confdef" -o Dpkg::Options::="--force-confold" install librbd1-dbg=0.94.2-156-g8355bda-1precise ceph=0.94.2-156-g8355bda-1precise ceph-test=0.94.2-156-g8355bda-1precise ceph-dbg=0.94.2-156-g8355bda-1precise rbd-fuse=0.94.2-156-g8355bda-1precise librados2-dbg=0.94.2-156-g8355bda-1precise ceph-fuse-dbg=0.94.2-156-g8355bda-1precise libcephfs-jni=0.94.2-156-g8355bda-1precise libcephfs1-dbg=0.94.2-156-g8355bda-1precise radosgw=0.94.2-156-g8355bda-1precise librados2=0.94.2-156-g8355bda-1precise libcephfs1=0.94.2-156-g8355bda-1precise ceph-mds=0.94.2-156-g8355bda-1precise radosgw-dbg=0.94.2-156-g8355bda-1precise librbd1=0.94.2-156-g8355bda-1precise python-ceph=0.94.2-156-g8355bda-1precise ceph-test-dbg=0.94.2-156-g8355bda-1precise ceph-fuse=0.94.2-156-g8355bda-1precise ceph-common=0.94.2-156-g8355bda-1precise libcephfs-java=0.94.2-156-g8355bda-1precise ceph-common-dbg=0.94.2-156-g8355bda-1precise ceph-mds-dbg=0.94.2-156-g8355bda-1precise'
next
master
vps
ubuntu 12.04
upgrade:hammer-x/stress-split/{0-cluster/start.yaml 1-hammer-install/hammer.yaml 2-partial-upgrade/firsthalf.yaml 3-thrash/default.yaml 4-mon/mona.yaml 5-workload/{rbd-cls.yaml rbd-import-export.yaml readwrite.yaml snaps-few-objects.yaml} 6-next-mon/monb.yaml 7-workload/{radosbench.yaml rbd_api.yaml} 8-next-mon/monc.yaml 9-workload/{rbd-python.yaml rgw-swift.yaml snaps-many-objects.yaml} distros/ubuntu_12.04.yaml}
'wait_until_healthy' reached maximum tries (150) after waiting for 900 seconds
next
master
vps
ubuntu 12.04
upgrade:hammer-x/stress-split-erasure-code/{0-cluster/start.yaml 1-hammer-install/hammer.yaml 2-partial-upgrade/firsthalf.yaml 3-thrash/default.yaml 4-mon/mona.yaml 5-workload/ec-rados-default.yaml 6-next-mon/monb.yaml 8-next-mon/monc.yaml 9-workload/ec-rados-plugin=jerasure-k=3-m=1.yaml distros/ubuntu_12.04.yaml}
SSH connection to vpm129 was lost: u'sudo DEBIAN_FRONTEND=noninteractive apt-get -y --force-yes -o Dpkg::Options::="--force-confdef" -o Dpkg::Options::="--force-confold" install ceph=0.94.2-156-g8355bda-1precise ceph-dbg=0.94.2-156-g8355bda-1precise ceph-mds=0.94.2-156-g8355bda-1precise ceph-mds-dbg=0.94.2-156-g8355bda-1precise ceph-common=0.94.2-156-g8355bda-1precise ceph-common-dbg=0.94.2-156-g8355bda-1precise ceph-fuse=0.94.2-156-g8355bda-1precise ceph-fuse-dbg=0.94.2-156-g8355bda-1precise ceph-test=0.94.2-156-g8355bda-1precise ceph-test-dbg=0.94.2-156-g8355bda-1precise radosgw=0.94.2-156-g8355bda-1precise radosgw-dbg=0.94.2-156-g8355bda-1precise python-ceph=0.94.2-156-g8355bda-1precise libcephfs1=0.94.2-156-g8355bda-1precise libcephfs1-dbg=0.94.2-156-g8355bda-1precise libcephfs-java=0.94.2-156-g8355bda-1precise libcephfs-jni=0.94.2-156-g8355bda-1precise librados2=0.94.2-156-g8355bda-1precise librados2-dbg=0.94.2-156-g8355bda-1precise librbd1=0.94.2-156-g8355bda-1precise librbd1-dbg=0.94.2-156-g8355bda-1precise rbd-fuse=0.94.2-156-g8355bda-1precise librados2=0.94.2-156-g8355bda-1precise librados2-dbg=0.94.2-156-g8355bda-1precise librbd1=0.94.2-156-g8355bda-1precise librbd1-dbg=0.94.2-156-g8355bda-1precise'
next
master
vps
centos 6.5
upgrade:hammer-x/parallel/{0-cluster/start.yaml 1-hammer-install/hammer.yaml 2-workload/{ec-rados-default.yaml rados_api.yaml rados_loadgenbig.yaml test_rbd_api.yaml test_rbd_python.yaml} 3-upgrade-sequence/upgrade-mon-osd-mds.yaml 4-final-workload/{rados-snaps-few-objects.yaml rados_loadgenmix.yaml rados_mon_thrash.yaml rbd_cls.yaml rbd_import_export.yaml rgw_swift.yaml} distros/centos_6.5.yaml}
Command failed with status 3: 'ansible-playbook -v --extra-vars \'{"ansible_ssh_user": "ubuntu"}\' -i /etc/ansible/hosts --limit vpm124.front.sepia.ceph.com,vpm136.front.sepia.ceph.com,vpm177.front.sepia.ceph.com /var/lib/teuthworker/src/ceph-cm-ansible_master/cephlab.yml'
next
master
vps
debian 7.0
upgrade:hammer-x/parallel/{0-cluster/start.yaml 1-hammer-install/hammer.yaml 2-workload/{ec-rados-default.yaml rados_api.yaml rados_loadgenbig.yaml test_rbd_api.yaml test_rbd_python.yaml} 3-upgrade-sequence/upgrade-all.yaml 4-final-workload/{rados-snaps-few-objects.yaml rados_loadgenmix.yaml rados_mon_thrash.yaml rbd_cls.yaml rbd_import_export.yaml rgw_swift.yaml} distros/debian_7.0.yaml}
next
master
vps
ubuntu 14.04
upgrade:hammer-x/point-to-point-x/{point-to-point.yaml distros/ubuntu_14.04.yaml}
Command failed on vpm118 with status 100: u'sudo DEBIAN_FRONTEND=noninteractive apt-get -y --force-yes -o Dpkg::Options::="--force-confdef" -o Dpkg::Options::="--force-confold" install librbd1-dbg=0.94.2-156-g8355bda-1trusty ceph=0.94.2-156-g8355bda-1trusty ceph-test=0.94.2-156-g8355bda-1trusty ceph-dbg=0.94.2-156-g8355bda-1trusty rbd-fuse=0.94.2-156-g8355bda-1trusty librados2-dbg=0.94.2-156-g8355bda-1trusty ceph-fuse-dbg=0.94.2-156-g8355bda-1trusty libcephfs-jni=0.94.2-156-g8355bda-1trusty libcephfs1-dbg=0.94.2-156-g8355bda-1trusty radosgw=0.94.2-156-g8355bda-1trusty librados2=0.94.2-156-g8355bda-1trusty libcephfs1=0.94.2-156-g8355bda-1trusty ceph-mds=0.94.2-156-g8355bda-1trusty radosgw-dbg=0.94.2-156-g8355bda-1trusty librbd1=0.94.2-156-g8355bda-1trusty python-ceph=0.94.2-156-g8355bda-1trusty ceph-test-dbg=0.94.2-156-g8355bda-1trusty ceph-fuse=0.94.2-156-g8355bda-1trusty ceph-common=0.94.2-156-g8355bda-1trusty libcephfs-java=0.94.2-156-g8355bda-1trusty ceph-common-dbg=0.94.2-156-g8355bda-1trusty ceph-mds-dbg=0.94.2-156-g8355bda-1trusty'
next
master
vps
ubuntu 14.04
upgrade:hammer-x/stress-split/{0-cluster/start.yaml 1-hammer-install/hammer.yaml 2-partial-upgrade/firsthalf.yaml 3-thrash/default.yaml 4-mon/mona.yaml 5-workload/{rbd-cls.yaml rbd-import-export.yaml readwrite.yaml snaps-few-objects.yaml} 6-next-mon/monb.yaml 7-workload/{radosbench.yaml rbd_api.yaml} 8-next-mon/monc.yaml 9-workload/{rbd-python.yaml rgw-swift.yaml snaps-many-objects.yaml} distros/ubuntu_14.04.yaml}
'wait_until_healthy' reached maximum tries (150) after waiting for 900 seconds
next
master
vps
ubuntu 14.04
upgrade:hammer-x/stress-split-erasure-code/{0-cluster/start.yaml 1-hammer-install/hammer.yaml 2-partial-upgrade/firsthalf.yaml 3-thrash/default.yaml 4-mon/mona.yaml 5-workload/ec-rados-default.yaml 6-next-mon/monb.yaml 8-next-mon/monc.yaml 9-workload/ec-rados-plugin=jerasure-k=3-m=1.yaml distros/ubuntu_14.04.yaml}
'wait_until_healthy' reached maximum tries (150) after waiting for 900 seconds
next
master
vps
ubuntu 12.04
upgrade:hammer-x/parallel/{0-cluster/start.yaml 1-hammer-install/hammer.yaml 2-workload/{ec-rados-default.yaml rados_api.yaml rados_loadgenbig.yaml test_rbd_api.yaml test_rbd_python.yaml} 3-upgrade-sequence/upgrade-mon-osd-mds.yaml 4-final-workload/{rados-snaps-few-objects.yaml rados_loadgenmix.yaml rados_mon_thrash.yaml rbd_cls.yaml rbd_import_export.yaml rgw_swift.yaml} distros/ubuntu_12.04.yaml}
Command failed with status 3: 'ansible-playbook -v --extra-vars \'{"ansible_ssh_user": "ubuntu"}\' -i /etc/ansible/hosts --limit vpm097.front.sepia.ceph.com,vpm158.front.sepia.ceph.com,vpm098.front.sepia.ceph.com /var/lib/teuthworker/src/ceph-cm-ansible_master/cephlab.yml'
next
master
vps
ubuntu 14.04
upgrade:hammer-x/parallel/{0-cluster/start.yaml 1-hammer-install/hammer.yaml 2-workload/{ec-rados-default.yaml rados_api.yaml rados_loadgenbig.yaml test_rbd_api.yaml test_rbd_python.yaml} 3-upgrade-sequence/upgrade-all.yaml 4-final-workload/{rados-snaps-few-objects.yaml rados_loadgenmix.yaml rados_mon_thrash.yaml rbd_cls.yaml rbd_import_export.yaml rgw_swift.yaml} distros/ubuntu_14.04.yaml}