ID
Status
Ceph Branch
Suite Branch
Teuthology Branch
Machine
OS
Nodes
Description
Failure Reason
firefly
vps
centos 
ceph-deploy/fs/{distro/centos_6.3.yaml tasks/cfuse_workunit_suites_blogbench.yaml}
Found coredumps on ubuntu@vpm010.front.sepia.ceph.com
firefly
vps
centos 
ceph-deploy/fs/{distro/centos_6.3.yaml tasks/cfuse_workunit_suites_dbench.yaml}
Found coredumps on ubuntu@vpm019.front.sepia.ceph.com
firefly
vps
centos 
ceph-deploy/fs/{distro/centos_6.3.yaml tasks/cfuse_workunit_suites_fsstress.yaml}
Found coredumps on ubuntu@vpm039.front.sepia.ceph.com
firefly
vps
centos 
ceph-deploy/fs/{distro/centos_6.4.yaml tasks/cfuse_workunit_suites_blogbench.yaml}
Found coredumps on ubuntu@vpm023.front.sepia.ceph.com
firefly
vps
centos 
ceph-deploy/fs/{distro/centos_6.4.yaml tasks/cfuse_workunit_suites_dbench.yaml}
Found coredumps on ubuntu@vpm031.front.sepia.ceph.com
firefly
vps
centos 
ceph-deploy/fs/{distro/centos_6.4.yaml tasks/cfuse_workunit_suites_fsstress.yaml}
Found coredumps on ubuntu@vpm047.front.sepia.ceph.com
firefly
vps
debian 
ceph-deploy/fs/{distro/debian_7.0.yaml tasks/cfuse_workunit_suites_blogbench.yaml}
Command failed on 10.214.138.64 with status 1: 'echo ttyS1 | sudo tee /sys/module/kgdboc/parameters/kgdboc'
firefly
vps
debian 
ceph-deploy/fs/{distro/debian_7.0.yaml tasks/cfuse_workunit_suites_dbench.yaml}
Found coredumps on ubuntu@vpm067.front.sepia.ceph.com
firefly
vps
debian 
ceph-deploy/fs/{distro/debian_7.0.yaml tasks/cfuse_workunit_suites_fsstress.yaml}
Found coredumps on ubuntu@vpm047.front.sepia.ceph.com
firefly
vps
fedora 
ceph-deploy/fs/{distro/fedora_19.yaml tasks/cfuse_workunit_suites_blogbench.yaml}
ceph health was unable to get 'HEALTH_OK' after waiting 15 minutes
firefly
vps
fedora 
ceph-deploy/fs/{distro/fedora_19.yaml tasks/cfuse_workunit_suites_dbench.yaml}
ceph health was unable to get 'HEALTH_OK' after waiting 15 minutes
firefly
vps
fedora 
ceph-deploy/fs/{distro/fedora_19.yaml tasks/cfuse_workunit_suites_fsstress.yaml}
ceph health was unable to get 'HEALTH_OK' after waiting 15 minutes
firefly
vps
rhel 
ceph-deploy/fs/{distro/rhel_6.3.yaml tasks/cfuse_workunit_suites_blogbench.yaml}
Found coredumps on ubuntu@vpm043.front.sepia.ceph.com
firefly
vps
rhel 
ceph-deploy/fs/{distro/rhel_6.3.yaml tasks/cfuse_workunit_suites_dbench.yaml}
Found coredumps on ubuntu@vpm063.front.sepia.ceph.com
firefly
vps
rhel 
ceph-deploy/fs/{distro/rhel_6.3.yaml tasks/cfuse_workunit_suites_fsstress.yaml}
Found coredumps on ubuntu@vpm107.front.sepia.ceph.com
firefly
vps
rhel 
ceph-deploy/fs/{distro/rhel_6.4.yaml tasks/cfuse_workunit_suites_blogbench.yaml}
firefly
vps
rhel 
ceph-deploy/fs/{distro/rhel_6.4.yaml tasks/cfuse_workunit_suites_dbench.yaml}
Found coredumps on ubuntu@vpm091.front.sepia.ceph.com
firefly
vps
rhel 
ceph-deploy/fs/{distro/rhel_6.4.yaml tasks/cfuse_workunit_suites_fsstress.yaml}
Found coredumps on ubuntu@vpm027.front.sepia.ceph.com
firefly
vps
ubuntu 
ceph-deploy/fs/{distro/ubuntu_12.04.yaml tasks/cfuse_workunit_suites_blogbench.yaml}
firefly
vps
ubuntu 
ceph-deploy/fs/{distro/ubuntu_12.04.yaml tasks/cfuse_workunit_suites_dbench.yaml}
firefly
vps
ubuntu 
ceph-deploy/fs/{distro/ubuntu_12.04.yaml tasks/cfuse_workunit_suites_fsstress.yaml}
firefly
vps
centos 
ceph-deploy/rados/{distro/centos_6.3.yaml tasks/rados_api_tests.yaml}
Command failed on 10.214.138.111 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=8fdfece9fd5419eeb1bc65b3ac4987f6e150fd9f TESTDIR="/home/ubuntu/cephtest" CEPH_ID="0" adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage /home/ubuntu/cephtest/workunit.client.0/rados/test.sh'
firefly
vps
centos 
ceph-deploy/rados/{distro/centos_6.3.yaml tasks/rados_python.yaml}
Found coredumps on ubuntu@vpm031.front.sepia.ceph.com
firefly
vps
centos 
ceph-deploy/rados/{distro/centos_6.3.yaml tasks/rados_workunit_loadgen_big.yaml}
firefly
vps
centos 
ceph-deploy/rados/{distro/centos_6.4.yaml tasks/rados_api_tests.yaml}
Command failed on 10.214.138.68 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=8fdfece9fd5419eeb1bc65b3ac4987f6e150fd9f TESTDIR="/home/ubuntu/cephtest" CEPH_ID="0" adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage /home/ubuntu/cephtest/workunit.client.0/rados/test.sh'
firefly
vps
centos 
ceph-deploy/rados/{distro/centos_6.4.yaml tasks/rados_python.yaml}
Found coredumps on ubuntu@vpm027.front.sepia.ceph.com
firefly
vps
centos 
ceph-deploy/rados/{distro/centos_6.4.yaml tasks/rados_workunit_loadgen_big.yaml}
firefly
vps
debian 
ceph-deploy/rados/{distro/debian_7.0.yaml tasks/rados_api_tests.yaml}
Command failed on 10.214.138.98 with status 1: 'echo ttyS1 | sudo tee /sys/module/kgdboc/parameters/kgdboc'
firefly
vps
debian 
ceph-deploy/rados/{distro/debian_7.0.yaml tasks/rados_python.yaml}
Found coredumps on ubuntu@vpm084.front.sepia.ceph.com
firefly
vps
debian 
ceph-deploy/rados/{distro/debian_7.0.yaml tasks/rados_workunit_loadgen_big.yaml}
Found coredumps on ubuntu@vpm107.front.sepia.ceph.com
firefly
vps
fedora 
ceph-deploy/rados/{distro/fedora_19.yaml tasks/rados_api_tests.yaml}
ceph health was unable to get 'HEALTH_OK' after waiting 15 minutes
firefly
vps
fedora 
ceph-deploy/rados/{distro/fedora_19.yaml tasks/rados_python.yaml}
ceph health was unable to get 'HEALTH_OK' after waiting 15 minutes
firefly
vps
fedora 
ceph-deploy/rados/{distro/fedora_19.yaml tasks/rados_workunit_loadgen_big.yaml}
ceph health was unable to get 'HEALTH_OK' after waiting 15 minutes
firefly
vps
rhel 
ceph-deploy/rados/{distro/rhel_6.3.yaml tasks/rados_api_tests.yaml}
Command failed on 10.214.138.104 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=8fdfece9fd5419eeb1bc65b3ac4987f6e150fd9f TESTDIR="/home/ubuntu/cephtest" CEPH_ID="0" adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage /home/ubuntu/cephtest/workunit.client.0/rados/test.sh'
firefly
vps
rhel 
ceph-deploy/rados/{distro/rhel_6.3.yaml tasks/rados_python.yaml}
Found coredumps on ubuntu@vpm084.front.sepia.ceph.com
firefly
vps
rhel 
ceph-deploy/rados/{distro/rhel_6.3.yaml tasks/rados_workunit_loadgen_big.yaml}
firefly
vps
rhel 
ceph-deploy/rados/{distro/rhel_6.4.yaml tasks/rados_api_tests.yaml}
firefly
vps
rhel 
ceph-deploy/rados/{distro/rhel_6.4.yaml tasks/rados_python.yaml}
Found coredumps on ubuntu@vpm051.front.sepia.ceph.com
firefly
vps
rhel 
ceph-deploy/rados/{distro/rhel_6.4.yaml tasks/rados_workunit_loadgen_big.yaml}
Found coredumps on ubuntu@vpm075.front.sepia.ceph.com
firefly
vps
ubuntu 
ceph-deploy/rados/{distro/ubuntu_12.04.yaml tasks/rados_api_tests.yaml}
Command failed on 10.214.138.147 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=8fdfece9fd5419eeb1bc65b3ac4987f6e150fd9f TESTDIR="/home/ubuntu/cephtest" CEPH_ID="0" adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage /home/ubuntu/cephtest/workunit.client.0/rados/test.sh'
firefly
vps
ubuntu 
ceph-deploy/rados/{distro/ubuntu_12.04.yaml tasks/rados_python.yaml}
firefly
vps
ubuntu 
ceph-deploy/rados/{distro/ubuntu_12.04.yaml tasks/rados_workunit_loadgen_big.yaml}
firefly
vps
centos 
ceph-deploy/rbd/{distro/centos_6.3.yaml tasks/rbd_api_tests_old_format.yaml}
Found coredumps on ubuntu@vpm035.front.sepia.ceph.com
firefly
vps
centos 
ceph-deploy/rbd/{distro/centos_6.3.yaml tasks/rbd_cli_tests.yaml}
Found coredumps on ubuntu@vpm095.front.sepia.ceph.com
firefly
vps
centos 
ceph-deploy/rbd/{distro/centos_6.3.yaml tasks/rbd_cls_test.yaml}
Found coredumps on ubuntu@vpm095.front.sepia.ceph.com
firefly
vps
centos 
ceph-deploy/rbd/{distro/centos_6.3.yaml tasks/rbd_python_api_tests.yaml}
Found coredumps on ubuntu@vpm063.front.sepia.ceph.com
firefly
vps
centos 
ceph-deploy/rbd/{distro/centos_6.4.yaml tasks/rbd_api_tests_old_format.yaml}
Found coredumps on ubuntu@vpm055.front.sepia.ceph.com
firefly
vps
centos 
ceph-deploy/rbd/{distro/centos_6.4.yaml tasks/rbd_cli_tests.yaml}
Found coredumps on ubuntu@vpm067.front.sepia.ceph.com
firefly
vps
centos 
ceph-deploy/rbd/{distro/centos_6.4.yaml tasks/rbd_cls_test.yaml}
firefly
vps
centos 
ceph-deploy/rbd/{distro/centos_6.4.yaml tasks/rbd_python_api_tests.yaml}
failed to install new distro kernel version within timeout
firefly
vps
debian 
ceph-deploy/rbd/{distro/debian_7.0.yaml tasks/rbd_api_tests_old_format.yaml}
Found coredumps on ubuntu@vpm055.front.sepia.ceph.com
firefly
vps
debian 
ceph-deploy/rbd/{distro/debian_7.0.yaml tasks/rbd_cli_tests.yaml}
Found coredumps on ubuntu@vpm106.front.sepia.ceph.com
firefly
vps
debian 
ceph-deploy/rbd/{distro/debian_7.0.yaml tasks/rbd_cls_test.yaml}
Found coredumps on ubuntu@vpm023.front.sepia.ceph.com
firefly
vps
debian 
ceph-deploy/rbd/{distro/debian_7.0.yaml tasks/rbd_python_api_tests.yaml}
firefly
vps
fedora 
ceph-deploy/rbd/{distro/fedora_19.yaml tasks/rbd_api_tests_old_format.yaml}
ceph health was unable to get 'HEALTH_OK' after waiting 15 minutes
firefly
vps
fedora 
ceph-deploy/rbd/{distro/fedora_19.yaml tasks/rbd_cli_tests.yaml}
ceph health was unable to get 'HEALTH_OK' after waiting 15 minutes
firefly
vps
fedora 
ceph-deploy/rbd/{distro/fedora_19.yaml tasks/rbd_cls_test.yaml}
ceph health was unable to get 'HEALTH_OK' after waiting 15 minutes
firefly
vps
fedora 
ceph-deploy/rbd/{distro/fedora_19.yaml tasks/rbd_python_api_tests.yaml}
ceph health was unable to get 'HEALTH_OK' after waiting 15 minutes
firefly
vps
rhel 
ceph-deploy/rbd/{distro/rhel_6.3.yaml tasks/rbd_api_tests_old_format.yaml}
Found coredumps on ubuntu@vpm084.front.sepia.ceph.com
firefly
vps
rhel 
ceph-deploy/rbd/{distro/rhel_6.3.yaml tasks/rbd_cli_tests.yaml}
Found coredumps on ubuntu@vpm071.front.sepia.ceph.com
firefly
vps
rhel 
ceph-deploy/rbd/{distro/rhel_6.3.yaml tasks/rbd_cls_test.yaml}
Found coredumps on ubuntu@vpm103.front.sepia.ceph.com
firefly
vps
rhel 
ceph-deploy/rbd/{distro/rhel_6.3.yaml tasks/rbd_python_api_tests.yaml}
firefly
vps
rhel 
ceph-deploy/rbd/{distro/rhel_6.4.yaml tasks/rbd_api_tests_old_format.yaml}
Found coredumps on ubuntu@vpm023.front.sepia.ceph.com
firefly
vps
rhel 
ceph-deploy/rbd/{distro/rhel_6.4.yaml tasks/rbd_cli_tests.yaml}
Found coredumps on ubuntu@vpm103.front.sepia.ceph.com
firefly
vps
rhel 
ceph-deploy/rbd/{distro/rhel_6.4.yaml tasks/rbd_cls_test.yaml}
Found coredumps on ubuntu@vpm055.front.sepia.ceph.com
firefly
vps
rhel 
ceph-deploy/rbd/{distro/rhel_6.4.yaml tasks/rbd_python_api_tests.yaml}
firefly
vps
ubuntu 
ceph-deploy/rbd/{distro/ubuntu_12.04.yaml tasks/rbd_api_tests_old_format.yaml}
firefly
vps
ubuntu 
ceph-deploy/rbd/{distro/ubuntu_12.04.yaml tasks/rbd_cli_tests.yaml}
firefly
vps
ubuntu 
ceph-deploy/rbd/{distro/ubuntu_12.04.yaml tasks/rbd_cls_test.yaml}
firefly
vps
ubuntu 
ceph-deploy/rbd/{distro/ubuntu_12.04.yaml tasks/rbd_python_api_tests.yaml}
firefly
vps
centos 
ceph-deploy/singleton/{all/basic-test.yaml distro/centos_6.3.yaml}
Found coredumps on ubuntu@vpm067.front.sepia.ceph.com
firefly
vps
centos 
ceph-deploy/singleton/{all/basic-test.yaml distro/centos_6.4.yaml}
Found coredumps on ubuntu@vpm079.front.sepia.ceph.com
firefly
vps
debian 
ceph-deploy/singleton/{all/basic-test.yaml distro/debian_7.0.yaml}
Found coredumps on ubuntu@vpm101.front.sepia.ceph.com
firefly
vps
fedora 
ceph-deploy/singleton/{all/basic-test.yaml distro/fedora_19.yaml}
ceph health was unable to get 'HEALTH_OK' after waiting 15 minutes
firefly
vps
rhel 
ceph-deploy/singleton/{all/basic-test.yaml distro/rhel_6.3.yaml}
Found coredumps on ubuntu@vpm039.front.sepia.ceph.com
firefly
vps
rhel 
ceph-deploy/singleton/{all/basic-test.yaml distro/rhel_6.4.yaml}
Found coredumps on ubuntu@vpm010.front.sepia.ceph.com
firefly
vps
ubuntu 
ceph-deploy/singleton/{all/basic-test.yaml distro/ubuntu_12.04.yaml}