ID
Status
Ceph Branch
Suite Branch
Teuthology Branch
Machine
OS
Nodes
Description
Failure Reason
wip-sage2-testing-2019-04-18-1339
wip-sage2-testing-2019-04-18-1339
master
smithi
ubuntu 16.04
rados:mgr/{clusters/{2-node-mgr.yaml openstack.yaml} debug/mgr.yaml objectstore/filestore-xfs.yaml supported-random-distro$/{ubuntu_16.04.yaml} tasks/orchestrator_cli.yaml}
"2019-04-19 03:11:08.720499 mon.b (mon.0) 102 : cluster [WRN] Health check failed: 3 osds down (OSD_DOWN)" in cluster log
wip-sage2-testing-2019-04-18-1339
wip-sage2-testing-2019-04-18-1339
master
smithi
rhel 7.5
rados:mgr/{clusters/{2-node-mgr.yaml openstack.yaml} debug/mgr.yaml objectstore/bluestore-bitmap.yaml supported-random-distro$/{rhel_latest.yaml} tasks/progress.yaml}
wip-sage2-testing-2019-04-18-1339
wip-sage2-testing-2019-04-18-1339
master
smithi
centos 7.5
rados:mgr/{clusters/{2-node-mgr.yaml openstack.yaml} debug/mgr.yaml objectstore/bluestore-comp.yaml supported-random-distro$/{centos_latest.yaml} tasks/prometheus.yaml}
Test failure: test_file_sd_command (tasks.mgr.test_prometheus.TestPrometheus)
wip-sage2-testing-2019-04-18-1339
wip-sage2-testing-2019-04-18-1339
master
smithi
centos 7.5
rados:mgr/{clusters/{2-node-mgr.yaml openstack.yaml} debug/mgr.yaml objectstore/bluestore-low-osd-mem-target.yaml supported-random-distro$/{centos_latest.yaml} tasks/ssh_orchestrator.yaml}
wip-sage2-testing-2019-04-18-1339
wip-sage2-testing-2019-04-18-1339
master
smithi
rhel 7.5
rados:mgr/{clusters/{2-node-mgr.yaml openstack.yaml} debug/mgr.yaml objectstore/bluestore-stupid.yaml supported-random-distro$/{rhel_latest.yaml} tasks/workunits.yaml}
Command failed (workunit test mgr/test_localpool.sh) on smithi058 with status 124: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=7c1ddb447f58a7c6ec8acdcd1c65284d108265de TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/mgr/test_localpool.sh'
wip-sage2-testing-2019-04-18-1339
wip-sage2-testing-2019-04-18-1339
master
smithi
centos 7.5
rados:mgr/{clusters/{2-node-mgr.yaml openstack.yaml} debug/mgr.yaml objectstore/bluestore-stupid.yaml supported-random-distro$/{centos_latest.yaml} tasks/crash.yaml}
wip-sage2-testing-2019-04-18-1339
wip-sage2-testing-2019-04-18-1339
master
smithi
centos 7.5
rados:mgr/{clusters/{2-node-mgr.yaml openstack.yaml} debug/mgr.yaml objectstore/filestore-xfs.yaml supported-random-distro$/{centos_latest.yaml} tasks/dashboard.yaml}
Command failed on smithi173 with status 1: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper kill ceph-osd -f --cluster ceph -i 0'
wip-sage2-testing-2019-04-18-1339
wip-sage2-testing-2019-04-18-1339
master
smithi
rhel 7.5
rados:mgr/{clusters/{2-node-mgr.yaml openstack.yaml} debug/mgr.yaml objectstore/bluestore-bitmap.yaml supported-random-distro$/{rhel_latest.yaml} tasks/failover.yaml}
wip-sage2-testing-2019-04-18-1339
wip-sage2-testing-2019-04-18-1339
master
smithi
ubuntu 16.04
rados:mgr/{clusters/{2-node-mgr.yaml openstack.yaml} debug/mgr.yaml objectstore/bluestore-comp.yaml supported-random-distro$/{ubuntu_16.04.yaml} tasks/insights.yaml}
Test failure: test_crash_history (tasks.mgr.test_insights.TestInsights)
wip-sage2-testing-2019-04-18-1339
wip-sage2-testing-2019-04-18-1339
master
smithi
ubuntu 16.04
rados:mgr/{clusters/{2-node-mgr.yaml openstack.yaml} debug/mgr.yaml objectstore/bluestore-low-osd-mem-target.yaml supported-random-distro$/{ubuntu_16.04.yaml} tasks/module_selftest.yaml}
Test failure: test_devicehealth (tasks.mgr.test_module_selftest.TestModuleSelftest)
wip-sage2-testing-2019-04-18-1339
wip-sage2-testing-2019-04-18-1339
master
smithi
ubuntu 18.04
rados:mgr/{clusters/{2-node-mgr.yaml openstack.yaml} debug/mgr.yaml objectstore/bluestore-stupid.yaml supported-random-distro$/{ubuntu_latest.yaml} tasks/orchestrator_cli.yaml}
wip-sage2-testing-2019-04-18-1339
wip-sage2-testing-2019-04-18-1339
master
smithi
centos 7.5
rados:mgr/{clusters/{2-node-mgr.yaml openstack.yaml} debug/mgr.yaml objectstore/filestore-xfs.yaml supported-random-distro$/{centos_latest.yaml} tasks/progress.yaml}
"2019-04-19 03:42:53.599109 mon.a (mon.0) 105 : cluster [WRN] Health check failed: 3 osds down (OSD_DOWN)" in cluster log
wip-sage2-testing-2019-04-18-1339
wip-sage2-testing-2019-04-18-1339
master
smithi
ubuntu 16.04
rados:mgr/{clusters/{2-node-mgr.yaml openstack.yaml} debug/mgr.yaml objectstore/bluestore-bitmap.yaml supported-random-distro$/{ubuntu_16.04.yaml} tasks/prometheus.yaml}
Test failure: test_file_sd_command (tasks.mgr.test_prometheus.TestPrometheus)
wip-sage2-testing-2019-04-18-1339
wip-sage2-testing-2019-04-18-1339
master
smithi
ubuntu 18.04
rados:mgr/{clusters/{2-node-mgr.yaml openstack.yaml} debug/mgr.yaml objectstore/bluestore-comp.yaml supported-random-distro$/{ubuntu_latest.yaml} tasks/ssh_orchestrator.yaml}
wip-sage2-testing-2019-04-18-1339
wip-sage2-testing-2019-04-18-1339
master
smithi
ubuntu 18.04
rados:mgr/{clusters/{2-node-mgr.yaml openstack.yaml} debug/mgr.yaml objectstore/bluestore-low-osd-mem-target.yaml supported-random-distro$/{ubuntu_latest.yaml} tasks/workunits.yaml}
Command failed (workunit test mgr/test_localpool.sh) on smithi049 with status 124: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=7c1ddb447f58a7c6ec8acdcd1c65284d108265de TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/mgr/test_localpool.sh'