Posted | Started | Updated | Runtime |
Duration |
In Waiting |
Machine | Teuthology Branch | OS Type | OS Version | Nodes | Status |
---|---|---|---|---|---|---|---|---|---|---|---|
2016-12-19 18:25:26 | 2016-12-19 18:26:05 | 2016-12-19 20:58:07 | 2:32:02 | 2:21:49 | 0:10:13 | vps | master | centos | 7.2 | 3 | fail |
Description: upgrade:hammer-jewel-x/stress-split/{0-cluster/{openstack.yaml start.yaml} 1-hammer-install-and-upgrade-to-jewel/hammer-to-jewel.yaml 2-partial-upgrade/firsthalf.yaml 3-thrash/default.yaml 4-mon/mona.yaml 5-workload/{rbd-cls.yaml rbd-import-export.yaml readwrite.yaml snaps-few-objects.yaml} 6-next-mon/monb.yaml 7-workload/{radosbench.yaml rbd_api.yaml} 8-next-mon/monc.yaml 9-workload/{rbd-python.yaml rgw-swift.yaml snaps-many-objects.yaml} distros/centos_7.2.yaml}
Sentry event: http://sentry.ceph.com/sepia/teuthology/?q=bd309b56f6af4685abe1e7abb260e6fa
Performance graphs: http://pcp.front.sepia.ceph.com:44323/grafana/index.html#/dashboard/script/index.js?time_to=2016-12-19T20%3A48%3A52&time_from=2016-12-19T18%3A34%3A04&hosts=vpm091%2Cvpm125%2Cvpm005
Command failed (workunit test rbd/test_librbd_python.sh) on vpm005 with status 134: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=jewel TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/rbd/test_librbd_python.sh'