User | Scheduled | Started | Updated | Runtime | Suite | Branch | Machine Type | Revision | Pass | Fail |
---|---|---|---|---|---|---|---|---|---|---|
teuthology | 2016-08-14 04:20:03 | 2016-08-14 04:57:09 | 2016-08-14 09:41:12 | 4:44:03 | upgrade:jewel-x | master | vps | ea63a4e | 3 | 8 |
Status | Job ID | Links | Posted | Started | Updated | Runtime |
Duration |
In Waiting |
Machine | Teuthology Branch | OS Type | OS Version | Description | Nodes |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
fail | 363721 | 2016-08-14 04:20:35 | 2016-08-14 04:53:05 | 2016-08-14 07:53:10 | 3:00:05 | 2:34:25 | 0:25:40 | vps | master | centos | 7.2 | upgrade:jewel-x/parallel/{kraken.yaml 0-cluster/{openstack.yaml start.yaml} 1-jewel-install/jewel.yaml 2-workload/{blogbench.yaml ec-rados-default.yaml rados_api.yaml rados_loadgenbig.yaml test_rbd_api.yaml test_rbd_python.yaml} 3-upgrade-sequence/upgrade-all.yaml 5-final-workload/{blogbench.yaml rados-snaps-few-objects.yaml rados_loadgenmix.yaml rados_mon_thrash.yaml rbd_cls.yaml rbd_import_export.yaml rgw_swift.yaml} distros/centos_7.2.yaml} | 3 | |
Failure Reason:
Command failed (workunit test rados/test.sh) on vpm199 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/mnt.1/client.1/tmp && cd -- /home/ubuntu/cephtest/mnt.1/client.1/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=ea63a4e1841f14d76e610993bd7c1ee6c21771aa TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="1" PATH=$PATH:/usr/sbin adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/workunit.client.1/rados/test.sh' |
||||||||||||||
fail | 363722 | 2016-08-14 04:20:35 | 2016-08-14 04:55:12 | 2016-08-14 07:19:16 | 2:24:04 | 0:44:58 | 1:39:06 | vps | master | centos | 7.2 | upgrade:jewel-x/point-to-point-x/{point-to-point-upgrade.yaml distros/centos_7.2.yaml} | 3 | |
Failure Reason:
Command failed (workunit test cls/test_cls_rbd.sh) on vpm197 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/mnt.1/client.1/tmp && cd -- /home/ubuntu/cephtest/mnt.1/client.1/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=jewel TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="1" PATH=$PATH:/usr/sbin adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/workunit.client.1/cls/test_cls_rbd.sh' |
||||||||||||||
fail | 363723 | 2016-08-14 04:20:36 | 2016-08-14 04:57:04 | 2016-08-14 09:41:12 | 4:44:08 | 3:00:10 | 1:43:58 | vps | master | centos | 7.2 | upgrade:jewel-x/stress-split/{0-cluster/{openstack.yaml start.yaml} 1-jewel-install/jewel.yaml 2-partial-upgrade/firsthalf.yaml 3-thrash/default.yaml 4-mon/mona.yaml 5-workload/{rbd-cls.yaml rbd-import-export.yaml readwrite.yaml snaps-few-objects.yaml} 6-next-mon/monb.yaml 7-workload/{radosbench.yaml rbd_api.yaml} 8-next-mon/monc.yaml 9-workload/{rbd-python.yaml rgw-swift.yaml snaps-many-objects.yaml} distros/centos_7.2.yaml} | 3 | |
Failure Reason:
Command failed on vpm121 with status 33: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph osd reweight-by-utilization 110 0.05' |
||||||||||||||
pass | 363724 | 2016-08-14 04:20:37 | 2016-08-14 04:57:04 | 2016-08-14 07:35:08 | 2:38:04 | 0:56:49 | 1:41:15 | vps | master | centos | 7.2 | upgrade:jewel-x/stress-split-erasure-code/{0-cluster/{openstack.yaml start.yaml} 1-jewel-install/jewel.yaml 2-partial-upgrade/firsthalf.yaml 3-thrash/default.yaml 4-mon/mona.yaml 5-workload/ec-rados-default.yaml 6-next-mon/monb.yaml 8-next-mon/monc.yaml 9-workload/ec-rados-plugin=jerasure-k=3-m=1.yaml distros/centos_7.2.yaml} | 3 | |
pass | 363725 | 2016-08-14 04:20:38 | 2016-08-14 04:57:04 | 2016-08-14 06:21:06 | 1:24:02 | 0:53:04 | 0:30:58 | vps | master | upgrade:jewel-x/stress-split-erasure-code-x86_64/{0-x86_64.yaml 0-cluster/{openstack.yaml start.yaml} 1-jewel-install/jewel.yaml 2-partial-upgrade/firsthalf.yaml 3-thrash/default.yaml 4-mon/mona.yaml 5-workload/ec-rados-default.yaml 6-next-mon/monb.yaml 8-next-mon/monc.yaml 9-workload/ec-rados-plugin=jerasure-k=3-m=1.yaml} | 3 | |||
fail | 363726 | 2016-08-14 04:20:38 | 2016-08-14 04:57:09 | 2016-08-14 08:13:15 | 3:16:06 | 2:56:41 | 0:19:25 | vps | master | ubuntu | 14.04 | upgrade:jewel-x/parallel/{kraken.yaml 0-cluster/{openstack.yaml start.yaml} 1-jewel-install/jewel.yaml 2-workload/{blogbench.yaml ec-rados-default.yaml rados_api.yaml rados_loadgenbig.yaml test_rbd_api.yaml test_rbd_python.yaml} 3-upgrade-sequence/upgrade-mon-osd-mds.yaml 5-final-workload/{blogbench.yaml rados-snaps-few-objects.yaml rados_loadgenmix.yaml rados_mon_thrash.yaml rbd_cls.yaml rbd_import_export.yaml rgw_swift.yaml} distros/ubuntu_14.04.yaml} | 3 | |
Failure Reason:
Command failed (workunit test rados/test.sh) on vpm155 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/mnt.1/client.1/tmp && cd -- /home/ubuntu/cephtest/mnt.1/client.1/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=ea63a4e1841f14d76e610993bd7c1ee6c21771aa TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="1" PATH=$PATH:/usr/sbin adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/workunit.client.1/rados/test.sh' |
||||||||||||||
fail | 363727 | 2016-08-14 04:20:39 | 2016-08-14 04:58:57 | 2016-08-14 07:59:02 | 3:00:05 | 2:28:46 | 0:31:19 | vps | master | centos | 7.2 | upgrade:jewel-x/parallel/{kraken.yaml 0-cluster/{openstack.yaml start.yaml} 1-jewel-install/jewel.yaml 2-workload/{blogbench.yaml ec-rados-default.yaml rados_api.yaml rados_loadgenbig.yaml test_rbd_api.yaml test_rbd_python.yaml} 3-upgrade-sequence/upgrade-mon-osd-mds.yaml 5-final-workload/{blogbench.yaml rados-snaps-few-objects.yaml rados_loadgenmix.yaml rados_mon_thrash.yaml rbd_cls.yaml rbd_import_export.yaml rgw_swift.yaml} distros/centos_7.2.yaml} | 3 | |
Failure Reason:
Command failed (workunit test rados/test.sh) on vpm175 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/mnt.1/client.1/tmp && cd -- /home/ubuntu/cephtest/mnt.1/client.1/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=ea63a4e1841f14d76e610993bd7c1ee6c21771aa TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="1" PATH=$PATH:/usr/sbin adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/workunit.client.1/rados/test.sh' |
||||||||||||||
fail | 363728 | 2016-08-14 04:20:40 | 2016-08-14 05:35:09 | 2016-08-14 07:13:12 | 1:38:03 | 0:38:28 | 0:59:35 | vps | master | ubuntu | 14.04 | upgrade:jewel-x/point-to-point-x/{point-to-point-upgrade.yaml distros/ubuntu_14.04.yaml} | 3 | |
Failure Reason:
Command failed (workunit test cls/test_cls_rbd.sh) on vpm085 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/mnt.1/client.1/tmp && cd -- /home/ubuntu/cephtest/mnt.1/client.1/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=jewel TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="1" PATH=$PATH:/usr/sbin adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/workunit.client.1/cls/test_cls_rbd.sh' |
||||||||||||||
fail | 363729 | 2016-08-14 04:20:41 | 2016-08-14 05:35:12 | 2016-08-14 06:53:13 | 1:18:01 | 0:12:49 | 1:05:12 | vps | master | ubuntu | 14.04 | upgrade:jewel-x/stress-split/{0-cluster/{openstack.yaml start.yaml} 1-jewel-install/jewel.yaml 2-partial-upgrade/firsthalf.yaml 3-thrash/default.yaml 4-mon/mona.yaml 5-workload/{rbd-cls.yaml rbd-import-export.yaml readwrite.yaml snaps-few-objects.yaml} 6-next-mon/monb.yaml 7-workload/{radosbench.yaml rbd_api.yaml} 8-next-mon/monc.yaml 9-workload/{rbd-python.yaml rgw-swift.yaml snaps-many-objects.yaml} distros/ubuntu_14.04.yaml} | 3 | |
Failure Reason:
Command failed on vpm029 with status 100: u'sudo DEBIAN_FRONTEND=noninteractive apt-get -y --force-yes -o Dpkg::Options::="--force-confdef" -o Dpkg::Options::="--force-confold" install ceph=10.2.2-243-g40fc75e-1trusty ceph-mds=10.2.2-243-g40fc75e-1trusty ceph-common=10.2.2-243-g40fc75e-1trusty ceph-fuse=10.2.2-243-g40fc75e-1trusty ceph-test=10.2.2-243-g40fc75e-1trusty radosgw=10.2.2-243-g40fc75e-1trusty python-ceph=10.2.2-243-g40fc75e-1trusty libcephfs1=10.2.2-243-g40fc75e-1trusty libcephfs-dev=10.2.2-243-g40fc75e-1trusty libcephfs-java=10.2.2-243-g40fc75e-1trusty libcephfs-jni=10.2.2-243-g40fc75e-1trusty librados2=10.2.2-243-g40fc75e-1trusty librbd1=10.2.2-243-g40fc75e-1trusty rbd-fuse=10.2.2-243-g40fc75e-1trusty' |
||||||||||||||
pass | 363730 | 2016-08-14 04:20:42 | 2016-08-14 05:36:57 | 2016-08-14 07:15:00 | 1:38:03 | 0:43:59 | 0:54:04 | vps | master | ubuntu | 14.04 | upgrade:jewel-x/stress-split-erasure-code/{0-cluster/{openstack.yaml start.yaml} 1-jewel-install/jewel.yaml 2-partial-upgrade/firsthalf.yaml 3-thrash/default.yaml 4-mon/mona.yaml 5-workload/ec-rados-default.yaml 6-next-mon/monb.yaml 8-next-mon/monc.yaml 9-workload/ec-rados-plugin=jerasure-k=3-m=1.yaml distros/ubuntu_14.04.yaml} | 3 | |
fail | 363731 | 2016-08-14 04:20:43 | 2016-08-14 05:41:02 | 2016-08-14 09:09:07 | 3:28:05 | 2:25:09 | 1:02:56 | vps | master | ubuntu | 14.04 | upgrade:jewel-x/parallel/{kraken.yaml 0-cluster/{openstack.yaml start.yaml} 1-jewel-install/jewel.yaml 2-workload/{blogbench.yaml ec-rados-default.yaml rados_api.yaml rados_loadgenbig.yaml test_rbd_api.yaml test_rbd_python.yaml} 3-upgrade-sequence/upgrade-all.yaml 5-final-workload/{blogbench.yaml rados-snaps-few-objects.yaml rados_loadgenmix.yaml rados_mon_thrash.yaml rbd_cls.yaml rbd_import_export.yaml rgw_swift.yaml} distros/ubuntu_14.04.yaml} | 3 | |
Failure Reason:
Command failed (workunit test rados/test.sh) on vpm117 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/mnt.1/client.1/tmp && cd -- /home/ubuntu/cephtest/mnt.1/client.1/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=ea63a4e1841f14d76e610993bd7c1ee6c21771aa TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="1" PATH=$PATH:/usr/sbin adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/workunit.client.1/rados/test.sh' |