User | Scheduled | Started | Updated | Runtime | Suite | Branch | Machine Type | Revision | Pass | Fail | Dead |
---|---|---|---|---|---|---|---|---|---|---|---|
teuthology | 2018-04-23 02:25:03 | 2018-04-23 02:28:49 | 2018-04-23 14:49:06 | 12:20:17 | upgrade:luminous-x | master | ovh | 4f73c60 | 5 | 9 | 4 |
Status | Job ID | Links | Posted | Started | Updated | Runtime |
Duration |
In Waiting |
Machine | Teuthology Branch | OS Type | OS Version | Description | Nodes |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
fail | 2428611 | 2018-04-23 02:25:47 | 2018-04-23 02:28:40 | 2018-04-23 03:32:40 | 1:04:00 | 0:44:52 | 0:19:08 | ovh | master | centos | 7.4 | upgrade:luminous-x/parallel/{0-cluster/{openstack.yaml start.yaml} 1-ceph-install/luminous.yaml 2-workload/{blogbench.yaml ec-rados-default.yaml rados_api.yaml rados_loadgenbig.yaml test_rbd_api.yaml test_rbd_python.yaml} 3-upgrade-sequence/upgrade-all.yaml 5-final-workload/{blogbench.yaml rados-snaps-few-objects.yaml rados_loadgenmix.yaml rados_mon_thrash.yaml rbd_cls.yaml rbd_import_export_no_upgrated.yaml rbd_import_export_upgrated.yaml rgw_swift.yaml} distros/centos_latest.yaml objectstore/bluestore.yaml} | 4 | |
Failure Reason:
Command failed (workunit test cls/test_cls_sdk.sh) on ovh015 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=luminous TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/cls/test_cls_sdk.sh' |
||||||||||||||
fail | 2428613 | 2018-04-23 02:25:48 | 2018-04-23 02:28:49 | 2018-04-23 03:56:50 | 1:28:01 | 1:16:46 | 0:11:15 | ovh | master | centos | 7.4 | upgrade:luminous-x/point-to-point-x/{distros/centos_latest.yaml point-to-point-upgrade.yaml} | 3 | |
Failure Reason:
Command failed (workunit test cls/test_cls_sdk.sh) on ovh052 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/mnt.1/client.1/tmp && cd -- /home/ubuntu/cephtest/mnt.1/client.1/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=luminous TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="1" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.1 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.1 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.1/qa/workunits/cls/test_cls_sdk.sh' |
||||||||||||||
fail | 2428615 | 2018-04-23 02:25:49 | 2018-04-23 02:30:11 | 2018-04-23 03:40:11 | 1:10:00 | 1:01:14 | 0:08:46 | ovh | master | centos | 7.4 | upgrade:luminous-x/stress-split/{0-cluster/{openstack.yaml start.yaml} 1-ceph-install/luminous.yaml 2-partial-upgrade/firsthalf.yaml 3-thrash/default.yaml 4-workload/{radosbench.yaml rbd-cls.yaml rbd-import-export.yaml rbd_api.yaml readwrite.yaml snaps-few-objects.yaml} 5-finish-upgrade.yaml 7-final-workload/{rbd-python.yaml rgw-swift.yaml snaps-many-objects.yaml} distros/centos_latest.yaml objectstore/bluestore.yaml thrashosds-health.yaml} | 3 | |
Failure Reason:
Command failed (workunit test rbd/test_librbd.sh) on ovh056 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=luminous TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/rbd/test_librbd.sh' |
||||||||||||||
pass | 2428617 | 2018-04-23 02:25:50 | 2018-04-23 02:32:33 | 2018-04-23 03:58:33 | 1:26:00 | 1:14:42 | 0:11:18 | ovh | master | centos | 7.4 | upgrade:luminous-x/stress-split-erasure-code/{0-cluster/{openstack.yaml start.yaml} 1-ceph-install/luminous.yaml 2-partial-upgrade/firsthalf.yaml 3-thrash/default.yaml 4-ec-workload.yaml 5-finish-upgrade.yaml 7-final-workload.yaml distros/centos_latest.yaml objectstore/bluestore.yaml thrashosds-health.yaml} | 3 | |
fail | 2428619 | 2018-04-23 02:25:51 | 2018-04-23 02:34:40 | 2018-04-23 03:42:40 | 1:08:00 | 0:44:43 | 0:23:17 | ovh | master | ubuntu | 16.04 | upgrade:luminous-x/parallel/{0-cluster/{openstack.yaml start.yaml} 1-ceph-install/luminous.yaml 2-workload/{blogbench.yaml ec-rados-default.yaml rados_api.yaml rados_loadgenbig.yaml test_rbd_api.yaml test_rbd_python.yaml} 3-upgrade-sequence/upgrade-all.yaml 5-final-workload/{blogbench.yaml rados-snaps-few-objects.yaml rados_loadgenmix.yaml rados_mon_thrash.yaml rbd_cls.yaml rbd_import_export_no_upgrated.yaml rbd_import_export_upgrated.yaml rgw_swift.yaml} distros/ubuntu_latest.yaml objectstore/bluestore.yaml} | 4 | |
Failure Reason:
Command failed (workunit test cls/test_cls_sdk.sh) on ovh059 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=luminous TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/cls/test_cls_sdk.sh' |
||||||||||||||
fail | 2428621 | 2018-04-23 02:25:52 | 2018-04-23 02:35:22 | 2018-04-23 03:57:23 | 1:22:01 | 0:42:49 | 0:39:12 | ovh | master | centos | 7.4 | upgrade:luminous-x/parallel/{0-cluster/{openstack.yaml start.yaml} 1-ceph-install/luminous.yaml 2-workload/{blogbench.yaml ec-rados-default.yaml rados_api.yaml rados_loadgenbig.yaml test_rbd_api.yaml test_rbd_python.yaml} 3-upgrade-sequence/upgrade-mon-osd-mds.yaml 5-final-workload/{blogbench.yaml rados-snaps-few-objects.yaml rados_loadgenmix.yaml rados_mon_thrash.yaml rbd_cls.yaml rbd_import_export_no_upgrated.yaml rbd_import_export_upgrated.yaml rgw_swift.yaml} distros/centos_latest.yaml objectstore/filestore-xfs.yaml} | 4 | |
Failure Reason:
Command failed (workunit test cls/test_cls_sdk.sh) on ovh075 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=luminous TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/cls/test_cls_sdk.sh' |
||||||||||||||
pass | 2428623 | 2018-04-23 02:25:53 | 2018-04-23 02:36:32 | 2018-04-23 05:04:34 | 2:28:02 | 2:15:53 | 0:12:09 | ovh | master | ubuntu | 16.04 | upgrade:luminous-x/stress-split/{0-cluster/{openstack.yaml start.yaml} 1-ceph-install/luminous.yaml 2-partial-upgrade/firsthalf.yaml 3-thrash/default.yaml 4-workload/{radosbench.yaml rbd-cls.yaml rbd-import-export.yaml rbd_api.yaml readwrite.yaml snaps-few-objects.yaml} 5-finish-upgrade.yaml 7-final-workload/{rbd-python.yaml rgw-swift.yaml snaps-many-objects.yaml} distros/ubuntu_latest.yaml objectstore/bluestore.yaml thrashosds-health.yaml} | 3 | |
pass | 2428625 | 2018-04-23 02:25:54 | 2018-04-23 02:36:33 | 2018-04-23 03:48:34 | 1:12:01 | 0:54:24 | 0:17:37 | ovh | master | ubuntu | 16.04 | upgrade:luminous-x/stress-split-erasure-code/{0-cluster/{openstack.yaml start.yaml} 1-ceph-install/luminous.yaml 2-partial-upgrade/firsthalf.yaml 3-thrash/default.yaml 4-ec-workload.yaml 5-finish-upgrade.yaml 7-final-workload.yaml distros/ubuntu_latest.yaml objectstore/bluestore.yaml thrashosds-health.yaml} | 3 | |
fail | 2428627 | 2018-04-23 02:25:54 | 2018-04-23 02:38:22 | 2018-04-23 04:52:24 | 2:14:02 | 0:43:14 | 1:30:48 | ovh | master | ubuntu | 16.04 | upgrade:luminous-x/parallel/{0-cluster/{openstack.yaml start.yaml} 1-ceph-install/luminous.yaml 2-workload/{blogbench.yaml ec-rados-default.yaml rados_api.yaml rados_loadgenbig.yaml test_rbd_api.yaml test_rbd_python.yaml} 3-upgrade-sequence/upgrade-mon-osd-mds.yaml 5-final-workload/{blogbench.yaml rados-snaps-few-objects.yaml rados_loadgenmix.yaml rados_mon_thrash.yaml rbd_cls.yaml rbd_import_export_no_upgrated.yaml rbd_import_export_upgrated.yaml rgw_swift.yaml} distros/ubuntu_latest.yaml objectstore/filestore-xfs.yaml} | 4 | |
Failure Reason:
Command failed (workunit test cls/test_cls_sdk.sh) on ovh032 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=luminous TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/cls/test_cls_sdk.sh' |
||||||||||||||
dead | 2428629 | 2018-04-23 02:25:55 | 2018-04-23 02:39:44 | 2018-04-23 14:42:18 | 12:02:34 | ovh | master | centos | 7.4 | upgrade:luminous-x/parallel/{0-cluster/{openstack.yaml start.yaml} 1-ceph-install/luminous.yaml 2-workload/{blogbench.yaml ec-rados-default.yaml rados_api.yaml rados_loadgenbig.yaml test_rbd_api.yaml test_rbd_python.yaml} 3-upgrade-sequence/upgrade-mon-osd-mds.yaml 5-final-workload/{blogbench.yaml rados-snaps-few-objects.yaml rados_loadgenmix.yaml rados_mon_thrash.yaml rbd_cls.yaml rbd_import_export_no_upgrated.yaml rbd_import_export_upgrated.yaml rgw_swift.yaml} distros/centos_latest.yaml objectstore/bluestore.yaml} | 4 | |||
fail | 2428631 | 2018-04-23 02:25:56 | 2018-04-23 02:40:33 | 2018-04-23 04:04:34 | 1:24:01 | 1:13:56 | 0:10:05 | ovh | master | ubuntu | 16.04 | upgrade:luminous-x/point-to-point-x/{distros/ubuntu_latest.yaml point-to-point-upgrade.yaml} | 3 | |
Failure Reason:
Command failed (workunit test cls/test_cls_sdk.sh) on ovh082 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/mnt.1/client.1/tmp && cd -- /home/ubuntu/cephtest/mnt.1/client.1/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=luminous TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="1" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.1 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.1 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.1/qa/workunits/cls/test_cls_sdk.sh' |
||||||||||||||
pass | 2428633 | 2018-04-23 02:25:57 | 2018-04-23 02:42:24 | 2018-04-23 05:04:26 | 2:22:02 | 2:11:00 | 0:11:02 | ovh | master | centos | 7.4 | upgrade:luminous-x/stress-split/{0-cluster/{openstack.yaml start.yaml} 1-ceph-install/luminous.yaml 2-partial-upgrade/firsthalf.yaml 3-thrash/default.yaml 4-workload/{radosbench.yaml rbd-cls.yaml rbd-import-export.yaml rbd_api.yaml readwrite.yaml snaps-few-objects.yaml} 5-finish-upgrade.yaml 7-final-workload/{rbd-python.yaml rgw-swift.yaml snaps-many-objects.yaml} distros/centos_latest.yaml objectstore/filestore-xfs.yaml thrashosds-health.yaml} | 3 | |
dead | 2428635 | 2018-04-23 02:25:58 | 2018-04-23 02:44:33 | 2018-04-23 14:47:08 | 12:02:35 | ovh | master | centos | 7.4 | upgrade:luminous-x/stress-split-erasure-code/{0-cluster/{openstack.yaml start.yaml} 1-ceph-install/luminous.yaml 2-partial-upgrade/firsthalf.yaml 3-thrash/default.yaml 4-ec-workload.yaml 5-finish-upgrade.yaml 7-final-workload.yaml distros/centos_latest.yaml objectstore/filestore-xfs.yaml thrashosds-health.yaml} | 3 | |||
dead | 2428637 | 2018-04-23 02:25:59 | 2018-04-23 02:44:38 | 2018-04-23 14:47:13 | 12:02:35 | ovh | master | ubuntu | 16.04 | upgrade:luminous-x/parallel/{0-cluster/{openstack.yaml start.yaml} 1-ceph-install/luminous.yaml 2-workload/{blogbench.yaml ec-rados-default.yaml rados_api.yaml rados_loadgenbig.yaml test_rbd_api.yaml test_rbd_python.yaml} 3-upgrade-sequence/upgrade-mon-osd-mds.yaml 5-final-workload/{blogbench.yaml rados-snaps-few-objects.yaml rados_loadgenmix.yaml rados_mon_thrash.yaml rbd_cls.yaml rbd_import_export_no_upgrated.yaml rbd_import_export_upgrated.yaml rgw_swift.yaml} distros/ubuntu_latest.yaml objectstore/bluestore.yaml} | 4 | |||
fail | 2428639 | 2018-04-23 02:26:00 | 2018-04-23 02:44:59 | 2018-04-23 04:07:00 | 1:22:01 | 0:45:24 | 0:36:37 | ovh | master | centos | 7.4 | upgrade:luminous-x/parallel/{0-cluster/{openstack.yaml start.yaml} 1-ceph-install/luminous.yaml 2-workload/{blogbench.yaml ec-rados-default.yaml rados_api.yaml rados_loadgenbig.yaml test_rbd_api.yaml test_rbd_python.yaml} 3-upgrade-sequence/upgrade-all.yaml 5-final-workload/{blogbench.yaml rados-snaps-few-objects.yaml rados_loadgenmix.yaml rados_mon_thrash.yaml rbd_cls.yaml rbd_import_export_no_upgrated.yaml rbd_import_export_upgrated.yaml rgw_swift.yaml} distros/centos_latest.yaml objectstore/filestore-xfs.yaml} | 4 | |
Failure Reason:
Command failed (workunit test cls/test_cls_sdk.sh) on ovh024 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=luminous TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/cls/test_cls_sdk.sh' |
||||||||||||||
pass | 2428641 | 2018-04-23 02:26:01 | 2018-04-23 02:46:28 | 2018-04-23 05:50:31 | 3:04:03 | 2:10:05 | 0:53:58 | ovh | master | ubuntu | 16.04 | upgrade:luminous-x/stress-split/{0-cluster/{openstack.yaml start.yaml} 1-ceph-install/luminous.yaml 2-partial-upgrade/firsthalf.yaml 3-thrash/default.yaml 4-workload/{radosbench.yaml rbd-cls.yaml rbd-import-export.yaml rbd_api.yaml readwrite.yaml snaps-few-objects.yaml} 5-finish-upgrade.yaml 7-final-workload/{rbd-python.yaml rgw-swift.yaml snaps-many-objects.yaml} distros/ubuntu_latest.yaml objectstore/filestore-xfs.yaml thrashosds-health.yaml} | 3 | |
dead | 2428643 | 2018-04-23 02:26:02 | 2018-04-23 02:46:28 | 2018-04-23 14:49:06 | 12:02:38 | ovh | master | ubuntu | 16.04 | upgrade:luminous-x/stress-split-erasure-code/{0-cluster/{openstack.yaml start.yaml} 1-ceph-install/luminous.yaml 2-partial-upgrade/firsthalf.yaml 3-thrash/default.yaml 4-ec-workload.yaml 5-finish-upgrade.yaml 7-final-workload.yaml distros/ubuntu_latest.yaml objectstore/filestore-xfs.yaml thrashosds-health.yaml} | 3 | |||
fail | 2428645 | 2018-04-23 02:26:03 | 2018-04-23 02:46:28 | 2018-04-23 05:30:31 | 2:44:03 | 0:44:03 | 2:00:00 | ovh | master | ubuntu | 16.04 | upgrade:luminous-x/parallel/{0-cluster/{openstack.yaml start.yaml} 1-ceph-install/luminous.yaml 2-workload/{blogbench.yaml ec-rados-default.yaml rados_api.yaml rados_loadgenbig.yaml test_rbd_api.yaml test_rbd_python.yaml} 3-upgrade-sequence/upgrade-all.yaml 5-final-workload/{blogbench.yaml rados-snaps-few-objects.yaml rados_loadgenmix.yaml rados_mon_thrash.yaml rbd_cls.yaml rbd_import_export_no_upgrated.yaml rbd_import_export_upgrated.yaml rgw_swift.yaml} distros/ubuntu_latest.yaml objectstore/filestore-xfs.yaml} | 4 | |
Failure Reason:
Command failed (workunit test cls/test_cls_sdk.sh) on ovh061 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=luminous TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/cls/test_cls_sdk.sh' |