User | Scheduled | Started | Updated | Runtime | Suite | Branch | Machine Type | Revision | Pass | Fail | Dead |
---|---|---|---|---|---|---|---|---|---|---|---|
teuthology | 2018-05-03 02:25:02 | 2018-05-03 02:42:20 | 2018-05-03 14:54:51 | 12:12:31 | upgrade:luminous-x | master | ovh | 135fe40 | 7 | 9 | 2 |
Status | Job ID | Links | Posted | Started | Updated | Runtime |
Duration |
In Waiting |
Machine | Teuthology Branch | OS Type | OS Version | Description | Nodes |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
fail | 2469852 | 2018-05-03 02:25:49 | 2018-05-03 02:30:09 | 2018-05-03 05:02:11 | 2:32:02 | 2:06:11 | 0:25:51 | ovh | master | centos | 7.4 | upgrade:luminous-x/parallel/{0-cluster/{openstack.yaml start.yaml} 1-ceph-install/luminous.yaml 2-workload/{blogbench.yaml ec-rados-default.yaml rados_api.yaml rados_loadgenbig.yaml test_rbd_api.yaml test_rbd_python.yaml} 3-upgrade-sequence/upgrade-all.yaml 5-final-workload/{blogbench.yaml rados-snaps-few-objects.yaml rados_loadgenmix.yaml rados_mon_thrash.yaml rbd_cls.yaml rbd_import_export_no_upgrated.yaml rbd_import_export_upgrated.yaml rgw_swift.yaml} distros/centos_latest.yaml objectstore/bluestore.yaml} | 4 | |
Failure Reason:
Command failed (workunit test rbd/import_export.sh) on ovh006 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/mnt.1/client.1/tmp && cd -- /home/ubuntu/cephtest/mnt.1/client.1/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=luminous TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="1" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.1 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.1 RBD_CREATE_ARGS=--new-format adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.1/qa/workunits/rbd/import_export.sh' |
||||||||||||||
fail | 2469853 | 2018-05-03 02:25:50 | 2018-05-03 02:42:20 | 2018-05-03 04:12:21 | 1:30:01 | 1:18:24 | 0:11:37 | ovh | master | centos | 7.4 | upgrade:luminous-x/point-to-point-x/{distros/centos_latest.yaml point-to-point-upgrade.yaml} | 3 | |
Failure Reason:
'wait_until_healthy' reached maximum tries (150) after waiting for 900 seconds |
||||||||||||||
pass | 2469855 | 2018-05-03 02:25:51 | 2018-05-03 02:44:14 | 2018-05-03 05:16:16 | 2:32:02 | 2:09:09 | 0:22:53 | ovh | master | centos | 7.4 | upgrade:luminous-x/stress-split/{0-cluster/{openstack.yaml start.yaml} 1-ceph-install/luminous.yaml 2-partial-upgrade/firsthalf.yaml 3-thrash/default.yaml 4-workload/{radosbench.yaml rbd-cls.yaml rbd-import-export.yaml rbd_api.yaml readwrite.yaml snaps-few-objects.yaml} 5-finish-upgrade.yaml 7-final-workload/{rbd-python.yaml rgw-swift.yaml snaps-many-objects.yaml} distros/centos_latest.yaml objectstore/bluestore.yaml thrashosds-health.yaml} | 3 | |
dead | 2469857 | 2018-05-03 02:25:52 | 2018-05-03 02:46:00 | 2018-05-03 14:48:40 | 12:02:40 | ovh | master | centos | 7.4 | upgrade:luminous-x/stress-split-erasure-code/{0-cluster/{openstack.yaml start.yaml} 1-ceph-install/luminous.yaml 2-partial-upgrade/firsthalf.yaml 3-thrash/default.yaml 4-ec-workload.yaml 5-finish-upgrade.yaml 7-final-workload.yaml distros/centos_latest.yaml objectstore/bluestore.yaml thrashosds-health.yaml} | 3 | |||
dead | 2469859 | 2018-05-03 02:25:52 | 2018-05-03 02:52:12 | 2018-05-03 14:54:51 | 12:02:39 | ovh | master | ubuntu | 16.04 | upgrade:luminous-x/parallel/{0-cluster/{openstack.yaml start.yaml} 1-ceph-install/luminous.yaml 2-workload/{blogbench.yaml ec-rados-default.yaml rados_api.yaml rados_loadgenbig.yaml test_rbd_api.yaml test_rbd_python.yaml} 3-upgrade-sequence/upgrade-all.yaml 5-final-workload/{blogbench.yaml rados-snaps-few-objects.yaml rados_loadgenmix.yaml rados_mon_thrash.yaml rbd_cls.yaml rbd_import_export_no_upgrated.yaml rbd_import_export_upgrated.yaml rgw_swift.yaml} distros/ubuntu_latest.yaml objectstore/bluestore.yaml} | 4 | |||
fail | 2469861 | 2018-05-03 02:25:53 | 2018-05-03 02:55:28 | 2018-05-03 06:09:32 | 3:14:04 | 2:11:09 | 1:02:55 | ovh | master | centos | 7.4 | upgrade:luminous-x/parallel/{0-cluster/{openstack.yaml start.yaml} 1-ceph-install/luminous.yaml 2-workload/{blogbench.yaml ec-rados-default.yaml rados_api.yaml rados_loadgenbig.yaml test_rbd_api.yaml test_rbd_python.yaml} 3-upgrade-sequence/upgrade-mon-osd-mds.yaml 5-final-workload/{blogbench.yaml rados-snaps-few-objects.yaml rados_loadgenmix.yaml rados_mon_thrash.yaml rbd_cls.yaml rbd_import_export_no_upgrated.yaml rbd_import_export_upgrated.yaml rgw_swift.yaml} distros/centos_latest.yaml objectstore/filestore-xfs.yaml} | 4 | |
Failure Reason:
Command failed (workunit test rbd/import_export.sh) on ovh015 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/mnt.1/client.1/tmp && cd -- /home/ubuntu/cephtest/mnt.1/client.1/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=luminous TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="1" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.1 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.1 RBD_CREATE_ARGS=--new-format adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.1/qa/workunits/rbd/import_export.sh' |
||||||||||||||
pass | 2469863 | 2018-05-03 02:25:54 | 2018-05-03 02:58:37 | 2018-05-03 05:58:40 | 3:00:03 | 2:04:37 | 0:55:26 | ovh | master | ubuntu | 16.04 | upgrade:luminous-x/stress-split/{0-cluster/{openstack.yaml start.yaml} 1-ceph-install/luminous.yaml 2-partial-upgrade/firsthalf.yaml 3-thrash/default.yaml 4-workload/{radosbench.yaml rbd-cls.yaml rbd-import-export.yaml rbd_api.yaml readwrite.yaml snaps-few-objects.yaml} 5-finish-upgrade.yaml 7-final-workload/{rbd-python.yaml rgw-swift.yaml snaps-many-objects.yaml} distros/ubuntu_latest.yaml objectstore/bluestore.yaml thrashosds-health.yaml} | 3 | |
pass | 2469865 | 2018-05-03 02:25:55 | 2018-05-03 03:00:20 | 2018-05-03 04:28:21 | 1:28:01 | 0:59:38 | 0:28:23 | ovh | master | ubuntu | 16.04 | upgrade:luminous-x/stress-split-erasure-code/{0-cluster/{openstack.yaml start.yaml} 1-ceph-install/luminous.yaml 2-partial-upgrade/firsthalf.yaml 3-thrash/default.yaml 4-ec-workload.yaml 5-finish-upgrade.yaml 7-final-workload.yaml distros/ubuntu_latest.yaml objectstore/bluestore.yaml thrashosds-health.yaml} | 3 | |
fail | 2469868 | 2018-05-03 02:25:56 | 2018-05-03 03:00:20 | 2018-05-03 07:18:25 | 4:18:05 | 2:04:40 | 2:13:25 | ovh | master | ubuntu | 16.04 | upgrade:luminous-x/parallel/{0-cluster/{openstack.yaml start.yaml} 1-ceph-install/luminous.yaml 2-workload/{blogbench.yaml ec-rados-default.yaml rados_api.yaml rados_loadgenbig.yaml test_rbd_api.yaml test_rbd_python.yaml} 3-upgrade-sequence/upgrade-mon-osd-mds.yaml 5-final-workload/{blogbench.yaml rados-snaps-few-objects.yaml rados_loadgenmix.yaml rados_mon_thrash.yaml rbd_cls.yaml rbd_import_export_no_upgrated.yaml rbd_import_export_upgrated.yaml rgw_swift.yaml} distros/ubuntu_latest.yaml objectstore/filestore-xfs.yaml} | 4 | |
Failure Reason:
Command failed (workunit test rbd/import_export.sh) on ovh006 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/mnt.1/client.1/tmp && cd -- /home/ubuntu/cephtest/mnt.1/client.1/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=luminous TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="1" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.1 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.1 RBD_CREATE_ARGS=--new-format adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.1/qa/workunits/rbd/import_export.sh' |
||||||||||||||
fail | 2469870 | 2018-05-03 02:25:56 | 2018-05-03 03:04:47 | 2018-05-03 08:52:53 | 5:48:06 | 2:11:38 | 3:36:28 | ovh | master | centos | 7.4 | upgrade:luminous-x/parallel/{0-cluster/{openstack.yaml start.yaml} 1-ceph-install/luminous.yaml 2-workload/{blogbench.yaml ec-rados-default.yaml rados_api.yaml rados_loadgenbig.yaml test_rbd_api.yaml test_rbd_python.yaml} 3-upgrade-sequence/upgrade-mon-osd-mds.yaml 5-final-workload/{blogbench.yaml rados-snaps-few-objects.yaml rados_loadgenmix.yaml rados_mon_thrash.yaml rbd_cls.yaml rbd_import_export_no_upgrated.yaml rbd_import_export_upgrated.yaml rgw_swift.yaml} distros/centos_latest.yaml objectstore/bluestore.yaml} | 4 | |
Failure Reason:
Command failed (workunit test rbd/import_export.sh) on ovh042 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/mnt.1/client.1/tmp && cd -- /home/ubuntu/cephtest/mnt.1/client.1/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=luminous TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="1" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.1 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.1 RBD_CREATE_ARGS=--new-format adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.1/qa/workunits/rbd/import_export.sh' |
||||||||||||||
fail | 2469872 | 2018-05-03 02:25:57 | 2018-05-03 03:11:47 | 2018-05-03 05:17:49 | 2:06:02 | 1:48:53 | 0:17:09 | ovh | master | ubuntu | 16.04 | upgrade:luminous-x/point-to-point-x/{distros/ubuntu_latest.yaml point-to-point-upgrade.yaml} | 3 | |
Failure Reason:
"2018-05-03 03:47:29.074717 mon.a mon.0 158.69.86.3:6789/0 235 : cluster [WRN] Health check failed: 1 filesystem is degraded (FS_DEGRADED)" in cluster log |
||||||||||||||
pass | 2469874 | 2018-05-03 02:25:58 | 2018-05-03 03:18:26 | 2018-05-03 05:34:28 | 2:16:02 | 2:01:53 | 0:14:09 | ovh | master | centos | 7.4 | upgrade:luminous-x/stress-split/{0-cluster/{openstack.yaml start.yaml} 1-ceph-install/luminous.yaml 2-partial-upgrade/firsthalf.yaml 3-thrash/default.yaml 4-workload/{radosbench.yaml rbd-cls.yaml rbd-import-export.yaml rbd_api.yaml readwrite.yaml snaps-few-objects.yaml} 5-finish-upgrade.yaml 7-final-workload/{rbd-python.yaml rgw-swift.yaml snaps-many-objects.yaml} distros/centos_latest.yaml objectstore/filestore-xfs.yaml thrashosds-health.yaml} | 3 | |
pass | 2469876 | 2018-05-03 02:25:59 | 2018-05-03 03:19:57 | 2018-05-03 04:45:59 | 1:26:02 | 0:48:17 | 0:37:45 | ovh | master | centos | 7.4 | upgrade:luminous-x/stress-split-erasure-code/{0-cluster/{openstack.yaml start.yaml} 1-ceph-install/luminous.yaml 2-partial-upgrade/firsthalf.yaml 3-thrash/default.yaml 4-ec-workload.yaml 5-finish-upgrade.yaml 7-final-workload.yaml distros/centos_latest.yaml objectstore/filestore-xfs.yaml thrashosds-health.yaml} | 3 | |
fail | 2469878 | 2018-05-03 02:26:00 | 2018-05-03 03:22:40 | 2018-05-03 09:34:48 | 6:12:08 | 2:04:53 | 4:07:15 | ovh | master | ubuntu | 16.04 | upgrade:luminous-x/parallel/{0-cluster/{openstack.yaml start.yaml} 1-ceph-install/luminous.yaml 2-workload/{blogbench.yaml ec-rados-default.yaml rados_api.yaml rados_loadgenbig.yaml test_rbd_api.yaml test_rbd_python.yaml} 3-upgrade-sequence/upgrade-mon-osd-mds.yaml 5-final-workload/{blogbench.yaml rados-snaps-few-objects.yaml rados_loadgenmix.yaml rados_mon_thrash.yaml rbd_cls.yaml rbd_import_export_no_upgrated.yaml rbd_import_export_upgrated.yaml rgw_swift.yaml} distros/ubuntu_latest.yaml objectstore/bluestore.yaml} | 4 | |
Failure Reason:
Command failed (workunit test rbd/import_export.sh) on ovh006 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/mnt.1/client.1/tmp && cd -- /home/ubuntu/cephtest/mnt.1/client.1/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=luminous TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="1" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.1 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.1 RBD_CREATE_ARGS=--new-format adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.1/qa/workunits/rbd/import_export.sh' |
||||||||||||||
fail | 2469880 | 2018-05-03 02:26:00 | 2018-05-03 03:46:30 | 2018-05-03 06:34:32 | 2:48:02 | 2:09:26 | 0:38:36 | ovh | master | centos | 7.4 | upgrade:luminous-x/parallel/{0-cluster/{openstack.yaml start.yaml} 1-ceph-install/luminous.yaml 2-workload/{blogbench.yaml ec-rados-default.yaml rados_api.yaml rados_loadgenbig.yaml test_rbd_api.yaml test_rbd_python.yaml} 3-upgrade-sequence/upgrade-all.yaml 5-final-workload/{blogbench.yaml rados-snaps-few-objects.yaml rados_loadgenmix.yaml rados_mon_thrash.yaml rbd_cls.yaml rbd_import_export_no_upgrated.yaml rbd_import_export_upgrated.yaml rgw_swift.yaml} distros/centos_latest.yaml objectstore/filestore-xfs.yaml} | 4 | |
Failure Reason:
Command failed (workunit test rbd/import_export.sh) on ovh058 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/mnt.1/client.1/tmp && cd -- /home/ubuntu/cephtest/mnt.1/client.1/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=luminous TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="1" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.1 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.1 RBD_CREATE_ARGS=--new-format adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.1/qa/workunits/rbd/import_export.sh' |
||||||||||||||
pass | 2469882 | 2018-05-03 02:26:01 | 2018-05-03 03:46:30 | 2018-05-03 06:06:32 | 2:20:02 | 2:00:04 | 0:19:58 | ovh | master | ubuntu | 16.04 | upgrade:luminous-x/stress-split/{0-cluster/{openstack.yaml start.yaml} 1-ceph-install/luminous.yaml 2-partial-upgrade/firsthalf.yaml 3-thrash/default.yaml 4-workload/{radosbench.yaml rbd-cls.yaml rbd-import-export.yaml rbd_api.yaml readwrite.yaml snaps-few-objects.yaml} 5-finish-upgrade.yaml 7-final-workload/{rbd-python.yaml rgw-swift.yaml snaps-many-objects.yaml} distros/ubuntu_latest.yaml objectstore/filestore-xfs.yaml thrashosds-health.yaml} | 3 | |
pass | 2469884 | 2018-05-03 02:26:02 | 2018-05-03 03:52:01 | 2018-05-03 04:54:02 | 1:02:01 | 0:51:07 | 0:10:54 | ovh | master | ubuntu | 16.04 | upgrade:luminous-x/stress-split-erasure-code/{0-cluster/{openstack.yaml start.yaml} 1-ceph-install/luminous.yaml 2-partial-upgrade/firsthalf.yaml 3-thrash/default.yaml 4-ec-workload.yaml 5-finish-upgrade.yaml 7-final-workload.yaml distros/ubuntu_latest.yaml objectstore/filestore-xfs.yaml thrashosds-health.yaml} | 3 | |
fail | 2469886 | 2018-05-03 02:26:03 | 2018-05-03 03:54:10 | 2018-05-03 08:24:16 | 4:30:06 | 2:04:51 | 2:25:15 | ovh | master | ubuntu | 16.04 | upgrade:luminous-x/parallel/{0-cluster/{openstack.yaml start.yaml} 1-ceph-install/luminous.yaml 2-workload/{blogbench.yaml ec-rados-default.yaml rados_api.yaml rados_loadgenbig.yaml test_rbd_api.yaml test_rbd_python.yaml} 3-upgrade-sequence/upgrade-all.yaml 5-final-workload/{blogbench.yaml rados-snaps-few-objects.yaml rados_loadgenmix.yaml rados_mon_thrash.yaml rbd_cls.yaml rbd_import_export_no_upgrated.yaml rbd_import_export_upgrated.yaml rgw_swift.yaml} distros/ubuntu_latest.yaml objectstore/filestore-xfs.yaml} | 4 | |
Failure Reason:
Command failed (workunit test rbd/import_export.sh) on ovh015 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/mnt.1/client.1/tmp && cd -- /home/ubuntu/cephtest/mnt.1/client.1/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=luminous TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="1" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.1 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.1 RBD_CREATE_ARGS=--new-format adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.1/qa/workunits/rbd/import_export.sh' |