Status Job ID Links Posted Started Updated
Runtime
Duration
In Waiting
Machine Teuthology Branch OS Type OS Version Description Nodes
fail 3442073 2019-01-10 07:00:24 2019-01-10 07:07:31 2019-01-10 07:23:30 0:15:59 0:02:37 0:13:22 ovh master ubuntu 16.04 smoke/1node/{clusters/{fixed-1.yaml openstack.yaml} distros/ubuntu_latest.yaml objectstore/filestore-xfs.yaml tasks/ceph-deploy.yaml} 1
Failure Reason:

{'ovh057.front.sepia.ceph.com': {'_ansible_parsed': True, 'invocation': {'module_args': {'comment': None, 'ssh_key_bits': 0, 'update_password': 'always', 'non_unique': False, 'force': False, 'skeleton': None, 'expires': None, 'ssh_key_passphrase': None, 'groups': ['fuse', 'kvm', 'disk'], 'createhome': True, 'home': None, 'move_home': False, 'password': None, 'generate_ssh_key': None, 'append': True, 'uid': None, 'ssh_key_comment': 'ansible-generated on ovh057', 'group': None, 'name': 'ubuntu', 'local': None, 'seuser': None, 'system': False, 'remove': False, 'state': 'present', 'ssh_key_file': None, 'login_class': None, 'shell': None, 'ssh_key_type': 'rsa'}}, 'changed': False, '_ansible_no_log': False, 'msg': 'Group kvm does not exist'}}

fail 3442074 2019-01-10 07:00:25 2019-01-10 07:17:35 2019-01-10 08:25:36 1:08:01 0:07:36 1:00:25 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/cfuse_workunit_suites_blogbench.yaml} 3
Failure Reason:

Command failed on ovh030 with status 1: '\n sudo yum -y install ceph\n '

fail 3442075 2019-01-10 07:00:26 2019-01-10 07:17:54 2019-01-10 11:17:57 4:00:03 0:20:23 3:39:40 ovh master centos 7.5 smoke/systemd/{clusters/{fixed-4.yaml openstack.yaml} distros/centos_latest.yaml objectstore/filestore-xfs.yaml tasks/systemd.yaml} 4
Failure Reason:

ceph-deploy: Failed during gather keys

fail 3442076 2019-01-10 07:00:26 2019-01-10 07:21:16 2019-01-10 08:13:16 0:52:00 0:08:06 0:43:54 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/cfuse_workunit_suites_fsstress.yaml} 3
Failure Reason:

Command failed on ovh083 with status 1: '\n sudo yum -y install ceph\n '

fail 3442077 2019-01-10 07:00:27 2019-01-10 07:22:46 2019-01-10 08:26:46 1:04:00 0:07:41 0:56:19 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/cfuse_workunit_suites_iozone.yaml} 3
Failure Reason:

Command failed on ovh063 with status 1: '\n sudo yum -y install ceph\n '

fail 3442078 2019-01-10 07:00:28 2019-01-10 07:23:31 2019-01-10 08:19:31 0:56:00 0:07:29 0:48:31 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/cfuse_workunit_suites_pjd.yaml} 3
Failure Reason:

Command failed on ovh057 with status 1: '\n sudo yum -y install ceph\n '

fail 3442079 2019-01-10 07:00:28 2019-01-10 07:23:34 2019-01-10 08:17:34 0:54:00 0:07:53 0:46:07 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/kclient_workunit_direct_io.yaml} 3
Failure Reason:

Command failed on ovh058 with status 1: '\n sudo yum -y install ceph\n '

fail 3442080 2019-01-10 07:00:29 2019-01-10 07:25:29 2019-01-10 08:45:30 1:20:01 0:07:29 1:12:32 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/kclient_workunit_suites_dbench.yaml} 3
Failure Reason:

Command failed on ovh093 with status 1: '\n sudo yum -y install ceph\n '

fail 3442081 2019-01-10 07:00:30 2019-01-10 07:25:33 2019-01-10 08:19:33 0:54:00 0:07:58 0:46:02 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/kclient_workunit_suites_fsstress.yaml} 3
Failure Reason:

Command failed on ovh026 with status 1: '\n sudo yum -y install ceph\n '

fail 3442082 2019-01-10 07:00:30 2019-01-10 07:27:06 2019-01-10 08:25:06 0:58:00 0:07:56 0:50:04 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/kclient_workunit_suites_pjd.yaml} 3
Failure Reason:

Command failed on ovh029 with status 1: '\n sudo yum -y install ceph\n '

fail 3442083 2019-01-10 07:00:31 2019-01-10 07:27:46 2019-01-10 08:23:46 0:56:00 0:07:48 0:48:12 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/libcephfs_interface_tests.yaml} 3
Failure Reason:

Command failed on ovh075 with status 1: '\n sudo yum -y install ceph\n '

fail 3442084 2019-01-10 07:00:32 2019-01-10 07:29:42 2019-01-10 08:27:42 0:58:00 0:07:40 0:50:20 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/mon_thrash.yaml} 3
Failure Reason:

Command failed on ovh081 with status 1: '\n sudo yum -y install ceph\n '

fail 3442085 2019-01-10 07:00:33 2019-01-10 07:29:52 2019-01-10 08:39:52 1:10:00 0:07:51 1:02:09 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rados_api_tests.yaml} 3
Failure Reason:

Command failed on ovh016 with status 1: '\n sudo yum -y install ceph\n '

fail 3442086 2019-01-10 07:00:33 2019-01-10 07:31:24 2019-01-10 08:35:24 1:04:00 0:07:48 0:56:12 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rados_bench.yaml} 3
Failure Reason:

Command failed on ovh045 with status 1: '\n sudo yum -y install ceph\n '

fail 3442087 2019-01-10 07:00:34 2019-01-10 07:31:48 2019-01-10 08:29:48 0:58:00 0:07:32 0:50:28 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rados_cache_snaps.yaml} 3
Failure Reason:

Command failed on ovh036 with status 1: '\n sudo yum -y install ceph\n '

fail 3442088 2019-01-10 07:00:35 2019-01-10 07:36:01 2019-01-10 11:44:04 4:08:03 0:02:42 4:05:21 ovh master ubuntu 16.04 smoke/systemd/{clusters/{fixed-4.yaml openstack.yaml} distros/ubuntu_latest.yaml objectstore/filestore-xfs.yaml tasks/systemd.yaml} 4
Failure Reason:

{'ovh097.front.sepia.ceph.com': {'_ansible_parsed': True, 'invocation': {'module_args': {'comment': None, 'ssh_key_bits': 0, 'update_password': 'always', 'non_unique': False, 'force': False, 'skeleton': None, 'expires': None, 'ssh_key_passphrase': None, 'groups': ['fuse', 'kvm', 'disk'], 'createhome': True, 'home': None, 'move_home': False, 'password': None, 'generate_ssh_key': None, 'append': True, 'uid': None, 'ssh_key_comment': 'ansible-generated on ovh097', 'group': None, 'name': 'ubuntu', 'local': None, 'seuser': None, 'system': False, 'remove': False, 'state': 'present', 'ssh_key_file': None, 'login_class': None, 'shell': None, 'ssh_key_type': 'rsa'}}, 'changed': False, '_ansible_no_log': False, 'msg': 'Group kvm does not exist'}, 'ovh072.front.sepia.ceph.com': {'_ansible_parsed': True, 'invocation': {'module_args': {'comment': None, 'ssh_key_bits': 0, 'update_password': 'always', 'non_unique': False, 'force': False, 'skeleton': None, 'expires': None, 'ssh_key_passphrase': None, 'groups': ['fuse', 'kvm', 'disk'], 'createhome': True, 'home': None, 'move_home': False, 'password': None, 'generate_ssh_key': None, 'append': True, 'uid': None, 'ssh_key_comment': 'ansible-generated on ovh072', 'group': None, 'name': 'ubuntu', 'local': None, 'seuser': None, 'system': False, 'remove': False, 'state': 'present', 'ssh_key_file': None, 'login_class': None, 'shell': None, 'ssh_key_type': 'rsa'}}, 'changed': False, '_ansible_no_log': False, 'msg': 'Group kvm does not exist'}, 'ovh073.front.sepia.ceph.com': {'_ansible_parsed': True, 'invocation': {'module_args': {'comment': None, 'ssh_key_bits': 0, 'update_password': 'always', 'non_unique': False, 'force': False, 'skeleton': None, 'expires': None, 'ssh_key_passphrase': None, 'groups': ['fuse', 'kvm', 'disk'], 'createhome': True, 'home': None, 'move_home': False, 'password': None, 'generate_ssh_key': None, 'append': True, 'uid': None, 'ssh_key_comment': 'ansible-generated on ovh073', 'group': None, 'name': 'ubuntu', 'local': None, 'seuser': None, 'system': False, 'remove': False, 'state': 'present', 'ssh_key_file': None, 'login_class': None, 'shell': None, 'ssh_key_type': 'rsa'}}, 'changed': False, '_ansible_no_log': False, 'msg': 'Group kvm does not exist'}, 'ovh100.front.sepia.ceph.com': {'_ansible_parsed': True, 'invocation': {'module_args': {'comment': None, 'ssh_key_bits': 0, 'update_password': 'always', 'non_unique': False, 'force': False, 'skeleton': None, 'expires': None, 'ssh_key_passphrase': None, 'groups': ['fuse', 'kvm', 'disk'], 'createhome': True, 'home': None, 'move_home': False, 'password': None, 'generate_ssh_key': None, 'append': True, 'uid': None, 'ssh_key_comment': 'ansible-generated on ovh100', 'group': None, 'name': 'ubuntu', 'local': None, 'seuser': None, 'system': False, 'remove': False, 'state': 'present', 'ssh_key_file': None, 'login_class': None, 'shell': None, 'ssh_key_type': 'rsa'}}, 'changed': False, '_ansible_no_log': False, 'msg': 'Group kvm does not exist'}}

fail 3442089 2019-01-10 07:00:35 2019-01-10 07:39:29 2019-01-10 08:47:29 1:08:00 0:08:08 0:59:52 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rados_cls_all.yaml} 3
Failure Reason:

Command failed on ovh072 with status 1: '\n sudo yum -y install ceph\n '

fail 3442090 2019-01-10 07:00:36 2019-01-10 07:39:29 2019-01-10 08:59:29 1:20:00 0:07:49 1:12:11 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rados_ec_snaps.yaml} 3
Failure Reason:

Command failed on ovh009 with status 1: '\n sudo yum -y install ceph\n '

fail 3442091 2019-01-10 07:00:37 2019-01-10 07:43:36 2019-01-10 08:47:36 1:04:00 0:07:29 0:56:31 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rados_python.yaml} 3
Failure Reason:

Command failed on ovh056 with status 1: '\n sudo yum -y install ceph\n '

fail 3442092 2019-01-10 07:00:37 2019-01-10 07:46:17 2019-01-10 09:00:17 1:14:00 0:07:56 1:06:04 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rados_workunit_loadgen_mix.yaml} 3
Failure Reason:

Command failed on ovh092 with status 1: '\n sudo yum -y install ceph\n '

fail 3442093 2019-01-10 07:00:38 2019-01-10 07:47:33 2019-01-10 09:03:34 1:16:01 0:07:49 1:08:12 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rbd_api_tests.yaml} 3
Failure Reason:

Command failed on ovh049 with status 1: '\n sudo yum -y install ceph\n '

fail 3442094 2019-01-10 07:00:39 2019-01-10 07:47:48 2019-01-10 09:17:49 1:30:01 0:07:57 1:22:04 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rbd_cli_import_export.yaml} 3
Failure Reason:

Command failed on ovh063 with status 1: '\n sudo yum -y install ceph\n '

fail 3442095 2019-01-10 07:00:40 2019-01-10 07:51:42 2019-01-10 09:03:42 1:12:00 0:07:41 1:04:19 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rbd_fsx.yaml} 3
Failure Reason:

Command failed on ovh086 with status 1: '\n sudo yum -y install ceph\n '

fail 3442096 2019-01-10 07:00:40 2019-01-10 07:59:37 2019-01-10 09:01:37 1:02:00 0:07:53 0:54:07 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rbd_python_api_tests.yaml} 3
Failure Reason:

Command failed on ovh011 with status 1: '\n sudo yum -y install ceph\n '

fail 3442097 2019-01-10 07:00:41 2019-01-10 08:05:40 2019-01-10 09:05:40 1:00:00 0:07:28 0:52:32 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rbd_workunit_suites_iozone.yaml} 3
Failure Reason:

Command failed on ovh058 with status 1: '\n sudo yum -y install ceph\n '

fail 3442098 2019-01-10 07:00:42 2019-01-10 08:07:55 2019-01-10 09:07:55 1:00:00 0:07:35 0:52:25 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rgw_ec_s3tests.yaml} 3
Failure Reason:

Command failed on ovh030 with status 1: '\n sudo yum -y install ceph\n '

fail 3442099 2019-01-10 07:00:42 2019-01-10 08:10:59 2019-01-10 09:18:59 1:08:00 0:07:48 1:00:12 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rgw_s3tests.yaml} 3
Failure Reason:

Command failed on ovh100 with status 1: '\n sudo yum -y install ceph\n '

fail 3442100 2019-01-10 07:00:43 2019-01-10 08:11:24 2019-01-10 09:07:24 0:56:00 0:07:57 0:48:03 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rgw_swift.yaml} 3
Failure Reason:

Command failed on ovh054 with status 1: '\n sudo yum -y install ceph\n '