Status Job ID Links Posted Started Updated
Runtime
Duration
In Waiting
Machine Teuthology Branch OS Type OS Version Description Nodes
fail 3414196 2019-01-02 07:00:25 2019-01-02 07:04:20 2019-01-02 07:12:19 0:07:59 0:03:09 0:04:50 ovh master ubuntu 16.04 smoke/1node/{clusters/{fixed-1.yaml openstack.yaml} distros/ubuntu_latest.yaml objectstore/filestore-xfs.yaml tasks/ceph-deploy.yaml} 1
Failure Reason:

{'ovh016.front.sepia.ceph.com': {'_ansible_parsed': True, 'invocation': {'module_args': {'comment': None, 'ssh_key_bits': 0, 'update_password': 'always', 'non_unique': False, 'force': False, 'skeleton': None, 'expires': None, 'ssh_key_passphrase': None, 'groups': ['fuse', 'kvm', 'disk'], 'createhome': True, 'home': None, 'move_home': False, 'password': None, 'generate_ssh_key': None, 'append': True, 'uid': None, 'ssh_key_comment': 'ansible-generated on ovh016', 'group': None, 'name': 'ubuntu', 'local': None, 'seuser': None, 'system': False, 'remove': False, 'state': 'present', 'ssh_key_file': None, 'login_class': None, 'shell': None, 'ssh_key_type': 'rsa'}}, 'changed': False, '_ansible_no_log': False, 'msg': 'Group kvm does not exist'}}

fail 3414197 2019-01-02 07:00:25 2019-01-02 07:04:20 2019-01-02 08:12:21 1:08:01 0:07:48 1:00:13 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/cfuse_workunit_suites_blogbench.yaml} 3
Failure Reason:

Command failed on ovh075 with status 1: '\n sudo yum -y install ceph\n '

fail 3414198 2019-01-02 07:00:26 2019-01-02 07:12:21 2019-01-02 10:46:28 3:34:07 0:24:07 3:10:00 ovh master centos 7.5 smoke/systemd/{clusters/{fixed-4.yaml openstack.yaml} distros/centos_latest.yaml objectstore/filestore-xfs.yaml tasks/systemd.yaml} 4
Failure Reason:

ceph-deploy: Failed during gather keys

fail 3414199 2019-01-02 07:00:27 2019-01-02 07:18:14 2019-01-02 09:52:15 2:34:01 0:07:24 2:26:37 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/cfuse_workunit_suites_fsstress.yaml} 3
Failure Reason:

Command failed on ovh100 with status 1: '\n sudo yum -y install ceph\n '

fail 3414200 2019-01-02 07:00:28 2019-01-02 07:26:27 2019-01-02 08:20:27 0:54:00 0:08:11 0:45:49 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/cfuse_workunit_suites_iozone.yaml} 3
Failure Reason:

Command failed on ovh061 with status 1: '\n sudo yum -y install ceph\n '

fail 3414201 2019-01-02 07:00:28 2019-01-02 07:28:15 2019-01-02 09:38:16 2:10:01 0:08:03 2:01:58 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/cfuse_workunit_suites_pjd.yaml} 3
Failure Reason:

Command failed on ovh069 with status 1: '\n sudo yum -y install ceph\n '

fail 3414202 2019-01-02 07:00:29 2019-01-02 07:28:16 2019-01-02 08:38:16 1:10:00 0:08:22 1:01:38 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/kclient_workunit_direct_io.yaml} 3
Failure Reason:

Command failed on ovh040 with status 1: '\n sudo yum -y install ceph\n '

fail 3414203 2019-01-02 07:00:30 2019-01-02 07:29:57 2019-01-02 09:35:58 2:06:01 0:07:52 1:58:09 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/kclient_workunit_suites_dbench.yaml} 3
Failure Reason:

Command failed on ovh063 with status 1: '\n sudo yum -y install ceph\n '

fail 3414204 2019-01-02 07:00:31 2019-01-02 07:30:28 2019-01-02 09:58:29 2:28:01 0:09:09 2:18:52 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/kclient_workunit_suites_fsstress.yaml} 3
Failure Reason:

Command failed on ovh041 with status 1: '\n sudo yum -y install ceph\n '

fail 3414205 2019-01-02 07:00:31 2019-01-02 07:31:58 2019-01-02 09:33:59 2:02:01 0:07:58 1:54:03 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/kclient_workunit_suites_pjd.yaml} 3
Failure Reason:

Command failed on ovh073 with status 1: '\n sudo yum -y install ceph\n '

fail 3414206 2019-01-02 07:00:32 2019-01-02 07:32:42 2019-01-02 09:16:42 1:44:00 0:08:06 1:35:54 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/libcephfs_interface_tests.yaml} 3
Failure Reason:

Command failed on ovh026 with status 1: '\n sudo yum -y install ceph\n '

fail 3414207 2019-01-02 07:00:33 2019-01-02 07:46:23 2019-01-02 08:50:23 1:04:00 0:08:06 0:55:54 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/mon_thrash.yaml} 3
Failure Reason:

Command failed on ovh066 with status 1: '\n sudo yum -y install ceph\n '

fail 3414208 2019-01-02 07:00:34 2019-01-02 07:47:13 2019-01-02 08:45:12 0:57:59 0:08:37 0:49:22 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rados_api_tests.yaml} 3
Failure Reason:

Command failed on ovh088 with status 1: '\n sudo yum -y install ceph\n '

fail 3414209 2019-01-02 07:00:34 2019-01-02 07:50:39 2019-01-02 09:04:40 1:14:01 0:08:21 1:05:40 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rados_bench.yaml} 3
Failure Reason:

Command failed on ovh038 with status 1: '\n sudo yum -y install ceph\n '

fail 3414210 2019-01-02 07:00:35 2019-01-02 07:53:01 2019-01-02 09:11:02 1:18:01 0:07:45 1:10:16 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rados_cache_snaps.yaml} 3
Failure Reason:

Command failed on ovh093 with status 1: '\n sudo yum -y install ceph\n '

fail 3414211 2019-01-02 07:00:36 2019-01-02 07:53:52 2019-01-02 14:51:58 6:58:06 0:02:47 6:55:19 ovh master ubuntu 16.04 smoke/systemd/{clusters/{fixed-4.yaml openstack.yaml} distros/ubuntu_latest.yaml objectstore/filestore-xfs.yaml tasks/systemd.yaml} 4
Failure Reason:

{'ovh001.front.sepia.ceph.com': {'_ansible_parsed': True, 'invocation': {'module_args': {'comment': None, 'ssh_key_bits': 0, 'update_password': 'always', 'non_unique': False, 'force': False, 'skeleton': None, 'expires': None, 'ssh_key_passphrase': None, 'groups': ['fuse', 'kvm', 'disk'], 'createhome': True, 'home': None, 'move_home': False, 'password': None, 'generate_ssh_key': None, 'append': True, 'uid': None, 'ssh_key_comment': 'ansible-generated on ovh001', 'group': None, 'name': 'ubuntu', 'local': None, 'seuser': None, 'system': False, 'remove': False, 'state': 'present', 'ssh_key_file': None, 'login_class': None, 'shell': None, 'ssh_key_type': 'rsa'}}, 'changed': False, '_ansible_no_log': False, 'msg': 'Group kvm does not exist'}, 'ovh070.front.sepia.ceph.com': {'_ansible_parsed': True, 'invocation': {'module_args': {'comment': None, 'ssh_key_bits': 0, 'update_password': 'always', 'non_unique': False, 'force': False, 'skeleton': None, 'expires': None, 'ssh_key_passphrase': None, 'groups': ['fuse', 'kvm', 'disk'], 'createhome': True, 'home': None, 'move_home': False, 'password': None, 'generate_ssh_key': None, 'append': True, 'uid': None, 'ssh_key_comment': 'ansible-generated on ovh070', 'group': None, 'name': 'ubuntu', 'local': None, 'seuser': None, 'system': False, 'remove': False, 'state': 'present', 'ssh_key_file': None, 'login_class': None, 'shell': None, 'ssh_key_type': 'rsa'}}, 'changed': False, '_ansible_no_log': False, 'msg': 'Group kvm does not exist'}, 'ovh066.front.sepia.ceph.com': {'_ansible_parsed': True, 'invocation': {'module_args': {'comment': None, 'ssh_key_bits': 0, 'update_password': 'always', 'non_unique': False, 'force': False, 'skeleton': None, 'expires': None, 'ssh_key_passphrase': None, 'groups': ['fuse', 'kvm', 'disk'], 'createhome': True, 'home': None, 'move_home': False, 'password': None, 'generate_ssh_key': None, 'append': True, 'uid': None, 'ssh_key_comment': 'ansible-generated on ovh066', 'group': None, 'name': 'ubuntu', 'local': None, 'seuser': None, 'system': False, 'remove': False, 'state': 'present', 'ssh_key_file': None, 'login_class': None, 'shell': None, 'ssh_key_type': 'rsa'}}, 'changed': False, '_ansible_no_log': False, 'msg': 'Group kvm does not exist'}, 'ovh068.front.sepia.ceph.com': {'_ansible_parsed': True, 'invocation': {'module_args': {'comment': None, 'ssh_key_bits': 0, 'update_password': 'always', 'non_unique': False, 'force': False, 'skeleton': None, 'expires': None, 'ssh_key_passphrase': None, 'groups': ['fuse', 'kvm', 'disk'], 'createhome': True, 'home': None, 'move_home': False, 'password': None, 'generate_ssh_key': None, 'append': True, 'uid': None, 'ssh_key_comment': 'ansible-generated on ovh068', 'group': None, 'name': 'ubuntu', 'local': None, 'seuser': None, 'system': False, 'remove': False, 'state': 'present', 'ssh_key_file': None, 'login_class': None, 'shell': None, 'ssh_key_type': 'rsa'}}, 'changed': False, '_ansible_no_log': False, 'msg': 'Group kvm does not exist'}}

fail 3414212 2019-01-02 07:00:37 2019-01-02 07:56:16 2019-01-02 08:50:16 0:54:00 0:08:45 0:45:15 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rados_cls_all.yaml} 3
Failure Reason:

Command failed on ovh091 with status 1: '\n sudo yum -y install ceph\n '

fail 3414213 2019-01-02 07:00:38 2019-01-02 07:58:09 2019-01-02 08:54:09 0:56:00 0:07:56 0:48:04 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rados_ec_snaps.yaml} 3
Failure Reason:

Command failed on ovh058 with status 1: '\n sudo yum -y install ceph\n '

fail 3414214 2019-01-02 07:00:38 2019-01-02 08:09:31 2019-01-02 09:17:31 1:08:00 0:08:14 0:59:46 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rados_python.yaml} 3
Failure Reason:

Command failed on ovh027 with status 1: '\n sudo yum -y install ceph\n '

fail 3414215 2019-01-02 07:00:39 2019-01-02 08:12:22 2019-01-02 09:20:22 1:08:00 0:07:48 1:00:12 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rados_workunit_loadgen_mix.yaml} 3
Failure Reason:

Command failed on ovh041 with status 1: '\n sudo yum -y install ceph\n '

fail 3414216 2019-01-02 07:00:40 2019-01-02 08:16:20 2019-01-02 09:42:21 1:26:01 0:07:25 1:18:36 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rbd_api_tests.yaml} 3
Failure Reason:

Command failed on ovh091 with status 1: '\n sudo yum -y install ceph\n '

fail 3414217 2019-01-02 07:00:40 2019-01-02 08:16:59 2019-01-02 09:55:00 1:38:01 0:08:49 1:29:12 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rbd_cli_import_export.yaml} 3
Failure Reason:

Command failed on ovh035 with status 1: '\n sudo yum -y install ceph\n '

fail 3414218 2019-01-02 07:00:41 2019-01-02 08:18:07 2019-01-02 09:20:07 1:02:00 0:07:46 0:54:14 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rbd_fsx.yaml} 3
Failure Reason:

Command failed on ovh022 with status 1: '\n sudo yum -y install ceph\n '

fail 3414219 2019-01-02 07:00:42 2019-01-02 08:20:29 2019-01-02 10:36:30 2:16:01 0:07:55 2:08:06 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rbd_python_api_tests.yaml} 3
Failure Reason:

Command failed on ovh035 with status 1: '\n sudo yum -y install ceph\n '

fail 3414220 2019-01-02 07:00:43 2019-01-02 08:22:51 2019-01-02 09:54:52 1:32:01 0:09:01 1:23:00 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rbd_workunit_suites_iozone.yaml} 3
Failure Reason:

Command failed on ovh098 with status 1: '\n sudo yum -y install ceph\n '

fail 3414221 2019-01-02 07:00:43 2019-01-02 08:26:30 2019-01-02 09:42:30 1:16:00 0:08:49 1:07:11 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rgw_ec_s3tests.yaml} 3
Failure Reason:

Command failed on ovh086 with status 1: '\n sudo yum -y install ceph\n '

fail 3414222 2019-01-02 07:00:44 2019-01-02 08:26:30 2019-01-02 09:20:29 0:53:59 0:07:56 0:46:03 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rgw_s3tests.yaml} 3
Failure Reason:

Command failed on ovh034 with status 1: '\n sudo yum -y install ceph\n '

fail 3414223 2019-01-02 07:00:45 2019-01-02 08:30:24 2019-01-02 09:44:24 1:14:00 0:07:58 1:06:02 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rgw_swift.yaml} 3
Failure Reason:

Command failed on ovh030 with status 1: '\n sudo yum -y install ceph\n '