Status Job ID Links Posted Started Updated
Runtime
Duration
In Waiting
Machine Teuthology Branch OS Type OS Version Description Nodes
fail 3462061 2019-01-14 07:00:27 2019-01-14 07:00:32 2019-01-14 07:18:30 0:17:58 0:02:45 0:15:13 ovh master ubuntu 16.04 smoke/1node/{clusters/{fixed-1.yaml openstack.yaml} distros/ubuntu_latest.yaml objectstore/filestore-xfs.yaml tasks/ceph-deploy.yaml} 1
Failure Reason:

{'ovh100.front.sepia.ceph.com': {'_ansible_parsed': True, 'invocation': {'module_args': {'comment': None, 'ssh_key_bits': 0, 'update_password': 'always', 'non_unique': False, 'force': False, 'skeleton': None, 'expires': None, 'ssh_key_passphrase': None, 'groups': ['fuse', 'kvm', 'disk'], 'createhome': True, 'home': None, 'move_home': False, 'password': None, 'generate_ssh_key': None, 'append': True, 'uid': None, 'ssh_key_comment': 'ansible-generated on ovh100', 'group': None, 'name': 'ubuntu', 'local': None, 'seuser': None, 'system': False, 'remove': False, 'state': 'present', 'ssh_key_file': None, 'login_class': None, 'shell': None, 'ssh_key_type': 'rsa'}}, 'changed': False, '_ansible_no_log': False, 'msg': 'Group kvm does not exist'}}

fail 3462062 2019-01-14 07:00:27 2019-01-14 07:00:32 2019-01-14 07:50:31 0:49:59 0:07:51 0:42:08 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/cfuse_workunit_suites_blogbench.yaml} 3
Failure Reason:

Command failed on ovh002 with status 1: '\n sudo yum -y install ceph\n '

fail 3462063 2019-01-14 07:00:28 2019-01-14 07:00:31 2019-01-14 10:14:33 3:14:02 0:18:19 2:55:43 ovh master centos 7.5 smoke/systemd/{clusters/{fixed-4.yaml openstack.yaml} distros/centos_latest.yaml objectstore/filestore-xfs.yaml tasks/systemd.yaml} 4
Failure Reason:

ceph-deploy: Failed to zap osds

fail 3462064 2019-01-14 07:00:29 2019-01-14 07:00:32 2019-01-14 08:02:31 1:01:59 0:07:53 0:54:06 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/cfuse_workunit_suites_fsstress.yaml} 3
Failure Reason:

Command failed on ovh056 with status 1: '\n sudo yum -y install ceph\n '

fail 3462065 2019-01-14 07:00:30 2019-01-14 07:02:57 2019-01-14 07:58:57 0:56:00 0:07:49 0:48:11 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/cfuse_workunit_suites_iozone.yaml} 3
Failure Reason:

Command failed on ovh033 with status 1: '\n sudo yum -y install ceph\n '

fail 3462066 2019-01-14 07:00:31 2019-01-14 07:04:29 2019-01-14 08:08:29 1:04:00 0:07:51 0:56:09 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/cfuse_workunit_suites_pjd.yaml} 3
Failure Reason:

Command failed on ovh042 with status 1: '\n sudo yum -y install ceph\n '

fail 3462067 2019-01-14 07:00:32 2019-01-14 07:06:25 2019-01-14 08:02:25 0:56:00 0:07:54 0:48:06 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/kclient_workunit_direct_io.yaml} 3
Failure Reason:

Command failed on ovh065 with status 1: '\n sudo yum -y install ceph\n '

fail 3462068 2019-01-14 07:00:32 2019-01-14 07:07:07 2019-01-14 08:13:08 1:06:01 0:07:46 0:58:15 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/kclient_workunit_suites_dbench.yaml} 3
Failure Reason:

Command failed on ovh038 with status 1: '\n sudo yum -y install ceph\n '

fail 3462069 2019-01-14 07:00:33 2019-01-14 07:15:00 2019-01-14 08:43:01 1:28:01 0:07:32 1:20:29 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/kclient_workunit_suites_fsstress.yaml} 3
Failure Reason:

Command failed on ovh024 with status 1: '\n sudo yum -y install ceph\n '

fail 3462070 2019-01-14 07:00:34 2019-01-14 07:16:29 2019-01-14 08:14:34 0:58:05 0:07:48 0:50:17 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/kclient_workunit_suites_pjd.yaml} 3
Failure Reason:

Command failed on ovh041 with status 1: '\n sudo yum -y install ceph\n '

fail 3462071 2019-01-14 07:00:35 2019-01-14 07:16:55 2019-01-14 08:16:55 1:00:00 0:08:05 0:51:55 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/libcephfs_interface_tests.yaml} 3
Failure Reason:

Command failed on ovh046 with status 1: '\n sudo yum -y install ceph\n '

fail 3462072 2019-01-14 07:00:36 2019-01-14 07:18:32 2019-01-14 08:12:32 0:54:00 0:07:38 0:46:22 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/mon_thrash.yaml} 3
Failure Reason:

Command failed on ovh012 with status 1: '\n sudo yum -y install ceph\n '

fail 3462073 2019-01-14 07:00:37 2019-01-14 07:20:08 2019-01-14 08:40:13 1:20:05 0:07:48 1:12:17 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rados_api_tests.yaml} 3
Failure Reason:

Command failed on ovh055 with status 1: '\n sudo yum -y install ceph\n '

fail 3462074 2019-01-14 07:00:37 2019-01-14 07:20:30 2019-01-14 08:22:30 1:02:00 0:08:09 0:53:51 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rados_bench.yaml} 3
Failure Reason:

Command failed on ovh060 with status 1: '\n sudo yum -y install ceph\n '

fail 3462075 2019-01-14 07:00:38 2019-01-14 07:20:34 2019-01-14 08:18:34 0:58:00 0:07:33 0:50:27 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rados_cache_snaps.yaml} 3
Failure Reason:

Command failed on ovh096 with status 1: '\n sudo yum -y install ceph\n '

fail 3462076 2019-01-14 07:00:39 2019-01-14 07:22:31 2019-01-14 08:06:31 0:44:00 0:02:58 0:41:02 ovh master ubuntu 16.04 smoke/systemd/{clusters/{fixed-4.yaml openstack.yaml} distros/ubuntu_latest.yaml objectstore/filestore-xfs.yaml tasks/systemd.yaml} 4
Failure Reason:

{'ovh026.front.sepia.ceph.com': {'_ansible_parsed': True, 'invocation': {'module_args': {'comment': None, 'ssh_key_bits': 0, 'update_password': 'always', 'non_unique': False, 'force': False, 'skeleton': None, 'expires': None, 'ssh_key_passphrase': None, 'groups': ['fuse', 'kvm', 'disk'], 'createhome': True, 'home': None, 'move_home': False, 'password': None, 'generate_ssh_key': None, 'append': True, 'uid': None, 'ssh_key_comment': 'ansible-generated on ovh026', 'group': None, 'name': 'ubuntu', 'local': None, 'seuser': None, 'system': False, 'remove': False, 'state': 'present', 'ssh_key_file': None, 'login_class': None, 'shell': None, 'ssh_key_type': 'rsa'}}, 'changed': False, '_ansible_no_log': False, 'msg': 'Group kvm does not exist'}, 'ovh037.front.sepia.ceph.com': {'_ansible_parsed': True, 'invocation': {'module_args': {'comment': None, 'ssh_key_bits': 0, 'update_password': 'always', 'non_unique': False, 'force': False, 'skeleton': None, 'expires': None, 'ssh_key_passphrase': None, 'groups': ['fuse', 'kvm', 'disk'], 'createhome': True, 'home': None, 'move_home': False, 'password': None, 'generate_ssh_key': None, 'append': True, 'uid': None, 'ssh_key_comment': 'ansible-generated on ovh037', 'group': None, 'name': 'ubuntu', 'local': None, 'seuser': None, 'system': False, 'remove': False, 'state': 'present', 'ssh_key_file': None, 'login_class': None, 'shell': None, 'ssh_key_type': 'rsa'}}, 'changed': False, '_ansible_no_log': False, 'msg': 'Group kvm does not exist'}, 'ovh070.front.sepia.ceph.com': {'_ansible_parsed': True, 'invocation': {'module_args': {'comment': None, 'ssh_key_bits': 0, 'update_password': 'always', 'non_unique': False, 'force': False, 'skeleton': None, 'expires': None, 'ssh_key_passphrase': None, 'groups': ['fuse', 'kvm', 'disk'], 'createhome': True, 'home': None, 'move_home': False, 'password': None, 'generate_ssh_key': None, 'append': True, 'uid': None, 'ssh_key_comment': 'ansible-generated on ovh070', 'group': None, 'name': 'ubuntu', 'local': None, 'seuser': None, 'system': False, 'remove': False, 'state': 'present', 'ssh_key_file': None, 'login_class': None, 'shell': None, 'ssh_key_type': 'rsa'}}, 'changed': False, '_ansible_no_log': False, 'msg': 'Group kvm does not exist'}, 'ovh016.front.sepia.ceph.com': {'_ansible_parsed': True, 'invocation': {'module_args': {'comment': None, 'ssh_key_bits': 0, 'update_password': 'always', 'non_unique': False, 'force': False, 'skeleton': None, 'expires': None, 'ssh_key_passphrase': None, 'groups': ['fuse', 'kvm', 'disk'], 'createhome': True, 'home': None, 'move_home': False, 'password': None, 'generate_ssh_key': None, 'append': True, 'uid': None, 'ssh_key_comment': 'ansible-generated on ovh016', 'group': None, 'name': 'ubuntu', 'local': None, 'seuser': None, 'system': False, 'remove': False, 'state': 'present', 'ssh_key_file': None, 'login_class': None, 'shell': None, 'ssh_key_type': 'rsa'}}, 'changed': False, '_ansible_no_log': False, 'msg': 'Group kvm does not exist'}}

fail 3462077 2019-01-14 07:00:40 2019-01-14 07:24:35 2019-01-14 08:24:35 1:00:00 0:07:39 0:52:21 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rados_cls_all.yaml} 3
Failure Reason:

Command failed on ovh011 with status 1: '\n sudo yum -y install ceph\n '

fail 3462078 2019-01-14 07:00:40 2019-01-14 07:24:43 2019-01-14 08:28:44 1:04:01 0:07:52 0:56:09 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rados_ec_snaps.yaml} 3
Failure Reason:

Command failed on ovh084 with status 1: '\n sudo yum -y install ceph\n '

fail 3462079 2019-01-14 07:00:41 2019-01-14 07:27:58 2019-01-14 08:29:58 1:02:00 0:08:16 0:53:44 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rados_python.yaml} 3
Failure Reason:

Command failed on ovh064 with status 1: '\n sudo yum -y install ceph\n '

fail 3462080 2019-01-14 07:00:42 2019-01-14 07:35:08 2019-01-14 08:33:08 0:58:00 0:08:08 0:49:52 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rados_workunit_loadgen_mix.yaml} 3
Failure Reason:

Command failed on ovh020 with status 1: '\n sudo yum -y install ceph\n '

fail 3462081 2019-01-14 07:00:43 2019-01-14 07:38:57 2019-01-14 09:00:57 1:22:00 0:07:43 1:14:17 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rbd_api_tests.yaml} 3
Failure Reason:

Command failed on ovh003 with status 1: '\n sudo yum -y install ceph\n '

fail 3462082 2019-01-14 07:00:44 2019-01-14 07:40:36 2019-01-14 09:04:36 1:24:00 0:07:30 1:16:30 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rbd_cli_import_export.yaml} 3
Failure Reason:

Command failed on ovh087 with status 1: '\n sudo yum -y install ceph\n '

fail 3462083 2019-01-14 07:00:44 2019-01-14 07:44:30 2019-01-14 08:38:30 0:54:00 0:07:46 0:46:14 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rbd_fsx.yaml} 3
Failure Reason:

Command failed on ovh093 with status 1: '\n sudo yum -y install ceph\n '

fail 3462084 2019-01-14 07:00:45 2019-01-14 07:46:55 2019-01-14 08:50:55 1:04:00 0:08:27 0:55:33 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rbd_python_api_tests.yaml} 3
Failure Reason:

Command failed on ovh001 with status 1: '\n sudo yum -y install ceph\n '

fail 3462085 2019-01-14 07:00:46 2019-01-14 07:50:33 2019-01-14 08:50:33 1:00:00 0:07:39 0:52:21 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rbd_workunit_suites_iozone.yaml} 3
Failure Reason:

Command failed on ovh065 with status 1: '\n sudo yum -y install ceph\n '

fail 3462086 2019-01-14 07:00:47 2019-01-14 07:54:42 2019-01-14 09:02:47 1:08:05 0:07:54 1:00:11 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rgw_ec_s3tests.yaml} 3
Failure Reason:

Command failed on ovh046 with status 1: '\n sudo yum -y install ceph\n '

fail 3462087 2019-01-14 07:00:47 2019-01-14 07:58:45 2019-01-14 09:08:45 1:10:00 0:07:53 1:02:07 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rgw_s3tests.yaml} 3
Failure Reason:

Command failed on ovh050 with status 1: '\n sudo yum -y install ceph\n '

fail 3462088 2019-01-14 07:00:48 2019-01-14 07:58:59 2019-01-14 08:52:59 0:54:00 0:07:30 0:46:30 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rgw_swift.yaml} 3
Failure Reason:

Command failed on ovh068 with status 1: '\n sudo yum -y install ceph\n '