Status Job ID Links Posted Started Updated
Runtime
Duration
In Waiting
Machine Teuthology Branch OS Type OS Version Description Nodes
fail 3844736 2019-04-14 00:00:25 2019-04-14 00:08:49 2019-04-14 00:20:48 0:11:59 0:03:27 0:08:32 ovh master ubuntu 18.04 smoke/1node/{clusters/{fixed-1.yaml openstack.yaml} distros/ubuntu_latest.yaml objectstore/filestore-xfs.yaml tasks/ceph-deploy.yaml} 1
Failure Reason:

{'ovh040.front.sepia.ceph.com': {'_ansible_parsed': True, 'invocation': {'module_args': {'comment': None, 'ssh_key_bits': 0, 'update_password': 'always', 'non_unique': False, 'force': False, 'skeleton': None, 'expires': None, 'ssh_key_passphrase': None, 'groups': ['fuse', 'kvm', 'disk'], 'createhome': True, 'home': None, 'move_home': False, 'password': None, 'generate_ssh_key': None, 'append': True, 'uid': None, 'ssh_key_comment': 'ansible-generated on ovh040', 'group': None, 'name': 'ubuntu', 'local': None, 'seuser': None, 'system': False, 'remove': False, 'state': 'present', 'ssh_key_file': None, 'login_class': None, 'shell': None, 'ssh_key_type': 'rsa'}}, 'changed': False, '_ansible_no_log': False, 'msg': 'Group kvm does not exist'}}

fail 3844737 2019-04-14 00:00:26 2019-04-14 00:10:47 2019-04-14 01:00:47 0:50:00 0:06:02 0:43:58 ovh master rhel 7.4 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/cfuse_workunit_suites_blogbench.yaml} 3
Failure Reason:

Command failed on ovh023 with status 1: '\n sudo yum -y install ceph-radosgw\n '

dead 3844738 2019-04-14 00:00:27 2019-04-14 00:21:02 2019-04-14 12:23:15 12:02:13 ovh master centos 7.6 smoke/systemd/{clusters/{fixed-4.yaml openstack.yaml} distros/centos_latest.yaml objectstore/filestore-xfs.yaml tasks/systemd.yaml}
fail 3844739 2019-04-14 00:00:27 2019-04-14 00:22:55 2019-04-14 01:20:55 0:58:00 0:06:17 0:51:43 ovh master rhel 7.4 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/cfuse_workunit_suites_fsstress.yaml} 3
Failure Reason:

Command failed on ovh060 with status 1: '\n sudo yum -y install ceph-radosgw\n '

fail 3844740 2019-04-14 00:00:28 2019-04-14 00:50:54 2019-04-14 01:40:59 0:50:05 0:05:57 0:44:08 ovh master rhel 7.4 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/cfuse_workunit_suites_iozone.yaml} 3
Failure Reason:

Command failed on ovh040 with status 1: '\n sudo yum -y install ceph-radosgw\n '

fail 3844741 2019-04-14 00:00:29 2019-04-14 01:00:53 2019-04-14 01:48:53 0:48:00 0:06:32 0:41:28 ovh master rhel 7.4 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/cfuse_workunit_suites_pjd.yaml} 3
Failure Reason:

Command failed on ovh023 with status 1: '\n sudo yum -y install ceph-radosgw\n '

fail 3844742 2019-04-14 00:00:30 2019-04-14 01:00:53 2019-04-14 01:58:53 0:58:00 0:06:17 0:51:43 ovh master rhel 7.4 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/kclient_workunit_direct_io.yaml} 3
Failure Reason:

Command failed on ovh037 with status 1: '\n sudo yum -y install ceph-radosgw\n '

fail 3844743 2019-04-14 00:00:30 2019-04-14 01:06:43 2019-04-14 01:56:42 0:49:59 0:05:59 0:44:00 ovh master rhel 7.4 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/kclient_workunit_suites_dbench.yaml} 3
Failure Reason:

Command failed on ovh068 with status 1: '\n sudo yum -y install ceph-radosgw\n '

fail 3844744 2019-04-14 00:00:31 2019-04-14 01:17:04 2019-04-14 02:13:04 0:56:00 0:06:04 0:49:56 ovh master rhel 7.4 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/kclient_workunit_suites_fsstress.yaml} 3
Failure Reason:

Command failed on ovh003 with status 1: '\n sudo yum -y install ceph-radosgw\n '

fail 3844745 2019-04-14 00:00:32 2019-04-14 01:21:07 2019-04-14 02:17:07 0:56:00 0:06:33 0:49:27 ovh master rhel 7.4 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/kclient_workunit_suites_pjd.yaml} 3
Failure Reason:

Command failed on ovh081 with status 1: '\n sudo yum -y install ceph-radosgw\n '

fail 3844746 2019-04-14 00:00:32 2019-04-14 01:41:04 2019-04-14 02:37:04 0:56:00 0:06:13 0:49:47 ovh master rhel 7.4 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/libcephfs_interface_tests.yaml} 3
Failure Reason:

Command failed on ovh028 with status 1: '\n sudo yum -y install ceph-radosgw\n '

fail 3844747 2019-04-14 00:00:33 2019-04-14 01:48:57 2019-04-14 02:34:56 0:45:59 0:06:23 0:39:36 ovh master rhel 7.4 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/mon_thrash.yaml} 3
Failure Reason:

Command failed on ovh079 with status 1: '\n sudo yum -y install ceph-radosgw\n '

fail 3844748 2019-04-14 00:00:34 2019-04-14 01:56:59 2019-04-14 02:48:59 0:52:00 0:06:29 0:45:31 ovh master rhel 7.4 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rados_api_tests.yaml} 3
Failure Reason:

Command failed on ovh069 with status 1: '\n sudo yum -y install ceph-radosgw\n '

fail 3844749 2019-04-14 00:00:35 2019-04-14 01:58:57 2019-04-14 02:50:58 0:52:01 0:06:21 0:45:40 ovh master rhel 7.4 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rados_bench.yaml} 3
Failure Reason:

Command failed on ovh036 with status 1: '\n sudo yum -y install ceph-radosgw\n '

fail 3844750 2019-04-14 00:00:35 2019-04-14 02:13:09 2019-04-14 03:01:09 0:48:00 0:06:13 0:41:47 ovh master rhel 7.4 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rados_cache_snaps.yaml} 3
Failure Reason:

Command failed on ovh080 with status 1: '\n sudo yum -y install ceph-radosgw\n '

fail 3844751 2019-04-14 00:00:36 2019-04-14 02:17:12 2019-04-14 02:43:11 0:25:59 0:04:14 0:21:45 ovh master ubuntu 18.04 smoke/systemd/{clusters/{fixed-4.yaml openstack.yaml} distros/ubuntu_latest.yaml objectstore/filestore-xfs.yaml tasks/systemd.yaml} 4
Failure Reason:

{'ovh048.front.sepia.ceph.com': {'_ansible_parsed': True, 'invocation': {'module_args': {'comment': None, 'ssh_key_bits': 0, 'update_password': 'always', 'non_unique': False, 'force': False, 'skeleton': None, 'expires': None, 'ssh_key_passphrase': None, 'groups': ['fuse', 'kvm', 'disk'], 'createhome': True, 'home': None, 'move_home': False, 'password': None, 'generate_ssh_key': None, 'append': True, 'uid': None, 'ssh_key_comment': 'ansible-generated on ovh048', 'group': None, 'name': 'ubuntu', 'local': None, 'seuser': None, 'system': False, 'remove': False, 'state': 'present', 'ssh_key_file': None, 'login_class': None, 'shell': None, 'ssh_key_type': 'rsa'}}, 'changed': False, '_ansible_no_log': False, 'msg': 'Group kvm does not exist'}, 'ovh029.front.sepia.ceph.com': {'_ansible_parsed': True, 'invocation': {'module_args': {'comment': None, 'ssh_key_bits': 0, 'update_password': 'always', 'non_unique': False, 'force': False, 'skeleton': None, 'expires': None, 'ssh_key_passphrase': None, 'groups': ['fuse', 'kvm', 'disk'], 'createhome': True, 'home': None, 'move_home': False, 'password': None, 'generate_ssh_key': None, 'append': True, 'uid': None, 'ssh_key_comment': 'ansible-generated on ovh029', 'group': None, 'name': 'ubuntu', 'local': None, 'seuser': None, 'system': False, 'remove': False, 'state': 'present', 'ssh_key_file': None, 'login_class': None, 'shell': None, 'ssh_key_type': 'rsa'}}, 'changed': False, '_ansible_no_log': False, 'msg': 'Group kvm does not exist'}, 'ovh081.front.sepia.ceph.com': {'_ansible_parsed': True, 'invocation': {'module_args': {'comment': None, 'ssh_key_bits': 0, 'update_password': 'always', 'non_unique': False, 'force': False, 'skeleton': None, 'expires': None, 'ssh_key_passphrase': None, 'groups': ['fuse', 'kvm', 'disk'], 'createhome': True, 'home': None, 'move_home': False, 'password': None, 'generate_ssh_key': None, 'append': True, 'uid': None, 'ssh_key_comment': 'ansible-generated on ovh081', 'group': None, 'name': 'ubuntu', 'local': None, 'seuser': None, 'system': False, 'remove': False, 'state': 'present', 'ssh_key_file': None, 'login_class': None, 'shell': None, 'ssh_key_type': 'rsa'}}, 'changed': False, '_ansible_no_log': False, 'msg': 'Group kvm does not exist'}, 'ovh042.front.sepia.ceph.com': {'_ansible_parsed': True, 'invocation': {'module_args': {'comment': None, 'ssh_key_bits': 0, 'update_password': 'always', 'non_unique': False, 'force': False, 'skeleton': None, 'expires': None, 'ssh_key_passphrase': None, 'groups': ['fuse', 'kvm', 'disk'], 'createhome': True, 'home': None, 'move_home': False, 'password': None, 'generate_ssh_key': None, 'append': True, 'uid': None, 'ssh_key_comment': 'ansible-generated on ovh042', 'group': None, 'name': 'ubuntu', 'local': None, 'seuser': None, 'system': False, 'remove': False, 'state': 'present', 'ssh_key_file': None, 'login_class': None, 'shell': None, 'ssh_key_type': 'rsa'}}, 'changed': False, '_ansible_no_log': False, 'msg': 'Group kvm does not exist'}}

fail 3844752 2019-04-14 00:00:37 2019-04-14 02:35:15 2019-04-14 03:25:15 0:50:00 0:06:30 0:43:30 ovh master rhel 7.4 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rados_cls_all.yaml} 3
Failure Reason:

Command failed on ovh044 with status 1: '\n sudo yum -y install ceph-radosgw\n '

fail 3844753 2019-04-14 00:00:38 2019-04-14 02:37:08 2019-04-14 03:27:08 0:50:00 0:06:12 0:43:48 ovh master rhel 7.4 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rados_ec_snaps.yaml} 3
Failure Reason:

Command failed on ovh078 with status 1: '\n sudo yum -y install ceph-radosgw\n '

fail 3844754 2019-04-14 00:00:38 2019-04-14 02:43:15 2019-04-14 03:39:15 0:56:00 0:05:51 0:50:09 ovh master rhel 7.4 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rados_python.yaml} 3
Failure Reason:

Command failed on ovh048 with status 1: '\n sudo yum -y install ceph-radosgw\n '

fail 3844755 2019-04-14 00:00:39 2019-04-14 02:49:09 2019-04-14 03:39:09 0:50:00 0:06:19 0:43:41 ovh master rhel 7.4 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rados_workunit_loadgen_mix.yaml} 3
Failure Reason:

Command failed on ovh051 with status 1: '\n sudo yum -y install ceph-radosgw\n '

fail 3844756 2019-04-14 00:00:40 2019-04-14 02:51:13 2019-04-14 03:47:13 0:56:00 0:06:19 0:49:41 ovh master rhel 7.4 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rbd_api_tests.yaml} 3
Failure Reason:

Command failed on ovh054 with status 1: '\n sudo yum -y install ceph-radosgw\n '

fail 3844757 2019-04-14 00:00:40 2019-04-14 03:01:13 2019-04-14 04:05:13 1:04:00 0:06:42 0:57:18 ovh master rhel 7.4 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rbd_cli_import_export.yaml} 3
Failure Reason:

Command failed on ovh027 with status 1: '\n sudo yum -y install ceph-radosgw\n '

fail 3844758 2019-04-14 00:00:41 2019-04-14 03:25:26 2019-04-14 04:09:26 0:44:00 0:05:55 0:38:05 ovh master rhel 7.4 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rbd_fsx.yaml} 3
Failure Reason:

Command failed on ovh023 with status 1: '\n sudo yum -y install ceph-radosgw\n '

fail 3844759 2019-04-14 00:00:42 2019-04-14 03:27:18 2019-04-14 04:17:17 0:49:59 0:06:11 0:43:48 ovh master rhel 7.4 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rbd_python_api_tests.yaml} 3
Failure Reason:

Command failed on ovh078 with status 1: '\n sudo yum -y install ceph-radosgw\n '

fail 3844760 2019-04-14 00:00:42 2019-04-14 03:39:21 2019-04-14 04:35:20 0:55:59 0:06:04 0:49:55 ovh master rhel 7.4 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rbd_workunit_suites_iozone.yaml} 3
Failure Reason:

Command failed on ovh065 with status 1: '\n sudo yum -y install ceph-radosgw\n '

fail 3844761 2019-04-14 00:00:43 2019-04-14 03:39:21 2019-04-14 04:33:20 0:53:59 0:06:08 0:47:51 ovh master rhel 7.4 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rgw_ec_s3tests.yaml} 3
Failure Reason:

Command failed on ovh037 with status 1: '\n sudo yum -y install ceph-radosgw\n '

fail 3844762 2019-04-14 00:00:44 2019-04-14 03:47:18 2019-04-14 04:47:18 1:00:00 0:06:11 0:53:49 ovh master rhel 7.4 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rgw_s3tests.yaml} 3
Failure Reason:

Command failed on ovh053 with status 1: '\n sudo yum -y install ceph-radosgw\n '

fail 3844763 2019-04-14 00:00:45 2019-04-14 04:05:17 2019-04-14 04:53:17 0:48:00 0:06:33 0:41:27 ovh master rhel 7.4 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rgw_swift.yaml} 3
Failure Reason:

Command failed on ovh036 with status 1: '\n sudo yum -y install ceph-radosgw\n '