Status Job ID Links Posted Started Updated
Runtime
Duration
In Waiting
Machine Teuthology Branch OS Type OS Version Description Nodes
fail 3471369 2019-01-16 07:00:25 2019-01-16 12:07:48 2019-01-16 12:29:48 0:22:00 0:02:50 0:19:10 ovh master ubuntu 16.04 smoke/1node/{clusters/{fixed-1.yaml openstack.yaml} distros/ubuntu_latest.yaml objectstore/filestore-xfs.yaml tasks/ceph-deploy.yaml} 1
Failure Reason:

{'ovh053.front.sepia.ceph.com': {'_ansible_parsed': True, 'invocation': {'module_args': {'comment': None, 'ssh_key_bits': 0, 'update_password': 'always', 'non_unique': False, 'force': False, 'skeleton': None, 'expires': None, 'ssh_key_passphrase': None, 'groups': ['fuse', 'kvm', 'disk'], 'createhome': True, 'home': None, 'move_home': False, 'password': None, 'generate_ssh_key': None, 'append': True, 'uid': None, 'ssh_key_comment': 'ansible-generated on ovh053', 'group': None, 'name': 'ubuntu', 'local': None, 'seuser': None, 'system': False, 'remove': False, 'state': 'present', 'ssh_key_file': None, 'login_class': None, 'shell': None, 'ssh_key_type': 'rsa'}}, 'changed': False, '_ansible_no_log': False, 'msg': 'Group kvm does not exist'}}

fail 3471370 2019-01-16 07:00:25 2019-01-16 12:09:48 2019-01-16 13:03:48 0:54:00 0:07:53 0:46:07 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/cfuse_workunit_suites_blogbench.yaml} 3
Failure Reason:

Command failed on ovh054 with status 1: '\n sudo yum -y install ceph\n '

fail 3471371 2019-01-16 07:00:26 2019-01-16 12:11:51 2019-01-16 13:29:51 1:18:00 0:17:04 1:00:56 ovh master centos 7.5 smoke/systemd/{clusters/{fixed-4.yaml openstack.yaml} distros/centos_latest.yaml objectstore/filestore-xfs.yaml tasks/systemd.yaml} 4
Failure Reason:

ceph-deploy: Failed to zap osds

fail 3471372 2019-01-16 07:00:27 2019-01-16 12:25:50 2019-01-16 13:19:50 0:54:00 0:07:49 0:46:11 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/cfuse_workunit_suites_fsstress.yaml} 3
Failure Reason:

Command failed on ovh067 with status 1: '\n sudo yum -y install ceph\n '

fail 3471373 2019-01-16 07:00:28 2019-01-16 12:29:50 2019-01-16 13:33:50 1:04:00 0:07:56 0:56:04 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/cfuse_workunit_suites_iozone.yaml} 3
Failure Reason:

Command failed on ovh020 with status 1: '\n sudo yum -y install ceph\n '

fail 3471374 2019-01-16 07:00:29 2019-01-16 13:03:51 2019-01-16 13:59:51 0:56:00 0:07:58 0:48:02 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/cfuse_workunit_suites_pjd.yaml} 3
Failure Reason:

Command failed on ovh072 with status 1: '\n sudo yum -y install ceph\n '

fail 3471375 2019-01-16 07:00:29 2019-01-16 13:19:52 2019-01-16 14:13:52 0:54:00 0:07:53 0:46:07 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/kclient_workunit_direct_io.yaml} 3
Failure Reason:

Command failed on ovh002 with status 1: '\n sudo yum -y install ceph\n '

fail 3471376 2019-01-16 07:00:30 2019-01-16 13:29:54 2019-01-16 14:21:53 0:51:59 0:07:44 0:44:15 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/kclient_workunit_suites_dbench.yaml} 3
Failure Reason:

Command failed on ovh020 with status 1: '\n sudo yum -y install ceph\n '

fail 3471377 2019-01-16 07:00:31 2019-01-16 13:33:52 2019-01-16 14:39:52 1:06:00 0:07:47 0:58:13 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/kclient_workunit_suites_fsstress.yaml} 3
Failure Reason:

Command failed on ovh037 with status 1: '\n sudo yum -y install ceph\n '

fail 3471378 2019-01-16 07:00:31 2019-01-16 13:59:52 2019-01-16 14:55:52 0:56:00 0:08:13 0:47:47 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/kclient_workunit_suites_pjd.yaml} 3
Failure Reason:

Command failed on ovh054 with status 1: '\n sudo yum -y install ceph\n '

fail 3471379 2019-01-16 07:00:32 2019-01-16 14:14:02 2019-01-16 15:08:02 0:54:00 0:07:54 0:46:06 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/libcephfs_interface_tests.yaml} 3
Failure Reason:

Command failed on ovh002 with status 1: '\n sudo yum -y install ceph\n '

fail 3471380 2019-01-16 07:00:33 2019-01-16 14:22:04 2019-01-16 15:18:04 0:56:00 0:08:01 0:47:59 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/mon_thrash.yaml} 3
Failure Reason:

Command failed on ovh083 with status 1: '\n sudo yum -y install ceph\n '

fail 3471381 2019-01-16 07:00:34 2019-01-16 14:40:01 2019-01-16 15:46:01 1:06:00 0:07:45 0:58:15 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rados_api_tests.yaml} 3
Failure Reason:

Command failed on ovh031 with status 1: '\n sudo yum -y install ceph\n '

fail 3471382 2019-01-16 07:00:34 2019-01-16 14:56:04 2019-01-16 15:52:04 0:56:00 0:08:02 0:47:58 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rados_bench.yaml} 3
Failure Reason:

Command failed on ovh054 with status 1: '\n sudo yum -y install ceph\n '

fail 3471383 2019-01-16 07:00:35 2019-01-16 15:08:14 2019-01-16 16:02:14 0:54:00 0:08:06 0:45:54 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rados_cache_snaps.yaml} 3
Failure Reason:

Command failed on ovh067 with status 1: '\n sudo yum -y install ceph\n '

fail 3471384 2019-01-16 07:00:36 2019-01-16 15:18:06 2019-01-16 15:38:05 0:19:59 0:03:00 0:16:59 ovh master ubuntu 16.04 smoke/systemd/{clusters/{fixed-4.yaml openstack.yaml} distros/ubuntu_latest.yaml objectstore/filestore-xfs.yaml tasks/systemd.yaml} 4
Failure Reason:

{'ovh040.front.sepia.ceph.com': {'_ansible_parsed': True, 'invocation': {'module_args': {'comment': None, 'ssh_key_bits': 0, 'update_password': 'always', 'non_unique': False, 'force': False, 'skeleton': None, 'expires': None, 'ssh_key_passphrase': None, 'groups': ['fuse', 'kvm', 'disk'], 'createhome': True, 'home': None, 'move_home': False, 'password': None, 'generate_ssh_key': None, 'append': True, 'uid': None, 'ssh_key_comment': 'ansible-generated on ovh040', 'group': None, 'name': 'ubuntu', 'local': None, 'seuser': None, 'system': False, 'remove': False, 'state': 'present', 'ssh_key_file': None, 'login_class': None, 'shell': None, 'ssh_key_type': 'rsa'}}, 'changed': False, '_ansible_no_log': False, 'msg': 'Group kvm does not exist'}, 'ovh083.front.sepia.ceph.com': {'_ansible_parsed': True, 'invocation': {'module_args': {'comment': None, 'ssh_key_bits': 0, 'update_password': 'always', 'non_unique': False, 'force': False, 'skeleton': None, 'expires': None, 'ssh_key_passphrase': None, 'groups': ['fuse', 'kvm', 'disk'], 'createhome': True, 'home': None, 'move_home': False, 'password': None, 'generate_ssh_key': None, 'append': True, 'uid': None, 'ssh_key_comment': 'ansible-generated on ovh083', 'group': None, 'name': 'ubuntu', 'local': None, 'seuser': None, 'system': False, 'remove': False, 'state': 'present', 'ssh_key_file': None, 'login_class': None, 'shell': None, 'ssh_key_type': 'rsa'}}, 'changed': False, '_ansible_no_log': False, 'msg': 'Group kvm does not exist'}, 'ovh020.front.sepia.ceph.com': {'_ansible_parsed': True, 'invocation': {'module_args': {'comment': None, 'ssh_key_bits': 0, 'update_password': 'always', 'non_unique': False, 'force': False, 'skeleton': None, 'expires': None, 'ssh_key_passphrase': None, 'groups': ['fuse', 'kvm', 'disk'], 'createhome': True, 'home': None, 'move_home': False, 'password': None, 'generate_ssh_key': None, 'append': True, 'uid': None, 'ssh_key_comment': 'ansible-generated on ovh020', 'group': None, 'name': 'ubuntu', 'local': None, 'seuser': None, 'system': False, 'remove': False, 'state': 'present', 'ssh_key_file': None, 'login_class': None, 'shell': None, 'ssh_key_type': 'rsa'}}, 'changed': False, '_ansible_no_log': False, 'msg': 'Group kvm does not exist'}, 'ovh059.front.sepia.ceph.com': {'_ansible_parsed': True, 'invocation': {'module_args': {'comment': None, 'ssh_key_bits': 0, 'update_password': 'always', 'non_unique': False, 'force': False, 'skeleton': None, 'expires': None, 'ssh_key_passphrase': None, 'groups': ['fuse', 'kvm', 'disk'], 'createhome': True, 'home': None, 'move_home': False, 'password': None, 'generate_ssh_key': None, 'append': True, 'uid': None, 'ssh_key_comment': 'ansible-generated on ovh059', 'group': None, 'name': 'ubuntu', 'local': None, 'seuser': None, 'system': False, 'remove': False, 'state': 'present', 'ssh_key_file': None, 'login_class': None, 'shell': None, 'ssh_key_type': 'rsa'}}, 'changed': False, '_ansible_no_log': False, 'msg': 'Group kvm does not exist'}}

fail 3471385 2019-01-16 07:00:36 2019-01-16 15:38:19 2019-01-16 16:36:19 0:58:00 0:08:14 0:49:46 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rados_cls_all.yaml} 3
Failure Reason:

Command failed on ovh059 with status 1: '\n sudo yum -y install ceph\n '

fail 3471386 2019-01-16 07:00:37 2019-01-16 15:46:03 2019-01-16 16:50:04 1:04:01 0:08:04 0:55:57 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rados_ec_snaps.yaml} 3
Failure Reason:

Command failed on ovh053 with status 1: '\n sudo yum -y install ceph\n '

fail 3471387 2019-01-16 07:00:38 2019-01-16 15:52:12 2019-01-16 16:46:12 0:54:00 0:07:50 0:46:10 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rados_python.yaml} 3
Failure Reason:

Command failed on ovh003 with status 1: '\n sudo yum -y install ceph\n '

fail 3471388 2019-01-16 07:00:39 2019-01-16 16:01:49 2019-01-16 16:57:49 0:56:00 0:08:05 0:47:55 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rados_workunit_loadgen_mix.yaml} 3
Failure Reason:

Command failed on ovh061 with status 1: '\n sudo yum -y install ceph\n '

fail 3471389 2019-01-16 07:00:39 2019-01-16 16:01:49 2019-01-16 16:59:48 0:57:59 0:07:53 0:50:06 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rbd_api_tests.yaml} 3
Failure Reason:

Command failed on ovh020 with status 1: '\n sudo yum -y install ceph\n '

fail 3471390 2019-01-16 07:00:40 2019-01-16 16:01:49 2019-01-16 17:01:49 1:00:00 0:08:04 0:51:56 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rbd_cli_import_export.yaml} 3
Failure Reason:

Command failed on ovh037 with status 1: '\n sudo yum -y install ceph\n '

fail 3471391 2019-01-16 07:00:41 2019-01-16 16:01:49 2019-01-16 16:57:49 0:56:00 0:08:00 0:48:00 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rbd_fsx.yaml} 3
Failure Reason:

Command failed on ovh076 with status 1: '\n sudo yum -y install ceph\n '

fail 3471392 2019-01-16 07:00:41 2019-01-16 16:01:49 2019-01-16 17:05:49 1:04:00 0:08:01 0:55:59 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rbd_python_api_tests.yaml} 3
Failure Reason:

Command failed on ovh045 with status 1: '\n sudo yum -y install ceph\n '

fail 3471393 2019-01-16 07:00:42 2019-01-16 16:01:49 2019-01-16 17:01:49 1:00:00 0:08:08 0:51:52 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rbd_workunit_suites_iozone.yaml} 3
Failure Reason:

Command failed on ovh084 with status 1: '\n sudo yum -y install ceph\n '

fail 3471394 2019-01-16 07:00:43 2019-01-16 16:02:15 2019-01-16 17:06:15 1:04:00 0:08:20 0:55:40 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rgw_ec_s3tests.yaml} 3
Failure Reason:

Command failed on ovh065 with status 1: '\n sudo yum -y install ceph\n '

fail 3471395 2019-01-16 07:00:44 2019-01-16 16:36:31 2019-01-16 17:32:31 0:56:00 0:08:14 0:47:46 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rgw_s3tests.yaml} 3
Failure Reason:

Command failed on ovh040 with status 1: '\n sudo yum -y install ceph\n '

fail 3471396 2019-01-16 07:00:44 2019-01-16 16:46:21 2019-01-16 17:38:21 0:52:00 0:07:45 0:44:15 ovh master rhel 7.5 smoke/basic/{clusters/{fixed-3-cephfs.yaml openstack.yaml} objectstore/bluestore-bitmap.yaml tasks/rgw_swift.yaml} 3
Failure Reason:

Command failed on ovh009 with status 1: '\n sudo yum -y install ceph\n '