Name Machine Type Up Locked Locked Since Locked By OS Type OS Version Arch Description
mira105.front.sepia.ceph.com mira True True 2020-06-29 21:09:17.118488 kyrylo.shatskyy@suse.com centos 7 x86_64 None
Status Job ID Links Posted Started Updated
Runtime
Duration
In Waiting
Machine Teuthology Branch OS Type OS Version Description Nodes
fail 5129668 2020-06-08 17:02:07 2020-06-08 17:51:53 2020-06-08 18:11:53 0:20:00 0:06:52 0:13:08 mira py2 centos 7.8 ceph-deploy/basic/{ceph-deploy-overrides/disable_diff_journal_disk config_options/cephdeploy_conf distros/centos_7.8 objectstore/bluestore-bitmap python_versions/python_3 tasks/ceph-admin-commands} 2
Failure Reason:

Command failed on mira075 with status 5: 'sudo stop ceph-all || sudo service ceph stop || sudo systemctl stop ceph.target'

fail 5129655 2020-06-08 17:01:56 2020-06-08 17:29:52 2020-06-08 17:51:52 0:22:00 0:07:01 0:14:59 mira py2 centos 7.8 ceph-deploy/basic/{ceph-deploy-overrides/enable_diff_journal_disk config_options/cephdeploy_conf distros/centos_7.8 objectstore/filestore-xfs python_versions/python_3 tasks/ceph-admin-commands} 2
Failure Reason:

Command failed on mira075 with status 5: 'sudo stop ceph-all || sudo service ceph stop || sudo systemctl stop ceph.target'

fail 5129650 2020-06-08 17:01:51 2020-06-08 17:15:50 2020-06-08 17:29:49 0:13:59 0:04:04 0:09:55 mira py2 ubuntu 16.04 ceph-deploy/basic/{ceph-deploy-overrides/disable_diff_journal_disk config_options/cephdeploy_conf distros/ubuntu_16.04 objectstore/bluestore-bitmap python_versions/python_3 tasks/ceph-admin-commands} 2
Failure Reason:

Command failed on mira075 with status 5: 'sudo stop ceph-all || sudo service ceph stop || sudo systemctl stop ceph.target'

fail 5129645 2020-06-08 17:01:46 2020-06-08 17:01:50 2020-06-08 17:17:48 0:15:58 0:04:09 0:11:49 mira py2 ubuntu 16.04 ceph-deploy/basic/{ceph-deploy-overrides/disable_diff_journal_disk config_options/cephdeploy_conf distros/ubuntu_16.04 objectstore/filestore-xfs python_versions/python_2 tasks/ceph-admin-commands} 2
Failure Reason:

Command failed on mira075 with status 5: 'sudo stop ceph-all || sudo service ceph stop || sudo systemctl stop ceph.target'

pass 5128363 2020-06-08 05:55:41 2020-06-08 06:39:48 2020-06-08 07:55:49 1:16:01 0:27:21 0:48:40 mira master centos 7.8 ceph-deploy/{cluster/4node config/ceph_volume_filestore distros/centos_latest python_versions/python_3 tasks/rbd_import_export} 4
pass 5128354 2020-06-08 05:55:33 2020-06-08 06:17:31 2020-06-08 06:59:31 0:42:00 0:26:04 0:15:56 mira master centos 7.8 ceph-deploy/{cluster/4node config/ceph_volume_dmcrypt_off distros/centos_latest python_versions/python_3 tasks/ceph-admin-commands} 4
pass 5128347 2020-06-08 05:55:27 2020-06-08 05:55:29 2020-06-08 06:19:29 0:24:00 0:15:37 0:08:23 mira master ubuntu 18.04 ceph-deploy/{cluster/4node config/ceph_volume_filestore distros/ubuntu_latest python_versions/python_3 tasks/rbd_import_export} 4
fail 5122584 2020-06-06 20:43:34 2020-06-06 21:40:13 2020-06-06 21:52:12 0:11:59 0:05:46 0:06:13 mira master ubuntu rgw:verify/{clusters/fixed-2 frontend/beast msgr-failures/few objectstore/bluestore-bitmap overrides proto/http rgw_pool_type/replicated tasks/{0-install cls ragweed s3tests-java s3tests} validater/valgrind} 2
Failure Reason:

No module named 'cStringIO'

fail 5122570 2020-06-06 20:43:21 2020-06-06 21:25:56 2020-06-06 21:39:55 0:13:59 0:06:12 0:07:47 mira master ubuntu rgw:verify/{clusters/fixed-2 frontend/beast msgr-failures/few objectstore/filestore-xfs overrides proto/http rgw_pool_type/ec-profile tasks/{0-install cls ragweed s3tests-java s3tests} validater/valgrind} 2
Failure Reason:

No module named 'cStringIO'

fail 5122564 2020-06-06 20:43:16 2020-06-06 21:11:39 2020-06-06 21:25:39 0:14:00 0:06:22 0:07:38 mira master ubuntu rgw:verify/{clusters/fixed-2 frontend/beast msgr-failures/few objectstore/filestore-xfs overrides proto/https rgw_pool_type/ec-profile tasks/{0-install cls ragweed s3tests-java s3tests} validater/valgrind} 2
Failure Reason:

No module named 'cStringIO'

fail 5122555 2020-06-06 20:43:08 2020-06-06 20:57:23 2020-06-06 21:11:23 0:14:00 0:06:19 0:07:41 mira master ubuntu rgw:verify/{clusters/fixed-2 frontend/beast msgr-failures/few objectstore/filestore-xfs overrides proto/https rgw_pool_type/ec-profile tasks/{0-install cls ragweed s3tests-java s3tests} validater/lockdep} 2
Failure Reason:

No module named 'cStringIO'

fail 5122539 2020-06-06 20:42:53 2020-06-06 20:43:09 2020-06-06 20:57:08 0:13:59 0:06:29 0:07:30 mira master ubuntu rgw:verify/{clusters/fixed-2 frontend/beast msgr-failures/few objectstore/bluestore-bitmap overrides proto/http rgw_pool_type/replicated tasks/{0-install cls ragweed s3tests-java s3tests} validater/lockdep} 2
Failure Reason:

No module named 'cStringIO'

fail 5119080 2020-06-05 05:55:49 2020-06-05 06:52:00 2020-06-05 07:05:59 0:13:59 0:04:38 0:09:21 mira py2 ubuntu 16.04 ceph-deploy/basic/{ceph-deploy-overrides/ceph_deploy_dmcrypt config_options/cephdeploy_conf distros/ubuntu_16.04 objectstore/filestore-xfs python_versions/python_3 tasks/ceph-admin-commands} 2
Failure Reason:

Command failed on mira105 with status 5: 'sudo stop ceph-all || sudo service ceph stop || sudo systemctl stop ceph.target'

fail 5119067 2020-06-05 05:55:37 2020-06-05 06:23:40 2020-06-05 06:51:40 0:28:00 0:08:10 0:19:50 mira py2 centos 7.4 ceph-deploy/ceph-volume/{cluster/4node config/ceph_volume_filestore distros/centos_latest tasks/rbd_import_export} 4
Failure Reason:

Command failed on mira081 with status 5: 'sudo stop ceph-all || sudo service ceph stop || sudo systemctl stop ceph.target'

fail 5119059 2020-06-05 05:55:31 2020-06-05 06:09:40 2020-06-05 06:23:40 0:14:00 0:03:35 0:10:25 mira py2 ubuntu 16.04 ceph-deploy/basic/{ceph-deploy-overrides/disable_diff_journal_disk config_options/cephdeploy_conf distros/ubuntu_16.04 objectstore/bluestore-bitmap python_versions/python_3 tasks/ceph-admin-commands} 2
Failure Reason:

Command failed on mira105 with status 5: 'sudo stop ceph-all || sudo service ceph stop || sudo systemctl stop ceph.target'

fail 5119050 2020-06-05 05:55:23 2020-06-05 05:55:40 2020-06-05 06:09:39 0:13:59 0:04:05 0:09:54 mira py2 ubuntu 16.04 ceph-deploy/basic/{ceph-deploy-overrides/disable_diff_journal_disk config_options/cephdeploy_conf distros/ubuntu_16.04 objectstore/filestore-xfs python_versions/python_3 tasks/ceph-admin-commands} 2
Failure Reason:

Command failed on mira105 with status 5: 'sudo stop ceph-all || sudo service ceph stop || sudo systemctl stop ceph.target'

pass 5112569 2020-06-02 14:42:18 2020-06-02 16:14:32 2020-06-02 16:30:30 0:15:58 0:09:37 0:06:21 mira master ubuntu rgw/multifs/{clusters/fixed-2.yaml frontend/civetweb.yaml objectstore/bluestore-bitmap.yaml overrides.yaml rgw_pool_type/replicated.yaml tasks/rgw_user_quota.yaml} 2
pass 5112535 2020-06-02 14:41:44 2020-06-02 14:54:08 2020-06-02 15:16:07 0:21:59 0:16:07 0:05:52 mira master ubuntu rgw/thrash/{civetweb.yaml clusters/fixed-2.yaml install.yaml objectstore/filestore-xfs.yaml thrasher/default.yaml thrashosds-health.yaml workload/rgw_multipart_upload.yaml} 2
dead 5112528 2020-06-02 14:41:37 2020-06-02 14:44:09 2020-06-02 14:54:08 0:09:59 0:02:06 0:07:53 mira master ubuntu rgw/multifs/{clusters/fixed-2.yaml frontend/civetweb.yaml objectstore/filestore-xfs.yaml overrides.yaml rgw_pool_type/ec.yaml tasks/rgw_multipart_upload.yaml} 2
Failure Reason:

{'Failure object was': {'mira105.front.sepia.ceph.com': {'results': [{'changed': True, 'end': '2020-06-02 14:52:01.088445', 'stdout': '', 'cmd': 'pvremove --force --force --yes [unknown]', 'failed': True, 'delta': '0:00:00.026314', 'stderr': ' WARNING: Device for PV ZZvdP6-eUFS-ZltY-TE1a-2Q6s-jzn2-dFXCZG not found or rejected by a filter.\\n WARNING: Device for PV Y6tC2B-5AGA-7Dud-Trfw-0ooG-FFRl-XbGGhx not found or rejected by a filter.\\n WARNING: Device for PV vpRazo-dHKM-KD7S-u4mt-83v1-XCy4-mOtIC5 not found or rejected by a filter.\\n WARNING: Device for PV siDUZw-8Gp2-RyaW-uhTt-ExzY-NKQm-plAx2g not found or rejected by a filter.\\n WARNING: Device for PV tmTpJM-qEsx-XRo3-IRC8-5rX7-8nE9-CS0OH0 not found or rejected by a filter.\\n Device [unknown] not found.', 'rc': 5, 'invocation': {'module_args': {'creates': 'None', 'executable': 'None', '_uses_shell': True, 'strip_empty_ends': True, '_raw_params': 'pvremove --force --force --yes [unknown]', 'removes': 'None', 'argv': 'None', 'warn': True, 'chdir': 'None', 'stdin_add_newline': True, 'stdin': 'None'}}, 'start': '2020-06-02 14:52:01.062131', 'msg': 'non-zero return code', 'stdout_lines': [], 'stderr_lines': [' WARNING: Device for PV ZZvdP6-eUFS-ZltY-TE1a-2Q6s-jzn2-dFXCZG not found or rejected by a filter.', ' WARNING: Device for PV Y6tC2B-5AGA-7Dud-Trfw-0ooG-FFRl-XbGGhx not found or rejected by a filter.', ' WARNING: Device for PV vpRazo-dHKM-KD7S-u4mt-83v1-XCy4-mOtIC5 not found or rejected by a filter.', ' WARNING: Device for PV siDUZw-8Gp2-RyaW-uhTt-ExzY-NKQm-plAx2g not found or rejected by a filter.', ' WARNING: Device for PV tmTpJM-qEsx-XRo3-IRC8-5rX7-8nE9-CS0OH0 not found or rejected by a filter.', ' Device [unknown] not found.'], '_ansible_no_log': False, 'item': '[unknown]', 'ansible_loop_var': 'item', '_ansible_item_label': '[unknown]'}, {'changed': True, 'end': '2020-06-02 14:52:01.270647', 'stdout': ' Labels on physical volume "/dev/sdc" successfully wiped.', 'cmd': 'pvremove --force --force --yes /dev/sdc', 'rc': 0, 'start': '2020-06-02 14:52:01.239638', 'stderr': ' WARNING: Device for PV ZZvdP6-eUFS-ZltY-TE1a-2Q6s-jzn2-dFXCZG not found or rejected by a filter.\\n WARNING: Device for PV Y6tC2B-5AGA-7Dud-Trfw-0ooG-FFRl-XbGGhx not found or rejected by a filter.\\n WARNING: Device for PV vpRazo-dHKM-KD7S-u4mt-83v1-XCy4-mOtIC5 not found or rejected by a filter.\\n WARNING: Device for PV siDUZw-8Gp2-RyaW-uhTt-ExzY-NKQm-plAx2g not found or rejected by a filter.\\n WARNING: Device for PV tmTpJM-qEsx-XRo3-IRC8-5rX7-8nE9-CS0OH0 not found or rejected by a filter.\\n WARNING: PV /dev/sdc is used by VG vg_hdd.\\n WARNING: Wiping physical volume label from /dev/sdc of volume group "vg_hdd".', 'delta': '0:00:00.031009', 'invocation': {'module_args': {'creates': 'None', 'executable': 'None', '_uses_shell': True, 'strip_empty_ends': True, '_raw_params': 'pvremove --force --force --yes /dev/sdc', 'removes': 'None', 'argv': 'None', 'warn': True, 'chdir': 'None', 'stdin_add_newline': True, 'stdin': 'None'}}, 'stdout_lines': [' Labels on physical volume "/dev/sdc" successfully wiped.'], 'stderr_lines': [' WARNING: Device for PV ZZvdP6-eUFS-ZltY-TE1a-2Q6s-jzn2-dFXCZG not found or rejected by a filter.', ' WARNING: Device for PV Y6tC2B-5AGA-7Dud-Trfw-0ooG-FFRl-XbGGhx not found or rejected by a filter.', ' WARNING: Device for PV vpRazo-dHKM-KD7S-u4mt-83v1-XCy4-mOtIC5 not found or rejected by a filter.', ' WARNING: Device for PV siDUZw-8Gp2-RyaW-uhTt-ExzY-NKQm-plAx2g not found or rejected by a filter.', ' WARNING: Device for PV tmTpJM-qEsx-XRo3-IRC8-5rX7-8nE9-CS0OH0 not found or rejected by a filter.', ' WARNING: PV /dev/sdc is used by VG vg_hdd.', ' WARNING: Wiping physical volume label from /dev/sdc of volume group "vg_hdd".'], '_ansible_no_log': False, 'failed': False, 'item': '/dev/sdc', 'ansible_loop_var': 'item', '_ansible_item_label': '/dev/sdc'}, {'changed': True, 'end': '2020-06-02 14:52:01.447292', 'stdout': '', 'cmd': 'pvremove --force --force --yes [unknown]', 'failed': True, 'delta': '0:00:00.028064', 'stderr': ' Device [unknown] not found.', 'rc': 5, 'invocation': {'module_args': {'creates': 'None', 'executable': 'None', '_uses_shell': True, 'strip_empty_ends': True, '_raw_params': 'pvremove --force --force --yes [unknown]', 'removes': 'None', 'argv': 'None', 'warn': True, 'chdir': 'None', 'stdin_add_newline': True, 'stdin': 'None'}}, 'start': '2020-06-02 14:52:01.419228', 'msg': 'non-zero return code', 'stdout_lines': [], 'stderr_lines': [' Device [unknown] not found.'], '_ansible_no_log': False, 'item': '[unknown]', 'ansible_loop_var': 'item', '_ansible_item_label': '[unknown]'}, {'changed': True, 'end': '2020-06-02 14:52:01.630973', 'stdout': '', 'cmd': 'pvremove --force --force --yes [unknown]', 'failed': True, 'delta': '0:00:00.027971', 'stderr': ' Device [unknown] not found.', 'rc': 5, 'invocation': {'module_args': {'creates': 'None', 'executable': 'None', '_uses_shell': True, 'strip_empty_ends': True, '_raw_params': 'pvremove --force --force --yes [unknown]', 'removes': 'None', 'argv': 'None', 'warn': True, 'chdir': 'None', 'stdin_add_newline': True, 'stdin': 'None'}}, 'start': '2020-06-02 14:52:01.603002', 'msg': 'non-zero return code', 'stdout_lines': [], 'stderr_lines': [' Device [unknown] not found.'], '_ansible_no_log': False, 'item': '[unknown]', 'ansible_loop_var': 'item', '_ansible_item_label': '[unknown]'}, {'changed': True, 'end': '2020-06-02 14:52:01.803041', 'stdout': '', 'cmd': 'pvremove --force --force --yes [unknown]', 'failed': True, 'delta': '0:00:00.025692', 'stderr': ' Device [unknown] not found.', 'rc': 5, 'invocation': {'module_args': {'creates': 'None', 'executable': 'None', '_uses_shell': True, 'strip_empty_ends': True, '_raw_params': 'pvremove --force --force --yes [unknown]', 'removes': 'None', 'argv': 'None', 'warn': True, 'chdir': 'None', 'stdin_add_newline': True, 'stdin': 'None'}}, 'start': '2020-06-02 14:52:01.777349', 'msg': 'non-zero return code', 'stdout_lines': [], 'stderr_lines': [' Device [unknown] not found.'], '_ansible_no_log': False, 'item': '[unknown]', 'ansible_loop_var': 'item', '_ansible_item_label': '[unknown]'}, {'changed': True, 'end': '2020-06-02 14:52:01.982268', 'stdout': '', 'cmd': 'pvremove --force --force --yes [unknown]', 'failed': True, 'delta': '0:00:00.027148', 'stderr': ' Device [unknown] not found.', 'rc': 5, 'invocation': {'module_args': {'creates': 'None', 'executable': 'None', '_uses_shell': True, 'strip_empty_ends': True, '_raw_params': 'pvremove --force --force --yes [unknown]', 'removes': 'None', 'argv': 'None', 'warn': True, 'chdir': 'None', 'stdin_add_newline': True, 'stdin': 'None'}}, 'start': '2020-06-02 14:52:01.955120', 'msg': 'non-zero return code', 'stdout_lines': [], 'stderr_lines': [' Device [unknown] not found.'], '_ansible_no_log': False, 'item': '[unknown]', 'ansible_loop_var': 'item', '_ansible_item_label': '[unknown]'}], 'changed': True, 'msg': 'All items completed'}}, 'Traceback (most recent call last)': 'File "/home/teuthworker/src/git.ceph.com_git_ceph-cm-ansible_master/callback_plugins/failure_log.py", line 44, in log_failure log.error(yaml.safe_dump(failure)) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/lib/python3.6/site-packages/yaml/__init__.py", line 306, in safe_dump return dump_all([data], stream, Dumper=SafeDumper, **kwds) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/lib/python3.6/site-packages/yaml/__init__.py", line 278, in dump_all dumper.represent(data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/lib/python3.6/site-packages/yaml/representer.py", line 27, in represent node = self.represent_data(data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/lib/python3.6/site-packages/yaml/representer.py", line 48, in represent_data node = self.yaml_representers[data_types[0]](self, data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/lib/python3.6/site-packages/yaml/representer.py", line 207, in represent_dict return self.represent_mapping(\'tag:yaml.org,2002:map\', data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/lib/python3.6/site-packages/yaml/representer.py", line 118, in represent_mapping node_value = self.represent_data(item_value) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/lib/python3.6/site-packages/yaml/representer.py", line 48, in represent_data node = self.yaml_representers[data_types[0]](self, data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/lib/python3.6/site-packages/yaml/representer.py", line 207, in represent_dict return self.represent_mapping(\'tag:yaml.org,2002:map\', data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/lib/python3.6/site-packages/yaml/representer.py", line 118, in represent_mapping node_value = self.represent_data(item_value) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/lib/python3.6/site-packages/yaml/representer.py", line 48, in represent_data node = self.yaml_representers[data_types[0]](self, data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/lib/python3.6/site-packages/yaml/representer.py", line 199, in represent_list return self.represent_sequence(\'tag:yaml.org,2002:seq\', data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/lib/python3.6/site-packages/yaml/representer.py", line 92, in represent_sequence node_item = self.represent_data(item) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/lib/python3.6/site-packages/yaml/representer.py", line 48, in represent_data node = self.yaml_representers[data_types[0]](self, data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/lib/python3.6/site-packages/yaml/representer.py", line 207, in represent_dict return self.represent_mapping(\'tag:yaml.org,2002:map\', data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/lib/python3.6/site-packages/yaml/representer.py", line 118, in represent_mapping node_value = self.represent_data(item_value) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/lib/python3.6/site-packages/yaml/representer.py", line 58, in represent_data node = self.yaml_representers[None](self, data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_master/virtualenv/lib/python3.6/site-packages/yaml/representer.py", line 231, in represent_undefined raise RepresenterError("cannot represent an object", data)', 'yaml.representer.RepresenterError': "('cannot represent an object', '[unknown]')"}

pass 5109662 2020-06-01 05:55:41 2020-06-01 14:52:27 2020-06-01 16:10:28 1:18:01 0:15:14 1:02:47 mira py2 ubuntu 18.04 ceph-deploy/{cluster/4node config/ceph_volume_bluestore distros/ubuntu_latest python_versions/python_3 tasks/rbd_import_export} 4