ID
Status
Ceph Branch
Suite Branch
Teuthology Branch
Machine
OS
Nodes
Description
Failure Reason
wip-rishabh-2023May15-1524
wip-rishabh-2023May15-1524
main
smithi
ubuntu 22.04
fs/functional/{begin/{0-install 1-ceph 2-logrotate} clusters/1a3s-mds-4c-client conf/{client mds mon osd} distro/{ubuntu_latest} mount/kclient/{mount-syntax/{v1} mount overrides/{distro/testing/k-testing ms-die-on-skipped}} objectstore/bluestore-ec-root overrides/{ignorelist_health ignorelist_wrongly_marked_down no_client_pidfile} tasks/snap-schedule}
"1684324202.6427338 mon.a (mon.0) 404 : cluster [WRN] Health check failed: Reduced data availability: 2 pgs peering (PG_AVAILABILITY)" in cluster log
wip-rishabh-2023May15-1524
wip-rishabh-2023May15-1524
main
smithi
ubuntu 20.04
fs/thrash/workloads/{begin/{0-install 1-ceph 2-logrotate} clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{ubuntu_20.04} mount/fuse msgr-failures/none objectstore-ec/bluestore-ec-root overrides/{frag ignorelist_health ignorelist_wrongly_marked_down prefetch_dirfrags/yes prefetch_entire_dirfrags/no races session_timeout thrashosds-health} ranks/3 tasks/{1-thrash/mon 2-workunit/fs/trivial_sync}}
wip-rishabh-2023May15-1524
wip-rishabh-2023May15-1524
main
smithi
rhel 8.6
fs/workload/{0-rhel_8 begin/{0-install 1-cephadm 2-logrotate} clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} mount/fuse objectstore-ec/bluestore-comp-ec-root omap_limit/10000 overrides/{cephsqlite-timeout frag ignorelist_health ignorelist_wrongly_marked_down osd-asserts session_timeout} ranks/1 standby-replay tasks/{0-subvolume/{with-namespace-isolated-and-quota} 1-check-counter 2-scrub/no 3-snaps/yes 4-flush/no 5-workunit/suites/pjd}}
Command failed (workunit test suites/pjd.sh) on smithi114 with status 124: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=27be51263d3d0c721b6d1ca9f42550bf0ab3c97b TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 CEPH_MNT=/home/ubuntu/cephtest/mnt.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 6h /home/ubuntu/cephtest/clone.client.0/qa/workunits/suites/pjd.sh'
wip-rishabh-2023May15-1524
wip-rishabh-2023May15-1524
main
smithi
centos 8.stream
fs/upgrade/mds_upgrade_sequence/{bluestore-bitmap centos_8.stream_container_tools conf/{client mds mon osd} fail_fs/yes overrides/{ignorelist_health ignorelist_wrongly_marked_down pg-warn syntax} roles tasks/{0-from/pacific 1-volume/{0-create 1-ranks/1 2-allow_standby_replay/yes 3-inline/yes 4-verify} 2-client 3-upgrade-mgr-staggered 4-config-upgrade/{fail_fs} 5-upgrade-with-workload 6-verify}}
wip-rishabh-2023May15-1524
wip-rishabh-2023May15-1524
main
smithi
ubuntu 22.04
fs/32bits/{begin/{0-install 1-ceph 2-logrotate} clusters/fixed-2-ucephfs conf/{client mds mon osd} distro/{ubuntu_latest} mount/fuse objectstore-ec/bluestore-bitmap overrides/{faked-ino ignorelist_health ignorelist_wrongly_marked_down} tasks/cfuse_workunit_suites_fsstress}
wip-rishabh-2023May15-1524
wip-rishabh-2023May15-1524
main
smithi
ubuntu 20.04
fs/full/{begin/{0-install 1-ceph 2-logrotate} clusters/1-node-1-mds-1-osd conf/{client mds mon osd} distro/{ubuntu_20.04} mount/fuse objectstore/bluestore-bitmap overrides overrides/{ignorelist_health ignorelist_wrongly_marked_down no_client_pidfile} tasks/mgr-osd-full}
wip-rishabh-2023May15-1524
wip-rishabh-2023May15-1524
main
smithi
centos 8.stream
fs/libcephfs/{begin/{0-install 1-ceph 2-logrotate} clusters/1-mds-1-client-coloc conf/{client mds mon osd} distro/{centos_8} objectstore/bluestore-bitmap overrides/{ignorelist_health ignorelist_wrongly_marked_down no_client_pidfile} tasks/client}
Command failed (workunit test client/test.sh) on smithi081 with status 124: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=27be51263d3d0c721b6d1ca9f42550bf0ab3c97b TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 CEPH_MNT=/home/ubuntu/cephtest/mnt.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/client/test.sh'
wip-rishabh-2023May15-1524
wip-rishabh-2023May15-1524
main
smithi
ubuntu 22.04
fs/multiclient/{begin/{0-install 1-ceph 2-logrotate} clusters/1-mds-3-client conf/{client mds mon osd} distros/ubuntu_latest mount/fuse objectstore-ec/bluestore-comp overrides/{ignorelist_health ignorelist_wrongly_marked_down} tasks/cephfs_misc_tests}
wip-rishabh-2023May15-1524
wip-rishabh-2023May15-1524
main
smithi
ubuntu 20.04
fs/shell/{begin/{0-install 1-ceph 2-logrotate} clusters/1-mds-1-client-coloc conf/{client mds mon osd} distro/{ubuntu_latest} mount/fuse objectstore/bluestore-bitmap overrides/{ignorelist_health ignorelist_wrongly_marked_down no_client_pidfile} tasks/cephfs-shell}
wip-rishabh-2023May15-1524
wip-rishabh-2023May15-1524
main
smithi
ubuntu 22.04
fs/snaps/{begin/{0-install 1-ceph 2-logrotate} clusters/1a3s-mds-1c-client conf/{client mds mon osd} distro/{ubuntu_latest} mount/fuse objectstore-ec/bluestore-comp overrides/{ignorelist_health ignorelist_wrongly_marked_down} tasks/workunit/snaps}
wip-rishabh-2023May15-1524
wip-rishabh-2023May15-1524
main
smithi
rhel 8.6
fs/workload/{0-rhel_8 begin/{0-install 1-cephadm 2-logrotate} clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} mount/kclient/{base/{mount-syntax/{v2} mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} ms_mode/secure wsync/no} objectstore-ec/bluestore-bitmap omap_limit/10 overrides/{cephsqlite-timeout frag ignorelist_health ignorelist_wrongly_marked_down osd-asserts session_timeout} ranks/multi/{export-check n/5 replication/always} standby-replay tasks/{0-subvolume/{with-no-extra-options} 1-check-counter 2-scrub/yes 3-snaps/no 4-flush/yes 5-workunit/kernel_untar_build}}
{'smithi194.front.sepia.ceph.com': {'changed': False, 'msg': 'All items completed', 'results': [{'_ansible_item_label': 'zcerza', '_ansible_no_log': False, 'ansible_loop_var': 'item', 'changed': False, 'failed': False, 'invocation': {'module_args': {'append': False, 'authorization': None, 'comment': None, 'create_home': True, 'expires': None, 'force': False, 'generate_ssh_key': None, 'group': None, 'groups': None, 'hidden': None, 'home': None, 'local': None, 'login_class': None, 'move_home': False, 'name': 'zcerza', 'non_unique': False, 'password': None, 'password_lock': None, 'profile': None, 'remove': False, 'role': None, 'seuser': None, 'shell': None, 'skeleton': None, 'ssh_key_bits': 0, 'ssh_key_comment': 'ansible-generated on smithi194', 'ssh_key_file': None, 'ssh_key_passphrase': None, 'ssh_key_type': 'rsa', 'state': 'absent', 'system': False, 'uid': None, 'update_password': 'always'}}, 'item': 'zcerza', 'name': 'zcerza', 'state': 'absent'}, {'_ansible_item_label': 'aschoen', '_ansible_no_log': False, 'ansible_loop_var': 'item', 'changed': False, 'failed': False, 'invocation': {'module_args': {'append': False, 'authorization': None, 'comment': None, 'create_home': True, 'expires': None, 'force': False, 'generate_ssh_key': None, 'group': None, 'groups': None, 'hidden': None, 'home': None, 'local': None, 'login_class': None, 'move_home': False, 'name': 'aschoen', 'non_unique': False, 'password': None, 'password_lock': None, 'profile': None, 'remove': False, 'role': None, 'seuser': None, 'shell': None, 'skeleton': None, 'ssh_key_bits': 0, 'ssh_key_comment': 'ansible-generated on smithi194', 'ssh_key_file': None, 'ssh_key_passphrase': None, 'ssh_key_type': 'rsa', 'state': 'absent', 'system': False, 'uid': None, 'update_password': 'always'}}, 'item': 'aschoen', 'name': 'aschoen', 'state': 'absent'}, {'_ansible_item_label': 'andrew', '_ansible_no_log': False, 'ansible_loop_var': 'item', 'changed': False, 'failed': False, 'invocation': {'module_args': {'append': False, 'authorization': None, 'comment': None, 'create_home': True, 'expires': None, 'force': False, 'generate_ssh_key': None, 'group': None, 'groups': None, 'hidden': None, 'home': None, 'local': None, 'login_class': None, 'move_home': False, 'name': 'andrew', 'non_unique': False, 'password': None, 'password_lock': None, 'profile': None, 'remove': False, 'role': None, 'seuser': None, 'shell': None, 'skeleton': None, 'ssh_key_bits': 0, 'ssh_key_comment': 'ansible-generated on smithi194', 'ssh_key_file': None, 'ssh_key_passphrase': None, 'ssh_key_type': 'rsa', 'state': 'absent', 'system': False, 'uid': None, 'update_password': 'always'}}, 'item': 'andrew', 'name': 'andrew', 'state': 'absent'}, {'_ansible_item_label': 'sweil', '_ansible_no_log': False, 'ansible_loop_var': 'item', 'changed': False, 'failed': False, 'invocation': {'module_args': {'append': False, 'authorization': None, 'comment': None, 'create_home': True, 'expires': None, 'force': False, 'generate_ssh_key': None, 'group': None, 'groups': None, 'hidden': None, 'home': None, 'local': None, 'login_class': None, 'move_home': False, 'name': 'sweil', 'non_unique': False, 'password': None, 'password_lock': None, 'profile': None, 'remove': False, 'role': None, 'seuser': None, 'shell': None, 'skeleton': None, 'ssh_key_bits': 0, 'ssh_key_comment': 'ansible-generated on smithi194', 'ssh_key_file': None, 'ssh_key_passphrase': None, 'ssh_key_type': 'rsa', 'state': 'absent', 'system': False, 'uid': None, 'update_password': 'always'}}, 'item': 'sweil', 'name': 'sweil', 'state': 'absent'}, {'_ansible_item_label': 'brad', '_ansible_no_log': False, 'ansible_loop_var': 'item', 'changed': False, 'failed': False, 'invocation': {'module_args': {'append': False, 'authorization': None, 'comment': None, 'create_home': True, 'expires': None, 'force': False, 'generate_ssh_key': None, 'group': None, 'groups': None, 'hidden': None, 'home': None, 'local': None, 'login_class': None, 'move_home': False, 'name': 'brad', 'non_unique': False, 'password': None, 'password_lock': None, 'profile': None, 'remove': False, 'role': None, 'seuser': None, 'shell': None, 'skeleton': None, 'ssh_key_bits': 0, 'ssh_key_comment': 'ansible-generated on smithi194', 'ssh_key_file': None, 'ssh_key_passphrase': None, 'ssh_key_type': 'rsa', 'state': 'absent', 'system': False, 'uid': None, 'update_password': 'always'}}, 'item': 'brad', 'name': 'brad', 'state': 'absent'}, {'_ansible_item_label': 'kefu', '_ansible_no_log': False, 'ansible_loop_var': 'item', 'changed': False, 'failed': False, 'invocation': {'module_args': {'append': False, 'authorization': None, 'comment': None, 'create_home': True, 'expires': None, 'force': False, 'generate_ssh_key': None, 'group': None, 'groups': None, 'hidden': None, 'home': None, 'local': None, 'login_class': None, 'move_home': False, 'name': 'kefu', 'non_unique': False, 'password': None, 'password_lock': None, 'profile': None, 'remove': False, 'role': None, 'seuser': None, 'shell': None, 'skeleton': None, 'ssh_key_bits': 0, 'ssh_key_comment': 'ansible-generated on smithi194', 'ssh_key_file': None, 'ssh_key_passphrase': None, 'ssh_key_type': 'rsa', 'state': 'absent', 'system': False, 'uid': None, 'update_password': 'always'}}, 'item': 'kefu', 'name': 'kefu', 'state': 'absent'}, {'_ansible_item_label': 'shylesh', '_ansible_no_log': False, 'ansible_loop_var': 'item', 'changed': False, 'failed': False, 'invocation': {'module_args': {'append': False, 'authorization': None, 'comment': None, 'create_home': True, 'expires': None, 'force': False, 'generate_ssh_key': None, 'group': None, 'groups': None, 'hidden': None, 'home': None, 'local': None, 'login_class': None, 'move_home': False, 'name': 'shylesh', 'non_unique': False, 'password': None, 'password_lock': None, 'profile': None, 'remove': False, 'role': None, 'seuser': None, 'shell': None, 'skeleton': None, 'ssh_key_bits': 0, 'ssh_key_comment': 'ansible-generated on smithi194', 'ssh_key_file': None, 'ssh_key_passphrase': None, 'ssh_key_type': 'rsa', 'state': 'absent', 'system': False, 'uid': None, 'update_password': 'always'}}, 'item': 'shylesh', 'name': 'shylesh', 'state': 'absent'}, {'_ansible_item_label': 'gmeno', '_ansible_no_log': False, 'ansible_loop_var': 'item', 'changed': False, 'failed': False, 'invocation': {'module_args': {'append': False, 'authorization': None, 'comment': None, 'create_home': True, 'expires': None, 'force': False, 'generate_ssh_key': None, 'group': None, 'groups': None, 'hidden': None, 'home': None, 'local': None, 'login_class': None, 'move_home': False, 'name': 'gmeno', 'non_unique': False, 'password': None, 'password_lock': None, 'profile': None, 'remove': False, 'role': None, 'seuser': None, 'shell': None, 'skeleton': None, 'ssh_key_bits': 0, 'ssh_key_comment': 'ansible-generated on smithi194', 'ssh_key_file': None, 'ssh_key_passphrase': None, 'ssh_key_type': 'rsa', 'state': 'absent', 'system': False, 'uid': None, 'update_password': 'always'}}, 'item': 'gmeno', 'name': 'gmeno', 'state': 'absent'}, {'_ansible_item_label': 'alfredodeza', '_ansible_no_log': False, 'ansible_loop_var': 'item', 'changed': False, 'failed': False, 'invocation': {'module_args': {'append': False, 'authorization': None, 'comment': None, 'create_home': True, 'expires': None, 'force': False, 'generate_ssh_key': None, 'group': None, 'groups': None, 'hidden': None, 'home': None, 'local': None, 'login_class': None, 'move_home': False, 'name': 'alfredodeza', 'non_unique': False, 'password': None, 'password_lock': None, 'profile': None, 'remove': False, 'role': None, 'seuser': None, 'shell': None, 'skeleton': None, 'ssh_key_bits': 0, 'ssh_key_comment': 'ansible-generated on smithi194', 'ssh_key_file': None, 'ssh_key_passphrase': None, 'ssh_key_type': 'rsa', 'state': 'absent', 'system': False, 'uid': None, 'update_password': 'always'}}, 'item': 'alfredodeza', 'name': 'alfredodeza', 'state': 'absent'}, {'_ansible_item_label': 'vumrao', '_ansible_no_log': False, 'ansible_loop_var': 'item', 'changed': False, 'failed': False, 'invocation': {'module_args': {'append': False, 'authorization': None, 'comment': None, 'create_home': True, 'expires': None, 'force': False, 'generate_ssh_key': None, 'group': None, 'groups': None, 'hidden': None, 'home': None, 'local': None, 'login_class': None, 'move_home': False, 'name': 'vumrao', 'non_unique': False, 'password': None, 'password_lock': None, 'profile': None, 'remove': False, 'role': None, 'seuser': None, 'shell': None, 'skeleton': None, 'ssh_key_bits': 0, 'ssh_key_comment': 'ansible-generated on smithi194', 'ssh_key_file': None, 'ssh_key_passphrase': None, 'ssh_key_type': 'rsa', 'state': 'absent', 'system': False, 'uid': None, 'update_password': 'always'}}, 'item': 'vumrao', 'name': 'vumrao', 'state': 'absent'}, {'_ansible_item_label': 'trhoden', '_ansible_no_log': False, 'ansible_loop_var': 'item', 'changed': False, 'failed': False, 'invocation': {'module_args': {'append': False, 'authorization': None, 'comment': None, 'create_home': True, 'expires': None, 'force': False, 'generate_ssh_key': None, 'group': None, 'groups': None, 'hidden': None, 'home': None, 'local': None, 'login_class': None, 'move_home': False, 'name': 'trhoden', 'non_unique': False, 'password': None, 'password_lock': None, 'profile': None, 'remove': False, 'role': None, 'seuser': None, 'shell': None, 'skeleton': None, 'ssh_key_bits': 0, 'ssh_key_comment': 'ansible-generated on smithi194', 'ssh_key_file': None, 'ssh_key_passphrase': None, 'ssh_key_type': 'rsa', 'state': 'absent', 'system': False, 'uid': None, 'update_password': 'always'}}, 'item': 'trhoden', 'name': 'trhoden', 'state': 'absent'}, {'_ansible_item_label': 'nishtha', '_ansible_no_log': False, 'ansible_loop_var': 'item', 'changed': False, 'failed': False, 'invocation': {'module_args': {'append': False, 'authorization': None, 'comment': None, 'create_home': True, 'expires': None, 'force': False, 'generate_ssh_key': None, 'group': None, 'groups': None, 'hidden': None, 'home': None, 'local': None, 'login_class': None, 'move_home': False, 'name': 'nishtha', 'non_unique': False, 'password': None, 'password_lock': None, 'profile': None, 'remove': False, 'role': None, 'seuser': None, 'shell': None, 'skeleton': None, 'ssh_key_bits': 0, 'ssh_key_comment': 'ansible-generated on smithi194', 'ssh_key_file': None, 'ssh_key_passphrase': None, 'ssh_key_type': 'rsa', 'state': 'absent', 'system': False, 'uid': None, 'update_password': 'always'}}, 'item': 'nishtha', 'name': 'nishtha', 'state': 'absent'}, {'_ansible_item_label': 'yguang', '_ansible_no_log': False, 'ansible_loop_var': 'item', 'changed': False, 'failed': False, 'invocation': {'module_args': {'append': False, 'authorization': None, 'comment': None, 'create_home': True, 'expires': None, 'force': False, 'generate_ssh_key': None, 'group': None, 'groups': None, 'hidden': None, 'home': None, 'local': None, 'login_class': None, 'move_home': False, 'name': 'yguang', 'non_unique': False, 'password': None, 'password_lock': None, 'profile': None, 'remove': False, 'role': None, 'seuser': None, 'shell': None, 'skeleton': None, 'ssh_key_bits': 0, 'ssh_key_comment': 'ansible-generated on smithi194', 'ssh_key_file': None, 'ssh_key_passphrase': None, 'ssh_key_type': 'rsa', 'state': 'absent', 'system': False, 'uid': None, 'update_password': 'always'}}, 'item': 'yguang', 'name': 'yguang', 'state': 'absent'}, {'_ansible_item_label': 'sdieffen', '_ansible_no_log': False, 'ansible_loop_var': 'item', 'changed': False, 'failed': False, 'invocation': {'module_args': {'append': False, 'authorization': None, 'comment': None, 'create_home': True, 'expires': None, 'force': False, 'generate_ssh_key': None, 'group': None, 'groups': None, 'hidden': None, 'home': None, 'local': None, 'login_class': None, 'move_home': False, 'name': 'sdieffen', 'non_unique': False, 'password': None, 'password_lock': None, 'profile': None, 'remove': False, 'role': None, 'seuser': None, 'shell': None, 'skeleton': None, 'ssh_key_bits': 0, 'ssh_key_comment': 'ansible-generated on smithi194', 'ssh_key_file': None, 'ssh_key_passphrase': None, 'ssh_key_type': 'rsa', 'state': 'absent', 'system': False, 'uid': None, 'update_password': 'always'}}, 'item': 'sdieffen', 'name': 'sdieffen', 'state': 'absent'}, {'_ansible_item_label': 'brian', '_ansible_no_log': False, 'ansible_loop_var': 'item', 'changed': False, 'failed': False, 'invocation': {'module_args': {'append': False, 'authorization': None, 'comment': None, 'create_home': True, 'expires': None, 'force': False, 'generate_ssh_key': None, 'group': None, 'groups': None, 'hidden': None, 'home': None, 'local': None, 'login_class': None, 'move_home': False, 'name': 'brian', 'non_unique': False, 'password': None, 'password_lock': None, 'profile': None, 'remove': False, 'role': None, 'seuser': None, 'shell': None, 'skeleton': None, 'ssh_key_bits': 0, 'ssh_key_comment': 'ansible-generated on smithi194', 'ssh_key_file': None, 'ssh_key_passphrase': None, 'ssh_key_type': 'rsa', 'state': 'absent', 'system': False, 'uid': None, 'update_password': 'always'}}, 'item': 'brian', 'name': 'brian', 'state': 'absent'}, {'_ansible_item_label': 'pmcgarry', '_ansible_no_log': False, 'ansible_loop_var': 'item', 'changed': False, 'failed': False, 'invocation': {'module_args': {'append': False, 'authorization': None, 'comment': None, 'create_home': True, 'expires': None, 'force': False, 'generate_ssh_key': None, 'group': None, 'groups': None, 'hidden': None, 'home': None, 'local': None, 'login_class': None, 'move_home': False, 'name': 'pmcgarry', 'non_unique': False, 'password': None, 'password_lock': None, 'profile': None, 'remove': False, 'role': None, 'seuser': None, 'shell': None, 'skeleton': None, 'ssh_key_bits': 0, 'ssh_key_comment': 'ansible-generated on smithi194', 'ssh_key_file': None, 'ssh_key_passphrase': None, 'ssh_key_type': 'rsa', 'state': 'absent', 'system': False, 'uid': None, 'update_password': 'always'}}, 'item': 'pmcgarry', 'name': 'pmcgarry', 'state': 'absent'}, {'_ansible_item_label': 'karnan', '_ansible_no_log': False, 'ansible_loop_var': 'item', 'changed': False, 'failed': False, 'invocation': {'module_args': {'append': False, 'authorization': None, 'comment': None, 'create_home': True, 'expires': None, 'force': False, 'generate_ssh_key': None, 'group': None, 'groups': None, 'hidden': None, 'home': None, 'local': None, 'login_class': None, 'move_home': False, 'name': 'karnan', 'non_unique': False, 'password': None, 'password_lock': None, 'profile': None, 'remove': False, 'role': None, 'seuser': None, 'shell': None, 'skeleton': None, 'ssh_key_bits': 0, 'ssh_key_comment': 'ansible-generated on smithi194', 'ssh_key_file': None, 'ssh_key_passphrase': None, 'ssh_key_type': 'rsa', 'state': 'absent', 'system': False, 'uid': None, 'update_password': 'always'}}, 'item': 'karnan', 'name': 'karnan', 'state': 'absent'}, {'_ansible_item_label': 'ryneli', '_ansible_no_log': False, 'ansible_loop_var': 'item', 'changed': False, 'failed': False, 'invocation': {'module_args': {'append': False, 'authorization': None, 'comment': None, 'create_home': True, 'expires': None, 'force': False, 'generate_ssh_key': None, 'group': None, 'groups': None, 'hidden': None, 'home': None, 'local': None, 'login_class': None, 'move_home': False, 'name': 'ryneli', 'non_unique': False, 'password': None, 'password_lock': None, 'profile': None, 'remove': False, 'role': None, 'seuser': None, 'shell': None, 'skeleton': None, 'ssh_key_bits': 0, 'ssh_key_comment': 'ansible-generated on smithi194', 'ssh_key_file': None, 'ssh_key_passphrase': None, 'ssh_key_type': 'rsa', 'state': 'absent', 'system': False, 'uid': None, 'update_password': 'always'}}, 'item': 'ryneli', 'name': 'ryneli', 'state': 'absent'}, {'_ansible_item_label': 'dlambrig', '_ansible_no_log': False, 'ansible_loop_var': 'item', 'changed': False, 'failed': False, 'invocation': {'module_args': {'append': False, 'authorization': None, 'comment': None, 'create_home': True, 'expires': None, 'force': False, 'generate_ssh_key': None, 'group': None, 'groups': None, 'hidden': None, 'home': None, 'local': None, 'login_class': None, 'move_home': False, 'name': 'dlambrig', 'non_unique': False, 'password': None, 'password_lock': None, 'profile': None, 'remove': False, 'role': None, 'seuser': None, 'shell': None, 'skeleton': None, 'ssh_key_bits': 0, 'ssh_key_comment': 'ansible-generated on smithi194', 'ssh_key_file': None, 'ssh_key_passphrase': None, 'ssh_key_type': 'rsa', 'state': 'absent', 'system': False, 'uid': None, 'update_password': 'always'}}, 'item': 'dlambrig', 'name': 'dlambrig', 'state': 'absent'}, {'_ansible_item_label': 'icolle', '_ansible_no_log': False, 'ansible_loop_var': 'item', 'changed': False, 'failed': False, 'invocation': {'module_args': {'append': False, 'authorization': None, 'comment': None, 'create_home': True, 'expires': None, 'force': False, 'generate_ssh_key': None, 'group': None, 'groups': None, 'hidden': None, 'home': None, 'local': None, 'login_class': None, 'move_home': False, 'name': 'icolle', 'non_unique': False, 'password': None, 'password_lock': None, 'profile': None, 'remove': False, 'role': None, 'seuser': None, 'shell': None, 'skeleton': None, 'ssh_key_bits': 0, 'ssh_key_comment': 'ansible-generated on smithi194', 'ssh_key_file': None, 'ssh_key_passphrase': None, 'ssh_key_type': 'rsa', 'state': 'absent', 'system': False, 'uid': None, 'update_password': 'always'}}, 'item': 'icolle', 'name': 'icolle', 'state': 'absent'}, {'_ansible_item_label': 'soumya', '_ansible_no_log': False, 'ansible_loop_var': 'item', 'changed': False, 'failed': False, 'invocation': {'module_args': {'append': False, 'authorization': None, 'comment': None, 'create_home': True, 'expires': None, 'force': False, 'generate_ssh_key': None, 'group': None, 'groups': None, 'hidden': None, 'home': None, 'local': None, 'login_class': None, 'move_home': False, 'name': 'soumya', 'non_unique': False, 'password': None, 'password_lock': None, 'profile': None, 'remove': False, 'role': None, 'seuser': None, 'shell': None, 'skeleton': None, 'ssh_key_bits': 0, 'ssh_key_comment': 'ansible-generated on smithi194', 'ssh_key_file': None, 'ssh_key_passphrase': None, 'ssh_key_type': 'rsa', 'state': 'absent', 'system': False, 'uid': None, 'update_password': 'always'}}, 'item': 'soumya', 'name': 'soumya', 'state': 'absent'}, {'_ansible_item_label': 'jspray', '_ansible_no_log': False, 'ansible_loop_var': 'item', 'changed': False, 'failed': False, 'invocation': {'module_args': {'append': False, 'authorization': None, 'comment': None, 'create_home': True, 'expires': None, 'force': False, 'generate_ssh_key': None, 'group': None, 'groups': None, 'hidden': None, 'home': None, 'local': None, 'login_class': None, 'move_home': False, 'name': 'jspray', 'non_unique': False, 'password': None, 'password_lock': None, 'profile': None, 'remove': False, 'role': None, 'seuser': None, 'shell': None, 'skeleton': None, 'ssh_key_bits': 0, 'ssh_key_comment': 'ansible-generated on smithi194', 'ssh_key_file': None, 'ssh_key_passphrase': None, 'ssh_key_type': 'rsa', 'state': 'absent', 'system': False, 'uid': None, 'update_password': 'always'}}, 'item': 'jspray', 'name': 'jspray', 'state': 'absent'}, {'_ansible_item_label': 'erwan', 'ansible_loop_var': 'item', 'item': 'erwan', 'msg': 'Failed to connect to the host via ssh: ssh: connect to host smithi194.front.sepia.ceph.com port 22: No route to host', 'unreachable': True}, {'_ansible_item_label': 'jj', 'ansible_loop_var': 'item', 'item': 'jj', 'msg': 'Failed to connect to the host via ssh: ssh: connect to host smithi194.front.sepia.ceph.com port 22: No route to host', 'unreachable': True}, {'_ansible_item_label': 'amarangone', 'ansible_loop_var': 'item', 'item': 'amarangone', 'msg': 'Failed to connect to the host via ssh: ssh: connect to host smithi194.front.sepia.ceph.com port 22: No route to host', 'unreachable': True}, {'_ansible_item_label': 'oprypin', 'ansible_loop_var': 'item', 'item': 'oprypin', 'msg': 'Failed to connect to the host via ssh: ssh: connect to host smithi194.front.sepia.ceph.com port 22: No route to host', 'unreachable': True}, {'_ansible_item_label': 'adamyanova', 'ansible_loop_var': 'item', 'item': 'adamyanova', 'msg': "Failed to connect to the host via ssh: Warning: Permanently added 'smithi194.front.sepia.ceph.com,172.21.15.194' (ECDSA) to the list of known hosts.\r\nubuntu@smithi194.front.sepia.ceph.com: Permission denied (publickey,password,keyboard-interactive).", 'unreachable': True}, {'_ansible_item_label': 'sbillah', 'ansible_loop_var': 'item', 'item': 'sbillah', 'msg': "Failed to connect to the host via ssh: Warning: Permanently added 'smithi194.front.sepia.ceph.com,172.21.15.194' (ECDSA) to the list of known hosts.\r\nubuntu@smithi194.front.sepia.ceph.com: Permission denied (publickey,password,keyboard-interactive).", 'unreachable': True}, {'_ansible_item_label': 'onyb', 'ansible_loop_var': 'item', 'item': 'onyb', 'msg': 'Failed to connect to the host via ssh: ssh: connect to host smithi194.front.sepia.ceph.com port 22: No route to host', 'unreachable': True}, {'_ansible_item_label': 'jwilliamson', '_ansible_no_log': False, 'ansible_loop_var': 'item', 'changed': False, 'failed': False, 'invocation': {'module_args': {'append': False, 'authorization': None, 'comment': None, 'create_home': True, 'expires': None, 'force': False, 'generate_ssh_key': None, 'group': None, 'groups': None, 'hidden': None, 'home': None, 'local': None, 'login_class': None, 'move_home': False, 'name': 'jwilliamson', 'non_unique': False, 'password': None, 'password_lock': None, 'profile': None, 'remove': False, 'role': None, 'seuser': None, 'shell': None, 'skeleton': None, 'ssh_key_bits': 0, 'ssh_key_comment': 'ansible-generated on smithi194', 'ssh_key_file': None, 'ssh_key_passphrase': None, 'ssh_key_type': 'rsa', 'state': 'absent', 'system': False, 'uid': None, 'update_password': 'always'}}, 'item': 'jwilliamson', 'name': 'jwilliamson', 'state': 'absent'}, {'_ansible_item_label': 'kmroz', '_ansible_no_log': False, 'ansible_loop_var': 'item', 'changed': False, 'failed': False, 'invocation': {'module_args': {'append': False, 'authorization': None, 'comment': None, 'create_home': True, 'expires': None, 'force': False, 'generate_ssh_key': None, 'group': None, 'groups': None, 'hidden': None, 'home': None, 'local': None, 'login_class': None, 'move_home': False, 'name': 'kmroz', 'non_unique': False, 'password': None, 'password_lock': None, 'profile': None, 'remove': False, 'role': None, 'seuser': None, 'shell': None, 'skeleton': None, 'ssh_key_bits': 0, 'ssh_key_comment': 'ansible-generated on smithi194', 'ssh_key_file': None, 'ssh_key_passphrase': None, 'ssh_key_type': 'rsa', 'state': 'absent', 'system': False, 'uid': None, 'update_password': 'always'}}, 'item': 'kmroz', 'name': 'kmroz', 'state': 'absent'}, {'_ansible_item_label': 'shehbazj', '_ansible_no_log': False, 'ansible_loop_var': 'item', 'changed': False, 'failed': False, 'invocation': {'module_args': {'append': False, 'authorization': None, 'comment': None, 'create_home': True, 'expires': None, 'force': False, 'generate_ssh_key': None, 'group': None, 'groups': None, 'hidden': None, 'home': None, 'local': None, 'login_class': None, 'move_home': False, 'name': 'shehbazj', 'non_unique': False, 'password': None, 'password_lock': None, 'profile': None, 'remove': False, 'role': None, 'seuser': None, 'shell': None, 'skeleton': None, 'ssh_key_bits': 0, 'ssh_key_comment': 'ansible-generated on smithi194', 'ssh_key_file': None, 'ssh_key_passphrase': None, 'ssh_key_type': 'rsa', 'state': 'absent', 'system': False, 'uid': None, 'update_password': 'always'}}, 'item': 'shehbazj', 'name': 'shehbazj', 'state': 'absent'}, {'_ansible_item_label': 'abhishekvrshny', '_ansible_no_log': False, 'ansible_loop_var': 'item', 'changed': False, 'failed': False, 'invocation': {'module_args': {'append': False, 'authorization': None, 'comment': None, 'create_home': True, 'expires': None, 'force': False, 'generate_ssh_key': None, 'group': None, 'groups': None, 'hidden': None, 'home': None, 'local': None, 'login_class': None, 'move_home': False, 'name': 'abhishekvrshny', 'non_unique': False, 'password': None, 'password_lock': None, 'profile': None, 'remove': False, 'role': None, 'seuser': None, 'shell': None, 'skeleton': None, 'ssh_key_bits': 0, 'ssh_key_comment': 'ansible-generated on smithi194.front.sepia.ceph.com', 'ssh_key_file': None, 'ssh_key_passphrase': None, 'ssh_key_type': 'rsa', 'state': 'absent', 'system': False, 'uid': None, 'update_password': 'always'}}, 'item': 'abhishekvrshny', 'name': 'abhishekvrshny', 'state': 'absent'}, {'_ansible_item_label': 'asheplyakov', '_ansible_no_log': False, 'ansible_loop_var': 'item', 'changed': False, 'failed': False, 'invocation': {'module_args': {'append': False, 'authorization': None, 'comment': None, 'create_home': True, 'expires': None, 'force': False, 'generate_ssh_key': None, 'group': None, 'groups': None, 'hidden': None, 'home': None, 'local': None, 'login_class': None, 'move_home': False, 'name': 'asheplyakov', 'non_unique': False, 'password': None, 'password_lock': None, 'profile': None, 'remove': False, 'role': None, 'seuser': None, 'shell': None, 'skeleton': None, 'ssh_key_bits': 0, 'ssh_key_comment': 'ansible-generated on smithi194.front.sepia.ceph.com', 'ssh_key_file': None, 'ssh_key_passphrase': None, 'ssh_key_type': 'rsa', 'state': 'absent', 'system': False, 'uid': None, 'update_password': 'always'}}, 'item': 'asheplyakov', 'name': 'asheplyakov', 'state': 'absent'}, {'_ansible_item_label': 'liupan', '_ansible_no_log': False, 'ansible_loop_var': 'item', 'changed': False, 'failed': False, 'invocation': {'module_args': {'append': False, 'authorization': None, 'comment': None, 'create_home': True, 'expires': None, 'force': False, 'generate_ssh_key': None, 'group': None, 'groups': None, 'hidden': None, 'home': None, 'local': None, 'login_class': None, 'move_home': False, 'name': 'liupan', 'non_unique': False, 'password': None, 'password_lock': None, 'profile': None, 'remove': False, 'role': None, 'seuser': None, 'shell': None, 'skeleton': None, 'ssh_key_bits': 0, 'ssh_key_comment': 'ansible-generated on smithi194.front.sepia.ceph.com', 'ssh_key_file': None, 'ssh_key_passphrase': None, 'ssh_key_type': 'rsa', 'state': 'absent', 'system': False, 'uid': None, 'update_password': 'always'}}, 'item': 'liupan', 'name': 'liupan', 'state': 'absent'}, {'_ansible_item_label': 'adeza', '_ansible_no_log': False, 'ansible_loop_var': 'item', 'changed': False, 'failed': False, 'invocation': {'module_args': {'append': False, 'authorization': None, 'comment': None, 'create_home': True, 'expires': None, 'force': False, 'generate_ssh_key': None, 'group': None, 'groups': None, 'hidden': None, 'home': None, 'local': None, 'login_class': None, 'move_home': False, 'name': 'adeza', 'non_unique': False, 'password': None, 'password_lock': None, 'profile': None, 'remove': False, 'role': None, 'seuser': None, 'shell': None, 'skeleton': None, 'ssh_key_bits': 0, 'ssh_key_comment': 'ansible-generated on smithi194.front.sepia.ceph.com', 'ssh_key_file': None, 'ssh_key_passphrase': None, 'ssh_key_type': 'rsa', 'state': 'absent', 'system': False, 'uid': None, 'update_password': 'always'}}, 'item': 'adeza', 'name': 'adeza', 'state': 'absent'}, {'_ansible_item_label': 'pranith', '_ansible_no_log': False, 'ansible_loop_var': 'item', 'changed': False, 'failed': False, 'invocation': {'module_args': {'append': False, 'authorization': None, 'comment': None, 'create_home': True, 'expires': None, 'force': False, 'generate_ssh_key': None, 'group': None, 'groups': None, 'hidden': None, 'home': None, 'local': None, 'login_class': None, 'move_home': False, 'name': 'pranith', 'non_unique': False, 'password': None, 'password_lock': None, 'profile': None, 'remove': False, 'role': None, 'seuser': None, 'shell': None, 'skeleton': None, 'ssh_key_bits': 0, 'ssh_key_comment': 'ansible-generated on smithi194.front.sepia.ceph.com', 'ssh_key_file': None, 'ssh_key_passphrase': None, 'ssh_key_type': 'rsa', 'state': 'absent', 'system': False, 'uid': None, 'update_password': 'always'}}, 'item': 'pranith', 'name': 'pranith', 'state': 'absent'}, {'_ansible_item_label': 'dorinda', '_ansible_no_log': False, 'ansible_loop_var': 'item', 'changed': False, 'failed': False, 'invocation': {'module_args': {'append': False, 'authorization': None, 'comment': None, 'create_home': True, 'expires': None, 'force': False, 'generate_ssh_key': None, 'group': None, 'groups': None, 'hidden': None, 'home': None, 'local': None, 'login_class': None, 'move_home': False, 'name': 'dorinda', 'non_unique': False, 'password': None, 'password_lock': None, 'profile': None, 'remove': False, 'role': None, 'seuser': None, 'shell': None, 'skeleton': None, 'ssh_key_bits': 0, 'ssh_key_comment': 'ansible-generated on smithi194.front.sepia.ceph.com', 'ssh_key_file': None, 'ssh_key_passphrase': None, 'ssh_key_type': 'rsa', 'state': 'absent', 'system': False, 'uid': None, 'update_password': 'always'}}, 'item': 'dorinda', 'name': 'dorinda', 'state': 'absent'}, {'_ansible_item_label': 'zyan', '_ansible_no_log': False, 'ansible_loop_var': 'item', 'changed': False, 'failed': False, 'invocation': {'module_args': {'append': False, 'authorization': None, 'comment': None, 'create_home': True, 'expires': None, 'force': False, 'generate_ssh_key': None, 'group': None, 'groups': None, 'hidden': None, 'home': None, 'local': None, 'login_class': None, 'move_home': False, 'name': 'zyan', 'non_unique': False, 'password': None, 'password_lock': None, 'profile': None, 'remove': False, 'role': None, 'seuser': None, 'shell': None, 'skeleton': None, 'ssh_key_bits': 0, 'ssh_key_comment': 'ansible-generated on smithi194.front.sepia.ceph.com', 'ssh_key_file': None, 'ssh_key_passphrase': None, 'ssh_key_type': 'rsa', 'state': 'absent', 'system': False, 'uid': None, 'update_password': 'always'}}, 'item': 'zyan', 'name': 'zyan', 'state': 'absent'}, {'_ansible_item_label': 'jdillaman', '_ansible_no_log': False, 'ansible_loop_var': 'item', 'changed': False, 'failed': False, 'invocation': {'module_args': {'append': False, 'authorization': None, 'comment': None, 'create_home': True, 'expires': None, 'force': False, 'generate_ssh_key': None, 'group': None, 'groups': None, 'hidden': None, 'home': None, 'local': None, 'login_class': None, 'move_home': False, 'name': 'jdillaman', 'non_unique': False, 'password': None, 'password_lock': None, 'profile': None, 'remove': False, 'role': None, 'seuser': None, 'shell': None, 'skeleton': None, 'ssh_key_bits': 0, 'ssh_key_comment': 'ansible-generated on smithi194.front.sepia.ceph.com', 'ssh_key_file': None, 'ssh_key_passphrase': None, 'ssh_key_type': 'rsa', 'state': 'absent', 'system': False, 'uid': None, 'update_password': 'always'}}, 'item': 'jdillaman', 'name': 'jdillaman', 'state': 'absent'}, {'_ansible_item_label': 'davidz', '_ansible_no_log': False, 'ansible_loop_var': 'item', 'changed': False, 'failed': False, 'invocation': {'module_args': {'append': False, 'authorization': None, 'comment': None, 'create_home': True, 'expires': None, 'force': False, 'generate_ssh_key': None, 'group': None, 'groups': None, 'hidden': None, 'home': None, 'local': None, 'login_class': None, 'move_home': False, 'name': 'davidz', 'non_unique': False, 'password': None, 'password_lock': None, 'profile': None, 'remove': False, 'role': None, 'seuser': None, 'shell': None, 'skeleton': None, 'ssh_key_bits': 0, 'ssh_key_comment': 'ansible-generated on smithi194.front.sepia.ceph.com', 'ssh_key_file': None, 'ssh_key_passphrase': None, 'ssh_key_type': 'rsa', 'state': 'absent', 'system': False, 'uid': None, 'update_password': 'always'}}, 'item': 'davidz', 'name': 'davidz', 'state': 'absent'}, {'_ansible_item_label': 'wusui', '_ansible_no_log': False, 'ansible_loop_var': 'item', 'changed': False, 'failed': False, 'invocation': {'module_args': {'append': False, 'authorization': None, 'comment': None, 'create_home': True, 'expires': None, 'force': False, 'generate_ssh_key': None, 'group': None, 'groups': None, 'hidden': None, 'home': None, 'local': None, 'login_class': None, 'move_home': False, 'name': 'wusui', 'non_unique': False, 'password': None, 'password_lock': None, 'profile': None, 'remove': False, 'role': None, 'seuser': None, 'shell': None, 'skeleton': None, 'ssh_key_bits': 0, 'ssh_key_comment': 'ansible-generated on smithi194.front.sepia.ceph.com', 'ssh_key_file': None, 'ssh_key_passphrase': None, 'ssh_key_type': 'rsa', 'state': 'absent', 'system': False, 'uid': None, 'update_password': 'always'}}, 'item': 'wusui', 'name': 'wusui', 'state': 'absent'}, {'_ansible_item_label': 'nwatkins', '_ansible_no_log': False, 'ansible_loop_var': 'item', 'changed': False, 'failed': False, 'invocation': {'module_args': {'append': False, 'authorization': None, 'comment': None, 'create_home': True, 'expires': None, 'force': False, 'generate_ssh_key': None, 'group': None, 'groups': None, 'hidden': None, 'home': None, 'local': None, 'login_class': None, 'move_home': False, 'name': 'nwatkins', 'non_unique': False, 'password': None, 'password_lock': None, 'profile': None, 'remove': False, 'role': None, 'seuser': None, 'shell': None, 'skeleton': None, 'ssh_key_bits': 0, 'ssh_key_comment': 'ansible-generated on smithi194.front.sepia.ceph.com', 'ssh_key_file': None, 'ssh_key_passphrase': None, 'ssh_key_type': 'rsa', 'state': 'absent', 'system': False, 'uid': None, 'update_password': 'always'}}, 'item': 'nwatkins', 'name': 'nwatkins', 'state': 'absent'}, {'_ansible_item_label': 'sidharthanup', '_ansible_no_log': False, 'ansible_loop_var': 'item', 'changed': False, 'failed': False, 'invocation': {'module_args': {'append': False, 'authorization': None, 'comment': None, 'create_home': True, 'expires': None, 'force': False, 'generate_ssh_key': None, 'group': None, 'groups': None, 'hidden': None, 'home': None, 'local': None, 'login_class': None, 'move_home': False, 'name': 'sidharthanup', 'non_unique': False, 'password': None, 'password_lock': None, 'profile': None, 'remove': False, 'role': None, 'seuser': None, 'shell': None, 'skeleton': None, 'ssh_key_bits': 0, 'ssh_key_comment': 'ansible-generated on smithi194.front.sepia.ceph.com', 'ssh_key_file': None, 'ssh_key_passphrase': None, 'ssh_key_type': 'rsa', 'state': 'absent', 'system': False, 'uid': None, 'update_password': 'always'}}, 'item': 'sidharthanup', 'name': 'sidharthanup', 'state': 'absent'}, {'_ansible_item_label': 'varsha', '_ansible_no_log': False, 'ansible_loop_var': 'item', 'changed': False, 'failed': False, 'invocation': {'module_args': {'append': False, 'authorization': None, 'comment': None, 'create_home': True, 'expires': None, 'force': False, 'generate_ssh_key': None, 'group': None, 'groups': None, 'hidden': None, 'home': None, 'local': None, 'login_class': None, 'move_home': False, 'name': 'varsha', 'non_unique': False, 'password': None, 'password_lock': None, 'profile': None, 'remove': False, 'role': None, 'seuser': None, 'shell': None, 'skeleton': None, 'ssh_key_bits': 0, 'ssh_key_comment': 'ansible-generated on smithi194.front.sepia.ceph.com', 'ssh_key_file': None, 'ssh_key_passphrase': None, 'ssh_key_type': 'rsa', 'state': 'absent', 'system': False, 'uid': None, 'update_password': 'always'}}, 'item': 'varsha', 'name': 'varsha', 'state': 'absent'}, {'_ansible_item_label': 'hmunjulu', '_ansible_no_log': False, 'ansible_loop_var': 'item', 'changed': False, 'failed': False, 'invocation': {'module_args': {'append': False, 'authorization': None, 'comment': None, 'create_home': True, 'expires': None, 'force': False, 'generate_ssh_key': None, 'group': None, 'groups': None, 'hidden': None, 'home': None, 'local': None, 'login_class': None, 'move_home': False, 'name': 'hmunjulu', 'non_unique': False, 'password': None, 'password_lock': None, 'profile': None, 'remove': False, 'role': None, 'seuser': None, 'shell': None, 'skeleton': None, 'ssh_key_bits': 0, 'ssh_key_comment': 'ansible-generated on smithi194.front.sepia.ceph.com', 'ssh_key_file': None, 'ssh_key_passphrase': None, 'ssh_key_type': 'rsa', 'state': 'absent', 'system': False, 'uid': None, 'update_password': 'always'}}, 'item': 'hmunjulu', 'name': 'hmunjulu', 'state': 'absent'}, {'_ansible_item_label': 'jlopez', '_ansible_no_log': False, 'ansible_loop_var': 'item', 'changed': False, 'failed': False, 'invocation': {'module_args': {'append': False, 'authorization': None, 'comment': None, 'create_home': True, 'expires': None, 'force': False, 'generate_ssh_key': None, 'group': None, 'groups': None, 'hidden': None, 'home': None, 'local': None, 'login_class': None, 'move_home': False, 'name': 'jlopez', 'non_unique': False, 'password': None, 'password_lock': None, 'profile': None, 'remove': False, 'role': None, 'seuser': None, 'shell': None, 'skeleton': None, 'ssh_key_bits': 0, 'ssh_key_comment': 'ansible-generated on smithi194.front.sepia.ceph.com', 'ssh_key_file': None, 'ssh_key_passphrase': None, 'ssh_key_type': 'rsa', 'state': 'absent', 'system': False, 'uid': None, 'update_password': 'always'}}, 'item': 'jlopez', 'name': 'jlopez', 'state': 'absent'}, {'_ansible_item_label': 'dfuller', '_ansible_no_log': False, 'ansible_loop_var': 'item', 'changed': False, 'failed': False, 'invocation': {'module_args': {'append': False, 'authorization': None, 'comment': None, 'create_home': True, 'expires': None, 'force': False, 'generate_ssh_key': None, 'group': None, 'groups': None, 'hidden': None, 'home': None, 'local': None, 'login_class': None, 'move_home': False, 'name': 'dfuller', 'non_unique': False, 'password': None, 'password_lock': None, 'profile': None, 'remove': False, 'role': None, 'seuser': None, 'shell': None, 'skeleton': None, 'ssh_key_bits': 0, 'ssh_key_comment': 'ansible-generated on smithi194.front.sepia.ceph.com', 'ssh_key_file': None, 'ssh_key_passphrase': None, 'ssh_key_type': 'rsa', 'state': 'absent', 'system': False, 'uid': None, 'update_password': 'always'}}, 'item': 'dfuller', 'name': 'dfuller', 'state': 'absent'}, {'_ansible_item_label': 'vasu', '_ansible_no_log': False, 'ansible_loop_var': 'item', 'changed': False, 'failed': False, 'invocation': {'module_args': {'append': False, 'authorization': None, 'comment': None, 'create_home': True, 'expires': None, 'force': False, 'generate_ssh_key': None, 'group': None, 'groups': None, 'hidden': None, 'home': None, 'local': None, 'login_class': None, 'move_home': False, 'name': 'vasu', 'non_unique': False, 'password': None, 'password_lock': None, 'profile': None, 'remove': False, 'role': None, 'seuser': None, 'shell': None, 'skeleton': None, 'ssh_key_bits': 0, 'ssh_key_comment': 'ansible-generated on smithi194.front.sepia.ceph.com', 'ssh_key_file': None, 'ssh_key_passphrase': None, 'ssh_key_type': 'rsa', 'state': 'absent', 'system': False, 'uid': None, 'update_password': 'always'}}, 'item': 'vasu', 'name': 'vasu', 'state': 'absent'}, {'_ansible_item_label': 'swagner', '_ansible_no_log': False, 'ansible_loop_var': 'item', 'changed': False, 'failed': False, 'invocation': {'module_args': {'append': False, 'authorization': None, 'comment': None, 'create_home': True, 'expires': None, 'force': False, 'generate_ssh_key': None, 'group': None, 'groups': None, 'hidden': None, 'home': None, 'local': None, 'login_class': None, 'move_home': False, 'name': 'swagner', 'non_unique': False, 'password': None, 'password_lock': None, 'profile': None, 'remove': False, 'role': None, 'seuser': None, 'shell': None, 'skeleton': None, 'ssh_key_bits': 0, 'ssh_key_comment': 'ansible-generated on smithi194.front.sepia.ceph.com', 'ssh_key_file': None, 'ssh_key_passphrase': None, 'ssh_key_type': 'rsa', 'state': 'absent', 'system': False, 'uid': None, 'update_password': 'always'}}, 'item': 'swagner', 'name': 'swagner', 'state': 'absent'}, {'_ansible_item_label': 'dpivonka', '_ansible_no_log': False, 'ansible_loop_var': 'item', 'changed': False, 'failed': False, 'invocation': {'module_args': {'append': False, 'authorization': None, 'comment': None, 'create_home': True, 'expires': None, 'force': False, 'generate_ssh_key': None, 'group': None, 'groups': None, 'hidden': None, 'home': None, 'local': None, 'login_class': None, 'move_home': False, 'name': 'dpivonka', 'non_unique': False, 'password': None, 'password_lock': None, 'profile': None, 'remove': False, 'role': None, 'seuser': None, 'shell': None, 'skeleton': None, 'ssh_key_bits': 0, 'ssh_key_comment': 'ansible-generated on smithi194.front.sepia.ceph.com', 'ssh_key_file': None, 'ssh_key_passphrase': None, 'ssh_key_type': 'rsa', 'state': 'absent', 'system': False, 'uid': None, 'update_password': 'always'}}, 'item': 'dpivonka', 'name': 'dpivonka', 'state': 'absent'}, {'_ansible_item_label': 'nlevine', '_ansible_no_log': False, 'ansible_loop_var': 'item', 'changed': False, 'failed': False, 'invocation': {'module_args': {'append': False, 'authorization': None, 'comment': None, 'create_home': True, 'expires': None, 'force': False, 'generate_ssh_key': None, 'group': None, 'groups': None, 'hidden': None, 'home': None, 'local': None, 'login_class': None, 'move_home': False, 'name': 'nlevine', 'non_unique': False, 'password': None, 'password_lock': None, 'profile': None, 'remove': False, 'role': None, 'seuser': None, 'shell': None, 'skeleton': None, 'ssh_key_bits': 0, 'ssh_key_comment': 'ansible-generated on smithi194.front.sepia.ceph.com', 'ssh_key_file': None, 'ssh_key_passphrase': None, 'ssh_key_type': 'rsa', 'state': 'absent', 'system': False, 'uid': None, 'update_password': 'always'}}, 'item': 'nlevine', 'name': 'nlevine', 'state': 'absent'}, {'_ansible_item_label': 'tbrekke', '_ansible_no_log': False, 'ansible_loop_var': 'item', 'changed': False, 'failed': False, 'invocation': {'module_args': {'append': False, 'authorization': None, 'comment': None, 'create_home': True, 'expires': None, 'force': False, 'generate_ssh_key': None, 'group': None, 'groups': None, 'hidden': None, 'home': None, 'local': None, 'login_class': None, 'move_home': False, 'name': 'tbrekke', 'non_unique': False, 'password': None, 'password_lock': None, 'profile': None, 'remove': False, 'role': None, 'seuser': None, 'shell': None, 'skeleton': None, 'ssh_key_bits': 0, 'ssh_key_comment': 'ansible-generated on smithi194.front.sepia.ceph.com', 'ssh_key_file': None, 'ssh_key_passphrase': None, 'ssh_key_type': 'rsa', 'state': 'absent', 'system': False, 'uid': None, 'update_password': 'always'}}, 'item': 'tbrekke', 'name': 'tbrekke', 'state': 'absent'}, {'_ansible_item_label': 'taco', '_ansible_no_log': False, 'ansible_loop_var': 'item', 'changed': False, 'failed': False, 'invocation': {'module_args': {'append': False, 'authorization': None, 'comment': None, 'create_home': True, 'expires': None, 'force': False, 'generate_ssh_key': None, 'group': None, 'groups': None, 'hidden': None, 'home': None, 'local': None, 'login_class': None, 'move_home': False, 'name': 'taco', 'non_unique': False, 'password': None, 'password_lock': None, 'profile': None, 'remove': False, 'role': None, 'seuser': None, 'shell': None, 'skeleton': None, 'ssh_key_bits': 0, 'ssh_key_comment': 'ansible-generated on smithi194.front.sepia.ceph.com', 'ssh_key_file': None, 'ssh_key_passphrase': None, 'ssh_key_type': 'rsa', 'state': 'absent', 'system': False, 'uid': None, 'update_password': 'always'}}, 'item': 'taco', 'name': 'taco', 'state': 'absent'}, {'_ansible_item_label': 'louis', '_ansible_no_log': False, 'ansible_loop_var': 'item', 'changed': False, 'failed': False, 'invocation': {'module_args': {'append': False, 'authorization': None, 'comment': None, 'create_home': True, 'expires': None, 'force': False, 'generate_ssh_key': None, 'group': None, 'groups': None, 'hidden': None, 'home': None, 'local': None, 'login_class': None, 'move_home': False, 'name': 'louis', 'non_unique': False, 'password': None, 'password_lock': None, 'profile': None, 'remove': False, 'role': None, 'seuser': None, 'shell': None, 'skeleton': None, 'ssh_key_bits': 0, 'ssh_key_comment': 'ansible-generated on smithi194.front.sepia.ceph.com', 'ssh_key_file': None, 'ssh_key_passphrase': None, 'ssh_key_type': 'rsa', 'state': 'absent', 'system': False, 'uid': None, 'update_password': 'always'}}, 'item': 'louis', 'name': 'louis', 'state': 'absent'}, {'_ansible_item_label': 'amarango', '_ansible_no_log': False, 'ansible_loop_var': 'item', 'changed': False, 'failed': False, 'invocation': {'module_args': {'append': False, 'authorization': None, 'comment': None, 'create_home': True, 'expires': None, 'force': False, 'generate_ssh_key': None, 'group': None, 'groups': None, 'hidden': None, 'home': None, 'local': None, 'login_class': None, 'move_home': False, 'name': 'amarango', 'non_unique': False, 'password': None, 'password_lock': None, 'profile': None, 'remove': False, 'role': None, 'seuser': None, 'shell': None, 'skeleton': None, 'ssh_key_bits': 0, 'ssh_key_comment': 'ansible-generated on smithi194.front.sepia.ceph.com', 'ssh_key_file': None, 'ssh_key_passphrase': None, 'ssh_key_type': 'rsa', 'state': 'absent', 'system': False, 'uid': None, 'update_password': 'always'}}, 'item': 'amarango', 'name': 'amarango', 'state': 'absent'}, {'_ansible_item_label': 'oobe', '_ansible_no_log': False, 'ansible_loop_var': 'item', 'changed': False, 'failed': False, 'invocation': {'module_args': {'append': False, 'authorization': None, 'comment': None, 'create_home': True, 'expires': None, 'force': False, 'generate_ssh_key': None, 'group': None, 'groups': None, 'hidden': None, 'home': None, 'local': None, 'login_class': None, 'move_home': False, 'name': 'oobe', 'non_unique': False, 'password': None, 'password_lock': None, 'profile': None, 'remove': False, 'role': None, 'seuser': None, 'shell': None, 'skeleton': None, 'ssh_key_bits': 0, 'ssh_key_comment': 'ansible-generated on smithi194.front.sepia.ceph.com', 'ssh_key_file': None, 'ssh_key_passphrase': None, 'ssh_key_type': 'rsa', 'state': 'absent', 'system': False, 'uid': None, 'update_password': 'always'}}, 'item': 'oobe', 'name': 'oobe', 'state': 'absent'}, {'_ansible_item_label': 'rturk', '_ansible_no_log': False, 'ansible_loop_var': 'item', 'changed': False, 'failed': False, 'invocation': {'module_args': {'append': False, 'authorization': None, 'comment': None, 'create_home': True, 'expires': None, 'force': False, 'generate_ssh_key': None, 'group': None, 'groups': None, 'hidden': None, 'home': None, 'local': None, 'login_class': None, 'move_home': False, 'name': 'rturk', 'non_unique': False, 'password': None, 'password_lock': None, 'profile': None, 'remove': False, 'role': None, 'seuser': None, 'shell': None, 'skeleton': None, 'ssh_key_bits': 0, 'ssh_key_comment': 'ansible-generated on smithi194.front.sepia.ceph.com', 'ssh_key_file': None, 'ssh_key_passphrase': None, 'ssh_key_type': 'rsa', 'state': 'absent', 'system': False, 'uid': None, 'update_password': 'always'}}, 'item': 'rturk', 'name': 'rturk', 'state': 'absent'}, {'_ansible_item_label': 'fche', '_ansible_no_log': False, 'ansible_loop_var': 'item', 'changed': False, 'failed': False, 'invocation': {'module_args': {'append': False, 'authorization': None, 'comment': None, 'create_home': True, 'expires': None, 'force': False, 'generate_ssh_key': None, 'group': None, 'groups': None, 'hidden': None, 'home': None, 'local': None, 'login_class': None, 'move_home': False, 'name': 'fche', 'non_unique': False, 'password': None, 'password_lock': None, 'profile': None, 'remove': False, 'role': None, 'seuser': None, 'shell': None, 'skeleton': None, 'ssh_key_bits': 0, 'ssh_key_comment': 'ansible-generated on smithi194.front.sepia.ceph.com', 'ssh_key_file': None, 'ssh_key_passphrase': None, 'ssh_key_type': 'rsa', 'state': 'absent', 'system': False, 'uid': None, 'update_password': 'always'}}, 'item': 'fche', 'name': 'fche', 'state': 'absent'}, {'_ansible_item_label': 'jbainbri', '_ansible_no_log': False, 'ansible_loop_var': 'item', 'changed': False, 'failed': False, 'invocation': {'module_args': {'append': False, 'authorization': None, 'comment': None, 'create_home': True, 'expires': None, 'force': False, 'generate_ssh_key': None, 'group': None, 'groups': None, 'hidden': None, 'home': None, 'local': None, 'login_class': None, 'move_home': False, 'name': 'jbainbri', 'non_unique': False, 'password': None, 'password_lock': None, 'profile': None, 'remove': False, 'role': None, 'seuser': None, 'shell': None, 'skeleton': None, 'ssh_key_bits': 0, 'ssh_key_comment': 'ansible-generated on smithi194.front.sepia.ceph.com', 'ssh_key_file': None, 'ssh_key_passphrase': None, 'ssh_key_type': 'rsa', 'state': 'absent', 'system': False, 'uid': None, 'update_password': 'always'}}, 'item': 'jbainbri', 'name': 'jbainbri', 'state': 'absent'}, {'_ansible_item_label': 'kdhananj', '_ansible_no_log': False, 'ansible_loop_var': 'item', 'changed': False, 'failed': False, 'invocation': {'module_args': {'append': False, 'authorization': None, 'comment': None, 'create_home': True, 'expires': None, 'force': False, 'generate_ssh_key': None, 'group': None, 'groups': None, 'hidden': None, 'home': None, 'local': None, 'login_class': None, 'move_home': False, 'name': 'kdhananj', 'non_unique': False, 'password': None, 'password_lock': None, 'profile': None, 'remove': False, 'role': None, 'seuser': None, 'shell': None, 'skeleton': None, 'ssh_key_bits': 0, 'ssh_key_comment': 'ansible-generated on smithi194.front.sepia.ceph.com', 'ssh_key_file': None, 'ssh_key_passphrase': None, 'ssh_key_type': 'rsa', 'state': 'absent', 'system': False, 'uid': None, 'update_password': 'always'}}, 'item': 'kdhananj', 'name': 'kdhananj', 'state': 'absent'}]}}
wip-rishabh-2023May15-1524
wip-rishabh-2023May15-1524
main
smithi
rhel 8.6
fs/workload/{0-rhel_8 begin/{0-install 1-cephadm 2-logrotate} clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} mount/fuse objectstore-ec/bluestore-comp-ec-root omap_limit/10000 overrides/{cephsqlite-timeout frag ignorelist_health ignorelist_wrongly_marked_down osd-asserts session_timeout} ranks/multi/{export-check n/3 replication/default} standby-replay tasks/{0-subvolume/{no-subvolume} 1-check-counter 2-scrub/no 3-snaps/yes 4-flush/no 5-workunit/postgres}}
Stale jobs detected, aborting.
wip-rishabh-2023May15-1524
wip-rishabh-2023May15-1524
main
smithi
centos 8.stream
fs/fscrypt/{begin/{0-install 1-ceph 2-logrotate} bluestore-bitmap clusters/1-mds-1-client conf/{client mds mon osd} distro/{centos_latest} mount/kclient/{mount-syntax/{v1} mount overrides/{distro/testing/k-testing ms-die-on-skipped}} overrides/{ignorelist_health ignorelist_health_more ignorelist_wrongly_marked_down pg-warn} tasks/{0-client 1-tests/fscrypt-iozone}}
Command failed (workunit test fs/fscrypt.sh) on smithi163 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=27be51263d3d0c721b6d1ca9f42550bf0ab3c97b TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 CEPH_MNT=/home/ubuntu/cephtest/mnt.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 6h /home/ubuntu/cephtest/clone.client.0/qa/workunits/fs/fscrypt.sh none iozone'
wip-rishabh-2023May15-1524
wip-rishabh-2023May15-1524
main
smithi
rhel 8.6
fs/workload/{0-rhel_8 begin/{0-install 1-cephadm 2-logrotate} clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} mount/kclient/{base/{mount-syntax/{v2} mount overrides/{distro/testing/k-testing ms-die-on-skipped}} ms_mode/legacy wsync/yes} objectstore-ec/bluestore-ec-root omap_limit/10000 overrides/{cephsqlite-timeout frag ignorelist_health ignorelist_wrongly_marked_down osd-asserts session_timeout} ranks/multi/{export-check n/3 replication/always} standby-replay tasks/{0-subvolume/{with-namespace-isolated-and-quota} 1-check-counter 2-scrub/no 3-snaps/yes 4-flush/no 5-workunit/suites/dbench}}
Command failed (workunit test suites/dbench.sh) on smithi072 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=27be51263d3d0c721b6d1ca9f42550bf0ab3c97b TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 CEPH_MNT=/home/ubuntu/cephtest/mnt.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/suites/dbench.sh'
wip-rishabh-2023May15-1524
wip-rishabh-2023May15-1524
main
smithi
centos 8.stream
fs/traceless/{begin/{0-install 1-ceph 2-logrotate} clusters/fixed-2-ucephfs conf/{client mds mon osd} distro/{centos_8} mount/fuse objectstore-ec/bluestore-comp overrides/{frag ignorelist_health ignorelist_wrongly_marked_down} tasks/cfuse_workunit_suites_fsstress traceless/50pc}
Command failed (workunit test suites/fsstress.sh) on smithi047 with status 124: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=27be51263d3d0c721b6d1ca9f42550bf0ab3c97b TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 CEPH_MNT=/home/ubuntu/cephtest/mnt.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 6h /home/ubuntu/cephtest/clone.client.0/qa/workunits/suites/fsstress.sh'
wip-rishabh-2023May15-1524
wip-rishabh-2023May15-1524
main
smithi
ubuntu 22.04
fs/thrash/workloads/{begin/{0-install 1-ceph 2-logrotate} clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{ubuntu_latest} mount/kclient/{mount-syntax/{v1} mount overrides/{distro/testing/k-testing ms-die-on-skipped}} msgr-failures/none objectstore-ec/bluestore-comp-ec-root overrides/{frag ignorelist_health ignorelist_wrongly_marked_down prefetch_dirfrags/no prefetch_entire_dirfrags/no races session_timeout thrashosds-health} ranks/1 tasks/{1-thrash/mds 2-workunit/suites/pjd}}
wip-rishabh-2023May15-1524
wip-rishabh-2023May15-1524
main
smithi
rhel 8.6
fs/thrash/workloads/{begin/{0-install 1-ceph 2-logrotate} clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{rhel_8} mount/fuse msgr-failures/osd-mds-delay objectstore-ec/bluestore-comp overrides/{frag ignorelist_health ignorelist_wrongly_marked_down prefetch_dirfrags/yes prefetch_entire_dirfrags/yes races session_timeout thrashosds-health} ranks/3 tasks/{1-thrash/mds 2-workunit/fs/snaps}}
Command failed (workunit test fs/snaps/untar_snap_rm.sh) on smithi023 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=27be51263d3d0c721b6d1ca9f42550bf0ab3c97b TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 CEPH_MNT=/home/ubuntu/cephtest/mnt.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 6h /home/ubuntu/cephtest/clone.client.0/qa/workunits/fs/snaps/untar_snap_rm.sh'
wip-rishabh-2023May15-1524
wip-rishabh-2023May15-1524
main
smithi
centos 8.stream
fs/32bits/{begin/{0-install 1-ceph 2-logrotate} clusters/fixed-2-ucephfs conf/{client mds mon osd} distro/{centos_8} mount/fuse objectstore-ec/bluestore-comp-ec-root overrides/{faked-ino ignorelist_health ignorelist_wrongly_marked_down} tasks/cfuse_workunit_suites_pjd}
Command failed (workunit test suites/pjd.sh) on smithi039 with status 124: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=27be51263d3d0c721b6d1ca9f42550bf0ab3c97b TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 CEPH_MNT=/home/ubuntu/cephtest/mnt.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 6h /home/ubuntu/cephtest/clone.client.0/qa/workunits/suites/pjd.sh'
wip-rishabh-2023May15-1524
wip-rishabh-2023May15-1524
main
smithi
rhel 8.6
fs/permission/{begin/{0-install 1-ceph 2-logrotate} clusters/fixed-2-ucephfs conf/{client mds mon osd} distro/{rhel_8} mount/fuse objectstore-ec/bluestore-comp-ec-root overrides/{ignorelist_health ignorelist_wrongly_marked_down} tasks/cfuse_workunit_suites_pjd}
Command failed (workunit test suites/pjd.sh) on smithi042 with status 124: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=27be51263d3d0c721b6d1ca9f42550bf0ab3c97b TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 CEPH_MNT=/home/ubuntu/cephtest/mnt.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 6h /home/ubuntu/cephtest/clone.client.0/qa/workunits/suites/pjd.sh'
wip-rishabh-2023May15-1524
wip-rishabh-2023May15-1524
main
smithi
rhel 8.6
fs/workload/{0-rhel_8 begin/{0-install 1-cephadm 2-logrotate} clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} mount/kclient/{base/{mount-syntax/{v1} mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} ms_mode/secure wsync/no} objectstore-ec/bluestore-bitmap omap_limit/10 overrides/{cephsqlite-timeout frag ignorelist_health ignorelist_wrongly_marked_down osd-asserts session_timeout} ranks/multi/{export-check n/3 replication/always} standby-replay tasks/{0-subvolume/{with-namespace-isolated} 1-check-counter 2-scrub/yes 3-snaps/no 4-flush/yes 5-workunit/suites/fsync-tester}}
wip-rishabh-2023May15-1524
wip-rishabh-2023May15-1524
main
smithi
ubuntu 22.04
fs/functional/{begin/{0-install 1-ceph 2-logrotate} clusters/1a3s-mds-4c-client conf/{client mds mon osd} distro/{ubuntu_latest} mount/fuse objectstore/bluestore-bitmap overrides/{ignorelist_health ignorelist_wrongly_marked_down no_client_pidfile} tasks/xfstests-dev}
hit max job timeout
wip-rishabh-2023May15-1524
wip-rishabh-2023May15-1524
main
smithi
rhel 8.6
fs/functional/{begin/{0-install 1-ceph 2-logrotate} clusters/1a3s-mds-4c-client conf/{client mds mon osd} distro/{ubuntu_latest} mount/kclient/{mount-syntax/{v1} mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} objectstore/bluestore-ec-root overrides/{ignorelist_health ignorelist_wrongly_marked_down no_client_pidfile} tasks/workunit/quota}
Command failed (workunit test fs/quota/quota.sh) on smithi062 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/mnt.3/client.3/tmp && cd -- /home/ubuntu/cephtest/mnt.3/client.3/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=27be51263d3d0c721b6d1ca9f42550bf0ab3c97b TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="3" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.3 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.3 CEPH_MNT=/home/ubuntu/cephtest/mnt.3 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.3/qa/workunits/fs/quota/quota.sh'
wip-rishabh-2023May15-1524
wip-rishabh-2023May15-1524
main
smithi
ubuntu 22.04
fs/multiclient/{begin/{0-install 1-ceph 2-logrotate} clusters/1-mds-2-client conf/{client mds mon osd} distros/ubuntu_latest mount/fuse objectstore-ec/bluestore-ec-root overrides/{ignorelist_health ignorelist_wrongly_marked_down} tasks/ior-shared-file}
Command failed on smithi097 with status 2: 'TESTDIR=/home/ubuntu/cephtest bash -s'
wip-rishabh-2023May15-1524
wip-rishabh-2023May15-1524
main
smithi
centos 8.stream
fs/upgrade/featureful_client/old_client/{bluestore-bitmap centos_latest clusters/1-mds-2-client-micro conf/{client mds mon osd} overrides/{ignorelist_health ignorelist_wrongly_marked_down multimds/yes pg-warn} tasks/{0-octopus 1-client 2-upgrade 3-compat_client/quincy}}
wip-rishabh-2023May15-1524
wip-rishabh-2023May15-1524
main
smithi
rhel 8.6
fs/workload/{0-rhel_8 begin/{0-install 1-cephadm 2-logrotate} clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} mount/kclient/{base/{mount-syntax/{v1} mount overrides/{distro/testing/k-testing ms-die-on-skipped}} ms_mode/legacy wsync/yes} objectstore-ec/bluestore-ec-root omap_limit/10000 overrides/{cephsqlite-timeout frag ignorelist_health ignorelist_wrongly_marked_down osd-asserts session_timeout} ranks/multi/{export-check n/5 replication/always} standby-replay tasks/{0-subvolume/{with-no-extra-options} 1-check-counter 2-scrub/yes 3-snaps/no 4-flush/no 5-workunit/kernel_untar_build}}
wip-rishabh-2023May15-1524
wip-rishabh-2023May15-1524
main
smithi
ubuntu 20.04
fs/verify/{begin/{0-install 1-ceph 2-logrotate} clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{ubuntu/{latest overrides}} mount/kclient/{k-testing mount ms-die-on-skipped} objectstore-ec/bluestore-bitmap overrides/{ignorelist_health ignorelist_wrongly_marked_down mon-debug session_timeout} ranks/5 tasks/dbench validater/valgrind}
saw valgrind issues
wip-rishabh-2023May15-1524
wip-rishabh-2023May15-1524
main
smithi
rhel 8.6
fs/workload/{0-rhel_8 begin/{0-install 1-cephadm 2-logrotate} clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} mount/kclient/{base/{mount-syntax/{v2} mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} ms_mode/secure wsync/no} objectstore-ec/bluestore-bitmap omap_limit/10 overrides/{cephsqlite-timeout frag ignorelist_health ignorelist_wrongly_marked_down osd-asserts session_timeout} ranks/multi/{export-check n/3 replication/default} standby-replay tasks/{0-subvolume/{no-subvolume} 1-check-counter 2-scrub/no 3-snaps/yes 4-flush/yes 5-workunit/postgres}}
Command failed on smithi006 with status 1: "sudo TESTDIR=/home/ubuntu/cephtest bash -c 'sudo -u postgres -- pgbench -s 500 -i'"
wip-rishabh-2023May15-1524
wip-rishabh-2023May15-1524
main
smithi
ubuntu 22.04
fs/traceless/{begin/{0-install 1-ceph 2-logrotate} clusters/fixed-2-ucephfs conf/{client mds mon osd} distro/{ubuntu_latest} mount/fuse objectstore-ec/bluestore-comp overrides/{frag ignorelist_health ignorelist_wrongly_marked_down} tasks/cfuse_workunit_suites_blogbench traceless/50pc}
hit max job timeout
wip-rishabh-2023May15-1524
wip-rishabh-2023May15-1524
main
smithi
rhel 8.6
fs/workload/{0-rhel_8 begin/{0-install 1-cephadm 2-logrotate} clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} mount/kclient/{base/{mount-syntax/{v1} mount overrides/{distro/testing/k-testing ms-die-on-skipped}} ms_mode/legacy wsync/yes} objectstore-ec/bluestore-ec-root omap_limit/10000 overrides/{cephsqlite-timeout frag ignorelist_health ignorelist_wrongly_marked_down osd-asserts session_timeout} ranks/multi/{export-check n/5 replication/default} standby-replay tasks/{0-subvolume/{no-subvolume} 1-check-counter 2-scrub/yes 3-snaps/no 4-flush/no 5-workunit/suites/ffsb}}
wip-rishabh-2023May15-1524
wip-rishabh-2023May15-1524
main
smithi
ubuntu 22.04
fs/functional/{begin/{0-install 1-ceph 2-logrotate} clusters/1a3s-mds-4c-client conf/{client mds mon osd} distro/{ubuntu_latest} mount/fuse objectstore/bluestore-ec-root overrides/{ignorelist_health ignorelist_wrongly_marked_down no_client_pidfile} tasks/backtrace}
wip-rishabh-2023May15-1524
wip-rishabh-2023May15-1524
main
smithi
ubuntu 20.04
fs/thrash/multifs/{begin/{0-install 1-ceph 2-logrotate} clusters/1a3s-mds-2c-client conf/{client mds mon osd} distro/{ubuntu_20.04} mount/fuse msgr-failures/osd-mds-delay objectstore/bluestore-bitmap overrides/{frag ignorelist_health ignorelist_wrongly_marked_down multifs session_timeout thrashosds-health} tasks/{1-thrash/mon 2-workunit/cfuse_workunit_trivial_sync}}
wip-rishabh-2023May15-1524
wip-rishabh-2023May15-1524
main
smithi
centos 8.stream
fs/upgrade/mds_upgrade_sequence/{bluestore-bitmap centos_8.stream_container_tools conf/{client mds mon osd} fail_fs/yes overrides/{ignorelist_health ignorelist_wrongly_marked_down pg-warn syntax} roles tasks/{0-from/pacific 1-volume/{0-create 1-ranks/1 2-allow_standby_replay/no 3-inline/no 4-verify} 2-client 3-upgrade-mgr-staggered 4-config-upgrade/{fail_fs} 5-upgrade-with-workload 6-verify}}
wip-rishabh-2023May15-1524
wip-rishabh-2023May15-1524
main
smithi
ubuntu 22.04
fs/multiclient/{begin/{0-install 1-ceph 2-logrotate} clusters/1-mds-3-client conf/{client mds mon osd} distros/ubuntu_latest mount/fuse objectstore-ec/bluestore-bitmap overrides/{ignorelist_health ignorelist_wrongly_marked_down} tasks/mdtest}
Command failed on smithi162 with status 2: 'TESTDIR=/home/ubuntu/cephtest bash -s'
wip-rishabh-2023May15-1524
wip-rishabh-2023May15-1524
main
smithi
ubuntu 20.04
fs/multifs/{begin/{0-install 1-ceph 2-logrotate} clusters/1a3s-mds-2c-client conf/{client mds mon osd} distro/{ubuntu_20.04} mount/kclient/{mount-syntax/{v2} mount overrides/{distro/testing/k-testing ms-die-on-skipped}} objectstore-ec/bluestore-bitmap overrides/{ignorelist_health ignorelist_wrongly_marked_down mon-debug} tasks/multifs-auth}
wip-rishabh-2023May15-1524
wip-rishabh-2023May15-1524
main
smithi
ubuntu 20.04
fs/snaps/{begin/{0-install 1-ceph 2-logrotate} clusters/1a3s-mds-1c-client conf/{client mds mon osd} distro/{ubuntu_20.04} mount/kclient/{mount-syntax/{v2} mount overrides/{distro/testing/k-testing ms-die-on-skipped}} objectstore-ec/bluestore-bitmap overrides/{ignorelist_health ignorelist_wrongly_marked_down} tasks/workunit/snaps}
wip-rishabh-2023May15-1524
wip-rishabh-2023May15-1524
main
smithi
ubuntu 20.04
fs/functional/{begin/{0-install 1-ceph 2-logrotate} clusters/1a3s-mds-4c-client conf/{client mds mon osd} distro/{ubuntu_20.04} mount/kclient/{mount-syntax/{v2} mount overrides/{distro/testing/k-testing ms-die-on-skipped}} objectstore/bluestore-ec-root overrides/{ignorelist_health ignorelist_wrongly_marked_down no_client_pidfile} tasks/client-limits}
wip-rishabh-2023May15-1524
wip-rishabh-2023May15-1524
main
smithi
rhel 8.6
fs/32bits/{begin/{0-install 1-ceph 2-logrotate} clusters/fixed-2-ucephfs conf/{client mds mon osd} distro/{rhel_8} mount/fuse objectstore-ec/bluestore-ec-root overrides/{faked-ino ignorelist_health ignorelist_wrongly_marked_down} tasks/cfuse_workunit_suites_pjd}
Command failed (workunit test suites/pjd.sh) on smithi107 with status 124: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=27be51263d3d0c721b6d1ca9f42550bf0ab3c97b TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 CEPH_MNT=/home/ubuntu/cephtest/mnt.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 6h /home/ubuntu/cephtest/clone.client.0/qa/workunits/suites/pjd.sh'
wip-rishabh-2023May15-1524
wip-rishabh-2023May15-1524
main
smithi
rhel 8.6
fs/permission/{begin/{0-install 1-ceph 2-logrotate} clusters/fixed-2-ucephfs conf/{client mds mon osd} distro/{rhel_8} mount/fuse objectstore-ec/bluestore-ec-root overrides/{ignorelist_health ignorelist_wrongly_marked_down} tasks/cfuse_workunit_suites_pjd}
Command failed (workunit test suites/pjd.sh) on smithi035 with status 124: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=27be51263d3d0c721b6d1ca9f42550bf0ab3c97b TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 CEPH_MNT=/home/ubuntu/cephtest/mnt.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 6h /home/ubuntu/cephtest/clone.client.0/qa/workunits/suites/pjd.sh'
wip-rishabh-2023May15-1524
wip-rishabh-2023May15-1524
main
smithi
rhel 8.6
fs/workload/{0-rhel_8 begin/{0-install 1-cephadm 2-logrotate} clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} mount/fuse objectstore-ec/bluestore-comp-ec-root omap_limit/10000 overrides/{cephsqlite-timeout frag ignorelist_health ignorelist_wrongly_marked_down osd-asserts session_timeout} ranks/multi/{export-check n/5 replication/default} standby-replay tasks/{0-subvolume/{with-namespace-isolated} 1-check-counter 2-scrub/yes 3-snaps/no 4-flush/no 5-workunit/fs/misc}}
error during scrub thrashing: reached maximum tries (30) after waiting for 900 seconds
wip-rishabh-2023May15-1524
wip-rishabh-2023May15-1524
main
smithi
rhel 8.6
fs/thrash/workloads/{begin/{0-install 1-ceph 2-logrotate} clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{rhel_8} mount/fuse msgr-failures/osd-mds-delay objectstore-ec/bluestore-bitmap overrides/{frag ignorelist_health ignorelist_wrongly_marked_down prefetch_dirfrags/no prefetch_entire_dirfrags/yes races session_timeout thrashosds-health} ranks/3 tasks/{1-thrash/osd 2-workunit/fs/snaps}}
SSH connection to smithi194 was lost: 'rm -f /tmp/kernel.x86_64.rpm && echo kernel-6.3.0_g2bcb17939031-1.x86_64.rpm | wget -nv -O /tmp/kernel.x86_64.rpm --base=https://4.chacra.ceph.com/r/kernel/testing/2bcb179390319a07fc43480888bd707374860ab3/centos/8/flavors/default/x86_64/ --input-file=-'
wip-rishabh-2023May15-1524
wip-rishabh-2023May15-1524
main
smithi
ubuntu 20.04
fs/traceless/{begin/{0-install 1-ceph 2-logrotate} clusters/fixed-2-ucephfs conf/{client mds mon osd} distro/{ubuntu_20.04} mount/fuse objectstore-ec/bluestore-ec-root overrides/{frag ignorelist_health ignorelist_wrongly_marked_down} tasks/cfuse_workunit_suites_dbench traceless/50pc}
Command failed (workunit test suites/dbench.sh) on smithi002 with status 124: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=27be51263d3d0c721b6d1ca9f42550bf0ab3c97b TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 CEPH_MNT=/home/ubuntu/cephtest/mnt.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/suites/dbench.sh'
wip-rishabh-2023May15-1524
wip-rishabh-2023May15-1524
main
smithi
ubuntu 22.04
fs/fscrypt/{begin/{0-install 1-ceph 2-logrotate} bluestore-bitmap clusters/1-mds-1-client conf/{client mds mon osd} distro/{ubuntu_latest} mount/kclient/{mount-syntax/{v1} mount overrides/{distro/testing/k-testing ms-die-on-skipped}} overrides/{ignorelist_health ignorelist_health_more ignorelist_wrongly_marked_down pg-warn} tasks/{0-client 1-tests/fscrypt-dbench}}
Command failed on smithi194 with status 128: 'cd /lib/firmware/updates && sudo git fetch origin && sudo git reset --hard origin/main'
wip-rishabh-2023May15-1524
wip-rishabh-2023May15-1524
main
smithi
ubuntu 20.04
fs/volumes/{begin/{0-install 1-ceph 2-logrotate} clusters/1a3s-mds-4c-client conf/{client mds mon osd} distro/{ubuntu_20.04} mount/kclient/{mount-syntax/{v2} mount overrides/{distro/testing/k-testing ms-die-on-skipped}} objectstore/bluestore-bitmap overrides/{ignorelist_health ignorelist_wrongly_marked_down no_client_pidfile} tasks/volumes/{overrides test/clone}}
SSH connection to smithi194 was lost: 'rm -f /tmp/linux-image.deb && echo linux-image-6.3.0-g2bcb17939031_6.3.0-g2bcb17939031-1_amd64.deb | wget -nv -O /tmp/linux-image.deb --base=https://2.chacra.ceph.com/r/kernel/testing/2bcb179390319a07fc43480888bd707374860ab3/ubuntu/focal/flavors/default/pool/main/l/linux-6.3.0-g2bcb17939031/ --input-file=-'
wip-rishabh-2023May15-1524
wip-rishabh-2023May15-1524
main
smithi
rhel 8.6
fs/workload/{0-rhel_8 begin/{0-install 1-cephadm 2-logrotate} clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} mount/kclient/{base/{mount-syntax/{v2} mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} ms_mode/crc wsync/no} objectstore-ec/bluestore-comp omap_limit/10 overrides/{cephsqlite-timeout frag ignorelist_health ignorelist_wrongly_marked_down osd-asserts session_timeout} ranks/multi/{export-check n/5 replication/default} standby-replay tasks/{0-subvolume/{with-namespace-isolated} 1-check-counter 2-scrub/no 3-snaps/yes 4-flush/yes 5-workunit/suites/ffsb}}
No module named 'tasks.fs'
wip-rishabh-2023May15-1524
wip-rishabh-2023May15-1524
main
smithi
ubuntu 22.04
fs/32bits/{begin/{0-install 1-ceph 2-logrotate} clusters/fixed-2-ucephfs conf/{client mds mon osd} distro/{ubuntu_latest} mount/fuse objectstore-ec/bluestore-bitmap overrides/{faked-ino ignorelist_health ignorelist_wrongly_marked_down} tasks/cfuse_workunit_suites_pjd}
Command failed on smithi202 with status 100: 'sudo DEBIAN_FRONTEND=noninteractive apt-get -y --force-yes -o Dpkg::Options::="--force-confdef" -o Dpkg::Options::="--force-confold" install ceph=18.0.0-3926-g27be5126-1jammy ceph-mds=18.0.0-3926-g27be5126-1jammy ceph-common=18.0.0-3926-g27be5126-1jammy ceph-fuse=18.0.0-3926-g27be5126-1jammy ceph-test=18.0.0-3926-g27be5126-1jammy radosgw=18.0.0-3926-g27be5126-1jammy python-ceph=18.0.0-3926-g27be5126-1jammy libcephfs1=18.0.0-3926-g27be5126-1jammy libcephfs-java=18.0.0-3926-g27be5126-1jammy libcephfs-jni=18.0.0-3926-g27be5126-1jammy librados2=18.0.0-3926-g27be5126-1jammy librbd1=18.0.0-3926-g27be5126-1jammy rbd-fuse=18.0.0-3926-g27be5126-1jammy python3-cephfs=18.0.0-3926-g27be5126-1jammy cephfs-shell=18.0.0-3926-g27be5126-1jammy cephfs-top=18.0.0-3926-g27be5126-1jammy cephfs-mirror=18.0.0-3926-g27be5126-1jammy'
wip-rishabh-2023May15-1524
wip-rishabh-2023May15-1524
main
smithi
ubuntu 22.04
fs/permission/{begin/{0-install 1-ceph 2-logrotate} clusters/fixed-2-ucephfs conf/{client mds mon osd} distro/{ubuntu_latest} mount/fuse objectstore-ec/bluestore-bitmap overrides/{ignorelist_health ignorelist_wrongly_marked_down} tasks/cfuse_workunit_suites_pjd}
Command failed on smithi154 with status 100: 'sudo DEBIAN_FRONTEND=noninteractive apt-get -y --force-yes -o Dpkg::Options::="--force-confdef" -o Dpkg::Options::="--force-confold" install ceph=18.0.0-3926-g27be5126-1jammy ceph-mds=18.0.0-3926-g27be5126-1jammy ceph-common=18.0.0-3926-g27be5126-1jammy ceph-fuse=18.0.0-3926-g27be5126-1jammy ceph-test=18.0.0-3926-g27be5126-1jammy radosgw=18.0.0-3926-g27be5126-1jammy python-ceph=18.0.0-3926-g27be5126-1jammy libcephfs1=18.0.0-3926-g27be5126-1jammy libcephfs-java=18.0.0-3926-g27be5126-1jammy libcephfs-jni=18.0.0-3926-g27be5126-1jammy librados2=18.0.0-3926-g27be5126-1jammy librbd1=18.0.0-3926-g27be5126-1jammy rbd-fuse=18.0.0-3926-g27be5126-1jammy python3-cephfs=18.0.0-3926-g27be5126-1jammy cephfs-shell=18.0.0-3926-g27be5126-1jammy cephfs-top=18.0.0-3926-g27be5126-1jammy cephfs-mirror=18.0.0-3926-g27be5126-1jammy'
wip-rishabh-2023May15-1524
wip-rishabh-2023May15-1524
main
smithi
ubuntu 22.04
fs/snaps/{begin/{0-install 1-ceph 2-logrotate} clusters/1a3s-mds-1c-client conf/{client mds mon osd} distro/{ubuntu_latest} mount/fuse objectstore-ec/bluestore-comp-ec-root overrides/{ignorelist_health ignorelist_wrongly_marked_down} tasks/workunit/snaps}
Command failed on smithi119 with status 100: 'sudo DEBIAN_FRONTEND=noninteractive apt-get -y --force-yes -o Dpkg::Options::="--force-confdef" -o Dpkg::Options::="--force-confold" install ceph=18.0.0-3926-g27be5126-1jammy ceph-mds=18.0.0-3926-g27be5126-1jammy ceph-common=18.0.0-3926-g27be5126-1jammy ceph-fuse=18.0.0-3926-g27be5126-1jammy ceph-test=18.0.0-3926-g27be5126-1jammy radosgw=18.0.0-3926-g27be5126-1jammy python-ceph=18.0.0-3926-g27be5126-1jammy libcephfs1=18.0.0-3926-g27be5126-1jammy libcephfs-java=18.0.0-3926-g27be5126-1jammy libcephfs-jni=18.0.0-3926-g27be5126-1jammy librados2=18.0.0-3926-g27be5126-1jammy librbd1=18.0.0-3926-g27be5126-1jammy rbd-fuse=18.0.0-3926-g27be5126-1jammy python3-cephfs=18.0.0-3926-g27be5126-1jammy cephfs-shell=18.0.0-3926-g27be5126-1jammy cephfs-top=18.0.0-3926-g27be5126-1jammy cephfs-mirror=18.0.0-3926-g27be5126-1jammy'
wip-rishabh-2023May15-1524
wip-rishabh-2023May15-1524
main
smithi
rhel 8.6
fs/workload/{0-rhel_8 begin/{0-install 1-cephadm 2-logrotate} clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} mount/kclient/{base/{mount-syntax/{v1} mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} ms_mode/secure wsync/no} objectstore-ec/bluestore-bitmap omap_limit/10 overrides/{cephsqlite-timeout frag ignorelist_health ignorelist_wrongly_marked_down osd-asserts session_timeout} ranks/multi/{export-check n/5 replication/always} standby-replay tasks/{0-subvolume/{with-namespace-isolated-and-quota} 1-check-counter 2-scrub/no 3-snaps/yes 4-flush/yes 5-workunit/suites/iogen}}
wip-rishabh-2023May15-1524
wip-rishabh-2023May15-1524
main
smithi
rhel 8.6
fs/workload/{0-rhel_8 begin/{0-install 1-cephadm 2-logrotate} clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} mount/kclient/{base/{mount-syntax/{v2} mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} ms_mode/secure wsync/no} objectstore-ec/bluestore-bitmap omap_limit/10 overrides/{cephsqlite-timeout frag ignorelist_health ignorelist_wrongly_marked_down osd-asserts session_timeout} ranks/multi/{export-check n/5 replication/default} standby-replay tasks/{0-subvolume/{with-namespace-isolated} 1-check-counter 2-scrub/no 3-snaps/no 4-flush/yes 5-workunit/fs/misc}}
wip-rishabh-2023May15-1524
wip-rishabh-2023May15-1524
main
smithi
rhel 8.6
fs/workload/{0-rhel_8 begin/{0-install 1-cephadm 2-logrotate} clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} mount/fuse objectstore-ec/bluestore-comp-ec-root omap_limit/10000 overrides/{cephsqlite-timeout frag ignorelist_health ignorelist_wrongly_marked_down osd-asserts session_timeout} ranks/multi/{export-check n/5 replication/always} standby-replay tasks/{0-subvolume/{with-namespace-isolated-and-quota} 1-check-counter 2-scrub/yes 3-snaps/yes 4-flush/no 5-workunit/kernel_untar_build}}
error during scrub thrashing: reached maximum tries (30) after waiting for 900 seconds
wip-rishabh-2023May15-1524
wip-rishabh-2023May15-1524
main
smithi
ubuntu 22.04
fs/traceless/{begin/{0-install 1-ceph 2-logrotate} clusters/fixed-2-ucephfs conf/{client mds mon osd} distro/{ubuntu_latest} mount/fuse objectstore-ec/bluestore-bitmap overrides/{frag ignorelist_health ignorelist_wrongly_marked_down} tasks/cfuse_workunit_suites_ffsb traceless/50pc}
Command failed (workunit test suites/ffsb.sh) on smithi123 with status 124: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=27be51263d3d0c721b6d1ca9f42550bf0ab3c97b TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 CEPH_MNT=/home/ubuntu/cephtest/mnt.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/suites/ffsb.sh'
wip-rishabh-2023May15-1524
wip-rishabh-2023May15-1524
main
smithi
rhel 8.6
fs/workload/{0-rhel_8 begin/{0-install 1-cephadm 2-logrotate} clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} mount/kclient/{base/{mount-syntax/{v2} mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} ms_mode/crc wsync/no} objectstore-ec/bluestore-comp omap_limit/10 overrides/{cephsqlite-timeout frag ignorelist_health ignorelist_wrongly_marked_down osd-asserts session_timeout} ranks/multi/{export-check n/3 replication/default} standby-replay tasks/{0-subvolume/{no-subvolume} 1-check-counter 2-scrub/no 3-snaps/no 4-flush/yes 5-workunit/postgres}}
wip-rishabh-2023May15-1524
wip-rishabh-2023May15-1524
main
smithi
ubuntu 22.04
fs/thrash/multifs/{begin/{0-install 1-ceph 2-logrotate} clusters/1a3s-mds-2c-client conf/{client mds mon osd} distro/{ubuntu_latest} mount/fuse msgr-failures/none objectstore/bluestore-bitmap overrides/{frag ignorelist_health ignorelist_wrongly_marked_down multifs session_timeout thrashosds-health} tasks/{1-thrash/mon 2-workunit/cfuse_workunit_snaptests}}
Command failed (workunit test fs/snaps/snaptest-multiple-capsnaps.sh) on smithi089 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=27be51263d3d0c721b6d1ca9f42550bf0ab3c97b TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 CEPH_MNT=/home/ubuntu/cephtest/mnt.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 6h /home/ubuntu/cephtest/clone.client.0/qa/workunits/fs/snaps/snaptest-multiple-capsnaps.sh'
wip-rishabh-2023May15-1524
wip-rishabh-2023May15-1524
main
smithi
ubuntu 22.04
fs/multiclient/{begin/{0-install 1-ceph 2-logrotate} clusters/1-mds-3-client conf/{client mds mon osd} distros/ubuntu_latest mount/fuse objectstore-ec/bluestore-comp overrides/{ignorelist_health ignorelist_wrongly_marked_down} tasks/ior-shared-file}
Command failed on smithi162 with status 2: 'TESTDIR=/home/ubuntu/cephtest bash -s'
wip-rishabh-2023May15-1524
wip-rishabh-2023May15-1524
main
smithi
centos 8.stream
fs/upgrade/featureful_client/old_client/{bluestore-bitmap centos_latest clusters/1-mds-2-client-micro conf/{client mds mon osd} overrides/{ignorelist_health ignorelist_wrongly_marked_down multimds/yes pg-warn} tasks/{0-octopus 1-client 2-upgrade 3-compat_client/quincy}}
wip-rishabh-2023May15-1524
wip-rishabh-2023May15-1524
main
smithi
rhel 8.6
fs/functional/{begin/{0-install 1-ceph 2-logrotate} clusters/1a3s-mds-4c-client conf/{client mds mon osd} distro/{rhel_8} mount/fuse objectstore/bluestore-ec-root overrides/{ignorelist_health ignorelist_wrongly_marked_down no_client_pidfile} tasks/quota}
Test failure: test_disable_enable_human_readable_quota_values (tasks.cephfs.test_quota.TestQuota)
wip-rishabh-2023May15-1524
wip-rishabh-2023May15-1524
main
smithi
rhel 8.6
fs/workload/{0-rhel_8 begin/{0-install 1-cephadm 2-logrotate} clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} mount/kclient/{base/{mount-syntax/{v2} mount overrides/{distro/testing/k-testing ms-die-on-skipped}} ms_mode/legacy wsync/yes} objectstore-ec/bluestore-ec-root omap_limit/10000 overrides/{cephsqlite-timeout frag ignorelist_health ignorelist_wrongly_marked_down osd-asserts session_timeout} ranks/multi/{export-check n/3 replication/default} standby-replay tasks/{0-subvolume/{with-quota} 1-check-counter 2-scrub/yes 3-snaps/yes 4-flush/no 5-workunit/suites/fsstress}}
wip-rishabh-2023May15-1524
wip-rishabh-2023May15-1524
main
smithi
centos 8.stream
fs/32bits/{begin/{0-install 1-ceph 2-logrotate} clusters/fixed-2-ucephfs conf/{client mds mon osd} distro/{centos_8} mount/fuse objectstore-ec/bluestore-comp overrides/{faked-ino ignorelist_health ignorelist_wrongly_marked_down} tasks/cfuse_workunit_suites_pjd}
Command failed (workunit test suites/pjd.sh) on smithi045 with status 124: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=27be51263d3d0c721b6d1ca9f42550bf0ab3c97b TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 CEPH_MNT=/home/ubuntu/cephtest/mnt.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 6h /home/ubuntu/cephtest/clone.client.0/qa/workunits/suites/pjd.sh'
wip-rishabh-2023May15-1524
wip-rishabh-2023May15-1524
main
smithi
ubuntu 20.04
fs/libcephfs/{begin/{0-install 1-ceph 2-logrotate} clusters/1-mds-1-client-coloc conf/{client mds mon osd} distro/{ubuntu_20.04} objectstore/bluestore-bitmap overrides/{ignorelist_health ignorelist_wrongly_marked_down no_client_pidfile} tasks/libcephfs_python}
Command failed (workunit test fs/test_python.sh) on smithi131 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=27be51263d3d0c721b6d1ca9f42550bf0ab3c97b TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 CEPH_MNT=/home/ubuntu/cephtest/mnt.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/fs/test_python.sh'
wip-rishabh-2023May15-1524
wip-rishabh-2023May15-1524
main
smithi
centos 8.stream
fs/permission/{begin/{0-install 1-ceph 2-logrotate} clusters/fixed-2-ucephfs conf/{client mds mon osd} distro/{centos_8} mount/fuse objectstore-ec/bluestore-comp overrides/{ignorelist_health ignorelist_wrongly_marked_down} tasks/cfuse_workunit_suites_pjd}
Command failed (workunit test suites/pjd.sh) on smithi046 with status 124: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=27be51263d3d0c721b6d1ca9f42550bf0ab3c97b TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 CEPH_MNT=/home/ubuntu/cephtest/mnt.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 6h /home/ubuntu/cephtest/clone.client.0/qa/workunits/suites/pjd.sh'
wip-rishabh-2023May15-1524
wip-rishabh-2023May15-1524
main
smithi
rhel 8.6
fs/workload/{0-rhel_8 begin/{0-install 1-cephadm 2-logrotate} clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} mount/fuse objectstore-ec/bluestore-comp-ec-root omap_limit/10000 overrides/{cephsqlite-timeout frag ignorelist_health ignorelist_wrongly_marked_down osd-asserts session_timeout} ranks/1 standby-replay tasks/{0-subvolume/{with-namespace-isolated} 1-check-counter 2-scrub/yes 3-snaps/yes 4-flush/no 5-workunit/suites/pjd}}
Command failed (workunit test suites/pjd.sh) on smithi037 with status 124: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=27be51263d3d0c721b6d1ca9f42550bf0ab3c97b TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 CEPH_MNT=/home/ubuntu/cephtest/mnt.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 6h /home/ubuntu/cephtest/clone.client.0/qa/workunits/suites/pjd.sh'
wip-rishabh-2023May15-1524
wip-rishabh-2023May15-1524
main
smithi
rhel 8.6
fs/verify/{begin/{0-install 1-ceph 2-logrotate} clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{rhel_8} mount/fuse objectstore-ec/bluestore-ec-root overrides/{ignorelist_health ignorelist_wrongly_marked_down mon-debug session_timeout} ranks/5 tasks/fsstress validater/lockdep}
hit max job timeout
wip-rishabh-2023May15-1524
wip-rishabh-2023May15-1524
main
smithi
rhel 8.6
fs/workload/{0-rhel_8 begin/{0-install 1-cephadm 2-logrotate} clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} mount/kclient/{base/{mount-syntax/{v2} mount overrides/{distro/testing/k-testing ms-die-on-skipped}} ms_mode/legacy wsync/yes} objectstore-ec/bluestore-ec-root omap_limit/10000 overrides/{cephsqlite-timeout frag ignorelist_health ignorelist_wrongly_marked_down osd-asserts session_timeout} ranks/multi/{export-check n/5 replication/default} standby-replay tasks/{0-subvolume/{with-no-extra-options} 1-check-counter 2-scrub/yes 3-snaps/yes 4-flush/no 5-workunit/fs/misc}}
error during scrub thrashing: reached maximum tries (30) after waiting for 900 seconds
wip-rishabh-2023May15-1524
wip-rishabh-2023May15-1524
main
smithi
ubuntu 22.04
fs/traceless/{begin/{0-install 1-ceph 2-logrotate} clusters/fixed-2-ucephfs conf/{client mds mon osd} distro/{ubuntu_latest} mount/fuse objectstore-ec/bluestore-comp-ec-root overrides/{frag ignorelist_health ignorelist_wrongly_marked_down} tasks/cfuse_workunit_suites_fsstress traceless/50pc}
Command failed (workunit test suites/fsstress.sh) on smithi112 with status 124: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=27be51263d3d0c721b6d1ca9f42550bf0ab3c97b TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 CEPH_MNT=/home/ubuntu/cephtest/mnt.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 6h /home/ubuntu/cephtest/clone.client.0/qa/workunits/suites/fsstress.sh'
wip-rishabh-2023May15-1524
wip-rishabh-2023May15-1524
main
smithi
rhel 8.6
fs/workload/{0-rhel_8 begin/{0-install 1-cephadm 2-logrotate} clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} mount/kclient/{base/{mount-syntax/{v2} mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} ms_mode/secure wsync/no} objectstore-ec/bluestore-bitmap omap_limit/10 overrides/{cephsqlite-timeout frag ignorelist_health ignorelist_wrongly_marked_down osd-asserts session_timeout} ranks/multi/{export-check n/5 replication/always} standby-replay tasks/{0-subvolume/{with-namespace-isolated-and-quota} 1-check-counter 2-scrub/no 3-snaps/no 4-flush/yes 5-workunit/kernel_untar_build}}
wip-rishabh-2023May15-1524
wip-rishabh-2023May15-1524
main
smithi
ubuntu 22.04
fs/multiclient/{begin/{0-install 1-ceph 2-logrotate} clusters/1-mds-2-client conf/{client mds mon osd} distros/ubuntu_latest mount/fuse objectstore-ec/bluestore-ec-root overrides/{ignorelist_health ignorelist_wrongly_marked_down} tasks/mdtest}
Command failed on smithi038 with status 2: 'TESTDIR=/home/ubuntu/cephtest bash -s'
wip-rishabh-2023May15-1524
wip-rishabh-2023May15-1524
main
smithi
rhel 8.6
fs/workload/{0-rhel_8 begin/{0-install 1-cephadm 2-logrotate} clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} mount/fuse objectstore-ec/bluestore-comp-ec-root omap_limit/10000 overrides/{cephsqlite-timeout frag ignorelist_health ignorelist_wrongly_marked_down osd-asserts session_timeout} ranks/multi/{export-check n/3 replication/default} standby-replay tasks/{0-subvolume/{with-namespace-isolated-and-quota} 1-check-counter 2-scrub/yes 3-snaps/yes 4-flush/no 5-workunit/postgres}}
error during scrub thrashing: reached maximum tries (30) after waiting for 900 seconds
wip-rishabh-2023May15-1524
wip-rishabh-2023May15-1524
main
smithi
rhel 8.6
fs/workload/{0-rhel_8 begin/{0-install 1-cephadm 2-logrotate} clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} mount/kclient/{base/{mount-syntax/{v2} mount overrides/{distro/testing/k-testing ms-die-on-skipped}} ms_mode/legacy wsync/yes} objectstore-ec/bluestore-ec-root omap_limit/10000 overrides/{cephsqlite-timeout frag ignorelist_health ignorelist_wrongly_marked_down osd-asserts session_timeout} ranks/multi/{export-check n/3 replication/always} standby-replay tasks/{0-subvolume/{with-namespace-isolated-and-quota} 1-check-counter 2-scrub/yes 3-snaps/yes 4-flush/no 5-workunit/suites/dbench}}
Command failed (workunit test suites/dbench.sh) on smithi115 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=27be51263d3d0c721b6d1ca9f42550bf0ab3c97b TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 CEPH_MNT=/home/ubuntu/cephtest/mnt.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/suites/dbench.sh'
wip-rishabh-2023May15-1524
wip-rishabh-2023May15-1524
main
smithi
ubuntu 20.04
fs/functional/{begin/{0-install 1-ceph 2-logrotate} clusters/1a3s-mds-4c-client conf/{client mds mon osd} distro/{ubuntu_20.04} mount/kclient/{mount-syntax/{v2} mount overrides/{distro/testing/k-testing ms-die-on-skipped}} objectstore/bluestore-bitmap overrides/{ignorelist_health ignorelist_wrongly_marked_down no_client_pidfile} tasks/xfstests-dev}
Test failure: test_generic (tasks.cephfs.tests_from_xfstests_dev.TestXFSTestsDev)
wip-rishabh-2023May15-1524
wip-rishabh-2023May15-1524
main
smithi
rhel 8.6
fs/workload/{0-rhel_8 begin/{0-install 1-cephadm 2-logrotate} clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} mount/kclient/{base/{mount-syntax/{v2} mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} ms_mode/crc wsync/no} objectstore-ec/bluestore-comp omap_limit/10 overrides/{cephsqlite-timeout frag ignorelist_health ignorelist_wrongly_marked_down osd-asserts session_timeout} ranks/multi/{export-check n/5 replication/always} standby-replay tasks/{0-subvolume/{no-subvolume} 1-check-counter 2-scrub/no 3-snaps/no 4-flush/yes 5-workunit/suites/iogen}}