ID
Status
Ceph Branch
Suite Branch
Teuthology Branch
Machine
OS
Nodes
Description
Failure Reason
master
master
master
gibba
centos 8.3
fs/functional/{begin clusters/1a3s-mds-4c-client conf/{client mds mon osd} distro/{rhel_8} mount/kclient/{mount overrides/{distro/testing/{flavor/centos_latest k-testing} ms-die-on-skipped}} objectstore/bluestore-ec-root overrides/{no_client_pidfile whitelist_health whitelist_wrongly_marked_down} tasks/backtrace}
master
master
master
gibba
rhel 8.4
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{centos_8} mount/kclient/{mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} ms_mode/{legacy} objectstore-ec/bluestore-comp omap_limit/10000 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/1 scrub/no standby-replay tasks/{0-check-counter workunit/suites/fsx} wsync/{no}}
master
master
master
gibba
centos 8.stream
fs/upgrade/mds_upgrade_sequence/{bluestore-bitmap centos_8.stream_container_tools conf/{client mds mon osd} overrides/{pg-warn whitelist_health whitelist_wrongly_marked_down} roles tasks/{0-from/pacific 1-volume/{0-create 1-ranks/1 2-allow_standby_replay/yes 3-verify} 2-client 3-upgrade-with-workload 4-verify}}
master
master
master
gibba
centos 8.3
fs/thrash/workloads/{begin clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{centos_8} mount/kclient/{mount overrides/{distro/testing/{flavor/centos_latest k-testing} ms-die-on-skipped}} msgr-failures/osd-mds-delay objectstore-ec/bluestore-comp-ec-root overrides/{frag session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} ranks/5 tasks/{1-thrash/mds 2-workunit/suites/fsstress}}
master
master
master
gibba
centos 8.3
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{centos_8.stream} mount/kclient/{mount overrides/{distro/testing/{flavor/centos_latest k-testing} ms-die-on-skipped}} ms_mode/{legacy} objectstore-ec/bluestore-ec-root omap_limit/10 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/3 scrub/yes standby-replay tasks/{0-check-counter workunit/suites/fsync-tester} wsync/{yes}}
master
master
master
gibba
ubuntu 20.04
fs/functional/{begin clusters/1a3s-mds-4c-client conf/{client mds mon osd} distro/{ubuntu_latest} mount/kclient/{mount overrides/{distro/testing/{flavor/ubuntu_latest k-testing} ms-die-on-skipped}} objectstore/bluestore-bitmap overrides/{no_client_pidfile whitelist_health whitelist_wrongly_marked_down} tasks/cap-flush}
master
master
master
gibba
ubuntu 20.04
fs/thrash/workloads/{begin clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{centos_8} mount/kclient/{mount overrides/{distro/testing/{flavor/ubuntu_latest k-testing} ms-die-on-skipped}} msgr-failures/none objectstore-ec/bluestore-comp overrides/{frag session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} ranks/1 tasks/{1-thrash/mon 2-workunit/fs/trivial_sync}}
master
master
master
gibba
ubuntu 20.04
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{centos_8.stream} mount/kclient/{mount overrides/{distro/testing/{flavor/ubuntu_latest k-testing} ms-die-on-skipped}} ms_mode/{secure} objectstore-ec/bluestore-bitmap omap_limit/10000 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/5 scrub/no standby-replay tasks/{0-check-counter workunit/fs/test_o_trunc} wsync/{yes}}
master
master
master
gibba
centos 8.3
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{centos_8} mount/fuse ms_mode/{legacy} objectstore-ec/bluestore-comp-ec-root omap_limit/10 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/1 scrub/yes standby-replay tasks/{0-check-counter workunit/suites/iogen} wsync/{yes}}
master
master
master
gibba
centos 8.3
fs/thrash/workloads/{begin clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{centos_8} mount/fuse msgr-failures/osd-mds-delay objectstore-ec/bluestore-ec-root overrides/{frag session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} ranks/3 tasks/{1-thrash/osd 2-workunit/suites/iozone}}
master
master
master
gibba
centos 8.stream
fs/functional/{begin clusters/1a3s-mds-4c-client conf/{client mds mon osd} distro/{centos_8.stream} mount/fuse objectstore/bluestore-ec-root overrides/{no_client_pidfile whitelist_health whitelist_wrongly_marked_down} tasks/client-limits}
master
master
master
gibba
rhel 8.4
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{centos_8} mount/kclient/{mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} ms_mode/{crc} objectstore-ec/bluestore-comp omap_limit/10000 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/3 scrub/no standby-replay tasks/{0-check-counter workunit/suites/iozone} wsync/{no}}
master
master
master
gibba
rhel 8.4
fs/thrash/workloads/{begin clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{centos_8} mount/kclient/{mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} msgr-failures/none objectstore-ec/bluestore-bitmap overrides/{frag session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} ranks/5 tasks/{1-thrash/mds 2-workunit/suites/pjd}}
master
master
master
gibba
centos 8.3
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{centos_8} mount/kclient/{mount overrides/{distro/testing/{flavor/centos_latest k-testing} ms-die-on-skipped}} ms_mode/{secure} objectstore-ec/bluestore-ec-root omap_limit/10 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/5 scrub/yes standby-replay tasks/{0-check-counter workunit/suites/pjd} wsync/{no}}
master
master
master
gibba
rhel 8.4
fs/functional/{begin clusters/1a3s-mds-4c-client conf/{client mds mon osd} distro/{ubuntu_latest} mount/kclient/{mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} objectstore/bluestore-bitmap overrides/{no_client_pidfile whitelist_health whitelist_wrongly_marked_down} tasks/client-readahead}
master
master
master
gibba
centos 8.stream
fs/32bits/{begin clusters/fixed-2-ucephfs conf/{client mds mon osd} distro/{centos_8.stream} mount/fuse objectstore-ec/bluestore-bitmap overrides/{faked-ino whitelist_health whitelist_wrongly_marked_down} tasks/cfuse_workunit_suites_fsstress}
master
master
master
gibba
centos 8.3
fs/bugs/client_trim_caps/{begin centos_latest clusters/small-cluster conf/{client mds mon osd} objectstore/bluestore-bitmap overrides/{no_client_pidfile whitelist_health whitelist_wrongly_marked_down} tasks/trim-i24137}
master
master
master
gibba
centos 8.3
fs/cephadm/multivolume/{0-start 1-mount 2-workload/dbench distro/centos_latest}
Command failed on gibba006 with status 5: 'sudo systemctl stop ceph-0841d3d6-4cb8-11ec-8c2d-001a4aab830c@mon.gibba006'
master
master
master
gibba
centos 8.stream
fs/full/{begin clusters/1-node-1-mds-1-osd conf/{client mds mon osd} distro/{centos_8.stream} mount/fuse objectstore/bluestore-bitmap overrides overrides/{no_client_pidfile whitelist_health whitelist_wrongly_marked_down} tasks/mgr-osd-full}
master
master
master
gibba
centos 8.stream
fs/libcephfs/{begin clusters/1-mds-1-client-coloc conf/{client mds mon osd} distro/{centos_8.stream} objectstore/bluestore-bitmap overrides/{no_client_pidfile whitelist_health whitelist_wrongly_marked_down} tasks/client}
master
master
master
gibba
rhel 8.4
fs/mirror/{begin cephfs-mirror/one-per-cluster clients/{mirror} cluster/{1-node} mount/fuse objectstore/bluestore-bitmap overrides/{whitelist_health} supported-random-distros$/{rhel_8} tasks/mirror}
master
master
master
gibba
centos 8.3
fs/mirror-ha/{begin cephfs-mirror/three-per-cluster clients/{mirror} cluster/{1-node} objectstore/bluestore-bitmap overrides/{whitelist_health} supported-random-distro$/{centos_8} workloads/cephfs-mirror-ha-workunit}
master
master
master
gibba
rhel 8.4
fs/mixed-clients/{begin clusters/1a3s-mds-2c-client conf/{client mds mon osd} kclient-overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped} objectstore-ec/bluestore-comp overrides/{osd-asserts whitelist_health whitelist_wrongly_marked_down} tasks/kernel_cfuse_workunits_dbench_iozone}
master
master
master
gibba
ubuntu 20.04
fs/multiclient/{begin clusters/1-mds-2-client conf/{client mds mon osd} distros/ubuntu_latest mount/fuse objectstore-ec/bluestore-comp overrides/{whitelist_health whitelist_wrongly_marked_down} tasks/cephfs_misc_tests}
master
master
master
gibba
rhel 8.4
fs/multifs/{begin clusters/1a3s-mds-2c-client conf/{client mds mon osd} distro/{centos_8.stream} mount/kclient/{mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} objectstore-ec/bluestore-comp overrides/{mon-debug whitelist_health whitelist_wrongly_marked_down} tasks/failover}
master
master
master
gibba
ubuntu 20.04
fs/permission/{begin clusters/fixed-2-ucephfs conf/{client mds mon osd} distro/{ubuntu_latest} mount/fuse objectstore-ec/bluestore-bitmap overrides/{whitelist_health whitelist_wrongly_marked_down} tasks/cfuse_workunit_misc}
master
master
master
gibba
ubuntu 20.04
fs/shell/{begin clusters/1-mds-1-client-coloc conf/{client mds mon osd} distro/{rhel_8} mount/fuse objectstore/bluestore-bitmap overrides/{no_client_pidfile whitelist_health whitelist_wrongly_marked_down} tasks/cephfs-shell}
master
master
master
gibba
rhel 8.4
fs/thrash/multifs/{begin clusters/1a3s-mds-2c-client conf/{client mds mon osd} distro/{ubuntu_latest} mount/kclient/{mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} msgr-failures/none objectstore/bluestore-bitmap overrides/{frag multifs session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} tasks/{1-thrash/mds 2-workunit/cfuse_workunit_trivial_sync}}
master
master
master
gibba
rhel 8.4
fs/top/{begin cluster/{1-node} mount/fuse objectstore/bluestore-bitmap overrides/whitelist_health supported-random-distros$/{rhel_8} tasks/fstop}
master
master
master
gibba
centos 8.3
fs/upgrade/featureful_client/old_client/{bluestore-bitmap centos_latest clusters/1-mds-2-client-micro conf/{client mds mon osd} overrides/{multimds/no pg-warn whitelist_health whitelist_wrongly_marked_down} tasks/{0-octopus 1-client 2-upgrade 3-compat_client/no}}
master
master
master
gibba
centos 8.3
fs/valgrind/{begin centos_latest debug mirror/{cephfs-mirror/one-per-cluster clients/mirror cluster/1-node mount/fuse overrides/whitelist_health tasks/mirror}}
Test failure: test_mirroring_init_failure_with_recovery (tasks.cephfs.test_mirroring.TestMirroring)
master
master
master
gibba
rhel 8.4
fs/verify/{begin clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{rhel_8} mount/fuse objectstore-ec/bluestore-comp overrides/{mon-debug session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/1 tasks/dbench validater/lockdep}
master
master
master
gibba
ubuntu 20.04
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{centos_8} mount/kclient/{mount overrides/{distro/testing/{flavor/ubuntu_latest k-testing} ms-die-on-skipped}} ms_mode/{secure} objectstore-ec/bluestore-ec-root omap_limit/10000 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/1 scrub/no standby-replay tasks/{0-check-counter workunit/direct_io} wsync/{yes}}
master
master
master
gibba
rhel 8.4
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{rhel_8} mount/fuse ms_mode/{secure} objectstore-ec/bluestore-bitmap omap_limit/10 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/3 scrub/yes standby-replay tasks/{0-check-counter workunit/fs/misc} wsync/{yes}}
master
master
master
gibba
centos 8.3
fs/thrash/workloads/{begin clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{centos_8.stream} mount/kclient/{mount overrides/{distro/testing/{flavor/centos_latest k-testing} ms-die-on-skipped}} msgr-failures/osd-mds-delay objectstore-ec/bluestore-comp-ec-root overrides/{frag session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} ranks/1 tasks/{1-thrash/mds 2-workunit/fs/snaps}}
master
master
master
gibba
centos 8.3
fs/functional/{begin clusters/1a3s-mds-4c-client conf/{client mds mon osd} distro/{rhel_8} mount/kclient/{mount overrides/{distro/testing/{flavor/centos_latest k-testing} ms-die-on-skipped}} objectstore/bluestore-ec-root overrides/{no_client_pidfile whitelist_health whitelist_wrongly_marked_down} tasks/client-recovery}
"2021-11-24T00:23:33.486183+0000 mon.a (mon.0) 576 : cluster [WRN] Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)" in cluster log
master
master
master
gibba
rhel 8.4
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{ubuntu_latest} mount/kclient/{mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} ms_mode/{crc} objectstore-ec/bluestore-comp-ec-root omap_limit/10000 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/5 scrub/no standby-replay tasks/{0-check-counter workunit/kernel_untar_build} wsync/{yes}}
master
master
master
gibba
ubuntu 20.04
fs/thrash/workloads/{begin clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{rhel_8} mount/kclient/{mount overrides/{distro/testing/{flavor/ubuntu_latest k-testing} ms-die-on-skipped}} msgr-failures/none objectstore-ec/bluestore-comp overrides/{frag session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} ranks/3 tasks/{1-thrash/mon 2-workunit/suites/ffsb}}
master
master
master
gibba
centos 8.3
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{centos_8.stream} mount/kclient/{mount overrides/{distro/testing/{flavor/centos_latest k-testing} ms-die-on-skipped}} ms_mode/{crc} objectstore-ec/bluestore-comp omap_limit/10 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/1 scrub/yes standby-replay tasks/{0-check-counter workunit/suites/blogbench} wsync/{yes}}
master
master
master
gibba
ubuntu 20.04
fs/functional/{begin clusters/1a3s-mds-4c-client conf/{client mds mon osd} distro/{centos_8} mount/kclient/{mount overrides/{distro/testing/{flavor/ubuntu_latest k-testing} ms-die-on-skipped}} objectstore/bluestore-bitmap overrides/{no_client_pidfile whitelist_health whitelist_wrongly_marked_down} tasks/damage}
master
master
master
gibba
centos 8.stream
fs/thrash/workloads/{begin clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{centos_8.stream} mount/fuse msgr-failures/osd-mds-delay objectstore-ec/bluestore-ec-root overrides/{frag session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} ranks/5 tasks/{1-thrash/osd 2-workunit/suites/fsstress}}
reached maximum tries (90) after waiting for 540 seconds
master
master
master
gibba
ubuntu 20.04
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{centos_8.stream} mount/kclient/{mount overrides/{distro/testing/{flavor/ubuntu_latest k-testing} ms-die-on-skipped}} ms_mode/{crc} objectstore-ec/bluestore-ec-root omap_limit/10000 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/3 scrub/no standby-replay tasks/{0-check-counter workunit/suites/dbench} wsync/{no}}
master
master
master
gibba
ubuntu 20.04
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{ubuntu_latest} mount/fuse ms_mode/{legacy} objectstore-ec/bluestore-bitmap omap_limit/10 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/5 scrub/yes standby-replay tasks/{0-check-counter workunit/suites/ffsb} wsync/{no}}
master
master
master
gibba
rhel 8.4
fs/thrash/workloads/{begin clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{rhel_8} mount/kclient/{mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} msgr-failures/none objectstore-ec/bluestore-bitmap overrides/{frag session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} ranks/1 tasks/{1-thrash/mds 2-workunit/fs/trivial_sync}}
master
master
master
gibba
centos 8.3
fs/functional/{begin clusters/1a3s-mds-4c-client conf/{client mds mon osd} distro/{centos_8} mount/fuse objectstore/bluestore-ec-root overrides/{no_client_pidfile whitelist_health whitelist_wrongly_marked_down} tasks/data-scan}
master
master
master
gibba
centos 8.stream
fs/upgrade/mds_upgrade_sequence/{bluestore-bitmap centos_8.stream_container_tools conf/{client mds mon osd} overrides/{pg-warn whitelist_health whitelist_wrongly_marked_down} roles tasks/{0-from/pacific 1-volume/{0-create 1-ranks/1 2-allow_standby_replay/no 3-verify} 2-client 3-upgrade-with-workload 4-verify}}
master
master
master
gibba
rhel 8.4
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{ubuntu_latest} mount/kclient/{mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} ms_mode/{crc} objectstore-ec/bluestore-comp-ec-root omap_limit/10000 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/1 scrub/no standby-replay tasks/{0-check-counter workunit/fs/norstats} wsync/{yes}}
master
master
master
gibba
centos 8.3
fs/thrash/workloads/{begin clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{centos_8} mount/kclient/{mount overrides/{distro/testing/{flavor/centos_latest k-testing} ms-die-on-skipped}} msgr-failures/osd-mds-delay objectstore-ec/bluestore-comp-ec-root overrides/{frag session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} ranks/3 tasks/{1-thrash/mon 2-workunit/suites/iozone}}
master
master
master
gibba
centos 8.3
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{ubuntu_latest} mount/kclient/{mount overrides/{distro/testing/{flavor/centos_latest k-testing} ms-die-on-skipped}} ms_mode/{secure} objectstore-ec/bluestore-comp omap_limit/10 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/3 scrub/yes standby-replay tasks/{0-check-counter workunit/suites/fsstress} wsync/{no}}
master
master
master
gibba
rhel 8.4
fs/functional/{begin clusters/1a3s-mds-4c-client conf/{client mds mon osd} distro/{rhel_8} mount/kclient/{mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} objectstore/bluestore-bitmap overrides/{no_client_pidfile whitelist_health whitelist_wrongly_marked_down} tasks/exports}
master
master
master
gibba
ubuntu 20.04
fs/snaps/{begin clusters/1a3s-mds-1c-client conf/{client mds mon osd} distro/{rhel_8} mount/kclient/{mount overrides/{distro/testing/{flavor/ubuntu_latest k-testing} ms-die-on-skipped}} objectstore-ec/bluestore-comp overrides/{whitelist_health whitelist_wrongly_marked_down} tasks/workunit/snaps}
master
master
master
gibba
centos 8.3
fs/traceless/{begin clusters/fixed-2-ucephfs conf/{client mds mon osd} distro/{centos_8} mount/fuse objectstore-ec/bluestore-ec-root overrides/{frag whitelist_health whitelist_wrongly_marked_down} tasks/cfuse_workunit_suites_ffsb traceless/50pc}
master
master
master
gibba
ubuntu 20.04
fs/volumes/{begin clusters/1a3s-mds-4c-client conf/{client mds mon osd} distro/{ubuntu_latest} mount/kclient/{mount overrides/{distro/testing/{flavor/ubuntu_latest k-testing} ms-die-on-skipped}} objectstore/bluestore-bitmap overrides/{no_client_pidfile whitelist_health whitelist_wrongly_marked_down} tasks/volumes/{overrides test/misc}}
master
master
master
gibba
ubuntu 20.04
fs/thrash/workloads/{begin clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{centos_8} mount/kclient/{mount overrides/{distro/testing/{flavor/ubuntu_latest k-testing} ms-die-on-skipped}} msgr-failures/none objectstore-ec/bluestore-comp overrides/{frag session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} ranks/5 tasks/{1-thrash/osd 2-workunit/suites/pjd}}
master
master
master
gibba
ubuntu 20.04
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{centos_8} mount/kclient/{mount overrides/{distro/testing/{flavor/ubuntu_latest k-testing} ms-die-on-skipped}} ms_mode/{legacy} objectstore-ec/bluestore-ec-root omap_limit/10000 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/5 scrub/no standby-replay tasks/{0-check-counter workunit/suites/fsx} wsync/{yes}}
master
master
master
gibba
centos 8.stream
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{centos_8.stream} mount/fuse ms_mode/{secure} objectstore-ec/bluestore-bitmap omap_limit/10 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/1 scrub/yes standby-replay tasks/{0-check-counter workunit/suites/fsync-tester} wsync/{no}}
master
master
master
gibba
centos 8.3
fs/thrash/multifs/{begin clusters/1a3s-mds-2c-client conf/{client mds mon osd} distro/{centos_8.stream} mount/kclient/{mount overrides/{distro/testing/{flavor/centos_latest k-testing} ms-die-on-skipped}} msgr-failures/osd-mds-delay objectstore/bluestore-bitmap overrides/{frag multifs session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} tasks/{1-thrash/mon 2-workunit/ffsb}}
master
master
master
gibba
centos 8.3
fs/functional/{begin clusters/1a3s-mds-4c-client conf/{client mds mon osd} distro/{centos_8.stream} mount/kclient/{mount overrides/{distro/testing/{flavor/centos_latest k-testing} ms-die-on-skipped}} objectstore/bluestore-ec-root overrides/{no_client_pidfile whitelist_health whitelist_wrongly_marked_down} tasks/forward-scrub}
master
master
master
gibba
rhel 8.4
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{ubuntu_latest} mount/kclient/{mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} ms_mode/{crc} objectstore-ec/bluestore-comp-ec-root omap_limit/10000 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/3 scrub/no standby-replay tasks/{0-check-counter workunit/fs/test_o_trunc} wsync/{yes}}
master
master
master
gibba
centos 8.3
fs/32bits/{begin clusters/fixed-2-ucephfs conf/{client mds mon osd} distro/{centos_8} mount/fuse objectstore-ec/bluestore-comp-ec-root overrides/{faked-ino whitelist_health whitelist_wrongly_marked_down} tasks/cfuse_workunit_suites_pjd}
master
master
master
gibba
centos 8.stream
fs/permission/{begin clusters/fixed-2-ucephfs conf/{client mds mon osd} distro/{centos_8.stream} mount/fuse objectstore-ec/bluestore-comp-ec-root overrides/{whitelist_health whitelist_wrongly_marked_down} tasks/cfuse_workunit_suites_pjd}
master
master
master
gibba
centos 8.3
fs/thrash/workloads/{begin clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{centos_8} mount/fuse msgr-failures/osd-mds-delay objectstore-ec/bluestore-ec-root overrides/{frag session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} ranks/1 tasks/{1-thrash/osd 2-workunit/fs/snaps}}
master
master
master
gibba
centos 8.3
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{rhel_8} mount/kclient/{mount overrides/{distro/testing/{flavor/centos_latest k-testing} ms-die-on-skipped}} ms_mode/{crc} objectstore-ec/bluestore-comp omap_limit/10 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/5 scrub/yes standby-replay tasks/{0-check-counter workunit/suites/iogen} wsync/{yes}}
The following counters failed to be set on mds daemons: {'mds.imported', 'mds.exported'}
master
master
master
gibba
centos 8.3
fs/upgrade/nofs/{bluestore-bitmap centos_latest conf/{client mds mon osd} no-mds-cluster overrides/{pg-warn whitelist_health whitelist_wrongly_marked_down} tasks/{0-octopus 1-upgrade}}
master
master
master
gibba
ubuntu 20.04
fs/functional/{begin clusters/1a3s-mds-4c-client conf/{client mds mon osd} distro/{rhel_8} mount/kclient/{mount overrides/{distro/testing/{flavor/ubuntu_latest k-testing} ms-die-on-skipped}} objectstore/bluestore-bitmap overrides/{no_client_pidfile whitelist_health whitelist_wrongly_marked_down} tasks/fragment}
master
master
master
gibba
rhel 8.4
fs/thrash/workloads/{begin clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{centos_8} mount/kclient/{mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} msgr-failures/none objectstore-ec/bluestore-bitmap overrides/{frag session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} ranks/3 tasks/{1-thrash/mds 2-workunit/suites/ffsb}}
master
master
master
gibba
ubuntu 20.04
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{rhel_8} mount/kclient/{mount overrides/{distro/testing/{flavor/ubuntu_latest k-testing} ms-die-on-skipped}} ms_mode/{crc} objectstore-ec/bluestore-ec-root omap_limit/10000 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/1 scrub/no standby-replay tasks/{0-check-counter workunit/suites/iozone} wsync/{yes}}
master
master
master
gibba
centos 8.stream
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{centos_8.stream} mount/fuse ms_mode/{secure} objectstore-ec/bluestore-bitmap omap_limit/10 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/3 scrub/yes standby-replay tasks/{0-check-counter workunit/suites/pjd} wsync/{no}}
master
master
master
gibba
centos 8.3
fs/thrash/workloads/{begin clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{centos_8.stream} mount/kclient/{mount overrides/{distro/testing/{flavor/centos_latest k-testing} ms-die-on-skipped}} msgr-failures/osd-mds-delay objectstore-ec/bluestore-comp-ec-root overrides/{frag session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} ranks/5 tasks/{1-thrash/mon 2-workunit/suites/fsstress}}
master
master
master
gibba
centos 8.3
fs/functional/{begin clusters/1a3s-mds-4c-client conf/{client mds mon osd} distro/{centos_8} mount/fuse objectstore/bluestore-ec-root overrides/{no_client_pidfile whitelist_health whitelist_wrongly_marked_down} tasks/journal-repair}
master
master
master
gibba
centos 8.3
fs/mixed-clients/{begin clusters/1a3s-mds-2c-client conf/{client mds mon osd} kclient-overrides/{distro/testing/{flavor/centos_latest k-testing} ms-die-on-skipped} objectstore-ec/bluestore-ec-root overrides/{osd-asserts whitelist_health whitelist_wrongly_marked_down} tasks/kernel_cfuse_workunits_untarbuild_blogbench}
master
master
master
gibba
ubuntu 20.04
fs/multiclient/{begin clusters/1-mds-3-client conf/{client mds mon osd} distros/ubuntu_latest mount/fuse objectstore-ec/bluestore-ec-root overrides/{whitelist_health whitelist_wrongly_marked_down} tasks/ior-shared-file}
master
master
master
gibba
centos 8.3
fs/multifs/{begin clusters/1a3s-mds-2c-client conf/{client mds mon osd} distro/{centos_8} mount/kclient/{mount overrides/{distro/testing/{flavor/centos_latest k-testing} ms-die-on-skipped}} objectstore-ec/bluestore-ec-root overrides/{mon-debug whitelist_health whitelist_wrongly_marked_down} tasks/multifs-auth}
master
master
master
gibba
centos 8.3
fs/verify/{begin clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{centos_8} mount/kclient/{k-testing mount ms-die-on-skipped} objectstore-ec/bluestore-ec-root overrides/{mon-debug session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/3 tasks/fsstress validater/valgrind}
master
master
master
gibba
rhel 8.4
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{centos_8.stream} mount/kclient/{mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} ms_mode/{crc} objectstore-ec/bluestore-comp-ec-root omap_limit/10000 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/5 scrub/no standby-replay tasks/{0-check-counter workunit/direct_io} wsync/{yes}}
master
master
master
gibba
ubuntu 20.04
fs/thrash/workloads/{begin clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{centos_8} mount/kclient/{mount overrides/{distro/testing/{flavor/ubuntu_latest k-testing} ms-die-on-skipped}} msgr-failures/none objectstore-ec/bluestore-comp overrides/{frag session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} ranks/1 tasks/{1-thrash/osd 2-workunit/fs/trivial_sync}}
master
master
master
gibba
centos 8.3
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{centos_8.stream} mount/kclient/{mount overrides/{distro/testing/{flavor/centos_latest k-testing} ms-die-on-skipped}} ms_mode/{secure} objectstore-ec/bluestore-comp omap_limit/10 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/1 scrub/yes standby-replay tasks/{0-check-counter workunit/fs/misc} wsync/{yes}}
master
master
master
gibba
rhel 8.4
fs/functional/{begin clusters/1a3s-mds-4c-client conf/{client mds mon osd} distro/{centos_8} mount/kclient/{mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} objectstore/bluestore-bitmap overrides/{no_client_pidfile whitelist_health whitelist_wrongly_marked_down} tasks/mds-flush}
master
master
master
gibba
centos 8.stream
fs/thrash/workloads/{begin clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{centos_8.stream} mount/fuse msgr-failures/osd-mds-delay objectstore-ec/bluestore-ec-root overrides/{frag session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} ranks/3 tasks/{1-thrash/mds 2-workunit/suites/iozone}}
master
master
master
gibba
ubuntu 20.04
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{centos_8.stream} mount/kclient/{mount overrides/{distro/testing/{flavor/ubuntu_latest k-testing} ms-die-on-skipped}} ms_mode/{legacy} objectstore-ec/bluestore-ec-root omap_limit/10000 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/3 scrub/no standby-replay tasks/{0-check-counter workunit/kernel_untar_build} wsync/{no}}
master
master
master
gibba
centos 8.stream
fs/upgrade/mds_upgrade_sequence/{bluestore-bitmap centos_8.stream_container_tools conf/{client mds mon osd} overrides/{pg-warn whitelist_health whitelist_wrongly_marked_down} roles tasks/{0-from/v16.2.4 1-volume/{0-create 1-ranks/2 2-allow_standby_replay/yes 3-verify} 2-client 3-upgrade-with-workload 4-verify}}
master
master
master
gibba
ubuntu 20.04
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{ubuntu_latest} mount/fuse ms_mode/{legacy} objectstore-ec/bluestore-bitmap omap_limit/10 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/5 scrub/yes standby-replay tasks/{0-check-counter workunit/suites/blogbench} wsync/{yes}}
master
master
master
gibba
rhel 8.4
fs/thrash/workloads/{begin clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{rhel_8} mount/kclient/{mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} msgr-failures/none objectstore-ec/bluestore-bitmap overrides/{frag session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} ranks/5 tasks/{1-thrash/mon 2-workunit/suites/pjd}}
master
master
master
gibba
centos 8.3
fs/functional/{begin clusters/1a3s-mds-4c-client conf/{client mds mon osd} distro/{centos_8.stream} mount/kclient/{mount overrides/{distro/testing/{flavor/centos_latest k-testing} ms-die-on-skipped}} objectstore/bluestore-ec-root overrides/{no_client_pidfile whitelist_health whitelist_wrongly_marked_down} tasks/mds-full}
master
master
master
gibba
rhel 8.4
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{centos_8.stream} mount/kclient/{mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} ms_mode/{legacy} objectstore-ec/bluestore-comp-ec-root omap_limit/10000 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/1 scrub/no standby-replay tasks/{0-check-counter workunit/suites/dbench} wsync/{yes}}
master
master
master
gibba
ubuntu 20.04
fs/thrash/multifs/{begin clusters/1a3s-mds-2c-client conf/{client mds mon osd} distro/{centos_8.stream} mount/kclient/{mount overrides/{distro/testing/{flavor/ubuntu_latest k-testing} ms-die-on-skipped}} msgr-failures/none objectstore/bluestore-bitmap overrides/{frag multifs session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} tasks/{1-thrash/mds 2-workunit/iozone}}
master
master
master
gibba
centos 8.3
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{centos_8} mount/kclient/{mount overrides/{distro/testing/{flavor/centos_latest k-testing} ms-die-on-skipped}} ms_mode/{legacy} objectstore-ec/bluestore-comp omap_limit/10 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/3 scrub/yes standby-replay tasks/{0-check-counter workunit/suites/ffsb} wsync/{yes}}
master
master
master
gibba
ubuntu 20.04
fs/functional/{begin clusters/1a3s-mds-4c-client conf/{client mds mon osd} distro/{rhel_8} mount/kclient/{mount overrides/{distro/testing/{flavor/ubuntu_latest k-testing} ms-die-on-skipped}} objectstore/bluestore-bitmap overrides/{no_client_pidfile whitelist_health whitelist_wrongly_marked_down} tasks/mds_creation_retry}
master
master
master
gibba
centos 8.3
fs/thrash/workloads/{begin clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{ubuntu_latest} mount/kclient/{mount overrides/{distro/testing/{flavor/centos_latest k-testing} ms-die-on-skipped}} msgr-failures/osd-mds-delay objectstore-ec/bluestore-comp-ec-root overrides/{frag session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} ranks/1 tasks/{1-thrash/mon 2-workunit/fs/snaps}}
master
master
master
gibba
ubuntu 20.04
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{centos_8.stream} mount/kclient/{mount overrides/{distro/testing/{flavor/ubuntu_latest k-testing} ms-die-on-skipped}} ms_mode/{legacy} objectstore-ec/bluestore-ec-root omap_limit/10000 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/5 scrub/no standby-replay tasks/{0-check-counter workunit/fs/norstats} wsync/{yes}}
master
master
master
gibba
rhel 8.4
fs/32bits/{begin clusters/fixed-2-ucephfs conf/{client mds mon osd} distro/{rhel_8} mount/fuse objectstore-ec/bluestore-comp overrides/{faked-ino whitelist_health whitelist_wrongly_marked_down} tasks/cfuse_workunit_suites_fsstress}
master
master
master
gibba
ubuntu 20.04
fs/libcephfs/{begin clusters/1-mds-1-client-coloc conf/{client mds mon osd} distro/{ubuntu_latest} objectstore/bluestore-bitmap overrides/{no_client_pidfile whitelist_health whitelist_wrongly_marked_down} tasks/ino_release_cb}
master
master
master
gibba
rhel 8.4
fs/permission/{begin clusters/fixed-2-ucephfs conf/{client mds mon osd} distro/{rhel_8} mount/fuse objectstore-ec/bluestore-comp overrides/{whitelist_health whitelist_wrongly_marked_down} tasks/cfuse_workunit_misc}
master
master
master
gibba
ubuntu 20.04
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{ubuntu_latest} mount/fuse ms_mode/{legacy} objectstore-ec/bluestore-bitmap omap_limit/10 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/1 scrub/yes standby-replay tasks/{0-check-counter workunit/suites/fsstress} wsync/{no}}
master
master
master
gibba
ubuntu 20.04
fs/thrash/workloads/{begin clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{ubuntu_latest} mount/kclient/{mount overrides/{distro/testing/{flavor/ubuntu_latest k-testing} ms-die-on-skipped}} msgr-failures/none objectstore-ec/bluestore-comp overrides/{frag session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} ranks/3 tasks/{1-thrash/osd 2-workunit/suites/ffsb}}
master
master
master
gibba
rhel 8.4
fs/functional/{begin clusters/1a3s-mds-4c-client conf/{client mds mon osd} distro/{rhel_8} mount/fuse objectstore/bluestore-ec-root overrides/{no_client_pidfile whitelist_health whitelist_wrongly_marked_down} tasks/metrics}
master
master
master
gibba
rhel 8.4
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{centos_8} mount/kclient/{mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} ms_mode/{legacy} objectstore-ec/bluestore-comp-ec-root omap_limit/10000 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/3 scrub/no standby-replay tasks/{0-check-counter workunit/suites/fsx} wsync/{no}}
master
master
master
gibba
centos 8.3
fs/upgrade/featureful_client/upgraded_client/{bluestore-bitmap centos_latest clusters/1-mds-2-client-micro conf/{client mds mon osd} overrides/{multimds/no pg-warn whitelist_health whitelist_wrongly_marked_down} tasks/{0-octopus 1-client 2-upgrade 3-client-upgrade 4-compat_client 5-client-sanity}}
master
master
master
gibba
ubuntu 20.04
fs/thrash/workloads/{begin clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{ubuntu_latest} mount/fuse msgr-failures/osd-mds-delay objectstore-ec/bluestore-ec-root overrides/{frag session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} ranks/5 tasks/{1-thrash/mds 2-workunit/suites/fsstress}}
reached maximum tries (90) after waiting for 540 seconds
master
master
master
gibba
centos 8.3
fs/snaps/{begin clusters/1a3s-mds-1c-client conf/{client mds mon osd} distro/{centos_8} mount/fuse objectstore-ec/bluestore-ec-root overrides/{whitelist_health whitelist_wrongly_marked_down} tasks/workunit/snaps}
master
master
master
gibba
centos 8.stream
fs/traceless/{begin clusters/fixed-2-ucephfs conf/{client mds mon osd} distro/{centos_8.stream} mount/fuse objectstore-ec/bluestore-bitmap overrides/{frag whitelist_health whitelist_wrongly_marked_down} tasks/cfuse_workunit_suites_fsstress traceless/50pc}
master
master
master
gibba
ubuntu 20.04
fs/volumes/{begin clusters/1a3s-mds-4c-client conf/{client mds mon osd} distro/{ubuntu_latest} mount/fuse objectstore/bluestore-bitmap overrides/{no_client_pidfile whitelist_health whitelist_wrongly_marked_down} tasks/volumes/{overrides test/snapshot}}
master
master
master
gibba
centos 8.3
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{centos_8.stream} mount/kclient/{mount overrides/{distro/testing/{flavor/centos_latest k-testing} ms-die-on-skipped}} ms_mode/{crc} objectstore-ec/bluestore-comp omap_limit/10 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/5 scrub/yes standby-replay tasks/{0-check-counter workunit/suites/fsync-tester} wsync/{yes}}
master
master
master
gibba
rhel 8.4
fs/functional/{begin clusters/1a3s-mds-4c-client conf/{client mds mon osd} distro/{ubuntu_latest} mount/kclient/{mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} objectstore/bluestore-bitmap overrides/{no_client_pidfile whitelist_health whitelist_wrongly_marked_down} tasks/multimds_misc}
master
master
master
gibba
rhel 8.4
fs/thrash/workloads/{begin clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{ubuntu_latest} mount/kclient/{mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} msgr-failures/none objectstore-ec/bluestore-bitmap overrides/{frag session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} ranks/1 tasks/{1-thrash/mon 2-workunit/fs/trivial_sync}}
master
master
master
gibba
ubuntu 20.04
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{centos_8} mount/kclient/{mount overrides/{distro/testing/{flavor/ubuntu_latest k-testing} ms-die-on-skipped}} ms_mode/{crc} objectstore-ec/bluestore-ec-root omap_limit/10000 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/1 scrub/no standby-replay tasks/{0-check-counter workunit/fs/test_o_trunc} wsync/{no}}
master
master
master
gibba
centos 8.stream
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{centos_8.stream} mount/fuse ms_mode/{crc} objectstore-ec/bluestore-bitmap omap_limit/10 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/3 scrub/yes standby-replay tasks/{0-check-counter workunit/suites/iogen} wsync/{yes}}
master
master
master
gibba
centos 8.3
fs/thrash/workloads/{begin clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{ubuntu_latest} mount/kclient/{mount overrides/{distro/testing/{flavor/centos_latest k-testing} ms-die-on-skipped}} msgr-failures/osd-mds-delay objectstore-ec/bluestore-comp-ec-root overrides/{frag session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} ranks/3 tasks/{1-thrash/osd 2-workunit/suites/iozone}}
master
master
master
gibba
centos 8.3
fs/functional/{begin clusters/1a3s-mds-4c-client conf/{client mds mon osd} distro/{centos_8} mount/kclient/{mount overrides/{distro/testing/{flavor/centos_latest k-testing} ms-die-on-skipped}} objectstore/bluestore-ec-root overrides/{no_client_pidfile whitelist_health whitelist_wrongly_marked_down} tasks/openfiletable}
master
master
master
gibba
rhel 8.4
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{centos_8.stream} mount/kclient/{mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} ms_mode/{crc} objectstore-ec/bluestore-comp-ec-root omap_limit/10000 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/5 scrub/no standby-replay tasks/{0-check-counter workunit/suites/iozone} wsync/{yes}}
master
master
master
gibba
ubuntu 20.04
fs/thrash/workloads/{begin clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{ubuntu_latest} mount/kclient/{mount overrides/{distro/testing/{flavor/ubuntu_latest k-testing} ms-die-on-skipped}} msgr-failures/none objectstore-ec/bluestore-comp overrides/{frag session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} ranks/5 tasks/{1-thrash/mds 2-workunit/suites/pjd}}
master
master
master
gibba
centos 8.3
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{ubuntu_latest} mount/kclient/{mount overrides/{distro/testing/{flavor/centos_latest k-testing} ms-die-on-skipped}} ms_mode/{legacy} objectstore-ec/bluestore-comp omap_limit/10 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/1 scrub/yes standby-replay tasks/{0-check-counter workunit/suites/pjd} wsync/{no}}
master
master
master
gibba
ubuntu 20.04
fs/functional/{begin clusters/1a3s-mds-4c-client conf/{client mds mon osd} distro/{centos_8.stream} mount/kclient/{mount overrides/{distro/testing/{flavor/ubuntu_latest k-testing} ms-die-on-skipped}} objectstore/bluestore-bitmap overrides/{no_client_pidfile whitelist_health whitelist_wrongly_marked_down} tasks/pool-perm}
master
master
master
gibba
ubuntu 20.04
fs/mixed-clients/{begin clusters/1a3s-mds-2c-client conf/{client mds mon osd} kclient-overrides/{distro/testing/{flavor/ubuntu_latest k-testing} ms-die-on-skipped} objectstore-ec/bluestore-bitmap overrides/{osd-asserts whitelist_health whitelist_wrongly_marked_down} tasks/kernel_cfuse_workunits_dbench_iozone}
master
master
master
gibba
ubuntu 20.04
fs/multiclient/{begin clusters/1-mds-2-client conf/{client mds mon osd} distros/ubuntu_latest mount/fuse objectstore-ec/bluestore-bitmap overrides/{whitelist_health whitelist_wrongly_marked_down} tasks/mdtest}
master
master
master
gibba
centos 8.3
fs/multifs/{begin clusters/1a3s-mds-2c-client conf/{client mds mon osd} distro/{centos_8.stream} mount/kclient/{mount overrides/{distro/testing/{flavor/centos_latest k-testing} ms-die-on-skipped}} objectstore-ec/bluestore-bitmap overrides/{mon-debug whitelist_health whitelist_wrongly_marked_down} tasks/failover}
master
master
master
gibba
ubuntu 20.04
fs/thrash/multifs/{begin clusters/1a3s-mds-2c-client conf/{client mds mon osd} distro/{rhel_8} mount/kclient/{mount overrides/{distro/testing/{flavor/ubuntu_latest k-testing} ms-die-on-skipped}} msgr-failures/osd-mds-delay objectstore/bluestore-bitmap overrides/{frag multifs session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} tasks/{1-thrash/mds 2-workunit/cfuse_workunit_snaptests}}
Command failed (workunit test fs/snaps/snaptest-git-ceph.sh) on gibba006 with status 128: 'mkdir -p -- /home/ubuntu/cephtest/mnt.1/client.1/tmp && cd -- /home/ubuntu/cephtest/mnt.1/client.1/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=292b9a3ce266acd070bcc99c4919beac2f88ae61 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="1" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.1 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.1 CEPH_MNT=/home/ubuntu/cephtest/mnt.1 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.1/qa/workunits/fs/snaps/snaptest-git-ceph.sh'
master
master
master
gibba
centos 8.stream
fs/upgrade/mds_upgrade_sequence/{bluestore-bitmap centos_8.stream_container_tools conf/{client mds mon osd} overrides/{pg-warn whitelist_health whitelist_wrongly_marked_down} roles tasks/{0-from/pacific 1-volume/{0-create 1-ranks/2 2-allow_standby_replay/no 3-verify} 2-client 3-upgrade-with-workload 4-verify}}
master
master
master
gibba
rhel 8.4
fs/verify/{begin clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{rhel_8} mount/fuse objectstore-ec/bluestore-bitmap overrides/{mon-debug session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/5 tasks/dbench validater/lockdep}
master
master
master
gibba
centos 8.3
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{rhel_8} mount/kclient/{mount overrides/{distro/testing/{flavor/centos_latest k-testing} ms-die-on-skipped}} ms_mode/{secure} objectstore-ec/bluestore-bitmap omap_limit/10000 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/3 scrub/no standby-replay tasks/{0-check-counter workunit/direct_io} wsync/{no}}
master
master
master
gibba
ubuntu 20.04
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{rhel_8} mount/kclient/{mount overrides/{distro/testing/{flavor/ubuntu_latest k-testing} ms-die-on-skipped}} ms_mode/{crc} objectstore-ec/bluestore-comp-ec-root omap_limit/10 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/5 scrub/yes standby-replay tasks/{0-check-counter workunit/fs/misc} wsync/{no}}
Command failed (workunit test fs/misc/multiple_rsync.sh) on gibba013 with status 124: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=292b9a3ce266acd070bcc99c4919beac2f88ae61 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 CEPH_MNT=/home/ubuntu/cephtest/mnt.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/fs/misc/multiple_rsync.sh'
master
master
master
gibba
ubuntu 20.04
fs/thrash/workloads/{begin clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{ubuntu_latest} mount/kclient/{mount overrides/{distro/testing/{flavor/ubuntu_latest k-testing} ms-die-on-skipped}} msgr-failures/osd-mds-delay objectstore-ec/bluestore-bitmap overrides/{frag session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} ranks/1 tasks/{1-thrash/mds 2-workunit/fs/snaps}}
master
master
master
gibba
centos 8.3
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{centos_8} mount/fuse ms_mode/{crc} objectstore-ec/bluestore-comp omap_limit/10000 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/1 scrub/no standby-replay tasks/{0-check-counter workunit/kernel_untar_build} wsync/{no}}
master
master
master
gibba
centos 8.3
fs/functional/{begin clusters/1a3s-mds-4c-client conf/{client mds mon osd} distro/{centos_8} mount/fuse objectstore/bluestore-ec-root overrides/{no_client_pidfile whitelist_health whitelist_wrongly_marked_down} tasks/quota}
master
master
master
gibba
centos 8.3
fs/thrash/workloads/{begin clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{centos_8} mount/fuse msgr-failures/none objectstore-ec/bluestore-comp-ec-root overrides/{frag session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} ranks/3 tasks/{1-thrash/mon 2-workunit/suites/ffsb}}
master
master
master
gibba
rhel 8.4
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{centos_8.stream} mount/kclient/{mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} ms_mode/{secure} objectstore-ec/bluestore-ec-root omap_limit/10 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/3 scrub/yes standby-replay tasks/{0-check-counter workunit/suites/blogbench} wsync/{yes}}
master
master
master
gibba
centos 8.stream
fs/32bits/{begin clusters/fixed-2-ucephfs conf/{client mds mon osd} distro/{centos_8.stream} mount/fuse objectstore-ec/bluestore-ec-root overrides/{faked-ino whitelist_health whitelist_wrongly_marked_down} tasks/cfuse_workunit_suites_pjd}
master
master
master
gibba
ubuntu 20.04
fs/permission/{begin clusters/fixed-2-ucephfs conf/{client mds mon osd} distro/{ubuntu_latest} mount/fuse objectstore-ec/bluestore-ec-root overrides/{whitelist_health whitelist_wrongly_marked_down} tasks/cfuse_workunit_suites_pjd}
master
master
master
gibba
rhel 8.4
fs/thrash/workloads/{begin clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{rhel_8} mount/kclient/{mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} msgr-failures/osd-mds-delay objectstore-ec/bluestore-comp overrides/{frag session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} ranks/5 tasks/{1-thrash/osd 2-workunit/suites/fsstress}}
master
master
master
gibba
centos 8.3
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{ubuntu_latest} mount/kclient/{mount overrides/{distro/testing/{flavor/centos_latest k-testing} ms-die-on-skipped}} ms_mode/{crc} objectstore-ec/bluestore-bitmap omap_limit/10000 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/5 scrub/no standby-replay tasks/{0-check-counter workunit/suites/dbench} wsync/{no}}
master
master
master
gibba
rhel 8.4
fs/functional/{begin clusters/1a3s-mds-4c-client conf/{client mds mon osd} distro/{centos_8.stream} mount/kclient/{mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} objectstore/bluestore-bitmap overrides/{no_client_pidfile whitelist_health whitelist_wrongly_marked_down} tasks/recovery-fs}
master
master
master
gibba
ubuntu 20.04
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{centos_8.stream} mount/kclient/{mount overrides/{distro/testing/{flavor/ubuntu_latest k-testing} ms-die-on-skipped}} ms_mode/{legacy} objectstore-ec/bluestore-comp-ec-root omap_limit/10 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/1 scrub/yes standby-replay tasks/{0-check-counter workunit/suites/ffsb} wsync/{no}}
master
master
master
gibba
centos 8.3
fs/thrash/workloads/{begin clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{centos_8} mount/kclient/{mount overrides/{distro/testing/{flavor/centos_latest k-testing} ms-die-on-skipped}} msgr-failures/none objectstore-ec/bluestore-ec-root overrides/{frag session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} ranks/1 tasks/{1-thrash/mds 2-workunit/fs/trivial_sync}}
master
master
master
gibba
centos 8.3
fs/upgrade/featureful_client/old_client/{bluestore-bitmap centos_latest clusters/1-mds-2-client-micro conf/{client mds mon osd} overrides/{multimds/yes pg-warn whitelist_health whitelist_wrongly_marked_down} tasks/{0-octopus 1-client 2-upgrade 3-compat_client/pacific}}
master
master
master
gibba
rhel 8.4
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{rhel_8} mount/fuse ms_mode/{crc} objectstore-ec/bluestore-comp omap_limit/10000 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/3 scrub/no standby-replay tasks/{0-check-counter workunit/fs/norstats} wsync/{yes}}
master
master
master
gibba
centos 8.3
fs/functional/{begin clusters/1a3s-mds-4c-client conf/{client mds mon osd} distro/{centos_8} mount/kclient/{mount overrides/{distro/testing/{flavor/centos_latest k-testing} ms-die-on-skipped}} objectstore/bluestore-ec-root overrides/{no_client_pidfile whitelist_health whitelist_wrongly_marked_down} tasks/scrub}
master
master
master
gibba
ubuntu 20.04
fs/thrash/workloads/{begin clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{rhel_8} mount/kclient/{mount overrides/{distro/testing/{flavor/ubuntu_latest k-testing} ms-die-on-skipped}} msgr-failures/osd-mds-delay objectstore-ec/bluestore-bitmap overrides/{frag session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} ranks/3 tasks/{1-thrash/mon 2-workunit/suites/iozone}}
master
master
master
gibba
rhel 8.4
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{centos_8.stream} mount/kclient/{mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} ms_mode/{legacy} objectstore-ec/bluestore-ec-root omap_limit/10 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/5 scrub/yes standby-replay tasks/{0-check-counter workunit/suites/fsstress} wsync/{no}}
master
master
master
gibba
centos 8.stream
fs/thrash/workloads/{begin clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{centos_8.stream} mount/fuse msgr-failures/none objectstore-ec/bluestore-comp-ec-root overrides/{frag session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} ranks/5 tasks/{1-thrash/osd 2-workunit/suites/pjd}}
master
master
master
gibba
centos 8.3
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{centos_8} mount/kclient/{mount overrides/{distro/testing/{flavor/centos_latest k-testing} ms-die-on-skipped}} ms_mode/{secure} objectstore-ec/bluestore-bitmap omap_limit/10000 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/1 scrub/no standby-replay tasks/{0-check-counter workunit/suites/fsx} wsync/{no}}
master
master
master
gibba
ubuntu 20.04
fs/functional/{begin clusters/1a3s-mds-4c-client conf/{client mds mon osd} distro/{rhel_8} mount/kclient/{mount overrides/{distro/testing/{flavor/ubuntu_latest k-testing} ms-die-on-skipped}} objectstore/bluestore-bitmap overrides/{no_client_pidfile whitelist_health whitelist_wrongly_marked_down} tasks/sessionmap}
master
master
master
gibba
ubuntu 20.04
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{centos_8.stream} mount/kclient/{mount overrides/{distro/testing/{flavor/ubuntu_latest k-testing} ms-die-on-skipped}} ms_mode/{legacy} objectstore-ec/bluestore-comp-ec-root omap_limit/10 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/3 scrub/yes standby-replay tasks/{0-check-counter workunit/suites/fsync-tester} wsync/{no}}
"2021-11-24T05:28:07.166862+0000 mon.a (mon.0) 156 : cluster [WRN] Health check failed: Degraded data redundancy: 1 pg degraded (PG_DEGRADED)" in cluster log
master
master
master
gibba
centos 8.stream
fs/thrash/multifs/{begin clusters/1a3s-mds-2c-client conf/{client mds mon osd} distro/{centos_8.stream} mount/fuse msgr-failures/none objectstore/bluestore-bitmap overrides/{frag multifs session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} tasks/{1-thrash/mon 2-workunit/cfuse_workunit_suites_fsstress}}
master
master
master
gibba
rhel 8.4
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{rhel_8} mount/fuse ms_mode/{legacy} objectstore-ec/bluestore-comp omap_limit/10000 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/5 scrub/no standby-replay tasks/{0-check-counter workunit/fs/test_o_trunc} wsync/{yes}}
master
master
master
gibba
ubuntu 20.04
fs/functional/{begin clusters/1a3s-mds-4c-client conf/{client mds mon osd} distro/{ubuntu_latest} mount/fuse objectstore/bluestore-ec-root overrides/{no_client_pidfile whitelist_health whitelist_wrongly_marked_down} tasks/snap-schedule}
"2021-11-24T05:41:50.675131+0000 mon.a (mon.0) 1015 : cluster [WRN] Health check failed: Reduced data availability: 2 pgs peering (PG_AVAILABILITY)" in cluster log
master
master
master
gibba
centos 8.3
fs/snaps/{begin clusters/1a3s-mds-1c-client conf/{client mds mon osd} distro/{centos_8} mount/fuse objectstore-ec/bluestore-bitmap overrides/{whitelist_health whitelist_wrongly_marked_down} tasks/workunit/snaps}
master
master
master
gibba
rhel 8.4
fs/traceless/{begin clusters/fixed-2-ucephfs conf/{client mds mon osd} distro/{rhel_8} mount/fuse objectstore-ec/bluestore-bitmap overrides/{frag whitelist_health whitelist_wrongly_marked_down} tasks/cfuse_workunit_suites_blogbench traceless/50pc}
master
master
master
gibba
centos 8.stream
fs/volumes/{begin clusters/1a3s-mds-4c-client conf/{client mds mon osd} distro/{centos_8.stream} mount/fuse objectstore/bluestore-bitmap overrides/{no_client_pidfile whitelist_health whitelist_wrongly_marked_down} tasks/volumes/{overrides test/basic}}
master
master
master
gibba
rhel 8.4
fs/thrash/workloads/{begin clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{ubuntu_latest} mount/kclient/{mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} msgr-failures/osd-mds-delay objectstore-ec/bluestore-comp overrides/{frag session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} ranks/1 tasks/{1-thrash/osd 2-workunit/fs/snaps}}
reached maximum tries (90) after waiting for 540 seconds
master
master
master
gibba
rhel 8.4
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{centos_8.stream} mount/kclient/{mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} ms_mode/{secure} objectstore-ec/bluestore-ec-root omap_limit/10 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/1 scrub/yes standby-replay tasks/{0-check-counter workunit/suites/iogen} wsync/{yes}}
master
master
master
gibba
centos 8.stream
fs/upgrade/mds_upgrade_sequence/{bluestore-bitmap centos_8.stream_container_tools conf/{client mds mon osd} overrides/{pg-warn whitelist_health whitelist_wrongly_marked_down} roles tasks/{0-from/v16.2.4 1-volume/{0-create 1-ranks/1 2-allow_standby_replay/yes 3-verify} 2-client 3-upgrade-with-workload 4-verify}}
master
master
master
gibba
centos 8.3
fs/thrash/workloads/{begin clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{centos_8} mount/kclient/{mount overrides/{distro/testing/{flavor/centos_latest k-testing} ms-die-on-skipped}} msgr-failures/none objectstore-ec/bluestore-ec-root overrides/{frag session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} ranks/3 tasks/{1-thrash/mds 2-workunit/suites/ffsb}}
master
master
master
gibba
centos 8.3
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{rhel_8} mount/kclient/{mount overrides/{distro/testing/{flavor/centos_latest k-testing} ms-die-on-skipped}} ms_mode/{legacy} objectstore-ec/bluestore-bitmap omap_limit/10000 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/3 scrub/no standby-replay tasks/{0-check-counter workunit/suites/iozone} wsync/{yes}}
master
master
master
gibba
rhel 8.4
fs/functional/{begin clusters/1a3s-mds-4c-client conf/{client mds mon osd} distro/{centos_8} mount/kclient/{mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} objectstore/bluestore-bitmap overrides/{no_client_pidfile whitelist_health whitelist_wrongly_marked_down} tasks/snapshots}
master
master
master
gibba
ubuntu 20.04
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{centos_8.stream} mount/kclient/{mount overrides/{distro/testing/{flavor/ubuntu_latest k-testing} ms-die-on-skipped}} ms_mode/{legacy} objectstore-ec/bluestore-comp-ec-root omap_limit/10 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/5 scrub/yes standby-replay tasks/{0-check-counter workunit/suites/pjd} wsync/{no}}
master
master
master
gibba
ubuntu 20.04
fs/thrash/workloads/{begin clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{rhel_8} mount/kclient/{mount overrides/{distro/testing/{flavor/ubuntu_latest k-testing} ms-die-on-skipped}} msgr-failures/osd-mds-delay objectstore-ec/bluestore-bitmap overrides/{frag session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} ranks/5 tasks/{1-thrash/mon 2-workunit/suites/fsstress}}
master
master
master
gibba
centos 8.3
fs/32bits/{begin clusters/fixed-2-ucephfs conf/{client mds mon osd} distro/{centos_8} mount/fuse objectstore-ec/bluestore-bitmap overrides/{faked-ino whitelist_health whitelist_wrongly_marked_down} tasks/cfuse_workunit_suites_pjd}
master
master
master
gibba
centos 8.stream
fs/libcephfs/{begin clusters/1-mds-1-client-coloc conf/{client mds mon osd} distro/{centos_8.stream} objectstore/bluestore-bitmap overrides/{no_client_pidfile whitelist_health whitelist_wrongly_marked_down} tasks/libcephfs/{frag test}}
Command failed (workunit test libcephfs/test.sh) on gibba011 with status 141: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=292b9a3ce266acd070bcc99c4919beac2f88ae61 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 CEPH_MNT=/home/ubuntu/cephtest/mnt.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/libcephfs/test.sh'
master
master
master
gibba
rhel 8.4
fs/mixed-clients/{begin clusters/1a3s-mds-2c-client conf/{client mds mon osd} kclient-overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped} objectstore-ec/bluestore-comp-ec-root overrides/{osd-asserts whitelist_health whitelist_wrongly_marked_down} tasks/kernel_cfuse_workunits_untarbuild_blogbench}
Error reimaging machines: reached maximum tries (100) after waiting for 600 seconds
master
master
master
gibba
ubuntu 20.04
fs/multiclient/{begin clusters/1-mds-3-client conf/{client mds mon osd} distros/ubuntu_latest mount/fuse objectstore-ec/bluestore-comp-ec-root overrides/{whitelist_health whitelist_wrongly_marked_down} tasks/cephfs_misc_tests}
"2021-11-24T06:15:45.348588+0000 mon.a (mon.0) 674 : cluster [WRN] Health check failed: Degraded data redundancy: 3/1140 objects degraded (0.263%), 1 pg degraded (PG_DEGRADED)" in cluster log
master
master
master
gibba
ubuntu 20.04
fs/multifs/{begin clusters/1a3s-mds-2c-client conf/{client mds mon osd} distro/{rhel_8} mount/kclient/{mount overrides/{distro/testing/{flavor/ubuntu_latest k-testing} ms-die-on-skipped}} objectstore-ec/bluestore-comp-ec-root overrides/{mon-debug whitelist_health whitelist_wrongly_marked_down} tasks/multifs-auth}
master
master
master
gibba
rhel 8.4
fs/permission/{begin clusters/fixed-2-ucephfs conf/{client mds mon osd} distro/{rhel_8} mount/fuse objectstore-ec/bluestore-bitmap overrides/{whitelist_health whitelist_wrongly_marked_down} tasks/cfuse_workunit_suites_pjd}
master
master
master
gibba
rhel 8.4
fs/verify/{begin clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{rhel_8} mount/kclient/{k-testing mount ms-die-on-skipped} objectstore-ec/bluestore-comp-ec-root overrides/{mon-debug session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/1 tasks/fsstress validater/valgrind}
master
master
master
gibba
centos 8.stream
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{centos_8.stream} mount/fuse ms_mode/{crc} objectstore-ec/bluestore-comp omap_limit/10000 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/1 scrub/no standby-replay tasks/{0-check-counter workunit/direct_io} wsync/{no}}
master
master
master
gibba
centos 8.3
fs/functional/{begin clusters/1a3s-mds-4c-client conf/{client mds mon osd} distro/{centos_8.stream} mount/kclient/{mount overrides/{distro/testing/{flavor/centos_latest k-testing} ms-die-on-skipped}} objectstore/bluestore-ec-root overrides/{no_client_pidfile whitelist_health whitelist_wrongly_marked_down} tasks/strays}
Test failure: test_dirfrag_limit (tasks.cephfs.test_strays.TestStrays)
master
master
master
gibba
centos 8.3
fs/thrash/workloads/{begin clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{centos_8} mount/fuse msgr-failures/none objectstore-ec/bluestore-comp-ec-root overrides/{frag session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} ranks/1 tasks/{1-thrash/osd 2-workunit/fs/trivial_sync}}
master
master
master
gibba
rhel 8.4
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{centos_8.stream} mount/kclient/{mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} ms_mode/{crc} objectstore-ec/bluestore-ec-root omap_limit/10 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/3 scrub/yes standby-replay tasks/{0-check-counter workunit/fs/misc} wsync/{no}}
master
master
master
gibba
rhel 8.4
fs/thrash/workloads/{begin clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{ubuntu_latest} mount/kclient/{mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} msgr-failures/osd-mds-delay objectstore-ec/bluestore-comp overrides/{frag session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} ranks/3 tasks/{1-thrash/mds 2-workunit/suites/iozone}}
master
master
master
gibba
centos 8.3
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{rhel_8} mount/kclient/{mount overrides/{distro/testing/{flavor/centos_latest k-testing} ms-die-on-skipped}} ms_mode/{legacy} objectstore-ec/bluestore-bitmap omap_limit/10000 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/5 scrub/no standby-replay tasks/{0-check-counter workunit/kernel_untar_build} wsync/{yes}}
master
master
master
gibba
ubuntu 20.04
fs/functional/{begin clusters/1a3s-mds-4c-client conf/{client mds mon osd} distro/{ubuntu_latest} mount/kclient/{mount overrides/{distro/testing/{flavor/ubuntu_latest k-testing} ms-die-on-skipped}} objectstore/bluestore-bitmap overrides/{no_client_pidfile whitelist_health whitelist_wrongly_marked_down} tasks/test_journal_migration}
"2021-11-24T06:26:15.325277+0000 mon.a (mon.0) 378 : cluster [WRN] Health check failed: Degraded data redundancy: 1/21644 objects degraded (0.005%), 1 pg degraded (PG_DEGRADED)" in cluster log
master
master
master
gibba
centos 8.3
fs/upgrade/featureful_client/old_client/{bluestore-bitmap centos_latest clusters/1-mds-2-client-micro conf/{client mds mon osd} overrides/{multimds/yes pg-warn whitelist_health whitelist_wrongly_marked_down} tasks/{0-octopus 1-client 2-upgrade 3-compat_client/no}}
master
master
master
gibba
ubuntu 20.04
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{rhel_8} mount/kclient/{mount overrides/{distro/testing/{flavor/ubuntu_latest k-testing} ms-die-on-skipped}} ms_mode/{crc} objectstore-ec/bluestore-comp-ec-root omap_limit/10 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/1 scrub/yes standby-replay tasks/{0-check-counter workunit/suites/blogbench} wsync/{yes}}
"2021-11-24T06:43:16.336041+0000 mon.a (mon.0) 157 : cluster [WRN] Health check failed: Degraded data redundancy: 2/4 objects degraded (50.000%), 1 pg degraded (PG_DEGRADED)" in cluster log
master
master
master
gibba
centos 8.3
fs/thrash/workloads/{begin clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{centos_8.stream} mount/kclient/{mount overrides/{distro/testing/{flavor/centos_latest k-testing} ms-die-on-skipped}} msgr-failures/none objectstore-ec/bluestore-ec-root overrides/{frag session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} ranks/5 tasks/{1-thrash/mon 2-workunit/suites/pjd}}
master
master
master
gibba
ubuntu 20.04
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{ubuntu_latest} mount/fuse ms_mode/{crc} objectstore-ec/bluestore-comp omap_limit/10000 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/3 scrub/no standby-replay tasks/{0-check-counter workunit/suites/dbench} wsync/{yes}}
master
master
master
gibba
rhel 8.4
fs/functional/{begin clusters/1a3s-mds-4c-client conf/{client mds mon osd} distro/{rhel_8} mount/fuse objectstore/bluestore-ec-root overrides/{no_client_pidfile whitelist_health whitelist_wrongly_marked_down} tasks/truncate_delay}
master
master
master
gibba
rhel 8.4
fs/thrash/multifs/{begin clusters/1a3s-mds-2c-client conf/{client mds mon osd} distro/{centos_8} mount/kclient/{mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} msgr-failures/osd-mds-delay objectstore/bluestore-bitmap overrides/{frag multifs session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} tasks/{1-thrash/mds 2-workunit/cfuse_workunit_suites_pjd}}
master
master
master
gibba
rhel 8.4
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{ubuntu_latest} mount/kclient/{mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} ms_mode/{legacy} objectstore-ec/bluestore-ec-root omap_limit/10 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/5 scrub/yes standby-replay tasks/{0-check-counter workunit/suites/ffsb} wsync/{no}}
master
master
master
gibba
ubuntu 20.04
fs/thrash/workloads/{begin clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{rhel_8} mount/kclient/{mount overrides/{distro/testing/{flavor/ubuntu_latest k-testing} ms-die-on-skipped}} msgr-failures/osd-mds-delay objectstore-ec/bluestore-bitmap overrides/{frag session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} ranks/1 tasks/{1-thrash/mon 2-workunit/fs/snaps}}
master
master
master
gibba
centos 8.3
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{centos_8.stream} mount/kclient/{mount overrides/{distro/testing/{flavor/centos_latest k-testing} ms-die-on-skipped}} ms_mode/{secure} objectstore-ec/bluestore-bitmap omap_limit/10000 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/1 scrub/no standby-replay tasks/{0-check-counter workunit/fs/norstats} wsync/{yes}}
master
master
master
gibba
rhel 8.4
fs/functional/{begin clusters/1a3s-mds-4c-client conf/{client mds mon osd} distro/{centos_8} mount/kclient/{mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} objectstore/bluestore-bitmap overrides/{no_client_pidfile whitelist_health whitelist_wrongly_marked_down} tasks/workunit/quota}
master
master
master
gibba
ubuntu 20.04
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{centos_8} mount/kclient/{mount overrides/{distro/testing/{flavor/ubuntu_latest k-testing} ms-die-on-skipped}} ms_mode/{secure} objectstore-ec/bluestore-comp-ec-root omap_limit/10 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/3 scrub/yes standby-replay tasks/{0-check-counter workunit/suites/fsstress} wsync/{yes}}
master
master
master
gibba
centos 8.3
fs/thrash/workloads/{begin clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{centos_8} mount/fuse msgr-failures/none objectstore-ec/bluestore-comp-ec-root overrides/{frag session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} ranks/3 tasks/{1-thrash/osd 2-workunit/suites/ffsb}}
master
master
master
gibba
centos 8.stream
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{centos_8.stream} mount/fuse ms_mode/{legacy} objectstore-ec/bluestore-comp omap_limit/10000 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/5 scrub/no standby-replay tasks/{0-check-counter workunit/suites/fsx} wsync/{yes}}
master
master
master
gibba
centos 8.3
fs/functional/{begin clusters/1a3s-mds-4c-client conf/{client mds mon osd} distro/{centos_8.stream} mount/kclient/{mount overrides/{distro/testing/{flavor/centos_latest k-testing} ms-die-on-skipped}} objectstore/bluestore-ec-root overrides/{no_client_pidfile whitelist_health whitelist_wrongly_marked_down} tasks/acls}
master
master
master
gibba
centos 8.stream
fs/upgrade/mds_upgrade_sequence/{bluestore-bitmap centos_8.stream_container_tools conf/{client mds mon osd} overrides/{pg-warn whitelist_health whitelist_wrongly_marked_down} roles tasks/{0-from/v16.2.4 1-volume/{0-create 1-ranks/1 2-allow_standby_replay/no 3-verify} 2-client 3-upgrade-with-workload 4-verify}}
master
master
master
gibba
rhel 8.4
fs/thrash/workloads/{begin clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{centos_8.stream} mount/kclient/{mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} msgr-failures/osd-mds-delay objectstore-ec/bluestore-comp overrides/{frag session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} ranks/5 tasks/{1-thrash/mds 2-workunit/suites/fsstress}}
master
master
master
gibba
rhel 8.4
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{rhel_8} mount/kclient/{mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} ms_mode/{legacy} objectstore-ec/bluestore-ec-root omap_limit/10 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/1 scrub/yes standby-replay tasks/{0-check-counter workunit/suites/fsync-tester} wsync/{no}}
master
master
master
gibba
centos 8.3
fs/thrash/workloads/{begin clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{rhel_8} mount/kclient/{mount overrides/{distro/testing/{flavor/centos_latest k-testing} ms-die-on-skipped}} msgr-failures/none objectstore-ec/bluestore-ec-root overrides/{frag session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} ranks/1 tasks/{1-thrash/mon 2-workunit/fs/trivial_sync}}
master
master
master
gibba
centos 8.3
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{centos_8} mount/kclient/{mount overrides/{distro/testing/{flavor/centos_latest k-testing} ms-die-on-skipped}} ms_mode/{secure} objectstore-ec/bluestore-bitmap omap_limit/10000 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/3 scrub/no standby-replay tasks/{0-check-counter workunit/fs/test_o_trunc} wsync/{no}}
master
master
master
gibba
ubuntu 20.04
fs/functional/{begin clusters/1a3s-mds-4c-client conf/{client mds mon osd} distro/{ubuntu_latest} mount/kclient/{mount overrides/{distro/testing/{flavor/ubuntu_latest k-testing} ms-die-on-skipped}} objectstore/bluestore-bitmap overrides/{no_client_pidfile whitelist_health whitelist_wrongly_marked_down} tasks/admin}
"2021-11-24T08:06:03.693775+0000 mon.a (mon.0) 1665 : cluster [WRN] Health check failed: Reduced data availability: 2 pgs inactive (PG_AVAILABILITY)" in cluster log
master
master
master
gibba
centos 8.stream
fs/32bits/{begin clusters/fixed-2-ucephfs conf/{client mds mon osd} distro/{centos_8.stream} mount/fuse objectstore-ec/bluestore-comp-ec-root overrides/{faked-ino whitelist_health whitelist_wrongly_marked_down} tasks/cfuse_workunit_suites_fsstress}
master
master
master
gibba
centos 8.stream
fs/permission/{begin clusters/fixed-2-ucephfs conf/{client mds mon osd} distro/{centos_8.stream} mount/fuse objectstore-ec/bluestore-comp-ec-root overrides/{whitelist_health whitelist_wrongly_marked_down} tasks/cfuse_workunit_misc}
master
master
master
gibba
ubuntu 20.04
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{centos_8.stream} mount/kclient/{mount overrides/{distro/testing/{flavor/ubuntu_latest k-testing} ms-die-on-skipped}} ms_mode/{crc} objectstore-ec/bluestore-comp-ec-root omap_limit/10 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/5 scrub/yes standby-replay tasks/{0-check-counter workunit/suites/iogen} wsync/{yes}}
"2021-11-24T08:08:32.942559+0000 mon.c (mon.0) 266 : cluster [WRN] Health check failed: 1/3 mons down, quorum c,b (MON_DOWN)" in cluster log
master
master
master
gibba
ubuntu 20.04
fs/thrash/workloads/{begin clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{rhel_8} mount/kclient/{mount overrides/{distro/testing/{flavor/ubuntu_latest k-testing} ms-die-on-skipped}} msgr-failures/osd-mds-delay objectstore-ec/bluestore-bitmap overrides/{frag session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} ranks/3 tasks/{1-thrash/osd 2-workunit/suites/iozone}}
master
master
master
gibba
rhel 8.4
fs/snaps/{begin clusters/1a3s-mds-1c-client conf/{client mds mon osd} distro/{centos_8} mount/kclient/{mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} objectstore-ec/bluestore-comp-ec-root overrides/{whitelist_health whitelist_wrongly_marked_down} tasks/workunit/snaps}
master
master
master
gibba
ubuntu 20.04
fs/traceless/{begin clusters/fixed-2-ucephfs conf/{client mds mon osd} distro/{ubuntu_latest} mount/fuse objectstore-ec/bluestore-comp-ec-root overrides/{frag whitelist_health whitelist_wrongly_marked_down} tasks/cfuse_workunit_suites_dbench traceless/50pc}
master
master
master
gibba
rhel 8.4
fs/volumes/{begin clusters/1a3s-mds-4c-client conf/{client mds mon osd} distro/{centos_8.stream} mount/kclient/{mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} objectstore/bluestore-bitmap overrides/{no_client_pidfile whitelist_health whitelist_wrongly_marked_down} tasks/volumes/{overrides test/clone}}
master
master
master
gibba
ubuntu 20.04
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{ubuntu_latest} mount/fuse ms_mode/{crc} objectstore-ec/bluestore-comp omap_limit/10000 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/1 scrub/no standby-replay tasks/{0-check-counter workunit/suites/iozone} wsync/{no}}
master
master
master
gibba
rhel 8.4
fs/functional/{begin clusters/1a3s-mds-4c-client conf/{client mds mon osd} distro/{rhel_8} mount/fuse objectstore/bluestore-ec-root overrides/{no_client_pidfile whitelist_health whitelist_wrongly_marked_down} tasks/alternate-pool}
master
master
master
gibba
rhel 8.4
fs/thrash/workloads/{begin clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{rhel_8} mount/fuse msgr-failures/none objectstore-ec/bluestore-comp-ec-root overrides/{frag session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} ranks/5 tasks/{1-thrash/mds 2-workunit/suites/pjd}}
master
master
master
gibba
rhel 8.4
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{rhel_8} mount/kclient/{mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} ms_mode/{crc} objectstore-ec/bluestore-ec-root omap_limit/10 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/3 scrub/yes standby-replay tasks/{0-check-counter workunit/suites/pjd} wsync/{no}}
master
master
master
gibba
centos 8.3
fs/mixed-clients/{begin clusters/1a3s-mds-2c-client conf/{client mds mon osd} kclient-overrides/{distro/testing/{flavor/centos_latest k-testing} ms-die-on-skipped} objectstore-ec/bluestore-comp overrides/{osd-asserts whitelist_health whitelist_wrongly_marked_down} tasks/kernel_cfuse_workunits_dbench_iozone}
master
master
master
gibba
ubuntu 20.04
fs/multiclient/{begin clusters/1-mds-2-client conf/{client mds mon osd} distros/ubuntu_latest mount/fuse objectstore-ec/bluestore-comp overrides/{whitelist_health whitelist_wrongly_marked_down} tasks/ior-shared-file}
master
master
master
gibba
rhel 8.4
fs/multifs/{begin clusters/1a3s-mds-2c-client conf/{client mds mon osd} distro/{rhel_8} mount/fuse objectstore-ec/bluestore-comp overrides/{mon-debug whitelist_health whitelist_wrongly_marked_down} tasks/failover}
master
master
master
gibba
centos 8.3
fs/thrash/multifs/{begin clusters/1a3s-mds-2c-client conf/{client mds mon osd} distro/{ubuntu_latest} mount/kclient/{mount overrides/{distro/testing/{flavor/centos_latest k-testing} ms-die-on-skipped}} msgr-failures/none objectstore/bluestore-bitmap overrides/{frag multifs session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} tasks/{1-thrash/mon 2-workunit/cfuse_workunit_trivial_sync}}
master
master
master
gibba
centos 8.stream
fs/upgrade/mds_upgrade_sequence/{bluestore-bitmap centos_8.stream_container_tools conf/{client mds mon osd} overrides/{pg-warn whitelist_health whitelist_wrongly_marked_down} roles tasks/{0-from/pacific 1-volume/{0-create 1-ranks/2 2-allow_standby_replay/yes 3-verify} 2-client 3-upgrade-with-workload 4-verify}}
master
master
master
gibba
ubuntu 20.04
fs/verify/{begin clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{ubuntu/{latest overrides}} mount/fuse objectstore-ec/bluestore-comp overrides/{mon-debug session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/3 tasks/dbench validater/lockdep}
master
master
master
gibba
centos 8.3
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{centos_8.stream} mount/kclient/{mount overrides/{distro/testing/{flavor/centos_latest k-testing} ms-die-on-skipped}} ms_mode/{legacy} objectstore-ec/bluestore-ec-root omap_limit/10000 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/5 scrub/no standby-replay tasks/{0-check-counter workunit/direct_io} wsync/{no}}
master
master
master
gibba
rhel 8.4
fs/functional/{begin clusters/1a3s-mds-4c-client conf/{client mds mon osd} distro/{centos_8.stream} mount/kclient/{mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} objectstore/bluestore-bitmap overrides/{no_client_pidfile whitelist_health whitelist_wrongly_marked_down} tasks/asok_dump_tree}
master
master
master
gibba
ubuntu 20.04
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{centos_8} mount/kclient/{mount overrides/{distro/testing/{flavor/ubuntu_latest k-testing} ms-die-on-skipped}} ms_mode/{secure} objectstore-ec/bluestore-bitmap omap_limit/10 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/1 scrub/yes standby-replay tasks/{0-check-counter workunit/fs/misc} wsync/{no}}
master
master
master
gibba
rhel 8.4
fs/thrash/workloads/{begin clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{rhel_8} mount/kclient/{mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} msgr-failures/osd-mds-delay objectstore-ec/bluestore-comp overrides/{frag session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} ranks/1 tasks/{1-thrash/mds 2-workunit/fs/snaps}}
master
master
master
gibba
rhel 8.4
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{rhel_8} mount/fuse ms_mode/{crc} objectstore-ec/bluestore-comp-ec-root omap_limit/10000 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/3 scrub/no standby-replay tasks/{0-check-counter workunit/kernel_untar_build} wsync/{yes}}
master
master
master
gibba
centos 8.3
fs/functional/{begin clusters/1a3s-mds-4c-client conf/{client mds mon osd} distro/{ubuntu_latest} mount/kclient/{mount overrides/{distro/testing/{flavor/centos_latest k-testing} ms-die-on-skipped}} objectstore/bluestore-ec-root overrides/{no_client_pidfile whitelist_health whitelist_wrongly_marked_down} tasks/auto-repair}
master
master
master
gibba
centos 8.3
fs/thrash/workloads/{begin clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{centos_8} mount/kclient/{mount overrides/{distro/testing/{flavor/centos_latest k-testing} ms-die-on-skipped}} msgr-failures/none objectstore-ec/bluestore-ec-root overrides/{frag session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} ranks/3 tasks/{1-thrash/mon 2-workunit/suites/ffsb}}
master
master
master
gibba
rhel 8.4
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{centos_8.stream} mount/kclient/{mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} ms_mode/{crc} objectstore-ec/bluestore-comp omap_limit/10 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/5 scrub/yes standby-replay tasks/{0-check-counter workunit/suites/blogbench} wsync/{no}}
hit max job timeout
master
master
master
gibba
ubuntu 20.04
fs/thrash/workloads/{begin clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{centos_8.stream} mount/kclient/{mount overrides/{distro/testing/{flavor/ubuntu_latest k-testing} ms-die-on-skipped}} msgr-failures/osd-mds-delay objectstore-ec/bluestore-bitmap overrides/{frag session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} ranks/5 tasks/{1-thrash/osd 2-workunit/suites/fsstress}}
master
master
master
gibba
centos 8.3
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{rhel_8} mount/kclient/{mount overrides/{distro/testing/{flavor/centos_latest k-testing} ms-die-on-skipped}} ms_mode/{crc} objectstore-ec/bluestore-ec-root omap_limit/10000 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/1 scrub/no standby-replay tasks/{0-check-counter workunit/suites/dbench} wsync/{yes}}
master
master
master
gibba
ubuntu 20.04
fs/functional/{begin clusters/1a3s-mds-4c-client conf/{client mds mon osd} distro/{centos_8} mount/kclient/{mount overrides/{distro/testing/{flavor/ubuntu_latest k-testing} ms-die-on-skipped}} objectstore/bluestore-bitmap overrides/{no_client_pidfile whitelist_health whitelist_wrongly_marked_down} tasks/backtrace}
master
master
master
gibba
ubuntu 20.04
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{rhel_8} mount/kclient/{mount overrides/{distro/testing/{flavor/ubuntu_latest k-testing} ms-die-on-skipped}} ms_mode/{secure} objectstore-ec/bluestore-bitmap omap_limit/10 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/3 scrub/yes standby-replay tasks/{0-check-counter workunit/suites/ffsb} wsync/{yes}}
Command failed (workunit test suites/ffsb.sh) on gibba017 with status 135: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=292b9a3ce266acd070bcc99c4919beac2f88ae61 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 CEPH_MNT=/home/ubuntu/cephtest/mnt.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/suites/ffsb.sh'
master
master
master
gibba
rhel 8.4
fs/thrash/workloads/{begin clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{rhel_8} mount/fuse msgr-failures/none objectstore-ec/bluestore-comp-ec-root overrides/{frag session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} ranks/1 tasks/{1-thrash/mds 2-workunit/fs/trivial_sync}}
master
master
master
gibba
centos 8.3
fs/upgrade/featureful_client/upgraded_client/{bluestore-bitmap centos_latest clusters/1-mds-2-client-micro conf/{client mds mon osd} overrides/{multimds/yes pg-warn whitelist_health whitelist_wrongly_marked_down} tasks/{0-octopus 1-client 2-upgrade 3-client-upgrade 4-compat_client 5-client-sanity}}
master
master
master
gibba
ubuntu 20.04
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{ubuntu_latest} mount/fuse ms_mode/{crc} objectstore-ec/bluestore-comp-ec-root omap_limit/10000 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/5 scrub/no standby-replay tasks/{0-check-counter workunit/fs/norstats} wsync/{no}}
master
master
master
gibba
centos 8.3
fs/functional/{begin clusters/1a3s-mds-4c-client conf/{client mds mon osd} distro/{centos_8} mount/fuse objectstore/bluestore-ec-root overrides/{no_client_pidfile whitelist_health whitelist_wrongly_marked_down} tasks/cap-flush}
master
master
master
gibba
rhel 8.4
fs/thrash/workloads/{begin clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{centos_8} mount/kclient/{mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} msgr-failures/osd-mds-delay objectstore-ec/bluestore-comp overrides/{frag session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} ranks/3 tasks/{1-thrash/mon 2-workunit/suites/iozone}}
master
master
master
gibba
ubuntu 20.04
fs/32bits/{begin clusters/fixed-2-ucephfs conf/{client mds mon osd} distro/{ubuntu_latest} mount/fuse objectstore-ec/bluestore-comp overrides/{faked-ino whitelist_health whitelist_wrongly_marked_down} tasks/cfuse_workunit_suites_pjd}
master
master
master
gibba
centos 8.stream
fs/libcephfs/{begin clusters/1-mds-1-client-coloc conf/{client mds mon osd} distro/{centos_8.stream} objectstore/bluestore-bitmap overrides/{no_client_pidfile whitelist_health whitelist_wrongly_marked_down} tasks/libcephfs_python}
master
master
master
gibba
rhel 8.4
fs/permission/{begin clusters/fixed-2-ucephfs conf/{client mds mon osd} distro/{rhel_8} mount/fuse objectstore-ec/bluestore-comp overrides/{whitelist_health whitelist_wrongly_marked_down} tasks/cfuse_workunit_suites_pjd}
master
master
master
gibba
rhel 8.4
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{ubuntu_latest} mount/kclient/{mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} ms_mode/{legacy} objectstore-ec/bluestore-comp omap_limit/10 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/1 scrub/yes standby-replay tasks/{0-check-counter workunit/suites/fsstress} wsync/{no}}
master
master
master
gibba
centos 8.3
fs/thrash/workloads/{begin clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{centos_8.stream} mount/kclient/{mount overrides/{distro/testing/{flavor/centos_latest k-testing} ms-die-on-skipped}} msgr-failures/none objectstore-ec/bluestore-ec-root overrides/{frag session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} ranks/5 tasks/{1-thrash/osd 2-workunit/suites/pjd}}
master
master
master
gibba
centos 8.3
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{ubuntu_latest} mount/kclient/{mount overrides/{distro/testing/{flavor/centos_latest k-testing} ms-die-on-skipped}} ms_mode/{secure} objectstore-ec/bluestore-ec-root omap_limit/10000 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/3 scrub/no standby-replay tasks/{0-check-counter workunit/suites/fsx} wsync/{no}}
master
master
master
gibba
rhel 8.4
fs/functional/{begin clusters/1a3s-mds-4c-client conf/{client mds mon osd} distro/{rhel_8} mount/kclient/{mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} objectstore/bluestore-bitmap overrides/{no_client_pidfile whitelist_health whitelist_wrongly_marked_down} tasks/client-limits}
master
master
master
gibba
ubuntu 20.04
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{centos_8} mount/kclient/{mount overrides/{distro/testing/{flavor/ubuntu_latest k-testing} ms-die-on-skipped}} ms_mode/{crc} objectstore-ec/bluestore-bitmap omap_limit/10 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/5 scrub/yes standby-replay tasks/{0-check-counter workunit/suites/fsync-tester} wsync/{no}}
master
master
master
gibba
ubuntu 20.04
fs/thrash/multifs/{begin clusters/1a3s-mds-2c-client conf/{client mds mon osd} distro/{centos_8} mount/kclient/{mount overrides/{distro/testing/{flavor/ubuntu_latest k-testing} ms-die-on-skipped}} msgr-failures/osd-mds-delay objectstore/bluestore-bitmap overrides/{frag multifs session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} tasks/{1-thrash/mds 2-workunit/ffsb}}
master
master
master
gibba
rhel 8.4
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{rhel_8} mount/fuse ms_mode/{crc} objectstore-ec/bluestore-comp-ec-root omap_limit/10000 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/1 scrub/no standby-replay tasks/{0-check-counter workunit/fs/test_o_trunc} wsync/{no}}
master
master
master
gibba
ubuntu 20.04
fs/thrash/workloads/{begin clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{ubuntu_latest} mount/kclient/{mount overrides/{distro/testing/{flavor/ubuntu_latest k-testing} ms-die-on-skipped}} msgr-failures/osd-mds-delay objectstore-ec/bluestore-bitmap overrides/{frag session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} ranks/1 tasks/{1-thrash/osd 2-workunit/fs/snaps}}
master
master
master
gibba
centos 8.3
fs/functional/{begin clusters/1a3s-mds-4c-client conf/{client mds mon osd} distro/{centos_8.stream} mount/kclient/{mount overrides/{distro/testing/{flavor/centos_latest k-testing} ms-die-on-skipped}} objectstore/bluestore-ec-root overrides/{no_client_pidfile whitelist_health whitelist_wrongly_marked_down} tasks/client-readahead}
Test failure: test_flush (tasks.cephfs.test_readahead.TestReadahead)
master
master
master
gibba
rhel 8.4
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{ubuntu_latest} mount/kclient/{mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} ms_mode/{legacy} objectstore-ec/bluestore-comp omap_limit/10 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/3 scrub/yes standby-replay tasks/{0-check-counter workunit/suites/iogen} wsync/{no}}
master
master
master
gibba
centos 8.stream
fs/upgrade/mds_upgrade_sequence/{bluestore-bitmap centos_8.stream_container_tools conf/{client mds mon osd} overrides/{pg-warn whitelist_health whitelist_wrongly_marked_down} roles tasks/{0-from/v16.2.4 1-volume/{0-create 1-ranks/2 2-allow_standby_replay/no 3-verify} 2-client 3-upgrade-with-workload 4-verify}}
master
master
master
gibba
ubuntu 20.04
fs/thrash/workloads/{begin clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{ubuntu_latest} mount/fuse msgr-failures/none objectstore-ec/bluestore-comp-ec-root overrides/{frag session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} ranks/3 tasks/{1-thrash/mds 2-workunit/suites/ffsb}}
master
master
master
gibba
centos 8.3
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{ubuntu_latest} mount/kclient/{mount overrides/{distro/testing/{flavor/centos_latest k-testing} ms-die-on-skipped}} ms_mode/{crc} objectstore-ec/bluestore-ec-root omap_limit/10000 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/5 scrub/no standby-replay tasks/{0-check-counter workunit/suites/iozone} wsync/{no}}
master
master
master
gibba
ubuntu 20.04
fs/functional/{begin clusters/1a3s-mds-4c-client conf/{client mds mon osd} distro/{centos_8} mount/kclient/{mount overrides/{distro/testing/{flavor/ubuntu_latest k-testing} ms-die-on-skipped}} objectstore/bluestore-bitmap overrides/{no_client_pidfile whitelist_health whitelist_wrongly_marked_down} tasks/client-recovery}
master
master
master
gibba
centos 8.3
fs/snaps/{begin clusters/1a3s-mds-1c-client conf/{client mds mon osd} distro/{ubuntu_latest} mount/kclient/{mount overrides/{distro/testing/{flavor/centos_latest k-testing} ms-die-on-skipped}} objectstore-ec/bluestore-comp overrides/{whitelist_health whitelist_wrongly_marked_down} tasks/workunit/snaps}
master
master
master
gibba
centos 8.3
fs/traceless/{begin clusters/fixed-2-ucephfs conf/{client mds mon osd} distro/{centos_8} mount/fuse objectstore-ec/bluestore-comp overrides/{frag whitelist_health whitelist_wrongly_marked_down} tasks/cfuse_workunit_suites_ffsb traceless/50pc}
master
master
master
gibba
centos 8.3
fs/volumes/{begin clusters/1a3s-mds-4c-client conf/{client mds mon osd} distro/{ubuntu_latest} mount/kclient/{mount overrides/{distro/testing/{flavor/centos_latest k-testing} ms-die-on-skipped}} objectstore/bluestore-bitmap overrides/{no_client_pidfile whitelist_health whitelist_wrongly_marked_down} tasks/volumes/{overrides test/misc}}
"2021-11-24T10:15:40.641016+0000 mon.b (mon.0) 248 : cluster [WRN] Health check failed: Reduced data availability: 1 pg peering (PG_AVAILABILITY)" in cluster log
master
master
master
gibba
ubuntu 20.04
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{ubuntu_latest} mount/kclient/{mount overrides/{distro/testing/{flavor/ubuntu_latest k-testing} ms-die-on-skipped}} ms_mode/{legacy} objectstore-ec/bluestore-bitmap omap_limit/10 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/1 scrub/yes standby-replay tasks/{0-check-counter workunit/suites/pjd} wsync/{yes}}
master
master
master
gibba
rhel 8.4
fs/thrash/workloads/{begin clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{centos_8} mount/kclient/{mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} msgr-failures/osd-mds-delay objectstore-ec/bluestore-comp overrides/{frag session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} ranks/5 tasks/{1-thrash/mon 2-workunit/suites/fsstress}}
master
master
master
gibba
ubuntu 20.04
fs/mixed-clients/{begin clusters/1a3s-mds-2c-client conf/{client mds mon osd} kclient-overrides/{distro/testing/{flavor/ubuntu_latest k-testing} ms-die-on-skipped} objectstore-ec/bluestore-ec-root overrides/{osd-asserts whitelist_health whitelist_wrongly_marked_down} tasks/kernel_cfuse_workunits_untarbuild_blogbench}
master
master
master
gibba
ubuntu 20.04
fs/multiclient/{begin clusters/1-mds-3-client conf/{client mds mon osd} distros/ubuntu_latest mount/fuse objectstore-ec/bluestore-ec-root overrides/{whitelist_health whitelist_wrongly_marked_down} tasks/mdtest}
master
master
master
gibba
rhel 8.4
fs/multifs/{begin clusters/1a3s-mds-2c-client conf/{client mds mon osd} distro/{centos_8.stream} mount/kclient/{mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} objectstore-ec/bluestore-ec-root overrides/{mon-debug whitelist_health whitelist_wrongly_marked_down} tasks/multifs-auth}
master
master
master
gibba
centos 8.3
fs/verify/{begin clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{centos_8} mount/kclient/{k-testing mount ms-die-on-skipped} objectstore-ec/bluestore-ec-root overrides/{mon-debug session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/5 tasks/fsstress validater/valgrind}
master
master
master
gibba
ubuntu 20.04
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{ubuntu_latest} mount/fuse ms_mode/{legacy} objectstore-ec/bluestore-comp-ec-root omap_limit/10000 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/3 scrub/no standby-replay tasks/{0-check-counter workunit/direct_io} wsync/{yes}}
master
master
master
gibba
centos 8.3
fs/thrash/workloads/{begin clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{ubuntu_latest} mount/kclient/{mount overrides/{distro/testing/{flavor/centos_latest k-testing} ms-die-on-skipped}} msgr-failures/none objectstore-ec/bluestore-ec-root overrides/{frag session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} ranks/1 tasks/{1-thrash/osd 2-workunit/fs/trivial_sync}}
master
master
master
gibba
centos 8.stream
fs/functional/{begin clusters/1a3s-mds-4c-client conf/{client mds mon osd} distro/{centos_8.stream} mount/fuse objectstore/bluestore-ec-root overrides/{no_client_pidfile whitelist_health whitelist_wrongly_marked_down} tasks/damage}
master
master
master
gibba
rhel 8.4
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{centos_8.stream} mount/kclient/{mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} ms_mode/{crc} objectstore-ec/bluestore-comp omap_limit/10 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/5 scrub/yes standby-replay tasks/{0-check-counter workunit/fs/misc} wsync/{yes}}
master
master
master
gibba
ubuntu 20.04
fs/thrash/workloads/{begin clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{centos_8} mount/kclient/{mount overrides/{distro/testing/{flavor/ubuntu_latest k-testing} ms-die-on-skipped}} msgr-failures/osd-mds-delay objectstore-ec/bluestore-bitmap overrides/{frag session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} ranks/3 tasks/{1-thrash/mds 2-workunit/suites/iozone}}
master
master
master
gibba
centos 8.3
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{rhel_8} mount/kclient/{mount overrides/{distro/testing/{flavor/centos_latest k-testing} ms-die-on-skipped}} ms_mode/{secure} objectstore-ec/bluestore-ec-root omap_limit/10000 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/1 scrub/no standby-replay tasks/{0-check-counter workunit/kernel_untar_build} wsync/{yes}}
master
master
master
gibba
rhel 8.4
fs/functional/{begin clusters/1a3s-mds-4c-client conf/{client mds mon osd} distro/{centos_8} mount/kclient/{mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} objectstore/bluestore-bitmap overrides/{no_client_pidfile whitelist_health whitelist_wrongly_marked_down} tasks/data-scan}
master
master
master
gibba
centos 8.3
fs/upgrade/featureful_client/old_client/{bluestore-bitmap centos_latest clusters/1-mds-2-client-micro conf/{client mds mon osd} overrides/{multimds/no pg-warn whitelist_health whitelist_wrongly_marked_down} tasks/{0-octopus 1-client 2-upgrade 3-compat_client/pacific}}
master
master
master
gibba
ubuntu 20.04
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{centos_8.stream} mount/kclient/{mount overrides/{distro/testing/{flavor/ubuntu_latest k-testing} ms-die-on-skipped}} ms_mode/{secure} objectstore-ec/bluestore-bitmap omap_limit/10 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/3 scrub/yes standby-replay tasks/{0-check-counter workunit/suites/blogbench} wsync/{yes}}
master
master
master
gibba
centos 8.stream
fs/thrash/workloads/{begin clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{centos_8.stream} mount/fuse msgr-failures/none objectstore-ec/bluestore-comp-ec-root overrides/{frag session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} ranks/5 tasks/{1-thrash/mon 2-workunit/suites/pjd}}
master
master
master
gibba
centos 8.3
fs/32bits/{begin clusters/fixed-2-ucephfs conf/{client mds mon osd} distro/{centos_8} mount/fuse objectstore-ec/bluestore-ec-root overrides/{faked-ino whitelist_health whitelist_wrongly_marked_down} tasks/cfuse_workunit_suites_fsstress}
master
master
master
gibba
ubuntu 20.04
fs/permission/{begin clusters/fixed-2-ucephfs conf/{client mds mon osd} distro/{ubuntu_latest} mount/fuse objectstore-ec/bluestore-ec-root overrides/{whitelist_health whitelist_wrongly_marked_down} tasks/cfuse_workunit_misc}
master
master
master
gibba
rhel 8.4
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{rhel_8} mount/fuse ms_mode/{secure} objectstore-ec/bluestore-comp-ec-root omap_limit/10000 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/5 scrub/no standby-replay tasks/{0-check-counter workunit/suites/dbench} wsync/{no}}
master
master
master
gibba
centos 8.3
fs/thrash/multifs/{begin clusters/1a3s-mds-2c-client conf/{client mds mon osd} distro/{centos_8} mount/fuse msgr-failures/none objectstore/bluestore-bitmap overrides/{frag multifs session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} tasks/{1-thrash/mon 2-workunit/iozone}}
master
master
master
gibba
centos 8.3
fs/functional/{begin clusters/1a3s-mds-4c-client conf/{client mds mon osd} distro/{centos_8.stream} mount/kclient/{mount overrides/{distro/testing/{flavor/centos_latest k-testing} ms-die-on-skipped}} objectstore/bluestore-ec-root overrides/{no_client_pidfile whitelist_health whitelist_wrongly_marked_down} tasks/exports}
master
master
master
gibba
rhel 8.4
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{centos_8.stream} mount/kclient/{mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} ms_mode/{crc} objectstore-ec/bluestore-comp omap_limit/10 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/1 scrub/yes standby-replay tasks/{0-check-counter workunit/suites/ffsb} wsync/{no}}
master
master
master
gibba
rhel 8.4
fs/thrash/workloads/{begin clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{ubuntu_latest} mount/kclient/{mount overrides/{distro/stock/{k-stock rhel_8} ms-die-on-skipped}} msgr-failures/osd-mds-delay objectstore-ec/bluestore-comp overrides/{frag session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} ranks/1 tasks/{1-thrash/mon 2-workunit/fs/snaps}}
master
master
master
gibba
centos 8.3
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{centos_8} mount/kclient/{mount overrides/{distro/testing/{flavor/centos_latest k-testing} ms-die-on-skipped}} ms_mode/{secure} objectstore-ec/bluestore-ec-root omap_limit/10000 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/3 scrub/no standby-replay tasks/{0-check-counter workunit/fs/norstats} wsync/{yes}}
{'Failure object was': {'gibba043.front.sepia.ceph.com': {'results': [{'name': 'zcerza', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'zcerza', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'zcerza', 'ansible_loop_var': 'item', '_ansible_item_label': 'zcerza'}, {'name': 'aschoen', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'aschoen', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'aschoen', 'ansible_loop_var': 'item', '_ansible_item_label': 'aschoen'}, {'name': 'andrew', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'andrew', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'andrew', 'ansible_loop_var': 'item', '_ansible_item_label': 'andrew'}, {'name': 'sweil', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'sweil', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'sweil', 'ansible_loop_var': 'item', '_ansible_item_label': 'sweil'}, {'name': 'brad', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'brad', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'brad', 'ansible_loop_var': 'item', '_ansible_item_label': 'brad'}, {'name': 'kefu', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'kefu', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'kefu', 'ansible_loop_var': 'item', '_ansible_item_label': 'kefu'}, {'name': 'shylesh', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'shylesh', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'shylesh', 'ansible_loop_var': 'item', '_ansible_item_label': 'shylesh'}, {'name': 'gmeno', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'gmeno', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'gmeno', 'ansible_loop_var': 'item', '_ansible_item_label': 'gmeno'}, {'name': 'alfredodeza', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'alfredodeza', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'alfredodeza', 'ansible_loop_var': 'item', '_ansible_item_label': 'alfredodeza'}, {'name': 'trhoden', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'trhoden', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'trhoden', 'ansible_loop_var': 'item', '_ansible_item_label': 'trhoden'}, {'name': 'nishtha', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'nishtha', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'nishtha', 'ansible_loop_var': 'item', '_ansible_item_label': 'nishtha'}, {'name': 'yguang', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'yguang', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'yguang', 'ansible_loop_var': 'item', '_ansible_item_label': 'yguang'}, {'name': 'sdieffen', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'sdieffen', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'sdieffen', 'ansible_loop_var': 'item', '_ansible_item_label': 'sdieffen'}, {'name': 'brian', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'brian', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'brian', 'ansible_loop_var': 'item', '_ansible_item_label': 'brian'}, {'name': 'pmcgarry', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'pmcgarry', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'pmcgarry', 'ansible_loop_var': 'item', '_ansible_item_label': 'pmcgarry'}, {'name': 'karnan', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'karnan', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'karnan', 'ansible_loop_var': 'item', '_ansible_item_label': 'karnan'}, {'name': 'ryneli', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'ryneli', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'ryneli', 'ansible_loop_var': 'item', '_ansible_item_label': 'ryneli'}, {'name': 'dlambrig', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'dlambrig', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'dlambrig', 'ansible_loop_var': 'item', '_ansible_item_label': 'dlambrig'}, {'name': 'icolle', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'icolle', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'icolle', 'ansible_loop_var': 'item', '_ansible_item_label': 'icolle'}, {'name': 'soumya', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'soumya', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'soumya', 'ansible_loop_var': 'item', '_ansible_item_label': 'soumya'}, {'name': 'jspray', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'jspray', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'jspray', 'ansible_loop_var': 'item', '_ansible_item_label': 'jspray'}, {'name': 'erwan', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'erwan', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'erwan', 'ansible_loop_var': 'item', '_ansible_item_label': 'erwan'}, {'name': 'jj', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'jj', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'jj', 'ansible_loop_var': 'item', '_ansible_item_label': 'jj'}, {'name': 'amarangone', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'amarangone', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'amarangone', 'ansible_loop_var': 'item', '_ansible_item_label': 'amarangone'}, {'name': 'oprypin', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'oprypin', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'oprypin', 'ansible_loop_var': 'item', '_ansible_item_label': 'oprypin'}, {'name': 'adamyanova', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'adamyanova', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'adamyanova', 'ansible_loop_var': 'item', '_ansible_item_label': 'adamyanova'}, {'name': 'sbillah', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'sbillah', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'sbillah', 'ansible_loop_var': 'item', '_ansible_item_label': 'sbillah'}, {'name': 'onyb', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'onyb', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'onyb', 'ansible_loop_var': 'item', '_ansible_item_label': 'onyb'}, {'name': 'jwilliamson', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'jwilliamson', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'jwilliamson', 'ansible_loop_var': 'item', '_ansible_item_label': 'jwilliamson'}, {'name': 'kmroz', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'kmroz', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'kmroz', 'ansible_loop_var': 'item', '_ansible_item_label': 'kmroz'}, {'name': 'shehbazj', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'shehbazj', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'shehbazj', 'ansible_loop_var': 'item', '_ansible_item_label': 'shehbazj'}, {'name': 'abhishekvrshny', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'abhishekvrshny', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'abhishekvrshny', 'ansible_loop_var': 'item', '_ansible_item_label': 'abhishekvrshny'}, {'name': 'asheplyakov', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'asheplyakov', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'asheplyakov', 'ansible_loop_var': 'item', '_ansible_item_label': 'asheplyakov'}, {'name': 'liupan', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'liupan', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'liupan', 'ansible_loop_var': 'item', '_ansible_item_label': 'liupan'}, {'name': 'adeza', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'adeza', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'adeza', 'ansible_loop_var': 'item', '_ansible_item_label': 'adeza'}, {'name': 'pranith', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'pranith', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'pranith', 'ansible_loop_var': 'item', '_ansible_item_label': 'pranith'}, {'name': 'dorinda', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'dorinda', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'dorinda', 'ansible_loop_var': 'item', '_ansible_item_label': 'dorinda'}, {'name': 'zyan', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'zyan', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'zyan', 'ansible_loop_var': 'item', '_ansible_item_label': 'zyan'}, {'name': 'jdillaman', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'jdillaman', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'jdillaman', 'ansible_loop_var': 'item', '_ansible_item_label': 'jdillaman'}, {'name': 'davidz', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'davidz', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'davidz', 'ansible_loop_var': 'item', '_ansible_item_label': 'davidz'}, {'name': 'wusui', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'wusui', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'wusui', 'ansible_loop_var': 'item', '_ansible_item_label': 'wusui'}, {'name': 'nwatkins', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'nwatkins', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'nwatkins', 'ansible_loop_var': 'item', '_ansible_item_label': 'nwatkins'}, {'name': 'sidharthanup', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'sidharthanup', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'sidharthanup', 'ansible_loop_var': 'item', '_ansible_item_label': 'sidharthanup'}, {'name': 'varsha', 'state': 'absent', 'force': False, 'remove': False, 'changed': True, 'invocation': {'module_args': {'name': 'varsha', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'varsha', 'ansible_loop_var': 'item', '_ansible_item_label': 'varsha'}, {'name': 'hmunjulu', 'msg': "userdel: user 'hmunjulu' does not exist\n", 'rc': 6, 'failed': True, 'invocation': {'module_args': {'name': 'hmunjulu', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'changed': False, 'item': 'hmunjulu', 'ansible_loop_var': 'item', '_ansible_item_label': 'hmunjulu'}, {'name': 'nlevine', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'nlevine', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'nlevine', 'ansible_loop_var': 'item', '_ansible_item_label': 'nlevine'}, {'name': 'tbrekke', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'tbrekke', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'tbrekke', 'ansible_loop_var': 'item', '_ansible_item_label': 'tbrekke'}, {'name': 'taco', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'taco', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'taco', 'ansible_loop_var': 'item', '_ansible_item_label': 'taco'}, {'name': 'louis', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'louis', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'louis', 'ansible_loop_var': 'item', '_ansible_item_label': 'louis'}, {'name': 'amarango', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'amarango', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'amarango', 'ansible_loop_var': 'item', '_ansible_item_label': 'amarango'}, {'name': 'oobe', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'oobe', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'oobe', 'ansible_loop_var': 'item', '_ansible_item_label': 'oobe'}, {'name': 'rturk', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'rturk', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'rturk', 'ansible_loop_var': 'item', '_ansible_item_label': 'rturk'}, {'name': 'fche', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'fche', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'fche', 'ansible_loop_var': 'item', '_ansible_item_label': 'fche'}, {'name': 'jbainbri', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'jbainbri', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'jbainbri', 'ansible_loop_var': 'item', '_ansible_item_label': 'jbainbri'}, {'name': 'kdhananj', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'kdhananj', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'kdhananj', 'ansible_loop_var': 'item', '_ansible_item_label': 'kdhananj'}], 'changed': True, 'msg': 'All items completed'}}, 'Traceback (most recent call last)': 'File "/home/teuthworker/src/git.ceph.com_git_ceph-cm-ansible_master/callback_plugins/failure_log.py", line 44, in log_failure log.error(yaml.safe_dump(failure)) File "/home/teuthworker/src/git.ceph.com_git_teuthology_6cfddd5e1c7d5ede1367cf3087c21af44eb88356/virtualenv/lib/python3.6/site-packages/yaml/__init__.py", line 306, in safe_dump return dump_all([data], stream, Dumper=SafeDumper, **kwds) File "/home/teuthworker/src/git.ceph.com_git_teuthology_6cfddd5e1c7d5ede1367cf3087c21af44eb88356/virtualenv/lib/python3.6/site-packages/yaml/__init__.py", line 278, in dump_all dumper.represent(data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_6cfddd5e1c7d5ede1367cf3087c21af44eb88356/virtualenv/lib/python3.6/site-packages/yaml/representer.py", line 27, in represent node = self.represent_data(data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_6cfddd5e1c7d5ede1367cf3087c21af44eb88356/virtualenv/lib/python3.6/site-packages/yaml/representer.py", line 48, in represent_data node = self.yaml_representers[data_types[0]](self, data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_6cfddd5e1c7d5ede1367cf3087c21af44eb88356/virtualenv/lib/python3.6/site-packages/yaml/representer.py", line 207, in represent_dict return self.represent_mapping(\'tag:yaml.org,2002:map\', data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_6cfddd5e1c7d5ede1367cf3087c21af44eb88356/virtualenv/lib/python3.6/site-packages/yaml/representer.py", line 118, in represent_mapping node_value = self.represent_data(item_value) File "/home/teuthworker/src/git.ceph.com_git_teuthology_6cfddd5e1c7d5ede1367cf3087c21af44eb88356/virtualenv/lib/python3.6/site-packages/yaml/representer.py", line 48, in represent_data node = self.yaml_representers[data_types[0]](self, data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_6cfddd5e1c7d5ede1367cf3087c21af44eb88356/virtualenv/lib/python3.6/site-packages/yaml/representer.py", line 207, in represent_dict return self.represent_mapping(\'tag:yaml.org,2002:map\', data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_6cfddd5e1c7d5ede1367cf3087c21af44eb88356/virtualenv/lib/python3.6/site-packages/yaml/representer.py", line 118, in represent_mapping node_value = self.represent_data(item_value) File "/home/teuthworker/src/git.ceph.com_git_teuthology_6cfddd5e1c7d5ede1367cf3087c21af44eb88356/virtualenv/lib/python3.6/site-packages/yaml/representer.py", line 48, in represent_data node = self.yaml_representers[data_types[0]](self, data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_6cfddd5e1c7d5ede1367cf3087c21af44eb88356/virtualenv/lib/python3.6/site-packages/yaml/representer.py", line 199, in represent_list return self.represent_sequence(\'tag:yaml.org,2002:seq\', data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_6cfddd5e1c7d5ede1367cf3087c21af44eb88356/virtualenv/lib/python3.6/site-packages/yaml/representer.py", line 92, in represent_sequence node_item = self.represent_data(item) File "/home/teuthworker/src/git.ceph.com_git_teuthology_6cfddd5e1c7d5ede1367cf3087c21af44eb88356/virtualenv/lib/python3.6/site-packages/yaml/representer.py", line 48, in represent_data node = self.yaml_representers[data_types[0]](self, data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_6cfddd5e1c7d5ede1367cf3087c21af44eb88356/virtualenv/lib/python3.6/site-packages/yaml/representer.py", line 207, in represent_dict return self.represent_mapping(\'tag:yaml.org,2002:map\', data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_6cfddd5e1c7d5ede1367cf3087c21af44eb88356/virtualenv/lib/python3.6/site-packages/yaml/representer.py", line 118, in represent_mapping node_value = self.represent_data(item_value) File "/home/teuthworker/src/git.ceph.com_git_teuthology_6cfddd5e1c7d5ede1367cf3087c21af44eb88356/virtualenv/lib/python3.6/site-packages/yaml/representer.py", line 58, in represent_data node = self.yaml_representers[None](self, data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_6cfddd5e1c7d5ede1367cf3087c21af44eb88356/virtualenv/lib/python3.6/site-packages/yaml/representer.py", line 231, in represent_undefined raise RepresenterError("cannot represent an object", data)', 'yaml.representer.RepresenterError': "('cannot represent an object', 'zcerza')"}
master
master
master
gibba
ubuntu 20.04
fs/functional/{begin clusters/1a3s-mds-4c-client conf/{client mds mon osd} distro/{centos_8.stream} mount/kclient/{mount overrides/{distro/testing/{flavor/ubuntu_latest k-testing} ms-die-on-skipped}} objectstore/bluestore-bitmap overrides/{no_client_pidfile whitelist_health whitelist_wrongly_marked_down} tasks/forward-scrub}
{'Failure object was': {'gibba043.front.sepia.ceph.com': {'results': [{'name': 'zcerza', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'zcerza', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'zcerza', 'ansible_loop_var': 'item', '_ansible_item_label': 'zcerza'}, {'name': 'aschoen', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'aschoen', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'aschoen', 'ansible_loop_var': 'item', '_ansible_item_label': 'aschoen'}, {'name': 'andrew', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'andrew', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'andrew', 'ansible_loop_var': 'item', '_ansible_item_label': 'andrew'}, {'name': 'sweil', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'sweil', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'sweil', 'ansible_loop_var': 'item', '_ansible_item_label': 'sweil'}, {'name': 'brad', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'brad', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'brad', 'ansible_loop_var': 'item', '_ansible_item_label': 'brad'}, {'name': 'kefu', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'kefu', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'kefu', 'ansible_loop_var': 'item', '_ansible_item_label': 'kefu'}, {'name': 'shylesh', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'shylesh', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'shylesh', 'ansible_loop_var': 'item', '_ansible_item_label': 'shylesh'}, {'name': 'gmeno', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'gmeno', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'gmeno', 'ansible_loop_var': 'item', '_ansible_item_label': 'gmeno'}, {'name': 'alfredodeza', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'alfredodeza', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'alfredodeza', 'ansible_loop_var': 'item', '_ansible_item_label': 'alfredodeza'}, {'name': 'trhoden', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'trhoden', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'trhoden', 'ansible_loop_var': 'item', '_ansible_item_label': 'trhoden'}, {'name': 'nishtha', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'nishtha', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'nishtha', 'ansible_loop_var': 'item', '_ansible_item_label': 'nishtha'}, {'name': 'yguang', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'yguang', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'yguang', 'ansible_loop_var': 'item', '_ansible_item_label': 'yguang'}, {'name': 'sdieffen', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'sdieffen', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'sdieffen', 'ansible_loop_var': 'item', '_ansible_item_label': 'sdieffen'}, {'name': 'brian', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'brian', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'brian', 'ansible_loop_var': 'item', '_ansible_item_label': 'brian'}, {'name': 'pmcgarry', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'pmcgarry', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'pmcgarry', 'ansible_loop_var': 'item', '_ansible_item_label': 'pmcgarry'}, {'name': 'karnan', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'karnan', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'karnan', 'ansible_loop_var': 'item', '_ansible_item_label': 'karnan'}, {'name': 'ryneli', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'ryneli', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'ryneli', 'ansible_loop_var': 'item', '_ansible_item_label': 'ryneli'}, {'name': 'dlambrig', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'dlambrig', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'dlambrig', 'ansible_loop_var': 'item', '_ansible_item_label': 'dlambrig'}, {'name': 'icolle', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'icolle', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'icolle', 'ansible_loop_var': 'item', '_ansible_item_label': 'icolle'}, {'name': 'soumya', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'soumya', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'soumya', 'ansible_loop_var': 'item', '_ansible_item_label': 'soumya'}, {'name': 'jspray', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'jspray', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'jspray', 'ansible_loop_var': 'item', '_ansible_item_label': 'jspray'}, {'name': 'erwan', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'erwan', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'erwan', 'ansible_loop_var': 'item', '_ansible_item_label': 'erwan'}, {'name': 'jj', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'jj', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'jj', 'ansible_loop_var': 'item', '_ansible_item_label': 'jj'}, {'name': 'amarangone', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'amarangone', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'amarangone', 'ansible_loop_var': 'item', '_ansible_item_label': 'amarangone'}, {'name': 'oprypin', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'oprypin', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'oprypin', 'ansible_loop_var': 'item', '_ansible_item_label': 'oprypin'}, {'name': 'adamyanova', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'adamyanova', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'adamyanova', 'ansible_loop_var': 'item', '_ansible_item_label': 'adamyanova'}, {'name': 'sbillah', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'sbillah', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'sbillah', 'ansible_loop_var': 'item', '_ansible_item_label': 'sbillah'}, {'name': 'onyb', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'onyb', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'onyb', 'ansible_loop_var': 'item', '_ansible_item_label': 'onyb'}, {'name': 'jwilliamson', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'jwilliamson', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'jwilliamson', 'ansible_loop_var': 'item', '_ansible_item_label': 'jwilliamson'}, {'name': 'kmroz', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'kmroz', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'kmroz', 'ansible_loop_var': 'item', '_ansible_item_label': 'kmroz'}, {'name': 'shehbazj', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'shehbazj', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'shehbazj', 'ansible_loop_var': 'item', '_ansible_item_label': 'shehbazj'}, {'name': 'abhishekvrshny', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'abhishekvrshny', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'abhishekvrshny', 'ansible_loop_var': 'item', '_ansible_item_label': 'abhishekvrshny'}, {'name': 'asheplyakov', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'asheplyakov', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'asheplyakov', 'ansible_loop_var': 'item', '_ansible_item_label': 'asheplyakov'}, {'name': 'liupan', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'liupan', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'liupan', 'ansible_loop_var': 'item', '_ansible_item_label': 'liupan'}, {'name': 'adeza', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'adeza', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'adeza', 'ansible_loop_var': 'item', '_ansible_item_label': 'adeza'}, {'name': 'pranith', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'pranith', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'pranith', 'ansible_loop_var': 'item', '_ansible_item_label': 'pranith'}, {'name': 'dorinda', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'dorinda', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'dorinda', 'ansible_loop_var': 'item', '_ansible_item_label': 'dorinda'}, {'name': 'zyan', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'zyan', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'zyan', 'ansible_loop_var': 'item', '_ansible_item_label': 'zyan'}, {'name': 'jdillaman', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'jdillaman', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'jdillaman', 'ansible_loop_var': 'item', '_ansible_item_label': 'jdillaman'}, {'name': 'davidz', 'state': 'absent', 'force': False, 'remove': False, 'changed': True, 'invocation': {'module_args': {'name': 'davidz', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'davidz', 'ansible_loop_var': 'item', '_ansible_item_label': 'davidz'}, {'name': 'wusui', 'state': 'absent', 'force': False, 'remove': False, 'changed': True, 'invocation': {'module_args': {'name': 'wusui', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'wusui', 'ansible_loop_var': 'item', '_ansible_item_label': 'wusui'}, {'name': 'nwatkins', 'state': 'absent', 'force': False, 'remove': False, 'changed': True, 'invocation': {'module_args': {'name': 'nwatkins', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'nwatkins', 'ansible_loop_var': 'item', '_ansible_item_label': 'nwatkins'}, {'name': 'sidharthanup', 'state': 'absent', 'force': False, 'remove': False, 'changed': True, 'invocation': {'module_args': {'name': 'sidharthanup', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'sidharthanup', 'ansible_loop_var': 'item', '_ansible_item_label': 'sidharthanup'}, {'name': 'varsha', 'msg': "userdel: user 'varsha' does not exist\n", 'rc': 6, 'failed': True, 'invocation': {'module_args': {'name': 'varsha', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'changed': False, 'item': 'varsha', 'ansible_loop_var': 'item', '_ansible_item_label': 'varsha'}, {'name': 'hmunjulu', 'state': 'absent', 'force': False, 'remove': False, 'changed': True, 'invocation': {'module_args': {'name': 'hmunjulu', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'hmunjulu', 'ansible_loop_var': 'item', '_ansible_item_label': 'hmunjulu'}, {'name': 'nlevine', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'nlevine', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'nlevine', 'ansible_loop_var': 'item', '_ansible_item_label': 'nlevine'}, {'name': 'tbrekke', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'tbrekke', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'tbrekke', 'ansible_loop_var': 'item', '_ansible_item_label': 'tbrekke'}, {'name': 'taco', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'taco', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'taco', 'ansible_loop_var': 'item', '_ansible_item_label': 'taco'}, {'name': 'louis', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'louis', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'louis', 'ansible_loop_var': 'item', '_ansible_item_label': 'louis'}, {'name': 'amarango', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'amarango', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'amarango', 'ansible_loop_var': 'item', '_ansible_item_label': 'amarango'}, {'name': 'oobe', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'oobe', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'oobe', 'ansible_loop_var': 'item', '_ansible_item_label': 'oobe'}, {'name': 'rturk', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'rturk', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'rturk', 'ansible_loop_var': 'item', '_ansible_item_label': 'rturk'}, {'name': 'fche', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'fche', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'fche', 'ansible_loop_var': 'item', '_ansible_item_label': 'fche'}, {'name': 'jbainbri', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'jbainbri', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'jbainbri', 'ansible_loop_var': 'item', '_ansible_item_label': 'jbainbri'}, {'name': 'kdhananj', 'state': 'absent', 'changed': False, 'invocation': {'module_args': {'name': 'kdhananj', 'state': 'absent', 'non_unique': False, 'force': False, 'remove': False, 'create_home': True, 'system': False, 'move_home': False, 'append': False, 'ssh_key_bits': 0, 'ssh_key_type': 'rsa', 'ssh_key_comment': 'ansible-generated on gibba043.front.sepia.ceph.com', 'update_password': 'always', 'uid': 'None', 'group': 'None', 'groups': 'None', 'comment': 'None', 'home': 'None', 'shell': 'None', 'password': 'None', 'login_class': 'None', 'hidden': 'None', 'seuser': 'None', 'skeleton': 'None', 'generate_ssh_key': 'None', 'ssh_key_file': 'None', 'ssh_key_passphrase': 'None', 'expires': 'None', 'password_lock': 'None', 'local': 'None', 'profile': 'None', 'authorization': 'None', 'role': 'None'}}, '_ansible_no_log': False, 'failed': False, 'item': 'kdhananj', 'ansible_loop_var': 'item', '_ansible_item_label': 'kdhananj'}], 'changed': True, 'msg': 'All items completed'}}, 'Traceback (most recent call last)': 'File "/home/teuthworker/src/git.ceph.com_git_ceph-cm-ansible_master/callback_plugins/failure_log.py", line 44, in log_failure log.error(yaml.safe_dump(failure)) File "/home/teuthworker/src/git.ceph.com_git_teuthology_6cfddd5e1c7d5ede1367cf3087c21af44eb88356/virtualenv/lib/python3.6/site-packages/yaml/__init__.py", line 306, in safe_dump return dump_all([data], stream, Dumper=SafeDumper, **kwds) File "/home/teuthworker/src/git.ceph.com_git_teuthology_6cfddd5e1c7d5ede1367cf3087c21af44eb88356/virtualenv/lib/python3.6/site-packages/yaml/__init__.py", line 278, in dump_all dumper.represent(data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_6cfddd5e1c7d5ede1367cf3087c21af44eb88356/virtualenv/lib/python3.6/site-packages/yaml/representer.py", line 27, in represent node = self.represent_data(data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_6cfddd5e1c7d5ede1367cf3087c21af44eb88356/virtualenv/lib/python3.6/site-packages/yaml/representer.py", line 48, in represent_data node = self.yaml_representers[data_types[0]](self, data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_6cfddd5e1c7d5ede1367cf3087c21af44eb88356/virtualenv/lib/python3.6/site-packages/yaml/representer.py", line 207, in represent_dict return self.represent_mapping(\'tag:yaml.org,2002:map\', data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_6cfddd5e1c7d5ede1367cf3087c21af44eb88356/virtualenv/lib/python3.6/site-packages/yaml/representer.py", line 118, in represent_mapping node_value = self.represent_data(item_value) File "/home/teuthworker/src/git.ceph.com_git_teuthology_6cfddd5e1c7d5ede1367cf3087c21af44eb88356/virtualenv/lib/python3.6/site-packages/yaml/representer.py", line 48, in represent_data node = self.yaml_representers[data_types[0]](self, data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_6cfddd5e1c7d5ede1367cf3087c21af44eb88356/virtualenv/lib/python3.6/site-packages/yaml/representer.py", line 207, in represent_dict return self.represent_mapping(\'tag:yaml.org,2002:map\', data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_6cfddd5e1c7d5ede1367cf3087c21af44eb88356/virtualenv/lib/python3.6/site-packages/yaml/representer.py", line 118, in represent_mapping node_value = self.represent_data(item_value) File "/home/teuthworker/src/git.ceph.com_git_teuthology_6cfddd5e1c7d5ede1367cf3087c21af44eb88356/virtualenv/lib/python3.6/site-packages/yaml/representer.py", line 48, in represent_data node = self.yaml_representers[data_types[0]](self, data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_6cfddd5e1c7d5ede1367cf3087c21af44eb88356/virtualenv/lib/python3.6/site-packages/yaml/representer.py", line 199, in represent_list return self.represent_sequence(\'tag:yaml.org,2002:seq\', data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_6cfddd5e1c7d5ede1367cf3087c21af44eb88356/virtualenv/lib/python3.6/site-packages/yaml/representer.py", line 92, in represent_sequence node_item = self.represent_data(item) File "/home/teuthworker/src/git.ceph.com_git_teuthology_6cfddd5e1c7d5ede1367cf3087c21af44eb88356/virtualenv/lib/python3.6/site-packages/yaml/representer.py", line 48, in represent_data node = self.yaml_representers[data_types[0]](self, data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_6cfddd5e1c7d5ede1367cf3087c21af44eb88356/virtualenv/lib/python3.6/site-packages/yaml/representer.py", line 207, in represent_dict return self.represent_mapping(\'tag:yaml.org,2002:map\', data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_6cfddd5e1c7d5ede1367cf3087c21af44eb88356/virtualenv/lib/python3.6/site-packages/yaml/representer.py", line 118, in represent_mapping node_value = self.represent_data(item_value) File "/home/teuthworker/src/git.ceph.com_git_teuthology_6cfddd5e1c7d5ede1367cf3087c21af44eb88356/virtualenv/lib/python3.6/site-packages/yaml/representer.py", line 58, in represent_data node = self.yaml_representers[None](self, data) File "/home/teuthworker/src/git.ceph.com_git_teuthology_6cfddd5e1c7d5ede1367cf3087c21af44eb88356/virtualenv/lib/python3.6/site-packages/yaml/representer.py", line 231, in represent_undefined raise RepresenterError("cannot represent an object", data)', 'yaml.representer.RepresenterError': "('cannot represent an object', 'zcerza')"}
master
master
master
gibba
ubuntu 20.04
fs/workload/{begin clusters/1a11s-mds-1c-client-3node conf/{client mds mon osd} distro/{ubuntu_latest} mount/kclient/{mount overrides/{distro/testing/{flavor/ubuntu_latest k-testing} ms-die-on-skipped}} ms_mode/{secure} objectstore-ec/bluestore-bitmap omap_limit/10 overrides/{frag osd-asserts session_timeout whitelist_health whitelist_wrongly_marked_down} ranks/5 scrub/yes standby-replay tasks/{0-check-counter workunit/suites/fsstress} wsync/{no}}
master
master
master
gibba
centos 8.3
fs/thrash/workloads/{begin clusters/1a5s-mds-1c-client conf/{client mds mon osd} distro/{centos_8.stream} mount/kclient/{mount overrides/{distro/testing/{flavor/centos_latest k-testing} ms-die-on-skipped}} msgr-failures/none objectstore-ec/bluestore-ec-root overrides/{frag session_timeout thrashosds-health whitelist_health whitelist_wrongly_marked_down} ranks/3 tasks/{1-thrash/osd 2-workunit/suites/ffsb}}