ID
Status
Ceph Branch
Suite Branch
Teuthology Branch
Machine
OS
Nodes
Description
Failure Reason
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore-ec/bluestore-comp-ec-root.yaml tasks/kclient_workunit_direct_io.yaml}
luminous
luminous
master
ovh
 
kcephfs/mixed-clients/{clusters/2-clients.yaml conf.yaml objectstore-ec/bluestore-comp-ec-root.yaml tasks/kernel_cfuse_workunits_dbench_iozone.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/bluestore-comp-ec-root.yaml tasks/auto-repair.yaml whitelist_health.yaml}
luminous
luminous
master
ovh
 
kcephfs/thrash/{clusters/fixed-3-cephfs.yaml conf.yaml objectstore-ec/bluestore-comp-ec-root.yaml thrashers/default.yaml thrashosds-health.yaml whitelist_health.yaml workloads/kclient_workunit_suites_ffsb.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore-ec/bluestore-comp.yaml tasks/kclient_workunit_kernel_untar_build.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/bluestore-comp.yaml tasks/backtrace.yaml whitelist_health.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore-ec/bluestore-ec-root.yaml tasks/kclient_workunit_misc.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore-ec/bluestore.yaml tasks/kclient_workunit_o_trunc.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/bluestore-ec-root.yaml tasks/client-limits.yaml whitelist_health.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore-ec/filestore-xfs.yaml tasks/kclient_workunit_snaps.yaml}
luminous
luminous
master
ovh
 
kcephfs/thrash/{clusters/fixed-3-cephfs.yaml conf.yaml objectstore-ec/bluestore-comp.yaml thrashers/mds.yaml thrashosds-health.yaml whitelist_health.yaml workloads/kclient_workunit_suites_iozone.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/bluestore.yaml tasks/client-recovery.yaml whitelist_health.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore-ec/bluestore-comp-ec-root.yaml tasks/kclient_workunit_suites_dbench.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore-ec/bluestore-comp.yaml tasks/kclient_workunit_suites_ffsb.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/filestore-xfs.yaml tasks/config-commands.yaml whitelist_health.yaml}
Command failed on ovh072 with status 1: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper kill ceph-osd -f --cluster ceph -i 7'
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore-ec/bluestore-ec-root.yaml tasks/kclient_workunit_suites_fsstress.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore-ec/bluestore.yaml tasks/kclient_workunit_suites_fsx.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/bluestore-comp-ec-root.yaml tasks/damage.yaml whitelist_health.yaml}
luminous
luminous
master
ovh
 
kcephfs/thrash/{clusters/fixed-3-cephfs.yaml conf.yaml objectstore-ec/bluestore-ec-root.yaml thrashers/mon.yaml thrashosds-health.yaml whitelist_health.yaml workloads/kclient_workunit_suites_ffsb.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore-ec/filestore-xfs.yaml tasks/kclient_workunit_suites_fsync.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/bluestore-comp.yaml tasks/data-scan.yaml whitelist_health.yaml}
Test failure: test_stashed_layout (tasks.cephfs.test_data_scan.TestDataScan)
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore-ec/bluestore-comp-ec-root.yaml tasks/kclient_workunit_suites_iozone.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore-ec/bluestore-comp.yaml tasks/kclient_workunit_suites_pjd.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/bluestore-ec-root.yaml tasks/failover.yaml whitelist_health.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore-ec/bluestore-ec-root.yaml tasks/kclient_workunit_trivial_sync.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore-ec/bluestore.yaml tasks/kclient_workunit_direct_io.yaml}
luminous
luminous
master
ovh
 
kcephfs/mixed-clients/{clusters/2-clients.yaml conf.yaml objectstore-ec/bluestore-comp.yaml tasks/kernel_cfuse_workunits_untarbuild_blogbench.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/bluestore.yaml tasks/forward-scrub.yaml whitelist_health.yaml}
luminous
luminous
master
ovh
 
kcephfs/thrash/{clusters/fixed-3-cephfs.yaml conf.yaml objectstore-ec/bluestore.yaml thrashers/default.yaml thrashosds-health.yaml whitelist_health.yaml workloads/kclient_workunit_suites_iozone.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore-ec/filestore-xfs.yaml tasks/kclient_workunit_kernel_untar_build.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/filestore-xfs.yaml tasks/journal-repair.yaml whitelist_health.yaml}
Command failed on ovh037 with status 1: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper kill ceph-osd -f --cluster ceph -i 3'
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore-ec/bluestore-comp-ec-root.yaml tasks/kclient_workunit_misc.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore-ec/bluestore-comp.yaml tasks/kclient_workunit_o_trunc.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/bluestore-comp-ec-root.yaml tasks/mds-flush.yaml whitelist_health.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore-ec/bluestore-ec-root.yaml tasks/kclient_workunit_snaps.yaml}
luminous
luminous
master
ovh
 
kcephfs/thrash/{clusters/fixed-3-cephfs.yaml conf.yaml objectstore-ec/filestore-xfs.yaml thrashers/mds.yaml thrashosds-health.yaml whitelist_health.yaml workloads/kclient_workunit_suites_ffsb.yaml}
Command failed (workunit test suites/ffsb.sh) on ovh094 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=luminous TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/suites/ffsb.sh'
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/bluestore-comp.yaml tasks/mds-full.yaml whitelist_health.yaml}
Test failure: test_full_different_file (tasks.cephfs.test_full.TestClusterFull)
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore-ec/bluestore.yaml tasks/kclient_workunit_suites_dbench.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore-ec/filestore-xfs.yaml tasks/kclient_workunit_suites_ffsb.yaml}
Command failed (workunit test suites/ffsb.sh) on ovh089 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=luminous TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/suites/ffsb.sh'
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/bluestore-ec-root.yaml tasks/pool-perm.yaml whitelist_health.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore-ec/bluestore-comp-ec-root.yaml tasks/kclient_workunit_suites_fsstress.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore-ec/bluestore-comp.yaml tasks/kclient_workunit_suites_fsx.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/bluestore.yaml tasks/sessionmap.yaml whitelist_health.yaml}
luminous
luminous
master
ovh
 
kcephfs/thrash/{clusters/fixed-3-cephfs.yaml conf.yaml objectstore-ec/bluestore-comp-ec-root.yaml thrashers/mon.yaml thrashosds-health.yaml whitelist_health.yaml workloads/kclient_workunit_suites_iozone.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore-ec/bluestore-ec-root.yaml tasks/kclient_workunit_suites_fsync.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/filestore-xfs.yaml tasks/strays.yaml whitelist_health.yaml}
Command failed on ovh039 with status 1: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper kill ceph-osd -f --cluster ceph -i 3'
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore-ec/bluestore.yaml tasks/kclient_workunit_suites_iozone.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore-ec/filestore-xfs.yaml tasks/kclient_workunit_suites_pjd.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/bluestore-comp-ec-root.yaml tasks/volume-client.yaml whitelist_health.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore-ec/bluestore-comp-ec-root.yaml tasks/kclient_workunit_trivial_sync.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore-ec/bluestore-comp.yaml tasks/kclient_workunit_direct_io.yaml}
luminous
luminous
master
ovh
 
kcephfs/mixed-clients/{clusters/2-clients.yaml conf.yaml objectstore-ec/bluestore-ec-root.yaml tasks/kernel_cfuse_workunits_dbench_iozone.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/bluestore-comp.yaml tasks/auto-repair.yaml whitelist_health.yaml}
luminous
luminous
master
ovh
 
kcephfs/thrash/{clusters/fixed-3-cephfs.yaml conf.yaml objectstore-ec/bluestore-comp.yaml thrashers/default.yaml thrashosds-health.yaml whitelist_health.yaml workloads/kclient_workunit_suites_ffsb.yaml}
Socket is closed
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore-ec/bluestore-ec-root.yaml tasks/kclient_workunit_kernel_untar_build.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/bluestore-ec-root.yaml tasks/backtrace.yaml whitelist_health.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore-ec/bluestore.yaml tasks/kclient_workunit_misc.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore-ec/filestore-xfs.yaml tasks/kclient_workunit_o_trunc.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/bluestore.yaml tasks/client-limits.yaml whitelist_health.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore-ec/bluestore-comp-ec-root.yaml tasks/kclient_workunit_snaps.yaml}
luminous
luminous
master
ovh
 
kcephfs/thrash/{clusters/fixed-3-cephfs.yaml conf.yaml objectstore-ec/bluestore-ec-root.yaml thrashers/mds.yaml thrashosds-health.yaml whitelist_health.yaml workloads/kclient_workunit_suites_iozone.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/filestore-xfs.yaml tasks/client-recovery.yaml whitelist_health.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore-ec/bluestore-comp.yaml tasks/kclient_workunit_suites_dbench.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore-ec/bluestore-ec-root.yaml tasks/kclient_workunit_suites_ffsb.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/bluestore-comp-ec-root.yaml tasks/config-commands.yaml whitelist_health.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore-ec/bluestore.yaml tasks/kclient_workunit_suites_fsstress.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore-ec/filestore-xfs.yaml tasks/kclient_workunit_suites_fsx.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/bluestore-comp.yaml tasks/damage.yaml whitelist_health.yaml}
luminous
luminous
master
ovh
 
kcephfs/thrash/{clusters/fixed-3-cephfs.yaml conf.yaml objectstore-ec/bluestore.yaml thrashers/mon.yaml thrashosds-health.yaml whitelist_health.yaml workloads/kclient_workunit_suites_ffsb.yaml}
Found coredumps on ubuntu@ovh072.front.sepia.ceph.com
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore-ec/bluestore-comp-ec-root.yaml tasks/kclient_workunit_suites_fsync.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/bluestore-ec-root.yaml tasks/data-scan.yaml whitelist_health.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore-ec/bluestore-comp.yaml tasks/kclient_workunit_suites_iozone.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore-ec/bluestore-ec-root.yaml tasks/kclient_workunit_suites_pjd.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/bluestore.yaml tasks/failover.yaml whitelist_health.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore-ec/bluestore.yaml tasks/kclient_workunit_trivial_sync.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore-ec/filestore-xfs.yaml tasks/kclient_workunit_direct_io.yaml}
luminous
luminous
master
ovh
 
kcephfs/mixed-clients/{clusters/2-clients.yaml conf.yaml objectstore-ec/bluestore.yaml tasks/kernel_cfuse_workunits_untarbuild_blogbench.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/filestore-xfs.yaml tasks/forward-scrub.yaml whitelist_health.yaml}
Command failed on ovh011 with status 1: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper kill ceph-osd -f --cluster ceph -i 7'
luminous
luminous
master
ovh
 
kcephfs/thrash/{clusters/fixed-3-cephfs.yaml conf.yaml objectstore-ec/filestore-xfs.yaml thrashers/default.yaml thrashosds-health.yaml whitelist_health.yaml workloads/kclient_workunit_suites_iozone.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore-ec/bluestore-comp-ec-root.yaml tasks/kclient_workunit_kernel_untar_build.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/bluestore-comp-ec-root.yaml tasks/journal-repair.yaml whitelist_health.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore-ec/bluestore-comp.yaml tasks/kclient_workunit_misc.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore-ec/bluestore-ec-root.yaml tasks/kclient_workunit_o_trunc.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/bluestore-comp.yaml tasks/mds-flush.yaml whitelist_health.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore-ec/bluestore.yaml tasks/kclient_workunit_snaps.yaml}
luminous
luminous
master
ovh
 
kcephfs/thrash/{clusters/fixed-3-cephfs.yaml conf.yaml objectstore-ec/bluestore-comp-ec-root.yaml thrashers/mds.yaml thrashosds-health.yaml whitelist_health.yaml workloads/kclient_workunit_suites_ffsb.yaml}
Command failed (workunit test suites/ffsb.sh) on ovh089 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=luminous TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/suites/ffsb.sh'
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/bluestore-ec-root.yaml tasks/mds-full.yaml whitelist_health.yaml}
"2018-03-03 23:04:10.593992 mon.a mon.0 158.69.92.7:6789/0 202 : cluster [WRN] Health check failed: 2 osds down (OSD_DOWN)" in cluster log
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore-ec/filestore-xfs.yaml tasks/kclient_workunit_suites_dbench.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore-ec/bluestore-comp-ec-root.yaml tasks/kclient_workunit_suites_ffsb.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/bluestore.yaml tasks/pool-perm.yaml whitelist_health.yaml}
"2018-03-04 05:59:35.674546 mon.a mon.0 158.69.68.70:6789/0 537 : cluster [WRN] daemon mds.c is not responding, replacing it as rank 0 with standby daemon mds.d" in cluster log
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore-ec/bluestore-comp.yaml tasks/kclient_workunit_suites_fsstress.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore-ec/bluestore-ec-root.yaml tasks/kclient_workunit_suites_fsx.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/filestore-xfs.yaml tasks/sessionmap.yaml whitelist_health.yaml}
Command failed on ovh098 with status 1: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper kill ceph-osd -f --cluster ceph -i 3'
luminous
luminous
master
ovh
 
kcephfs/thrash/{clusters/fixed-3-cephfs.yaml conf.yaml objectstore-ec/bluestore-comp.yaml thrashers/mon.yaml thrashosds-health.yaml whitelist_health.yaml workloads/kclient_workunit_suites_iozone.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore-ec/bluestore.yaml tasks/kclient_workunit_suites_fsync.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/bluestore-comp-ec-root.yaml tasks/strays.yaml whitelist_health.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore-ec/filestore-xfs.yaml tasks/kclient_workunit_suites_iozone.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore-ec/bluestore-comp-ec-root.yaml tasks/kclient_workunit_suites_pjd.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/bluestore-comp.yaml tasks/volume-client.yaml whitelist_health.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore-ec/bluestore-comp.yaml tasks/kclient_workunit_trivial_sync.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore-ec/bluestore-ec-root.yaml tasks/kclient_workunit_direct_io.yaml}
luminous
luminous
master
ovh
 
kcephfs/mixed-clients/{clusters/2-clients.yaml conf.yaml objectstore-ec/filestore-xfs.yaml tasks/kernel_cfuse_workunits_dbench_iozone.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/bluestore-ec-root.yaml tasks/auto-repair.yaml whitelist_health.yaml}
luminous
luminous
master
ovh
 
kcephfs/thrash/{clusters/fixed-3-cephfs.yaml conf.yaml objectstore-ec/bluestore-ec-root.yaml thrashers/default.yaml thrashosds-health.yaml whitelist_health.yaml workloads/kclient_workunit_suites_ffsb.yaml}
timed out waiting for admin_socket to appear after osd.2 restart
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore-ec/bluestore.yaml tasks/kclient_workunit_kernel_untar_build.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/bluestore.yaml tasks/backtrace.yaml whitelist_health.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore-ec/filestore-xfs.yaml tasks/kclient_workunit_misc.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore-ec/bluestore-comp-ec-root.yaml tasks/kclient_workunit_o_trunc.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/filestore-xfs.yaml tasks/client-limits.yaml whitelist_health.yaml}
Command failed on ovh078 with status 1: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper kill ceph-osd -f --cluster ceph -i 3'
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore-ec/bluestore-comp.yaml tasks/kclient_workunit_snaps.yaml}
luminous
luminous
master
ovh
 
kcephfs/thrash/{clusters/fixed-3-cephfs.yaml conf.yaml objectstore-ec/bluestore.yaml thrashers/mds.yaml thrashosds-health.yaml whitelist_health.yaml workloads/kclient_workunit_suites_iozone.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/bluestore-comp-ec-root.yaml tasks/client-recovery.yaml whitelist_health.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore-ec/bluestore-ec-root.yaml tasks/kclient_workunit_suites_dbench.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore-ec/bluestore.yaml tasks/kclient_workunit_suites_ffsb.yaml}
"2018-03-04 00:48:02.451967 mon.a mon.0 158.69.94.30:6789/0 162 : cluster [WRN] Health check failed: 1 osds down (OSD_DOWN)" in cluster log
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/bluestore-comp.yaml tasks/config-commands.yaml whitelist_health.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore-ec/filestore-xfs.yaml tasks/kclient_workunit_suites_fsstress.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore-ec/bluestore-comp-ec-root.yaml tasks/kclient_workunit_suites_fsx.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/bluestore-ec-root.yaml tasks/damage.yaml whitelist_health.yaml}
luminous
luminous
master
ovh
 
kcephfs/thrash/{clusters/fixed-3-cephfs.yaml conf.yaml objectstore-ec/filestore-xfs.yaml thrashers/mon.yaml thrashosds-health.yaml whitelist_health.yaml workloads/kclient_workunit_suites_ffsb.yaml}
Command failed (workunit test suites/ffsb.sh) on ovh032 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=luminous TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/suites/ffsb.sh'
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore-ec/bluestore-comp.yaml tasks/kclient_workunit_suites_fsync.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/bluestore.yaml tasks/data-scan.yaml whitelist_health.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore-ec/bluestore-ec-root.yaml tasks/kclient_workunit_suites_iozone.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore-ec/bluestore.yaml tasks/kclient_workunit_suites_pjd.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/filestore-xfs.yaml tasks/failover.yaml whitelist_health.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore-ec/filestore-xfs.yaml tasks/kclient_workunit_trivial_sync.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore-ec/bluestore-comp-ec-root.yaml tasks/kclient_workunit_direct_io.yaml}
luminous
luminous
master
ovh
 
kcephfs/mixed-clients/{clusters/2-clients.yaml conf.yaml objectstore-ec/bluestore-comp-ec-root.yaml tasks/kernel_cfuse_workunits_untarbuild_blogbench.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/bluestore-comp-ec-root.yaml tasks/forward-scrub.yaml whitelist_health.yaml}
"2018-03-04 04:25:28.524488 mon.a mon.0 158.69.67.1:6789/0 977 : cluster [WRN] daemon mds.c is not responding, replacing it as rank 0 with standby daemon mds.b" in cluster log
luminous
luminous
master
ovh
 
kcephfs/thrash/{clusters/fixed-3-cephfs.yaml conf.yaml objectstore-ec/bluestore-comp-ec-root.yaml thrashers/default.yaml thrashosds-health.yaml whitelist_health.yaml workloads/kclient_workunit_suites_iozone.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore-ec/bluestore-comp.yaml tasks/kclient_workunit_kernel_untar_build.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/bluestore-comp.yaml tasks/journal-repair.yaml whitelist_health.yaml}
Test failure: test_reset (tasks.cephfs.test_journal_repair.TestJournalRepair), test_reset (tasks.cephfs.test_journal_repair.TestJournalRepair)
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore-ec/bluestore-ec-root.yaml tasks/kclient_workunit_misc.yaml}
"2018-03-04 02:23:18.888273 mon.a mon.0 158.69.65.114:6789/0 159 : cluster [WRN] Health check failed: 2 osds down (OSD_DOWN)" in cluster log
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore-ec/bluestore.yaml tasks/kclient_workunit_o_trunc.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/bluestore-ec-root.yaml tasks/mds-flush.yaml whitelist_health.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore-ec/filestore-xfs.yaml tasks/kclient_workunit_snaps.yaml}
luminous
luminous
master
ovh
 
kcephfs/thrash/{clusters/fixed-3-cephfs.yaml conf.yaml objectstore-ec/bluestore-comp.yaml thrashers/mds.yaml thrashosds-health.yaml whitelist_health.yaml workloads/kclient_workunit_suites_ffsb.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/bluestore.yaml tasks/mds-full.yaml whitelist_health.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore-ec/bluestore-comp-ec-root.yaml tasks/kclient_workunit_suites_dbench.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore-ec/bluestore-comp.yaml tasks/kclient_workunit_suites_ffsb.yaml}
"2018-03-04 10:15:08.174100 mon.b mon.0 158.69.71.201:6789/0 256 : cluster [WRN] Health check failed: 1 osds down (OSD_DOWN)" in cluster log
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/filestore-xfs.yaml tasks/pool-perm.yaml whitelist_health.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore-ec/bluestore-ec-root.yaml tasks/kclient_workunit_suites_fsstress.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore-ec/bluestore.yaml tasks/kclient_workunit_suites_fsx.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/bluestore-comp-ec-root.yaml tasks/sessionmap.yaml whitelist_health.yaml}
"2018-03-04 11:01:46.836790 mon.a mon.0 158.69.74.0:6789/0 1009 : cluster [WRN] daemon mds.a is not responding, replacing it as rank 0 with standby daemon mds.b" in cluster log
luminous
luminous
master
ovh
 
kcephfs/thrash/{clusters/fixed-3-cephfs.yaml conf.yaml objectstore-ec/bluestore-ec-root.yaml thrashers/mon.yaml thrashosds-health.yaml whitelist_health.yaml workloads/kclient_workunit_suites_iozone.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore-ec/filestore-xfs.yaml tasks/kclient_workunit_suites_fsync.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/bluestore-comp.yaml tasks/strays.yaml whitelist_health.yaml}
Test failure: test_purge_queue_op_rate (tasks.cephfs.test_strays.TestStrays)
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore-ec/bluestore-comp-ec-root.yaml tasks/kclient_workunit_suites_iozone.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore-ec/bluestore-comp.yaml tasks/kclient_workunit_suites_pjd.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/bluestore-ec-root.yaml tasks/volume-client.yaml whitelist_health.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore-ec/bluestore-ec-root.yaml tasks/kclient_workunit_trivial_sync.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore-ec/bluestore.yaml tasks/kclient_workunit_direct_io.yaml}
luminous
luminous
master
ovh
 
kcephfs/mixed-clients/{clusters/2-clients.yaml conf.yaml objectstore-ec/bluestore-comp.yaml tasks/kernel_cfuse_workunits_dbench_iozone.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/bluestore.yaml tasks/auto-repair.yaml whitelist_health.yaml}
"2018-03-04 12:14:05.374944 mon.a mon.0 158.69.75.56:6789/0 655 : cluster [WRN] daemon mds.a is not responding, replacing it as rank 0 with standby daemon mds.c" in cluster log
luminous
luminous
master
ovh
 
kcephfs/thrash/{clusters/fixed-3-cephfs.yaml conf.yaml objectstore-ec/bluestore.yaml thrashers/default.yaml thrashosds-health.yaml whitelist_health.yaml workloads/kclient_workunit_suites_ffsb.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore-ec/filestore-xfs.yaml tasks/kclient_workunit_kernel_untar_build.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/filestore-xfs.yaml tasks/backtrace.yaml whitelist_health.yaml}
Command failed on ovh052 with status 1: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper kill ceph-osd -f --cluster ceph -i 3'
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore-ec/bluestore-comp-ec-root.yaml tasks/kclient_workunit_misc.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore-ec/bluestore-comp.yaml tasks/kclient_workunit_o_trunc.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/bluestore-comp-ec-root.yaml tasks/client-limits.yaml whitelist_health.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore-ec/bluestore-ec-root.yaml tasks/kclient_workunit_snaps.yaml}
luminous
luminous
master
ovh
 
kcephfs/thrash/{clusters/fixed-3-cephfs.yaml conf.yaml objectstore-ec/filestore-xfs.yaml thrashers/mds.yaml thrashosds-health.yaml whitelist_health.yaml workloads/kclient_workunit_suites_iozone.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/bluestore-comp.yaml tasks/client-recovery.yaml whitelist_health.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore-ec/bluestore.yaml tasks/kclient_workunit_suites_dbench.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore-ec/filestore-xfs.yaml tasks/kclient_workunit_suites_ffsb.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/bluestore-ec-root.yaml tasks/config-commands.yaml whitelist_health.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore-ec/bluestore-comp-ec-root.yaml tasks/kclient_workunit_suites_fsstress.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore-ec/bluestore-comp.yaml tasks/kclient_workunit_suites_fsx.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/bluestore.yaml tasks/damage.yaml whitelist_health.yaml}
luminous
luminous
master
ovh
 
kcephfs/thrash/{clusters/fixed-3-cephfs.yaml conf.yaml objectstore-ec/bluestore-comp-ec-root.yaml thrashers/mon.yaml thrashosds-health.yaml whitelist_health.yaml workloads/kclient_workunit_suites_ffsb.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore-ec/bluestore-ec-root.yaml tasks/kclient_workunit_suites_fsync.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/filestore-xfs.yaml tasks/data-scan.yaml whitelist_health.yaml}
"2018-03-04 13:17:07.732241 mon.a mon.0 158.69.77.73:6789/0 131 : cluster [WRN] Health check failed: 1 osds down (OSD_DOWN)" in cluster log
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore-ec/bluestore.yaml tasks/kclient_workunit_suites_iozone.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore-ec/filestore-xfs.yaml tasks/kclient_workunit_suites_pjd.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/bluestore-comp-ec-root.yaml tasks/failover.yaml whitelist_health.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore-ec/bluestore-comp-ec-root.yaml tasks/kclient_workunit_trivial_sync.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore-ec/bluestore-comp.yaml tasks/kclient_workunit_direct_io.yaml}
luminous
luminous
master
ovh
 
kcephfs/mixed-clients/{clusters/2-clients.yaml conf.yaml objectstore-ec/bluestore-ec-root.yaml tasks/kernel_cfuse_workunits_untarbuild_blogbench.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/bluestore-comp.yaml tasks/forward-scrub.yaml whitelist_health.yaml}
luminous
luminous
master
ovh
 
kcephfs/thrash/{clusters/fixed-3-cephfs.yaml conf.yaml objectstore-ec/bluestore-comp.yaml thrashers/default.yaml thrashosds-health.yaml whitelist_health.yaml workloads/kclient_workunit_suites_iozone.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore-ec/bluestore-ec-root.yaml tasks/kclient_workunit_kernel_untar_build.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/bluestore-ec-root.yaml tasks/journal-repair.yaml whitelist_health.yaml}
Test failure: test_reset (tasks.cephfs.test_journal_repair.TestJournalRepair), test_reset (tasks.cephfs.test_journal_repair.TestJournalRepair)
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore-ec/bluestore.yaml tasks/kclient_workunit_misc.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore-ec/filestore-xfs.yaml tasks/kclient_workunit_o_trunc.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/bluestore.yaml tasks/mds-flush.yaml whitelist_health.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore-ec/bluestore-comp-ec-root.yaml tasks/kclient_workunit_snaps.yaml}
luminous
luminous
master
ovh
 
kcephfs/thrash/{clusters/fixed-3-cephfs.yaml conf.yaml objectstore-ec/bluestore-ec-root.yaml thrashers/mds.yaml thrashosds-health.yaml whitelist_health.yaml workloads/kclient_workunit_suites_ffsb.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/filestore-xfs.yaml tasks/mds-full.yaml whitelist_health.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore-ec/bluestore-comp.yaml tasks/kclient_workunit_suites_dbench.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore-ec/bluestore-ec-root.yaml tasks/kclient_workunit_suites_ffsb.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/bluestore-comp-ec-root.yaml tasks/pool-perm.yaml whitelist_health.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore-ec/bluestore.yaml tasks/kclient_workunit_suites_fsstress.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore-ec/filestore-xfs.yaml tasks/kclient_workunit_suites_fsx.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/bluestore-comp.yaml tasks/sessionmap.yaml whitelist_health.yaml}
luminous
luminous
master
ovh
 
kcephfs/thrash/{clusters/fixed-3-cephfs.yaml conf.yaml objectstore-ec/bluestore.yaml thrashers/mon.yaml thrashosds-health.yaml whitelist_health.yaml workloads/kclient_workunit_suites_iozone.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore-ec/bluestore-comp-ec-root.yaml tasks/kclient_workunit_suites_fsync.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/bluestore-ec-root.yaml tasks/strays.yaml whitelist_health.yaml}
"2018-03-04 17:38:36.607758 mon.a mon.0 158.69.82.93:6789/0 2833 : cluster [WRN] daemon mds.c is not responding, replacing it as rank 0 with standby daemon mds.d" in cluster log
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore-ec/bluestore-comp.yaml tasks/kclient_workunit_suites_iozone.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore-ec/bluestore-ec-root.yaml tasks/kclient_workunit_suites_pjd.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/bluestore.yaml tasks/volume-client.yaml whitelist_health.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore-ec/bluestore.yaml tasks/kclient_workunit_trivial_sync.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore-ec/filestore-xfs.yaml tasks/kclient_workunit_direct_io.yaml}
luminous
luminous
master
ovh
 
kcephfs/mixed-clients/{clusters/2-clients.yaml conf.yaml objectstore-ec/bluestore.yaml tasks/kernel_cfuse_workunits_dbench_iozone.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/filestore-xfs.yaml tasks/auto-repair.yaml whitelist_health.yaml}
Command failed on ovh034 with status 1: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper kill ceph-osd -f --cluster ceph -i 2'
luminous
luminous
master
ovh
 
kcephfs/thrash/{clusters/fixed-3-cephfs.yaml conf.yaml objectstore-ec/filestore-xfs.yaml thrashers/default.yaml thrashosds-health.yaml whitelist_health.yaml workloads/kclient_workunit_suites_ffsb.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore-ec/bluestore-comp-ec-root.yaml tasks/kclient_workunit_kernel_untar_build.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/bluestore-comp-ec-root.yaml tasks/backtrace.yaml whitelist_health.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore-ec/bluestore-comp.yaml tasks/kclient_workunit_misc.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore-ec/bluestore-ec-root.yaml tasks/kclient_workunit_o_trunc.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/bluestore-comp.yaml tasks/client-limits.yaml whitelist_health.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore-ec/bluestore.yaml tasks/kclient_workunit_snaps.yaml}
luminous
luminous
master
ovh
 
kcephfs/thrash/{clusters/fixed-3-cephfs.yaml conf.yaml objectstore-ec/bluestore-comp-ec-root.yaml thrashers/mds.yaml thrashosds-health.yaml whitelist_health.yaml workloads/kclient_workunit_suites_iozone.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/bluestore-ec-root.yaml tasks/client-recovery.yaml whitelist_health.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore-ec/filestore-xfs.yaml tasks/kclient_workunit_suites_dbench.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore-ec/bluestore-comp-ec-root.yaml tasks/kclient_workunit_suites_ffsb.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/bluestore.yaml tasks/config-commands.yaml whitelist_health.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore-ec/bluestore-comp.yaml tasks/kclient_workunit_suites_fsstress.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore-ec/bluestore-ec-root.yaml tasks/kclient_workunit_suites_fsx.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/filestore-xfs.yaml tasks/damage.yaml whitelist_health.yaml}
"2018-03-04 20:14:27.132188 mon.a mon.0 158.69.87.217:6789/0 136 : cluster [WRN] Health check failed: 1 osds down (OSD_DOWN)" in cluster log
luminous
luminous
master
ovh
 
kcephfs/thrash/{clusters/fixed-3-cephfs.yaml conf.yaml objectstore-ec/bluestore-comp.yaml thrashers/mon.yaml thrashosds-health.yaml whitelist_health.yaml workloads/kclient_workunit_suites_ffsb.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore-ec/bluestore.yaml tasks/kclient_workunit_suites_fsync.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/bluestore-comp-ec-root.yaml tasks/data-scan.yaml whitelist_health.yaml}
"2018-03-04 21:14:06.174012 mon.a mon.0 158.69.88.202:6789/0 2666 : cluster [WRN] daemon mds.c is not responding, replacing it as rank 0 with standby daemon mds.d" in cluster log
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore-ec/filestore-xfs.yaml tasks/kclient_workunit_suites_iozone.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore-ec/bluestore-comp-ec-root.yaml tasks/kclient_workunit_suites_pjd.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/bluestore-comp.yaml tasks/failover.yaml whitelist_health.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore-ec/bluestore-comp.yaml tasks/kclient_workunit_trivial_sync.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore-ec/bluestore-ec-root.yaml tasks/kclient_workunit_direct_io.yaml}
luminous
luminous
master
ovh
 
kcephfs/mixed-clients/{clusters/2-clients.yaml conf.yaml objectstore-ec/filestore-xfs.yaml tasks/kernel_cfuse_workunits_untarbuild_blogbench.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/bluestore-ec-root.yaml tasks/forward-scrub.yaml whitelist_health.yaml}
luminous
luminous
master
ovh
 
kcephfs/thrash/{clusters/fixed-3-cephfs.yaml conf.yaml objectstore-ec/bluestore-ec-root.yaml thrashers/default.yaml thrashosds-health.yaml whitelist_health.yaml workloads/kclient_workunit_suites_iozone.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore-ec/bluestore.yaml tasks/kclient_workunit_kernel_untar_build.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/bluestore.yaml tasks/journal-repair.yaml whitelist_health.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore-ec/filestore-xfs.yaml tasks/kclient_workunit_misc.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore-ec/bluestore-comp-ec-root.yaml tasks/kclient_workunit_o_trunc.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/filestore-xfs.yaml tasks/mds-flush.yaml whitelist_health.yaml}
Command failed on ovh040 with status 1: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper kill ceph-osd -f --cluster ceph -i 3'
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore-ec/bluestore-comp.yaml tasks/kclient_workunit_snaps.yaml}
luminous
luminous
master
ovh
 
kcephfs/thrash/{clusters/fixed-3-cephfs.yaml conf.yaml objectstore-ec/bluestore.yaml thrashers/mds.yaml thrashosds-health.yaml whitelist_health.yaml workloads/kclient_workunit_suites_ffsb.yaml}
Found coredumps on ubuntu@ovh028.front.sepia.ceph.com
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/bluestore-comp-ec-root.yaml tasks/mds-full.yaml whitelist_health.yaml}
"2018-03-04 18:36:00.523342 mon.a mon.0 158.69.85.165:6789/0 193 : cluster [WRN] Health check failed: 1 osds down (OSD_DOWN)" in cluster log
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore-ec/bluestore-ec-root.yaml tasks/kclient_workunit_suites_dbench.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore-ec/bluestore.yaml tasks/kclient_workunit_suites_ffsb.yaml}
"2018-03-04 18:21:01.199442 mon.b mon.0 158.69.84.128:6789/0 152 : cluster [WRN] Health check failed: 1 osds down (OSD_DOWN)" in cluster log
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/bluestore-comp.yaml tasks/pool-perm.yaml whitelist_health.yaml}
"2018-03-04 19:48:24.922098 mon.a mon.0 158.69.87.117:6789/0 565 : cluster [WRN] daemon mds.c is not responding, replacing it as rank 0 with standby daemon mds.d" in cluster log
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore-ec/filestore-xfs.yaml tasks/kclient_workunit_suites_fsstress.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore-ec/bluestore-comp-ec-root.yaml tasks/kclient_workunit_suites_fsx.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/bluestore-ec-root.yaml tasks/sessionmap.yaml whitelist_health.yaml}
luminous
luminous
master
ovh
 
kcephfs/thrash/{clusters/fixed-3-cephfs.yaml conf.yaml objectstore-ec/filestore-xfs.yaml thrashers/mon.yaml thrashosds-health.yaml whitelist_health.yaml workloads/kclient_workunit_suites_iozone.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore-ec/bluestore-comp.yaml tasks/kclient_workunit_suites_fsync.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/bluestore.yaml tasks/strays.yaml whitelist_health.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore-ec/bluestore-ec-root.yaml tasks/kclient_workunit_suites_iozone.yaml}
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore-ec/bluestore.yaml tasks/kclient_workunit_suites_pjd.yaml}
luminous
luminous
master
ovh
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/filestore-xfs.yaml tasks/volume-client.yaml whitelist_health.yaml}
Command failed on ovh040 with status 1: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper kill ceph-osd -f --cluster ceph -i 3'
luminous
luminous
master
ovh
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore-ec/filestore-xfs.yaml tasks/kclient_workunit_trivial_sync.yaml}