ID
Status
Ceph Branch
Suite Branch
Teuthology Branch
Machine
OS
Nodes
Description
Failure Reason
luminous
luminous
master
smithi
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore/bluestore.yaml tasks/kclient_workunit_direct_io.yaml}
luminous
luminous
master
smithi
 
kcephfs/mixed-clients/{clusters/2-clients.yaml conf.yaml objectstore/bluestore.yaml tasks/kernel_cfuse_workunits_dbench_iozone.yaml}
luminous
luminous
master
smithi
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/bluestore.yaml tasks/auto-repair.yaml}
"2017-09-02 05:02:53.802194 mon.a mon.0 172.21.15.55:6789/0 205 : cluster [WRN] Health check failed: 1 filesystem is degraded (FS_DEGRADED)" in cluster log
luminous
luminous
master
smithi
 
kcephfs/thrash/{clusters/fixed-3-cephfs.yaml conf.yaml objectstore/bluestore.yaml thrashers/default.yaml workloads/kclient_workunit_suites_ffsb.yaml}
"2017-09-02 04:46:21.740862 mon.b mon.0 172.21.15.59:6789/0 168 : cluster [WRN] Health check failed: noscrub flag(s) set (OSDMAP_FLAGS)" in cluster log
luminous
luminous
master
smithi
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore/filestore-xfs.yaml tasks/kclient_workunit_kernel_untar_build.yaml}
luminous
luminous
master
smithi
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/filestore-xfs.yaml tasks/backtrace.yaml}
"2017-09-02 04:47:01.384280 mon.a mon.0 172.21.15.138:6789/0 197 : cluster [WRN] Health check failed: 1 filesystem is degraded (FS_DEGRADED)" in cluster log
luminous
luminous
master
smithi
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore/bluestore.yaml tasks/kclient_workunit_misc.yaml}
luminous
luminous
master
smithi
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore/filestore-xfs.yaml tasks/kclient_workunit_o_trunc.yaml}
luminous
luminous
master
smithi
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/bluestore.yaml tasks/client-limits.yaml}
"2017-09-02 05:35:53.628170 mon.a mon.0 172.21.15.169:6789/0 207 : cluster [WRN] Health check failed: 1 filesystem is degraded (FS_DEGRADED)" in cluster log
luminous
luminous
master
smithi
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore/bluestore.yaml tasks/kclient_workunit_snaps.yaml}
Command failed (workunit test fs/snaps/untar_snap_rm.sh) on smithi066 with status 124: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=luminous TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/fs/snaps/untar_snap_rm.sh'
luminous
luminous
master
smithi
 
kcephfs/thrash/{clusters/fixed-3-cephfs.yaml conf.yaml objectstore/filestore-xfs.yaml thrashers/mds.yaml workloads/kclient_workunit_suites_iozone.yaml}
"2017-09-02 04:49:23.664917 mon.b mon.0 172.21.15.38:6789/0 117 : cluster [WRN] Health check failed: 1 filesystem is degraded (FS_DEGRADED)" in cluster log
luminous
luminous
master
smithi
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/filestore-xfs.yaml tasks/client-recovery.yaml}
luminous
luminous
master
smithi
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore/filestore-xfs.yaml tasks/kclient_workunit_suites_dbench.yaml}
"2017-09-02 05:32:32.596741 mon.b mon.0 172.21.15.100:6789/0 288 : cluster [WRN] Health check failed: 1/3 mons down, quorum b,c (MON_DOWN)" in cluster log
luminous
luminous
master
smithi
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore/bluestore.yaml tasks/kclient_workunit_suites_ffsb.yaml}
"2017-09-02 04:52:46.918399 mon.b mon.0 172.21.15.137:6789/0 243 : cluster [WRN] Health check failed: 1/3 mons down, quorum b,c (MON_DOWN)" in cluster log
luminous
luminous
master
smithi
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/bluestore.yaml tasks/config-commands.yaml}
"2017-09-02 07:32:09.014760 mon.a mon.0 172.21.15.134:6789/0 208 : cluster [WRN] Health check failed: 1 filesystem is degraded (FS_DEGRADED)" in cluster log
luminous
luminous
master
smithi
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore/filestore-xfs.yaml tasks/kclient_workunit_suites_fsstress.yaml}
luminous
luminous
master
smithi
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore/bluestore.yaml tasks/kclient_workunit_suites_fsx.yaml}
luminous
luminous
master
smithi
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/filestore-xfs.yaml tasks/damage.yaml}
"2017-09-02 06:40:04.334766 mon.a mon.0 172.21.15.176:6789/0 381 : cluster [ERR] Health check failed: 1 MDSs report damaged metadata (MDS_DAMAGE)" in cluster log
luminous
luminous
master
smithi
 
kcephfs/thrash/{clusters/fixed-3-cephfs.yaml conf.yaml objectstore/bluestore.yaml thrashers/mon.yaml workloads/kclient_workunit_suites_ffsb.yaml}
"2017-09-02 04:47:46.597559 mon.b mon.1 172.21.15.193:6789/0 92 : cluster [WRN] Health check failed: 1/3 mons down, quorum b,c (MON_DOWN)" in cluster log
luminous
luminous
master
smithi
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore/filestore-xfs.yaml tasks/kclient_workunit_suites_fsync.yaml}
luminous
luminous
master
smithi
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/bluestore.yaml tasks/data-scan.yaml}
Test failure: test_rebuild_simple_altpool (tasks.cephfs.test_data_scan.TestDataScan)
luminous
luminous
master
smithi
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore/bluestore.yaml tasks/kclient_workunit_suites_iozone.yaml}
luminous
luminous
master
smithi
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore/filestore-xfs.yaml tasks/kclient_workunit_suites_pjd.yaml}
luminous
luminous
master
smithi
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/filestore-xfs.yaml tasks/failover.yaml}
"2017-09-02 09:59:52.393855 mon.a mon.0 172.21.15.158:6789/0 195 : cluster [WRN] Health check failed: 1 filesystem is degraded (FS_DEGRADED)" in cluster log
luminous
luminous
master
smithi
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore/bluestore.yaml tasks/kclient_workunit_trivial_sync.yaml}
luminous
luminous
master
smithi
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore/filestore-xfs.yaml tasks/kclient_workunit_direct_io.yaml}
luminous
luminous
master
smithi
 
kcephfs/mixed-clients/{clusters/2-clients.yaml conf.yaml objectstore/filestore-xfs.yaml tasks/kernel_cfuse_workunits_untarbuild_blogbench.yaml}
"2017-09-02 04:59:12.310616 mon.b mon.0 172.21.15.18:6789/0 147 : cluster [WRN] overall HEALTH_WARN 1/3 mons down, quorum b,c" in cluster log
luminous
luminous
master
smithi
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/bluestore.yaml tasks/forward-scrub.yaml}
"2017-09-02 05:03:23.310544 mon.a mon.0 172.21.15.21:6789/0 465 : cluster [ERR] Health check failed: 1 MDSs report damaged metadata (MDS_DAMAGE)" in cluster log
luminous
luminous
master
smithi
 
kcephfs/thrash/{clusters/fixed-3-cephfs.yaml conf.yaml objectstore/filestore-xfs.yaml thrashers/default.yaml workloads/kclient_workunit_suites_iozone.yaml}
"2017-09-02 05:02:46.015141 mon.a mon.0 172.21.15.50:6789/0 192 : cluster [WRN] Health check failed: noscrub flag(s) set (OSDMAP_FLAGS)" in cluster log
luminous
luminous
master
smithi
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore/bluestore.yaml tasks/kclient_workunit_kernel_untar_build.yaml}
luminous
luminous
master
smithi
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/filestore-xfs.yaml tasks/journal-repair.yaml}
"2017-09-02 04:58:41.311982 mon.a mon.0 172.21.15.37:6789/0 386 : cluster [ERR] Health check failed: 1 MDSs report damaged metadata (MDS_DAMAGE)" in cluster log
luminous
luminous
master
smithi
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore/filestore-xfs.yaml tasks/kclient_workunit_misc.yaml}
luminous
luminous
master
smithi
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore/bluestore.yaml tasks/kclient_workunit_o_trunc.yaml}
luminous
luminous
master
smithi
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/bluestore.yaml tasks/mds-flush.yaml}
"2017-09-02 06:09:18.801094 mon.a mon.0 172.21.15.39:6789/0 191 : cluster [WRN] Health check failed: 1 filesystem is degraded (FS_DEGRADED)" in cluster log
luminous
luminous
master
smithi
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore/filestore-xfs.yaml tasks/kclient_workunit_snaps.yaml}
"2017-09-02 07:26:55.102847 mon.a mon.0 172.21.15.2:6789/0 354 : cluster [WRN] Health check failed: 1 osds down (OSD_DOWN)" in cluster log
luminous
luminous
master
smithi
 
kcephfs/thrash/{clusters/fixed-3-cephfs.yaml conf.yaml objectstore/bluestore.yaml thrashers/mds.yaml workloads/kclient_workunit_suites_ffsb.yaml}
"2017-09-02 05:13:01.122206 mon.b mon.0 172.21.15.51:6789/0 186 : cluster [WRN] Health check failed: 1 filesystem is degraded (FS_DEGRADED)" in cluster log
luminous
luminous
master
smithi
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/filestore-xfs.yaml tasks/mds-full.yaml}
Command failed on smithi182 with status 1: 'sudo cp /var/lib/ceph/osd/ceph-0/fsid /tmp/tmpEYJaec'
luminous
luminous
master
smithi
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore/bluestore.yaml tasks/kclient_workunit_suites_dbench.yaml}
luminous
luminous
master
smithi
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore/filestore-xfs.yaml tasks/kclient_workunit_suites_ffsb.yaml}
luminous
luminous
master
smithi
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/bluestore.yaml tasks/pool-perm.yaml}
"2017-09-02 08:21:33.033139 mon.a mon.0 172.21.15.82:6789/0 206 : cluster [WRN] Health check failed: 1 filesystem is degraded (FS_DEGRADED)" in cluster log
luminous
luminous
master
smithi
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore/bluestore.yaml tasks/kclient_workunit_suites_fsstress.yaml}
luminous
luminous
master
smithi
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore/filestore-xfs.yaml tasks/kclient_workunit_suites_fsx.yaml}
luminous
luminous
master
smithi
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/filestore-xfs.yaml tasks/sessionmap.yaml}
"2017-09-02 06:41:53.291321 mon.a mon.0 172.21.15.92:6789/0 215 : cluster [WRN] Health check failed: 1 filesystem is degraded (FS_DEGRADED)" in cluster log
luminous
luminous
master
smithi
 
kcephfs/thrash/{clusters/fixed-3-cephfs.yaml conf.yaml objectstore/filestore-xfs.yaml thrashers/mon.yaml workloads/kclient_workunit_suites_iozone.yaml}
"2017-09-02 05:07:01.633043 mon.b mon.0 172.21.15.63:6789/0 113 : cluster [WRN] Health check failed: 1/3 mons down, quorum b,a (MON_DOWN)" in cluster log
luminous
luminous
master
smithi
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore/bluestore.yaml tasks/kclient_workunit_suites_fsync.yaml}
luminous
luminous
master
smithi
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/bluestore.yaml tasks/strays.yaml}
Test failure: test_files_throttle (tasks.cephfs.test_strays.TestStrays)
luminous
luminous
master
smithi
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore/filestore-xfs.yaml tasks/kclient_workunit_suites_iozone.yaml}
luminous
luminous
master
smithi
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore/bluestore.yaml tasks/kclient_workunit_suites_pjd.yaml}
luminous
luminous
master
smithi
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/filestore-xfs.yaml tasks/volume-client.yaml}
"2017-09-02 06:39:50.838902 mon.a mon.0 172.21.15.179:6789/0 206 : cluster [WRN] Health check failed: 1 filesystem is degraded (FS_DEGRADED)" in cluster log
luminous
luminous
master
smithi
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore/filestore-xfs.yaml tasks/kclient_workunit_trivial_sync.yaml}
luminous
luminous
master
smithi
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore/bluestore.yaml tasks/kclient_workunit_direct_io.yaml}
luminous
luminous
master
smithi
 
kcephfs/mixed-clients/{clusters/2-clients.yaml conf.yaml objectstore/filestore-xfs.yaml tasks/kernel_cfuse_workunits_dbench_iozone.yaml}
"2017-09-02 06:05:15.160059 mon.b mon.0 172.21.15.9:6789/0 171 : cluster [WRN] overall HEALTH_WARN 1/3 mons down, quorum b,c" in cluster log
luminous
luminous
master
smithi
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/filestore-xfs.yaml tasks/auto-repair.yaml}
"2017-09-02 09:00:05.926144 mon.a mon.0 172.21.15.26:6789/0 197 : cluster [WRN] Health check failed: 1 filesystem is degraded (FS_DEGRADED)" in cluster log
luminous
luminous
master
smithi
 
kcephfs/thrash/{clusters/fixed-3-cephfs.yaml conf.yaml objectstore/filestore-xfs.yaml thrashers/default.yaml workloads/kclient_workunit_suites_ffsb.yaml}
"2017-09-02 05:15:56.746112 mon.a mon.0 172.21.15.1:6789/0 187 : cluster [WRN] Health check failed: noscrub flag(s) set (OSDMAP_FLAGS)" in cluster log
luminous
luminous
master
smithi
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore/filestore-xfs.yaml tasks/kclient_workunit_kernel_untar_build.yaml}
luminous
luminous
master
smithi
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/bluestore.yaml tasks/backtrace.yaml}
"2017-09-02 05:43:41.845121 mon.a mon.0 172.21.15.6:6789/0 221 : cluster [WRN] Health check failed: 1 filesystem is degraded (FS_DEGRADED)" in cluster log
luminous
luminous
master
smithi
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore/bluestore.yaml tasks/kclient_workunit_misc.yaml}
luminous
luminous
master
smithi
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore/filestore-xfs.yaml tasks/kclient_workunit_o_trunc.yaml}
luminous
luminous
master
smithi
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/filestore-xfs.yaml tasks/client-limits.yaml}
"2017-09-02 05:37:26.925566 mon.a mon.0 172.21.15.53:6789/0 205 : cluster [WRN] Health check failed: 1 filesystem is degraded (FS_DEGRADED)" in cluster log
luminous
luminous
master
smithi
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore/bluestore.yaml tasks/kclient_workunit_snaps.yaml}
Command failed (workunit test fs/snaps/untar_snap_rm.sh) on smithi071 with status 124: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=luminous TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/fs/snaps/untar_snap_rm.sh'
luminous
luminous
master
smithi
 
kcephfs/thrash/{clusters/fixed-3-cephfs.yaml conf.yaml objectstore/bluestore.yaml thrashers/mds.yaml workloads/kclient_workunit_suites_iozone.yaml}
"2017-09-02 05:23:21.311889 mon.a mon.0 172.21.15.38:6789/0 189 : cluster [WRN] Health check failed: 1 filesystem is degraded (FS_DEGRADED)" in cluster log
luminous
luminous
master
smithi
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/bluestore.yaml tasks/client-recovery.yaml}
luminous
luminous
master
smithi
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore/filestore-xfs.yaml tasks/kclient_workunit_suites_dbench.yaml}
"2017-09-02 05:48:22.606848 mon.b mon.1 172.21.15.131:6789/0 108 : cluster [WRN] Health check failed: 1/3 mons down, quorum b,c (MON_DOWN)" in cluster log
luminous
luminous
master
smithi
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore/bluestore.yaml tasks/kclient_workunit_suites_ffsb.yaml}
"2017-09-02 06:11:40.822839 mon.b mon.0 172.21.15.59:6789/0 161 : cluster [WRN] Health check failed: 1/3 mons down, quorum b,c (MON_DOWN)" in cluster log
luminous
luminous
master
smithi
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/filestore-xfs.yaml tasks/config-commands.yaml}
"2017-09-02 07:37:18.005081 mon.a mon.0 172.21.15.28:6789/0 198 : cluster [WRN] Health check failed: 1 filesystem is degraded (FS_DEGRADED)" in cluster log
luminous
luminous
master
smithi
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore/filestore-xfs.yaml tasks/kclient_workunit_suites_fsstress.yaml}
luminous
luminous
master
smithi
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore/bluestore.yaml tasks/kclient_workunit_suites_fsx.yaml}
luminous
luminous
master
smithi
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/bluestore.yaml tasks/damage.yaml}
"2017-09-02 05:48:52.490952 mon.a mon.0 172.21.15.203:6789/0 395 : cluster [ERR] Health check failed: 1 MDSs report damaged metadata (MDS_DAMAGE)" in cluster log
luminous
luminous
master
smithi
 
kcephfs/thrash/{clusters/fixed-3-cephfs.yaml conf.yaml objectstore/filestore-xfs.yaml thrashers/mon.yaml workloads/kclient_workunit_suites_ffsb.yaml}
"2017-09-02 06:02:14.937716 mon.a mon.0 172.21.15.37:6789/0 189 : cluster [WRN] Health check failed: 1/3 mons down, quorum a,c (MON_DOWN)" in cluster log
luminous
luminous
master
smithi
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore/filestore-xfs.yaml tasks/kclient_workunit_suites_fsync.yaml}
luminous
luminous
master
smithi
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/filestore-xfs.yaml tasks/data-scan.yaml}
Test failure: test_rebuild_simple_altpool (tasks.cephfs.test_data_scan.TestDataScan)
luminous
luminous
master
smithi
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore/bluestore.yaml tasks/kclient_workunit_suites_iozone.yaml}
luminous
luminous
master
smithi
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore/filestore-xfs.yaml tasks/kclient_workunit_suites_pjd.yaml}
luminous
luminous
master
smithi
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/bluestore.yaml tasks/failover.yaml}
"2017-09-02 05:37:26.248671 mon.a mon.0 172.21.15.114:6789/0 215 : cluster [WRN] Health check failed: 1 filesystem is degraded (FS_DEGRADED)" in cluster log
luminous
luminous
master
smithi
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore/bluestore.yaml tasks/kclient_workunit_trivial_sync.yaml}
luminous
luminous
master
smithi
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore/filestore-xfs.yaml tasks/kclient_workunit_direct_io.yaml}
luminous
luminous
master
smithi
 
kcephfs/mixed-clients/{clusters/2-clients.yaml conf.yaml objectstore/bluestore.yaml tasks/kernel_cfuse_workunits_untarbuild_blogbench.yaml}
"2017-09-02 06:40:14.819093 mon.b mon.0 172.21.15.1:6789/0 183 : cluster [WRN] overall HEALTH_WARN 1/3 mons down, quorum b,c" in cluster log
luminous
luminous
master
smithi
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/filestore-xfs.yaml tasks/forward-scrub.yaml}
"2017-09-02 08:31:37.025133 mon.a mon.0 172.21.15.67:6789/0 448 : cluster [ERR] Health check failed: 1 MDSs report damaged metadata (MDS_DAMAGE)" in cluster log
luminous
luminous
master
smithi
 
kcephfs/thrash/{clusters/fixed-3-cephfs.yaml conf.yaml objectstore/bluestore.yaml thrashers/default.yaml workloads/kclient_workunit_suites_iozone.yaml}
"2017-09-02 05:40:06.160090 mon.b mon.0 172.21.15.132:6789/0 174 : cluster [WRN] Health check failed: noscrub flag(s) set (OSDMAP_FLAGS)" in cluster log
luminous
luminous
master
smithi
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore/bluestore.yaml tasks/kclient_workunit_kernel_untar_build.yaml}
luminous
luminous
master
smithi
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/bluestore.yaml tasks/journal-repair.yaml}
"2017-09-02 05:42:50.186033 mon.a mon.0 172.21.15.202:6789/0 405 : cluster [ERR] Health check failed: 1 MDSs report damaged metadata (MDS_DAMAGE)" in cluster log
luminous
luminous
master
smithi
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore/filestore-xfs.yaml tasks/kclient_workunit_misc.yaml}
"2017-09-02 05:44:54.506431 mon.a mon.0 172.21.15.21:6789/0 216 : cluster [WRN] daemon mds.a-s is not responding, replacing it as rank 0 with standby daemon mds.a" in cluster log
luminous
luminous
master
smithi
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore/bluestore.yaml tasks/kclient_workunit_o_trunc.yaml}
luminous
luminous
master
smithi
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/filestore-xfs.yaml tasks/mds-flush.yaml}
"2017-09-02 07:48:18.962230 mon.a mon.0 172.21.15.130:6789/0 221 : cluster [WRN] Health check failed: 1 filesystem is degraded (FS_DEGRADED)" in cluster log
luminous
luminous
master
smithi
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore/filestore-xfs.yaml tasks/kclient_workunit_snaps.yaml}
Command failed (workunit test fs/snaps/untar_snap_rm.sh) on smithi155 with status 124: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=luminous TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/fs/snaps/untar_snap_rm.sh'
luminous
luminous
master
smithi
 
kcephfs/thrash/{clusters/fixed-3-cephfs.yaml conf.yaml objectstore/filestore-xfs.yaml thrashers/mds.yaml workloads/kclient_workunit_suites_ffsb.yaml}
"2017-09-02 05:39:24.153281 mon.b mon.0 172.21.15.95:6789/0 169 : cluster [WRN] Health check failed: 1 filesystem is degraded (FS_DEGRADED)" in cluster log
luminous
luminous
master
smithi
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/bluestore.yaml tasks/mds-full.yaml}
Command failed on smithi156 with status 1: 'sudo cp /var/lib/ceph/osd/ceph-1/fsid /tmp/tmpboOgWP'
luminous
luminous
master
smithi
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore/bluestore.yaml tasks/kclient_workunit_suites_dbench.yaml}
"2017-09-02 05:51:44.359796 mon.b mon.0 172.21.15.85:6789/0 243 : cluster [WRN] overall HEALTH_WARN 1/3 mons down, quorum b,c" in cluster log
luminous
luminous
master
smithi
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore/filestore-xfs.yaml tasks/kclient_workunit_suites_ffsb.yaml}
SELinux denials found on ubuntu@smithi003.front.sepia.ceph.com: ['type=AVC msg=audit(1504331484.463:4380): avc: denied { getattr } for pid=22751 comm="ceph-osd" path="/dev/nvme0n1p2" dev="devtmpfs" ino=58265 scontext=system_u:system_r:ceph_t:s0 tcontext=system_u:object_r:nvme_device_t:s0 tclass=blk_file permissive=1']
luminous
luminous
master
smithi
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/filestore-xfs.yaml tasks/pool-perm.yaml}
"2017-09-02 05:44:16.799595 mon.a mon.0 172.21.15.82:6789/0 198 : cluster [WRN] Health check failed: 1 filesystem is degraded (FS_DEGRADED)" in cluster log
luminous
luminous
master
smithi
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore/bluestore.yaml tasks/kclient_workunit_suites_fsstress.yaml}
luminous
luminous
master
smithi
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore/filestore-xfs.yaml tasks/kclient_workunit_suites_fsx.yaml}
luminous
luminous
master
smithi
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/bluestore.yaml tasks/sessionmap.yaml}
"2017-09-02 09:09:09.061145 mon.a mon.0 172.21.15.12:6789/0 218 : cluster [WRN] Health check failed: 1 filesystem is degraded (FS_DEGRADED)" in cluster log
luminous
luminous
master
smithi
 
kcephfs/thrash/{clusters/fixed-3-cephfs.yaml conf.yaml objectstore/bluestore.yaml thrashers/mon.yaml workloads/kclient_workunit_suites_iozone.yaml}
"2017-09-02 05:53:08.978933 mon.a mon.1 172.21.15.138:6789/0 49 : cluster [WRN] Health check failed: 1/3 mons down, quorum a,c (MON_DOWN)" in cluster log
luminous
luminous
master
smithi
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore/bluestore.yaml tasks/kclient_workunit_suites_fsync.yaml}
luminous
luminous
master
smithi
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/filestore-xfs.yaml tasks/strays.yaml}
Test failure: test_ops_throttle (tasks.cephfs.test_strays.TestStrays)
luminous
luminous
master
smithi
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore/filestore-xfs.yaml tasks/kclient_workunit_suites_iozone.yaml}
luminous
luminous
master
smithi
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore/bluestore.yaml tasks/kclient_workunit_suites_pjd.yaml}
luminous
luminous
master
smithi
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/bluestore.yaml tasks/volume-client.yaml}
Command failed on smithi079 with status 100: u'sudo DEBIAN_FRONTEND=noninteractive apt-get -y --force-yes -o Dpkg::Options::="--force-confdef" -o Dpkg::Options::="--force-confold" install ceph=12.2.0-22-gfa99676-1trusty ceph-mds=12.2.0-22-gfa99676-1trusty ceph-mgr=12.2.0-22-gfa99676-1trusty ceph-common=12.2.0-22-gfa99676-1trusty ceph-fuse=12.2.0-22-gfa99676-1trusty ceph-test=12.2.0-22-gfa99676-1trusty radosgw=12.2.0-22-gfa99676-1trusty python-ceph=12.2.0-22-gfa99676-1trusty libcephfs2=12.2.0-22-gfa99676-1trusty libcephfs-dev=12.2.0-22-gfa99676-1trusty libcephfs-java=12.2.0-22-gfa99676-1trusty libcephfs-jni=12.2.0-22-gfa99676-1trusty librados2=12.2.0-22-gfa99676-1trusty librbd1=12.2.0-22-gfa99676-1trusty rbd-fuse=12.2.0-22-gfa99676-1trusty'
luminous
luminous
master
smithi
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore/filestore-xfs.yaml tasks/kclient_workunit_trivial_sync.yaml}