ID
Status
Ceph Branch
Suite Branch
Teuthology Branch
Machine
OS
Nodes
Description
Failure Reason
luminous
luminous
master
smithi
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/bluestore.yaml tasks/auto-repair.yaml whitelist_health.yaml}
"2017-09-16 22:49:36.901032 mon.a mon.0 172.21.15.158:6789/0 517 : cluster [WRN] Health check failed: 1 MDSs are read only (MDS_READ_ONLY)" in cluster log
luminous
luminous
master
smithi
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/bluestore.yaml tasks/client-limits.yaml whitelist_health.yaml}
"2017-09-16 22:50:30.597410 mon.a mon.0 172.21.15.188:6789/0 619 : cluster [WRN] MDS health message (mds.0): MDS cache is too large (521kB/1GB); 202 inodes in use by clients, 0 stray files" in cluster log
luminous
luminous
master
smithi
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/filestore-xfs.yaml tasks/client-recovery.yaml whitelist_health.yaml}
luminous
luminous
master
smithi
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/bluestore.yaml tasks/data-scan.yaml whitelist_health.yaml}
Test failure: test_rebuild_simple_altpool (tasks.cephfs.test_data_scan.TestDataScan)
luminous
luminous
master
smithi
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/filestore-xfs.yaml tasks/failover.yaml whitelist_health.yaml}
"2017-09-16 22:50:23.730379 mon.a mon.0 172.21.15.169:6789/0 460 : cluster [WRN] daemon mds.b is not responding, replacing it as rank 0 with standby daemon mds.d" in cluster log
luminous
luminous
master
smithi
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore/bluestore.yaml tasks/kclient_workunit_kernel_untar_build.yaml}
luminous
luminous
master
smithi
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/filestore-xfs.yaml tasks/mds-full.yaml whitelist_health.yaml}
Command failed on smithi073 with status 1: 'sudo cp /var/lib/ceph/osd/ceph-0/fsid /tmp/tmp75M0gm'
luminous
luminous
master
smithi
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/bluestore.yaml tasks/strays.yaml whitelist_health.yaml}
Test failure: test_migration_on_shutdown (tasks.cephfs.test_strays.TestStrays)
luminous
luminous
master
smithi
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/filestore-xfs.yaml tasks/auto-repair.yaml whitelist_health.yaml}
"2017-09-16 22:56:24.070876 mon.a mon.0 172.21.15.56:6789/0 511 : cluster [WRN] Health check failed: 1 MDSs are read only (MDS_READ_ONLY)" in cluster log
luminous
luminous
master
smithi
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/filestore-xfs.yaml tasks/client-limits.yaml whitelist_health.yaml}
"2017-09-16 22:59:42.852012 mon.a mon.0 172.21.15.95:6789/0 643 : cluster [WRN] MDS health message (mds.0): MDS cache is too large (521kB/1GB); 202 inodes in use by clients, 0 stray files" in cluster log
luminous
luminous
master
smithi
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/bluestore.yaml tasks/client-recovery.yaml whitelist_health.yaml}
"2017-09-16 23:09:38.811653 mds.d mds.0 172.21.15.84:6810/1429441865 1 : cluster [WRN] evicting unresponsive client smithi156: (4323), after waiting 45 seconds during MDS startup" in cluster log
luminous
luminous
master
smithi
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore/bluestore.yaml tasks/kclient_workunit_suites_ffsb.yaml}
luminous
luminous
master
smithi
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/filestore-xfs.yaml tasks/data-scan.yaml whitelist_health.yaml}
Test failure: test_rebuild_simple_altpool (tasks.cephfs.test_data_scan.TestDataScan)
luminous
luminous
master
smithi
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/bluestore.yaml tasks/failover.yaml whitelist_health.yaml}
"2017-09-16 23:01:45.688492 mon.a mon.0 172.21.15.100:6789/0 465 : cluster [WRN] daemon mds.b is not responding, replacing it as rank 0 with standby daemon mds.d" in cluster log
luminous
luminous
master
smithi
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore/filestore-xfs.yaml tasks/kclient_workunit_misc.yaml}
luminous
luminous
master
smithi
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/bluestore.yaml tasks/mds-full.yaml whitelist_health.yaml}
Command failed on smithi099 with status 1: 'sudo cp /var/lib/ceph/osd/ceph-0/fsid /tmp/tmpKdEKSW'
luminous
luminous
master
smithi
 
kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore/bluestore.yaml tasks/kclient_workunit_suites_dbench.yaml}
luminous
luminous
master
smithi
 
kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/filestore-xfs.yaml tasks/strays.yaml whitelist_health.yaml}
Test failure: test_replicated_delete_speed (tasks.cephfs.test_strays.TestStrays)