Status Job ID Links Posted Started Updated
Runtime
Duration
In Waiting
Machine Teuthology Branch OS Type OS Version Description Nodes
pass 1113128 2017-05-08 03:25:27 2017-05-08 13:22:42 2017-05-08 13:40:42 0:18:00 0:10:37 0:07:23 smithi master kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore/bluestore.yaml tasks/kclient_workunit_direct_io.yaml} 3
pass 1113131 2017-05-08 03:25:27 2017-05-08 13:23:06 2017-05-08 14:17:06 0:54:00 0:47:13 0:06:47 smithi master kcephfs/mixed-clients/{clusters/2-clients.yaml conf.yaml objectstore/bluestore.yaml tasks/kernel_cfuse_workunits_dbench_iozone.yaml} 4
fail 1113134 2017-05-08 03:25:28 2017-05-08 13:23:06 2017-05-08 14:35:07 1:12:01 0:13:31 0:58:30 smithi master kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/bluestore.yaml tasks/auto-repair.yaml} 4
Failure Reason:

"2017-05-08 14:29:34.393891 mon.0 172.21.15.9:6789/0 460 : cluster [WRN] MDS health message (mds.0): MDS in read-only mode" in cluster log

pass 1113137 2017-05-08 03:25:29 2017-05-08 13:23:13 2017-05-08 14:09:13 0:46:00 0:37:24 0:08:36 smithi master kcephfs/thrash/{clusters/fixed-3-cephfs.yaml conf.yaml objectstore/bluestore.yaml thrashers/default.yaml workloads/kclient_workunit_suites_ffsb.yaml} 3
pass 1113140 2017-05-08 03:25:29 2017-05-08 13:24:29 2017-05-08 14:00:29 0:36:00 0:22:24 0:13:36 smithi master kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore/filestore-xfs.yaml tasks/kclient_workunit_kernel_untar_build.yaml} 3
pass 1113143 2017-05-08 03:25:30 2017-05-08 13:24:33 2017-05-08 13:38:34 0:14:01 0:08:53 0:05:08 smithi master kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/filestore-xfs.yaml tasks/backtrace.yaml} 4
pass 1113146 2017-05-08 03:25:31 2017-05-08 13:25:04 2017-05-08 14:05:04 0:40:00 0:29:22 0:10:38 smithi master kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore/bluestore.yaml tasks/kclient_workunit_misc.yaml} 3
pass 1113149 2017-05-08 03:25:31 2017-05-08 13:26:45 2017-05-08 13:52:45 0:26:00 0:17:42 0:08:18 smithi master kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore/filestore-xfs.yaml tasks/kclient_workunit_o_trunc.yaml} 3
fail 1113152 2017-05-08 03:25:32 2017-05-08 13:26:47 2017-05-08 13:58:47 0:32:00 0:15:51 0:16:09 smithi master kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/bluestore.yaml tasks/client-limits.yaml} 4
Failure Reason:

"2017-05-08 13:48:56.521192 mon.0 172.21.15.91:6789/0 601 : cluster [WRN] MDS health message (mds.0): Too many inodes in cache (163/100), 153 inodes in use by clients, 0 stray files" in cluster log

pass 1113154 2017-05-08 03:25:33 2017-05-08 13:26:55 2017-05-08 14:18:55 0:52:00 0:44:03 0:07:57 smithi master kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore/bluestore.yaml tasks/kclient_workunit_snaps.yaml} 3
pass 1113157 2017-05-08 03:25:33 2017-05-08 13:26:58 2017-05-08 13:54:58 0:28:00 0:22:02 0:05:58 smithi master kcephfs/thrash/{clusters/fixed-3-cephfs.yaml conf.yaml objectstore/filestore-xfs.yaml thrashers/mds.yaml workloads/kclient_workunit_suites_iozone.yaml} 3
pass 1113159 2017-05-08 03:25:34 2017-05-08 13:28:51 2017-05-08 14:06:51 0:38:00 0:33:38 0:04:22 smithi master kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/filestore-xfs.yaml tasks/client-recovery.yaml} 4
pass 1113162 2017-05-08 03:25:35 2017-05-08 13:31:18 2017-05-08 14:17:18 0:46:00 0:38:20 0:07:40 smithi master kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore/filestore-xfs.yaml tasks/kclient_workunit_suites_dbench.yaml} 3
pass 1113165 2017-05-08 03:25:35 2017-05-08 13:31:18 2017-05-08 14:13:18 0:42:00 0:32:51 0:09:09 smithi master kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore/bluestore.yaml tasks/kclient_workunit_suites_ffsb.yaml} 3
pass 1113168 2017-05-08 03:25:36 2017-05-08 13:31:18 2017-05-08 13:59:18 0:28:00 0:11:42 0:16:18 smithi master kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/bluestore.yaml tasks/config-commands.yaml} 4
pass 1113171 2017-05-08 03:25:37 2017-05-08 13:31:18 2017-05-08 13:47:18 0:16:00 0:07:48 0:08:12 smithi master kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore/filestore-xfs.yaml tasks/kclient_workunit_suites_fsstress.yaml} 3
pass 1113174 2017-05-08 03:25:37 2017-05-08 13:32:51 2017-05-08 14:00:50 0:27:59 0:21:09 0:06:50 smithi master kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore/bluestore.yaml tasks/kclient_workunit_suites_fsx.yaml} 3
fail 1113177 2017-05-08 03:25:38 2017-05-08 13:33:02 2017-05-08 14:01:02 0:28:00 0:18:18 0:09:42 smithi master kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/filestore-xfs.yaml tasks/damage.yaml} 4
Failure Reason:

"2017-05-08 13:43:54.886051 mon.0 172.21.15.105:6789/0 347 : cluster [ERR] MDS health message (mds.0): Metadata damage detected" in cluster log

pass 1113180 2017-05-08 03:25:38 2017-05-08 13:33:04 2017-05-08 14:11:04 0:38:00 0:32:32 0:05:28 smithi master kcephfs/thrash/{clusters/fixed-3-cephfs.yaml conf.yaml objectstore/bluestore.yaml thrashers/mon.yaml workloads/kclient_workunit_suites_ffsb.yaml} 3
pass 1113181 2017-05-08 03:25:39 2017-05-08 13:34:42 2017-05-08 14:04:41 0:29:59 0:09:12 0:20:47 smithi master kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore/filestore-xfs.yaml tasks/kclient_workunit_suites_fsync.yaml} 3
fail 1113182 2017-05-08 03:25:40 2017-05-08 13:34:42 2017-05-08 14:44:44 1:10:02 0:28:20 0:41:42 smithi master kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/bluestore.yaml tasks/data-scan.yaml} 4
Failure Reason:

Test failure: test_rebuild_simple_altpool (tasks.cephfs.test_data_scan.TestDataScan)

pass 1113183 2017-05-08 03:25:40 2017-05-08 13:34:42 2017-05-08 13:54:41 0:19:59 0:12:17 0:07:42 smithi master kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore/bluestore.yaml tasks/kclient_workunit_suites_iozone.yaml} 3
pass 1113184 2017-05-08 03:25:41 2017-05-08 13:34:42 2017-05-08 13:52:41 0:17:59 0:11:04 0:06:55 smithi master kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore/filestore-xfs.yaml tasks/kclient_workunit_suites_pjd.yaml} 3
pass 1113185 2017-05-08 03:25:42 2017-05-08 13:34:42 2017-05-08 14:36:42 1:02:00 0:17:51 0:44:09 smithi master kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/filestore-xfs.yaml tasks/failover.yaml} 4
pass 1113186 2017-05-08 03:25:42 2017-05-08 13:34:42 2017-05-08 14:02:41 0:27:59 0:09:38 0:18:21 smithi master kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore/bluestore.yaml tasks/kclient_workunit_trivial_sync.yaml} 3
pass 1113187 2017-05-08 03:25:43 2017-05-08 13:34:42 2017-05-08 13:56:41 0:21:59 0:10:13 0:11:46 smithi master kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore/filestore-xfs.yaml tasks/kclient_workunit_direct_io.yaml} 3
pass 1113188 2017-05-08 03:25:43 2017-05-08 13:34:42 2017-05-08 14:14:42 0:40:00 0:31:28 0:08:32 smithi master kcephfs/mixed-clients/{clusters/2-clients.yaml conf.yaml objectstore/filestore-xfs.yaml tasks/kernel_cfuse_workunits_untarbuild_blogbench.yaml} 4
fail 1113189 2017-05-08 03:25:44 2017-05-08 13:34:42 2017-05-08 14:26:43 0:52:01 0:14:14 0:37:47 smithi master kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/bluestore.yaml tasks/forward-scrub.yaml} 4
Failure Reason:

"2017-05-08 14:17:17.792743 mon.0 172.21.15.97:6789/0 417 : cluster [ERR] MDS health message (mds.0): Metadata damage detected" in cluster log

pass 1113190 2017-05-08 03:25:45 2017-05-08 13:34:44 2017-05-08 14:08:44 0:34:00 0:25:49 0:08:11 smithi master kcephfs/thrash/{clusters/fixed-3-cephfs.yaml conf.yaml objectstore/filestore-xfs.yaml thrashers/default.yaml workloads/kclient_workunit_suites_iozone.yaml} 3
pass 1113191 2017-05-08 03:25:45 2017-05-08 13:34:46 2017-05-08 14:02:46 0:28:00 0:22:53 0:05:07 smithi master kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore/bluestore.yaml tasks/kclient_workunit_kernel_untar_build.yaml} 3
fail 1113192 2017-05-08 03:25:47 2017-05-08 13:34:49 2017-05-08 15:06:51 1:32:02 0:24:28 1:07:34 smithi master kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/filestore-xfs.yaml tasks/journal-repair.yaml} 4
Failure Reason:

"2017-05-08 14:46:34.026608 mon.0 172.21.15.73:6789/0 359 : cluster [ERR] MDS health message (mds.0): Metadata damage detected" in cluster log

pass 1113193 2017-05-08 03:25:47 2017-05-08 13:34:58 2017-05-08 14:22:58 0:48:00 0:33:48 0:14:12 smithi master kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore/filestore-xfs.yaml tasks/kclient_workunit_misc.yaml} 3
pass 1113194 2017-05-08 03:25:48 2017-05-08 13:34:58 2017-05-08 14:08:58 0:34:00 0:19:47 0:14:13 smithi master kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore/bluestore.yaml tasks/kclient_workunit_o_trunc.yaml} 3
pass 1113195 2017-05-08 03:25:49 2017-05-08 13:34:59 2017-05-08 15:39:01 2:04:02 0:08:58 1:55:04 smithi master kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/bluestore.yaml tasks/mds-flush.yaml} 4
pass 1113196 2017-05-08 03:25:49 2017-05-08 13:35:01 2017-05-08 14:17:01 0:42:00 0:36:50 0:05:10 smithi master kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore/filestore-xfs.yaml tasks/kclient_workunit_snaps.yaml} 3
pass 1113197 2017-05-08 03:25:50 2017-05-08 13:36:54 2017-05-08 14:16:55 0:40:01 0:32:30 0:07:31 smithi master kcephfs/thrash/{clusters/fixed-3-cephfs.yaml conf.yaml objectstore/bluestore.yaml thrashers/mds.yaml workloads/kclient_workunit_suites_ffsb.yaml} 3
fail 1113198 2017-05-08 03:25:50 2017-05-08 13:36:54 2017-05-08 16:08:58 2:32:04 0:12:20 2:19:44 smithi master kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/filestore-xfs.yaml tasks/mds-full.yaml} 4
Failure Reason:

Test failure: test_full_different_file (tasks.cephfs.test_full.TestClusterFull)

pass 1113199 2017-05-08 03:25:51 2017-05-08 13:36:54 2017-05-08 15:02:56 1:26:02 0:42:23 0:43:39 smithi master kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore/bluestore.yaml tasks/kclient_workunit_suites_dbench.yaml} 3
pass 1113200 2017-05-08 03:25:52 2017-05-08 13:36:54 2017-05-08 14:16:55 0:40:01 0:25:53 0:14:08 smithi master kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore/filestore-xfs.yaml tasks/kclient_workunit_suites_ffsb.yaml} 3
pass 1113201 2017-05-08 03:25:52 2017-05-08 13:36:54 2017-05-08 14:42:56 1:06:02 0:12:18 0:53:44 smithi master kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/bluestore.yaml tasks/pool-perm.yaml} 4
pass 1113202 2017-05-08 03:25:53 2017-05-08 13:38:45 2017-05-08 14:10:45 0:32:00 0:10:58 0:21:02 smithi master kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore/bluestore.yaml tasks/kclient_workunit_suites_fsstress.yaml} 3
pass 1113203 2017-05-08 03:25:53 2017-05-08 13:38:45 2017-05-08 14:14:45 0:36:00 0:16:14 0:19:46 smithi master kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore/filestore-xfs.yaml tasks/kclient_workunit_suites_fsx.yaml} 3
pass 1113204 2017-05-08 03:25:54 2017-05-08 13:38:53 2017-05-08 14:06:53 0:28:00 0:13:03 0:14:57 smithi master kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/filestore-xfs.yaml tasks/sessionmap.yaml} 4
pass 1113205 2017-05-08 03:25:54 2017-05-08 13:40:36 2017-05-08 14:30:36 0:50:00 0:25:25 0:24:35 smithi master kcephfs/thrash/{clusters/fixed-3-cephfs.yaml conf.yaml objectstore/filestore-xfs.yaml thrashers/mon.yaml workloads/kclient_workunit_suites_iozone.yaml} 3
pass 1113206 2017-05-08 03:25:55 2017-05-08 13:40:41 2017-05-08 14:22:41 0:42:00 0:10:54 0:31:06 smithi master kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore/bluestore.yaml tasks/kclient_workunit_suites_fsync.yaml} 3
pass 1113207 2017-05-08 03:25:56 2017-05-08 13:40:44 2017-05-08 15:08:45 1:28:01 0:28:28 0:59:33 smithi master kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/bluestore.yaml tasks/strays.yaml} 4
pass 1113208 2017-05-08 03:25:56 2017-05-08 13:41:05 2017-05-08 14:11:04 0:29:59 0:24:34 0:05:25 smithi master kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore/filestore-xfs.yaml tasks/kclient_workunit_suites_iozone.yaml} 3
pass 1113209 2017-05-08 03:25:57 2017-05-08 13:42:17 2017-05-08 14:02:15 0:19:58 0:11:47 0:08:11 smithi master kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore/bluestore.yaml tasks/kclient_workunit_suites_pjd.yaml} 3
pass 1113210 2017-05-08 03:25:58 2017-05-08 13:42:29 2017-05-08 14:20:30 0:38:01 0:19:52 0:18:09 smithi master kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/filestore-xfs.yaml tasks/volume-client.yaml} 4
pass 1113211 2017-05-08 03:25:58 2017-05-08 13:42:38 2017-05-08 13:58:37 0:15:59 0:09:13 0:06:46 smithi master kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore/filestore-xfs.yaml tasks/kclient_workunit_trivial_sync.yaml} 3
pass 1113212 2017-05-08 03:25:59 2017-05-08 13:42:49 2017-05-08 13:58:50 0:16:01 0:11:03 0:04:58 smithi master kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore/bluestore.yaml tasks/kclient_workunit_direct_io.yaml} 3
pass 1113213 2017-05-08 03:26:00 2017-05-08 13:43:02 2017-05-08 14:51:03 1:08:01 0:42:17 0:25:44 smithi master kcephfs/mixed-clients/{clusters/2-clients.yaml conf.yaml objectstore/filestore-xfs.yaml tasks/kernel_cfuse_workunits_dbench_iozone.yaml} 4
fail 1113214 2017-05-08 03:26:01 2017-05-08 13:43:02 2017-05-08 15:33:04 1:50:02 0:13:19 1:36:43 smithi master kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/filestore-xfs.yaml tasks/auto-repair.yaml} 4
Failure Reason:

"2017-05-08 15:25:31.065114 mon.0 172.21.15.150:6789/0 457 : cluster [WRN] MDS health message (mds.0): MDS in read-only mode" in cluster log

pass 1113215 2017-05-08 03:26:01 2017-05-08 13:44:33 2017-05-08 14:24:33 0:40:00 0:28:36 0:11:24 smithi master kcephfs/thrash/{clusters/fixed-3-cephfs.yaml conf.yaml objectstore/filestore-xfs.yaml thrashers/default.yaml workloads/kclient_workunit_suites_ffsb.yaml} 3
pass 1113216 2017-05-08 03:26:02 2017-05-08 13:44:43 2017-05-08 14:22:43 0:38:00 0:21:32 0:16:28 smithi master kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore/filestore-xfs.yaml tasks/kclient_workunit_kernel_untar_build.yaml} 3
pass 1113217 2017-05-08 03:26:02 2017-05-08 13:44:53 2017-05-08 14:16:53 0:32:00 0:10:49 0:21:11 smithi master kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/bluestore.yaml tasks/backtrace.yaml} 4
pass 1113218 2017-05-08 03:26:03 2017-05-08 13:44:54 2017-05-08 14:42:56 0:58:02 0:32:10 0:25:52 smithi master kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore/bluestore.yaml tasks/kclient_workunit_misc.yaml} 3
pass 1113219 2017-05-08 03:26:04 2017-05-08 13:46:04 2017-05-08 14:20:04 0:34:00 0:19:34 0:14:26 smithi master kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore/filestore-xfs.yaml tasks/kclient_workunit_o_trunc.yaml} 3
fail 1113220 2017-05-08 03:26:04 2017-05-08 13:46:35 2017-05-08 14:18:35 0:32:00 0:16:00 0:16:00 smithi master kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/filestore-xfs.yaml tasks/client-limits.yaml} 4
Failure Reason:

"2017-05-08 14:08:08.630082 mon.0 172.21.15.64:6789/0 582 : cluster [WRN] MDS health message (mds.0): Too many inodes in cache (211/100), 202 inodes in use by clients, 0 stray files" in cluster log

pass 1113222 2017-05-08 03:26:05 2017-05-08 13:46:48 2017-05-08 14:42:51 0:56:03 0:45:25 0:10:38 smithi master kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore/bluestore.yaml tasks/kclient_workunit_snaps.yaml} 3
pass 1113224 2017-05-08 03:26:06 2017-05-08 13:47:31 2017-05-08 14:11:31 0:24:00 0:13:07 0:10:53 smithi master kcephfs/thrash/{clusters/fixed-3-cephfs.yaml conf.yaml objectstore/bluestore.yaml thrashers/mds.yaml workloads/kclient_workunit_suites_iozone.yaml} 3
pass 1113226 2017-05-08 03:26:06 2017-05-08 13:51:15 2017-05-08 15:21:16 1:30:01 0:32:28 0:57:33 smithi master kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/bluestore.yaml tasks/client-recovery.yaml} 4
pass 1113228 2017-05-08 03:26:07 2017-05-08 13:52:58 2017-05-08 14:42:58 0:50:00 0:37:41 0:12:19 smithi master kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore/filestore-xfs.yaml tasks/kclient_workunit_suites_dbench.yaml} 3
pass 1113230 2017-05-08 03:26:08 2017-05-08 13:52:58 2017-05-08 14:36:58 0:44:00 0:32:15 0:11:45 smithi master kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore/bluestore.yaml tasks/kclient_workunit_suites_ffsb.yaml} 3
pass 1113232 2017-05-08 03:26:08 2017-05-08 13:52:58 2017-05-08 14:26:58 0:34:00 0:11:25 0:22:35 smithi master kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/filestore-xfs.yaml tasks/config-commands.yaml} 4
pass 1113234 2017-05-08 03:26:09 2017-05-08 13:52:58 2017-05-08 14:08:58 0:16:00 0:10:39 0:05:21 smithi master kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore/filestore-xfs.yaml tasks/kclient_workunit_suites_fsstress.yaml} 3
pass 1113236 2017-05-08 03:26:09 2017-05-08 13:52:58 2017-05-08 14:36:58 0:44:00 0:20:17 0:23:43 smithi master kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore/bluestore.yaml tasks/kclient_workunit_suites_fsx.yaml} 3
fail 1113238 2017-05-08 03:26:10 2017-05-08 13:54:47 2017-05-08 14:58:48 1:04:01 0:18:02 0:45:59 smithi master kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/bluestore.yaml tasks/damage.yaml} 4
Failure Reason:

"2017-05-08 14:42:43.190531 mon.0 172.21.15.70:6789/0 340 : cluster [ERR] MDS health message (mds.0): Metadata damage detected" in cluster log

pass 1113239 2017-05-08 03:26:11 2017-05-08 13:54:47 2017-05-08 14:26:47 0:32:00 0:23:28 0:08:32 smithi master kcephfs/thrash/{clusters/fixed-3-cephfs.yaml conf.yaml objectstore/filestore-xfs.yaml thrashers/mon.yaml workloads/kclient_workunit_suites_ffsb.yaml} 3
pass 1113241 2017-05-08 03:26:11 2017-05-08 13:54:59 2017-05-08 14:26:59 0:32:00 0:08:26 0:23:34 smithi master kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore/filestore-xfs.yaml tasks/kclient_workunit_suites_fsync.yaml} 3
fail 1113243 2017-05-08 03:26:12 2017-05-08 13:55:02 2017-05-08 14:31:02 0:36:00 0:27:55 0:08:05 smithi master kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/filestore-xfs.yaml tasks/data-scan.yaml} 4
Failure Reason:

Test failure: test_rebuild_simple_altpool (tasks.cephfs.test_data_scan.TestDataScan)

pass 1113245 2017-05-08 03:26:12 2017-05-08 13:55:03 2017-05-08 14:11:02 0:15:59 0:10:11 0:05:48 smithi master kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore/bluestore.yaml tasks/kclient_workunit_suites_iozone.yaml} 3
pass 1113247 2017-05-08 03:26:13 2017-05-08 13:57:03 2017-05-08 14:13:02 0:15:59 0:10:28 0:05:31 smithi master kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore/filestore-xfs.yaml tasks/kclient_workunit_suites_pjd.yaml} 3
pass 1113249 2017-05-08 03:26:14 2017-05-08 13:58:51 2017-05-08 15:40:52 1:42:01 0:23:58 1:18:03 smithi master kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/bluestore.yaml tasks/failover.yaml} 4
pass 1113251 2017-05-08 03:26:14 2017-05-08 13:58:51 2017-05-08 14:28:51 0:30:00 0:07:08 0:22:52 smithi master kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore/bluestore.yaml tasks/kclient_workunit_trivial_sync.yaml} 3
pass 1113253 2017-05-08 03:26:15 2017-05-08 13:58:51 2017-05-08 14:40:51 0:42:00 0:09:44 0:32:16 smithi master kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore/filestore-xfs.yaml tasks/kclient_workunit_direct_io.yaml} 3
pass 1113254 2017-05-08 03:26:16 2017-05-08 13:58:53 2017-05-08 15:18:55 1:20:02 0:35:05 0:44:57 smithi master kcephfs/mixed-clients/{clusters/2-clients.yaml conf.yaml objectstore/bluestore.yaml tasks/kernel_cfuse_workunits_untarbuild_blogbench.yaml} 4
fail 1113256 2017-05-08 03:26:16 2017-05-08 13:59:07 2017-05-08 14:25:07 0:26:00 0:14:11 0:11:49 smithi master kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/filestore-xfs.yaml tasks/forward-scrub.yaml} 4
Failure Reason:

"2017-05-08 14:16:33.075852 mon.0 172.21.15.153:6789/0 398 : cluster [ERR] MDS health message (mds.0): Metadata damage detected" in cluster log

pass 1113258 2017-05-08 03:26:17 2017-05-08 13:59:19 2017-05-08 14:17:19 0:18:00 0:10:52 0:07:08 smithi master kcephfs/thrash/{clusters/fixed-3-cephfs.yaml conf.yaml objectstore/bluestore.yaml thrashers/default.yaml workloads/kclient_workunit_suites_iozone.yaml} 3
pass 1113259 2017-05-08 03:26:17 2017-05-08 14:00:33 2017-05-08 14:38:33 0:38:00 0:26:06 0:11:54 smithi master kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore/bluestore.yaml tasks/kclient_workunit_kernel_untar_build.yaml} 3
fail 1113260 2017-05-08 03:26:18 2017-05-08 14:00:33 2017-05-08 14:36:33 0:36:00 0:26:58 0:09:02 smithi master kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/bluestore.yaml tasks/journal-repair.yaml} 4
Failure Reason:

"2017-05-08 14:13:43.508310 mon.0 172.21.15.114:6789/0 362 : cluster [ERR] MDS health message (mds.0): Metadata damage detected" in cluster log

pass 1113261 2017-05-08 03:26:19 2017-05-08 14:00:51 2017-05-08 15:06:52 1:06:01 0:30:53 0:35:08 smithi master kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore/filestore-xfs.yaml tasks/kclient_workunit_misc.yaml} 3
pass 1113262 2017-05-08 03:26:19 2017-05-08 14:00:52 2017-05-08 14:24:52 0:24:00 0:19:22 0:04:38 smithi master kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore/bluestore.yaml tasks/kclient_workunit_o_trunc.yaml} 3
pass 1113263 2017-05-08 03:26:20 2017-05-08 14:01:04 2017-05-08 14:35:04 0:34:00 0:08:42 0:25:18 smithi master kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/filestore-xfs.yaml tasks/mds-flush.yaml} 4
pass 1113264 2017-05-08 03:26:21 2017-05-08 14:02:30 2017-05-08 15:10:31 1:08:01 0:41:04 0:26:57 smithi master kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore/filestore-xfs.yaml tasks/kclient_workunit_snaps.yaml} 3
pass 1113265 2017-05-08 03:26:21 2017-05-08 14:02:38 2017-05-08 14:34:38 0:32:00 0:25:26 0:06:34 smithi master kcephfs/thrash/{clusters/fixed-3-cephfs.yaml conf.yaml objectstore/filestore-xfs.yaml thrashers/mds.yaml workloads/kclient_workunit_suites_ffsb.yaml} 3
fail 1113266 2017-05-08 03:26:22 2017-05-08 14:02:42 2017-05-08 14:52:44 0:50:02 0:12:20 0:37:42 smithi master kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/bluestore.yaml tasks/mds-full.yaml} 4
Failure Reason:

Test failure: test_full_different_file (tasks.cephfs.test_full.TestClusterFull)

pass 1113267 2017-05-08 03:26:22 2017-05-08 14:02:47 2017-05-08 15:00:48 0:58:01 0:42:51 0:15:10 smithi master kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore/bluestore.yaml tasks/kclient_workunit_suites_dbench.yaml} 3
pass 1113268 2017-05-08 03:26:23 2017-05-08 14:05:06 2017-05-08 14:39:06 0:34:00 0:23:29 0:10:31 smithi master kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore/filestore-xfs.yaml tasks/kclient_workunit_suites_ffsb.yaml} 3
pass 1113269 2017-05-08 03:26:24 2017-05-08 14:05:06 2017-05-08 15:41:07 1:36:01 0:11:29 1:24:32 smithi master kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/filestore-xfs.yaml tasks/pool-perm.yaml} 4
pass 1113270 2017-05-08 03:26:25 2017-05-08 14:07:05 2017-05-08 14:19:05 0:12:00 0:08:34 0:03:26 smithi master kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore/bluestore.yaml tasks/kclient_workunit_suites_fsstress.yaml} 3
pass 1113271 2017-05-08 03:26:25 2017-05-08 14:07:05 2017-05-08 14:39:05 0:32:00 0:16:56 0:15:04 smithi master kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore/filestore-xfs.yaml tasks/kclient_workunit_suites_fsx.yaml} 3
pass 1113272 2017-05-08 03:26:26 2017-05-08 14:09:07 2017-05-08 15:41:09 1:32:02 0:10:54 1:21:08 smithi master kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/bluestore.yaml tasks/sessionmap.yaml} 4
pass 1113273 2017-05-08 03:26:26 2017-05-08 14:09:07 2017-05-08 15:03:08 0:54:01 0:13:35 0:40:26 smithi master kcephfs/thrash/{clusters/fixed-3-cephfs.yaml conf.yaml objectstore/bluestore.yaml thrashers/mon.yaml workloads/kclient_workunit_suites_iozone.yaml} 3
pass 1113274 2017-05-08 03:26:27 2017-05-08 14:09:07 2017-05-08 14:51:07 0:42:00 0:07:58 0:34:02 smithi master kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore/bluestore.yaml tasks/kclient_workunit_suites_fsync.yaml} 3
fail 1113275 2017-05-08 03:26:28 2017-05-08 14:09:15 2017-05-08 15:21:16 1:12:01 0:29:24 0:42:37 smithi master kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/filestore-xfs.yaml tasks/strays.yaml} 4
Failure Reason:

Test failure: test_purge_queue_op_rate (tasks.cephfs.test_strays.TestStrays)

pass 1113276 2017-05-08 03:26:28 2017-05-08 14:11:05 2017-05-08 14:55:05 0:44:00 0:23:50 0:20:10 smithi master kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore/filestore-xfs.yaml tasks/kclient_workunit_suites_iozone.yaml} 3
pass 1113277 2017-05-08 03:26:29 2017-05-08 14:11:05 2017-05-08 14:27:05 0:16:00 0:11:44 0:04:16 smithi master kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/yes.yaml objectstore/bluestore.yaml tasks/kclient_workunit_suites_pjd.yaml} 3
pass 1113278 2017-05-08 03:26:30 2017-05-08 14:11:05 2017-05-08 15:17:06 1:06:01 0:22:14 0:43:47 smithi master kcephfs/recovery/{clusters/4-remote-clients.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore/bluestore.yaml tasks/volume-client.yaml} 4
pass 1113279 2017-05-08 03:26:30 2017-05-08 14:11:06 2017-05-08 14:31:05 0:19:59 0:09:46 0:10:13 smithi master kcephfs/cephfs/{clusters/fixed-3-cephfs.yaml conf.yaml inline/no.yaml objectstore/filestore-xfs.yaml tasks/kclient_workunit_trivial_sync.yaml} 3