Status Job ID Links Posted Started Updated
Runtime
Duration
In Waiting
Machine Teuthology Branch OS Type OS Version Description Nodes
fail 7421047 2023-10-11 02:17:22 2023-10-11 04:04:24 2023-10-11 07:27:54 3:23:30 3:12:41 0:10:49 smithi main centos 9.stream crimson-rados/basic/{clusters/fixed-2 crimson-supported-all-distro/centos_latest crimson_qa_overrides deploy/ceph tasks/rados_api_tests} 2
Failure Reason:

Command failed (workunit test rados/test.sh) on smithi050 with status 124: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=9587ffdfc8c52bcf06abbe89d2f527c434a3e9f5 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 CEPH_MNT=/home/ubuntu/cephtest/mnt.0 CRIMSON_COMPAT=1 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/rados/test.sh'

pass 7421048 2023-10-11 02:17:23 2023-10-11 04:04:25 2023-10-11 04:29:15 0:24:50 0:16:04 0:08:46 smithi main centos 9.stream crimson-rados/perf/{clusters/{fixed-2} crimson-supported-all-distro/centos_latest crimson_qa_overrides deploy/ceph objectstore/bluestore settings/optimized workloads/fio_4K_rand_read} 1
fail 7421049 2023-10-11 02:17:24 2023-10-11 04:04:25 2023-10-11 07:27:30 3:23:05 3:11:57 0:11:08 smithi main centos 9.stream crimson-rados/rbd/{clusters/fixed-1 crimson-supported-all-distro/centos_latest crimson_qa_overrides deploy/ceph tasks/rbd_api_tests} 1
Failure Reason:

Command failed (workunit test rbd/crimson/test_crimson_librbd.sh) on smithi094 with status 124: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=9587ffdfc8c52bcf06abbe89d2f527c434a3e9f5 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 CEPH_MNT=/home/ubuntu/cephtest/mnt.0 RBD_FEATURES=61 CRIMSON_COMPAT=1 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/rbd/crimson/test_crimson_librbd.sh'

pass 7421050 2023-10-11 02:17:24 2023-10-11 04:06:26 2023-10-11 04:31:41 0:25:15 0:15:40 0:09:35 smithi main centos 9.stream crimson-rados/singleton/{all/osd-backfill crimson-supported-all-distro/centos_latest crimson_qa_overrides objectstore/bluestore rados} 1
pass 7421051 2023-10-11 02:17:25 2023-10-11 04:06:56 2023-10-11 04:33:01 0:26:05 0:15:45 0:10:20 smithi main centos 9.stream crimson-rados/thrash/{0-size-min-size-overrides/3-size-2-min-size 1-pg-log-overrides/normal_pg_log 2-recovery-overrides/{default} clusters/{fixed-2} crimson-supported-all-distro/centos_latest crimson_qa_overrides deploy/ceph objectstore/bluestore thrashers/default thrashosds-health workloads/admin_socket_objecter_requests} 2
pass 7421052 2023-10-11 02:17:26 2023-10-11 04:07:57 2023-10-11 04:34:01 0:26:04 0:15:37 0:10:27 smithi main centos 9.stream crimson-rados/perf/{clusters/{fixed-2} crimson-supported-all-distro/centos_latest crimson_qa_overrides deploy/ceph objectstore/bluestore settings/optimized workloads/fio_4K_rand_rw} 1
dead 7421053 2023-10-11 02:17:26 2023-10-11 04:08:47 2023-10-11 16:18:42 12:09:55 smithi main centos 9.stream crimson-rados/thrash/{0-size-min-size-overrides/3-size-2-min-size 1-pg-log-overrides/normal_pg_log 2-recovery-overrides/{default} clusters/{fixed-2} crimson-supported-all-distro/centos_latest crimson_qa_overrides deploy/ceph objectstore/bluestore thrashers/default thrashosds-health workloads/pool-snaps-few-objects} 2
Failure Reason:

hit max job timeout

fail 7421054 2023-10-11 02:17:27 2023-10-11 04:09:37 2023-10-11 04:34:56 0:25:19 0:15:05 0:10:14 smithi main centos 9.stream crimson-rados/perf/{clusters/{fixed-2} crimson-supported-all-distro/centos_latest crimson_qa_overrides deploy/ceph objectstore/bluestore settings/optimized workloads/fio_4M_rand_read} 1
Failure Reason:

"2023-10-11T04:28:48.853393+0000 mon.a (mon.0) 127 : cluster [WRN] Health check failed: Reduced data availability: 9 pgs inactive (PG_AVAILABILITY)" in cluster log

fail 7421055 2023-10-11 02:17:28 2023-10-11 04:09:38 2023-10-11 07:31:07 3:21:29 3:12:05 0:09:24 smithi main centos 9.stream crimson-rados/rbd/{clusters/fixed-1 crimson-supported-all-distro/centos_latest crimson_qa_overrides deploy/ceph tasks/rbd_api_tests_old_format} 1
Failure Reason:

Command failed (workunit test rbd/crimson/test_crimson_librbd.sh) on smithi169 with status 124: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=9587ffdfc8c52bcf06abbe89d2f527c434a3e9f5 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 CEPH_MNT=/home/ubuntu/cephtest/mnt.0 CRIMSON_COMPAT=1 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/rbd/crimson/test_crimson_librbd.sh'

pass 7421056 2023-10-11 02:17:28 2023-10-11 04:09:38 2023-10-11 04:46:20 0:36:42 0:25:45 0:10:57 smithi main centos 9.stream crimson-rados/thrash/{0-size-min-size-overrides/3-size-2-min-size 1-pg-log-overrides/normal_pg_log 2-recovery-overrides/{default} clusters/{fixed-2} crimson-supported-all-distro/centos_latest crimson_qa_overrides deploy/ceph objectstore/bluestore thrashers/default thrashosds-health workloads/radosbench-high-concurrency} 2
pass 7421057 2023-10-11 02:17:29 2023-10-11 04:10:49 2023-10-11 04:36:24 0:25:35 0:16:10 0:09:25 smithi main centos 9.stream crimson-rados/perf/{clusters/{fixed-2} crimson-supported-all-distro/centos_latest crimson_qa_overrides deploy/ceph objectstore/bluestore settings/optimized workloads/fio_4M_rand_rw} 1
pass 7421058 2023-10-11 02:17:29 2023-10-11 04:11:29 2023-10-11 04:46:33 0:35:04 0:26:31 0:08:33 smithi main centos 9.stream crimson-rados/thrash/{0-size-min-size-overrides/3-size-2-min-size 1-pg-log-overrides/normal_pg_log 2-recovery-overrides/{default} clusters/{fixed-2} crimson-supported-all-distro/centos_latest crimson_qa_overrides deploy/ceph objectstore/bluestore thrashers/default thrashosds-health workloads/radosbench} 2
pass 7421059 2023-10-11 02:17:30 2023-10-11 04:11:30 2023-10-11 04:36:20 0:24:50 0:15:51 0:08:59 smithi main centos 9.stream crimson-rados/perf/{clusters/{fixed-2} crimson-supported-all-distro/centos_latest crimson_qa_overrides deploy/ceph objectstore/bluestore settings/optimized workloads/fio_4M_rand_write} 1
fail 7421060 2023-10-11 02:17:31 2023-10-11 04:11:30 2023-10-11 05:33:53 1:22:23 1:11:52 0:10:31 smithi main centos 9.stream crimson-rados/basic/{clusters/fixed-2 crimson-supported-all-distro/centos_latest crimson_qa_overrides deploy/ceph tasks/rados_python} 2
Failure Reason:

Command failed (workunit test rados/test_python.sh) on smithi005 with status 124: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=9587ffdfc8c52bcf06abbe89d2f527c434a3e9f5 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 CEPH_MNT=/home/ubuntu/cephtest/mnt.0 CRIMSON_COMPAT=1 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 1h /home/ubuntu/cephtest/clone.client.0/qa/workunits/rados/test_python.sh -m \'not (wait or tier or ec or bench or stats)\''

fail 7421061 2023-10-11 02:17:31 2023-10-11 04:12:00 2023-10-11 07:33:44 3:21:44 3:12:26 0:09:18 smithi main centos 9.stream crimson-rados/rbd/{clusters/fixed-1 crimson-supported-all-distro/centos_latest crimson_qa_overrides deploy/ceph tasks/rbd_cls_tests} 1
Failure Reason:

Command failed (workunit test cls/test_cls_rbd.sh) on smithi138 with status 124: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=9587ffdfc8c52bcf06abbe89d2f527c434a3e9f5 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 CEPH_MNT=/home/ubuntu/cephtest/mnt.0 CRIMSON_COMPAT=1 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/cls/test_cls_rbd.sh'

dead 7421062 2023-10-11 02:17:32 2023-10-11 04:12:31 2023-10-11 16:21:49 12:09:18 smithi main centos 9.stream crimson-rados/thrash/{0-size-min-size-overrides/3-size-2-min-size 1-pg-log-overrides/normal_pg_log 2-recovery-overrides/{default} clusters/{fixed-2} crimson-supported-all-distro/centos_latest crimson_qa_overrides deploy/ceph objectstore/bluestore thrashers/default thrashosds-health workloads/small-objects-balanced} 2
Failure Reason:

hit max job timeout

fail 7421063 2023-10-11 02:17:33 2023-10-11 04:13:31 2023-10-11 04:36:50 0:23:19 0:12:58 0:10:21 smithi main centos 9.stream crimson-rados/perf/{clusters/{fixed-2} crimson-supported-all-distro/centos_latest crimson_qa_overrides deploy/ceph objectstore/bluestore settings/optimized workloads/radosbench_4K_rand_read} 1
Failure Reason:

"2023-10-11T04:32:42.229603+0000 mon.a (mon.0) 152 : cluster [WRN] Health check failed: Degraded data redundancy: 117/9556 objects degraded (1.224%), 6 pgs degraded (PG_DEGRADED)" in cluster log

dead 7421064 2023-10-11 02:17:33 2023-10-11 04:13:52 2023-10-11 16:25:07 12:11:15 smithi main centos 9.stream crimson-rados/thrash/{0-size-min-size-overrides/3-size-2-min-size 1-pg-log-overrides/normal_pg_log 2-recovery-overrides/{default} clusters/{fixed-2} crimson-supported-all-distro/centos_latest crimson_qa_overrides deploy/ceph objectstore/bluestore thrashers/default thrashosds-health workloads/small-objects-localized} 2
Failure Reason:

hit max job timeout

fail 7421065 2023-10-11 02:17:34 2023-10-11 04:16:32 2023-10-11 04:39:32 0:23:00 0:13:38 0:09:22 smithi main centos 9.stream crimson-rados/perf/{clusters/{fixed-2} crimson-supported-all-distro/centos_latest crimson_qa_overrides deploy/ceph objectstore/bluestore settings/optimized workloads/radosbench_4K_seq_read} 1
Failure Reason:

"2023-10-11T04:37:05.179879+0000 mon.a (mon.0) 154 : cluster [WRN] Health check failed: Degraded data redundancy: 66/5914 objects degraded (1.116%), 6 pgs degraded (PG_DEGRADED)" in cluster log

fail 7421066 2023-10-11 02:17:35 2023-10-11 04:16:33 2023-10-11 07:37:14 3:20:41 3:12:37 0:08:04 smithi main centos 9.stream crimson-rados/rbd/{clusters/fixed-1 crimson-supported-all-distro/centos_latest crimson_qa_overrides deploy/ceph tasks/rbd_lock_and_fence} 1
Failure Reason:

Command failed (workunit test rbd/test_lock_fence.sh) on smithi037 with status 124: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=9587ffdfc8c52bcf06abbe89d2f527c434a3e9f5 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 CEPH_MNT=/home/ubuntu/cephtest/mnt.0 CRIMSON_COMPAT=1 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/rbd/test_lock_fence.sh'

fail 7421067 2023-10-11 02:17:35 2023-10-11 04:16:33 2023-10-11 04:41:14 0:24:41 0:14:21 0:10:20 smithi main centos 9.stream crimson-rados/perf/{clusters/{fixed-2} crimson-supported-all-distro/centos_latest crimson_qa_overrides deploy/ceph objectstore/bluestore settings/optimized workloads/radosbench_4M_rand_read} 1
Failure Reason:

"2023-10-11T04:37:03.372468+0000 mon.a (mon.0) 153 : cluster [WRN] Health check failed: Degraded data redundancy: 15/1788 objects degraded (0.839%), 7 pgs degraded (PG_DEGRADED)" in cluster log

dead 7421068 2023-10-11 02:17:36 2023-10-11 04:16:33 2023-10-11 16:25:02 12:08:29 smithi main centos 9.stream crimson-rados/thrash/{0-size-min-size-overrides/3-size-2-min-size 1-pg-log-overrides/normal_pg_log 2-recovery-overrides/{default} clusters/{fixed-2} crimson-supported-all-distro/centos_latest crimson_qa_overrides deploy/ceph objectstore/bluestore thrashers/default thrashosds-health workloads/small-objects} 2
Failure Reason:

hit max job timeout

fail 7421069 2023-10-11 02:17:37 2023-10-11 04:16:34 2023-10-11 04:37:43 0:21:09 0:12:33 0:08:36 smithi main centos 9.stream crimson-rados/perf/{clusters/{fixed-2} crimson-supported-all-distro/centos_latest crimson_qa_overrides deploy/ceph objectstore/bluestore settings/optimized workloads/radosbench_4M_seq_read} 1
Failure Reason:

"2023-10-11T04:35:53.096967+0000 mon.a (mon.0) 149 : cluster [WRN] Health check failed: Degraded data redundancy: 4/1696 objects degraded (0.236%), 3 pgs degraded (PG_DEGRADED)" in cluster log

dead 7421070 2023-10-11 02:17:37 2023-10-11 04:16:44 2023-10-11 16:25:03 12:08:19 smithi main centos 9.stream crimson-rados/thrash/{0-size-min-size-overrides/3-size-2-min-size 1-pg-log-overrides/normal_pg_log 2-recovery-overrides/{default} clusters/{fixed-2} crimson-supported-all-distro/centos_latest crimson_qa_overrides deploy/ceph objectstore/bluestore thrashers/default thrashosds-health workloads/snaps-few-objects-balanced} 2
Failure Reason:

hit max job timeout

dead 7421071 2023-10-11 02:17:38 2023-10-11 04:16:44 2023-10-11 16:25:28 12:08:44 smithi main centos 9.stream crimson-rados/basic/{clusters/fixed-2 crimson-supported-all-distro/centos_latest crimson_qa_overrides deploy/ceph tasks/readwrite} 2
Failure Reason:

hit max job timeout

fail 7421072 2023-10-11 02:17:39 2023-10-11 04:16:55 2023-10-11 07:38:40 3:21:45 3:12:06 0:09:39 smithi main centos 9.stream crimson-rados/rbd/{clusters/fixed-1 crimson-supported-all-distro/centos_latest crimson_qa_overrides deploy/ceph tasks/rbd_python_api_tests} 1
Failure Reason:

Command failed (workunit test rbd/test_librbd_python.sh) on smithi006 with status 124: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=9587ffdfc8c52bcf06abbe89d2f527c434a3e9f5 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 CEPH_MNT=/home/ubuntu/cephtest/mnt.0 RBD_FEATURES=61 CRIMSON_COMPAT=1 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/rbd/test_librbd_python.sh -m \'not skip_if_crimson\''

fail 7421073 2023-10-11 02:17:39 2023-10-11 04:17:35 2023-10-11 04:39:53 0:22:18 0:12:14 0:10:04 smithi main centos 9.stream crimson-rados/perf/{clusters/{fixed-2} crimson-supported-all-distro/centos_latest crimson_qa_overrides deploy/ceph objectstore/bluestore settings/optimized workloads/radosbench_4M_write} 1
Failure Reason:

"2023-10-11T04:38:53.189066+0000 mon.a (mon.0) 152 : cluster [WRN] Health check failed: Degraded data redundancy: 9/1590 objects degraded (0.566%), 5 pgs degraded (PG_DEGRADED)" in cluster log

dead 7421074 2023-10-11 02:17:40 2023-10-11 04:18:06 2023-10-11 16:26:30 12:08:24 smithi main centos 9.stream crimson-rados/thrash/{0-size-min-size-overrides/3-size-2-min-size 1-pg-log-overrides/normal_pg_log 2-recovery-overrides/{default} clusters/{fixed-2} crimson-supported-all-distro/centos_latest crimson_qa_overrides deploy/ceph objectstore/bluestore thrashers/default thrashosds-health workloads/snaps-few-objects-localized} 2
Failure Reason:

hit max job timeout

pass 7421075 2023-10-11 02:17:41 2023-10-11 04:18:06 2023-10-11 04:51:20 0:33:14 0:22:55 0:10:19 smithi main centos 9.stream crimson-rados/perf/{clusters/{fixed-2} crimson-supported-all-distro/centos_latest crimson_qa_overrides deploy/ceph objectstore/bluestore settings/optimized workloads/radosbench_omap_write} 1
dead 7421076 2023-10-11 02:17:41 2023-10-11 04:18:26 2023-10-11 16:27:12 12:08:46 smithi main centos 9.stream crimson-rados/thrash/{0-size-min-size-overrides/3-size-2-min-size 1-pg-log-overrides/normal_pg_log 2-recovery-overrides/{default} clusters/{fixed-2} crimson-supported-all-distro/centos_latest crimson_qa_overrides deploy/ceph objectstore/bluestore thrashers/default thrashosds-health workloads/snaps-few-objects} 2
Failure Reason:

hit max job timeout

fail 7421077 2023-10-11 02:17:42 2023-10-11 04:18:57 2023-10-11 07:39:54 3:20:57 3:11:32 0:09:25 smithi main centos 9.stream crimson-rados/rbd/{clusters/fixed-1 crimson-supported-all-distro/centos_latest crimson_qa_overrides deploy/ceph tasks/rbd_python_api_tests_old_format} 1
Failure Reason:

Command failed (workunit test rbd/test_librbd_python.sh) on smithi110 with status 124: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=9587ffdfc8c52bcf06abbe89d2f527c434a3e9f5 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 CEPH_MNT=/home/ubuntu/cephtest/mnt.0 CRIMSON_COMPAT=1 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/rbd/test_librbd_python.sh -m \'not skip_if_crimson\''

pass 7421078 2023-10-11 02:17:43 2023-10-11 04:18:57 2023-10-11 04:44:46 0:25:49 0:16:00 0:09:49 smithi main centos 9.stream crimson-rados/perf/{clusters/{fixed-2} crimson-supported-all-distro/centos_latest crimson_qa_overrides deploy/ceph objectstore/bluestore settings/optimized workloads/sample_fio} 1
dead 7421079 2023-10-11 02:17:43 2023-10-11 04:19:07 2023-10-11 16:27:34 12:08:27 smithi main centos 9.stream crimson-rados/thrash/{0-size-min-size-overrides/3-size-2-min-size 1-pg-log-overrides/normal_pg_log 2-recovery-overrides/{default} clusters/{fixed-2} crimson-supported-all-distro/centos_latest crimson_qa_overrides deploy/ceph objectstore/bluestore thrashers/default thrashosds-health workloads/write_fadvise_dontneed} 2
Failure Reason:

hit max job timeout

fail 7421080 2023-10-11 02:17:44 2023-10-11 04:19:28 2023-10-11 04:40:46 0:21:18 0:11:46 0:09:32 smithi main centos 9.stream crimson-rados/perf/{clusters/{fixed-2} crimson-supported-all-distro/centos_latest crimson_qa_overrides deploy/ceph objectstore/bluestore settings/optimized workloads/sample_radosbench} 1
Failure Reason:

"2023-10-11T04:39:03.596272+0000 mon.a (mon.0) 150 : cluster [WRN] Health check failed: Degraded data redundancy: 194/10372 objects degraded (1.870%), 1 pg degraded (PG_DEGRADED)" in cluster log