ID
Status
Ceph Branch
Suite Branch
Teuthology Branch
Machine
OS
Nodes
Description
Failure Reason
wip-crimson-scrub-testing-2023-10-16
wip-crimson-scrub-testing-2023-10-16
main
smithi
centos 9.stream
crimson-rados/basic/{clusters/fixed-2 crimson-supported-all-distro/centos_latest crimson_qa_overrides deploy/ceph tasks/rados_api_tests}
wip-crimson-scrub-testing-2023-10-16
wip-crimson-scrub-testing-2023-10-16
main
smithi
centos 9.stream
crimson-rados/perf/{clusters/{fixed-2} crimson-supported-all-distro/centos_latest crimson_qa_overrides deploy/ceph objectstore/bluestore settings/optimized workloads/fio_4K_rand_read}
wip-crimson-scrub-testing-2023-10-16
wip-crimson-scrub-testing-2023-10-16
main
smithi
centos 9.stream
crimson-rados/rbd/{clusters/fixed-1 crimson-supported-all-distro/centos_latest crimson_qa_overrides deploy/ceph tasks/rbd_api_tests}
Command failed (workunit test rbd/crimson/test_crimson_librbd.sh) on smithi049 with status 124: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=a83d484198747052d1ed9eeed0a72e20b007e411 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 CEPH_MNT=/home/ubuntu/cephtest/mnt.0 RBD_FEATURES=61 CRIMSON_COMPAT=1 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/rbd/crimson/test_crimson_librbd.sh'
wip-crimson-scrub-testing-2023-10-16
wip-crimson-scrub-testing-2023-10-16
main
smithi
centos 9.stream
crimson-rados/singleton/{all/osd-backfill crimson-supported-all-distro/centos_latest crimson_qa_overrides objectstore/bluestore rados}
wip-crimson-scrub-testing-2023-10-16
wip-crimson-scrub-testing-2023-10-16
main
smithi
centos 9.stream
crimson-rados/thrash/{0-size-min-size-overrides/3-size-2-min-size 1-pg-log-overrides/normal_pg_log 2-recovery-overrides/{default} clusters/{fixed-2} crimson-supported-all-distro/centos_latest crimson_qa_overrides deploy/ceph objectstore/bluestore thrashers/default thrashosds-health workloads/admin_socket_objecter_requests}
wip-crimson-scrub-testing-2023-10-16
wip-crimson-scrub-testing-2023-10-16
main
smithi
centos 9.stream
crimson-rados/perf/{clusters/{fixed-2} crimson-supported-all-distro/centos_latest crimson_qa_overrides deploy/ceph objectstore/bluestore settings/optimized workloads/fio_4K_rand_rw}
wip-crimson-scrub-testing-2023-10-16
wip-crimson-scrub-testing-2023-10-16
main
smithi
centos 9.stream
crimson-rados/thrash/{0-size-min-size-overrides/3-size-2-min-size 1-pg-log-overrides/normal_pg_log 2-recovery-overrides/{default} clusters/{fixed-2} crimson-supported-all-distro/centos_latest crimson_qa_overrides deploy/ceph objectstore/bluestore thrashers/default thrashosds-health workloads/pool-snaps-few-objects}
hit max job timeout
wip-crimson-scrub-testing-2023-10-16
wip-crimson-scrub-testing-2023-10-16
main
smithi
centos 9.stream
crimson-rados/perf/{clusters/{fixed-2} crimson-supported-all-distro/centos_latest crimson_qa_overrides deploy/ceph objectstore/bluestore settings/optimized workloads/fio_4M_rand_read}
"2023-10-17T07:17:50.565784+0000 mon.a (mon.0) 132 : cluster [WRN] Health check failed: Reduced data availability: 3 pgs inactive (PG_AVAILABILITY)" in cluster log
wip-crimson-scrub-testing-2023-10-16
wip-crimson-scrub-testing-2023-10-16
main
smithi
centos 9.stream
crimson-rados/rbd/{clusters/fixed-1 crimson-supported-all-distro/centos_latest crimson_qa_overrides deploy/ceph tasks/rbd_api_tests_old_format}
Command failed (workunit test rbd/crimson/test_crimson_librbd.sh) on smithi071 with status 124: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=a83d484198747052d1ed9eeed0a72e20b007e411 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 CEPH_MNT=/home/ubuntu/cephtest/mnt.0 CRIMSON_COMPAT=1 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/rbd/crimson/test_crimson_librbd.sh'
wip-crimson-scrub-testing-2023-10-16
wip-crimson-scrub-testing-2023-10-16
main
smithi
centos 9.stream
crimson-rados/thrash/{0-size-min-size-overrides/3-size-2-min-size 1-pg-log-overrides/normal_pg_log 2-recovery-overrides/{default} clusters/{fixed-2} crimson-supported-all-distro/centos_latest crimson_qa_overrides deploy/ceph objectstore/bluestore thrashers/default thrashosds-health workloads/radosbench-high-concurrency}
wip-crimson-scrub-testing-2023-10-16
wip-crimson-scrub-testing-2023-10-16
main
smithi
centos 9.stream
crimson-rados/perf/{clusters/{fixed-2} crimson-supported-all-distro/centos_latest crimson_qa_overrides deploy/ceph objectstore/bluestore settings/optimized workloads/fio_4M_rand_rw}
"2023-10-17T07:17:57.290219+0000 mon.a (mon.0) 126 : cluster [WRN] Health check failed: Reduced data availability: 20 pgs inactive (PG_AVAILABILITY)" in cluster log
wip-crimson-scrub-testing-2023-10-16
wip-crimson-scrub-testing-2023-10-16
main
smithi
centos 9.stream
crimson-rados/thrash/{0-size-min-size-overrides/3-size-2-min-size 1-pg-log-overrides/normal_pg_log 2-recovery-overrides/{default} clusters/{fixed-2} crimson-supported-all-distro/centos_latest crimson_qa_overrides deploy/ceph objectstore/bluestore thrashers/default thrashosds-health workloads/radosbench}
wip-crimson-scrub-testing-2023-10-16
wip-crimson-scrub-testing-2023-10-16
main
smithi
centos 9.stream
crimson-rados/perf/{clusters/{fixed-2} crimson-supported-all-distro/centos_latest crimson_qa_overrides deploy/ceph objectstore/bluestore settings/optimized workloads/fio_4M_rand_write}
wip-crimson-scrub-testing-2023-10-16
wip-crimson-scrub-testing-2023-10-16
main
smithi
centos 9.stream
crimson-rados/basic/{clusters/fixed-2 crimson-supported-all-distro/centos_latest crimson_qa_overrides deploy/ceph tasks/rados_python}
wip-crimson-scrub-testing-2023-10-16
wip-crimson-scrub-testing-2023-10-16
main
smithi
centos 9.stream
crimson-rados/rbd/{clusters/fixed-1 crimson-supported-all-distro/centos_latest crimson_qa_overrides deploy/ceph tasks/rbd_cls_tests}
Command failed (workunit test cls/test_cls_lock.sh) on smithi084 with status 124: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=a83d484198747052d1ed9eeed0a72e20b007e411 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 CEPH_MNT=/home/ubuntu/cephtest/mnt.0 CRIMSON_COMPAT=1 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/cls/test_cls_lock.sh'
wip-crimson-scrub-testing-2023-10-16
wip-crimson-scrub-testing-2023-10-16
main
smithi
centos 9.stream
crimson-rados/thrash/{0-size-min-size-overrides/3-size-2-min-size 1-pg-log-overrides/normal_pg_log 2-recovery-overrides/{default} clusters/{fixed-2} crimson-supported-all-distro/centos_latest crimson_qa_overrides deploy/ceph objectstore/bluestore thrashers/default thrashosds-health workloads/small-objects-balanced}
Command crashed: 'CEPH_CLIENT_ID=0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph_test_rados --balance-reads --max-ops 400000 --objects 1024 --max-in-flight 64 --size 4000000 --min-stride-size 400000 --max-stride-size 800000 --max-seconds 600 --op read 100 --op write 50 --op delete 50 --op snap_create 50 --op snap_remove 50 --op rollback 0 --op setattr 25 --op rmattr 25 --op copy_from 0 --op write_excl 50 --pool unique_pool_0'
wip-crimson-scrub-testing-2023-10-16
wip-crimson-scrub-testing-2023-10-16
main
smithi
centos 9.stream
crimson-rados/perf/{clusters/{fixed-2} crimson-supported-all-distro/centos_latest crimson_qa_overrides deploy/ceph objectstore/bluestore settings/optimized workloads/radosbench_4K_rand_read}
"2023-10-17T07:42:26.162809+0000 mon.a (mon.0) 155 : cluster [WRN] Health check failed: Degraded data redundancy: 106/7804 objects degraded (1.358%), 8 pgs degraded (PG_DEGRADED)" in cluster log
wip-crimson-scrub-testing-2023-10-16
wip-crimson-scrub-testing-2023-10-16
main
smithi
centos 9.stream
crimson-rados/thrash/{0-size-min-size-overrides/3-size-2-min-size 1-pg-log-overrides/normal_pg_log 2-recovery-overrides/{default} clusters/{fixed-2} crimson-supported-all-distro/centos_latest crimson_qa_overrides deploy/ceph objectstore/bluestore thrashers/default thrashosds-health workloads/small-objects-localized}
Command crashed: 'CEPH_CLIENT_ID=0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph_test_rados --localize-reads --max-ops 400000 --objects 1024 --max-in-flight 64 --size 4000000 --min-stride-size 400000 --max-stride-size 800000 --max-seconds 600 --op read 100 --op write 50 --op delete 50 --op snap_create 50 --op snap_remove 50 --op rollback 0 --op setattr 25 --op rmattr 25 --op copy_from 0 --op write_excl 50 --pool unique_pool_0'
wip-crimson-scrub-testing-2023-10-16
wip-crimson-scrub-testing-2023-10-16
main
smithi
centos 9.stream
crimson-rados/perf/{clusters/{fixed-2} crimson-supported-all-distro/centos_latest crimson_qa_overrides deploy/ceph objectstore/bluestore settings/optimized workloads/radosbench_4K_seq_read}
"2023-10-17T07:43:04.099027+0000 mon.a (mon.0) 153 : cluster [WRN] Health check failed: Degraded data redundancy: 109/9096 objects degraded (1.198%), 6 pgs degraded (PG_DEGRADED)" in cluster log
wip-crimson-scrub-testing-2023-10-16
wip-crimson-scrub-testing-2023-10-16
main
smithi
centos 9.stream
crimson-rados/rbd/{clusters/fixed-1 crimson-supported-all-distro/centos_latest crimson_qa_overrides deploy/ceph tasks/rbd_lock_and_fence}
wip-crimson-scrub-testing-2023-10-16
wip-crimson-scrub-testing-2023-10-16
main
smithi
centos 9.stream
crimson-rados/perf/{clusters/{fixed-2} crimson-supported-all-distro/centos_latest crimson_qa_overrides deploy/ceph objectstore/bluestore settings/optimized workloads/radosbench_4M_rand_read}
"2023-10-17T07:46:20.129720+0000 mon.a (mon.0) 150 : cluster [WRN] Health check failed: Degraded data redundancy: 8/1662 objects degraded (0.481%), 3 pgs degraded (PG_DEGRADED)" in cluster log
wip-crimson-scrub-testing-2023-10-16
wip-crimson-scrub-testing-2023-10-16
main
smithi
centos 9.stream
crimson-rados/thrash/{0-size-min-size-overrides/3-size-2-min-size 1-pg-log-overrides/normal_pg_log 2-recovery-overrides/{default} clusters/{fixed-2} crimson-supported-all-distro/centos_latest crimson_qa_overrides deploy/ceph objectstore/bluestore thrashers/default thrashosds-health workloads/small-objects}
hit max job timeout
wip-crimson-scrub-testing-2023-10-16
wip-crimson-scrub-testing-2023-10-16
main
smithi
centos 9.stream
crimson-rados/perf/{clusters/{fixed-2} crimson-supported-all-distro/centos_latest crimson_qa_overrides deploy/ceph objectstore/bluestore settings/optimized workloads/radosbench_4M_seq_read}
"2023-10-17T07:52:54.844054+0000 mon.a (mon.0) 155 : cluster [WRN] Health check failed: Degraded data redundancy: 25/1412 objects degraded (1.771%), 9 pgs degraded (PG_DEGRADED)" in cluster log
wip-crimson-scrub-testing-2023-10-16
wip-crimson-scrub-testing-2023-10-16
main
smithi
centos 9.stream
crimson-rados/thrash/{0-size-min-size-overrides/3-size-2-min-size 1-pg-log-overrides/normal_pg_log 2-recovery-overrides/{default} clusters/{fixed-2} crimson-supported-all-distro/centos_latest crimson_qa_overrides deploy/ceph objectstore/bluestore thrashers/default thrashosds-health workloads/snaps-few-objects-balanced}
hit max job timeout
wip-crimson-scrub-testing-2023-10-16
wip-crimson-scrub-testing-2023-10-16
main
smithi
centos 9.stream
crimson-rados/basic/{clusters/fixed-2 crimson-supported-all-distro/centos_latest crimson_qa_overrides deploy/ceph tasks/readwrite}
"2023-10-17T08:00:59.382322+0000 mon.a (mon.0) 153 : cluster [WRN] Health check failed: 1 slow ops, oldest one blocked for 32 sec, mon.a has slow ops (SLOW_OPS)" in cluster log