ID
Status
Ceph Branch
Suite Branch
Teuthology Branch
Machine
OS
Nodes
Description
Failure Reason
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
ubuntu 18.04
rados/singleton-nomsgr/{all/full-tiering.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}}
Exiting scrub checking -- not all pgs scrubbed.
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
centos 7.6
rados/monthrash/{ceph.yaml clusters/9-mons.yaml msgr-failures/mon-delay.yaml msgr/async.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{centos_7.yaml} thrashers/one.yaml workloads/rados_mon_workunits.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
centos 7.6
rados/thrash-old-clients/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-install/mimic-v1only.yaml backoff/peering.yaml ceph.yaml clusters/{openstack.yaml three-plus-one.yaml} d-balancer/off.yaml distro$/{centos_latest.yaml} msgr-failures/fastclose.yaml rados.yaml thrashers/pggrow.yaml thrashosds-health.yaml workloads/rbd_cls.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
rhel 7.6
rados/thrash-erasure-code/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} fast/fast.yaml msgr-failures/osd-delay.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml recovery-overrides/{more-active-recovery.yaml} supported-random-distro$/{rhel_7.yaml} thrashers/careful.yaml thrashosds-health.yaml workloads/ec-small-objects.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
rhel 7.6
rados/objectstore/{backends/filestore-idempotent-aio-journal.yaml supported-random-distro$/{rhel_7.yaml}}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
centos 7.6
rados/standalone/{supported-random-distro$/{centos_7.yaml} workloads/misc.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
ubuntu 18.04
rados/thrash-erasure-code-shec/{ceph.yaml clusters/{fixed-4.yaml openstack.yaml} msgr-failures/few.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml recovery-overrides/{default.yaml} supported-random-distro$/{ubuntu_latest.yaml} thrashers/careful.yaml thrashosds-health.yaml workloads/ec-rados-plugin=shec-k=4-m=3-c=2.yaml}
SSH connection to smithi012 was lost: 'CEPH_CLIENT_ID=0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph_test_rados --no-omap --ec-pool --max-ops 400 --objects 50 --max-in-flight 16 --size 4000000 --min-stride-size 400000 --max-stride-size 800000 --max-seconds 0 --op snap_remove 50 --op snap_create 50 --op rollback 50 --op setattr 25 --op read 100 --op copy_from 50 --op write 0 --op rmattr 25 --op append 100 --op delete 50 --pool unique_pool_0'
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
rhel 7.6
rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml 2-recovery-overrides/{default.yaml} backoff/peering_and_degraded.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/off.yaml msgr-failures/few.yaml msgr/async-v2only.yaml objectstore/bluestore-stupid.yaml rados.yaml supported-random-distro$/{rhel_7.yaml} thrashers/morepggrow.yaml thrashosds-health.yaml workloads/cache-pool-snaps.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
centos 7.6
rados/singleton/{all/admin-socket.yaml msgr-failures/many.yaml msgr/async.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{centos_7.yaml}}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
ubuntu 18.04
rados/basic/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/many.yaml msgr/async.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} tasks/rados_stress_watch.yaml}
Command failed (workunit test rados/stress_watch.sh) on smithi140 with status 124: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=475286020efbcbe469ec6169e1c477e69c61c6c6 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/rados/stress_watch.sh'
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
centos 7.6
rados/mgr/{clusters/{2-node-mgr.yaml openstack.yaml} debug/mgr.yaml objectstore/bluestore-low-osd-mem-target.yaml supported-random-distro$/{centos_7.yaml} tasks/progress.yaml}
Test failure: test_osd_came_back (tasks.mgr.test_progress.TestProgress)
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
centos 7.6
rados/perf/{ceph.yaml objectstore/bluestore-bitmap.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{centos_7.yaml} workloads/radosbench_4M_rand_read.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
rhel 7.6
rados/multimon/{clusters/6.yaml msgr-failures/many.yaml msgr/async.yaml no_pools.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml supported-random-distro$/{rhel_7.yaml} tasks/mon_recovery.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
 
rados/verify/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-thrash/default/{default.yaml thrashosds-health.yaml} msgr-failures/few.yaml msgr/async.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml tasks/rados_cls_all.yaml validater/lockdep.yaml}
Command failed (workunit test cls/test_cls_journal.sh) on smithi088 with status 124: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=475286020efbcbe469ec6169e1c477e69c61c6c6 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/cls/test_cls_journal.sh'
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
rhel 7.6
rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml 2-recovery-overrides/{default.yaml} backoff/normal.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/upmap.yaml msgr-failures/osd-delay.yaml msgr/async.yaml objectstore/filestore-xfs.yaml rados.yaml supported-random-distro$/{rhel_7.yaml} thrashers/none.yaml thrashosds-health.yaml workloads/cache-snaps.yaml}
"2019-08-19T18:56:15.874421+0000 mon.a (mon.0) 776 : cluster [WRN] Health check failed: 1 daemons have recently crashed (RECENT_CRASH)" in cluster log
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
rhel 7.6
rados/singleton/{all/deduptool.yaml msgr-failures/few.yaml msgr/async-v1only.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
rhel 7.6
rados/singleton-nomsgr/{all/health-warnings.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
centos 7.6
rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml 2-recovery-overrides/{default.yaml} backoff/peering.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/crush-compat.yaml msgr-failures/fastclose.yaml msgr/async-v1only.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{centos_7.yaml} thrashers/pggrow.yaml thrashosds-health.yaml workloads/cache.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
centos 7.6
rados/singleton/{all/divergent_priors.yaml msgr-failures/many.yaml msgr/async-v2only.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml supported-random-distro$/{centos_7.yaml}}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
rhel 7.6
rados/perf/{ceph.yaml objectstore/bluestore-comp.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{rhel_7.yaml} workloads/radosbench_4M_seq_read.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
centos 7.6
rados/thrash-old-clients/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-install/mimic.yaml backoff/peering_and_degraded.yaml ceph.yaml clusters/{openstack.yaml three-plus-one.yaml} d-balancer/crush-compat.yaml distro$/{centos_latest.yaml} msgr-failures/few.yaml rados.yaml thrashers/careful.yaml thrashosds-health.yaml workloads/snaps-few-objects.yaml}
psutil.NoSuchProcess process no longer exists (pid=16210)
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
rhel 7.6
rados/thrash-erasure-code-isa/{arch/x86_64.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/fastclose.yaml objectstore/bluestore-stupid.yaml rados.yaml recovery-overrides/{more-active-recovery.yaml} supported-random-distro$/{rhel_7.yaml} thrashers/careful.yaml thrashosds-health.yaml workloads/ec-rados-plugin=isa-k=2-m=1.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
ubuntu 18.04
rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml 2-recovery-overrides/{more-active-recovery.yaml} backoff/peering_and_degraded.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/off.yaml msgr-failures/few.yaml msgr/async-v2only.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} thrashers/careful.yaml thrashosds-health.yaml workloads/dedup_tier.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
ubuntu 18.04
rados/basic/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/few.yaml msgr/async-v1only.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} tasks/rados_striper.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
rhel 7.6
rados/thrash-erasure-code-big/{ceph.yaml cluster/{12-osds.yaml openstack.yaml} msgr-failures/few.yaml objectstore/bluestore-comp.yaml rados.yaml recovery-overrides/{more-active-recovery.yaml} supported-random-distro$/{rhel_7.yaml} thrashers/mapgap.yaml thrashosds-health.yaml workloads/ec-rados-plugin=jerasure-k=4-m=2.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
centos 7.6
rados/thrash-erasure-code-overwrites/{bluestore-bitmap.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} fast/normal.yaml msgr-failures/osd-delay.yaml rados.yaml recovery-overrides/{default.yaml} supported-random-distro$/{centos_7.yaml} thrashers/minsize_recovery.yaml thrashosds-health.yaml workloads/ec-snaps-few-objects-overwrites.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
ubuntu 18.04
rados/singleton/{all/divergent_priors2.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/bluestore-stupid.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
rhel 7.6
rados/mgr/{clusters/{2-node-mgr.yaml openstack.yaml} debug/mgr.yaml objectstore/bluestore-stupid.yaml supported-random-distro$/{rhel_7.yaml} tasks/prometheus.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
rhel 7.6
rados/singleton-nomsgr/{all/large-omap-object-warnings.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}}
"2019-08-19T19:00:55.177775+0000 mon.a (mon.0) 82 : cluster [WRN] Health check failed: 1 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED)" in cluster log
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
ubuntu 18.04
rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml 2-recovery-overrides/{more-active-recovery.yaml} backoff/normal.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/upmap.yaml msgr-failures/osd-delay.yaml msgr/async.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} thrashers/default.yaml thrashosds-health.yaml workloads/pool-snaps-few-objects.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
rhel 7.6
rados/objectstore/{backends/filestore-idempotent.yaml supported-random-distro$/{rhel_7.yaml}}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
ubuntu 18.04
rados/thrash-erasure-code/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} fast/normal.yaml msgr-failures/fastclose.yaml objectstore/bluestore-stupid.yaml rados.yaml recovery-overrides/{more-active-recovery.yaml} supported-random-distro$/{ubuntu_latest.yaml} thrashers/default.yaml thrashosds-health.yaml workloads/ec-rados-plugin=clay-k=4-m=2.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
centos 7.6
rados/singleton/{all/dump-stuck.yaml msgr-failures/many.yaml msgr/async-v1only.yaml objectstore/filestore-xfs.yaml rados.yaml supported-random-distro$/{centos_7.yaml}}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
centos 7.6
rados/perf/{ceph.yaml objectstore/bluestore-low-osd-mem-target.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{centos_7.yaml} workloads/radosbench_4M_write.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
ubuntu 18.04
rados/monthrash/{ceph.yaml clusters/3-mons.yaml msgr-failures/few.yaml msgr/async-v1only.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} thrashers/sync-many.yaml workloads/snaps-few-objects.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
ubuntu 18.04
rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml 2-recovery-overrides/{default.yaml} backoff/peering.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/crush-compat.yaml msgr-failures/fastclose.yaml msgr/async-v1only.yaml objectstore/bluestore-stupid.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} thrashers/mapgap.yaml thrashosds-health.yaml workloads/rados_api_tests.yaml}
Command failed (workunit test rados/test.sh) on smithi191 with status 124: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=475286020efbcbe469ec6169e1c477e69c61c6c6 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/rados/test.sh'
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
ubuntu 18.04
rados/singleton/{all/ec-lost-unfound.yaml msgr-failures/few.yaml msgr/async-v2only.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
centos 7.6
rados/thrash-old-clients/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-install/nautilus-v1only.yaml backoff/normal.yaml ceph.yaml clusters/{openstack.yaml three-plus-one.yaml} d-balancer/off.yaml distro$/{centos_latest.yaml} msgr-failures/osd-delay.yaml rados.yaml thrashers/default.yaml thrashosds-health.yaml workloads/test_rbd_api.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
rhel 7.6
rados/basic/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/many.yaml msgr/async-v2only.yaml objectstore/bluestore-stupid.yaml rados.yaml supported-random-distro$/{rhel_7.yaml} tasks/rados_workunit_loadgen_big.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
ubuntu 18.04
rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml 2-recovery-overrides/{default.yaml} backoff/peering_and_degraded.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/off.yaml msgr-failures/few.yaml msgr/async-v2only.yaml objectstore/filestore-xfs.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} thrashers/morepggrow.yaml thrashosds-health.yaml workloads/radosbench.yaml}
reached maximum tries (500) after waiting for 3000 seconds
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
rhel 7.6
rados/singleton-bluestore/{all/cephtool.yaml msgr-failures/many.yaml msgr/async-v2only.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}}
Command failed (workunit test cephtool/test.sh) on smithi092 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=475286020efbcbe469ec6169e1c477e69c61c6c6 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/cephtool/test.sh'
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
ubuntu 18.04
rados/singleton-nomsgr/{all/lazy_omap_stats_output.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
rhel 7.6
rados/standalone/{supported-random-distro$/{rhel_7.yaml} workloads/mon.yaml}
Command failed (workunit test mon/health-mute.sh) on smithi169 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=475286020efbcbe469ec6169e1c477e69c61c6c6 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/standalone/mon/health-mute.sh'
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
ubuntu 18.04
rados/upgrade/nautilus-x-singleton/{0-cluster/{openstack.yaml start.yaml} 1-install/nautilus.yaml 2-partial-upgrade/firsthalf.yaml 3-thrash/default.yaml 4-workload/{rbd-cls.yaml rbd-import-export.yaml readwrite.yaml snaps-few-objects.yaml} 5-workload/{radosbench.yaml rbd_api.yaml} 6-finish-upgrade.yaml 7-octopus.yaml 8-workload/{rbd-python.yaml rgw-swift.yaml snaps-many-objects.yaml} bluestore-bitmap.yaml supported-random-distro$/{ubuntu_latest.yaml} thrashosds-health.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
rhel 7.6
rados/thrash-erasure-code-shec/{ceph.yaml clusters/{fixed-4.yaml openstack.yaml} msgr-failures/osd-delay.yaml objectstore/bluestore-stupid.yaml rados.yaml recovery-overrides/{default.yaml} supported-random-distro$/{rhel_7.yaml} thrashers/default.yaml thrashosds-health.yaml workloads/ec-rados-plugin=shec-k=4-m=3-c=2.yaml}
Command failed on smithi012 with status 1: 'sudo package-cleanup -y --oldkernels'
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
ubuntu 18.04
rados/mgr/{clusters/{2-node-mgr.yaml openstack.yaml} debug/mgr.yaml objectstore/filestore-xfs.yaml supported-random-distro$/{ubuntu_latest.yaml} tasks/ssh_orchestrator.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
rhel 7.6
rados/perf/{ceph.yaml objectstore/bluestore-stupid.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{rhel_7.yaml} workloads/radosbench_omap_write.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
rhel 7.6
rados/singleton/{all/erasure-code-nonregression.yaml msgr-failures/many.yaml msgr/async.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
ubuntu 18.04
rados/multimon/{clusters/9.yaml msgr-failures/few.yaml msgr/async-v1only.yaml no_pools.yaml objectstore/bluestore-stupid.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} tasks/mon_clock_no_skews.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
centos 
rados/verify/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-thrash/none.yaml msgr-failures/few.yaml msgr/async-v1only.yaml objectstore/bluestore-stupid.yaml rados.yaml tasks/mon_recovery.yaml validater/valgrind.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
rhel 7.6
rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml 2-recovery-overrides/{more-active-recovery.yaml} backoff/normal.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/upmap.yaml msgr-failures/osd-delay.yaml msgr/async.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{rhel_7.yaml} thrashers/none.yaml thrashosds-health.yaml workloads/redirect.yaml}
"2019-08-19T19:24:17.893799+0000 mon.a (mon.0) 339 : cluster [WRN] Health check failed: 2 daemons have recently crashed (RECENT_CRASH)" in cluster log
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
centos 7.6
rados/singleton/{all/lost-unfound-delete.yaml msgr-failures/few.yaml msgr/async-v1only.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml supported-random-distro$/{centos_7.yaml}}
timed out waiting for admin_socket to appear after osd.1 restart
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
ubuntu 18.04
rados/thrash-erasure-code-isa/{arch/x86_64.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/few.yaml objectstore/filestore-xfs.yaml rados.yaml recovery-overrides/{more-active-recovery.yaml} supported-random-distro$/{ubuntu_latest.yaml} thrashers/default.yaml thrashosds-health.yaml workloads/ec-rados-plugin=isa-k=2-m=1.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
centos 7.6
rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml 2-recovery-overrides/{default.yaml} backoff/peering.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/crush-compat.yaml msgr-failures/fastclose.yaml msgr/async-v1only.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{centos_7.yaml} thrashers/pggrow.yaml thrashosds-health.yaml workloads/redirect_promote_tests.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
ubuntu 18.04
rados/objectstore/{backends/fusestore.yaml supported-random-distro$/{ubuntu_latest.yaml}}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
ubuntu 18.04
rados/thrash-erasure-code-big/{ceph.yaml cluster/{12-osds.yaml openstack.yaml} msgr-failures/osd-delay.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml recovery-overrides/{more-active-recovery.yaml} supported-random-distro$/{ubuntu_latest.yaml} thrashers/morepggrow.yaml thrashosds-health.yaml workloads/ec-rados-plugin=lrc-k=4-m=2-l=3.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
centos 7.6
rados/thrash-erasure-code-overwrites/{bluestore-bitmap.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} fast/fast.yaml msgr-failures/fastclose.yaml rados.yaml recovery-overrides/{default.yaml} supported-random-distro$/{centos_7.yaml} thrashers/morepggrow.yaml thrashosds-health.yaml workloads/ec-pool-snaps-few-objects-overwrites.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
ubuntu 18.04
rados/perf/{ceph.yaml objectstore/filestore-xfs.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{ubuntu_latest.yaml} workloads/sample_fio.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
rhel 7.6
rados/singleton-nomsgr/{all/librados_hello_world.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
rhel 7.6
rados/singleton/{all/lost-unfound.yaml msgr-failures/many.yaml msgr/async-v2only.yaml objectstore/bluestore-stupid.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}}
timed out waiting for admin_socket to appear after osd.1 restart
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
ubuntu 18.04
rados/basic/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/few.yaml msgr/async.yaml objectstore/filestore-xfs.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} tasks/rados_workunit_loadgen_mix.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
centos 7.6
rados/thrash-old-clients/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-install/nautilus-v2only.yaml backoff/peering.yaml ceph.yaml clusters/{openstack.yaml three-plus-one.yaml} d-balancer/crush-compat.yaml distro$/{centos_latest.yaml} msgr-failures/fastclose.yaml rados.yaml thrashers/mapgap.yaml thrashosds-health.yaml workloads/cache-snaps.yaml}
psutil.NoSuchProcess process no longer exists (pid=25868)
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
centos 7.6
rados/thrash-erasure-code/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} fast/fast.yaml msgr-failures/few.yaml objectstore/filestore-xfs.yaml rados.yaml recovery-overrides/{default.yaml} supported-random-distro$/{centos_7.yaml} thrashers/fastread.yaml thrashosds-health.yaml workloads/ec-rados-plugin=jerasure-k=2-m=1.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
centos 7.6
rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml 2-recovery-overrides/{default.yaml} backoff/peering_and_degraded.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/off.yaml msgr-failures/few.yaml msgr/async-v2only.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml supported-random-distro$/{centos_7.yaml} thrashers/careful.yaml thrashosds-health.yaml workloads/redirect_set_object.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
centos 7.6
rados/mgr/{clusters/{2-node-mgr.yaml openstack.yaml} debug/mgr.yaml objectstore/bluestore-bitmap.yaml supported-random-distro$/{centos_7.yaml} tasks/workunits.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
rhel 7.6
rados/singleton/{all/max-pg-per-osd.from-mon.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/filestore-xfs.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
rhel 7.6
rados/monthrash/{ceph.yaml clusters/9-mons.yaml msgr-failures/mon-delay.yaml msgr/async-v2only.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml supported-random-distro$/{rhel_7.yaml} thrashers/sync.yaml workloads/pool-create-delete.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
centos 7.6
rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml 2-recovery-overrides/{default.yaml} backoff/normal.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/upmap.yaml msgr-failures/osd-delay.yaml msgr/async.yaml objectstore/bluestore-stupid.yaml rados.yaml supported-random-distro$/{centos_7.yaml} thrashers/default.yaml thrashosds-health.yaml workloads/set-chunk-promote-flush.yaml}
psutil.NoSuchProcess process no longer exists (pid=5452)
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
rhel 7.6
rados/perf/{ceph.yaml objectstore/bluestore-bitmap.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{rhel_7.yaml} workloads/sample_radosbench.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
centos 7.6
rados/singleton/{all/max-pg-per-osd.from-primary.yaml msgr-failures/many.yaml msgr/async-v1only.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{centos_7.yaml}}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
centos 7.6
rados/singleton-nomsgr/{all/msgr.yaml rados.yaml supported-random-distro$/{centos_7.yaml}}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
rhel 7.6
rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml 2-recovery-overrides/{more-active-recovery.yaml} backoff/peering.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/crush-compat.yaml msgr-failures/fastclose.yaml msgr/async-v1only.yaml objectstore/filestore-xfs.yaml rados.yaml supported-random-distro$/{rhel_7.yaml} thrashers/mapgap.yaml thrashosds-health.yaml workloads/set-chunks-read.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
rhel 7.6
rados/basic/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/many.yaml msgr/async-v1only.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{rhel_7.yaml} tasks/rados_workunit_loadgen_mostlyread.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
centos 7.6
rados/singleton/{all/max-pg-per-osd.from-replica.yaml msgr-failures/few.yaml msgr/async-v2only.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{centos_7.yaml}}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
centos 7.6
rados/thrash-old-clients/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-install/nautilus.yaml backoff/peering_and_degraded.yaml ceph.yaml clusters/{openstack.yaml three-plus-one.yaml} d-balancer/off.yaml distro$/{centos_latest.yaml} msgr-failures/few.yaml rados.yaml thrashers/morepggrow.yaml thrashosds-health.yaml workloads/radosbench.yaml}
reached maximum tries (500) after waiting for 3000 seconds
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
ubuntu 18.04
rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml 2-recovery-overrides/{default.yaml} backoff/peering_and_degraded.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/off.yaml msgr-failures/few.yaml msgr/async-v2only.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} thrashers/morepggrow.yaml thrashosds-health.yaml workloads/small-objects.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
rhel 7.6
rados/objectstore/{backends/keyvaluedb.yaml supported-random-distro$/{rhel_7.yaml}}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
rhel 7.6
rados/standalone/{supported-random-distro$/{rhel_7.yaml} workloads/osd.yaml}
psutil.NoSuchProcess process no longer exists (pid=25206)
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
centos 7.6
rados/mgr/{clusters/{2-node-mgr.yaml openstack.yaml} debug/mgr.yaml objectstore/bluestore-bitmap.yaml supported-random-distro$/{centos_7.yaml} tasks/crash.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
ubuntu 18.04
rados/perf/{ceph.yaml objectstore/bluestore-bitmap.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{ubuntu_latest.yaml} workloads/cosbench_64K_read_write.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
rhel 7.6
rados/thrash-erasure-code-shec/{ceph.yaml clusters/{fixed-4.yaml openstack.yaml} msgr-failures/fastclose.yaml objectstore/filestore-xfs.yaml rados.yaml recovery-overrides/{more-active-recovery.yaml} supported-random-distro$/{rhel_7.yaml} thrashers/careful.yaml thrashosds-health.yaml workloads/ec-rados-plugin=shec-k=4-m=3-c=2.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
ubuntu 18.04
rados/thrash-erasure-code-isa/{arch/x86_64.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/osd-delay.yaml objectstore/bluestore-bitmap.yaml rados.yaml recovery-overrides/{default.yaml} supported-random-distro$/{ubuntu_latest.yaml} thrashers/mapgap.yaml thrashosds-health.yaml workloads/ec-rados-plugin=isa-k=2-m=1.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
centos 7.6
rados/singleton/{all/mon-auth-caps.yaml msgr-failures/many.yaml msgr/async.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml supported-random-distro$/{centos_7.yaml}}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
ubuntu 18.04
rados/multimon/{clusters/21.yaml msgr-failures/many.yaml msgr/async-v2only.yaml no_pools.yaml objectstore/filestore-xfs.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} tasks/mon_clock_with_skews.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
 
rados/verify/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-thrash/default/{default.yaml thrashosds-health.yaml} msgr-failures/few.yaml msgr/async-v2only.yaml objectstore/filestore-xfs.yaml rados.yaml tasks/rados_api_tests.yaml validater/lockdep.yaml}
Command failed (workunit test rados/test.sh) on smithi037 with status 124: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=475286020efbcbe469ec6169e1c477e69c61c6c6 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 6h /home/ubuntu/cephtest/clone.client.0/qa/workunits/rados/test.sh'
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
ubuntu 18.04
rados/singleton-nomsgr/{all/multi-backfill-reject.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}}
"2019-08-19T19:43:17.056648+0000 mon.a (mon.0) 244 : cluster [WRN] Health check failed: 2 daemons have recently crashed (RECENT_CRASH)" in cluster log
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
rhel 7.6
rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml 2-recovery-overrides/{default.yaml} backoff/normal.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/upmap.yaml msgr-failures/osd-delay.yaml msgr/async.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{rhel_7.yaml} thrashers/none.yaml thrashosds-health.yaml workloads/snaps-few-objects.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
ubuntu 18.04
rados/thrash-erasure-code-big/{ceph.yaml cluster/{12-osds.yaml openstack.yaml} msgr-failures/fastclose.yaml objectstore/bluestore-stupid.yaml rados.yaml recovery-overrides/{default.yaml} supported-random-distro$/{ubuntu_latest.yaml} thrashers/pggrow.yaml thrashosds-health.yaml workloads/ec-rados-plugin=jerasure-k=4-m=2.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
ubuntu 18.04
rados/thrash-erasure-code-overwrites/{bluestore-bitmap.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} fast/normal.yaml msgr-failures/few.yaml rados.yaml recovery-overrides/{default.yaml} supported-random-distro$/{ubuntu_latest.yaml} thrashers/pggrow.yaml thrashosds-health.yaml workloads/ec-small-objects-fast-read-overwrites.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
ubuntu 18.04
rados/thrash-erasure-code/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} fast/normal.yaml msgr-failures/osd-delay.yaml objectstore/bluestore-bitmap.yaml rados.yaml recovery-overrides/{more-active-recovery.yaml} supported-random-distro$/{ubuntu_latest.yaml} thrashers/minsize_recovery.yaml thrashosds-health.yaml workloads/ec-rados-plugin=jerasure-k=3-m=1.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
centos 7.6
rados/singleton/{all/mon-config-key-caps.yaml msgr-failures/few.yaml msgr/async-v1only.yaml objectstore/bluestore-stupid.yaml rados.yaml supported-random-distro$/{centos_7.yaml}}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
rhel 7.6
rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml 2-recovery-overrides/{more-active-recovery.yaml} backoff/peering.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/crush-compat.yaml msgr-failures/fastclose.yaml msgr/async-v1only.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml supported-random-distro$/{rhel_7.yaml} thrashers/pggrow.yaml thrashosds-health.yaml workloads/write_fadvise_dontneed.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
rhel 7.6
rados/basic/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/few.yaml msgr/async-v2only.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{rhel_7.yaml} tasks/readwrite.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
rhel 7.6
rados/perf/{ceph.yaml objectstore/bluestore-comp.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{rhel_7.yaml} workloads/cosbench_64K_write.yaml}
Command failed on smithi124 with status 1: 'find /home/ubuntu/cephtest -ls ; rmdir -- /home/ubuntu/cephtest'
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
centos 7.6
rados/singleton/{all/mon-config-keys.yaml msgr-failures/many.yaml msgr/async-v2only.yaml objectstore/filestore-xfs.yaml rados.yaml supported-random-distro$/{centos_7.yaml}}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
rhel 7.6
rados/monthrash/{ceph.yaml clusters/3-mons.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/bluestore-stupid.yaml rados.yaml supported-random-distro$/{rhel_7.yaml} thrashers/force-sync-many.yaml workloads/rados_5925.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
centos 7.6
rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml 2-recovery-overrides/{default.yaml} backoff/peering_and_degraded.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/off.yaml msgr-failures/few.yaml msgr/async-v2only.yaml objectstore/bluestore-stupid.yaml rados.yaml supported-random-distro$/{centos_7.yaml} thrashers/careful.yaml thrashosds-health.yaml workloads/admin_socket_objecter_requests.yaml}
reached maximum tries (800) after waiting for 4800 seconds
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
centos 7.6
rados/thrash-old-clients/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-install/hammer.yaml backoff/normal.yaml ceph.yaml clusters/{openstack.yaml three-plus-one.yaml} d-balancer/crush-compat.yaml distro$/{centos_latest.yaml} msgr-failures/osd-delay.yaml rados.yaml thrashers/none.yaml thrashosds-health.yaml workloads/rbd_cls.yaml}
psutil.NoSuchProcess process no longer exists (pid=4904)
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
ubuntu 18.04
rados/mgr/{clusters/{2-node-mgr.yaml openstack.yaml} debug/mgr.yaml objectstore/bluestore-comp.yaml supported-random-distro$/{ubuntu_latest.yaml} tasks/dashboard.yaml}
Test failure: test_create_get_update_delete_w_tenant (tasks.mgr.dashboard.test_rgw.RgwBucketTest)
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
centos 7.6
rados/singleton-bluestore/{all/cephtool.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{centos_7.yaml}}
Command failed (workunit test cephtool/test.sh) on smithi068 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=475286020efbcbe469ec6169e1c477e69c61c6c6 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/cephtool/test.sh'
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
ubuntu 18.04
rados/singleton-nomsgr/{all/osd_stale_reads.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}}
'NoneType' object is not iterable
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
rhel 7.6
rados/singleton/{all/mon-config.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
ubuntu 18.04
rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml 2-recovery-overrides/{more-active-recovery.yaml} backoff/normal.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/upmap.yaml msgr-failures/osd-delay.yaml msgr/async.yaml objectstore/filestore-xfs.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} thrashers/default.yaml thrashosds-health.yaml workloads/cache-agent-big.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
centos 7.6
rados/perf/{ceph.yaml objectstore/bluestore-low-osd-mem-target.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{centos_7.yaml} workloads/fio_4K_rand_read.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
centos 7.6
rados/objectstore/{backends/objectcacher-stress.yaml supported-random-distro$/{centos_7.yaml}}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
centos 7.6
rados/basic/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/many.yaml msgr/async.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml supported-random-distro$/{centos_7.yaml} tasks/repair_test.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
rhel 7.6
rados/singleton/{all/osd-backfill.yaml msgr-failures/many.yaml msgr/async-v1only.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
ubuntu 18.04
rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml 2-recovery-overrides/{default.yaml} backoff/peering.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/crush-compat.yaml msgr-failures/fastclose.yaml msgr/async-v1only.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} thrashers/mapgap.yaml thrashosds-health.yaml workloads/cache-agent-small.yaml}
timed out waiting for admin_socket to appear after osd.1 restart
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
centos 7.6
rados/thrash-erasure-code-isa/{arch/x86_64.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/fastclose.yaml objectstore/bluestore-comp.yaml rados.yaml recovery-overrides/{more-active-recovery.yaml} supported-random-distro$/{centos_7.yaml} thrashers/morepggrow.yaml thrashosds-health.yaml workloads/ec-rados-plugin=isa-k=2-m=1.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
ubuntu 18.04
rados/singleton-nomsgr/{all/pool-access.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
rhel 7.6
rados/singleton/{all/osd-recovery-incomplete.yaml msgr-failures/few.yaml msgr/async-v2only.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}}
Command failed on smithi065 with status 6: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph tell osd.0 flush_pg_stats'
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
centos 7.6
rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml 2-recovery-overrides/{more-active-recovery.yaml} backoff/peering_and_degraded.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/off.yaml msgr-failures/few.yaml msgr/async-v2only.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{centos_7.yaml} thrashers/morepggrow.yaml thrashosds-health.yaml workloads/cache-pool-snaps-readproxy.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
rhel 7.6
rados/thrash-erasure-code/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} fast/fast.yaml msgr-failures/fastclose.yaml objectstore/bluestore-comp.yaml rados.yaml recovery-overrides/{default.yaml} supported-random-distro$/{rhel_7.yaml} thrashers/morepggrow.yaml thrashosds-health.yaml workloads/ec-radosbench.yaml}
reached maximum tries (800) after waiting for 4800 seconds
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
centos 7.6
rados/thrash-old-clients/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-install/jewel-v1only.yaml backoff/peering.yaml ceph.yaml clusters/{openstack.yaml three-plus-one.yaml} d-balancer/off.yaml distro$/{centos_latest.yaml} msgr-failures/fastclose.yaml rados.yaml thrashers/pggrow.yaml thrashosds-health.yaml workloads/snaps-few-objects.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
centos 7.6
rados/mgr/{clusters/{2-node-mgr.yaml openstack.yaml} debug/mgr.yaml objectstore/bluestore-low-osd-mem-target.yaml supported-random-distro$/{centos_7.yaml} tasks/failover.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
ubuntu 18.04
rados/perf/{ceph.yaml objectstore/bluestore-stupid.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{ubuntu_latest.yaml} workloads/fio_4K_rand_rw.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
ubuntu 18.04
rados/standalone/{supported-random-distro$/{ubuntu_latest.yaml} workloads/scrub.yaml}
Command failed (workunit test scrub/osd-scrub-repair.sh) on smithi162 with status 124: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=475286020efbcbe469ec6169e1c477e69c61c6c6 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/standalone/scrub/osd-scrub-repair.sh'
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
ubuntu 18.04
rados/thrash-erasure-code-big/{ceph.yaml cluster/{12-osds.yaml openstack.yaml} msgr-failures/few.yaml objectstore/filestore-xfs.yaml rados.yaml recovery-overrides/{default.yaml} supported-random-distro$/{ubuntu_latest.yaml} thrashers/careful.yaml thrashosds-health.yaml workloads/ec-rados-plugin=lrc-k=4-m=2-l=3.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
ubuntu 18.04
rados/thrash-erasure-code-overwrites/{bluestore-bitmap.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} fast/fast.yaml msgr-failures/osd-delay.yaml rados.yaml recovery-overrides/{default.yaml} supported-random-distro$/{ubuntu_latest.yaml} thrashers/careful.yaml thrashosds-health.yaml workloads/ec-small-objects-overwrites.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
rhel 7.6
rados/thrash-erasure-code-shec/{ceph.yaml clusters/{fixed-4.yaml openstack.yaml} msgr-failures/few.yaml objectstore/bluestore-bitmap.yaml rados.yaml recovery-overrides/{default.yaml} supported-random-distro$/{rhel_7.yaml} thrashers/default.yaml thrashosds-health.yaml workloads/ec-rados-plugin=shec-k=4-m=3-c=2.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
rhel 7.6
rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml 2-recovery-overrides/{default.yaml} backoff/normal.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/upmap.yaml msgr-failures/osd-delay.yaml msgr/async.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml supported-random-distro$/{rhel_7.yaml} thrashers/none.yaml thrashosds-health.yaml workloads/cache-pool-snaps.yaml}
"2019-08-19T20:24:23.154627+0000 mon.a (mon.0) 738 : cluster [WRN] Health check failed: 1 daemons have recently crashed (RECENT_CRASH)" in cluster log
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
ubuntu 18.04
rados/singleton/{all/osd-recovery.yaml msgr-failures/many.yaml msgr/async.yaml objectstore/bluestore-stupid.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
centos 7.6
rados/multimon/{clusters/3.yaml msgr-failures/few.yaml msgr/async.yaml no_pools.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{centos_7.yaml} tasks/mon_recovery.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
centos 
rados/verify/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-thrash/none.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/bluestore-bitmap.yaml rados.yaml tasks/rados_cls_all.yaml validater/valgrind.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
centos 7.6
rados/basic/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/few.yaml msgr/async-v1only.yaml objectstore/bluestore-stupid.yaml rados.yaml supported-random-distro$/{centos_7.yaml} tasks/rgw_snaps.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
rhel 7.6
rados/singleton-nomsgr/{all/recovery-unfound-found.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
ubuntu 18.04
rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml 2-recovery-overrides/{more-active-recovery.yaml} backoff/peering.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/crush-compat.yaml msgr-failures/fastclose.yaml msgr/async-v1only.yaml objectstore/bluestore-stupid.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} thrashers/pggrow.yaml thrashosds-health.yaml workloads/cache-snaps.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
ubuntu 18.04
rados/singleton/{all/peer.yaml msgr-failures/few.yaml msgr/async-v1only.yaml objectstore/filestore-xfs.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
ubuntu 18.04
rados/perf/{ceph.yaml objectstore/filestore-xfs.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{ubuntu_latest.yaml} workloads/fio_4M_rand_read.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
centos 7.6
rados/monthrash/{ceph.yaml clusters/9-mons.yaml msgr-failures/mon-delay.yaml msgr/async-v1only.yaml objectstore/filestore-xfs.yaml rados.yaml supported-random-distro$/{centos_7.yaml} thrashers/many.yaml workloads/rados_api_tests.yaml}
Command failed (workunit test rados/test.sh) on smithi114 with status 124: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=475286020efbcbe469ec6169e1c477e69c61c6c6 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/rados/test.sh'
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
rhel 7.6
rados/objectstore/{backends/objectstore.yaml supported-random-distro$/{rhel_7.yaml}}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
ubuntu 18.04
rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml 2-recovery-overrides/{default.yaml} backoff/peering_and_degraded.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/off.yaml msgr-failures/few.yaml msgr/async-v2only.yaml objectstore/filestore-xfs.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} thrashers/careful.yaml thrashosds-health.yaml workloads/cache.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
ubuntu 18.04
rados/singleton/{all/pg-autoscaler.yaml msgr-failures/many.yaml msgr/async-v2only.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}}
Found coredumps on ubuntu@smithi060.front.sepia.ceph.com
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
ubuntu 18.04
rados/mgr/{clusters/{2-node-mgr.yaml openstack.yaml} debug/mgr.yaml objectstore/bluestore-stupid.yaml supported-random-distro$/{ubuntu_latest.yaml} tasks/insights.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
centos 7.6
rados/thrash-old-clients/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-install/jewel.yaml backoff/peering_and_degraded.yaml ceph.yaml clusters/{openstack.yaml three-plus-one.yaml} d-balancer/crush-compat.yaml distro$/{centos_latest.yaml} msgr-failures/few.yaml rados.yaml thrashers/careful.yaml thrashosds-health.yaml workloads/test_rbd_api.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
rhel 7.6
rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml 2-recovery-overrides/{default.yaml} backoff/normal.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/upmap.yaml msgr-failures/osd-delay.yaml msgr/async.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{rhel_7.yaml} thrashers/default.yaml thrashosds-health.yaml workloads/dedup_tier.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
rhel 7.6
rados/perf/{ceph.yaml objectstore/bluestore-bitmap.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{rhel_7.yaml} workloads/fio_4M_rand_rw.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
centos 7.6
rados/singleton/{all/pg-removal-interruption.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{centos_7.yaml}}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
centos 7.6
rados/singleton-nomsgr/{all/version-number-sanity.yaml rados.yaml supported-random-distro$/{centos_7.yaml}}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
rhel 7.6
rados/basic/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/many.yaml msgr/async-v2only.yaml objectstore/filestore-xfs.yaml rados.yaml supported-random-distro$/{rhel_7.yaml} tasks/scrub_test.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
ubuntu 18.04
rados/thrash-erasure-code-isa/{arch/x86_64.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/few.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml recovery-overrides/{default.yaml} supported-random-distro$/{ubuntu_latest.yaml} thrashers/none.yaml thrashosds-health.yaml workloads/ec-rados-plugin=isa-k=2-m=1.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
ubuntu 18.04
rados/thrash-erasure-code/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} fast/normal.yaml msgr-failures/few.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml recovery-overrides/{more-active-recovery.yaml} supported-random-distro$/{ubuntu_latest.yaml} thrashers/pggrow.yaml thrashosds-health.yaml workloads/ec-small-objects-fast-read.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
centos 7.6
rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml 2-recovery-overrides/{more-active-recovery.yaml} backoff/peering.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/crush-compat.yaml msgr-failures/fastclose.yaml msgr/async-v1only.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{centos_7.yaml} thrashers/mapgap.yaml thrashosds-health.yaml workloads/pool-snaps-few-objects.yaml}
failed to become clean before timeout expired
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
centos 7.6
rados/singleton/{all/radostool.yaml msgr-failures/many.yaml msgr/async-v1only.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml supported-random-distro$/{centos_7.yaml}}
Command failed (workunit test rados/test_rados_tool.sh) on smithi083 with status 124: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=475286020efbcbe469ec6169e1c477e69c61c6c6 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/rados/test_rados_tool.sh'
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
rhel 7.6
rados/thrash-erasure-code-big/{ceph.yaml cluster/{12-osds.yaml openstack.yaml} msgr-failures/osd-delay.yaml objectstore/bluestore-bitmap.yaml rados.yaml recovery-overrides/{default.yaml} supported-random-distro$/{rhel_7.yaml} thrashers/default.yaml thrashosds-health.yaml workloads/ec-rados-plugin=jerasure-k=4-m=2.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
rhel 7.6
rados/thrash-erasure-code-overwrites/{bluestore-bitmap.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} fast/normal.yaml msgr-failures/fastclose.yaml rados.yaml recovery-overrides/{default.yaml} supported-random-distro$/{rhel_7.yaml} thrashers/default.yaml thrashosds-health.yaml workloads/ec-snaps-few-objects-overwrites.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
ubuntu 18.04
rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml 2-recovery-overrides/{default.yaml} backoff/peering_and_degraded.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/off.yaml msgr-failures/few.yaml msgr/async-v2only.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} thrashers/morepggrow.yaml thrashosds-health.yaml workloads/rados_api_tests.yaml}
Command failed (workunit test rados/test.sh) on smithi190 with status 124: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=475286020efbcbe469ec6169e1c477e69c61c6c6 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/rados/test.sh'
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
ubuntu 18.04
rados/mgr/{clusters/{2-node-mgr.yaml openstack.yaml} debug/mgr.yaml objectstore/filestore-xfs.yaml supported-random-distro$/{ubuntu_latest.yaml} tasks/module_selftest.yaml}
"2019-08-19T20:41:15.642908+0000 mon.a (mon.0) 1192 : cluster [WRN] Health check failed: 1 osds down (OSD_DOWN)" in cluster log
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
ubuntu 18.04
rados/perf/{ceph.yaml objectstore/bluestore-comp.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{ubuntu_latest.yaml} workloads/fio_4M_rand_write.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
ubuntu 18.04
rados/singleton/{all/random-eio.yaml msgr-failures/few.yaml msgr/async-v2only.yaml objectstore/bluestore-stupid.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
centos 7.6
rados/thrash-old-clients/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-install/luminous-v1only.yaml backoff/normal.yaml ceph.yaml clusters/{openstack.yaml three-plus-one.yaml} d-balancer/off.yaml distro$/{centos_latest.yaml} msgr-failures/osd-delay.yaml rados.yaml thrashers/default.yaml thrashosds-health.yaml workloads/cache-snaps.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
ubuntu 18.04
rados/objectstore/{backends/alloc-hint.yaml supported-random-distro$/{ubuntu_latest.yaml}}
Command failed (workunit test rados/test_alloc_hint.sh) on smithi113 with status 124: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=475286020efbcbe469ec6169e1c477e69c61c6c6 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/rados/test_alloc_hint.sh'
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
rhel 7.6
rados/rest/{mgr-restful.yaml supported-random-distro$/{rhel_7.yaml}}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
rhel 7.6
rados/singleton-bluestore/{all/cephtool.yaml msgr-failures/few.yaml msgr/async-v1only.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}}
Command failed (workunit test cephtool/test.sh) on smithi065 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=475286020efbcbe469ec6169e1c477e69c61c6c6 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/cephtool/test.sh'
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
centos 
rados/singleton-flat/valgrind-leaks.yaml
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
rhel 7.6
rados/singleton-nomsgr/{all/admin_socket_output.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
ubuntu 18.04
rados/standalone/{supported-random-distro$/{ubuntu_latest.yaml} workloads/crush.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
ubuntu 18.04
rados/upgrade/mimic-x-singleton/{0-cluster/{openstack.yaml start.yaml} 1-install/mimic.yaml 2-partial-upgrade/firsthalf.yaml 3-thrash/default.yaml 4-workload/{rbd-cls.yaml rbd-import-export.yaml readwrite.yaml snaps-few-objects.yaml} 5-workload/{radosbench.yaml rbd_api.yaml} 6-finish-upgrade.yaml 7-nautilus.yaml 8-workload/{rbd-python.yaml rgw-swift.yaml snaps-many-objects.yaml} bluestore-bitmap.yaml supported-random-distro$/{ubuntu_latest.yaml} thrashosds-health.yaml}
failed to become clean before timeout expired
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
rhel 7.6
rados/thrash-erasure-code-shec/{ceph.yaml clusters/{fixed-4.yaml openstack.yaml} msgr-failures/osd-delay.yaml objectstore/bluestore-comp.yaml rados.yaml recovery-overrides/{more-active-recovery.yaml} supported-random-distro$/{rhel_7.yaml} thrashers/careful.yaml thrashosds-health.yaml workloads/ec-rados-plugin=shec-k=4-m=3-c=2.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
centos 7.6
rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml 2-recovery-overrides/{more-active-recovery.yaml} backoff/normal.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/upmap.yaml msgr-failures/osd-delay.yaml msgr/async.yaml objectstore/bluestore-stupid.yaml rados.yaml supported-random-distro$/{centos_7.yaml} thrashers/none.yaml thrashosds-health.yaml workloads/radosbench.yaml}
reached maximum tries (500) after waiting for 3000 seconds
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
rhel 7.6
rados/basic/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/few.yaml msgr/async-v2only.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{rhel_7.yaml} tasks/rados_api_tests.yaml}
Command failed (workunit test rados/test.sh) on smithi002 with status 124: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=475286020efbcbe469ec6169e1c477e69c61c6c6 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/rados/test.sh'
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
rhel 7.6
rados/multimon/{clusters/6.yaml msgr-failures/many.yaml msgr/async-v1only.yaml no_pools.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{rhel_7.yaml} tasks/mon_clock_no_skews.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
 
rados/verify/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-thrash/default/{default.yaml thrashosds-health.yaml} msgr-failures/few.yaml msgr/async-v1only.yaml objectstore/bluestore-comp.yaml rados.yaml tasks/mon_recovery.yaml validater/lockdep.yaml}
"2019-08-19T21:03:53.254475+0000 mon.a (mon.0) 527 : cluster [WRN] Health check failed: 1 daemons have recently crashed (RECENT_CRASH)" in cluster log
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
rhel 7.6
rados/singleton/{all/rebuild-mondb.yaml msgr-failures/many.yaml msgr/async.yaml objectstore/filestore-xfs.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}}
+ sudo adjust-ulimits ceph-objectstore-tool --data-path /var/lib/ceph/osd/ceph-0 --journal-path /var/lib/ceph/osd/ceph-0/journal --no-mon-config --op update-mon-db --mon-store-path /home/ubuntu/cephtest/mon-store *** stack smashing detected ***: ceph-objectstore-tool terminated ======= Backtrace: ========= /lib64/libc.so.6(__fortify_fail+0x37)[0x7fadca990b67] /lib64/libc.so.6(+0x117b22)[0x7fadca990b22] ceph-objectstore-tool(_Z13update_mon_dbR11ObjectStoreR13OSDSuperblockRKSsS4_+0x2c51)[0x5581bb233a91] ceph-objectstore-tool(main+0x557e)[0x5581bb1b805e] /lib64/libc.so.6(__libc_start_main+0xf5)[0x7fadca89b495] ceph-objectstore-tool(+0x3cfe50)[0x5581bb1e7e50] ======= Memory map: ======== 5581bae18000-5581bbcf9000 r-xp 00000000 08:01 1728 /usr/bin/ceph-objectstore-tool 5581bbef8000-5581bbf30000 r--p 00ee0000 08:01 1728 /usr/bin/ceph-objectstore-tool 5581bbf30000-5581bbf44000 rw-p 00f18000 08:01 1728 /usr/bin/ceph-objectstore-tool 5581bbf44000-5581bc05c000 rw-p 00000000 00:00 0 5581bd85f000-5581bfdc2000 rw-p 00000000 00:00 0 [heap] 7fadb36ed000-7fadb36ee000 ---p 00000000 00:00 0 7fadb36ee000-7fadb3eee000 rw-p 00000000 00:00 0 7fadb3eee000-7fadb3eef000 ---p 00000000 00:00 0 7fadb3eef000-7fadb46ef000 rw-p 00000000 00:00 0 7fadb46ef000-7fadb46f0000 ---p 00000000 00:00 0 7fadb46f0000-7fadb4ef0000 rw-p 00000000 00:00 0 7fadb4ef0000-7fadb4ef1000 ---p 00000000 00:00 0 7fadb4ef1000-7fadb56f1000 rw-p 00000000 00:00 0 7fadbaefc000-7fadbaefd000 ---p 00000000 00:00 0 7fadbaefd000-7fadbb6fd000 rw-p 00000000 00:00 0 7fadbb6fd000-7fadbb6fe000 ---p 00000000 00:00 0 7fadbb6fe000-7fadbbefe000 rw-p 00000000 00:00 0 7fadbbefe000-7fadbbeff000 ---p 00000000 00:00 0 7fadbbeff000-7fadbc6ff000 rw-p 00000000 00:00 0 7fadbc6ff000-7fadbc700000 ---p 00000000 00:00 0 7fadbc700000-7fadbcf00000 rw-p 00000000 00:00 0 7fadbcf00000-7fadbcf01000 ---p 00000000 00:00 0 7fadbcf01000-7fadbd701000 rw-p 00000000 00:00 0 7fadbd701000-7fadbd702000 ---p 00000000 00:00 0 7fadbd702000-7fadbdf02000 rw-p 00000000 00:00 0 7fadbdf02000-7fadbdf03000 ---p 00000000 00:00 0 7fadbdf03000-7fadbe703000 rw-p 00000000 00:00 0 7fadbe703000-7fadbe704000 ---p 00000000 00:00 0 7fadbe704000-7fadbef04000 rw-p 00000000 00:00 0 7fadbef04000-7fadbef05000 ---p 00000000 00:00 0 7fadbef05000-7fadbf705000 rw-p 00000000 00:00 0 7fadbf705000-7fadbf706000 ---p 00000000 00:00 0 7fadbf706000-7fadbff06000 rw-p 00000000 00:00 0 7fadbff06000-7fadbff07000 ---p 00000000 00:00 0 7fadbff07000-7fadc0707000 rw-p 00000000 00:00 0 7fadc0707000-7fadc0708000 ---p 00000000 00:00 0 7fadc0708000-7fadc0f08000 rw-p 00000000 00:00 0 7fadc0f08000-7fadc0f09000 ---p 00000000 00:00 0 7fadc0f09000-7fadc1709000 rw-p 00000000 00:00 0 7fadc1709000-7fadc170a000 ---p 00000000 00:00 0 7fadc170a000-7fadc1f0a000 rw-p 00000000 00:00 0 7fadc1f0a000-7fadc1f0b000 ---p 00000000 00:00 0 7fadc1f0b000-7fadc270b000 rw-p 00000000 00:00 0 7fadc270b000-7fadc270c000 ---p 00000000 00:00 0 7fadc270c000-7fadc2f0c000 rw-p 00000000 00:00 0 7fadc2f0c000-7fadc2f0d000 ---p 00000000 00:00 0 7fadc2f0d000-7fadc370d000 rw-p 00000000 00:00 0 7fadc370d000-7fadc370e000 ---p 00000000 00:00 0 7fadc370e000-7fadc3f0e000 rw-p 00000000 00:00 0 7fadc3f0e000-7fadc3f0f000 ---p 00000000 00:00 0 7fadc3f0f000-7fadc470f000 rw-p 00000000 00:00 0 7fadc470f000-7fadc4710000 ---p 00000000 00:00 0 7fadc4710000-7fadc4f10000 rw-p 00000000 00:00 0 7fadc4f10000-7fadc4f11000 ---p 00000000 00:00 0 7fadc4f11000-7fadc5711000 rw-p 00000000 00:00 0 7fadc5711000-7fadc5712000 ---p 00000000 00:00 0 7fadc5712000-7fadc5f12000 rw-p 00000000 00:00 0 7fadc5f12000-7fadc5f13000 ---p 00000000 00:00 0 7fadc5f13000-7fadc6713000 rw-p 00000000 00:00 0 7fadc6713000-7fadc6714000 ---p 00000000 00:00 0 7fadc6714000-7fadc6f14000 rw-p 00000000 00:00 0 7fadc6f14000-7fadc6f15000 ---p 00000000 00:00 0 7fadc6f15000-7fadc7715000 rw-p 00000000 00:00 0 7fadc7715000-7fadc7716000 ---p 00000000 00:00 0 7fadc7716000-7fadc7f16000 rw-p 00000000 00:00 0 7fadc7f16000-7fadc7f17000 ---p 00000000 00:00 0 7fadc7f17000-7fadc8717000 rw-p 00000000 00:00 0 7fadc8717000-7fadc8718000 ---p 00000000 00:00 0 7fadc8718000-7fadc8f18000 rw-p 00000000 00:00 0 7fadc8f18000-7fadc8f1a000 r-xp 00000000 08:01 1406 /usr/lib64/liburcu-common.so.1.0.0 7fadc8f1a000-7fadc9119000 ---p 00002000 08:01 1406 /usr/lib64/liburcu-common.so.1.0.0 7fadc9119000-7fadc911a000 r--p 00001000 08:01 1406 /usr/lib64/liburcu-common.so.1.0.0 7fadc911a000-7fadc911b000 rw-p 00002000 08:01 1406 /usr/lib64/liburcu-common.so.1.0.0 7fadc911b000-7fadc9121000 r-xp 00000000 08:01 1404 /usr/lib64/liburcu-cds.so.1.0.0 7fadc9121000-7fadc9320000 ---p 00006000 08:01 1404 /usr/lib64/liburcu-cds.so.1.0.0 7fadc9320000-7fadc9321000 r--p 00005000 08:01 1404 /usr/lib64/liburcu-cds.so.1.0.0 7fadc9321000-7fadc9322000 rw-p 00006000 08:01 1404 /usr/lib64/liburcu-cds.so.1.0.0 7fadc9322000-7fadc9328000 r-xp 00000000 08:01 1402 /usr/lib64/liburcu-bp.so.1.0.0 7fadc9328000-7fadc9528000 ---p 00006000 08:01 1402 /usr/lib64/liburcu-bp.so.1.0.0 7fadc9528000-7fadc9529000 r--p 00006000 08:01 1402 /usr/lib64/liburcu-bp.so.1.0.0 7fadc9529000-7fadc952a000 rw-p 00007000 08:01 1402 /usr/lib64/liburcu-bp.so.1.0.0 7fadc952a000-7fadc9533000 r-xp 00000000 08:01 1430 /usr/lib64/liblttng-ust-tracepoint.so.0.0.0 7fadc9533000-7fadc9732000 ---p 00009000 08:01 1430 /usr/lib64/liblttng-ust-tracepoint.so.0.0.0 7fadc9732000-7fadc9733000 r--p 00008000 08:01 1430 /usr/lib64/liblttng-ust-tracepoint.so.0.0.0 7fadc9733000-7fadc9734000 rw-p 00009000 08:01 1430 /usr/lib64/liblttng-ust-tracepoint.so.0.0.0 7fadc9734000-7fadc9744000 rw-p 00000000 00:00 0 7fadc9744000-7fadc9753000 r-xp 00000000 08:01 9223 /usr/lib64/libbz2.so.1.0.6 7fadc9753000-7fadc9952000 ---p 0000f000 08:01 9223 /usr/lib64/libbz2.so.1.0.6 7fadc9952000-7fadc9953000 r--p 0000e000 08:01 9223 /usr/lib64/libbz2.so.1.0.6 7fadc9953000-7fadc9954000 rw-p 0000f000 08:01 9223 /usr/lib64/libbz2.so.1.0.6 7fadc9954000-7fadc9979000 r-xp 00000000 08:01 9216 /usr/lib64/liblzma.so.5.2.2 7fadc9979000-7fadc9b78000 ---p 00025000 08:01 9216 /usr/lib64/liblzma.so.5.2.2 7fadc9b78000-7fadc9b79000 r--p 00024000 08:01 9216 /usr/lib64/liblzma.so.5.2.2 7fadc9b79000-7fadc9b7a000 rw-p 00025000 08:01 9216 /usr/lib64/liblzma.so.5.2.2 7fadc9b7a000-7fadc9b91000 r-xp 00000000 08:01 9317 /usr/lib64/libelf-0.172.so 7fadc9b91000-7fadc9d90000 ---p 00017000 08:01 9317 /usr/lib64/libelf-0.172.so 7fadc9d90000-7fadc9d91000 r--p 00016000 08:01 9317 /usr/lib64/libelf-0.172.so 7fadc9d91000-7fadc9d92000 rw-p 00017000 08:01 9317 /usr/lib64/libelf-0.172.so 7fadc9d92000-7fadc9d96000 r-xp 00000000 08:01 9364 /usr/lib64/libattr.so.1.1.0 7fadc9d96000-7fadc9f95000 ---p 00004000 08:01 9364 /usr/lib64/libattr.so.1.1.0 7fadc9f95000-7fadc9f96000 r--p 00003000 08:01 9364 /usr/lib64/libattr.so.1.1.0 7fadc9f96000-7fadc9f97000 rw-p 00004000 08:01 9364 /usr/lib64/libattr.so.1.1.0 7fadc9f97000-7fadc9fb5000 r-xp 00000000 08:01 9529 /usr/lib64/libnl-3.so.200.23.0 7fadc9fb5000-7fadca1b5000 ---p 0001e000 08:01 9529 /usr/lib64/libnl-3.so.200.23.0 7fadca1b5000-7fadca1b7000 r--p 0001e000 08:01 9529 /usr/lib64/libnl-3.so.200.23.0 7fadca1b7000-7fadca1b8000 rw-p 00020000 08:01 9529 /usr/lib64/libnl-3.so.200.23.0 7fadca1b8000-7fadca21c000 r-xp 00000000 08:01 9537 /usr/lib64/libnl-route-3.so.200.23.0 7fadca21c000-7fadca41b000 ---p 00064000 08:01 9537 /usr/lib64/libnl-route-3.so.200.23.0 7fadca41b000-7fadca41e000 r--p 00063000 08:01 9537 /usr/lib64/libnl-route-3.so.200.23.0 7fadca41e000-7fadca423000 rw-p 00066000 08:01 9537 /usr/lib64/libnl-route-3.so.200.23.0 7fadca423000-7fadca425000 rw-p 00000000 00:00 0 7fadca425000-7fadca471000 r-xp 00000000 08:01 11388 /usr/lib64/libdw-0.172.so 7fadca471000-7fadca671000 ---p 0004c000 08:01 11388 /usr/lib64/libdw-0.172.so 7fadca671000-7fadca673000 r--p 0004c000 08:01 11388 /usr/lib64/libdw-0.172.so 7fadca673000-7fadca674000 rw-p 0004e000 08:01 11388 /usr/lib64/libdw-0.172.so 7fadca674000-7fadca678000 r-xp 00000000 08:01 9366 /usr/lib64/libcap.so.2.22 7fadca678000-7fadca877000 ---p 00004000 08:01 9366 /usr/lib64/libcap.so.2.22 7fadca877000-7fadca878000 r--p 00003000 08:01 9366 /usr/lib64/libcap.so.2.22 7fadca878000-7fadca879000 rw-p 00004000 08:01 9366 /usr/lib64/libcap.so.2.22 7fadca879000-7fadcaa3b000 r-xp 00000000 08:01 9086 /usr/lib64/libc-2.17.so 7fadcaa3b000-7fadcac3b000 ---p 001c2000 08:01 9086 /usr/lib64/libc-2.17.so 7fadcac3b000-7fadcac3f000 r--p 001c2000 08:01 9086 /usr/lib64/libc-2.17.so 7fadcac3f000-7fadcac41000 rw-p 001c6000 08:01 9086 /usr/lib64/libc-2.17.so 7fadcac41000-7fadcac46000 rw-p 00000000 00:00 0 7fadcac46000-7fadcac5b000 r-xp 00000000 08:01 9799 /usr/lib64/libgcc_s-4.8.5-20150702.so.1 7fadcac5b000-7fadcae5a000 ---p 00015000 08:01 9799 /usr/lib64/libgcc_s-4.8.5-20150702.so.1 7fadcae5a000-7fadcae5b000 r--p 00014000 08:01 9799 /usr/lib64/libgcc_s-4.8.5-20150702.so.1 7fadcae5b000-7fadcae5c000 rw-p 00015000 08:01 9799 /usr/lib64/libgcc_s-4.8.5-20150702.so.1 7fadcae5c000-7fadcaf5d000 r-xp 00000000 08:01 9091 /usr/lib64/libm-2.17.so 7fadcaf5d000-7fadcb15c000 ---p 00101000 08:01 9091 /usr/lib64/libm-2.17.so 7fadcb15c000-7fadcb15d000 r--p 00100000 08:01 9091 /usr/lib64/libm-2.17.so 7fadcb15d000-7fadcb15e000 rw-p 00101000 08:01 9091 /usr/lib64/libm-2.17.so 7fadcb15e000-7fadcb247000 r-xp 00000000 08:01 9119 /usr/lib64/libstdc++.so.6.0.19 7fadcb247000-7fadcb446000 ---p 000e9000 08:01 9119 /usr/lib64/libstdc++.so.6.0.19 7fadcb446000-7fadcb44e000 r--p 000e8000 08:01 9119 /usr/lib64/libstdc++.so.6.0.19 7fadcb44e000-7fadcb450000 rw-p 000f0000 08:01 9119 /usr/lib64/libstdc++.so.6.0.19 7fadcb450000-7fadcb465000 rw-p 00000000 00:00 0 7fadcb465000-7fadcb47b000 r-xp 00000000 08:01 9106 /usr/lib64/libresolv-2.17.so 7fadcb47b000-7fadcb67a000 ---p 00016000 08:01 9106 /usr/lib64/libresolv-2.17.so 7fadcb67a000-7fadcb67b000 r--p 00015000 08:01 9106 /usr/lib64/libresolv-2.17.so 7fadcb67b000-7fadcb67c000 rw-p 00016000 08:01 9106 /usr/lib64/libresolv-2.17.so 7fadcb67c000-7fadcb67e000 rw-p 00000000 00:00 0 7fadcb67e000-7fadcb685000 r-xp 00000000 08:01 9107 /usr/lib64/librt-2.17.so 7fadcb685000-7fadcb884000 ---p 00007000 08:01 9107 /usr/lib64/librt-2.17.so 7fadcb884000-7fadcb885000 r--p 00006000 08:01 9107 /usr/lib64/librt-2.17.so 7fadcb885000-7fadcb886000 rw-p 00007000 08:01 9107 /usr/lib64/librt-2.17.so 7fadcb886000-7fadcb89b000 r-xp 00000000 08:01 14974 /usr/lib64/librdmacm.so.1.1.17.2 7fadcb89b000-7fadcba9a000 ---p 00015000 08:01 14974 /usr/lib64/librdmacm.so.1.1.17.2 7fadcba9a000-7fadcba9b000 r--p 00014000 08:01 14974 /usr/lib64/librdmacm.so.1.1.17.2 7fadcba9b000-7fadcba9c000 rw-p 00015000 08:01 14974 /usr/lib64/librdmacm.so.1.1.17.2 7fadcba9c000-7fadcba9d000 rw-p 00000000 00:00 0 7fadcba9d000-7fadcbab2000 r-xp 00000000 08:01 14968 /usr/lib64/libibverbs.so.1.1.17.2 7fadcbab2000-7fadcbcb2000 ---p 00015000 08:01 14968 /usr/lib64/libibverbs.so.1.1.17.2 7fadcbcb2000-7fadcbcb3000 r--p 00015000 08:01 14968 /usr/lib64/libibverbs.so.1.1.17.2 7fadcbcb3000-7fadcbcb4000 rw-p 00016000 08:01 14968 /usr/lib64/libibverbs.so.1.1.17.2 7fadcbcb4000-7fadcbcc9000 r-xp 00000000 08:01 11403 /usr/lib64/libudev.so.1.6.2 7fadcbcc9000-7fadcbec8000 ---p 00015000 08:01 11403 /usr/lib64/libudev.so.1.6.2 7fadcbec8000-7fadcbec9000 r--p 00014000 08:01 11403 /usr/lib64/libudev.so.1.6.2 7fadcbec9000-7fadcbeca000 rw-p 00015000 08:01 11403 /usr/lib64/libudev.so.1.6.2 7fadcbeca000-7fadcbee1000 r-xp 00000000 08:01 9105 /usr/lib64/libpthread-2.17.so 7fadcbee1000-7fadcc0e0000 ---p 00017000 08:01 9105 /usr/lib64/libpthread-2.17.so 7fadcc0e0000-7fadcc0e1000 r--p 00016000 08:01 9105 /usr/lib64/libpthread-2.17.so 7fadcc0e1000-7fadcc0e2000 rw-p 00017000 08:01 9105 /usr/lib64/libpthread-2.17.so 7fadcc0e2000-7fadcc0e6000 rw-p 00000000 00:00 0 7fadcc0e6000-7fadcc31a000 r-xp 00000000 08:01 10149 /usr/lib64/libcrypto.so.1.0.2k 7fadcc31a000-7fadcc51a000 ---p 00234000 08:01 10149 /usr/lib64/libcrypto.so.1.0.2k 7fadcc51a000-7fadcc536000 r--p 00234000 08:01 10149 /usr/lib64/libcrypto.so.1.0.2k 7fadcc536000-7fadcc543000 rw-p 00250000 08:01 10149 /usr/lib64/libcrypto.so.1.0.2k 7fadcc543000-7fadcc547000 rw-p 00000000 00:00 0 7fadcc547000-7fadcc583000 r-xp 00000000 08:01 10273 /usr/lib64/libblkid.so.1.1.0 7fadcc583000-7fadcc782000 ---p 0003c000 08:01 10273 /usr/lib64/libblkid.so.1.1.0 7fadcc782000-7fadcc785000 r--p 0003b000 08:01 10273 /usr/lib64/libblkid.so.1.1.0 7fadcc785000-7fadcc786000 rw-p 0003e000 08:01 10273 /usr/lib64/libblkid.so.1.1.0 7fadcc786000-7fadcc787000 rw-p 00000000 00:00 0 7fadcc787000-7fadcd10d000 r-xp 00000000 08:01 393372 /usr/lib64/ceph/libceph-common.so.0 7fadcd10d000-7fadcd30d000 ---p 00986000 08:01 393372 /usr/lib64/ceph/libceph-common.so.0 7fadcd30d000-7fadcd32c000 r--p 00986000 08:01 393372 /usr/lib64/ceph/libceph-common.so.0 7fadcd32c000-7fadcd335000 rw-p 009a5000 08:01 393372 /usr/lib64/ceph/libceph-common.so.0 7fadcd335000-7fadd5555000 rw-p 00000000 00:00 0 7fadd5555000-7fadd559b000 r-xp 00000000 08:01 14944 /usr/lib64/libtcmalloc.so.4.4.5 7fadd559b000-7fadd579b000 ---p 00046000 08:01 14944 /usr/lib64/libtcmalloc.so.4.4.5 7fadd579b000-7fadd579c000 r--p 00046000 08:01 14944 /usr/lib64/libtcmalloc.so.4.4.5 7fadd579c000-7fadd579d000 rw-p 00047000 08:01 14944 /usr/lib64/libtcmalloc.so.4.4.5 7fadd579d000-7fadd594a000 rw-p 00000000 00:00 0 7fadd594a000-7fadd594e000 r-xp 00000000 08:01 9221 /usr/lib64/libuuid.so.1.3.0 7fadd594e000-7fadd5b4d000 ---p 00004000 08:01 9221 /usr/lib64/libuuid.so.1.3.0 7fadd5b4d000-7fadd5b4e000 r--p 00003000 08:01 9221 /usr/lib64/libuuid.so.1.3.0 7fadd5b4e000-7fadd5b4f000 rw-p 00004000 08:01 9221 /usr/lib64/libuuid.so.1.3.0 7fadd5b4f000-7fadd5b50000 r-xp 00000000 08:01 9579 /usr/lib64/libaio.so.1.0.1 7fadd5b50000-7fadd5d4f000 ---p 00001000 08:01 9579 /usr/lib64/libaio.so.1.0.1 7fadd5d4f000-7fadd5d50000 r--p 00000000 08:01 9579 /usr/lib64/libaio.so.1.0.1 7fadd5d50000-7fadd5d51000 rw-p 00001000 08:01 9579 /usr/lib64/libaio.so.1.0.1 7fadd5d51000-7fadd5d66000 r-xp 00000000 08:01 9206 /usr/lib64/libz.so.1.2.7 7fadd5d66000-7fadd5f65000 ---p 00015000 08:01 9206 /usr/lib64/libz.so.1.2.7 7fadd5f65000-7fadd5f66000 r--p 00014000 08:01 9206 /usr/lib64/libz.so.1.2.7 7fadd5f66000-7fadd5f67000 rw-p 00015000 08:01 9206 /usr/lib64/libz.so.1.2.7 7fadd5f67000-7fadd5f7b000 r-xp 00000000 08:01 9677 /usr/lib64/liblz4.so.1.7.5 7fadd5f7b000-7fadd617a000 ---p 00014000 08:01 9677 /usr/lib64/liblz4.so.1.7.5 7fadd617a000-7fadd617b000 r--p 00013000 08:01 9677 /usr/lib64/liblz4.so.1.7.5 7fadd617b000-7fadd617c000 rw-p 00014000 08:01 9677 /usr/lib64/liblz4.so.1.7.5 7fadd617c000-7fadd6181000 r-xp 00000000 08:01 10098 /usr/lib64/libsnappy.so.1.1.4 7fadd6181000-7fadd6380000 ---p 00005000 08:01 10098 /usr/lib64/libsnappy.so.1.1.4 7fadd6380000-7fadd6381000 r--p 00004000 08:01 10098 /usr/lib64/libsnappy.so.1.1.4 7fadd6381000-7fadd6382000 rw-p 00005000 08:01 10098 /usr/lib64/libsnappy.so.1.1.4 7fadd6382000-7fadd63d4000 r-xp 00000000 08:01 1270 /usr/lib64/libleveldb.so.1.0.7 7fadd63d4000-7fadd65d3000 ---p 00052000 08:01 1270 /usr/lib64/libleveldb.so.1.0.7 7fadd65d3000-7fadd65d5000 r--p 00051000 08:01 1270 /usr/lib64/libleveldb.so.1.0.7 7fadd65d5000-7fadd65d6000 rw-p 00053000 08:01 1270 /usr/lib64/libleveldb.so.1.0.7 7fadd65d6000-7fadd6601000 r-xp 00000000 08:01 14619 /usr/lib64/libfuse.so.2.9.2 7fadd6601000-7fadd6801000 ---p 0002b000 08:01 14619 /usr/lib64/libfuse.so.2.9.2 7fadd6801000-7fadd6813000 r--p 0002b000 08:01 14619 /usr/lib64/libfuse.so.2.9.2 7fadd6813000-7fadd6814000 rw-p 0003d000 08:01 14619 /usr/lib64/libfuse.so.2.9.2 7fadd6814000-7fadd6816000 r-xp 00000000 08:01 9090 /usr/lib64/libdl-2.17.so 7fadd6816000-7fadd6a16000 ---p 00002000 08:01 9090 /usr/lib64/libdl-2.17.so 7fadd6a16000-7fadd6a17000 r--p 00002000 08:01 9090 /usr/lib64/libdl-2.17.so 7fadd6a17000-7fadd6a18000 rw-p 00003000 08:01 9090 /usr/lib64/libdl-2.17.so 7fadd6a18000-7fadd6a3a000 r-xp 00000000 08:01 9080 /usr/lib64/ld-2.17.so 7fadd6c07000-7fadd6c2b000 rw-p 00000000 00:00 0 7fadd6c37000-7fadd6c39000 rw-p 00000000 00:00 0 7fadd6c39000-7fadd6c3a000 r--p 00021000 08:01 9080 /usr/lib64/ld-2.17.so 7fadd6c3a000-7fadd6c3b000 rw-p 00022000 08:01 9080 /usr/lib64/ld-2.17.so 7fadd6c3b000-7fadd6c3c000 rw-p 00000000 00:00 0 7ffc1265e000-7ffc12717000 rw-p 00000000 00:00 0 [stack] 7ffc1272b000-7ffc1272d000 r-xp 00000000 00:00 0 [vdso] ffffffffff600000-ffffffffff601000 r-xp 00000000 00:00 0 [vsyscall] *** Caught signal (Aborted) ** in thread 7fadd6c1a980 thread_name:ceph-objectstor ceph version 15.0.0-4067-g4752860 (475286020efbcbe469ec6169e1c477e69c61c6c6) octopus (dev) 1: (()+0xf5d0) [0x7fadcbed95d0] 2: (gsignal()+0x37) [0x7fadca8af2c7] 3: (abort()+0x148) [0x7fadca8b09b8] 4: (()+0x78e17) [0x7fadca8f1e17] 5: (__fortify_fail()+0x37) [0x7fadca990b67] 6: (()+0x117b22) [0x7fadca990b22] 7: (update_mon_db(ObjectStore&, OSDSuperblock&, std::string const&, std::string const&)+0x2c51) [0x5581bb233a91] 8: (main()+0x557e) [0x5581bb1b805e] 9: (__libc_start_main()+0xf5) [0x7fadca89b495] 10: (()+0x3cfe50) [0x5581bb1e7e50]
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
ubuntu 18.04
rados/monthrash/{ceph.yaml clusters/3-mons.yaml msgr-failures/few.yaml msgr/async-v2only.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} thrashers/one.yaml workloads/rados_mon_osdmap_prune.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
centos 7.6
rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml 2-recovery-overrides/{default.yaml} backoff/peering.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/crush-compat.yaml msgr-failures/fastclose.yaml msgr/async-v1only.yaml objectstore/filestore-xfs.yaml rados.yaml supported-random-distro$/{centos_7.yaml} thrashers/pggrow.yaml thrashosds-health.yaml workloads/redirect.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
rhel 7.6
rados/perf/{ceph.yaml objectstore/bluestore-low-osd-mem-target.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{rhel_7.yaml} workloads/radosbench_4K_rand_read.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
centos 7.6
rados/singleton/{all/recovery-preemption.yaml msgr-failures/few.yaml msgr/async-v1only.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{centos_7.yaml}}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
rhel 7.6
rados/singleton-nomsgr/{all/cache-fs-trunc.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
centos 7.6
rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml 2-recovery-overrides/{default.yaml} backoff/peering_and_degraded.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/off.yaml msgr-failures/few.yaml msgr/async-v2only.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{centos_7.yaml} thrashers/careful.yaml thrashosds-health.yaml workloads/redirect_promote_tests.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
ubuntu 18.04
rados/mgr/{clusters/{2-node-mgr.yaml openstack.yaml} debug/mgr.yaml objectstore/bluestore-bitmap.yaml supported-random-distro$/{ubuntu_latest.yaml} tasks/orchestrator_cli.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
rhel 7.6
rados/thrash-erasure-code/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} fast/fast.yaml msgr-failures/osd-delay.yaml objectstore/bluestore-stupid.yaml rados.yaml recovery-overrides/{more-active-recovery.yaml} supported-random-distro$/{rhel_7.yaml} thrashers/careful.yaml thrashosds-health.yaml workloads/ec-small-objects-many-deletes.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
ubuntu 18.04
rados/thrash-erasure-code-isa/{arch/x86_64.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/osd-delay.yaml objectstore/bluestore-stupid.yaml rados.yaml recovery-overrides/{default.yaml} supported-random-distro$/{ubuntu_latest.yaml} thrashers/pggrow.yaml thrashosds-health.yaml workloads/ec-rados-plugin=isa-k=2-m=1.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
rhel 7.6
rados/singleton/{all/resolve_stuck_peering.yaml msgr-failures/many.yaml msgr/async-v2only.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
centos 7.6
rados/thrash-old-clients/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-install/luminous.yaml backoff/peering.yaml ceph.yaml clusters/{openstack.yaml three-plus-one.yaml} d-balancer/crush-compat.yaml distro$/{centos_latest.yaml} msgr-failures/fastclose.yaml rados.yaml thrashers/mapgap.yaml thrashosds-health.yaml workloads/radosbench.yaml}
SSH connection to smithi115 was lost: 'rpm -q kernel --last | head -n 1'
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
centos 7.6
rados/basic/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/many.yaml msgr/async.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{centos_7.yaml} tasks/rados_cls_all.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
centos 7.6
rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml 2-recovery-overrides/{more-active-recovery.yaml} backoff/normal.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/upmap.yaml msgr-failures/osd-delay.yaml msgr/async.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{centos_7.yaml} thrashers/default.yaml thrashosds-health.yaml workloads/redirect_set_object.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
ubuntu 18.04
rados/perf/{ceph.yaml objectstore/bluestore-stupid.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{ubuntu_latest.yaml} workloads/radosbench_4K_seq_read.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
ubuntu 18.04
rados/singleton/{all/test-crash.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
rhel 7.6
rados/objectstore/{backends/ceph_objectstore_tool.yaml supported-random-distro$/{rhel_7.yaml}}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
centos 7.6
rados/thrash-erasure-code-big/{ceph.yaml cluster/{12-osds.yaml openstack.yaml} msgr-failures/fastclose.yaml objectstore/bluestore-comp.yaml rados.yaml recovery-overrides/{more-active-recovery.yaml} supported-random-distro$/{centos_7.yaml} thrashers/fastread.yaml thrashosds-health.yaml workloads/ec-rados-plugin=lrc-k=4-m=2-l=3.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
centos 7.6
rados/thrash-erasure-code-overwrites/{bluestore-bitmap.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} fast/fast.yaml msgr-failures/few.yaml rados.yaml recovery-overrides/{more-active-recovery.yaml} supported-random-distro$/{centos_7.yaml} thrashers/fastread.yaml thrashosds-health.yaml workloads/ec-pool-snaps-few-objects-overwrites.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
rhel 7.6
rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml 2-recovery-overrides/{more-active-recovery.yaml} backoff/peering.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/crush-compat.yaml msgr-failures/fastclose.yaml msgr/async-v1only.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml supported-random-distro$/{rhel_7.yaml} thrashers/mapgap.yaml thrashosds-health.yaml workloads/set-chunk-promote-flush.yaml}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
rhel 7.6
rados/singleton-nomsgr/{all/ceph-kvstore-tool.yaml rados.yaml supported-random-distro$/{rhel_7.yaml}}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
centos 7.6
rados/singleton/{all/test_envlibrados_for_rocksdb.yaml msgr-failures/many.yaml msgr/async-v1only.yaml objectstore/bluestore-stupid.yaml rados.yaml supported-random-distro$/{centos_7.yaml}}
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
centos 7.6
rados/mgr/{clusters/{2-node-mgr.yaml openstack.yaml} debug/mgr.yaml objectstore/bluestore-comp.yaml supported-random-distro$/{centos_7.yaml} tasks/progress.yaml}
Test failure: test_osd_came_back (tasks.mgr.test_progress.TestProgress)
wip-sage3-testing-2019-08-19-1043
wip-sage3-testing-2019-08-19-1043
master
smithi
centos 7.6
rados/perf/{ceph.yaml objectstore/filestore-xfs.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{centos_7.yaml} workloads/radosbench_4M_rand_read.yaml}