Status Job ID Links Posted Started Updated
Runtime
Duration
In Waiting
Machine Teuthology Branch OS Type OS Version Description Nodes
fail 4615005 2019-12-19 17:26:41 2019-12-19 17:28:05 2019-12-19 18:14:05 0:46:00 0:31:43 0:14:17 smithi master rhel 8.0 rados/basic/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/few.yaml msgr/async-v2only.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{rhel_8.yaml} tasks/rados_workunit_loadgen_mostlyread.yaml} 2
Failure Reason:

SELinux denials found on ubuntu@smithi097.front.sepia.ceph.com: ['type=AVC msg=audit(1576778806.304:6884): avc: denied { open } for pid=30757 comm="rhsmcertd-worke" path="/etc/dnf/modules.d/satellite-5-client.module" dev="sda1" ino=57237 scontext=system_u:system_r:rhsmcertd_t:s0 tcontext=system_u:object_r:root_t:s0 tclass=file permissive=1', 'type=AVC msg=audit(1576778805.891:6878): avc: denied { map } for pid=30757 comm="rhsmcertd-worke" path="/var/lib/rpm/__db.001" dev="sda1" ino=262271 scontext=system_u:system_r:rhsmcertd_t:s0 tcontext=unconfined_u:object_r:var_lib_t:s0 tclass=file permissive=1', 'type=AVC msg=audit(1576778805.967:6880): avc: denied { create } for pid=30757 comm="rhsmcertd-worke" name="metadata_lock.pid" scontext=system_u:system_r:rhsmcertd_t:s0 tcontext=system_u:object_r:rpm_var_cache_t:s0 tclass=file permissive=1', 'type=AVC msg=audit(1576778806.515:6885): avc: denied { read } for pid=30811 comm="setroubleshootd" name="Packages" dev="sda1" ino=262250 scontext=system_u:system_r:setroubleshootd_t:s0-s0:c0.c1023 tcontext=unconfined_u:object_r:var_lib_t:s0 tclass=file permissive=1', 'type=AVC msg=audit(1576778805.967:6879): avc: denied { open } for pid=30757 comm="rhsmcertd-worke" path="/var/log/hawkey.log" dev="sda1" ino=60817 scontext=system_u:system_r:rhsmcertd_t:s0 tcontext=system_u:object_r:var_log_t:s0 tclass=file permissive=1', 'type=AVC msg=audit(1576778805.891:6875): avc: denied { open } for pid=30757 comm="rhsmcertd-worke" path="/var/lib/rpm/.dbenv.lock" dev="sda1" ino=262270 scontext=system_u:system_r:rhsmcertd_t:s0 tcontext=unconfined_u:object_r:var_lib_t:s0 tclass=file permissive=1', 'type=AVC msg=audit(1576778806.220:6883): avc: denied { remove_name } for pid=30757 comm="rhsmcertd-worke" name="metadata_lock.pid" dev="sda1" ino=56725 scontext=system_u:system_r:rhsmcertd_t:s0 tcontext=system_u:object_r:rpm_var_cache_t:s0 tclass=dir permissive=1', 'type=AVC msg=audit(1576778805.891:6875): avc: denied { read write } for pid=30757 comm="rhsmcertd-worke" name=".dbenv.lock" dev="sda1" ino=262270 scontext=system_u:system_r:rhsmcertd_t:s0 tcontext=unconfined_u:object_r:var_lib_t:s0 tclass=file permissive=1', 'type=AVC msg=audit(1576778806.515:6886): avc: denied { lock } for pid=30811 comm="setroubleshootd" path="/var/lib/rpm/Packages" dev="sda1" ino=262250 scontext=system_u:system_r:setroubleshootd_t:s0-s0:c0.c1023 tcontext=unconfined_u:object_r:var_lib_t:s0 tclass=file permissive=1', 'type=AVC msg=audit(1576778806.076:6881): avc: denied { open } for pid=30757 comm="rhsmcertd-worke" path="/var/cache/dnf/ceph-eb28c24780af4654/repodata/repomd.xml" dev="sda1" ino=262154 scontext=system_u:system_r:rhsmcertd_t:s0 tcontext=unconfined_u:object_r:rpm_var_cache_t:s0 tclass=file permissive=1', 'type=AVC msg=audit(1576778806.220:6883): avc: denied { unlink } for pid=30757 comm="rhsmcertd-worke" name="metadata_lock.pid" dev="sda1" ino=56725 scontext=system_u:system_r:rhsmcertd_t:s0 tcontext=system_u:object_r:rpm_var_cache_t:s0 tclass=file permissive=1', 'type=AVC msg=audit(1576778805.967:6880): avc: denied { add_name } for pid=30757 comm="rhsmcertd-worke" name="metadata_lock.pid" scontext=system_u:system_r:rhsmcertd_t:s0 tcontext=system_u:object_r:rpm_var_cache_t:s0 tclass=dir permissive=1', 'type=AVC msg=audit(1576778805.891:6877): avc: denied { getattr } for pid=30757 comm="rhsmcertd-worke" path="/var/lib/rpm/__db.001" dev="sda1" ino=262271 scontext=system_u:system_r:rhsmcertd_t:s0 tcontext=unconfined_u:object_r:var_lib_t:s0 tclass=file permissive=1', 'type=AVC msg=audit(1576778806.515:6887): avc: denied { map } for pid=30811 comm="setroubleshootd" path="/var/lib/rpm/Name" dev="sda1" ino=262251 scontext=system_u:system_r:setroubleshootd_t:s0-s0:c0.c1023 tcontext=unconfined_u:object_r:var_lib_t:s0 tclass=file permissive=1', 'type=AVC msg=audit(1576778806.304:6884): avc: denied { read } for pid=30757 comm="rhsmcertd-worke" name="satellite-5-client.module" dev="sda1" ino=57237 scontext=system_u:system_r:rhsmcertd_t:s0 tcontext=system_u:object_r:root_t:s0 tclass=file permissive=1', 'type=AVC msg=audit(1576778806.076:6882): avc: denied { setattr } for pid=30757 comm="rhsmcertd-worke" name="beb9905f58b0597637f875222723f63fc02df6af242782946b359057f8da6f84-primary.xml.gz" dev="sda1" ino=262180 scontext=system_u:system_r:rhsmcertd_t:s0 tcontext=unconfined_u:object_r:rpm_var_cache_t:s0 tclass=file permissive=1', 'type=AVC msg=audit(1576778805.891:6876): avc: denied { lock } for pid=30757 comm="rhsmcertd-worke" path="/var/lib/rpm/.dbenv.lock" dev="sda1" ino=262270 scontext=system_u:system_r:rhsmcertd_t:s0 tcontext=unconfined_u:object_r:var_lib_t:s0 tclass=file permissive=1', 'type=AVC msg=audit(1576778805.967:6880): avc: denied { write } for pid=30757 comm="rhsmcertd-worke" name="dnf" dev="sda1" ino=60792 scontext=system_u:system_r:rhsmcertd_t:s0 tcontext=system_u:object_r:rpm_var_cache_t:s0 tclass=dir permissive=1', 'type=AVC msg=audit(1576778806.515:6885): avc: denied { open } for pid=30811 comm="setroubleshootd" path="/var/lib/rpm/Packages" dev="sda1" ino=262250 scontext=system_u:system_r:setroubleshootd_t:s0-s0:c0.c1023 tcontext=unconfined_u:object_r:var_lib_t:s0 tclass=file permissive=1', 'type=AVC msg=audit(1576778805.967:6880): avc: denied { open } for pid=30757 comm="rhsmcertd-worke" path="/var/cache/dnf/metadata_lock.pid" dev="sda1" ino=56725 scontext=system_u:system_r:rhsmcertd_t:s0 tcontext=system_u:object_r:rpm_var_cache_t:s0 tclass=file permissive=1']

pass 4615006 2019-12-19 17:26:43 2019-12-19 17:28:39 2019-12-19 17:54:39 0:26:00 0:19:35 0:06:25 smithi master ubuntu 18.04 rados/thrash-erasure-code/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} fast/fast.yaml msgr-failures/few.yaml objectstore/bluestore-comp.yaml rados.yaml recovery-overrides/{default.yaml} supported-random-distro$/{ubuntu_latest.yaml} thrashers/minsize_recovery.yaml thrashosds-health.yaml workloads/ec-small-objects-fast-read.yaml} 2
pass 4615007 2019-12-19 17:26:44 2019-12-19 17:28:57 2019-12-19 17:58:56 0:29:59 0:18:31 0:11:28 smithi master rhel 8.0 rados/mgr/{clusters/{2-node-mgr.yaml} debug/mgr.yaml objectstore/bluestore-comp.yaml supported-random-distro$/{rhel_8.yaml} tasks/workunits.yaml} 2
pass 4615008 2019-12-19 17:26:45 2019-12-19 17:29:37 2019-12-19 17:57:37 0:28:00 0:11:49 0:16:11 smithi master centos 8.0 rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml 2-recovery-overrides/{more-async-recovery.yaml} backoff/normal.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/off.yaml msgr-failures/osd-delay.yaml msgr/async.yaml objectstore/filestore-xfs.yaml rados.yaml supported-random-distro$/{centos_8.yaml} thrashers/none.yaml thrashosds-health.yaml workloads/redirect.yaml} 2
pass 4615009 2019-12-19 17:26:46 2019-12-19 17:29:51 2019-12-19 17:47:50 0:17:59 0:10:14 0:07:45 smithi master rhel 8.0 rados/singleton/{all/pg-removal-interruption.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{rhel_8.yaml}} 1
fail 4615010 2019-12-19 17:26:47 2019-12-19 17:33:15 2019-12-19 17:49:15 0:16:00 0:07:48 0:08:12 smithi master centos 8.0 rados/perf/{ceph.yaml objectstore/bluestore-stupid.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{centos_8.yaml} workloads/radosbench_4K_rand_read.yaml} 1
Failure Reason:

Command failed on smithi174 with status 1: 'sudo yum -y install python3-yaml python3-lxml librbd-devel pdsh collectl'

fail 4615011 2019-12-19 17:26:48 2019-12-19 17:33:15 2019-12-19 18:11:16 0:38:01 0:04:05 0:33:56 smithi master centos 8.0 rados/thrash-old-clients/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-install/mimic.yaml backoff/normal.yaml ceph.yaml clusters/{openstack.yaml three-plus-one.yaml} d-balancer/crush-compat.yaml distro$/{centos_latest.yaml} msgr-failures/few.yaml rados.yaml thrashers/default.yaml thrashosds-health.yaml workloads/rbd_cls.yaml} 4
Failure Reason:

Failed to fetch package version from https://shaman.ceph.com/api/search/?status=ready&project=ceph&flavor=default&distros=centos%2F8%2Fx86_64&ref=mimic

fail 4615012 2019-12-19 17:26:49 2019-12-19 17:33:33 2019-12-19 18:21:35 0:48:02 0:35:46 0:12:16 smithi master centos 8.0 rados/dashboard/{clusters/{2-node-mgr.yaml} debug/mgr.yaml objectstore/bluestore-stupid.yaml supported-random-distro$/{centos_8.yaml} tasks/dashboard.yaml} 2
Failure Reason:

Test failure: test_all (tasks.mgr.dashboard.test_rgw.RgwBucketTest)

pass 4615013 2019-12-19 17:26:50 2019-12-19 17:35:50 2019-12-19 17:51:49 0:15:59 0:09:16 0:06:43 smithi master rhel 8.0 rados/objectstore/{backends/keyvaluedb.yaml supported-random-distro$/{rhel_8.yaml}} 1
pass 4615014 2019-12-19 17:26:51 2019-12-19 17:36:37 2019-12-19 18:00:37 0:24:00 0:17:05 0:06:55 smithi master rhel 8.0 rados/singleton-nomsgr/{all/msgr.yaml rados.yaml supported-random-distro$/{rhel_8.yaml}} 1
pass 4615015 2019-12-19 17:26:52 2019-12-19 17:37:50 2019-12-19 18:23:51 0:46:01 0:38:31 0:07:30 smithi master ubuntu 18.04 rados/singleton/{all/radostool.yaml msgr-failures/many.yaml msgr/async-v1only.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}} 1
pass 4615016 2019-12-19 17:26:54 2019-12-19 17:37:51 2019-12-19 17:59:50 0:21:59 0:15:20 0:06:39 smithi master rhel 8.0 rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml 2-recovery-overrides/{more-async-recovery.yaml} backoff/peering.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/upmap.yaml msgr-failures/fastclose.yaml msgr/async-v1only.yaml objectstore/bluestore-avl.yaml rados.yaml supported-random-distro$/{rhel_8.yaml} thrashers/pggrow.yaml thrashosds-health.yaml workloads/redirect_promote_tests.yaml} 2
pass 4615017 2019-12-19 17:26:55 2019-12-19 17:37:51 2019-12-19 17:55:50 0:17:59 0:11:08 0:06:51 smithi master ubuntu 18.04 rados/basic/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/many.yaml msgr/async.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} tasks/readwrite.yaml} 2
pass 4615018 2019-12-19 17:26:56 2019-12-19 17:37:51 2019-12-19 18:09:50 0:31:59 0:24:12 0:07:47 smithi master rhel 8.0 rados/singleton/{all/random-eio.yaml msgr-failures/few.yaml msgr/async-v2only.yaml objectstore/bluestore-stupid.yaml rados.yaml supported-random-distro$/{rhel_8.yaml}} 2
pass 4615019 2019-12-19 17:26:57 2019-12-19 17:39:07 2019-12-19 20:13:11 2:34:04 2:22:57 0:11:07 smithi master ubuntu 18.04 rados/standalone/{supported-random-distro$/{ubuntu_latest.yaml} workloads/osd.yaml} 1
fail 4615020 2019-12-19 17:26:58 2019-12-19 17:39:07 2019-12-19 17:59:07 0:20:00 0:07:37 0:12:23 smithi master centos 8.0 rados/perf/{ceph.yaml objectstore/bluestore-basic-min-osd-mem-target.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{centos_8.yaml} workloads/radosbench_4K_seq_read.yaml} 1
Failure Reason:

Command failed on smithi195 with status 1: 'sudo yum -y install python3-yaml python3-lxml librbd-devel pdsh collectl'

pass 4615021 2019-12-19 17:26:59 2019-12-19 17:39:07 2019-12-19 17:55:06 0:15:59 0:09:13 0:06:46 smithi master rhel 8.0 rados/multimon/{clusters/3.yaml msgr-failures/few.yaml msgr/async-v1only.yaml no_pools.yaml objectstore/bluestore-stupid.yaml rados.yaml supported-random-distro$/{rhel_8.yaml} tasks/mon_clock_with_skews.yaml} 2
pass 4615022 2019-12-19 17:27:00 2019-12-19 17:39:21 2019-12-19 18:23:20 0:43:59 0:35:17 0:08:42 smithi master rhel 8.0 rados/monthrash/{ceph.yaml clusters/9-mons.yaml msgr-failures/few.yaml msgr/async-v1only.yaml objectstore/filestore-xfs.yaml rados.yaml supported-random-distro$/{rhel_8.yaml} thrashers/sync-many.yaml workloads/rados_mon_workunits.yaml} 2
pass 4615023 2019-12-19 17:27:01 2019-12-19 17:41:02 2019-12-19 18:13:03 0:32:01 0:21:00 0:11:01 smithi master ubuntu 18.04 rados/singleton-nomsgr/{all/multi-backfill-reject.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml}} 2
fail 4615024 2019-12-19 17:27:02 2019-12-19 17:43:51 2019-12-19 17:59:50 0:15:59 0:07:52 0:08:07 smithi master ubuntu 18.04 rados/mgr/{clusters/{2-node-mgr.yaml} debug/mgr.yaml objectstore/bluestore-low-osd-mem-target.yaml supported-random-distro$/{ubuntu_latest.yaml} tasks/cephadm_orchestrator.yaml} 2
Failure Reason:

Test failure: test_host_ls (tasks.mgr.test_cephadm_orchestrator.TestOrchestratorCli)

pass 4615025 2019-12-19 17:27:03 2019-12-19 17:43:52 2019-12-19 18:13:51 0:29:59 0:20:39 0:09:20 smithi master ubuntu 18.04 rados/thrash-erasure-code-overwrites/{bluestore-bitmap.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} fast/normal.yaml msgr-failures/fastclose.yaml rados.yaml recovery-overrides/{more-partial-recovery.yaml} supported-random-distro$/{ubuntu_latest.yaml} thrashers/careful.yaml thrashosds-health.yaml workloads/ec-small-objects-overwrites.yaml} 2
pass 4615026 2019-12-19 17:27:04 2019-12-19 17:43:56 2019-12-19 18:17:56 0:34:00 0:11:31 0:22:29 smithi master ubuntu 18.04 rados/thrash-erasure-code-shec/{ceph.yaml clusters/{fixed-4.yaml openstack.yaml} msgr-failures/osd-delay.yaml objectstore/bluestore-avl.yaml rados.yaml recovery-overrides/{more-async-partial-recovery.yaml} supported-random-distro$/{ubuntu_latest.yaml} thrashers/careful.yaml thrashosds-health.yaml workloads/ec-rados-plugin=shec-k=4-m=3-c=2.yaml} 4
fail 4615027 2019-12-19 17:27:05 2019-12-19 17:46:14 2019-12-19 18:32:14 0:46:00 0:36:51 0:09:09 smithi master centos 8.0 rados/thrash-erasure-code-isa/{arch/x86_64.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/few.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml recovery-overrides/{more-async-partial-recovery.yaml} supported-random-distro$/{centos_8.yaml} thrashers/none.yaml thrashosds-health.yaml workloads/ec-rados-plugin=isa-k=2-m=1.yaml} 2
Failure Reason:

Scrubbing terminated -- not all pgs were active and clean.

fail 4615028 2019-12-19 17:27:07 2019-12-19 17:46:14 2019-12-19 18:30:14 0:44:00 0:28:27 0:15:33 smithi master rhel 8.0 rados/thrash-erasure-code-big/{ceph.yaml cluster/{12-osds.yaml openstack.yaml} msgr-failures/few.yaml objectstore/bluestore-bitmap.yaml rados.yaml recovery-overrides/{more-active-recovery.yaml} supported-random-distro$/{rhel_8.yaml} thrashers/morepggrow.yaml thrashosds-health.yaml workloads/ec-rados-plugin=jerasure-k=4-m=2.yaml} 3
Failure Reason:

SELinux denials found on ubuntu@smithi161.front.sepia.ceph.com: ['type=AVC msg=audit(1576779494.205:15575): avc: denied { map } for pid=51760 comm="rhsmcertd-worke" path="/var/lib/rpm/__db.001" dev="sda1" ino=262271 scontext=system_u:system_r:rhsmcertd_t:s0 tcontext=unconfined_u:object_r:var_lib_t:s0 tclass=file permissive=1', 'type=AVC msg=audit(1576779494.283:15580): avc: denied { add_name } for pid=51760 comm="rhsmcertd-worke" name="metadata_lock.pid" scontext=system_u:system_r:rhsmcertd_t:s0 tcontext=system_u:object_r:rpm_var_cache_t:s0 tclass=dir permissive=1', 'type=AVC msg=audit(1576779494.822:15585): avc: denied { read } for pid=51833 comm="setroubleshootd" name="Packages" dev="sda1" ino=262250 scontext=system_u:system_r:setroubleshootd_t:s0-s0:c0.c1023 tcontext=unconfined_u:object_r:var_lib_t:s0 tclass=file permissive=1', 'type=AVC msg=audit(1576779494.397:15581): avc: denied { open } for pid=51760 comm="rhsmcertd-worke" path="/var/cache/dnf/ceph-eb28c24780af4654/repodata/repomd.xml" dev="sda1" ino=262154 scontext=system_u:system_r:rhsmcertd_t:s0 tcontext=unconfined_u:object_r:rpm_var_cache_t:s0 tclass=file permissive=1', 'type=AVC msg=audit(1576779494.822:15585): avc: denied { open } for pid=51833 comm="setroubleshootd" path="/var/lib/rpm/Packages" dev="sda1" ino=262250 scontext=system_u:system_r:setroubleshootd_t:s0-s0:c0.c1023 tcontext=unconfined_u:object_r:var_lib_t:s0 tclass=file permissive=1', 'type=AVC msg=audit(1576779494.283:15580): avc: denied { open } for pid=51760 comm="rhsmcertd-worke" path="/var/cache/dnf/metadata_lock.pid" dev="sda1" ino=59149 scontext=system_u:system_r:rhsmcertd_t:s0 tcontext=system_u:object_r:rpm_var_cache_t:s0 tclass=file permissive=1', 'type=AVC msg=audit(1576779494.205:15572): avc: denied { open } for pid=51760 comm="rhsmcertd-worke" path="/var/lib/rpm/.dbenv.lock" dev="sda1" ino=262270 scontext=system_u:system_r:rhsmcertd_t:s0 tcontext=unconfined_u:object_r:var_lib_t:s0 tclass=file permissive=1', 'type=AVC msg=audit(1576779494.574:15583): avc: denied { remove_name } for pid=51760 comm="rhsmcertd-worke" name="metadata_lock.pid" dev="sda1" ino=59149 scontext=system_u:system_r:rhsmcertd_t:s0 tcontext=system_u:object_r:rpm_var_cache_t:s0 tclass=dir permissive=1', 'type=AVC msg=audit(1576779494.205:15573): avc: denied { lock } for pid=51760 comm="rhsmcertd-worke" path="/var/lib/rpm/.dbenv.lock" dev="sda1" ino=262270 scontext=system_u:system_r:rhsmcertd_t:s0 tcontext=unconfined_u:object_r:var_lib_t:s0 tclass=file permissive=1', 'type=AVC msg=audit(1576779494.283:15580): avc: denied { create } for pid=51760 comm="rhsmcertd-worke" name="metadata_lock.pid" scontext=system_u:system_r:rhsmcertd_t:s0 tcontext=system_u:object_r:rpm_var_cache_t:s0 tclass=file permissive=1', 'type=AVC msg=audit(1576779494.656:15584): avc: denied { read } for pid=51760 comm="rhsmcertd-worke" name="satellite-5-client.module" dev="sda1" ino=57237 scontext=system_u:system_r:rhsmcertd_t:s0 tcontext=system_u:object_r:root_t:s0 tclass=file permissive=1', 'type=AVC msg=audit(1576779494.822:15587): avc: denied { map } for pid=51833 comm="setroubleshootd" path="/var/lib/rpm/Name" dev="sda1" ino=262251 scontext=system_u:system_r:setroubleshootd_t:s0-s0:c0.c1023 tcontext=unconfined_u:object_r:var_lib_t:s0 tclass=file permissive=1', 'type=AVC msg=audit(1576779494.656:15584): avc: denied { open } for pid=51760 comm="rhsmcertd-worke" path="/etc/dnf/modules.d/satellite-5-client.module" dev="sda1" ino=57237 scontext=system_u:system_r:rhsmcertd_t:s0 tcontext=system_u:object_r:root_t:s0 tclass=file permissive=1', 'type=AVC msg=audit(1576779494.822:15586): avc: denied { lock } for pid=51833 comm="setroubleshootd" path="/var/lib/rpm/Packages" dev="sda1" ino=262250 scontext=system_u:system_r:setroubleshootd_t:s0-s0:c0.c1023 tcontext=unconfined_u:object_r:var_lib_t:s0 tclass=file permissive=1', 'type=AVC msg=audit(1576779494.205:15572): avc: denied { read write } for pid=51760 comm="rhsmcertd-worke" name=".dbenv.lock" dev="sda1" ino=262270 scontext=system_u:system_r:rhsmcertd_t:s0 tcontext=unconfined_u:object_r:var_lib_t:s0 tclass=file permissive=1', 'type=AVC msg=audit(1576779494.397:15582): avc: denied { setattr } for pid=51760 comm="rhsmcertd-worke" name="beb9905f58b0597637f875222723f63fc02df6af242782946b359057f8da6f84-primary.xml.gz" dev="sda1" ino=262180 scontext=system_u:system_r:rhsmcertd_t:s0 tcontext=unconfined_u:object_r:rpm_var_cache_t:s0 tclass=file permissive=1', 'type=AVC msg=audit(1576779494.574:15583): avc: denied { unlink } for pid=51760 comm="rhsmcertd-worke" name="metadata_lock.pid" dev="sda1" ino=59149 scontext=system_u:system_r:rhsmcertd_t:s0 tcontext=system_u:object_r:rpm_var_cache_t:s0 tclass=file permissive=1', 'type=AVC msg=audit(1576779494.283:15580): avc: denied { write } for pid=51760 comm="rhsmcertd-worke" name="dnf" dev="sda1" ino=60792 scontext=system_u:system_r:rhsmcertd_t:s0 tcontext=system_u:object_r:rpm_var_cache_t:s0 tclass=dir permissive=1', 'type=AVC msg=audit(1576779494.282:15579): avc: denied { open } for pid=51760 comm="rhsmcertd-worke" path="/var/log/hawkey.log" dev="sda1" ino=60817 scontext=system_u:system_r:rhsmcertd_t:s0 tcontext=system_u:object_r:var_log_t:s0 tclass=file permissive=1', 'type=AVC msg=audit(1576779494.205:15574): avc: denied { getattr } for pid=51760 comm="rhsmcertd-worke" path="/var/lib/rpm/__db.001" dev="sda1" ino=262271 scontext=system_u:system_r:rhsmcertd_t:s0 tcontext=unconfined_u:object_r:var_lib_t:s0 tclass=file permissive=1']

pass 4615029 2019-12-19 17:27:08 2019-12-19 17:47:59 2019-12-19 18:21:58 0:33:59 0:27:59 0:06:00 smithi master rados/verify/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-thrash/default/{default.yaml thrashosds-health.yaml} msgr-failures/few.yaml msgr/async-v1only.yaml objectstore/bluestore-stupid.yaml rados.yaml tasks/rados_api_tests.yaml validater/lockdep.yaml} 2
fail 4615030 2019-12-19 17:27:09 2019-12-19 17:47:59 2019-12-19 18:03:57 0:15:58 0:10:52 0:05:06 smithi master rhel 8.0 rados/singleton/{all/rebuild-mondb.yaml msgr-failures/many.yaml msgr/async.yaml objectstore/filestore-xfs.yaml rados.yaml supported-random-distro$/{rhel_8.yaml}} 1
Failure Reason:

Command failed on smithi174 with status 13: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph quorum_status'

pass 4615031 2019-12-19 17:27:10 2019-12-19 17:49:39 2019-12-19 18:11:38 0:21:59 0:11:56 0:10:03 smithi master centos 8.0 rados/thrash/{0-size-min-size-overrides/2-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml 2-recovery-overrides/{default.yaml} backoff/peering_and_degraded.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/crush-compat.yaml msgr-failures/few.yaml msgr/async-v2only.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{centos_8.yaml} thrashers/careful.yaml thrashosds-health.yaml workloads/redirect_set_object.yaml} 2
fail 4615032 2019-12-19 17:27:11 2019-12-19 17:49:39 2019-12-19 18:41:38 0:51:59 0:04:16 0:47:43 smithi master centos 8.0 rados/thrash-old-clients/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-install/nautilus-v1only.yaml backoff/peering.yaml ceph.yaml clusters/{openstack.yaml three-plus-one.yaml} d-balancer/off.yaml distro$/{centos_latest.yaml} msgr-failures/osd-delay.yaml rados.yaml thrashers/mapgap.yaml thrashosds-health.yaml workloads/snaps-few-objects.yaml} 4
Failure Reason:

Failed to fetch package version from https://shaman.ceph.com/api/search/?status=ready&project=ceph&flavor=default&distros=centos%2F8%2Fx86_64&ref=nautilus

fail 4615033 2019-12-19 17:27:12 2019-12-19 17:52:16 2019-12-19 18:06:15 0:13:59 0:08:27 0:05:32 smithi master rhel 8.0 rados/cephadm/{fixed-2.yaml mode/root.yaml msgr/async-v1only.yaml start.yaml supported-random-distro$/{rhel_8.yaml} tasks/rados_python.yaml} 2
Failure Reason:

Command failed on smithi085 with status 1: 'sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid ff52d338-2289-11ea-827c-001a4aab830c --force'

fail 4615034 2019-12-19 17:27:13 2019-12-19 17:52:16 2019-12-19 18:32:16 0:40:00 0:34:41 0:05:19 smithi master rhel 8.0 rados/singleton-bluestore/{all/cephtool.yaml msgr-failures/many.yaml msgr/async-v1only.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{rhel_8.yaml}} 1
Failure Reason:

Command failed (workunit test cephtool/test.sh) on smithi057 with status 127: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=757a6055117c5c4493ba811ba9becbaeaad22fe5 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/cephtool/test.sh'

fail 4615035 2019-12-19 17:27:14 2019-12-19 17:52:21 2019-12-19 18:24:21 0:32:00 0:23:11 0:08:49 smithi master centos 8.0 rados/thrash-erasure-code/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} fast/normal.yaml msgr-failures/osd-delay.yaml objectstore/bluestore-low-osd-mem-target.yaml rados.yaml recovery-overrides/{more-async-recovery.yaml} supported-random-distro$/{centos_8.yaml} thrashers/morepggrow.yaml thrashosds-health.yaml workloads/ec-small-objects-many-deletes.yaml} 2
Failure Reason:

Command failed on smithi172 with status 11: u'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg deep-scrub 1.82'

pass 4615036 2019-12-19 17:27:15 2019-12-19 17:53:55 2019-12-19 18:15:55 0:22:00 0:17:07 0:04:53 smithi master rhel 8.0 rados/singleton/{all/recovery-preemption.yaml msgr-failures/few.yaml msgr/async-v1only.yaml objectstore/bluestore-avl.yaml rados.yaml supported-random-distro$/{rhel_8.yaml}} 1
fail 4615037 2019-12-19 17:27:16 2019-12-19 17:54:40 2019-12-19 18:10:40 0:16:00 0:09:22 0:06:38 smithi master rhel 8.0 rados/perf/{ceph.yaml objectstore/bluestore-bitmap.yaml openstack.yaml settings/optimized.yaml supported-random-distro$/{rhel_8.yaml} workloads/radosbench_4M_rand_read.yaml} 1
Failure Reason:

Command failed on smithi096 with status 1: 'sudo yum -y install python3-yaml python3-lxml librbd-devel pdsh collectl'

pass 4615038 2019-12-19 17:27:18 2019-12-19 17:55:30 2019-12-19 18:17:29 0:21:59 0:15:30 0:06:29 smithi master ubuntu 18.04 rados/basic/{ceph.yaml clusters/{fixed-2.yaml openstack.yaml} msgr-failures/few.yaml msgr/async-v1only.yaml objectstore/bluestore-stupid.yaml rados.yaml supported-random-distro$/{ubuntu_latest.yaml} tasks/repair_test.yaml} 2
pass 4615039 2019-12-19 17:27:19 2019-12-19 17:55:50 2019-12-19 18:15:49 0:19:59 0:11:37 0:08:22 smithi master centos 8.0 rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml 2-recovery-overrides/{default.yaml} backoff/normal.yaml ceph.yaml clusters/{fixed-2.yaml openstack.yaml} d-balancer/off.yaml msgr-failures/osd-delay.yaml msgr/async.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{centos_8.yaml} thrashers/default.yaml thrashosds-health.yaml workloads/set-chunk-promote-flush.yaml} 2
pass 4615040 2019-12-19 17:27:20 2019-12-19 17:55:51 2019-12-19 18:33:51 0:38:00 0:30:09 0:07:51 smithi master centos 8.0 rados/objectstore/{backends/objectcacher-stress.yaml supported-random-distro$/{centos_8.yaml}} 1
pass 4615041 2019-12-19 17:27:21 2019-12-19 17:57:35 2019-12-19 18:23:32 0:25:57 0:18:03 0:07:54 smithi master centos 8.0 rados/singleton-nomsgr/{all/osd_stale_reads.yaml rados.yaml supported-random-distro$/{centos_8.yaml}} 1
pass 4615042 2019-12-19 17:27:22 2019-12-19 17:57:38 2019-12-19 18:15:37 0:17:59 0:08:57 0:09:02 smithi master centos 8.0 rados/singleton/{all/resolve_stuck_peering.yaml msgr-failures/many.yaml msgr/async-v2only.yaml objectstore/bluestore-bitmap.yaml rados.yaml supported-random-distro$/{centos_8.yaml}} 2
fail 4615043 2019-12-19 17:27:23 2019-12-19 17:58:25 2019-12-19 18:34:24 0:35:59 0:27:34 0:08:25 smithi master centos 8.0 rados/singleton/{all/test-crash.yaml msgr-failures/few.yaml msgr/async.yaml objectstore/bluestore-comp.yaml rados.yaml supported-random-distro$/{centos_8.yaml}} 1
Failure Reason:

Command failed (workunit test rados/test_crash.sh) on smithi178 with status 1: 'mkdir -p -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=757a6055117c5c4493ba811ba9becbaeaad22fe5 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 CEPH_ROOT=/home/ubuntu/cephtest/clone.client.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/rados/test_crash.sh'

pass 4615044 2019-12-19 17:27:24 2019-12-19 17:59:02 2019-12-19 18:17:01 0:17:59 0:09:57 0:08:02 smithi master centos 8.0 rados/mgr/{clusters/{2-node-mgr.yaml} debug/mgr.yaml objectstore/bluestore-stupid.yaml supported-random-distro$/{centos_8.yaml} tasks/crash.yaml} 2