Description: rbd/maintenance/{base/install clusters/{fixed-3 openstack} objectstore/bluestore-comp-zstd qemu/xfstests supported-random-distro$/{rhel_8} workloads/rebuild_object_map}

Log: http://qa-proxy.ceph.com/teuthology/teuthology-2020-11-22_02:01:03-rbd-master-distro-basic-smithi/5644989/teuthology.log

Failure Reason:

[Errno 28] No space left on device

  • log_href: http://qa-proxy.ceph.com/teuthology/teuthology-2020-11-22_02:01:03-rbd-master-distro-basic-smithi/5644989/teuthology.log
  • archive_path: /home/teuthworker/archive/teuthology-2020-11-22_02:01:03-rbd-master-distro-basic-smithi/5644989
  • description: rbd/maintenance/{base/install clusters/{fixed-3 openstack} objectstore/bluestore-comp-zstd qemu/xfstests supported-random-distro$/{rhel_8} workloads/rebuild_object_map}
  • duration: 1:51:18
  • email: ceph-qa@ceph.io
  • failure_reason: [Errno 28] No space left on device
  • flavor: basic
  • job_id: 5644989
  • kernel:
    • sha1: distro
    • kdb: True
  • last_in_suite: False
  • machine_type: smithi
  • name: teuthology-2020-11-22_02:01:03-rbd-master-distro-basic-smithi
  • nuke_on_error: True
  • os_type: rhel
  • os_version: 8.3
  • overrides:
    • ceph-deploy:
      • conf:
        • client:
          • log file: /var/log/ceph/ceph-$name.$pid.log
        • mon:
          • osd default pool size: 2
    • selinux:
      • whitelist:
        • scontext=system_u:system_r:logrotate_t:s0
    • workunit:
      • sha1: 2d5830650cd847384a4f9e95d040a197a48dd5ed
      • branch: master
    • ceph:
      • log-whitelist:
        • \(MDS_ALL_DOWN\)
        • \(MDS_UP_LESS_THAN_MAX\)
      • sha1: 0ba966763e22c896b164f58fd3df4d640286dfe3
      • fs: xfs
      • conf:
        • mgr:
          • debug ms: 1
          • debug mgr: 20
        • mon:
          • debug paxos: 20
          • debug mon: 20
          • debug ms: 1
        • osd:
          • mon osd full ratio: 0.9
          • bluestore compression algorithm: zstd
          • debug ms: 20
          • bluestore fsck on mount: True
          • debug osd: 25
          • bluestore compression mode: aggressive
          • debug bluestore: 20
          • debug bluefs: 20
          • osd objectstore: bluestore
          • mon osd backfillfull_ratio: 0.85
          • bluestore block size: 96636764160
          • osd shutdown pgref assert: True
          • debug filestore: 20
          • debug rocksdb: 10
          • mon osd nearfull ratio: 0.8
          • osd failsafe full ratio: 0.95
          • debug journal: 20
      • log-ignorelist:
        • \(MDS_ALL_DOWN\)
        • \(MDS_UP_LESS_THAN_MAX\)
    • install:
      • ceph:
        • sha1: 0ba966763e22c896b164f58fd3df4d640286dfe3
    • admin_socket:
      • branch: master
    • thrashosds:
      • bdev_inject_crash_probability: 0.5
      • bdev_inject_crash: 2
  • owner: scheduled_teuthology@teuthology
  • pid:
  • roles:
    • ['mon.a', 'mon.c', 'mgr.x', 'osd.0', 'osd.1', 'osd.2', 'osd.3']
    • ['mon.b', 'mgr.y', 'osd.4', 'osd.5', 'osd.6', 'osd.7']
    • ['client.0']
  • sentry_event:
  • status: fail
  • success: False
  • branch: master
  • seed:
  • sha1: 0ba966763e22c896b164f58fd3df4d640286dfe3
  • subset:
  • suite:
  • suite_branch: master
  • suite_path:
  • suite_relpath:
  • suite_repo:
  • suite_sha1: 2d5830650cd847384a4f9e95d040a197a48dd5ed
  • targets:
    • smithi117.front.sepia.ceph.com: ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQC6SasoTHxFY8K5hIk5E5OIjGGsiV1g0/ulWUznAGDYPlaf4QLJKKyjfbBP6RZ4c0fDw1g6b87jmm0xMUNdkn6HGa1S1zGJqykgmOOzBtyWktPcrnHEW9gljJsmdtY5nHxjOHBSTeropxMRrlTus83IXKFOlVcNGFQM2S0TIpadYdo3pXM5INvJxwNGXjx5ZifYDVOATjNr2Cp+Elyb7msT4cnmFeNEoHHiozI/zGXqOazHWI0oAb+t9dy8U4FeHZh4l1g1rFW1rpXHRNaj4gXY3q9mPBqVKY944GX4+6LuGU2UxOKH2oUmLUbQinRESmCYB9PlcB7Sumf+2OzAY3y9MaFwlolTFE+plsmfhPSURKkKhG7W1M1eG1KdwiPAHvPwl8qDwkLpxMYlHBbiniRYqqTvOei6byZkjlLRhGX8bGcPDPlyACwrKn5L+fIruXsc25CuZmPZD/2KJZA62X08n3LET/Et64Uf6zL7rocWoEzfNd9D+2z+HCLr1wxSnF0=
    • smithi184.front.sepia.ceph.com: ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCmyYKvWnl5OmL08PS4G5LMj9igfa5GKIuUWOTud4AxHJ7Io9BYQCw+yaToPTExH5nPUvyctz47uLAdD0ONPwPzjueMlgCqv6JV8YERb1gidoIJt/aSSPWkc2f4GHb6FF3TDUcgnFimoOZTSr+Q3k6eEBDknNgt4Xv14MtbzDcf9uiva1575BLodwRosw3iMY7Y6zaw8jHz1bQCQ+7lYdWfR86/WXHqmC8HU1YYbTDdnf8wB7EABTS8gw9jmqyFR/tGWrN7aGuzt8RXeHw/2QOgLahUVBYVvwXCx5RmcLl5pYl9B/p+IJCzBHVXlU7MuzACuKRD42sLk6TIUmOPRvUBo2qrq8RPYRa/0IgTeGaBUxZ5O8jqM6cQXcFhdIZ+HdG4OVumFr8h/bKgU3tbhIj6l4XVpQ4n35DHSiPlldiENdpEh113088F2eTbCk2r7Xvs+O0ZwVnxGhytFETT3kn+IE9XWeZU+9T/RW+QeQV7MksGd355mEE5ByIGQ6xW3Ik=
    • smithi043.front.sepia.ceph.com: ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQC58KSgjUxvwq+0UY57E97kox5hI0GqFtCa1xj3DlmIZWxws8h+1nrKjd88rsXm1xeek5P9IjqrZTFsbnkXHmrUU0Cvw3aH1DDw6ub7f9qCs0rn+qXZanx8oPR8AR4RNH5RQRk9iyM5MBT2RzGkp7ox2LL29ilB5J9G39+IK/U5oP6WXbDzfeL94M5IjQbCuN5Mnzb3uA0N5KJPrqts48Us0hW4W8q9T2wxct3T9eZNYmTABORiJBxyRCVf8OnPGMXl7IRN4u7SYLW7yr6ra5XydoZ3TokWLHiaKgoi7VW+aTstPUj/JKIzZccnqa/tgKuGNQWn9ntYHFRb2wlXeYjupaP+kU05H9IHUm7kjxbjujJzC0Ly3JJ1x1GQ2tvdKhwuscxmNL6/ngn0WcSRqktRg9OnXg+A57Gij+TWnXwc6Iv4lGcxRUL02ZN+bNurIc8QQ8Xshrsj5Wziz4nqUogn35JIpbGmiWpm87MW9jTuSRc4rp1H3bdes0q3XFVV7Ws=
  • tasks:
    • internal.check_packages:
    • internal.buildpackages_prep:
    • internal.lock_machines:
      • 3
      • smithi
    • internal.save_config:
    • internal.check_lock:
    • internal.add_remotes:
    • console_log:
    • internal.connect:
    • internal.push_inventory:
    • internal.serialize_remote_roles:
    • internal.check_conflict:
    • internal.check_ceph_data:
    • internal.vm_setup:
    • kernel:
      • sha1: distro
      • kdb: True
    • internal.base:
    • internal.archive_upload:
    • internal.archive:
    • internal.coredump:
    • internal.sudo:
    • internal.syslog:
    • internal.timer:
    • pcp:
    • selinux:
    • ansible.cephlab:
    • clock:
    • install:
    • ceph:
    • parallel:
      • io_workload
      • op_workload
  • teuthology_branch: master
  • verbose: True
  • pcp_grafana_url:
  • priority:
  • user:
  • queue:
  • posted: 2020-11-22 02:04:35
  • started: 2020-11-22 06:00:05
  • updated: 2020-11-22 08:34:11
  • status_class: danger
  • runtime: 2:34:06
  • wait_time: 0:42:48