Description: kcephfs/recovery/{clusters/1-mds-4-client.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/filestore-xfs.yaml overrides/{debug.yaml frag_enable.yaml log-config.yaml osd-asserts.yaml whitelist_health.yaml whitelist_wrongly_marked_down.yaml} tasks/damage.yaml whitelist_health.yaml}

Log: http://qa-proxy.ceph.com/teuthology/yuriw-2018-04-06_21:41:01-kcephfs-wip-yuri5-testing-2018-04-06-1944-luminous-testing-basic-smithi/2365504/teuthology.log

  • log_href: http://qa-proxy.ceph.com/teuthology/yuriw-2018-04-06_21:41:01-kcephfs-wip-yuri5-testing-2018-04-06-1944-luminous-testing-basic-smithi/2365504/teuthology.log
  • archive_path: /home/teuthworker/archive/yuriw-2018-04-06_21:41:01-kcephfs-wip-yuri5-testing-2018-04-06-1944-luminous-testing-basic-smithi/2365504
  • description: kcephfs/recovery/{clusters/1-mds-4-client.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml mounts/kmounts.yaml objectstore-ec/filestore-xfs.yaml overrides/{debug.yaml frag_enable.yaml log-config.yaml osd-asserts.yaml whitelist_health.yaml whitelist_wrongly_marked_down.yaml} tasks/damage.yaml whitelist_health.yaml}
  • duration: 0:18:31
  • email: ceph-qa@ceph.com
  • failure_reason:
  • flavor: basic
  • job_id: 2365504
  • kernel:
    • flavor: basic
    • sha1: 9815bce65e8113cd416b3b39dafa16e76b0d000b
    • kdb: True
  • last_in_suite: False
  • machine_type: smithi
  • name: yuriw-2018-04-06_21:41:01-kcephfs-wip-yuri5-testing-2018-04-06-1944-luminous-testing-basic-smithi
  • nuke_on_error: True
  • os_type:
  • os_version:
  • overrides:
    • ceph:
      • log-whitelist:
        • slow request
        • overall HEALTH_
        • \(FS_DEGRADED\)
        • \(MDS_FAILED\)
        • \(MDS_DEGRADED\)
        • \(FS_WITH_FAILED_MDS\)
        • \(MDS_DAMAGE\)
        • overall HEALTH_
        • \(OSD_DOWN\)
        • \(OSD_
        • but it is still running
        • is not responding
        • bad backtrace
        • object missing on disk
        • error reading table object
        • error reading sessionmap
        • Error loading MDS rank
        • missing journal object
        • Error recovering journal
        • error decoding table object
        • failed to read JournalPointer
        • Corrupt directory entry
        • Corrupt fnode header
        • corrupt sessionmap header
        • Corrupt dentry
        • Scrub error on inode
        • Metadata damage detected
        • overall HEALTH_
        • \(FS_DEGRADED\)
        • \(MDS_FAILED\)
        • \(MDS_DEGRADED\)
        • \(FS_WITH_FAILED_MDS\)
        • \(MDS_DAMAGE\)
      • fs: xfs
      • conf:
        • client.1:
          • debug ms: 1
          • debug client: 20
        • client.0:
          • debug ms: 1
          • debug client: 20
        • client:
          • debug ms: 1
          • debug client: 10
        • mon:
          • debug mon: 20
          • debug paxos: 20
          • debug ms: 1
        • mds:
          • mds bal split bits: 3
          • mds bal split size: 100
          • debug mds: 20
          • mds bal merge size: 5
          • debug ms: 1
          • mds bal frag: True
          • mds bal fragment size max: 10000
        • osd:
          • debug ms: 1
          • debug journal: 20
          • debug osd: 25
          • osd objectstore: filestore
          • osd sloppy crc: True
          • debug filestore: 20
          • osd shutdown pgref assert: True
      • sha1: 6a533d43db1af912c5a701428f9b43e10b52b620
    • ceph-deploy:
      • fs: xfs
      • filestore: True
      • conf:
        • client:
          • log file: /var/log/ceph/ceph-$name.$pid.log
        • mon:
          • osd default pool size: 2
        • osd:
          • osd sloppy crc: True
          • osd objectstore: filestore
    • workunit:
      • sha1: 6a533d43db1af912c5a701428f9b43e10b52b620
      • branch: wip-yuri5-testing-2018-04-06-1944-luminous
    • install:
      • ceph:
        • sha1: 6a533d43db1af912c5a701428f9b43e10b52b620
    • admin_socket:
      • branch: wip-yuri5-testing-2018-04-06-1944-luminous
  • owner: scheduled_yuriw@teuthology
  • pid:
  • roles:
    • ['mon.a', 'mgr.y', 'mds.a', 'osd.0', 'osd.1', 'osd.2', 'osd.3']
    • ['mon.b', 'mon.c', 'mgr.x', 'mds.a-s', 'osd.4', 'osd.5', 'osd.6', 'osd.7']
    • ['client.0']
    • ['client.1']
    • ['client.2']
    • ['client.3']
  • sentry_event:
  • status: pass
  • success: True
  • branch: wip-yuri5-testing-2018-04-06-1944-luminous
  • seed:
  • sha1: 6a533d43db1af912c5a701428f9b43e10b52b620
  • subset:
  • suite:
  • suite_branch: wip-yuri5-testing-2018-04-06-1944-luminous
  • suite_path:
  • suite_relpath:
  • suite_repo:
  • suite_sha1: 6a533d43db1af912c5a701428f9b43e10b52b620
  • targets:
    • smithi200.front.sepia.ceph.com: ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQDB4PLC3XtRtP99cUgltbzcktVHw3Y6azzTtDs7I6VA7Hgi6trHU1AIAOaZcYSdnBKSz6E9kBqeIhe6wxS8m2JjVwcDea2UkbeIH0U9ng9IO5QTZrl3p2hp73dUEbkVnVpO0Se/cEYIQ8+EtVWMyGMsKXTz2VP7sHhZxDvC/f3+HOnJ3J4c4Imidol9gpQoxms9nCWKBWeFtFQOjiWFLj4ckEiiNhp+2x2Kdroqr2yO24VrLDIOi+IHaXxMoDZrqsL+EnKPxkWZwyUlVDdfHSBmWAY1VMmETQOeSn3D3o/lN13ssZTsIvwciYsBdrsNsa/SHZIp9WCWlNLc+Wz+vGst
    • smithi169.front.sepia.ceph.com: ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQC5tNsh+MhkoDCO61DEfTdwNE7b/KP4FaI9npB4hmP+MtObs1GHmAjvhhRoe3RatL2HtNpDITzC437rvtAIqXQbxYXTn57HvQVxe4VdTa3K1vH0a+KY+YS8+4TNIuQekaoNTNrVyqXBmBlkx+Jt+SxzR1LLOXDQ97IrN6m1ccG5ngPBDojM9vJXffCDrkKVWKO7vHJ1T8PUad2mivb8VYXt4n/PI+so5b5+N77JMuWXP+IPs2LeKe8QG+JNHV2diBcjmjemuCymkN1ykCOuN6bWCbxP7IzwZmdd3lb8y1/Z51wL1Eh9JaBJScwnuoOUaxA4NbAl98hMwE4czw1pSnKx
    • smithi197.front.sepia.ceph.com: ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQDTdBmUEK1KgB1otM0Y4zEdAXAYYQFuAV9KMmCmAtWdPEekYV5TO23W1UPaFTLdnxze0ZEIr1oLZzG9VPNX0+7DaqdkTXDhmh0QBqYn5G4qY4fzz9GwY6MkVU85hUWXERfcQ252mbhROWYERFN/iIR5tE6NDQ9vRDB3xjmIBaTgH/nVmf8SINAT6XRM8fUvTK0yv103N0+7+6oNnGRLFKqZLR1r8VQuriYTcxI48e2Iqq1GqzcrCvV9YPk3JbKlKSXrr7dzuS22XSvdibE2/1DkvxIL5J2OexNKS48ZIQleB7jfhpVvuZ/DiA1WPMxMriCjG5JH7OY0xl6zLb9gSznt
    • smithi008.front.sepia.ceph.com: ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQC8dRLCogIpAjlFt9FrUov4SuG2iZHNojNgTta2YZ6iGTeDWOqmWudEQ/VTk3880W9ujQ7oVBHTjJDJE8clmqjOzpVp0u0CLUL+MlHf6/zAUn0L8M70geTkwQCNdpZKx2DY0VlAZO/cZSUj/cnRqmYEjFZT4LgAnTBIfC2r61ReDqLoEP3+U4VhIA18w5uKHr4GGi6ILVLimivNkx1Oa8ffhhuwHPylXK4GAyhIDKUJB+WrDHQcSFg1+7W4m9czQgLH//upOOMZ6qRIVh0IJ2iKwpeNsrcIBjBGFlfwM8qpc0ttp8ZbaPNoc5Rot24jLzuVGyH1bVQdeNNNEm5Tlcuv
    • smithi077.front.sepia.ceph.com: ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQDtPTe9MzVTI2oODEf74/nQ6a3DIgOUyd577Nhgob9W32YvEjeJ8xl5MeWuI+Rh24rGHOh2JR9gvEuUod/H5DEh4tSh3nPa/xw3dtcmTQC7AFSE5tE0Zen4ik1i9nr5/M4y2iAWefTSbQID4NjHCeUmvJwgJ88gStwPi2sgJitFAeqcX1qWVvchOcBvbf4gFi+27Ic0V1E9H5qLpdSWQ3/F5Uvn1dzMPSIIu9Ymqcx1010v1FBmkkHXo6JR3nhagMZnZbirYTs+vz3HqQVUscItR4jU3csWJdjA1Disn+Tg50nZBsx6gIe739CN5lK4TXJOP0rqfOPtZOI7z2+/zrTz
    • smithi108.front.sepia.ceph.com: ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQChGtCsDe+hYGZN8MqSMond0URJ528lqXKo0Jbtdjkutf1fikVu42IDdWJjD+Ad8PrumtK7YT1xF1EoGII5bMBQUBxDgdmunHpMljIAY75POQd0REgfRmeGQeyJAxD0WD/XaWa0vQITrSGMJqeK2KE/HuTC6tolaw9PpU6JSCAWgen2bRtlac1QvdWfD1QJ/lJbiWXPNcmiZyflgi+jUAO9+5dIuY6Y0Dtekrc9JhU/tVchjP8nkcBBkNLpbOVE9mYJpMuxpcYDxPuSp8icxEY1Sdb4tgXHY4/+6pG/uEejvITyuEaQwzUbo/YtXbRU/VZZQQyQFUUubgBUyD0VsnmV
  • tasks:
    • internal.check_packages:
    • internal.buildpackages_prep:
    • internal.lock_machines:
      • 6
      • smithi
    • internal.save_config:
    • internal.check_lock:
    • internal.add_remotes:
    • console_log:
    • internal.connect:
    • internal.push_inventory:
    • internal.serialize_remote_roles:
    • internal.check_conflict:
    • internal.check_ceph_data:
    • internal.vm_setup:
    • kernel:
      • flavor: basic
      • sha1: 9815bce65e8113cd416b3b39dafa16e76b0d000b
      • kdb: True
    • internal.base:
    • internal.archive_upload:
    • internal.archive:
    • internal.coredump:
    • internal.sudo:
    • internal.syslog:
    • internal.timer:
    • pcp:
    • selinux:
    • ansible.cephlab:
    • clock:
    • install:
    • ceph:
    • kclient:
    • cephfs_test_runner:
      • modules:
        • tasks.cephfs.test_damage
  • teuthology_branch: master
  • verbose: True
  • pcp_grafana_url:
  • priority:
  • user:
  • queue:
  • posted: 2018-04-06 21:41:36
  • started: 2018-04-07 00:52:25
  • updated: 2018-04-07 02:20:26
  • status_class: success
  • runtime: 1:28:01
  • wait_time: 1:09:30