Description: kcephfs/recovery/{clusters/1-mds-4-client.yaml conf/{client.yaml mds.yaml mon.yaml osd.yaml} mounts/kmounts.yaml objectstore-ec/filestore-xfs.yaml overrides/{frag_enable.yaml log-config.yaml osd-asserts.yaml whitelist_health.yaml whitelist_wrongly_marked_down.yaml} tasks/failover.yaml whitelist_health.yaml}

Log: http://qa-proxy.ceph.com/teuthology/yuriw-2019-01-11_13:48:59-kcephfs-wip_luminous_1.10.2019-testing-basic-smithi/3448001/teuthology.log

  • log_href: http://qa-proxy.ceph.com/teuthology/yuriw-2019-01-11_13:48:59-kcephfs-wip_luminous_1.10.2019-testing-basic-smithi/3448001/teuthology.log
  • archive_path: /home/teuthworker/archive/yuriw-2019-01-11_13:48:59-kcephfs-wip_luminous_1.10.2019-testing-basic-smithi/3448001
  • description: kcephfs/recovery/{clusters/1-mds-4-client.yaml conf/{client.yaml mds.yaml mon.yaml osd.yaml} mounts/kmounts.yaml objectstore-ec/filestore-xfs.yaml overrides/{frag_enable.yaml log-config.yaml osd-asserts.yaml whitelist_health.yaml whitelist_wrongly_marked_down.yaml} tasks/failover.yaml whitelist_health.yaml}
  • duration: 0:14:25
  • email: ceph-qa@lists.ceph.com
  • failure_reason:
  • flavor: basic
  • job_id: 3448001
  • kernel:
    • flavor: basic
    • sha1: 9e78dcdd94782ed234d6c3e148674b799a2c4a8b
    • kdb: True
  • last_in_suite: False
  • machine_type: smithi
  • name: yuriw-2019-01-11_13:48:59-kcephfs-wip_luminous_1.10.2019-testing-basic-smithi
  • nuke_on_error: True
  • os_type:
  • os_version:
  • overrides:
    • ceph:
      • log-whitelist:
        • \(MDS_ALL_DOWN\)
        • \(MDS_UP_LESS_THAN_MAX\)
        • overall HEALTH_
        • \(FS_DEGRADED\)
        • \(MDS_FAILED\)
        • \(MDS_DEGRADED\)
        • \(FS_WITH_FAILED_MDS\)
        • \(MDS_DAMAGE\)
        • overall HEALTH_
        • \(OSD_DOWN\)
        • \(OSD_
        • but it is still running
        • is not responding
        • not responding, replacing
        • \(MDS_INSUFFICIENT_STANDBY\)
        • overall HEALTH_
        • \(FS_DEGRADED\)
        • \(MDS_FAILED\)
        • \(MDS_DEGRADED\)
        • \(FS_WITH_FAILED_MDS\)
        • \(MDS_DAMAGE\)
      • fs: xfs
      • conf:
        • mds:
          • mds bal split bits: 3
          • mds bal split size: 100
          • osd op complaint time: 180
          • debug mds: 20
          • mds bal merge size: 5
          • debug ms: 1
          • mds bal frag: True
          • mds verify scatter: True
          • mds bal fragment size max: 10000
          • mds op complaint time: 180
          • mds debug scatterstat: True
          • mds debug frag: True
        • client:
          • debug ms: 1
          • debug client: 20
          • client mount timeout: 600
        • mon:
          • debug ms: 1
          • debug mon: 20
          • debug paxos: 20
          • mon op complaint time: 120
        • osd:
          • debug ms: 1
          • debug journal: 20
          • debug osd: 25
          • osd objectstore: filestore
          • osd sloppy crc: True
          • debug filestore: 20
          • osd shutdown pgref assert: True
          • osd op complaint time: 180
      • sha1: 185f2c4ccf4822a9efe29d09b4321a6a2ea37dae
    • ceph-deploy:
      • fs: xfs
      • filestore: True
      • conf:
        • client:
          • log file: /var/log/ceph/ceph-$name.$pid.log
        • mon:
          • osd default pool size: 2
        • osd:
          • osd sloppy crc: True
          • osd objectstore: filestore
    • workunit:
      • sha1: 185f2c4ccf4822a9efe29d09b4321a6a2ea37dae
      • branch: wip_luminous_1.10.2019
    • install:
      • ceph:
        • sha1: 185f2c4ccf4822a9efe29d09b4321a6a2ea37dae
    • admin_socket:
      • branch: wip_luminous_1.10.2019
  • owner: scheduled_yuriw@teuthology
  • pid:
  • roles:
    • ['mon.a', 'mgr.y', 'mds.a', 'osd.0', 'osd.1', 'osd.2', 'osd.3']
    • ['mon.b', 'mon.c', 'mgr.x', 'mds.a-s', 'osd.4', 'osd.5', 'osd.6', 'osd.7']
    • ['client.0']
    • ['client.1']
    • ['client.2']
    • ['client.3']
  • sentry_event:
  • status: pass
  • success: True
  • branch: wip_luminous_1.10.2019
  • seed:
  • sha1: 185f2c4ccf4822a9efe29d09b4321a6a2ea37dae
  • subset:
  • suite:
  • suite_branch: wip_luminous_1.10.2019
  • suite_path:
  • suite_relpath:
  • suite_repo:
  • suite_sha1: 185f2c4ccf4822a9efe29d09b4321a6a2ea37dae
  • targets:
    • smithi121.front.sepia.ceph.com: ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQC0NE9TH6pENM3Cgvxs3pp38fFYR/CMbvang7rQLWOUe4rhFP8wJICNgHDrNlejOV9LxGVQc25ODI8lFSFEB4TPEDFJdCk7aT5YT74ZB4lGFWFLKxUOV73Khi7Q+Tt/qHHjjYZYuQJeFE6xm81m7zFQzguM39JDxF0t2BQTlpDFaUKiRXSy7Xliojys7bQbZycFVuW6dhhQ3Rd6uN8v066UhJoPgSrEILkL3kgrWaqK4Xdvfeu/mGc3a2ittSlJNeewolQhN4u9XItWHWTo6gS0g38slHJDAiQSgbQ2oLTHWqBtzHBpc48jBnbODyPHdTZUBv+W72El9S3WZz9uLpm9
    • smithi174.front.sepia.ceph.com: ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQCyTG4NhhIkCzT8HLBZGzJoXJPialf73zyPeL6pLE/Dje1PDzRf5/Jb0EyKVU/ImkUbJKhR3fGS17ALcWjWZjotmXmGbrLtNZFiHKTmorimKOOUJSguVc8qxtySsiMJOvCiZHj2R8ICZaYTSY+j6bt3h2/xXwT+2Tf0Iu/WS4FMxYgTfyFHyHF4lWzJXpn6cJTN1nYKhUzrsxQA0soZW8f0+ZvLxSUIjtQM14MpNAhwm8Acp3MEelKJOTSK+5bF9PQAbr9SXe3inC8ZOsQ8QoPmFiQ6VzEaebRgeFMXBfY++FDMcdBCyyg9TNpD5gDYsZauq+DBvOp3f1xGwIPrQuot
    • smithi192.front.sepia.ceph.com: ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQCzksVDGO7p+WySM66Ki0/E2vXOBDXV9LPjP4FCG1XY5WJJH1qStPiBrgpTgDacK78XUoXyZGtbP0DcO40Qme+ZVwMhB0daV5atLipOECNCpTzEZTaSyUdGEUmNYjhle2LaX+adLhCdbconDcyYuCOhpjxOn21ecFJ/Cf1w8lP/COSTGdj4CYSARfW45XYnr9c8tXP3YhnnrejwhH0d4WbO2fP/em4nKzUNgM4KkbywNAuV+Ttao0JIqpWC0X3pZbNwk2IUavhqmaZpo4a/Bwck1/hyQxzkSMjtSUat3SHU25L6p/EEz+vY7GBHQ74D1oysDUugrNsaQRvYa1IMNuhz
    • smithi175.front.sepia.ceph.com: ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQD6Xk0TnLWVhPfcPeB4zkY6tUQE6579oa6EnLyJYNJxTJ8gUKN1PfFSzR6d5A9y1mZsdM3QUrvXIjs5Iu+YVDsPfSbgdCCgdWknG/71J5N+V30q118+vE+xo0ln7AbtEOYUj1A6UYNiJSf54bOBJFEpIgRt18ZV4ac288CCwKyDJMEhEYEyrs83CD//LCzch2GKon1X5e3ND1fxOGmEuiLXD37VerhEFdsPOS3UuWPnzIt27SfUIKKi1bJH3ILYzvalGj8WdZpck06SwSxg2dLsvEJ9s16D00bNb0vb2yxYmEPZet+TsfldxqNd1AqsKL0YlSruaiZz4P4cUyl069cH
    • smithi191.front.sepia.ceph.com: ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQDurXNG21zElgocOEMInSx9QwD7o2ocu7F0HSD4JfTveiq9TbRBXTt9KhSKzayjnjNV8jODUOSX2TFM5xy9lqeGVD/jvTAwG1IJ6pFVVRSgRKysI+tpz6S9BKm4+Kai/J5SiIyoJufOvwDZW28E4ulGgPUPJT+SIpNM5Z2EJDOTAr4gH2ZGHpYO4IZIRPEFIVfNdO/mxA9V1AZOOq5S6T7QRE6rarpdFRnnGQ75xV+S9hjpLbUHj5521oxCyQqSMK0kaVCOAeG5IcN9Te1zBOV8TWB9QMWy0VEk1M+XN/JsPJn8dc1NXSt3K/CPxpZFGo4AGYOHAl1uU+nXBYmXMMQV
    • smithi205.front.sepia.ceph.com: ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQC91vrCpGxvjh6FgypjglamIhfp0mgp9Tnr37D/HXCq5VbP5Hr3gArnSR/OSrawWHyLKewTXt7jUnDSWvnhjX8/2AcIc/fkZLHk0FEgacBLYjX4SzFgOmIvNNmAlg1UXQcE0dSiXfFX7acsfCZHB5mL/EaMzEVkuVf/O33XoT4K+4RyhCUpGDjle+DuZUlHaWxjPFdLGcQMRglgNcbJ1hj+4PDl9RIinopFdAaHreAxfcJlljhxvUNcYUx1lNSdgaQ9FVuDS71qsJDJv17vH/6Cg4s1qO+gp4r0tIiqD8Dazg1apME8ZBGJjjuzlxr9Ald+3+EmxNzoMONfJ8Wxjzmx
  • tasks:
    • internal.check_packages:
    • internal.buildpackages_prep:
    • internal.lock_machines:
      • 6
      • smithi
    • internal.save_config:
    • internal.check_lock:
    • internal.add_remotes:
    • console_log:
    • internal.connect:
    • internal.push_inventory:
    • internal.serialize_remote_roles:
    • internal.check_conflict:
    • internal.check_ceph_data:
    • internal.vm_setup:
    • kernel:
      • flavor: basic
      • sha1: 9e78dcdd94782ed234d6c3e148674b799a2c4a8b
      • kdb: True
    • internal.base:
    • internal.archive_upload:
    • internal.archive:
    • internal.coredump:
    • internal.sudo:
    • internal.syslog:
    • internal.timer:
    • pcp:
    • selinux:
    • ansible.cephlab:
    • clock:
    • install:
    • ceph:
    • kclient:
    • cephfs_test_runner:
      • modules:
        • tasks.cephfs.test_failover
      • fail_on_skip: False
  • teuthology_branch: master
  • verbose: True
  • pcp_grafana_url:
  • priority:
  • user:
  • queue:
  • posted: 2019-01-11 13:49:35
  • started: 2019-01-11 16:42:57
  • updated: 2019-01-11 18:26:58
  • status_class: success
  • runtime: 1:44:01
  • wait_time: 1:29:36