# This playbook migrates an OSD from filestore to bluestore backend. # # Use it like this: # ansible-playbook infrastructure-playbooks/filestore-to-bluestore.yml --limit # If all osds on the node are using filestore backend, then *ALL* of them will be shrinked and redeployed using bluestore backend with ceph-volume. # # If a mix of filestore and bluestore OSDs is detected on the node, the node will be skipped unless you pass `force_filestore_to_bluestore=True` as an extra var. # ie: ansible-playbook infrastructure-playbooks/filestore-to-bluestore.yml --limit -e force_filestore_to_bluestore=True - hosts: "{{ osd_group_name }}" become: true serial: 1 vars: delegate_facts_host: true tasks: - name: gather and delegate facts setup: gather_subset: - 'all' - '!facter' - '!ohai' delegate_to: "{{ item }}" delegate_facts: True with_items: "{{ groups[mon_group_name] }}" run_once: true when: delegate_facts_host | bool - import_role: name: ceph-defaults - name: import_role ceph-facts import_role: name: ceph-facts tasks_from: container_binary.yml - name: set_fact ceph_cmd set_fact: ceph_cmd: "{{ container_binary + ' run --rm --net=host -v /etc/ceph:/etc/ceph:z -v /var/lib/ceph:/var/lib/ceph:z -v /var/run/ceph:/var/run/ceph:z --entrypoint=ceph ' + ceph_docker_registry + '/' + ceph_docker_image + ':' + ceph_docker_image_tag if containerized_deployment | bool else 'ceph' }}" - name: get ceph osd tree data command: "{{ ceph_cmd }} --cluster {{ cluster }} osd tree -f json" delegate_to: "{{ groups[mon_group_name][0] }}" register: osd_tree changed_when: false run_once: true - name: set_fact osd_ids set_fact: osd_ids: "{{ osd_ids | default([]) | union(item) }}" with_items: - "{{ ((osd_tree.stdout | default('{}') | trim | from_json).nodes | selectattr('name', 'match', '^' + inventory_hostname + '$') | map(attribute='children') | list) }}" - name: get osd metadata command: "{{ ceph_cmd }} --cluster {{ cluster }} osd metadata osd.{{ item }} -f json" register: osd_metadata delegate_to: "{{ groups[mon_group_name][0] }}" run_once: true changed_when: false with_items: "{{ osd_ids }}" - name: set_fact _osd_objectstore set_fact: _osd_objectstore: "{{ _osd_objectstore | default([]) | union([(item.stdout | from_json).osd_objectstore]) }}" with_items: "{{ osd_metadata.results }}" - name: set_fact skip_this_node set_fact: skip_this_node: "{{ ('filestore' in _osd_objectstore and 'bluestore' in _osd_objectstore and not force_filestore_to_bluestore | default(False)) or ('filestore' not in _osd_objectstore) }}" - name: filestore to bluestore migration workflow when: not skip_this_node | bool block: - name: get ceph-volume lvm inventory data ceph_volume: cluster: "{{ cluster }}" action: inventory environment: CEPH_CONTAINER_IMAGE: "{{ ceph_docker_registry + '/' + ceph_docker_image + ':' + ceph_docker_image_tag if containerized_deployment | bool else None }}" CEPH_CONTAINER_BINARY: "{{ container_binary }}" register: ceph_volume_inventory - name: set_fact inventory set_fact: inventory: "{{ ceph_volume_inventory.stdout | from_json }}" - name: set_fact ceph_disk_osds set_fact: ceph_disk_osds_devices: "{{ ceph_disk_osds_devices | default([]) + [item.path] }}" with_items: "{{ inventory }}" when: - not item.available | bool - "'Used by ceph-disk' in item.rejected_reasons" - name: ceph-disk prepared OSDs related tasks when: ceph_disk_osds_devices | default([]) | length > 0 block: - name: get partlabel command: blkid "{{ item + 'p' if item is match('/dev/(cciss/c[0-9]d[0-9]|nvme[0-9]n[0-9]){1,2}$') else item }}"1 -s PARTLABEL -o value register: partlabel with_items: "{{ ceph_disk_osds_devices | default([]) }}" - name: get simple scan data ceph_volume_simple_scan: path: "{{ item.item + 'p1' if item.item is match('/dev/(cciss/c[0-9]d[0-9]|nvme[0-9]n[0-9]){1,2}$') else item.item + '1' }}" cluster: "{{ cluster }" stdout: true environment: CEPH_CONTAINER_IMAGE: "{{ ceph_docker_registry + '/' + ceph_docker_image + ':' + ceph_docker_image_tag if containerized_deployment | bool else None }}" CEPH_CONTAINER_BINARY: "{{ container_binary }}" register: simple_scan with_items: "{{ partlabel.results | default([]) }}" when: item.stdout == 'ceph data' ignore_errors: true - name: mark out osds ceph_osd: ids: "{{ (item.0.stdout | from_json).whoami }}" cluster: "{{ cluster }}" state: out environment: CEPH_CONTAINER_IMAGE: "{{ ceph_docker_registry + '/' + ceph_docker_image + ':' + ceph_docker_image_tag if containerized_deployment | bool else None }}" CEPH_CONTAINER_BINARY: "{{ container_binary }}" with_together: - "{{ simple_scan.results }}" - "{{ partlabel.results }}" delegate_to: "{{ groups[mon_group_name][0] }}" run_once: true when: item.1.stdout == 'ceph data' - name: stop and disable old osd services service: name: "ceph-osd@{{ (item.0.stdout | from_json).whoami }}" state: stopped enabled: no with_together: - "{{ simple_scan.results }}" - "{{ partlabel.results }}" when: item.1.stdout == 'ceph data' - name: umount osd data mount: path: "/var/lib/ceph/osd/{{ cluster }}-{{ (item.0.stdout | from_json).whoami }}" state: unmounted with_together: - "{{ simple_scan.results }}" - "{{ partlabel.results }}" when: item.1.stdout == 'ceph data' - name: umount osd lockbox mount: path: "/var/lib/ceph/osd-lockbox/{{ (item.0.stdout | from_json).data.uuid }}" state: unmounted with_together: - "{{ simple_scan.results }}" - "{{ partlabel.results }}" when: - item.1.stdout == 'ceph data' - (item.0.stdout | from_json).encrypted | default(False) | bool - name: ensure dmcrypt for data device is closed command: cryptsetup close "{{ (item.0.stdout | from_json).data.uuid }}" with_together: - "{{ simple_scan.results }}" - "{{ partlabel.results }}" failed_when: false changed_when: false when: - item.1.stdout == 'ceph data' - (item.0.stdout | from_json).encrypted | default(False) | bool - name: ensure dmcrypt for journal device is closed command: cryptsetup close "{{ (item.0.stdout | from_json).journal.uuid }}" with_together: - "{{ simple_scan.results }}" - "{{ partlabel.results }}" failed_when: false changed_when: false when: - item.1.stdout == 'ceph data' - (item.0.stdout | from_json).encrypted | default(False) | bool - name: zap data devices ceph_volume: cluster: "{{ cluster }}" action: zap destroy: true data: "{{ (item.0.stdout | from_json).data.path }}" environment: CEPH_CONTAINER_IMAGE: "{{ ceph_docker_registry + '/' + ceph_docker_image + ':' + ceph_docker_image_tag if containerized_deployment | bool else None }}" CEPH_CONTAINER_BINARY: "{{ container_binary }}" with_together: - "{{ simple_scan.results }}" - "{{ partlabel.results }}" when: item.1.stdout == 'ceph data' - name: zap journal devices ceph_volume: cluster: "{{ cluster }}" action: zap destroy: true journal: "{{ (item.0.stdout | from_json).journal.path }}" environment: CEPH_CONTAINER_IMAGE: "{{ ceph_docker_registry + '/' + ceph_docker_image + ':' + ceph_docker_image_tag if containerized_deployment | bool else None }}" CEPH_CONTAINER_BINARY: "{{ container_binary }}" with_together: - "{{ simple_scan.results }}" - "{{ partlabel.results }}" when: - item.1.stdout == 'ceph data' - (item.0.stdout | from_json).journal.path is defined - name: get ceph-volume lvm list data ceph_volume: cluster: "{{ cluster }}" action: list environment: CEPH_CONTAINER_IMAGE: "{{ ceph_docker_registry + '/' + ceph_docker_image + ':' + ceph_docker_image_tag if containerized_deployment | bool else None }}" CEPH_CONTAINER_BINARY: "{{ container_binary }}" register: ceph_volume_lvm_list - name: set_fact _lvm_list set_fact: _lvm_list: "{{ _lvm_list | default([]) + item.value }}" with_dict: "{{ (ceph_volume_lvm_list.stdout | default('{}') | from_json) }}" - name: ceph-volume prepared OSDs related tasks block: - name: mark out osds ceph_osd: ids: "{{ (ceph_volume_lvm_list.stdout | default('{}') | from_json).keys() | list }}" cluster: "{{ cluster }}" state: out environment: CEPH_CONTAINER_IMAGE: "{{ ceph_docker_registry + '/' + ceph_docker_image + ':' + ceph_docker_image_tag if containerized_deployment | bool else None }}" CEPH_CONTAINER_BINARY: "{{ container_binary }}" delegate_to: "{{ groups[mon_group_name][0] }}" run_once: true - name: stop and disable old osd services service: name: "ceph-osd@{{ item }}" state: stopped enabled: no with_items: "{{ (ceph_volume_lvm_list.stdout | default('{}') | from_json).keys() | list }}" - name: stop and disable ceph-volume services service: name: "ceph-volume@lvm-{{ item.tags['ceph.osd_id'] }}-{{ item.tags['ceph.osd_fsid'] }}" state: stopped enabled: no with_items: "{{ _lvm_list }}" when: - not containerized_deployment | bool - item.type == 'data' - name: mark down osds ceph_osd: ids: "{{ (ceph_volume_lvm_list.stdout | default('{}') | from_json).keys() | list }}" cluster: "{{ cluster }}" state: down environment: CEPH_CONTAINER_IMAGE: "{{ ceph_docker_registry + '/' + ceph_docker_image + ':' + ceph_docker_image_tag if containerized_deployment | bool else None }}" CEPH_CONTAINER_BINARY: "{{ container_binary }}" delegate_to: "{{ groups[mon_group_name][0] }}" run_once: true - name: ensure all dmcrypt for data and journal are closed command: cryptsetup close "{{ item['lv_uuid'] }}" with_items: "{{ _lvm_list }}" changed_when: false failed_when: false when: item['tags'].get('ceph.encrypted', 0) | int == 1 - name: set_fact osd_fsid_list set_fact: osd_fsid_list: "{{ osd_fsid_list | default([]) + [{'osd_fsid': item.tags['ceph.osd_fsid'], 'destroy': (item.lv_name.startswith('osd-data-') and item.vg_name.startswith('ceph-')) | ternary(true, false), 'device': item.devices[0], 'journal': item['tags']['ceph.journal_device'] }] }}" with_items: "{{ _lvm_list }}" when: item.type == 'data' - name: zap ceph-volume prepared OSDs ceph_volume: action: "zap" osd_fsid: "{{ item.osd_fsid }}" destroy: false environment: CEPH_VOLUME_DEBUG: "{{ ceph_volume_debug }}" CEPH_CONTAINER_IMAGE: "{{ ceph_docker_registry + '/' + ceph_docker_image + ':' + ceph_docker_image_tag if containerized_deployment | bool else None }}" CEPH_CONTAINER_BINARY: "{{ container_binary }}" loop: "{{ osd_fsid_list }}" when: osd_fsid_list is defined - name: zap destroy ceph-volume prepared devices ceph_volume: action: "zap" data: "{{ item.device }}" destroy: true environment: CEPH_VOLUME_DEBUG: "{{ ceph_volume_debug }}" CEPH_CONTAINER_IMAGE: "{{ ceph_docker_registry + '/' + ceph_docker_image + ':' + ceph_docker_image_tag if containerized_deployment | bool else None }}" CEPH_CONTAINER_BINARY: "{{ container_binary }}" loop: "{{ osd_fsid_list }}" when: - osd_fsid_list is defined - item.destroy | bool - name: test if the journal device hasn't been already destroyed because of collocation stat: path: "{{ item.journal }}" loop: "{{ osd_fsid_list }}" register: journal_path when: - osd_fsid_list is defined - item.destroy | bool - item.journal is defined - item.journal not in (lvm_volumes | selectattr('journal', 'defined') | map(attribute='journal') | list) - name: zap destroy ceph-volume prepared journal devices ceph_volume: action: "zap" data: "{{ item.0.journal }}" destroy: true environment: CEPH_VOLUME_DEBUG: "{{ ceph_volume_debug }}" CEPH_CONTAINER_IMAGE: "{{ ceph_docker_registry + '/' + ceph_docker_image + ':' + ceph_docker_image_tag if containerized_deployment else None }}" CEPH_CONTAINER_BINARY: "{{ container_binary }}" loop: "{{ osd_fsid_list | zip(journal_path.results) | list }}" when: - osd_fsid_list is defined - item.0.destroy | bool - item.0.journal is defined - item.0.journal not in (lvm_volumes | selectattr('journal', 'defined') | map(attribute='journal') | list) - item.1.stat.exists | bool - name: ensure all dm are closed command: dmsetup remove "{{ item['lv_path'] }}" with_items: "{{ _lvm_list }}" changed_when: false failed_when: false when: - item['lv_path'] is defined # Do not close mappers for non 'lvm batch' devices - devices | default([]) | length > 0 - name: ensure all pv are removed command: "pvremove --yes {{ item.devices[0] }}" with_items: "{{ _lvm_list }}" failed_when: false when: - item.type == 'data' - item.lv_name.startswith('osd-data-') | bool - item.vg_name.startswith('ceph-') | bool when: _lvm_list is defined - name: set_fact osd_ids set_fact: osd_ids: "{{ osd_ids | default([]) + [item] }}" with_items: - "{{ ((osd_tree.stdout | default('{}') | from_json).nodes | selectattr('name', 'match', inventory_hostname) | map(attribute='children') | list) }}" - name: purge osd(s) from the cluster ceph_osd: ids: "{{ item }}" cluster: "{{ cluster }}" state: purge environment: CEPH_CONTAINER_IMAGE: "{{ ceph_docker_registry + '/' + ceph_docker_image + ':' + ceph_docker_image_tag if containerized_deployment | bool else None }}" CEPH_CONTAINER_BINARY: "{{ container_binary }}" run_once: true delegate_to: "{{ groups[mon_group_name][0] }}" with_items: "{{ osd_ids }}" - name: purge /var/lib/ceph/osd directories file: path: "/var/lib/ceph/osd/{{ cluster }}-{{ item }}" state: absent with_items: "{{ osd_ids }}" - name: force osd_objectstore to bluestore set_fact: osd_objectstore: bluestore - name: refresh ansible devices fact setup: gather_subset: - 'all' - '!facter' - '!ohai' filter: ansible_devices when: osd_auto_discovery | bool - import_role: name: ceph-defaults - import_role: name: ceph-facts - name: remove gpt header command: parted -s "{{ item }}" mklabel msdos with_items: "{{ (devices + dedicated_devices | default([]) + ceph_disk_osds_devices | default([])) | unique }}" - name: update lvm_volumes configuration for bluestore when: - lvm_volumes | length > 0 - not osd_auto_discovery | bool block: - name: reuse filestore journal partition for bluestore db set_fact: config_part: "{{ config_part | default([]) + [item | combine({'db': item.journal})] }}" with_items: "{{ lvm_volumes | selectattr('journal_vg', 'undefined') | list }}" - name: reuse filestore journal vg/lv for bluestore db set_fact: config_vglv: "{{ config_vglv | default([]) + [item | combine({'db': item.journal, 'db_vg': item.journal_vg})] }}" with_items: "{{ lvm_volumes | selectattr('journal_vg', 'defined') | list }}" - name: override lvm_volumes with bluestore configuration set_fact: lvm_volumes: "{{ config_part | default([]) + config_vglv | default([]) }}" - import_role: name: ceph-handler - import_role: name: ceph-container-common when: containerized_deployment | bool - import_role: name: ceph-config - import_role: name: ceph-osd - name: final play hosts: "{{ osd_group_name }}" become: true gather_facts: false tasks: - import_role: name: ceph-defaults - name: report any skipped node during this playbook debug: msg: | "WARNING:" "This node has been skipped because OSDs are either" "all bluestore or there's a mix of filestore and bluestore OSDs" when: - skip_this_node | bool