--- - name: include create_ceph_initial_dirs.yml include_tasks: create_ceph_initial_dirs.yml when: containerized_deployment | bool - name: include_tasks rgw_systemd_environment_file.yml include_tasks: rgw_systemd_environment_file.yml when: inventory_hostname in groups.get(rgw_group_name, []) - name: config file operations related to OSDs when: - inventory_hostname in groups.get(osd_group_name, []) # the rolling_update.yml playbook sets num_osds to the number of currently # running osds - not rolling_update | bool block: - name: count number of osds for lvm scenario set_fact: num_osds: "{{ lvm_volumes | length | int }}" when: lvm_volumes | default([]) | length > 0 - block: - name: look up for ceph-volume rejected devices ceph_volume: cluster: "{{ cluster }}" action: "inventory" register: rejected_devices environment: CEPH_VOLUME_DEBUG: "{{ ceph_volume_debug }}" CEPH_CONTAINER_IMAGE: "{{ ceph_docker_registry + '/' + ceph_docker_image + ':' + ceph_docker_image_tag if containerized_deployment | bool else None }}" CEPH_CONTAINER_BINARY: "{{ container_binary }}" PYTHONIOENCODING: utf-8 - name: set_fact rejected_devices set_fact: _rejected_devices: "{{ _rejected_devices | default([]) + [item.path] }}" with_items: "{{ rejected_devices.stdout | default('{}') | from_json }}" when: "'Used by ceph-disk' in item.rejected_reasons" - name: set_fact _devices set_fact: _devices: "{{ devices | difference(_rejected_devices | default([])) }}" - name: run 'ceph-volume lvm batch --report' to see how many osds are to be created ceph_volume: cluster: "{{ cluster }}" objectstore: "{{ osd_objectstore }}" batch_devices: "{{ _devices }}" osds_per_device: "{{ osds_per_device | default(1) | int }}" journal_size: "{{ journal_size }}" block_db_size: "{{ block_db_size }}" report: true action: "batch" register: lvm_batch_report environment: CEPH_VOLUME_DEBUG: "{{ ceph_volume_debug }}" CEPH_CONTAINER_IMAGE: "{{ ceph_docker_registry + '/' + ceph_docker_image + ':' + ceph_docker_image_tag if containerized_deployment | bool else None }}" CEPH_CONTAINER_BINARY: "{{ container_binary }}" PYTHONIOENCODING: utf-8 when: _devices | default([]) | length > 0 when: - devices | default([]) | length > 0 - name: set_fact num_osds from the output of 'ceph-volume lvm batch --report' set_fact: num_osds: "{{ ((lvm_batch_report.stdout | default('{}') | from_json).osds | default([]) | length | int) + (_rejected_devices | default([]) | length | int) }}" when: - devices | default([]) | length > 0 - (lvm_batch_report.stdout | default('{}') | from_json).changed | default(true) | bool - name: run 'ceph-volume lvm list' to see how many osds have already been created ceph_volume: action: "list" register: lvm_list environment: CEPH_VOLUME_DEBUG: "{{ ceph_volume_debug }}" CEPH_CONTAINER_IMAGE: "{{ ceph_docker_registry + '/' + ceph_docker_image + ':' + ceph_docker_image_tag if containerized_deployment | bool else None }}" CEPH_CONTAINER_BINARY: "{{ container_binary }}" PYTHONIOENCODING: utf-8 changed_when: false when: - devices | default([]) | length > 0 - not (lvm_batch_report.stdout | default('{}') | from_json).changed | default(false) | bool - name: set_fact num_osds from the output of 'ceph-volume lvm list' set_fact: num_osds: "{{ lvm_list.stdout | default('{}') | from_json | length | int }}" when: - devices | default([]) | length > 0 - not (lvm_batch_report.stdout | default('{}') | from_json).changed | default(false) | bool - name: create ceph conf directory file: path: "/etc/ceph" state: directory owner: "ceph" group: "ceph" mode: "{{ ceph_directories_mode }}" when: not containerized_deployment | bool - name: "generate {{ cluster }}.conf configuration file" action: config_template args: src: "ceph.conf.j2" dest: "{{ ceph_conf_key_directory }}/{{ cluster }}.conf" owner: "{{ ceph_uid if containerized_deployment | bool else 'ceph' }}" group: "{{ ceph_uid if containerized_deployment | bool else 'ceph' }}" mode: "0644" config_overrides: "{{ ceph_conf_overrides }}" config_type: ini notify: - restart ceph mons - restart ceph osds - restart ceph mdss - restart ceph rgws - restart ceph mgrs - restart ceph rbdmirrors - restart ceph rbd-target-api-gw