--- # # This playbook does a cephadm adopt for all the Ceph services # - name: confirm whether user really meant to adopt the cluster by cephadm hosts: localhost connection: local become: false gather_facts: false vars_prompt: - name: ireallymeanit prompt: Are you sure you want to adopt the cluster by cephadm ? default: 'no' private: no tasks: - name: exit playbook, if user did not mean to adopt the cluster by cephadm fail: msg: > Exiting cephadm-adopt playbook, cluster was NOT adopted. To adopt the cluster, either say 'yes' on the prompt or use `-e ireallymeanit=yes` on the command line when invoking the playbook when: ireallymeanit != 'yes' - name: gather facts and prepare system for cephadm hosts: - "{{ mon_group_name|default('mons') }}" - "{{ osd_group_name|default('osds') }}" - "{{ mds_group_name|default('mdss') }}" - "{{ rgw_group_name|default('rgws') }}" - "{{ mgr_group_name|default('mgrs') }}" - "{{ rbdmirror_group_name|default('rbdmirrors') }}" - "{{ nfs_group_name|default('nfss') }}" - "{{ iscsi_gw_group_name|default('iscsigws') }}" - "{{ monitoring_group_name|default('monitoring') }}" become: true gather_facts: false vars: delegate_facts_host: true tasks: - import_role: name: ceph-defaults - name: gather facts setup: gather_subset: - 'all' - '!facter' - '!ohai' when: not delegate_facts_host | bool or inventory_hostname in groups.get(client_group_name, []) - name: gather and delegate facts setup: gather_subset: - 'all' - '!facter' - '!ohai' delegate_to: "{{ item }}" delegate_facts: true with_items: "{{ groups['all'] | difference(groups.get('clients', [])) }}" run_once: true when: delegate_facts_host | bool - name: fail if one osd node is using filestore fail: msg: > filestore OSDs are not supported with cephadm. Please convert them with the filestore-to-bluestore.yml playbook first. when: - osd_group_name in group_names - osd_objectstore == 'filestore' - import_role: name: ceph-facts tasks_from: container_binary.yml - name: get the ceph version command: "{{ container_binary + ' run --rm --entrypoint=ceph ' + ceph_docker_registry + '/' + ceph_docker_image + ':' + ceph_docker_image_tag if containerized_deployment | bool else 'ceph' }} --version" changed_when: false register: ceph_version_out - name: set_fact ceph_version set_fact: ceph_version: "{{ ceph_version_out.stdout.split(' ')[2] }}" - name: fail on pre octopus ceph releases fail: msg: > Your Ceph version {{ ceph_version }} is not supported for this operation. Please upgrade your cluster with the rolling_update.yml playbook first. when: ceph_version is version('15.2', '<') - name: check if it is atomic host stat: path: /run/ostree-booted register: stat_ostree - name: set_fact is_atomic set_fact: is_atomic: "{{ stat_ostree.stat.exists }}" - import_role: name: ceph-container-engine when: not containerized_deployment | bool - import_role: name: ceph-container-common tasks_from: registry.yml when: - not containerized_deployment | bool - ceph_docker_registry_auth | bool - name: "pulling {{ ceph_docker_registry }}/{{ ceph_docker_image }}:{{ ceph_docker_image_tag }} image" command: "{{ timeout_command }} {{ container_binary }} pull {{ ceph_docker_registry }}/{{ ceph_docker_image }}:{{ ceph_docker_image_tag }}" changed_when: false register: docker_image until: docker_image.rc == 0 retries: "{{ docker_pull_retry }}" delay: 10 when: - not containerized_deployment | bool - inventory_hostname in groups.get(mon_group_name, []) or inventory_hostname in groups.get(osd_group_name, []) or inventory_hostname in groups.get(mds_group_name, []) or inventory_hostname in groups.get(rgw_group_name, []) or inventory_hostname in groups.get(mgr_group_name, []) or inventory_hostname in groups.get(rbdmirror_group_name, []) or inventory_hostname in groups.get(iscsi_gw_group_name, []) or inventory_hostname in groups.get(nfs_group_name, []) - name: install cephadm requirements package: name: ['python3', 'lvm2'] register: result until: result is succeeded - name: install cephadm package: name: cephadm register: result until: result is succeeded when: not containerized_deployment | bool - name: install cephadm mgr module package: name: ceph-mgr-cephadm register: result until: result is succeeded when: - not containerized_deployment | bool - mgr_group_name in group_names - name: get cephadm from the container image when: containerized_deployment | bool block: - name: create a cephadm container command: "{{ container_binary }} create --name cephadm {{ ceph_docker_registry }}/{{ ceph_docker_image }}:{{ ceph_docker_image_tag }}" changed_when: false - name: cp the cephadm cli file command: "{{ container_binary }} cp cephadm:/usr/sbin/cephadm /usr/sbin/cephadm" args: creates: /usr/sbin/cephadm - name: remove the cephadm container command: "{{ container_binary }} rm cephadm" changed_when: false - name: set_fact container_exec_cmd set_fact: container_exec_cmd: "{{ container_binary }} exec ceph-mon-{{ hostvars[groups[mon_group_name][0]]['ansible_hostname'] }}" when: containerized_deployment | bool - name: get current fsid command: "{{ container_exec_cmd | default('') }} ceph --admin-daemon /var/run/ceph/{{ cluster }}-mon.{{ hostvars[groups[mon_group_name][0]]['ansible_hostname'] }}.asok config get fsid --format json" register: current_fsid run_once: true changed_when: false delegate_to: "{{ groups[mon_group_name][0] }}" - name: set_fact fsid set_fact: fsid: "{{ (current_fsid.stdout | from_json).fsid }}" - name: enable cephadm mgr module ceph_mgr_module: name: cephadm cluster: "{{ cluster }}" state: enable environment: CEPH_CONTAINER_IMAGE: "{{ ceph_docker_registry + '/' + ceph_docker_image + ':' + ceph_docker_image_tag if containerized_deployment | bool else None }}" CEPH_CONTAINER_BINARY: "{{ container_binary }}" run_once: true delegate_to: '{{ groups[mon_group_name][0] }}' - name: set cephadm as orchestrator backend command: "{{ container_exec_cmd | default('') }} ceph --cluster {{ cluster }} orch set backend cephadm" changed_when: false run_once: true delegate_to: '{{ groups[mon_group_name][0] }}' - name: generate cephadm ssh key command: "{{ container_exec_cmd | default('') }} ceph --cluster {{ cluster }} cephadm generate-key" changed_when: false run_once: true delegate_to: '{{ groups[mon_group_name][0] }}' - name: get the cephadm ssh pub key command: "{{ container_exec_cmd | default('') }} ceph --cluster {{ cluster }} cephadm get-pub-key" changed_when: false run_once: true register: cephadm_pubpkey delegate_to: '{{ groups[mon_group_name][0] }}' - name: allow cephadm key for root account authorized_key: user: root key: '{{ cephadm_pubpkey.stdout }}' - name: run cephadm prepare-host command: cephadm prepare-host changed_when: false environment: CEPHADM_IMAGE: '{{ ceph_docker_registry }}/{{ ceph_docker_image }}:{{ ceph_docker_image_tag }}' - name: set default container image in ceph configuration command: "{{ container_exec_cmd | default('') }} ceph --cluster {{ cluster }} config set global container_image {{ ceph_docker_registry }}/{{ ceph_docker_image }}:{{ ceph_docker_image_tag }}" changed_when: false run_once: true delegate_to: '{{ groups[mon_group_name][0] }}' - name: set container image base in ceph configuration command: "{{ container_exec_cmd | default('') }} ceph --cluster {{ cluster }} config set mgr mgr/cephadm/container_image_base {{ ceph_docker_registry }}/{{ ceph_docker_image }}" changed_when: false run_once: true delegate_to: '{{ groups[mon_group_name][0] }}' - name: set dashboard container image in ceph mgr configuration when: dashboard_enabled | bool run_once: true block: - name: set alertmanager container image in ceph configuration command: "{{ container_exec_cmd | default('') }} ceph --cluster {{ cluster }} config set mgr mgr/cephadm/container_image_alertmanager {{ alertmanager_container_image }}" changed_when: false delegate_to: '{{ groups[mon_group_name][0] }}' - name: set grafana container image in ceph configuration command: "{{ container_exec_cmd | default('') }} ceph --cluster {{ cluster }} config set mgr mgr/cephadm/container_image_grafana {{ grafana_container_image }}" changed_when: false delegate_to: '{{ groups[mon_group_name][0] }}' - name: set node-exporter container image in ceph configuration command: "{{ container_exec_cmd | default('') }} ceph --cluster {{ cluster }} config set mgr mgr/cephadm/container_image_node_exporter {{ node_exporter_container_image }}" changed_when: false delegate_to: '{{ groups[mon_group_name][0] }}' - name: set prometheus container image in ceph configuration command: "{{ container_exec_cmd | default('') }} ceph --cluster {{ cluster }} config set mgr mgr/cephadm/container_image_prometheus {{ prometheus_container_image }}" changed_when: false delegate_to: '{{ groups[mon_group_name][0] }}' - name: manage nodes with cephadm command: "{{ container_exec_cmd | default('') }} ceph --cluster {{ cluster }} orch host add {{ hostvars[item]['ansible_hostname'] }} {{ hostvars[item]['ansible_default_ipv4']['address'] }} {{ hostvars[item]['group_names'] | join(' ') }}" changed_when: false run_once: true loop: '{{ ansible_play_hosts_all }}' delegate_to: '{{ groups[mon_group_name][0] }}' - name: add ceph label for core component command: "{{ container_exec_cmd | default('') }} ceph --cluster {{ cluster }} orch host label add {{ hostvars[item]['ansible_hostname'] }} ceph" changed_when: false run_once: true loop: '{{ ansible_play_hosts_all }}' delegate_to: '{{ groups[mon_group_name][0] }}' when: item in groups.get(mon_group_name, []) or item in groups.get(osd_group_name, []) or item in groups.get(mds_group_name, []) or item in groups.get(rgw_group_name, []) or item in groups.get(mgr_group_name, []) or item in groups.get(rbdmirror_group_name, []) - name: set_fact ceph_cmd set_fact: ceph_cmd: "{{ container_binary + ' run --rm --net=host -v /etc/ceph:/etc/ceph:z -v /var/lib/ceph:/var/lib/ceph:z -v /var/run/ceph:/var/run/ceph:z --entrypoint=ceph ' + ceph_docker_registry + '/' + ceph_docker_image + ':' + ceph_docker_image_tag if containerized_deployment | bool else 'ceph' }}" - name: get the client.admin keyring ceph_key: name: client.admin cluster: "{{ cluster }}" output_format: plain state: info environment: CEPH_CONTAINER_IMAGE: "{{ ceph_docker_registry + '/' + ceph_docker_image + ':' + ceph_docker_image_tag if containerized_deployment | bool else None }}" CEPH_CONTAINER_BINARY: "{{ container_binary }}" run_once: true delegate_to: '{{ groups[mon_group_name][0] }}' register: client_admin_keyring - name: copy the client.admin keyring copy: dest: "/etc/ceph/{{ cluster }}.client.admin.keyring" content: "{{ client_admin_keyring.stdout + '\n' }}" owner: "{{ ceph_uid | int if containerized_deployment | bool else 'ceph' }}" group: "{{ ceph_uid | int if containerized_deployment | bool else 'ceph' }}" mode: "{{ ceph_keyring_permissions }}" run_once: true delegate_to: "{{ item }}" with_items: - "{{ groups.get(osd_group_name, []) }}" - "{{ groups.get(mds_group_name, []) }}" - "{{ groups.get(rgw_group_name, []) }}" - "{{ groups.get(mgr_group_name, []) }}" - "{{ groups.get(rbdmirror_group_name, []) }}" - name: assimilate ceph configuration command: "{{ ceph_cmd }} --cluster {{ cluster }} config assimilate-conf -i /etc/ceph/{{ cluster }}.conf" changed_when: false when: inventory_hostname in groups.get(mon_group_name, []) or inventory_hostname in groups.get(osd_group_name, []) or inventory_hostname in groups.get(mds_group_name, []) or inventory_hostname in groups.get(rgw_group_name, []) or inventory_hostname in groups.get(mgr_group_name, []) or inventory_hostname in groups.get(rbdmirror_group_name, []) - name: set_fact cephadm_cmd set_fact: cephadm_cmd: "cephadm {{ '--docker' if container_binary == 'docker' else '' }}" - name: adopt ceph mon daemons hosts: "{{ mon_group_name|default('mons') }}" serial: 1 become: true gather_facts: false tasks: - import_role: name: ceph-defaults - name: adopt mon daemon cephadm_adopt: name: "mon.{{ ansible_hostname }}" cluster: "{{ cluster }}" image: "{{ ceph_docker_registry }}/{{ ceph_docker_image }}:{{ ceph_docker_image_tag }}" docker: "{{ true if container_binary == 'docker' else false }}" pull: false firewalld: "{{ true if configure_firewall | bool else false }}" - name: reset failed ceph-mon systemd unit command: 'systemctl reset-failed ceph-mon@{{ ansible_hostname }}' # noqa 303 changed_when: false failed_when: false when: containerized_deployment | bool - name: remove ceph-mon systemd unit file file: path: /etc/systemd/system/ceph-mon@.service state: absent when: containerized_deployment | bool - name: remove ceph-mon systemd override directory file: path: /etc/systemd/system/ceph-mon@.service.d state: absent when: not containerized_deployment | bool - name: waiting for the monitor to join the quorum... command: "{{ cephadm_cmd }} shell --fsid {{ fsid }} -- ceph --cluster {{ cluster }} quorum_status --format json" changed_when: false register: ceph_health_raw until: > ansible_hostname in (ceph_health_raw.stdout | from_json)["quorum_names"] retries: "{{ health_mon_check_retries }}" delay: "{{ health_mon_check_delay }}" environment: CEPHADM_IMAGE: '{{ ceph_docker_registry }}/{{ ceph_docker_image }}:{{ ceph_docker_image_tag }}' - name: adopt ceph mgr daemons hosts: "{{ mgr_group_name|default('mgrs') }}" serial: 1 become: true gather_facts: false tasks: - import_role: name: ceph-defaults - name: adopt mgr daemon cephadm_adopt: name: "mgr.{{ ansible_hostname }}" cluster: "{{ cluster }}" image: "{{ ceph_docker_registry }}/{{ ceph_docker_image }}:{{ ceph_docker_image_tag }}" docker: "{{ true if container_binary == 'docker' else false }}" pull: false firewalld: "{{ true if configure_firewall | bool else false }}" - name: reset failed ceph-mgr systemd unit command: 'systemctl reset-failed ceph-mgr@{{ ansible_hostname }}' # noqa 303 changed_when: false failed_when: false when: containerized_deployment | bool - name: remove ceph-mgr systemd unit file file: path: /etc/systemd/system/ceph-mgr@.service state: absent when: containerized_deployment | bool - name: remove ceph-mgr systemd override directory file: path: /etc/systemd/system/ceph-mgr@.service.d state: absent when: not containerized_deployment | bool - name: set osd flags hosts: "{{ mon_group_name|default('mons') }}[0]" become: true gather_facts: false tasks: - import_role: name: ceph-defaults - name: set osd flags command: "{{ cephadm_cmd }} shell --fsid {{ fsid }} -- ceph --cluster {{ cluster }} osd set {{ item }}" changed_when: false with_items: - noout - nodeep-scrub environment: CEPHADM_IMAGE: '{{ ceph_docker_registry }}/{{ ceph_docker_image }}:{{ ceph_docker_image_tag }}' - name: adopt ceph osd daemons hosts: "{{ osd_group_name|default('osd') }}" serial: 1 become: true gather_facts: false tasks: - import_role: name: ceph-defaults - import_role: name: ceph-facts tasks_from: container_binary.yml when: containerized_deployment | bool - name: get osd list ceph_volume: cluster: "{{ cluster }}" action: list environment: CEPH_CONTAINER_IMAGE: "{{ ceph_docker_registry + '/' + ceph_docker_image + ':' + ceph_docker_image_tag if containerized_deployment | bool else None }}" CEPH_CONTAINER_BINARY: "{{ container_binary }}" register: osd_list - name: set osd fsid for containerized deployment lineinfile: path: '/var/lib/ceph/osd/{{ cluster }}-{{ item.key }}/fsid' line: "{{ (item.value | selectattr('type', 'equalto', 'block') | map(attribute='tags') | first)['ceph.osd_fsid'] }}" owner: '{{ ceph_uid }}' group: '{{ ceph_uid }}' create: true with_dict: '{{ osd_list.stdout | from_json }}' when: containerized_deployment | bool - name: set osd type for containerized deployment lineinfile: path: '/var/lib/ceph/osd/{{ cluster }}-{{ item }}/type' line: 'bluestore' owner: '{{ ceph_uid }}' group: '{{ ceph_uid }}' create: true loop: '{{ (osd_list.stdout | from_json).keys() | list }}' when: containerized_deployment | bool - name: adopt osd daemon cephadm_adopt: name: "osd.{{ item }}" cluster: "{{ cluster }}" image: "{{ ceph_docker_registry }}/{{ ceph_docker_image }}:{{ ceph_docker_image_tag }}" docker: "{{ true if container_binary == 'docker' else false }}" pull: false firewalld: "{{ true if configure_firewall | bool else false }}" loop: '{{ (osd_list.stdout | from_json).keys() | list }}' - name: remove ceph-osd systemd unit and ceph-osd-run.sh files file: path: '{{ item }}' state: absent loop: - /etc/systemd/system/ceph-osd@.service - "{{ ceph_osd_docker_run_script_path | default('/usr/share') }}/ceph-osd-run.sh" when: containerized_deployment | bool - name: remove ceph-osd systemd override directory file: path: /etc/systemd/system/ceph-osd@.service.d state: absent when: not containerized_deployment | bool - name: waiting for clean pgs... command: "{{ cephadm_cmd }} shell --fsid {{ fsid }} -- ceph --cluster {{ cluster }} pg stat --format json" changed_when: false register: ceph_health_post until: > (((ceph_health_post.stdout | from_json).pg_summary.num_pg_by_state | length) > 0) and (((ceph_health_post.stdout | from_json).pg_summary.num_pg_by_state | selectattr('name', 'search', '^active\\+clean') | map(attribute='num') | list | sum) == (ceph_health_post.stdout | from_json).pg_summary.num_pgs) delegate_to: "{{ groups[mon_group_name][0] }}" retries: "{{ health_osd_check_retries }}" delay: "{{ health_osd_check_delay }}" environment: CEPHADM_IMAGE: '{{ ceph_docker_registry }}/{{ ceph_docker_image }}:{{ ceph_docker_image_tag }}' - name: unset osd flags hosts: "{{ mon_group_name|default('mons') }}[0]" become: true gather_facts: false tasks: - import_role: name: ceph-defaults - name: unset osd flags command: "{{ cephadm_cmd }} shell --fsid {{ fsid }} -- ceph --cluster {{ cluster }} osd unset {{ item }}" changed_when: false with_items: - noout - nodeep-scrub environment: CEPHADM_IMAGE: '{{ ceph_docker_registry }}/{{ ceph_docker_image }}:{{ ceph_docker_image_tag }}' - name: redeploy mds daemons hosts: "{{ mds_group_name|default('mdss') }}" become: true gather_facts: false tasks: - import_role: name: ceph-defaults - name: update the placement of metadata hosts command: "{{ cephadm_cmd }} shell --fsid {{ fsid }} -- ceph --cluster {{ cluster }} orch apply mds {{ cephfs }} --placement='{{ groups.get(mds_group_name, []) | length }} label:{{ mds_group_name }}'" run_once: true changed_when: false delegate_to: "{{ groups[mon_group_name][0] }}" environment: CEPHADM_IMAGE: '{{ ceph_docker_registry }}/{{ ceph_docker_image }}:{{ ceph_docker_image_tag }}' - name: stop and remove legacy ceph mds daemons hosts: "{{ mds_group_name|default('mdss') }}" serial: 1 become: true gather_facts: false tasks: - import_role: name: ceph-defaults - name: stop and disable ceph-mds systemd service service: name: 'ceph-mds@{{ ansible_hostname }}' state: stopped enabled: false - name: stop and disable ceph-mds systemd target service: name: ceph-mds.target state: stopped enabled: false when: not containerized_deployment | bool - name: reset failed ceph-mds systemd unit command: 'systemctl reset-failed ceph-mds@{{ ansible_hostname }}' # noqa 303 changed_when: false failed_when: false when: containerized_deployment | bool - name: remove ceph-mds systemd unit file file: path: /etc/systemd/system/ceph-mds@.service state: absent when: containerized_deployment | bool - name: remove ceph-mds systemd override directory file: path: /etc/systemd/system/ceph-mds@.service.d state: absent when: not containerized_deployment | bool - name: remove legacy ceph mds data file: path: '/var/lib/ceph/mds/{{ cluster }}-{{ ansible_hostname }}' state: absent - name: rgw realm/zonegroup/zone requirements hosts: "{{ rgw_group_name|default('rgws') }}" become: true gather_facts: false tasks: - import_role: name: ceph-defaults - name: for non multisite setup when: not rgw_multisite | bool run_once: true delegate_to: "{{ groups[mon_group_name][0] }}" block: - name: create a default realm command: "{{ cephadm_cmd }} shell --fsid {{ fsid }} -- radosgw-admin --cluster {{ cluster }} realm create --rgw-realm=default --default" run_once: true changed_when: false environment: CEPHADM_IMAGE: '{{ ceph_docker_registry }}/{{ ceph_docker_image }}:{{ ceph_docker_image_tag }}' - name: modify the default zonegroup command: "{{ cephadm_cmd }} shell --fsid {{ fsid }} -- radosgw-admin --cluster {{ cluster }} zonegroup modify --rgw-realm=default --rgw-zonegroup=default" run_once: true changed_when: false environment: CEPHADM_IMAGE: '{{ ceph_docker_registry }}/{{ ceph_docker_image }}:{{ ceph_docker_image_tag }}' - name: modify the default zone command: "{{ cephadm_cmd }} shell --fsid {{ fsid }} -- radosgw-admin --cluster {{ cluster }} zone modify --rgw-realm=default --rgw-zonegroup=default --rgw-zone=default" run_once: true changed_when: false environment: CEPHADM_IMAGE: '{{ ceph_docker_registry }}/{{ ceph_docker_image }}:{{ ceph_docker_image_tag }}' - name: commit the period command: "{{ cephadm_cmd }} shell --fsid {{ fsid }} -- radosgw-admin --cluster {{ cluster }} period update --commit" run_once: true changed_when: false environment: CEPHADM_IMAGE: '{{ ceph_docker_registry }}/{{ ceph_docker_image }}:{{ ceph_docker_image_tag }}' - name: update the placement of radosgw hosts command: "{{ cephadm_cmd }} shell --fsid {{ fsid }} -- ceph --cluster {{ cluster }} orch apply rgw {{ rgw_realm | default('default') }} {{ rgw_zone | default('default') }} --placement='{{ groups.get(rgw_group_name, []) | length }} label:{{ rgw_group_name }}' --port={{ radosgw_frontend_port }} {{ '--ssl' if radosgw_frontend_ssl_certificate else '' }}" run_once: true changed_when: false delegate_to: "{{ groups[mon_group_name][0] }}" environment: CEPHADM_IMAGE: '{{ ceph_docker_registry }}/{{ ceph_docker_image }}:{{ ceph_docker_image_tag }}' - name: redeploy rgw daemons hosts: "{{ rgw_group_name|default('rgws') }}" serial: 1 become: true gather_facts: false tasks: - import_role: name: ceph-defaults - import_role: name: ceph-facts tasks_from: set_radosgw_address.yml - name: stop and disable ceph-radosgw systemd service service: name: 'ceph-radosgw@rgw.{{ ansible_hostname }}.{{ item.instance_name }}' state: stopped enabled: false loop: '{{ rgw_instances }}' - name: stop and disable ceph-radosgw systemd target service: name: ceph-rgw.target state: stopped enabled: false when: not containerized_deployment | bool - name: reset failed ceph-radosgw systemd unit command: 'systemctl reset-failed ceph-radosgw@rgw.{{ ansible_hostname }}.{{ item.instance_name }}' # noqa 303 changed_when: false failed_when: false loop: '{{ rgw_instances }}' when: containerized_deployment | bool - name: remove ceph-radosgw systemd unit file file: path: /etc/systemd/system/ceph-radosgw@.service state: absent when: containerized_deployment | bool - name: remove ceph-radosgw systemd override directory file: path: /etc/systemd/system/ceph-radosgw@.service.d state: absent when: not containerized_deployment | bool - name: remove legacy ceph radosgw data file: path: '/var/lib/ceph/radosgw/{{ cluster }}-rgw.{{ ansible_hostname }}.{{ item.instance_name }}' state: absent loop: '{{ rgw_instances }}' - name: remove legacy ceph radosgw directory file: path: '/var/lib/ceph/radosgw/{{ cluster }}-rgw.{{ ansible_hostname }}' state: absent - name: redeploy rbd-mirror daemons hosts: "{{ rbdmirror_group_name|default('rbdmirrors') }}" become: true gather_facts: false tasks: - import_role: name: ceph-defaults - name: update the placement of rbd-mirror hosts command: "{{ cephadm_cmd }} shell --fsid {{ fsid }} -- ceph --cluster {{ cluster }} orch apply rbd-mirror --placement='{{ groups.get(rbdmirror_group_name, []) | length }} label:{{ rbdmirror_group_name }}'" run_once: true changed_when: false delegate_to: "{{ groups[mon_group_name][0] }}" environment: CEPHADM_IMAGE: '{{ ceph_docker_registry }}/{{ ceph_docker_image }}:{{ ceph_docker_image_tag }}' - name: stop and remove legacy rbd-mirror daemons hosts: "{{ rbdmirror_group_name|default('rbdmirrors') }}" serial: 1 become: true gather_facts: false tasks: - import_role: name: ceph-defaults - name: stop and disable rbd-mirror systemd service service: name: 'ceph-rbd-mirror@rbd-mirror.{{ ansible_hostname }}' state: stopped enabled: false - name: stop and disable rbd-mirror systemd target service: name: ceph-rbd-mirror.target state: stopped enabled: false when: not containerized_deployment | bool - name: reset failed rbd-mirror systemd unit command: 'systemctl reset-failed ceph-rbd-mirror@rbd-mirror.{{ ansible_hostname }}' # noqa 303 changed_when: false failed_when: false when: containerized_deployment | bool - name: remove rbd-mirror systemd unit file file: path: /etc/systemd/system/ceph-rbd-mirror@.service state: absent when: containerized_deployment | bool - name: remove rbd-mirror systemd override directory file: path: /etc/systemd/system/ceph-rbd-mirror@.service.d state: absent when: not containerized_deployment | bool - name: redeploy iscsigw daemons hosts: "{{ iscsi_gw_group_name|default('iscsigws') }}" become: true gather_facts: false tasks: - import_role: name: ceph-defaults - name: update the placement of iscsigw hosts command: "{{ cephadm_cmd }} shell --fsid {{ fsid }} -- ceph --cluster {{ cluster }} orch apply iscsi {{ iscsi_pool_name | default('rbd') }} {{ api_user | default('admin') }} {{ api_password | default('admin') }} {{ trusted_ip_list | default('192.168.122.1') }} --placement='{{ groups.get(iscsi_gw_group_name, []) | length }} label:{{ iscsi_gw_group_name }}'" run_once: true changed_when: false delegate_to: '{{ groups[mon_group_name][0] }}' environment: CEPHADM_IMAGE: '{{ ceph_docker_registry }}/{{ ceph_docker_image }}:{{ ceph_docker_image_tag }}' - name: stop and remove legacy iscsigw daemons hosts: "{{ iscsi_gw_group_name|default('iscsigws') }}" serial: 1 become: true gather_facts: false tasks: - import_role: name: ceph-defaults - name: stop and disable iscsigw systemd services service: name: '{{ item }}' state: stopped enabled: false with_items: - rbd-target-api - rbd-target-gw - tcmu-runner - name: reset failed iscsigw systemd units command: 'systemctl reset-failed {{ item }}' # noqa 303 changed_when: false failed_when: false with_items: - rbd-target-api - rbd-target-gw - tcmu-runner when: containerized_deployment | bool - name: remove iscsigw systemd unit files file: path: '/etc/systemd/system/{{ item }}.service' state: absent with_items: - rbd-target-api - rbd-target-gw - tcmu-runner when: containerized_deployment | bool - name: redeploy ceph-crash daemons hosts: - "{{ mon_group_name|default('mons') }}" - "{{ osd_group_name|default('osds') }}" - "{{ mds_group_name|default('mdss') }}" - "{{ rgw_group_name|default('rgws') }}" - "{{ mgr_group_name|default('mgrs') }}" - "{{ rbdmirror_group_name|default('rbdmirrors') }}" become: true gather_facts: false tasks: - import_role: name: ceph-defaults - name: stop and disable ceph-crash systemd service service: name: ceph-crash state: stopped enabled: false failed_when: false when: not containerized_deployment | bool - name: update the placement of ceph-crash hosts command: "{{ cephadm_cmd }} shell --fsid {{ fsid }} -- ceph --cluster {{ cluster }} orch apply crash --placement='label:ceph'" run_once: true changed_when: false delegate_to: '{{ groups[mon_group_name][0] }}' environment: CEPHADM_IMAGE: '{{ ceph_docker_registry }}/{{ ceph_docker_image }}:{{ ceph_docker_image_tag }}' - name: redeploy alertmanager/grafana/prometheus daemons hosts: "{{ monitoring_group_name|default('monitoring') }}" serial: 1 become: true gather_facts: false tasks: - import_role: name: ceph-defaults - name: with dashboard enabled when: dashboard_enabled | bool block: # (workaround) cephadm adopt alertmanager only stops prometheus-alertmanager systemd service - name: stop and disable alertmanager systemd unit service: name: alertmanager state: stopped enabled: false # (workaround) cephadm adopt alertmanager only uses /etc/prometheus/alertmanager.yml - name: create alertmanager config symlink file: path: /etc/prometheus/alertmanager.yml src: '{{ alertmanager_conf_dir }}/alertmanager.yml' state: link # (workaround) cephadm adopt alertmanager only uses /var/lib/prometheus/alertmanager/ - name: create alertmanager data symlink file: path: '{{ prometheus_data_dir }}/alertmanager' src: '{{ alertmanager_data_dir }}' state: link - name: adopt alertmanager daemon cephadm_adopt: name: "alertmanager.{{ ansible_hostname }}" cluster: "{{ cluster }}" image: "{{ alertmanager_container_image }}" docker: "{{ true if container_binary == 'docker' else false }}" pull: false firewalld: "{{ true if configure_firewall | bool else false }}" - name: remove alertmanager systemd unit file file: path: /etc/systemd/system/alertmanager.service state: absent - name: remove the legacy alertmanager data file: path: '{{ alertmanager_data_dir }}' state: absent - name: stop and disable prometheus systemd unit service: name: prometheus state: stopped enabled: false - name: remove alertmanager data symlink file: path: '{{ prometheus_data_dir }}/alertmanager' state: absent # (workaround) cephadm adopt prometheus only uses /var/lib/prometheus/metrics/ - name: tmp copy the prometheus data copy: src: '{{ prometheus_data_dir }}/' dest: /var/lib/prom_metrics owner: 65534 group: 65534 remote_src: true # (workaround) cephadm adopt prometheus only uses /var/lib/prometheus/metrics/ - name: restore the prometheus data copy: src: /var/lib/prom_metrics/ dest: /var/lib/prometheus/metrics owner: 65534 group: 65534 remote_src: true - name: remove the tmp prometheus data copy file: path: /var/lib/prom_metrics state: absent # (workaround) https://tracker.ceph.com/issues/45120 - name: create missing prometheus target directory file: path: '/var/lib/ceph/{{ fsid }}/prometheus.{{ ansible_hostname }}/etc/prometheus' state: directory owner: 65534 group: 65534 recurse: true - name: adopt prometheus daemon cephadm_adopt: name: "prometheus.{{ ansible_hostname }}" cluster: "{{ cluster }}" image: "{{ prometheus_container_image }}" docker: "{{ true if container_binary == 'docker' else false }}" pull: false firewalld: "{{ true if configure_firewall | bool else false }}" - name: remove prometheus systemd unit file file: path: /etc/systemd/system/prometheus.service state: absent - name: remove the legacy prometheus data file: path: '{{ prometheus_data_dir }}' state: absent # (workaround) cephadm adopt grafana only stops grafana systemd service - name: stop and disable grafana systemd unit service: name: grafana-server state: stopped enabled: false - name: adopt grafana daemon cephadm_adopt: name: "grafana.{{ ansible_hostname }}" cluster: "{{ cluster }}" image: "{{ grafana_container_image }}" docker: "{{ true if container_binary == 'docker' else false }}" pull: false firewalld: "{{ true if configure_firewall | bool else false }}" - name: remove grafana systemd unit file file: path: /etc/systemd/system/grafana-server.service state: absent - name: remove the legacy grafana data file: path: /var/lib/grafana state: absent - name: redeploy node-exporter daemons hosts: - "{{ mon_group_name|default('mons') }}" - "{{ osd_group_name|default('osds') }}" - "{{ mds_group_name|default('mdss') }}" - "{{ rgw_group_name|default('rgws') }}" - "{{ mgr_group_name|default('mgrs') }}" - "{{ rbdmirror_group_name|default('rbdmirrors') }}" - "{{ nfs_group_name|default('nfss') }}" - "{{ iscsi_gw_group_name|default('iscsigws') }}" - "{{ monitoring_group_name|default('monitoring') }}" become: true gather_facts: false tasks: - import_role: name: ceph-defaults - name: with dashboard enabled when: dashboard_enabled | bool block: - name: stop and disable node-exporter systemd service service: name: node_exporter state: stopped enabled: false - name: remove node_exporter systemd unit file file: path: /etc/systemd/system/node_exporter.service state: absent - name: update the placement of node-exporter hosts command: "{{ cephadm_cmd }} shell --fsid {{ fsid }} -- ceph --cluster {{ cluster }} orch apply node-exporter --placement='*'" run_once: true changed_when: false delegate_to: '{{ groups[mon_group_name][0] }}' environment: CEPHADM_IMAGE: '{{ ceph_docker_registry }}/{{ ceph_docker_image }}:{{ ceph_docker_image_tag }}' - name: adjust placement daemons hosts: "{{ mon_group_name|default('mons') }}[0]" become: true gather_facts: false tasks: - import_role: name: ceph-defaults - name: update the placement of monitor hosts command: "{{ cephadm_cmd }} shell --fsid {{ fsid }} -- ceph --cluster {{ cluster }} orch apply mon --placement='{{ groups.get(mon_group_name, []) | length }} label:{{ mon_group_name }}'" changed_when: false environment: CEPHADM_IMAGE: '{{ ceph_docker_registry }}/{{ ceph_docker_image }}:{{ ceph_docker_image_tag }}' - name: update the placement of manager hosts command: "{{ cephadm_cmd }} shell --fsid {{ fsid }} -- ceph --cluster {{ cluster }} orch apply mgr --placement='{{ groups.get(mgr_group_name, []) | length }} label:{{ mgr_group_name }}'" changed_when: false environment: CEPHADM_IMAGE: '{{ ceph_docker_registry }}/{{ ceph_docker_image }}:{{ ceph_docker_image_tag }}' - name: with dashboard enabled when: dashboard_enabled | bool block: - name: update the placement of alertmanager hosts command: "{{ cephadm_cmd }} shell --fsid {{ fsid }} -- ceph --cluster {{ cluster }} orch apply alertmanager --placement='{{ groups.get(monitoring_group_name, []) | length }} label:{{ monitoring_group_name }}'" changed_when: false environment: CEPHADM_IMAGE: '{{ ceph_docker_registry }}/{{ ceph_docker_image }}:{{ ceph_docker_image_tag }}' - name: update the placement of grafana hosts command: "{{ cephadm_cmd }} shell --fsid {{ fsid }} -- ceph --cluster {{ cluster }} orch apply grafana --placement='{{ groups.get(monitoring_group_name, []) | length }} label:{{ monitoring_group_name }}'" changed_when: false environment: CEPHADM_IMAGE: '{{ ceph_docker_registry }}/{{ ceph_docker_image }}:{{ ceph_docker_image_tag }}' - name: update the placement of prometheus hosts command: "{{ cephadm_cmd }} shell --fsid {{ fsid }} -- ceph --cluster {{ cluster }} orch apply prometheus --placement='{{ groups.get(monitoring_group_name, []) | length }} label:{{ monitoring_group_name }}'" changed_when: false environment: CEPHADM_IMAGE: '{{ ceph_docker_registry }}/{{ ceph_docker_image }}:{{ ceph_docker_image_tag }}' - name: show ceph orchestrator status hosts: "{{ mon_group_name|default('mons') }}[0]" become: true gather_facts: false tasks: - import_role: name: ceph-defaults - name: show ceph orchestrator services command: "{{ cephadm_cmd }} shell --fsid {{ fsid }} -- ceph --cluster {{ cluster }} orch ls --refresh" changed_when: false environment: CEPHADM_IMAGE: '{{ ceph_docker_registry }}/{{ ceph_docker_image }}:{{ ceph_docker_image_tag }}' - name: show ceph orchestrator daemons command: "{{ cephadm_cmd }} shell --fsid {{ fsid }} -- ceph --cluster {{ cluster }} orch ps --refresh" changed_when: false environment: CEPHADM_IMAGE: '{{ ceph_docker_registry }}/{{ ceph_docker_image }}:{{ ceph_docker_image_tag }}' - name: inform users about cephadm debug: msg: | This Ceph cluster is now managed by cephadm. Any new changes to the cluster need to be achieved by using the cephadm CLI and you don't need to use ceph-ansible playbooks anymore.