mirror of https://github.com/ceph/ceph-ansible.git
shrink-osd: admin key not needed for container shrink
Also do some clean Signed-off-by: Sébastien Han <seb@redhat.com>pull/2009/head
parent
779f642fa8
commit
2fb4981ca9
|
@ -91,12 +91,15 @@
|
||||||
with_items: "{{ osd_hosts }}"
|
with_items: "{{ osd_hosts }}"
|
||||||
delegate_to: "{{ item }}"
|
delegate_to: "{{ item }}"
|
||||||
failed_when: false
|
failed_when: false
|
||||||
|
when:
|
||||||
|
- not containerized_deployment
|
||||||
|
|
||||||
- name: fail when admin key is not present
|
- name: fail when admin key is not present
|
||||||
fail:
|
fail:
|
||||||
msg: "The Ceph admin key is not present on the OSD node, please add it and remove it after the playbook is done."
|
msg: "The Ceph admin key is not present on the OSD node, please add it and remove it after the playbook is done."
|
||||||
with_items: "{{ ceph_admin_key.results }}"
|
with_items: "{{ ceph_admin_key.results }}"
|
||||||
when:
|
when:
|
||||||
|
- not containerized_deployment
|
||||||
- item.stat.exists == false
|
- item.stat.exists == false
|
||||||
|
|
||||||
# NOTE(leseb): using '>' is the only way I could have the command working
|
# NOTE(leseb): using '>' is the only way I could have the command working
|
||||||
|
@ -127,8 +130,6 @@
|
||||||
|
|
||||||
- name: deactivating osd(s)
|
- name: deactivating osd(s)
|
||||||
command: ceph-disk deactivate --cluster {{ cluster }} --deactivate-by-id {{ item.0 }} --mark-out
|
command: ceph-disk deactivate --cluster {{ cluster }} --deactivate-by-id {{ item.0 }} --mark-out
|
||||||
register: deactivate
|
|
||||||
ignore_errors: yes
|
|
||||||
run_once: true
|
run_once: true
|
||||||
with_together:
|
with_together:
|
||||||
- "{{ osd_to_kill.split(',') }}"
|
- "{{ osd_to_kill.split(',') }}"
|
||||||
|
@ -137,21 +138,8 @@
|
||||||
when:
|
when:
|
||||||
- not containerized_deployment
|
- not containerized_deployment
|
||||||
|
|
||||||
- name: set osd(s) out when ceph-disk deactivating fail
|
|
||||||
command: "{{ docker_exec_cmd }} ceph --cluster {{ cluster }} osd out osd.{{ item.0 }}"
|
|
||||||
delegate_to: "{{ groups[mon_group_name][0] }}"
|
|
||||||
with_together:
|
|
||||||
- "{{ osd_to_kill.split(',') }}"
|
|
||||||
- "{{ deactivate.results }}"
|
|
||||||
when:
|
|
||||||
- not containerized_deployment
|
|
||||||
- not item.1.get("skipped")
|
|
||||||
- item.1.stderr|length > 0
|
|
||||||
|
|
||||||
- name: destroying osd(s)
|
- name: destroying osd(s)
|
||||||
command: ceph-disk destroy --cluster {{ cluster }} --destroy-by-id {{ item.0 }} --zap
|
command: ceph-disk destroy --cluster {{ cluster }} --destroy-by-id {{ item.0 }} --zap
|
||||||
register: destroy
|
|
||||||
ignore_errors: yes
|
|
||||||
run_once: true
|
run_once: true
|
||||||
with_together:
|
with_together:
|
||||||
- "{{ osd_to_kill.split(',') }}"
|
- "{{ osd_to_kill.split(',') }}"
|
||||||
|
@ -161,32 +149,20 @@
|
||||||
- not containerized_deployment
|
- not containerized_deployment
|
||||||
|
|
||||||
- name: remove osd(s) from crush_map when ceph-disk destroy fail
|
- name: remove osd(s) from crush_map when ceph-disk destroy fail
|
||||||
command: "{{ docker_exec_cmd }} ceph --cluster {{ cluster }} osd crush remove osd.{{ item.0 }}"
|
command: "{{ docker_exec_cmd }} ceph --cluster {{ cluster }} osd crush remove osd.{{ item }}"
|
||||||
run_once: true
|
run_once: true
|
||||||
delegate_to: "{{ groups[mon_group_name][0] }}"
|
delegate_to: "{{ groups[mon_group_name][0] }}"
|
||||||
with_together:
|
with_items: "{{ osd_to_kill.split(',') }}"
|
||||||
- "{{ osd_to_kill.split(',') }}"
|
|
||||||
- "{{ destroy.results }}"
|
|
||||||
when:
|
|
||||||
- (item.1.get("skipped") or item.1.stderr|length > 0)
|
|
||||||
|
|
||||||
- name: delete osd(s) auth key when ceph-disk destroy fail
|
- name: delete osd(s) auth key when ceph-disk destroy fail
|
||||||
command: "{{ docker_exec_cmd }} ceph --cluster {{ cluster }} auth del osd.{{ item.0 }}"
|
command: "{{ docker_exec_cmd }} ceph --cluster {{ cluster }} auth del osd.{{ item }}"
|
||||||
delegate_to: "{{ groups[mon_group_name][0] }}"
|
delegate_to: "{{ groups[mon_group_name][0] }}"
|
||||||
with_together:
|
with_items: "{{ osd_to_kill.split(',') }}"
|
||||||
- "{{ osd_to_kill.split(',') }}"
|
|
||||||
- "{{ destroy.results }}"
|
|
||||||
when:
|
|
||||||
- (item.1.get("skipped") or item.1.stderr|length > 0)
|
|
||||||
|
|
||||||
- name: deallocate osd(s) id when ceph-disk destroy fail
|
- name: deallocate osd(s) id when ceph-disk destroy fail
|
||||||
command: "{{ docker_exec_cmd }} ceph --cluster {{ cluster }} osd rm {{ item.0 }}"
|
command: "{{ docker_exec_cmd }} ceph --cluster {{ cluster }} osd rm {{ item }}"
|
||||||
delegate_to: "{{ groups[mon_group_name][0] }}"
|
delegate_to: "{{ groups[mon_group_name][0] }}"
|
||||||
with_together:
|
with_items: "{{ osd_to_kill.split(',') }}"
|
||||||
- "{{ osd_to_kill.split(',') }}"
|
|
||||||
- "{{ destroy.results }}"
|
|
||||||
when:
|
|
||||||
- (item.1.get("skipped") or item.1.stderr|length > 0)
|
|
||||||
|
|
||||||
- name: show ceph health
|
- name: show ceph health
|
||||||
command: "{{ docker_exec_cmd }} ceph --cluster {{ cluster }} -s"
|
command: "{{ docker_exec_cmd }} ceph --cluster {{ cluster }} -s"
|
||||||
|
|
Loading…
Reference in New Issue