ceph-ansible/roles/ceph-osd
Guillaume Abrioux 873fc8ec0f osd: ensure /var/lib/ceph/osd/{cluster}-{id} is present
This commit ensures that the `/var/lib/ceph/osd/{{ cluster }}-{{ osd_id }}` is
present before starting OSDs.

This is needed specificly when redeploying an OSD in case of OS upgrade
failure.
Since ceph data are still present on its devices then the node can be
redeployed, however those directories aren't present since they are
initially created by ceph-volume. We could recreate them manually but
for better user experience we can ask ceph-ansible to recreate them.

NOTE:
this only works for OSDs that were deployed with ceph-volume.
ceph-disk deployed OSDs would have to get those directories recreated
manually.

Closes: https://bugzilla.redhat.com/show_bug.cgi?id=1898486

Signed-off-by: Guillaume Abrioux <gabrioux@redhat.com>
2020-11-19 09:20:28 +01:00
..
defaults ceph-osd: remove ceph-osd-run.sh script 2020-06-18 17:51:13 +02:00
meta meta: set the right minimum ansible version required for galaxy 2018-12-11 09:59:25 +01:00
tasks osd: ensure /var/lib/ceph/osd/{cluster}-{id} is present 2020-11-19 09:20:28 +01:00
templates container: force rm --storage on ExecStartPre 2020-11-16 10:38:40 -05:00
vars osd: backward compatibility with old disk_list.sh location 2019-03-18 17:25:51 +00:00
LICENSE Add READMEs for each roles 2015-07-25 10:51:53 +02:00
README.md Cleanup readme files in roles directories 2017-10-17 11:22:06 +02:00

README.md

Ansible role: ceph-osd

Documentation is available at http://docs.ceph.com/ceph-ansible/.