ceph-ansible/roles/ceph-osd
Dimitri Savineau 62d98971f2 ceph-volume: Set max open files limit on container
The ceph-volume lvm list command takes ages to complete when having
a lot of LV devices on containerized deployment.
For instance, with 25 OSDs on a node it takes 3 mins 44s to list the
OSD.
Adding the max open files limit to the container engine cli when
executing the ceph-volume command seems to improve a lot thee
execution time ~30s.

This was impacting the OSDs creation with ceph-volume (both filestore
and bluestore) when using multiple LV devices.

Closes: https://bugzilla.redhat.com/show_bug.cgi?id=1702285

Signed-off-by: Dimitri Savineau <dsavinea@redhat.com>
(cherry picked from commit b987534881)
2019-06-20 20:00:53 -04:00
..
defaults ceph-osd: Increase cpu limit to 4 2019-04-30 12:11:42 -04:00
meta meta: set the right minimum ansible version required for galaxy 2018-12-11 09:59:25 +01:00
tasks ceph-volume: Set max open files limit on container 2019-06-20 20:00:53 -04:00
templates podman: Add systemd dependency on network.target 2019-06-07 16:06:26 +02:00
vars osd: backward compatibility with old disk_list.sh location 2019-03-18 17:25:51 +00:00
LICENSE Add READMEs for each roles 2015-07-25 10:51:53 +02:00
README.md Cleanup readme files in roles directories 2017-10-17 11:22:06 +02:00

README.md

Ansible role: ceph-osd

Documentation is available at http://docs.ceph.com/ceph-ansible/.