ceph-ansible/roles/ceph-osd
Dimitri Savineau f4212b20e5 ceph-volume: Set max open files limit on container
The ceph-volume lvm list command takes ages to complete when having
a lot of LV devices on containerized deployment.
For instance, with 25 OSDs on a node it takes 3 mins 44s to list the
OSD.
Adding the max open files limit to the container engine cli when
executing the ceph-volume command seems to improve a lot thee
execution time ~30s.

This was impacting the OSDs creation with ceph-volume (both filestore
and bluestore) when using multiple LV devices.

Closes: https://bugzilla.redhat.com/show_bug.cgi?id=1702285

Signed-off-by: Dimitri Savineau <dsavinea@redhat.com>
(cherry picked from commit b987534881)
2019-06-20 20:01:13 -04:00
..
defaults update default rhcs values and docs 2019-06-04 14:18:23 +02:00
meta ceph-osd: Add support for openSUSE Leap distributions 2017-11-14 10:51:23 +00:00
tasks ceph-volume: Set max open files limit on container 2019-06-20 20:01:13 -04:00
templates ceph-osd: do not relabel /run/udev in containerized context 2019-06-19 23:46:46 +02:00
vars osd: backward compatibility with old disk_list.sh location 2019-03-18 21:56:53 +00:00
LICENSE Add READMEs for each roles 2015-07-25 10:51:53 +02:00
README.md Cleanup readme files in roles directories 2017-10-17 11:22:06 +02:00

README.md

Ansible role: ceph-osd

Documentation is available at http://docs.ceph.com/ceph-ansible/.