The recently added Trusty box was build with a Controller named
"SATAController", we must reuse the same name while adding our devices.
Signed-off-by: Sébastien Han <sebastien.han@enovance.com>
Without this plugin if a Ceph version is present in a repo (let's say
epel) it will install the epel version and not the ICE version.
We install yum-plugin-priorities.noarch to honor the 'priority=1' flag.
Signed-off-by: Sébastien Han <sebastien.han@enovance.com>
In storage world it's often recommended to disable transparent hugepages
as they will tend to lower performance.
Note that this change won't survive reboot. There are several ways to
disable this permanently such as:
* rc.local
* grub boot line
It's a bit tricky to do this in Ansible since it really depends on the
OS you're running on.
Signed-off-by: Sébastien Han <sebastien.han@enovance.com>
Default behavior is to fail if a variable is not declared however this
can be disable in your ansible.cfg so we force this variable as
mandatory.
Signed-off-by: Sébastien Han <sebastien.han@enovance.com>
Still WIP, @mwheckmann free to test
As requested by #162
Current known issue, since ceph.conf gets modified during every single
run (at the end during the merge) so this will restart ceph daemons.
Signed-off-by: Sébastien Han <sebastien.han@enovance.com>
Depending on the OS you are runnning on you should be able to configure
these values.
Re-ordering file for clarity as well.
Signed-off-by: Sébastien Han <sebastien.han@enovance.com>
Updating each group_vars file to reflect the content of each role
default variables.
Closes: #169
Signed-off-by: Sébastien Han <sebastien.han@enovance.com>
Big cluster will easily reach the default limit so we need to increase
it and make it configurable.
Signed-off-by: Sébastien Han <sebastien.han@enovance.com>
Add a fetch/ceph_cluster_uuid.conf file so we keep the same UUID from
the Vagrantfile and from the ceph-common run.
Prior to that change the Vagrantfile was setting
4a158d27-f750-41d5-9e7f-26ce4c9d2d45 but while playing the ceph-common
role we check if fetch/ceph_cluster_uuid.conf exists, if not we generate
an UUID. Here we ended up with 2 UUIDs...
IMPORTANT NOTE FOR NON-VAGRANT DEPLOYMENT. IF YOU WANT TO USE YOUR OWN
UUID PLEASE REMOVE THAT FILE BEFORE RUNNING THE PLAYBOOK.
Signed-off-by: Sébastien Han <sebastien.han@enovance.com>
We remove all the partitions, label and re-create something clean prior
to prepare the design. This will help solving many issues with existing
disks or while scratching/deploy test environments often.
Signed-off-by: Sébastien Han <sebastien.han@enovance.com>