mirror of https://github.com/ceph/ceph-ansible.git
c7175c1400
Multi distro installation support |
||
---|---|---|
fetch | ||
group_vars | ||
roles | ||
LICENSE | ||
README.md | ||
Vagrantfile | ||
hosts | ||
rc | ||
rolling_update.yml | ||
site.yml |
README.md
ceph-ansible
Ansible playbook for Ceph!
What does it do?
- Authentication (cephx), this can be disabled.
- Supports cluster public and private network.
- Monitors deployment. You can easily start with one monitor and then progressively add new nodes. So can deploy one monitor for testing purpose. For production, I recommend to a
- Object Storage Daemons. Like the monitors you can start with a certain amount of nodes and then grow this number. The playbook either supports a dedicated device for storing th
- Metadata daemons.
- Collocation. The playbook supports collocating Monitors, OSDs and MDSs on the same machine.
- The playbook was validated on both Debian Wheezy and Ubuntu 12.04 LTS.
- Tested on Ceph Dumpling and Emperor.
- A rolling upgrade playbook was written, an upgrade from Dumpling to Emperor was performed and worked.
Setup with Vagrant
First modify the rc
file we your home directory:
export ANSIBLE_CONFIG=<whatever_path>/.ansible.cfg
Do the same for the .ansible.cfg
file:
[defaults]
host_key_checking = False
remote_user = vagrant
hostfile = <whatever_path>/hosts
log_path = <whatever_path>/ansible.log
ansible_managed = Ansible managed: modified on %Y-%m-%d %H:%M:%S by {uid}
private_key_file = ~/.vagrant.d/insecure_private_key
error_on_undefined_vars = False
Edit your /etc/hosts
file with:
# Ansible hosts
127.0.0.1 ceph-mon0
127.0.0.1 ceph-mon1
127.0.0.1 ceph-mon2
127.0.0.1 ceph-osd0
127.0.0.1 ceph-osd1
127.0.0.1 ceph-osd2
Now since we use Vagrant and port forwarding, don't forget to grab the SSH local port of your VMs.
Then edit your hosts
file accordingly.
Ok let's get serious now. Run your virtual machines:
$ vagrant up
...
...
...
Test if Ansible can access the virtual machines:
$ ansible all -m ping
ceph-mon0 | success >> {
"changed": false,
"ping": "pong"
}
ceph-mon1 | success >> {
"changed": false,
"ping": "pong"
}
ceph-osd0 | success >> {
"changed": false,
"ping": "pong"
}
ceph-osd2 | success >> {
"changed": false,
"ping": "pong"
}
ceph-mon2 | success >> {
"changed": false,
"ping": "pong"
}
ceph-osd1 | success >> {
"changed": false,
"ping": "pong"
}
Ready to deploy? Let's go!
$ ansible-playbook -f 6 -v site.yml
...
...
____________
< PLAY RECAP >
------------
\ ^__^
\ (oo)\_______
(__)\ )\/\
||----w |
|| ||
ceph-mon0 : ok=13 changed=10 unreachable=0 failed=0
ceph-mon1 : ok=13 changed=9 unreachable=0 failed=0
ceph-mon2 : ok=13 changed=9 unreachable=0 failed=0
ceph-osd0 : ok=19 changed=12 unreachable=0 failed=0
ceph-osd1 : ok=19 changed=12 unreachable=0 failed=0
ceph-osd2 : ok=19 changed=12 unreachable=0 failed=0
Check the status:
$ vagrant ssh mon0 -c "sudo ceph -s"
cluster 4a158d27-f750-41d5-9e7f-26ce4c9d2d45
health HEALTH_OK
monmap e3: 3 mons at {ceph-mon0=192.168.0.10:6789/0,ceph-mon1=192.168.0.11:6789/0,ceph-mon2=192.168.0.12:6789/0}, election epoch 6, quorum 0,1,2 ceph-mon0,ceph-mon1,ceph-mon
mdsmap e6: 1/1/1 up {0=ceph-osd0=up:active}, 2 up:standby
osdmap e10: 6 osds: 6 up, 6 in
pgmap v17: 192 pgs, 3 pools, 9470 bytes data, 21 objects
205 MB used, 29728 MB / 29933 MB avail
192 active+clean