Ansible playbooks to deploy Ceph, the distributed filesystem.
 
 
 
 
Go to file
Leseb 3b68622f9d Merge pull request #81 from leseb/enh-testing-mode
Enable single host mode
2014-05-01 01:27:20 +02:00
fetch Add Ceph Playbook 2014-03-03 19:08:51 +01:00
group_vars Enable single host mode 2014-05-01 01:24:20 +02:00
roles Enable single host mode 2014-05-01 01:24:20 +02:00
.ansible.cfg Ability to reduce verbosity 2014-03-12 10:46:41 +01:00
.gitignore Ignore keyrings, they should never be commited to git 2014-03-11 10:29:25 +00:00
LICENSE Add Ceph Playbook 2014-03-03 19:08:51 +01:00
README.md Update README 2014-03-23 23:20:09 +01:00
Vagrantfile Reduce VM memory consumption 2014-03-14 10:47:06 +01:00
hosts Add maintenance playbook 2014-04-09 17:51:59 +02:00
maintenance.yml Add maintenance playbook 2014-04-09 17:51:59 +02:00
purge.yml Fix #73 by using `ansible_fqdn` 2014-04-18 10:55:31 +02:00
rc Improve vagrant testing and ansible setup 2014-03-11 14:14:37 +01:00
rolling_update.yml Add Ceph Playbook 2014-03-03 19:08:51 +01:00
site.yml Add RADOS Gateway support 2014-03-13 23:43:57 +01:00

README.md

ceph-ansible

Ansible playbook for Ceph!

What does it do?

General support for:

  • Monitors
  • OSDs
  • MDSs
  • RGW

More details:

  • Authentication (cephx), this can be disabled.
  • Supports cluster public and private network.
  • Monitors deployment. You can easily start with one monitor and then progressively add new nodes. So can deploy one monitor for testing purpose. For production, I recommend to a
  • Object Storage Daemons. Like the monitors you can start with a certain amount of nodes and then grow this number. The playbook either supports a dedicated device for storing th
  • Metadata daemons.
  • Collocation. The playbook supports collocating Monitors, OSDs and MDSs on the same machine.
  • The playbook was validated on Debian Wheezy, Ubuntu 12.04 LTS and CentOS 6.4.
  • Tested on Ceph Dumpling and Emperor.
  • A rolling upgrade playbook was written, an upgrade from Dumpling to Emperor was performed and worked.

Setup with Vagrant

First source the rc file:

$ source rc

Edit your /etc/hosts file with:

# Ansible hosts
127.0.0.1   ceph-mon0
127.0.0.1   ceph-mon1
127.0.0.1   ceph-mon2
127.0.0.1   ceph-osd0
127.0.0.1   ceph-osd1
127.0.0.1   ceph-osd2
127.0.0.1   ceph-rgw

Now since we use Vagrant and port forwarding, don't forget to collect the SSH local port of your VMs. Then edit your hosts file accordingly.

Ok let's get serious now. Run your virtual machines:

$ vagrant up
...
...
...

Test if Ansible can access the virtual machines:

$ ansible all -m ping
ceph-mon0 | success >> {
    "changed": false,
    "ping": "pong"
}

ceph-mon1 | success >> {
    "changed": false,
    "ping": "pong"
}

ceph-osd0 | success >> {
    "changed": false,
    "ping": "pong"
}

ceph-osd2 | success >> {
    "changed": false,
    "ping": "pong"
}

ceph-mon2 | success >> {
    "changed": false,
    "ping": "pong"
}

ceph-osd1 | success >> {
    "changed": false,
    "ping": "pong"
}

ceph-rgw | success >> {
    "changed": false,
    "ping": "pong"
}

DON'T FORGET TO GENERATE A FSID FOR THE CLUSTER AND A KEY FOR THE MONITOR

For this go to group_vars/all and group_vars/mons and append the fsid and key.

These are ONLY examples, DON'T USE THEM IN PRODUCTION:

  • fsid: 4a158d27-f750-41d5-9e7f-26ce4c9d2d45
  • monitor: AQAWqilTCDh7CBAAawXt6kyTgLFCxSvJhTEmuw==

Ready to deploy? Let's go!

$ ansible-playbook -f 7 -v site.yml
...
...
 ____________
< PLAY RECAP >
 ------------
        \   ^__^
         \  (oo)\_______
            (__)\       )\/\
                ||----w |
                ||     ||


ceph-mon0                  : ok=13   changed=10   unreachable=0    failed=0
ceph-mon1                  : ok=13   changed=9    unreachable=0    failed=0
ceph-mon2                  : ok=13   changed=9    unreachable=0    failed=0
ceph-osd0                  : ok=19   changed=12   unreachable=0    failed=0
ceph-osd1                  : ok=19   changed=12   unreachable=0    failed=0
ceph-osd2                  : ok=19   changed=12   unreachable=0    failed=0
ceph-rgw                   : ok=23   changed=16   unreachable=0    failed=0

Check the status:

$ vagrant ssh mon0 -c "sudo ceph -s"
    cluster 4a158d27-f750-41d5-9e7f-26ce4c9d2d45
     health HEALTH_OK
     monmap e3: 3 mons at {ceph-mon0=192.168.0.10:6789/0,ceph-mon1=192.168.0.11:6789/0,ceph-mon2=192.168.0.12:6789/0}, election epoch 6, quorum 0,1,2 ceph-mon0,ceph-mon1,ceph-mon
     mdsmap e6: 1/1/1 up {0=ceph-osd0=up:active}, 2 up:standby
     osdmap e10: 6 osds: 6 up, 6 in
      pgmap v17: 192 pgs, 3 pools, 9470 bytes data, 21 objects
            205 MB used, 29728 MB / 29933 MB avail
                 192 active+clean