Ansible playbooks to deploy Ceph, the distributed filesystem.
 
 
 
 
Go to file
Matthew Vernon 04f4991648 restart_osd_daemon.sh.j2 - consider active+clean+* pgs as OK
After restarting each OSD, restart_osd_daemon.sh checks that the
cluster is in a good state before moving on to the next one. One of
the checks it does is that the number of pgs in the state
"active+clean" is equal to the total number of pgs in the cluster.

On large clusters (e.g. we have 173,696 pgs), it is likely that at
least one pg will be scrubbing and/or deep-scrubbing at any one
time. These pgs are in state "active+clean+scrubbing" or
"active+clean+scrubbing+deep", so the script was erroneously not
including them in the "good" count. Similar concerns apply to
"active+clean+snaptrim" and "active+clean+snaptrim_wait".

Fix this by considering as good any pg whose state contains
active+clean. Do this as an integer comparison to num_pgs in pgmap.

(could this be backported to at least stable-3.0 please?)

Closes: #2008
Signed-off-by: Matthew Vernon <mv3@sanger.ac.uk>
2018-09-24 10:33:46 +00:00
.github/ISSUE_TEMPLATE Update issue templates 2018-07-12 14:10:15 +02:00
contrib vagrant: move variable samples to contrib 2018-08-21 23:54:24 +02:00
docker docker: add README documenting directory purpose 2016-09-21 18:21:50 -04:00
docs doc: update lvm doc 2018-09-13 15:29:24 +00:00
group_vars osd: add osd memory target option 2018-09-18 10:12:46 +00:00
infrastructure-playbooks shrink-osd: follow up on 36fb3cde 2018-09-18 07:27:41 +00:00
library ceph_volume: adds the osds_per_device parameter 2018-09-12 20:27:14 +00:00
plugins ceph-volume: implement the 'lvm batch' subcommand 2018-08-09 09:41:58 -04:00
profiles Common: Add profiles 2017-07-19 11:50:03 +02:00
roles restart_osd_daemon.sh.j2 - consider active+clean+* pgs as OK 2018-09-24 10:33:46 +00:00
tests tests: fix monitor_address for shrink_osd scenario 2018-09-13 18:14:01 +02:00
.gitignore Revert "add .vscode/ to gitignore" 2018-04-27 13:20:04 +02:00
.mergify.yml ci: remove DCO 2018-07-02 16:08:18 +02:00
CONTRIBUTING.md doc: Update CONTRIBUTING.md 2018-06-22 14:48:23 +02:00
LICENSE Add Ceph Playbook 2014-03-03 19:08:51 +01:00
Makefile Makefile: followup on #2585 2018-06-01 12:50:03 +02:00
README-MULTISITE.md Experimental RGW Multisite Support 2016-08-16 14:57:01 -04:00
README.rst Update Documentation example link to 3.0 2018-02-07 16:34:45 +01:00
Vagrantfile vagrantfile: fix references to OpenStack settings 2018-09-21 07:00:03 +00:00
ansible.cfg tests: increase ssh timeout and retries in ansible.cfg 2018-06-07 18:00:52 +02:00
ceph-ansible.spec.in set the python-notario version to >= 0.0.13 in ceph-ansible.spec.in 2018-05-18 17:58:24 +02:00
dummy-ansible-hosts Fix Travis 2015-01-21 16:33:26 +01:00
example-ansible-role-requirements.yml add a galaxy dependency file example 2016-02-12 15:55:42 +01:00
generate_group_vars_sample.sh contrib: fix generate group_vars samples 2018-06-08 10:21:54 +02:00
requirements.txt main: update requirements.txt 2018-07-24 18:23:04 +02:00
rhcs_edits.txt Resync rhcs_edits.txt 2018-08-08 15:53:52 +02:00
rundep.sample common: remove hdparm 2018-07-07 08:53:47 +00:00
rundep_installer.sh Allow ceph-ansible to be run on a locally built/installed Ceph 2016-08-12 10:02:15 -04:00
site-docker.yml.sample site-docker: fix undefined variable error 2018-09-18 07:27:41 +00:00
site.yml.sample sites: fix conditonnal 2018-08-27 22:11:15 +02:00
test.yml remove rbd-mirror as well from testing. No packages are available yet 2016-04-12 07:38:13 -04:00
tox.ini tests: do not upgrade ceph release for switch_to_containers scenario 2018-09-09 14:25:14 +02:00
vagrant_variables.yml.sample remove kernel.pid_max 2018-02-23 13:57:57 +01:00

README.rst

ceph-ansible
============
Ansible playbooks for Ceph, the distributed filesystem.

Please refer to our hosted documentation here: http://docs.ceph.com/ceph-ansible/master/

You can view documentation for our ``stable-*`` branches by substituting ``master`` in the link
above for the name of the branch. For example: http://docs.ceph.com/ceph-ansible/stable-3.0/