Merged with Upstream Master

pull/924/head
James Saint-Rossy 2016-08-17 12:00:36 -04:00
commit 32f6ef7747
79 changed files with 1235 additions and 191 deletions

3
Vagrantfile vendored
View File

@ -18,6 +18,7 @@ CLIENTS = settings['client_vms']
SUBNET = settings['subnet']
BOX = settings['vagrant_box']
BOX_URL = settings['vagrant_box_url']
SYNC_DIR = settings['vagrant_sync_dir']
MEMORY = settings['memory']
STORAGECTL = settings['vagrant_storagectl']
ETH = settings['eth']
@ -112,7 +113,7 @@ Vagrant.configure(VAGRANTFILE_API_VERSION) do |config|
# Faster bootup. Disable if you need this for libvirt
config.vm.provider :libvirt do |v,override|
override.vm.synced_folder '.', '/home/vagrant/sync', disabled: true
override.vm.synced_folder '.', SYNC_DIR, disabled: true
end
if BOX == 'openstack'

View File

@ -84,12 +84,24 @@ dummy:
## Configure package origin
#
#ceph_origin: 'upstream' # or 'distro'
#ceph_origin: 'upstream' #'distro' or 'local'
# 'distro' means that no separate repo file will be added
# you will get whatever version of Ceph is included in your Linux distro.
#
#ceph_use_distro_backports: false # DEBIAN ONLY
# 'local' means that the ceph binaries will be copied over from the local machine
# LOCAL CEPH INSTALLATION (ceph_origin==local)
#
# Path to DESTDIR of the ceph install
#ceph_installation_dir: "/path/to/ceph_installation/"
# Whether or not to use installer script rundep_installer.sh
# This script takes in rundep and installs the packages line by line onto the machine
# If this is set to false then it is assumed that the machine ceph is being copied onto will already have
# all runtime dependencies installed
#use_installer: false
# Root directory for ceph-ansible
#ansible_dir: "/path/to/ceph-ansible"
#ceph_use_distro_backports: false # DEBIAN ONLY
# STABLE
########
@ -132,15 +144,15 @@ dummy:
# time. Please use ELRepo's latest upstream 4.x kernels if you want to run CephFS
# on RHEL 7.
#
#ceph_stable_rh_storage: false
#ceph_rhcs: false
# This will affect how/what repositories are enabled depending on the desired
# version. The next version will use "2" not "2.0" which would not work.
#ceph_stable_rh_storage_version: 1.3 # next version is 2
#ceph_stable_rh_storage_cdn_install: false # assumes all the nodes can connect to cdn.redhat.com
#ceph_stable_rh_storage_iso_install: false # usually used when nodes don't have access to cdn.redhat.com
#ceph_stable_rh_storage_iso_path:
#ceph_stable_rh_storage_mount_path: /tmp/rh-storage-mount
#ceph_stable_rh_storage_repository_path: /tmp/rh-storage-repo # where to copy iso's content
#ceph_rhcs_version: 1.3 # next version is 2
#ceph_rhcs_cdn_install: false # assumes all the nodes can connect to cdn.redhat.com
#ceph_rhcs_iso_install: false # usually used when nodes don't have access to cdn.redhat.com
#ceph_rhcs_iso_path:
#ceph_rhcs_mount_path: /tmp/rh-storage-mount
#ceph_rhcs_repository_path: /tmp/rh-storage-repo # where to copy iso's content
# UBUNTU CLOUD ARCHIVE
@ -352,6 +364,8 @@ dummy:
#raw_multi_journal: False
#osd_directory: False
#bluestore: False
#dmcrypt_journal_collocation: False
#dmcrypt_dedicated_journal: False
#osd_auto_discovery: False

View File

@ -35,6 +35,28 @@ dummy:
#ceph_nfs_pseudo_path: "/ceph"
#ceph_nfs_protocols: "3,4"
#ceph_nfs_access_type: "RW"
#ceph_nfs_log_file: "/var/log/ganesha.log"
####################
# FSAL Ceph Config #
####################
#ceph_nfs_ceph_export_id: 20134
#ceph_nfs_ceph_pseudo_path: "/ceph"
#ceph_nfs_ceph_protocols: "3,4"
#ceph_nfs_ceph_access_type: "RW"
###################
# FSAL RGW Config #
###################
#ceph_nfs_rgw_export_id: 20134
#ceph_nfs_rgw_pseudo_path: "/ceph"
#ceph_nfs_rgw_protocols: "3,4"
#ceph_nfs_rgw_access_type: "RW"
#ceph_nfs_rgw_user: "cephnfs"
# Note: keys are optional and can be generated, but not on containerized, where
# they must be configered.
#ceph_nfs_rgw_access_key: "QFAMEDSJP5DEKJO0DDXY"
#ceph_nfs_rgw_secret_key: "iaSFLDVvDdQt6lkNzHyW4fPLZugBAI1g17LO0+87[MAC[M#C"
###################

View File

@ -121,7 +121,7 @@ dummy:
# Keys are stored into the monitors k/v store
# Use 'true' to enable this scenario
# Both journal and data are stored on the same dm-crypt encrypted device
#dmcrypt_journal_colocation: false
#dmcrypt_journal_collocation: false
# VI. Encrypt osd data and/or journal devices with dm-crypt.

View File

@ -0,0 +1,5 @@
Infrastructure playbooks
========================
This directory contains a variety of playbooks that can be used independently of the Ceph roles we have.
They aim to perform infrastructure related tasks that would help use managing a Ceph cluster or performing certain operational tasks.

View File

@ -261,7 +261,7 @@
rbdmirror_group_name in group_names
- name: check for anything running ceph
shell: "ps awux | grep -v grep | grep -q -- ceph-"
shell: "ps awux | grep -- [c]eph-"
register: check_for_running_ceph
failed_when: check_for_running_ceph.rc == 0
@ -293,15 +293,18 @@
osd_group_name in group_names
- name: remove osd mountpoint tree
shell: rm -rf /var/lib/ceph/osd
file:
path: /var/lib/ceph/osd/
state: absent
register: remove_osd_mountpoints
failed_when: false
ignore_errors: true
when:
osd_group_name in group_names
- name: remove monitor store and bootstrap keys
shell: rm -rf /var/lib/ceph/
failed_when: false
file:
path: /var/lib/ceph/
state: absent
when:
mon_group_name in group_names
@ -313,7 +316,7 @@
- remove data
when:
osd_group_name in group_names and
remove_osd_mountpoints.rc != 0
remove_osd_mountpoints.failed is defined
- name: see if ceph-disk is installed
shell: "which ceph-disk"
@ -322,7 +325,7 @@
- name: zap osd disks
shell: ceph-disk zap "{{ item }}"
with_items: devices
with_items: "{{ devices | default([]) }}"
when:
osd_group_name in group_names and
ceph_disk_present.rc == 0 and
@ -423,6 +426,7 @@
- name: request data removal
local_action: shell echo requesting data removal
become: false
notify:
- remove data

View File

@ -0,0 +1,142 @@
---
# This playbook shrinks the Ceph monitors from your cluster
# It can remove any number of monitor(s) from the cluster and ALL THEIR DATA
#
# Use it like this:
# ansible-playbook shrink-mon.yml -e mon_host=ceph-mon01,ceph-mon02
# Prompts for confirmation to shrink, defaults to no and
# doesn't shrink the cluster. yes shrinks the cluster.
#
# ansible-playbook -e ireallymeanit=yes|no shrink-cluster.yml
# Overrides the prompt using -e option. Can be used in
# automation scripts to avoid interactive prompt.
- name: confirm whether user really meant to remove monitor(s) from the ceph cluster
hosts:
- localhost
gather_facts: false
become: true
vars_prompt:
- name: ireallymeanit
prompt: Are you sure you want to shrink the cluster?
default: 'no'
private: no
tasks:
- include_vars: roles/ceph-common/defaults/main.yml
- include_vars: group_vars/all
- name: exit playbook, if user did not mean to shrink cluster
fail:
msg: "Exiting shrink-mon playbook, no monitor(s) was/were removed.
To shrink the cluster, either say 'yes' on the prompt or
or use `-e ireallymeanit=yes` on the command line when
invoking the playbook"
when: ireallymeanit != 'yes'
- name: exit playbook, if no monitor(s) was/were given
fail:
msg: "mon_host must be declared
Exiting shrink-cluster playbook, no monitor(s) was/were removed.
On the command line when invoking the playbook, you can use
-e mon_host=ceph-mon01,ceph-mon02 argument."
when: mon_host is not defined
- name: test if ceph command exist
command: command -v ceph
changed_when: false
failed_when: false
register: ceph_command
- name: exit playbook, if ceph command does not exist
debug:
msg: "The ceph command is not available, please install it :("
run_once: true
when:
- ceph_command.rc != 0
- name: exit playbook, if cluster files do not exist
stat:
path: "{{ item }}"
register: ceph_conf_key
with_items:
- /etc/ceph/{{ cluster }}.conf
- /etc/ceph/{{ cluster }}.client.admin.keyring
failed_when: false
- fail:
msg: "Ceph's configuration file is not present in /etc/ceph"
with_items: "{{ceph_conf_key.results}}"
when:
- item.stat.exists == false
- name: exit playbook, if can not connect to the cluster
command: timeout 5 ceph --cluster {{ cluster }} health
register: ceph_health
until: ceph_health.stdout.find("HEALTH") > -1
retries: 5
delay: 2
- name: verify given monitors are reachable
command: ping -c 1 {{ item }}
with_items: "{{mon_host.split(',')}}"
register: mon_reachable
failed_when: false
- fail:
msg: "One or more monitors are not reachable, please check your /etc/hosts or your DNS"
with_items: "{{mon_reachable.results}}"
when:
- item.rc != 0
- name: stop monitor service (systemd)
service:
name: ceph-mon@{{ item }}
state: stopped
enabled: no
with_items: "{{mon_host.split(',')}}"
delegate_to: "{{item}}"
failed_when: false
- name: purge monitor store
file:
path: /var/lib/ceph/mon/{{ cluster }}-{{ item }}
state: absent
with_items: "{{mon_host.split(',')}}"
delegate_to: "{{item}}"
- name: remove monitor from the quorum
command: ceph --cluster {{ cluster }} mon remove {{ item }}
failed_when: false
with_items: "{{mon_host.split(',')}}"
# NOTE (leseb): sorry for the 'sleep' command
# but it will take a couple of seconds for other monitors
# to notice that one member has left.
# 'sleep 5' is not that bad and should be sufficient
- name: verify the monitor is out of the cluster
shell: "sleep 5 && ceph --cluster {{ cluster }} -s | grep monmap | sed 's/.*quorum//' | egrep -sq {{ item }}"
with_items: "{{mon_host.split(',')}}"
failed_when: false
register: ceph_health_mon
- name: please remove the monitor from your ceph configuration file
debug:
msg: "The monitor(s) has/have been successfully removed from the cluster.
Please remove the monitor(s) entry(ies) from the rest of your ceph configuration files, cluster wide."
run_once: true
with_items: "{{ceph_health_mon.results}}"
when:
- item.rc != 0
- name: please remove the monitor from your ceph configuration file
fail:
msg: "Monitor(s) appear(s) to still be part of the cluster, please check what happened."
run_once: true
with_items: "{{ceph_health_mon.results}}"
when:
- item.rc == 0

View File

@ -0,0 +1,131 @@
---
# This playbook shrinks Ceph OSDs.
# It can remove any number of OSD(s) from the cluster and ALL THEIR DATA
#
# Use it like this:
# ansible-playbook shrink-osd.yml -e osd_id=0,2,6
# Prompts for confirmation to shrink, defaults to no and
# doesn't shrink the cluster. yes shrinks the cluster.
#
# ansible-playbook -e ireallymeanit=yes|no shrink-osd.yml
# Overrides the prompt using -e option. Can be used in
# automation scripts to avoid interactive prompt.
- name: confirm whether user really meant to remove osd(s) from the cluster
hosts:
- localhost
gather_facts: false
become: true
vars_prompt:
- name: ireallymeanit
prompt: Are you sure you want to shrink the cluster?
default: 'no'
private: no
tasks:
- include_vars: roles/ceph-common/defaults/main.yml
- include_vars: group_vars/all
- name: exit playbook, if user did not mean to shrink cluster
fail:
msg: "Exiting shrink-osd playbook, no osd(s) was/were removed..
To shrink the cluster, either say 'yes' on the prompt or
or use `-e ireallymeanit=yes` on the command line when
invoking the playbook"
when: ireallymeanit != 'yes'
- name: exit playbook, if no osd(s) was/were given
fail:
msg: "osd_ids must be declared
Exiting shrink-osd playbook, no OSD()s was/were removed.
On the command line when invoking the playbook, you can use
-e osd_ids=0,1,2,3 argument."
when: osd_ids is not defined
- name: test if ceph command exist
command: command -v ceph
changed_when: false
failed_when: false
register: ceph_command
- name: exit playbook, if ceph command does not exist
debug:
msg: "The ceph command is not available, please install it :("
run_once: true
when:
- ceph_command.rc != 0
- name: exit playbook, if cluster files do not exist
stat:
path: "{{ item }}"
register: ceph_conf_key
with_items:
- /etc/ceph/{{ cluster }}.conf
- /etc/ceph/{{ cluster }}.client.admin.keyring
failed_when: false
- fail:
msg: "Ceph's configuration file is not present in /etc/ceph"
with_items: "{{ceph_conf_key.results}}"
when:
- item.stat.exists == false
- name: exit playbook, if can not connect to the cluster
command: timeout 5 ceph --cluster {{ cluster }} health
register: ceph_health
until: ceph_health.stdout.find("HEALTH") > -1
retries: 5
delay: 2
# NOTE (leseb): just in case, the complex filters mechanism below does not work anymore.
# This will be a quick and easy fix but will require using the shell module.
# - name: find the host where the osd(s) is/are running on
# shell: |
# ceph --cluster {{ cluster }} osd find {{ item }} | grep -Po '(?<="ip": ")[^:]*'
# with_items: "{{osd_ids.split(',')}}"
# register: osd_hosts
#
- name: find the host where the osd(s) is/are running on
command: ceph --cluster {{ cluster }} osd find {{ item }}
with_items: "{{osd_ids.split(',')}}"
register: osd_hosts
- set_fact: ip_item="{{(item.stdout | from_json).ip}}"
with_items: "{{osd_hosts.results}}"
register: ip_result
- set_fact: ips="{{ ip_result.results | map(attribute='ansible_facts.ip_item') | list }}"
- set_fact: real_ips="{{ ips | regex_replace(':[0-9][0-9][0-9][0-9]\/[0-9][0-9][0-9][0-9]', '') }}"
- name: check if ceph admin key exists on the osd nodes
stat:
path: "/etc/ceph/{{ cluster }}.client.admin.keyring"
register: ceph_admin_key
with_items: "{{real_ips}}"
delegate_to: "{{item}}"
failed_when: false
- fail:
msg: "The Ceph admin key is not present on the OSD node, please add it and remove it after the playbook is done."
with_items: "{{ceph_admin_key.results}}"
when:
- item.stat.exists == false
- name: deactivating osd(s)
command: ceph-disk deactivate --cluster {{ cluster }} --deactivate-by-id {{ item.0 }} --mark-out
with_together:
- "{{osd_ids.split(',')}}"
- "{{real_ips}}"
delegate_to: "{{item.1}}"
- name: destroying osd(s)
command: ceph-disk destroy --cluster {{ cluster }} --destroy-by-id {{ item.0 }} --zap
with_together:
- "{{osd_ids.split(',')}}"
- "{{real_ips}}"
delegate_to: "{{item.1}}"

View File

@ -25,7 +25,7 @@ Have a look at `defaults/main.yml`.
* Install source, choose one of these:
* `ceph_stable`
* `ceph_dev`
* `ceph_stable_rh_storage`
* `ceph_rhcs`
* `ceph_custom`
* `journal_size`
* `monitor_interface`

View File

@ -76,12 +76,24 @@ ceph_test: False
## Configure package origin
#
ceph_origin: 'upstream' # or 'distro'
ceph_origin: 'upstream' # or 'distro' or 'local'
# 'distro' means that no separate repo file will be added
# you will get whatever version of Ceph is included in your Linux distro.
#
ceph_use_distro_backports: false # DEBIAN ONLY
# 'local' means that the ceph binaries will be copied over from the local machine
# LOCAL CEPH INSTALLATION (ceph_origin==local)
#
# Path to DESTDIR of the ceph install
#ceph_installation_dir: "/path/to/ceph_installation/"
# Whether or not to use installer script rundep_installer.sh
# This script takes in rundep and installs the packages line by line onto the machine
# If this is set to false then it is assumed that the machine ceph is being copied onto will already have
# all runtime dependencies installed
#use_installer: false
# Root directory for ceph-ansible
#ansible_dir: "/path/to/ceph-ansible"
ceph_use_distro_backports: false # DEBIAN ONLY
# STABLE
########
@ -124,15 +136,15 @@ ceph_stable_redhat_distro: el7
# time. Please use ELRepo's latest upstream 4.x kernels if you want to run CephFS
# on RHEL 7.
#
ceph_stable_rh_storage: false
ceph_rhcs: false
# This will affect how/what repositories are enabled depending on the desired
# version. The next version will use "2" not "2.0" which would not work.
ceph_stable_rh_storage_version: 1.3 # next version is 2
ceph_stable_rh_storage_cdn_install: false # assumes all the nodes can connect to cdn.redhat.com
ceph_stable_rh_storage_iso_install: false # usually used when nodes don't have access to cdn.redhat.com
#ceph_stable_rh_storage_iso_path:
ceph_stable_rh_storage_mount_path: /tmp/rh-storage-mount
ceph_stable_rh_storage_repository_path: /tmp/rh-storage-repo # where to copy iso's content
ceph_rhcs_version: 1.3 # next version is 2
ceph_rhcs_cdn_install: false # assumes all the nodes can connect to cdn.redhat.com
ceph_rhcs_iso_install: false # usually used when nodes don't have access to cdn.redhat.com
#ceph_rhcs_iso_path:
ceph_rhcs_mount_path: /tmp/rh-storage-mount
ceph_rhcs_repository_path: /tmp/rh-storage-repo # where to copy iso's content
# UBUNTU CLOUD ARCHIVE
@ -344,5 +356,7 @@ journal_collocation: False
raw_multi_journal: False
osd_directory: False
bluestore: False
dmcrypt_journal_collocation: False
dmcrypt_dedicated_journal: False
osd_auto_discovery: False

View File

@ -5,6 +5,7 @@
when:
- ceph_origin != 'upstream'
- ceph_origin != 'distro'
- ceph_origin != 'local'
tags:
- package-install
@ -15,18 +16,18 @@
- ceph_origin == 'upstream'
- not ceph_stable
- not ceph_dev
- not ceph_stable_rh_storage
- not ceph_rhcs
- not ceph_stable_uca
tags:
- package-install
- name: verify that a method was chosen for red hat storage
fail:
msg: "choose between ceph_stable_rh_storage_cdn_install and ceph_stable_rh_storage_iso_install"
msg: "choose between ceph_rhcs_cdn_install and ceph_rhcs_iso_install"
when:
- ceph_stable_rh_storage
- not ceph_stable_rh_storage_cdn_install
- not ceph_stable_rh_storage_iso_install
- ceph_rhcs
- not ceph_rhcs_cdn_install
- not ceph_rhcs_iso_install
- ceph_origin == "upstream"
tags:
- package-install
@ -71,7 +72,7 @@
- not raw_multi_journal
- not osd_directory
- not bluestore
- not dmcrypt_journal_colocation
- not dmcrypt_journal_collocation
- not dmcrypt_dedicated_journal
- name: verify only one osd scenario was chosen
@ -86,14 +87,15 @@
or (raw_multi_journal and osd_directory)
or (raw_multi_journal and bluestore)
or (osd_directory and bluestore)
or (dmcrypt_journal_colocation and journal_collocation)
or (dmcrypt_journal_colocation and raw_multi_journal)
or (dmcrypt_journal_colocation and osd_directory)
or (dmcrypt_journal_colocation and bluestore)
or (dmcrypt_journal_collocation and journal_collocation)
or (dmcrypt_journal_collocation and raw_multi_journal)
or (dmcrypt_journal_collocation and osd_directory)
or (dmcrypt_journal_collocation and bluestore)
or (dmcrypt_dedicated_journal and journal_collocation)
or (dmcrypt_dedicated_journal and raw_multi_journal)
or (dmcrypt_dedicated_journal and osd_directory)
or (dmcrypt_dedicated_journal and bluestore)
or (dmcrypt_dedicated_journal and dmcrypt_journal_collocation)
- name: verify devices have been provided
fail:

View File

@ -0,0 +1,6 @@
---
- name: check ntp installation on atomic
command: rpm -q chrony
register: ntp_pkg_query
ignore_errors: true
changed_when: false

View File

@ -18,7 +18,7 @@
fail:
msg: "Distribution not supported {{ ansible_distribution_version }} by Red Hat Ceph Storage, only RHEL 7.1"
when:
- ceph_stable_rh_storage
- ceph_rhcs
- ansible_distribution_version | version_compare('7.1', '<')
- name: fail on unsupported distribution for ubuntu cloud archive

View File

@ -39,13 +39,67 @@
include: redhat_ceph_repository.yml
when: ceph_origin == 'upstream'
- name: make sure /tmp exists
file:
path: /tmp
state: directory
when:
- ceph_origin == 'local'
- use_installer
- name: use mktemp to create name for rundep
command: "mktemp /tmp/rundep.XXXXXXXX"
register: rundep_location
when:
- ceph_origin == 'local'
- use_installer
- name: copy rundep
copy:
src: "{{ansible_dir}}/rundep"
dest: "{{ item }}"
with_items: rundep_location.stdout_lines
when:
- ceph_origin == 'local'
- use_installer
- name: install ceph dependencies
script: "{{ ansible_dir }}/rundep_installer.sh {{ item }}"
become: true
with_items: rundep_location.stdout_lines
when:
- ceph_origin == 'local'
- use_installer
- name: install ceph
yum:
name: ceph
state: "{{ (upgrade_ceph_packages|bool) | ternary('latest','present') }}"
when: not use_server_package_split
when:
- not use_server_package_split
- ansible_pkg_mgr == "yum"
- ceph_origin != 'local'
- name: install distro or red hat storage ceph mon
- name: synchronize ceph install
synchronize:
src: "{{ceph_installation_dir}}/"
dest: "/"
when:
- ceph_origin == 'local'
- name: create user group ceph
group:
name: 'ceph'
when:
- ceph_origin == 'local'
- name: create user ceph
user:
name: 'ceph'
when:
- ceph_origin == 'local'
- name: install distro or red hat storage ceph mon via yum
yum:
name: "ceph-mon"
state: "{{ (upgrade_ceph_packages|bool) | ternary('latest','present') }}"
@ -57,7 +111,7 @@
or ceph_origin == "distro"
or ceph_custom
- name: install distro or red hat storage ceph mon
- name: install distro or red hat storage ceph mon via dnf
dnf:
name: "ceph-mon"
state: "{{ (upgrade_ceph_packages|bool) | ternary('latest','present') }}"
@ -69,7 +123,7 @@
or ceph_dev
or ceph_custom
- name: install distro or red hat storage ceph osd
- name: install distro or red hat storage ceph osd via yum
yum:
name: "ceph-osd"
state: "{{ (upgrade_ceph_packages|bool) | ternary('latest','present') }}"
@ -81,7 +135,7 @@
or ceph_dev
or ceph_custom
- name: install distro or red hat storage ceph osd
- name: install distro or red hat storage ceph osd via dnf
dnf:
name: "ceph-osd"
state: "{{ (upgrade_ceph_packages|bool) | ternary('latest','present') }}"
@ -93,7 +147,7 @@
or ceph_dev
or ceph_custom
- name: install distro or red hat storage ceph mds
- name: install distro or red hat storage ceph mds via yum
yum:
name: "ceph-mds"
state: "{{ (upgrade_ceph_packages|bool) | ternary('latest','present') }}"
@ -105,7 +159,7 @@
or ceph_dev
or ceph_custom
- name: install distro or red hat storage ceph mds
- name: install distro or red hat storage ceph mds via dnf
dnf:
name: "ceph-mds"
state: "{{ (upgrade_ceph_packages|bool) | ternary('latest','present') }}"
@ -117,7 +171,7 @@
or ceph_dev
or ceph_custom
- name: install distro or red hat storage ceph base
- name: install distro or red hat storage ceph base via yum
yum:
name: "ceph-base"
state: "{{ (upgrade_ceph_packages|bool) | ternary('latest','present') }}"
@ -129,7 +183,7 @@
or ceph_dev
or ceph_custom
- name: install distro or red hat storage ceph base
- name: install distro or red hat storage ceph base via dnf
dnf:
name: "ceph-base"
state: "{{ (upgrade_ceph_packages|bool) | ternary('latest','present') }}"
@ -173,18 +227,40 @@
- rgw_group_name in group_names
- ansible_pkg_mgr == "dnf"
- name: install NFS gateway
- name: install nfs ceph gateway
yum:
name: nfs-ganesha-ceph
state: "{{ (upgrade_ceph_packages|bool) | ternary('latest','present') }}"
when:
- nfs_group_name in group_names
- ansible_pkg_mgr == "yum"
- fsal_ceph
- name: install NFS gateway
- name: install nfs ceph gateway
dnf:
name: nfs-ganesha-ceph
state: "{{ (upgrade_ceph_packages|bool) | ternary('latest','present') }}"
when:
- nfs_group_name in group_names
- ansible_pkg_mgr == "dnf"
- fsal_ceph
- name: install nfs rgw gateway
yum:
name: "{{ item }}"
with_items:
- nfs-ganesha-rgw
- ceph-radosgw
when:
- nfs_group_name in group_names
- ansible_pkg_mgr == "yum"
- fsal_rgw
- name: install nfs rgw gateway
dnf:
name: "{{ item }}"
with_items:
- nfs-ganesha-rgw
- ceph-radosgw
when:
- nfs_group_name in group_names
- ansible_pkg_mgr == "dnf"
- fsal_rgw

View File

@ -5,15 +5,15 @@
# intelligent way to determine the location of the key.
- name: install the rh ceph storage repository key
apt_key:
file: "{{ ceph_stable_rh_storage_repository_path }}/MON/release.asc"
file: "{{ ceph_rhcs_repository_path }}/MON/release.asc"
state: present
when:
- ceph_stable_rh_storage
- ceph_stable_rh_storage_iso_install
- ceph_rhcs
- ceph_rhcs_iso_install
- name: add rh ceph storage repositories
apt_repository:
repo: "deb file://{{ ceph_stable_rh_storage_repository_path }}/{{ item }}/ {{ ansible_lsb.codename }} main"
repo: "deb file://{{ ceph_rhcs_repository_path }}/{{ item }}/ {{ ansible_lsb.codename }} main"
state: present
changed_when: false
with_items:
@ -22,14 +22,14 @@
- "Tools"
- "Agent"
when:
- ceph_stable_rh_storage
- ceph_stable_rh_storage_iso_install
- ceph_rhcs
- ceph_rhcs_iso_install
- name: add the red hat storage apt-key
apt_key:
data: "{{ lookup('file', role_path+'/files/cephstablerhcs.asc') }}"
state: present
when: not ceph_stable_rh_storage_iso_install
when: not ceph_rhcs_iso_install
- name: install dependencies
apt:

View File

@ -1,10 +1,10 @@
---
- name: install red hat storage repository key
rpm_key:
key: "{{ ceph_stable_rh_storage_repository_path }}/RPM-GPG-KEY-redhat-release"
key: "{{ ceph_rhcs_repository_path }}/RPM-GPG-KEY-redhat-release"
state: present
when:
- ceph_stable_rh_storage_iso_install
- ceph_rhcs_iso_install
- name: add red hat storage repository
template:
@ -14,7 +14,7 @@
group: root
mode: 0644
when:
- ceph_stable_rh_storage_iso_install
- ceph_rhcs_iso_install
- name: install dependencies
yum:

View File

@ -11,15 +11,15 @@
- include: ./pre_requisites/prerequisite_rh_storage_iso_install.yml
when:
- ceph_stable_rh_storage
- ceph_stable_rh_storage_iso_install
- ceph_rhcs
- ceph_rhcs_iso_install
tags:
- package-install
- include: ./pre_requisites/prerequisite_rh_storage_cdn_install.yml
when:
- ceph_stable_rh_storage
- ceph_stable_rh_storage_cdn_install
- ceph_rhcs
- ceph_rhcs_cdn_install
- ansible_os_family == "RedHat"
tags:
- package-install
@ -27,28 +27,28 @@
- include: ./installs/install_on_redhat.yml
when:
ansible_os_family == 'RedHat' and
not ceph_stable_rh_storage_iso_install
not ceph_rhcs_iso_install
tags:
- package-install
- include: ./installs/install_rh_storage_on_redhat.yml
when:
- ansible_distribution == "RedHat"
- ceph_stable_rh_storage
- ceph_rhcs
tags:
- package-install
- include: ./installs/install_on_debian.yml
when:
- ansible_os_family == 'Debian'
- not ceph_stable_rh_storage
- not ceph_rhcs
tags:
- package-install
- include: ./installs/install_rh_storage_on_debian.yml
when:
- ansible_os_family == 'Debian'
- ceph_stable_rh_storage
- ceph_rhcs
tags:
- package-install

View File

@ -0,0 +1,11 @@
---
- include: ../checks/check_ntp_atomic.yml
when: is_atomic
- name: start the ntp service
service:
name: chronyd
enabled: yes
state: started
when:
- ntp_pkg_query.rc == 0

View File

@ -5,42 +5,42 @@
changed_when: false
- name: check if the red hat storage monitor repo is already present
shell: yum --noplugins --cacheonly repolist | grep -sq rhel-7-server-rhceph-{{ ceph_stable_rh_storage_version }}-mon-rpms
shell: yum --noplugins --cacheonly repolist | grep -sq rhel-7-server-rhceph-{{ ceph_rhcs_version }}-mon-rpms
changed_when: false
failed_when: false
register: rh_storage_mon_repo
when: mon_group_name in group_names
- name: enable red hat storage monitor repository
command: subscription-manager repos --enable rhel-7-server-rhceph-{{ ceph_stable_rh_storage_version }}-mon-rpms
command: subscription-manager repos --enable rhel-7-server-rhceph-{{ ceph_rhcs_version }}-mon-rpms
changed_when: false
when:
- mon_group_name in group_names
- rh_storage_mon_repo.rc != 0
- name: check if the red hat storage osd repo is already present
shell: yum --noplugins --cacheonly repolist | grep -sq rhel-7-server-rhceph-{{ ceph_stable_rh_storage_version }}-osd-rpms
shell: yum --noplugins --cacheonly repolist | grep -sq rhel-7-server-rhceph-{{ ceph_rhcs_version }}-osd-rpms
changed_when: false
failed_when: false
register: rh_storage_osd_repo
when: osd_group_name in group_names
- name: enable red hat storage osd repository
command: subscription-manager repos --enable rhel-7-server-rhceph-{{ ceph_stable_rh_storage_version }}-osd-rpms
command: subscription-manager repos --enable rhel-7-server-rhceph-{{ ceph_rhcs_version }}-osd-rpms
changed_when: false
when:
- osd_group_name in group_names
- rh_storage_osd_repo.rc != 0
- name: check if the red hat storage rados gateway repo is already present
shell: yum --noplugins --cacheonly repolist | grep -sq rhel-7-server-rhceph-{{ ceph_stable_rh_storage_version }}-tools-rpms
shell: yum --noplugins --cacheonly repolist | grep -sq rhel-7-server-rhceph-{{ ceph_rhcs_version }}-tools-rpms
changed_when: false
failed_when: false
register: rh_storage_rgw_repo
when: rgw_group_name in group_names
- name: enable red hat storage rados gateway repository
command: subscription-manager repos --enable rhel-7-server-rhceph-{{ ceph_stable_rh_storage_version }}-tools-rpms
command: subscription-manager repos --enable rhel-7-server-rhceph-{{ ceph_rhcs_version }}-tools-rpms
changed_when: false
when:
- rgw_group_name in group_names

View File

@ -4,40 +4,40 @@
path: "{{ item }}"
state: directory
with_items:
- "{{ ceph_stable_rh_storage_mount_path }}"
- "{{ ceph_stable_rh_storage_repository_path }}"
- "{{ ceph_rhcs_mount_path }}"
- "{{ ceph_rhcs_repository_path }}"
- name: ensure destination iso directory exists
file:
path: "{{ ceph_stable_rh_storage_iso_path | dirname }}"
path: "{{ ceph_rhcs_iso_path | dirname }}"
state: directory
recurse: yes
when: "'{{ ceph_stable_rh_storage_iso_path | dirname }}' != '/'"
when: "'{{ ceph_rhcs_iso_path | dirname }}' != '/'"
- name: fetch the red hat storage iso from the ansible server
copy:
src: "{{ ceph_stable_rh_storage_iso_path }}"
dest: "{{ ceph_stable_rh_storage_iso_path }}"
src: "{{ ceph_rhcs_iso_path }}"
dest: "{{ ceph_rhcs_iso_path }}"
# assumption: ceph_stable_rh_storage_mount_path does not specify directory
# assumption: ceph_rhcs_mount_path does not specify directory
- name: mount red hat storage iso file
mount:
name: "{{ ceph_stable_rh_storage_mount_path }}"
src: "{{ ceph_stable_rh_storage_iso_path }}"
name: "{{ ceph_rhcs_mount_path }}"
src: "{{ ceph_rhcs_iso_path }}"
fstype: iso9660
opts: ro,loop,noauto
passno: 2
state: mounted
- name: copy red hat storage iso content
shell: cp -r {{ ceph_stable_rh_storage_mount_path }}/* {{ ceph_stable_rh_storage_repository_path }}
shell: cp -r {{ ceph_rhcs_mount_path }}/* {{ ceph_rhcs_repository_path }}
args:
creates: "{{ ceph_stable_rh_storage_repository_path }}/README"
creates: "{{ ceph_rhcs_repository_path }}/README"
- name: unmount red hat storage iso file
mount:
name: "{{ ceph_stable_rh_storage_mount_path }}"
src: "{{ ceph_stable_rh_storage_iso_path }}"
name: "{{ ceph_rhcs_mount_path }}"
src: "{{ ceph_rhcs_iso_path }}"
fstype: iso9660
state: unmounted

View File

@ -8,59 +8,65 @@ auth service required = none
auth client required = none
auth supported = none
{% endif %}
{% if not mon_containerized_deployment_with_kv %}
{% if not mon_containerized_deployment_with_kv and not mon_containerized_deployment %}
fsid = {{ fsid }}
{% endif %}
max open files = {{ max_open_files }}
{% if common_single_host_mode is defined %}
osd crush chooseleaf type = 0
{% endif %}
[client.libvirt]
admin socket = {{ rbd_client_admin_socket_path }}/$cluster-$type.$id.$pid.$cctid.asok # must be writable by QEMU and allowed by SELinux or AppArmor
log file = {{ rbd_client_log_file }} # must be writable by QEMU and allowed by SELinux or AppArmor
{# NOTE (leseb): the blank lines in-between are needed otherwise we won't get any line break #}
{% if groups[mon_group_name] is defined %}
mon_initial_members = {% if groups[mon_group_name] is defined %}{% for host in groups[mon_group_name] %}{% if hostvars[host]['ansible_fqdn'] is defined and mon_use_fqdn %}{{ hostvars[host]['ansible_fqdn'] }}{% if not loop.last %},{% endif %}{% elif hostvars[host]['ansible_hostname'] is defined %}{{ hostvars[host]['ansible_hostname'] }}{% if not loop.last %},{% endif %}{% endif %}{% endfor %}{% endif %}
{% endif %}
[mon]
{% if not mon_containerized_deployment_with_kv %}
{% for host in groups[mon_group_name] %}
{% if hostvars[host]['ansible_fqdn'] is defined and mon_use_fqdn %}
[mon.{{ hostvars[host]['ansible_fqdn'] }}]
host = {{ hostvars[host]['ansible_fqdn'] }}
{% elif hostvars[host]['ansible_hostname'] is defined %}
[mon.{{ hostvars[host]['ansible_hostname'] }}]
host = {{ hostvars[host]['ansible_hostname'] }}
{% endif %}
# we need to check if monitor_interface is defined in the inventory per host or if it's set in a group_vars file
{% if mon_containerized_deployment %}
{% set interface = ["ansible_",ceph_mon_docker_interface]|join %}
{% if interface in hostvars[host] and 'ipv4' in hostvars[host][interface] %}
# user address from interface {{ ceph_mon_docker_interface }}
mon addr = {{ hostvars[host][interface]['ipv4']['address'] }}
{% elif hostvars[host]['monitor_address'] is defined %}
# use host monitor address
mon addr = {{ hostvars[host]['monitor_address'] }}
{% elif monitor_address != "0.0.0.0" %}
# use group_var monitor address
mon addr = monitor_address
{% endif %}
{% if not mon_containerized_deployment and not mon_containerized_deployment_with_kv %}
{% if groups[mon_group_name] is defined %}
mon host = {% for host in groups[mon_group_name] %}{{ hostvars[host]['ansible_' + monitor_interface]['ipv4']['address'] }}{% if not loop.last %},{% endif %}{% endfor %}
{% elif (hostvars[host]['monitor_interface'] is defined and hostvars[host]['monitor_interface'] != "interface") or monitor_interface != "interface" %}
{% include 'mon_addr_interface.j2' %}
{% else %}
{% include 'mon_addr_address.j2' %}
{% endif %}
{% endif %}
{% if mon_containerized_deployment %}
fsid = {{ fsid }}
{% if groups[mon_group_name] is defined %}
{% for host in groups[mon_group_name] %}
{% if mon_containerized_deployment %}
{% set interface = ["ansible_",ceph_mon_docker_interface]|join %}
mon_host = {{ hostvars[host][interface]['ipv4']['address'] }}
{% if not loop.last %},{% endif %}
{% elif hostvars[host]['monitor_address'] is defined %}
mon_host = {{ hostvars[host]['monitor_address'] }}
{% if not loop.last %},{% endif %}
{% elif monitor_address != "0.0.0.0" %}
mon_host = monitor_address
{% if not loop.last %},{% endif %}
{% endif %}
{% endfor %}
{% endif %}
{% endif %}
{% if public_network is defined %}
public_network = {{ public_network }}
{% endif %}
{% if cluster_network is defined %}
cluster_network = {{ cluster_network }}
{% endif %}
{% if common_single_host_mode is defined %}
osd crush chooseleaf type = 0
{% endif %}
[client.libvirt]
admin socket = {{ rbd_client_admin_socket_path }}/$cluster-$type.$id.$pid.$cctid.asok # must be writable by QEMU and allowed by SELinux or AppArmor
log file = {{ rbd_client_log_file }} # must be writable by QEMU and allowed by SELinux or AppArmor
[osd]
osd mkfs type = {{ osd_mkfs_type }}
osd mkfs options xfs = {{ osd_mkfs_options_xfs }}
osd mount options xfs = {{ osd_mount_options_xfs }}
osd journal size = {{ journal_size }}
{% if cluster_network is defined %}
cluster_network = {{ cluster_network }}
{% endif %}
{% if public_network is defined %}
public_network = {{ public_network }}
{% endif %}
{% if filestore_xattr_use_omap != None %}
filestore xattr use omap = {{ filestore_xattr_use_omap }}
{% elif osd_mkfs_type == "ext4" %}

View File

@ -1,17 +1,18 @@
#jinja2: trim_blocks: "true", lstrip_blocks: "true"
# {{ ansible_managed }}
{% if fsal_ceph %}
EXPORT
{
Export_ID={{ ceph_nfs_export_id }};
Export_ID={{ ceph_nfs_ceph_export_id }};
Path = "/";
Pseudo = {{ ceph_nfs_pseudo_path }};
Pseudo = {{ ceph_nfs_ceph_pseudo_path }};
Access_Type = {{ ceph_nfs_access_type }};
Access_Type = {{ ceph_nfs_ceph_access_type }};
NFS_Protocols = {{ ceph_nfs_protocols }};
NFS_Protocols = {{ ceph_nfs_ceph_protocols }};
Transport_Protocols = TCP;
@ -21,3 +22,37 @@ EXPORT
Name = CEPH;
}
}
{% endif %}
{% if fsal_rgw %}
EXPORT
{
Export_ID={{ ceph_nfs_rgw_export_id }};
Path = "/";
Pseudo = {{ ceph_nfs_rgw_pseudo_path }};
Access_Type = {{ ceph_nfs_rgw_access_type }};
NFS_Protocols = {{ ceph_nfs_rgw_protocols }};
Transport_Protocols = TCP;
Sectype = sys,krb5,krb5i,krb5p;
FSAL {
Name = RGW;
User_Id = "{{ ceph_nfs_rgw_user }}";
Access_Key_Id ="{{ ceph_nfs_rgw_access_key }}";
Secret_Access_Key = "{{ ceph_nfs_rgw_secret_key }}";
}
}
{% endif %}
LOG {
Facility {
name = FILE;
destination = "{{ ceph_nfs_log_file }}";
enable = active;
}
}

View File

@ -1,2 +1 @@
mon addr = {{ hostvars[host]['monitor_address'] if hostvars[host]['monitor_address'] is defined else monitor_address }}
mon host = {% for host in groups[mon_group_name] %}{{ hostvars[host]['monitor_address'] if hostvars[host]['monitor_address'] is defined else monitor_address }}{% if not loop.last %},{% endif %}{% endfor %}

View File

@ -1,2 +1 @@
mon addr = {{ hostvars[host]['ansible_' + (hostvars[host]['monitor_interface'] if hostvars[host]['monitor_interface'] is defined else monitor_interface) ]['ipv4']['address'] }}
mon host = {% for host in groups[mon_group_name] %}{{ hostvars[host]['ansible_' + (hostvars[host]['monitor_interface'] if hostvars[host]['monitor_interface'] is defined else monitor_interface) ]['ipv4']['address'] }}{% if not loop.last %},{% endif %}{% endfor %}

View File

@ -1,21 +1,21 @@
# {{ ansible_managed }}
[rh_storage_mon]
name=Red Hat Ceph Storage - local packages for Ceph monitor
baseurl=file://{{ ceph_stable_rh_storage_repository_path }}/MON
baseurl=file://{{ ceph_rhcs_repository_path }}/MON
enabled=1
gpgcheck=1
priority=1
[rh_storage_osd]
name=Red Hat Ceph Storage - local packages for Ceph OSD
baseurl=file://{{ ceph_stable_rh_storage_repository_path }}/OSD
baseurl=file://{{ ceph_rhcs_repository_path }}/OSD
enabled=1
gpgcheck=1
priority=1
[rh_storage_tools]
name=Red Hat Ceph Storage - local packages for Ceph client, MDS, and RGW
baseurl=file://{{ ceph_stable_rh_storage_repository_path }}/Tools
baseurl=file://{{ ceph_rhcs_repository_path }}/Tools
enabled=1
gpgcheck=1
priority=1

View File

@ -17,6 +17,24 @@
when: ceph_health.rc != 0
- include: pre_requisite.yml
- include: "{{ playbook_dir }}/roles/ceph-common/tasks/misc/ntp_atomic.yml"
when:
- is_atomic
- ansible_os_family == 'RedHat'
- ntp_service_enabled
- include: "{{ playbook_dir }}/roles/ceph-common/tasks/misc/ntp_redhat.yml"
when:
- not is_atomic
- ansible_os_family == 'RedHat'
- ntp_service_enabled
- include: "{{ playbook_dir }}/roles/ceph-common/tasks/misc/ntp_debian.yml"
when:
- ansible_os_family == 'Debian'
- ntp_service_enabled
- include: "{{ playbook_dir }}/roles/ceph-common/tasks/docker/fetch_image.yml"
vars:
ceph_docker_username: "{{ ceph_mds_docker_username }}"

View File

@ -125,3 +125,34 @@
with_pkg
when: ansible_version['full'] | version_compare('2.1.0.0', '>=')
- name: install ntp on redhat using yum
yum:
name: ntp
state: present
when:
- ansible_os_family == 'RedHat'
- ansible_pkg_mgr == 'yum'
- ntp_service_enabled
tags:
with_pkg
- name: install ntp on redhat using dnf
dnf:
name: ntp
state: present
when:
- ansible_os_family == 'RedHat'
- ansible_pkg_mgr == 'dnf'
- ntp_service_enabled
tags:
with_pkg
- name: install ntp on debian
apt:
name: ntp
state: present
when:
- ansible_os_family == 'Debian'
- ntp_service_enabled
tags:
with_pkg

View File

@ -44,5 +44,5 @@
net: host
state: running
env: "CEPH_DAEMON=MDS,CEPHFS_CREATE=1,{{ ceph_mds_docker_extra_env }}"
volumes: "/var/lib/ceph:/var/lib/ceph,/etc/ceph:/etc/ceph"
volumes: "/var/lib/ceph:/var/lib/ceph,/etc/ceph:/etc/ceph,/etc/localtime:/etc/localtime:ro"
when: ansible_os_family != 'RedHat' and ansible_os_family != 'CoreOS'

View File

@ -14,6 +14,7 @@ ExecStart=/usr/bin/docker run --rm --net=host \
-e KV_TYPE={{kv_type}} \
-e KV_IP={{kv_endpoint}} \
{% endif -%}
-v /etc/localtime:/etc/localtime:ro \
--privileged \
-e CEPH_DAEMON=MDS \
-e CEPHFS_CREATE=1 \

View File

@ -53,7 +53,7 @@
- is_after_hammer
- name: ceph monitor mkfs without keyring (for or after infernalis release)
command: ceph-mon --setuser ceph --setgroup ceph --mkfs -i {{ monitor_name }} --fsid {{ fsid }}
command: ceph-mon --cluster {{ cluster }} --setuser ceph --setgroup ceph --mkfs -i {{ monitor_name }} --fsid {{ fsid }}
args:
creates: /var/lib/ceph/mon/{{ cluster }}-{{ monitor_name }}/store.db
when:

View File

@ -20,6 +20,23 @@
- include: pre_requisite.yml
- include: "{{ playbook_dir }}/roles/ceph-common/tasks/misc/ntp_atomic.yml"
when:
- is_atomic
- ansible_os_family == 'RedHat'
- ntp_service_enabled
- include: "{{ playbook_dir }}/roles/ceph-common/tasks/misc/ntp_redhat.yml"
when:
- not is_atomic
- ansible_os_family == 'RedHat'
- ntp_service_enabled
- include: "{{ playbook_dir }}/roles/ceph-common/tasks/misc/ntp_debian.yml"
when:
- ansible_os_family == 'Debian'
- ntp_service_enabled
- include: "{{ playbook_dir }}/roles/ceph-common/tasks/docker/fetch_image.yml"
vars:
ceph_docker_username: "{{ ceph_mon_docker_username }}"

View File

@ -126,3 +126,35 @@
tags:
with_pkg
when: ansible_version['full'] | version_compare('2.1.0.0', '>=')
- name: install ntp on redhat using yum
yum:
name: ntp
state: present
when:
- ansible_os_family == 'RedHat'
- ansible_pkg_mgr == 'yum'
- ntp_service_enabled
tags:
with_pkg
- name: install ntp on redhat using dnf
dnf:
name: ntp
state: present
when:
- ansible_os_family == 'RedHat'
- ansible_pkg_mgr == 'dnf'
- ntp_service_enabled
tags:
with_pkg
- name: install ntp on debian
apt:
name: ntp
state: present
when:
- ansible_os_family == 'Debian'
- ntp_service_enabled
tags:
with_pkg

View File

@ -85,7 +85,7 @@
state: "running"
privileged: "{{ mon_docker_privileged }}"
env: "MON_IP={{ hostvars[inventory_hostname]['ansible_' + ceph_mon_docker_interface]['ipv4']['address'] }},CEPH_DAEMON=MON,CEPH_PUBLIC_NETWORK={{ ceph_mon_docker_subnet }},CEPH_FSID={{ fsid }},{{ ceph_mon_extra_envs }}"
volumes: "/var/lib/ceph:/var/lib/ceph,/etc/ceph:/etc/ceph"
volumes: "/var/lib/ceph:/var/lib/ceph,/etc/ceph:/etc/ceph,/etc/localtime:/etc/localtime:ro"
when:
- ansible_os_family != 'RedHat'
- ansible_os_family != 'CoreOS'
@ -99,6 +99,7 @@
state: "running"
privileged: "{{ mon_docker_privileged }}"
env: "KV_TYPE={{kv_type}},KV_IP={{kv_endpoint}},MON_IP={{ hostvars[inventory_hostname]['ansible_' + ceph_mon_docker_interface]['ipv4']['address'] }},CEPH_DAEMON=MON,CEPH_PUBLIC_NETWORK={{ ceph_mon_docker_subnet }},{{ ceph_mon_extra_envs }}"
volumes: "/etc/localtime:/etc/localtime:ro"
when:
- ansible_os_family != 'RedHat'
- ansible_os_family != 'CoreOS'

View File

@ -10,13 +10,15 @@
with_items:
- done
- upstart
when: not use_systemd
when:
- not use_systemd
- name: start and add that the monitor service to the init sequence (ubuntu)
command: initctl emit ceph-mon cluster={{ cluster }} id={{ monitor_name }}
changed_when: false
failed_when: false
when: not use_systemd
when:
- not use_systemd
# NOTE (leseb): somehow the service ansible module is messing things up
# as a safety measure we run the raw command

View File

@ -15,6 +15,7 @@ ExecStart=/usr/bin/docker run --rm --name %i --net=host \
-e KV_IP={{kv_endpoint}}\
-e KV_PORT={{kv_port}} \
{% endif -%}
-v /etc/localtime:/etc/localtime:ro \
{% if mon_docker_privileged -%}
--privileged \
{% endif -%}

View File

@ -27,6 +27,28 @@ ceph_nfs_export_id: 20134
ceph_nfs_pseudo_path: "/ceph"
ceph_nfs_protocols: "3,4"
ceph_nfs_access_type: "RW"
ceph_nfs_log_file: "/var/log/ganesha.log"
####################
# FSAL Ceph Config #
####################
ceph_nfs_ceph_export_id: 20134
ceph_nfs_ceph_pseudo_path: "/ceph"
ceph_nfs_ceph_protocols: "3,4"
ceph_nfs_ceph_access_type: "RW"
###################
# FSAL RGW Config #
###################
ceph_nfs_rgw_export_id: 20134
ceph_nfs_rgw_pseudo_path: "/ceph"
ceph_nfs_rgw_protocols: "3,4"
ceph_nfs_rgw_access_type: "RW"
ceph_nfs_rgw_user: "cephnfs"
# Note: keys are optional and can be generated, but not on containerized, where
# they must be configered.
#ceph_nfs_rgw_access_key: "QFAMEDSJP5DEKJO0DDXY"
#ceph_nfs_rgw_secret_key: "iaSFLDVvDdQt6lkNzHyW4fPLZugBAI1g17LO0+87[MAC[M#C"
###################

View File

@ -1,10 +1,28 @@
---
- name: push ceph files to the ansible server
fetch:
src: "{{ item.0 }}"
dest: "{{ fetch_directory }}/docker_mon_files/{{ item.0 }}"
flat: yes
- name: set config and keys paths
set_fact:
ceph_config_keys:
- /etc/ceph/ceph.conf
- /etc/ceph/ceph.client.admin.keyring
- /var/lib/ceph/radosgw/keyring
- name: stat for config and keys
local_action: stat path={{ fetch_directory }}/docker_mon_files/{{ item }}
with_items: ceph_config_keys
changed_when: false
become: false
failed_when: false
register: statconfig
- name: try to fetch config and keys
copy:
src: "{{ fetch_directory }}/docker_mon_files/{{ item.0 }}"
dest: "{{ item.0 }}"
owner: "64045"
group: "64045"
mode: 0644
changed_when: false
with_together:
- ceph_config_keys
- statconfig.results
when: item.1.stat.exists == false
when: item.1.stat.exists == true

View File

@ -7,6 +7,21 @@
group: root
mode: 0644
- name: create the nfs rgw user
docker:
image: "{{ ceph_rgw_docker_username }}/{{ ceph_rgw_docker_imagename }}:{{ ceph_rgw_docker_image_tag }}"
name: ceph-{{ ansible_hostname }}-rgw-user
expose: "{{ ceph_rgw_civetweb_port }}"
ports: "{{ ceph_rgw_civetweb_port }}:{{ ceph_rgw_civetweb_port }}"
state: running
env: "CEPH_DAEMON=RGW_USER,RGW_USERNAME={{ ceph_nfs_rgw_user }},RGW_USER_ACCESS_KEY={{ ceph_nfs_rgw_access_key }},RGW_USER_SECRET_KEY={{ ceph_nfs_rgw_secret_key }}"
volumes: "/var/lib/ceph:/var/lib/ceph,/etc/ceph:/etc/ceph"
when: fsal_rgw
- name: get user create output
command: docker logs ceph-{{ ansible_hostname }}-rgw-user
register: rgwuser
- name: generate ganesha configuration file
action: config_template
args:

View File

@ -22,6 +22,7 @@
with_items:
- /etc/ceph/
- /var/lib/ceph/
- /var/lib/ceph/radosgw
when: not after_hammer
- name: create bootstrap directories (after hammer)
@ -34,6 +35,7 @@
with_items:
- /etc/ceph/
- /var/lib/ceph/
- /var/lib/ceph/radosgw
when: after_hammer
- name: create ganesha directories

View File

@ -1,13 +1,12 @@
---
- name: set config and keys paths
- name: set config paths
set_fact:
ceph_config_keys:
- /etc/ceph/ceph.conf
nfs_config_keys:
- /etc/ganesha/ganesha.conf
- name: stat for config and keys
local_action: stat path={{ fetch_directory }}/docker_mon_files/{{ item }}
with_items: ceph_config_keys
with_items: nfs_config_keys
changed_when: false
become: false
failed_when: false
@ -22,6 +21,16 @@
mode: 0644
changed_when: false
with_together:
- ceph_config_keys
- nfs_config_keys
- statconfig.results
when: item.1.stat.exists == true
- name: push ganesha files to the ansible server
fetch:
src: "{{ item.0 }}"
dest: "{{ fetch_directory }}/docker_mon_files/{{ item.0 }}"
flat: yes
with_together:
- nfs_config_keys
- statconfig.results
when: item.1.stat.exists == false

View File

@ -20,6 +20,23 @@
- include: pre_requisite.yml
- include: "{{ playbook_dir }}/roles/ceph-common/tasks/misc/ntp_atomic.yml"
when:
- is_atomic
- ansible_os_family == 'RedHat'
- ntp_service_enabled
- include: "{{ playbook_dir }}/roles/ceph-common/tasks/misc/ntp_redhat.yml"
when:
- not is_atomic
- ansible_os_family == 'RedHat'
- ntp_service_enabled
- include: "{{ playbook_dir }}/roles/ceph-common/tasks/misc/ntp_debian.yml"
when:
- ansible_os_family == 'Debian'
- ntp_service_enabled
- include: "{{ playbook_dir }}/roles/ceph-common/tasks/docker/fetch_image.yml"
vars:
ceph_docker_username: "{{ ceph_nfs_docker_username }}"
@ -28,12 +45,19 @@
- include: dirs_permissions.yml
# let the first ganesha create configs and keyrings
# Copy Ceph configs to host
- include: copy_configs.yml
- include: selinux.yml
when: ansible_os_family == 'RedHat'
# let the first ganesha create configs and users
- include: create_configs.yml
when:
inventory_hostname == groups.nfss[0] and
mon_containerized_default_ceph_conf_with_kv
# Copy Ganesha configs to host
- include: fetch_configs.yml
when: not mon_containerized_deployment_with_kv
@ -42,5 +66,3 @@
- include: start_docker_nfs.yml
- include: copy_configs.yml
when: not mon_containerized_deployment_with_kv

View File

@ -97,3 +97,35 @@
enabled: yes
tags:
with_pkg
- name: install ntp on redhat using yum
yum:
name: ntp
state: present
when:
- ansible_os_family == 'RedHat'
- ansible_pkg_mgr == 'yum'
- ntp_service_enabled
tags:
with_pkg
- name: install ntp on redhat using dnf
dnf:
name: ntp
state: present
when:
- ansible_os_family == 'RedHat'
- ansible_pkg_mgr == 'dnf'
- ntp_service_enabled
tags:
with_pkg
- name: install ntp on debian
apt:
name: ntp
state: present
when:
- ansible_os_family == 'Debian'
- ntp_service_enabled
tags:
with_pkg

View File

@ -61,7 +61,7 @@
privileged: true
ports: "{{ ceph_nfs_port }}:{{ ceph_nfs_port }},111:111"
env: "CEPH_DAEMON=NFS,CEPH_PUBLIC_NETWORK={{ ceph_nfs_docker_subnet }},{{ ceph_nfs_extra_envs }}"
volumes: "/etc/ceph:/etc/ceph,/etc/ganesha:/etc/ganesha"
volumes: "/etc/ceph:/etc/ceph,/etc/ganesha:/etc/ganesha,/etc/localtime:/etc/localtime:ro"
when:
not is_atomic and
ansible_os_family != 'CoreOS' and
@ -75,7 +75,7 @@
state: "running"
privileged: true
env: "CEPH_DAEMON=NFS,CEPH_PUBLIC_NETWORK={{ ceph_nfs_docker_subnet }},{{ ceph_nfs_extra_envs }}"
volumes: "/etc/ganesha:/etc/ganesha"
volumes: "/etc/ganesha:/etc/ganesha,/etc/localtime:/etc/localtime:ro"
when:
not is_atomic and
ansible_os_family != 'CoreOS' and

View File

@ -1,5 +1,5 @@
---
- name: create NFS gateway directories
- name: create nfs gateway directories
file:
path: "{{ item }}"
state: directory
@ -10,7 +10,33 @@
- /var/lib/nfs/ganesha
- /var/run/ganesha
- name: start NFS gateway service
- name: create rgw nfs user
command: radosgw-admin user create --uid={{ ceph_nfs_rgw_user }} --display-name="RGW NFS User"
register: rgwuser
when: fsal_rgw
- name: set access key
set_fact:
ceph_nfs_rgw_access_key: "{{ (rgwuser.stdout | from_json)['keys'][0]['access_key'] }}"
when: fsal_rgw
- name: set secret key
set_fact:
ceph_nfs_rgw_secret_key: "{{(rgwuser.stdout | from_json)['keys'][0]['secret_key']}}"
when: fsal_rgw
- name: generate ganesha configuration file
action: config_template
args:
src: "{{ playbook_dir }}/roles/ceph-common/templates/ganesha.conf.j2"
dest: /etc/ganesha/ganesha.conf
owner: "root"
group: "root"
mode: "0644"
config_overrides: "{{ ganesha_conf_overrides }}"
config_type: ini
- name: start nfs gateway service
service:
name: nfs-ganesha
state: started

View File

@ -7,7 +7,7 @@ After=docker.service
EnvironmentFile=-/etc/environment
ExecStartPre=-/usr/bin/docker rm %i
ExecStartPre=/usr/bin/mkdir -p /etc/ceph /etc/ganesha /var/lib/nfs/ganesha
ExecStart=/usr/bin/docker run --rm --name %i --net=host \
ExecStart=/usr/bin/docker run --rm --net=host \
{% if not mon_containerized_deployment_with_kv -%}
-v /etc/ceph:/etc/ceph \
-v /etc/ganesha:/etc/ganesha \
@ -15,10 +15,10 @@ ExecStart=/usr/bin/docker run --rm --name %i --net=host \
-e KV_TYPE={{kv_type}} \
-e KV_IP={{kv_endpoint}}\
{% endif -%}
-v /etc/localtime:/etc/localtime:ro \
--privileged \
-e CEPH_DAEMON=NFS \
-e CEPH_PUBLIC_NETWORK={{ ceph_mon_docker_subnet }} \
--name={{ ansible_hostname }} \
--name=nfs-{{ ansible_hostname }} \
{{ ceph_nfs_docker_username }}/{{ ceph_nfs_docker_imagename }}:{{ ceph_nfs_docker_image_tag }}
ExecStopPost=-/usr/bin/docker stop %i
Restart=always

View File

@ -113,7 +113,7 @@ bluestore: false
# Keys are stored into the monitors k/v store
# Use 'true' to enable this scenario
# Both journal and data are stored on the same dm-crypt encrypted device
dmcrypt_journal_colocation: false
dmcrypt_journal_collocation: false
# VI. Encrypt osd data and/or journal devices with dm-crypt.

View File

@ -36,7 +36,7 @@
- item.value.removable == "0"
- item.value.partitions|count == 0
- osd_auto_discovery
- dmcrypt_journal_colocation
- dmcrypt_journal_collocation
- name: activate osd(s) when device is a disk (dmcrypt)
command: ceph-disk activate --dmcrypt {{ item.1 | regex_replace('^(\/dev\/cciss\/c[0-9]{1}d[0-9]{1})$', '\\1p') }}1
@ -56,7 +56,7 @@
# https://github.com/ansible/ansible/issues/4297
- name: combine ispartition results
set_fact:
combined_activate_osd_disk_results: "{{ activate_osd_disk if not dmcrypt_journal_colocation else activate_osd_disk_dmcrypt }}"
combined_activate_osd_disk_results: "{{ activate_osd_disk if not dmcrypt_journal_collocation else activate_osd_disk_dmcrypt }}"
- name: fail if ceph-disk cannot create an OSD
fail:

View File

@ -20,6 +20,23 @@
- include: pre_requisite.yml
- include: "{{ playbook_dir }}/roles/ceph-common/tasks/misc/ntp_atomic.yml"
when:
- is_atomic
- ansible_os_family == 'RedHat'
- ntp_service_enabled
- include: "{{ playbook_dir }}/roles/ceph-common/tasks/misc/ntp_redhat.yml"
when:
- not is_atomic
- ansible_os_family == 'RedHat'
- ntp_service_enabled
- include: "{{ playbook_dir }}/roles/ceph-common/tasks/misc/ntp_debian.yml"
when:
- ansible_os_family == 'Debian'
- ntp_service_enabled
- include: "{{ playbook_dir }}/roles/ceph-common/tasks/docker/fetch_image.yml"
vars:
ceph_docker_username: '{{ ceph_osd_docker_username }}'

View File

@ -125,3 +125,35 @@
tags:
with_pkg
when: ansible_version['full'] | version_compare('2.1.0.0', '>=')
- name: install ntp on redhat using yum
yum:
name: ntp
state: present
when:
- ansible_os_family == 'RedHat'
- ansible_pkg_mgr == 'yum'
- ntp_service_enabled
tags:
with_pkg
- name: install ntp on redhat using dnf
dnf:
name: ntp
state: present
when:
- ansible_os_family == 'RedHat'
- ansible_pkg_mgr == 'dnf'
- ntp_service_enabled
tags:
with_pkg
- name: install ntp on debian
apt:
name: ntp
state: present
when:
- ansible_os_family == 'Debian'
- ntp_service_enabled
tags:
with_pkg

View File

@ -28,10 +28,11 @@
-v /etc/ceph:/etc/ceph \
-v /var/lib/ceph/:/var/lib/ceph/ \
-v /dev:/dev \
-v /etc/localtime:/etc/localtime:ro \
-e "OSD_DEVICE={{ item.0 }}" \
-e "{{ ceph_osd_docker_prepare_env }}" \
"{{ ceph_osd_docker_username }}/{{ ceph_osd_docker_imagename }}:{{ ceph_osd_docker_image_tag }}" \
OSD_CEPH_DISK_PREPARE
-e CEPH_DAEMON=osd_ceph_disk_prepare \
"{{ ceph_osd_docker_username }}/{{ ceph_osd_docker_imagename }}:{{ ceph_osd_docker_image_tag }}"
with_together:
- ceph_osd_docker_devices
- osd_prepared.results
@ -48,6 +49,7 @@
--name="{{ ansible_hostname }}-osd-prepare-{{ item.0 |
regex_replace('/', '') }}" \
-v /dev:/dev \
-v /etc/localtime:/etc/localtime:ro \
-e "OSD_DEVICE={{ item.0 }}" \
-e "{{ ceph_osd_docker_prepare_env }}" \
-e CEPH_DAEMON=osd_ceph_disk_prepare \
@ -106,7 +108,7 @@
state: started
privileged: yes
env: "OSD_DEVICE={{ item }},{{ ceph_osd_docker_extra_env }}"
volumes: "/var/lib/ceph:/var/lib/ceph,/etc/ceph:/etc/ceph,/dev:/dev,/run:/run"
volumes: "/var/lib/ceph:/var/lib/ceph,/etc/ceph:/etc/ceph,/etc/localtime:/etc/localtime:ro,/dev:/dev,/run:/run"
with_items: ceph_osd_docker_devices
when:
- ansible_os_family != 'RedHat'
@ -122,7 +124,7 @@
state: running
privileged: yes
env: "KV_TYPE={{kv_type}},KV_IP={{kv_endpoint}},OSD_DEVICE={{ item }},{{ ceph_osd_docker_extra_env }}"
volumes: "/dev/:/dev/"
volumes: "/etc/localtime:/etc/localtime:ro,/dev/:/dev/"
with_items: ceph_osd_docker_devices
when:
- ansible_os_family != 'RedHat'

View File

@ -32,9 +32,9 @@
# Hard code this so we will skip the entire file instead of individual tasks (Default isn't Consistent)
static: False
- include: ./scenarios/dmcrypt-journal-colocation.yml
- include: ./scenarios/dmcrypt-journal-collocation.yml
when:
- dmcrypt_journal_colocation
- dmcrypt_journal_collocation
- not osd_containerized_deployment
# Hard code this so we will skip the entire file instead of individual tasks (Default isn't Consistent)
static: False

View File

@ -35,7 +35,7 @@
set_fact:
copy_admin_key: true
when:
- dmcrypt_journal_colocation or dmcrypt_dedicated_journal
- dmcrypt_journal_collocation or dmcrypt_dedicated_journal
- name: copy osd bootstrap key
copy:

View File

@ -16,7 +16,7 @@
- ansible_devices is defined
- item.value.removable == "0"
- item.value.partitions|count == 0
- dmcrypt_journal_colocation
- dmcrypt_journal_collocation
- osd_auto_discovery
- name: manually prepare osd disk(s) (dmcrypt)
@ -30,7 +30,7 @@
- not item.1.get("skipped")
- item.0.get("rc", 0) != 0
- item.1.get("rc", 0) != 0
- dmcrypt_journal_colocation
- dmcrypt_journal_collocation
- not osd_auto_discovery
- include: ../activate_osds.yml

View File

@ -15,6 +15,7 @@ ExecStart=/usr/bin/docker run --rm --net=host --pid=host\
-e KV_IP={{kv_endpoint}} \
-e KV_PORT={{kv_port}} \
{% endif -%}
-v /etc/localtime:/etc/localtime:ro \
-v /dev:/dev \
--privileged \
-e CEPH_DAEMON=OSD_CEPH_DISK_ACTIVATE \

View File

@ -17,6 +17,24 @@
when: ceph_health.rc != 0
- include: pre_requisite.yml
- include: "{{ playbook_dir }}/roles/ceph-common/tasks/misc/ntp_atomic.yml"
when:
- is_atomic
- ansible_os_family == 'RedHat'
- ntp_service_enabled
- include: "{{ playbook_dir }}/roles/ceph-common/tasks/misc/ntp_redhat.yml"
when:
- not is_atomic
- ansible_os_family == 'RedHat'
- ntp_service_enabled
- include: "{{ playbook_dir }}/roles/ceph-common/tasks/misc/ntp_debian.yml"
when:
- ansible_os_family == 'Debian'
- ntp_service_enabled
- include: "{{ playbook_dir }}/roles/ceph-common/tasks/docker/fetch_image.yml"
vars:
ceph_docker_username: "{{ ceph_rbd_mirror_docker_username }}"

View File

@ -124,3 +124,35 @@
tags:
with_pkg
when: ansible_version['full'] | version_compare('2.1.0.0', '>=')
- name: install ntp on redhat using yum
yum:
name: ntp
state: present
when:
- ansible_os_family == 'RedHat'
- ansible_pkg_mgr == 'yum'
- ntp_service_enabled
tags:
with_pkg
- name: install ntp on redhat using dnf
dnf:
name: ntp
state: present
when:
- ansible_os_family == 'RedHat'
- ansible_pkg_mgr == 'dnf'
- ntp_service_enabled
tags:
with_pkg
- name: install ntp on debian
apt:
name: ntp
state: present
when:
- ansible_os_family == 'Debian'
- ntp_service_enabled
tags:
with_pkg

View File

@ -8,6 +8,5 @@
shell: chcon -Rt svirt_sandbox_file_t {{ item }}
with_items:
- /etc/ceph
- /var/lib/ceph
changed_when: false
when: sestatus.stdout != 'Disabled'

View File

@ -4,21 +4,14 @@
become: true
template:
src: "{{ role_path }}/templates/ceph-rbd-mirror.service.j2"
dest: /var/lib/ceph/ceph-rbd-mirror@.service
dest: /etc/systemd/system/ceph-rbd-mirror@.service
owner: "root"
group: "root"
mode: "0644"
when: ansible_os_family == 'RedHat' or ansible_os_family == 'CoreOS'
- name: link systemd unit file for rbd mirror instance
file:
src: /var/lib/ceph/ceph-rbd-mirror@.service
dest: /etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror@{{ ansible_hostname }}.service
state: link
when: ansible_os_family == 'RedHat' or ansible_os_family == 'CoreOS'
- name: enable systemd unit file for rbd mirror instance
command: systemctl enable /etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror@{{ ansible_hostname }}.service
command: systemctl enable ceph-rbd-mirror@{{ ansible_hostname }}.service
failed_when: false
changed_when: false
when: ansible_os_family == 'RedHat' or ansible_os_family == 'CoreOS'
@ -39,9 +32,9 @@
- name: run the ceph rbd mirror docker image
docker:
image: "{{ ceph_rbd_mirror_docker_username }}/{{ ceph_rbd_mirror_docker_imagename }}"
name: ceph-{{ ansible_hostname }}-rbd-mirror
image: "{{ ceph_rbd_mirror_docker_username }}/{{ ceph_rbd_mirror_docker_imagename }}:{{ ceph_rbd_mirror_docker_image_tag }}"
name: "{{ ansible_hostname }}"
net: host
state: running
volumes: "/var/lib/ceph:/var/lib/ceph,/etc/ceph:/etc/ceph"
volumes: "/etc/ceph:/etc/ceph,/etc/localtime:/etc/localtime:ro"
when: ansible_os_family != 'RedHat' and ansible_os_family != 'CoreOS'

View File

@ -8,16 +8,16 @@ ExecStartPre=-/usr/bin/docker stop {{ ansible_hostname }}
ExecStartPre=-/usr/bin/docker rm {{ ansible_hostname }}
ExecStart=/usr/bin/docker run --rm --net=host \
{% if not rbd_mirror_containerized_deployment_with_kv -%}
-v /var/lib/ceph:/var/lib/ceph \
-v /etc/ceph:/etc/ceph \
{% else -%}
-e KV_TYPE={{kv_type}} \
-e KV_IP={{kv_endpoint}} \
{% endif -%}
-v /etc/localtime:/etc/localtime:ro \
--privileged \
-e CEPH_DAEMON=RBD_MIRROR \
--name={{ ansible_hostname }} \
{{ ceph_rbd_mirror_docker_username }}/{{ ceph_rbd_mirror_docker_imagename }}
{{ ceph_rbd_mirror_docker_username }}/{{ ceph_rbd_mirror_docker_imagename }}:{{ ceph_rbd_mirror_docker_image_tag }}
ExecStopPost=-/usr/bin/docker stop {{ ansible_hostname }}
Restart=always
RestartSec=10s

View File

@ -1,5 +1,31 @@
---
- name: check if it is Atomic host
stat: path=/run/ostree-booted
register: stat_ostree
- name: set fact for using Atomic host
set_fact:
is_atomic: '{{ stat_ostree.stat.exists }}'
- include: pre_requisite.yml
- include: "{{ playbook_dir }}/roles/ceph-common/tasks/misc/ntp_atomic.yml"
when:
- is_atomic
- ansible_os_family == 'RedHat'
- ntp_service_enabled
- include: "{{ playbook_dir }}/roles/ceph-common/tasks/misc/ntp_redhat.yml"
when:
- not is_atomic
- ansible_os_family == 'RedHat'
- ntp_service_enabled
- include: "{{ playbook_dir }}/roles/ceph-common/tasks/misc/ntp_debian.yml"
when:
- ansible_os_family == 'Debian'
- ntp_service_enabled
- include: "{{ playbook_dir }}/roles/ceph-common/tasks/docker/fetch_image.yml"
vars:
ceph_docker_username: "{{ ceph_restapi_docker_username }}"

View File

@ -122,3 +122,35 @@
enabled: yes
tags:
with_pkg
- name: install ntp on redhat using yum
yum:
name: ntp
state: present
when:
- ansible_os_family == 'RedHat'
- ansible_pkg_mgr == 'yum'
- ntp_service_enabled
tags:
with_pkg
- name: install ntp on redhat using dnf
dnf:
name: ntp
state: present
when:
- ansible_os_family == 'RedHat'
- ansible_pkg_mgr == 'dnf'
- ntp_service_enabled
tags:
with_pkg
- name: install ntp on debian
apt:
name: ntp
state: present
when:
- ansible_os_family == 'Debian'
- ntp_service_enabled
tags:
with_pkg

View File

@ -7,4 +7,4 @@
expose: "{{ ceph_restapi_port }}"
state: running
env: "RESTAPI_IP={{ hostvars[inventory_hostname]['ansible_' + ceph_restapi_docker_interface]['ipv4']['address'] }},CEPH_DAEMON=RESTAPI,{{ ceph_restapi_docker_extra_env }}"
volumes: "/etc/ceph:/etc/ceph"
volumes: "/etc/ceph:/etc/ceph,/etc/localtime:/etc/localtime:ro"

View File

@ -0,0 +1,36 @@
---
- name: set config and keys paths
set_fact:
rgw_config_keys:
- "/var/lib/ceph/radosgw/{{ ansible_hostname }}/keyring"
when: fsal_rgw
- name: wait for rgw keyring
wait_for: path="/var/lib/ceph/radosgw/{{ ansible_hostname }}/keyring"
when:
- fsal_rgw
- inventory_hostname == groups.rgws[0]
- name: stat for config and keys
local_action: stat path={{ fetch_directory }}/docker_mon_files/{{ item }}
with_items: rgw_config_keys
changed_when: false
become: false
failed_when: false
register: statconfig
when:
- fsal_rgw
- inventory_hostname == groups.rgws[0]
- name: push ceph files to the ansible server
fetch:
src: "{{ item.0 }}"
dest: "{{ fetch_directory }}/docker_mon_files/var/lib/ceph/radosgw/keyring"
flat: yes
with_together:
- rgw_config_keys
- statconfig.results
when:
- item.1.stat.exists == false
- fsal_rgw
- inventory_hostname == groups.rgws[0]

View File

@ -17,6 +17,24 @@
when: ceph_health.rc != 0
- include: pre_requisite.yml
- include: "{{ playbook_dir }}/roles/ceph-common/tasks/misc/ntp_atomic.yml"
when:
- is_atomic
- ansible_os_family == 'RedHat'
- ntp_service_enabled
- include: "{{ playbook_dir }}/roles/ceph-common/tasks/misc/ntp_redhat.yml"
when:
- not is_atomic
- ansible_os_family == 'RedHat'
- ntp_service_enabled
- include: "{{ playbook_dir }}/roles/ceph-common/tasks/misc/ntp_debian.yml"
when:
- ansible_os_family == 'Debian'
- ntp_service_enabled
- include: "{{ playbook_dir }}/roles/ceph-common/tasks/docker/fetch_image.yml"
vars:
ceph_docker_username: "{{ ceph_rgw_docker_username }}"
@ -31,3 +49,5 @@
when: ansible_os_family == 'RedHat'
- include: start_docker_rgw.yml
- include: copy_configs.yml

View File

@ -110,3 +110,35 @@
enabled: yes
tags:
with_pkg
- name: install ntp on redhat using yum
yum:
name: ntp
state: present
when:
- ansible_os_family == 'RedHat'
- ansible_pkg_mgr == 'yum'
- ntp_service_enabled
tags:
with_pkg
- name: install ntp on redhat using dnf
dnf:
name: ntp
state: present
when:
- ansible_os_family == 'RedHat'
- ansible_pkg_mgr == 'dnf'
- ntp_service_enabled
tags:
with_pkg
- name: install ntp on debian
apt:
name: ntp
state: present
when:
- ansible_os_family == 'Debian'
- ntp_service_enabled
tags:
with_pkg

View File

@ -45,5 +45,5 @@
ports: "{{ ceph_rgw_civetweb_port }}:{{ ceph_rgw_civetweb_port }}"
state: running
env: "CEPH_DAEMON=RGW,{{ ceph_rgw_docker_extra_env }}"
volumes: "/var/lib/ceph:/var/lib/ceph,/etc/ceph:/etc/ceph"
volumes: "/var/lib/ceph:/var/lib/ceph,/etc/ceph:/etc/ceph,/etc/localtime:/etc/localtime:ro"
when: ansible_os_family != 'RedHat' and ansible_os_family != 'CoreOS'

View File

@ -41,6 +41,14 @@
group: "{{ key_group }}"
when: cephx
- name: ensure ceph-radosgw systemd unit file is present
command: chkconfig --add ceph-radosgw
args:
creates: /var/run/systemd/generator.late/ceph-radosgw.service
when:
- ansible_os_family == "RedHat"
- is_before_infernalis
- name: activate rados gateway with upstart
file:
path: /var/lib/ceph/radosgw/{{ cluster }}-rgw.{{ ansible_hostname }}/{{ item }}

View File

@ -14,6 +14,7 @@ ExecStart=/usr/bin/docker run --rm --net=host \
-e KV_TYPE={{kv_type}} \
-e KV_IP={{kv_endpoint}} \
{% endif -%}
-v /etc/localtime:/etc/localtime:ro \
--privileged \
-e CEPH_DAEMON=RGW \
--name={{ ansible_hostname }} \

45
rundep.sample 100644
View File

@ -0,0 +1,45 @@
#Package lines can be commented out with '#'
#
#boost-atomic
#boost-chrono
#boost-date-time
#boost-iostreams
#boost-program
#boost-random
#boost-regex
#boost-system
#boost-thread
#bzip2-libs
#cyrus-sasl-lib
#expat
#fcgi
#fuse-libs
#glibc
#hdparm
#keyutils-libs
#leveldb
#libaio
#libatomic_ops
#libattr
#libblkid
#libcap
#libcom_err
#libcurl
#libgcc
#libicu
#libidn
#libnghttp2
#libpsl
#libselinux
#libssh2
#libstdc++
#libunistring
#nss-softokn-freebl
#openldap
#openssl-libs
#pcre
#python-nose
#python-sphinx
#snappy
#systemd-libs
#zlib

View File

@ -0,0 +1,27 @@
#!/bin/bash -e
#
# Copyright (C) 2014, 2015 Red Hat <contact@redhat.com>
#
# Author: Daniel Lin <danielin@umich.edu>
#
# This library is free software; you can redistribute it and/or
# modify it under the terms of the GNU Lesser General Public
# License as published by the Free Software Foundation; either
# version 2.1 of the License, or (at your option) any later version.
#
if test -f /etc/redhat-release ; then
PACKAGE_INSTALLER=yum
elif type apt-get > /dev/null 2>&1 ; then
PACKAGE_INSTALLER=apt-get
else
echo "ERROR: Package Installer could not be determined"
exit 1
fi
while read p; do
if [[ $p =~ ^#.* ]] ; then
continue
fi
$PACKAGE_INSTALLER install $p -y
done < $1

View File

@ -8,6 +8,7 @@ osd_vms: 1
mds_vms: 0
rgw_vms: 0
nfs_vms: 0
rbd_mirror_vms: 0
client_vms: 0
# Deploy RESTAPI on each of the Monitors
@ -23,6 +24,9 @@ disks: "[ '/dev/sdb', '/dev/sdc' ]"
eth: 'enp0s8'
vagrant_box: centos/atomic-host
# The sync directory changes based on vagrant box
# Set to /home/vagrant/sync for Centos/7, /home/{ user }/vagrant for openstack and defaults to /vagrant
vagrant_sync_dir: /home/vagrant/sync
# if vagrant fails to attach storage controller, add the storage controller name by:
# VBoxManage storagectl `VBoxManage list vms |grep ceph-ansible_osd0|awk '{print $1}'|tr \" ' '` --name "SATA" --add sata

View File

@ -44,6 +44,9 @@ disks: "[ '/dev/sdb', '/dev/sdc' ]"
# - https://atlas.hashicorp.com/boxes/search?utf8=✓&sort=&provider=virtualbox&q=
# - https://download.gluster.org/pub/gluster/purpleidea/vagrant/
vagrant_box: ubuntu/trusty64
# The sync directory changes based on vagrant box
# Set to /home/vagrant/sync for Centos/7, /home/{ user }/vagrant for openstack and defaults to /vagrant
vagrant_sync_dir: /home/vagrant/sync
# VAGRANT URL
# This is a URL to download an image from an alternate location. vagrant_box
# above should be set to the filename of the image.