From 7ddbe7471228b410d56d36b00a449d1cb8434d4a Mon Sep 17 00:00:00 2001 From: Alex Schultz Date: Wed, 3 Mar 2021 07:43:50 -0700 Subject: [PATCH] Use ansible_facts It has come to our attention that using ansible_* vars that are populated with INJECT_FACTS_AS_VARS=True is not very performant. In order to be able to support setting that to off, we need to update the references to use ansible_facts[] instead of ansible_. Related: ansible#73654 Closes: https://bugzilla.redhat.com/show_bug.cgi?id=1935406 Signed-off-by: Alex Schultz (cherry picked from commit a7f2fa73e63e69dba2e41aaac9732397eec437c9) --- group_vars/all.yml.sample | 12 ++--- group_vars/iscsigws.yml.sample | 6 +-- group_vars/mdss.yml.sample | 4 +- group_vars/mgrs.yml.sample | 2 +- group_vars/mons.yml.sample | 2 +- group_vars/nfss.yml.sample | 4 +- group_vars/osds.yml.sample | 2 +- group_vars/rbdmirrors.yml.sample | 2 +- group_vars/rhcs.yml.sample | 12 ++--- .../filestore-to-bluestore.yml | 2 +- infrastructure-playbooks/purge-cluster.yml | 46 ++++++++-------- .../purge-container-cluster.yml | 30 +++++------ .../purge-iscsi-gateways.yml | 2 +- infrastructure-playbooks/rolling_update.yml | 54 +++++++++---------- infrastructure-playbooks/shrink-mds.yml | 4 +- infrastructure-playbooks/shrink-mgr.yml | 4 +- infrastructure-playbooks/shrink-mon.yml | 4 +- infrastructure-playbooks/shrink-osd.yml | 4 +- infrastructure-playbooks/shrink-rbdmirror.yml | 4 +- infrastructure-playbooks/shrink-rgw.yml | 4 +- ...inerized-to-containerized-ceph-daemons.yml | 18 +++---- .../untested-by-ci/cluster-os-migration.yml | 54 +++++++++---------- .../untested-by-ci/make-osd-partitions.yml | 8 +-- .../untested-by-ci/replace-osd.yml | 2 +- profiles/rgw-keystone-v2 | 4 +- profiles/rgw-keystone-v3 | 4 +- profiles/rgw-radosgw-static-website | 2 +- profiles/rgw-usage-log | 2 +- roles/ceph-client/tasks/create_users_keys.yml | 2 +- .../tasks/configure_cluster_name.yml | 4 +- .../tasks/configure_memory_allocator.yml | 4 +- ...nfigure_redhat_repository_installation.yml | 2 +- .../installs/debian_community_repository.yml | 2 +- .../installs/debian_custom_repository.yml | 2 +- .../tasks/installs/debian_dev_repository.yml | 2 +- .../installs/install_debian_packages.yml | 2 +- .../installs/install_redhat_packages.yml | 4 +- .../prerequisite_rhcs_cdn_install.yml | 10 ++-- .../installs/redhat_community_repository.yml | 6 +-- .../tasks/installs/redhat_dev_repository.yml | 2 +- roles/ceph-common/tasks/main.yml | 8 +-- .../tasks/rgw_systemd_environment_file.yml | 4 +- roles/ceph-config/templates/ceph.conf.j2 | 26 ++++----- .../pre_requisites/debian_prerequisites.yml | 6 +-- .../tasks/pre_requisites/prerequisites.yml | 8 +-- roles/ceph-crash/tasks/main.yml | 2 +- .../tasks/configure_dashboard.yml | 10 ++-- .../tasks/configure_dashboard_backends.yml | 6 +-- roles/ceph-dashboard/tasks/main.yml | 2 +- roles/ceph-defaults/defaults/main.yml | 14 ++--- roles/ceph-facts/tasks/container_binary.yml | 2 +- roles/ceph-facts/tasks/facts.yml | 40 +++++++------- roles/ceph-facts/tasks/grafana.yml | 8 +-- .../ceph-facts/tasks/set_monitor_address.yml | 8 +-- .../ceph-facts/tasks/set_radosgw_address.yml | 10 ++-- .../ceph-grafana/tasks/configure_grafana.yml | 4 +- roles/ceph-grafana/templates/grafana.ini.j2 | 2 +- .../tasks/check_running_containers.yml | 14 ++--- roles/ceph-handler/tasks/handler_crash.yml | 2 +- .../templates/restart_mds_daemon.sh.j2 | 8 +-- .../templates/restart_mgr_daemon.sh.j2 | 10 ++-- .../templates/restart_mon_daemon.sh.j2 | 14 ++--- .../templates/restart_nfs_daemon.sh.j2 | 6 +-- .../templates/restart_rbd_mirror_daemon.sh.j2 | 8 +-- .../templates/restart_rgw_daemon.sh.j2 | 2 +- roles/ceph-infra/tasks/configure_firewall.yml | 2 +- roles/ceph-infra/tasks/main.yml | 4 +- roles/ceph-iscsi-gw/defaults/main.yml | 6 +-- roles/ceph-iscsi-gw/tasks/common.yml | 6 +-- roles/ceph-iscsi-gw/tasks/deploy_ssl_keys.yml | 2 +- .../tasks/non-container/prerequisites.yml | 4 +- roles/ceph-mds/defaults/main.yml | 4 +- roles/ceph-mds/tasks/containerized.yml | 4 +- roles/ceph-mds/tasks/non_containerized.yml | 10 ++-- roles/ceph-mds/templates/ceph-mds.service.j2 | 12 ++--- roles/ceph-mgr/defaults/main.yml | 2 +- roles/ceph-mgr/tasks/common.yml | 16 +++--- roles/ceph-mgr/tasks/main.yml | 2 +- roles/ceph-mgr/tasks/pre_requisite.yml | 12 ++--- roles/ceph-mgr/tasks/start_mgr.yml | 6 +-- roles/ceph-mgr/templates/ceph-mgr.service.j2 | 10 ++-- roles/ceph-mon/defaults/main.yml | 2 +- roles/ceph-mon/tasks/ceph_keys.yml | 2 +- roles/ceph-mon/tasks/deploy_monitors.yml | 2 +- roles/ceph-mon/tasks/main.yml | 2 +- roles/ceph-mon/tasks/start_monitor.yml | 6 +-- roles/ceph-mon/templates/ceph-mon.service.j2 | 2 +- roles/ceph-nfs/defaults/main.yml | 4 +- roles/ceph-nfs/tasks/create_rgw_nfs_user.yml | 2 +- roles/ceph-nfs/tasks/ganesha_selinux_fix.yml | 4 +- roles/ceph-nfs/tasks/main.yml | 8 +-- .../tasks/pre_requisite_container.yml | 2 +- .../tasks/pre_requisite_non_container.yml | 14 ++--- .../pre_requisite_non_container_debian.yml | 4 +- .../pre_requisite_non_container_red_hat.yml | 2 +- roles/ceph-nfs/tasks/start_nfs.yml | 4 +- roles/ceph-nfs/templates/ceph-nfs.service.j2 | 4 +- roles/ceph-osd/defaults/main.yml | 2 +- roles/ceph-osd/tasks/main.yml | 4 +- roles/ceph-osd/tasks/start_osds.yml | 4 +- roles/ceph-osd/tasks/system_tuning.yml | 4 +- roles/ceph-osd/templates/ceph-osd.service.j2 | 4 +- .../templates/alertmanager.service.j2 | 2 +- .../templates/alertmanager.yml.j2 | 2 +- .../templates/prometheus.service.j2 | 2 +- .../templates/prometheus.yml.j2 | 6 +-- roles/ceph-rbd-mirror/defaults/main.yml | 2 +- roles/ceph-rbd-mirror/tasks/common.yml | 8 +-- .../tasks/configure_mirroring.yml | 6 +-- roles/ceph-rbd-mirror/tasks/main.yml | 2 +- .../tasks/start_container_rbd_mirror.yml | 2 +- .../tasks/start_rbd_mirror.yml | 6 +-- .../templates/ceph-rbd-mirror.service.j2 | 10 ++-- .../templates/haproxy.cfg.j2 | 2 +- .../templates/keepalived.conf.j2 | 4 +- roles/ceph-rgw/handlers/main.yml | 2 +- roles/ceph-rgw/tasks/openstack-keystone.yml | 4 +- roles/ceph-rgw/tasks/pre_requisite.yml | 4 +- roles/ceph-rgw/tasks/start_docker_rgw.yml | 2 +- roles/ceph-rgw/tasks/start_radosgw.yml | 2 +- .../templates/ceph-radosgw.service.j2 | 16 +++--- roles/ceph-validate/tasks/check_eth_mon.yml | 8 +-- roles/ceph-validate/tasks/check_eth_rgw.yml | 8 +-- .../ceph-validate/tasks/check_ipaddr_mon.yml | 2 +- roles/ceph-validate/tasks/check_iscsi.yml | 6 +-- roles/ceph-validate/tasks/check_nfs.yml | 2 +- roles/ceph-validate/tasks/check_system.yml | 32 +++++------ roles/ceph-validate/tasks/main.yml | 4 +- site-container.yml.sample | 4 +- .../add-osds/container/group_vars/all | 4 +- tests/functional/add-osds/group_vars/all | 4 +- .../add-rbdmirrors/container/group_vars/all | 4 +- .../functional/add-rbdmirrors/group_vars/all | 4 +- .../add-rgws/container/group_vars/all | 4 +- tests/functional/add-rgws/group_vars/all | 4 +- .../all-in-one/container/group_vars/all | 4 +- tests/functional/all-in-one/group_vars/all | 4 +- .../all_daemons/container/group_vars/all | 4 +- tests/functional/all_daemons/container/hosts | 2 +- .../all_daemons/container/hosts-ubuntu | 2 +- .../container/hosts-upgrade-to-octopus | 2 +- tests/functional/all_daemons/group_vars/all | 2 +- tests/functional/all_daemons/hosts | 2 +- .../all_daemons/hosts-switch-to-containers | 2 +- .../all_daemons/hosts-upgrade-to-octopus | 2 +- .../collocation/container/group_vars/all | 4 +- tests/functional/collocation/group_vars/all | 4 +- tests/functional/docker2podman/group_vars/all | 4 +- .../container/inventory/group_vars/all | 4 +- .../external_clients/inventory/group_vars/all | 4 +- .../container/group_vars/all | 2 +- .../filestore-to-bluestore/group_vars/all | 2 +- .../container/group_vars/all | 4 +- .../lvm-auto-discovery/group_vars/all | 4 +- .../lvm-batch/container/group_vars/all | 4 +- tests/functional/lvm-batch/group_vars/all | 4 +- .../lvm-osds/container/group_vars/all | 2 +- tests/functional/lvm-osds/group_vars/all | 2 +- tests/functional/podman/group_vars/all | 4 +- .../rgw-multisite/container/group_vars/all | 4 +- .../container/secondary/group_vars/all | 4 +- tests/functional/rgw-multisite/group_vars/all | 4 +- .../rgw-multisite/secondary/group_vars/all | 4 +- tests/functional/rhcs_setup.yml | 2 +- tests/functional/setup.yml | 4 +- .../shrink_mds/container/group_vars/all | 2 +- .../shrink_mgr/container/group_vars/all | 2 +- .../shrink_mon/container/group_vars/all | 2 +- tests/functional/shrink_mon/hosts | 2 +- .../shrink_mon/hosts-switch-to-containers | 2 +- tests/functional/shrink_mon/hosts-ubuntu | 2 +- .../shrink_osd/container/group_vars/all | 2 +- .../shrink_rbdmirror/container/group_vars/all | 2 +- .../shrink_rgw/container/group_vars/all | 4 +- tests/functional/shrink_rgw/group_vars/all | 4 +- 175 files changed, 521 insertions(+), 521 deletions(-) diff --git a/group_vars/all.yml.sample b/group_vars/all.yml.sample index 200e036f1..a72a7ef2f 100644 --- a/group_vars/all.yml.sample +++ b/group_vars/all.yml.sample @@ -80,7 +80,7 @@ dummy: #centos_package_dependencies: # - epel-release -# - "{{ 'python3-libselinux' if ansible_distribution_major_version | int >= 8 else 'libselinux-python' }}" +# - "{{ (ansible_facts['distribution_major_version'] is version('8', '>=')) | ternary('python3-libselinux', 'libselinux-python') }}" #redhat_package_dependencies: [] @@ -150,7 +150,7 @@ dummy: # Use the option below to specify your applicable package tree, eg. when using non-LTS Ubuntu versions # # for a list of available Debian distributions, visit http://download.ceph.com/debian-{{ ceph_stable_release }}/dists/ # for more info read: https://github.com/ceph/ceph-ansible/issues/305 -#ceph_stable_distro_source: "{{ ansible_distribution_release }}" +#ceph_stable_distro_source: "{{ ansible_facts['distribution_release'] }}" # REPOSITORY: RHCS VERSION RED HAT STORAGE (from 4.0) @@ -178,7 +178,7 @@ dummy: # #ceph_stable_repo_uca: "http://ubuntu-cloud.archive.canonical.com/ubuntu" #ceph_stable_openstack_release_uca: queens -#ceph_stable_release_uca: "{{ ansible_distribution_release }}-updates/{{ ceph_stable_openstack_release_uca }}" +#ceph_stable_release_uca: "{{ ansible_facts['distribution_release'] }}-updates/{{ ceph_stable_openstack_release_uca }}" # REPOSITORY: openSUSE OBS # @@ -188,7 +188,7 @@ dummy: # usually has newer Ceph releases than the normal distro repository. # # -#ceph_obs_repo: "https://download.opensuse.org/repositories/filesystems:/ceph:/{{ ceph_stable_release }}/openSUSE_Leap_{{ ansible_distribution_version }}/" +#ceph_obs_repo: "https://download.opensuse.org/repositories/filesystems:/ceph:/{{ ceph_stable_release }}/openSUSE_Leap_{{ ansible_facts['distribution_version'] }}/" # REPOSITORY: DEV # @@ -251,7 +251,7 @@ dummy: #ceph_conf_key_directory: /etc/ceph -#ceph_uid: "{{ '64045' if not containerized_deployment | bool and ansible_os_family == 'Debian' else '167' }}" +#ceph_uid: "{{ '64045' if not containerized_deployment | bool and ansible_facts['os_family'] == 'Debian' else '167' }}" # Permissions for keyring files in /etc/ceph #ceph_keyring_permissions: '0600' @@ -549,7 +549,7 @@ dummy: # global: # foo: 1234 # bar: 5678 -# "client.rgw.{{ hostvars[groups.get(rgw_group_name)[0]]['ansible_hostname'] }}": +# "client.rgw.{{ hostvars[groups.get(rgw_group_name)[0]]['ansible_facts']['hostname'] }}": # rgw_zone: zone1 # #ceph_conf_overrides: {} diff --git a/group_vars/iscsigws.yml.sample b/group_vars/iscsigws.yml.sample index 8fcfdfbb0..67c1bb26e 100644 --- a/group_vars/iscsigws.yml.sample +++ b/group_vars/iscsigws.yml.sample @@ -43,14 +43,14 @@ dummy: # These options can be passed using the 'ceph_mds_docker_extra_env' variable. # TCMU_RUNNER resource limitation -#ceph_tcmu_runner_docker_memory_limit: "{{ ansible_memtotal_mb }}m" +#ceph_tcmu_runner_docker_memory_limit: "{{ ansible_facts['memtotal_mb'] }}m" #ceph_tcmu_runner_docker_cpu_limit: 1 # RBD_TARGET_GW resource limitation -#ceph_rbd_target_gw_docker_memory_limit: "{{ ansible_memtotal_mb }}m" +#ceph_rbd_target_gw_docker_memory_limit: "{{ ansible_facts['memtotal_mb'] }}m" #ceph_rbd_target_gw_docker_cpu_limit: 1 # RBD_TARGET_API resource limitation -#ceph_rbd_target_api_docker_memory_limit: "{{ ansible_memtotal_mb }}m" +#ceph_rbd_target_api_docker_memory_limit: "{{ ansible_facts['memtotal_mb'] }}m" #ceph_rbd_target_api_docker_cpu_limit: 1 diff --git a/group_vars/mdss.yml.sample b/group_vars/mdss.yml.sample index 2412ea730..14b1bfb8a 100644 --- a/group_vars/mdss.yml.sample +++ b/group_vars/mdss.yml.sample @@ -27,13 +27,13 @@ dummy: # For the whole list of limits you can apply see: docs.docker.com/engine/admin/resource_constraints # Default values are based from: https://access.redhat.com/documentation/en-us/red_hat_ceph_storage/2/html/red_hat_ceph_storage_hardware_guide/minimum_recommendations # These options can be passed using the 'ceph_mds_docker_extra_env' variable. -#ceph_mds_docker_memory_limit: "{{ ansible_memtotal_mb }}m" +#ceph_mds_docker_memory_limit: "{{ ansible_facts['memtotal_mb'] }}m" #ceph_mds_docker_cpu_limit: 4 # we currently for MDS_NAME to hostname because of a bug in ceph-docker # fix here: https://github.com/ceph/ceph-docker/pull/770 # this will go away soon. -#ceph_mds_docker_extra_env: -e MDS_NAME={{ ansible_hostname }} +#ceph_mds_docker_extra_env: -e MDS_NAME={{ ansible_facts['hostname'] }} #ceph_config_keys: [] # DON'T TOUCH ME diff --git a/group_vars/mgrs.yml.sample b/group_vars/mgrs.yml.sample index ab0d11164..55a01f41d 100644 --- a/group_vars/mgrs.yml.sample +++ b/group_vars/mgrs.yml.sample @@ -43,7 +43,7 @@ dummy: # For the whole list of limits you can apply see: docs.docker.com/engine/admin/resource_constraints # Default values are based from: https://access.redhat.com/documentation/en-us/red_hat_ceph_storage/2/html/red_hat_ceph_storage_hardware_guide/minimum_recommendations # These options can be passed using the 'ceph_mgr_docker_extra_env' variable. -#ceph_mgr_docker_memory_limit: "{{ ansible_memtotal_mb }}m" +#ceph_mgr_docker_memory_limit: "{{ ansible_facts['memtotal_mb'] }}m" #ceph_mgr_docker_cpu_limit: 1 #ceph_mgr_docker_extra_env: diff --git a/group_vars/mons.yml.sample b/group_vars/mons.yml.sample index 002d62b22..ad59172b7 100644 --- a/group_vars/mons.yml.sample +++ b/group_vars/mons.yml.sample @@ -45,7 +45,7 @@ dummy: # For the whole list of limits you can apply see: docs.docker.com/engine/admin/resource_constraints # Default values are based from: https://access.redhat.com/documentation/en-us/red_hat_ceph_storage/2/html/red_hat_ceph_storage_hardware_guide/minimum_recommendations # These options can be passed using the 'ceph_mon_docker_extra_env' variable. -#ceph_mon_docker_memory_limit: "{{ ansible_memtotal_mb }}m" +#ceph_mon_docker_memory_limit: "{{ ansible_facts['memtotal_mb'] }}m" #ceph_mon_docker_cpu_limit: 1 #ceph_mon_container_listen_port: 3300 diff --git a/group_vars/nfss.yml.sample b/group_vars/nfss.yml.sample index e3820bdb7..2abf4554d 100644 --- a/group_vars/nfss.yml.sample +++ b/group_vars/nfss.yml.sample @@ -25,7 +25,7 @@ dummy: #ceph_nfs_enable_service: true # ceph-nfs systemd service uses ansible's hostname as an instance id, -# so service name is ceph-nfs@{{ ansible_hostname }}, this is not +# so service name is ceph-nfs@{{ ansible_facts['hostname'] }}, this is not # ideal when ceph-nfs is managed by pacemaker across multiple hosts - in # such case it's better to have constant instance id instead which # can be set by 'ceph_nfs_service_suffix' @@ -95,7 +95,7 @@ dummy: # they must be configered. #ceph_nfs_rgw_access_key: "QFAMEDSJP5DEKJO0DDXY" #ceph_nfs_rgw_secret_key: "iaSFLDVvDdQt6lkNzHyW4fPLZugBAI1g17LO0+87[MAC[M#C" -#rgw_client_name: client.rgw.{{ ansible_hostname }} +#rgw_client_name: client.rgw.{{ ansible_facts['hostname'] }} ################### # CONFIG OVERRIDE # diff --git a/group_vars/osds.yml.sample b/group_vars/osds.yml.sample index 60b2bfbd2..6e418a7ce 100644 --- a/group_vars/osds.yml.sample +++ b/group_vars/osds.yml.sample @@ -169,7 +169,7 @@ dummy: # For the whole list of limits you can apply see: docs.docker.com/engine/admin/resource_constraints # Default values are based from: https://access.redhat.com/documentation/en-us/red_hat_ceph_storage/2/html/red_hat_ceph_storage_hardware_guide/minimum_recommendations # These options can be passed using the 'ceph_osd_docker_extra_env' variable. -#ceph_osd_docker_memory_limit: "{{ ansible_memtotal_mb }}m" +#ceph_osd_docker_memory_limit: "{{ ansible_facts['memtotal_mb'] }}m" #ceph_osd_docker_cpu_limit: 4 # The next two variables are undefined, and thus, unused by default. diff --git a/group_vars/rbdmirrors.yml.sample b/group_vars/rbdmirrors.yml.sample index 2c565fe01..009841361 100644 --- a/group_vars/rbdmirrors.yml.sample +++ b/group_vars/rbdmirrors.yml.sample @@ -50,7 +50,7 @@ dummy: # For the whole list of limits you can apply see: docs.docker.com/engine/admin/resource_constraints # Default values are based from: https://access.redhat.com/documentation/en-us/red_hat_ceph_storage/2/html/red_hat_ceph_storage_hardware_guide/minimum_recommendations # These options can be passed using the 'ceph_rbd_mirror_docker_extra_env' variable. -#ceph_rbd_mirror_docker_memory_limit: "{{ ansible_memtotal_mb }}m" +#ceph_rbd_mirror_docker_memory_limit: "{{ ansible_facts['memtotal_mb'] }}m" #ceph_rbd_mirror_docker_cpu_limit: 1 #ceph_rbd_mirror_docker_extra_env: diff --git a/group_vars/rhcs.yml.sample b/group_vars/rhcs.yml.sample index 8b52434e3..3d5f36ecc 100644 --- a/group_vars/rhcs.yml.sample +++ b/group_vars/rhcs.yml.sample @@ -80,7 +80,7 @@ dummy: #centos_package_dependencies: # - epel-release -# - "{{ 'python3-libselinux' if ansible_distribution_major_version | int >= 8 else 'libselinux-python' }}" +# - "{{ (ansible_facts['distribution_major_version'] is version('8', '>=')) | ternary('python3-libselinux', 'libselinux-python') }}" #redhat_package_dependencies: [] @@ -150,7 +150,7 @@ ceph_repository: rhcs # Use the option below to specify your applicable package tree, eg. when using non-LTS Ubuntu versions # # for a list of available Debian distributions, visit http://download.ceph.com/debian-{{ ceph_stable_release }}/dists/ # for more info read: https://github.com/ceph/ceph-ansible/issues/305 -#ceph_stable_distro_source: "{{ ansible_distribution_release }}" +#ceph_stable_distro_source: "{{ ansible_facts['distribution_release'] }}" # REPOSITORY: RHCS VERSION RED HAT STORAGE (from 4.0) @@ -178,7 +178,7 @@ ceph_rhcs_version: 4 # #ceph_stable_repo_uca: "http://ubuntu-cloud.archive.canonical.com/ubuntu" #ceph_stable_openstack_release_uca: queens -#ceph_stable_release_uca: "{{ ansible_distribution_release }}-updates/{{ ceph_stable_openstack_release_uca }}" +#ceph_stable_release_uca: "{{ ansible_facts['distribution_release'] }}-updates/{{ ceph_stable_openstack_release_uca }}" # REPOSITORY: openSUSE OBS # @@ -188,7 +188,7 @@ ceph_rhcs_version: 4 # usually has newer Ceph releases than the normal distro repository. # # -#ceph_obs_repo: "https://download.opensuse.org/repositories/filesystems:/ceph:/{{ ceph_stable_release }}/openSUSE_Leap_{{ ansible_distribution_version }}/" +#ceph_obs_repo: "https://download.opensuse.org/repositories/filesystems:/ceph:/{{ ceph_stable_release }}/openSUSE_Leap_{{ ansible_facts['distribution_version'] }}/" # REPOSITORY: DEV # @@ -251,7 +251,7 @@ ceph_iscsi_config_dev: false #ceph_conf_key_directory: /etc/ceph -#ceph_uid: "{{ '64045' if not containerized_deployment | bool and ansible_os_family == 'Debian' else '167' }}" +#ceph_uid: "{{ '64045' if not containerized_deployment | bool and ansible_facts['os_family'] == 'Debian' else '167' }}" # Permissions for keyring files in /etc/ceph #ceph_keyring_permissions: '0600' @@ -549,7 +549,7 @@ ceph_iscsi_config_dev: false # global: # foo: 1234 # bar: 5678 -# "client.rgw.{{ hostvars[groups.get(rgw_group_name)[0]]['ansible_hostname'] }}": +# "client.rgw.{{ hostvars[groups.get(rgw_group_name)[0]]['ansible_facts']['hostname'] }}": # rgw_zone: zone1 # #ceph_conf_overrides: {} diff --git a/infrastructure-playbooks/filestore-to-bluestore.yml b/infrastructure-playbooks/filestore-to-bluestore.yml index b3be2eaec..149de7a8c 100644 --- a/infrastructure-playbooks/filestore-to-bluestore.yml +++ b/infrastructure-playbooks/filestore-to-bluestore.yml @@ -36,7 +36,7 @@ - name: set_fact container_run_cmd, container_exec_cmd set_fact: container_run_cmd: "{{ container_binary + ' run --rm --privileged=true --net=host --pid=host --ipc=host -v /dev:/dev -v /etc/ceph:/etc/ceph -v /var/lib/ceph:/var/lib/ceph -v /var/run:/var/run --entrypoint=' if containerized_deployment | bool else '' }}ceph-volume {{ ceph_docker_registry + '/' + ceph_docker_image + ':' + ceph_docker_image_tag if containerized_deployment | bool else '' }}" - container_exec_cmd: "{{ container_binary + ' exec ceph-mon-' + hostvars[groups[mon_group_name][0]]['ansible_hostname'] if containerized_deployment | bool else '' }}" + container_exec_cmd: "{{ container_binary + ' exec ceph-mon-' + hostvars[groups[mon_group_name][0]]['ansible_facts']['hostname'] if containerized_deployment | bool else '' }}" - name: get ceph osd tree data command: "{{ container_exec_cmd }} ceph --cluster {{ cluster }} osd tree -f json" diff --git a/infrastructure-playbooks/purge-cluster.yml b/infrastructure-playbooks/purge-cluster.yml index b29dd93a9..5bed86875 100644 --- a/infrastructure-playbooks/purge-cluster.yml +++ b/infrastructure-playbooks/purge-cluster.yml @@ -71,7 +71,7 @@ run_once: true - name: get all nfs-ganesha mount points - command: grep "{{ hostvars[item]['ansible_all_ipv4_addresses'] | ips_in_ranges(public_network.split(',')) | first }}" /proc/mounts + command: grep "{{ hostvars[item]['ansible_facts']['all_ipv4_addresses'] | ips_in_ranges(public_network.split(',')) | first }}" /proc/mounts register: nfs_ganesha_mount_points failed_when: false with_items: "{{ groups[nfs_group_name] }}" @@ -127,7 +127,7 @@ name: nfs-ganesha state: stopped failed_when: false - when: ansible_service_mgr == 'systemd' + when: ansible_facts['service_mgr'] == 'systemd' - name: purge node-exporter hosts: @@ -249,7 +249,7 @@ - name: stop ceph mdss with systemd service: - name: ceph-mds@{{ ansible_hostname }} + name: ceph-mds@{{ ansible_facts['hostname'] }} state: stopped enabled: no failed_when: false @@ -270,11 +270,11 @@ - name: stop ceph mgrs with systemd service: - name: ceph-mgr@{{ ansible_hostname }} + name: ceph-mgr@{{ ansible_facts['hostname'] }} state: stopped enabled: no failed_when: false - when: ansible_service_mgr == 'systemd' + when: ansible_facts['service_mgr'] == 'systemd' - name: purge rgwloadbalancer cluster @@ -318,7 +318,7 @@ - name: stop ceph rgws with systemd service: - name: "ceph-radosgw@rgw.{{ ansible_hostname }}.{{ item.instance_name }}" + name: "ceph-radosgw@rgw.{{ ansible_facts['hostname'] }}.{{ item.instance_name }}" state: stopped enabled: no failed_when: false @@ -340,7 +340,7 @@ - name: stop ceph rbd mirror with systemd service: - name: "ceph-rbd-mirror@rbd-mirror.{{ ansible_hostname }}" + name: "ceph-rbd-mirror@rbd-mirror.{{ ansible_facts['hostname'] }}" state: stopped failed_when: false @@ -368,7 +368,7 @@ become: false wait_for: port: 22 - host: "{{ hostvars[inventory_hostname]['ansible_default_ipv4']['address'] }}" + host: "{{ hostvars[inventory_hostname]['ansible_facts']['default_ipv4']['address'] }}" state: started delay: 10 timeout: 500 @@ -398,7 +398,7 @@ state: stopped enabled: no with_items: "{{ osd_ids.stdout_lines }}" - when: ansible_service_mgr == 'systemd' + when: ansible_facts['service_mgr'] == 'systemd' - name: remove ceph udev rules file: @@ -613,7 +613,7 @@ - name: stop ceph mons with systemd service: - name: "ceph-{{ item }}@{{ ansible_hostname }}" + name: "ceph-{{ item }}@{{ ansible_facts['hostname'] }}" state: stopped enabled: no failed_when: false @@ -732,27 +732,27 @@ yum: name: "{{ ceph_packages }}" state: absent - when: ansible_pkg_mgr == 'yum' + when: ansible_facts['pkg_mgr'] == 'yum' - name: purge ceph packages with dnf dnf: name: "{{ ceph_packages }}" state: absent - when: ansible_pkg_mgr == 'dnf' + when: ansible_facts['pkg_mgr'] == 'dnf' - name: purge ceph packages with apt apt: name: "{{ ceph_packages }}" state: absent purge: true - when: ansible_pkg_mgr == 'apt' + when: ansible_facts['pkg_mgr'] == 'apt' - name: purge remaining ceph packages with yum yum: name: "{{ ceph_remaining_packages }}" state: absent when: - - ansible_pkg_mgr == 'yum' + - ansible_facts['pkg_mgr'] == 'yum' - purge_all_packages | bool - name: purge remaining ceph packages with dnf @@ -760,7 +760,7 @@ name: "{{ ceph_remaining_packages }}" state: absent when: - - ansible_pkg_mgr == 'dnf' + - ansible_facts['pkg_mgr'] == 'dnf' - purge_all_packages | bool - name: purge remaining ceph packages with apt @@ -768,7 +768,7 @@ name: "{{ ceph_remaining_packages }}" state: absent when: - - ansible_pkg_mgr == 'apt' + - ansible_facts['pkg_mgr'] == 'apt' - purge_all_packages | bool - name: purge extra packages with yum @@ -776,7 +776,7 @@ name: "{{ extra_packages }}" state: absent when: - - ansible_pkg_mgr == 'yum' + - ansible_facts['pkg_mgr'] == 'yum' - purge_all_packages | bool - name: purge extra packages with dnf @@ -784,7 +784,7 @@ name: "{{ extra_packages }}" state: absent when: - - ansible_pkg_mgr == 'dnf' + - ansible_facts['pkg_mgr'] == 'dnf' - purge_all_packages | bool - name: purge extra packages with apt @@ -792,7 +792,7 @@ name: "{{ extra_packages }}" state: absent when: - - ansible_pkg_mgr == 'apt' + - ansible_facts['pkg_mgr'] == 'apt' - purge_all_packages | bool - name: remove config and any ceph socket left @@ -818,7 +818,7 @@ - name: purge dnf cache command: dnf clean all - when: ansible_pkg_mgr == 'dnf' + when: ansible_facts['pkg_mgr'] == 'dnf' - name: purge rpm cache in /tmp file: @@ -827,7 +827,7 @@ - name: clean apt command: apt-get clean # noqa 303 - when: ansible_pkg_mgr == 'apt' + when: ansible_facts['pkg_mgr'] == 'apt' - name: purge ceph repo file in /etc/yum.repos.d file: @@ -837,7 +837,7 @@ - ceph-dev - ceph_stable - rh_storage - when: ansible_os_family == 'RedHat' + when: ansible_facts['os_family'] == 'RedHat' - name: check for anything running ceph command: "ps -u ceph -U ceph" @@ -858,7 +858,7 @@ path: "{{ item.path }}" state: absent with_items: "{{ systemd_files.files }}" - when: ansible_service_mgr == 'systemd' + when: ansible_facts['service_mgr'] == 'systemd' - name: purge fetch directory diff --git a/infrastructure-playbooks/purge-container-cluster.yml b/infrastructure-playbooks/purge-container-cluster.yml index 358cee783..0935dd2ea 100644 --- a/infrastructure-playbooks/purge-container-cluster.yml +++ b/infrastructure-playbooks/purge-container-cluster.yml @@ -55,7 +55,7 @@ run_once: true - name: get all nfs-ganesha mount points - command: grep "{{ hostvars[item]['ansible_all_ipv4_addresses'] | ips_in_ranges(public_network.split(',')) | first }}" /proc/mounts + command: grep "{{ hostvars[item]['ansible_facts']['all_ipv4_addresses'] | ips_in_ranges(public_network.split(',')) | first }}" /proc/mounts register: nfs_ganesha_mount_points failed_when: false with_items: "{{ groups[nfs_group_name] }}" @@ -104,7 +104,7 @@ - name: disable ceph nfs service service: - name: "ceph-nfs@{{ ansible_hostname }}" + name: "ceph-nfs@{{ ansible_facts['hostname'] }}" state: stopped enabled: no ignore_errors: true @@ -114,7 +114,7 @@ path: /etc/systemd/system/ceph-nfs@.service state: absent - - name: remove ceph nfs directories for "{{ ansible_hostname }}" + - name: remove ceph nfs directories for "{{ ansible_facts['hostname'] }}" file: path: "{{ item }}" state: absent @@ -134,7 +134,7 @@ - name: disable ceph mds service service: - name: "ceph-mds@{{ ansible_hostname }}" + name: "ceph-mds@{{ ansible_facts['hostname'] }}" state: stopped enabled: no ignore_errors: true @@ -181,7 +181,7 @@ - name: disable ceph mgr service service: - name: "ceph-mgr@{{ ansible_hostname }}" + name: "ceph-mgr@{{ ansible_facts['hostname'] }}" state: stopped enabled: no ignore_errors: true @@ -208,7 +208,7 @@ - name: disable ceph rgw service service: - name: "ceph-radosgw@rgw.{{ ansible_hostname }}.{{ item.instance_name }}" + name: "ceph-radosgw@rgw.{{ ansible_facts['hostname'] }}.{{ item.instance_name }}" state: stopped enabled: no failed_when: false @@ -230,7 +230,7 @@ - name: disable ceph rbd-mirror service service: - name: "ceph-rbd-mirror@rbd-mirror.{{ ansible_hostname }}" + name: "ceph-rbd-mirror@rbd-mirror.{{ ansible_facts['hostname'] }}" state: stopped enabled: no ignore_errors: true @@ -352,8 +352,8 @@ enabled: no ignore_errors: true with_items: - - "ceph-mgr@{{ ansible_hostname }}" - - "ceph-mon@{{ ansible_hostname }}" + - "ceph-mgr@{{ ansible_facts['hostname'] }}" + - "ceph-mon@{{ ansible_facts['hostname'] }}" - name: remove ceph mon and mgr service file: @@ -486,7 +486,7 @@ tasks: - name: stop ceph-crash container service: - name: "ceph-crash@{{ ansible_hostname }}" + name: "ceph-crash@{{ ansible_facts['hostname'] }}" state: stopped enabled: no failed_when: false @@ -595,7 +595,7 @@ state: absent update_cache: yes autoremove: yes - when: ansible_os_family == 'Debian' + when: ansible_facts['os_family'] == 'Debian' - name: red hat based systems tasks block: @@ -616,7 +616,7 @@ args: warn: no when: - ansible_pkg_mgr == "yum" + ansible_facts['pkg_mgr'] == "yum" - name: dnf related tasks on red hat block: @@ -635,9 +635,9 @@ args: warn: no when: - ansible_pkg_mgr == "dnf" + ansible_facts['pkg_mgr'] == "dnf" when: - ansible_os_family == 'RedHat' and + ansible_facts['os_family'] == 'RedHat' and not is_atomic - name: find any service-cid file left @@ -677,7 +677,7 @@ become: true tasks: - - name: purge ceph directories for "{{ ansible_hostname }}" and ceph socket + - name: purge ceph directories for "{{ ansible_facts['hostname'] }}" and ceph socket file: path: "{{ item }}" state: absent diff --git a/infrastructure-playbooks/purge-iscsi-gateways.yml b/infrastructure-playbooks/purge-iscsi-gateways.yml index 7c2482bd2..ec2247d2f 100644 --- a/infrastructure-playbooks/purge-iscsi-gateways.yml +++ b/infrastructure-playbooks/purge-iscsi-gateways.yml @@ -83,7 +83,7 @@ - name: set_fact container_exec_cmd set_fact: - container_exec_cmd: "{{ container_binary }} exec ceph-mon-{{ ansible_hostname }}" + container_exec_cmd: "{{ container_binary }} exec ceph-mon-{{ ansible_facts['hostname'] }}" when: containerized_deployment | bool - name: get iscsi gateway list diff --git a/infrastructure-playbooks/rolling_update.yml b/infrastructure-playbooks/rolling_update.yml index c84397099..5627b5af4 100644 --- a/infrastructure-playbooks/rolling_update.yml +++ b/infrastructure-playbooks/rolling_update.yml @@ -194,7 +194,7 @@ # after the package gets upgraded - name: stop ceph mon - shortname systemd: - name: ceph-mon@{{ ansible_hostname }} + name: ceph-mon@{{ ansible_facts['hostname'] }} state: stopped enabled: no masked: yes @@ -204,7 +204,7 @@ # after the package gets upgraded - name: stop ceph mon - fqdn systemd: - name: ceph-mon@{{ ansible_fqdn }} + name: ceph-mon@{{ ansible_facts['fqdn'] }} state: stopped enabled: no masked: yes @@ -214,7 +214,7 @@ # after ALL monitors, even when collocated - name: mask the mgr service systemd: - name: ceph-mgr@{{ ansible_hostname }} + name: ceph-mgr@{{ ansible_facts['hostname'] }} masked: yes when: inventory_hostname in groups[mgr_group_name] | default([]) or groups[mgr_group_name] | default([]) | length == 0 @@ -234,7 +234,7 @@ - name: start ceph mgr systemd: - name: ceph-mgr@{{ ansible_hostname }} + name: ceph-mgr@{{ ansible_facts['hostname'] }} state: started enabled: yes ignore_errors: True # if no mgr collocated with mons @@ -244,20 +244,20 @@ register: ceph_health_raw until: - ceph_health_raw.rc == 0 - - (hostvars[inventory_hostname]['ansible_hostname'] in (ceph_health_raw.stdout | default('{}') | from_json)["quorum_names"] or - hostvars[inventory_hostname]['ansible_fqdn'] in (ceph_health_raw.stdout | default('{}') | from_json)["quorum_names"]) + - (hostvars[inventory_hostname]['ansible_facts']['hostname'] in (ceph_health_raw.stdout | default('{}') | from_json)["quorum_names"] or + hostvars[inventory_hostname]['ansible_facts']['fqdn'] in (ceph_health_raw.stdout | default('{}') | from_json)["quorum_names"]) retries: "{{ health_mon_check_retries }}" delay: "{{ health_mon_check_delay }}" when: not containerized_deployment | bool - name: container | waiting for the containerized monitor to join the quorum... command: > - {{ container_binary }} exec ceph-mon-{{ ansible_hostname }} ceph --cluster "{{ cluster }}" -m "{{ hostvars[groups[mon_group_name][0]]['_current_monitor_address'] }}" quorum_status --format json + {{ container_binary }} exec ceph-mon-{{ ansible_facts['hostname'] }} ceph --cluster "{{ cluster }}" -m "{{ hostvars[groups[mon_group_name][0]]['_current_monitor_address'] }}" quorum_status --format json register: ceph_health_raw until: - ceph_health_raw.rc == 0 - - (hostvars[inventory_hostname]['ansible_hostname'] in (ceph_health_raw.stdout | default('{}') | from_json)["quorum_names"] or - hostvars[inventory_hostname]['ansible_fqdn'] in (ceph_health_raw.stdout | default('{}') | from_json)["quorum_names"]) + - (hostvars[inventory_hostname]['ansible_facts']['hostname'] in (ceph_health_raw.stdout | default('{}') | from_json)["quorum_names"] or + hostvars[inventory_hostname]['ansible_facts']['fqdn'] in (ceph_health_raw.stdout | default('{}') | from_json)["quorum_names"]) retries: "{{ health_mon_check_retries }}" delay: "{{ health_mon_check_delay }}" when: containerized_deployment | bool @@ -289,7 +289,7 @@ block: - name: stop ceph mgr systemd: - name: ceph-mgr@{{ ansible_hostname }} + name: ceph-mgr@{{ ansible_facts['hostname'] }} state: stopped masked: yes @@ -323,7 +323,7 @@ # or if we run a Ceph cluster before Luminous - name: stop ceph mgr systemd: - name: ceph-mgr@{{ ansible_hostname }} + name: ceph-mgr@{{ ansible_facts['hostname'] }} state: stopped enabled: no masked: yes @@ -389,7 +389,7 @@ - name: set_fact container_exec_cmd_osd set_fact: - container_exec_cmd_update_osd: "{{ container_binary }} exec ceph-mon-{{ hostvars[groups[mon_group_name][0]]['ansible_hostname'] }}" + container_exec_cmd_update_osd: "{{ container_binary }} exec ceph-mon-{{ hostvars[groups[mon_group_name][0]]['ansible_facts']['hostname'] }}" when: containerized_deployment | bool - name: stop ceph osd @@ -477,7 +477,7 @@ - name: set_fact container_exec_cmd_osd set_fact: - container_exec_cmd_update_osd: "{{ container_binary }} exec ceph-mon-{{ ansible_hostname }}" + container_exec_cmd_update_osd: "{{ container_binary }} exec ceph-mon-{{ ansible_facts['hostname'] }}" when: containerized_deployment | bool - name: get osd versions @@ -539,7 +539,7 @@ set_fact: mds_active_host: "{{ [hostvars[item]['inventory_hostname']] }}" with_items: "{{ groups[mds_group_name] }}" - when: hostvars[item]['ansible_hostname'] == mds_active_name + when: hostvars[item]['ansible_facts']['hostname'] == mds_active_name - name: create standby_mdss group add_host: @@ -551,7 +551,7 @@ - name: stop standby ceph mds systemd: - name: "ceph-mds@{{ hostvars[item]['ansible_hostname'] }}" + name: "ceph-mds@{{ hostvars[item]['ansible_facts']['hostname'] }}" state: stopped enabled: no delegate_to: "{{ item }}" @@ -562,7 +562,7 @@ # somehow, having a single task doesn't work in containerized context - name: mask systemd units for standby ceph mds systemd: - name: "ceph-mds@{{ hostvars[item]['ansible_hostname'] }}" + name: "ceph-mds@{{ hostvars[item]['ansible_facts']['hostname'] }}" masked: yes delegate_to: "{{ item }}" with_items: "{{ groups['standby_mdss'] }}" @@ -598,7 +598,7 @@ - name: prevent restart from the packaging systemd: - name: ceph-mds@{{ ansible_hostname }} + name: ceph-mds@{{ ansible_facts['hostname'] }} enabled: no masked: yes when: not containerized_deployment | bool @@ -618,14 +618,14 @@ - name: restart ceph mds systemd: - name: ceph-mds@{{ ansible_hostname }} + name: ceph-mds@{{ ansible_facts['hostname'] }} state: restarted enabled: yes masked: no when: not containerized_deployment | bool - name: restart active mds - command: "{{ container_binary }} stop ceph-mds-{{ ansible_hostname }}" + command: "{{ container_binary }} stop ceph-mds-{{ ansible_facts['hostname'] }}" changed_when: false when: containerized_deployment | bool @@ -644,7 +644,7 @@ - name: prevent restarts from the packaging systemd: - name: ceph-mds@{{ ansible_hostname }} + name: ceph-mds@{{ ansible_facts['hostname'] }} enabled: no masked: yes when: not containerized_deployment | bool @@ -684,7 +684,7 @@ - name: stop ceph rgw when upgrading from stable-3.2 systemd: - name: ceph-radosgw@rgw.{{ ansible_hostname }} + name: ceph-radosgw@rgw.{{ ansible_facts['hostname'] }} state: stopped enabled: no masked: yes @@ -692,7 +692,7 @@ - name: stop ceph rgw systemd: - name: ceph-radosgw@rgw.{{ ansible_hostname }}.{{ item.instance_name }} + name: ceph-radosgw@rgw.{{ ansible_facts['hostname'] }}.{{ item.instance_name }} state: stopped enabled: no masked: yes @@ -721,7 +721,7 @@ tasks: - name: stop ceph rbd mirror systemd: - name: "ceph-rbd-mirror@rbd-mirror.{{ ansible_hostname }}" + name: "ceph-rbd-mirror@rbd-mirror.{{ ansible_facts['hostname'] }}" state: stopped enabled: no masked: yes @@ -765,7 +765,7 @@ - name: systemd stop nfs container systemd: - name: ceph-nfs@{{ ceph_nfs_service_suffix | default(ansible_hostname) }} + name: ceph-nfs@{{ ceph_nfs_service_suffix | default(ansible_facts['hostname']) }} state: stopped enabled: no masked: yes @@ -899,7 +899,7 @@ name: ceph-facts - name: container | disallow pre-nautilus OSDs and enable all new nautilus-only functionality - command: "{{ container_binary }} exec ceph-mon-{{ hostvars[groups[mon_group_name][0]]['ansible_hostname'] }} ceph --cluster {{ cluster }} osd require-osd-release nautilus" + command: "{{ container_binary }} exec ceph-mon-{{ hostvars[groups[mon_group_name][0]]['ansible_facts']['hostname'] }} ceph --cluster {{ cluster }} osd require-osd-release nautilus" delegate_to: "{{ groups[mon_group_name][0] }}" run_once: True when: @@ -915,7 +915,7 @@ - groups.get(mon_group_name, []) | length > 0 - name: container | enable msgr2 protocol - command: "{{ container_binary }} exec ceph-mon-{{ hostvars[groups[mon_group_name][0]]['ansible_hostname'] }} ceph --cluster {{ cluster }} mon enable-msgr2" + command: "{{ container_binary }} exec ceph-mon-{{ hostvars[groups[mon_group_name][0]]['ansible_facts']['hostname'] }} ceph --cluster {{ cluster }} mon enable-msgr2" delegate_to: "{{ groups[mon_group_name][0] }}" run_once: True when: @@ -1036,7 +1036,7 @@ - name: set_fact container_exec_cmd_status set_fact: - container_exec_cmd_status: "{{ container_binary }} exec ceph-mon-{{ hostvars[groups[mon_group_name][0]]['ansible_hostname'] }}" + container_exec_cmd_status: "{{ container_binary }} exec ceph-mon-{{ hostvars[groups[mon_group_name][0]]['ansible_facts']['hostname'] }}" when: containerized_deployment | bool - name: show ceph status diff --git a/infrastructure-playbooks/shrink-mds.yml b/infrastructure-playbooks/shrink-mds.yml index 3441e1960..afbcd0df3 100644 --- a/infrastructure-playbooks/shrink-mds.yml +++ b/infrastructure-playbooks/shrink-mds.yml @@ -61,7 +61,7 @@ - name: set_fact container_exec_cmd for mon0 set_fact: - container_exec_cmd: "{{ container_binary }} exec ceph-mon-{{ ansible_hostname }}" + container_exec_cmd: "{{ container_binary }} exec ceph-mon-{{ ansible_facts['hostname'] }}" when: containerized_deployment | bool - name: exit playbook, if can not connect to the cluster @@ -74,7 +74,7 @@ - name: set_fact mds_to_kill_hostname set_fact: - mds_to_kill_hostname: "{{ hostvars[mds_to_kill]['ansible_hostname'] }}" + mds_to_kill_hostname: "{{ hostvars[mds_to_kill]['ansible_facts']['hostname'] }}" tasks: # get rid of this as soon as "systemctl stop ceph-msd@$HOSTNAME" also diff --git a/infrastructure-playbooks/shrink-mgr.yml b/infrastructure-playbooks/shrink-mgr.yml index 02de185ec..f99726890 100644 --- a/infrastructure-playbooks/shrink-mgr.yml +++ b/infrastructure-playbooks/shrink-mgr.yml @@ -39,7 +39,7 @@ - name: set_fact container_exec_cmd when: containerized_deployment | bool set_fact: - container_exec_cmd: "{{ container_binary }} exec ceph-mon-{{ ansible_hostname }}" + container_exec_cmd: "{{ container_binary }} exec ceph-mon-{{ ansible_facts['hostname'] }}" - name: exit playbook, if can not connect to the cluster command: "{{ container_exec_cmd | default('') }} timeout 5 ceph --cluster {{ cluster }} health" @@ -92,7 +92,7 @@ - name: set_fact mgr_to_kill_hostname set_fact: - mgr_to_kill_hostname: "{{ hostvars[mgr_to_kill]['ansible_hostname'] }}" + mgr_to_kill_hostname: "{{ hostvars[mgr_to_kill]['ansible_facts']['hostname'] }}" tasks: - name: stop manager services and verify it diff --git a/infrastructure-playbooks/shrink-mon.yml b/infrastructure-playbooks/shrink-mon.yml index df2ae60bd..05d6c2be5 100644 --- a/infrastructure-playbooks/shrink-mon.yml +++ b/infrastructure-playbooks/shrink-mon.yml @@ -76,7 +76,7 @@ - name: "set_fact container_exec_cmd build {{ container_binary }} exec command (containerized)" set_fact: - container_exec_cmd: "{{ container_binary }} exec ceph-mon-{{ hostvars[mon_host]['ansible_hostname'] }}" + container_exec_cmd: "{{ container_binary }} exec ceph-mon-{{ hostvars[mon_host]['ansible_facts']['hostname'] }}" when: containerized_deployment | bool - name: exit playbook, if can not connect to the cluster @@ -90,7 +90,7 @@ - name: set_fact mon_to_kill_hostname set_fact: - mon_to_kill_hostname: "{{ hostvars[mon_to_kill]['ansible_hostname'] }}" + mon_to_kill_hostname: "{{ hostvars[mon_to_kill]['ansible_facts']['hostname'] }}" - name: stop monitor service(s) service: diff --git a/infrastructure-playbooks/shrink-osd.yml b/infrastructure-playbooks/shrink-osd.yml index 9c0454db6..6dd00b3b0 100644 --- a/infrastructure-playbooks/shrink-osd.yml +++ b/infrastructure-playbooks/shrink-osd.yml @@ -65,7 +65,7 @@ post_tasks: - name: set_fact container_exec_cmd build docker exec command (containerized) set_fact: - container_exec_cmd: "{{ container_binary }} exec ceph-mon-{{ ansible_hostname }}" + container_exec_cmd: "{{ container_binary }} exec ceph-mon-{{ ansible_facts['hostname'] }}" when: containerized_deployment | bool - name: set_fact container_run_cmd @@ -97,7 +97,7 @@ with_nested: - "{{ groups.get(osd_group_name) }}" - "{{ osd_hosts }}" - when: hostvars[item.0]['ansible_hostname'] == item.1 + when: hostvars[item.0]['ansible_facts']['hostname'] == item.1 - name: get ceph-volume lvm list data command: "{{ container_run_cmd }} lvm list --format json" diff --git a/infrastructure-playbooks/shrink-rbdmirror.yml b/infrastructure-playbooks/shrink-rbdmirror.yml index 3e6539330..db7cc3a21 100644 --- a/infrastructure-playbooks/shrink-rbdmirror.yml +++ b/infrastructure-playbooks/shrink-rbdmirror.yml @@ -64,7 +64,7 @@ - name: set_fact container_exec_cmd for mon0 when: containerized_deployment | bool set_fact: - container_exec_cmd: "{{ container_binary }} exec ceph-mon-{{ ansible_hostname }}" + container_exec_cmd: "{{ container_binary }} exec ceph-mon-{{ ansible_facts['hostname'] }}" - name: exit playbook, if can not connect to the cluster command: "{{ container_exec_cmd | default('') }} timeout 5 ceph --cluster {{ cluster }} service dump -f json" @@ -76,7 +76,7 @@ - name: set_fact rbdmirror_to_kill_hostname set_fact: - rbdmirror_to_kill_hostname: "{{ hostvars[rbdmirror_to_kill]['ansible_hostname'] }}" + rbdmirror_to_kill_hostname: "{{ hostvars[rbdmirror_to_kill]['ansible_facts']['hostname'] }}" - name: set_fact rbdmirror_gids set_fact: diff --git a/infrastructure-playbooks/shrink-rgw.yml b/infrastructure-playbooks/shrink-rgw.yml index 7739628fb..b41628975 100644 --- a/infrastructure-playbooks/shrink-rgw.yml +++ b/infrastructure-playbooks/shrink-rgw.yml @@ -66,7 +66,7 @@ - name: set_fact container_exec_cmd for mon0 set_fact: - container_exec_cmd: "{{ container_binary }} exec ceph-mon-{{ ansible_hostname }}" + container_exec_cmd: "{{ container_binary }} exec ceph-mon-{{ ansible_facts['hostname'] }}" when: containerized_deployment | bool - name: exit playbook, if can not connect to the cluster @@ -95,7 +95,7 @@ set_fact: rgw_host: '{{ item }}' with_items: '{{ groups[rgw_group_name] }}' - when: hostvars[item]['ansible_hostname'] == rgw_to_kill.split('.')[0] + when: hostvars[item]['ansible_facts']['hostname'] == rgw_to_kill.split('.')[0] - name: stop rgw service service: diff --git a/infrastructure-playbooks/switch-from-non-containerized-to-containerized-ceph-daemons.yml b/infrastructure-playbooks/switch-from-non-containerized-to-containerized-ceph-daemons.yml index 8b0190fa7..e554b4293 100644 --- a/infrastructure-playbooks/switch-from-non-containerized-to-containerized-ceph-daemons.yml +++ b/infrastructure-playbooks/switch-from-non-containerized-to-containerized-ceph-daemons.yml @@ -74,7 +74,7 @@ - name: stop non-containerized ceph mon service: - name: "ceph-mon@{{ ansible_hostname }}" + name: "ceph-mon@{{ ansible_facts['hostname'] }}" state: stopped enabled: no @@ -114,7 +114,7 @@ when: ldb_files.rc == 0 - name: copy mon initial keyring in /etc/ceph to satisfy fetch config task in ceph-container-common - command: cp /var/lib/ceph/mon/{{ cluster }}-{{ ansible_hostname }}/keyring /etc/ceph/{{ cluster }}.mon.keyring + command: cp /var/lib/ceph/mon/{{ cluster }}-{{ ansible_facts['hostname'] }}/keyring /etc/ceph/{{ cluster }}.mon.keyring args: creates: /etc/ceph/{{ cluster }}.mon.keyring changed_when: false @@ -137,8 +137,8 @@ - name: waiting for the monitor to join the quorum... command: "{{ container_binary }} run --rm -v /etc/ceph:/etc/ceph:z --entrypoint=ceph {{ ceph_docker_registry }}/{{ ceph_docker_image }}:{{ ceph_docker_image_tag }} --cluster {{ cluster }} quorum_status --format json" register: ceph_health_raw - until: ansible_hostname in (ceph_health_raw.stdout | trim | from_json)["quorum_names"] changed_when: false + until: hostvars[mon_host]['ansible_facts']['hostname'] in (ceph_health_raw.stdout | from_json)["quorum_names"] retries: "{{ health_mon_check_retries }}" delay: "{{ health_mon_check_delay }}" @@ -158,7 +158,7 @@ # will not exist - name: stop non-containerized ceph mgr(s) service: - name: "ceph-mgr@{{ ansible_hostname }}" + name: "ceph-mgr@{{ ansible_facts['hostname'] }}" state: stopped enabled: no failed_when: false @@ -338,14 +338,14 @@ post_tasks: - name: get num_pgs command: > - {{ container_binary }} exec ceph-mon-{{ hostvars[groups[mon_group_name][0]]['ansible_hostname'] }} ceph --cluster {{ cluster }} pg stat --format json + {{ container_binary }} exec ceph-mon-{{ hostvars[groups[mon_group_name][0]]['ansible_facts']['hostname'] }} ceph --cluster {{ cluster }} pg stat --format json register: ceph_pgs delegate_to: "{{ groups[mon_group_name][0] }}" changed_when: false - name: container - waiting for clean pgs... command: > - {{ container_binary }} exec ceph-mon-{{ hostvars[groups[mon_group_name][0]]['ansible_hostname'] }} ceph --cluster {{ cluster }} pg stat --format json + {{ container_binary }} exec ceph-mon-{{ hostvars[groups[mon_group_name][0]]['ansible_facts']['hostname'] }} ceph --cluster {{ cluster }} pg stat --format json register: ceph_health_post until: > (((ceph_health_post.stdout | from_json).pg_summary.num_pg_by_state | length) > 0) @@ -393,7 +393,7 @@ - name: stop non-containerized ceph mds(s) service: - name: "ceph-mds@{{ ansible_hostname }}" + name: "ceph-mds@{{ ansible_facts['hostname'] }}" state: stopped enabled: no @@ -465,7 +465,7 @@ tasks: - name: stop non-containerized ceph rgw(s) service: - name: "ceph-radosgw@rgw.{{ ansible_hostname }}.{{ item.instance_name }}" + name: "ceph-radosgw@rgw.{{ ansible_facts['hostname'] }}.{{ item.instance_name }}" state: stopped enabled: no with_items: "{{ rgw_instances }}" @@ -506,7 +506,7 @@ pre_tasks: - name: stop non-containerized ceph rbd mirror(s) service: - name: "ceph-rbd-mirror@rbd-mirror.{{ ansible_hostname }}" + name: "ceph-rbd-mirror@rbd-mirror.{{ ansible_facts['hostname'] }}" state: stopped enabled: no diff --git a/infrastructure-playbooks/untested-by-ci/cluster-os-migration.yml b/infrastructure-playbooks/untested-by-ci/cluster-os-migration.yml index ab92b7351..e7f9485b9 100644 --- a/infrastructure-playbooks/untested-by-ci/cluster-os-migration.yml +++ b/infrastructure-playbooks/untested-by-ci/cluster-os-migration.yml @@ -19,31 +19,31 @@ - name: Check if the node has be migrated already stat: > - path=/var/lib/ceph/mon/ceph-{{ ansible_hostname }}/migration_completed + path=/var/lib/ceph/mon/ceph-{{ ansible_facts['hostname'] }}/migration_completed register: migration_completed failed_when: false - name: Check for failed run stat: > - path=/var/lib/ceph/{{ ansible_hostname }}.tar + path=/var/lib/ceph/{{ ansible_facts['hostname'] }}.tar register: mon_archive_leftover - fail: msg="Looks like an archive is already there, please remove it!" when: migration_completed.stat.exists == False and mon_archive_leftover.stat.exists == True - name: Compress the store as much as possible - command: ceph tell mon.{{ ansible_hostname }} compact + command: ceph tell mon.{{ ansible_facts['hostname'] }} compact when: migration_completed.stat.exists == False - name: Check if sysvinit stat: > - path=/var/lib/ceph/mon/ceph-{{ ansible_hostname }}/sysvinit + path=/var/lib/ceph/mon/ceph-{{ ansible_facts['hostname'] }}/sysvinit register: monsysvinit changed_when: False - name: Check if upstart stat: > - path=/var/lib/ceph/mon/ceph-{{ ansible_hostname }}/upstart + path=/var/lib/ceph/mon/ceph-{{ ansible_facts['hostname'] }}/upstart register: monupstart changed_when: False @@ -70,7 +70,7 @@ service: > name=ceph-mon state=restarted - args=id={{ ansible_hostname }} + args=id={{ ansible_facts['hostname'] }} when: monupstart.stat.exists == True and migration_completed.stat.exists == False - name: Restart the Monitor after compaction (Sysvinit) @@ -92,7 +92,7 @@ service: > name=ceph-mon state=stopped - args=id={{ ansible_hostname }} + args=id={{ ansible_facts['hostname'] }} when: monupstart.stat.exists == True and migration_completed.stat.exists == False - name: Stop the monitor (Sysvinit) @@ -125,15 +125,15 @@ # NOTE (leseb): should we convert upstart to sysvinit here already? - name: Archive monitor stores shell: > - tar -cpvzf - --one-file-system . /etc/ceph/* | cat > {{ ansible_hostname }}.tar + tar -cpvzf - --one-file-system . /etc/ceph/* | cat > {{ ansible_facts['hostname'] }}.tar chdir=/var/lib/ceph/ - creates={{ ansible_hostname }}.tar + creates={{ ansible_facts['hostname'] }}.tar when: migration_completed.stat.exists == False - name: Scp the Monitor store fetch: > - src=/var/lib/ceph/{{ ansible_hostname }}.tar - dest={{ backup_dir }}/monitors-backups/{{ ansible_hostname }}.tar + src=/var/lib/ceph/{{ ansible_facts['hostname'] }}.tar + dest={{ backup_dir }}/monitors-backups/{{ ansible_facts['hostname'] }}.tar flat=yes when: migration_completed.stat.exists == False @@ -155,13 +155,13 @@ - name: Check if sysvinit stat: > - path=/var/lib/ceph/mon/ceph-{{ ansible_hostname }}/sysvinit + path=/var/lib/ceph/mon/ceph-{{ ansible_facts['hostname'] }}/sysvinit register: monsysvinit changed_when: False - name: Check if upstart stat: > - path=/var/lib/ceph/mon/ceph-{{ ansible_hostname }}/upstart + path=/var/lib/ceph/mon/ceph-{{ ansible_facts['hostname'] }}/upstart register: monupstart changed_when: False @@ -169,7 +169,7 @@ service: > name=ceph-mon state=stopped - args=id={{ ansible_hostname }} + args=id={{ ansible_facts['hostname'] }} when: monupstart.stat.exists == True and migration_completed.stat.exists == False - name: Make sure the monitor is stopped (Sysvinit) @@ -182,7 +182,7 @@ # NOTE (leseb): 'creates' was added in Ansible 1.6 - name: Copy and unarchive the monitor store unarchive: > - src={{ backup_dir }}/monitors-backups/{{ ansible_hostname }}.tar + src={{ backup_dir }}/monitors-backups/{{ ansible_facts['hostname'] }}.tar dest=/var/lib/ceph/ copy=yes mode=0600 @@ -218,7 +218,7 @@ - name: Waiting for the monitor to join the quorum... shell: > - ceph -s | grep monmap | sed 's/.*quorum//' | egrep -q {{ ansible_hostname }} + ceph -s | grep monmap | sed 's/.*quorum//' | egrep -q {{ ansible_facts['hostname'] }} register: result until: result.rc == 0 retries: 5 @@ -229,7 +229,7 @@ - name: Done moving to the next monitor file: > - path=/var/lib/ceph/mon/ceph-{{ ansible_hostname }}/migration_completed + path=/var/lib/ceph/mon/ceph-{{ ansible_facts['hostname'] }}/migration_completed state=touch owner=root group=root @@ -252,7 +252,7 @@ - name: Check for failed run stat: > - path=/var/lib/ceph/{{ ansible_hostname }}.tar + path=/var/lib/ceph/{{ ansible_facts['hostname'] }}.tar register: osd_archive_leftover - fail: msg="Looks like an archive is already there, please remove it!" @@ -297,9 +297,9 @@ - name: Archive ceph configs shell: > - tar -cpvzf - --one-file-system . /etc/ceph/ceph.conf | cat > {{ ansible_hostname }}.tar + tar -cpvzf - --one-file-system . /etc/ceph/ceph.conf | cat > {{ ansible_facts['hostname'] }}.tar chdir=/var/lib/ceph/ - creates={{ ansible_hostname }}.tar + creates={{ ansible_facts['hostname'] }}.tar when: migration_completed.stat.exists == False - name: Create backup directory @@ -315,7 +315,7 @@ - name: Scp OSDs dirs and configs fetch: > - src=/var/lib/ceph/{{ ansible_hostname }}.tar + src=/var/lib/ceph/{{ ansible_facts['hostname'] }}.tar dest={{ backup_dir }}/osds-backups/ flat=yes when: migration_completed.stat.exists == False @@ -371,7 +371,7 @@ # NOTE (leseb): 'creates' was added in Ansible 1.6 - name: Copy and unarchive the OSD configs unarchive: > - src={{ backup_dir }}/osds-backups/{{ ansible_hostname }}.tar + src={{ backup_dir }}/osds-backups/{{ ansible_facts['hostname'] }}.tar dest=/var/lib/ceph/ copy=yes mode=0600 @@ -446,7 +446,7 @@ - name: Check for failed run stat: > - path=/var/lib/ceph/{{ ansible_hostname }}.tar + path=/var/lib/ceph/{{ ansible_facts['hostname'] }}.tar register: rgw_archive_leftover - fail: msg="Looks like an archive is already there, please remove it!" @@ -454,9 +454,9 @@ - name: Archive rados gateway configs shell: > - tar -cpvzf - --one-file-system . /etc/ceph/* | cat > {{ ansible_hostname }}.tar + tar -cpvzf - --one-file-system . /etc/ceph/* | cat > {{ ansible_facts['hostname'] }}.tar chdir=/var/lib/ceph/ - creates={{ ansible_hostname }}.tar + creates={{ ansible_facts['hostname'] }}.tar when: migration_completed.stat.exists == False - name: Create backup directory @@ -472,7 +472,7 @@ - name: Scp RGWs dirs and configs fetch: > - src=/var/lib/ceph/{{ ansible_hostname }}.tar + src=/var/lib/ceph/{{ ansible_facts['hostname'] }}.tar dest={{ backup_dir }}/rgws-backups/ flat=yes when: migration_completed.stat.exists == False @@ -512,7 +512,7 @@ # NOTE (leseb): 'creates' was added in Ansible 1.6 - name: Copy and unarchive the OSD configs unarchive: > - src={{ backup_dir }}/rgws-backups/{{ ansible_hostname }}.tar + src={{ backup_dir }}/rgws-backups/{{ ansible_facts['hostname'] }}.tar dest=/var/lib/ceph/ copy=yes mode=0600 diff --git a/infrastructure-playbooks/untested-by-ci/make-osd-partitions.yml b/infrastructure-playbooks/untested-by-ci/make-osd-partitions.yml index c6aadfe18..29f40433c 100644 --- a/infrastructure-playbooks/untested-by-ci/make-osd-partitions.yml +++ b/infrastructure-playbooks/untested-by-ci/make-osd-partitions.yml @@ -35,13 +35,13 @@ include_vars: "{{ item }}" with_first_found: - files: - - "host_vars/{{ ansible_hostname }}.yml" + - "host_vars/{{ ansible_facts['hostname'] }}.yml" - "host_vars/default.yml" skip: true - name: exit playbook, if devices not defined fail: - msg: "devices must be define in host_vars/default.yml or host_vars/{{ ansible_hostname }}.yml" + msg: "devices must be define in host_vars/default.yml or host_vars/{{ ansible_facts['hostname'] }}.yml" when: devices is not defined - name: install sgdisk(gdisk) @@ -67,12 +67,12 @@ - set_fact: owner: 167 group: 167 - when: ansible_os_family == "RedHat" + when: ansible_facts['os_family'] == "RedHat" - set_fact: owner: 64045 group: 64045 - when: ansible_os_family == "Debian" + when: ansible_facts['os_family'] == "Debian" - name: change partitions ownership file: diff --git a/infrastructure-playbooks/untested-by-ci/replace-osd.yml b/infrastructure-playbooks/untested-by-ci/replace-osd.yml index 9184598f9..11f386914 100644 --- a/infrastructure-playbooks/untested-by-ci/replace-osd.yml +++ b/infrastructure-playbooks/untested-by-ci/replace-osd.yml @@ -60,7 +60,7 @@ post_tasks: - name: set_fact container_exec_cmd build docker exec command (containerized) set_fact: - container_exec_cmd: "docker exec ceph-mon-{{ hostvars[groups[mon_group_name][0]]['ansible_hostname'] }}" + container_exec_cmd: "docker exec ceph-mon-{{ hostvars[groups[mon_group_name][0]]['ansible_facts']['hostname'] }}" when: containerized_deployment | bool - name: exit playbook, if can not connect to the cluster diff --git a/profiles/rgw-keystone-v2 b/profiles/rgw-keystone-v2 index 6ba9fcf43..0d1d0abfa 100644 --- a/profiles/rgw-keystone-v2 +++ b/profiles/rgw-keystone-v2 @@ -7,7 +7,7 @@ ceph_conf_overrides: - "client.rgw.{{ hostvars[inventory_hostname]['ansible_hostname'] }}": + "client.rgw.{{ hostvars[inventory_hostname]['ansible_facts']['hostname'] }}": "rgw keystone api version": "2" "rgw keystone url": "http://192.168.0.1:35357" "rgw keystone admin token": "password" @@ -16,7 +16,7 @@ ceph_conf_overrides: "rgw keystone token cache size": "10000" "rgw keystone revocation interval": "900" "rgw s3 auth use keystone": "true" - "nss db path": "/var/lib/ceph/radosgw/ceph-radosgw.{{ ansible_hostname }}/nss" + "nss db path": "/var/lib/ceph/radosgw/ceph-radosgw.{{ ansible_facts['hostname'] }}/nss" # NOTE (leseb): to authentivate with Keystone you have two options: diff --git a/profiles/rgw-keystone-v3 b/profiles/rgw-keystone-v3 index 1044e6d44..327de816c 100644 --- a/profiles/rgw-keystone-v3 +++ b/profiles/rgw-keystone-v3 @@ -7,7 +7,7 @@ ceph_conf_overrides: - "client.rgw.{{ hostvars[inventory_hostname]['ansible_hostname'] }}": + "client.rgw.{{ hostvars[inventory_hostname]['ansible_facts']['hostname'] }}": "rgw keystone api version": "3" "rgw keystone url": "http://192.168.0.1:35357" "rgw keystone admin token": "password" @@ -17,7 +17,7 @@ ceph_conf_overrides: "rgw keystone token cache size": "10000" "rgw keystone revocation interval": "900" "rgw s3 auth use keystone": "true" - "nss db path": "/var/lib/ceph/radosgw/ceph-radosgw.{{ ansible_hostname }}/nss" + "nss db path": "/var/lib/ceph/radosgw/ceph-radosgw.{{ ansible_facts['hostname'] }}/nss" # NOTE (leseb): to authentivate with Keystone you have two options: diff --git a/profiles/rgw-radosgw-static-website b/profiles/rgw-radosgw-static-website index cf2dd230d..1cd18c260 100644 --- a/profiles/rgw-radosgw-static-website +++ b/profiles/rgw-radosgw-static-website @@ -6,6 +6,6 @@ # The double quotes are important, do NOT remove them. ceph_conf_overrides: - "client.rgw.{{ hostvars[inventory_hostname]['ansible_hostname'] }}": + "client.rgw.{{ hostvars[inventory_hostname]['ansible_facts']['hostname'] }}": rgw enable static website = true rgw dns s3website name = objects-website-region.domain.com diff --git a/profiles/rgw-usage-log b/profiles/rgw-usage-log index 40645281c..af704fb72 100644 --- a/profiles/rgw-usage-log +++ b/profiles/rgw-usage-log @@ -6,7 +6,7 @@ # The double quotes are important, do NOT remove them. ceph_conf_overrides: - "client.rgw.{{ hostvars[inventory_hostname]['ansible_hostname'] }}": + "client.rgw.{{ hostvars[inventory_hostname]['ansible_facts']['hostname'] }}": rgw enable usage log = true rgw usage log tick interval = 30 rgw usage log flush threshold = 1024 diff --git a/roles/ceph-client/tasks/create_users_keys.yml b/roles/ceph-client/tasks/create_users_keys.yml index 64d68ab27..b98565330 100644 --- a/roles/ceph-client/tasks/create_users_keys.yml +++ b/roles/ceph-client/tasks/create_users_keys.yml @@ -20,7 +20,7 @@ name: "{{ item }}" groups: _filtered_clients with_items: "{{ groups[client_group_name] | intersect(ansible_play_batch) }}" - when: (hostvars[item]['ansible_architecture'] == 'x86_64') or (not containerized_deployment | bool) + when: (hostvars[item]['ansible_facts']['architecture'] == 'x86_64') or (not containerized_deployment | bool) - name: set_fact delegated_node set_fact: diff --git a/roles/ceph-common/tasks/configure_cluster_name.yml b/roles/ceph-common/tasks/configure_cluster_name.yml index ab7e51853..7b7cb4293 100644 --- a/roles/ceph-common/tasks/configure_cluster_name.yml +++ b/roles/ceph-common/tasks/configure_cluster_name.yml @@ -6,7 +6,7 @@ create: yes line: "CLUSTER={{ cluster }}" regexp: "^CLUSTER=" - when: ansible_os_family in ["RedHat", "Suse"] + when: ansible_facts['os_family'] in ["RedHat", "Suse"] # NOTE(leseb): we are performing the following check # to ensure any Jewel installation will not fail. @@ -19,7 +19,7 @@ # - All previous versions from Canonical # - Infernalis from ceph.com - name: debian based systems - configure cluster name - when: ansible_os_family == "Debian" + when: ansible_facts['os_family'] == "Debian" block: - name: check /etc/default/ceph exist stat: diff --git a/roles/ceph-common/tasks/configure_memory_allocator.yml b/roles/ceph-common/tasks/configure_memory_allocator.yml index b76145777..115ecf59e 100644 --- a/roles/ceph-common/tasks/configure_memory_allocator.yml +++ b/roles/ceph-common/tasks/configure_memory_allocator.yml @@ -7,7 +7,7 @@ regexp: "^TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=" line: "TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES={{ ceph_tcmalloc_max_total_thread_cache }}" when: - - ansible_os_family == 'Debian' + - ansible_facts['os_family'] == 'Debian' - etc_default_ceph.stat.exists notify: restart ceph osds @@ -18,5 +18,5 @@ create: yes regexp: "^TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=" line: "TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES={{ ceph_tcmalloc_max_total_thread_cache }}" - when: ansible_os_family == 'RedHat' + when: ansible_facts['os_family'] == 'RedHat' notify: restart ceph osds diff --git a/roles/ceph-common/tasks/installs/configure_redhat_repository_installation.yml b/roles/ceph-common/tasks/installs/configure_redhat_repository_installation.yml index a18801120..a36355a1b 100644 --- a/roles/ceph-common/tasks/installs/configure_redhat_repository_installation.yml +++ b/roles/ceph-common/tasks/installs/configure_redhat_repository_installation.yml @@ -21,4 +21,4 @@ args: warn: no changed_when: false - when: ansible_pkg_mgr == 'yum' + when: ansible_facts['pkg_mgr'] == 'yum' diff --git a/roles/ceph-common/tasks/installs/debian_community_repository.yml b/roles/ceph-common/tasks/installs/debian_community_repository.yml index 0afa23bef..6832a3f5b 100644 --- a/roles/ceph-common/tasks/installs/debian_community_repository.yml +++ b/roles/ceph-common/tasks/installs/debian_community_repository.yml @@ -15,6 +15,6 @@ - name: configure debian ceph stable community repository apt_repository: - repo: "deb {{ ceph_stable_repo }} {{ ceph_stable_distro_source | default(ansible_distribution_release) }} main" + repo: "deb {{ ceph_stable_repo }} {{ ceph_stable_distro_source | default(ansible_facts['distribution_release']) }} main" state: present update_cache: yes diff --git a/roles/ceph-common/tasks/installs/debian_custom_repository.yml b/roles/ceph-common/tasks/installs/debian_custom_repository.yml index 10c5ea38f..607ce0896 100644 --- a/roles/ceph-common/tasks/installs/debian_custom_repository.yml +++ b/roles/ceph-common/tasks/installs/debian_custom_repository.yml @@ -9,6 +9,6 @@ - name: configure debian custom repository apt_repository: - repo: "deb {{ ceph_custom_repo }} {{ ansible_distribution_release }} main" + repo: "deb {{ ceph_custom_repo }} {{ ansible_facts['distribution_release'] }} main" state: present update_cache: yes diff --git a/roles/ceph-common/tasks/installs/debian_dev_repository.yml b/roles/ceph-common/tasks/installs/debian_dev_repository.yml index 9bde8bd2b..c5c8c5c3a 100644 --- a/roles/ceph-common/tasks/installs/debian_dev_repository.yml +++ b/roles/ceph-common/tasks/installs/debian_dev_repository.yml @@ -1,7 +1,7 @@ --- - name: fetch ceph debian development repository uri: - url: https://shaman.ceph.com/api/repos/ceph/{{ ceph_dev_branch }}/{{ ceph_dev_sha1 }}/{{ ansible_distribution | lower }}/{{ ansible_distribution_release }}/repo + url: https://shaman.ceph.com/api/repos/ceph/{{ ceph_dev_branch }}/{{ ceph_dev_sha1 }}/{{ ansible_facts['distribution'] | lower }}/{{ ansible_facts['distribution_release'] }}/repo return_content: yes register: ceph_dev_deb_repo diff --git a/roles/ceph-common/tasks/installs/install_debian_packages.yml b/roles/ceph-common/tasks/installs/install_debian_packages.yml index 6bdc03f4a..3c6db106f 100644 --- a/roles/ceph-common/tasks/installs/install_debian_packages.yml +++ b/roles/ceph-common/tasks/installs/install_debian_packages.yml @@ -4,6 +4,6 @@ name: "{{ debian_ceph_pkgs | unique }}" update_cache: no state: "{{ (upgrade_ceph_packages|bool) | ternary('latest','present') }}" - default_release: "{{ ceph_stable_release_uca | default('') if ceph_origin == 'repository' and ceph_repository == 'uca' else '' }}{{ ansible_distribution_release ~ '-backports' if ceph_origin == 'distro' and ceph_use_distro_backports else '' }}" + default_release: "{{ ceph_stable_release_uca | default('') if ceph_origin == 'repository' and ceph_repository == 'uca' else '' }}{{ ansible_facts['distribution_release'] ~ '-backports' if ceph_origin == 'distro' and ceph_use_distro_backports | bool else '' }}" register: result until: result is succeeded diff --git a/roles/ceph-common/tasks/installs/install_redhat_packages.yml b/roles/ceph-common/tasks/installs/install_redhat_packages.yml index 28b56fd9a..7e9f6d62f 100644 --- a/roles/ceph-common/tasks/installs/install_redhat_packages.yml +++ b/roles/ceph-common/tasks/installs/install_redhat_packages.yml @@ -5,7 +5,7 @@ state: present register: result until: result is succeeded - when: ansible_distribution == 'RedHat' + when: ansible_facts['distribution'] == 'RedHat' - name: install centos dependencies yum: @@ -13,7 +13,7 @@ state: present register: result until: result is succeeded - when: ansible_distribution == 'CentOS' + when: ansible_facts['distribution'] == 'CentOS' - name: install redhat ceph packages package: diff --git a/roles/ceph-common/tasks/installs/prerequisite_rhcs_cdn_install.yml b/roles/ceph-common/tasks/installs/prerequisite_rhcs_cdn_install.yml index 51835a956..f53302ef3 100644 --- a/roles/ceph-common/tasks/installs/prerequisite_rhcs_cdn_install.yml +++ b/roles/ceph-common/tasks/installs/prerequisite_rhcs_cdn_install.yml @@ -1,24 +1,24 @@ --- - name: "rhcs {{ ceph_rhcs_version }} on rhel 8" - when: ansible_distribution_major_version | int == 8 + when: ansible_facts['distribution_major_version'] | int == 8 block: - name: enable red hat storage monitor repository rhsm_repository: - name: "rhceph-{{ ceph_rhcs_version }}-mon-for-rhel-8-{{ ansible_architecture }}-rpms" + name: "rhceph-{{ ceph_rhcs_version }}-mon-for-rhel-8-{{ ansible_facts['architecture'] }}-rpms" when: (mon_group_name in group_names or mgr_group_name in group_names) - name: enable red hat storage osd repository rhsm_repository: - name: "rhceph-{{ ceph_rhcs_version }}-osd-for-rhel-8-{{ ansible_architecture }}-rpms" + name: "rhceph-{{ ceph_rhcs_version }}-osd-for-rhel-8-{{ ansible_facts['architecture'] }}-rpms" when: osd_group_name in group_names - name: enable red hat storage tools repository rhsm_repository: - name: "rhceph-{{ ceph_rhcs_version }}-tools-for-rhel-8-{{ ansible_architecture }}-rpms" + name: "rhceph-{{ ceph_rhcs_version }}-tools-for-rhel-8-{{ ansible_facts['architecture'] }}-rpms" when: (mgr_group_name in group_names or rgw_group_name in group_names or mds_group_name in group_names or nfs_group_name in group_names or iscsi_gw_group_name in group_names or client_group_name in group_names or monitoring_group_name in group_names) - name: "rhcs {{ ceph_rhcs_version }} on rhel 7" - when: ansible_distribution_major_version | int == 7 + when: ansible_facts['distribution_major_version'] | int == 7 block: - name: enable red hat storage monitor repository rhsm_repository: diff --git a/roles/ceph-common/tasks/installs/redhat_community_repository.yml b/roles/ceph-common/tasks/installs/redhat_community_repository.yml index 5fd817d75..4db2737cc 100644 --- a/roles/ceph-common/tasks/installs/redhat_community_repository.yml +++ b/roles/ceph-common/tasks/installs/redhat_community_repository.yml @@ -5,7 +5,7 @@ register: result until: result is succeeded tags: with_pkg - when: ansible_distribution_major_version | int == 7 + when: ansible_facts['distribution_major_version'] | int == 7 - name: configure red hat ceph community repository stable key rpm_key: @@ -21,7 +21,7 @@ gpgcheck: yes state: present gpgkey: "{{ ceph_stable_key }}" - baseurl: "{{ ceph_mirror }}/rpm-{{ ceph_stable_release }}/el{{ ansible_distribution_major_version }}/$basearch" + baseurl: "{{ ceph_mirror }}/rpm-{{ ceph_stable_release }}/el{{ ansible_facts['distribution_major_version'] }}/$basearch" file: ceph_stable priority: 2 register: result @@ -34,7 +34,7 @@ gpgcheck: yes state: present gpgkey: "{{ ceph_stable_key }}" - baseurl: "{{ ceph_mirror }}/rpm-{{ ceph_stable_release }}/el{{ ansible_distribution_major_version }}/noarch" + baseurl: "{{ ceph_mirror }}/rpm-{{ ceph_stable_release }}/el{{ ansible_facts['distribution_major_version'] }}/noarch" file: ceph_stable priority: 2 register: result diff --git a/roles/ceph-common/tasks/installs/redhat_dev_repository.yml b/roles/ceph-common/tasks/installs/redhat_dev_repository.yml index f6b7a0674..1b0af582f 100644 --- a/roles/ceph-common/tasks/installs/redhat_dev_repository.yml +++ b/roles/ceph-common/tasks/installs/redhat_dev_repository.yml @@ -1,7 +1,7 @@ --- - name: get latest available build uri: - url: "https://shaman.ceph.com/api/search/?status=ready&project=ceph&flavor=default&distros=centos/{{ ansible_distribution_major_version }}/{{ ansible_architecture }}&ref={{ ceph_dev_branch }}&sha1={{ ceph_dev_sha1 }}" + url: "https://shaman.ceph.com/api/search/?status=ready&project=ceph&flavor=default&distros=centos/{{ ansible_facts['distribution_major_version'] }}/{{ ansible_facts['architecture'] }}&ref={{ ceph_dev_branch }}&sha1={{ ceph_dev_sha1 }}" return_content: yes run_once: true register: latest_build diff --git a/roles/ceph-common/tasks/main.yml b/roles/ceph-common/tasks/main.yml index 2239aad06..31ff66a88 100644 --- a/roles/ceph-common/tasks/main.yml +++ b/roles/ceph-common/tasks/main.yml @@ -1,22 +1,22 @@ --- - name: include_tasks installs/install_on_redhat.yml include_tasks: installs/install_on_redhat.yml - when: ansible_os_family == 'RedHat' + when: ansible_facts['os_family'] == 'RedHat' tags: package-install - name: include_tasks installs/install_on_suse.yml include_tasks: installs/install_on_suse.yml - when: ansible_os_family == 'Suse' + when: ansible_facts['os_family'] == 'Suse' tags: package-install - name: include installs/install_on_debian.yml include_tasks: installs/install_on_debian.yml tags: package-install - when: ansible_os_family == 'Debian' + when: ansible_facts['os_family'] == 'Debian' - name: include_tasks installs/install_on_clear.yml include_tasks: installs/install_on_clear.yml - when: ansible_os_family == 'ClearLinux' + when: ansible_facts['os_family'] == 'ClearLinux' tags: package-install - name: get ceph version diff --git a/roles/ceph-config/tasks/rgw_systemd_environment_file.yml b/roles/ceph-config/tasks/rgw_systemd_environment_file.yml index ee310f110..d59ba4593 100644 --- a/roles/ceph-config/tasks/rgw_systemd_environment_file.yml +++ b/roles/ceph-config/tasks/rgw_systemd_environment_file.yml @@ -1,7 +1,7 @@ --- - name: create rados gateway instance directories file: - path: "/var/lib/ceph/radosgw/{{ cluster }}-rgw.{{ ansible_hostname }}.{{ item.instance_name }}" + path: "/var/lib/ceph/radosgw/{{ cluster }}-rgw.{{ ansible_facts['hostname'] }}.{{ item.instance_name }}" state: directory owner: "{{ ceph_uid if containerized_deployment else 'ceph' }}" group: "{{ ceph_uid if containerized_deployment else 'ceph' }}" @@ -10,7 +10,7 @@ - name: generate environment file copy: - dest: "/var/lib/ceph/radosgw/{{ cluster }}-rgw.{{ ansible_hostname }}.{{ item.instance_name }}/EnvironmentFile" + dest: "/var/lib/ceph/radosgw/{{ cluster }}-rgw.{{ ansible_facts['hostname'] }}.{{ item.instance_name }}/EnvironmentFile" owner: "root" group: "root" mode: "0644" diff --git a/roles/ceph-config/templates/ceph.conf.j2 b/roles/ceph-config/templates/ceph.conf.j2 index 09e3bd785..fef76271e 100644 --- a/roles/ceph-config/templates/ceph.conf.j2 +++ b/roles/ceph-config/templates/ceph.conf.j2 @@ -24,10 +24,10 @@ osd crush chooseleaf type = 0 {% if nb_mon > 0 and inventory_hostname in groups.get(mon_group_name, []) %} mon initial members = {% for host in groups[mon_group_name] %} - {% if hostvars[host]['ansible_fqdn'] is defined and mon_use_fqdn -%} - {{ hostvars[host]['ansible_fqdn'] }} - {%- elif hostvars[host]['ansible_hostname'] is defined -%} - {{ hostvars[host]['ansible_hostname'] }} + {% if hostvars[host]['ansible_facts']['fqdn'] is defined and mon_use_fqdn -%} + {{ hostvars[host]['ansible_facts']['fqdn'] }} + {%- elif hostvars[host]['ansible_facts']['hostname'] is defined -%} + {{ hostvars[host]['ansible_facts']['hostname'] }} {%- endif %} {%- if not loop.last %},{% endif %} {% endfor %} @@ -92,13 +92,13 @@ filestore xattr use omap = true [osd] {% if is_hci | bool and _num_osds > 0 %} {# hci_safety_factor is the safety factor for HCI deployments #} -{% if ansible_memtotal_mb * 1048576 * hci_safety_factor / _num_osds > osd_memory_target %} -{% set _osd_memory_target = (ansible_memtotal_mb * 1048576 * hci_safety_factor / _num_osds) | int %} +{% if ansible_facts['memtotal_mb'] * 1048576 * hci_safety_factor / _num_osds > osd_memory_target %} +{% set _osd_memory_target = (ansible_facts['memtotal_mb'] * 1048576 * hci_safety_factor / _num_osds) | int %} {% endif %} {% elif _num_osds > 0 %} {# non_hci_safety_factor is the safety factor for dedicated nodes #} -{% if ansible_memtotal_mb * 1048576 * non_hci_safety_factor / _num_osds > osd_memory_target %} -{% set _osd_memory_target = (ansible_memtotal_mb * 1048576 * non_hci_safety_factor / _num_osds) | int %} +{% if ansible_facts['memtotal_mb'] * 1048576 * non_hci_safety_factor / _num_osds > osd_memory_target %} +{% set _osd_memory_target = (ansible_facts['memtotal_mb'] * 1048576 * non_hci_safety_factor / _num_osds) | int %} {% endif %} {% endif %} osd memory target = {{ _osd_memory_target | default(osd_memory_target) }} @@ -106,14 +106,14 @@ osd memory target = {{ _osd_memory_target | default(osd_memory_target) }} {% endif %} {% if inventory_hostname in groups.get(rgw_group_name, []) %} -{% set _rgw_hostname = hostvars[inventory_hostname]['rgw_hostname'] | default(hostvars[inventory_hostname]['ansible_hostname']) %} +{% set _rgw_hostname = hostvars[inventory_hostname]['rgw_hostname'] | default(hostvars[inventory_hostname]['ansible_facts']['hostname']) %} {# {{ hostvars[host]['rgw_hostname'] }} for backward compatibility, fqdn issues. See bz1580408 #} {% if hostvars[inventory_hostname]['rgw_instances'] is defined %} {% for instance in hostvars[inventory_hostname]['rgw_instances'] %} [client.rgw.{{ _rgw_hostname + '.' + instance['instance_name'] }}] host = {{ _rgw_hostname }} keyring = /var/lib/ceph/radosgw/{{ cluster }}-rgw.{{ _rgw_hostname + '.' + instance['instance_name'] }}/keyring -log file = /var/log/ceph/{{ cluster }}-rgw-{{ hostvars[inventory_hostname]['ansible_hostname'] + '.' + instance['instance_name'] }}.log +log file = /var/log/ceph/{{ cluster }}-rgw-{{ hostvars[inventory_hostname]['ansible_facts']['hostname'] + '.' + instance['instance_name'] }}.log {% set _rgw_binding_socket = instance['radosgw_address'] | default(_radosgw_address) | string + ':' + instance['radosgw_frontend_port'] | default(radosgw_frontend_port) | string %} {%- macro frontend_line(frontend_type) -%} {%- if frontend_type == 'civetweb' -%} @@ -139,12 +139,12 @@ rgw_zone = {{ instance['rgw_zone'] }} {% if inventory_hostname in groups.get(nfs_group_name, []) and inventory_hostname not in groups.get(rgw_group_name, []) %} {% for host in groups[nfs_group_name] %} -{% set _rgw_hostname = hostvars[host]['rgw_hostname'] | default(hostvars[host]['ansible_hostname']) %} -{% if nfs_obj_gw %} +{% set _rgw_hostname = hostvars[host]['rgw_hostname'] | default(hostvars[host]['ansible_facts']['hostname']) %} +{% if nfs_obj_gw | bool %} [client.rgw.{{ _rgw_hostname }}] host = {{ _rgw_hostname }} keyring = /var/lib/ceph/radosgw/{{ cluster }}-rgw.{{ _rgw_hostname }}/keyring -log file = /var/log/ceph/{{ cluster }}-rgw-{{ hostvars[host]['ansible_hostname'] }}.log +log file = /var/log/ceph/{{ cluster }}-rgw-{{ hostvars[host]['ansible_facts']['hostname'] }}.log {% endif %} {% endfor %} {% endif %} diff --git a/roles/ceph-container-engine/tasks/pre_requisites/debian_prerequisites.yml b/roles/ceph-container-engine/tasks/pre_requisites/debian_prerequisites.yml index 1adf86ca9..7a965da98 100644 --- a/roles/ceph-container-engine/tasks/pre_requisites/debian_prerequisites.yml +++ b/roles/ceph-container-engine/tasks/pre_requisites/debian_prerequisites.yml @@ -14,14 +14,14 @@ - name: add docker's gpg key apt_key: - url: 'https://download.docker.com/linux/{{ ansible_distribution | lower }}/gpg' + url: "https://download.docker.com/linux/{{ ansible_facts['distribution'] | lower }}/gpg" register: result until: result is succeeded when: container_package_name == 'docker-ce' - name: add docker repository apt_repository: - repo: "deb https://download.docker.com/linux/{{ ansible_distribution | lower }} {{ ansible_distribution_release }} stable" + repo: "deb https://download.docker.com/linux/{{ ansible_facts['distribution'] | lower }} {{ ansible_facts['distribution_release'] }} stable" when: container_package_name == 'docker-ce' - name: add podman ppa repository @@ -29,4 +29,4 @@ repo: "ppa:projectatomic/ppa" when: - container_package_name == 'podman' - - ansible_distribution == 'Ubuntu' + - ansible_facts['distribution'] == 'Ubuntu' diff --git a/roles/ceph-container-engine/tasks/pre_requisites/prerequisites.yml b/roles/ceph-container-engine/tasks/pre_requisites/prerequisites.yml index 5ed0127e7..baa373bfc 100644 --- a/roles/ceph-container-engine/tasks/pre_requisites/prerequisites.yml +++ b/roles/ceph-container-engine/tasks/pre_requisites/prerequisites.yml @@ -2,13 +2,13 @@ - name: include specific variables include_vars: "{{ item }}" with_first_found: - - "{{ ansible_distribution }}-{{ ansible_distribution_major_version }}.yml" - - "{{ ansible_os_family }}.yml" + - "{{ ansible_facts['distribution'] }}-{{ ansible_facts['distribution_major_version'] }}.yml" + - "{{ ansible_facts['os_family'] }}.yml" - name: debian based systems tasks include_tasks: debian_prerequisites.yml when: - - ansible_os_family == 'Debian' + - ansible_facts['os_family'] == 'Debian' tags: with_pkg # ensure extras enabled for docker @@ -18,7 +18,7 @@ state: present enabled: yes when: - - ansible_distribution == 'CentOS' + - ansible_facts['distribution'] == 'CentOS' - ceph_docker_enable_centos_extra_repo | bool tags: with_pkg diff --git a/roles/ceph-crash/tasks/main.yml b/roles/ceph-crash/tasks/main.yml index 4bcb62274..e2e0806ce 100644 --- a/roles/ceph-crash/tasks/main.yml +++ b/roles/ceph-crash/tasks/main.yml @@ -52,7 +52,7 @@ - name: start the ceph-crash service systemd: - name: "{{ 'ceph-crash@' + ansible_hostname if containerized_deployment | bool else 'ceph-crash.service' }}" + name: "{{ 'ceph-crash@' + ansible_facts['hostname'] if containerized_deployment | bool else 'ceph-crash.service' }}" state: started enabled: yes masked: no diff --git a/roles/ceph-dashboard/tasks/configure_dashboard.yml b/roles/ceph-dashboard/tasks/configure_dashboard.yml index 869b03a69..3e4139c37 100644 --- a/roles/ceph-dashboard/tasks/configure_dashboard.yml +++ b/roles/ceph-dashboard/tasks/configure_dashboard.yml @@ -1,7 +1,7 @@ --- - name: set_fact container_exec_cmd set_fact: - container_exec_cmd: "{{ container_binary }} exec ceph-mon-{{ hostvars[groups[mon_group_name][0]]['ansible_hostname'] }}" + container_exec_cmd: "{{ container_binary }} exec ceph-mon-{{ hostvars[groups[mon_group_name][0]]['ansible_facts']['hostname'] }}" when: containerized_deployment | bool - name: set_fact container_run_cmd @@ -378,7 +378,7 @@ - name: add iscsi gateways - ipv4 command: "{{ ceph_cmd }} --cluster {{ cluster }} dashboard iscsi-gateway-add -i -" # noqa 304 args: - stdin: "{{ 'https' if hostvars[item]['api_secure'] | default(false) | bool else 'http' }}://{{ hostvars[item]['api_user'] | default('admin') }}:{{ hostvars[item]['api_password'] | default('admin') }}@{{ hostvars[item]['ansible_all_ipv4_addresses'] | ips_in_ranges(public_network.split(',')) | first }}:{{ hostvars[item]['api_port'] | default(5000) }}" + stdin: "{{ 'https' if hostvars[item]['api_secure'] | default(false) | bool else 'http' }}://{{ hostvars[item]['api_user'] | default('admin') }}:{{ hostvars[item]['api_password'] | default('admin') }}@{{ hostvars[item]['ansible_facts']['all_ipv4_addresses'] | ips_in_ranges(public_network.split(',')) | first }}:{{ hostvars[item]['api_port'] | default(5000) }}" stdin_add_newline: no changed_when: false delegate_to: "{{ groups[mon_group_name][0] }}" @@ -389,7 +389,7 @@ - dashboard_password_from_stdin | bool - name: add iscsi gateways - ipv4 (legacy) - command: "{{ container_exec_cmd }} ceph --cluster {{ cluster }} dashboard iscsi-gateway-add {{ 'https' if hostvars[item]['api_secure'] | default(false) | bool else 'http' }}://{{ hostvars[item]['api_user'] | default('admin') }}:{{ hostvars[item]['api_password'] | default('admin') }}@{{ hostvars[item]['ansible_all_ipv4_addresses'] | ips_in_ranges(public_network.split(',')) | first }}:{{ hostvars[item]['api_port'] | default(5000) }}" + command: "{{ container_exec_cmd }} ceph --cluster {{ cluster }} dashboard iscsi-gateway-add {{ 'https' if hostvars[item]['api_secure'] | default(false) | bool else 'http' }}://{{ hostvars[item]['api_user'] | default('admin') }}:{{ hostvars[item]['api_password'] | default('admin') }}@{{ hostvars[item]['ansible_facts']['all_ipv4_addresses'] | ips_in_ranges(public_network.split(',')) | first }}:{{ hostvars[item]['api_port'] | default(5000) }}" changed_when: false delegate_to: "{{ groups[mon_group_name][0] }}" with_items: "{{ groups[iscsi_gw_group_name] }}" @@ -401,7 +401,7 @@ - name: add iscsi gateways - ipv6 command: "{{ ceph_cmd }} --cluster {{ cluster }} dashboard iscsi-gateway-add -i -" # noqa 304 args: - stdin: "{{ 'https' if hostvars[item]['api_secure'] | default(false) | bool else 'http' }}://{{ hostvars[item]['api_user'] | default('admin') }}:{{ hostvars[item]['api_password'] | default('admin') }}@{{ hostvars[item]['ansible_all_ipv6_addresses'] | ips_in_ranges(public_network.split(',')) | last | ipwrap }}:{{ hostvars[item]['api_port'] | default(5000) }}" + stdin: "{{ 'https' if hostvars[item]['api_secure'] | default(false) | bool else 'http' }}://{{ hostvars[item]['api_user'] | default('admin') }}:{{ hostvars[item]['api_password'] | default('admin') }}@{{ hostvars[item]['ansible_facts']['all_ipv6_addresses'] | ips_in_ranges(public_network.split(',')) | last | ipwrap }}:{{ hostvars[item]['api_port'] | default(5000) }}" stdin_add_newline: no changed_when: false delegate_to: "{{ groups[mon_group_name][0] }}" @@ -412,7 +412,7 @@ - dashboard_password_from_stdin | bool - name: add iscsi gateways - ipv6 (legacy) - command: "{{ container_exec_cmd }} ceph --cluster {{ cluster }} dashboard iscsi-gateway-add {{ 'https' if hostvars[item]['api_secure'] | default(false) | bool else 'http' }}://{{ hostvars[item]['api_user'] | default('admin') }}:{{ hostvars[item]['api_password'] | default('admin') }}@{{ hostvars[item]['ansible_all_ipv6_addresses'] | ips_in_ranges(public_network.split(',')) | last | ipwrap }}:{{ hostvars[item]['api_port'] | default(5000) }}" + command: "{{ container_exec_cmd }} ceph --cluster {{ cluster }} dashboard iscsi-gateway-add {{ 'https' if hostvars[item]['api_secure'] | default(false) | bool else 'http' }}://{{ hostvars[item]['api_user'] | default('admin') }}:{{ hostvars[item]['api_password'] | default('admin') }}@{{ hostvars[item]['ansible_facts']['all_ipv6_addresses'] | ips_in_ranges(public_network.split(',')) | last | ipwrap }}:{{ hostvars[item]['api_port'] | default(5000) }}" changed_when: false delegate_to: "{{ groups[mon_group_name][0] }}" with_items: "{{ groups[iscsi_gw_group_name] }}" diff --git a/roles/ceph-dashboard/tasks/configure_dashboard_backends.yml b/roles/ceph-dashboard/tasks/configure_dashboard_backends.yml index 9538142a6..e2f7774b4 100644 --- a/roles/ceph-dashboard/tasks/configure_dashboard_backends.yml +++ b/roles/ceph-dashboard/tasks/configure_dashboard_backends.yml @@ -1,16 +1,16 @@ --- - name: get current mgr backend - ipv4 set_fact: - mgr_server_addr: "{{ hostvars[dashboard_backend]['ansible_all_ipv4_addresses'] | ips_in_ranges(public_network.split(',')) | first }}" + mgr_server_addr: "{{ hostvars[dashboard_backend]['ansible_facts']['all_ipv4_addresses'] | ips_in_ranges(public_network.split(',')) | first }}" when: ip_version == 'ipv4' - name: get current mgr backend - ipv6 set_fact: - mgr_server_addr: "{{ hostvars[dashboard_backend]['ansible_all_ipv6_addresses'] | ips_in_ranges(public_network.split(',')) | last }}" + mgr_server_addr: "{{ hostvars[dashboard_backend]['ansible_facts']['all_ipv6_addresses'] | ips_in_ranges(public_network.split(',')) | last }}" when: ip_version == 'ipv6' - name: config the current dashboard backend - command: "{{ container_exec_cmd }} ceph --cluster {{ cluster }} config set mgr mgr/dashboard/{{ hostvars[dashboard_backend]['ansible_hostname'] }}/server_addr {{ mgr_server_addr }}" + command: "{{ container_exec_cmd }} ceph --cluster {{ cluster }} config set mgr mgr/dashboard/{{ hostvars[dashboard_backend]['ansible_facts']['hostname'] }}/server_addr {{ mgr_server_addr }}" delegate_to: "{{ groups[mon_group_name][0] }}" changed_when: false run_once: true diff --git a/roles/ceph-dashboard/tasks/main.yml b/roles/ceph-dashboard/tasks/main.yml index ee3cbc528..723b31649 100644 --- a/roles/ceph-dashboard/tasks/main.yml +++ b/roles/ceph-dashboard/tasks/main.yml @@ -4,5 +4,5 @@ - name: print dashboard URL debug: - msg: "The dashboard has been deployed! You can access your dashboard web UI at {{ dashboard_protocol }}://{{ ansible_fqdn }}:{{ dashboard_port }}/ as an '{{ dashboard_admin_user }}' user with '{{ dashboard_admin_password }}' password." + msg: "The dashboard has been deployed! You can access your dashboard web UI at {{ dashboard_protocol }}://{{ ansible_facts['fqdn'] }}:{{ dashboard_port }}/ as an '{{ dashboard_admin_user }}' user with '{{ dashboard_admin_password }}' password." run_once: true diff --git a/roles/ceph-defaults/defaults/main.yml b/roles/ceph-defaults/defaults/main.yml index d16a0f8c0..8b86752c4 100644 --- a/roles/ceph-defaults/defaults/main.yml +++ b/roles/ceph-defaults/defaults/main.yml @@ -72,7 +72,7 @@ debian_package_dependencies: [] centos_package_dependencies: - epel-release - - "{{ 'python3-libselinux' if ansible_distribution_major_version | int >= 8 else 'libselinux-python' }}" + - "{{ (ansible_facts['distribution_major_version'] is version('8', '>=')) | ternary('python3-libselinux', 'libselinux-python') }}" redhat_package_dependencies: [] @@ -142,7 +142,7 @@ nfs_ganesha_stable_deb_repo: "{{ ceph_mirror }}/nfs-ganesha/deb-{{ nfs_ganesha_s # Use the option below to specify your applicable package tree, eg. when using non-LTS Ubuntu versions # # for a list of available Debian distributions, visit http://download.ceph.com/debian-{{ ceph_stable_release }}/dists/ # for more info read: https://github.com/ceph/ceph-ansible/issues/305 -#ceph_stable_distro_source: "{{ ansible_distribution_release }}" +#ceph_stable_distro_source: "{{ ansible_facts['distribution_release'] }}" # REPOSITORY: RHCS VERSION RED HAT STORAGE (from 4.0) @@ -170,7 +170,7 @@ ceph_rhcs_repository_path: "{{ ceph_stable_rh_storage_repository_path | default( # ceph_stable_repo_uca: "http://ubuntu-cloud.archive.canonical.com/ubuntu" ceph_stable_openstack_release_uca: queens -ceph_stable_release_uca: "{{ ansible_distribution_release }}-updates/{{ ceph_stable_openstack_release_uca }}" +ceph_stable_release_uca: "{{ ansible_facts['distribution_release'] }}-updates/{{ ceph_stable_openstack_release_uca }}" # REPOSITORY: openSUSE OBS # @@ -180,7 +180,7 @@ ceph_stable_release_uca: "{{ ansible_distribution_release }}-updates/{{ ceph_sta # usually has newer Ceph releases than the normal distro repository. # # -ceph_obs_repo: "https://download.opensuse.org/repositories/filesystems:/ceph:/{{ ceph_stable_release }}/openSUSE_Leap_{{ ansible_distribution_version }}/" +ceph_obs_repo: "https://download.opensuse.org/repositories/filesystems:/ceph:/{{ ceph_stable_release }}/openSUSE_Leap_{{ ansible_facts['distribution_version'] }}/" # REPOSITORY: DEV # @@ -243,7 +243,7 @@ generate_fsid: true ceph_conf_key_directory: /etc/ceph -ceph_uid: "{{ '64045' if not containerized_deployment | bool and ansible_os_family == 'Debian' else '167' }}" +ceph_uid: "{{ '64045' if not containerized_deployment | bool and ansible_facts['os_family'] == 'Debian' else '167' }}" # Permissions for keyring files in /etc/ceph ceph_keyring_permissions: '0600' @@ -541,7 +541,7 @@ rgw_zone: default # global: # foo: 1234 # bar: 5678 -# "client.rgw.{{ hostvars[groups.get(rgw_group_name)[0]]['ansible_hostname'] }}": +# "client.rgw.{{ hostvars[groups.get(rgw_group_name)[0]]['ansible_facts']['hostname'] }}": # rgw_zone: zone1 # ceph_conf_overrides: {} @@ -877,4 +877,4 @@ use_fqdn_yes_i_am_sure: false container_exec_cmd: docker: false -ceph_volume_debug: "{{ enable_ceph_volume_debug | ternary(1, 0) }}" \ No newline at end of file +ceph_volume_debug: "{{ enable_ceph_volume_debug | ternary(1, 0) }}" diff --git a/roles/ceph-facts/tasks/container_binary.yml b/roles/ceph-facts/tasks/container_binary.yml index 8e866bcbf..44249419f 100644 --- a/roles/ceph-facts/tasks/container_binary.yml +++ b/roles/ceph-facts/tasks/container_binary.yml @@ -6,4 +6,4 @@ - name: set_fact container_binary set_fact: - container_binary: "{{ 'podman' if (podman_binary.stat.exists and ansible_distribution == 'Fedora') or (ansible_os_family == 'RedHat' and ansible_distribution_major_version == '8') else 'docker' }}" \ No newline at end of file + container_binary: "{{ 'podman' if (podman_binary.stat.exists and ansible_facts['distribution'] == 'Fedora') or (ansible_facts['os_family'] == 'RedHat' and ansible_facts['distribution_major_version'] == '8') else 'docker' }}" \ No newline at end of file diff --git a/roles/ceph-facts/tasks/facts.yml b/roles/ceph-facts/tasks/facts.yml index 3bb2f39c5..a288b82ac 100644 --- a/roles/ceph-facts/tasks/facts.yml +++ b/roles/ceph-facts/tasks/facts.yml @@ -28,9 +28,9 @@ set_fact: ceph_release: "{{ ceph_stable_release }}" -- name: set_fact monitor_name ansible_hostname +- name: set_fact monitor_name ansible_facts['hostname'] set_fact: - monitor_name: "{{ hostvars[item]['ansible_hostname'] }}" + monitor_name: "{{ hostvars[item]['ansible_facts']['hostname'] }}" delegate_to: "{{ item }}" delegate_facts: true with_items: "{{ groups.get(mon_group_name, []) }}" @@ -39,9 +39,9 @@ - groups.get(mon_group_name, []) | length > 0 - not mon_use_fqdn | bool -- name: set_fact monitor_name ansible_fqdn +- name: set_fact monitor_name ansible_facts['fqdn'] set_fact: - monitor_name: "{{ hostvars[item]['ansible_fqdn'] }}" + monitor_name: "{{ hostvars[item]['ansible_facts']['fqdn'] }}" delegate_to: "{{ item }}" delegate_facts: true with_items: "{{ groups.get(mon_group_name, []) }}" @@ -55,12 +55,12 @@ block: - name: set_fact container_exec_cmd set_fact: - container_exec_cmd: "{{ container_binary }} exec ceph-mon-{{ hostvars[groups[mon_group_name][0]]['ansible_hostname'] if not rolling_update else hostvars[mon_host | default(groups[mon_group_name][0])]['ansible_hostname'] }}" + container_exec_cmd: "{{ container_binary }} exec ceph-mon-{{ hostvars[groups[mon_group_name][0]]['ansible_facts']['hostname'] if not rolling_update | bool else hostvars[mon_host | default(groups[mon_group_name][0])]['ansible_facts']['hostname'] }}" when: - containerized_deployment | bool - name: find a running mon container - command: "{{ container_binary }} ps -q --filter name=ceph-mon-{{ hostvars[item]['ansible_hostname'] }}" + command: "{{ container_binary }} ps -q --filter name=ceph-mon-{{ hostvars[item]['ansible_facts']['hostname'] }}" register: find_running_mon_container failed_when: false run_once: true @@ -115,7 +115,7 @@ - name: set_fact _container_exec_cmd set_fact: - _container_exec_cmd: "{{ container_binary }} exec ceph-mon-{{ hostvars[groups[mon_group_name][0] if running_mon is undefined else running_mon]['ansible_hostname'] }}" + _container_exec_cmd: "{{ container_binary }} exec ceph-mon-{{ hostvars[groups[mon_group_name][0] if running_mon is undefined else running_mon]['ansible_facts']['hostname'] }}" when: - containerized_deployment | bool @@ -142,7 +142,7 @@ when: rolling_update or groups.get(mon_group_name, []) | length == 0 - name: get current fsid - command: "{{ timeout_command }} {{ container_exec_cmd }} ceph --admin-daemon /var/run/ceph/{{ cluster }}-mon.{{ hostvars[mon_host | default(groups[mon_group_name][0])]['ansible_hostname'] }}.asok config get fsid" + command: "{{ timeout_command }} {{ container_exec_cmd }} ceph --admin-daemon /var/run/ceph/{{ cluster }}-mon.{{ hostvars[mon_host | default(groups[mon_group_name][0])]['ansible_facts']['hostname'] }}.asok config get fsid" register: rolling_update_fsid delegate_to: "{{ mon_host | default(groups[mon_group_name][0]) }}" until: rolling_update_fsid is succeeded @@ -179,14 +179,14 @@ set_fact: fsid: "{{ cluster_uuid.stdout }}" -- name: set_fact mds_name ansible_hostname +- name: set_fact mds_name ansible_facts['hostname'] set_fact: - mds_name: "{{ ansible_hostname }}" + mds_name: "{{ ansible_facts['hostname'] }}" when: not mds_use_fqdn | bool -- name: set_fact mds_name ansible_fqdn +- name: set_fact mds_name ansible_facts['fqdn'] set_fact: - mds_name: "{{ ansible_fqdn }}" + mds_name: "{{ ansible_facts['fqdn'] }}" when: mds_use_fqdn | bool - name: resolve device link(s) @@ -276,10 +276,10 @@ - name: set_fact devices generate device list when osd_auto_discovery set_fact: devices: "{{ (devices | default([]) + [ item.key | regex_replace('^', '/dev/') ]) | unique }}" - with_dict: "{{ ansible_devices }}" + with_dict: "{{ ansible_facts['devices'] }}" when: - osd_auto_discovery | default(False) | bool - - ansible_devices is defined + - ansible_facts['devices'] is defined - item.value.removable == "0" - item.value.sectors != "0" - item.value.partitions|count == 0 @@ -308,9 +308,9 @@ - name: set_fact rgw_hostname set_fact: - rgw_hostname: "{% set _value = ansible_hostname -%} + rgw_hostname: "{% set _value = ansible_facts['hostname'] -%} {% for key in (ceph_current_status['services']['rgw']['daemons'] | list) -%} - {% if key == ansible_fqdn -%} + {% if key == ansible_facts['fqdn'] -%} {% set _value = key -%} {% endif -%} {% endfor -%} @@ -378,24 +378,24 @@ - name: set ntp service name for Debian family set_fact: ntp_service_name: ntp - when: ansible_os_family == 'Debian' + when: ansible_facts['os_family'] == 'Debian' - name: set ntp service name for Red Hat family set_fact: ntp_service_name: ntpd - when: ansible_os_family in ['RedHat', 'Suse'] + when: ansible_facts['os_family'] in ['RedHat', 'Suse'] - name: set chrony daemon name RedHat and Ubuntu based OSs block: - name: set chronyd daemon name for RedHat based OSs set_fact: chrony_daemon_name: chronyd - when: ansible_os_family in ["RedHat", "Suse"] + when: ansible_facts['os_family'] in ["RedHat", "Suse"] - name: set chronyd daemon name for Ubuntu based OSs set_fact: chrony_daemon_name: chrony - when: ansible_os_family == "Debian" + when: ansible_facts['os_family'] == "Debian" - name: set_fact use_new_ceph_iscsi package or old ceph-iscsi-config/cli set_fact: diff --git a/roles/ceph-facts/tasks/grafana.yml b/roles/ceph-facts/tasks/grafana.yml index 47bae2eac..b0f9b7842 100644 --- a/roles/ceph-facts/tasks/grafana.yml +++ b/roles/ceph-facts/tasks/grafana.yml @@ -1,6 +1,6 @@ - name: set grafana_server_addr fact - ipv4 set_fact: - grafana_server_addr: "{{ hostvars[inventory_hostname]['ansible_all_ipv4_addresses'] | ips_in_ranges(public_network.split(',')) | first }}" + grafana_server_addr: "{{ hostvars[inventory_hostname]['ansible_facts']['all_ipv4_addresses'] | ips_in_ranges(public_network.split(',')) | first }}" when: - groups.get(grafana_server_group_name, []) | length > 0 - ip_version == 'ipv4' @@ -9,7 +9,7 @@ - name: set grafana_server_addr fact - ipv6 set_fact: - grafana_server_addr: "{{ hostvars[inventory_hostname]['ansible_all_ipv6_addresses'] | ips_in_ranges(public_network.split(',')) | last | ipwrap }}" + grafana_server_addr: "{{ hostvars[inventory_hostname]['ansible_facts']['all_ipv6_addresses'] | ips_in_ranges(public_network.split(',')) | last | ipwrap }}" when: - groups.get(grafana_server_group_name, []) | length > 0 - ip_version == 'ipv6' @@ -18,7 +18,7 @@ - name: set grafana_server_addrs fact - ipv4 set_fact: - grafana_server_addrs: "{{ (grafana_server_addrs | default([]) + [hostvars[item]['ansible_all_ipv4_addresses'] | ips_in_ranges(public_network.split(',')) | first]) | unique }}" + grafana_server_addrs: "{{ (grafana_server_addrs | default([]) + [hostvars[item]['ansible_facts']['all_ipv4_addresses'] | ips_in_ranges(public_network.split(',')) | first]) | unique }}" with_items: "{{ groups.get(grafana_server_group_name, []) }}" when: - groups.get(grafana_server_group_name, []) | length > 0 @@ -27,7 +27,7 @@ - name: set grafana_server_addrs fact - ipv6 set_fact: - grafana_server_addrs: "{{ (grafana_server_addrs | default([]) + [hostvars[item]['ansible_all_ipv6_addresses'] | ips_in_ranges(public_network.split(',')) | last | ipwrap]) | unique }}" + grafana_server_addrs: "{{ (grafana_server_addrs | default([]) + [hostvars[item]['ansible_facts']['all_ipv6_addresses'] | ips_in_ranges(public_network.split(',')) | last | ipwrap]) | unique }}" with_items: "{{ groups.get(grafana_server_group_name, []) }}" when: - groups.get(grafana_server_group_name, []) | length > 0 diff --git a/roles/ceph-facts/tasks/set_monitor_address.yml b/roles/ceph-facts/tasks/set_monitor_address.yml index aee025aab..fb8b4690d 100644 --- a/roles/ceph-facts/tasks/set_monitor_address.yml +++ b/roles/ceph-facts/tasks/set_monitor_address.yml @@ -1,7 +1,7 @@ --- - name: set_fact _monitor_address to monitor_address_block ipv4 set_fact: - _monitor_addresses: "{{ _monitor_addresses | default([]) + [{ 'name': item, 'addr': hostvars[item]['ansible_all_ipv4_addresses'] | ips_in_ranges(hostvars[item]['monitor_address_block'].split(',')) | first }] }}" + _monitor_addresses: "{{ _monitor_addresses | default([]) + [{ 'name': item, 'addr': hostvars[item]['ansible_facts']['all_ipv4_addresses'] | ips_in_ranges(hostvars[item]['monitor_address_block'].split(',')) | first }] }}" with_items: "{{ groups.get(mon_group_name, []) }}" when: - "item not in _monitor_addresses | default([]) | selectattr('name', 'defined') | map(attribute='name') | list" @@ -11,7 +11,7 @@ - name: set_fact _monitor_address to monitor_address_block ipv6 set_fact: - _monitor_addresses: "{{ _monitor_addresses | default([]) + [{ 'name': item, 'addr': hostvars[item]['ansible_all_ipv6_addresses'] | ips_in_ranges(hostvars[item]['monitor_address_block'].split(',')) | last | ipwrap }] }}" + _monitor_addresses: "{{ _monitor_addresses | default([]) + [{ 'name': item, 'addr': hostvars[item]['ansible_facts']['all_ipv6_addresses'] | ips_in_ranges(hostvars[item]['monitor_address_block'].split(',')) | last | ipwrap }] }}" with_items: "{{ groups.get(mon_group_name, []) }}" when: - "item not in _monitor_addresses | default([]) | selectattr('name', 'defined') | map(attribute='name') | list" @@ -30,7 +30,7 @@ - name: set_fact _monitor_address to monitor_interface - ipv4 set_fact: - _monitor_addresses: "{{ _monitor_addresses | default([]) + [{ 'name': item, 'addr': hostvars[item]['ansible_' + (hostvars[item]['monitor_interface']|replace('-', '_'))][ip_version]['address'] | ipwrap }] }}" + _monitor_addresses: "{{ _monitor_addresses | default([]) + [{ 'name': item, 'addr': hostvars[item]['ansible_facts'][(hostvars[item]['monitor_interface']|replace('-', '_'))][ip_version]['address'] | ipwrap }] }}" with_items: "{{ groups.get(mon_group_name, []) }}" when: - "item not in _monitor_addresses | default([]) | selectattr('name', 'defined') | map(attribute='name') | list" @@ -41,7 +41,7 @@ - name: set_fact _monitor_address to monitor_interface - ipv6 set_fact: - _monitor_addresses: "{{ _monitor_addresses | default([]) + [{ 'name': item, 'addr': hostvars[item]['ansible_' + (hostvars[item]['monitor_interface']|replace('-', '_'))][ip_version][0]['address'] | ipwrap }] }}" + _monitor_addresses: "{{ _monitor_addresses | default([]) + [{ 'name': item, 'addr': hostvars[item]['ansible_facts'][(hostvars[item]['monitor_interface']|replace('-', '_'))][ip_version][0]['address'] | ipwrap }] }}" with_items: "{{ groups.get(mon_group_name, []) }}" when: - "item not in _monitor_addresses | default([]) | selectattr('name', 'defined') | map(attribute='name') | list" diff --git a/roles/ceph-facts/tasks/set_radosgw_address.yml b/roles/ceph-facts/tasks/set_radosgw_address.yml index e586a0401..b662340b7 100644 --- a/roles/ceph-facts/tasks/set_radosgw_address.yml +++ b/roles/ceph-facts/tasks/set_radosgw_address.yml @@ -1,7 +1,7 @@ --- - name: set_fact _radosgw_address to radosgw_address_block ipv4 set_fact: - _radosgw_address: "{{ hostvars[inventory_hostname]['ansible_all_ipv4_addresses'] | ips_in_ranges(hostvars[inventory_hostname]['radosgw_address_block'].split(',')) | first }}" + _radosgw_address: "{{ hostvars[inventory_hostname]['ansible_facts']['all_ipv4_addresses'] | ips_in_ranges(hostvars[inventory_hostname]['radosgw_address_block'].split(',')) | first }}" when: - radosgw_address_block is defined - radosgw_address_block != 'subnet' @@ -9,7 +9,7 @@ - name: set_fact _radosgw_address to radosgw_address_block ipv6 set_fact: - _radosgw_address: "{{ hostvars[inventory_hostname]['ansible_all_ipv6_addresses'] | ips_in_ranges(hostvars[inventory_hostname]['radosgw_address_block'].split(',')) | last | ipwrap }}" + _radosgw_address: "{{ hostvars[inventory_hostname]['ansible_facts']['all_ipv6_addresses'] | ips_in_ranges(hostvars[inventory_hostname]['radosgw_address_block'].split(',')) | last | ipwrap }}" when: - radosgw_address_block is defined - radosgw_address_block != 'subnet' @@ -30,16 +30,16 @@ block: - name: set_fact _interface set_fact: - _interface: "{{ 'ansible_' + (radosgw_interface | replace('-', '_')) }}" + _interface: "{{ (radosgw_interface | replace('-', '_')) }}" - name: set_fact _radosgw_address to radosgw_interface - ipv4 set_fact: - _radosgw_address: "{{ hostvars[inventory_hostname][_interface][ip_version]['address'] }}" + _radosgw_address: "{{ hostvars[inventory_hostname]['ansible_facts'][_interface][ip_version]['address'] }}" when: ip_version == 'ipv4' - name: set_fact _radosgw_address to radosgw_interface - ipv6 set_fact: - _radosgw_address: "{{ hostvars[inventory_hostname][_interface][ip_version][0]['address'] | ipwrap }}" + _radosgw_address: "{{ hostvars[inventory_hostname]['ansible_facts'][_interface][ip_version][0]['address'] | ipwrap }}" when: ip_version == 'ipv6' - name: set_fact rgw_instances without rgw multisite diff --git a/roles/ceph-grafana/tasks/configure_grafana.yml b/roles/ceph-grafana/tasks/configure_grafana.yml index 5275b89f9..57928d321 100644 --- a/roles/ceph-grafana/tasks/configure_grafana.yml +++ b/roles/ceph-grafana/tasks/configure_grafana.yml @@ -7,7 +7,7 @@ until: result is succeeded when: - not containerized_deployment | bool - - ansible_os_family in ['RedHat', 'Suse'] + - ansible_facts['os_family'] in ['RedHat', 'Suse'] tags: package-install - name: make sure grafana is down @@ -41,7 +41,7 @@ with_items: "{{ grafana_dashboard_files }}" when: - not containerized_deployment | bool - - not ansible_os_family in ['RedHat', 'Suse'] + - not ansible_facts['os_family'] in ['RedHat', 'Suse'] - name: write grafana.ini template: diff --git a/roles/ceph-grafana/templates/grafana.ini.j2 b/roles/ceph-grafana/templates/grafana.ini.j2 index 4c897b2de..61ef46c09 100644 --- a/roles/ceph-grafana/templates/grafana.ini.j2 +++ b/roles/ceph-grafana/templates/grafana.ini.j2 @@ -18,7 +18,7 @@ org_role = Viewer [server] cert_file = /etc/grafana/ceph-dashboard.crt cert_key = /etc/grafana/ceph-dashboard.key -domain = {{ ansible_fqdn }} +domain = {{ ansible_facts['fqdn'] }} protocol = {{ dashboard_protocol }} http_port = {{ grafana_port }} http_addr = {{ grafana_server_addr }} diff --git a/roles/ceph-handler/tasks/check_running_containers.yml b/roles/ceph-handler/tasks/check_running_containers.yml index 32d90f4a1..bcb98f857 100644 --- a/roles/ceph-handler/tasks/check_running_containers.yml +++ b/roles/ceph-handler/tasks/check_running_containers.yml @@ -1,6 +1,6 @@ --- - name: check for a mon container - command: "{{ container_binary }} ps -q --filter='name=ceph-mon-{{ ansible_hostname }}'" + command: "{{ container_binary }} ps -q --filter='name=ceph-mon-{{ ansible_facts['hostname'] }}'" register: ceph_mon_container_stat changed_when: false failed_when: false @@ -16,7 +16,7 @@ when: inventory_hostname in groups.get(osd_group_name, []) - name: check for a mds container - command: "{{ container_binary }} ps -q --filter='name=ceph-mds-{{ ansible_hostname }}'" + command: "{{ container_binary }} ps -q --filter='name=ceph-mds-{{ ansible_facts['hostname'] }}'" register: ceph_mds_container_stat changed_when: false failed_when: false @@ -24,7 +24,7 @@ when: inventory_hostname in groups.get(mds_group_name, []) - name: check for a rgw container - command: "{{ container_binary }} ps -q --filter='name=ceph-rgw-{{ ansible_hostname }}'" + command: "{{ container_binary }} ps -q --filter='name=ceph-rgw-{{ ansible_facts['hostname'] }}'" register: ceph_rgw_container_stat changed_when: false failed_when: false @@ -32,7 +32,7 @@ when: inventory_hostname in groups.get(rgw_group_name, []) - name: check for a mgr container - command: "{{ container_binary }} ps -q --filter='name=ceph-mgr-{{ ansible_hostname }}'" + command: "{{ container_binary }} ps -q --filter='name=ceph-mgr-{{ ansible_facts['hostname'] }}'" register: ceph_mgr_container_stat changed_when: false failed_when: false @@ -40,7 +40,7 @@ when: inventory_hostname in groups.get(mgr_group_name, []) - name: check for a rbd mirror container - command: "{{ container_binary }} ps -q --filter='name=ceph-rbd-mirror-{{ ansible_hostname }}'" + command: "{{ container_binary }} ps -q --filter='name=ceph-rbd-mirror-{{ ansible_facts['hostname'] }}'" register: ceph_rbd_mirror_container_stat changed_when: false failed_when: false @@ -48,7 +48,7 @@ when: inventory_hostname in groups.get(rbdmirror_group_name, []) - name: check for a nfs container - command: "{{ container_binary }} ps -q --filter='name=ceph-nfs-{{ ceph_nfs_service_suffix | default(ansible_hostname) }}'" + command: "{{ container_binary }} ps -q --filter='name=ceph-nfs-{{ ceph_nfs_service_suffix | default(ansible_facts['hostname']) }}'" register: ceph_nfs_container_stat changed_when: false failed_when: false @@ -80,7 +80,7 @@ when: inventory_hostname in groups.get(iscsi_gw_group_name, []) - name: check for a ceph-crash container - command: "{{ container_binary }} ps -q --filter='name=ceph-crash-{{ ansible_hostname }}'" + command: "{{ container_binary }} ps -q --filter='name=ceph-crash-{{ ansible_facts['hostname'] }}'" register: ceph_crash_container_stat changed_when: false failed_when: false diff --git a/roles/ceph-handler/tasks/handler_crash.yml b/roles/ceph-handler/tasks/handler_crash.yml index 6187fca79..b4039b10e 100644 --- a/roles/ceph-handler/tasks/handler_crash.yml +++ b/roles/ceph-handler/tasks/handler_crash.yml @@ -5,7 +5,7 @@ - name: restart the ceph-crash service systemd: - name: ceph-crash@{{ ansible_hostname }} + name: ceph-crash@{{ ansible_facts['hostname'] }} state: restarted enabled: yes masked: no diff --git a/roles/ceph-handler/templates/restart_mds_daemon.sh.j2 b/roles/ceph-handler/templates/restart_mds_daemon.sh.j2 index b40d579b2..441ddd8f1 100644 --- a/roles/ceph-handler/templates/restart_mds_daemon.sh.j2 +++ b/roles/ceph-handler/templates/restart_mds_daemon.sh.j2 @@ -3,13 +3,13 @@ RETRIES="{{ handler_health_mds_check_retries }}" DELAY="{{ handler_health_mds_check_delay }}" MDS_NAME="{{ mds_name }}" -{% if containerized_deployment %} -DOCKER_EXEC="{{ container_binary }} exec ceph-mds-{{ ansible_hostname }}" +{% if containerized_deployment | bool %} +DOCKER_EXEC="{{ container_binary }} exec ceph-mds-{{ ansible_facts['hostname'] }}" {% endif %} # Backward compatibility -$DOCKER_EXEC test -S /var/run/ceph/{{ cluster }}-mds.{{ ansible_fqdn }}.asok && SOCKET=/var/run/ceph/{{ cluster }}-mds.{{ ansible_fqdn }}.asok -$DOCKER_EXEC test -S /var/run/ceph/{{ cluster }}-mds.{{ ansible_hostname }}.asok && SOCKET=/var/run/ceph/{{ cluster }}-mds.{{ ansible_hostname }}.asok +$DOCKER_EXEC test -S /var/run/ceph/{{ cluster }}-mds.{{ ansible_facts['fqdn'] }}.asok && SOCKET=/var/run/ceph/{{ cluster }}-mds.{{ ansible_facts['fqdn'] }}.asok +$DOCKER_EXEC test -S /var/run/ceph/{{ cluster }}-mds.{{ ansible_facts['hostname'] }}.asok && SOCKET=/var/run/ceph/{{ cluster }}-mds.{{ ansible_facts['hostname'] }}.asok # First, restart the daemon systemctl restart ceph-mds@${MDS_NAME} diff --git a/roles/ceph-handler/templates/restart_mgr_daemon.sh.j2 b/roles/ceph-handler/templates/restart_mgr_daemon.sh.j2 index c9f355400..e20c61901 100644 --- a/roles/ceph-handler/templates/restart_mgr_daemon.sh.j2 +++ b/roles/ceph-handler/templates/restart_mgr_daemon.sh.j2 @@ -2,14 +2,14 @@ RETRIES="{{ handler_health_mgr_check_retries }}" DELAY="{{ handler_health_mgr_check_delay }}" -MGR_NAME="{{ ansible_hostname }}" -{% if containerized_deployment %} -DOCKER_EXEC="{{ container_binary }} exec ceph-mgr-{{ ansible_hostname }}" +MGR_NAME="{{ ansible_facts['hostname'] }}" +{% if containerized_deployment | bool %} +DOCKER_EXEC="{{ container_binary }} exec ceph-mgr-{{ ansible_facts['hostname'] }}" {% endif %} # Backward compatibility -$DOCKER_EXEC test -S /var/run/ceph/{{ cluster }}-mgr.{{ ansible_fqdn }}.asok && SOCKET=/var/run/ceph/{{ cluster }}-mgr.{{ ansible_fqdn }}.asok -$DOCKER_EXEC test -S /var/run/ceph/{{ cluster }}-mgr.{{ ansible_hostname }}.asok && SOCKET=/var/run/ceph/{{ cluster }}-mgr.{{ ansible_hostname }}.asok +$DOCKER_EXEC test -S /var/run/ceph/{{ cluster }}-mgr.{{ ansible_facts['fqdn'] }}.asok && SOCKET=/var/run/ceph/{{ cluster }}-mgr.{{ ansible_facts['fqdn'] }}.asok +$DOCKER_EXEC test -S /var/run/ceph/{{ cluster }}-mgr.{{ ansible_facts['hostname'] }}.asok && SOCKET=/var/run/ceph/{{ cluster }}-mgr.{{ ansible_facts['hostname'] }}.asok systemctl reset-failed ceph-mgr@${MGR_NAME} # First, restart the daemon diff --git a/roles/ceph-handler/templates/restart_mon_daemon.sh.j2 b/roles/ceph-handler/templates/restart_mon_daemon.sh.j2 index 06383eae1..2beade15a 100644 --- a/roles/ceph-handler/templates/restart_mon_daemon.sh.j2 +++ b/roles/ceph-handler/templates/restart_mon_daemon.sh.j2 @@ -3,19 +3,19 @@ RETRIES="{{ handler_health_mon_check_retries }}" DELAY="{{ handler_health_mon_check_delay }}" MONITOR_NAME="{{ monitor_name }}" -{% if containerized_deployment %} -DOCKER_EXEC="{{ container_binary }} exec ceph-mon-{{ ansible_hostname }}" +{% if containerized_deployment | bool %} +DOCKER_EXEC="{{ container_binary }} exec ceph-mon-{{ ansible_facts['hostname'] }}" {% endif %} # if daemon is uninstalled, no restarting is needed; so exit with success -systemctl status ceph-mon@{{ ansible_hostname }} > /dev/null +systemctl status ceph-mon@{{ ansible_facts['hostname'] }} > /dev/null if [[ $? -ne 0 ]]; then exit 0 fi # Backward compatibility -$DOCKER_EXEC test -S /var/run/ceph/{{ cluster }}-mon.{{ ansible_fqdn }}.asok && SOCKET=/var/run/ceph/{{ cluster }}-mon.{{ ansible_fqdn }}.asok -$DOCKER_EXEC test -S /var/run/ceph/{{ cluster }}-mon.{{ ansible_hostname }}.asok && SOCKET=/var/run/ceph/{{ cluster }}-mon.{{ ansible_hostname }}.asok +$DOCKER_EXEC test -S /var/run/ceph/{{ cluster }}-mon.{{ ansible_facts['fqdn'] }}.asok && SOCKET=/var/run/ceph/{{ cluster }}-mon.{{ ansible_facts['fqdn'] }}.asok +$DOCKER_EXEC test -S /var/run/ceph/{{ cluster }}-mon.{{ ansible_facts['hostname'] }}.asok && SOCKET=/var/run/ceph/{{ cluster }}-mon.{{ ansible_facts['hostname'] }}.asok check_quorum() { while [ $RETRIES -ne 0 ]; do @@ -34,7 +34,7 @@ exit 1 } # First, restart the daemon -systemctl restart ceph-mon@{{ ansible_hostname }} +systemctl restart ceph-mon@{{ ansible_facts['hostname'] }} COUNT=10 # Wait and ensure the socket exists after restarting the daemon @@ -45,5 +45,5 @@ while [ $COUNT -ne 0 ]; do done # If we reach this point, it means the socket is not present. echo "Socket file ${SOCKET} could not be found, which means the monitor is not running. Showing ceph-mon unit logs now:" -journalctl -u ceph-mon@{{ ansible_hostname }} +journalctl -u ceph-mon@{{ ansible_facts['hostname'] }} exit 1 diff --git a/roles/ceph-handler/templates/restart_nfs_daemon.sh.j2 b/roles/ceph-handler/templates/restart_nfs_daemon.sh.j2 index 0fab87cc0..d7432a3a7 100644 --- a/roles/ceph-handler/templates/restart_nfs_daemon.sh.j2 +++ b/roles/ceph-handler/templates/restart_nfs_daemon.sh.j2 @@ -2,10 +2,10 @@ RETRIES="{{ handler_health_nfs_check_retries }}" DELAY="{{ handler_health_nfs_check_delay }}" -NFS_NAME="ceph-nfs@{{ ceph_nfs_service_suffix | default(ansible_hostname) }}" +NFS_NAME="ceph-nfs@{{ ceph_nfs_service_suffix | default(ansible_facts['hostname']) }}" PID=/var/run/ganesha.pid -{% if containerized_deployment %} -DOCKER_EXEC="{{ container_binary }} exec ceph-nfs-{{ ceph_nfs_service_suffix | default(ansible_hostname) }}" +{% if containerized_deployment | bool %} +DOCKER_EXEC="{{ container_binary }} exec ceph-nfs-{{ ceph_nfs_service_suffix | default(ansible_facts['hostname']) }}" {% endif %} # First, restart the daemon diff --git a/roles/ceph-handler/templates/restart_rbd_mirror_daemon.sh.j2 b/roles/ceph-handler/templates/restart_rbd_mirror_daemon.sh.j2 index 5fb23454a..249346935 100644 --- a/roles/ceph-handler/templates/restart_rbd_mirror_daemon.sh.j2 +++ b/roles/ceph-handler/templates/restart_rbd_mirror_daemon.sh.j2 @@ -2,14 +2,14 @@ RETRIES="{{ handler_health_rbd_mirror_check_retries }}" DELAY="{{ handler_health_rbd_mirror_check_delay }}" -RBD_MIRROR_NAME="{{ ansible_hostname }}" +RBD_MIRROR_NAME="{{ ansible_facts['hostname'] }}" {% if containerized_deployment %} -DOCKER_EXEC="{{ container_binary }} exec ceph-rbd-mirror-{{ ansible_hostname }}" +DOCKER_EXEC="{{ container_binary }} exec ceph-rbd-mirror-{{ ansible_facts['hostname'] }}" {% endif %} # Backward compatibility -$DOCKER_EXEC test -S /var/run/ceph/{{ cluster }}-client.rbd-mirror.{{ ansible_fqdn }}.asok && SOCKET=/var/run/ceph/{{ cluster }}-client.rbd-mirror.{{ ansible_fqdn }}.asok -$DOCKER_EXEC test -S /var/run/ceph/{{ cluster }}-client.rbd-mirror.{{ ansible_hostname }}.asok && SOCKET=/var/run/ceph/{{ cluster }}-client.rbd-mirror.{{ ansible_hostname }}.asok +$DOCKER_EXEC test -S /var/run/ceph/{{ cluster }}-client.rbd-mirror.{{ ansible_facts['fqdn'] }}.asok && SOCKET=/var/run/ceph/{{ cluster }}-client.rbd-mirror.{{ ansible_facts['fqdn'] }}.asok +$DOCKER_EXEC test -S /var/run/ceph/{{ cluster }}-client.rbd-mirror.{{ ansible_facts['hostname'] }}.asok && SOCKET=/var/run/ceph/{{ cluster }}-client.rbd-mirror.{{ ansible_facts['hostname'] }}.asok # First, restart the daemon systemctl restart ceph-rbd-mirror@rbd-mirror.${RBD_MIRROR_NAME} diff --git a/roles/ceph-handler/templates/restart_rgw_daemon.sh.j2 b/roles/ceph-handler/templates/restart_rgw_daemon.sh.j2 index 62a5c3118..93eda64b8 100644 --- a/roles/ceph-handler/templates/restart_rgw_daemon.sh.j2 +++ b/roles/ceph-handler/templates/restart_rgw_daemon.sh.j2 @@ -2,7 +2,7 @@ RETRIES="{{ handler_health_rgw_check_retries }}" DELAY="{{ handler_health_rgw_check_delay }}" -HOST_NAME="{{ ansible_hostname }}" +HOST_NAME="{{ ansible_facts['hostname'] }}" RGW_NUMS={{ rgw_instances | length | int }} RGW_FRONTEND_SSL_CERT={{ radosgw_frontend_ssl_certificate }} if [ -n "$RGW_FRONTEND_SSL_CERT" ]; then diff --git a/roles/ceph-infra/tasks/configure_firewall.yml b/roles/ceph-infra/tasks/configure_firewall.yml index c43919c08..9be6c670e 100644 --- a/roles/ceph-infra/tasks/configure_firewall.yml +++ b/roles/ceph-infra/tasks/configure_firewall.yml @@ -15,7 +15,7 @@ block: - name: install firewalld python binding package: - name: "python{{ ansible_python.version.major }}-firewall" + name: "python{{ ansible_facts['python']['version']['major'] }}-firewall" tags: with_pkg when: not is_atomic | bool diff --git a/roles/ceph-infra/tasks/main.yml b/roles/ceph-infra/tasks/main.yml index 40d5876a1..e9620327b 100644 --- a/roles/ceph-infra/tasks/main.yml +++ b/roles/ceph-infra/tasks/main.yml @@ -2,7 +2,7 @@ - name: update cache for Debian based OSs apt: update_cache: yes - when: ansible_os_family == "Debian" + when: ansible_facts['os_family'] == "Debian" register: result until: result is succeeded @@ -10,7 +10,7 @@ include_tasks: configure_firewall.yml when: - configure_firewall | bool - - ansible_os_family in ['RedHat', 'Suse'] + - ansible_facts['os_family'] in ['RedHat', 'Suse'] tags: configure_firewall - name: include_tasks setup_ntp.yml diff --git a/roles/ceph-iscsi-gw/defaults/main.yml b/roles/ceph-iscsi-gw/defaults/main.yml index 50d8c9766..c161e3647 100644 --- a/roles/ceph-iscsi-gw/defaults/main.yml +++ b/roles/ceph-iscsi-gw/defaults/main.yml @@ -35,13 +35,13 @@ trusted_ip_list: 192.168.122.1 # These options can be passed using the 'ceph_mds_docker_extra_env' variable. # TCMU_RUNNER resource limitation -ceph_tcmu_runner_docker_memory_limit: "{{ ansible_memtotal_mb }}m" +ceph_tcmu_runner_docker_memory_limit: "{{ ansible_facts['memtotal_mb'] }}m" ceph_tcmu_runner_docker_cpu_limit: 1 # RBD_TARGET_GW resource limitation -ceph_rbd_target_gw_docker_memory_limit: "{{ ansible_memtotal_mb }}m" +ceph_rbd_target_gw_docker_memory_limit: "{{ ansible_facts['memtotal_mb'] }}m" ceph_rbd_target_gw_docker_cpu_limit: 1 # RBD_TARGET_API resource limitation -ceph_rbd_target_api_docker_memory_limit: "{{ ansible_memtotal_mb }}m" +ceph_rbd_target_api_docker_memory_limit: "{{ ansible_facts['memtotal_mb'] }}m" ceph_rbd_target_api_docker_cpu_limit: 1 diff --git a/roles/ceph-iscsi-gw/tasks/common.yml b/roles/ceph-iscsi-gw/tasks/common.yml index e566e16df..292903e35 100644 --- a/roles/ceph-iscsi-gw/tasks/common.yml +++ b/roles/ceph-iscsi-gw/tasks/common.yml @@ -24,7 +24,7 @@ - name: add mgr ip address to trusted list with dashboard - ipv4 set_fact: - trusted_ip_list: '{{ trusted_ip_list }},{{ hostvars[item]["ansible_all_ipv4_addresses"] | ips_in_ranges(public_network.split(",")) | first }}' + trusted_ip_list: '{{ trusted_ip_list }},{{ hostvars[item]["ansible_facts"]["all_ipv4_addresses"] | ips_in_ranges(public_network.split(",")) | first }}' with_items: '{{ groups[mgr_group_name] | default(groups[mon_group_name]) }}' when: - dashboard_enabled | bool @@ -32,7 +32,7 @@ - name: add mgr ip address to trusted list with dashboard - ipv6 set_fact: - trusted_ip_list: '{{ trusted_ip_list }},{{ hostvars[item]["ansible_all_ipv6_addresses"] | ips_in_ranges(public_network.split(",")) | last }}' + trusted_ip_list: '{{ trusted_ip_list }},{{ hostvars[item]["ansible_facts"]["all_ipv6_addresses"] | ips_in_ranges(public_network.split(",")) | last }}' with_items: '{{ groups[mgr_group_name] | default(groups[mon_group_name]) }}' when: - dashboard_enabled | bool @@ -49,7 +49,7 @@ - name: set_fact container_exec_cmd set_fact: - container_exec_cmd: "{{ container_binary }} exec ceph-mon-{{ hostvars[groups[mon_group_name][0]]['ansible_hostname'] }}" + container_exec_cmd: "{{ container_binary }} exec ceph-mon-{{ hostvars[groups[mon_group_name][0]]['ansible_facts']['hostname'] }}" delegate_to: "{{ groups[mon_group_name][0] }}" when: containerized_deployment | bool diff --git a/roles/ceph-iscsi-gw/tasks/deploy_ssl_keys.yml b/roles/ceph-iscsi-gw/tasks/deploy_ssl_keys.yml index cd98ff13a..83b03e799 100644 --- a/roles/ceph-iscsi-gw/tasks/deploy_ssl_keys.yml +++ b/roles/ceph-iscsi-gw/tasks/deploy_ssl_keys.yml @@ -33,7 +33,7 @@ command: > openssl req -newkey rsa:2048 -nodes -keyout {{ iscsi_ssl_tmp_dir.path }}/iscsi-gateway.key -x509 -days 365 -out {{ iscsi_ssl_tmp_dir.path }}/iscsi-gateway.crt - -subj "/C=US/ST=./L=./O=RedHat/OU=Linux/CN={{ ansible_hostname }}" + -subj "/C=US/ST=./L=./O=RedHat/OU=Linux/CN={{ ansible_facts['hostname'] }}" delegate_to: localhost run_once: True with_items: "{{ crt_files_exist.results }}" diff --git a/roles/ceph-iscsi-gw/tasks/non-container/prerequisites.yml b/roles/ceph-iscsi-gw/tasks/non-container/prerequisites.yml index 0d8296daf..0dcad9f23 100644 --- a/roles/ceph-iscsi-gw/tasks/non-container/prerequisites.yml +++ b/roles/ceph-iscsi-gw/tasks/non-container/prerequisites.yml @@ -1,6 +1,6 @@ --- - name: red hat based systems tasks - when: ansible_os_family == 'RedHat' + when: ansible_facts['os_family'] == 'RedHat' block: - name: set_fact common_pkgs set_fact: @@ -76,7 +76,7 @@ - name: ceph-iscsi stable repository get_url: - url: 'https://download.ceph.com/ceph-iscsi/{{ "3" if use_new_ceph_iscsi | bool else "2" }}/rpm/el{{ ansible_distribution_major_version }}/ceph-iscsi.repo' + url: "https://download.ceph.com/ceph-iscsi/{{ '3' if use_new_ceph_iscsi | bool else '2' }}/rpm/el{{ ansible_facts['distribution_major_version'] }}/ceph-iscsi.repo" dest: /etc/yum.repos.d/ceph-iscsi.repo force: true when: ceph_repository == 'community' diff --git a/roles/ceph-mds/defaults/main.yml b/roles/ceph-mds/defaults/main.yml index 0b443ea9a..e249d02eb 100644 --- a/roles/ceph-mds/defaults/main.yml +++ b/roles/ceph-mds/defaults/main.yml @@ -19,13 +19,13 @@ copy_admin_key: false # For the whole list of limits you can apply see: docs.docker.com/engine/admin/resource_constraints # Default values are based from: https://access.redhat.com/documentation/en-us/red_hat_ceph_storage/2/html/red_hat_ceph_storage_hardware_guide/minimum_recommendations # These options can be passed using the 'ceph_mds_docker_extra_env' variable. -ceph_mds_docker_memory_limit: "{{ ansible_memtotal_mb }}m" +ceph_mds_docker_memory_limit: "{{ ansible_facts['memtotal_mb'] }}m" ceph_mds_docker_cpu_limit: 4 # we currently for MDS_NAME to hostname because of a bug in ceph-docker # fix here: https://github.com/ceph/ceph-docker/pull/770 # this will go away soon. -ceph_mds_docker_extra_env: -e MDS_NAME={{ ansible_hostname }} +ceph_mds_docker_extra_env: -e MDS_NAME={{ ansible_facts['hostname'] }} ceph_config_keys: [] # DON'T TOUCH ME diff --git a/roles/ceph-mds/tasks/containerized.yml b/roles/ceph-mds/tasks/containerized.yml index 08c4b0e79..8919366ba 100644 --- a/roles/ceph-mds/tasks/containerized.yml +++ b/roles/ceph-mds/tasks/containerized.yml @@ -4,14 +4,14 @@ - name: systemd start mds container systemd: - name: ceph-mds@{{ ansible_hostname }} + name: ceph-mds@{{ ansible_facts['hostname'] }} state: started enabled: yes masked: no daemon_reload: yes - name: wait for mds socket to exist - command: "{{ container_binary }} exec ceph-mds-{{ ansible_hostname }} sh -c 'stat /var/run/ceph/{{ cluster }}-mds.{{ ansible_hostname }}.asok || stat /var/run/ceph/{{ cluster }}-mds.{{ ansible_fqdn }}.asok'" + command: "{{ container_binary }} exec ceph-mds-{{ ansible_facts['hostname'] }} sh -c 'stat /var/run/ceph/{{ cluster }}-mds.{{ ansible_facts['hostname'] }}.asok || stat /var/run/ceph/{{ cluster }}-mds.{{ ansible_facts['fqdn'] }}.asok'" changed_when: false register: multi_mds_socket retries: 5 diff --git a/roles/ceph-mds/tasks/non_containerized.yml b/roles/ceph-mds/tasks/non_containerized.yml index 6f8f15a61..0e2314d7a 100644 --- a/roles/ceph-mds/tasks/non_containerized.yml +++ b/roles/ceph-mds/tasks/non_containerized.yml @@ -3,10 +3,10 @@ apt: name: ceph-mds state: "{{ (upgrade_ceph_packages|bool) | ternary('latest','present') }}" - default_release: "{{ ceph_stable_release_uca | default('') if ceph_origin == 'repository' and ceph_repository == 'uca' else '' }}{{ ansible_distribution_release ~ '-backports' if ceph_origin == 'distro' and ceph_use_distro_backports else '' }}" + default_release: "{{ ceph_stable_release_uca | default('') if ceph_origin == 'repository' and ceph_repository == 'uca' else '' }}{{ ansible_facts['distribution_release'] ~ '-backports' if ceph_origin == 'distro' and ceph_use_distro_backports | bool else '' }}" when: - mds_group_name in group_names - - ansible_os_family == 'Debian' + - ansible_facts['os_family'] == 'Debian' register: result until: result is succeeded @@ -18,7 +18,7 @@ until: result is succeeded when: - mds_group_name in group_names - - ansible_os_family in ['Suse', 'RedHat'] + - ansible_facts['os_family'] in ['Suse', 'RedHat'] - name: create mds keyring command: ceph --cluster {{ cluster }} --name client.bootstrap-mds --keyring /var/lib/ceph/bootstrap-mds/{{ cluster }}.keyring auth get-or-create mds.{{ mds_name }} osd 'allow rwx' mds 'allow' mon 'allow profile mds' -o /var/lib/ceph/mds/{{ cluster }}-{{ mds_name }}/keyring @@ -41,7 +41,7 @@ path: "/etc/systemd/system/ceph-mds@.service.d/" when: - ceph_mds_systemd_overrides is defined - - ansible_service_mgr == 'systemd' + - ansible_facts['service_mgr'] == 'systemd' - name: add ceph-mds systemd service overrides config_template: @@ -51,7 +51,7 @@ config_type: "ini" when: - ceph_mds_systemd_overrides is defined - - ansible_service_mgr == 'systemd' + - ansible_facts['service_mgr'] == 'systemd' - name: start and add that the metadata service to the init sequence service: diff --git a/roles/ceph-mds/templates/ceph-mds.service.j2 b/roles/ceph-mds/templates/ceph-mds.service.j2 index 6d60fd25a..27dc48e87 100644 --- a/roles/ceph-mds/templates/ceph-mds.service.j2 +++ b/roles/ceph-mds/templates/ceph-mds.service.j2 @@ -6,17 +6,17 @@ Requires=docker.service {% else %} After=network.target {% endif %} -{% set cpu_limit = ansible_processor_vcpus|int if ceph_mds_docker_cpu_limit|int > ansible_processor_vcpus|int else ceph_mds_docker_cpu_limit|int %} +{% set cpu_limit = ansible_facts['processor_vcpus']|int if ceph_mds_docker_cpu_limit|int > ansible_facts['processor_vcpus']|int else ceph_mds_docker_cpu_limit|int %} [Service] EnvironmentFile=-/etc/environment {% if container_binary == 'podman' %} ExecStartPre=-/usr/bin/rm -f /%t/%n-pid /%t/%n-cid -ExecStartPre=-/usr/bin/{{ container_binary }} rm --storage ceph-mds-{{ ansible_hostname }} +ExecStartPre=-/usr/bin/{{ container_binary }} rm --storage ceph-mds-{{ ansible_facts['hostname'] }} {% else %} -ExecStartPre=-/usr/bin/{{ container_binary }} stop ceph-mds-{{ ansible_hostname }} +ExecStartPre=-/usr/bin/{{ container_binary }} stop ceph-mds-{{ ansible_facts['hostname'] }} {% endif %} -ExecStartPre=-/usr/bin/{{ container_binary }} rm ceph-mds-{{ ansible_hostname }} +ExecStartPre=-/usr/bin/{{ container_binary }} rm ceph-mds-{{ ansible_facts['hostname'] }} ExecStart=/usr/bin/{{ container_binary }} run --rm --net=host \ {% if container_binary == 'podman' %} -d --log-driver journald --conmon-pidfile /%t/%n-pid --cidfile /%t/%n-cid \ @@ -32,12 +32,12 @@ ExecStart=/usr/bin/{{ container_binary }} run --rm --net=host \ -e CEPH_DAEMON=MDS \ -e CONTAINER_IMAGE={{ ceph_docker_registry }}/{{ ceph_docker_image }}:{{ ceph_docker_image_tag }} \ {{ ceph_mds_docker_extra_env }} \ - --name=ceph-mds-{{ ansible_hostname }} \ + --name=ceph-mds-{{ ansible_facts['hostname'] }} \ {{ ceph_docker_registry }}/{{ ceph_docker_image }}:{{ ceph_docker_image_tag }} {% if container_binary == 'podman' %} ExecStop=-/usr/bin/sh -c "/usr/bin/{{ container_binary }} rm -f `cat /%t/%n-cid`" {% else %} -ExecStopPost=-/usr/bin/{{ container_binary }} stop ceph-mds-{{ ansible_hostname }} +ExecStopPost=-/usr/bin/{{ container_binary }} stop ceph-mds-{{ ansible_facts['hostname'] }} {% endif %} KillMode=none Restart=always diff --git a/roles/ceph-mgr/defaults/main.yml b/roles/ceph-mgr/defaults/main.yml index 053f623d7..7455043d5 100644 --- a/roles/ceph-mgr/defaults/main.yml +++ b/roles/ceph-mgr/defaults/main.yml @@ -35,7 +35,7 @@ ceph_mgr_packages: # For the whole list of limits you can apply see: docs.docker.com/engine/admin/resource_constraints # Default values are based from: https://access.redhat.com/documentation/en-us/red_hat_ceph_storage/2/html/red_hat_ceph_storage_hardware_guide/minimum_recommendations # These options can be passed using the 'ceph_mgr_docker_extra_env' variable. -ceph_mgr_docker_memory_limit: "{{ ansible_memtotal_mb }}m" +ceph_mgr_docker_memory_limit: "{{ ansible_facts['memtotal_mb'] }}m" ceph_mgr_docker_cpu_limit: 1 ceph_mgr_docker_extra_env: diff --git a/roles/ceph-mgr/tasks/common.yml b/roles/ceph-mgr/tasks/common.yml index e924dcac2..9d2d82ec4 100644 --- a/roles/ceph-mgr/tasks/common.yml +++ b/roles/ceph-mgr/tasks/common.yml @@ -1,7 +1,7 @@ --- - name: create mgr directory file: - path: /var/lib/ceph/mgr/{{ cluster }}-{{ ansible_hostname }} + path: /var/lib/ceph/mgr/{{ cluster }}-{{ ansible_facts['hostname'] }} state: directory owner: "{{ ceph_uid if containerized_deployment else 'ceph' }}" group: "{{ ceph_uid if containerized_deployment else 'ceph' }}" @@ -9,7 +9,7 @@ - name: fetch ceph mgr keyring ceph_key: - name: "mgr.{{ ansible_hostname }}" + name: "mgr.{{ ansible_facts['hostname'] }}" caps: mon: allow profile mgr osd: allow * @@ -19,7 +19,7 @@ owner: "{{ ceph_uid if containerized_deployment else 'ceph' }}" group: "{{ ceph_uid if containerized_deployment else 'ceph' }}" mode: "0400" - dest: "/var/lib/ceph/mgr/{{ cluster }}-{{ ansible_hostname }}/keyring" + dest: "/var/lib/ceph/mgr/{{ cluster }}-{{ ansible_facts['hostname'] }}/keyring" environment: CEPH_CONTAINER_IMAGE: "{{ ceph_docker_registry + '/' + ceph_docker_image + ':' + ceph_docker_image_tag if containerized_deployment else None }}" CEPH_CONTAINER_BINARY: "{{ container_binary }}" @@ -30,7 +30,7 @@ block: - name: create ceph mgr keyring(s) on a mon node ceph_key: - name: "mgr.{{ hostvars[item]['ansible_hostname'] }}" + name: "mgr.{{ hostvars[item]['ansible_facts']['hostname'] }}" caps: mon: allow profile mgr osd: allow * @@ -51,7 +51,7 @@ set_fact: _mgr_keys: - { 'name': 'client.admin', 'path': "/etc/ceph/{{ cluster }}.client.admin.keyring", 'copy_key': copy_admin_key } - - { 'name': "mgr.{{ ansible_hostname }}", 'path': "/var/lib/ceph/mgr/{{ cluster }}-{{ ansible_hostname }}/keyring", 'copy_key': true } + - { 'name': "mgr.{{ ansible_facts['hostname'] }}", 'path': "/var/lib/ceph/mgr/{{ cluster }}-{{ ansible_facts['hostname'] }}/keyring", 'copy_key': true } - name: get keys from monitors command: "{{ _container_exec_cmd | default('') }} ceph --cluster {{ cluster }} auth get {{ item.name }}" @@ -77,8 +77,8 @@ - name: set mgr key permissions file: - path: /var/lib/ceph/mgr/{{ cluster }}-{{ ansible_hostname }}/keyring - owner: "{{ ceph_uid if containerized_deployment else 'ceph' }}" - group: "{{ ceph_uid if containerized_deployment else 'ceph' }}" + path: /var/lib/ceph/mgr/{{ cluster }}-{{ ansible_facts['hostname'] }}/keyring + owner: "{{ ceph_uid if containerized_deployment | bool else 'ceph' }}" + group: "{{ ceph_uid if containerized_deployment | bool else 'ceph' }}" mode: "{{ ceph_keyring_permissions }}" when: cephx | bool diff --git a/roles/ceph-mgr/tasks/main.yml b/roles/ceph-mgr/tasks/main.yml index d5cf2ef1e..0c60b026a 100644 --- a/roles/ceph-mgr/tasks/main.yml +++ b/roles/ceph-mgr/tasks/main.yml @@ -1,7 +1,7 @@ --- - name: set_fact container_exec_cmd set_fact: - container_exec_cmd: "{{ container_binary }} exec ceph-mon-{{ hostvars[item]['ansible_hostname'] }}" + container_exec_cmd: "{{ container_binary }} exec ceph-mon-{{ hostvars[item]['ansible_facts']['hostname'] }}" with_items: "{{ groups.get(mon_group_name, []) }}" delegate_to: "{{ item }}" delegate_facts: true diff --git a/roles/ceph-mgr/tasks/pre_requisite.yml b/roles/ceph-mgr/tasks/pre_requisite.yml index 0cfe8d1fb..6c441ea3b 100644 --- a/roles/ceph-mgr/tasks/pre_requisite.yml +++ b/roles/ceph-mgr/tasks/pre_requisite.yml @@ -1,10 +1,10 @@ --- - name: set_fact ceph_mgr_packages for sso set_fact: - ceph_mgr_packages: "{{ ceph_mgr_packages | union(['python3-saml' if ansible_distribution_major_version | int == 8 else 'python-saml']) }}" + ceph_mgr_packages: "{{ ceph_mgr_packages | union(['python3-saml' if ansible_facts['distribution_major_version'] | int == 8 else 'python-saml']) }}" when: - dashboard_enabled | bool - - ansible_distribution == 'RedHat' + - ansible_facts['distribution'] == 'RedHat' - name: install ceph-mgr packages on RedHat or SUSE package: @@ -12,16 +12,16 @@ state: "{{ (upgrade_ceph_packages|bool) | ternary('latest','present') }}" register: result until: result is succeeded - when: ansible_os_family in ['RedHat', 'Suse'] + when: ansible_facts['os_family'] in ['RedHat', 'Suse'] - name: install ceph-mgr packages for debian apt: name: '{{ ceph_mgr_packages }}' state: "{{ (upgrade_ceph_packages|bool) | ternary('latest','present') }}" - default_release: "{{ ceph_stable_release_uca | default('') if ceph_origin == 'repository' and ceph_repository == 'uca' else ''}}{{ ansible_distribution_release ~ '-backports' if ceph_origin == 'distro' and ceph_use_distro_backports else '' }}" + default_release: "{{ ceph_stable_release_uca | default('') if ceph_origin == 'repository' and ceph_repository == 'uca' else ''}}{{ ansible_facts['distribution_release'] ~ '-backports' if ceph_origin == 'distro' and ceph_use_distro_backports | bool else '' }}" register: result until: result is succeeded - when: ansible_os_family == 'Debian' + when: ansible_facts['os_family'] == 'Debian' - name: install routes python library for dashboard module apt: @@ -29,5 +29,5 @@ register: result until: result is succeeded when: - - ansible_os_family == 'Debian' + - ansible_facts['os_family'] == 'Debian' - "'ceph-mgr-dashboard' in ceph_mgr_packages" diff --git a/roles/ceph-mgr/tasks/start_mgr.yml b/roles/ceph-mgr/tasks/start_mgr.yml index 348a718dd..658ca9758 100644 --- a/roles/ceph-mgr/tasks/start_mgr.yml +++ b/roles/ceph-mgr/tasks/start_mgr.yml @@ -5,7 +5,7 @@ path: "/etc/systemd/system/ceph-mgr@.service.d/" when: - ceph_mgr_systemd_overrides is defined - - ansible_service_mgr == 'systemd' + - ansible_facts['service_mgr'] == 'systemd' - name: add ceph-mgr systemd service overrides config_template: @@ -15,7 +15,7 @@ config_type: "ini" when: - ceph_mgr_systemd_overrides is defined - - ansible_service_mgr == 'systemd' + - ansible_facts['service_mgr'] == 'systemd' - name: include_tasks systemd.yml include_tasks: systemd.yml @@ -23,7 +23,7 @@ - name: systemd start mgr systemd: - name: ceph-mgr@{{ ansible_hostname }} + name: ceph-mgr@{{ ansible_facts['hostname'] }} state: started enabled: yes masked: no diff --git a/roles/ceph-mgr/templates/ceph-mgr.service.j2 b/roles/ceph-mgr/templates/ceph-mgr.service.j2 index 5e4874837..5f7f9d9ce 100644 --- a/roles/ceph-mgr/templates/ceph-mgr.service.j2 +++ b/roles/ceph-mgr/templates/ceph-mgr.service.j2 @@ -11,11 +11,11 @@ After=network.target EnvironmentFile=-/etc/environment {% if container_binary == 'podman' %} ExecStartPre=-/usr/bin/rm -f /%t/%n-pid /%t/%n-cid -ExecStartPre=-/usr/bin/{{ container_binary }} rm --storage ceph-mgr-{{ ansible_hostname }} +ExecStartPre=-/usr/bin/{{ container_binary }} rm --storage ceph-mgr-{{ ansible_facts['hostname'] }} {% else %} -ExecStartPre=-/usr/bin/{{ container_binary }} stop ceph-mgr-{{ ansible_hostname }} +ExecStartPre=-/usr/bin/{{ container_binary }} stop ceph-mgr-{{ ansible_facts['hostname'] }} {% endif %} -ExecStartPre=-/usr/bin/{{ container_binary }} rm ceph-mgr-{{ ansible_hostname }} +ExecStartPre=-/usr/bin/{{ container_binary }} rm ceph-mgr-{{ ansible_facts['hostname'] }} ExecStart=/usr/bin/{{ container_binary }} run --rm --net=host \ {% if container_binary == 'podman' %} -d --log-driver journald --conmon-pidfile /%t/%n-pid --cidfile /%t/%n-cid \ @@ -31,12 +31,12 @@ ExecStart=/usr/bin/{{ container_binary }} run --rm --net=host \ -e CEPH_DAEMON=MGR \ -e CONTAINER_IMAGE={{ ceph_docker_registry }}/{{ ceph_docker_image }}:{{ ceph_docker_image_tag }} \ {{ ceph_mgr_docker_extra_env }} \ - --name=ceph-mgr-{{ ansible_hostname }} \ + --name=ceph-mgr-{{ ansible_facts['hostname'] }} \ {{ ceph_docker_registry }}/{{ ceph_docker_image }}:{{ ceph_docker_image_tag }} {% if container_binary == 'podman' %} ExecStop=-/usr/bin/sh -c "/usr/bin/{{ container_binary }} rm -f `cat /%t/%n-cid`" {% else %} -ExecStopPost=-/usr/bin/{{ container_binary }} stop ceph-mgr-{{ ansible_hostname }} +ExecStopPost=-/usr/bin/{{ container_binary }} stop ceph-mgr-{{ ansible_facts['hostname'] }} {% endif %} KillMode=none Restart=always diff --git a/roles/ceph-mon/defaults/main.yml b/roles/ceph-mon/defaults/main.yml index ac362b412..636bc4506 100644 --- a/roles/ceph-mon/defaults/main.yml +++ b/roles/ceph-mon/defaults/main.yml @@ -37,7 +37,7 @@ client_admin_ceph_authtool_cap: # For the whole list of limits you can apply see: docs.docker.com/engine/admin/resource_constraints # Default values are based from: https://access.redhat.com/documentation/en-us/red_hat_ceph_storage/2/html/red_hat_ceph_storage_hardware_guide/minimum_recommendations # These options can be passed using the 'ceph_mon_docker_extra_env' variable. -ceph_mon_docker_memory_limit: "{{ ansible_memtotal_mb }}m" +ceph_mon_docker_memory_limit: "{{ ansible_facts['memtotal_mb'] }}m" ceph_mon_docker_cpu_limit: 1 ceph_mon_container_listen_port: 3300 diff --git a/roles/ceph-mon/tasks/ceph_keys.yml b/roles/ceph-mon/tasks/ceph_keys.yml index 7d8ac9f71..e55504ec9 100644 --- a/roles/ceph-mon/tasks/ceph_keys.yml +++ b/roles/ceph-mon/tasks/ceph_keys.yml @@ -5,7 +5,7 @@ ceph --cluster {{ cluster }} -n mon. - -k /var/lib/ceph/mon/{{ cluster }}-{{ ansible_hostname }}/keyring + -k /var/lib/ceph/mon/{{ cluster }}-{{ ansible_facts['hostname'] }}/keyring mon_status --format json register: ceph_health_raw diff --git a/roles/ceph-mon/tasks/deploy_monitors.yml b/roles/ceph-mon/tasks/deploy_monitors.yml index 9f7cf893b..70e29fb71 100644 --- a/roles/ceph-mon/tasks/deploy_monitors.yml +++ b/roles/ceph-mon/tasks/deploy_monitors.yml @@ -7,7 +7,7 @@ name: mon. cluster: "{{ cluster }}" user: mon. - user_key: "/var/lib/ceph/mon/{{ cluster }}-{{ hostvars[running_mon]['ansible_hostname'] }}/keyring" + user_key: "/var/lib/ceph/mon/{{ cluster }}-{{ hostvars[running_mon]['ansible_facts']['hostname'] }}/keyring" state: info environment: CEPH_CONTAINER_IMAGE: "{{ ceph_docker_registry + '/' + ceph_docker_image + ':' + ceph_docker_image_tag if containerized_deployment | bool else None }}" diff --git a/roles/ceph-mon/tasks/main.yml b/roles/ceph-mon/tasks/main.yml index 4aa692506..848b744f2 100644 --- a/roles/ceph-mon/tasks/main.yml +++ b/roles/ceph-mon/tasks/main.yml @@ -1,7 +1,7 @@ --- - name: set_fact container_exec_cmd set_fact: - container_exec_cmd: "{{ container_binary }} exec ceph-mon-{{ ansible_hostname }}" + container_exec_cmd: "{{ container_binary }} exec ceph-mon-{{ ansible_facts['hostname'] }}" when: containerized_deployment | bool - name: include deploy_monitors.yml diff --git a/roles/ceph-mon/tasks/start_monitor.yml b/roles/ceph-mon/tasks/start_monitor.yml index c6a4e74ea..8c3b14314 100644 --- a/roles/ceph-mon/tasks/start_monitor.yml +++ b/roles/ceph-mon/tasks/start_monitor.yml @@ -6,7 +6,7 @@ when: - not containerized_deployment | bool - ceph_mon_systemd_overrides is defined - - ansible_service_mgr == 'systemd' + - ansible_facts['service_mgr'] == 'systemd' - name: add ceph-mon systemd service overrides config_template: @@ -17,7 +17,7 @@ when: - not containerized_deployment | bool - ceph_mon_systemd_overrides is defined - - ansible_service_mgr == 'systemd' + - ansible_facts['service_mgr'] == 'systemd' - name: include_tasks systemd.yml include_tasks: systemd.yml @@ -25,7 +25,7 @@ - name: start the monitor service systemd: - name: ceph-mon@{{ monitor_name if not containerized_deployment else ansible_hostname }} + name: ceph-mon@{{ monitor_name if not containerized_deployment | bool else ansible_facts['hostname'] }} state: started enabled: yes masked: no diff --git a/roles/ceph-mon/templates/ceph-mon.service.j2 b/roles/ceph-mon/templates/ceph-mon.service.j2 index ef8c84668..390c50445 100644 --- a/roles/ceph-mon/templates/ceph-mon.service.j2 +++ b/roles/ceph-mon/templates/ceph-mon.service.j2 @@ -28,7 +28,7 @@ ExecStart=/usr/bin/{{ container_binary }} run --rm --name ceph-mon-%i \ -v /var/run/ceph:/var/run/ceph:z \ -v /etc/localtime:/etc/localtime:ro \ -v /var/log/ceph:/var/log/ceph:z \ -{% if ansible_distribution == 'RedHat' -%} +{% if ansible_facts['distribution'] == 'RedHat' -%} -v /etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:z \ {% endif -%} {% if mon_docker_privileged -%} diff --git a/roles/ceph-nfs/defaults/main.yml b/roles/ceph-nfs/defaults/main.yml index adc9f5b14..5f058525e 100644 --- a/roles/ceph-nfs/defaults/main.yml +++ b/roles/ceph-nfs/defaults/main.yml @@ -17,7 +17,7 @@ copy_admin_key: false ceph_nfs_enable_service: true # ceph-nfs systemd service uses ansible's hostname as an instance id, -# so service name is ceph-nfs@{{ ansible_hostname }}, this is not +# so service name is ceph-nfs@{{ ansible_facts['hostname'] }}, this is not # ideal when ceph-nfs is managed by pacemaker across multiple hosts - in # such case it's better to have constant instance id instead which # can be set by 'ceph_nfs_service_suffix' @@ -87,7 +87,7 @@ ceph_nfs_rgw_sectype: "sys,krb5,krb5i,krb5p" # they must be configered. #ceph_nfs_rgw_access_key: "QFAMEDSJP5DEKJO0DDXY" #ceph_nfs_rgw_secret_key: "iaSFLDVvDdQt6lkNzHyW4fPLZugBAI1g17LO0+87[MAC[M#C" -rgw_client_name: client.rgw.{{ ansible_hostname }} +rgw_client_name: client.rgw.{{ ansible_facts['hostname'] }} ################### # CONFIG OVERRIDE # diff --git a/roles/ceph-nfs/tasks/create_rgw_nfs_user.yml b/roles/ceph-nfs/tasks/create_rgw_nfs_user.yml index 5acacb08e..06860798c 100644 --- a/roles/ceph-nfs/tasks/create_rgw_nfs_user.yml +++ b/roles/ceph-nfs/tasks/create_rgw_nfs_user.yml @@ -1,7 +1,7 @@ --- - name: set_fact container_exec_cmd_nfs set_fact: - container_exec_cmd_nfs: "{{ container_binary }} exec ceph-mon-{{ hostvars[groups[mon_group_name][0]]['ansible_hostname'] }}" + container_exec_cmd_nfs: "{{ container_binary }} exec ceph-mon-{{ hostvars[groups[mon_group_name][0]]['ansible_facts']['hostname'] }}" when: containerized_deployment | bool - name: check if "{{ ceph_nfs_rgw_user }}" exists diff --git a/roles/ceph-nfs/tasks/ganesha_selinux_fix.yml b/roles/ceph-nfs/tasks/ganesha_selinux_fix.yml index f739b32d7..eb51e526b 100644 --- a/roles/ceph-nfs/tasks/ganesha_selinux_fix.yml +++ b/roles/ceph-nfs/tasks/ganesha_selinux_fix.yml @@ -15,7 +15,7 @@ state: present register: result until: result is succeeded - when: ansible_distribution_major_version == '7' + when: ansible_facts['distribution_major_version'] == '7' - name: install nfs-ganesha-selinux and python3-policycoreutils on RHEL 8 package: @@ -23,7 +23,7 @@ state: present register: result until: result is succeeded - when: ansible_distribution_major_version == '8' + when: ansible_facts['distribution_major_version'] == '8' - name: add ganesha_t to permissive domain selinux_permissive: diff --git a/roles/ceph-nfs/tasks/main.yml b/roles/ceph-nfs/tasks/main.yml index 4cd3bd3e2..4ba9f96a0 100644 --- a/roles/ceph-nfs/tasks/main.yml +++ b/roles/ceph-nfs/tasks/main.yml @@ -2,7 +2,7 @@ # global/common requirement - name: stop nfs server service systemd: - name: "{{ 'nfs-server' if ansible_os_family == 'RedHat' else 'nfsserver' if ansible_os_family == 'Suse' else 'nfs-kernel-server' if ansible_os_family == 'Debian' }}" + name: "{{ 'nfs-server' if ansible_facts['os_family'] == 'RedHat' else 'nfsserver' if ansible_facts['os_family'] == 'Suse' else 'nfs-kernel-server' if ansible_facts['os_family'] == 'Debian' }}" state: stopped enabled: no failed_when: false @@ -24,7 +24,7 @@ import_tasks: ganesha_selinux_fix.yml when: - not containerized_deployment | bool - - ansible_os_family == 'RedHat' + - ansible_facts['os_family'] == 'RedHat' - name: nfs with external ceph cluster task related when: @@ -40,7 +40,7 @@ mode: "0755" with_items: - "{{ ceph_nfs_ceph_user }}" - - "{{ ansible_hostname }}" + - "{{ ansible_facts['hostname'] }}" - name: set_fact rgw_client_name set_fact: @@ -55,7 +55,7 @@ group: "{{ ceph_uid if containerized_deployment else 'ceph' }}" with_nested: - "{{ hostvars[groups['_filtered_clients'][0]]['slurp_client_keys']['results'] | default([]) }}" - - ['/var/lib/ceph/radosgw/{{ cluster }}-rgw.{{ ceph_nfs_ceph_user }}/keyring', '/var/lib/ceph/radosgw/{{ cluster }}-rgw.{{ ansible_hostname }}/keyring'] + - ['/var/lib/ceph/radosgw/{{ cluster }}-rgw.{{ ceph_nfs_ceph_user }}/keyring', "/var/lib/ceph/radosgw/{{ cluster }}-rgw.{{ ansible_facts['hostname'] }}/keyring"] when: - not item.0.get('skipped', False) - item.0.item.name == 'client.' + ceph_nfs_ceph_user or item.0.item.name == rgw_client_name diff --git a/roles/ceph-nfs/tasks/pre_requisite_container.yml b/roles/ceph-nfs/tasks/pre_requisite_container.yml index c2529d8c7..65e5fe5b5 100644 --- a/roles/ceph-nfs/tasks/pre_requisite_container.yml +++ b/roles/ceph-nfs/tasks/pre_requisite_container.yml @@ -3,7 +3,7 @@ block: - name: set_fact container_exec_cmd set_fact: - container_exec_cmd: "{{ container_binary }} exec ceph-mon-{{ hostvars[item]['ansible_hostname'] }}" + container_exec_cmd: "{{ container_binary }} exec ceph-mon-{{ hostvars[item]['ansible_facts']['hostname'] }}" with_items: "{{ groups.get(mon_group_name, []) }}" delegate_to: "{{ item }}" delegate_facts: true diff --git a/roles/ceph-nfs/tasks/pre_requisite_non_container.yml b/roles/ceph-nfs/tasks/pre_requisite_non_container.yml index 69efda56b..0bf796a05 100644 --- a/roles/ceph-nfs/tasks/pre_requisite_non_container.yml +++ b/roles/ceph-nfs/tasks/pre_requisite_non_container.yml @@ -1,11 +1,11 @@ --- - name: include red hat based system related tasks include_tasks: pre_requisite_non_container_red_hat.yml - when: ansible_os_family == 'RedHat' + when: ansible_facts['os_family'] == 'RedHat' - name: include debian based system related tasks include_tasks: pre_requisite_non_container_debian.yml - when: ansible_os_family == 'Debian' + when: ansible_facts['os_family'] == 'Debian' - name: install nfs rgw/cephfs gateway - suse zypper: @@ -18,7 +18,7 @@ when: - (ceph_origin == 'repository' or ceph_origin == 'distro') - ceph_repository != 'rhcs' - - ansible_os_family == 'Suse' + - ansible_facts['os_family'] == 'Suse' - item.install | bool register: result until: result is succeeded @@ -35,7 +35,7 @@ with_items: - { name: "/var/lib/ceph/bootstrap-rgw", create: "{{ nfs_obj_gw }}" } - { name: "/var/lib/ceph/radosgw", create: "{{ nfs_obj_gw }}" } - - { name: "/var/lib/ceph/radosgw/{{ cluster }}-rgw.{{ ansible_hostname }}", create: "{{ nfs_obj_gw }}" } + - { name: "/var/lib/ceph/radosgw/{{ cluster }}-rgw.{{ ansible_facts['hostname'] }}", create: "{{ nfs_obj_gw }}" } - { name: "{{ rbd_client_admin_socket_path }}", create: "{{ nfs_obj_gw }}" } - { name: "/var/log/ceph", create: true } - { name: "/var/run/ceph", create: true } @@ -74,14 +74,14 @@ when: nfs_obj_gw | bool block: - name: create rados gateway keyring - command: ceph --cluster {{ cluster }} --name client.bootstrap-rgw --keyring /var/lib/ceph/bootstrap-rgw/{{ cluster }}.keyring auth get-or-create client.rgw.{{ ansible_hostname }} osd 'allow rwx' mon 'allow rw' -o /var/lib/ceph/radosgw/{{ cluster }}-rgw.{{ ansible_hostname }}/keyring + command: ceph --cluster {{ cluster }} --name client.bootstrap-rgw --keyring /var/lib/ceph/bootstrap-rgw/{{ cluster }}.keyring auth get-or-create client.rgw.{{ ansible_facts['hostname'] }} osd 'allow rwx' mon 'allow rw' -o /var/lib/ceph/radosgw/{{ cluster }}-rgw.{{ ansible_facts['hostname'] }}/keyring args: - creates: /var/lib/ceph/radosgw/{{ cluster }}-rgw.{{ ansible_hostname }}/keyring + creates: /var/lib/ceph/radosgw/{{ cluster }}-rgw.{{ ansible_facts['hostname'] }}/keyring changed_when: false - name: set rados gateway key permissions file: - path: /var/lib/ceph/radosgw/{{ cluster }}-rgw.{{ ansible_hostname }}/keyring + path: /var/lib/ceph/radosgw/{{ cluster }}-rgw.{{ ansible_facts['hostname'] }}/keyring owner: "ceph" group: "ceph" mode: "0600" diff --git a/roles/ceph-nfs/tasks/pre_requisite_non_container_debian.yml b/roles/ceph-nfs/tasks/pre_requisite_non_container_debian.yml index 6db66b9bf..d2c4dd4bd 100644 --- a/roles/ceph-nfs/tasks/pre_requisite_non_container_debian.yml +++ b/roles/ceph-nfs/tasks/pre_requisite_non_container_debian.yml @@ -9,7 +9,7 @@ block: - name: add nfs-ganesha stable repository apt_repository: - repo: "deb {{ nfs_ganesha_stable_deb_repo }} {{ ceph_stable_distro_source | default(ansible_distribution_release) }} main" + repo: "deb {{ nfs_ganesha_stable_deb_repo }} {{ ceph_stable_distro_source | default(ansible_facts['distribution_release']) }} main" state: present update_cache: no register: add_ganesha_apt_repo @@ -30,7 +30,7 @@ block: - name: fetch nfs-ganesha development repository uri: - url: https://shaman.ceph.com/api/repos/nfs-ganesha/next/latest/{{ ansible_distribution | lower }}/{{ ansible_distribution_release }}/flavors/{{ nfs_ganesha_flavor }}/repo + url: https://shaman.ceph.com/api/repos/nfs-ganesha/next/latest/{{ ansible_facts['distribution'] | lower }}/{{ ansible_facts['distribution_release'] }}/flavors/{{ nfs_ganesha_flavor }}/repo return_content: yes register: nfs_ganesha_dev_apt_repo diff --git a/roles/ceph-nfs/tasks/pre_requisite_non_container_red_hat.yml b/roles/ceph-nfs/tasks/pre_requisite_non_container_red_hat.yml index d2c1ff20e..527de75f5 100644 --- a/roles/ceph-nfs/tasks/pre_requisite_non_container_red_hat.yml +++ b/roles/ceph-nfs/tasks/pre_requisite_non_container_red_hat.yml @@ -18,7 +18,7 @@ block: - name: fetch nfs-ganesha red hat development repository uri: - url: https://shaman.ceph.com/api/repos/nfs-ganesha/next/latest/{{ ansible_distribution | lower }}/{{ ansible_distribution_major_version }}/flavors/{{ nfs_ganesha_flavor }}/repo + url: https://shaman.ceph.com/api/repos/nfs-ganesha/next/latest/{{ ansible_facts['distribution'] | lower }}/{{ ansible_facts['distribution_major_version'] }}/flavors/{{ nfs_ganesha_flavor }}/repo return_content: yes register: nfs_ganesha_dev_yum_repo diff --git a/roles/ceph-nfs/tasks/start_nfs.yml b/roles/ceph-nfs/tasks/start_nfs.yml index d3e4ac1f8..c7b304eb8 100644 --- a/roles/ceph-nfs/tasks/start_nfs.yml +++ b/roles/ceph-nfs/tasks/start_nfs.yml @@ -8,7 +8,7 @@ - name: set_fact container_exec_cmd_nfs - internal set_fact: - exec_cmd_nfs: "{{ container_binary + ' exec ceph-mon-' + hostvars[groups[mon_group_name][0]]['ansible_hostname'] if containerized_deployment | bool else '' }} rados" + exec_cmd_nfs: "{{ container_binary + ' exec ceph-mon-' + hostvars[groups[mon_group_name][0]]['ansible_facts']['hostname'] if containerized_deployment | bool else '' }} rados" delegate_node: "{{ groups[mon_group_name][0] }}" when: groups.get(mon_group_name, []) | length > 0 @@ -72,7 +72,7 @@ - name: systemd start nfs container systemd: - name: ceph-nfs@{{ ceph_nfs_service_suffix | default(ansible_hostname) }} + name: ceph-nfs@{{ ceph_nfs_service_suffix | default(ansible_facts['hostname']) }} state: started enabled: yes masked: no diff --git a/roles/ceph-nfs/templates/ceph-nfs.service.j2 b/roles/ceph-nfs/templates/ceph-nfs.service.j2 index 1f749cd5a..e168b3e0e 100644 --- a/roles/ceph-nfs/templates/ceph-nfs.service.j2 +++ b/roles/ceph-nfs/templates/ceph-nfs.service.j2 @@ -15,7 +15,7 @@ ExecStartPre=-/usr/bin/rm -f /%t/%n-pid /%t/%n-cid ExecStartPre=-/usr/bin/{{ container_binary }} rm --storage ceph-nfs-%i {% endif %} ExecStartPre=-/usr/bin/{{ container_binary }} rm ceph-nfs-%i -ExecStartPre={{ '/bin/mkdir' if ansible_os_family == 'Debian' else '/usr/bin/mkdir' }} -p /etc/ceph /etc/ganesha /var/lib/nfs/ganesha /var/log/ganesha +ExecStartPre={{ '/bin/mkdir' if ansible_facts['os_family'] == 'Debian' else '/usr/bin/mkdir' }} -p /etc/ceph /etc/ganesha /var/lib/nfs/ganesha /var/log/ganesha ExecStart=/usr/bin/{{ container_binary }} run --rm --net=host \ {% if container_binary == 'podman' %} -d --log-driver journald --conmon-pidfile /%t/%n-pid --cidfile /%t/%n-cid \ @@ -36,7 +36,7 @@ ExecStart=/usr/bin/{{ container_binary }} run --rm --net=host \ -e CEPH_DAEMON=NFS \ -e CONTAINER_IMAGE={{ ceph_docker_registry }}/{{ ceph_docker_image }}:{{ ceph_docker_image_tag }} \ {{ ceph_nfs_docker_extra_env }} \ - --name=ceph-nfs-{{ ceph_nfs_service_suffix | default(ansible_hostname) }} \ + --name=ceph-nfs-{{ ceph_nfs_service_suffix | default(ansible_facts['hostname']) }} \ {{ ceph_docker_registry }}/{{ ceph_docker_image }}:{{ ceph_docker_image_tag }} {% if container_binary == 'podman' %} ExecStop=-/usr/bin/sh -c "/usr/bin/{{ container_binary }} rm -f `cat /%t/%n-cid`" diff --git a/roles/ceph-osd/defaults/main.yml b/roles/ceph-osd/defaults/main.yml index 465840186..8084140d4 100644 --- a/roles/ceph-osd/defaults/main.yml +++ b/roles/ceph-osd/defaults/main.yml @@ -161,7 +161,7 @@ ceph_config_keys: [] # DON'T TOUCH ME # For the whole list of limits you can apply see: docs.docker.com/engine/admin/resource_constraints # Default values are based from: https://access.redhat.com/documentation/en-us/red_hat_ceph_storage/2/html/red_hat_ceph_storage_hardware_guide/minimum_recommendations # These options can be passed using the 'ceph_osd_docker_extra_env' variable. -ceph_osd_docker_memory_limit: "{{ ansible_memtotal_mb }}m" +ceph_osd_docker_memory_limit: "{{ ansible_facts['memtotal_mb'] }}m" ceph_osd_docker_cpu_limit: 4 # The next two variables are undefined, and thus, unused by default. diff --git a/roles/ceph-osd/tasks/main.yml b/roles/ceph-osd/tasks/main.yml index bc56ae143..185950553 100644 --- a/roles/ceph-osd/tasks/main.yml +++ b/roles/ceph-osd/tasks/main.yml @@ -5,7 +5,7 @@ - name: set_fact container_exec_cmd set_fact: - container_exec_cmd: "{{ container_binary }} exec ceph-mon-{{ hostvars[item]['ansible_hostname'] }}" + container_exec_cmd: "{{ container_binary }} exec ceph-mon-{{ hostvars[item]['ansible_facts']['hostname'] }}" with_items: "{{ groups.get(mon_group_name, []) }}" delegate_to: "{{ item }}" delegate_facts: true @@ -23,7 +23,7 @@ until: result is succeeded when: - not containerized_deployment | bool - - ansible_os_family != 'ClearLinux' + - ansible_facts['os_family'] != 'ClearLinux' - name: install numactl when needed package: diff --git a/roles/ceph-osd/tasks/start_osds.yml b/roles/ceph-osd/tasks/start_osds.yml index 69bce22a5..087b4fac9 100644 --- a/roles/ceph-osd/tasks/start_osds.yml +++ b/roles/ceph-osd/tasks/start_osds.yml @@ -38,7 +38,7 @@ path: "/etc/systemd/system/ceph-osd@.service.d/" when: - ceph_osd_systemd_overrides is defined - - ansible_service_mgr == 'systemd' + - ansible_facts['service_mgr'] == 'systemd' - name: add ceph-osd systemd service overrides config_template: @@ -48,7 +48,7 @@ config_type: "ini" when: - ceph_osd_systemd_overrides is defined - - ansible_service_mgr == 'systemd' + - ansible_facts['service_mgr'] == 'systemd' - name: ensure "/var/lib/ceph/osd/{{ cluster }}-{{ item }}" is present file: diff --git a/roles/ceph-osd/tasks/system_tuning.yml b/roles/ceph-osd/tasks/system_tuning.yml index 86a157a8d..48c7d665e 100644 --- a/roles/ceph-osd/tasks/system_tuning.yml +++ b/roles/ceph-osd/tasks/system_tuning.yml @@ -2,7 +2,7 @@ - name: debian based systems tasks when: - osd_objectstore == 'filestore' - - ansible_os_family == "Debian" + - ansible_facts['os_family'] == "Debian" block: - name: disable osd directory parsing by updatedb command: updatedb -e /var/lib/ceph @@ -45,7 +45,7 @@ - name: set_fact vm_min_free_kbytes set_fact: - vm_min_free_kbytes: "{{ 4194303 if ansible_memtotal_mb >= 49152 else default_vm_min_free_kbytes.stdout }}" + vm_min_free_kbytes: "{{ 4194303 if ansible_facts['memtotal_mb'] >= 49152 else default_vm_min_free_kbytes.stdout }}" - name: apply operating system tuning sysctl: diff --git a/roles/ceph-osd/templates/ceph-osd.service.j2 b/roles/ceph-osd/templates/ceph-osd.service.j2 index 32d78a566..216db6b17 100644 --- a/roles/ceph-osd/templates/ceph-osd.service.j2 +++ b/roles/ceph-osd/templates/ceph-osd.service.j2 @@ -7,7 +7,7 @@ Requires=docker.service {% else %} After=network.target {% endif %} -{% set cpu_limit = ansible_processor_vcpus|int if ceph_osd_docker_cpu_limit|int > ansible_processor_vcpus|int else ceph_osd_docker_cpu_limit|int %} +{% set cpu_limit = ansible_facts['processor_vcpus']|int if ceph_osd_docker_cpu_limit|int > ansible_facts['processor_vcpus']|int else ceph_osd_docker_cpu_limit|int %} [Service] EnvironmentFile=-/etc/environment @@ -48,7 +48,7 @@ numactl \ -v /var/run/ceph:/var/run/ceph:z \ -v /var/run/udev/:/var/run/udev/ \ -v /var/log/ceph:/var/log/ceph:z \ - {% if ansible_distribution == 'Ubuntu' -%} + {% if ansible_facts['distribution'] == 'Ubuntu' -%} --security-opt apparmor:unconfined \ {% endif -%} {{ container_env_args }} \ diff --git a/roles/ceph-prometheus/templates/alertmanager.service.j2 b/roles/ceph-prometheus/templates/alertmanager.service.j2 index 6aba7709e..28280c3d6 100644 --- a/roles/ceph-prometheus/templates/alertmanager.service.j2 +++ b/roles/ceph-prometheus/templates/alertmanager.service.j2 @@ -35,7 +35,7 @@ ExecStart=/usr/bin/{{ container_binary }} run --rm --name=alertmanager \ --cluster.peer={{ peer }}:{{ alertmanager_cluster_port }} \ {% endfor %} --storage.path=/alertmanager \ - --web.external-url=http://{{ ansible_fqdn }}:{{ alertmanager_port }}/ \ + --web.external-url=http://{{ ansible_facts['fqdn'] }}:{{ alertmanager_port }}/ \ --web.listen-address={{ grafana_server_addr }}:{{ alertmanager_port }} {% if container_binary == 'podman' %} ExecStop=/usr/bin/sh -c "/usr/bin/{{ container_binary }} rm -f `cat /%t/%n-cid`" diff --git a/roles/ceph-prometheus/templates/alertmanager.yml.j2 b/roles/ceph-prometheus/templates/alertmanager.yml.j2 index c2600acec..2ca58f417 100644 --- a/roles/ceph-prometheus/templates/alertmanager.yml.j2 +++ b/roles/ceph-prometheus/templates/alertmanager.yml.j2 @@ -11,5 +11,5 @@ receivers: - name: 'ceph-dashboard' webhook_configs: {% for host in groups['mgrs'] | default(groups['mons']) %} - - url: '{{ dashboard_protocol }}://{{ hostvars[host]['ansible_fqdn'] }}:{{ dashboard_port }}/api/prometheus_receiver' + - url: '{{ dashboard_protocol }}://{{ hostvars[host]['ansible_facts']['fqdn'] }}:{{ dashboard_port }}/api/prometheus_receiver' {% endfor %} diff --git a/roles/ceph-prometheus/templates/prometheus.service.j2 b/roles/ceph-prometheus/templates/prometheus.service.j2 index 3a17da6f0..601f19a1d 100644 --- a/roles/ceph-prometheus/templates/prometheus.service.j2 +++ b/roles/ceph-prometheus/templates/prometheus.service.j2 @@ -31,7 +31,7 @@ ExecStart=/usr/bin/{{ container_binary }} run --rm --name=prometheus \ {{ prometheus_container_image }} \ --config.file=/etc/prometheus/prometheus.yml \ --storage.tsdb.path=/prometheus \ - --web.external-url=http://{{ ansible_fqdn }}:{{ prometheus_port }}/ \ + --web.external-url=http://{{ ansible_facts['fqdn'] }}:{{ prometheus_port }}/ \ --web.listen-address={{ grafana_server_addr }}:{{ prometheus_port }} {% if container_binary == 'podman' %} ExecStop=/usr/bin/sh -c "/usr/bin/{{ container_binary }} rm -f `cat /%t/%n-cid`" diff --git a/roles/ceph-prometheus/templates/prometheus.yml.j2 b/roles/ceph-prometheus/templates/prometheus.yml.j2 index 29d0461e1..28882f3db 100644 --- a/roles/ceph-prometheus/templates/prometheus.yml.j2 +++ b/roles/ceph-prometheus/templates/prometheus.yml.j2 @@ -22,14 +22,14 @@ scrape_configs: {% for host in (groups['all'] | difference(groups[grafana_server_group_name]|union(groups.get(client_group_name, [])))) %} - targets: ['{{ host }}:{{ node_exporter_port }}'] labels: - instance: "{{ hostvars[host]['ansible_nodename'] }}" + instance: "{{ hostvars[host]['ansible_facts']['nodename'] }}" {% endfor %} - job_name: 'grafana' static_configs: {% for host in groups[grafana_server_group_name] %} - targets: ['{{ host }}:{{ node_exporter_port }}'] labels: - instance: "{{ hostvars[host]['ansible_nodename'] }}" + instance: "{{ hostvars[host]['ansible_facts']['nodename'] }}" {% endfor %} {% if iscsi_gw_group_name in groups %} - job_name: 'iscsi-gws' @@ -37,7 +37,7 @@ scrape_configs: {% for host in groups[iscsi_gw_group_name] %} - targets: ['{{ host }}:9287'] labels: - instance: "{{ hostvars[host]['ansible_nodename'] }}" + instance: "{{ hostvars[host]['ansible_facts']['nodename'] }}" {% endfor %} {% endif %} alerting: diff --git a/roles/ceph-rbd-mirror/defaults/main.yml b/roles/ceph-rbd-mirror/defaults/main.yml index 75af947ff..a99f2ba8a 100644 --- a/roles/ceph-rbd-mirror/defaults/main.yml +++ b/roles/ceph-rbd-mirror/defaults/main.yml @@ -42,7 +42,7 @@ ceph_rbd_mirror_remote_user: "" # For the whole list of limits you can apply see: docs.docker.com/engine/admin/resource_constraints # Default values are based from: https://access.redhat.com/documentation/en-us/red_hat_ceph_storage/2/html/red_hat_ceph_storage_hardware_guide/minimum_recommendations # These options can be passed using the 'ceph_rbd_mirror_docker_extra_env' variable. -ceph_rbd_mirror_docker_memory_limit: "{{ ansible_memtotal_mb }}m" +ceph_rbd_mirror_docker_memory_limit: "{{ ansible_facts['memtotal_mb'] }}m" ceph_rbd_mirror_docker_cpu_limit: 1 ceph_rbd_mirror_docker_extra_env: diff --git a/roles/ceph-rbd-mirror/tasks/common.yml b/roles/ceph-rbd-mirror/tasks/common.yml index e350808ae..5fee414d5 100644 --- a/roles/ceph-rbd-mirror/tasks/common.yml +++ b/roles/ceph-rbd-mirror/tasks/common.yml @@ -28,17 +28,17 @@ ceph --cluster {{ cluster }} --name client.bootstrap-rbd-mirror --keyring /var/lib/ceph/bootstrap-rbd-mirror/{{ cluster }}.keyring - auth get-or-create client.rbd-mirror.{{ ansible_hostname }} + auth get-or-create client.rbd-mirror.{{ ansible_facts['hostname'] }} mon 'profile rbd-mirror' osd 'profile rbd' - -o /etc/ceph/{{ cluster }}.client.rbd-mirror.{{ ansible_hostname }}.keyring + -o /etc/ceph/{{ cluster }}.client.rbd-mirror.{{ ansible_facts['hostname'] }}.keyring args: - creates: /etc/ceph/{{ cluster }}.client.rbd-mirror.{{ ansible_hostname }}.keyring + creates: /etc/ceph/{{ cluster }}.client.rbd-mirror.{{ ansible_facts['hostname'] }}.keyring when: not containerized_deployment | bool - name: set rbd-mirror key permissions file: - path: /etc/ceph/{{ cluster }}.client.rbd-mirror.{{ ansible_hostname }}.keyring + path: /etc/ceph/{{ cluster }}.client.rbd-mirror.{{ ansible_facts['hostname'] }}.keyring owner: "ceph" group: "ceph" mode: "{{ ceph_keyring_permissions }}" diff --git a/roles/ceph-rbd-mirror/tasks/configure_mirroring.yml b/roles/ceph-rbd-mirror/tasks/configure_mirroring.yml index 4da988c1b..07a1384fd 100644 --- a/roles/ceph-rbd-mirror/tasks/configure_mirroring.yml +++ b/roles/ceph-rbd-mirror/tasks/configure_mirroring.yml @@ -1,14 +1,14 @@ --- - name: enable mirroring on the pool - command: "{{ container_exec_cmd | default('') }} rbd --cluster {{ cluster }} --keyring /etc/ceph/{{ cluster }}.client.rbd-mirror.{{ ansible_hostname }}.keyring --name client.rbd-mirror.{{ ansible_hostname }} mirror pool enable {{ ceph_rbd_mirror_pool }} {{ ceph_rbd_mirror_mode }}" + command: "{{ container_exec_cmd | default('') }} rbd --cluster {{ cluster }} --keyring /etc/ceph/{{ cluster }}.client.rbd-mirror.{{ ansible_facts['hostname'] }}.keyring --name client.rbd-mirror.{{ ansible_facts['hostname'] }} mirror pool enable {{ ceph_rbd_mirror_pool }} {{ ceph_rbd_mirror_mode }}" changed_when: false - name: list mirroring peer - command: "{{ container_exec_cmd | default('') }} rbd --cluster {{ cluster }} --keyring /etc/ceph/{{ cluster }}.client.rbd-mirror.{{ ansible_hostname }}.keyring --name client.rbd-mirror.{{ ansible_hostname }} mirror pool info {{ ceph_rbd_mirror_pool }}" + command: "{{ container_exec_cmd | default('') }} rbd --cluster {{ cluster }} --keyring /etc/ceph/{{ cluster }}.client.rbd-mirror.{{ ansible_facts['hostname'] }}.keyring --name client.rbd-mirror.{{ ansible_facts['hostname'] }} mirror pool info {{ ceph_rbd_mirror_pool }}" changed_when: false register: mirror_peer - name: add a mirroring peer - command: "{{ container_exec_cmd | default('') }} rbd --cluster {{ cluster }} --keyring /etc/ceph/{{ cluster }}.client.rbd-mirror.{{ ansible_hostname }}.keyring --name client.rbd-mirror.{{ ansible_hostname }} mirror pool peer add {{ ceph_rbd_mirror_pool }} {{ ceph_rbd_mirror_remote_user }}@{{ ceph_rbd_mirror_remote_cluster }}" + command: "{{ container_exec_cmd | default('') }} rbd --cluster {{ cluster }} --keyring /etc/ceph/{{ cluster }}.client.rbd-mirror.{{ ansible_facts['hostname'] }}.keyring --name client.rbd-mirror.{{ ansible_facts['hostname'] }} mirror pool peer add {{ ceph_rbd_mirror_pool }} {{ ceph_rbd_mirror_remote_user }}@{{ ceph_rbd_mirror_remote_cluster }}" changed_when: false when: ceph_rbd_mirror_remote_user in mirror_peer.stdout diff --git a/roles/ceph-rbd-mirror/tasks/main.yml b/roles/ceph-rbd-mirror/tasks/main.yml index 1544c17b7..98a6404ae 100644 --- a/roles/ceph-rbd-mirror/tasks/main.yml +++ b/roles/ceph-rbd-mirror/tasks/main.yml @@ -16,7 +16,7 @@ block: - name: set_fact container_exec_cmd set_fact: - container_exec_cmd: "{{ container_binary }} exec ceph-rbd-mirror-{{ ansible_hostname }}" + container_exec_cmd: "{{ container_binary }} exec ceph-rbd-mirror-{{ ansible_facts['hostname'] }}" - name: include start_container_rbd_mirror.yml include_tasks: start_container_rbd_mirror.yml diff --git a/roles/ceph-rbd-mirror/tasks/start_container_rbd_mirror.yml b/roles/ceph-rbd-mirror/tasks/start_container_rbd_mirror.yml index cebeccff4..c1ac8cd37 100644 --- a/roles/ceph-rbd-mirror/tasks/start_container_rbd_mirror.yml +++ b/roles/ceph-rbd-mirror/tasks/start_container_rbd_mirror.yml @@ -5,7 +5,7 @@ - name: systemd start rbd mirror container systemd: - name: ceph-rbd-mirror@rbd-mirror.{{ ansible_hostname }} + name: ceph-rbd-mirror@rbd-mirror.{{ ansible_facts['hostname'] }} state: started enabled: yes masked: no diff --git a/roles/ceph-rbd-mirror/tasks/start_rbd_mirror.yml b/roles/ceph-rbd-mirror/tasks/start_rbd_mirror.yml index 70140c5d1..cbc32748b 100644 --- a/roles/ceph-rbd-mirror/tasks/start_rbd_mirror.yml +++ b/roles/ceph-rbd-mirror/tasks/start_rbd_mirror.yml @@ -5,7 +5,7 @@ path: "/etc/systemd/system/ceph-rbd-mirror@.service.d/" when: - ceph_rbd_mirror_systemd_overrides is defined - - ansible_service_mgr == 'systemd' + - ansible_facts['service_mgr'] == 'systemd' - name: add ceph-rbd-mirror systemd service overrides config_template: @@ -15,7 +15,7 @@ config_type: "ini" when: - ceph_rbd_mirror_systemd_overrides is defined - - ansible_service_mgr == 'systemd' + - ansible_facts['service_mgr'] == 'systemd' - name: stop and remove the generic rbd-mirror service instance service: @@ -34,7 +34,7 @@ - name: start and add the rbd-mirror service instance service: - name: "ceph-rbd-mirror@rbd-mirror.{{ ansible_hostname }}" + name: "ceph-rbd-mirror@rbd-mirror.{{ ansible_facts['hostname'] }}" state: started enabled: yes masked: no diff --git a/roles/ceph-rbd-mirror/templates/ceph-rbd-mirror.service.j2 b/roles/ceph-rbd-mirror/templates/ceph-rbd-mirror.service.j2 index 830d29ae2..8dd83bd8e 100644 --- a/roles/ceph-rbd-mirror/templates/ceph-rbd-mirror.service.j2 +++ b/roles/ceph-rbd-mirror/templates/ceph-rbd-mirror.service.j2 @@ -11,11 +11,11 @@ After=network.target EnvironmentFile=-/etc/environment {% if container_binary == 'podman' %} ExecStartPre=-/usr/bin/rm -f /%t/%n-pid /%t/%n-cid -ExecStartPre=-/usr/bin/{{ container_binary }} rm --storage ceph-rbd-mirror-{{ ansible_hostname }} +ExecStartPre=-/usr/bin/{{ container_binary }} rm --storage ceph-rbd-mirror-{{ ansible_facts['hostname'] }} {% else %} -ExecStartPre=-/usr/bin/{{ container_binary }} stop ceph-rbd-mirror-{{ ansible_hostname }} +ExecStartPre=-/usr/bin/{{ container_binary }} stop ceph-rbd-mirror-{{ ansible_facts['hostname'] }} {% endif %} -ExecStartPre=-/usr/bin/{{ container_binary }} rm ceph-rbd-mirror-{{ ansible_hostname }} +ExecStartPre=-/usr/bin/{{ container_binary }} rm ceph-rbd-mirror-{{ ansible_facts['hostname'] }} ExecStart=/usr/bin/{{ container_binary }} run --rm --net=host \ {% if container_binary == 'podman' %} -d --log-driver journald --conmon-pidfile /%t/%n-pid --cidfile /%t/%n-cid \ @@ -30,13 +30,13 @@ ExecStart=/usr/bin/{{ container_binary }} run --rm --net=host \ -e CLUSTER={{ cluster }} \ -e CEPH_DAEMON=RBD_MIRROR \ -e CONTAINER_IMAGE={{ ceph_docker_registry }}/{{ ceph_docker_image }}:{{ ceph_docker_image_tag }} \ - --name=ceph-rbd-mirror-{{ ansible_hostname }} \ + --name=ceph-rbd-mirror-{{ ansible_facts['hostname'] }} \ {{ ceph_rbd_mirror_docker_extra_env }} \ {{ ceph_docker_registry }}/{{ ceph_docker_image }}:{{ ceph_docker_image_tag }} {% if container_binary == 'podman' %} ExecStop=-/usr/bin/sh -c "/usr/bin/{{ container_binary }} rm -f `cat /%t/%n-cid`" {% else %} -ExecStopPost=-/usr/bin/{{ container_binary }} stop ceph-rbd-mirror-{{ ansible_hostname }} +ExecStopPost=-/usr/bin/{{ container_binary }} stop ceph-rbd-mirror-{{ ansible_facts['hostname'] }} {% endif %} KillMode=none Restart=always diff --git a/roles/ceph-rgw-loadbalancer/templates/haproxy.cfg.j2 b/roles/ceph-rgw-loadbalancer/templates/haproxy.cfg.j2 index 65bd79186..8f448160d 100644 --- a/roles/ceph-rgw-loadbalancer/templates/haproxy.cfg.j2 +++ b/roles/ceph-rgw-loadbalancer/templates/haproxy.cfg.j2 @@ -46,6 +46,6 @@ backend rgw-backend option httpchk HEAD / {% for host in groups[rgw_group_name] %} {% for instance in hostvars[host]['rgw_instances'] %} - server {{ 'server-' + hostvars[host]['ansible_hostname'] + '-' + instance['instance_name'] }} {{ instance['radosgw_address'] }}:{{ instance['radosgw_frontend_port'] }} weight 100 check + server {{ 'server-' + hostvars[host]['ansible_facts']['hostname'] + '-' + instance['instance_name'] }} {{ instance['radosgw_address'] }}:{{ instance['radosgw_frontend_port'] }} weight 100 check {% endfor %} {% endfor %} diff --git a/roles/ceph-rgw-loadbalancer/templates/keepalived.conf.j2 b/roles/ceph-rgw-loadbalancer/templates/keepalived.conf.j2 index 0c5bd6135..8793680f8 100644 --- a/roles/ceph-rgw-loadbalancer/templates/keepalived.conf.j2 +++ b/roles/ceph-rgw-loadbalancer/templates/keepalived.conf.j2 @@ -15,8 +15,8 @@ vrrp_script check_haproxy { {% for instance in vrrp_instances %} vrrp_instance {{ instance['name'] }} { - state {{ 'MASTER' if ansible_hostname == instance['master'] else 'BACKUP' }} - priority {{ '100' if ansible_hostname == instance['master'] else '90' }} + state {{ 'MASTER' if ansible_facts['hostname'] == instance['master'] else 'BACKUP' }} + priority {{ '100' if ansible_facts['hostname'] == instance['master'] else '90' }} interface {{ virtual_ip_interface }} virtual_router_id {{ 50 + loop.index }} advert_int 1 diff --git a/roles/ceph-rgw/handlers/main.yml b/roles/ceph-rgw/handlers/main.yml index f4bab87c1..5d993b360 100644 --- a/roles/ceph-rgw/handlers/main.yml +++ b/roles/ceph-rgw/handlers/main.yml @@ -1,6 +1,6 @@ --- - name: restart rgw service: - name: "ceph-radosgw@rgw.{{ ansible_hostname }}.{{ item.instance_name }}" + name: "ceph-radosgw@rgw.{{ ansible_facts['hostname'] }}.{{ item.instance_name }}" state: restarted with_items: "{{ rgw_instances }}" diff --git a/roles/ceph-rgw/tasks/openstack-keystone.yml b/roles/ceph-rgw/tasks/openstack-keystone.yml index 3af8405da..acfe50bb4 100644 --- a/roles/ceph-rgw/tasks/openstack-keystone.yml +++ b/roles/ceph-rgw/tasks/openstack-keystone.yml @@ -5,7 +5,7 @@ state: present register: result until: result is succeeded - when: ansible_pkg_mgr == 'yum' or ansible_pkg_mgr == 'dnf' + when: ansible_facts['pkg_mgr'] == 'yum' or ansible_facts['pkg_mgr'] == 'dnf' - name: install libnss3-tools on debian package: @@ -13,7 +13,7 @@ state: present register: result until: result is succeeded - when: ansible_pkg_mgr == 'apt' + when: ansible_facts['pkg_mgr'] == 'apt' - name: create nss directory for keystone certificates file: diff --git a/roles/ceph-rgw/tasks/pre_requisite.yml b/roles/ceph-rgw/tasks/pre_requisite.yml index 68a2f429d..cd489e0e4 100644 --- a/roles/ceph-rgw/tasks/pre_requisite.yml +++ b/roles/ceph-rgw/tasks/pre_requisite.yml @@ -2,11 +2,11 @@ - name: create rgw keyrings ceph_key: state: present - name: "client.rgw.{{ ansible_hostname }}.{{ item.instance_name }}" + name: "client.rgw.{{ ansible_facts['hostname'] }}.{{ item.instance_name }}" cluster: "{{ cluster }}" user: "client.bootstrap-rgw" user_key: /var/lib/ceph/bootstrap-rgw/{{ cluster }}.keyring - dest: "/var/lib/ceph/radosgw/{{ cluster }}-rgw.{{ ansible_hostname }}.{{ item.instance_name }}/keyring" + dest: "/var/lib/ceph/radosgw/{{ cluster }}-rgw.{{ ansible_facts['hostname'] }}.{{ item.instance_name }}/keyring" caps: osd: 'allow rwx' mon: 'allow rw' diff --git a/roles/ceph-rgw/tasks/start_docker_rgw.yml b/roles/ceph-rgw/tasks/start_docker_rgw.yml index d0d96314e..f4e3296ad 100644 --- a/roles/ceph-rgw/tasks/start_docker_rgw.yml +++ b/roles/ceph-rgw/tasks/start_docker_rgw.yml @@ -4,7 +4,7 @@ - name: systemd start rgw container systemd: - name: ceph-radosgw@rgw.{{ ansible_hostname }}.{{ item.instance_name }} + name: ceph-radosgw@rgw.{{ ansible_facts['hostname'] }}.{{ item.instance_name }} state: started enabled: yes masked: no diff --git a/roles/ceph-rgw/tasks/start_radosgw.yml b/roles/ceph-rgw/tasks/start_radosgw.yml index 1431d988b..564540a86 100644 --- a/roles/ceph-rgw/tasks/start_radosgw.yml +++ b/roles/ceph-rgw/tasks/start_radosgw.yml @@ -15,7 +15,7 @@ - name: start rgw instance service: - name: ceph-radosgw@rgw.{{ ansible_hostname }}.{{ item.instance_name }} + name: ceph-radosgw@rgw.{{ ansible_facts['hostname'] }}.{{ item.instance_name }} state: started enabled: yes masked: no diff --git a/roles/ceph-rgw/templates/ceph-radosgw.service.j2 b/roles/ceph-rgw/templates/ceph-radosgw.service.j2 index 54bd5b0c1..d0f8d5358 100644 --- a/roles/ceph-rgw/templates/ceph-radosgw.service.j2 +++ b/roles/ceph-rgw/templates/ceph-radosgw.service.j2 @@ -6,17 +6,17 @@ Requires=docker.service {% else %} After=network.target {% endif %} -{% set cpu_limit = ansible_processor_vcpus|int if ceph_rgw_docker_cpu_limit|int > ansible_processor_vcpus|int else ceph_rgw_docker_cpu_limit|int %} +{% set cpu_limit = ansible_facts['processor_vcpus']|int if ceph_rgw_docker_cpu_limit|int > ansible_facts['processor_vcpus']|int else ceph_rgw_docker_cpu_limit|int %} [Service] EnvironmentFile=/var/lib/ceph/radosgw/{{ cluster }}-%i/EnvironmentFile {% if container_binary == 'podman' %} ExecStartPre=-/usr/bin/rm -f /%t/%n-pid /%t/%n-cid -ExecStartPre=-/usr/bin/{{ container_binary }} rm --storage ceph-rgw-{{ ansible_hostname }}-${INST_NAME} +ExecStartPre=-/usr/bin/{{ container_binary }} rm --storage ceph-rgw-{{ ansible_facts['hostname'] }}-${INST_NAME} {% else %} -ExecStartPre=-/usr/bin/{{ container_binary }} stop ceph-rgw-{{ ansible_hostname }}-${INST_NAME} +ExecStartPre=-/usr/bin/{{ container_binary }} stop ceph-rgw-{{ ansible_facts['hostname'] }}-${INST_NAME} {% endif %} -ExecStartPre=-/usr/bin/{{ container_binary }} rm ceph-rgw-{{ ansible_hostname }}-${INST_NAME} +ExecStartPre=-/usr/bin/{{ container_binary }} rm ceph-rgw-{{ ansible_facts['hostname'] }}-${INST_NAME} ExecStart=/usr/bin/{{ container_binary }} run --rm --net=host \ {% if container_binary == 'podman' %} -d --log-driver journald --conmon-pidfile /%t/%n-pid --cidfile /%t/%n-cid \ @@ -34,7 +34,7 @@ ExecStart=/usr/bin/{{ container_binary }} run --rm --net=host \ -v /var/run/ceph:/var/run/ceph:z \ -v /etc/localtime:/etc/localtime:ro \ -v /var/log/ceph:/var/log/ceph:z \ - {% if ansible_distribution == 'RedHat' -%} + {% if ansible_facts['distribution'] == 'RedHat' -%} -v /etc/pki/ca-trust/extracted:/etc/pki/ca-trust/extracted:z \ {% endif -%} {% if radosgw_frontend_ssl_certificate -%} @@ -42,15 +42,15 @@ ExecStart=/usr/bin/{{ container_binary }} run --rm --net=host \ {% endif -%} -e CEPH_DAEMON=RGW \ -e CLUSTER={{ cluster }} \ - -e RGW_NAME={{ ansible_hostname }}.${INST_NAME} \ + -e RGW_NAME={{ ansible_facts['hostname'] }}.${INST_NAME} \ -e CONTAINER_IMAGE={{ ceph_docker_registry }}/{{ ceph_docker_image }}:{{ ceph_docker_image_tag }} \ - --name=ceph-rgw-{{ ansible_hostname }}-${INST_NAME} \ + --name=ceph-rgw-{{ ansible_facts['hostname'] }}-${INST_NAME} \ {{ ceph_rgw_docker_extra_env }} \ {{ ceph_docker_registry }}/{{ ceph_docker_image }}:{{ ceph_docker_image_tag }} {% if container_binary == 'podman' %} ExecStop=-/usr/bin/sh -c "/usr/bin/{{ container_binary }} rm -f `cat /%t/%n-cid`" {% else %} -ExecStopPost=-/usr/bin/{{ container_binary }} stop ceph-rgw-{{ ansible_hostname }}-${INST_NAME} +ExecStopPost=-/usr/bin/{{ container_binary }} stop ceph-rgw-{{ ansible_facts['hostname'] }}-${INST_NAME} {% endif %} KillMode=none Restart=always diff --git a/roles/ceph-validate/tasks/check_eth_mon.yml b/roles/ceph-validate/tasks/check_eth_mon.yml index 5f805473c..17251127e 100644 --- a/roles/ceph-validate/tasks/check_eth_mon.yml +++ b/roles/ceph-validate/tasks/check_eth_mon.yml @@ -2,23 +2,23 @@ - name: "fail if {{ monitor_interface }} does not exist on {{ inventory_hostname }}" fail: msg: "{{ monitor_interface }} does not exist on {{ inventory_hostname }}" - when: monitor_interface not in ansible_interfaces + when: monitor_interface not in ansible_facts['interfaces'] - name: "fail if {{ monitor_interface }} is not active on {{ inventory_hostname }}" fail: msg: "{{ monitor_interface }} is not active on {{ inventory_hostname }}" - when: not hostvars[inventory_hostname]['ansible_' + (monitor_interface | replace('-', '_'))]['active'] + when: not hostvars[inventory_hostname]['ansible_facts'][(monitor_interface | replace('-', '_'))]['active'] - name: "fail if {{ monitor_interface }} does not have any ip v4 address on {{ inventory_hostname }}" fail: msg: "{{ monitor_interface }} does not have any IPv4 address on {{ inventory_hostname }}" when: - ip_version == "ipv4" - - hostvars[inventory_hostname]['ansible_' + (monitor_interface | replace('-', '_'))]['ipv4'] is not defined + - hostvars[inventory_hostname]['ansible_facts'][(monitor_interface | replace('-', '_'))]['ipv4'] is not defined - name: "fail if {{ monitor_interface }} does not have any ip v6 address on {{ inventory_hostname }}" fail: msg: "{{ monitor_interface }} does not have any IPv6 address on {{ inventory_hostname }}" when: - ip_version == "ipv6" - - hostvars[inventory_hostname]['ansible_' + (monitor_interface | replace('-', '_'))]['ipv6'] is not defined + - hostvars[inventory_hostname]['ansible_facts'][(monitor_interface | replace('-', '_'))]['ipv6'] is not defined diff --git a/roles/ceph-validate/tasks/check_eth_rgw.yml b/roles/ceph-validate/tasks/check_eth_rgw.yml index 73db6078e..c2438cf3d 100644 --- a/roles/ceph-validate/tasks/check_eth_rgw.yml +++ b/roles/ceph-validate/tasks/check_eth_rgw.yml @@ -2,23 +2,23 @@ - name: "fail if {{ radosgw_interface }} does not exist on {{ inventory_hostname }}" fail: msg: "{{ radosgw_interface }} does not exist on {{ inventory_hostname }}" - when: radosgw_interface not in ansible_interfaces + when: radosgw_interface not in ansible_facts['interfaces'] - name: "fail if {{ radosgw_interface }} is not active on {{ inventory_hostname }}" fail: msg: "{{ radosgw_interface }} is not active on {{ inventory_hostname }}" - when: hostvars[inventory_hostname]['ansible_' + (radosgw_interface | replace('-', '_'))]['active'] == "false" + when: hostvars[inventory_hostname]['ansible_facts'][(radosgw_interface | replace('-', '_'))]['active'] == "false" - name: "fail if {{ radosgw_interface }} does not have any ip v4 address on {{ inventory_hostname }}" fail: msg: "{{ radosgw_interface }} does not have any IPv4 address on {{ inventory_hostname }}" when: - ip_version == "ipv4" - - hostvars[inventory_hostname]['ansible_' + (radosgw_interface | replace('-', '_'))]['ipv4'] is not defined + - hostvars[inventory_hostname]['ansible_facts'][(radosgw_interface | replace('-', '_'))]['ipv4'] is not defined - name: "fail if {{ radosgw_interface }} does not have any ip v6 address on {{ inventory_hostname }}" fail: msg: "{{ radosgw_interface }} does not have any IPv6 address on {{ inventory_hostname }}" when: - ip_version == "ipv6" - - hostvars[inventory_hostname]['ansible_' + (radosgw_interface | replace('-', '_'))]['ipv6'] is not defined + - hostvars[inventory_hostname]['ansible_facts'][(radosgw_interface | replace('-', '_'))]['ipv6'] is not defined diff --git a/roles/ceph-validate/tasks/check_ipaddr_mon.yml b/roles/ceph-validate/tasks/check_ipaddr_mon.yml index 3bcd46269..734cd6970 100644 --- a/roles/ceph-validate/tasks/check_ipaddr_mon.yml +++ b/roles/ceph-validate/tasks/check_ipaddr_mon.yml @@ -2,4 +2,4 @@ - name: "fail if {{ inventory_hostname }} does not have any {{ ip_version }} address on {{ monitor_address_block }}" fail: msg: "{{ inventory_hostname }} does not have any {{ ip_version }} address on {{ monitor_address_block }}" - when: hostvars[inventory_hostname]['ansible_all_' + ip_version + '_addresses'] | ips_in_ranges(hostvars[inventory_hostname]['monitor_address_block'].split(',')) | length == 0 + when: hostvars[inventory_hostname]['ansible_facts']['all_' + ip_version + '_addresses'] | ips_in_ranges(hostvars[inventory_hostname]['monitor_address_block'].split(',')) | length == 0 diff --git a/roles/ceph-validate/tasks/check_iscsi.yml b/roles/ceph-validate/tasks/check_iscsi.yml index 0b06afe43..85b27456a 100644 --- a/roles/ceph-validate/tasks/check_iscsi.yml +++ b/roles/ceph-validate/tasks/check_iscsi.yml @@ -30,10 +30,10 @@ - name: fail on unsupported distribution for iscsi gateways fail: msg: "iSCSI gateways can only be deployed on Red Hat Enterprise Linux, CentOS or Fedora" - when: ansible_distribution not in ['RedHat', 'CentOS', 'Fedora'] + when: ansible_facts['distribution'] not in ['RedHat', 'CentOS', 'Fedora'] - name: fail on unsupported distribution version for iscsi gateways - command: 'grep -q {{ item }}=m {% if is_atomic|bool %}/usr/lib/ostree-boot{% else %}/boot{% endif %}/config-{{ ansible_kernel }}' + command: 'grep -q {{ item }}=m {% if is_atomic|bool %}/usr/lib/ostree-boot{% else %}/boot{% endif %}/config-{{ ansible_facts["kernel"] }}' register: iscsi_kernel changed_when: false failed_when: iscsi_kernel.rc != 0 @@ -41,5 +41,5 @@ - CONFIG_TARGET_CORE - CONFIG_TCM_USER2 - CONFIG_ISCSI_TARGET - when: ansible_distribution in ['RedHat', 'CentOS'] + when: ansible_facts['distribution'] in ['RedHat', 'CentOS'] when: iscsi_gw_group_name in group_names diff --git a/roles/ceph-validate/tasks/check_nfs.yml b/roles/ceph-validate/tasks/check_nfs.yml index 99895dcbb..6c7b82eca 100644 --- a/roles/ceph-validate/tasks/check_nfs.yml +++ b/roles/ceph-validate/tasks/check_nfs.yml @@ -12,4 +12,4 @@ msg: "ceph-nfs packages are not available from openSUSE Leap 15.x repositories (ceph_origin = 'distro')" when: - ceph_origin == 'distro' - - ansible_distribution == 'openSUSE Leap' + - ansible_facts['distribution'] == 'openSUSE Leap' diff --git a/roles/ceph-validate/tasks/check_system.yml b/roles/ceph-validate/tasks/check_system.yml index 5592b7b00..330199f90 100644 --- a/roles/ceph-validate/tasks/check_system.yml +++ b/roles/ceph-validate/tasks/check_system.yml @@ -13,29 +13,29 @@ - name: fail on unsupported system fail: - msg: "System not supported {{ ansible_system }}" - when: ansible_system not in ['Linux'] + msg: "System not supported {{ ansible_facts['system'] }}" + when: ansible_facts['system'] not in ['Linux'] - name: fail on unsupported architecture fail: - msg: "Architecture not supported {{ ansible_architecture }}" - when: ansible_architecture not in ['x86_64', 'ppc64le', 'armv7l', 'aarch64'] + msg: "Architecture not supported {{ ansible_facts['architecture'] }}" + when: ansible_facts['architecture'] not in ['x86_64', 'ppc64le', 'armv7l', 'aarch64'] - name: fail on unsupported distribution fail: - msg: "Distribution not supported {{ ansible_os_family }}" - when: ansible_os_family not in ['Debian', 'RedHat', 'ClearLinux', 'Suse'] + msg: "Distribution not supported {{ ansible_facts['os_family'] }}" + when: ansible_facts['os_family'] not in ['Debian', 'RedHat', 'ClearLinux', 'Suse'] - name: red hat based systems tasks when: - ceph_repository == 'rhcs' - - ansible_distribution == 'RedHat' + - ansible_facts['distribution'] == 'RedHat' block: - name: fail on unsupported distribution for red hat ceph storage fail: - msg: "Distribution not supported {{ ansible_distribution_version }} by Red Hat Ceph Storage, only RHEL 8 (>= 8.1) or RHEL 7 (>= 7.7)" - when: (ansible_distribution_major_version | int == 8 and ansible_distribution_version is version('8.1', '<')) or - (ansible_distribution_major_version | int == 7 and ansible_distribution_version is version('7.7', '<')) + msg: "Distribution not supported {{ ansible_facts['distribution_version'] }} by Red Hat Ceph Storage, only RHEL 8 (>= 8.1) or RHEL 7 (>= 7.7)" + when: (ansible_facts['distribution_major_version'] | int == 8 and ansible_facts['distribution_version'] is version('8.1', '<')) or + (ansible_facts['distribution_major_version'] | int == 7 and ansible_facts['distribution_version'] is version('7.7', '<')) - name: subscription manager related tasks when: ceph_repository_type == 'cdn' @@ -54,19 +54,19 @@ - name: fail on unsupported distribution for ubuntu cloud archive fail: - msg: "Distribution not supported by Ubuntu Cloud Archive: {{ ansible_distribution }}" + msg: "Distribution not supported by Ubuntu Cloud Archive: {{ ansible_facts['distribution'] }}" when: - ceph_repository == 'uca' - - ansible_distribution != 'Ubuntu' + - ansible_facts['distribution'] != 'Ubuntu' - name: "fail on unsupported openSUSE distribution (only 15.x supported)" fail: - msg: "Distribution not supported: {{ ansible_distribution }}" + msg: "Distribution not supported: {{ ansible_facts['distribution'] }}" when: - - ansible_distribution == 'openSUSE Leap' - - ansible_distribution_major_version != '15' + - ansible_facts['distribution'] == 'openSUSE Leap' + - ansible_facts['distribution_major_version'] != '15' - name: fail if systemd is not present fail: msg: "Systemd must be present" - when: ansible_service_mgr != 'systemd' + when: ansible_facts['service_mgr'] != 'systemd' diff --git a/roles/ceph-validate/tasks/main.yml b/roles/ceph-validate/tasks/main.yml index a2c77049b..07c2e95e5 100644 --- a/roles/ceph-validate/tasks/main.yml +++ b/roles/ceph-validate/tasks/main.yml @@ -97,7 +97,7 @@ - not use_fqdn_yes_i_am_sure | bool - name: debian based systems tasks - when: ansible_os_family == 'Debian' + when: ansible_facts['os_family'] == 'Debian' block: - name: fail if local scenario is enabled on debian fail: @@ -124,7 +124,7 @@ msg: installation can't happen on Atomic and ntpd needs to be installed when: - is_atomic | default(False) | bool - - ansible_os_family == 'RedHat' + - ansible_facts['os_family'] == 'RedHat' - ntp_daemon_type == 'ntpd' - name: make sure journal_size configured diff --git a/site-container.yml.sample b/site-container.yml.sample index af9c317f2..a3e97d404 100644 --- a/site-container.yml.sample +++ b/site-container.yml.sample @@ -467,11 +467,11 @@ - name: set_fact container_binary set_fact: - container_binary: "{{ 'podman' if (podman_binary.stat.exists and ansible_distribution == 'Fedora') or (ansible_os_family == 'RedHat' and ansible_distribution_major_version == '8') else 'docker' }}" + container_binary: "{{ 'podman' if (podman_binary.stat.exists and ansible_facts['distribution'] == 'Fedora') or (ansible_facts['os_family'] == 'RedHat' and ansible_facts['distribution_major_version'] == '8') else 'docker' }}" - name: get ceph status from the first monitor command: > - {{ container_binary }} exec ceph-mon-{{ hostvars[groups[mon_group_name][0]]['ansible_hostname'] }} ceph --cluster {{ cluster }} -s + {{ container_binary }} exec ceph-mon-{{ hostvars[groups[mon_group_name][0]]['ansible_facts']['hostname'] }} ceph --cluster {{ cluster }} -s register: ceph_status changed_when: false delegate_to: "{{ groups[mon_group_name][0] }}" diff --git a/tests/functional/add-osds/container/group_vars/all b/tests/functional/add-osds/container/group_vars/all index 021f20c4f..4c577d417 100644 --- a/tests/functional/add-osds/container/group_vars/all +++ b/tests/functional/add-osds/container/group_vars/all @@ -6,8 +6,8 @@ containerized_deployment: true cluster: ceph public_network: "192.168.55.0/24" cluster_network: "192.168.56.0/24" -monitor_interface: "{{ 'eth1' if ansible_distribution == 'CentOS' else 'ens6' }}" -radosgw_interface: "{{ 'eth1' if ansible_distribution == 'CentOS' else 'ens6' }}" +monitor_interface: "{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" +radosgw_interface: "{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" journal_size: 100 osd_objectstore: "bluestore" copy_admin_key: true diff --git a/tests/functional/add-osds/group_vars/all b/tests/functional/add-osds/group_vars/all index 663a4b1d0..391ce1ac5 100644 --- a/tests/functional/add-osds/group_vars/all +++ b/tests/functional/add-osds/group_vars/all @@ -4,8 +4,8 @@ ceph_repository: community cluster: ceph public_network: "192.168.53.0/24" cluster_network: "192.168.54.0/24" -monitor_interface: "{{ 'eth1' if ansible_distribution == 'CentOS' else 'ens6' }}" -radosgw_interface: "{{ 'eth1' if ansible_distribution == 'CentOS' else 'ens6' }}" +monitor_interface: "{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" +radosgw_interface: "{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" journal_size: 100 osd_objectstore: "bluestore" copy_admin_key: true diff --git a/tests/functional/add-rbdmirrors/container/group_vars/all b/tests/functional/add-rbdmirrors/container/group_vars/all index 460c5c71e..8675fabc6 100644 --- a/tests/functional/add-rbdmirrors/container/group_vars/all +++ b/tests/functional/add-rbdmirrors/container/group_vars/all @@ -6,8 +6,8 @@ containerized_deployment: true cluster: ceph public_network: "192.168.67.0/24" cluster_network: "192.168.68.0/24" -monitor_interface: "{{ 'eth1' if ansible_distribution == 'CentOS' else 'ens6' }}" -radosgw_interface: "{{ 'eth1' if ansible_distribution == 'CentOS' else 'ens6' }}" +monitor_interface: "{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" +radosgw_interface: "{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" journal_size: 100 osd_objectstore: "bluestore" osd_scenario: lvm diff --git a/tests/functional/add-rbdmirrors/group_vars/all b/tests/functional/add-rbdmirrors/group_vars/all index 7bf0324e8..2d192f9cd 100644 --- a/tests/functional/add-rbdmirrors/group_vars/all +++ b/tests/functional/add-rbdmirrors/group_vars/all @@ -4,8 +4,8 @@ ceph_repository: community cluster: ceph public_network: "192.168.65.0/24" cluster_network: "192.168.66.0/24" -monitor_interface: "{{ 'eth1' if ansible_distribution == 'CentOS' else 'ens6' }}" -radosgw_interface: "{{ 'eth1' if ansible_distribution == 'CentOS' else 'ens6' }}" +monitor_interface: "{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" +radosgw_interface: "{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" journal_size: 100 osd_objectstore: "bluestore" osd_scenario: lvm diff --git a/tests/functional/add-rgws/container/group_vars/all b/tests/functional/add-rgws/container/group_vars/all index 33f386334..1f2c0a636 100644 --- a/tests/functional/add-rgws/container/group_vars/all +++ b/tests/functional/add-rgws/container/group_vars/all @@ -6,8 +6,8 @@ containerized_deployment: true cluster: ceph public_network: "192.168.71.0/24" cluster_network: "192.168.72.0/24" -monitor_interface: "{{ 'eth1' if ansible_distribution == 'CentOS' else 'ens6' }}" -radosgw_interface: "{{ 'eth1' if ansible_distribution == 'CentOS' else 'ens6' }}" +monitor_interface: "{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" +radosgw_interface: "{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" journal_size: 100 osd_objectstore: "bluestore" osd_scenario: lvm diff --git a/tests/functional/add-rgws/group_vars/all b/tests/functional/add-rgws/group_vars/all index 819b3bbd0..b3158ff69 100644 --- a/tests/functional/add-rgws/group_vars/all +++ b/tests/functional/add-rgws/group_vars/all @@ -4,8 +4,8 @@ ceph_repository: community cluster: ceph public_network: "192.168.69.0/24" cluster_network: "192.168.70.0/24" -monitor_interface: "{{ 'eth1' if ansible_distribution == 'CentOS' else 'ens6' }}" -radosgw_interface: "{{ 'eth1' if ansible_distribution == 'CentOS' else 'ens6' }}" +monitor_interface: "{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" +radosgw_interface: "{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" journal_size: 100 osd_objectstore: "bluestore" osd_scenario: lvm diff --git a/tests/functional/all-in-one/container/group_vars/all b/tests/functional/all-in-one/container/group_vars/all index 455cc7d66..1e50b133d 100644 --- a/tests/functional/all-in-one/container/group_vars/all +++ b/tests/functional/all-in-one/container/group_vars/all @@ -4,8 +4,8 @@ docker: True containerized_deployment: True -monitor_interface: "{{ 'eth1' if ansible_distribution == 'CentOS' else 'ens6' }}" -radosgw_interface: "{{ 'eth1' if ansible_distribution == 'CentOS' else 'ens6' }}" +monitor_interface: "{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" +radosgw_interface: "{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" radosgw_num_instances: 2 ceph_mon_docker_subnet: "{{ public_network }}" ceph_docker_on_openstack: False diff --git a/tests/functional/all-in-one/group_vars/all b/tests/functional/all-in-one/group_vars/all index 72d14dd5b..a6d220b48 100644 --- a/tests/functional/all-in-one/group_vars/all +++ b/tests/functional/all-in-one/group_vars/all @@ -2,8 +2,8 @@ containerized_deployment: False ceph_origin: repository ceph_repository: community -monitor_interface: "{{ 'eth1' if ansible_distribution == 'CentOS' else 'ens6' }}" -radosgw_interface: "{{ 'eth1' if ansible_distribution == 'CentOS' else 'ens6' }}" +monitor_interface: "{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" +radosgw_interface: "{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" ceph_mon_docker_subnet: "{{ public_network }}" ceph_docker_on_openstack: False openstack_config: True diff --git a/tests/functional/all_daemons/container/group_vars/all b/tests/functional/all_daemons/container/group_vars/all index 95814b0b0..9b9310614 100644 --- a/tests/functional/all_daemons/container/group_vars/all +++ b/tests/functional/all_daemons/container/group_vars/all @@ -4,8 +4,8 @@ docker: True containerized_deployment: True -monitor_interface: "{{ 'eth1' if ansible_distribution == 'CentOS' else 'ens6' }}" -radosgw_interface: "{{ 'eth1' if ansible_distribution == 'CentOS' else 'ens6' }}" +monitor_interface: "{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" +radosgw_interface: "{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" ceph_mon_docker_subnet: "{{ public_network }}" ceph_docker_on_openstack: False public_network: "192.168.17.0/24" diff --git a/tests/functional/all_daemons/container/hosts b/tests/functional/all_daemons/container/hosts index fe64c8e8b..996d78076 100644 --- a/tests/functional/all_daemons/container/hosts +++ b/tests/functional/all_daemons/container/hosts @@ -1,6 +1,6 @@ [mons] mon0 monitor_address=192.168.17.10 -mon1 monitor_interface="{{ 'eth1' if ansible_distribution == 'CentOS' else 'ens6' }}" +mon1 monitor_interface="{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" mon2 monitor_address=192.168.17.12 [mgrs] diff --git a/tests/functional/all_daemons/container/hosts-ubuntu b/tests/functional/all_daemons/container/hosts-ubuntu index 5e22f90e3..daf24cfe5 100644 --- a/tests/functional/all_daemons/container/hosts-ubuntu +++ b/tests/functional/all_daemons/container/hosts-ubuntu @@ -1,6 +1,6 @@ [mons] mon0 monitor_address=192.168.17.10 -mon1 monitor_interface="{{ 'eth1' if ansible_distribution == 'CentOS' else 'ens6' }}" +mon1 monitor_interface="{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" mon2 monitor_address=192.168.17.12 [mgrs] diff --git a/tests/functional/all_daemons/container/hosts-upgrade-to-octopus b/tests/functional/all_daemons/container/hosts-upgrade-to-octopus index 95e020424..9f967a21d 100644 --- a/tests/functional/all_daemons/container/hosts-upgrade-to-octopus +++ b/tests/functional/all_daemons/container/hosts-upgrade-to-octopus @@ -1,6 +1,6 @@ [mons] mon0 monitor_address=192.168.17.10 -mon1 monitor_interface="{{ 'eth1' if ansible_distribution == 'CentOS' else 'ens6' }}" +mon1 monitor_interface="{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" mon2 monitor_address=192.168.17.12 [mgrs] diff --git a/tests/functional/all_daemons/group_vars/all b/tests/functional/all_daemons/group_vars/all index 574adfaf1..9a80e566f 100644 --- a/tests/functional/all_daemons/group_vars/all +++ b/tests/functional/all_daemons/group_vars/all @@ -3,7 +3,7 @@ ceph_origin: repository ceph_repository: community public_network: "192.168.1.0/24" cluster_network: "192.168.2.0/24" -radosgw_interface: "{{ 'eth1' if ansible_distribution == 'CentOS' else 'ens6' }}" +radosgw_interface: "{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" ceph_conf_overrides: global: osd_pool_default_size: 1 diff --git a/tests/functional/all_daemons/hosts b/tests/functional/all_daemons/hosts index 1c3bc54a0..85c64c9a3 100644 --- a/tests/functional/all_daemons/hosts +++ b/tests/functional/all_daemons/hosts @@ -1,6 +1,6 @@ [mons] mon0 monitor_address=192.168.1.10 -mon1 monitor_interface="{{ 'eth1' if ansible_distribution == 'CentOS' else 'ens6' }}" +mon1 monitor_interface="{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" mon2 monitor_address=192.168.1.12 [mgrs] diff --git a/tests/functional/all_daemons/hosts-switch-to-containers b/tests/functional/all_daemons/hosts-switch-to-containers index 8c3aed16a..2f31ca9ce 100644 --- a/tests/functional/all_daemons/hosts-switch-to-containers +++ b/tests/functional/all_daemons/hosts-switch-to-containers @@ -3,7 +3,7 @@ docker=True [mons] mon0 monitor_address=192.168.1.10 -mon1 monitor_interface="{{ 'eth1' if ansible_distribution == 'CentOS' else 'ens6' }}" +mon1 monitor_interface="{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" mon2 monitor_address=192.168.1.12 [mgrs] diff --git a/tests/functional/all_daemons/hosts-upgrade-to-octopus b/tests/functional/all_daemons/hosts-upgrade-to-octopus index e0e64ba69..7d2bd6dd4 100644 --- a/tests/functional/all_daemons/hosts-upgrade-to-octopus +++ b/tests/functional/all_daemons/hosts-upgrade-to-octopus @@ -1,6 +1,6 @@ [mons] mon0 monitor_address=192.168.1.10 -mon1 monitor_interface="{{ 'eth1' if ansible_distribution == 'CentOS' else 'ens6' }}" +mon1 monitor_interface="{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" mon2 monitor_address=192.168.1.12 [mgrs] diff --git a/tests/functional/collocation/container/group_vars/all b/tests/functional/collocation/container/group_vars/all index 6bc61511f..d8cd6aa73 100644 --- a/tests/functional/collocation/container/group_vars/all +++ b/tests/functional/collocation/container/group_vars/all @@ -4,8 +4,8 @@ docker: True containerized_deployment: True -monitor_interface: "{{ 'eth1' if ansible_distribution == 'CentOS' else 'ens6' }}" -radosgw_interface: "{{ 'eth1' if ansible_distribution == 'CentOS' else 'ens6' }}" +monitor_interface: "{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" +radosgw_interface: "{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" radosgw_num_instances: 2 ceph_mon_docker_subnet: "{{ public_network }}" ceph_docker_on_openstack: False diff --git a/tests/functional/collocation/group_vars/all b/tests/functional/collocation/group_vars/all index 6a34387f2..9fdbd55b4 100644 --- a/tests/functional/collocation/group_vars/all +++ b/tests/functional/collocation/group_vars/all @@ -2,8 +2,8 @@ containerized_deployment: False ceph_origin: repository ceph_repository: community -monitor_interface: "{{ 'eth1' if ansible_distribution == 'CentOS' else 'ens6' }}" -radosgw_interface: "{{ 'eth1' if ansible_distribution == 'CentOS' else 'ens6' }}" +monitor_interface: "{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" +radosgw_interface: "{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" ceph_mon_docker_subnet: "{{ public_network }}" ceph_docker_on_openstack: False public_network: "192.168.15.0/24" diff --git a/tests/functional/docker2podman/group_vars/all b/tests/functional/docker2podman/group_vars/all index 4a3134f07..23205e24c 100644 --- a/tests/functional/docker2podman/group_vars/all +++ b/tests/functional/docker2podman/group_vars/all @@ -4,8 +4,8 @@ docker: True containerized_deployment: True -monitor_interface: "{{ 'eth1' if ansible_distribution == 'CentOS' else 'ens6' }}" -radosgw_interface: "{{ 'eth1' if ansible_distribution == 'CentOS' else 'ens6' }}" +monitor_interface: "{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" +radosgw_interface: "{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" ceph_mon_docker_subnet: "{{ public_network }}" ceph_docker_on_openstack: False public_network: "192.168.58.0/24" diff --git a/tests/functional/external_clients/container/inventory/group_vars/all b/tests/functional/external_clients/container/inventory/group_vars/all index 2bd0e1a35..b61c80808 100644 --- a/tests/functional/external_clients/container/inventory/group_vars/all +++ b/tests/functional/external_clients/container/inventory/group_vars/all @@ -4,8 +4,8 @@ docker: True containerized_deployment: True -monitor_interface: "{{ 'eth1' if ansible_distribution == 'CentOS' else 'ens6' }}" -radosgw_interface: "{{ 'eth1' if ansible_distribution == 'CentOS' else 'ens6' }}" +monitor_interface: "{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" +radosgw_interface: "{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" radosgw_num_instances: 2 ceph_mon_docker_subnet: "{{ public_network }}" ceph_docker_on_openstack: False diff --git a/tests/functional/external_clients/inventory/group_vars/all b/tests/functional/external_clients/inventory/group_vars/all index 276b8dfaf..6a58eddcf 100644 --- a/tests/functional/external_clients/inventory/group_vars/all +++ b/tests/functional/external_clients/inventory/group_vars/all @@ -2,8 +2,8 @@ containerized_deployment: False ceph_origin: repository ceph_repository: community -monitor_interface: "{{ 'eth1' if ansible_distribution == 'CentOS' else 'ens6' }}" -radosgw_interface: "{{ 'eth1' if ansible_distribution == 'CentOS' else 'ens6' }}" +monitor_interface: "{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" +radosgw_interface: "{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" ceph_mon_docker_subnet: "{{ public_network }}" ceph_docker_on_openstack: False openstack_config: True diff --git a/tests/functional/filestore-to-bluestore/container/group_vars/all b/tests/functional/filestore-to-bluestore/container/group_vars/all index 8f3fa777f..fc9067dc5 100644 --- a/tests/functional/filestore-to-bluestore/container/group_vars/all +++ b/tests/functional/filestore-to-bluestore/container/group_vars/all @@ -8,7 +8,7 @@ ceph_origin: repository ceph_repository: community public_network: "192.168.43.0/24" cluster_network: "192.168.44.0/24" -monitor_interface: "{{ 'eth1' if ansible_distribution == 'CentOS' else 'ens6' }}" +monitor_interface: "{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" journal_size: 100 copy_admin_key: true containerized_deployment: true diff --git a/tests/functional/filestore-to-bluestore/group_vars/all b/tests/functional/filestore-to-bluestore/group_vars/all index 23799f822..d19c75aca 100644 --- a/tests/functional/filestore-to-bluestore/group_vars/all +++ b/tests/functional/filestore-to-bluestore/group_vars/all @@ -4,7 +4,7 @@ ceph_origin: repository ceph_repository: community public_network: "192.168.41.0/24" cluster_network: "192.168.42.0/24" -monitor_interface: "{{ 'eth1' if ansible_distribution == 'CentOS' else 'ens6' }}" +monitor_interface: "{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" journal_size: 100 copy_admin_key: true # test-volume is created by tests/functional/lvm_setup.yml from /dev/sdb diff --git a/tests/functional/lvm-auto-discovery/container/group_vars/all b/tests/functional/lvm-auto-discovery/container/group_vars/all index ac6994009..531fd6446 100644 --- a/tests/functional/lvm-auto-discovery/container/group_vars/all +++ b/tests/functional/lvm-auto-discovery/container/group_vars/all @@ -10,8 +10,8 @@ ceph_repository: community cluster: ceph public_network: "192.168.39.0/24" cluster_network: "192.168.40.0/24" -monitor_interface: "{{ 'eth1' if ansible_distribution == 'CentOS' else 'ens6' }}" -radosgw_interface: "{{ 'eth1' if ansible_distribution == 'CentOS' else 'ens6' }}" +monitor_interface: "{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" +radosgw_interface: "{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" journal_size: 100 osd_objectstore: "bluestore" crush_device_class: test diff --git a/tests/functional/lvm-auto-discovery/group_vars/all b/tests/functional/lvm-auto-discovery/group_vars/all index aa15d6851..0284affca 100644 --- a/tests/functional/lvm-auto-discovery/group_vars/all +++ b/tests/functional/lvm-auto-discovery/group_vars/all @@ -5,8 +5,8 @@ ceph_repository: community cluster: ceph public_network: "192.168.39.0/24" cluster_network: "192.168.40.0/24" -monitor_interface: "{{ 'eth1' if ansible_distribution == 'CentOS' else 'ens6' }}" -radosgw_interface: "{{ 'eth1' if ansible_distribution == 'CentOS' else 'ens6' }}" +monitor_interface: "{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" +radosgw_interface: "{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" osd_objectstore: "bluestore" crush_device_class: test copy_admin_key: true diff --git a/tests/functional/lvm-batch/container/group_vars/all b/tests/functional/lvm-batch/container/group_vars/all index 77fa7a97d..513bea3c3 100644 --- a/tests/functional/lvm-batch/container/group_vars/all +++ b/tests/functional/lvm-batch/container/group_vars/all @@ -10,8 +10,8 @@ ceph_repository: community cluster: ceph public_network: "192.168.39.0/24" cluster_network: "192.168.40.0/24" -monitor_interface: "{{ 'eth1' if ansible_distribution == 'CentOS' else 'ens6' }}" -radosgw_interface: "{{ 'eth1' if ansible_distribution == 'CentOS' else 'ens6' }}" +monitor_interface: "{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" +radosgw_interface: "{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" journal_size: 100 osd_objectstore: "bluestore" crush_device_class: test diff --git a/tests/functional/lvm-batch/group_vars/all b/tests/functional/lvm-batch/group_vars/all index ad72b72c5..aed954110 100644 --- a/tests/functional/lvm-batch/group_vars/all +++ b/tests/functional/lvm-batch/group_vars/all @@ -5,8 +5,8 @@ ceph_repository: community cluster: ceph public_network: "192.168.39.0/24" cluster_network: "192.168.40.0/24" -monitor_interface: "{{ 'eth1' if ansible_distribution == 'CentOS' else 'ens6' }}" -radosgw_interface: "{{ 'eth1' if ansible_distribution == 'CentOS' else 'ens6' }}" +monitor_interface: "{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" +radosgw_interface: "{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" osd_objectstore: "bluestore" crush_device_class: test copy_admin_key: true diff --git a/tests/functional/lvm-osds/container/group_vars/all b/tests/functional/lvm-osds/container/group_vars/all index 1d483e1e8..bc5941341 100644 --- a/tests/functional/lvm-osds/container/group_vars/all +++ b/tests/functional/lvm-osds/container/group_vars/all @@ -8,7 +8,7 @@ ceph_origin: repository ceph_repository: community public_network: "192.168.39.0/24" cluster_network: "192.168.40.0/24" -monitor_interface: "{{ 'eth1' if ansible_distribution == 'CentOS' else 'ens6' }}" +monitor_interface: "{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" journal_size: 100 copy_admin_key: true containerized_deployment: true diff --git a/tests/functional/lvm-osds/group_vars/all b/tests/functional/lvm-osds/group_vars/all index 8cae342f0..2cd64374f 100644 --- a/tests/functional/lvm-osds/group_vars/all +++ b/tests/functional/lvm-osds/group_vars/all @@ -4,7 +4,7 @@ ceph_origin: repository ceph_repository: community public_network: "192.168.39.0/24" cluster_network: "192.168.40.0/24" -monitor_interface: "{{ 'eth1' if ansible_distribution == 'CentOS' else 'ens6' }}" +monitor_interface: "{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" journal_size: 100 copy_admin_key: true # test-volume is created by tests/functional/lvm_setup.yml from /dev/sdb diff --git a/tests/functional/podman/group_vars/all b/tests/functional/podman/group_vars/all index 00c75a47b..aaf6f2636 100644 --- a/tests/functional/podman/group_vars/all +++ b/tests/functional/podman/group_vars/all @@ -4,8 +4,8 @@ docker: True containerized_deployment: True -monitor_interface: "{{ 'eth1' if ansible_distribution == 'CentOS' else 'ens6' }}" -radosgw_interface: "{{ 'eth1' if ansible_distribution == 'CentOS' else 'ens6' }}" +monitor_interface: "{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" +radosgw_interface: "{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" ceph_mon_docker_subnet: "{{ public_network }}" ceph_docker_on_openstack: False public_network: "192.168.30.0/24" diff --git a/tests/functional/rgw-multisite/container/group_vars/all b/tests/functional/rgw-multisite/container/group_vars/all index f60231934..1e47a8d48 100644 --- a/tests/functional/rgw-multisite/container/group_vars/all +++ b/tests/functional/rgw-multisite/container/group_vars/all @@ -6,8 +6,8 @@ ceph_repository: community cluster: ceph public_network: "192.168.105.0/24" cluster_network: "192.168.106.0/24" -monitor_interface: "{{ 'eth1' if ansible_distribution == 'CentOS' else 'ens6' }}" -radosgw_interface: "{{ 'eth1' if ansible_distribution == 'CentOS' else 'ens6' }}" +monitor_interface: "{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" +radosgw_interface: "{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" journal_size: 100 osd_objectstore: "bluestore" copy_admin_key: true diff --git a/tests/functional/rgw-multisite/container/secondary/group_vars/all b/tests/functional/rgw-multisite/container/secondary/group_vars/all index 8b118a178..f37fd1c38 100644 --- a/tests/functional/rgw-multisite/container/secondary/group_vars/all +++ b/tests/functional/rgw-multisite/container/secondary/group_vars/all @@ -6,8 +6,8 @@ ceph_repository: community cluster: ceph public_network: "192.168.107.0/24" cluster_network: "192.168.108.0/24" -monitor_interface: "{{ 'eth1' if ansible_distribution == 'CentOS' else 'ens6' }}" -radosgw_interface: "{{ 'eth1' if ansible_distribution == 'CentOS' else 'ens6' }}" +monitor_interface: "{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" +radosgw_interface: "{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" journal_size: 100 osd_objectstore: "bluestore" copy_admin_key: true diff --git a/tests/functional/rgw-multisite/group_vars/all b/tests/functional/rgw-multisite/group_vars/all index 9ad488a0b..4fe73d43a 100644 --- a/tests/functional/rgw-multisite/group_vars/all +++ b/tests/functional/rgw-multisite/group_vars/all @@ -4,8 +4,8 @@ ceph_repository: community cluster: ceph public_network: "192.168.101.0/24" cluster_network: "192.168.102.0/24" -monitor_interface: "{{ 'eth1' if ansible_distribution == 'CentOS' else 'ens6' }}" -radosgw_interface: "{{ 'eth1' if ansible_distribution == 'CentOS' else 'ens6' }}" +monitor_interface: "{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" +radosgw_interface: "{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" journal_size: 100 osd_objectstore: "bluestore" copy_admin_key: true diff --git a/tests/functional/rgw-multisite/secondary/group_vars/all b/tests/functional/rgw-multisite/secondary/group_vars/all index 97f8440dc..c247169b2 100644 --- a/tests/functional/rgw-multisite/secondary/group_vars/all +++ b/tests/functional/rgw-multisite/secondary/group_vars/all @@ -4,8 +4,8 @@ ceph_repository: community cluster: ceph public_network: "192.168.103.0/24" cluster_network: "192.168.104.0/24" -monitor_interface: "{{ 'eth1' if ansible_distribution == 'CentOS' else 'ens6' }}" -radosgw_interface: "{{ 'eth1' if ansible_distribution == 'CentOS' else 'ens6' }}" +monitor_interface: "{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" +radosgw_interface: "{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" journal_size: 100 osd_objectstore: "bluestore" copy_admin_key: true diff --git a/tests/functional/rhcs_setup.yml b/tests/functional/rhcs_setup.yml index 14804669b..7bc97ec14 100644 --- a/tests/functional/rhcs_setup.yml +++ b/tests/functional/rhcs_setup.yml @@ -81,7 +81,7 @@ state: present register: result until: result is succeeded - when: ansible_os_family == 'RedHat' + when: ansible_facts['os_family'] == 'RedHat' - name: allow insecure docker registries lineinfile: diff --git a/tests/functional/setup.yml b/tests/functional/setup.yml index 96e62b233..0413e7337 100644 --- a/tests/functional/setup.yml +++ b/tests/functional/setup.yml @@ -50,8 +50,8 @@ option: metalink state: absent when: - - ansible_distribution == 'CentOS' - - ansible_distribution_major_version | int == 7 + - ansible_facts['distribution'] == 'CentOS' + - ansible_facts['distribution_major_version'] | int == 7 - not is_atomic | bool - name: resize logical volume for root partition to fill remaining free space diff --git a/tests/functional/shrink_mds/container/group_vars/all b/tests/functional/shrink_mds/container/group_vars/all index e621da450..7e85132c7 100644 --- a/tests/functional/shrink_mds/container/group_vars/all +++ b/tests/functional/shrink_mds/container/group_vars/all @@ -4,7 +4,7 @@ docker: True containerized_deployment: True -monitor_interface: "{{ 'eth1' if ansible_distribution == 'CentOS' else 'ens6' }}" +monitor_interface: "{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" ceph_mon_docker_subnet: "{{ public_network }}" ceph_docker_on_openstack: False public_network: "192.168.79.0/24" diff --git a/tests/functional/shrink_mgr/container/group_vars/all b/tests/functional/shrink_mgr/container/group_vars/all index c26ebb3d6..44753fcd6 100644 --- a/tests/functional/shrink_mgr/container/group_vars/all +++ b/tests/functional/shrink_mgr/container/group_vars/all @@ -4,7 +4,7 @@ docker: True containerized_deployment: True -monitor_interface: "{{ 'eth1' if ansible_distribution == 'CentOS' else 'ens6' }}" +monitor_interface: "{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" ceph_mon_docker_subnet: "{{ public_network }}" ceph_docker_on_openstack: False public_network: "192.168.83.0/24" diff --git a/tests/functional/shrink_mon/container/group_vars/all b/tests/functional/shrink_mon/container/group_vars/all index 9801d081a..f549cebe0 100644 --- a/tests/functional/shrink_mon/container/group_vars/all +++ b/tests/functional/shrink_mon/container/group_vars/all @@ -4,7 +4,7 @@ docker: True containerized_deployment: True -monitor_interface: "{{ 'eth1' if ansible_distribution == 'CentOS' else 'ens6' }}" +monitor_interface: "{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" ceph_mon_docker_subnet: "{{ public_network }}" ceph_docker_on_openstack: False public_network: "192.168.17.0/24" diff --git a/tests/functional/shrink_mon/hosts b/tests/functional/shrink_mon/hosts index 541558a7a..c02302146 100644 --- a/tests/functional/shrink_mon/hosts +++ b/tests/functional/shrink_mon/hosts @@ -1,6 +1,6 @@ [mons] mon0 monitor_address=192.168.1.10 -mon1 monitor_interface="{{ 'eth1' if ansible_distribution == 'CentOS' else 'ens6' }}" +mon1 monitor_interface="{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" mon2 monitor_address=192.168.1.12 [osds] diff --git a/tests/functional/shrink_mon/hosts-switch-to-containers b/tests/functional/shrink_mon/hosts-switch-to-containers index c95d9d2f1..b995e9b99 100644 --- a/tests/functional/shrink_mon/hosts-switch-to-containers +++ b/tests/functional/shrink_mon/hosts-switch-to-containers @@ -3,7 +3,7 @@ docker=True [mons] mon0 monitor_address=192.168.1.10 -mon1 monitor_interface="{{ 'eth1' if ansible_distribution == 'CentOS' else 'ens6' }}" +mon1 monitor_interface="{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" mon2 monitor_address=192.168.1.12 [osds] diff --git a/tests/functional/shrink_mon/hosts-ubuntu b/tests/functional/shrink_mon/hosts-ubuntu index f2cc0b125..e0bf3eb2e 100644 --- a/tests/functional/shrink_mon/hosts-ubuntu +++ b/tests/functional/shrink_mon/hosts-ubuntu @@ -1,6 +1,6 @@ [mons] mon0 monitor_address=192.168.1.10 -mon1 monitor_interface="{{ 'eth1' if ansible_distribution == 'CentOS' else 'ens6' }}" +mon1 monitor_interface="{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" mon2 monitor_address=192.168.1.12 [osds] diff --git a/tests/functional/shrink_osd/container/group_vars/all b/tests/functional/shrink_osd/container/group_vars/all index c257f6103..e99fedfe7 100644 --- a/tests/functional/shrink_osd/container/group_vars/all +++ b/tests/functional/shrink_osd/container/group_vars/all @@ -4,7 +4,7 @@ docker: True containerized_deployment: True -monitor_interface: "{{ 'eth1' if ansible_distribution == 'CentOS' else 'ens6' }}" +monitor_interface: "{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" ceph_mon_docker_subnet: "{{ public_network }}" ceph_docker_on_openstack: False public_network: "192.168.73.0/24" diff --git a/tests/functional/shrink_rbdmirror/container/group_vars/all b/tests/functional/shrink_rbdmirror/container/group_vars/all index abdd93bd4..a95a0cf6c 100644 --- a/tests/functional/shrink_rbdmirror/container/group_vars/all +++ b/tests/functional/shrink_rbdmirror/container/group_vars/all @@ -5,7 +5,7 @@ docker: True public_network: "192.168.87.0/24" cluster_network: "192.168.88.0/24" containerized_deployment: True -monitor_interface: "{{ 'eth1' if ansible_distribution == 'CentOS' else 'ens6' }}" +monitor_interface: "{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" ceph_mon_docker_subnet: "{{ public_network }}" ceph_docker_on_openstack: False ceph_conf_overrides: diff --git a/tests/functional/shrink_rgw/container/group_vars/all b/tests/functional/shrink_rgw/container/group_vars/all index e49c96ba8..291e80ec5 100644 --- a/tests/functional/shrink_rgw/container/group_vars/all +++ b/tests/functional/shrink_rgw/container/group_vars/all @@ -6,8 +6,8 @@ docker: True containerized_deployment: True public_network: "192.168.91.0/24" cluster_network: "192.168.92.0/24" -monitor_interface: "{{ 'eth1' if ansible_distribution == 'CentOS' else 'ens6' }}" -radosgw_interface: "{{ 'eth1' if ansible_distribution == 'CentOS' else 'ens6' }}" +monitor_interface: "{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" +radosgw_interface: "{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" ceph_mon_docker_subnet: "{{ public_network }}" ceph_docker_on_openstack: False ceph_conf_overrides: diff --git a/tests/functional/shrink_rgw/group_vars/all b/tests/functional/shrink_rgw/group_vars/all index 406a0cf1c..3a78f052f 100644 --- a/tests/functional/shrink_rgw/group_vars/all +++ b/tests/functional/shrink_rgw/group_vars/all @@ -3,8 +3,8 @@ ceph_origin: repository ceph_repository: community public_network: "192.168.89.0/24" cluster_network: "192.168.90.0/24" -monitor_interface: "{{ 'eth1' if ansible_distribution == 'CentOS' else 'ens6' }}" -radosgw_interface: "{{ 'eth1' if ansible_distribution == 'CentOS' else 'ens6' }}" +monitor_interface: "{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" +radosgw_interface: "{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" osd_objectstore: "bluestore" osd_scenario: lvm copy_admin_key: true