diff --git a/group_vars/all.yml.sample b/group_vars/all.yml.sample index 05d1b3f75..72a5bc6b8 100644 --- a/group_vars/all.yml.sample +++ b/group_vars/all.yml.sample @@ -554,64 +554,6 @@ dummy: #docker_pull_timeout: "300s" -############# -# OPENSTACK # -############# -#openstack_config: false -# When pg_autoscale_mode is set to True, you must add the target_size_ratio key with a correct value -# `pg_num` and `pgp_num` keys will be ignored, even if specified. -# eg: -# openstack_glance_pool: -# name: "images" -# rule_name: "my_replicated_rule" -# application: "rbd" -# pg_autoscale_mode: false -# pg_num: 16 -# pgp_num: 16 -# target_size_ratio: 0.2 -#openstack_glance_pool: -# name: "images" -# application: "rbd" -#openstack_cinder_pool: -# name: "volumes" -# application: "rbd" -#openstack_nova_pool: -# name: "vms" -# application: "rbd" -#openstack_cinder_backup_pool: -# name: "backups" -# application: "rbd" -#openstack_gnocchi_pool: -# name: "metrics" -# application: "rbd" -#openstack_cephfs_data_pool: -# name: "manila_data" -# application: "cephfs" -#openstack_cephfs_metadata_pool: -# name: "manila_metadata" -# application: "cephfs" -#openstack_pools: -# - "{{ openstack_glance_pool }}" -# - "{{ openstack_cinder_pool }}" -# - "{{ openstack_nova_pool }}" -# - "{{ openstack_cinder_backup_pool }}" -# - "{{ openstack_gnocchi_pool }}" -# - "{{ openstack_cephfs_data_pool }}" -# - "{{ openstack_cephfs_metadata_pool }}" - - -# The value for 'key' can be a pre-generated key, -# e.g key: "AQDC2UxZH4yeLhAAgTaZb+4wDUlYOsr1OfZSpQ==" -# By default, keys will be auto-generated. -# -#openstack_keys: -# - { name: client.glance, caps: { mon: "profile rbd", osd: "profile rbd pool={{ openstack_cinder_pool.name }}, profile rbd pool={{ openstack_glance_pool.name }}"}, mode: "0600" } -# - { name: client.cinder, caps: { mon: "profile rbd", osd: "profile rbd pool={{ openstack_cinder_pool.name }}, profile rbd pool={{ openstack_nova_pool.name }}, profile rbd pool={{ openstack_glance_pool.name }}"}, mode: "0600" } -# - { name: client.cinder-backup, caps: { mon: "profile rbd", osd: "profile rbd pool={{ openstack_cinder_backup_pool.name }}"}, mode: "0600" } -# - { name: client.gnocchi, caps: { mon: "profile rbd", osd: "profile rbd pool={{ openstack_gnocchi_pool.name }}"}, mode: "0600", } -# - { name: client.openstack, caps: { mon: "profile rbd", osd: "profile rbd pool={{ openstack_glance_pool.name }}, profile rbd pool={{ openstack_nova_pool.name }}, profile rbd pool={{ openstack_cinder_pool.name }}, profile rbd pool={{ openstack_cinder_backup_pool.name }}"}, mode: "0600" } - - ############# # DASHBOARD # ############# diff --git a/roles/ceph-defaults/defaults/main.yml b/roles/ceph-defaults/defaults/main.yml index 489022dc0..c22f63b5e 100644 --- a/roles/ceph-defaults/defaults/main.yml +++ b/roles/ceph-defaults/defaults/main.yml @@ -546,64 +546,6 @@ docker_pull_retry: 3 docker_pull_timeout: "300s" -############# -# OPENSTACK # -############# -openstack_config: false -# When pg_autoscale_mode is set to True, you must add the target_size_ratio key with a correct value -# `pg_num` and `pgp_num` keys will be ignored, even if specified. -# eg: -# openstack_glance_pool: -# name: "images" -# rule_name: "my_replicated_rule" -# application: "rbd" -# pg_autoscale_mode: false -# pg_num: 16 -# pgp_num: 16 -# target_size_ratio: 0.2 -openstack_glance_pool: - name: "images" - application: "rbd" -openstack_cinder_pool: - name: "volumes" - application: "rbd" -openstack_nova_pool: - name: "vms" - application: "rbd" -openstack_cinder_backup_pool: - name: "backups" - application: "rbd" -openstack_gnocchi_pool: - name: "metrics" - application: "rbd" -openstack_cephfs_data_pool: - name: "manila_data" - application: "cephfs" -openstack_cephfs_metadata_pool: - name: "manila_metadata" - application: "cephfs" -openstack_pools: - - "{{ openstack_glance_pool }}" - - "{{ openstack_cinder_pool }}" - - "{{ openstack_nova_pool }}" - - "{{ openstack_cinder_backup_pool }}" - - "{{ openstack_gnocchi_pool }}" - - "{{ openstack_cephfs_data_pool }}" - - "{{ openstack_cephfs_metadata_pool }}" - - -# The value for 'key' can be a pre-generated key, -# e.g key: "AQDC2UxZH4yeLhAAgTaZb+4wDUlYOsr1OfZSpQ==" -# By default, keys will be auto-generated. -# -openstack_keys: - - { name: client.glance, caps: { mon: "profile rbd", osd: "profile rbd pool={{ openstack_cinder_pool.name }}, profile rbd pool={{ openstack_glance_pool.name }}"}, mode: "0600" } - - { name: client.cinder, caps: { mon: "profile rbd", osd: "profile rbd pool={{ openstack_cinder_pool.name }}, profile rbd pool={{ openstack_nova_pool.name }}, profile rbd pool={{ openstack_glance_pool.name }}"}, mode: "0600" } - - { name: client.cinder-backup, caps: { mon: "profile rbd", osd: "profile rbd pool={{ openstack_cinder_backup_pool.name }}"}, mode: "0600" } - - { name: client.gnocchi, caps: { mon: "profile rbd", osd: "profile rbd pool={{ openstack_gnocchi_pool.name }}"}, mode: "0600", } - - { name: client.openstack, caps: { mon: "profile rbd", osd: "profile rbd pool={{ openstack_glance_pool.name }}, profile rbd pool={{ openstack_nova_pool.name }}, profile rbd pool={{ openstack_cinder_pool.name }}, profile rbd pool={{ openstack_cinder_backup_pool.name }}"}, mode: "0600" } - - ############# # DASHBOARD # ############# diff --git a/roles/ceph-osd/tasks/main.yml b/roles/ceph-osd/tasks/main.yml index f51fc9845..8987f57cf 100644 --- a/roles/ceph-osd/tasks/main.yml +++ b/roles/ceph-osd/tasks/main.yml @@ -99,13 +99,3 @@ ansible.builtin.include_tasks: crush_rules.yml when: hostvars[groups[mon_group_name][0]]['crush_rule_config'] | default(crush_rule_config) | bool tags: wait_all_osds_up - -# Create the pools listed in openstack_pools -- name: Include openstack_config.yml - ansible.builtin.include_tasks: openstack_config.yml - when: - - not add_osd | bool - - not rolling_update | default(False) | bool - - openstack_config | bool - - inventory_hostname == groups[osd_group_name] | last - tags: wait_all_osds_up diff --git a/roles/ceph-osd/tasks/openstack_config.yml b/roles/ceph-osd/tasks/openstack_config.yml deleted file mode 100644 index 6b276c2c0..000000000 --- a/roles/ceph-osd/tasks/openstack_config.yml +++ /dev/null @@ -1,68 +0,0 @@ ---- -- name: Pool related tasks - block: - - name: Create openstack pool(s) - ceph_pool: - name: "{{ item.name }}" - cluster: "{{ cluster }}" - pg_num: "{{ item.pg_num | default(omit) }}" - pgp_num: "{{ item.pgp_num | default(omit) }}" - size: "{{ item.size | default(omit) }}" - min_size: "{{ item.min_size | default(omit) }}" - pool_type: "{{ item.type | default('replicated') }}" - rule_name: "{{ item.rule_name | default(omit) }}" - erasure_profile: "{{ item.erasure_profile | default(omit) }}" - pg_autoscale_mode: "{{ item.pg_autoscale_mode | default(omit) }}" - target_size_ratio: "{{ item.target_size_ratio | default(omit) }}" - application: "{{ item.application | default(omit) }}" - with_items: "{{ openstack_pools }}" - delegate_to: "{{ groups[mon_group_name][0] }}" - environment: - CEPH_CONTAINER_IMAGE: "{{ ceph_docker_registry + '/' + ceph_docker_image + ':' + ceph_docker_image_tag if containerized_deployment | bool else None }}" - CEPH_CONTAINER_BINARY: "{{ container_binary }}" - -- name: Create openstack cephx key(s) - when: - - cephx | bool - - openstack_config | bool - block: - - name: Generate keys - ceph_key: - name: "{{ item.name }}" - caps: "{{ item.caps }}" - secret: "{{ item.key | default('') }}" - cluster: "{{ cluster }}" - mode: "{{ item.mode | default(ceph_keyring_permissions) }}" - environment: - CEPH_CONTAINER_IMAGE: "{{ ceph_docker_registry + '/' + ceph_docker_image + ':' + ceph_docker_image_tag if containerized_deployment | bool else None }}" - CEPH_CONTAINER_BINARY: "{{ container_binary }}" - with_items: "{{ openstack_keys }}" - delegate_to: "{{ groups[mon_group_name][0] }}" - no_log: "{{ no_log_on_ceph_key_tasks }}" - - - name: Get keys from monitors - ceph_key: - name: "{{ item.name }}" - cluster: "{{ cluster }}" - output_format: plain - state: info - environment: - CEPH_CONTAINER_IMAGE: "{{ ceph_docker_registry + '/' + ceph_docker_image + ':' + ceph_docker_image_tag if containerized_deployment | bool else None }}" - CEPH_CONTAINER_BINARY: "{{ container_binary }}" - register: _osp_keys - with_items: "{{ openstack_keys }}" - delegate_to: "{{ groups.get(mon_group_name)[0] }}" - no_log: "{{ no_log_on_ceph_key_tasks }}" - - - name: Copy ceph key(s) if needed - ansible.builtin.copy: - dest: "/etc/ceph/{{ cluster }}.{{ item.0.item.name }}.keyring" - content: "{{ item.0.stdout + '\n' }}" - owner: "{{ ceph_uid if containerized_deployment | bool else 'ceph' }}" - group: "{{ ceph_uid if containerized_deployment | bool else 'ceph' }}" - mode: "{{ item.0.item.mode | default(ceph_keyring_permissions) }}" - with_nested: - - "{{ _osp_keys.results }}" - - "{{ groups[mon_group_name] }}" - delegate_to: "{{ item.1 }}" - no_log: "{{ no_log_on_ceph_key_tasks }}" diff --git a/roles/ceph-validate/tasks/check_pools.yml b/roles/ceph-validate/tasks/check_pools.yml index bca9d26ff..0acf7c957 100644 --- a/roles/ceph-validate/tasks/check_pools.yml +++ b/roles/ceph-validate/tasks/check_pools.yml @@ -3,7 +3,6 @@ ansible.builtin.fail: msg: "You must set a target_size_ratio value on following pool: {{ item.name }}." with_items: - - "{{ openstack_pools | default([]) }}" - "{{ cephfs_pools | default([]) }}" - "{{ pools | default([]) }}" when: diff --git a/tests/functional/all-in-one/container/group_vars/all b/tests/functional/all-in-one/container/group_vars/all index 1a9a4c9f6..bed6ae593 100644 --- a/tests/functional/all-in-one/container/group_vars/all +++ b/tests/functional/all-in-one/container/group_vars/all @@ -11,7 +11,6 @@ public_network: "192.168.19.0/24" cluster_network: "192.168.20.0/24" rgw_override_bucket_index_max_shards: 16 rgw_bucket_default_quota_max_objects: 1638400 -openstack_config: True dashboard_enabled: false ceph_conf_overrides: global: diff --git a/tests/functional/all-in-one/group_vars/all b/tests/functional/all-in-one/group_vars/all index 1b5ddd634..e4967a8b1 100644 --- a/tests/functional/all-in-one/group_vars/all +++ b/tests/functional/all-in-one/group_vars/all @@ -4,7 +4,6 @@ ceph_origin: repository ceph_repository: community radosgw_interface: "{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" ceph_mon_docker_subnet: "{{ public_network }}" -openstack_config: True dashboard_enabled: False public_network: "192.168.17.0/24" cluster_network: "192.168.18.0/24" diff --git a/tests/functional/all_daemons/container/group_vars/all b/tests/functional/all_daemons/container/group_vars/all index a33e13d1e..65d71f384 100644 --- a/tests/functional/all_daemons/container/group_vars/all +++ b/tests/functional/all_daemons/container/group_vars/all @@ -17,18 +17,6 @@ ceph_conf_overrides: mon_warn_on_pool_no_redundancy: false osd_pool_default_size: 1 mon_max_pg_per_osd: 300 -openstack_config: True -openstack_glance_pool: - name: "images" - size: 1 - target_size_ratio: 0.2 -openstack_cinder_pool: - name: "volumes" - rule_name: "HDD" - size: 1 -openstack_pools: - - "{{ openstack_glance_pool }}" - - "{{ openstack_cinder_pool }}" docker_pull_timeout: 600s handler_health_mon_check_delay: 10 handler_health_osd_check_delay: 10 diff --git a/tests/functional/all_daemons/group_vars/all b/tests/functional/all_daemons/group_vars/all index 51663bc86..75c50ecb0 100644 --- a/tests/functional/all_daemons/group_vars/all +++ b/tests/functional/all_daemons/group_vars/all @@ -11,20 +11,6 @@ ceph_conf_overrides: mon_warn_on_pool_no_redundancy: false osd_pool_default_size: 1 mon_max_pg_per_osd: 300 -openstack_config: True -openstack_glance_pool: - name: "images" - size: 1 - application: rbd - target_size_ratio: 0.2 -openstack_cinder_pool: - name: "volumes" - rule_name: "HDD" - size: 1 - application: rbd -openstack_pools: - - "{{ openstack_glance_pool }}" - - "{{ openstack_cinder_pool }}" handler_health_mon_check_delay: 10 handler_health_osd_check_delay: 10 mds_max_mds: 2 diff --git a/tests/functional/all_daemons_ipv6/container/group_vars/all b/tests/functional/all_daemons_ipv6/container/group_vars/all index e4503e0ee..4e3ef7de4 100644 --- a/tests/functional/all_daemons_ipv6/container/group_vars/all +++ b/tests/functional/all_daemons_ipv6/container/group_vars/all @@ -18,18 +18,6 @@ ceph_conf_overrides: mon_warn_on_pool_no_redundancy: false osd_pool_default_size: 1 mon_max_pg_per_osd: 300 -openstack_config: True -openstack_glance_pool: - name: "images" - size: 1 - target_size_ratio: 0.2 -openstack_cinder_pool: - name: "volumes" - rule_name: "HDD" - size: 1 -openstack_pools: - - "{{ openstack_glance_pool }}" - - "{{ openstack_cinder_pool }}" docker_pull_timeout: 600s handler_health_mon_check_delay: 10 handler_health_osd_check_delay: 10 diff --git a/tests/functional/all_daemons_ipv6/group_vars/all b/tests/functional/all_daemons_ipv6/group_vars/all index aa5b93f66..1748266d1 100644 --- a/tests/functional/all_daemons_ipv6/group_vars/all +++ b/tests/functional/all_daemons_ipv6/group_vars/all @@ -12,20 +12,6 @@ ceph_conf_overrides: mon_warn_on_pool_no_redundancy: false osd_pool_default_size: 1 mon_max_pg_per_osd: 300 -openstack_config: True -openstack_glance_pool: - name: "images" - size: 1 - application: rbd - target_size_ratio: 0.2 -openstack_cinder_pool: - name: "volumes" - rule_name: "HDD" - size: 1 - application: rbd -openstack_pools: - - "{{ openstack_glance_pool }}" - - "{{ openstack_cinder_pool }}" handler_health_mon_check_delay: 10 handler_health_osd_check_delay: 10 mds_max_mds: 2 diff --git a/tests/functional/docker2podman/group_vars/all b/tests/functional/docker2podman/group_vars/all index 362172266..966ab6854 100644 --- a/tests/functional/docker2podman/group_vars/all +++ b/tests/functional/docker2podman/group_vars/all @@ -16,18 +16,6 @@ ceph_conf_overrides: mon_allow_pool_size_one: true mon_warn_on_pool_no_redundancy: false osd_pool_default_size: 1 -openstack_config: False -openstack_glance_pool: - name: "images" - rule_name: "HDD" - size: 1 -openstack_cinder_pool: - name: "volumes" - rule_name: "HDD" - size: 1 -openstack_pools: - - "{{ openstack_glance_pool }}" - - "{{ openstack_cinder_pool }}" handler_health_mon_check_delay: 10 handler_health_osd_check_delay: 10 dashboard_admin_password: $sX!cD$rYU6qR^B! @@ -38,4 +26,4 @@ ceph_docker_image_tag: latest-reef node_exporter_container_image: "quay.io/prometheus/node-exporter:v0.17.0" prometheus_container_image: "quay.io/prometheus/prometheus:v2.7.2" alertmanager_container_image: "quay.io/prometheus/alertmanager:v0.16.2" -grafana_container_image: "quay.io/ceph/ceph-grafana:6.7.4" \ No newline at end of file +grafana_container_image: "quay.io/ceph/ceph-grafana:6.7.4" diff --git a/tests/functional/external_clients/container/inventory/group_vars/all b/tests/functional/external_clients/container/inventory/group_vars/all index 84474d502..beb09c357 100644 --- a/tests/functional/external_clients/container/inventory/group_vars/all +++ b/tests/functional/external_clients/container/inventory/group_vars/all @@ -11,7 +11,6 @@ public_network: "192.168.31.0/24" cluster_network: "192.168.32.0/24" rgw_override_bucket_index_max_shards: 16 rgw_bucket_default_quota_max_objects: 1638400 -openstack_config: True dashboard_enabled: false ceph_conf_overrides: global: diff --git a/tests/functional/external_clients/inventory/group_vars/all b/tests/functional/external_clients/inventory/group_vars/all index 1152b04ce..29c0ed43b 100644 --- a/tests/functional/external_clients/inventory/group_vars/all +++ b/tests/functional/external_clients/inventory/group_vars/all @@ -4,7 +4,6 @@ ceph_origin: repository ceph_repository: community radosgw_interface: "{{ 'eth1' if ansible_facts['distribution'] == 'CentOS' else 'ens6' }}" ceph_mon_docker_subnet: "{{ public_network }}" -openstack_config: True dashboard_enabled: False public_network: "192.168.31.0/24" cluster_network: "192.168.32.0/24" diff --git a/tests/functional/lvm-osds/container/group_vars/all b/tests/functional/lvm-osds/container/group_vars/all index 087d0a5e4..64286b499 100644 --- a/tests/functional/lvm-osds/container/group_vars/all +++ b/tests/functional/lvm-osds/container/group_vars/all @@ -21,20 +21,6 @@ ceph_conf_overrides: dashboard_enabled: False handler_health_mon_check_delay: 10 handler_health_osd_check_delay: 10 -openstack_config: True -openstack_glance_pool: - name: "images" - type: 3 - size: 1 - application: rbd - target_size_ratio: 0.2 -openstack_cinder_pool: - name: "volumes" - size: 1 - application: rbd -openstack_pools: - - "{{ openstack_glance_pool }}" - - "{{ openstack_cinder_pool }}" ceph_docker_registry: quay.io ceph_docker_image: ceph/daemon-base -ceph_docker_image_tag: latest-reef \ No newline at end of file +ceph_docker_image_tag: latest-reef diff --git a/tests/functional/lvm-osds/group_vars/all b/tests/functional/lvm-osds/group_vars/all index c8066f013..c39cfd433 100644 --- a/tests/functional/lvm-osds/group_vars/all +++ b/tests/functional/lvm-osds/group_vars/all @@ -16,17 +16,4 @@ ceph_conf_overrides: dashboard_enabled: False handler_health_mon_check_delay: 10 handler_health_osd_check_delay: 10 -openstack_config: True -openstack_glance_pool: - name: "images" - type: 3 - size: 1 - application: rbd - target_size_ratio: 0.2 -openstack_cinder_pool: - name: "volumes" - size: 1 - application: rbd -openstack_pools: - - "{{ openstack_glance_pool }}" - - "{{ openstack_cinder_pool }}" \ No newline at end of file + diff --git a/tests/functional/podman/group_vars/all b/tests/functional/podman/group_vars/all index 3a8328a79..427f6bed5 100644 --- a/tests/functional/podman/group_vars/all +++ b/tests/functional/podman/group_vars/all @@ -15,18 +15,6 @@ ceph_conf_overrides: mon_allow_pool_size_one: true mon_warn_on_pool_no_redundancy: false osd_pool_default_size: 1 -openstack_config: True -openstack_glance_pool: - name: "images" - rule_name: "HDD" - size: 1 -openstack_cinder_pool: - name: "volumes" - rule_name: "HDD" - size: 1 -openstack_pools: - - "{{ openstack_glance_pool }}" - - "{{ openstack_cinder_pool }}" handler_health_mon_check_delay: 10 handler_health_osd_check_delay: 10 dashboard_admin_password: $sX!cD$rYU6qR^B! @@ -37,4 +25,4 @@ ceph_docker_image_tag: latest-reef node_exporter_container_image: "quay.io/prometheus/node-exporter:v0.17.0" prometheus_container_image: "quay.io/prometheus/prometheus:v2.7.2" alertmanager_container_image: "quay.io/prometheus/alertmanager:v0.16.2" -grafana_container_image: "quay.io/ceph/ceph-grafana:6.7.4" \ No newline at end of file +grafana_container_image: "quay.io/ceph/ceph-grafana:6.7.4" diff --git a/tests/functional/shrink_mds/container/group_vars/all b/tests/functional/shrink_mds/container/group_vars/all index 3e9787332..97a6ea8f4 100644 --- a/tests/functional/shrink_mds/container/group_vars/all +++ b/tests/functional/shrink_mds/container/group_vars/all @@ -12,7 +12,6 @@ ceph_conf_overrides: mon_allow_pool_size_one: true mon_warn_on_pool_no_redundancy: false osd_pool_default_size: 1 -openstack_config: False dashboard_enabled: False copy_admin_key: True ceph_docker_registry: quay.io diff --git a/tests/functional/shrink_mgr/container/group_vars/all b/tests/functional/shrink_mgr/container/group_vars/all index e38b9e43d..367c38b70 100644 --- a/tests/functional/shrink_mgr/container/group_vars/all +++ b/tests/functional/shrink_mgr/container/group_vars/all @@ -12,7 +12,6 @@ ceph_conf_overrides: mon_allow_pool_size_one: true mon_warn_on_pool_no_redundancy: false osd_pool_default_size: 1 -openstack_config: False dashboard_enabled: False ceph_docker_registry: quay.io ceph_docker_image: ceph/daemon-base diff --git a/tests/functional/shrink_mon/container/group_vars/all b/tests/functional/shrink_mon/container/group_vars/all index 8522ee710..2d75b6fa6 100644 --- a/tests/functional/shrink_mon/container/group_vars/all +++ b/tests/functional/shrink_mon/container/group_vars/all @@ -12,7 +12,6 @@ ceph_conf_overrides: mon_allow_pool_size_one: true mon_warn_on_pool_no_redundancy: false osd_pool_default_size: 1 -openstack_config: False dashboard_enabled: False ceph_docker_registry: quay.io ceph_docker_image: ceph/daemon-base diff --git a/tests/functional/shrink_osd/container/group_vars/all b/tests/functional/shrink_osd/container/group_vars/all index 2f778f363..5cc5f6e81 100644 --- a/tests/functional/shrink_osd/container/group_vars/all +++ b/tests/functional/shrink_osd/container/group_vars/all @@ -12,7 +12,6 @@ ceph_conf_overrides: mon_allow_pool_size_one: true mon_warn_on_pool_no_redundancy: false osd_pool_default_size: 1 -openstack_config: False dashboard_enabled: False copy_admin_key: True ceph_docker_registry: quay.io diff --git a/tests/functional/shrink_osd/group_vars/all b/tests/functional/shrink_osd/group_vars/all index d6999b17e..956f325f8 100644 --- a/tests/functional/shrink_osd/group_vars/all +++ b/tests/functional/shrink_osd/group_vars/all @@ -6,6 +6,5 @@ cluster_network: "192.168.72.0/24" ceph_conf_overrides: global: osd_pool_default_size: 3 -openstack_config: False dashboard_enabled: False copy_admin_key: True \ No newline at end of file diff --git a/tests/functional/shrink_rbdmirror/container/group_vars/all b/tests/functional/shrink_rbdmirror/container/group_vars/all index f253911dd..684d5b5c2 100644 --- a/tests/functional/shrink_rbdmirror/container/group_vars/all +++ b/tests/functional/shrink_rbdmirror/container/group_vars/all @@ -11,7 +11,6 @@ ceph_conf_overrides: mon_allow_pool_size_one: true mon_warn_on_pool_no_redundancy: false osd_pool_default_size: 1 -openstack_config: False dashboard_enabled: False copy_admin_key: True ceph_docker_registry: quay.io diff --git a/tests/functional/shrink_rgw/container/group_vars/all b/tests/functional/shrink_rgw/container/group_vars/all index 1451936c2..900211e8d 100644 --- a/tests/functional/shrink_rgw/container/group_vars/all +++ b/tests/functional/shrink_rgw/container/group_vars/all @@ -13,7 +13,6 @@ ceph_conf_overrides: mon_allow_pool_size_one: true mon_warn_on_pool_no_redundancy: false osd_pool_default_size: 1 -openstack_config: False dashboard_enabled: False copy_admin_key: True ceph_docker_registry: quay.io diff --git a/tests/functional/subset_update/container/group_vars/all b/tests/functional/subset_update/container/group_vars/all index 70115443a..a9b38103a 100644 --- a/tests/functional/subset_update/container/group_vars/all +++ b/tests/functional/subset_update/container/group_vars/all @@ -17,7 +17,6 @@ ceph_conf_overrides: mon_warn_on_pool_no_redundancy: false osd_pool_default_size: 1 mon_max_pg_per_osd: 300 -openstack_config: false docker_pull_timeout: 600s handler_health_mon_check_delay: 10 handler_health_osd_check_delay: 10 diff --git a/tests/functional/subset_update/group_vars/all b/tests/functional/subset_update/group_vars/all index 13725dd32..4161ddea3 100644 --- a/tests/functional/subset_update/group_vars/all +++ b/tests/functional/subset_update/group_vars/all @@ -11,7 +11,6 @@ ceph_conf_overrides: mon_warn_on_pool_no_redundancy: false osd_pool_default_size: 1 mon_max_pg_per_osd: 300 -openstack_config: false handler_health_mon_check_delay: 10 handler_health_osd_check_delay: 10 mds_max_mds: 2