diff --git a/group_vars/all.yml.sample b/group_vars/all.yml.sample index e6f95f4fa..fa96792ff 100644 --- a/group_vars/all.yml.sample +++ b/group_vars/all.yml.sample @@ -554,6 +554,7 @@ dummy: # erasure_profile: "" # expected_num_objects: "" # application: "rbd" +# size: "" #openstack_cinder_pool: # name: "volumes" # pg_num: "{{ hostvars[groups[mon_group_name][0]]['osd_pool_default_pg_num'] }}" @@ -563,6 +564,7 @@ dummy: # erasure_profile: "" # expected_num_objects: "" # application: "rbd" +# size: "" #openstack_nova_pool: # name: "vms" # pg_num: "{{ hostvars[groups[mon_group_name][0]]['osd_pool_default_pg_num'] }}" @@ -572,6 +574,7 @@ dummy: # erasure_profile: "" # expected_num_objects: "" # application: "rbd" +# size: "" #openstack_cinder_backup_pool: # name: "backups" # pg_num: "{{ hostvars[groups[mon_group_name][0]]['osd_pool_default_pg_num'] }}" @@ -581,6 +584,7 @@ dummy: # erasure_profile: "" # expected_num_objects: "" # application: "rbd" +# size: "" #openstack_gnocchi_pool: # name: "metrics" # pg_num: "{{ hostvars[groups[mon_group_name][0]]['osd_pool_default_pg_num'] }}" @@ -590,6 +594,7 @@ dummy: # erasure_profile: "" # expected_num_objects: "" # application: "rbd" +# size: "" #openstack_pools: # - "{{ openstack_glance_pool }}" diff --git a/group_vars/rhcs.yml.sample b/group_vars/rhcs.yml.sample index cabb77279..2135893c4 100644 --- a/group_vars/rhcs.yml.sample +++ b/group_vars/rhcs.yml.sample @@ -554,6 +554,7 @@ ceph_docker_registry: "registry.access.redhat.com/rhceph/" # erasure_profile: "" # expected_num_objects: "" # application: "rbd" +# size: "" #openstack_cinder_pool: # name: "volumes" # pg_num: "{{ hostvars[groups[mon_group_name][0]]['osd_pool_default_pg_num'] }}" @@ -563,6 +564,7 @@ ceph_docker_registry: "registry.access.redhat.com/rhceph/" # erasure_profile: "" # expected_num_objects: "" # application: "rbd" +# size: "" #openstack_nova_pool: # name: "vms" # pg_num: "{{ hostvars[groups[mon_group_name][0]]['osd_pool_default_pg_num'] }}" @@ -572,6 +574,7 @@ ceph_docker_registry: "registry.access.redhat.com/rhceph/" # erasure_profile: "" # expected_num_objects: "" # application: "rbd" +# size: "" #openstack_cinder_backup_pool: # name: "backups" # pg_num: "{{ hostvars[groups[mon_group_name][0]]['osd_pool_default_pg_num'] }}" @@ -581,6 +584,7 @@ ceph_docker_registry: "registry.access.redhat.com/rhceph/" # erasure_profile: "" # expected_num_objects: "" # application: "rbd" +# size: "" #openstack_gnocchi_pool: # name: "metrics" # pg_num: "{{ hostvars[groups[mon_group_name][0]]['osd_pool_default_pg_num'] }}" @@ -590,6 +594,7 @@ ceph_docker_registry: "registry.access.redhat.com/rhceph/" # erasure_profile: "" # expected_num_objects: "" # application: "rbd" +# size: "" #openstack_pools: # - "{{ openstack_glance_pool }}" diff --git a/roles/ceph-client/defaults/main.yml b/roles/ceph-client/defaults/main.yml index ec477f299..80f9a5552 100644 --- a/roles/ceph-client/defaults/main.yml +++ b/roles/ceph-client/defaults/main.yml @@ -18,6 +18,7 @@ test: type: 1 erasure_profile: "" expected_num_objects: "" + size: "" test2: name: "test2" pg_num: "{{ hostvars[groups[mon_group_name][0]]['osd_pool_default_pg_num'] }}" @@ -26,6 +27,7 @@ test2: type: 1 erasure_profile: "" expected_num_objects: "" + size: "" pools: - "{{ test }}" - "{{ test2 }}" diff --git a/roles/ceph-client/tasks/create_users_keys.yml b/roles/ceph-client/tasks/create_users_keys.yml index 95ac72577..790831f44 100644 --- a/roles/ceph-client/tasks/create_users_keys.yml +++ b/roles/ceph-client/tasks/create_users_keys.yml @@ -84,40 +84,51 @@ - keys | length > 0 - inventory_hostname == groups.get('_filtered_clients') | first -- name: list existing pool(s) - command: > - {{ docker_exec_cmd | default('') }} ceph --cluster {{ cluster }} - osd pool get {{ item.name }} size - with_items: "{{ pools }}" - register: created_pools - failed_when: false - delegate_to: "{{ delegated_node }}" +- name: pool related tasks when: - condition_copy_admin_key - inventory_hostname == groups.get('_filtered_clients', []) | first + block: + - name: list existing pool(s) + command: > + {{ docker_exec_cmd | default('') }} ceph --cluster {{ cluster }} + osd pool get {{ item.name }} size + with_items: "{{ pools }}" + register: created_pools + failed_when: false + delegate_to: "{{ delegated_node }}" -- name: create ceph pool(s) - command: > - {{ docker_exec_cmd | default('') }} ceph --cluster {{ cluster }} - osd pool create {{ item.0.name }} - {{ item.0.pg_num }} - {{ item.0.pgp_num }} - {{ 'replicated_rule' if item.0.rule_name | default('replicated_rule') == '' else item.0.rule_name | default('replicated_rule') }} - {{ 1 if item.0.type|default(1) == 'replicated' else 3 if item.0.type|default(1) == 'erasure' else item.0.type|default(1) }} - {%- if (item.0.type | default("1") == '3' or item.0.type | default("1") == 'erasure') and item.0.erasure_profile != '' %} - {{ item.0.erasure_profile }} - {%- endif %} - {{ item.0.expected_num_objects | default('') }} - with_together: - - "{{ pools }}" - - "{{ created_pools.results }}" - changed_when: false - delegate_to: "{{ delegated_node }}" - when: - - pools | length > 0 - - condition_copy_admin_key - - inventory_hostname in groups.get('_filtered_clients') | first - - item.1.rc != 0 + - name: create ceph pool(s) + command: > + {{ docker_exec_cmd | default('') }} ceph --cluster {{ cluster }} + osd pool create {{ item.0.name }} + {{ item.0.pg_num }} + {{ item.0.pgp_num }} + {{ 'replicated_rule' if item.0.rule_name | default('replicated_rule') == '' else item.0.rule_name | default('replicated_rule') }} + {{ 1 if item.0.type|default(1) == 'replicated' else 3 if item.0.type|default(1) == 'erasure' else item.0.type|default(1) }} + {%- if (item.0.type | default("1") == '3' or item.0.type | default("1") == 'erasure') and item.0.erasure_profile != '' %} + {{ item.0.erasure_profile }} + {%- endif %} + {{ item.0.expected_num_objects | default('') }} + with_together: + - "{{ pools }}" + - "{{ created_pools.results }}" + changed_when: false + delegate_to: "{{ delegated_node }}" + when: + - pools | length > 0 + - item.1.rc != 0 + + - name: customize pool size + command: > + {{ docker_exec_cmd | default('') }} ceph --cluster {{ cluster }} + osd pool set {{ item.name }} size {{ item.size | default('') }} + with_items: "{{ pools | unique }}" + delegate_to: "{{ delegate_node }}" + changed_when: false + when: + - pools | length > 0 + - item.size | default ("") != "" - name: get client cephx keys copy: diff --git a/roles/ceph-defaults/defaults/main.yml b/roles/ceph-defaults/defaults/main.yml index 511b2294b..dbddf6d1e 100644 --- a/roles/ceph-defaults/defaults/main.yml +++ b/roles/ceph-defaults/defaults/main.yml @@ -346,8 +346,8 @@ cephfs_data: cephfs_data # name of the data pool for a given filesystem cephfs_metadata: cephfs_metadata # name of the metadata pool for a given filesystem cephfs_pools: - - { name: "{{ cephfs_data }}", pgs: "{{ hostvars[groups[mon_group_name][0]]['osd_pool_default_pg_num'] }}" } - - { name: "{{ cephfs_metadata }}", pgs: "{{ hostvars[groups[mon_group_name][0]]['osd_pool_default_pg_num'] }}" } + - { name: "{{ cephfs_data }}", pgs: "{{ hostvars[groups[mon_group_name][0]]['osd_pool_default_pg_num'] }}", size: "" } + - { name: "{{ cephfs_metadata }}", pgs: "{{ hostvars[groups[mon_group_name][0]]['osd_pool_default_pg_num'] }}", size: "" } ## OSD options # @@ -546,6 +546,7 @@ openstack_glance_pool: erasure_profile: "" expected_num_objects: "" application: "rbd" + size: "" openstack_cinder_pool: name: "volumes" pg_num: "{{ hostvars[groups[mon_group_name][0]]['osd_pool_default_pg_num'] }}" @@ -555,6 +556,7 @@ openstack_cinder_pool: erasure_profile: "" expected_num_objects: "" application: "rbd" + size: "" openstack_nova_pool: name: "vms" pg_num: "{{ hostvars[groups[mon_group_name][0]]['osd_pool_default_pg_num'] }}" @@ -564,6 +566,7 @@ openstack_nova_pool: erasure_profile: "" expected_num_objects: "" application: "rbd" + size: "" openstack_cinder_backup_pool: name: "backups" pg_num: "{{ hostvars[groups[mon_group_name][0]]['osd_pool_default_pg_num'] }}" @@ -573,6 +576,7 @@ openstack_cinder_backup_pool: erasure_profile: "" expected_num_objects: "" application: "rbd" + size: "" openstack_gnocchi_pool: name: "metrics" pg_num: "{{ hostvars[groups[mon_group_name][0]]['osd_pool_default_pg_num'] }}" @@ -582,6 +586,27 @@ openstack_gnocchi_pool: erasure_profile: "" expected_num_objects: "" application: "rbd" + size: "" +openstack_cephfs_data_pool: + name: "manila_data" + pg_num: "{{ hostvars[groups[mon_group_name][0]]['osd_pool_default_pg_num'] }}" + pgp_num: "{{ hostvars[groups[mon_group_name][0]]['osd_pool_default_pg_num'] }}" + rule_name: "replicated_rule" + type: 1 + erasure_profile: "" + expected_num_objects: "" + application: "rbd" + size: "" +openstack_cephfs_metadata_pool: + name: "manila_metadata" + pg_num: "{{ hostvars[groups[mon_group_name][0]]['osd_pool_default_pg_num'] }}" + pgp_num: "{{ hostvars[groups[mon_group_name][0]]['osd_pool_default_pg_num'] }}" + rule_name: "replicated_rule" + type: 1 + erasure_profile: "" + expected_num_objects: "" + application: "rbd" + size: "" openstack_pools: - "{{ openstack_glance_pool }}" @@ -589,6 +614,8 @@ openstack_pools: - "{{ openstack_nova_pool }}" - "{{ openstack_cinder_backup_pool }}" - "{{ openstack_gnocchi_pool }}" + - "{{ openstack_cephfs_data_pool }}" + - "{{ openstack_cephfs_metadata_pool }}" # The value for 'key' can be a pre-generated key, diff --git a/roles/ceph-iscsi-gw/defaults/main.yml b/roles/ceph-iscsi-gw/defaults/main.yml index 94309e74a..5f8b8a703 100644 --- a/roles/ceph-iscsi-gw/defaults/main.yml +++ b/roles/ceph-iscsi-gw/defaults/main.yml @@ -56,6 +56,7 @@ client_connections: {} # Whether or not to generate secure certificate to iSCSI gateway nodes generate_crt: False +rbd_pool_size: "" ################## # RBD-TARGET-API # diff --git a/roles/ceph-iscsi-gw/tasks/common.yml b/roles/ceph-iscsi-gw/tasks/common.yml index 8e9383c9d..9f54ebde2 100644 --- a/roles/ceph-iscsi-gw/tasks/common.yml +++ b/roles/ceph-iscsi-gw/tasks/common.yml @@ -32,15 +32,23 @@ register: rbd_pool_exists delegate_to: "{{ groups[mon_group_name][0] }}" -- name: get default value for osd_pool_default_pg_num - command: "{{ docker_exec_cmd | default('') }} ceph --cluster {{ cluster }} daemon mon.{{ hostvars[groups[mon_group_name][0]]['ansible_hostname'] }} config get osd_pool_default_pg_num" - changed_when: false - register: osd_pool_default_pg_num - delegate_to: "{{ groups[mon_group_name][0] }}" +- name: rbd pool related tasks when: "'rbd' not in (rbd_pool_exists.stdout | from_json)" + block: + - name: get default value for osd_pool_default_pg_num + command: "{{ docker_exec_cmd | default('') }} ceph --cluster {{ cluster }} daemon mon.{{ hostvars[groups[mon_group_name][0]]['ansible_hostname'] }} config get osd_pool_default_pg_num" + changed_when: false + register: osd_pool_default_pg_num + delegate_to: "{{ groups[mon_group_name][0] }}" -- name: create a rbd pool if it doesn't exist - command: "{{ docker_exec_cmd | default('') }} ceph --cluster {{ cluster }} osd pool create rbd {{ (osd_pool_default_pg_num.stdout | from_json).osd_pool_default_pg_num }}" - changed_when: false - delegate_to: "{{ groups[mon_group_name][0] }}" - when: "'rbd' not in (rbd_pool_exists.stdout | from_json)" + - name: create a rbd pool if it doesn't exist + command: "{{ docker_exec_cmd | default('') }} ceph --cluster {{ cluster }} osd pool create rbd {{ (osd_pool_default_pg_num.stdout | from_json).osd_pool_default_pg_num }}" + changed_when: false + delegate_to: "{{ groups[mon_group_name][0] }}" + + - name: customize pool size + command: "{{ docker_exec_cmd | default('') }} ceph --cluster {{ cluster }} osd pool set rbd size {{ rbd_pool_size | default('') }}" + delegate_to: "{{ groups[mon_group_name][0] }}" + changed_when: false + when: + - rbd_pool_size | default ("") != "" diff --git a/roles/ceph-mds/tasks/create_mds_filesystems.yml b/roles/ceph-mds/tasks/create_mds_filesystems.yml index 8418a5cc7..787fbe2bd 100644 --- a/roles/ceph-mds/tasks/create_mds_filesystems.yml +++ b/roles/ceph-mds/tasks/create_mds_filesystems.yml @@ -1,10 +1,19 @@ --- -- name: create filesystem pools - command: "{{ hostvars[groups[mon_group_name][0]]['docker_exec_cmd'] | default('') }} ceph --cluster {{ cluster }} osd pool create {{ item.name }} {{ item.pgs }}" - changed_when: false - delegate_to: "{{ groups[mon_group_name][0] }}" - with_items: - - "{{ cephfs_pools }}" +- name: filesystem pools related tasks + block: + - name: create filesystem pools + command: "{{ hostvars[groups[mon_group_name][0]]['docker_exec_cmd'] | default('') }} ceph --cluster {{ cluster }} osd pool create {{ item.name }} {{ item.pgs }}" + changed_when: false + delegate_to: "{{ groups[mon_group_name][0] }}" + with_items: + - "{{ cephfs_pools }}" + + - name: customize pool size + command: "{{ hostvars[groups[mon_group_name][0]]['docker_exec_cmd'] | default('') }} ceph --cluster {{ cluster }} osd pool set {{ item.name }} size {{ item.size | default('') }}" + with_items: "{{ cephfs_pools | unique }}" + delegate_to: "{{ groups[mon_group_name][0] }}" + changed_when: false + when: item.size | default ("") != "" - name: check if ceph filesystem already exists command: "{{ hostvars[groups[mon_group_name][0]]['docker_exec_cmd'] | default('') }} ceph --cluster {{ cluster }} fs get {{ cephfs }}" diff --git a/roles/ceph-osd/tasks/openstack_config.yml b/roles/ceph-osd/tasks/openstack_config.yml index 80fb571ad..052345aef 100644 --- a/roles/ceph-osd/tasks/openstack_config.yml +++ b/roles/ceph-osd/tasks/openstack_config.yml @@ -9,42 +9,53 @@ delegate_to: "{{ groups[mon_group_name][0] }}" until: wait_for_all_osds_up.rc == 0 -- name: list existing pool(s) - command: > - {{ hostvars[groups[mon_group_name][0]]['docker_exec_cmd'] | default('') }} ceph --cluster {{ cluster }} - osd pool get {{ item.name }} size - with_items: "{{ openstack_pools | unique }}" - register: created_pools - delegate_to: "{{ groups[mon_group_name][0] }}" - failed_when: false +- name: pool related tasks + block: + - name: list existing pool(s) + command: > + {{ hostvars[groups[mon_group_name][0]]['docker_exec_cmd'] | default('') }} ceph --cluster {{ cluster }} + osd pool get {{ item.name }} size + with_items: "{{ openstack_pools | unique }}" + register: created_pools + delegate_to: "{{ groups[mon_group_name][0] }}" + failed_when: false -- name: create openstack pool(s) - command: > - {{ hostvars[groups[mon_group_name][0]]['docker_exec_cmd'] | default('') }} ceph --cluster {{ cluster }} - osd pool create {{ item.0.name }} - {{ item.0.pg_num }} - {{ item.0.pgp_num | default(item.0.pg_num) }} - {{ 'replicated_rule' if item.0.rule_name | default('replicated_rule') == '' else item.0.rule_name | default('replicated_rule') }} - {{ 1 if item.0.type|default(1) == 'replicated' else 3 if item.0.type|default(1) == 'erasure' else item.0.type|default(1) }} - {%- if (item.0.type | default("1") == '3' or item.0.type | default("1") == 'erasure') and item.0.erasure_profile != '' %} - {{ item.0.erasure_profile }} - {%- endif %} - {{ item.0.expected_num_objects | default('') }} - with_together: - - "{{ openstack_pools | unique }}" - - "{{ created_pools.results }}" - changed_when: false - delegate_to: "{{ groups[mon_group_name][0] }}" - when: - - item.1.get('rc', 0) != 0 + - name: create openstack pool(s) + command: > + {{ hostvars[groups[mon_group_name][0]]['docker_exec_cmd'] | default('') }} ceph --cluster {{ cluster }} + osd pool create {{ item.0.name }} + {{ item.0.pg_num }} + {{ item.0.pgp_num | default(item.0.pg_num) }} + {{ 'replicated_rule' if item.0.rule_name | default('replicated_rule') == '' else item.0.rule_name | default('replicated_rule') }} + {{ 1 if item.0.type|default(1) == 'replicated' else 3 if item.0.type|default(1) == 'erasure' else item.0.type|default(1) }} + {%- if (item.0.type | default("1") == '3' or item.0.type | default("1") == 'erasure') and item.0.erasure_profile != '' %} + {{ item.0.erasure_profile }} + {%- endif %} + {{ item.0.expected_num_objects | default('') }} + with_together: + - "{{ openstack_pools | unique }}" + - "{{ created_pools.results }}" + changed_when: false + delegate_to: "{{ groups[mon_group_name][0] }}" + when: + - item.1.get('rc', 0) != 0 -- name: assign application to pool(s) - command: "{{ hostvars[groups[mon_group_name][0]]['docker_exec_cmd'] | default('') }} ceph --cluster {{ cluster }} osd pool application enable {{ item.name }} {{ item.application }}" - with_items: "{{ openstack_pools | unique }}" - changed_when: false - delegate_to: "{{ groups[mon_group_name][0] }}" - when: - - item.application is defined + - name: customize pool size + command: > + {{ hostvars[groups[mon_group_name][0]]['docker_exec_cmd'] | default('') }} ceph --cluster {{ cluster }} + osd pool set {{ item.name }} size {{ item.size | default('') }} + with_items: "{{ openstack_pools | unique }}" + delegate_to: "{{ groups[mon_group_name][0] }}" + changed_when: false + when: item.size | default ("") != "" + + - name: assign application to pool(s) + command: "{{ hostvars[groups[mon_group_name][0]]['docker_exec_cmd'] | default('') }} ceph --cluster {{ cluster }} osd pool application enable {{ item.name }} {{ item.application }}" + with_items: "{{ openstack_pools | unique }}" + changed_when: false + delegate_to: "{{ groups[mon_group_name][0] }}" + when: + - item.application is defined - name: create openstack cephx key(s) ceph_key: @@ -81,4 +92,5 @@ when: - cephx - openstack_config - - item.0 != groups[mon_group_name] \ No newline at end of file + - item.0 != groups[mon_group_name] + diff --git a/roles/ceph-rgw/defaults/main.yml b/roles/ceph-rgw/defaults/main.yml index b3ff65643..1e108be44 100644 --- a/roles/ceph-rgw/defaults/main.yml +++ b/roles/ceph-rgw/defaults/main.yml @@ -37,10 +37,13 @@ rgw_pull_proto: "http" #rgw_create_pools: # defaults.rgw.buckets.data: # pg_num: 16 +# size: "" # defaults.rgw.buckets.index: # pg_num: 32 +# size: "" # foo: # pg_num: 4 +# size: "" ########## diff --git a/roles/ceph-rgw/tasks/main.yml b/roles/ceph-rgw/tasks/main.yml index 88935e0f7..9a1a65ae5 100644 --- a/roles/ceph-rgw/tasks/main.yml +++ b/roles/ceph-rgw/tasks/main.yml @@ -24,11 +24,21 @@ include_tasks: docker/main.yml when: containerized_deployment -- name: create rgw pools if rgw_create_pools is defined - command: "{{ docker_exec_cmd }} ceph --connect-timeout 5 --cluster {{ cluster }} osd pool create {{ item.key }} {{ item.value.pg_num }}" - changed_when: false - with_dict: "{{ rgw_create_pools }}" - delegate_to: "{{ groups[mon_group_name][0] }}" - run_once: true +- name: rgw pool realted tasks when: - rgw_create_pools is defined + block: + - name: create rgw pools if rgw_create_pools is defined + command: "{{ docker_exec_cmd }} ceph --connect-timeout 5 --cluster {{ cluster }} osd pool create {{ item.key }} {{ item.value.pg_num }}" + changed_when: false + with_dict: "{{ rgw_create_pools }}" + delegate_to: "{{ groups[mon_group_name][0] }}" + run_once: true + + - name: customize pool size + command: "{{ docker_exec_cmd }} ceph --connect-timeout 5 --cluster {{ cluster }} osd pool set {{ item.key }} size {{ item.size | default('') }}" + with_dict: "{{ rgw_create_pools }}" + delegate_to: "{{ groups[mon_group_name][0] }}" + changed_when: false + run_once: true + when: item.size | default ("") != "" diff --git a/tests/functional/centos/7/cluster/ceph-override.json b/tests/functional/centos/7/cluster/ceph-override.json index 1a9600a14..f2ec97114 100644 --- a/tests/functional/centos/7/cluster/ceph-override.json +++ b/tests/functional/centos/7/cluster/ceph-override.json @@ -1,9 +1,20 @@ { "ceph_conf_overrides": { "global": { - "osd_pool_default_pg_num": 12, - "osd_pool_default_size": 1 + "osd_pool_default_pg_num": 12 } }, + "cephfs_pools": [ + { + "name": "cephfs_metadata", + "pgs": 8, + "size": 2 + }, + { + "name": "cephfs_data", + "pgs": 8, + "size": 2 + } + ], "ceph_mon_docker_memory_limit": "2g" } diff --git a/tests/functional/centos/7/cluster/group_vars/all b/tests/functional/centos/7/cluster/group_vars/all index 46934a5b0..5f9c83052 100644 --- a/tests/functional/centos/7/cluster/group_vars/all +++ b/tests/functional/centos/7/cluster/group_vars/all @@ -27,6 +27,7 @@ openstack_glance_pool: type: 1 erasure_profile: "" expected_num_objects: "" + size: "" openstack_cinder_pool: name: "volumes" pg_num: "{{ hostvars[groups[mon_group_name][0]]['osd_pool_default_pg_num'] }}" @@ -35,6 +36,7 @@ openstack_cinder_pool: type: 1 erasure_profile: "" expected_num_objects: "" + size: "" openstack_pools: - "{{ openstack_glance_pool }}" - - "{{ openstack_cinder_pool }}" \ No newline at end of file + - "{{ openstack_cinder_pool }}" diff --git a/tests/functional/centos/7/cluster/group_vars/clients b/tests/functional/centos/7/cluster/group_vars/clients index 21e5c5691..55180053b 100644 --- a/tests/functional/centos/7/cluster/group_vars/clients +++ b/tests/functional/centos/7/cluster/group_vars/clients @@ -9,6 +9,7 @@ test: type: 1 erasure_profile: "" expected_num_objects: "" + size: "" test2: name: "test2" pg_num: "{{ hostvars[groups[mon_group_name][0]]['osd_pool_default_pg_num'] }}" @@ -17,6 +18,7 @@ test2: type: 1 erasure_profile: "" expected_num_objects: "" + size: "" pools: - "{{ test }}" - "{{ test2 }}" diff --git a/tests/functional/centos/7/docker/group_vars/all b/tests/functional/centos/7/docker/group_vars/all index c3b5a92d0..57e385766 100644 --- a/tests/functional/centos/7/docker/group_vars/all +++ b/tests/functional/centos/7/docker/group_vars/all @@ -28,6 +28,7 @@ openstack_glance_pool: type: 1 erasure_profile: "" expected_num_objects: "" + size: "" openstack_cinder_pool: name: "volumes" pg_num: "{{ hostvars[groups[mon_group_name][0]]['osd_pool_default_pg_num'] }}" @@ -36,6 +37,7 @@ openstack_cinder_pool: type: 1 erasure_profile: "" expected_num_objects: "" + size: "" openstack_pools: - "{{ openstack_glance_pool }}" - - "{{ openstack_cinder_pool }}" \ No newline at end of file + - "{{ openstack_cinder_pool }}"