284 lines
8.9 KiB
YAML
284 lines
8.9 KiB
YAML
---
|
|
# Kubernetes dashboard
|
|
# RBAC required. see docs/getting-started.md for access details.
|
|
# dashboard_enabled: false
|
|
|
|
# Helm deployment
|
|
helm_enabled: false
|
|
|
|
# Registry deployment
|
|
registry_enabled: false
|
|
# registry_namespace: kube-system
|
|
# registry_storage_class: ""
|
|
# registry_disk_size: "10Gi"
|
|
|
|
# Metrics Server deployment
|
|
metrics_server_enabled: false
|
|
# metrics_server_container_port: 10250
|
|
# metrics_server_kubelet_insecure_tls: true
|
|
# metrics_server_metric_resolution: 15s
|
|
# metrics_server_kubelet_preferred_address_types: "InternalIP,ExternalIP,Hostname"
|
|
# metrics_server_host_network: false
|
|
# metrics_server_replicas: 1
|
|
|
|
# Rancher Local Path Provisioner
|
|
local_path_provisioner_enabled: false
|
|
# local_path_provisioner_namespace: "local-path-storage"
|
|
# local_path_provisioner_storage_class: "local-path"
|
|
# local_path_provisioner_reclaim_policy: Delete
|
|
# local_path_provisioner_claim_root: /opt/local-path-provisioner/
|
|
# local_path_provisioner_debug: false
|
|
# local_path_provisioner_image_repo: "{{ docker_image_repo }}/rancher/local-path-provisioner"
|
|
# local_path_provisioner_image_tag: "v0.0.24"
|
|
# local_path_provisioner_helper_image_repo: "busybox"
|
|
# local_path_provisioner_helper_image_tag: "latest"
|
|
|
|
# Local volume provisioner deployment
|
|
local_volume_provisioner_enabled: false
|
|
# local_volume_provisioner_namespace: kube-system
|
|
# local_volume_provisioner_nodelabels:
|
|
# - kubernetes.io/hostname
|
|
# - topology.kubernetes.io/region
|
|
# - topology.kubernetes.io/zone
|
|
# local_volume_provisioner_storage_classes:
|
|
# local-storage:
|
|
# host_dir: /mnt/disks
|
|
# mount_dir: /mnt/disks
|
|
# volume_mode: Filesystem
|
|
# fs_type: ext4
|
|
# fast-disks:
|
|
# host_dir: /mnt/fast-disks
|
|
# mount_dir: /mnt/fast-disks
|
|
# block_cleaner_command:
|
|
# - "/scripts/shred.sh"
|
|
# - "2"
|
|
# volume_mode: Filesystem
|
|
# fs_type: ext4
|
|
# local_volume_provisioner_tolerations:
|
|
# - effect: NoSchedule
|
|
# operator: Exists
|
|
|
|
# CSI Volume Snapshot Controller deployment, set this to true if your CSI is able to manage snapshots
|
|
# currently, setting cinder_csi_enabled=true would automatically enable the snapshot controller
|
|
# Longhorn is an external CSI that would also require setting this to true but it is not included in kubespray
|
|
# csi_snapshot_controller_enabled: false
|
|
# csi snapshot namespace
|
|
# snapshot_controller_namespace: kube-system
|
|
|
|
# CephFS provisioner deployment
|
|
cephfs_provisioner_enabled: false
|
|
# cephfs_provisioner_namespace: "cephfs-provisioner"
|
|
# cephfs_provisioner_cluster: ceph
|
|
# cephfs_provisioner_monitors: "172.24.0.1:6789,172.24.0.2:6789,172.24.0.3:6789"
|
|
# cephfs_provisioner_admin_id: admin
|
|
# cephfs_provisioner_secret: secret
|
|
# cephfs_provisioner_storage_class: cephfs
|
|
# cephfs_provisioner_reclaim_policy: Delete
|
|
# cephfs_provisioner_claim_root: /volumes
|
|
# cephfs_provisioner_deterministic_names: true
|
|
|
|
# RBD provisioner deployment
|
|
rbd_provisioner_enabled: false
|
|
# rbd_provisioner_namespace: rbd-provisioner
|
|
# rbd_provisioner_replicas: 2
|
|
# rbd_provisioner_monitors: "172.24.0.1:6789,172.24.0.2:6789,172.24.0.3:6789"
|
|
# rbd_provisioner_pool: kube
|
|
# rbd_provisioner_admin_id: admin
|
|
# rbd_provisioner_secret_name: ceph-secret-admin
|
|
# rbd_provisioner_secret: ceph-key-admin
|
|
# rbd_provisioner_user_id: kube
|
|
# rbd_provisioner_user_secret_name: ceph-secret-user
|
|
# rbd_provisioner_user_secret: ceph-key-user
|
|
# rbd_provisioner_user_secret_namespace: rbd-provisioner
|
|
# rbd_provisioner_fs_type: ext4
|
|
# rbd_provisioner_image_format: "2"
|
|
# rbd_provisioner_image_features: layering
|
|
# rbd_provisioner_storage_class: rbd
|
|
# rbd_provisioner_reclaim_policy: Delete
|
|
|
|
# Gateway API CRDs
|
|
gateway_api_enabled: false
|
|
# gateway_api_experimental_channel: false
|
|
|
|
# Nginx ingress controller deployment
|
|
ingress_nginx_enabled: false
|
|
# ingress_nginx_host_network: false
|
|
# ingress_nginx_service_type: LoadBalancer
|
|
# ingress_nginx_service_annotations:
|
|
# example.io/loadbalancerIPs: 1.2.3.4
|
|
# ingress_nginx_service_nodeport_http: 30080
|
|
# ingress_nginx_service_nodeport_https: 30081
|
|
ingress_publish_status_address: ""
|
|
# ingress_nginx_nodeselector:
|
|
# kubernetes.io/os: "linux"
|
|
# ingress_nginx_tolerations:
|
|
# - key: "node-role.kubernetes.io/control-plane"
|
|
# operator: "Equal"
|
|
# value: ""
|
|
# effect: "NoSchedule"
|
|
# ingress_nginx_namespace: "ingress-nginx"
|
|
# ingress_nginx_insecure_port: 80
|
|
# ingress_nginx_secure_port: 443
|
|
# ingress_nginx_configmap:
|
|
# map-hash-bucket-size: "128"
|
|
# ssl-protocols: "TLSv1.2 TLSv1.3"
|
|
# ingress_nginx_configmap_tcp_services:
|
|
# 9000: "default/example-go:8080"
|
|
# ingress_nginx_configmap_udp_services:
|
|
# 53: "kube-system/coredns:53"
|
|
# ingress_nginx_extra_args:
|
|
# - --default-ssl-certificate=default/foo-tls
|
|
# ingress_nginx_termination_grace_period_seconds: 300
|
|
# ingress_nginx_class: nginx
|
|
# ingress_nginx_without_class: true
|
|
# ingress_nginx_default: false
|
|
|
|
# ALB ingress controller deployment
|
|
ingress_alb_enabled: false
|
|
# alb_ingress_aws_region: "us-east-1"
|
|
# alb_ingress_restrict_scheme: "false"
|
|
# Enables logging on all outbound requests sent to the AWS API.
|
|
# If logging is desired, set to true.
|
|
# alb_ingress_aws_debug: "false"
|
|
|
|
# Cert manager deployment
|
|
cert_manager_enabled: false
|
|
# cert_manager_namespace: "cert-manager"
|
|
# cert_manager_tolerations:
|
|
# - key: node-role.kubernetes.io/control-plane
|
|
# effect: NoSchedule
|
|
# cert_manager_affinity:
|
|
# nodeAffinity:
|
|
# preferredDuringSchedulingIgnoredDuringExecution:
|
|
# - weight: 100
|
|
# preference:
|
|
# matchExpressions:
|
|
# - key: node-role.kubernetes.io/control-plane
|
|
# operator: In
|
|
# values:
|
|
# - ""
|
|
# cert_manager_nodeselector:
|
|
# kubernetes.io/os: "linux"
|
|
|
|
# cert_manager_trusted_internal_ca: |
|
|
# -----BEGIN CERTIFICATE-----
|
|
# [REPLACE with your CA certificate]
|
|
# -----END CERTIFICATE-----
|
|
# cert_manager_leader_election_namespace: kube-system
|
|
|
|
# cert_manager_dns_policy: "ClusterFirst"
|
|
# cert_manager_dns_config:
|
|
# nameservers:
|
|
# - "1.1.1.1"
|
|
# - "8.8.8.8"
|
|
|
|
# cert_manager_controller_extra_args:
|
|
# - "--dns01-recursive-nameservers-only=true"
|
|
# - "--dns01-recursive-nameservers=1.1.1.1:53,8.8.8.8:53"
|
|
|
|
# MetalLB deployment
|
|
metallb_enabled: false
|
|
metallb_speaker_enabled: "{{ metallb_enabled }}"
|
|
metallb_namespace: "metallb-system"
|
|
# metallb_version: v0.13.9
|
|
# metallb_protocol: "layer2"
|
|
# metallb_port: "7472"
|
|
# metallb_memberlist_port: "7946"
|
|
# metallb_config:
|
|
# speaker:
|
|
# nodeselector:
|
|
# kubernetes.io/os: "linux"
|
|
# tolerations:
|
|
# - key: "node-role.kubernetes.io/control-plane"
|
|
# operator: "Equal"
|
|
# value: ""
|
|
# effect: "NoSchedule"
|
|
# controller:
|
|
# nodeselector:
|
|
# kubernetes.io/os: "linux"
|
|
# tolerations:
|
|
# - key: "node-role.kubernetes.io/control-plane"
|
|
# operator: "Equal"
|
|
# value: ""
|
|
# effect: "NoSchedule"
|
|
# address_pools:
|
|
# primary:
|
|
# ip_range:
|
|
# - 10.5.0.0/16
|
|
# auto_assign: true
|
|
# pool1:
|
|
# ip_range:
|
|
# - 10.6.0.0/16
|
|
# auto_assign: true
|
|
# pool2:
|
|
# ip_range:
|
|
# - 10.10.0.0/16
|
|
# auto_assign: true
|
|
# layer2:
|
|
# - primary
|
|
# layer3:
|
|
# defaults:
|
|
# peer_port: 179
|
|
# hold_time: 120s
|
|
# communities:
|
|
# vpn-only: "1234:1"
|
|
# NO_ADVERTISE: "65535:65282"
|
|
# metallb_peers:
|
|
# peer1:
|
|
# peer_address: 10.6.0.1
|
|
# peer_asn: 64512
|
|
# my_asn: 4200000000
|
|
# communities:
|
|
# - vpn-only
|
|
# address_pool:
|
|
# - pool1
|
|
# peer2:
|
|
# peer_address: 10.10.0.1
|
|
# peer_asn: 64513
|
|
# my_asn: 4200000000
|
|
# communities:
|
|
# - NO_ADVERTISE
|
|
# address_pool:
|
|
# - pool2
|
|
|
|
argocd_enabled: false
|
|
# argocd_version: v2.11.0
|
|
# argocd_namespace: argocd
|
|
# Default password:
|
|
# - https://argo-cd.readthedocs.io/en/stable/getting_started/#4-login-using-the-cli
|
|
# ---
|
|
# The initial password is autogenerated and stored in `argocd-initial-admin-secret` in the argocd namespace defined above.
|
|
# Using the argocd CLI the generated password can be automatically be fetched from the current kubectl context with the command:
|
|
# argocd admin initial-password -n argocd
|
|
# ---
|
|
# Use the following var to set admin password
|
|
# argocd_admin_password: "password"
|
|
|
|
# The plugin manager for kubectl
|
|
krew_enabled: false
|
|
krew_root_dir: "/usr/local/krew"
|
|
|
|
# Kube VIP
|
|
kube_vip_enabled: false
|
|
# kube_vip_arp_enabled: true
|
|
# kube_vip_controlplane_enabled: true
|
|
# kube_vip_address: 192.168.56.120
|
|
# loadbalancer_apiserver:
|
|
# address: "{{ kube_vip_address }}"
|
|
# port: 6443
|
|
# kube_vip_interface: eth0
|
|
# kube_vip_services_enabled: false
|
|
# kube_vip_dns_mode: first
|
|
# kube_vip_cp_detect: false
|
|
# kube_vip_leasename: plndr-cp-lock
|
|
# kube_vip_enable_node_labeling: false
|
|
|
|
# Node Feature Discovery
|
|
node_feature_discovery_enabled: false
|
|
# node_feature_discovery_gc_sa_name: node-feature-discovery
|
|
# node_feature_discovery_gc_sa_create: false
|
|
# node_feature_discovery_worker_sa_name: node-feature-discovery
|
|
# node_feature_discovery_worker_sa_create: false
|
|
# node_feature_discovery_master_config:
|
|
# extraLabelNs: ["nvidia.com"]
|