2016-07-04 20:13:18 +08:00
Getting started
===============
2016-12-09 17:38:38 +08:00
Building your own inventory
2017-01-11 00:02:28 +08:00
---------------------------
2016-12-09 17:38:38 +08:00
2017-02-14 18:08:27 +08:00
Ansible inventory can be stored in 3 formats: YAML, JSON, or INI-like. There is
2016-12-09 17:38:38 +08:00
an example inventory located
2019-04-21 02:09:45 +08:00
[here ](https://github.com/kubernetes-sigs/kubespray/blob/master/inventory/sample/inventory.ini ).
2016-12-09 17:38:38 +08:00
2017-01-11 23:15:04 +08:00
You can use an
2018-11-28 16:15:25 +08:00
[inventory generator ](https://github.com/kubernetes-sigs/kubespray/blob/master/contrib/inventory_builder/inventory.py )
2016-12-09 17:38:38 +08:00
to create or modify an Ansible inventory. Currently, it is limited in
2017-09-25 19:19:38 +08:00
functionality and is only used for configuring a basic Kubespray cluster inventory, but it does
support creating inventory file for large clusters as well. It now supports
2017-01-11 23:15:04 +08:00
separated ETCD and Kubernetes master roles from node role if the size exceeds a
2017-09-25 19:19:38 +08:00
certain threshold. Run `python3 contrib/inventory_builder/inventory.py help` help for more information.
2017-01-11 23:15:04 +08:00
Example inventory generator usage:
2017-01-11 19:46:44 +08:00
2018-02-16 20:53:35 +08:00
cp -r inventory/sample inventory/mycluster
declare -a IPS=(10.10.1.3 10.10.1.4 10.10.1.5)
2019-04-21 02:09:45 +08:00
CONFIG_FILE=inventory/mycluster/hosts.yml python3 contrib/inventory_builder/inventory.py ${IPS[@]}
Then use `inventory/mycluster/hosts.yml` as inventory file.
2017-01-11 19:46:44 +08:00
Starting custom deployment
--------------------------
Once you have an inventory, you may want to customize deployment data vars
and start the deployment:
2018-02-16 20:53:35 +08:00
**IMPORTANT**: Edit my\_inventory/groups\_vars/\*.yaml to override data vars:
2017-02-14 18:09:18 +08:00
2019-04-21 02:09:45 +08:00
ansible-playbook -i inventory/mycluster/hosts.yml cluster.yml -b -v \
2018-02-16 20:53:35 +08:00
--private-key=~/.ssh/private_key
2017-01-11 19:46:44 +08:00
See more details in the [ansible guide ](ansible.md ).
2017-05-25 03:49:21 +08:00
Adding nodes
2017-09-07 01:20:25 +08:00
------------
2017-05-25 03:49:21 +08:00
2018-07-12 23:22:11 +08:00
You may want to add worker, master or etcd nodes to your existing cluster. This can be done by re-running the `cluster.yml` playbook, or you can target the bare minimum needed to get kubelet installed on the worker and talking to your masters. This is especially helpful when doing something like autoscaling your clusters.
2017-05-25 03:49:21 +08:00
2018-07-12 23:22:11 +08:00
- Add the new worker node to your inventory in the appropriate group (or utilize a [dynamic inventory ](https://docs.ansible.com/ansible/intro_dynamic_inventory.html )).
2018-11-08 00:18:03 +08:00
- Run the ansible-playbook command, substituting `cluster.yml` for `scale.yml` :
2018-02-16 20:53:35 +08:00
2019-04-21 02:09:45 +08:00
ansible-playbook -i inventory/mycluster/hosts.yml scale.yml -b -v \
2018-02-16 20:53:35 +08:00
--private-key=~/.ssh/private_key
2017-09-07 01:20:25 +08:00
2018-03-13 14:05:03 +08:00
Remove nodes
------------
You may want to remove **worker** nodes to your existing cluster. This can be done by re-running the `remove-node.yml` playbook. First, all nodes will be drained, then stop some kubernetes services and delete some certificates, and finally execute the kubectl command to delete these nodes. This can be combined with the add node function, This is generally helpful when doing something like autoscaling your clusters. Of course if a node is not working, you can remove the node and install it again.
2018-06-29 13:51:04 +08:00
Add worker nodes to the list under kube-node if you want to delete them (or utilize a [dynamic inventory ](https://docs.ansible.com/ansible/intro_dynamic_inventory.html )).
2019-04-21 02:09:45 +08:00
ansible-playbook -i inventory/mycluster/hosts.yml remove-node.yml -b -v \
2018-06-29 13:51:04 +08:00
--private-key=~/.ssh/private_key
2019-01-08 19:09:18 +08:00
Use `--extra-vars "node=<nodename>,<nodename2>"` to select the node you want to delete.
2018-06-29 13:51:04 +08:00
```
2019-04-21 02:09:45 +08:00
ansible-playbook -i inventory/mycluster/hosts.yml remove-node.yml -b -v \
2018-06-29 13:51:04 +08:00
--private-key=~/.ssh/private_key \
--extra-vars "node=nodename,nodename2"
```
2018-03-13 14:05:03 +08:00
2017-09-07 01:20:25 +08:00
Connecting to Kubernetes
------------------------
2018-02-16 20:53:35 +08:00
2017-09-07 01:20:25 +08:00
By default, Kubespray configures kube-master hosts with insecure access to
kube-apiserver via port 8080. A kubeconfig file is not necessary in this case,
2018-02-16 20:53:35 +08:00
because kubectl will use < http: / / localhost:8080 > to connect. The kubeconfig files
2017-09-07 01:20:25 +08:00
generated will point to localhost (on kube-masters) and kube-node hosts will
connect either to a localhost nginx proxy or to a loadbalancer if configured.
2017-11-03 15:11:36 +08:00
More details on this process are in the [HA guide ](ha-mode.md ).
2017-09-07 01:20:25 +08:00
2018-02-16 20:53:35 +08:00
Kubespray permits connecting to the cluster remotely on any IP of any
kube-master host on port 6443 by default. However, this requires
authentication. One could generate a kubeconfig based on one installed
2017-09-07 01:20:25 +08:00
kube-master hosts (needs improvement) or connect with a username and password.
2017-09-11 17:47:04 +08:00
By default, a user with admin rights is created, named `kube` .
2018-02-16 20:53:35 +08:00
The password can be viewed after deployment by looking at the file
2018-09-04 00:04:50 +08:00
`{{ credentials_dir }}/kube_user.creds` (`credentials_dir` is set to `{{ inventory_dir }}/credentials` by default). This contains a randomly generated
2017-09-07 01:20:25 +08:00
password. If you wish to set your own password, just precreate/modify this
2018-02-16 20:53:35 +08:00
file yourself.
2017-09-07 01:20:25 +08:00
For more information on kubeconfig and accessing a Kubernetes cluster, refer to
the Kubernetes [documentation ](https://kubernetes.io/docs/tasks/access-application-cluster/configure-access-multiple-clusters/ ).
2017-09-10 04:38:03 +08:00
Accessing Kubernetes Dashboard
------------------------------
2017-11-10 05:59:30 +08:00
As of kubernetes-dashboard v1.7.x:
2018-02-16 20:53:35 +08:00
- New login options that use apiserver auth proxying of token/basic/kubeconfig by default
- Requires RBAC in authorization\_modes
- Only serves over https
- No longer available at < https: // first_master:6443 / ui > until apiserver is updated with the https proxy URL
2017-09-10 04:38:03 +08:00
2017-11-10 05:59:30 +08:00
If the variable `dashboard_enabled` is set (default is true), then you can access the Kubernetes Dashboard at the following URL, You will be prompted for credentials:
2018-02-16 20:53:35 +08:00
< https: / / first_master:6443 / api / v1 / namespaces / kube-system / services / https:kubernetes-dashboard: / proxy / # ! / login >
2017-09-10 04:38:03 +08:00
2017-11-10 05:59:30 +08:00
Or you can run 'kubectl proxy' from your local machine to access dashboard in your browser from:
2018-02-16 20:53:35 +08:00
< http: / / localhost:8001 / api / v1 / namespaces / kube-system / services / https:kubernetes-dashboard: / proxy / # ! / login >
2017-09-18 20:30:57 +08:00
2018-02-16 20:53:35 +08:00
It is recommended to access dashboard from behind a gateway (like Ingress Controller) that enforces an authentication token. Details and other access options here: < https: / / github . com / kubernetes / dashboard / wiki / Accessing-Dashboard---1 . 7 . X-and-above >
2017-10-16 03:41:17 +08:00
2017-09-18 20:30:57 +08:00
Accessing Kubernetes API
------------------------
The main client of Kubernetes is `kubectl` . It is installed on each kube-master
host and can optionally be configured on your ansible host by setting
2018-02-16 20:53:35 +08:00
`kubectl_localhost: true` and `kubeconfig_localhost: true` in the configuration:
- If `kubectl_localhost` enabled, `kubectl` will download onto `/usr/local/bin/` and setup with bash completion. A helper script `inventory/mycluster/artifacts/kubectl.sh` also created for setup with below `admin.conf` .
- If `kubeconfig_localhost` enabled `admin.conf` will appear in the `inventory/mycluster/artifacts/` directory after deployment.
You can see a list of nodes by running the following commands:
2017-09-18 20:30:57 +08:00
2018-02-16 20:53:35 +08:00
cd inventory/mycluster/artifacts
./kubectl.sh get nodes
2017-09-18 20:30:57 +08:00
2018-02-16 20:53:35 +08:00
If desired, copy admin.conf to ~/.kube/config.