mirror of
https://github.com/techno-tim/k3s-ansible.git
synced 2025-12-28 19:52:40 +01:00
Compare commits
44 Commits
v1.29.0+k3
...
39361d2d8e
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
39361d2d8e | ||
|
|
a1b108a651 | ||
|
|
aa2738ae83 | ||
|
|
16fb511d49 | ||
|
|
62b80237ef | ||
|
|
c7c727c3dc | ||
|
|
0422bfa2ac | ||
|
|
0333406725 | ||
|
|
f4a19d368b | ||
|
|
02d212c007 | ||
|
|
80095250e9 | ||
|
|
4fe2c92795 | ||
|
|
b3f2a4addc | ||
|
|
cb03ee829e | ||
|
|
9e2e82faeb | ||
|
|
7c1f6cbe42 | ||
|
|
604eb7a6e6 | ||
|
|
a204ed5169 | ||
|
|
b6608ca3e4 | ||
|
|
8252a45dfd | ||
|
|
c99f098c2e | ||
|
|
7867b87d85 | ||
|
|
dfe19f3731 | ||
|
|
a46d97a28d | ||
|
|
dc9d571f17 | ||
|
|
6742551e5c | ||
|
|
fb3478a086 | ||
|
|
518c5bb62a | ||
|
|
3f5d8dfe9f | ||
|
|
efbfadcb93 | ||
|
|
f81ec04ba2 | ||
|
|
8432d3bc66 | ||
|
|
14ae9df1bc | ||
|
|
f175716339 | ||
|
|
955c6f6b4a | ||
|
|
3b74985767 | ||
|
|
9ace193ade | ||
|
|
83a0be3afd | ||
|
|
029eba6102 | ||
|
|
0c8253b3a5 | ||
|
|
326b71dfa2 | ||
|
|
b95d6dd2cc | ||
|
|
e4146b4ca9 | ||
|
|
1fb10faf7f |
17
.github/download-boxes.sh
vendored
17
.github/download-boxes.sh
vendored
@@ -9,12 +9,17 @@ set -euo pipefail
|
||||
GIT_ROOT=$(git rev-parse --show-toplevel)
|
||||
PROVIDER=virtualbox
|
||||
|
||||
# Read all boxes for all platforms from the "molecule.yml" files
|
||||
all_boxes=$(cat "${GIT_ROOT}"/molecule/*/molecule.yml |
|
||||
yq -r '.platforms[].box' | # Read the "box" property of each node under "platforms"
|
||||
grep --invert-match --regexp=--- | # Filter out file separators
|
||||
sort |
|
||||
uniq)
|
||||
yq --version
|
||||
|
||||
# Define the path to the molecule.yml files
|
||||
MOLECULE_YML_PATH="${GIT_ROOT}/molecule/*/molecule.yml"
|
||||
|
||||
# Extract and sort unique boxes from all molecule.yml files
|
||||
all_boxes=$(for file in $MOLECULE_YML_PATH; do
|
||||
yq eval '.platforms[].box' "$file"
|
||||
done | sort -u)
|
||||
|
||||
echo all_boxes: "$all_boxes"
|
||||
|
||||
# Read the boxes that are currently present on the system (for the current provider)
|
||||
present_boxes=$(
|
||||
|
||||
6
.github/workflows/cache.yml
vendored
6
.github/workflows/cache.yml
vendored
@@ -11,19 +11,19 @@ jobs:
|
||||
|
||||
steps:
|
||||
- name: Check out the codebase
|
||||
uses: actions/checkout@b4ffde65f46336ab88eb53be808477a3936bae11 # 4.1.1
|
||||
uses: actions/checkout@0ad4b8fadaa221de15dcec353f45205ec38ea70b # 4.1.4
|
||||
with:
|
||||
ref: ${{ github.event.pull_request.head.sha }}
|
||||
|
||||
- name: Set up Python ${{ env.PYTHON_VERSION }}
|
||||
uses: actions/setup-python@0a5c61591373683505ea898e09a3ea4f39ef2b9c # 5.0.0
|
||||
uses: actions/setup-python@82c7e631bb3cdc910f68e0081d67478d79c6982d # 5.1.0
|
||||
with:
|
||||
python-version: ${{ env.PYTHON_VERSION }}
|
||||
cache: 'pip' # caching pip dependencies
|
||||
|
||||
- name: Cache Vagrant boxes
|
||||
id: cache-vagrant
|
||||
uses: actions/cache@13aacd865c20de90d75de3b17ebe84f7a17d57d2 # 4.0
|
||||
uses: actions/cache@0c45773b623bea8c8e75f6c82b208c3cf94ea4f9 # 4.0
|
||||
with:
|
||||
lookup-only: true #if it exists, we don't need to restore and can skip the next step
|
||||
path: |
|
||||
|
||||
12
.github/workflows/lint.yml
vendored
12
.github/workflows/lint.yml
vendored
@@ -11,18 +11,18 @@ jobs:
|
||||
|
||||
steps:
|
||||
- name: Check out the codebase
|
||||
uses: actions/checkout@b4ffde65f46336ab88eb53be808477a3936bae11 # 4.1.1
|
||||
uses: actions/checkout@0ad4b8fadaa221de15dcec353f45205ec38ea70b # 4.1.4
|
||||
with:
|
||||
ref: ${{ github.event.pull_request.head.sha }}
|
||||
|
||||
- name: Set up Python ${{ env.PYTHON_VERSION }}
|
||||
uses: actions/setup-python@0a5c61591373683505ea898e09a3ea4f39ef2b9c # 5.0.0
|
||||
uses: actions/setup-python@82c7e631bb3cdc910f68e0081d67478d79c6982d # 5.1.0
|
||||
with:
|
||||
python-version: ${{ env.PYTHON_VERSION }}
|
||||
cache: 'pip' # caching pip dependencies
|
||||
|
||||
- name: Restore Ansible cache
|
||||
uses: actions/cache/restore@13aacd865c20de90d75de3b17ebe84f7a17d57d2 # 4.0
|
||||
uses: actions/cache/restore@0c45773b623bea8c8e75f6c82b208c3cf94ea4f9 # 4.0
|
||||
with:
|
||||
path: ~/.ansible/collections
|
||||
key: ansible-${{ hashFiles('collections/requirements.yml') }}
|
||||
@@ -38,16 +38,16 @@ jobs:
|
||||
echo "::endgroup::"
|
||||
|
||||
- name: Run pre-commit
|
||||
uses: pre-commit/action@646c83fcd040023954eafda54b4db0192ce70507 # 3.0.0
|
||||
uses: pre-commit/action@2c7b3805fd2a0fd8c1884dcaebf91fc102a13ecd # 3.0.1
|
||||
|
||||
ensure-pinned-actions:
|
||||
name: Ensure SHA Pinned Actions
|
||||
runs-on: self-hosted
|
||||
steps:
|
||||
- name: Checkout code
|
||||
uses: actions/checkout@b4ffde65f46336ab88eb53be808477a3936bae11 # 4.1.1
|
||||
uses: actions/checkout@0ad4b8fadaa221de15dcec353f45205ec38ea70b # 4.1.4
|
||||
- name: Ensure SHA pinned actions
|
||||
uses: zgosalvez/github-actions-ensure-sha-pinned-actions@ba37328d4ea95eaf8b3bd6c6cef308f709a5f2ec # 3.0.3
|
||||
uses: zgosalvez/github-actions-ensure-sha-pinned-actions@40e45e738b3cad2729f599d8afc6ed02184e1dbd # 3.0.5
|
||||
with:
|
||||
allowlist: |
|
||||
aws-actions/
|
||||
|
||||
10
.github/workflows/test.yml
vendored
10
.github/workflows/test.yml
vendored
@@ -10,7 +10,7 @@ jobs:
|
||||
matrix:
|
||||
scenario:
|
||||
- default
|
||||
- ipv6
|
||||
# - ipv6
|
||||
- single_node
|
||||
- calico
|
||||
- cilium
|
||||
@@ -21,7 +21,7 @@ jobs:
|
||||
|
||||
steps:
|
||||
- name: Check out the codebase
|
||||
uses: actions/checkout@b4ffde65f46336ab88eb53be808477a3936bae11 # 4.1.1
|
||||
uses: actions/checkout@0ad4b8fadaa221de15dcec353f45205ec38ea70b # 4.1.4
|
||||
with:
|
||||
ref: ${{ github.event.pull_request.head.sha }}
|
||||
|
||||
@@ -59,13 +59,13 @@ jobs:
|
||||
EOF
|
||||
|
||||
- name: Set up Python ${{ env.PYTHON_VERSION }}
|
||||
uses: actions/setup-python@0a5c61591373683505ea898e09a3ea4f39ef2b9c # 5.0.0
|
||||
uses: actions/setup-python@82c7e631bb3cdc910f68e0081d67478d79c6982d # 5.1.0
|
||||
with:
|
||||
python-version: ${{ env.PYTHON_VERSION }}
|
||||
cache: 'pip' # caching pip dependencies
|
||||
|
||||
- name: Restore vagrant Boxes cache
|
||||
uses: actions/cache/restore@13aacd865c20de90d75de3b17ebe84f7a17d57d2 # 4.0
|
||||
uses: actions/cache/restore@0c45773b623bea8c8e75f6c82b208c3cf94ea4f9 # 4.0
|
||||
with:
|
||||
path: ~/.vagrant.d/boxes
|
||||
key: vagrant-boxes-${{ hashFiles('**/molecule.yml') }}
|
||||
@@ -118,7 +118,7 @@ jobs:
|
||||
|
||||
- name: Upload log files
|
||||
if: always() # do this even if a step before has failed
|
||||
uses: actions/upload-artifact@26f96dfa697d77e81fd5907df203aa23a56210a8 # 4.3.0
|
||||
uses: actions/upload-artifact@65462800fd760344b1a7b4382951275a0abb4808 # 4.3.3
|
||||
with:
|
||||
name: logs
|
||||
path: |
|
||||
|
||||
16
README.md
16
README.md
@@ -96,8 +96,22 @@ ansible-playbook reset.yml -i inventory/my-cluster/hosts.ini
|
||||
To copy your `kube config` locally so that you can access your **Kubernetes** cluster run:
|
||||
|
||||
```bash
|
||||
scp debian@master_ip:~/.kube/config ~/.kube/config
|
||||
scp debian@master_ip:/etc/rancher/k3s/k3s.yaml ~/.kube/config
|
||||
```
|
||||
If you get file Permission denied, go into the node and temporarly run:
|
||||
```bash
|
||||
sudo chmod 777 /etc/rancher/k3s/k3s.yaml
|
||||
```
|
||||
Then copy with the scp command and reset the permissions back to:
|
||||
```bash
|
||||
sudo chmod 600 /etc/rancher/k3s/k3s.yaml
|
||||
```
|
||||
|
||||
You'll then want to modify the config to point to master IP by running:
|
||||
```bash
|
||||
sudo nano ~/.kube/config
|
||||
```
|
||||
Then change `server: https://127.0.0.1:6443` to match your master IP: `server: https://192.168.1.222:6443`
|
||||
|
||||
### 🔨 Testing your cluster
|
||||
|
||||
|
||||
@@ -1,5 +1,5 @@
|
||||
---
|
||||
k3s_version: v1.29.0+k3s1
|
||||
k3s_version: v1.29.2+k3s1
|
||||
# this is the user that has ssh access to these machines
|
||||
ansible_user: ansibleuser
|
||||
systemd_dir: /etc/systemd/system
|
||||
@@ -13,13 +13,13 @@ flannel_iface: "eth0"
|
||||
# uncomment calico_iface to use tigera operator/calico cni instead of flannel https://docs.tigera.io/calico/latest/about
|
||||
# calico_iface: "eth0"
|
||||
calico_ebpf: false # use eBPF dataplane instead of iptables
|
||||
calico_tag: "v3.27.0" # calico version tag
|
||||
calico_tag: "v3.27.2" # calico version tag
|
||||
|
||||
# uncomment cilium_iface to use cilium cni instead of flannel or calico
|
||||
# ensure v4.19.57, v5.1.16, v5.2.0 or more recent kernel
|
||||
# cilium_iface: "eth0"
|
||||
cilium_mode: "native" # native when nodes on same subnet or using bgp, else set routed
|
||||
cilium_tag: "v1.14.6" # cilium version tag
|
||||
cilium_tag: "v1.15.2" # cilium version tag
|
||||
cilium_hubble: true # enable hubble observability relay and ui
|
||||
|
||||
# if using calico or cilium, you may specify the cluster pod cidr pool
|
||||
@@ -72,7 +72,7 @@ extra_agent_args: >-
|
||||
{{ extra_args }}
|
||||
|
||||
# image tag for kube-vip
|
||||
kube_vip_tag_version: "v0.6.4"
|
||||
kube_vip_tag_version: "v0.7.2"
|
||||
|
||||
# tag for kube-vip-cloud-provider manifest
|
||||
# kube_vip_cloud_provider_tag_version: "main"
|
||||
@@ -93,8 +93,8 @@ metal_lb_mode: "layer2"
|
||||
# metal_lb_bgp_peer_address: "192.168.30.1"
|
||||
|
||||
# image tag for metal lb
|
||||
metal_lb_speaker_tag_version: "v0.13.12"
|
||||
metal_lb_controller_tag_version: "v0.13.12"
|
||||
metal_lb_speaker_tag_version: "v0.14.3"
|
||||
metal_lb_controller_tag_version: "v0.14.3"
|
||||
|
||||
# metallb ip range for load balancer
|
||||
metal_lb_ip_range: "192.168.30.80-192.168.30.90"
|
||||
@@ -160,6 +160,10 @@ custom_registries_yaml: |
|
||||
username: yourusername
|
||||
password: yourpassword
|
||||
|
||||
# On some distros like Diet Pi, there is no dbus installed. dbus required by the default reboot command.
|
||||
# Uncomment if you need a custom reboot command
|
||||
# custom_reboot_command: /usr/sbin/shutdown -r now
|
||||
|
||||
# Only enable and configure these if you access the internet through a proxy
|
||||
# proxy_env:
|
||||
# HTTP_PROXY: "http://proxy.domain.local:3128"
|
||||
|
||||
@@ -6,4 +6,5 @@
|
||||
- name: Reboot the nodes (and Wait upto 5 mins max)
|
||||
become: true
|
||||
reboot:
|
||||
reboot_command: "{{ custom_reboot_command | default(omit) }}"
|
||||
reboot_timeout: 300
|
||||
|
||||
@@ -6,7 +6,7 @@
|
||||
#
|
||||
ansible-compat==4.1.11
|
||||
# via molecule
|
||||
ansible-core==2.16.3
|
||||
ansible-core==2.16.6
|
||||
# via
|
||||
# -r requirements.in
|
||||
# ansible-compat
|
||||
@@ -77,7 +77,7 @@ molecule==6.0.3
|
||||
# via
|
||||
# -r requirements.in
|
||||
# molecule-plugins
|
||||
molecule-plugins[vagrant]==23.5.0
|
||||
molecule-plugins[vagrant]==23.5.3
|
||||
# via -r requirements.in
|
||||
netaddr==0.10.1
|
||||
# via -r requirements.in
|
||||
@@ -96,9 +96,9 @@ platformdirs==4.1.0
|
||||
# via virtualenv
|
||||
pluggy==1.3.0
|
||||
# via molecule
|
||||
pre-commit==3.6.0
|
||||
pre-commit==3.7.0
|
||||
# via -r requirements.in
|
||||
pre-commit-hooks==4.5.0
|
||||
pre-commit-hooks==4.6.0
|
||||
# via -r requirements.in
|
||||
pyasn1==0.5.1
|
||||
# via
|
||||
|
||||
@@ -12,6 +12,7 @@
|
||||
- name: Reboot and wait for node to come back up
|
||||
become: true
|
||||
reboot:
|
||||
reboot_command: "{{ custom_reboot_command | default(omit) }}"
|
||||
reboot_timeout: 3600
|
||||
|
||||
- name: Revert changes to Proxmox cluster
|
||||
|
||||
8
roles/download/meta/main.yml
Normal file
8
roles/download/meta/main.yml
Normal file
@@ -0,0 +1,8 @@
|
||||
---
|
||||
argument_specs:
|
||||
main:
|
||||
short_description: Manage the downloading of K3S binaries
|
||||
options:
|
||||
k3s_version:
|
||||
description: The desired version of K3S
|
||||
required: true
|
||||
4
roles/k3s_agent/defaults/main.yml
Normal file
4
roles/k3s_agent/defaults/main.yml
Normal file
@@ -0,0 +1,4 @@
|
||||
---
|
||||
extra_agent_args: ""
|
||||
group_name_master: master
|
||||
systemd_dir: /etc/systemd/system
|
||||
34
roles/k3s_agent/meta/main.yml
Normal file
34
roles/k3s_agent/meta/main.yml
Normal file
@@ -0,0 +1,34 @@
|
||||
---
|
||||
argument_specs:
|
||||
main:
|
||||
short_description: Setup k3s agents
|
||||
options:
|
||||
apiserver_endpoint:
|
||||
description: Virtual ip-address configured on each master
|
||||
required: true
|
||||
|
||||
extra_agent_args:
|
||||
description: Extra arguments for agents nodes
|
||||
|
||||
group_name_master:
|
||||
description: Name of the master group
|
||||
default: master
|
||||
|
||||
k3s_token:
|
||||
description: Token used to communicate between masters
|
||||
|
||||
proxy_env:
|
||||
type: dict
|
||||
description: Internet proxy configurations
|
||||
default: ~
|
||||
options:
|
||||
HTTP_PROXY:
|
||||
required: true
|
||||
HTTPS_PROXY:
|
||||
required: true
|
||||
NO_PROXY:
|
||||
required: true
|
||||
|
||||
systemd_dir:
|
||||
description: Path to systemd services
|
||||
default: /etc/systemd/system
|
||||
@@ -1,5 +1,4 @@
|
||||
---
|
||||
|
||||
- name: Create k3s-node.service.d directory
|
||||
file:
|
||||
path: '{{ systemd_dir }}/k3s-node.service.d'
|
||||
@@ -7,7 +6,7 @@
|
||||
owner: root
|
||||
group: root
|
||||
mode: '0755'
|
||||
|
||||
when: proxy_env is defined
|
||||
|
||||
- name: Copy K3s http_proxy conf file
|
||||
template:
|
||||
@@ -16,3 +15,4 @@
|
||||
owner: root
|
||||
group: root
|
||||
mode: '0755'
|
||||
when: proxy_env is defined
|
||||
|
||||
@@ -1,19 +1,35 @@
|
||||
---
|
||||
- name: Check for PXE-booted system
|
||||
block:
|
||||
- name: Check if system is PXE-booted
|
||||
ansible.builtin.command:
|
||||
cmd: cat /proc/cmdline
|
||||
register: boot_cmdline
|
||||
changed_when: false
|
||||
check_mode: false
|
||||
|
||||
- name: Set fact for PXE-booted system
|
||||
ansible.builtin.set_fact:
|
||||
is_pxe_booted: "{{ 'root=/dev/nfs' in boot_cmdline.stdout }}"
|
||||
when: boot_cmdline.stdout is defined
|
||||
|
||||
- name: Include http_proxy configuration tasks
|
||||
ansible.builtin.include_tasks: http_proxy.yml
|
||||
|
||||
- name: Deploy K3s http_proxy conf
|
||||
include_tasks: http_proxy.yml
|
||||
when: proxy_env is defined
|
||||
|
||||
- name: Copy K3s service file
|
||||
template:
|
||||
- name: Configure the k3s service
|
||||
ansible.builtin.template:
|
||||
src: "k3s.service.j2"
|
||||
dest: "{{ systemd_dir }}/k3s-node.service"
|
||||
owner: root
|
||||
group: root
|
||||
mode: 0755
|
||||
mode: '0755'
|
||||
|
||||
- name: Enable and check K3s service
|
||||
systemd:
|
||||
- name: Manage k3s service
|
||||
ansible.builtin.systemd:
|
||||
name: k3s-node
|
||||
daemon_reload: true
|
||||
state: restarted
|
||||
|
||||
@@ -7,11 +7,14 @@ After=network-online.target
|
||||
Type=notify
|
||||
ExecStartPre=-/sbin/modprobe br_netfilter
|
||||
ExecStartPre=-/sbin/modprobe overlay
|
||||
ExecStart=/usr/local/bin/k3s agent --server https://{{ apiserver_endpoint | ansible.utils.ipwrap }}:6443 --token {{ hostvars[groups[group_name_master | default('master')][0]]['token'] | default(k3s_token) }} {{ extra_agent_args | default("") }}
|
||||
# Conditional snapshotter based on PXE boot status
|
||||
ExecStart=/usr/local/bin/k3s agent \
|
||||
--server https://{{ apiserver_endpoint | ansible.utils.ipwrap }}:6443 \
|
||||
{% if is_pxe_booted | default(false) %}--snapshotter native \
|
||||
{% endif %}--token {{ hostvars[groups[group_name_master | default('master')][0]]['token'] | default(k3s_token) }} \
|
||||
{{ extra_agent_args }}
|
||||
KillMode=process
|
||||
Delegate=yes
|
||||
# Having non-zero Limit*s causes performance problems due to accounting overhead
|
||||
# in the kernel. We recommend using cgroups to do container-local accounting.
|
||||
LimitNOFILE=1048576
|
||||
LimitNPROC=infinity
|
||||
LimitCORE=infinity
|
||||
|
||||
@@ -1,6 +0,0 @@
|
||||
---
|
||||
# Indicates whether custom registries for k3s should be configured
|
||||
# Possible values:
|
||||
# - present
|
||||
# - absent
|
||||
state: present
|
||||
20
roles/k3s_custom_registries/meta/main.yml
Normal file
20
roles/k3s_custom_registries/meta/main.yml
Normal file
@@ -0,0 +1,20 @@
|
||||
---
|
||||
argument_specs:
|
||||
main:
|
||||
short_description: Configure the use of a custom container registry
|
||||
options:
|
||||
custom_registries_yaml:
|
||||
description:
|
||||
- YAML block defining custom registries.
|
||||
- >
|
||||
The following is an example that pulls all images used in
|
||||
this playbook through your private registries.
|
||||
- >
|
||||
It also allows you to pull your own images from your private
|
||||
registry, without having to use imagePullSecrets in your
|
||||
deployments.
|
||||
- >
|
||||
If all you need is your own images and you don't care about
|
||||
caching the docker/quay/ghcr.io images, you can just remove
|
||||
those from the mirrors: section.
|
||||
required: true
|
||||
@@ -1,12 +1,18 @@
|
||||
---
|
||||
# If you want to explicitly define an interface that ALL control nodes
|
||||
# should use to propagate the VIP, define it here. Otherwise, kube-vip
|
||||
# will determine the right interface automatically at runtime.
|
||||
kube_vip_iface: null
|
||||
extra_server_args: ""
|
||||
|
||||
# Name of the master group
|
||||
group_name_master: master
|
||||
|
||||
kube_vip_iface: ~
|
||||
kube_vip_cloud_provider_tag_version: main
|
||||
kube_vip_tag_version: v0.7.2
|
||||
|
||||
metal_lb_controller_tag_version: v0.14.3
|
||||
metal_lb_speaker_tag_version: v0.14.3
|
||||
metal_lb_type: native
|
||||
|
||||
retry_count: 20
|
||||
|
||||
# yamllint disable rule:line-length
|
||||
server_init_args: >-
|
||||
{% if groups[group_name_master | default('master')] | length > 1 %}
|
||||
@@ -17,4 +23,6 @@ server_init_args: >-
|
||||
{% endif %}
|
||||
--token {{ k3s_token }}
|
||||
{% endif %}
|
||||
{{ extra_server_args | default('') }}
|
||||
{{ extra_server_args }}
|
||||
|
||||
systemd_dir: /etc/systemd/system
|
||||
|
||||
85
roles/k3s_server/meta/main.yml
Normal file
85
roles/k3s_server/meta/main.yml
Normal file
@@ -0,0 +1,85 @@
|
||||
---
|
||||
argument_specs:
|
||||
main:
|
||||
short_description: Setup k3s servers
|
||||
options:
|
||||
apiserver_endpoint:
|
||||
description: Virtual ip-address configured on each master
|
||||
required: true
|
||||
|
||||
cilium_bgp:
|
||||
description:
|
||||
- Enable cilium BGP control plane for LB services and pod cidrs.
|
||||
- Disables the use of MetalLB.
|
||||
type: bool
|
||||
default: ~
|
||||
|
||||
cilium_iface:
|
||||
description: The network interface used for when Cilium is enabled
|
||||
default: ~
|
||||
|
||||
extra_server_args:
|
||||
description: Extra arguments for server nodes
|
||||
default: ""
|
||||
|
||||
group_name_master:
|
||||
description: Name of the master group
|
||||
default: master
|
||||
|
||||
kube_vip_iface:
|
||||
description:
|
||||
- Explicitly define an interface that ALL control nodes
|
||||
- should use to propagate the VIP, define it here.
|
||||
- Otherwise, kube-vip will determine the right interface
|
||||
- automatically at runtime.
|
||||
default: ~
|
||||
|
||||
kube_vip_tag_version:
|
||||
description: Image tag for kube-vip
|
||||
default: v0.7.2
|
||||
|
||||
kube_vip_cloud_provider_tag_version:
|
||||
description: Tag for kube-vip-cloud-provider manifest when enabled
|
||||
default: main
|
||||
|
||||
kube_vip_lb_ip_range:
|
||||
description: IP range for kube-vip load balancer
|
||||
default: ~
|
||||
|
||||
metal_lb_controller_tag_version:
|
||||
description: Image tag for MetalLB
|
||||
default: v0.14.3
|
||||
|
||||
metal_lb_speaker_tag_version:
|
||||
description: Image tag for MetalLB
|
||||
default: v0.14.3
|
||||
|
||||
metal_lb_type:
|
||||
choices:
|
||||
- frr
|
||||
- native
|
||||
default: native
|
||||
|
||||
proxy_env:
|
||||
type: dict
|
||||
description: Internet proxy configurations
|
||||
default: ~
|
||||
options:
|
||||
HTTP_PROXY:
|
||||
required: true
|
||||
HTTPS_PROXY:
|
||||
required: true
|
||||
NO_PROXY:
|
||||
required: true
|
||||
|
||||
retry_count:
|
||||
description: Amount of retries when verifying that nodes joined
|
||||
type: int
|
||||
default: 20
|
||||
|
||||
server_init_args:
|
||||
description: Arguments for server nodes
|
||||
|
||||
systemd_dir:
|
||||
description: Path to systemd services
|
||||
default: /etc/systemd/system
|
||||
@@ -1,5 +1,4 @@
|
||||
---
|
||||
|
||||
- name: Create k3s.service.d directory
|
||||
file:
|
||||
path: '{{ systemd_dir }}/k3s.service.d'
|
||||
|
||||
@@ -1,5 +1,4 @@
|
||||
---
|
||||
|
||||
- name: Stop k3s-init
|
||||
systemd:
|
||||
name: k3s-init
|
||||
|
||||
@@ -10,7 +10,7 @@
|
||||
|
||||
- name: Download vip rbac manifest to first master
|
||||
ansible.builtin.get_url:
|
||||
url: "https://raw.githubusercontent.com/kube-vip/kube-vip/{{ kube_vip_tag_version }}/docs/manifests/rbac.yaml"
|
||||
url: "https://kube-vip.io/manifests/rbac.yaml"
|
||||
dest: "/var/lib/rancher/k3s/server/manifests/vip-rbac.yaml"
|
||||
owner: root
|
||||
group: root
|
||||
|
||||
@@ -1,6 +1,20 @@
|
||||
---
|
||||
# Timeout to wait for MetalLB services to come up
|
||||
metal_lb_available_timeout: 240s
|
||||
bpf_lb_algorithm: maglev
|
||||
bpf_lb_mode: hybrid
|
||||
|
||||
# Name of the master group
|
||||
calico_ebpf: false
|
||||
calico_tag: v3.27.2
|
||||
|
||||
cilium_bgp: false
|
||||
cilium_exportPodCIDR: true
|
||||
cilium_hubble: true
|
||||
cilium_mode: native
|
||||
|
||||
cluster_cidr: 10.52.0.0/16
|
||||
enable_bpf_masquerade: true
|
||||
kube_proxy_replacement: true
|
||||
group_name_master: master
|
||||
|
||||
metal_lb_mode: layer2
|
||||
metal_lb_available_timeout: 240s
|
||||
metal_lb_controller_tag_version: v0.14.3
|
||||
|
||||
94
roles/k3s_server_post/meta/main.yml
Normal file
94
roles/k3s_server_post/meta/main.yml
Normal file
@@ -0,0 +1,94 @@
|
||||
---
|
||||
argument_specs:
|
||||
main:
|
||||
short_description: Configure k3s cluster
|
||||
options:
|
||||
apiserver_endpoint:
|
||||
description: Virtual ip-address configured on each master
|
||||
required: true
|
||||
|
||||
bpf_lb_algorithm:
|
||||
description: BPF lb algorithm
|
||||
default: maglev
|
||||
|
||||
bpf_lb_mode:
|
||||
description: BPF lb mode
|
||||
default: hybrid
|
||||
|
||||
calico_ebpf:
|
||||
description: Use eBPF dataplane instead of iptables
|
||||
type: bool
|
||||
default: false
|
||||
|
||||
calico_iface:
|
||||
description: The network interface used for when Calico is enabled
|
||||
default: ~
|
||||
|
||||
calico_tag:
|
||||
description: Calico version tag
|
||||
default: v3.27.2
|
||||
|
||||
cilium_bgp:
|
||||
description:
|
||||
- Enable cilium BGP control plane for LB services and pod cidrs.
|
||||
- Disables the use of MetalLB.
|
||||
type: bool
|
||||
default: false
|
||||
|
||||
cilium_exportPodCIDR:
|
||||
description: Export pod CIDR
|
||||
type: bool
|
||||
default: true
|
||||
|
||||
cilium_hubble:
|
||||
description: Enable Cilium Hubble
|
||||
type: bool
|
||||
default: true
|
||||
|
||||
cilium_iface:
|
||||
description: The network interface used for when Cilium is enabled
|
||||
default: ~
|
||||
|
||||
cilium_mode:
|
||||
description: Inner-node communication mode
|
||||
default: native
|
||||
choices:
|
||||
- native
|
||||
- routed
|
||||
|
||||
cluster_cidr:
|
||||
description: Inner-cluster IP range
|
||||
default: 10.52.0.0/16
|
||||
|
||||
enable_bpf_masquerade:
|
||||
description: Use IP masquerading
|
||||
type: bool
|
||||
default: true
|
||||
|
||||
group_name_master:
|
||||
description: Name of the master group
|
||||
default: master
|
||||
|
||||
kube_proxy_replacement:
|
||||
description: Replace the native kube-proxy with Cilium
|
||||
type: bool
|
||||
default: true
|
||||
|
||||
kube_vip_lb_ip_range:
|
||||
description: IP range for kube-vip load balancer
|
||||
default: ~
|
||||
|
||||
metal_lb_available_timeout:
|
||||
description: Wait for MetalLB resources
|
||||
default: 240s
|
||||
|
||||
metal_lb_controller_tag_version:
|
||||
description: Image tag for MetalLB
|
||||
default: v0.14.3
|
||||
|
||||
metal_lb_mode:
|
||||
description: Metallb mode
|
||||
default: layer2
|
||||
choices:
|
||||
- bgp
|
||||
- layer2
|
||||
@@ -48,7 +48,7 @@
|
||||
k3s kubectl wait {{ item.type }}/{{ item.name }}
|
||||
--namespace='tigera-operator'
|
||||
--for=condition=Available=True
|
||||
--timeout=7s
|
||||
--timeout=30s
|
||||
register: tigera_result
|
||||
changed_when: false
|
||||
until: tigera_result is succeeded
|
||||
@@ -87,7 +87,7 @@
|
||||
--namespace='{{ item.namespace }}'
|
||||
--for=condition=Available
|
||||
{% endif %}
|
||||
--timeout=7s
|
||||
--timeout=30s
|
||||
register: cr_result
|
||||
changed_when: false
|
||||
until: cr_result is succeeded
|
||||
|
||||
@@ -170,22 +170,22 @@
|
||||
{% if cilium_mode == "native" or (cilium_bgp and cilium_exportPodCIDR != 'false') %}
|
||||
--helm-set ipv4NativeRoutingCIDR={{ cluster_cidr }}
|
||||
{% endif %}
|
||||
--helm-set k8sServiceHost={{ apiserver_endpoint }}
|
||||
--helm-set k8sServicePort="6443"
|
||||
--helm-set routingMode={{ cilium_mode | default("native") }}
|
||||
--helm-set k8sServiceHost="127.0.0.1"
|
||||
--helm-set k8sServicePort="6444"
|
||||
--helm-set routingMode={{ cilium_mode }}
|
||||
--helm-set autoDirectNodeRoutes={{ "true" if cilium_mode == "native" else "false" }}
|
||||
--helm-set kubeProxyReplacement={{ kube_proxy_replacement | default("true") }}
|
||||
--helm-set bpf.masquerade={{ enable_bpf_masquerade | default("true") }}
|
||||
--helm-set kubeProxyReplacement={{ kube_proxy_replacement }}
|
||||
--helm-set bpf.masquerade={{ enable_bpf_masquerade }}
|
||||
--helm-set bgpControlPlane.enabled={{ cilium_bgp | default("false") }}
|
||||
--helm-set hubble.enabled={{ "true" if cilium_hubble else "false" }}
|
||||
--helm-set hubble.relay.enabled={{ "true" if cilium_hubble else "false" }}
|
||||
--helm-set hubble.ui.enabled={{ "true" if cilium_hubble else "false" }}
|
||||
{% if kube_proxy_replacement is not false %}
|
||||
--helm-set bpf.loadBalancer.algorithm={{ bpf_lb_algorithm | default("maglev") }}
|
||||
--helm-set bpf.loadBalancer.mode={{ bpf_lb_mode | default("hybrid") }}
|
||||
--helm-set bpf.loadBalancer.algorithm={{ bpf_lb_algorithm }}
|
||||
--helm-set bpf.loadBalancer.mode={{ bpf_lb_mode }}
|
||||
{% endif %}
|
||||
environment:
|
||||
KUBECONFIG: /home/{{ ansible_user }}/.kube/config
|
||||
KUBECONFIG: "{{ ansible_user_dir }}/.kube/config"
|
||||
register: cilium_install_result
|
||||
changed_when: cilium_install_result.rc == 0
|
||||
when: cilium_installed.rc != 0 or cilium_needs_update
|
||||
@@ -202,7 +202,7 @@
|
||||
--namespace=kube-system
|
||||
--for=condition=Available
|
||||
{% endif %}
|
||||
--timeout=7s
|
||||
--timeout=30s
|
||||
register: cr_result
|
||||
changed_when: false
|
||||
until: cr_result is succeeded
|
||||
|
||||
@@ -2,4 +2,5 @@
|
||||
- name: Reboot server
|
||||
become: true
|
||||
reboot:
|
||||
reboot_command: "{{ custom_reboot_command | default(omit) }}"
|
||||
listen: reboot server
|
||||
|
||||
@@ -1,4 +1,5 @@
|
||||
---
|
||||
- name: Reboot
|
||||
reboot:
|
||||
reboot_command: "{{ custom_reboot_command | default(omit) }}"
|
||||
listen: reboot
|
||||
|
||||
@@ -1,7 +1,27 @@
|
||||
---
|
||||
- name: Test for cmdline path
|
||||
stat:
|
||||
path: /boot/firmware/cmdline.txt
|
||||
register: boot_cmdline_path
|
||||
failed_when: false
|
||||
changed_when: false
|
||||
|
||||
- name: Set cmdline path based on Debian version and command result
|
||||
set_fact:
|
||||
cmdline_path: >-
|
||||
{{
|
||||
(
|
||||
boot_cmdline_path.stat.exists and
|
||||
ansible_facts.lsb.description | default('') is match('Debian.*(?!(bookworm|sid))')
|
||||
) | ternary(
|
||||
'/boot/firmware/cmdline.txt',
|
||||
'/boot/cmdline.txt'
|
||||
)
|
||||
}}
|
||||
|
||||
- name: Activating cgroup support
|
||||
lineinfile:
|
||||
path: /boot/cmdline.txt
|
||||
path: "{{ cmdline_path }}"
|
||||
regexp: '^((?!.*\bcgroup_enable=cpuset cgroup_memory=1 cgroup_enable=memory\b).*)$'
|
||||
line: '\1 cgroup_enable=cpuset cgroup_memory=1 cgroup_enable=memory'
|
||||
backrefs: true
|
||||
|
||||
2
roles/reset/defaults/main.yml
Normal file
2
roles/reset/defaults/main.yml
Normal file
@@ -0,0 +1,2 @@
|
||||
---
|
||||
systemd_dir: /etc/systemd/system
|
||||
9
site.yml
9
site.yml
@@ -1,4 +1,13 @@
|
||||
---
|
||||
- name: Pre tasks
|
||||
hosts: all
|
||||
pre_tasks:
|
||||
- name: Verify Ansible is version 2.11 or above. (If this fails you may need to update Ansible)
|
||||
assert:
|
||||
that: "ansible_version.full is version_compare('2.11', '>=')"
|
||||
msg: >
|
||||
"Ansible is out of date. See here for more info: https://docs.technotim.live/posts/ansible-automation/"
|
||||
|
||||
- name: Prepare Proxmox cluster
|
||||
hosts: proxmox
|
||||
gather_facts: true
|
||||
|
||||
Reference in New Issue
Block a user