forked from tim/k3s-ansible
Compare commits
57 Commits
k3s-1-27
...
self-hoste
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
2ae9ee82f0 | ||
|
|
5380f93b64 | ||
|
|
a414453ad4 | ||
|
|
23c5d9ec89 | ||
|
|
6b339e1985 | ||
|
|
a031807660 | ||
|
|
7dd305aabc | ||
|
|
500931e2fd | ||
|
|
cf357cf164 | ||
|
|
215e0d10ed | ||
|
|
c6ed680dc1 | ||
|
|
8343a6199e | ||
|
|
b524f97552 | ||
|
|
f741040e44 | ||
|
|
09bc628ba6 | ||
|
|
71ff6b86cd | ||
|
|
23729ddbbe | ||
|
|
e254c407f0 | ||
|
|
713b4694e1 | ||
|
|
952d513124 | ||
|
|
dd1e596332 | ||
|
|
6af47f96d0 | ||
|
|
664deec6c3 | ||
|
|
646459e7f5 | ||
|
|
64242d9729 | ||
|
|
f4864ddb64 | ||
|
|
6a83cde0c6 | ||
|
|
77ac928c0d | ||
|
|
8300a7aaac | ||
|
|
bdc6af5f46 | ||
|
|
dc8276157a | ||
|
|
37f0cb11d2 | ||
|
|
68e7c77b22 | ||
|
|
d82c4feac8 | ||
|
|
9217d8607b | ||
|
|
fbc15aa1a1 | ||
|
|
b55ec046ad | ||
|
|
b3cc178045 | ||
|
|
13be424187 | ||
|
|
d9cecd5364 | ||
|
|
afb96dbee2 | ||
|
|
30ffc69192 | ||
|
|
94e385c28e | ||
|
|
dbb2cda17a | ||
|
|
d24cdb97db | ||
|
|
5bebec930b | ||
|
|
ac52acdec1 | ||
|
|
105b2c2f1e | ||
|
|
d20f485fca | ||
|
|
f9bb9dabae | ||
|
|
6f15ef260e | ||
|
|
de1966fe02 | ||
|
|
fc823122d8 | ||
|
|
2f8d94bb5e | ||
|
|
9c3814ce72 | ||
|
|
0e60f4643b | ||
|
|
bb20514a6a |
2
.github/PULL_REQUEST_TEMPLATE.md
vendored
2
.github/PULL_REQUEST_TEMPLATE.md
vendored
@@ -11,5 +11,5 @@
|
|||||||
- [ ] Ran `site.yml` playbook
|
- [ ] Ran `site.yml` playbook
|
||||||
- [ ] Ran `reset.yml` playbook
|
- [ ] Ran `reset.yml` playbook
|
||||||
- [ ] Did not add any unnecessary changes
|
- [ ] Did not add any unnecessary changes
|
||||||
- [ ] Ran pre-commit install at least once before committing
|
|
||||||
- [ ] 🚀
|
- [ ] 🚀
|
||||||
|
- [ ] Ran pre-commit install at least once before committing
|
||||||
|
|||||||
7
.github/download-boxes.sh
vendored
7
.github/download-boxes.sh
vendored
@@ -5,10 +5,15 @@
|
|||||||
# already present on the system.
|
# already present on the system.
|
||||||
|
|
||||||
set -euo pipefail
|
set -euo pipefail
|
||||||
|
YQ_VERSION=v4.29.2
|
||||||
|
YQ_BINARY=yq_linux_amd64
|
||||||
GIT_ROOT=$(git rev-parse --show-toplevel)
|
GIT_ROOT=$(git rev-parse --show-toplevel)
|
||||||
PROVIDER=virtualbox
|
PROVIDER=virtualbox
|
||||||
|
|
||||||
|
# get yq used for filtering
|
||||||
|
sudo wget https://github.com/mikefarah/yq/releases/download/${YQ_VERSION}/${YQ_BINARY} -O /usr/bin/yq &&\
|
||||||
|
sudo chmod +x /usr/bin/yq
|
||||||
|
|
||||||
# Read all boxes for all platforms from the "molecule.yml" files
|
# Read all boxes for all platforms from the "molecule.yml" files
|
||||||
all_boxes=$(cat "${GIT_ROOT}"/molecule/*/molecule.yml |
|
all_boxes=$(cat "${GIT_ROOT}"/molecule/*/molecule.yml |
|
||||||
yq -r '.platforms[].box' | # Read the "box" property of each node under "platforms"
|
yq -r '.platforms[].box' | # Read the "box" property of each node under "platforms"
|
||||||
|
|||||||
15
.github/workflows/ci.yml
vendored
15
.github/workflows/ci.yml
vendored
@@ -1,15 +0,0 @@
|
|||||||
---
|
|
||||||
name: "CI"
|
|
||||||
on:
|
|
||||||
pull_request:
|
|
||||||
push:
|
|
||||||
branches:
|
|
||||||
- master
|
|
||||||
paths-ignore:
|
|
||||||
- '**/README.md'
|
|
||||||
jobs:
|
|
||||||
lint:
|
|
||||||
uses: ./.github/workflows/lint.yml
|
|
||||||
test:
|
|
||||||
uses: ./.github/workflows/test.yml
|
|
||||||
needs: [lint]
|
|
||||||
21
.github/workflows/lint.yml
vendored
21
.github/workflows/lint.yml
vendored
@@ -1,22 +1,27 @@
|
|||||||
---
|
---
|
||||||
name: Linting
|
name: Linting
|
||||||
on:
|
on:
|
||||||
workflow_call:
|
pull_request:
|
||||||
|
push:
|
||||||
|
branches:
|
||||||
|
- master
|
||||||
|
paths-ignore:
|
||||||
|
- '**/README.md'
|
||||||
jobs:
|
jobs:
|
||||||
pre-commit-ci:
|
pre-commit-ci:
|
||||||
name: Pre-Commit
|
name: Pre-Commit
|
||||||
runs-on: ubuntu-latest
|
runs-on: self-hosted
|
||||||
env:
|
env:
|
||||||
PYTHON_VERSION: "3.11"
|
PYTHON_VERSION: "3.10"
|
||||||
|
|
||||||
steps:
|
steps:
|
||||||
- name: Check out the codebase
|
- name: Check out the codebase
|
||||||
uses: actions/checkout@e2f20e631ae6d7dd3b768f56a5d2af784dd54791 # v3 2.5.0
|
uses: actions/checkout@2541b1294d2704b0964813337f33b291d3f8596b # 3.0.2
|
||||||
with:
|
with:
|
||||||
ref: ${{ github.event.pull_request.head.sha }}
|
ref: ${{ github.event.pull_request.head.sha }}
|
||||||
|
|
||||||
- name: Set up Python ${{ env.PYTHON_VERSION }}
|
- name: Set up Python ${{ env.PYTHON_VERSION }}
|
||||||
uses: actions/setup-python@75f3110429a8c05be0e1bf360334e4cced2b63fa # 2.3.3
|
uses: actions/setup-python@13ae5bb136fac2878aff31522b9efb785519f984 # 4.3.0
|
||||||
with:
|
with:
|
||||||
python-version: ${{ env.PYTHON_VERSION }}
|
python-version: ${{ env.PYTHON_VERSION }}
|
||||||
cache: 'pip' # caching pip dependencies
|
cache: 'pip' # caching pip dependencies
|
||||||
@@ -56,12 +61,12 @@ jobs:
|
|||||||
|
|
||||||
ensure-pinned-actions:
|
ensure-pinned-actions:
|
||||||
name: Ensure SHA Pinned Actions
|
name: Ensure SHA Pinned Actions
|
||||||
runs-on: ubuntu-latest
|
runs-on: self-hosted
|
||||||
steps:
|
steps:
|
||||||
- name: Checkout code
|
- name: Checkout code
|
||||||
uses: actions/checkout@e2f20e631ae6d7dd3b768f56a5d2af784dd54791 # v3 2.5.0
|
uses: actions/checkout@2541b1294d2704b0964813337f33b291d3f8596b # 3.0.2
|
||||||
- name: Ensure SHA pinned actions
|
- name: Ensure SHA pinned actions
|
||||||
uses: zgosalvez/github-actions-ensure-sha-pinned-actions@af2eb3226618e2494e3d9084f515ad6dcf16e229 # 2.0.1
|
uses: zgosalvez/github-actions-ensure-sha-pinned-actions@6ca5574367befbc9efdb2fa25978084159c5902d # 1.3.0
|
||||||
with:
|
with:
|
||||||
allowlist: |
|
allowlist: |
|
||||||
aws-actions/
|
aws-actions/
|
||||||
|
|||||||
44
.github/workflows/test.yml
vendored
44
.github/workflows/test.yml
vendored
@@ -1,11 +1,17 @@
|
|||||||
---
|
---
|
||||||
name: Test
|
name: Test
|
||||||
on:
|
on:
|
||||||
workflow_call:
|
pull_request:
|
||||||
|
push:
|
||||||
|
branches:
|
||||||
|
- master
|
||||||
|
paths-ignore:
|
||||||
|
- '**/README.md'
|
||||||
jobs:
|
jobs:
|
||||||
molecule:
|
molecule:
|
||||||
name: Molecule
|
name: Molecule
|
||||||
runs-on: macos-12
|
runs-on: self-hosted
|
||||||
|
|
||||||
strategy:
|
strategy:
|
||||||
matrix:
|
matrix:
|
||||||
scenario:
|
scenario:
|
||||||
@@ -14,14 +20,41 @@ jobs:
|
|||||||
- single_node
|
- single_node
|
||||||
fail-fast: false
|
fail-fast: false
|
||||||
env:
|
env:
|
||||||
PYTHON_VERSION: "3.11"
|
PYTHON_VERSION: "3.10"
|
||||||
|
VAGRANT_DEFAULT_PROVIDER: virtualbox
|
||||||
|
|
||||||
steps:
|
steps:
|
||||||
- name: Check out the codebase
|
- name: Check out the codebase
|
||||||
uses: actions/checkout@e2f20e631ae6d7dd3b768f56a5d2af784dd54791 # v3 2.5.0
|
uses: actions/checkout@2541b1294d2704b0964813337f33b291d3f8596b # 3.0.2
|
||||||
with:
|
with:
|
||||||
ref: ${{ github.event.pull_request.head.sha }}
|
ref: ${{ github.event.pull_request.head.sha }}
|
||||||
|
|
||||||
|
- name: Install Virtual Box from Oracle
|
||||||
|
run: |
|
||||||
|
echo "::group::Virtual Box"
|
||||||
|
wget -O- https://www.virtualbox.org/download/oracle_vbox_2016.asc | sudo gpg --dearmor --yes --output /usr/share/keyrings/oracle-virtualbox-2016.gpg
|
||||||
|
echo "deb [arch=amd64 signed-by=/usr/share/keyrings/oracle-virtualbox-2016.gpg] https://download.virtualbox.org/virtualbox/debian $(lsb_release -cs) contrib" | sudo tee -a /etc/apt/sources.list.d/virtualbox.list
|
||||||
|
sudo apt update && sudo apt install -y linux-headers-generic linux-headers-5.15.0-52-generic build-essential dkms virtualbox-dkms virtualbox-6.1
|
||||||
|
echo "::endgroup::"
|
||||||
|
echo "::group::Virtual Box Test"
|
||||||
|
vboxmanage --version
|
||||||
|
sudo /sbin/vboxconfig
|
||||||
|
sudo modprobe vboxdrv
|
||||||
|
vboxmanage --version
|
||||||
|
echo "::endgroup::"
|
||||||
|
|
||||||
|
- name: Install Vagrant
|
||||||
|
run: |
|
||||||
|
echo "::group::Install Vagrant"
|
||||||
|
wget -O- https://apt.releases.hashicorp.com/gpg | gpg --dearmor | sudo tee /usr/share/keyrings/hashicorp-archive-keyring.gpg
|
||||||
|
echo "deb [signed-by=/usr/share/keyrings/hashicorp-archive-keyring.gpg] https://apt.releases.hashicorp.com $(lsb_release -cs) main" | sudo tee /etc/apt/sources.list.d/hashicorp.list
|
||||||
|
sudo apt update && sudo apt install -y vagrant
|
||||||
|
vagrant version
|
||||||
|
vagrant plugin list
|
||||||
|
vagrant plugin install vagrant-vbguest
|
||||||
|
vagrant plugin list
|
||||||
|
echo "::endgroup::"
|
||||||
|
|
||||||
- name: Configure VirtualBox
|
- name: Configure VirtualBox
|
||||||
run: |-
|
run: |-
|
||||||
sudo mkdir -p /etc/vbox
|
sudo mkdir -p /etc/vbox
|
||||||
@@ -54,7 +87,7 @@ jobs:
|
|||||||
run: ./.github/download-boxes.sh
|
run: ./.github/download-boxes.sh
|
||||||
|
|
||||||
- name: Set up Python ${{ env.PYTHON_VERSION }}
|
- name: Set up Python ${{ env.PYTHON_VERSION }}
|
||||||
uses: actions/setup-python@75f3110429a8c05be0e1bf360334e4cced2b63fa # 2.3.3
|
uses: actions/setup-python@13ae5bb136fac2878aff31522b9efb785519f984 # 4.3.0
|
||||||
with:
|
with:
|
||||||
python-version: ${{ env.PYTHON_VERSION }}
|
python-version: ${{ env.PYTHON_VERSION }}
|
||||||
cache: 'pip' # caching pip dependencies
|
cache: 'pip' # caching pip dependencies
|
||||||
@@ -71,7 +104,6 @@ jobs:
|
|||||||
|
|
||||||
- name: Test with molecule
|
- name: Test with molecule
|
||||||
run: molecule test --scenario-name ${{ matrix.scenario }}
|
run: molecule test --scenario-name ${{ matrix.scenario }}
|
||||||
timeout-minutes: 90
|
|
||||||
env:
|
env:
|
||||||
ANSIBLE_K3S_LOG_DIR: ${{ runner.temp }}/logs/k3s-ansible/${{ matrix.scenario }}
|
ANSIBLE_K3S_LOG_DIR: ${{ runner.temp }}/logs/k3s-ansible/${{ matrix.scenario }}
|
||||||
ANSIBLE_SSH_RETRIES: 4
|
ANSIBLE_SSH_RETRIES: 4
|
||||||
|
|||||||
2
.gitignore
vendored
2
.gitignore
vendored
@@ -1,3 +1 @@
|
|||||||
.env/
|
.env/
|
||||||
*.log
|
|
||||||
ansible.cfg
|
|
||||||
|
|||||||
@@ -1,35 +1,21 @@
|
|||||||
---
|
---
|
||||||
repos:
|
repos:
|
||||||
- repo: https://github.com/pre-commit/pre-commit-hooks
|
- repo: https://github.com/pre-commit/pre-commit-hooks
|
||||||
rev: f71fa2c1f9cf5cb705f73dffe4b21f7c61470ba9 # frozen: v4.4.0
|
rev: v4.3.0
|
||||||
hooks:
|
hooks:
|
||||||
- id: requirements-txt-fixer
|
- id: requirements-txt-fixer
|
||||||
- id: sort-simple-yaml
|
- id: sort-simple-yaml
|
||||||
- id: detect-private-key
|
- id: detect-private-key
|
||||||
- id: check-merge-conflict
|
|
||||||
- id: end-of-file-fixer
|
|
||||||
- id: mixed-line-ending
|
|
||||||
- id: trailing-whitespace
|
|
||||||
args: [--markdown-linebreak-ext=md]
|
|
||||||
- repo: https://github.com/adrienverge/yamllint.git
|
- repo: https://github.com/adrienverge/yamllint.git
|
||||||
rev: b05e028c5881819161d11cb543fd96a30c06cceb # frozen: v1.32.0
|
rev: v1.28.0
|
||||||
hooks:
|
hooks:
|
||||||
- id: yamllint
|
- id: yamllint
|
||||||
args: [-c=.yamllint]
|
args: [-c=.yamllint]
|
||||||
- repo: https://github.com/ansible-community/ansible-lint.git
|
- repo: https://github.com/ansible-community/ansible-lint.git
|
||||||
rev: 3293b64b939c0de16ef8cb81dd49255e475bf89a # frozen: v6.17.2
|
rev: v6.8.2
|
||||||
hooks:
|
hooks:
|
||||||
- id: ansible-lint
|
- id: ansible-lint
|
||||||
- repo: https://github.com/shellcheck-py/shellcheck-py
|
- repo: https://github.com/shellcheck-py/shellcheck-py
|
||||||
rev: 375289a39f5708101b1f916eb729e8d6da96993f # frozen: v0.9.0.5
|
rev: v0.8.0.4
|
||||||
hooks:
|
hooks:
|
||||||
- id: shellcheck
|
- id: shellcheck
|
||||||
- repo: https://github.com/Lucas-C/pre-commit-hooks
|
|
||||||
rev: 12885e376b93dc4536ad68d156065601e4433665 # frozen: v1.5.1
|
|
||||||
hooks:
|
|
||||||
- id: remove-crlf
|
|
||||||
- id: remove-tabs
|
|
||||||
- repo: https://github.com/sirosen/texthooks
|
|
||||||
rev: c4ffd3e31669dd4fa4d31a23436cc13839730084 # frozen: 0.5.0
|
|
||||||
hooks:
|
|
||||||
- id: fix-smartquotes
|
|
||||||
|
|||||||
20
README.md
20
README.md
@@ -8,9 +8,9 @@ This is based on the work from [this fork](https://github.com/212850a/k3s-ansibl
|
|||||||
|
|
||||||
If you want more context on how this works, see:
|
If you want more context on how this works, see:
|
||||||
|
|
||||||
📄 [Documentation](https://technotim.live/posts/k3s-etcd-ansible/) (including example commands)
|
📄 [Documentation](https://docs.technotim.live/posts/k3s-etcd-ansible/) (including example commands)
|
||||||
|
|
||||||
📺 [Watch the Video](https://www.youtube.com/watch?v=CbkEWcUZ7zM)
|
📺 [Video](https://www.youtube.com/watch?v=CbkEWcUZ7zM)
|
||||||
|
|
||||||
## 📖 k3s Ansible Playbook
|
## 📖 k3s Ansible Playbook
|
||||||
|
|
||||||
@@ -28,14 +28,14 @@ on processor architecture:
|
|||||||
|
|
||||||
## ✅ System requirements
|
## ✅ System requirements
|
||||||
|
|
||||||
- Control Node (the machine you are running `ansible` commands) must have Ansible 2.11+ If you need a quick primer on Ansible [you can check out my docs and setting up Ansible](https://technotim.live/posts/ansible-automation/).
|
- Deployment environment must have Ansible 2.4.0+. If you need a quick primer on Ansible [you can check out my docs and setting up Ansible](https://docs.technotim.live/posts/ansible-automation/).
|
||||||
|
|
||||||
- You will also need to install collections that this playbook uses by running `ansible-galaxy collection install -r ./collections/requirements.yml` (important❗)
|
|
||||||
|
|
||||||
- [`netaddr` package](https://pypi.org/project/netaddr/) must be available to Ansible. If you have installed Ansible via apt, this is already taken care of. If you have installed Ansible via `pip`, make sure to install `netaddr` into the respective virtual environment.
|
- [`netaddr` package](https://pypi.org/project/netaddr/) must be available to Ansible. If you have installed Ansible via apt, this is already taken care of. If you have installed Ansible via `pip`, make sure to install `netaddr` into the respective virtual environment.
|
||||||
|
|
||||||
- `server` and `agent` nodes should have passwordless SSH access, if not you can supply arguments to provide credentials `--ask-pass --ask-become-pass` to each command.
|
- `server` and `agent` nodes should have passwordless SSH access, if not you can supply arguments to provide credentials `--ask-pass --ask-become-pass` to each command.
|
||||||
|
|
||||||
|
- You will also need to install collections that this playbook uses by running `ansible-galaxy collection install -r ./collections/requirements.yml`
|
||||||
|
|
||||||
## 🚀 Getting Started
|
## 🚀 Getting Started
|
||||||
|
|
||||||
### 🍴 Preparation
|
### 🍴 Preparation
|
||||||
@@ -67,8 +67,6 @@ node
|
|||||||
|
|
||||||
If multiple hosts are in the master group, the playbook will automatically set up k3s in [HA mode with etcd](https://rancher.com/docs/k3s/latest/en/installation/ha-embedded/).
|
If multiple hosts are in the master group, the playbook will automatically set up k3s in [HA mode with etcd](https://rancher.com/docs/k3s/latest/en/installation/ha-embedded/).
|
||||||
|
|
||||||
Finally, copy `ansible.example.cfg` to `ansible.cfg` and adapt the inventory path to match the files that you just created.
|
|
||||||
|
|
||||||
This requires at least k3s version `1.19.1` however the version is configurable by using the `k3s_version` variable.
|
This requires at least k3s version `1.19.1` however the version is configurable by using the `k3s_version` variable.
|
||||||
|
|
||||||
If needed, you can also edit `inventory/my-cluster/group_vars/all.yml` to match your environment.
|
If needed, you can also edit `inventory/my-cluster/group_vars/all.yml` to match your environment.
|
||||||
@@ -101,7 +99,7 @@ scp debian@master_ip:~/.kube/config ~/.kube/config
|
|||||||
|
|
||||||
### 🔨 Testing your cluster
|
### 🔨 Testing your cluster
|
||||||
|
|
||||||
See the commands [here](https://technotim.live/posts/k3s-etcd-ansible/#testing-your-cluster).
|
See the commands [here](https://docs.technotim.live/posts/k3s-etcd-ansible/#testing-your-cluster).
|
||||||
|
|
||||||
### Troubleshooting
|
### Troubleshooting
|
||||||
|
|
||||||
@@ -114,13 +112,9 @@ It is run automatically in CI, but you can also run the tests locally.
|
|||||||
This might be helpful for quick feedback in a few cases.
|
This might be helpful for quick feedback in a few cases.
|
||||||
You can find more information about it [here](molecule/README.md).
|
You can find more information about it [here](molecule/README.md).
|
||||||
|
|
||||||
### Pre-commit Hooks
|
|
||||||
|
|
||||||
This repo uses `pre-commit` and `pre-commit-hooks` to lint and fix common style and syntax errors. Be sure to install python packages and then run `pre-commit install`. For more information, see [pre-commit](https://pre-commit.com/)
|
|
||||||
|
|
||||||
## Thanks 🤝
|
## Thanks 🤝
|
||||||
|
|
||||||
This repo is really standing on the shoulders of giants. Thank you to all those who have contributed and thanks to these repos for code and ideas:
|
This repo is really standing on the shoulders of giants. Thank you to all those who have contributed and tanks to these repos for code and ideas:
|
||||||
|
|
||||||
- [k3s-io/k3s-ansible](https://github.com/k3s-io/k3s-ansible)
|
- [k3s-io/k3s-ansible](https://github.com/k3s-io/k3s-ansible)
|
||||||
- [geerlingguy/turing-pi-cluster](https://github.com/geerlingguy/turing-pi-cluster)
|
- [geerlingguy/turing-pi-cluster](https://github.com/geerlingguy/turing-pi-cluster)
|
||||||
|
|||||||
12
ansible.cfg
Normal file
12
ansible.cfg
Normal file
@@ -0,0 +1,12 @@
|
|||||||
|
[defaults]
|
||||||
|
nocows = True
|
||||||
|
roles_path = ./roles
|
||||||
|
inventory = ./hosts.ini
|
||||||
|
|
||||||
|
remote_tmp = $HOME/.ansible/tmp
|
||||||
|
local_tmp = $HOME/.ansible/tmp
|
||||||
|
pipelining = True
|
||||||
|
become = True
|
||||||
|
host_key_checking = False
|
||||||
|
deprecation_warnings = False
|
||||||
|
callback_whitelist = profile_tasks
|
||||||
@@ -1,2 +0,0 @@
|
|||||||
[defaults]
|
|
||||||
inventory = inventory/my-cluster/hosts.ini ; Adapt this to the path to your inventory file
|
|
||||||
@@ -1,3 +1,3 @@
|
|||||||
#!/bin/bash
|
#!/bin/bash
|
||||||
|
|
||||||
ansible-playbook site.yml
|
ansible-playbook site.yml -i inventory/my-cluster/hosts.ini
|
||||||
@@ -1,5 +1,5 @@
|
|||||||
---
|
---
|
||||||
k3s_version: v1.27.4+k3s1
|
k3s_version: v1.24.6+k3s1
|
||||||
# this is the user that has ssh access to these machines
|
# this is the user that has ssh access to these machines
|
||||||
ansible_user: ansibleuser
|
ansible_user: ansibleuser
|
||||||
systemd_dir: /etc/systemd/system
|
systemd_dir: /etc/systemd/system
|
||||||
@@ -41,43 +41,11 @@ extra_agent_args: >-
|
|||||||
{{ extra_args }}
|
{{ extra_args }}
|
||||||
|
|
||||||
# image tag for kube-vip
|
# image tag for kube-vip
|
||||||
kube_vip_tag_version: "v0.5.12"
|
kube_vip_tag_version: "v0.5.5"
|
||||||
|
|
||||||
# metallb type frr or native
|
|
||||||
metal_lb_type: "native"
|
|
||||||
|
|
||||||
# metallb mode layer2 or bgp
|
|
||||||
metal_lb_mode: "layer2"
|
|
||||||
|
|
||||||
# bgp options
|
|
||||||
# metal_lb_bgp_my_asn: "64513"
|
|
||||||
# metal_lb_bgp_peer_asn: "64512"
|
|
||||||
# metal_lb_bgp_peer_address: "192.168.30.1"
|
|
||||||
|
|
||||||
# image tag for metal lb
|
# image tag for metal lb
|
||||||
metal_lb_speaker_tag_version: "v0.13.9"
|
metal_lb_speaker_tag_version: "v0.13.6"
|
||||||
metal_lb_controller_tag_version: "v0.13.9"
|
metal_lb_controller_tag_version: "v0.13.6"
|
||||||
|
|
||||||
# metallb ip range for load balancer
|
# metallb ip range for load balancer
|
||||||
metal_lb_ip_range: "192.168.30.80-192.168.30.90"
|
metal_lb_ip_range: "192.168.30.80-192.168.30.90"
|
||||||
|
|
||||||
# Only enable if your nodes are proxmox LXC nodes, make sure to configure your proxmox nodes
|
|
||||||
# in your hosts.ini file.
|
|
||||||
# Please read https://gist.github.com/triangletodd/02f595cd4c0dc9aac5f7763ca2264185 before using this.
|
|
||||||
# Most notably, your containers must be privileged, and must not have nesting set to true.
|
|
||||||
# Please note this script disables most of the security of lxc containers, with the trade off being that lxc
|
|
||||||
# containers are significantly more resource efficent compared to full VMs.
|
|
||||||
# Mixing and matching VMs and lxc containers is not supported, ymmv if you want to do this.
|
|
||||||
# I would only really recommend using this if you have partiularly low powered proxmox nodes where the overhead of
|
|
||||||
# VMs would use a significant portion of your available resources.
|
|
||||||
proxmox_lxc_configure: false
|
|
||||||
# the user that you would use to ssh into the host, for example if you run ssh some-user@my-proxmox-host,
|
|
||||||
# set this value to some-user
|
|
||||||
proxmox_lxc_ssh_user: root
|
|
||||||
# the unique proxmox ids for all of the containers in the cluster, both worker and master nodes
|
|
||||||
proxmox_lxc_ct_ids:
|
|
||||||
- 200
|
|
||||||
- 201
|
|
||||||
- 202
|
|
||||||
- 203
|
|
||||||
- 204
|
|
||||||
|
|||||||
@@ -1,2 +0,0 @@
|
|||||||
---
|
|
||||||
ansible_user: '{{ proxmox_lxc_ssh_user }}'
|
|
||||||
@@ -7,11 +7,6 @@
|
|||||||
192.168.30.41
|
192.168.30.41
|
||||||
192.168.30.42
|
192.168.30.42
|
||||||
|
|
||||||
# only required if proxmox_lxc_configure: true
|
|
||||||
# must contain all proxmox instances that have a master or worker node
|
|
||||||
# [proxmox]
|
|
||||||
# 192.168.30.43
|
|
||||||
|
|
||||||
[k3s_cluster:children]
|
[k3s_cluster:children]
|
||||||
master
|
master
|
||||||
node
|
node
|
||||||
|
|||||||
@@ -3,73 +3,56 @@ dependency:
|
|||||||
name: galaxy
|
name: galaxy
|
||||||
driver:
|
driver:
|
||||||
name: vagrant
|
name: vagrant
|
||||||
platforms:
|
.platform_presets:
|
||||||
|
- &control
|
||||||
- name: control1
|
|
||||||
box: generic/ubuntu2204
|
|
||||||
memory: 2048
|
memory: 2048
|
||||||
cpus: 2
|
cpus: 2
|
||||||
groups:
|
groups:
|
||||||
- k3s_cluster
|
- k3s_cluster
|
||||||
- master
|
- master
|
||||||
|
- &node
|
||||||
|
memory: 2048
|
||||||
|
cpus: 2
|
||||||
|
groups:
|
||||||
|
- k3s_cluster
|
||||||
|
- node
|
||||||
|
- &debian
|
||||||
|
box: generic/debian11
|
||||||
|
- &rocky
|
||||||
|
box: generic/rocky9
|
||||||
|
- &ubuntu
|
||||||
|
box: generic/ubuntu2204
|
||||||
|
config_options:
|
||||||
|
# We currently can not use public-key based authentication on Ubuntu 22.04,
|
||||||
|
# see: https://github.com/chef/bento/issues/1405
|
||||||
|
ssh.username: "vagrant"
|
||||||
|
ssh.password: "vagrant"
|
||||||
|
platforms:
|
||||||
|
- <<: [*control, *ubuntu]
|
||||||
|
name: control1
|
||||||
interfaces:
|
interfaces:
|
||||||
- network_name: private_network
|
- network_name: private_network
|
||||||
ip: 192.168.30.38
|
ip: 192.168.30.38
|
||||||
config_options:
|
- <<: [*control, *debian]
|
||||||
# We currently can not use public-key based authentication on Ubuntu 22.04,
|
name: control2
|
||||||
# see: https://github.com/chef/bento/issues/1405
|
|
||||||
ssh.username: "vagrant"
|
|
||||||
ssh.password: "vagrant"
|
|
||||||
|
|
||||||
- name: control2
|
|
||||||
box: generic/debian11
|
|
||||||
memory: 2048
|
|
||||||
cpus: 2
|
|
||||||
groups:
|
|
||||||
- k3s_cluster
|
|
||||||
- master
|
|
||||||
interfaces:
|
interfaces:
|
||||||
- network_name: private_network
|
- network_name: private_network
|
||||||
ip: 192.168.30.39
|
ip: 192.168.30.39
|
||||||
|
- <<: [*control, *rocky]
|
||||||
- name: control3
|
name: control3
|
||||||
box: generic/rocky9
|
|
||||||
memory: 2048
|
|
||||||
cpus: 2
|
|
||||||
groups:
|
|
||||||
- k3s_cluster
|
|
||||||
- master
|
|
||||||
interfaces:
|
interfaces:
|
||||||
- network_name: private_network
|
- network_name: private_network
|
||||||
ip: 192.168.30.40
|
ip: 192.168.30.40
|
||||||
|
- <<: [*node, *ubuntu]
|
||||||
- name: node1
|
name: node1
|
||||||
box: generic/ubuntu2204
|
|
||||||
memory: 2048
|
|
||||||
cpus: 2
|
|
||||||
groups:
|
|
||||||
- k3s_cluster
|
|
||||||
- node
|
|
||||||
interfaces:
|
interfaces:
|
||||||
- network_name: private_network
|
- network_name: private_network
|
||||||
ip: 192.168.30.41
|
ip: 192.168.30.41
|
||||||
config_options:
|
- <<: [*node, *rocky]
|
||||||
# We currently can not use public-key based authentication on Ubuntu 22.04,
|
name: node2
|
||||||
# see: https://github.com/chef/bento/issues/1405
|
|
||||||
ssh.username: "vagrant"
|
|
||||||
ssh.password: "vagrant"
|
|
||||||
|
|
||||||
- name: node2
|
|
||||||
box: generic/rocky9
|
|
||||||
memory: 2048
|
|
||||||
cpus: 2
|
|
||||||
groups:
|
|
||||||
- k3s_cluster
|
|
||||||
- node
|
|
||||||
interfaces:
|
interfaces:
|
||||||
- network_name: private_network
|
- network_name: private_network
|
||||||
ip: 192.168.30.42
|
ip: 192.168.30.42
|
||||||
|
|
||||||
provisioner:
|
provisioner:
|
||||||
name: ansible
|
name: ansible
|
||||||
playbooks:
|
playbooks:
|
||||||
|
|||||||
@@ -4,8 +4,7 @@
|
|||||||
tasks:
|
tasks:
|
||||||
- name: Override host variables
|
- name: Override host variables
|
||||||
ansible.builtin.set_fact:
|
ansible.builtin.set_fact:
|
||||||
# See:
|
# See: https://github.com/flannel-io/flannel/blob/67d603aaf45ef80f5dd39f43714fc5e6f8a637eb/Documentation/troubleshooting.md#Vagrant # noqa yaml[line-length]
|
||||||
# https://github.com/flannel-io/flannel/blob/67d603aaf45ef80f5dd39f43714fc5e6f8a637eb/Documentation/troubleshooting.md#Vagrant
|
|
||||||
flannel_iface: eth1
|
flannel_iface: eth1
|
||||||
|
|
||||||
# The test VMs might be a bit slow, so we give them more time to join the cluster:
|
# The test VMs might be a bit slow, so we give them more time to join the cluster:
|
||||||
|
|||||||
@@ -1,3 +0,0 @@
|
|||||||
---
|
|
||||||
node_ipv4: 192.168.123.12
|
|
||||||
node_ipv6: fdad:bad:ba55::de:12
|
|
||||||
@@ -3,54 +3,37 @@ dependency:
|
|||||||
name: galaxy
|
name: galaxy
|
||||||
driver:
|
driver:
|
||||||
name: vagrant
|
name: vagrant
|
||||||
platforms:
|
.platform_presets:
|
||||||
- name: control1
|
- &control
|
||||||
box: generic/ubuntu2204
|
|
||||||
memory: 2048
|
memory: 2048
|
||||||
cpus: 2
|
cpus: 2
|
||||||
groups:
|
groups:
|
||||||
- k3s_cluster
|
- k3s_cluster
|
||||||
- master
|
- master
|
||||||
interfaces:
|
- &node
|
||||||
- network_name: private_network
|
|
||||||
ip: fdad:bad:ba55::de:11
|
|
||||||
config_options:
|
|
||||||
# We currently can not use public-key based authentication on Ubuntu 22.04,
|
|
||||||
# see: https://github.com/chef/bento/issues/1405
|
|
||||||
ssh.username: "vagrant"
|
|
||||||
ssh.password: "vagrant"
|
|
||||||
|
|
||||||
- name: control2
|
|
||||||
box: generic/ubuntu2204
|
|
||||||
memory: 2048
|
|
||||||
cpus: 2
|
|
||||||
groups:
|
|
||||||
- k3s_cluster
|
|
||||||
- master
|
|
||||||
interfaces:
|
|
||||||
- network_name: private_network
|
|
||||||
ip: fdad:bad:ba55::de:12
|
|
||||||
config_options:
|
|
||||||
# We currently can not use public-key based authentication on Ubuntu 22.04,
|
|
||||||
# see: https://github.com/chef/bento/issues/1405
|
|
||||||
ssh.username: "vagrant"
|
|
||||||
ssh.password: "vagrant"
|
|
||||||
|
|
||||||
- name: node1
|
|
||||||
box: generic/ubuntu2204
|
|
||||||
memory: 2048
|
memory: 2048
|
||||||
cpus: 2
|
cpus: 2
|
||||||
groups:
|
groups:
|
||||||
- k3s_cluster
|
- k3s_cluster
|
||||||
- node
|
- node
|
||||||
interfaces:
|
- &ubuntu
|
||||||
- network_name: private_network
|
box: generic/ubuntu2204
|
||||||
ip: fdad:bad:ba55::de:21
|
|
||||||
config_options:
|
config_options:
|
||||||
# We currently can not use public-key based authentication on Ubuntu 22.04,
|
# We currently can not use public-key based authentication on Ubuntu 22.04,
|
||||||
# see: https://github.com/chef/bento/issues/1405
|
# see: https://github.com/chef/bento/issues/1405
|
||||||
ssh.username: "vagrant"
|
ssh.username: "vagrant"
|
||||||
ssh.password: "vagrant"
|
ssh.password: "vagrant"
|
||||||
|
platforms:
|
||||||
|
- <<: [*control, *ubuntu]
|
||||||
|
name: control1
|
||||||
|
interfaces:
|
||||||
|
- network_name: private_network
|
||||||
|
ip: fdad:bad:ba55::de:11
|
||||||
|
- <<: [*node, *ubuntu]
|
||||||
|
name: node1
|
||||||
|
interfaces:
|
||||||
|
- network_name: private_network
|
||||||
|
ip: fdad:bad:ba55::de:21
|
||||||
provisioner:
|
provisioner:
|
||||||
name: ansible
|
name: ansible
|
||||||
playbooks:
|
playbooks:
|
||||||
|
|||||||
@@ -4,15 +4,9 @@
|
|||||||
tasks:
|
tasks:
|
||||||
- name: Override host variables (1/2)
|
- name: Override host variables (1/2)
|
||||||
ansible.builtin.set_fact:
|
ansible.builtin.set_fact:
|
||||||
# See:
|
# See: https://github.com/flannel-io/flannel/blob/67d603aaf45ef80f5dd39f43714fc5e6f8a637eb/Documentation/troubleshooting.md#Vagrant # noqa yaml[line-length]
|
||||||
# https://github.com/flannel-io/flannel/blob/67d603aaf45ef80f5dd39f43714fc5e6f8a637eb/Documentation/troubleshooting.md#Vagrant
|
|
||||||
flannel_iface: eth1
|
flannel_iface: eth1
|
||||||
|
|
||||||
# In this scenario, we have multiple interfaces that the VIP could be
|
|
||||||
# broadcasted on. Since we have assigned a dedicated private network
|
|
||||||
# here, let's make sure that it is used.
|
|
||||||
kube_vip_iface: eth1
|
|
||||||
|
|
||||||
# The test VMs might be a bit slow, so we give them more time to join the cluster:
|
# The test VMs might be a bit slow, so we give them more time to join the cluster:
|
||||||
retry_count: 45
|
retry_count: 45
|
||||||
|
|
||||||
|
|||||||
@@ -2,4 +2,4 @@
|
|||||||
- name: Verify
|
- name: Verify
|
||||||
hosts: all
|
hosts: all
|
||||||
roles:
|
roles:
|
||||||
- verify_from_outside
|
- verify/from_outside
|
||||||
|
|||||||
@@ -6,4 +6,4 @@ outside_host: localhost
|
|||||||
testing_namespace: molecule-verify-from-outside
|
testing_namespace: molecule-verify-from-outside
|
||||||
|
|
||||||
# The directory in which the example manifests reside
|
# The directory in which the example manifests reside
|
||||||
example_manifests_path: ../../../example
|
example_manifests_path: ../../../../example
|
||||||
@@ -34,14 +34,14 @@
|
|||||||
|
|
||||||
- name: Assert that the nginx welcome page is available
|
- name: Assert that the nginx welcome page is available
|
||||||
ansible.builtin.uri:
|
ansible.builtin.uri:
|
||||||
url: http://{{ ip | ansible.utils.ipwrap }}:{{ port_ }}/
|
url: http://{{ ip | ansible.utils.ipwrap }}:{{ port }}/
|
||||||
return_content: yes
|
return_content: yes
|
||||||
register: result
|
register: result
|
||||||
failed_when: "'Welcome to nginx!' not in result.content"
|
failed_when: "'Welcome to nginx!' not in result.content"
|
||||||
vars:
|
vars:
|
||||||
ip: >-
|
ip: >-
|
||||||
{{ nginx_services.resources[0].status.loadBalancer.ingress[0].ip }}
|
{{ nginx_services.resources[0].status.loadBalancer.ingress[0].ip }}
|
||||||
port_: >-
|
port: >-
|
||||||
{{ nginx_services.resources[0].spec.ports[0].port }}
|
{{ nginx_services.resources[0].spec.ports[0].port }}
|
||||||
# Deactivated linter rules:
|
# Deactivated linter rules:
|
||||||
# - jinja[invalid]: As of version 6.6.0, ansible-lint complains that the input to ipwrap
|
# - jinja[invalid]: As of version 6.6.0, ansible-lint complains that the input to ipwrap
|
||||||
@@ -4,8 +4,7 @@
|
|||||||
tasks:
|
tasks:
|
||||||
- name: Override host variables
|
- name: Override host variables
|
||||||
ansible.builtin.set_fact:
|
ansible.builtin.set_fact:
|
||||||
# See:
|
# See: https://github.com/flannel-io/flannel/blob/67d603aaf45ef80f5dd39f43714fc5e6f8a637eb/Documentation/troubleshooting.md#Vagrant # noqa yaml[line-length]
|
||||||
# https://github.com/flannel-io/flannel/blob/67d603aaf45ef80f5dd39f43714fc5e6f8a637eb/Documentation/troubleshooting.md#Vagrant
|
|
||||||
flannel_iface: eth1
|
flannel_iface: eth1
|
||||||
|
|
||||||
# The test VMs might be a bit slow, so we give them more time to join the cluster:
|
# The test VMs might be a bit slow, so we give them more time to join the cluster:
|
||||||
|
|||||||
@@ -1,9 +0,0 @@
|
|||||||
---
|
|
||||||
- name: Reboot k3s_cluster
|
|
||||||
hosts: k3s_cluster
|
|
||||||
gather_facts: yes
|
|
||||||
tasks:
|
|
||||||
- name: Reboot the nodes (and Wait upto 5 mins max)
|
|
||||||
become: true
|
|
||||||
reboot:
|
|
||||||
reboot_timeout: 300
|
|
||||||
@@ -1,10 +0,0 @@
|
|||||||
ansible-core>=2.13.5
|
|
||||||
jmespath>=1.0.1
|
|
||||||
jsonpatch>=1.32
|
|
||||||
kubernetes>=25.3.0
|
|
||||||
molecule-vagrant>=1.0.0
|
|
||||||
molecule>=4.0.3
|
|
||||||
netaddr>=0.8.0
|
|
||||||
pre-commit>=2.20.0
|
|
||||||
pre-commit-hooks>=1.3.1
|
|
||||||
pyyaml>=6.0
|
|
||||||
168
requirements.txt
168
requirements.txt
@@ -1,178 +1,72 @@
|
|||||||
#
|
ansible-compat==2.2.1
|
||||||
# This file is autogenerated by pip-compile with Python 3.11
|
ansible-core==2.13.5
|
||||||
# by the following command:
|
ansible-lint==6.8.4
|
||||||
#
|
|
||||||
# pip-compile requirements.in
|
|
||||||
#
|
|
||||||
ansible-compat==3.0.1
|
|
||||||
# via molecule
|
|
||||||
ansible-core==2.15.2
|
|
||||||
# via
|
|
||||||
# -r requirements.in
|
|
||||||
# ansible-compat
|
|
||||||
arrow==1.2.3
|
arrow==1.2.3
|
||||||
# via jinja2-time
|
|
||||||
attrs==22.1.0
|
attrs==22.1.0
|
||||||
# via jsonschema
|
|
||||||
binaryornot==0.4.4
|
binaryornot==0.4.4
|
||||||
# via cookiecutter
|
black==22.10.0
|
||||||
|
bracex==2.3.post1
|
||||||
cachetools==5.2.0
|
cachetools==5.2.0
|
||||||
# via google-auth
|
Cerberus==1.3.2
|
||||||
certifi==2022.9.24
|
certifi==2022.9.24
|
||||||
# via
|
|
||||||
# kubernetes
|
|
||||||
# requests
|
|
||||||
cffi==1.15.1
|
cffi==1.15.1
|
||||||
# via cryptography
|
|
||||||
cfgv==3.3.1
|
|
||||||
# via pre-commit
|
|
||||||
chardet==5.0.0
|
chardet==5.0.0
|
||||||
# via binaryornot
|
|
||||||
charset-normalizer==2.1.1
|
charset-normalizer==2.1.1
|
||||||
# via requests
|
|
||||||
click==8.1.3
|
click==8.1.3
|
||||||
# via
|
|
||||||
# click-help-colors
|
|
||||||
# cookiecutter
|
|
||||||
# molecule
|
|
||||||
click-help-colors==0.9.1
|
click-help-colors==0.9.1
|
||||||
# via molecule
|
|
||||||
commonmark==0.9.1
|
commonmark==0.9.1
|
||||||
# via rich
|
|
||||||
cookiecutter==2.1.1
|
cookiecutter==2.1.1
|
||||||
# via molecule
|
cryptography==38.0.1
|
||||||
cryptography==38.0.3
|
|
||||||
# via ansible-core
|
|
||||||
distlib==0.3.6
|
|
||||||
# via virtualenv
|
|
||||||
distro==1.8.0
|
distro==1.8.0
|
||||||
# via selinux
|
|
||||||
enrich==1.2.7
|
enrich==1.2.7
|
||||||
# via molecule
|
|
||||||
filelock==3.8.0
|
filelock==3.8.0
|
||||||
# via virtualenv
|
google-auth==2.13.0
|
||||||
google-auth==2.14.0
|
|
||||||
# via kubernetes
|
|
||||||
identify==2.5.8
|
|
||||||
# via pre-commit
|
|
||||||
idna==3.4
|
idna==3.4
|
||||||
# via requests
|
importlib-resources==5.10.0
|
||||||
jinja2==3.1.2
|
Jinja2==3.1.2
|
||||||
# via
|
|
||||||
# ansible-core
|
|
||||||
# cookiecutter
|
|
||||||
# jinja2-time
|
|
||||||
# molecule
|
|
||||||
# molecule-vagrant
|
|
||||||
jinja2-time==0.2.0
|
jinja2-time==0.2.0
|
||||||
# via cookiecutter
|
|
||||||
jmespath==1.0.1
|
jmespath==1.0.1
|
||||||
# via -r requirements.in
|
jsonpatch==1.32
|
||||||
jsonpatch==1.33
|
|
||||||
# via -r requirements.in
|
|
||||||
jsonpointer==2.3
|
jsonpointer==2.3
|
||||||
# via jsonpatch
|
jsonschema==4.16.0
|
||||||
jsonschema==4.17.0
|
kubernetes==24.2.0
|
||||||
# via
|
MarkupSafe==2.1.1
|
||||||
# ansible-compat
|
molecule==4.0.1
|
||||||
# molecule
|
|
||||||
kubernetes==25.3.0
|
|
||||||
# via -r requirements.in
|
|
||||||
markupsafe==2.1.1
|
|
||||||
# via jinja2
|
|
||||||
molecule==4.0.4
|
|
||||||
# via
|
|
||||||
# -r requirements.in
|
|
||||||
# molecule-vagrant
|
|
||||||
molecule-vagrant==1.0.0
|
molecule-vagrant==1.0.0
|
||||||
# via -r requirements.in
|
mypy-extensions==0.4.3
|
||||||
netaddr==0.8.0
|
netaddr==0.8.0
|
||||||
# via -r requirements.in
|
|
||||||
nodeenv==1.7.0
|
|
||||||
# via pre-commit
|
|
||||||
oauthlib==3.2.2
|
oauthlib==3.2.2
|
||||||
# via requests-oauthlib
|
|
||||||
packaging==21.3
|
packaging==21.3
|
||||||
# via
|
pathspec==0.10.1
|
||||||
# ansible-compat
|
pkgutil-resolve-name==1.3.10
|
||||||
# ansible-core
|
|
||||||
# molecule
|
|
||||||
platformdirs==2.5.2
|
platformdirs==2.5.2
|
||||||
# via virtualenv
|
|
||||||
pluggy==1.0.0
|
pluggy==1.0.0
|
||||||
# via molecule
|
pre-commit==2.20.0
|
||||||
pre-commit==2.21.0
|
|
||||||
# via -r requirements.in
|
|
||||||
pre-commit-hooks==4.4.0
|
|
||||||
# via -r requirements.in
|
|
||||||
pyasn1==0.4.8
|
pyasn1==0.4.8
|
||||||
# via
|
|
||||||
# pyasn1-modules
|
|
||||||
# rsa
|
|
||||||
pyasn1-modules==0.2.8
|
pyasn1-modules==0.2.8
|
||||||
# via google-auth
|
|
||||||
pycparser==2.21
|
pycparser==2.21
|
||||||
# via cffi
|
Pygments==2.13.0
|
||||||
pygments==2.13.0
|
|
||||||
# via rich
|
|
||||||
pyparsing==3.0.9
|
pyparsing==3.0.9
|
||||||
# via packaging
|
pyrsistent==0.18.1
|
||||||
pyrsistent==0.19.2
|
|
||||||
# via jsonschema
|
|
||||||
python-dateutil==2.8.2
|
python-dateutil==2.8.2
|
||||||
# via
|
|
||||||
# arrow
|
|
||||||
# kubernetes
|
|
||||||
python-slugify==6.1.2
|
python-slugify==6.1.2
|
||||||
# via cookiecutter
|
|
||||||
python-vagrant==1.0.0
|
python-vagrant==1.0.0
|
||||||
# via molecule-vagrant
|
PyYAML==6.0
|
||||||
pyyaml==6.0.1
|
|
||||||
# via
|
|
||||||
# -r requirements.in
|
|
||||||
# ansible-compat
|
|
||||||
# ansible-core
|
|
||||||
# cookiecutter
|
|
||||||
# kubernetes
|
|
||||||
# molecule
|
|
||||||
# molecule-vagrant
|
|
||||||
# pre-commit
|
|
||||||
requests==2.28.1
|
requests==2.28.1
|
||||||
# via
|
|
||||||
# cookiecutter
|
|
||||||
# kubernetes
|
|
||||||
# requests-oauthlib
|
|
||||||
requests-oauthlib==1.3.1
|
requests-oauthlib==1.3.1
|
||||||
# via kubernetes
|
|
||||||
resolvelib==0.8.1
|
resolvelib==0.8.1
|
||||||
# via ansible-core
|
|
||||||
rich==12.6.0
|
rich==12.6.0
|
||||||
# via
|
|
||||||
# enrich
|
|
||||||
# molecule
|
|
||||||
rsa==4.9
|
rsa==4.9
|
||||||
# via google-auth
|
ruamel.yaml==0.17.21
|
||||||
ruamel-yaml==0.17.21
|
ruamel.yaml.clib==0.2.7
|
||||||
# via pre-commit-hooks
|
|
||||||
selinux==0.2.1
|
selinux==0.2.1
|
||||||
# via molecule-vagrant
|
|
||||||
six==1.16.0
|
six==1.16.0
|
||||||
# via
|
subprocess-tee==0.3.5
|
||||||
# google-auth
|
|
||||||
# kubernetes
|
|
||||||
# python-dateutil
|
|
||||||
subprocess-tee==0.4.1
|
|
||||||
# via ansible-compat
|
|
||||||
text-unidecode==1.3
|
text-unidecode==1.3
|
||||||
# via python-slugify
|
tomli==2.0.1
|
||||||
|
typing-extensions==4.4.0
|
||||||
urllib3==1.26.12
|
urllib3==1.26.12
|
||||||
# via
|
wcmatch==8.4.1
|
||||||
# kubernetes
|
websocket-client==1.4.1
|
||||||
# requests
|
yamllint==1.28.0
|
||||||
virtualenv==20.16.6
|
zipp==3.10.0
|
||||||
# via pre-commit
|
|
||||||
websocket-client==1.4.2
|
|
||||||
# via kubernetes
|
|
||||||
|
|
||||||
# The following packages are considered to be unsafe in a requirements file:
|
|
||||||
# setuptools
|
|
||||||
|
|||||||
2
reset.sh
2
reset.sh
@@ -1,3 +1,3 @@
|
|||||||
#!/bin/bash
|
#!/bin/bash
|
||||||
|
|
||||||
ansible-playbook reset.yml
|
ansible-playbook reset.yml -i inventory/my-cluster/hosts.ini
|
||||||
|
|||||||
17
reset.yml
17
reset.yml
@@ -1,24 +1,13 @@
|
|||||||
---
|
---
|
||||||
- name: Reset k3s cluster
|
|
||||||
hosts: k3s_cluster
|
- hosts: k3s_cluster
|
||||||
gather_facts: yes
|
gather_facts: yes
|
||||||
|
become: yes
|
||||||
roles:
|
roles:
|
||||||
- role: reset
|
- role: reset
|
||||||
become: true
|
|
||||||
- role: raspberrypi
|
- role: raspberrypi
|
||||||
become: true
|
|
||||||
vars: {state: absent}
|
vars: {state: absent}
|
||||||
post_tasks:
|
post_tasks:
|
||||||
- name: Reboot and wait for node to come back up
|
- name: Reboot and wait for node to come back up
|
||||||
become: true
|
|
||||||
reboot:
|
reboot:
|
||||||
reboot_timeout: 3600
|
reboot_timeout: 3600
|
||||||
|
|
||||||
- name: Revert changes to Proxmox cluster
|
|
||||||
hosts: proxmox
|
|
||||||
gather_facts: true
|
|
||||||
become: yes
|
|
||||||
remote_user: "{{ proxmox_lxc_ssh_user }}"
|
|
||||||
roles:
|
|
||||||
- role: reset_proxmox_lxc
|
|
||||||
when: proxmox_lxc_configure
|
|
||||||
|
|||||||
12
roles/k3s/master/defaults/main.yml
Normal file
12
roles/k3s/master/defaults/main.yml
Normal file
@@ -0,0 +1,12 @@
|
|||||||
|
---
|
||||||
|
ansible_user: root
|
||||||
|
server_init_args: >-
|
||||||
|
{% if groups['master'] | length > 1 %}
|
||||||
|
{% if ansible_hostname == hostvars[groups['master'][0]]['ansible_hostname'] %}
|
||||||
|
--cluster-init
|
||||||
|
{% else %}
|
||||||
|
--server https://{{ hostvars[groups['master'][0]].k3s_node_ip }}:6443
|
||||||
|
{% endif %}
|
||||||
|
--token {{ k3s_token }}
|
||||||
|
{% endif %}
|
||||||
|
{{ extra_server_args | default('') }}
|
||||||
@@ -1,22 +1,63 @@
|
|||||||
---
|
---
|
||||||
|
|
||||||
- name: Stop k3s-init
|
- name: Clean previous runs of k3s-init
|
||||||
systemd:
|
systemd:
|
||||||
name: k3s-init
|
name: k3s-init
|
||||||
state: stopped
|
state: stopped
|
||||||
failed_when: false
|
failed_when: false
|
||||||
|
|
||||||
- name: Clean previous runs of k3s-init # noqa command-instead-of-module
|
- name: Clean previous runs of k3s-init
|
||||||
# The systemd module does not support "reset-failed", so we need to resort to command.
|
|
||||||
command: systemctl reset-failed k3s-init
|
command: systemctl reset-failed k3s-init
|
||||||
failed_when: false
|
failed_when: false
|
||||||
changed_when: false
|
changed_when: false
|
||||||
|
args:
|
||||||
|
warn: false # The ansible systemd module does not support reset-failed
|
||||||
|
|
||||||
- name: Deploy vip manifest
|
- name: Create manifests directory on first master
|
||||||
include_tasks: vip.yml
|
file:
|
||||||
|
path: /var/lib/rancher/k3s/server/manifests
|
||||||
|
state: directory
|
||||||
|
owner: root
|
||||||
|
group: root
|
||||||
|
mode: 0644
|
||||||
|
when: ansible_hostname == hostvars[groups['master'][0]]['ansible_hostname']
|
||||||
|
|
||||||
- name: Deploy metallb manifest
|
- name: Copy vip rbac manifest to first master
|
||||||
include_tasks: metallb.yml
|
template:
|
||||||
|
src: "vip.rbac.yaml.j2"
|
||||||
|
dest: "/var/lib/rancher/k3s/server/manifests/vip-rbac.yaml"
|
||||||
|
owner: root
|
||||||
|
group: root
|
||||||
|
mode: 0644
|
||||||
|
when: ansible_hostname == hostvars[groups['master'][0]]['ansible_hostname']
|
||||||
|
|
||||||
|
- name: Copy vip manifest to first master
|
||||||
|
template:
|
||||||
|
src: "vip.yaml.j2"
|
||||||
|
dest: "/var/lib/rancher/k3s/server/manifests/vip.yaml"
|
||||||
|
owner: root
|
||||||
|
group: root
|
||||||
|
mode: 0644
|
||||||
|
when: ansible_hostname == hostvars[groups['master'][0]]['ansible_hostname']
|
||||||
|
|
||||||
|
# these will be copied and installed now, then tested later and apply config
|
||||||
|
- name: Copy metallb namespace to first master
|
||||||
|
template:
|
||||||
|
src: "metallb.namespace.j2"
|
||||||
|
dest: "/var/lib/rancher/k3s/server/manifests/metallb-namespace.yaml"
|
||||||
|
owner: root
|
||||||
|
group: root
|
||||||
|
mode: 0644
|
||||||
|
when: ansible_hostname == hostvars[groups['master'][0]]['ansible_hostname']
|
||||||
|
|
||||||
|
- name: Copy metallb namespace to first master
|
||||||
|
template:
|
||||||
|
src: "metallb.crds.j2"
|
||||||
|
dest: "/var/lib/rancher/k3s/server/manifests/metallb-crds.yaml"
|
||||||
|
owner: root
|
||||||
|
group: root
|
||||||
|
mode: 0644
|
||||||
|
when: ansible_hostname == hostvars[groups['master'][0]]['ansible_hostname']
|
||||||
|
|
||||||
- name: Init cluster inside the transient k3s-init service
|
- name: Init cluster inside the transient k3s-init service
|
||||||
command:
|
command:
|
||||||
@@ -25,15 +66,16 @@
|
|||||||
--unit=k3s-init \
|
--unit=k3s-init \
|
||||||
k3s server {{ server_init_args }}"
|
k3s server {{ server_init_args }}"
|
||||||
creates: "{{ systemd_dir }}/k3s.service"
|
creates: "{{ systemd_dir }}/k3s.service"
|
||||||
|
args:
|
||||||
|
warn: false # The ansible systemd module does not support transient units
|
||||||
|
|
||||||
- name: Verification
|
- name: Verification
|
||||||
when: not ansible_check_mode
|
|
||||||
block:
|
block:
|
||||||
- name: Verify that all nodes actually joined (check k3s-init.service if this fails)
|
- name: Verify that all nodes actually joined (check k3s-init.service if this fails)
|
||||||
command:
|
command:
|
||||||
cmd: k3s kubectl get nodes -l "node-role.kubernetes.io/master=true" -o=jsonpath="{.items[*].metadata.name}"
|
cmd: k3s kubectl get nodes -l "node-role.kubernetes.io/master=true" -o=jsonpath="{.items[*].metadata.name}"
|
||||||
register: nodes
|
register: nodes
|
||||||
until: nodes.rc == 0 and (nodes.stdout.split() | length) == (groups[group_name_master | default('master')] | length) # yamllint disable-line rule:line-length
|
until: nodes.rc == 0 and (nodes.stdout.split() | length) == (groups['master'] | length)
|
||||||
retries: "{{ retry_count | default(20) }}"
|
retries: "{{ retry_count | default(20) }}"
|
||||||
delay: 10
|
delay: 10
|
||||||
changed_when: false
|
changed_when: false
|
||||||
@@ -49,6 +91,7 @@
|
|||||||
name: k3s-init
|
name: k3s-init
|
||||||
state: stopped
|
state: stopped
|
||||||
failed_when: false
|
failed_when: false
|
||||||
|
when: not ansible_check_mode
|
||||||
|
|
||||||
- name: Copy K3s service file
|
- name: Copy K3s service file
|
||||||
register: k3s_service
|
register: k3s_service
|
||||||
@@ -96,24 +139,24 @@
|
|||||||
|
|
||||||
- name: Create directory .kube
|
- name: Create directory .kube
|
||||||
file:
|
file:
|
||||||
path: "{{ ansible_user_dir }}/.kube"
|
path: ~{{ ansible_user }}/.kube
|
||||||
state: directory
|
state: directory
|
||||||
owner: "{{ ansible_user_id }}"
|
owner: "{{ ansible_user }}"
|
||||||
mode: "u=rwx,g=rx,o="
|
mode: "u=rwx,g=rx,o="
|
||||||
|
|
||||||
- name: Copy config file to user home directory
|
- name: Copy config file to user home directory
|
||||||
copy:
|
copy:
|
||||||
src: /etc/rancher/k3s/k3s.yaml
|
src: /etc/rancher/k3s/k3s.yaml
|
||||||
dest: "{{ ansible_user_dir }}/.kube/config"
|
dest: ~{{ ansible_user }}/.kube/config
|
||||||
remote_src: yes
|
remote_src: yes
|
||||||
owner: "{{ ansible_user_id }}"
|
owner: "{{ ansible_user }}"
|
||||||
mode: "u=rw,g=,o="
|
mode: "u=rw,g=,o="
|
||||||
|
|
||||||
- name: Configure kubectl cluster to {{ endpoint_url }}
|
- name: Configure kubectl cluster to {{ endpoint_url }}
|
||||||
command: >-
|
command: >-
|
||||||
k3s kubectl config set-cluster default
|
k3s kubectl config set-cluster default
|
||||||
--server={{ endpoint_url }}
|
--server={{ endpoint_url }}
|
||||||
--kubeconfig {{ ansible_user_dir }}/.kube/config
|
--kubeconfig ~{{ ansible_user }}/.kube/config
|
||||||
changed_when: true
|
changed_when: true
|
||||||
vars:
|
vars:
|
||||||
endpoint_url: >-
|
endpoint_url: >-
|
||||||
1797
roles/k3s/master/templates/metallb.crds.j2
Normal file
1797
roles/k3s/master/templates/metallb.crds.j2
Normal file
File diff suppressed because it is too large
Load Diff
6
roles/k3s/master/templates/metallb.namespace.j2
Normal file
6
roles/k3s/master/templates/metallb.namespace.j2
Normal file
@@ -0,0 +1,6 @@
|
|||||||
|
apiVersion: v1
|
||||||
|
kind: Namespace
|
||||||
|
metadata:
|
||||||
|
name: metallb-system
|
||||||
|
labels:
|
||||||
|
app: metallb
|
||||||
32
roles/k3s/master/templates/vip.rbac.yaml.j2
Normal file
32
roles/k3s/master/templates/vip.rbac.yaml.j2
Normal file
@@ -0,0 +1,32 @@
|
|||||||
|
apiVersion: v1
|
||||||
|
kind: ServiceAccount
|
||||||
|
metadata:
|
||||||
|
name: kube-vip
|
||||||
|
namespace: kube-system
|
||||||
|
---
|
||||||
|
apiVersion: rbac.authorization.k8s.io/v1
|
||||||
|
kind: ClusterRole
|
||||||
|
metadata:
|
||||||
|
annotations:
|
||||||
|
rbac.authorization.kubernetes.io/autoupdate: "true"
|
||||||
|
name: system:kube-vip-role
|
||||||
|
rules:
|
||||||
|
- apiGroups: [""]
|
||||||
|
resources: ["services", "services/status", "nodes", "endpoints"]
|
||||||
|
verbs: ["list","get","watch", "update"]
|
||||||
|
- apiGroups: ["coordination.k8s.io"]
|
||||||
|
resources: ["leases"]
|
||||||
|
verbs: ["list", "get", "watch", "update", "create"]
|
||||||
|
---
|
||||||
|
kind: ClusterRoleBinding
|
||||||
|
apiVersion: rbac.authorization.k8s.io/v1
|
||||||
|
metadata:
|
||||||
|
name: system:kube-vip-binding
|
||||||
|
roleRef:
|
||||||
|
apiGroup: rbac.authorization.k8s.io
|
||||||
|
kind: ClusterRole
|
||||||
|
name: system:kube-vip-role
|
||||||
|
subjects:
|
||||||
|
- kind: ServiceAccount
|
||||||
|
name: kube-vip
|
||||||
|
namespace: kube-system
|
||||||
@@ -30,10 +30,8 @@ spec:
|
|||||||
value: "true"
|
value: "true"
|
||||||
- name: port
|
- name: port
|
||||||
value: "6443"
|
value: "6443"
|
||||||
{% if kube_vip_iface %}
|
|
||||||
- name: vip_interface
|
- name: vip_interface
|
||||||
value: {{ kube_vip_iface }}
|
value: {{ flannel_iface }}
|
||||||
{% endif %}
|
|
||||||
- name: vip_cidr
|
- name: vip_cidr
|
||||||
value: "{{ apiserver_endpoint | ansible.utils.ipsubnet | ansible.utils.ipaddr('prefix') }}"
|
value: "{{ apiserver_endpoint | ansible.utils.ipsubnet | ansible.utils.ipaddr('prefix') }}"
|
||||||
- name: cp_enable
|
- name: cp_enable
|
||||||
@@ -1,3 +0,0 @@
|
|||||||
---
|
|
||||||
# Name of the master group
|
|
||||||
group_name_master: master
|
|
||||||
@@ -7,7 +7,7 @@ After=network-online.target
|
|||||||
Type=notify
|
Type=notify
|
||||||
ExecStartPre=-/sbin/modprobe br_netfilter
|
ExecStartPre=-/sbin/modprobe br_netfilter
|
||||||
ExecStartPre=-/sbin/modprobe overlay
|
ExecStartPre=-/sbin/modprobe overlay
|
||||||
ExecStart=/usr/local/bin/k3s agent --server https://{{ apiserver_endpoint | ansible.utils.ipwrap }}:6443 --token {{ hostvars[groups[group_name_master | default('master')][0]]['token'] | default(k3s_token) }} {{ extra_agent_args | default("") }}
|
ExecStart=/usr/local/bin/k3s agent --server https://{{ apiserver_endpoint | ansible.utils.ipwrap }}:6443 --token {{ hostvars[groups['master'][0]]['token'] | default(k3s_token) }} {{ extra_agent_args | default("") }}
|
||||||
KillMode=process
|
KillMode=process
|
||||||
Delegate=yes
|
Delegate=yes
|
||||||
# Having non-zero Limit*s causes performance problems due to accounting overhead
|
# Having non-zero Limit*s causes performance problems due to accounting overhead
|
||||||
@@ -1,6 +1,3 @@
|
|||||||
---
|
---
|
||||||
# Timeout to wait for MetalLB services to come up
|
# Timeout to wait for MetalLB services to come up
|
||||||
metal_lb_available_timeout: 120s
|
metal_lb_available_timeout: 120s
|
||||||
|
|
||||||
# Name of the master group
|
|
||||||
group_name_master: master
|
|
||||||
@@ -3,25 +3,25 @@
|
|||||||
file:
|
file:
|
||||||
path: /tmp/k3s
|
path: /tmp/k3s
|
||||||
state: directory
|
state: directory
|
||||||
owner: "{{ ansible_user_id }}"
|
owner: "{{ ansible_user }}"
|
||||||
mode: 0755
|
mode: 0755
|
||||||
with_items: "{{ groups[group_name_master | default('master')] }}"
|
with_items: "{{ groups['master'] }}"
|
||||||
run_once: true
|
run_once: true
|
||||||
|
|
||||||
- name: Copy metallb CRs manifest to first master
|
- name: Copy metallb CRs manifest to first master
|
||||||
template:
|
template:
|
||||||
src: "metallb.crs.j2"
|
src: "metallb.crs.j2"
|
||||||
dest: "/tmp/k3s/metallb-crs.yaml"
|
dest: "/tmp/k3s/metallb-crs.yaml"
|
||||||
owner: "{{ ansible_user_id }}"
|
owner: "{{ ansible_user }}"
|
||||||
mode: 0755
|
mode: 0755
|
||||||
with_items: "{{ groups[group_name_master | default('master')] }}"
|
with_items: "{{ groups['master'] }}"
|
||||||
run_once: true
|
run_once: true
|
||||||
|
|
||||||
- name: Test metallb-system namespace
|
- name: Test metallb-system namespace
|
||||||
command: >-
|
command: >-
|
||||||
k3s kubectl -n metallb-system
|
k3s kubectl -n metallb-system
|
||||||
changed_when: false
|
changed_when: false
|
||||||
with_items: "{{ groups[group_name_master | default('master')] }}"
|
with_items: "{{ groups['master'] }}"
|
||||||
run_once: true
|
run_once: true
|
||||||
|
|
||||||
- name: Wait for MetalLB resources
|
- name: Wait for MetalLB resources
|
||||||
@@ -66,7 +66,7 @@
|
|||||||
command: >-
|
command: >-
|
||||||
k3s kubectl -n metallb-system get endpoints webhook-service
|
k3s kubectl -n metallb-system get endpoints webhook-service
|
||||||
changed_when: false
|
changed_when: false
|
||||||
with_items: "{{ groups[group_name_master | default('master')] }}"
|
with_items: "{{ groups['master'] }}"
|
||||||
run_once: true
|
run_once: true
|
||||||
|
|
||||||
- name: Apply metallb CRs
|
- name: Apply metallb CRs
|
||||||
@@ -79,23 +79,16 @@
|
|||||||
until: this.rc == 0
|
until: this.rc == 0
|
||||||
retries: 5
|
retries: 5
|
||||||
|
|
||||||
- name: Test metallb-system resources for Layer 2 configuration
|
- name: Test metallb-system resources
|
||||||
command: >-
|
command: >-
|
||||||
k3s kubectl -n metallb-system get {{ item }}
|
k3s kubectl -n metallb-system get {{ item }}
|
||||||
changed_when: false
|
changed_when: false
|
||||||
run_once: true
|
run_once: true
|
||||||
when: metal_lb_mode == "layer2"
|
|
||||||
with_items:
|
with_items:
|
||||||
- IPAddressPool
|
- IPAddressPool
|
||||||
- L2Advertisement
|
- L2Advertisement
|
||||||
|
|
||||||
- name: Test metallb-system resources for BGP configuration
|
- name: Remove tmp directory used for manifests
|
||||||
command: >-
|
file:
|
||||||
k3s kubectl -n metallb-system get {{ item }}
|
path: /tmp/k3s
|
||||||
changed_when: false
|
state: absent
|
||||||
run_once: true
|
|
||||||
when: metal_lb_mode == "bgp"
|
|
||||||
with_items:
|
|
||||||
- IPAddressPool
|
|
||||||
- BGPPeer
|
|
||||||
- BGPAdvertisement
|
|
||||||
@@ -13,31 +13,9 @@ spec:
|
|||||||
{% for range in metal_lb_ip_range %}
|
{% for range in metal_lb_ip_range %}
|
||||||
- {{ range }}
|
- {{ range }}
|
||||||
{% endfor %}
|
{% endfor %}
|
||||||
|
|
||||||
{% if metal_lb_mode == "layer2" %}
|
|
||||||
---
|
---
|
||||||
apiVersion: metallb.io/v1beta1
|
apiVersion: metallb.io/v1beta1
|
||||||
kind: L2Advertisement
|
kind: L2Advertisement
|
||||||
metadata:
|
metadata:
|
||||||
name: default
|
name: default
|
||||||
namespace: metallb-system
|
namespace: metallb-system
|
||||||
{% endif %}
|
|
||||||
{% if metal_lb_mode == "bgp" %}
|
|
||||||
---
|
|
||||||
apiVersion: metallb.io/v1beta2
|
|
||||||
kind: BGPPeer
|
|
||||||
metadata:
|
|
||||||
name: default
|
|
||||||
namespace: metallb-system
|
|
||||||
spec:
|
|
||||||
myASN: {{ metal_lb_bgp_my_asn }}
|
|
||||||
peerASN: {{ metal_lb_bgp_peer_asn }}
|
|
||||||
peerAddress: {{ metal_lb_bgp_peer_address }}
|
|
||||||
|
|
||||||
---
|
|
||||||
apiVersion: metallb.io/v1beta1
|
|
||||||
kind: BGPAdvertisement
|
|
||||||
metadata:
|
|
||||||
name: default
|
|
||||||
namespace: metallb-system
|
|
||||||
{% endif %}
|
|
||||||
@@ -1,20 +0,0 @@
|
|||||||
---
|
|
||||||
# If you want to explicitly define an interface that ALL control nodes
|
|
||||||
# should use to propagate the VIP, define it here. Otherwise, kube-vip
|
|
||||||
# will determine the right interface automatically at runtime.
|
|
||||||
kube_vip_iface: null
|
|
||||||
|
|
||||||
# Name of the master group
|
|
||||||
group_name_master: master
|
|
||||||
|
|
||||||
# yamllint disable rule:line-length
|
|
||||||
server_init_args: >-
|
|
||||||
{% if groups[group_name_master | default('master')] | length > 1 %}
|
|
||||||
{% if ansible_hostname == hostvars[groups[group_name_master | default('master')][0]]['ansible_hostname'] %}
|
|
||||||
--cluster-init
|
|
||||||
{% else %}
|
|
||||||
--server https://{{ hostvars[groups[group_name_master | default('master')][0]].k3s_node_ip | split(",") | first | ansible.utils.ipwrap }}:6443
|
|
||||||
{% endif %}
|
|
||||||
--token {{ k3s_token }}
|
|
||||||
{% endif %}
|
|
||||||
{{ extra_server_args | default('') }}
|
|
||||||
@@ -1,30 +0,0 @@
|
|||||||
---
|
|
||||||
- name: Create manifests directory on first master
|
|
||||||
file:
|
|
||||||
path: /var/lib/rancher/k3s/server/manifests
|
|
||||||
state: directory
|
|
||||||
owner: root
|
|
||||||
group: root
|
|
||||||
mode: 0644
|
|
||||||
when: ansible_hostname == hostvars[groups[group_name_master | default('master')][0]]['ansible_hostname']
|
|
||||||
|
|
||||||
- name: "Download to first master: manifest for metallb-{{ metal_lb_type }}"
|
|
||||||
ansible.builtin.get_url:
|
|
||||||
url: "https://raw.githubusercontent.com/metallb/metallb/{{ metal_lb_controller_tag_version }}/config/manifests/metallb-{{ metal_lb_type }}.yaml" # noqa yaml[line-length]
|
|
||||||
dest: "/var/lib/rancher/k3s/server/manifests/metallb-crds.yaml"
|
|
||||||
owner: root
|
|
||||||
group: root
|
|
||||||
mode: 0644
|
|
||||||
when: ansible_hostname == hostvars[groups[group_name_master | default('master')][0]]['ansible_hostname']
|
|
||||||
|
|
||||||
- name: Set image versions in manifest for metallb-{{ metal_lb_type }}
|
|
||||||
ansible.builtin.replace:
|
|
||||||
path: "/var/lib/rancher/k3s/server/manifests/metallb-crds.yaml"
|
|
||||||
regexp: "{{ item.change | ansible.builtin.regex_escape }}"
|
|
||||||
replace: "{{ item.to }}"
|
|
||||||
with_items:
|
|
||||||
- change: "metallb/speaker:{{ metal_lb_controller_tag_version }}"
|
|
||||||
to: "metallb/speaker:{{ metal_lb_speaker_tag_version }}"
|
|
||||||
loop_control:
|
|
||||||
label: "{{ item.change }} => {{ item.to }}"
|
|
||||||
when: ansible_hostname == hostvars[groups[group_name_master | default('master')][0]]['ansible_hostname']
|
|
||||||
@@ -1,27 +0,0 @@
|
|||||||
---
|
|
||||||
- name: Create manifests directory on first master
|
|
||||||
file:
|
|
||||||
path: /var/lib/rancher/k3s/server/manifests
|
|
||||||
state: directory
|
|
||||||
owner: root
|
|
||||||
group: root
|
|
||||||
mode: 0644
|
|
||||||
when: ansible_hostname == hostvars[groups[group_name_master | default('master')][0]]['ansible_hostname']
|
|
||||||
|
|
||||||
- name: Download vip rbac manifest to first master
|
|
||||||
ansible.builtin.get_url:
|
|
||||||
url: "https://raw.githubusercontent.com/kube-vip/kube-vip/{{ kube_vip_tag_version }}/docs/manifests/rbac.yaml"
|
|
||||||
dest: "/var/lib/rancher/k3s/server/manifests/vip-rbac.yaml"
|
|
||||||
owner: root
|
|
||||||
group: root
|
|
||||||
mode: 0644
|
|
||||||
when: ansible_hostname == hostvars[groups[group_name_master | default('master')][0]]['ansible_hostname']
|
|
||||||
|
|
||||||
- name: Copy vip manifest to first master
|
|
||||||
template:
|
|
||||||
src: "vip.yaml.j2"
|
|
||||||
dest: "/var/lib/rancher/k3s/server/manifests/vip.yaml"
|
|
||||||
owner: root
|
|
||||||
group: root
|
|
||||||
mode: 0644
|
|
||||||
when: ansible_hostname == hostvars[groups[group_name_master | default('master')][0]]['ansible_hostname']
|
|
||||||
@@ -1,8 +0,0 @@
|
|||||||
---
|
|
||||||
- name: Deploy metallb pool
|
|
||||||
include_tasks: metallb.yml
|
|
||||||
|
|
||||||
- name: Remove tmp directory used for manifests
|
|
||||||
file:
|
|
||||||
path: /tmp/k3s
|
|
||||||
state: absent
|
|
||||||
@@ -1,4 +0,0 @@
|
|||||||
---
|
|
||||||
- name: Reboot server
|
|
||||||
become: true
|
|
||||||
reboot:
|
|
||||||
@@ -1,21 +0,0 @@
|
|||||||
---
|
|
||||||
- name: Check for rc.local file
|
|
||||||
stat:
|
|
||||||
path: /etc/rc.local
|
|
||||||
register: rcfile
|
|
||||||
|
|
||||||
- name: Create rc.local if needed
|
|
||||||
lineinfile:
|
|
||||||
path: /etc/rc.local
|
|
||||||
line: "#!/bin/sh -e"
|
|
||||||
create: true
|
|
||||||
insertbefore: BOF
|
|
||||||
mode: "u=rwx,g=rx,o=rx"
|
|
||||||
when: not rcfile.stat.exists
|
|
||||||
|
|
||||||
- name: Write rc.local file
|
|
||||||
blockinfile:
|
|
||||||
path: /etc/rc.local
|
|
||||||
content: "{{ lookup('template', 'templates/rc.local.j2') }}"
|
|
||||||
state: present
|
|
||||||
notify: reboot server
|
|
||||||
@@ -1,30 +1,30 @@
|
|||||||
---
|
---
|
||||||
- name: Set same timezone on every Server
|
- name: Set same timezone on every Server
|
||||||
community.general.timezone:
|
timezone:
|
||||||
name: "{{ system_timezone }}"
|
name: "{{ system_timezone }}"
|
||||||
when: (system_timezone is defined) and (system_timezone != "Your/Timezone")
|
when: (system_timezone is defined) and (system_timezone != "Your/Timezone")
|
||||||
|
|
||||||
- name: Set SELinux to disabled state
|
- name: Set SELinux to disabled state
|
||||||
ansible.posix.selinux:
|
selinux:
|
||||||
state: disabled
|
state: disabled
|
||||||
when: ansible_os_family == "RedHat"
|
when: ansible_os_family == "RedHat"
|
||||||
|
|
||||||
- name: Enable IPv4 forwarding
|
- name: Enable IPv4 forwarding
|
||||||
ansible.posix.sysctl:
|
sysctl:
|
||||||
name: net.ipv4.ip_forward
|
name: net.ipv4.ip_forward
|
||||||
value: "1"
|
value: "1"
|
||||||
state: present
|
state: present
|
||||||
reload: yes
|
reload: yes
|
||||||
|
|
||||||
- name: Enable IPv6 forwarding
|
- name: Enable IPv6 forwarding
|
||||||
ansible.posix.sysctl:
|
sysctl:
|
||||||
name: net.ipv6.conf.all.forwarding
|
name: net.ipv6.conf.all.forwarding
|
||||||
value: "1"
|
value: "1"
|
||||||
state: present
|
state: present
|
||||||
reload: yes
|
reload: yes
|
||||||
|
|
||||||
- name: Enable IPv6 router advertisements
|
- name: Enable IPv6 router advertisements
|
||||||
ansible.posix.sysctl:
|
sysctl:
|
||||||
name: net.ipv6.conf.all.accept_ra
|
name: net.ipv6.conf.all.accept_ra
|
||||||
value: "2"
|
value: "2"
|
||||||
state: present
|
state: present
|
||||||
@@ -38,13 +38,13 @@
|
|||||||
when: ansible_os_family == "RedHat"
|
when: ansible_os_family == "RedHat"
|
||||||
|
|
||||||
- name: Load br_netfilter
|
- name: Load br_netfilter
|
||||||
community.general.modprobe:
|
modprobe:
|
||||||
name: br_netfilter
|
name: br_netfilter
|
||||||
state: present
|
state: present
|
||||||
when: ansible_os_family == "RedHat"
|
when: ansible_os_family == "RedHat"
|
||||||
|
|
||||||
- name: Set bridge-nf-call-iptables (just to be sure)
|
- name: Set bridge-nf-call-iptables (just to be sure)
|
||||||
ansible.posix.sysctl:
|
sysctl:
|
||||||
name: "{{ item }}"
|
name: "{{ item }}"
|
||||||
value: "1"
|
value: "1"
|
||||||
state: present
|
state: present
|
||||||
|
|||||||
@@ -1,11 +0,0 @@
|
|||||||
---
|
|
||||||
- name: Reboot containers
|
|
||||||
block:
|
|
||||||
- name: Get container ids from filtered files
|
|
||||||
set_fact:
|
|
||||||
proxmox_lxc_filtered_ids: >-
|
|
||||||
{{ proxmox_lxc_filtered_files | map("split", "/") | map("last") | map("split", ".") | map("first") }}
|
|
||||||
- name: Reboot container
|
|
||||||
command: "pct reboot {{ item }}"
|
|
||||||
loop: "{{ proxmox_lxc_filtered_ids }}"
|
|
||||||
changed_when: true
|
|
||||||
@@ -1,44 +0,0 @@
|
|||||||
---
|
|
||||||
- name: Check for container files that exist on this host
|
|
||||||
stat:
|
|
||||||
path: "/etc/pve/lxc/{{ item }}.conf"
|
|
||||||
loop: "{{ proxmox_lxc_ct_ids }}"
|
|
||||||
register: stat_results
|
|
||||||
|
|
||||||
- name: Filter out files that do not exist
|
|
||||||
set_fact:
|
|
||||||
proxmox_lxc_filtered_files:
|
|
||||||
'{{ stat_results.results | rejectattr("stat.exists", "false") | map(attribute="stat.path") }}'
|
|
||||||
|
|
||||||
# https://gist.github.com/triangletodd/02f595cd4c0dc9aac5f7763ca2264185
|
|
||||||
- name: Ensure lxc config has the right apparmor profile
|
|
||||||
lineinfile:
|
|
||||||
dest: "{{ item }}"
|
|
||||||
regexp: "^lxc.apparmor.profile"
|
|
||||||
line: "lxc.apparmor.profile: unconfined"
|
|
||||||
loop: "{{ proxmox_lxc_filtered_files }}"
|
|
||||||
notify: reboot containers
|
|
||||||
|
|
||||||
- name: Ensure lxc config has the right cgroup
|
|
||||||
lineinfile:
|
|
||||||
dest: "{{ item }}"
|
|
||||||
regexp: "^lxc.cgroup.devices.allow"
|
|
||||||
line: "lxc.cgroup.devices.allow: a"
|
|
||||||
loop: "{{ proxmox_lxc_filtered_files }}"
|
|
||||||
notify: reboot containers
|
|
||||||
|
|
||||||
- name: Ensure lxc config has the right cap drop
|
|
||||||
lineinfile:
|
|
||||||
dest: "{{ item }}"
|
|
||||||
regexp: "^lxc.cap.drop"
|
|
||||||
line: "lxc.cap.drop: "
|
|
||||||
loop: "{{ proxmox_lxc_filtered_files }}"
|
|
||||||
notify: reboot containers
|
|
||||||
|
|
||||||
- name: Ensure lxc config has the right mounts
|
|
||||||
lineinfile:
|
|
||||||
dest: "{{ item }}"
|
|
||||||
regexp: "^lxc.mount.auto"
|
|
||||||
line: 'lxc.mount.auto: "proc:rw sys:rw"'
|
|
||||||
loop: "{{ proxmox_lxc_filtered_files }}"
|
|
||||||
notify: reboot containers
|
|
||||||
@@ -47,16 +47,20 @@
|
|||||||
- raspberry_pi|default(false)
|
- raspberry_pi|default(false)
|
||||||
- ansible_facts.lsb.description|default("") is match("Debian.*bullseye")
|
- ansible_facts.lsb.description|default("") is match("Debian.*bullseye")
|
||||||
|
|
||||||
- name: Execute OS related tasks on the Raspberry Pi - {{ action_ }}
|
- name: execute OS related tasks on the Raspberry Pi - {{ action }}
|
||||||
include_tasks: "{{ item }}"
|
include_tasks: "{{ item }}"
|
||||||
with_first_found:
|
with_first_found:
|
||||||
- "{{ action_ }}/{{ detected_distribution }}-{{ detected_distribution_major_version }}.yml"
|
- "{{ action }}/{{ detected_distribution }}-{{ detected_distribution_major_version }}.yml"
|
||||||
- "{{ action_ }}/{{ detected_distribution }}.yml"
|
- "{{ action }}/{{ detected_distribution }}.yml"
|
||||||
- "{{ action_ }}/{{ ansible_distribution }}-{{ ansible_distribution_major_version }}.yml"
|
- "{{ action }}/{{ ansible_distribution }}-{{ ansible_distribution_major_version }}.yml"
|
||||||
- "{{ action_ }}/{{ ansible_distribution }}.yml"
|
- "{{ action }}/{{ ansible_distribution }}.yml"
|
||||||
- "{{ action_ }}/default.yml"
|
- "{{ action }}/default.yml"
|
||||||
vars:
|
vars:
|
||||||
action_: >-
|
action: >-
|
||||||
{% if state == "present" %}setup{% else %}teardown{% endif %}
|
{% if state == "present" -%}
|
||||||
|
setup
|
||||||
|
{%- else -%}
|
||||||
|
teardown
|
||||||
|
{%- endif %}
|
||||||
when:
|
when:
|
||||||
- raspberry_pi|default(false)
|
- raspberry_pi|default(false)
|
||||||
|
|||||||
@@ -8,22 +8,20 @@
|
|||||||
notify: reboot
|
notify: reboot
|
||||||
|
|
||||||
- name: Install iptables
|
- name: Install iptables
|
||||||
apt:
|
apt: name=iptables state=present
|
||||||
name: iptables
|
|
||||||
state: present
|
|
||||||
|
|
||||||
- name: Flush iptables before changing to iptables-legacy
|
- name: Flush iptables before changing to iptables-legacy
|
||||||
iptables:
|
iptables:
|
||||||
flush: true
|
flush: true
|
||||||
|
|
||||||
- name: Changing to iptables-legacy
|
- name: Changing to iptables-legacy
|
||||||
community.general.alternatives:
|
alternatives:
|
||||||
path: /usr/sbin/iptables-legacy
|
path: /usr/sbin/iptables-legacy
|
||||||
name: iptables
|
name: iptables
|
||||||
register: ip4_legacy
|
register: ip4_legacy
|
||||||
|
|
||||||
- name: Changing to ip6tables-legacy
|
- name: Changing to ip6tables-legacy
|
||||||
community.general.alternatives:
|
alternatives:
|
||||||
path: /usr/sbin/ip6tables-legacy
|
path: /usr/sbin/ip6tables-legacy
|
||||||
name: ip6tables
|
name: ip6tables
|
||||||
register: ip6_legacy
|
register: ip6_legacy
|
||||||
|
|||||||
@@ -54,31 +54,3 @@
|
|||||||
file:
|
file:
|
||||||
path: /tmp/k3s
|
path: /tmp/k3s
|
||||||
state: absent
|
state: absent
|
||||||
|
|
||||||
- name: Check if rc.local exists
|
|
||||||
stat:
|
|
||||||
path: /etc/rc.local
|
|
||||||
register: rcfile
|
|
||||||
|
|
||||||
- name: Remove rc.local modifications for proxmox lxc containers
|
|
||||||
become: true
|
|
||||||
blockinfile:
|
|
||||||
path: /etc/rc.local
|
|
||||||
content: "{{ lookup('template', 'templates/rc.local.j2') }}"
|
|
||||||
create: false
|
|
||||||
state: absent
|
|
||||||
when: proxmox_lxc_configure and rcfile.stat.exists
|
|
||||||
|
|
||||||
- name: Check rc.local for cleanup
|
|
||||||
become: true
|
|
||||||
slurp:
|
|
||||||
src: /etc/rc.local
|
|
||||||
register: rcslurp
|
|
||||||
when: proxmox_lxc_configure and rcfile.stat.exists
|
|
||||||
|
|
||||||
- name: Cleanup rc.local if we only have a Shebang line
|
|
||||||
become: true
|
|
||||||
file:
|
|
||||||
path: /etc/rc.local
|
|
||||||
state: absent
|
|
||||||
when: proxmox_lxc_configure and rcfile.stat.exists and ((rcslurp.content | b64decode).splitlines() | length) <= 1
|
|
||||||
|
|||||||
@@ -9,7 +9,7 @@
|
|||||||
check_mode: false
|
check_mode: false
|
||||||
|
|
||||||
- name: Umount filesystem
|
- name: Umount filesystem
|
||||||
ansible.posix.mount:
|
mount:
|
||||||
path: "{{ item }}"
|
path: "{{ item }}"
|
||||||
state: unmounted
|
state: unmounted
|
||||||
with_items:
|
with_items:
|
||||||
|
|||||||
@@ -1 +0,0 @@
|
|||||||
../../proxmox_lxc/handlers/main.yml
|
|
||||||
@@ -1,47 +0,0 @@
|
|||||||
---
|
|
||||||
- name: Check for container files that exist on this host
|
|
||||||
stat:
|
|
||||||
path: "/etc/pve/lxc/{{ item }}.conf"
|
|
||||||
loop: "{{ proxmox_lxc_ct_ids }}"
|
|
||||||
register: stat_results
|
|
||||||
|
|
||||||
- name: Filter out files that do not exist
|
|
||||||
set_fact:
|
|
||||||
proxmox_lxc_filtered_files:
|
|
||||||
'{{ stat_results.results | rejectattr("stat.exists", "false") | map(attribute="stat.path") }}'
|
|
||||||
|
|
||||||
- name: Remove LXC apparmor profile
|
|
||||||
lineinfile:
|
|
||||||
dest: "{{ item }}"
|
|
||||||
regexp: "^lxc.apparmor.profile"
|
|
||||||
line: "lxc.apparmor.profile: unconfined"
|
|
||||||
state: absent
|
|
||||||
loop: "{{ proxmox_lxc_filtered_files }}"
|
|
||||||
notify: reboot containers
|
|
||||||
|
|
||||||
- name: Remove lxc cgroups
|
|
||||||
lineinfile:
|
|
||||||
dest: "{{ item }}"
|
|
||||||
regexp: "^lxc.cgroup.devices.allow"
|
|
||||||
line: "lxc.cgroup.devices.allow: a"
|
|
||||||
state: absent
|
|
||||||
loop: "{{ proxmox_lxc_filtered_files }}"
|
|
||||||
notify: reboot containers
|
|
||||||
|
|
||||||
- name: Remove lxc cap drop
|
|
||||||
lineinfile:
|
|
||||||
dest: "{{ item }}"
|
|
||||||
regexp: "^lxc.cap.drop"
|
|
||||||
line: "lxc.cap.drop: "
|
|
||||||
state: absent
|
|
||||||
loop: "{{ proxmox_lxc_filtered_files }}"
|
|
||||||
notify: reboot containers
|
|
||||||
|
|
||||||
- name: Remove lxc mounts
|
|
||||||
lineinfile:
|
|
||||||
dest: "{{ item }}"
|
|
||||||
regexp: "^lxc.mount.auto"
|
|
||||||
line: 'lxc.mount.auto: "proc:rw sys:rw"'
|
|
||||||
state: absent
|
|
||||||
loop: "{{ proxmox_lxc_filtered_files }}"
|
|
||||||
notify: reboot containers
|
|
||||||
40
site.yml
40
site.yml
@@ -1,40 +1,24 @@
|
|||||||
---
|
---
|
||||||
- name: Prepare Proxmox cluster
|
|
||||||
hosts: proxmox
|
- hosts: k3s_cluster
|
||||||
gather_facts: true
|
gather_facts: yes
|
||||||
become: yes
|
become: yes
|
||||||
roles:
|
roles:
|
||||||
- role: proxmox_lxc
|
|
||||||
when: proxmox_lxc_configure
|
|
||||||
|
|
||||||
- name: Prepare k3s nodes
|
|
||||||
hosts: k3s_cluster
|
|
||||||
gather_facts: yes
|
|
||||||
roles:
|
|
||||||
- role: lxc
|
|
||||||
become: true
|
|
||||||
when: proxmox_lxc_configure
|
|
||||||
- role: prereq
|
- role: prereq
|
||||||
become: true
|
|
||||||
- role: download
|
- role: download
|
||||||
become: true
|
|
||||||
- role: raspberrypi
|
- role: raspberrypi
|
||||||
become: true
|
|
||||||
|
|
||||||
- name: Setup k3s servers
|
- hosts: master
|
||||||
hosts: master
|
become: yes
|
||||||
roles:
|
roles:
|
||||||
- role: k3s_server
|
- role: k3s/master
|
||||||
become: true
|
|
||||||
|
|
||||||
- name: Setup k3s agents
|
- hosts: node
|
||||||
hosts: node
|
become: yes
|
||||||
roles:
|
roles:
|
||||||
- role: k3s_agent
|
- role: k3s/node
|
||||||
become: true
|
|
||||||
|
|
||||||
- name: Configure k3s cluster
|
- hosts: master
|
||||||
hosts: master
|
become: yes
|
||||||
roles:
|
roles:
|
||||||
- role: k3s_server_post
|
- role: k3s/post
|
||||||
become: true
|
|
||||||
|
|||||||
@@ -1,8 +0,0 @@
|
|||||||
# Kubeadm 1.15 needs /dev/kmsg to be there, but it's not in lxc, but we can just use /dev/console instead
|
|
||||||
# see: https://github.com/kubernetes-sigs/kind/issues/662
|
|
||||||
if [ ! -e /dev/kmsg ]; then
|
|
||||||
ln -s /dev/console /dev/kmsg
|
|
||||||
fi
|
|
||||||
|
|
||||||
# https://medium.com/@kvaps/run-kubernetes-in-lxc-container-f04aa94b6c9c
|
|
||||||
mount --make-rshared /
|
|
||||||
Reference in New Issue
Block a user