3 Commits

Author SHA1 Message Date
7f72ab53f4 fix: Incorrect indentation causing malformed PEM file
All checks were successful
continuous-integration/drone/push Build is passing
2023-07-07 10:18:06 +02:00
042d3b0eab fix: Upgrade K8s patch version
All checks were successful
continuous-integration/drone/push Build is passing
2023-07-06 13:44:05 +02:00
7482054904 fix: Refactor kustomize templates
Some checks failed
continuous-integration/drone/push Build is failing
2023-07-06 12:58:31 +02:00
40 changed files with 354 additions and 1039 deletions

View File

@ -10,11 +10,6 @@ volumes:
claim: claim:
name: flexvolsmb-drone-scratch name: flexvolsmb-drone-scratch
trigger:
event:
exclude:
- tag
steps: steps:
- name: Debugging information - name: Debugging information
image: bv11-cr01.bessems.eu/library/packer-extended image: bv11-cr01.bessems.eu/library/packer-extended
@ -26,6 +21,8 @@ steps:
- yamllint --version - yamllint --version
- name: Linting - name: Linting
depends_on:
- Debugging information
image: bv11-cr01.bessems.eu/library/packer-extended image: bv11-cr01.bessems.eu/library/packer-extended
pull: always pull: always
commands: commands:
@ -35,48 +32,9 @@ steps:
packer/preseed/UbuntuServer22.04/user-data \ packer/preseed/UbuntuServer22.04/user-data \
scripts scripts
- name: Semantic Release (Dry-run)
image: bv11-cr01.bessems.eu/proxy/library/node:20-slim
pull: always
commands:
- |
apt-get update
- |
apt-get install -y --no-install-recommends \
curl \
git-core \
jq \
ca-certificates
- |
curl -L https://api.github.com/repos/mikefarah/yq/releases/latest | \
jq -r '.assets[] | select(.name | endswith("yq_linux_amd64")) | .browser_download_url' | \
xargs -I {} curl -L -o /bin/yq {} && \
chmod +x /bin/yq
- |
npm install \
semantic-release \
@semantic-release/commit-analyzer \
@semantic-release/exec \
- |
export K8S_VERSION=$(yq '.components.clusterapi.workload.version.k8s' < ./ansible/vars/metacluster.yml)
export GIT_CREDENTIALS=$${GIT_USERNAME}:$${GIT_APIKEY}
- |
npx semantic-release \
--package @semantic-release/exec \
--package semantic-release \
--branches ${DRONE_BRANCH} \
--tag-format "K8s_$${K8S_VERSION}-v\$${version}" \
--dry-run \
--plugins @semantic-release/commit-analyzer,@semantic-release/exec \
--analyzeCommits @semantic-release/commit-analyzer \
--verifyRelease @semantic-release/exec \
--verifyReleaseCmd 'echo "$${nextRelease.version}" > .version'
environment:
GIT_APIKEY:
from_secret: git_apikey
GIT_USERNAME: djpbessems
- name: Install Ansible Galaxy collections - name: Install Ansible Galaxy collections
depends_on:
- Linting
image: bv11-cr01.bessems.eu/library/packer-extended image: bv11-cr01.bessems.eu/library/packer-extended
pull: always pull: always
commands: commands:
@ -84,8 +42,13 @@ steps:
ansible-galaxy collection install \ ansible-galaxy collection install \
-r ansible/requirements.yml \ -r ansible/requirements.yml \
-p ./ansible/collections -p ./ansible/collections
volumes:
- name: scratch
path: /scratch
- name: Kubernetes Bootstrap Appliance - name: Kubernetes Bootstrap Appliance
depends_on:
- Install Ansible Galaxy collections
image: bv11-cr01.bessems.eu/library/packer-extended image: bv11-cr01.bessems.eu/library/packer-extended
pull: always pull: always
commands: commands:
@ -94,7 +57,6 @@ steps:
packer/preseed/UbuntuServer22.04/user-data packer/preseed/UbuntuServer22.04/user-data
- | - |
export K8S_VERSION=$(yq '.components.clusterapi.workload.version.k8s' < ./ansible/vars/metacluster.yml) export K8S_VERSION=$(yq '.components.clusterapi.workload.version.k8s' < ./ansible/vars/metacluster.yml)
export APPLIANCE_VERSION=$(cat .version)
- | - |
packer init -upgrade \ packer init -upgrade \
./packer ./packer
@ -109,7 +71,6 @@ steps:
-var ssh_password=$${SSH_PASSWORD} \ -var ssh_password=$${SSH_PASSWORD} \
-var vsphere_password=$${VSPHERE_PASSWORD} \ -var vsphere_password=$${VSPHERE_PASSWORD} \
-var k8s_version=$K8S_VERSION \ -var k8s_version=$K8S_VERSION \
-var appliance_version=$APPLIANCE_VERSION \
./packer ./packer
- | - |
packer build \ packer build \
@ -123,7 +84,6 @@ steps:
-var ssh_password=$${SSH_PASSWORD} \ -var ssh_password=$${SSH_PASSWORD} \
-var vsphere_password=$${VSPHERE_PASSWORD} \ -var vsphere_password=$${VSPHERE_PASSWORD} \
-var k8s_version=$K8S_VERSION \ -var k8s_version=$K8S_VERSION \
-var appliance_version=$APPLIANCE_VERSION \
./packer ./packer
environment: environment:
DOCKER_USERNAME: DOCKER_USERNAME:
@ -146,6 +106,8 @@ steps:
path: /scratch path: /scratch
- name: Kubernetes Upgrade Appliance - name: Kubernetes Upgrade Appliance
depends_on:
- Install Ansible Galaxy collections
image: bv11-cr01.bessems.eu/library/packer-extended image: bv11-cr01.bessems.eu/library/packer-extended
pull: alwaysquery( pull: alwaysquery(
commands: commands:
@ -154,7 +116,6 @@ steps:
packer/preseed/UbuntuServer22.04/user-data packer/preseed/UbuntuServer22.04/user-data
- | - |
export K8S_VERSION=$(yq '.components.clusterapi.workload.version.k8s' < ./ansible/vars/metacluster.yml) export K8S_VERSION=$(yq '.components.clusterapi.workload.version.k8s' < ./ansible/vars/metacluster.yml)
export APPLIANCE_VERSION=$(cat .version)
- | - |
packer init -upgrade \ packer init -upgrade \
./packer ./packer
@ -169,7 +130,6 @@ steps:
-var ssh_password=$${SSH_PASSWORD} \ -var ssh_password=$${SSH_PASSWORD} \
-var vsphere_password=$${VSPHERE_PASSWORD} \ -var vsphere_password=$${VSPHERE_PASSWORD} \
-var k8s_version=$K8S_VERSION \ -var k8s_version=$K8S_VERSION \
-var appliance_version=$APPLIANCE_VERSION \
./packer ./packer
- | - |
packer build \ packer build \
@ -183,7 +143,6 @@ steps:
-var ssh_password=$${SSH_PASSWORD} \ -var ssh_password=$${SSH_PASSWORD} \
-var vsphere_password=$${VSPHERE_PASSWORD} \ -var vsphere_password=$${VSPHERE_PASSWORD} \
-var k8s_version=$K8S_VERSION \ -var k8s_version=$K8S_VERSION \
-var appliance_version=$APPLIANCE_VERSION \
./packer ./packer
environment: environment:
DOCKER_USERNAME: DOCKER_USERNAME:
@ -206,6 +165,9 @@ steps:
path: /scratch path: /scratch
- name: Remove temporary resources - name: Remove temporary resources
depends_on:
- Kubernetes Bootstrap Appliance
- Kubernetes Upgrade Appliance
image: bv11-cr01.bessems.eu/library/packer-extended image: bv11-cr01.bessems.eu/library/packer-extended
commands: commands:
- | - |

View File

@ -1,151 +0,0 @@
name: Container & Helm chart
on: [push]
jobs:
linting:
name: Linting
runs-on: dind-rootless
steps:
- name: Check out repository code
uses: actions/checkout@v4
- name: yamllint
uses: bewuethr/yamllint-action@v1
with:
config-file: .yamllint.yaml
semrel_dryrun:
name: Semantic Release (Dry-run)
runs-on: dind-rootless
outputs:
version: ${{ steps.sem_rel.outputs.version }}
steps:
- name: Check out repository code
uses: actions/checkout@v4
- name: Setup Node
uses: actions/setup-node@v3
with:
node-version: 20
- name: Install dependencies
run: |
npm install \
semantic-release \
@semantic-release/commit-analyzer \
@semantic-release/exec
- name: Semantic Release (dry-run)
id: sem_rel
run: |
npx semantic-release \
--package @semantic-release/exec \
--package semantic-release \
--branches ${{ gitea.refname }} \
--tag-format 'v${version}' \
--dry-run \
--plugins @semantic-release/commit-analyzer,@semantic-release/exec \
--analyzeCommits @semantic-release/commit-analyzer \
--verifyRelease @semantic-release/exec \
--verifyReleaseCmd 'echo "version=${nextRelease.version}" >> $GITHUB_OUTPUT'
env:
GIT_CREDENTIALS: ${{ secrets.GIT_USERNAME }}:${{ secrets.GIT_APIKEY }}
- name: Assert semantic release output
run: |
[[ -z "${{ steps.sem_rel.outputs.version }}" ]] && {
echo 'No release tag - exiting'; exit 1
} || {
echo 'Release tag set correctly: ${{ steps.sem_rel.outputs.version }}'; exit 0
}
build_image:
name: Kubernetes Bootstrap Appliance
container: code.spamasaurus.com/djpbessems/packer-extended:1.3.0
runs-on: dind-rootless
needs: [semrel_dryrun, linting]
steps:
- name: Checkout
uses: actions/checkout@v4
- name: Parse Kubernetes version
uses: mikefarah/yq@master
id: get_k8sversion
with:
cmd: yq '.components.clusterapi.workload.version.k8s' ansible/vars/metacluster.yml
- name: Set up packer
uses: hashicorp/setup-packer@main
id: setup
with:
version: "latest"
- name: Prepare build environment
id: init
run: |
packer init -upgrade ./packer
ansible-galaxy collection install \
-r ansible/requirements.yml \
-p ./ansible/collections
echo "BUILD_COMMIT=$(echo ${{ gitea.sha }} | cut -c 1-10)" >> $GITHUB_ENV
echo "BUILD_SUFFIX=$(openssl rand -hex 3)" >> $GITHUB_ENV
- name: Run `packer validate`
id: validate
run: |
# BUILD_COMMIT=$(echo "${{ gitea.sha }}" | cut -c 1-10)
# BUILD_SUFFIX=$(openssl rand -hex 3)
packer validate \
-only=vsphere-iso.bootstrap \
-var vm_name=${{ gitea.run_number }}-${BUILD_COMMIT}-${BUILD_SUFFIX} \
-var docker_username=${{ secrets.DOCKER_USERNAME }} \
-var docker_password=${{ secrets.DOCKER_PASSWORD }} \
-var repo_username=${{ secrets.REPO_USERNAME }} \
-var repo_password=${{ secrets.REPO_PASSWORD }} \
-var ssh_password=${{ secrets.SSH_PASSWORD }} \
-var hv_password=${{ secrets.HV_PASSWORD }} \
-var k8s_version=${{ steps.get_k8sversion.outputs.result }} \
-var appliance_version=${{ needs.semrel_dryrun.outputs.version }} \
./packer
- name: Run `packer build`
run: |
# BUILD_COMMIT=$(echo "${{ gitea.sha }}" | cut -c 1-10)
# BUILD_SUFFIX=$(openssl rand -hex 3)
packer build \
-only=vsphere-iso.bootstrap \
-var vm_name=${{ gitea.run_number }}-${BUILD_COMMIT}-${BUILD_SUFFIX} \
-var docker_username=${{ secrets.DOCKER_USERNAME }} \
-var docker_password=${{ secrets.DOCKER_PASSWORD }} \
-var repo_username=${{ secrets.REPO_USERNAME }} \
-var repo_password=${{ secrets.REPO_PASSWORD }} \
-var ssh_password=${{ secrets.SSH_PASSWORD }} \
-var hv_password=${{ secrets.HV_PASSWORD }} \
-var k8s_version=${{ steps.get_k8sversion.outputs.result }} \
-var appliance_version=${{ needs.semrel_dryrun.outputs.version }} \
./packer
# env:
# PACKER_LOG: 1
# semrel:
# name: Semantic Release
# runs-on: dind-rootless
# needs: [build_container, build_chart]
# steps:
# - name: Check out repository code
# uses: actions/checkout@v3
# - name: Setup Node
# uses: actions/setup-node@v3
# with:
# node-version: 20
# - name: Install dependencies
# run: |
# npm install \
# semantic-release \
# @semantic-release/changelog \
# @semantic-release/commit-analyzer \
# @semantic-release/git \
# @semantic-release/release-notes-generator
# - name: Semantic Release
# run: |
# npx semantic-release \
# --branches ${{ gitea.refname }} \
# --tag-format 'v${version}' \
# --plugins @semantic-release/commit-analyzer,@semantic-release/release-notes-generator,@semantic-release/changelog,@semantic-release/git
# env:
# GIT_CREDENTIALS: ${{ secrets.GIT_USERNAME }}:${{ secrets.GIT_APIKEY }}

View File

@ -1,4 +0,0 @@
extends: relaxed
rules:
line-length: disable

View File

@ -1,4 +1,4 @@
- name: Parse Cluster-API manifests for container images - name: Parse manifests for container images
ansible.builtin.shell: ansible.builtin.shell:
# This set of commands is necessary to deal with multi-line scalar values # This set of commands is necessary to deal with multi-line scalar values
# eg.: # eg.:
@ -9,17 +9,11 @@
cat {{ item.dest }} | yq --no-doc eval '.. | .image? | select(.)' | awk '!/ /'; cat {{ item.dest }} | yq --no-doc eval '.. | .image? | select(.)' | awk '!/ /';
cat {{ item.dest }} | yq eval '.data.data' | yq --no-doc eval '.. | .image? | select(.)'; cat {{ item.dest }} | yq eval '.data.data' | yq --no-doc eval '.. | .image? | select(.)';
cat {{ item.dest }} | yq --no-doc eval '.. | .files? | with_entries(select(.value.path == "*.yaml")).[0].content' | awk '!/null/' | yq eval '.. | .image? | select(.)' cat {{ item.dest }} | yq --no-doc eval '.. | .files? | with_entries(select(.value.path == "*.yaml")).[0].content' | awk '!/null/' | yq eval '.. | .image? | select(.)'
register: clusterapi_parsedmanifests register: parsedmanifests
loop: "{{ clusterapi_manifests.results }}" loop: "{{ clusterapi_manifests.results }}"
loop_control: loop_control:
label: "{{ item.dest | basename }}" label: "{{ item.dest | basename }}"
- name: Parse pinniped manifest for container images
ansible.builtin.shell:
cmd: >-
cat {{ pinniped_manifest.dest }} | yq --no-doc eval '.. | .image? | select(.)' | awk '!/ /';
register: pinniped_parsedmanifest
- name: Parse metacluster helm charts for container images - name: Parse metacluster helm charts for container images
ansible.builtin.shell: ansible.builtin.shell:
cmd: "{{ item.value.helm.parse_logic }}" cmd: "{{ item.value.helm.parse_logic }}"
@ -47,10 +41,8 @@
results: "{{ (chartimages_metacluster | json_query('results[*].stdout_lines')) + (chartimages_workloadcluster | json_query('results[*].stdout_lines')) | select() | flatten | list }}" results: "{{ (chartimages_metacluster | json_query('results[*].stdout_lines')) + (chartimages_workloadcluster | json_query('results[*].stdout_lines')) | select() | flatten | list }}"
- source: kubeadm - source: kubeadm
results: "{{ kubeadmimages.stdout_lines }}" results: "{{ kubeadmimages.stdout_lines }}"
- source: clusterapi - source: manifests
results: "{{ clusterapi_parsedmanifests | json_query('results[*].stdout_lines') | select() | flatten | list }}" results: "{{ parsedmanifests | json_query('results[*].stdout_lines') | select() | flatten | list }}"
- source: pinniped
results: "{{ pinniped_parsedmanifest.stdout_lines }}"
loop_control: loop_control:
label: "{{ item.source }}" label: "{{ item.source }}"
@ -72,4 +64,4 @@
docker://{{ item }} \ docker://{{ item }} \
docker-archive:./{{ ( item | regex_findall('[^/:]+'))[-2] }}_{{ lookup('ansible.builtin.password', '/dev/null length=5 chars=ascii_lowercase,digits seed={{ item }}') }}.tar:{{ item }} docker-archive:./{{ ( item | regex_findall('[^/:]+'))[-2] }}_{{ lookup('ansible.builtin.password', '/dev/null length=5 chars=ascii_lowercase,digits seed={{ item }}') }}.tar:{{ item }}
chdir: /opt/metacluster/container-images chdir: /opt/metacluster/container-images
loop: "{{ (containerimages_charts + containerimages_kubeadm + containerimages_clusterapi + containerimages_pinniped + dependencies.container_images) | flatten | unique | sort }}" loop: "{{ (containerimages_charts + containerimages_kubeadm + containerimages_manifests + dependencies.container_images) | flatten | unique | sort }}"

View File

@ -16,7 +16,8 @@
- /opt/metacluster/helm-charts - /opt/metacluster/helm-charts
- /opt/metacluster/k3s - /opt/metacluster/k3s
- /opt/metacluster/kube-vip - /opt/metacluster/kube-vip
- /opt/metacluster/pinniped - /opt/workloadcluster/git-repositories/gitops/charts
- /opt/workloadcluster/git-repositories/gitops/values
- /opt/workloadcluster/helm-charts - /opt/workloadcluster/helm-charts
- /opt/workloadcluster/node-templates - /opt/workloadcluster/node-templates
- /var/lib/rancher/k3s/agent/images - /var/lib/rancher/k3s/agent/images

View File

@ -1,6 +1,6 @@
- block: - block:
- name: Aggregate meta-cluster chart_values into dict - name: Aggregate chart_values into dict
ansible.builtin.set_fact: ansible.builtin.set_fact:
metacluster_chartvalues: "{{ metacluster_chartvalues | default({}) | combine({ item.key: { 'chart_values': (item.value.helm.chart_values | from_yaml) } }) }}" metacluster_chartvalues: "{{ metacluster_chartvalues | default({}) | combine({ item.key: { 'chart_values': (item.value.helm.chart_values | from_yaml) } }) }}"
when: item.value.helm.chart_values is defined when: item.value.helm.chart_values is defined
@ -8,34 +8,22 @@
loop_control: loop_control:
label: "{{ item.key }}" label: "{{ item.key }}"
- name: Combine and write dict to vars_file - name: Write dict to vars_file
ansible.builtin.copy: ansible.builtin.copy:
dest: /opt/firstboot/ansible/vars/metacluster.yml dest: /opt/firstboot/ansible/vars/metacluster.yml
content: >- content: >-
{{ {{
{ 'components': ( { 'components': (
metacluster_chartvalues | metacluster_chartvalues |
combine({ 'clusterapi' : components['clusterapi'] }) | combine({ 'clusterapi': components.clusterapi }) |
combine({ 'kubevip' : components['kubevip'] }) | combine({ 'kubevip' : components.kubevip }) )
combine({ 'localuserauthenticator': components['pinniped']['local-user-authenticator'] })),
'appliance': {
'version': (applianceversion)
}
} | to_nice_yaml(indent=2, width=4096) } | to_nice_yaml(indent=2, width=4096)
}} }}
- name: Aggregate workload-cluster chart_values into dict - name: Aggregate chart_values into dict
ansible.builtin.set_fact: ansible.builtin.set_fact:
workloadcluster_chartvalues: | workloadcluster_chartvalues: "{{ workloadcluster_chartvalues | default({}) | combine({ item.key: { 'chart_values': (item.value.chart_values | default('') | from_yaml) } }) }}"
{{ # when: item.value.chart_values is defined
workloadcluster_chartvalues | default({}) | combine({
item.key: {
'chart_values': (item.value.chart_values | default('') | from_yaml),
'extra_manifests': (item.value.extra_manifests | default([])),
'namespace': (item.value.namespace)
}
})
}}
loop: "{{ query('ansible.builtin.dict', downstream.helm_charts) }}" loop: "{{ query('ansible.builtin.dict', downstream.helm_charts) }}"
loop_control: loop_control:
label: "{{ item.key }}" label: "{{ item.key }}"
@ -49,7 +37,7 @@
} | to_nice_yaml(indent=2, width=4096) } | to_nice_yaml(indent=2, width=4096)
}} }}
- name: Download Cluster-API manifests - name: Download ClusterAPI manifests
ansible.builtin.get_url: ansible.builtin.get_url:
url: "{{ item.url }}" url: "{{ item.url }}"
dest: /opt/metacluster/cluster-api/{{ item.dest }} dest: /opt/metacluster/cluster-api/{{ item.dest }}
@ -109,22 +97,6 @@
delay: 5 delay: 5
until: kubevip_manifest is not failed until: kubevip_manifest is not failed
- name: Download pinniped local-user-authenticator manifest
ansible.builtin.get_url:
url: https://get.pinniped.dev/{{ components.pinniped['local-user-authenticator'].version }}/install-local-user-authenticator.yaml
dest: /opt/metacluster/pinniped/local-user-authenticator.yaml
register: pinniped_manifest
retries: 5
delay: 5
until: pinniped_manifest is not failed
- name: Trim image hash from manifest
ansible.builtin.replace:
path: /opt/metacluster/pinniped/local-user-authenticator.yaml
regexp: '([ ]*image: .*)@.*'
replace: '\1'
no_log: true
# - name: Inject manifests # - name: Inject manifests
# ansible.builtin.template: # ansible.builtin.template:
# src: "{{ item.type }}.j2" # src: "{{ item.type }}.j2"

View File

@ -2,9 +2,6 @@
- hosts: 127.0.0.1 - hosts: 127.0.0.1
connection: local connection: local
gather_facts: true gather_facts: true
vars:
# Needed by some templating in various tasks
_newline: "\n"
vars_files: vars_files:
- defaults.yml - defaults.yml
- metacluster.yml - metacluster.yml

View File

@ -1,176 +0,0 @@
- block:
- name: Install dex
kubernetes.core.helm:
name: dex
chart_ref: /opt/metacluster/helm-charts/dex
release_namespace: dex
create_namespace: true
wait: false
kubeconfig: "{{ kubeconfig.path }}"
values: "{{ components['dex'].chart_values }}"
- block:
- name: Install pinniped local-user-authenticator
kubernetes.core.k8s:
src: /opt/metacluster/pinniped/local-user-authenticator.yaml
state: present
kubeconfig: "{{ kubeconfig.path }}"
- name: Create local-user-authenticator accounts
kubernetes.core.k8s:
template: secret.j2
state: present
kubeconfig: "{{ kubeconfig.path }}"
vars:
_template:
name: "{{ item.username }}"
namespace: local-user-authenticator
type: ''
data:
- key: groups
value: "{{ 'group1,group2' | b64encode }}"
- key: passwordHash
value: "{{ item.password | b64encode }}"
loop: "{{ components['localuserauthenticator'].users }}"
- block:
- name: Install pinniped chart
kubernetes.core.helm:
name: pinniped
chart_ref: /opt/metacluster/helm-charts/pinniped
release_namespace: pinniped-supervisor
create_namespace: true
wait: false
kubeconfig: "{{ kubeconfig.path }}"
values: "{{ components['pinniped'].chart_values }}"
- name: Add ingress for supervisor
kubernetes.core.k8s:
template: "{{ item.kind }}.j2"
state: present
kubeconfig: "{{ kubeconfig.path }}"
vars:
_template:
name: "{{ item.name }}"
namespace: "{{ item.namespace }}"
spec: "{{ item.spec }}"
loop:
- kind: ingressroute
name: pinniped-supervisor
namespace: pinniped-supervisor
spec: |2
entryPoints:
- web
- websecure
routes:
- kind: Rule
match: Host(`auth.{{ vapp['metacluster.fqdn'] }}`)
services:
- kind: Service
name: pinniped-supervisor
namespace: pinniped-supervisor
port: 443
scheme: https
serversTransport: pinniped-supervisor
- kind: serverstransport
name: pinniped-supervisor
namespace: pinniped-supervisor
spec: |2
insecureSkipVerify: true
serverName: auth.{{ vapp['metacluster.fqdn'] }}
loop_control:
label: "{{ item.kind ~ '/' ~ item.name ~ ' (' ~ item.namespace ~ ')' }}"
- name: Ensure pinniped API availability
ansible.builtin.uri:
url: https://auth.{{ vapp['metacluster.fqdn'] }}/healthz
method: GET
register: api_readycheck
until:
- api_readycheck.status == 200
- api_readycheck.msg is search("OK")
retries: "{{ playbook.retries }}"
delay: "{{ ((storage_benchmark | float) * playbook.delay.short) | int }}"
# TODO: Migrate to step-ca
- name: Initialize tempfile
ansible.builtin.tempfile:
state: directory
register: certificate
- name: Create private key (RSA, 4096 bits)
community.crypto.openssl_privatekey:
path: "{{ certificate.path }}/certificate.key"
- name: Create self-signed certificate
community.crypto.x509_certificate:
path: "{{ certificate.path }}/certificate.crt"
privatekey_path: "{{ certificate.path }}/certificate.key"
provider: selfsigned
- name: Store self-signed certificate for use by pinniped supervisor
kubernetes.core.k8s:
template: secret.j2
state: present
kubeconfig: "{{ kubeconfig.path }}"
vars:
_template:
name: pinniped-supervisor-tls
namespace: pinniped-supervisor
type: kubernetes.io/tls
data:
- key: tls.crt
value: "{{ lookup('ansible.builtin.file', certificate.path ~ '/certificate.crt') | b64encode }}"
- key: tls.key
value: "{{ lookup('ansible.builtin.file', certificate.path ~ '/certificate.key') | b64encode }}"
# TODO: Migrate to step-ca
- name: Create pinniped resources
kubernetes.core.k8s:
template: "{{ item.kind }}.j2"
state: present
kubeconfig: "{{ kubeconfig.path }}"
vars:
_template:
name: "{{ item.name }}"
namespace: "{{ item.namespace }}"
type: "{{ item.type | default('') }}"
data: "{{ item.data | default(omit) }}"
spec: "{{ item.spec | default(omit) }}"
loop:
- kind: oidcidentityprovider
name: dex-staticpasswords
namespace: pinniped-supervisor
spec: |2
issuer: https://idps.{{ vapp['metacluster.fqdn'] }}
tls:
certificateAuthorityData: "{{ (stepca_cm_certs.resources[0].data['intermediate_ca.crt'] ~ _newline ~ stepca_cm_certs.resources[0].data['root_ca.crt']) | b64encode }}"
authorizationConfig:
additionalScopes: [offline_access, groups, email]
allowPasswordGrant: false
claims:
username: email
groups: groups
client:
secretName: dex-clientcredentials
- kind: secret
name: dex-clientcredentials
namespace: pinniped-supervisor
type: secrets.pinniped.dev/oidc-client
data:
- key: clientID
value: "{{ 'pinniped-supervisor' | b64encode }}"
- key: clientSecret
value: "{{ lookup('ansible.builtin.password', '/dev/null length=64 chars=ascii_lowercase,digits seed=' ~ vapp['metacluster.fqdn']) | b64encode }}"
- kind: federationdomain
name: metacluster-sso
namespace: pinniped-supervisor
spec: |2
issuer: https://auth.{{ vapp['metacluster.fqdn'] }}/sso
tls:
secretName: pinniped-supervisor-tls
loop_control:
label: "{{ item.kind ~ '/' ~ item.name }}"

View File

@ -1,9 +1,14 @@
- block: - block:
- name: Inject password into values file - name: Initialize tempfile
ansible.builtin.tempfile:
state: file
register: values_file
- name: Write chart values w/ password to tempfile
ansible.builtin.copy: ansible.builtin.copy:
dest: "{{ stepconfig.path }}" dest: "{{ values_file.path }}"
content: "{{ lookup('ansible.builtin.file', stepconfig.path) | regex_replace('(ca_password|provisioner_password):[ ]?\n', '\\1: ' ~ (vapp['metacluster.password'] | b64encode) ~ '\n') }}" content: "{{ stepca_values.stdout | regex_replace('(ca_password|provisioner_password): ', '\\1: ' ~ (vapp['metacluster.password'] | b64encode)) }}"
no_log: true no_log: true
- name: Install step-ca chart - name: Install step-ca chart
@ -16,7 +21,13 @@
wait: true wait: true
kubeconfig: "{{ kubeconfig.path }}" kubeconfig: "{{ kubeconfig.path }}"
values_files: values_files:
- "{{ stepconfig.path }}" - "{{ values_file.path }}"
- name: Cleanup tempfile
ansible.builtin.file:
path: "{{ values_file.path }}"
state: absent
when: values_file.path is defined
- name: Retrieve configmap w/ root certificate - name: Retrieve configmap w/ root certificate
kubernetes.core.k8s_info: kubernetes.core.k8s_info:
@ -34,7 +45,6 @@
kubeconfig: "{{ kubeconfig.path }}" kubeconfig: "{{ kubeconfig.path }}"
loop: loop:
- argo-cd - argo-cd
- gitea
# - kube-system # - kube-system
- name: Store root certificate in namespaced configmaps/secrets - name: Store root certificate in namespaced configmaps/secrets
@ -48,7 +58,6 @@
namespace: "{{ item.namespace }}" namespace: "{{ item.namespace }}"
annotations: "{{ item.annotations | default('{}') | indent(width=4, first=True) }}" annotations: "{{ item.annotations | default('{}') | indent(width=4, first=True) }}"
labels: "{{ item.labels | default('{}') | indent(width=4, first=True) }}" labels: "{{ item.labels | default('{}') | indent(width=4, first=True) }}"
type: "{{ item.type | default('') }}"
data: "{{ item.data }}" data: "{{ item.data }}"
loop: loop:
- name: argocd-tls-certs-cm - name: argocd-tls-certs-cm
@ -64,12 +73,6 @@
data: data:
- key: git.{{ vapp['metacluster.fqdn'] }} - key: git.{{ vapp['metacluster.fqdn'] }}
value: "{{ stepca_cm_certs.resources[0].data['root_ca.crt'] }}" value: "{{ stepca_cm_certs.resources[0].data['root_ca.crt'] }}"
- name: step-certificates-certs
namespace: gitea
kind: secret
data:
- key: ca_chain.crt
value: "{{ (stepca_cm_certs.resources[0].data['intermediate_ca.crt'] ~ _newline ~ stepca_cm_certs.resources[0].data['root_ca.crt']) | b64encode }}"
- name: step-certificates-certs - name: step-certificates-certs
namespace: kube-system namespace: kube-system
kind: secret kind: secret
@ -90,7 +93,7 @@
_template: _template:
name: step-ca name: step-ca
namespace: step-ca namespace: step-ca
spec: |2 config: |2
entryPoints: entryPoints:
- websecure - websecure
routes: routes:

View File

@ -32,7 +32,7 @@
_template: _template:
name: gitea-ssh name: gitea-ssh
namespace: gitea namespace: gitea
spec: |2 config: |2
entryPoints: entryPoints:
- ssh - ssh
routes: routes:
@ -55,7 +55,6 @@
force_basic_auth: yes force_basic_auth: yes
body: body:
name: token_init_{{ lookup('password', '/dev/null length=5 chars=ascii_letters,digits') }} name: token_init_{{ lookup('password', '/dev/null length=5 chars=ascii_letters,digits') }}
scopes: ["write:user","write:organization"]
register: gitea_api_token register: gitea_api_token
- name: Retrieve existing gitea configuration - name: Retrieve existing gitea configuration
@ -111,8 +110,8 @@
- organization: mc - organization: mc
body: body:
name: GitOps.ClusterAPI name: GitOps.ClusterAPI
auto_init: true # auto_init: true
default_branch: main # default_branch: main
description: ClusterAPI manifests description: ClusterAPI manifests
- organization: mc - organization: mc
body: body:
@ -123,15 +122,15 @@
- organization: wl - organization: wl
body: body:
name: GitOps.Config name: GitOps.Config
auto_init: true # auto_init: true
default_branch: main # default_branch: main
description: GitOps manifests description: GitOps manifests
- organization: wl - organization: wl
body: body:
name: ClusterAccess.Store name: GitOps.HelmCharts
auto_init: true # auto_init: true
default_branch: main # default_branch: main
description: Kubeconfig files description: Helm charts
loop_control: loop_control:
label: "{{ item.organization ~ '/' ~ item.body.name }}" label: "{{ item.organization ~ '/' ~ item.body.name }}"

View File

@ -27,7 +27,7 @@
_template: _template:
name: traefik-dashboard name: traefik-dashboard
namespace: kube-system namespace: kube-system
spec: |2 config: |2
entryPoints: entryPoints:
- web - web
- websecure - websecure

View File

@ -12,15 +12,6 @@
- registry - registry
- storage - storage
- name: Create step-ca config dictionary
ansible.builtin.set_fact:
stepconfig: "{{ { 'path': ansible_env.HOME ~ '/.step/config/values.yaml' } }}"
- name: Create step-ca target folder
ansible.builtin.file:
path: "{{ stepconfig.path | dirname }}"
state: directory
- name: Initialize tempfile - name: Initialize tempfile
ansible.builtin.tempfile: ansible.builtin.tempfile:
state: file state: file
@ -45,8 +36,8 @@
--address=:9000 \ --address=:9000 \
--provisioner=admin \ --provisioner=admin \
--acme \ --acme \
--password-file={{ stepca_password.path }} | tee {{ stepconfig.path }} --password-file={{ stepca_password.path }}
creates: "{{ stepconfig.path }}" register: stepca_values
- name: Cleanup tempfile - name: Cleanup tempfile
ansible.builtin.file: ansible.builtin.file:
@ -57,20 +48,12 @@
- name: Store root CA certificate - name: Store root CA certificate
ansible.builtin.copy: ansible.builtin.copy:
dest: /usr/local/share/ca-certificates/root_ca.crt dest: /usr/local/share/ca-certificates/root_ca.crt
content: "{{ (lookup('ansible.builtin.file', stepconfig.path) | from_yaml).inject.certificates.root_ca }}" content: "{{ (stepca_values.stdout | from_yaml).inject.certificates.root_ca }}"
- name: Update certificate truststore - name: Update certificate truststore
ansible.builtin.command: ansible.builtin.command:
cmd: update-ca-certificates cmd: update-ca-certificates
- name: Extract container images (for idempotency purposes)
ansible.builtin.unarchive:
src: /opt/metacluster/container-images/image-tarballs.tgz
dest: /opt/metacluster/container-images
remote_src: no
when:
- lookup('ansible.builtin.fileglob', '/opt/metacluster/container-images/*.tgz') is match('.*image-tarballs.tgz')
- name: Get all stored fully qualified container image names - name: Get all stored fully qualified container image names
ansible.builtin.shell: ansible.builtin.shell:
cmd: >- cmd: >-

View File

@ -42,30 +42,19 @@
retries: "{{ playbook.retries }}" retries: "{{ playbook.retries }}"
delay: "{{ (storage_benchmark | int) * (playbook.delay.medium | int) }}" delay: "{{ (storage_benchmark | int) * (playbook.delay.medium | int) }}"
- name: Install tab-completion - name: Install kubectl tab-completion
ansible.builtin.shell: ansible.builtin.shell:
cmd: |- cmd: kubectl completion bash | tee /etc/bash_completion.d/kubectl
{{ item }} completion bash > /etc/bash_completion.d/{{ item }}
creates: /etc/bash_completion.d/{{ item }}
loop:
- kubectl
- helm
- step
- name: Create kubeconfig dictionary - name: Initialize tempfile
ansible.builtin.set_fact: ansible.builtin.tempfile:
kubeconfig: "{{ { 'path': ansible_env.HOME ~ '/.kube/config' } }}" state: file
register: kubeconfig
- name: Create kubeconfig target folder
ansible.builtin.file:
path: "{{ kubeconfig.path | dirname }}"
state: directory
- name: Retrieve kubeconfig - name: Retrieve kubeconfig
ansible.builtin.command: ansible.builtin.command:
cmd: kubectl config view --raw cmd: kubectl config view --raw
register: kubectl_config register: kubectl_config
no_log: true
- name: Store kubeconfig in tempfile - name: Store kubeconfig in tempfile
ansible.builtin.copy: ansible.builtin.copy:

View File

@ -1,12 +1,10 @@
- import_tasks: init.yml - import_tasks: init.yml
- import_tasks: k3s.yml - import_tasks: k3s.yml
- import_tasks: assets.yml - import_tasks: assets.yml
- import_tasks: virtualip.yml - import_tasks: kube-vip.yml
- import_tasks: metadata.yml
- import_tasks: storage.yml - import_tasks: storage.yml
- import_tasks: ingress.yml - import_tasks: ingress.yml
- import_tasks: certauthority.yml - import_tasks: certauthority.yml
- import_tasks: registry.yml - import_tasks: registry.yml
- import_tasks: git.yml - import_tasks: git.yml
- import_tasks: gitops.yml - import_tasks: gitops.yml
- import_tasks: authentication.yml

View File

@ -1,57 +0,0 @@
- block:
- name: Aggregate manifest-component versions into dictionary
ansible.builtin.set_fact:
manifest_versions: "{{ manifest_versions | default([]) + [ item | combine( {'type': 'manifest', 'id': index } ) ] }}"
loop:
- name: cluster-api
versions:
management:
base: "{{ components.clusterapi.management.version.base }}"
cert_manager: "{{ components.clusterapi.management.version.cert_manager }}"
infrastructure_vsphere: "{{ components.clusterapi.management.version.infrastructure_vsphere }}"
ipam_incluster: "{{ components.clusterapi.management.version.ipam_incluster }}"
cpi_vsphere: "{{ components.clusterapi.management.version.cpi_vsphere }}"
workload:
calico: "{{ components.clusterapi.workload.version.calico }}"
k8s: "{{ components.clusterapi.workload.version.k8s }}"
- name: kube-vip
version: "{{ components.kubevip.version }}"
loop_control:
label: "{{ item.name }}"
index_var: index
- name: Install json-server chart
kubernetes.core.helm:
name: json-server
chart_ref: /opt/metacluster/helm-charts/json-server
release_namespace: json-server
create_namespace: true
wait: false
kubeconfig: "{{ kubeconfig.path }}"
values: |
{{
components['json-server'].chart_values |
combine(
{ 'jsonServer': { 'seedData': { 'configInline': (
{ 'appliance': { "version": appliance.version }, 'components': manifest_versions, 'healthz': { 'status': 'running' } }
) | to_json } } }
)
}}
- name: Ensure json-server API availability
ansible.builtin.uri:
url: https://version.{{ vapp['metacluster.fqdn'] }}/healthz
method: GET
# This mock REST API -ironically- does not support json encoded body argument
body_format: raw
register: api_readycheck
until:
- api_readycheck.json.status is defined
- api_readycheck.json.status == 'running'
retries: "{{ playbook.retries }}"
delay: "{{ (storage_benchmark | int) * (playbook.delay.long | int) }}"
module_defaults:
ansible.builtin.uri:
validate_certs: no
status_code: [200, 201]

View File

@ -1,40 +0,0 @@
- name: Initialize tempfolder
ansible.builtin.tempfile:
state: directory
register: pinniped_kubeconfig
- name: Pull existing repository
ansible.builtin.git:
repo: https://git.{{ vapp['metacluster.fqdn'] }}/wl/ClusterAccess.Store.git
dest: "{{ pinniped_kubeconfig.path }}"
version: main
- name: Generate kubeconfig
ansible.builtin.shell:
cmd: pinniped get kubeconfig --kubeconfig {{ capi_kubeconfig.path }}
register: pinniped_config
until:
- pinniped_config is not failed
retries: "{{ playbook.retries }}"
delay: "{{ ((storage_benchmark | float) * playbook.delay.short) | int }}"
- name: Store kubeconfig in tempfile
ansible.builtin.copy:
dest: "{{ pinniped_kubeconfig.path }}/kubeconfig"
content: "{{ pinniped_config.stdout }}"
mode: 0600
no_log: true
- name: Push git repository
lvrfrc87.git_acp.git_acp:
path: "{{ pinniped_kubeconfig.path }}"
branch: main
comment: "Upload kubeconfig files"
add:
- .
url: https://administrator:{{ vapp['metacluster.password'] | urlencode }}@git.{{ vapp['metacluster.fqdn'] }}/wl/ClusterAccess.Store.git
environment:
GIT_AUTHOR_NAME: administrator
GIT_AUTHOR_EMAIL: administrator@{{ vapp['metacluster.fqdn'] }}
GIT_COMMITTER_NAME: administrator
GIT_COMMITTER_EMAIL: administrator@{{ vapp['metacluster.fqdn'] }}

View File

@ -85,40 +85,6 @@
--kubeconfig {{ kubeconfig.path }} --kubeconfig {{ kubeconfig.path }}
chdir: /opt/metacluster/cluster-api chdir: /opt/metacluster/cluster-api
- name: Initialize tempfolder
ansible.builtin.tempfile:
state: directory
register: capi_clustermanifest
- name: Pull existing repository
ansible.builtin.git:
repo: https://git.{{ vapp['metacluster.fqdn'] }}/mc/GitOps.ClusterAPI.git
dest: "{{ capi_clustermanifest.path }}"
version: main
- name: Generate Cluster API provider manifests
ansible.builtin.shell:
cmd: >-
clusterctl generate provider \
-v5 \
--{{ item.type }} {{ item.name }}:{{ item.version }} \
--config ./clusterctl.yaml > {{ capi_clustermanifest.path }}/provider-{{ item.name }}.yaml
chdir: /opt/metacluster/cluster-api
loop:
- type: infrastructure
name: vsphere
version: "{{ components.clusterapi.management.version.infrastructure_vsphere }}"
- type: ipam
name: in-cluster
version: "{{ components.clusterapi.management.version.ipam_incluster }}"
- name: Split cluster API provider manifests into separate files
ansible.builtin.shell:
cmd: >-
awk 'BEGINFILE {print "---"}{print}' {{ capi_clustermanifest.path }}/provider-*.yaml |
kubectl slice \
-o {{ capi_clustermanifest.path }}/providers
- name: Ensure controller availability - name: Ensure controller availability
kubernetes.core.k8s_info: kubernetes.core.k8s_info:
kind: Deployment kind: Deployment
@ -158,17 +124,22 @@
chdir: /opt/metacluster/cluster-api chdir: /opt/metacluster/cluster-api
register: clusterctl_newcluster register: clusterctl_newcluster
- name: Initialize tempfolder
ansible.builtin.tempfile:
state: directory
register: capi_clustermanifest
- name: Save workload cluster manifest - name: Save workload cluster manifest
ansible.builtin.copy: ansible.builtin.copy:
dest: "{{ capi_clustermanifest.path }}/new-cluster.yaml" dest: "{{ capi_clustermanifest.path }}/new-cluster.yaml"
content: "{{ clusterctl_newcluster.stdout }}" content: "{{ clusterctl_newcluster.stdout }}"
- name: Split workload cluster manifest into separate files - name: Split manifest into separate files
ansible.builtin.shell: ansible.builtin.shell:
cmd: >- cmd: >-
kubectl slice \ kubectl slice \
-f {{ capi_clustermanifest.path }}/new-cluster.yaml \ -f {{ capi_clustermanifest.path }}/new-cluster.yaml \
-o {{ capi_clustermanifest.path }}/downstream-cluster -o {{ capi_clustermanifest.path }}/manifests
- name: Generate nodepool kustomization manifest - name: Generate nodepool kustomization manifest
ansible.builtin.template: ansible.builtin.template:
@ -184,20 +155,13 @@
- name: Store nodepool manifest - name: Store nodepool manifest
ansible.builtin.copy: ansible.builtin.copy:
dest: "{{ capi_clustermanifest.path }}/nodepool-worker-storage.yaml" dest: "{{ capi_clustermanifest.path }}/manifests/nodepool-worker-storage.yaml"
content: "{{ lookup('kubernetes.core.kustomize', dir=capi_clustermanifest.path) }}" content: "{{ lookup('kubernetes.core.kustomize', dir=capi_clustermanifest.path) }}"
- name: Split nodepool manifest into separate files
ansible.builtin.shell:
cmd: >-
kubectl slice \
-f {{ capi_clustermanifest.path }}/nodepool-worker-storage.yaml \
-o {{ capi_clustermanifest.path }}/downstream-cluster
- name: Create in-cluster IpPool - name: Create in-cluster IpPool
ansible.builtin.template: ansible.builtin.template:
src: ippool.j2 src: ippool.j2
dest: "{{ capi_clustermanifest.path }}/downstream-cluster/inclusterippool-{{ _template.cluster.name }}.yml" dest: "{{ capi_clustermanifest.path }}/manifests/inclusterippool-{{ _template.cluster.name }}.yml"
vars: vars:
_template: _template:
cluster: cluster:
@ -209,27 +173,24 @@
prefix: "{{ vapp['guestinfo.prefixlength'] }}" prefix: "{{ vapp['guestinfo.prefixlength'] }}"
gateway: "{{ vapp['guestinfo.gateway'] }}" gateway: "{{ vapp['guestinfo.gateway'] }}"
- name: Push git repository - name: Initialize/Push git repository
lvrfrc87.git_acp.git_acp: ansible.builtin.shell:
path: "{{ capi_clustermanifest.path }}" cmd: |
branch: main git init
comment: "Upload manifests" git config --global user.email "administrator@{{ vapp['metacluster.fqdn'] }}"
add: git config --global user.name "administrator"
- ./downstream-cluster git checkout -b main
- ./providers git add ./manifests
clean: untracked git commit -m "Upload manifests"
url: https://administrator:{{ vapp['metacluster.password'] | urlencode }}@git.{{ vapp['metacluster.fqdn'] }}/mc/GitOps.ClusterAPI.git git remote add origin https://git.{{ vapp['metacluster.fqdn'] }}/mc/GitOps.ClusterAPI.git
environment: git push https://administrator:{{ vapp['metacluster.password'] | urlencode }}@git.{{ vapp['metacluster.fqdn'] }}/mc/GitOps.ClusterAPI.git --all
GIT_AUTHOR_NAME: administrator chdir: "{{ capi_clustermanifest.path }}"
GIT_AUTHOR_EMAIL: administrator@{{ vapp['metacluster.fqdn'] }}
GIT_COMMITTER_NAME: administrator
GIT_COMMITTER_EMAIL: administrator@{{ vapp['metacluster.fqdn'] }}
# - name: Cleanup tempfolder - name: Cleanup tempfolder
# ansible.builtin.file: ansible.builtin.file:
# path: "{{ capi_clustermanifest.path }}" path: "{{ capi_clustermanifest.path }}"
# state: absent state: absent
# when: capi_clustermanifest.path is defined when: capi_clustermanifest.path is defined
- name: Configure Cluster API repository - name: Configure Cluster API repository
ansible.builtin.template: ansible.builtin.template:
@ -274,7 +235,7 @@
namespace: default namespace: default
repository: repository:
url: https://git.{{ vapp['metacluster.fqdn'] }}/mc/GitOps.ClusterAPI.git url: https://git.{{ vapp['metacluster.fqdn'] }}/mc/GitOps.ClusterAPI.git
path: downstream-cluster path: manifests
revision: main revision: main
notify: notify:
- Apply manifests - Apply manifests
@ -316,12 +277,7 @@
# TODO: move to git repo # TODO: move to git repo
- name: Apply cni plugin manifest - name: Apply cni plugin manifest
kubernetes.core.k8s: kubernetes.core.k8s:
definition: | src: /opt/metacluster/cluster-api/cni-calico/{{ components.clusterapi.workload.version.calico }}/calico.yaml
{{
lookup('ansible.builtin.file', '/opt/metacluster/cluster-api/cni-calico/' ~ components.clusterapi.workload.version.calico ~ '/calico.yaml') |
regex_replace('# - name: CALICO_IPV4POOL_CIDR', '- name: CALICO_IPV4POOL_CIDR') |
regex_replace('# value: "192.168.0.0/16"', ' value: "172.30.0.0/16"')
}}
state: present state: present
wait: true wait: true
kubeconfig: "{{ capi_kubeconfig.path }}" kubeconfig: "{{ capi_kubeconfig.path }}"

View File

@ -5,20 +5,6 @@
recurse: false recurse: false
register: helm_charts register: helm_charts
- name: Pull existing repository
ansible.builtin.git:
repo: https://git.{{ vapp['metacluster.fqdn'] }}/wl/GitOps.Config.git
dest: /opt/workloadcluster/git-repositories/gitops
version: main
- name: Create folder structure within new git-repository
ansible.builtin.file:
path: "{{ item }}"
state: directory
loop:
- /opt/workloadcluster/git-repositories/gitops/charts
- /opt/workloadcluster/git-repositories/gitops/values
- name: Create hard-links to populate new git-repository - name: Create hard-links to populate new git-repository
ansible.builtin.shell: ansible.builtin.shell:
cmd: >- cmd: >-
@ -27,18 +13,6 @@
loop_control: loop_control:
label: "{{ item.path | basename }}" label: "{{ item.path | basename }}"
- name: Write custom manifests to respective chart templates store
ansible.builtin.template:
src: "{{ src }}"
dest: /opt/workloadcluster/git-repositories/gitops/charts/{{ manifest.value.namespace }}/{{ manifest.key }}/templates/{{ (src | split('.'))[0] ~ '-' ~ _template.name ~ '.yaml' }}
vars:
manifest: "{{ item.0 }}"
src: "{{ item.1.src }}"
_template: "{{ item.1._template }}"
loop: "{{ query('ansible.builtin.subelements', query('ansible.builtin.dict', downstream_components), 'value.extra_manifests') }}"
loop_control:
label: "{{ (src | split('.'))[0] ~ '-' ~ _template.name }}"
- name: Create subfolders - name: Create subfolders
ansible.builtin.file: ansible.builtin.file:
path: /opt/workloadcluster/git-repositories/gitops/values/{{ item.key }} path: /opt/workloadcluster/git-repositories/gitops/values/{{ item.key }}
@ -55,19 +29,18 @@
loop_control: loop_control:
label: "{{ item.key }}" label: "{{ item.key }}"
- name: Push git repository - name: Initialize/Push git repository
lvrfrc87.git_acp.git_acp: ansible.builtin.shell:
path: /opt/workloadcluster/git-repositories/gitops cmd: |
branch: main git init
comment: "Upload charts" git config --global user.email "administrator@{{ vapp['metacluster.fqdn'] }}"
add: git config --global user.name "administrator"
- . git checkout -b main
url: https://administrator:{{ vapp['metacluster.password'] | urlencode }}@git.{{ vapp['metacluster.fqdn'] }}/wl/GitOps.Config.git git add .
environment: git commit -m "Upload charts"
GIT_AUTHOR_NAME: administrator git remote add origin https://git.{{ vapp['metacluster.fqdn'] }}/wl/GitOps.Config.git
GIT_AUTHOR_EMAIL: administrator@{{ vapp['metacluster.fqdn'] }} git push https://administrator:{{ vapp['metacluster.password'] | urlencode }}@git.{{ vapp['metacluster.fqdn'] }}/wl/GitOps.Config.git --all
GIT_COMMITTER_NAME: administrator chdir: /opt/workloadcluster/git-repositories/gitops
GIT_COMMITTER_EMAIL: administrator@{{ vapp['metacluster.fqdn'] }}
- name: Retrieve workload-cluster kubeconfig - name: Retrieve workload-cluster kubeconfig
kubernetes.core.k8s_info: kubernetes.core.k8s_info:

View File

@ -6,7 +6,6 @@
- import_tasks: clusterapi.yml - import_tasks: clusterapi.yml
- import_tasks: gitops.yml - import_tasks: gitops.yml
- import_tasks: authentication.yml
when: when:
- vapp['deployment.type'] != 'core' - vapp['deployment.type'] != 'core'

View File

@ -1,7 +0,0 @@
apiVersion: config.supervisor.pinniped.dev/v1alpha1
kind: FederationDomain
metadata:
name: {{ _template.name }}
namespace: {{ _template.namespace }}
spec:
{{ _template.spec }}

View File

@ -4,4 +4,4 @@ metadata:
name: {{ _template.name }} name: {{ _template.name }}
namespace: {{ _template.namespace }} namespace: {{ _template.namespace }}
spec: spec:
{{ _template.spec }} {{ _template.config }}

View File

@ -4,4 +4,4 @@ metadata:
name: {{ _template.name }} name: {{ _template.name }}
namespace: {{ _template.namespace }} namespace: {{ _template.namespace }}
spec: spec:
{{ _template.spec }} {{ _template.config }}

View File

@ -1,10 +1,10 @@
apiVersion: ipam.cluster.x-k8s.io/v1alpha2 apiVersion: ipam.cluster.x-k8s.io/v1alpha1
kind: InClusterIPPool kind: InClusterIPPool
metadata: metadata:
name: inclusterippool-{{ _template.cluster.name }} name: inclusterippool-{{ _template.cluster.name }}
namespace: {{ _template.cluster.namespace }} namespace: {{ _template.cluster.namespace }}
spec: spec:
addresses: start: {{ _template.cluster.network.startip }}
- {{ _template.cluster.network.startip }}-{{ _template.cluster.network.endip }} end: {{ _template.cluster.network.endip }}
prefix: {{ _template.cluster.network.prefix }} prefix: {{ _template.cluster.network.prefix }}
gateway: {{ _template.cluster.network.gateway }} gateway: {{ _template.cluster.network.gateway }}

View File

@ -1,6 +0,0 @@
apiVersion: authentication.concierge.pinniped.dev/v1alpha1
kind: JWTAuthenticator
metadata:
name: {{ _template.name }}
spec:
{{ _template.spec }}

View File

@ -216,11 +216,6 @@ patches:
kind: Cluster kind: Cluster
name: \${CLUSTER_NAME} name: \${CLUSTER_NAME}
patch: |- patch: |-
- op: replace
path: /spec/clusterNetwork/pods
value:
cidrBlocks:
- 172.30.0.0/16
- op: replace - op: replace
path: /spec/controlPlaneRef/name path: /spec/controlPlaneRef/name
value: ${CLUSTER_NAME}-master value: ${CLUSTER_NAME}-master

View File

@ -1,9 +1,9 @@
apiVersion: kustomize.config.k8s.io/v1beta1 apiVersion: kustomize.config.k8s.io/v1beta1
kind: Kustomization kind: Kustomization
resources: resources:
- downstream-cluster/kubeadmconfigtemplate-{{ _template.cluster.name }}-worker.yaml - manifests/kubeadmconfigtemplate-{{ _template.cluster.name }}-worker.yaml
- downstream-cluster/machinedeployment-{{ _template.cluster.name }}-worker.yaml - manifests/machinedeployment-{{ _template.cluster.name }}-worker.yaml
- downstream-cluster/vspheremachinetemplate-{{ _template.cluster.name }}-worker.yaml - manifests/vspheremachinetemplate-{{ _template.cluster.name }}-worker.yaml
patches: patches:
- patch: |- - patch: |-

View File

@ -1,7 +0,0 @@
apiVersion: idp.supervisor.pinniped.dev/v1alpha1
kind: OIDCIdentityProvider
metadata:
name: {{ _template.name }}
namespace: {{ _template.namespace }}
spec:
{{ _template.spec }}

View File

@ -3,7 +3,6 @@ kind: Secret
metadata: metadata:
name: {{ _template.name }} name: {{ _template.name }}
namespace: {{ _template.namespace }} namespace: {{ _template.namespace }}
type: {{ _template.type }}
data: data:
{% for kv_pair in _template.data %} {% for kv_pair in _template.data %}
"{{ kv_pair.key }}": {{ kv_pair.value }} "{{ kv_pair.key }}": {{ kv_pair.value }}

View File

@ -1,7 +0,0 @@
apiVersion: traefik.containo.us/v1alpha1
kind: ServersTransport
metadata:
name: {{ _template.name }}
namespace: {{ _template.namespace }}
spec:
{{ _template.spec }}

View File

@ -1,6 +1,12 @@
- import_tasks: service.yml - import_tasks: service.yml
- import_tasks: cron.yml - import_tasks: cron.yml
- name: Cleanup tempfile
ansible.builtin.file:
path: "{{ kubeconfig.path }}"
state: absent
when: kubeconfig.path is defined
# - name: Reboot host # - name: Reboot host
# ansible.builtin.shell: # ansible.builtin.shell:
# cmd: systemctl reboot # cmd: systemctl reboot

View File

@ -1,6 +1,2 @@
[defaults] [defaults]
callbacks_enabled = ansible.posix.profile_tasks callbacks_enabled = ansible.posix.profile_tasks
force_color = true
[callback_profile_tasks]
task_output_limit = 5

View File

@ -1,8 +1,7 @@
platform: platform:
k3s: k3s:
version: v1.30.0+k3s1 version: v1.25.9+k3s1
# version: v1.27.1+k3s1
packaged_components: packaged_components:
- name: traefik - name: traefik
@ -23,8 +22,7 @@ platform:
port: 8022 port: 8022
protocol: TCP protocol: TCP
web: web:
redirectTo: redirectTo: websecure
port: websecure
websecure: websecure:
tls: tls:
certResolver: stepca certResolver: stepca
@ -35,10 +33,12 @@ platform:
helm_repositories: helm_repositories:
- name: argo - name: argo
url: https://argoproj.github.io/argo-helm url: https://argoproj.github.io/argo-helm
- name: bitnami - name: authentik
url: https://charts.bitnami.com/bitnami url: https://charts.goauthentik.io
- name: dexidp # - name: codecentric
url: https://charts.dexidp.io # url: https://codecentric.github.io/helm-charts
# - name: dex
# url: https://charts.dexidp.io
- name: gitea-charts - name: gitea-charts
url: https://dl.gitea.io/charts/ url: https://dl.gitea.io/charts/
- name: harbor - name: harbor
@ -51,140 +51,151 @@ platform:
url: https://prometheus-community.github.io/helm-charts url: https://prometheus-community.github.io/helm-charts
- name: smallstep - name: smallstep
url: https://smallstep.github.io/helm-charts/ url: https://smallstep.github.io/helm-charts/
- name: spamasaurus
url: https://code.spamasaurus.com/api/packages/djpbessems/helm
components: components:
argo-cd: argo-cd:
helm: helm:
version: 6.7.7 # (=ArgoCD v.2.10.5) version: 5.27.4 # (= ArgoCD v2.6.7)
chart: argo/argo-cd chart: argo/argo-cd
parse_logic: helm template . | yq --no-doc eval '.. | .image? | select(.)' | sort -u | awk '!/ /' parse_logic: helm template . | yq --no-doc eval '.. | .image? | select(.)' | sort -u | awk '!/ /'
chart_values: !unsafe | chart_values: !unsafe |
configs: configs:
cm:
resource.compareoptions: |
ignoreAggregatedRoles: true
resource.customizations.ignoreDifferences.all: |
jsonPointers:
- /spec/conversion/webhook/clientConfig/caBundle
params:
server.insecure: true
secret: secret:
argocdServerAdminPassword: "{{ vapp['metacluster.password'] | password_hash('bcrypt') }}" argocdServerAdminPassword: "{{ vapp['metacluster.password'] | password_hash('bcrypt') }}"
global:
domain: gitops.{{ vapp['metacluster.fqdn'] | lower }}
server: server:
extraArgs:
- --insecure
ingress: ingress:
enabled: true enabled: true
hosts:
- gitops.{{ vapp['metacluster.fqdn'] }}
authentik:
helm:
version: 2023.3.1
chart: authentik/authentik
parse_logic: helm template . --set postgresql.enabled=true,redis.enabled=true | yq --no-doc eval '.. | .image? | select(.)' | sort -u | awk '!/ /'
chart_values: !unsafe |
authentik:
avatars: none
secret_key: "{{ lookup('ansible.builtin.password', '/dev/null length=64 chars=ascii_lowercase,digits seed=' ~ vapp['guestinfo.hostname']) }}"
postgresql:
password: "{{ lookup('ansible.builtin.password', '/dev/null length=32 chars=ascii_lowercase,digits seed=' ~ vapp['guestinfo.hostname']) }}"
env:
AUTHENTIK_BOOTSTRAP_PASSWORD: "{{ vapp['metacluster.password'] }}"
ingress:
enabled: true
hosts:
- host: auth.{{ vapp['metacluster.fqdn'] }}
paths:
- path: "/"
pathType: Prefix
postgresql:
enabled: true
postgresqlPassword: "{{ lookup('ansible.builtin.password', '/dev/null length=32 chars=ascii_lowercase,digits seed=' ~ vapp['guestinfo.hostname']) }}"
redis:
enabled: true
cert-manager: cert-manager:
helm: helm:
version: 1.14.4 version: 1.11.0
chart: jetstack/cert-manager chart: jetstack/cert-manager
parse_logic: helm template . | yq --no-doc eval '.. | .image? | select(.)' | sort -u | awk '!/ /' parse_logic: helm template . | yq --no-doc eval '.. | .image? | select(.)' | sort -u | awk '!/ /'
chart_values: !unsafe | # chart_values: !unsafe |
installCRDs: true # installCRDs: true
clusterapi: clusterapi:
management: management:
version: version:
# Must match the version referenced at `dependencies.static_binaries[.filename==clusterctl].url` # Must match the version referenced at `dependencies.static_binaries[.filename==clusterctl].url`
base: v1.6.3 base: v1.4.0
# Must match the version referenced at `components.cert-manager.helm.version` # Must match the version referenced at `components.cert-manager.helm.version`
cert_manager: v1.14.4 cert_manager: v1.11.0
infrastructure_vsphere: v1.9.2 infrastructure_vsphere: v1.6.0
ipam_incluster: v0.1.0 ipam_incluster: v0.1.0-alpha.2
# Refer to `https://console.cloud.google.com/gcr/images/cloud-provider-vsphere/GLOBAL/cpi/release/manager` for available tags # Refer to `https://console.cloud.google.com/gcr/images/cloud-provider-vsphere/GLOBAL/cpi/release/manager` for available tags
cpi_vsphere: v1.30.1 cpi_vsphere: v1.25.2
workload: workload:
version: version:
calico: v3.27.3 calico: v3.25.0
k8s: v1.30.1 k8s: v1.25.10
node_template: node_template:
# url: https://{{ repo_username }}:{{ repo_password }}@sn.itch.fyi/Repository/rel/ubuntu-2204-kube-v1.27.1.ova url: https://{{ repo_username }}:{{ repo_password }}@sn.itch.fyi/Repository/rel/ubuntu-2204-kube-v1.25.10.ova
url: https://github.com/kubernetes-sigs/cluster-api-provider-vsphere/releases/download/templates%2Fv1.30.0/ubuntu-2204-kube-v1.30.0.ova
dex: # dex:
helm: # helm:
version: 0.15.3 # (= Dex 2.37.0) # version: 0.13.0 # (= Dex 2.35.3)
chart: dexidp/dex # chart: dex/dex
parse_logic: helm template . | yq --no-doc eval '.. | .image? | select(.)' | sort -u | awk '!/ /' # parse_logic: helm template . | yq --no-doc eval '.. | .image? | select(.)' | sort -u | awk '!/ /'
chart_values: !unsafe | # chart_values: !unsafe |
config: # config:
issuer: https://idps.{{ vapp['metacluster.fqdn'] }} # connectors:
storage: # - type: ldap
type: kubernetes # id: ldap
config: # name: "LDAP"
inCluster: true # config:
staticClients: # host: "{{ vapp['ldap.fqdn'] }}:636"
- id: pinniped-supervisor # insecureNoSSL: false
secret: "{{ lookup('ansible.builtin.password', '/dev/null length=64 chars=ascii_lowercase,digits seed=' ~ vapp['metacluster.fqdn']) }}" # insecureSkipVerify: true
name: Pinniped Supervisor client # bindDN: "{{ vapp['ldap.dn'] }}"
redirectURIs: # bindPW: "{{ vapp['ldap.password'] }}"
- https://auth.{{ vapp['metacluster.fqdn'] }}/sso/callback
enablePasswordDB: true # usernamePrompt: "Username"
staticPasswords: # userSearch:
- email: user@{{ vapp['metacluster.fqdn'] }} # baseDN: OU=Administrators,OU=Useraccounts,DC=bessems,DC=eu
hash: "{{ vapp['metacluster.password'] | password_hash('bcrypt') }}" # filter: "(objectClass=person)"
username: user # username: userPrincipalName
userID: "{{ lookup('ansible.builtin.password', '/dev/null length=64 chars=ascii_lowercase,digits seed=' ~ vapp['metacluster.fqdn']) | to_uuid }}" # idAttr: DN
ingress: # emailAttr: userPrincipalName
enabled: true # nameAttr: cn
hosts:
- host: idps.{{ vapp['metacluster.fqdn'] }} # groupSearch:
paths: # baseDN: OU=Roles,OU=Groups,DC=bessems,DC=eu
- path: / # filter: "(objectClass=group)"
pathType: Prefix # userMatchers:
# - userAttr: DN
# groupAttr: member
# nameAttr: cn
# enablePasswordDB: true
# issuer: https://oidc.{{ vapp['metacluster.fqdn'] }}
# storage:
# type: kubernetes
# config:
# inCluster: true
# ingress:
# enabled: true
# hosts:
# - host: oidc.{{ vapp['metacluster.fqdn'] }}
# paths:
# - path: /
# pathType: Prefix
gitea: gitea:
helm: helm:
version: v10.1.3 # (= Gitea v1.21.7) version: v7.0.2 # (= Gitea v1.18.3)
chart: gitea-charts/gitea chart: gitea-charts/gitea
parse_logic: helm template . | yq --no-doc eval '.. | .image? | select(.)' | sort -u | sed '/:/!s/$/:latest/' parse_logic: helm template . | yq --no-doc eval '.. | .image? | select(.)' | sort -u | sed '/:/!s/$/:latest/'
chart_values: !unsafe | chart_values: !unsafe |
extraVolumes:
- secret:
defaultMode: 420
secretName: step-certificates-certs
name: step-certificates-certs
extraVolumeMounts:
- mountPath: /etc/ssl/certs/ca-chain.crt
name: step-certificates-certs
readOnly: true
subPath: ca_chain.crt
gitea: gitea:
admin: admin:
username: administrator username: administrator
password: "{{ vapp['metacluster.password'] }}" password: "{{ vapp['metacluster.password'] }}"
email: administrator@{{ vapp['metacluster.fqdn'] | lower }} email: admin@{{ vapp['metacluster.fqdn'] }}
config: config:
cache:
ADAPTER: memory
server: server:
OFFLINE_MODE: true OFFLINE_MODE: true
PROTOCOL: http PROTOCOL: http
ROOT_URL: https://git.{{ vapp['metacluster.fqdn'] | lower }}/ ROOT_URL: https://git.{{ vapp['metacluster.fqdn'] }}/
session:
PROVIDER: db
image: image:
pullPolicy: IfNotPresent pullPolicy: IfNotPresent
ingress: ingress:
enabled: true enabled: true
hosts: hosts:
- host: git.{{ vapp['metacluster.fqdn'] | lower }} - host: git.{{ vapp['metacluster.fqdn'] }}
paths: paths:
- path: / - path: /
pathType: Prefix pathType: Prefix
postgresql:
enabled: true
image:
tag: 16.1.0-debian-11-r25
postgresql-ha:
enabled: false
redis-cluster:
enabled: false
service: service:
ssh: ssh:
type: ClusterIP type: ClusterIP
@ -193,7 +204,7 @@ components:
harbor: harbor:
helm: helm:
version: 1.14.1 # (= Harbor v2.10.1) version: 1.11.0 # (= Harbor v2.7.0)
chart: harbor/harbor chart: harbor/harbor
parse_logic: helm template . | yq --no-doc eval '.. | .image? | select(.)' | sort -u | awk '!/ /' parse_logic: helm template . | yq --no-doc eval '.. | .image? | select(.)' | sort -u | awk '!/ /'
chart_values: !unsafe | chart_values: !unsafe |
@ -201,11 +212,11 @@ components:
ingress: ingress:
annotations: {} annotations: {}
hosts: hosts:
core: registry.{{ vapp['metacluster.fqdn'] | lower }} core: registry.{{ vapp['metacluster.fqdn'] }}
tls: tls:
certSource: none certSource: none
enabled: false enabled: false
externalURL: https://registry.{{ vapp['metacluster.fqdn'] | lower }} externalURL: https://registry.{{ vapp['metacluster.fqdn'] }}
harborAdminPassword: "{{ vapp['metacluster.password'] }}" harborAdminPassword: "{{ vapp['metacluster.password'] }}"
notary: notary:
enabled: false enabled: false
@ -214,28 +225,37 @@ components:
registry: registry:
size: 25Gi size: 25Gi
json-server: # keycloakx:
helm: # helm:
version: v0.8.4 # version: 2.1.1 # (= Keycloak 20.0.3)
chart: spamasaurus/json-server # chart: codecentric/keycloakx
parse_logic: helm template . | yq --no-doc eval '.. | .image? | select(.)' | sort -u | awk '!/ /' # parse_logic: helm template . | yq --no-doc eval '.. | .image? | select(.)' | sort -u | awk '!/ /'
chart_values: !unsafe | # chart_values: !unsafe |
ingress: # command:
enabled: true # - "/opt/keycloak/bin/kc.sh"
hosts: # - "start"
- host: version.{{ vapp['metacluster.fqdn'] }} # - "--http-enabled=true"
paths: # - "--http-port=8080"
- path: / # - "--hostname-strict=false"
pathType: Prefix # - "--hostname-strict-https=false"
jsonServer: # extraEnv: |
image: # - name: KEYCLOAK_ADMIN
repository: code.spamasaurus.com/djpbessems/json-server # value: admin
seedData: # - name: KEYCLOAK_ADMIN_PASSWORD
configInline: {} # value: {{ vapp['metacluster.password'] }}
sidecar: # - name: KC_PROXY
targetUrl: version.{{ vapp['metacluster.fqdn'] }} # value: "passthrough"
image: # - name: JAVA_OPTS_APPEND
repository: code.spamasaurus.com/djpbessems/json-server # value: >-
# -Djgroups.dns.query={% raw %}{{ include "keycloak.fullname" . }}{% endraw %}-headless
# ingress:
# enabled: true
# rules:
# - host: keycloak.{{ vapp['metacluster.fqdn'] }}
# paths:
# - path: /
# pathType: Prefix
# tls: []
kube-prometheus-stack: kube-prometheus-stack:
helm: helm:
@ -250,62 +270,42 @@ components:
kubevip: kubevip:
# Must match the version referenced at `dependencies.container_images` # Must match the version referenced at `dependencies.container_images`
version: v0.6.3 version: v0.5.8
longhorn: longhorn:
helm: helm:
version: 1.5.4 version: 1.4.1
chart: longhorn/longhorn chart: longhorn/longhorn
parse_logic: cat values.yaml | yq eval '.. | select(has("repository")) | .repository + ":" + .tag' parse_logic: cat values.yaml | yq eval '.. | select(has("repository")) | .repository + ":" + .tag'
chart_values: !unsafe | chart_values: !unsafe |
defaultSettings: defaultSettings:
concurrentReplicaRebuildPerNodeLimit: 10 allowNodeDrainWithLastHealthyReplica: true
defaultDataPath: /mnt/blockstorage defaultDataPath: /mnt/blockstorage
logLevel: Info defaultReplicaCount: 1
nodeDrainPolicy: block-for-eviction-if-contains-last-replica
replicaSoftAntiAffinity: true
priorityClass: system-node-critical
storageOverProvisioningPercentage: 200
storageReservedPercentageForDefaultDisk: 0
ingress: ingress:
enabled: true enabled: true
host: storage.{{ vapp['metacluster.fqdn'] | lower }} host: storage.{{ vapp['metacluster.fqdn'] }}
longhornManager: persistence:
priorityClass: system-node-critical defaultClassReplicaCount: 1
longhornDriver:
priorityClass: system-node-critical
pinniped:
helm:
version: 1.3.10 # (= Pinniped v0.27.0)
chart: bitnami/pinniped
parse_logic: helm template . | yq --no-doc eval '.. | .image? | select(.)' | sort -u | awk '!/ /'
chart_values: !unsafe |
concierge:
enabled: false
supervisor:
service:
public:
type: ClusterIP
local-user-authenticator:
# Must match the appVersion (!=chart version) referenced at `components.pinniped.helm.version`
version: v0.27.0
users:
- username: metauser
password: !unsafe "{{ vapp['metacluster.password'] | password_hash('bcrypt') }}"
- username: metaguest
password: !unsafe "{{ vapp['metacluster.password'] | password_hash('bcrypt') }}"
step-certificates: step-certificates:
helm: helm:
version: 1.25.2 # (= step-ca v0.25.2) version: 1.23.0
chart: smallstep/step-certificates chart: smallstep/step-certificates
parse_logic: helm template . | yq --no-doc eval '.. | .image? | select(.)' | sed '/:/!s/$/:latest/' | sort -u parse_logic: helm template . | yq --no-doc eval '.. | .image? | select(.)' | sed '/:/!s/$/:latest/' | sort -u
chart_values: !unsafe | chart_values: !unsafe |
ca:
dns: ca.{{ vapp['metacluster.fqdn'] }},step-certificates.step-ca.svc.cluster.local,127.0.0.1
password: "{{ vapp['metacluster.password'] }}"
provisioner:
name: admin
password: "{{ vapp['metacluster.password'] }}"
inject: inject:
secrets: secrets:
ca_password: "{{ vapp['metacluster.password'] | b64encode }}" ca_password: "{{ vapp['metacluster.password'] | b64encode }}"
provisioner_password: "{{ vapp['metacluster.password'] | b64encode }}" provisioner_password: "{{ vapp['metacluster.password'] | b64encode }}"
service:
targetPort: 9000
dependencies: dependencies:
@ -316,48 +316,43 @@ dependencies:
- community.general - community.general
- community.vmware - community.vmware
- kubernetes.core - kubernetes.core
- lvrfrc87.git_acp
container_images: container_images:
# This should match the image tag referenced at `platform.packaged_components[.name==traefik].config` # This should match the image tag referenced at `platform.packaged_components[.name==traefik].config`
- busybox:1 - busybox:1
- ghcr.io/kube-vip/kube-vip:v0.6.3 - ghcr.io/kube-vip/kube-vip:v0.5.8
# The following list is generated by running the following commands: # The following list is generated by running the following commands:
# $ clusterctl init -i vsphere:<version> [...] # $ clusterctl init -i vsphere:<version> [...]
# $ clusterctl generate cluster <name> [...] | yq eval '.data.data' | yq --no-doc eval '.. | .image? | select(.)' | sort -u # $ clusterctl generate cluster <name> [...] | yq eval '.data.data' | yq --no-doc eval '.. | .image? | select(.)' | sort -u
- gcr.io/cloud-provider-vsphere/cpi/release/manager:v1.27.0 - gcr.io/cloud-provider-vsphere/cpi/release/manager:v1.18.1
- gcr.io/cloud-provider-vsphere/csi/release/driver:v3.1.0 - gcr.io/cloud-provider-vsphere/csi/release/driver:v2.1.0
- gcr.io/cloud-provider-vsphere/csi/release/syncer:v3.1.0 - gcr.io/cloud-provider-vsphere/csi/release/syncer:v2.1.0
- registry.k8s.io/sig-storage/csi-attacher:v4.3.0 - quay.io/k8scsi/csi-attacher:v3.0.0
- registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.8.0 - quay.io/k8scsi/csi-node-driver-registrar:v2.0.1
- registry.k8s.io/sig-storage/csi-provisioner:v3.5.0 - quay.io/k8scsi/csi-provisioner:v2.0.0
- registry.k8s.io/sig-storage/csi-resizer:v1.8.0 - quay.io/k8scsi/livenessprobe:v2.1.0
- registry.k8s.io/sig-storage/csi-snapshotter:v6.2.2
- registry.k8s.io/sig-storage/livenessprobe:v2.10.0
static_binaries: static_binaries:
- filename: clusterctl - filename: clusterctl
url: https://github.com/kubernetes-sigs/cluster-api/releases/download/v1.6.3/clusterctl-linux-amd64 url: https://github.com/kubernetes-sigs/cluster-api/releases/download/v1.4.0/clusterctl-linux-amd64
- filename: govc - filename: govc
url: https://github.com/vmware/govmomi/releases/download/v0.36.3/govc_Linux_x86_64.tar.gz url: https://github.com/vmware/govmomi/releases/download/v0.29.0/govc_Linux_x86_64.tar.gz
archive: compressed archive: compressed
- filename: helm - filename: helm
url: https://get.helm.sh/helm-v3.14.3-linux-amd64.tar.gz url: https://get.helm.sh/helm-v3.10.2-linux-amd64.tar.gz
archive: compressed archive: compressed
extra_opts: --strip-components=1 extra_opts: --strip-components=1
- filename: kubectl-slice - filename: kubectl-slice
url: https://github.com/patrickdappollonio/kubectl-slice/releases/download/v1.2.9/kubectl-slice_linux_x86_64.tar.gz url: https://github.com/patrickdappollonio/kubectl-slice/releases/download/v1.2.5/kubectl-slice_linux_x86_64.tar.gz
archive: compressed archive: compressed
- filename: pinniped
url: https://github.com/vmware-tanzu/pinniped/releases/download/v0.25.0/pinniped-cli-linux-amd64
- filename: skopeo - filename: skopeo
url: https://code.spamasaurus.com/api/packages/djpbessems/generic/skopeo/v1.12.0/skopeo_linux_amd64 url: https://code.spamasaurus.com/api/packages/djpbessems/generic/skopeo/v1.12.0/skopeo_linux_amd64
- filename: step - filename: step
url: https://dl.step.sm/gh-release/cli/gh-release-header/v0.25.2/step_linux_0.25.2_amd64.tar.gz url: https://dl.step.sm/gh-release/cli/gh-release-header/v0.23.0/step_linux_0.23.0_amd64.tar.gz
archive: compressed archive: compressed
extra_opts: --strip-components=2 extra_opts: --strip-components=2
- filename: yq - filename: yq
url: https://github.com/mikefarah/yq/releases/download/v4.43.1/yq_linux_amd64 url: http://github.com/mikefarah/yq/releases/download/v4.30.5/yq_linux_amd64
packages: packages:
apt: apt:

View File

@ -1,8 +1,6 @@
downstream: downstream:
helm_repositories: helm_repositories:
- name: bitnami
url: https://charts.bitnami.com/bitnami
- name: longhorn - name: longhorn
url: https://charts.longhorn.io url: https://charts.longhorn.io
- name: sealed-secrets - name: sealed-secrets
@ -11,7 +9,7 @@ downstream:
helm_charts: helm_charts:
longhorn: longhorn:
version: 1.5.4 version: 1.4.1
chart: longhorn/longhorn chart: longhorn/longhorn
namespace: longhorn-system namespace: longhorn-system
parse_logic: cat values.yaml | yq eval '.. | select(has("repository")) | .repository + ":" + .tag' parse_logic: cat values.yaml | yq eval '.. | select(has("repository")) | .repository + ":" + .tag'
@ -20,24 +18,6 @@ downstream:
createDefaultDiskLabeledNodes: true createDefaultDiskLabeledNodes: true
defaultDataPath: /mnt/blockstorage defaultDataPath: /mnt/blockstorage
pinniped:
version: 1.3.10 # (= Pinniped v0.27.0)
chart: bitnami/pinniped
namespace: pinniped-concierge
parse_logic: helm template . | yq --no-doc eval '.. | .image? | select(.)' | sort -u | awk '!/ /'
chart_values: !unsafe |
supervisor:
enabled: false
extra_manifests:
- src: jwtauthenticator.j2
_template:
name: metacluster-sso
spec: !unsafe |2
issuer: https://auth.{{ vapp['metacluster.fqdn'] }}/sso
audience: "{{ vapp['workloadcluster.name'] | lower }}"
tls:
certificateAuthorityData: "{{ (stepca_cm_certs.resources[0].data['intermediate_ca.crt'] ~ _newline ~ stepca_cm_certs.resources[0].data['root_ca.crt']) | b64encode }}"
sealed-secrets: sealed-secrets:
version: 2.8.1 # (= Sealed Secrets v0.20.2) version: 2.8.1 # (= Sealed Secrets v0.20.2)
chart: sealed-secrets/sealed-secrets chart: sealed-secrets/sealed-secrets

View File

@ -1,14 +1,5 @@
packer { packer {
required_plugins { required_plugins {
vsphere = {
source = "github.com/hashicorp/vsphere"
version = "~> 1"
}
ansible = {
source = "github.com/hashicorp/ansible"
version = "~> 1"
}
} }
} }
@ -37,7 +28,6 @@ build {
extra_arguments = [ extra_arguments = [
"--extra-vars", "appliancetype=${source.name}", "--extra-vars", "appliancetype=${source.name}",
"--extra-vars", "applianceversion=${var.appliance_version}",
"--extra-vars", "ansible_ssh_pass=${var.ssh_password}", "--extra-vars", "ansible_ssh_pass=${var.ssh_password}",
"--extra-vars", "docker_username=${var.docker_username}", "--extra-vars", "docker_username=${var.docker_username}",
"--extra-vars", "docker_password=${var.docker_password}", "--extra-vars", "docker_password=${var.docker_password}",
@ -50,12 +40,12 @@ build {
inline = [ inline = [
"pwsh -command \"& scripts/Update-OvfConfiguration.ps1 \\", "pwsh -command \"& scripts/Update-OvfConfiguration.ps1 \\",
" -ApplianceType '${source.name}' \\", " -ApplianceType '${source.name}' \\",
" -OVFFile '/data/scratch/bld_${var.vm_name}_${source.name}.ovf' \"", " -OVFFile '/scratch/bld_${var.vm_name}_${source.name}.ovf' \"",
"pwsh -file scripts/Update-Manifest.ps1 \\", "pwsh -file scripts/Update-Manifest.ps1 \\",
" -ManifestFileName '/data/scratch/bld_${var.vm_name}_${source.name}.mf'", " -ManifestFileName '/scratch/bld_${var.vm_name}_${source.name}.mf'",
"ovftool --acceptAllEulas --allowExtraConfig --overwrite \\", "ovftool --acceptAllEulas --allowExtraConfig --overwrite \\",
" '/data/scratch/bld_${var.vm_name}_${source.name}.ovf' \\", " '/scratch/bld_${var.vm_name}_${source.name}.ovf' \\",
" /output/airgapped-k8s-${var.appliance_version}+${var.k8s_version}-${source.name}.ova" " /output/airgapped-k8s-${var.k8s_version}.${source.name}.ova"
] ]
} }
} }

View File

@ -1,5 +1,5 @@
iso_url = "sn.itch.fyi/Repository/iso/Canonical/Ubuntu%20Server%2022.04/ubuntu-22.04.3-live-server-amd64.iso" iso_url = "sn.itch.fyi/Repository/iso/Canonical/Ubuntu%20Server%2022.04/ubuntu-22.04.1-live-server-amd64.iso"
iso_checksum = "sha256:A4ACFDA10B18DA50E2EC50CCAF860D7F20B389DF8765611142305C0E911D16FD" iso_checksum = "sha256:10F19C5B2B8D6DB711582E0E27F5116296C34FE4B313BA45F9B201A5007056CB"
// iso_url = "sn.itch.fyi/Repository/iso/Canonical/Ubuntu%20Server%2022.04/ubuntu-22.04.1-live-server-amd64.iso" // iso_url = "sn.itch.fyi/Repository/iso/Canonical/Ubuntu%20Server%2022.04/ubuntu-22.04-live-server-amd64.iso"
// iso_checksum = "sha256:10F19C5B2B8D6DB711582E0E27F5116296C34FE4B313BA45F9B201A5007056CB" // iso_checksum = "sha256:84AEAF7823C8C61BAA0AE862D0A06B03409394800000B3235854A6B38EB4856F"

View File

@ -1,19 +1,10 @@
#cloud-config #cloud-config
autoinstall: autoinstall:
version: 1 version: 1
apt:
geoip: true
preserve_sources_list: false
primary:
- arches: [amd64, i386]
uri: http://archive.ubuntu.com/ubuntu
- arches: [default]
uri: http://ports.ubuntu.com/ubuntu-ports
early-commands:
- sudo systemctl stop ssh
locale: en_US locale: en_US
keyboard: keyboard:
layout: us layout: en
variant: us
network: network:
network: network:
version: 2 version: 2
@ -25,18 +16,14 @@ autoinstall:
layout: layout:
name: direct name: direct
identity: identity:
hostname: ubuntu-server hostname: packer-template
username: ubuntu username: ubuntu
# password: $6$ZThRyfmSMh9499ar$KSZus58U/l58Efci0tiJEqDKFCpoy.rv25JjGRv5.iL33AQLTY2aljumkGiDAiX6LsjzVsGTgH85Tx4S.aTfx0
password: $6$rounds=4096$ZKfzRoaQOtc$M.fhOsI0gbLnJcCONXz/YkPfSoefP4i2/PQgzi2xHEi2x9CUhush.3VmYKL0XVr5JhoYvnLfFwqwR/1YYEqZy/ password: $6$rounds=4096$ZKfzRoaQOtc$M.fhOsI0gbLnJcCONXz/YkPfSoefP4i2/PQgzi2xHEi2x9CUhush.3VmYKL0XVr5JhoYvnLfFwqwR/1YYEqZy/
ssh: ssh:
install-server: true install-server: yes
allow-pw: true allow-pw: true
packages:
- openssh-server
- open-vm-tools
- cloud-init
user-data: user-data:
disable_root: false disable_root: false
late-commands: late-commands:
- echo 'ubuntu ALL=(ALL) NOPASSWD:ALL' > /target/etc/sudoers.d/ubuntu - echo 'ubuntu ALL=(ALL) NOPASSWD:ALL' > /target/etc/sudoers.d/ubuntu
- curtin in-target --target=/target -- chmod 440 /etc/sudoers.d/ubuntu

View File

@ -1,61 +1,61 @@
source "vsphere-iso" "ubuntu" { source "vsphere-iso" "ubuntu" {
vcenter_server = var.hv_fqdn vcenter_server = var.vcenter_server
username = var.hv_username username = var.vsphere_username
password = var.hv_password password = var.vsphere_password
insecure_connection = "true" insecure_connection = "true"
datacenter = var.hv_datacenter datacenter = var.vsphere_datacenter
cluster = var.hv_cluster cluster = var.vsphere_cluster
host = var.hv_host host = var.vsphere_host
folder = var.hv_folder folder = var.vsphere_folder
datastore = var.hv_datastore datastore = var.vsphere_datastore
guest_os_type = "ubuntu64Guest" guest_os_type = "ubuntu64Guest"
boot_order = "disk,cdrom" boot_order = "disk,cdrom"
boot_command = [ boot_command = [
"e<down><down><down><end>", "e<down><down><down><end>",
" autoinstall network-config=disabled ds=nocloud;", " autoinstall ds=nocloud;",
"<F10>" "<F10>"
] ]
boot_wait = "2s" boot_wait = "2s"
communicator = "ssh" communicator = "ssh"
ssh_username = "ubuntu" ssh_username = "ubuntu"
ssh_password = var.ssh_password ssh_password = var.ssh_password
ssh_timeout = "20m" ssh_timeout = "20m"
ssh_handshake_attempts = "100" ssh_handshake_attempts = "100"
ssh_pty = true ssh_pty = true
CPUs = 4 CPUs = 4
RAM = 8192 RAM = 8192
network_adapters { network_adapters {
network = var.hv_network network = var.vsphere_network
network_card = "vmxnet3" network_card = "vmxnet3"
} }
storage { storage {
disk_size = 76800 disk_size = 76800
disk_thin_provisioned = true disk_thin_provisioned = true
} }
disk_controller_type = ["pvscsi"] disk_controller_type = ["pvscsi"]
usb_controller = ["xhci"] usb_controller = ["xhci"]
set_host_for_datastore_uploads = true cd_files = [
cd_files = [
"packer/preseed/UbuntuServer22.04/user-data", "packer/preseed/UbuntuServer22.04/user-data",
"packer/preseed/UbuntuServer22.04/meta-data" "packer/preseed/UbuntuServer22.04/meta-data"
] ]
cd_label = "cidata" cd_label = "cidata"
iso_url = local.iso_authenticatedurl iso_url = local.iso_authenticatedurl
iso_checksum = var.iso_checksum iso_checksum = var.iso_checksum
shutdown_command = "echo '${var.ssh_password}' | sudo -S shutdown -P now" shutdown_command = "echo '${var.ssh_password}' | sudo -S shutdown -P now"
shutdown_timeout = "5m" shutdown_timeout = "5m"
remove_cdrom = true remove_cdrom = true
export { export {
output_directory = "/data/scratch" images = false
output_directory = "/scratch"
} }
} }

View File

@ -1,17 +1,17 @@
variable "hv_fqdn" {} variable "vcenter_server" {}
variable "hv_username" {} variable "vsphere_username" {}
variable "hv_password" { variable "vsphere_password" {
sensitive = true sensitive = true
} }
variable "hv_host" {} variable "vsphere_host" {}
variable "hv_datacenter" {} variable "vsphere_datacenter" {}
variable "hv_cluster" {} variable "vsphere_cluster" {}
variable "hv_templatefolder" {} variable "vsphere_templatefolder" {}
variable "hv_folder" {} variable "vsphere_folder" {}
variable "hv_datastore" {} variable "vsphere_datastore" {}
variable "hv_network" {} variable "vsphere_network" {}
variable "vm_name" {} variable "vm_name" {}
variable "ssh_password" { variable "ssh_password" {
@ -34,5 +34,4 @@ variable "docker_password" {
sensitive = true sensitive = true
} }
variable "appliance_version" {}
variable "k8s_version" {} variable "k8s_version" {}

View File

@ -1,10 +1,9 @@
hv_fqdn = "lab-vc-01.bessems.lan" vcenter_server = "bv11-vc.bessems.lan"
hv_username = "administrator@vsphere.local" vsphere_username = "administrator@vsphere.local"
# urlencoded "4/55-Clydebank-Rd" vsphere_datacenter = "DeSchakel"
hv_datacenter = "4%2f55-Clydebank-Rd" vsphere_cluster = "Cluster.01"
hv_cluster = "Cluster.01" vsphere_host = "bv11-esx02.bessems.lan"
hv_host = "lab-esx-02.bessems.lan" vsphere_datastore = "ESX02.SSD02"
hv_datastore = "ESX02.SSD02" vsphere_folder = "/Packer"
hv_folder = "/Packer" vsphere_templatefolder = "/Templates"
hv_templatefolder = "/Templates" vsphere_network = "LAN"
hv_network = "LAN"