Compare commits
1 Commits
4bcb1198f3
...
UbuntuServ
| Author | SHA1 | Date | |
|---|---|---|---|
| d29f7b3223 |
15
.drone.yml
15
.drone.yml
@@ -14,31 +14,29 @@ steps:
|
|||||||
- name: Debugging information
|
- name: Debugging information
|
||||||
image: bv11-cr01.bessems.eu/library/packer-extended
|
image: bv11-cr01.bessems.eu/library/packer-extended
|
||||||
commands:
|
commands:
|
||||||
|
- yamllint --version
|
||||||
|
- packer --version
|
||||||
- ansible --version
|
- ansible --version
|
||||||
- ovftool --version
|
- ovftool --version
|
||||||
- packer --version
|
- name: Ubuntu Server 20.04
|
||||||
- yamllint --version
|
|
||||||
- name: Kubernetes Bootstrap Appliance
|
|
||||||
image: bv11-cr01.bessems.eu/library/packer-extended
|
image: bv11-cr01.bessems.eu/library/packer-extended
|
||||||
pull: always
|
pull: always
|
||||||
commands:
|
commands:
|
||||||
- |
|
- |
|
||||||
sed -i -e "s/<<img-password>>/$${SSH_PASSWORD}/g" \
|
sed -i -e "s/<<img-password>>/$${SSH_PASSWORD}/g" \
|
||||||
packer/preseed/UbuntuServer22.04/user-data
|
packer/preseed/UbuntuServer20.04/user-data
|
||||||
- |
|
- |
|
||||||
yamllint -d "{extends: relaxed, rules: {line-length: disable}}" \
|
yamllint -d "{extends: relaxed, rules: {line-length: disable}}" \
|
||||||
ansible \
|
ansible \
|
||||||
packer/preseed/UbuntuServer22.04/user-data \
|
packer/preseed/UbuntuServer20.04/user-data \
|
||||||
scripts
|
scripts
|
||||||
- |
|
|
||||||
ansible-galaxy install \
|
|
||||||
-r ansible/requirements.yml
|
|
||||||
- |
|
- |
|
||||||
packer init -upgrade \
|
packer init -upgrade \
|
||||||
./packer
|
./packer
|
||||||
- |
|
- |
|
||||||
packer validate \
|
packer validate \
|
||||||
-var vm_name=$DRONE_BUILD_NUMBER-${DRONE_COMMIT_SHA:0:10} \
|
-var vm_name=$DRONE_BUILD_NUMBER-${DRONE_COMMIT_SHA:0:10} \
|
||||||
|
-var vm_guestos=ubuntuserver20.04 \
|
||||||
-var repo_username=$${REPO_USERNAME} \
|
-var repo_username=$${REPO_USERNAME} \
|
||||||
-var repo_password=$${REPO_PASSWORD} \
|
-var repo_password=$${REPO_PASSWORD} \
|
||||||
-var vsphere_password=$${VSPHERE_PASSWORD} \
|
-var vsphere_password=$${VSPHERE_PASSWORD} \
|
||||||
@@ -48,6 +46,7 @@ steps:
|
|||||||
packer build \
|
packer build \
|
||||||
-on-error=cleanup -timestamp-ui \
|
-on-error=cleanup -timestamp-ui \
|
||||||
-var vm_name=$DRONE_BUILD_NUMBER-${DRONE_COMMIT_SHA:0:10} \
|
-var vm_name=$DRONE_BUILD_NUMBER-${DRONE_COMMIT_SHA:0:10} \
|
||||||
|
-var vm_guestos=ubuntuserver20.04 \
|
||||||
-var repo_username=$${REPO_USERNAME} \
|
-var repo_username=$${REPO_USERNAME} \
|
||||||
-var repo_password=$${REPO_PASSWORD} \
|
-var repo_password=$${REPO_PASSWORD} \
|
||||||
-var vsphere_password=$${VSPHERE_PASSWORD} \
|
-var vsphere_password=$${VSPHERE_PASSWORD} \
|
||||||
|
|||||||
@@ -1 +1 @@
|
|||||||
# Packer.Images [](https://ci.spamasaurus.com/djpbessems/Packer.Images)
|
# Packer.Images [](https://ci.spamasaurus.com/djpbessems/Packer.Images)
|
||||||
|
|||||||
@@ -1,10 +1,7 @@
|
|||||||
---
|
---
|
||||||
- hosts: all
|
- hosts: all
|
||||||
gather_facts: false
|
gather_facts: false
|
||||||
vars_files:
|
|
||||||
- metacluster.yml
|
|
||||||
become: true
|
become: true
|
||||||
roles:
|
roles:
|
||||||
- os
|
- os
|
||||||
- firstboot
|
- firstboot
|
||||||
- assets
|
|
||||||
|
|||||||
@@ -1,7 +0,0 @@
|
|||||||
collections:
|
|
||||||
- name: https://github.com/ansible-collections/ansible.utils
|
|
||||||
type: git
|
|
||||||
- name: https://github.com/ansible-collections/community.general
|
|
||||||
type: git
|
|
||||||
- name: https://github.com/ansible-collections/kubernetes.core
|
|
||||||
type: git
|
|
||||||
@@ -1,49 +0,0 @@
|
|||||||
- name: Parse manifests for container images
|
|
||||||
ansible.builtin.shell:
|
|
||||||
# This set of commands is necessary to deal with multi-line scalar values
|
|
||||||
# eg.:
|
|
||||||
# key: |
|
|
||||||
# multi-line
|
|
||||||
# value
|
|
||||||
cmd: >-
|
|
||||||
cat {{ item.dest }} | yq --no-doc eval '.. | .image? | select(.)' | awk '!/ /';
|
|
||||||
cat {{ item.dest }} | yq eval '.data.data' | yq --no-doc eval '.. | .image? | select(.)';
|
|
||||||
cat {{ item.dest }} | yq --no-doc eval '.. | .files? | with_entries(select(.value.path == "*.yaml")).[0].content' | awk '!/null/' | yq eval '.. | .image? | select(.)'
|
|
||||||
register: parsedmanifests
|
|
||||||
loop: "{{ clusterapi_manifests.results }}"
|
|
||||||
loop_control:
|
|
||||||
label: "{{ item.dest | basename }}"
|
|
||||||
|
|
||||||
- name: Parse helm charts for container images
|
|
||||||
ansible.builtin.shell:
|
|
||||||
cmd: "{{ item.value.helm.parse_logic }}"
|
|
||||||
chdir: /opt/metacluster/helm-charts/{{ item.key }}
|
|
||||||
register: chartimages
|
|
||||||
when: item.value.helm is defined
|
|
||||||
loop: "{{ lookup('ansible.builtin.dict', components) }}"
|
|
||||||
loop_control:
|
|
||||||
label: "{{ item.key }}"
|
|
||||||
|
|
||||||
- name: Store container images in dicts
|
|
||||||
ansible.builtin.set_fact:
|
|
||||||
containerimages_{{ item.source }}: "{{ item.results }}"
|
|
||||||
loop:
|
|
||||||
- source: charts
|
|
||||||
results: "{{ chartimages | json_query('results[*].stdout_lines') | select() | flatten | list }}"
|
|
||||||
- source: kubeadm
|
|
||||||
results: "{{ kubeadmimages.stdout_lines }}"
|
|
||||||
- source: manifests
|
|
||||||
results: "{{ parsedmanifests | json_query('results[*].stdout_lines') | select() | flatten | list }}"
|
|
||||||
loop_control:
|
|
||||||
label: "{{ item.source }}"
|
|
||||||
|
|
||||||
- name: Pull and store containerimages
|
|
||||||
ansible.builtin.shell:
|
|
||||||
cmd: >-
|
|
||||||
skopeo copy \
|
|
||||||
--insecure-policy \
|
|
||||||
--retry-times=5 \
|
|
||||||
docker://{{ item }} \
|
|
||||||
docker-archive:./{{ ( item | regex_findall('[^/:]+'))[-2] }}_{{ lookup('ansible.builtin.password', '/dev/null length=5 chars=ascii_lowercase,digits seed={{ item }}') }}.tar:{{ item }}
|
|
||||||
chdir: /opt/metacluster/container-images
|
|
||||||
loop: "{{ (containerimages_charts + containerimages_kubeadm + containerimages_manifests + dependencies.container_images) | flatten | unique | sort }}"
|
|
||||||
@@ -1,31 +0,0 @@
|
|||||||
---
|
|
||||||
- name: Initialize tempfolder
|
|
||||||
ansible.builtin.tempfile:
|
|
||||||
state: directory
|
|
||||||
register: archive
|
|
||||||
|
|
||||||
- name: Download & extract archived static binary
|
|
||||||
ansible.builtin.unarchive:
|
|
||||||
src: "{{ item.url }}"
|
|
||||||
dest: "{{ archive.path }}"
|
|
||||||
remote_src: yes
|
|
||||||
extra_opts: "{{ item.extra_opts | default(omit) }}"
|
|
||||||
register: staticbinary_download
|
|
||||||
retries: 5
|
|
||||||
delay: 5
|
|
||||||
until: staticbinary_download is not failed
|
|
||||||
|
|
||||||
- name: Install extracted binary
|
|
||||||
ansible.builtin.copy:
|
|
||||||
src: "{{ archive.path }}/{{ item.filename }}"
|
|
||||||
dest: /usr/local/bin/{{ item.filename }}
|
|
||||||
remote_src: yes
|
|
||||||
owner: root
|
|
||||||
group: root
|
|
||||||
mode: 0755
|
|
||||||
|
|
||||||
- name: Cleanup tempfolder
|
|
||||||
ansible.builtin.file:
|
|
||||||
path: "{{ archive.path }}"
|
|
||||||
state: absent
|
|
||||||
when: archive.path is defined
|
|
||||||
@@ -1,54 +0,0 @@
|
|||||||
- name: Download & install static binaries
|
|
||||||
ansible.builtin.get_url:
|
|
||||||
url: "{{ item.url }}"
|
|
||||||
url_username: "{{ item.username | default(omit) }}"
|
|
||||||
url_password: "{{ item.password | default(omit) }}"
|
|
||||||
dest: /usr/local/bin/{{ item.filename }}
|
|
||||||
owner: root
|
|
||||||
group: root
|
|
||||||
mode: 0755
|
|
||||||
register: staticbinary_download
|
|
||||||
loop: "{{ dependencies.static_binaries | selectattr('archive', 'undefined') }}"
|
|
||||||
loop_control:
|
|
||||||
label: "{{ item.filename }}"
|
|
||||||
retries: 5
|
|
||||||
delay: 5
|
|
||||||
until: staticbinary_download is not failed
|
|
||||||
|
|
||||||
- name: Download, extract & install archived static binaries
|
|
||||||
include_tasks: dependencies.archive_compressed.yml
|
|
||||||
loop: "{{ dependencies.static_binaries | rejectattr('archive', 'undefined') | selectattr('archive', 'equalto', 'compressed') }}"
|
|
||||||
loop_control:
|
|
||||||
label: "{{ item.filename }}"
|
|
||||||
|
|
||||||
- name: Install ansible-galaxy collections
|
|
||||||
ansible.builtin.shell:
|
|
||||||
cmd: ansible-galaxy collection install {{ item }}
|
|
||||||
register: collections
|
|
||||||
loop: "{{ dependencies.ansible_galaxy_collections }}"
|
|
||||||
retries: 5
|
|
||||||
delay: 5
|
|
||||||
until: collections is not failed
|
|
||||||
|
|
||||||
- name: Install distro packages
|
|
||||||
ansible.builtin.apt:
|
|
||||||
pkg: "{{ dependencies.packages.apt }}"
|
|
||||||
state: latest
|
|
||||||
update_cache: yes
|
|
||||||
install_recommends: no
|
|
||||||
|
|
||||||
- name: Upgrade all packages
|
|
||||||
ansible.builtin.apt:
|
|
||||||
name: '*'
|
|
||||||
state: latest
|
|
||||||
update_cache: yes
|
|
||||||
|
|
||||||
- name: Install additional python packages
|
|
||||||
ansible.builtin.pip:
|
|
||||||
name: "{{ dependencies.packages.pip }}"
|
|
||||||
state: latest
|
|
||||||
|
|
||||||
- name: Cleanup apt cache
|
|
||||||
ansible.builtin.apt:
|
|
||||||
autoremove: yes
|
|
||||||
purge: yes
|
|
||||||
@@ -1,5 +0,0 @@
|
|||||||
- name: Clone git repository
|
|
||||||
ansible.builtin.git:
|
|
||||||
repo: "{{ platform.gitops.repository.uri }}"
|
|
||||||
version: "{{ platform.gitops.repository.revision }}"
|
|
||||||
dest: /opt/metacluster/git-repositories/gitops
|
|
||||||
@@ -1,19 +0,0 @@
|
|||||||
- name: Add helm repositories
|
|
||||||
kubernetes.core.helm_repository:
|
|
||||||
name: "{{ item.name }}"
|
|
||||||
repo_url: "{{ item.url }}"
|
|
||||||
state: present
|
|
||||||
loop: "{{ platform.helm_repositories }}"
|
|
||||||
|
|
||||||
- name: Fetch helm charts
|
|
||||||
ansible.builtin.command:
|
|
||||||
cmd: helm fetch {{ item.value.helm.chart }} --untar --version {{ item.value.helm.version }}
|
|
||||||
chdir: /opt/metacluster/helm-charts
|
|
||||||
when: item.value.helm is defined
|
|
||||||
register: helmcharts
|
|
||||||
loop: "{{ lookup('ansible.builtin.dict', components) }}"
|
|
||||||
loop_control:
|
|
||||||
label: "{{ item.key }}"
|
|
||||||
retries: 5
|
|
||||||
delay: 5
|
|
||||||
until: helmcharts is not failed
|
|
||||||
@@ -1,43 +0,0 @@
|
|||||||
- name: Download & install K3s binary
|
|
||||||
ansible.builtin.get_url:
|
|
||||||
url: https://github.com/k3s-io/k3s/releases/download/{{ platform.k3s.version }}/k3s
|
|
||||||
dest: /usr/local/bin/k3s
|
|
||||||
owner: root
|
|
||||||
group: root
|
|
||||||
mode: 0755
|
|
||||||
register: download
|
|
||||||
until: download is not failed
|
|
||||||
retries: 3
|
|
||||||
delay: 10
|
|
||||||
|
|
||||||
- name: Download K3s images tarball
|
|
||||||
ansible.builtin.get_url:
|
|
||||||
url: https://github.com/k3s-io/k3s/releases/download/{{ platform.k3s.version }}/k3s-airgap-images-amd64.tar.gz
|
|
||||||
dest: /var/lib/rancher/k3s/agent/images
|
|
||||||
register: download
|
|
||||||
until: download is not failed
|
|
||||||
retries: 3
|
|
||||||
delay: 10
|
|
||||||
|
|
||||||
- name: Download K3s install script
|
|
||||||
ansible.builtin.get_url:
|
|
||||||
url: https://get.k3s.io
|
|
||||||
dest: /opt/metacluster/k3s/install.sh
|
|
||||||
owner: root
|
|
||||||
group: root
|
|
||||||
mode: 0755
|
|
||||||
register: download
|
|
||||||
until: download is not failed
|
|
||||||
retries: 3
|
|
||||||
delay: 10
|
|
||||||
|
|
||||||
- name: Inject manifests
|
|
||||||
ansible.builtin.template:
|
|
||||||
src: helmchartconfig.j2
|
|
||||||
dest: /var/lib/rancher/k3s/server/manifests/{{ item.name }}-config.yaml
|
|
||||||
owner: root
|
|
||||||
group: root
|
|
||||||
mode: 0600
|
|
||||||
loop: "{{ platform.packaged_components }}"
|
|
||||||
loop_control:
|
|
||||||
label: "{{ item.name }}"
|
|
||||||
@@ -1,26 +0,0 @@
|
|||||||
- name: Initialize tempfile
|
|
||||||
ansible.builtin.tempfile:
|
|
||||||
state: directory
|
|
||||||
register: kubeadm
|
|
||||||
|
|
||||||
- name: Download kubeadm binary
|
|
||||||
ansible.builtin.get_url:
|
|
||||||
url: https://dl.k8s.io/release/{{ components.clusterapi.workload.version.k8s }}/bin/linux/amd64/kubeadm
|
|
||||||
dest: "{{ kubeadm.path }}/kubeadm"
|
|
||||||
mode: u+x
|
|
||||||
|
|
||||||
- name: Retrieve container images list
|
|
||||||
ansible.builtin.shell:
|
|
||||||
cmd: "{{ kubeadm.path }}/kubeadm config images list --kubernetes-version {{ components.clusterapi.workload.version.k8s }}"
|
|
||||||
register: kubeadmimages
|
|
||||||
|
|
||||||
- name: Store list of container images for reference
|
|
||||||
ansible.builtin.copy:
|
|
||||||
dest: /opt/metacluster/cluster-api/imagelist
|
|
||||||
content: "{{ kubeadmimages.stdout }}"
|
|
||||||
|
|
||||||
- name: Cleanup tempfile
|
|
||||||
ansible.builtin.file:
|
|
||||||
path: "{{ kubeadm.path }}"
|
|
||||||
state: absent
|
|
||||||
when: kubeadm.path is defined
|
|
||||||
@@ -1,30 +0,0 @@
|
|||||||
- name: Create folder structure(s)
|
|
||||||
ansible.builtin.file:
|
|
||||||
path: "{{ item }}"
|
|
||||||
state: directory
|
|
||||||
loop:
|
|
||||||
- /etc/rancher/k3s
|
|
||||||
- /opt/metacluster/cluster-api/bootstrap-kubeadm/{{ components.clusterapi.management.version.base }}
|
|
||||||
- /opt/metacluster/cluster-api/cert-manager/{{ components.clusterapi.management.version.cert_manager }}
|
|
||||||
- /opt/metacluster/cluster-api/cluster-api/{{ components.clusterapi.management.version.base }}
|
|
||||||
- /opt/metacluster/cluster-api/cni-calico/{{ components.clusterapi.workload.version.calico }}
|
|
||||||
- /opt/metacluster/cluster-api/control-plane-kubeadm/{{ components.clusterapi.management.version.base }}
|
|
||||||
- /opt/metacluster/cluster-api/infrastructure-vsphere/{{ components.clusterapi.management.version.infrastructure_vsphere }}
|
|
||||||
- /opt/metacluster/cluster-api/ipam-in-cluster/{{ components.clusterapi.management.version.ipam_incluster }}
|
|
||||||
- /opt/metacluster/container-images
|
|
||||||
- /opt/metacluster/git-repositories/gitops
|
|
||||||
- /opt/metacluster/helm-charts
|
|
||||||
- /opt/metacluster/k3s
|
|
||||||
- /opt/metacluster/kube-vip
|
|
||||||
- /opt/workloadcluster/node-templates
|
|
||||||
- /var/lib/rancher/k3s/agent/images
|
|
||||||
- /var/lib/rancher/k3s/server/manifests
|
|
||||||
|
|
||||||
- import_tasks: dependencies.yml
|
|
||||||
- import_tasks: k3s.yml
|
|
||||||
- import_tasks: helm.yml
|
|
||||||
- import_tasks: git.yml
|
|
||||||
- import_tasks: manifests.yml
|
|
||||||
- import_tasks: kubeadm.yml
|
|
||||||
- import_tasks: containerimages.yml
|
|
||||||
- import_tasks: nodetemplates.yml
|
|
||||||
@@ -1,86 +0,0 @@
|
|||||||
- block:
|
|
||||||
|
|
||||||
- name: Aggregate chart_values into dict
|
|
||||||
ansible.builtin.set_fact:
|
|
||||||
chart_values: "{{ chart_values | default({}) | combine({ (item.key | regex_replace('[^A-Za-z0-9]', '')): { 'chart_values': (item.value.helm.chart_values | from_yaml) } }) }}"
|
|
||||||
when: item.value.helm.chart_values is defined
|
|
||||||
loop: "{{ lookup('ansible.builtin.dict', components) }}"
|
|
||||||
loop_control:
|
|
||||||
label: "{{ item.key }}"
|
|
||||||
|
|
||||||
- name: Write dict to vars_file
|
|
||||||
ansible.builtin.copy:
|
|
||||||
dest: /opt/firstboot/ansible/vars/metacluster.yml
|
|
||||||
content: >-
|
|
||||||
{{
|
|
||||||
{ 'components': (
|
|
||||||
chart_values |
|
|
||||||
combine({ 'clusterapi': components.clusterapi }) |
|
|
||||||
combine({ 'kubevip' : components.kubevip }) )
|
|
||||||
} | to_nice_yaml(indent=2, width=4096)
|
|
||||||
}}
|
|
||||||
|
|
||||||
- name: Download ClusterAPI manifests
|
|
||||||
ansible.builtin.get_url:
|
|
||||||
url: "{{ item.url }}"
|
|
||||||
dest: /opt/metacluster/cluster-api/{{ item.dest }}
|
|
||||||
register: clusterapi_manifests
|
|
||||||
loop:
|
|
||||||
# This list is based on `clusterctl config repositories`
|
|
||||||
# Note: Each manifest also needs a `metadata.yaml` file stored in the respective folder
|
|
||||||
- url: https://github.com/kubernetes-sigs/cluster-api/releases/download/{{ components.clusterapi.management.version.base }}/bootstrap-components.yaml
|
|
||||||
dest: bootstrap-kubeadm/{{ components.clusterapi.management.version.base }}/bootstrap-components.yaml
|
|
||||||
- url: https://github.com/kubernetes-sigs/cluster-api/releases/download/{{ components.clusterapi.management.version.base }}/core-components.yaml
|
|
||||||
dest: cluster-api/{{ components.clusterapi.management.version.base }}/core-components.yaml
|
|
||||||
- url: https://github.com/kubernetes-sigs/cluster-api/releases/download/{{ components.clusterapi.management.version.base }}/control-plane-components.yaml
|
|
||||||
dest: control-plane-kubeadm/{{ components.clusterapi.management.version.base }}/control-plane-components.yaml
|
|
||||||
# This downloads the same `metadata.yaml` file to three separate folders
|
|
||||||
- url: https://github.com/kubernetes-sigs/cluster-api/releases/download/{{ components.clusterapi.management.version.base }}/metadata.yaml
|
|
||||||
dest: bootstrap-kubeadm/{{ components.clusterapi.management.version.base }}/metadata.yaml
|
|
||||||
- url: https://github.com/kubernetes-sigs/cluster-api/releases/download/{{ components.clusterapi.management.version.base }}/metadata.yaml
|
|
||||||
dest: cluster-api/{{ components.clusterapi.management.version.base }}/metadata.yaml
|
|
||||||
- url: https://github.com/kubernetes-sigs/cluster-api/releases/download/{{ components.clusterapi.management.version.base }}/metadata.yaml
|
|
||||||
dest: control-plane-kubeadm/{{ components.clusterapi.management.version.base }}/metadata.yaml
|
|
||||||
# The vsphere infrastructure provider requires multiple files (`cluster-template.yaml` and `metadata.yaml` on top of default files)
|
|
||||||
- url: https://github.com/kubernetes-sigs/cluster-api-provider-vsphere/releases/download/{{ components.clusterapi.management.version.infrastructure_vsphere }}/infrastructure-components.yaml
|
|
||||||
dest: infrastructure-vsphere/{{ components.clusterapi.management.version.infrastructure_vsphere }}/infrastructure-components.yaml
|
|
||||||
- url: https://github.com/kubernetes-sigs/cluster-api-provider-vsphere/releases/download/{{ components.clusterapi.management.version.infrastructure_vsphere }}/cluster-template.yaml
|
|
||||||
dest: infrastructure-vsphere/{{ components.clusterapi.management.version.infrastructure_vsphere }}/cluster-template.yaml
|
|
||||||
- url: https://github.com/kubernetes-sigs/cluster-api-provider-vsphere/releases/download/{{ components.clusterapi.management.version.infrastructure_vsphere }}/metadata.yaml
|
|
||||||
dest: infrastructure-vsphere/{{ components.clusterapi.management.version.infrastructure_vsphere }}/metadata.yaml
|
|
||||||
# Additionally, cert-manager is a prerequisite
|
|
||||||
- url: https://github.com/cert-manager/cert-manager/releases/download/{{ components.clusterapi.management.version.cert_manager }}/cert-manager.yaml
|
|
||||||
dest: cert-manager/{{ components.clusterapi.management.version.cert_manager }}/cert-manager.yaml
|
|
||||||
# Finally, workload clusters will need a CNI plugin
|
|
||||||
- url: https://raw.githubusercontent.com/projectcalico/calico/{{ components.clusterapi.workload.version.calico }}/manifests/calico.yaml
|
|
||||||
dest: cni-calico/{{ components.clusterapi.workload.version.calico }}/calico.yaml
|
|
||||||
# IPAM in-cluster provider (w/ metadata.yaml)
|
|
||||||
- url: https://github.com/telekom/cluster-api-ipam-provider-in-cluster/releases/download/{{ components.clusterapi.management.version.ipam_incluster }}/ipam-components.yaml
|
|
||||||
dest: ipam-in-cluster/{{ components.clusterapi.management.version.ipam_incluster }}/ipam-components.yaml
|
|
||||||
- url: https://github.com/telekom/cluster-api-ipam-provider-in-cluster/releases/download/{{ components.clusterapi.management.version.ipam_incluster }}/metadata.yaml
|
|
||||||
dest: ipam-in-cluster/{{ components.clusterapi.management.version.ipam_incluster }}/metadata.yaml
|
|
||||||
loop_control:
|
|
||||||
label: "{{ item.url | basename }}"
|
|
||||||
retries: 5
|
|
||||||
delay: 5
|
|
||||||
until: clusterapi_manifests is not failed
|
|
||||||
|
|
||||||
- name: Download kube-vip RBAC manifest
|
|
||||||
ansible.builtin.get_url:
|
|
||||||
url: https://kube-vip.io/manifests/rbac.yaml
|
|
||||||
dest: /opt/metacluster/kube-vip/rbac.yaml
|
|
||||||
register: kubevip_manifest
|
|
||||||
retries: 5
|
|
||||||
delay: 5
|
|
||||||
until: kubevip_manifest is not failed
|
|
||||||
|
|
||||||
# - name: Inject manifests
|
|
||||||
# ansible.builtin.template:
|
|
||||||
# src: "{{ item.type }}.j2"
|
|
||||||
# dest: /var/lib/rancher/k3s/server/manifests/{{ item.name }}-manifest.yaml
|
|
||||||
# owner: root
|
|
||||||
# group: root
|
|
||||||
# mode: 0600
|
|
||||||
# loop: "{{ lookup('ansible.builtin.dict', components) | map(attribute='value.manifests') | list | select('defined') | flatten }}"
|
|
||||||
# loop_control:
|
|
||||||
# label: "{{ item.type ~ '/' ~ item.name }}"
|
|
||||||
@@ -1,4 +0,0 @@
|
|||||||
- name: Download node-template image
|
|
||||||
ansible.builtin.uri:
|
|
||||||
url: "{{ components.clusterapi.workload.node_template.url }}"
|
|
||||||
dest: /opt/workloadcluster/node-templates/{{ components.clusterapi.workload.node_template.url | basename}}
|
|
||||||
@@ -1,8 +0,0 @@
|
|||||||
apiVersion: helm.cattle.io/v1
|
|
||||||
kind: HelmChartConfig
|
|
||||||
metadata:
|
|
||||||
name: {{ item.name }}
|
|
||||||
namespace: {{ item.namespace }}
|
|
||||||
spec:
|
|
||||||
valuesContent: |-
|
|
||||||
{{ item.config }}
|
|
||||||
@@ -1,26 +0,0 @@
|
|||||||
---
|
|
||||||
- hosts: 127.0.0.1
|
|
||||||
connection: local
|
|
||||||
gather_facts: true
|
|
||||||
vars_files:
|
|
||||||
- defaults.yml
|
|
||||||
- metacluster.yml
|
|
||||||
# become: true
|
|
||||||
roles:
|
|
||||||
- vapp
|
|
||||||
- network
|
|
||||||
- preflight
|
|
||||||
- users
|
|
||||||
- disks
|
|
||||||
- metacluster
|
|
||||||
- workloadcluster
|
|
||||||
- tty
|
|
||||||
- cleanup
|
|
||||||
handlers:
|
|
||||||
- name: Apply manifests
|
|
||||||
kubernetes.core.k8s:
|
|
||||||
src: "{{ item }}"
|
|
||||||
state: present
|
|
||||||
kubeconfig: "{{ kubeconfig.path }}"
|
|
||||||
loop: "{{ query('ansible.builtin.fileglob', '/var/lib/rancher/k3s/server/manifests/*.yaml') | sort }}"
|
|
||||||
ignore_errors: yes
|
|
||||||
@@ -1,14 +0,0 @@
|
|||||||
import netaddr
|
|
||||||
|
|
||||||
def netaddr_iter_iprange(ip_start, ip_end):
|
|
||||||
return [str(ip) for ip in netaddr.iter_iprange(ip_start, ip_end)]
|
|
||||||
|
|
||||||
class FilterModule(object):
|
|
||||||
''' Ansible filter. Interface to netaddr methods.
|
|
||||||
https://pypi.org/project/netaddr/
|
|
||||||
'''
|
|
||||||
|
|
||||||
def filters(self):
|
|
||||||
return {
|
|
||||||
'netaddr_iter_iprange': netaddr_iter_iprange
|
|
||||||
}
|
|
||||||
@@ -1,131 +0,0 @@
|
|||||||
- block:
|
|
||||||
|
|
||||||
- name: Install step-ca chart
|
|
||||||
kubernetes.core.helm:
|
|
||||||
name: step-certificates
|
|
||||||
chart_ref: /opt/metacluster/helm-charts/step-certificates
|
|
||||||
release_namespace: step-ca
|
|
||||||
create_namespace: yes
|
|
||||||
# Unable to use REST api based readycheck due to missing ingress
|
|
||||||
wait: yes
|
|
||||||
kubeconfig: "{{ kubeconfig.path }}"
|
|
||||||
values: "{{ components.stepcertificates.chart_values }}"
|
|
||||||
|
|
||||||
- name: Retrieve configmap w/ root certificate
|
|
||||||
kubernetes.core.k8s_info:
|
|
||||||
kind: ConfigMap
|
|
||||||
name: step-certificates-certs
|
|
||||||
namespace: step-ca
|
|
||||||
kubeconfig: "{{ kubeconfig.path }}"
|
|
||||||
register: stepca_cm_certs
|
|
||||||
|
|
||||||
- name: Create target namespaces
|
|
||||||
kubernetes.core.k8s:
|
|
||||||
kind: Namespace
|
|
||||||
name: "{{ item }}"
|
|
||||||
state: present
|
|
||||||
kubeconfig: "{{ kubeconfig.path }}"
|
|
||||||
loop:
|
|
||||||
- argo-cd
|
|
||||||
# - kube-system
|
|
||||||
|
|
||||||
- name: Store root certificate in namespaced configmaps/secrets
|
|
||||||
kubernetes.core.k8s:
|
|
||||||
state: present
|
|
||||||
template: "{{ item.kind }}.j2"
|
|
||||||
kubeconfig: "{{ kubeconfig.path }}"
|
|
||||||
vars:
|
|
||||||
_template:
|
|
||||||
name: "{{ item.name }}"
|
|
||||||
namespace: "{{ item.namespace }}"
|
|
||||||
annotations: "{{ item.annotations | default('{}') | indent(width=4, first=True) }}"
|
|
||||||
labels: "{{ item.labels | default('{}') | indent(width=4, first=True) }}"
|
|
||||||
data: "{{ item.data }}"
|
|
||||||
loop:
|
|
||||||
- name: argocd-tls-certs-cm
|
|
||||||
namespace: argo-cd
|
|
||||||
kind: configmap
|
|
||||||
annotations: |
|
|
||||||
meta.helm.sh/release-name: argo-cd
|
|
||||||
meta.helm.sh/release-namespace: argo-cd
|
|
||||||
labels: |
|
|
||||||
app.kubernetes.io/managed-by: Helm
|
|
||||||
app.kubernetes.io/name: argocd-cm
|
|
||||||
app.kubernetes.io/part-of: argocd
|
|
||||||
data:
|
|
||||||
- key: git.{{ vapp['metacluster.fqdn'] }}
|
|
||||||
value: "{{ stepca_cm_certs.resources[0].data['root_ca.crt'] }}"
|
|
||||||
- name: step-certificates-certs
|
|
||||||
namespace: kube-system
|
|
||||||
kind: secret
|
|
||||||
data:
|
|
||||||
- key: root_ca.crt
|
|
||||||
value: "{{ stepca_cm_certs.resources[0].data['root_ca.crt'] | b64encode }}"
|
|
||||||
loop_control:
|
|
||||||
label: "{{ item.kind ~ '/' ~ item.name ~ ' (' ~ item.namespace ~ ')' }}"
|
|
||||||
|
|
||||||
- name: Configure step-ca passthrough ingress
|
|
||||||
ansible.builtin.template:
|
|
||||||
src: ingressroutetcp.j2
|
|
||||||
dest: /var/lib/rancher/k3s/server/manifests/{{ _template.name }}-manifest.yaml
|
|
||||||
owner: root
|
|
||||||
group: root
|
|
||||||
mode: 0600
|
|
||||||
vars:
|
|
||||||
_template:
|
|
||||||
name: step-ca
|
|
||||||
namespace: step-ca
|
|
||||||
config: |2
|
|
||||||
entryPoints:
|
|
||||||
- websecure
|
|
||||||
routes:
|
|
||||||
- match: HostSNI(`ca.{{ vapp['metacluster.fqdn'] }}`)
|
|
||||||
services:
|
|
||||||
- name: step-certificates
|
|
||||||
port: 443
|
|
||||||
tls:
|
|
||||||
passthrough: true
|
|
||||||
notify:
|
|
||||||
- Apply manifests
|
|
||||||
|
|
||||||
- name: Inject step-ca certificate into traefik container
|
|
||||||
ansible.builtin.blockinfile:
|
|
||||||
path: /var/lib/rancher/k3s/server/manifests/traefik-config.yaml
|
|
||||||
block: |2
|
|
||||||
volumes:
|
|
||||||
- name: step-certificates-certs
|
|
||||||
mountPath: /step-ca
|
|
||||||
type: secret
|
|
||||||
env:
|
|
||||||
- name: LEGO_CA_CERTIFICATES
|
|
||||||
value: /step-ca/root_ca.crt
|
|
||||||
marker: ' # {mark} ANSIBLE MANAGED BLOCK'
|
|
||||||
notify:
|
|
||||||
- Apply manifests
|
|
||||||
|
|
||||||
- name: Trigger handlers
|
|
||||||
ansible.builtin.meta: flush_handlers
|
|
||||||
|
|
||||||
- name: Retrieve step-ca configuration
|
|
||||||
kubernetes.core.k8s_info:
|
|
||||||
kind: ConfigMap
|
|
||||||
name: step-certificates-config
|
|
||||||
namespace: step-ca
|
|
||||||
kubeconfig: "{{ kubeconfig.path }}"
|
|
||||||
register: stepca_cm_config
|
|
||||||
|
|
||||||
- name: Install root CA in system truststore
|
|
||||||
ansible.builtin.shell:
|
|
||||||
cmd: >-
|
|
||||||
step ca bootstrap \
|
|
||||||
--ca-url=https://ca.{{ vapp['metacluster.fqdn'] }} \
|
|
||||||
--fingerprint={{ stepca_cm_config.resources[0].data['defaults.json'] | from_json | json_query('fingerprint') }} \
|
|
||||||
--install \
|
|
||||||
--force
|
|
||||||
update-ca-certificates
|
|
||||||
|
|
||||||
module_defaults:
|
|
||||||
ansible.builtin.uri:
|
|
||||||
validate_certs: no
|
|
||||||
status_code: [200, 201]
|
|
||||||
body_format: json
|
|
||||||
@@ -1,139 +0,0 @@
|
|||||||
- block:
|
|
||||||
|
|
||||||
- name: Install gitea chart
|
|
||||||
kubernetes.core.helm:
|
|
||||||
name: gitea
|
|
||||||
chart_ref: /opt/metacluster/helm-charts/gitea
|
|
||||||
release_namespace: gitea
|
|
||||||
create_namespace: yes
|
|
||||||
wait: no
|
|
||||||
kubeconfig: "{{ kubeconfig.path }}"
|
|
||||||
values: "{{ components.gitea.chart_values }}"
|
|
||||||
|
|
||||||
- name: Ensure gitea API availability
|
|
||||||
ansible.builtin.uri:
|
|
||||||
url: https://git.{{ vapp['metacluster.fqdn'] }}/api/healthz
|
|
||||||
method: GET
|
|
||||||
register: api_readycheck
|
|
||||||
until:
|
|
||||||
- api_readycheck.json.status is defined
|
|
||||||
- api_readycheck.json.status == 'pass'
|
|
||||||
retries: "{{ playbook.retries }}"
|
|
||||||
delay: "{{ playbook.delays.long }}"
|
|
||||||
|
|
||||||
- name: Configure additional SSH ingress
|
|
||||||
ansible.builtin.template:
|
|
||||||
src: ingressroutetcp.j2
|
|
||||||
dest: /var/lib/rancher/k3s/server/manifests/{{ _template.name }}-manifest.yaml
|
|
||||||
owner: root
|
|
||||||
group: root
|
|
||||||
mode: 0600
|
|
||||||
vars:
|
|
||||||
_template:
|
|
||||||
name: gitea-ssh
|
|
||||||
namespace: gitea
|
|
||||||
config: |2
|
|
||||||
entryPoints:
|
|
||||||
- ssh
|
|
||||||
routes:
|
|
||||||
- match: HostSNI(`*`)
|
|
||||||
services:
|
|
||||||
- name: gitea-ssh
|
|
||||||
port: 22
|
|
||||||
notify:
|
|
||||||
- Apply manifests
|
|
||||||
|
|
||||||
- name: Trigger handlers
|
|
||||||
ansible.builtin.meta: flush_handlers
|
|
||||||
|
|
||||||
- name: Generate gitea API token
|
|
||||||
ansible.builtin.uri:
|
|
||||||
url: https://git.{{ vapp['metacluster.fqdn'] }}/api/v1/users/administrator/tokens
|
|
||||||
method: POST
|
|
||||||
user: administrator
|
|
||||||
password: "{{ vapp['metacluster.password'] }}"
|
|
||||||
force_basic_auth: yes
|
|
||||||
body:
|
|
||||||
name: token_init_{{ lookup('password', '/dev/null length=5 chars=ascii_letters,digits') }}
|
|
||||||
register: gitea_api_token
|
|
||||||
|
|
||||||
- name: Retrieve existing gitea configuration
|
|
||||||
ansible.builtin.uri:
|
|
||||||
url: https://git.{{ vapp['metacluster.fqdn'] }}/api/v1/repos/search
|
|
||||||
method: GET
|
|
||||||
register: gitea_existing_config
|
|
||||||
|
|
||||||
- block:
|
|
||||||
|
|
||||||
- name: Register SSH public key
|
|
||||||
ansible.builtin.uri:
|
|
||||||
url: https://git.{{ vapp['metacluster.fqdn'] }}/api/v1/user/keys
|
|
||||||
method: POST
|
|
||||||
headers:
|
|
||||||
Authorization: token {{ gitea_api_token.json.sha1 }}
|
|
||||||
body:
|
|
||||||
key: "{{ gitops_sshkey.public_key }}"
|
|
||||||
read_only: false
|
|
||||||
title: GitOps
|
|
||||||
|
|
||||||
- name: Create organization(s)
|
|
||||||
ansible.builtin.uri:
|
|
||||||
url: https://git.{{ vapp['metacluster.fqdn'] }}/api/v1/orgs
|
|
||||||
method: POST
|
|
||||||
headers:
|
|
||||||
Authorization: token {{ gitea_api_token.json.sha1 }}
|
|
||||||
body: "{{ item }}"
|
|
||||||
loop:
|
|
||||||
- full_name: Meta-cluster
|
|
||||||
description: Meta-cluster configuration items
|
|
||||||
username: mc
|
|
||||||
website: https://git.{{ vapp['metacluster.fqdn'] }}/mc
|
|
||||||
location: '[...]'
|
|
||||||
visibility: public
|
|
||||||
- full_name: Workload-cluster
|
|
||||||
description: Workload-cluster configuration items
|
|
||||||
username: wl
|
|
||||||
website: https://git.{{ vapp['metacluster.fqdn'] }}/wl
|
|
||||||
location: '[...]'
|
|
||||||
visibility: public
|
|
||||||
loop_control:
|
|
||||||
label: "{{ item.full_name }}"
|
|
||||||
|
|
||||||
- name: Create repositories
|
|
||||||
ansible.builtin.uri:
|
|
||||||
url: https://git.{{ vapp['metacluster.fqdn'] }}/api/v1/orgs/{{ item.organization }}/repos
|
|
||||||
method: POST
|
|
||||||
headers:
|
|
||||||
Authorization: token {{ gitea_api_token.json.sha1 }}
|
|
||||||
body: "{{ item.body }}"
|
|
||||||
loop:
|
|
||||||
- organization: mc
|
|
||||||
body:
|
|
||||||
name: GitOps.Config
|
|
||||||
# auto_init: true
|
|
||||||
# default_branch: main
|
|
||||||
description: GitOps manifests
|
|
||||||
- organization: wl
|
|
||||||
body:
|
|
||||||
name: Template.GitOps.Config
|
|
||||||
# auto_init: true
|
|
||||||
# default_branch: main
|
|
||||||
description: GitOps manifests
|
|
||||||
loop_control:
|
|
||||||
label: "{{ item.organization ~ '/' ~ item.body.name }}"
|
|
||||||
|
|
||||||
- name: Rebase/Push source gitops repository
|
|
||||||
ansible.builtin.shell:
|
|
||||||
cmd: |
|
|
||||||
git config --local http.sslVerify false
|
|
||||||
git remote set-url origin https://administrator:{{ vapp['metacluster.password'] | urlencode }}@git.{{ vapp['metacluster.fqdn'] }}/mc/GitOps.Config.git
|
|
||||||
git push
|
|
||||||
chdir: /opt/metacluster/git-repositories/gitops
|
|
||||||
|
|
||||||
when: (gitea_existing_config.json is undefined) or (gitea_existing_config.json.data | length == 0)
|
|
||||||
|
|
||||||
module_defaults:
|
|
||||||
ansible.builtin.uri:
|
|
||||||
validate_certs: no
|
|
||||||
status_code: [200, 201]
|
|
||||||
body_format: json
|
|
||||||
@@ -1,70 +0,0 @@
|
|||||||
- block:
|
|
||||||
|
|
||||||
- name: Install argo-cd chart
|
|
||||||
kubernetes.core.helm:
|
|
||||||
name: argo-cd
|
|
||||||
chart_ref: /opt/metacluster/helm-charts/argo-cd
|
|
||||||
release_namespace: argo-cd
|
|
||||||
create_namespace: yes
|
|
||||||
wait: no
|
|
||||||
kubeconfig: "{{ kubeconfig.path }}"
|
|
||||||
values: "{{ components.argocd.chart_values }}"
|
|
||||||
|
|
||||||
- name: Ensure argo-cd API availability
|
|
||||||
ansible.builtin.uri:
|
|
||||||
url: https://gitops.{{ vapp['metacluster.fqdn'] }}/api/version
|
|
||||||
method: GET
|
|
||||||
register: api_readycheck
|
|
||||||
until:
|
|
||||||
- api_readycheck.json.Version is defined
|
|
||||||
retries: "{{ playbook.retries }}"
|
|
||||||
delay: "{{ playbook.delays.long }}"
|
|
||||||
|
|
||||||
- name: Generate argo-cd API token
|
|
||||||
ansible.builtin.uri:
|
|
||||||
url: https://gitops.{{ vapp['metacluster.fqdn'] }}/api/v1/session
|
|
||||||
method: POST
|
|
||||||
force_basic_auth: yes
|
|
||||||
body:
|
|
||||||
username: admin
|
|
||||||
password: "{{ vapp['metacluster.password'] }}"
|
|
||||||
register: argocd_api_token
|
|
||||||
|
|
||||||
- name: Configure metacluster-gitops repository
|
|
||||||
ansible.builtin.template:
|
|
||||||
src: gitrepo.j2
|
|
||||||
dest: /var/lib/rancher/k3s/server/manifests/{{ _template.name }}-manifest.yaml
|
|
||||||
owner: root
|
|
||||||
group: root
|
|
||||||
mode: 0600
|
|
||||||
vars:
|
|
||||||
_template:
|
|
||||||
name: argocd-gitrepo-metacluster
|
|
||||||
namespace: argo-cd
|
|
||||||
uid: "{{ lookup('ansible.builtin.password', '/dev/null length=5 chars=ascii_lowercase,digits seed=inventory_hostname') }}"
|
|
||||||
privatekey: "{{ lookup('ansible.builtin.file', '~/.ssh/git_rsa_id') | indent(4, true) }}"
|
|
||||||
notify:
|
|
||||||
- Apply manifests
|
|
||||||
|
|
||||||
- name: Create applicationset
|
|
||||||
ansible.builtin.template:
|
|
||||||
src: applicationset.j2
|
|
||||||
dest: /var/lib/rancher/k3s/server/manifests/{{ _template.name }}-manifest.yaml
|
|
||||||
owner: root
|
|
||||||
group: root
|
|
||||||
mode: 0600
|
|
||||||
vars:
|
|
||||||
_template:
|
|
||||||
name: argocd-applicationset-metacluster
|
|
||||||
namespace: argo-cd
|
|
||||||
notify:
|
|
||||||
- Apply manifests
|
|
||||||
|
|
||||||
- name: Trigger handlers
|
|
||||||
ansible.builtin.meta: flush_handlers
|
|
||||||
|
|
||||||
module_defaults:
|
|
||||||
ansible.builtin.uri:
|
|
||||||
validate_certs: no
|
|
||||||
status_code: [200, 201]
|
|
||||||
body_format: json
|
|
||||||
@@ -1,26 +0,0 @@
|
|||||||
- name: Configure traefik dashboard ingress
|
|
||||||
ansible.builtin.template:
|
|
||||||
src: ingressroute.j2
|
|
||||||
dest: /var/lib/rancher/k3s/server/manifests/{{ _template.name }}-manifest.yaml
|
|
||||||
owner: root
|
|
||||||
group: root
|
|
||||||
mode: 0600
|
|
||||||
vars:
|
|
||||||
_template:
|
|
||||||
name: traefik-dashboard
|
|
||||||
namespace: kube-system
|
|
||||||
config: |2
|
|
||||||
entryPoints:
|
|
||||||
- web
|
|
||||||
- websecure
|
|
||||||
routes:
|
|
||||||
- kind: Rule
|
|
||||||
match: Host(`ingress.{{ vapp['metacluster.fqdn'] }}`)
|
|
||||||
services:
|
|
||||||
- kind: TraefikService
|
|
||||||
name: api@internal
|
|
||||||
notify:
|
|
||||||
- Apply manifests
|
|
||||||
|
|
||||||
- name: Trigger handlers
|
|
||||||
ansible.builtin.meta: flush_handlers
|
|
||||||
@@ -1,13 +0,0 @@
|
|||||||
- name: Configure fallback name resolution
|
|
||||||
ansible.builtin.lineinfile:
|
|
||||||
path: /etc/hosts
|
|
||||||
line: "{{ vapp['guestinfo.ipaddress'] }} {{ item ~ '.' ~ vapp['metacluster.fqdn'] }}"
|
|
||||||
state: present
|
|
||||||
loop:
|
|
||||||
# TODO: Make this list dynamic
|
|
||||||
- ca
|
|
||||||
- git
|
|
||||||
- gitops
|
|
||||||
- ingress
|
|
||||||
- registry
|
|
||||||
- storage
|
|
||||||
@@ -1,74 +0,0 @@
|
|||||||
- name: Store custom configuration files
|
|
||||||
ansible.builtin.copy:
|
|
||||||
dest: "{{ item.filename }}"
|
|
||||||
content: "{{ item.content }}"
|
|
||||||
loop:
|
|
||||||
- filename: /etc/rancher/k3s/config.yaml
|
|
||||||
content: |
|
|
||||||
kubelet-arg:
|
|
||||||
- "config=/etc/rancher/k3s/kubelet.config"
|
|
||||||
- filename: /etc/rancher/k3s/kubelet.config
|
|
||||||
content: |
|
|
||||||
apiVersion: kubelet.config.k8s.io/v1beta1
|
|
||||||
kind: KubeletConfiguration
|
|
||||||
|
|
||||||
shutdownGracePeriod: 180s
|
|
||||||
shtudownGracePeriodCriticalPods: 60s
|
|
||||||
loop_control:
|
|
||||||
label: "{{ item.filename }}"
|
|
||||||
|
|
||||||
- name: Gather service facts
|
|
||||||
ansible.builtin.service_facts:
|
|
||||||
# Module requires no attributes
|
|
||||||
|
|
||||||
- name: Install K3s
|
|
||||||
ansible.builtin.command:
|
|
||||||
cmd: ./install.sh
|
|
||||||
chdir: /opt/metacluster/k3s
|
|
||||||
environment:
|
|
||||||
INSTALL_K3S_SKIP_DOWNLOAD: 'true'
|
|
||||||
INSTALL_K3S_EXEC: "server --cluster-init --token {{ vapp['metacluster.token'] | trim }} --tls-san {{ vapp['metacluster.vip'] }} --disable local-storage --config /etc/rancher/k3s/config.yaml"
|
|
||||||
when: ansible_facts.services['k3s.service'] is undefined
|
|
||||||
|
|
||||||
- name: Debug possible taints on k3s node
|
|
||||||
ansible.builtin.shell:
|
|
||||||
cmd: >-
|
|
||||||
while true;
|
|
||||||
do
|
|
||||||
kubectl get nodes -o custom-columns=NAME:.metadata.name,TAINTS:.spec.taints --no-headers | awk '{print strftime("%H:%M:%S"),$0;fflush();}' >> /var/log/taintlog
|
|
||||||
sleep 1
|
|
||||||
done
|
|
||||||
async: 1800
|
|
||||||
poll: 0
|
|
||||||
|
|
||||||
- name: Ensure API availability
|
|
||||||
ansible.builtin.uri:
|
|
||||||
url: https://{{ vapp['guestinfo.ipaddress'] }}:6443/livez?verbose
|
|
||||||
method: GET
|
|
||||||
validate_certs: no
|
|
||||||
status_code: [200, 401]
|
|
||||||
register: api_readycheck
|
|
||||||
until: api_readycheck.json.apiVersion is defined
|
|
||||||
retries: "{{ playbook.retries }}"
|
|
||||||
delay: "{{ playbook.delays.medium }}"
|
|
||||||
|
|
||||||
- name: Install kubectl tab-completion
|
|
||||||
ansible.builtin.shell:
|
|
||||||
cmd: kubectl completion bash | tee /etc/bash_completion.d/kubectl
|
|
||||||
|
|
||||||
- name: Initialize tempfile
|
|
||||||
ansible.builtin.tempfile:
|
|
||||||
state: file
|
|
||||||
register: kubeconfig
|
|
||||||
|
|
||||||
- name: Retrieve kubeconfig
|
|
||||||
ansible.builtin.command:
|
|
||||||
cmd: kubectl config view --raw
|
|
||||||
register: kubectl_config
|
|
||||||
|
|
||||||
- name: Store kubeconfig in tempfile
|
|
||||||
ansible.builtin.copy:
|
|
||||||
dest: "{{ kubeconfig.path }}"
|
|
||||||
content: "{{ kubectl_config.stdout }}"
|
|
||||||
mode: 0600
|
|
||||||
no_log: true
|
|
||||||
@@ -1,27 +0,0 @@
|
|||||||
- name: Generate kube-vip manifest
|
|
||||||
ansible.builtin.shell:
|
|
||||||
cmd: >-
|
|
||||||
ctr run --rm --net-host ghcr.io/kube-vip/kube-vip:{{ components.kubevip.version }} vip \
|
|
||||||
/kube-vip manifest daemonset \
|
|
||||||
--interface eth0 \
|
|
||||||
--address {{ vapp['metacluster.vip'] }} \
|
|
||||||
--inCluster \
|
|
||||||
--taint \
|
|
||||||
--controlplane \
|
|
||||||
--services \
|
|
||||||
--arp \
|
|
||||||
--leaderElection
|
|
||||||
register: kubevip_manifest
|
|
||||||
|
|
||||||
- name: Inject manifests
|
|
||||||
ansible.builtin.copy:
|
|
||||||
dest: /var/lib/rancher/k3s/server/manifests/kubevip-manifest.yaml
|
|
||||||
content: |
|
|
||||||
{{ lookup('ansible.builtin.file', '/opt/metacluster/kube-vip/rbac.yaml') }}
|
|
||||||
---
|
|
||||||
{{ kubevip_manifest.stdout | replace('imagePullPolicy: Always', 'imagePullPolicy: IfNotPresent') }}
|
|
||||||
notify:
|
|
||||||
- Apply manifests
|
|
||||||
|
|
||||||
- name: Trigger handlers
|
|
||||||
ansible.builtin.meta: flush_handlers
|
|
||||||
@@ -1,10 +0,0 @@
|
|||||||
- import_tasks: init.yml
|
|
||||||
- import_tasks: k3s.yml
|
|
||||||
- import_tasks: assets.yml
|
|
||||||
- import_tasks: kube-vip.yml
|
|
||||||
- import_tasks: storage.yml
|
|
||||||
- import_tasks: ingress.yml
|
|
||||||
- import_tasks: certauthority.yml
|
|
||||||
- import_tasks: registry.yml
|
|
||||||
- import_tasks: git.yml
|
|
||||||
- import_tasks: gitops.yml
|
|
||||||
@@ -1,71 +0,0 @@
|
|||||||
- block:
|
|
||||||
|
|
||||||
- name: Install harbor chart
|
|
||||||
kubernetes.core.helm:
|
|
||||||
name: harbor
|
|
||||||
chart_ref: /opt/metacluster/helm-charts/harbor
|
|
||||||
release_namespace: harbor
|
|
||||||
create_namespace: yes
|
|
||||||
wait: no
|
|
||||||
kubeconfig: "{{ kubeconfig.path }}"
|
|
||||||
values: "{{ components.harbor.chart_values }}"
|
|
||||||
|
|
||||||
- name: Ensure harbor API availability
|
|
||||||
ansible.builtin.uri:
|
|
||||||
url: https://registry.{{ vapp['metacluster.fqdn'] }}/api/v2.0/health
|
|
||||||
method: GET
|
|
||||||
register: api_readycheck
|
|
||||||
until:
|
|
||||||
- api_readycheck.json.status is defined
|
|
||||||
- api_readycheck.json.status == 'healthy'
|
|
||||||
retries: "{{ playbook.retries }}"
|
|
||||||
delay: "{{ playbook.delays.long }}"
|
|
||||||
|
|
||||||
- name: Push images to registry
|
|
||||||
ansible.builtin.shell:
|
|
||||||
cmd: >-
|
|
||||||
skopeo copy \
|
|
||||||
--insecure-policy \
|
|
||||||
--dest-tls-verify=false \
|
|
||||||
--dest-creds admin:{{ vapp['metacluster.password'] }} \
|
|
||||||
docker-archive:./{{ item | basename }} \
|
|
||||||
docker://registry.{{ vapp['metacluster.fqdn'] }}/library/$( \
|
|
||||||
skopeo list-tags \
|
|
||||||
--insecure-policy \
|
|
||||||
docker-archive:./{{ item | basename }} | \
|
|
||||||
jq -r '.Tags[0]')
|
|
||||||
chdir: /opt/metacluster/container-images/
|
|
||||||
register: push_result
|
|
||||||
loop: "{{ query('ansible.builtin.fileglob', '/opt/metacluster/container-images/*.tar') | sort }}"
|
|
||||||
loop_control:
|
|
||||||
label: "{{ item | basename }}"
|
|
||||||
retries: "{{ playbook.retries }}"
|
|
||||||
delay: "{{ playbook.delays.short }}"
|
|
||||||
until: push_result is not failed
|
|
||||||
|
|
||||||
- name: Get all stored container images (=artifacts)
|
|
||||||
ansible.builtin.uri:
|
|
||||||
url: https://registry.{{ vapp['metacluster.fqdn'] }}/api/v2.0/search?q=library
|
|
||||||
method: GET
|
|
||||||
register: registry_artifacts
|
|
||||||
|
|
||||||
- name: Get source registries of all artifacts
|
|
||||||
ansible.builtin.set_fact:
|
|
||||||
source_registries: "{{ (source_registries | default([]) + [(item | split('/'))[1]]) | unique | sort }}"
|
|
||||||
loop: "{{ registry_artifacts.json.repository | json_query('[*].repository_name') }}"
|
|
||||||
|
|
||||||
- name: Configure K3s node for private registry
|
|
||||||
ansible.builtin.template:
|
|
||||||
dest: /etc/rancher/k3s/registries.yaml
|
|
||||||
src: registries.j2
|
|
||||||
vars:
|
|
||||||
_template:
|
|
||||||
data: "{{ source_registries }}"
|
|
||||||
hv:
|
|
||||||
fqdn: "{{ vapp['metacluster.fqdn'] }}"
|
|
||||||
|
|
||||||
module_defaults:
|
|
||||||
ansible.builtin.uri:
|
|
||||||
validate_certs: no
|
|
||||||
status_code: [200, 201, 401]
|
|
||||||
body_format: json
|
|
||||||
@@ -1,26 +0,0 @@
|
|||||||
- block:
|
|
||||||
- name: Install longhorn chart
|
|
||||||
kubernetes.core.helm:
|
|
||||||
name: longhorn
|
|
||||||
chart_ref: /opt/metacluster/helm-charts/longhorn
|
|
||||||
release_namespace: longhorn-system
|
|
||||||
create_namespace: yes
|
|
||||||
wait: no
|
|
||||||
kubeconfig: "{{ kubeconfig.path }}"
|
|
||||||
values: "{{ components.longhorn.chart_values }}"
|
|
||||||
|
|
||||||
- name: Ensure longhorn API availability
|
|
||||||
ansible.builtin.uri:
|
|
||||||
url: https://storage.{{ vapp['metacluster.fqdn'] }}/v1
|
|
||||||
method: GET
|
|
||||||
register: api_readycheck
|
|
||||||
until:
|
|
||||||
- api_readycheck is not failed
|
|
||||||
retries: "{{ playbook.retries }}"
|
|
||||||
delay: "{{ playbook.delays.long }}"
|
|
||||||
|
|
||||||
module_defaults:
|
|
||||||
ansible.builtin.uri:
|
|
||||||
validate_certs: no
|
|
||||||
status_code: [200, 201]
|
|
||||||
body_format: json
|
|
||||||
@@ -1,188 +0,0 @@
|
|||||||
- block:
|
|
||||||
# Below tasks circumvent usernames with `<domain>\<username>` format, which causes CAPV to
|
|
||||||
# incorrectly interpret the backslash (despite automatic escaping) as an escape sequence.
|
|
||||||
# `vcenter_session.user` will instead contain the username in `<username>@<domain>` format.
|
|
||||||
|
|
||||||
- name: Generate vCenter API token
|
|
||||||
ansible.builtin.uri:
|
|
||||||
url: https://{{ vapp['hv.fqdn'] }}/api/session
|
|
||||||
method: POST
|
|
||||||
headers:
|
|
||||||
Authorization: Basic {{ ( vapp['hv.username'] ~ ':' ~ vapp['hv.password'] ) | b64encode }}
|
|
||||||
register: vcenterapi_token
|
|
||||||
|
|
||||||
- name: Retrieve vCenter API session details
|
|
||||||
ansible.builtin.uri:
|
|
||||||
url: https://{{ vapp['hv.fqdn'] }}/api/session
|
|
||||||
method: GET
|
|
||||||
headers:
|
|
||||||
vmware-api-session-id: "{{ vcenterapi_token.json }}"
|
|
||||||
register: vcenter_session
|
|
||||||
|
|
||||||
module_defaults:
|
|
||||||
ansible.builtin.uri:
|
|
||||||
validate_certs: no
|
|
||||||
status_code: [200, 201]
|
|
||||||
body_format: json
|
|
||||||
|
|
||||||
- name: Configure clusterctl
|
|
||||||
ansible.builtin.template:
|
|
||||||
src: clusterctl.j2
|
|
||||||
dest: /opt/metacluster/cluster-api/clusterctl.yaml
|
|
||||||
vars:
|
|
||||||
_template:
|
|
||||||
version:
|
|
||||||
base: "{{ components.clusterapi.management.version.base }}"
|
|
||||||
cert_manager: "{{ components.clusterapi.management.version.cert_manager }}"
|
|
||||||
infrastructure_vsphere: "{{ components.clusterapi.management.version.infrastructure_vsphere }}"
|
|
||||||
ipam_incluster: "{{ components.clusterapi.management.version.ipam_incluster }}"
|
|
||||||
hv:
|
|
||||||
fqdn: "{{ vapp['hv.fqdn'] }}"
|
|
||||||
tlsthumbprint: "{{ tls_thumbprint.stdout }}"
|
|
||||||
username: "{{ vcenter_session.json.user }}"
|
|
||||||
password: "{{ vapp['hv.password'] }}"
|
|
||||||
datacenter: "{{ vcenter_info.datacenter }}"
|
|
||||||
datastore: "{{ vcenter_info.datastore }}"
|
|
||||||
network: "{{ vcenter_info.network }}"
|
|
||||||
resourcepool: "{{ vcenter_info.resourcepool }}"
|
|
||||||
folder: "{{ vcenter_info.folder }}"
|
|
||||||
cluster:
|
|
||||||
nodetemplate: "{{ (components.clusterapi.workload.node_template.url | basename | split('.'))[:-1] | join('.') }}"
|
|
||||||
publickey: "{{ vapp['guestinfo.rootsshkey'] }}"
|
|
||||||
version: "{{ components.clusterapi.workload.version.k8s }}"
|
|
||||||
vip: "{{ vapp['workloadcluster.vip'] }}"
|
|
||||||
|
|
||||||
- name: Update image references to use local registry
|
|
||||||
ansible.builtin.replace:
|
|
||||||
dest: "{{ item.root ~ '/' ~ item.path }}"
|
|
||||||
regexp: '([ ]+image:[ "]+)(?!({{ _template.pattern }}|"{{ _template.pattern }}))'
|
|
||||||
replace: '\1{{ _template.pattern }}'
|
|
||||||
vars:
|
|
||||||
_template:
|
|
||||||
pattern: registry.{{ vapp['metacluster.fqdn'] }}/library/
|
|
||||||
loop: "{{ lookup('community.general.filetree', '/opt/metacluster/cluster-api') }}"
|
|
||||||
loop_control:
|
|
||||||
label: "{{ item.path }}"
|
|
||||||
when:
|
|
||||||
- item.path is search('.yaml')
|
|
||||||
- item.path is not search("clusterctl.yaml|metadata.yaml")
|
|
||||||
|
|
||||||
- name: Generate kustomization template
|
|
||||||
ansible.builtin.template:
|
|
||||||
src: kustomization.cluster-template.j2
|
|
||||||
dest: /opt/metacluster/cluster-api/infrastructure-vsphere/{{ components.clusterapi.management.version.infrastructure_vsphere }}/kustomization.yaml
|
|
||||||
vars:
|
|
||||||
_template:
|
|
||||||
fqdn: "{{ vapp['metacluster.fqdn'] }}"
|
|
||||||
rootca: "{{ stepca_cm_certs.resources[0].data['root_ca.crt'] }}"
|
|
||||||
script:
|
|
||||||
# Base64 encoded; to avoid variable substitution when clusterctl parses the cluster-template.yml
|
|
||||||
encoded: IyEvYmluL2Jhc2gKdm10b29sc2QgLS1jbWQgJ2luZm8tZ2V0IGd1ZXN0aW5mby5vdmZFbnYnID4gL3RtcC9vdmZlbnYKCklQQWRkcmVzcz0kKHNlZCAtbiAncy8uKlByb3BlcnR5IG9lOmtleT0iZ3Vlc3RpbmZvLmludGVyZmFjZS4wLmlwLjAuYWRkcmVzcyIgb2U6dmFsdWU9IlwoW14iXSpcKS4qL1wxL3AnIC90bXAvb3ZmZW52KQpTdWJuZXRNYXNrPSQoc2VkIC1uICdzLy4qUHJvcGVydHkgb2U6a2V5PSJndWVzdGluZm8uaW50ZXJmYWNlLjAuaXAuMC5uZXRtYXNrIiBvZTp2YWx1ZT0iXChbXiJdKlwpLiovXDEvcCcgL3RtcC9vdmZlbnYpCkdhdGV3YXk9JChzZWQgLW4gJ3MvLipQcm9wZXJ0eSBvZTprZXk9Imd1ZXN0aW5mby5pbnRlcmZhY2UuMC5yb3V0ZS4wLmdhdGV3YXkiIG9lOnZhbHVlPSJcKFteIl0qXCkuKi9cMS9wJyAvdG1wL292ZmVudikKRE5TPSQoc2VkIC1uICdzLy4qUHJvcGVydHkgb2U6a2V5PSJndWVzdGluZm8uZG5zLnNlcnZlcnMiIG9lOnZhbHVlPSJcKFteIl0qXCkuKi9cMS9wJyAvdG1wL292ZmVudikKTUFDQWRkcmVzcz0kKHNlZCAtbiAncy8uKnZlOkFkYXB0ZXIgdmU6bWFjPSJcKFteIl0qXCkuKi9cMS9wJyAvdG1wL292ZmVudikKCm1hc2syY2lkcigpIHsKICBjPTAKICB4PTAkKCBwcmludGYgJyVvJyAkezEvLy4vIH0gKQoKICB3aGlsZSBbICR4IC1ndCAwIF07IGRvCiAgICBsZXQgYys9JCgoeCUyKSkgJ3g+Pj0xJwogIGRvbmUKCiAgZWNobyAkYwp9CgpQcmVmaXg9JChtYXNrMmNpZHIgJFN1Ym5ldE1hc2spCgpjYXQgPiAvZXRjL25ldHBsYW4vMDEtbmV0Y2ZnLnlhbWwgPDxFT0YKbmV0d29yazoKICB2ZXJzaW9uOiAyCiAgcmVuZGVyZXI6IG5ldHdvcmtkCiAgZXRoZXJuZXRzOgogICAgaWQwOgogICAgICBzZXQtbmFtZTogZXRoMAogICAgICBtYXRjaDoKICAgICAgICBtYWNhZGRyZXNzOiAkTUFDQWRkcmVzcwogICAgICBhZGRyZXNzZXM6CiAgICAgICAgLSAkSVBBZGRyZXNzLyRQcmVmaXgKICAgICAgZ2F0ZXdheTQ6ICRHYXRld2F5CiAgICAgIG5hbWVzZXJ2ZXJzOgogICAgICAgIGFkZHJlc3NlcyA6IFskRE5TXQpFT0YKcm0gL2V0Yy9uZXRwbGFuLzUwKi55YW1sIC1mCgpzdWRvIG5ldHBsYW4gYXBwbHk=
|
|
||||||
runcmds:
|
|
||||||
- update-ca-certificates
|
|
||||||
- bash /root/network.sh
|
|
||||||
|
|
||||||
- name: Store custom cluster-template
|
|
||||||
ansible.builtin.copy:
|
|
||||||
dest: /opt/metacluster/cluster-api/custom-cluster-template.yaml
|
|
||||||
content: "{{ lookup('kubernetes.core.kustomize', dir='/opt/metacluster/cluster-api/infrastructure-vsphere/' ~ components.clusterapi.management.version.infrastructure_vsphere ) }}"
|
|
||||||
|
|
||||||
- name: Initialize Cluster API management cluster
|
|
||||||
ansible.builtin.shell:
|
|
||||||
cmd: >-
|
|
||||||
clusterctl init \
|
|
||||||
-v5 \
|
|
||||||
--infrastructure vsphere:{{ components.clusterapi.management.version.infrastructure_vsphere }} \
|
|
||||||
--ipam in-cluster:{{ components.clusterapi.management.version.ipam_incluster }} \
|
|
||||||
--config ./clusterctl.yaml \
|
|
||||||
--kubeconfig {{ kubeconfig.path }}
|
|
||||||
chdir: /opt/metacluster/cluster-api
|
|
||||||
|
|
||||||
- name: Ensure CAPI/CAPV controller availability
|
|
||||||
kubernetes.core.k8s_info:
|
|
||||||
kind: Deployment
|
|
||||||
name: "{{ item.name }}"
|
|
||||||
namespace: "{{ item.namespace }}"
|
|
||||||
wait: true
|
|
||||||
kubeconfig: "{{ kubeconfig.path }}"
|
|
||||||
loop:
|
|
||||||
- name: capi-controller-manager
|
|
||||||
namespace: capi-system
|
|
||||||
- name: capv-controller-manager
|
|
||||||
namespace: capv-system
|
|
||||||
loop_control:
|
|
||||||
label: "{{ item.name }}"
|
|
||||||
|
|
||||||
- name: Parse vApp for workload cluster sizing
|
|
||||||
ansible.builtin.set_fact:
|
|
||||||
clustersize: >-
|
|
||||||
{{ {
|
|
||||||
'controlplane': vapp['deployment.type'] | regex_findall('^cp(\d)+') | first,
|
|
||||||
'workers': vapp['deployment.type'] | regex_findall('w(\d)+$') | first
|
|
||||||
} }}
|
|
||||||
|
|
||||||
- name: Generate workload cluster manifest
|
|
||||||
ansible.builtin.shell:
|
|
||||||
cmd: >-
|
|
||||||
clusterctl generate cluster \
|
|
||||||
{{ vapp['workloadcluster.name'] | lower }} \
|
|
||||||
--control-plane-machine-count {{ clustersize.controlplane }} \
|
|
||||||
--worker-machine-count {{ clustersize.workers }} \
|
|
||||||
--from ./custom-cluster-template.yaml \
|
|
||||||
--config ./clusterctl.yaml \
|
|
||||||
--kubeconfig {{ kubeconfig.path }}
|
|
||||||
chdir: /opt/metacluster/cluster-api
|
|
||||||
register: clusterctl_newcluster
|
|
||||||
|
|
||||||
# TODO: move to git repo
|
|
||||||
- name: Save workload cluster manifest
|
|
||||||
ansible.builtin.copy:
|
|
||||||
dest: /opt/metacluster/cluster-api/new-cluster.yaml
|
|
||||||
content: "{{ clusterctl_newcluster.stdout }}"
|
|
||||||
- name: Apply workload cluster manifest
|
|
||||||
kubernetes.core.k8s:
|
|
||||||
definition: >-
|
|
||||||
{{ clusterctl_newcluster.stdout }}
|
|
||||||
wait: yes
|
|
||||||
kubeconfig: "{{ kubeconfig.path }}"
|
|
||||||
# TODO: move to git repo
|
|
||||||
|
|
||||||
- name: Wait for cluster to be available
|
|
||||||
ansible.builtin.shell:
|
|
||||||
cmd: >-
|
|
||||||
kubectl wait clusters.cluster.x-k8s.io/{{ vapp['workloadcluster.name'] | lower }} \
|
|
||||||
--for=condition=Ready \
|
|
||||||
--timeout 0s
|
|
||||||
register: cluster_readycheck
|
|
||||||
until: cluster_readycheck is succeeded
|
|
||||||
retries: "{{ playbook.retries }}"
|
|
||||||
delay: "{{ playbook.delays.long }}"
|
|
||||||
|
|
||||||
- name: Initialize tempfile
|
|
||||||
ansible.builtin.tempfile:
|
|
||||||
state: file
|
|
||||||
register: capi_kubeconfig
|
|
||||||
|
|
||||||
- name: Retrieve kubeconfig
|
|
||||||
ansible.builtin.shell:
|
|
||||||
cmd: >-
|
|
||||||
clusterctl get kubeconfig \
|
|
||||||
{{ vapp['workloadcluster.name'] | lower }} \
|
|
||||||
--kubeconfig {{ kubeconfig.path }}
|
|
||||||
register: capi_kubectl_config
|
|
||||||
|
|
||||||
- name: Store kubeconfig in tempfile
|
|
||||||
ansible.builtin.copy:
|
|
||||||
dest: "{{ capi_kubeconfig.path }}"
|
|
||||||
content: "{{ capi_kubectl_config.stdout }}"
|
|
||||||
mode: 0600
|
|
||||||
no_log: true
|
|
||||||
|
|
||||||
# TODO: move to git repo
|
|
||||||
- name: Apply cni plugin manifest
|
|
||||||
kubernetes.core.k8s:
|
|
||||||
src: /opt/metacluster/cluster-api/cni-calico/{{ components.clusterapi.workload.version.calico }}/calico.yaml
|
|
||||||
state: present
|
|
||||||
wait: yes
|
|
||||||
kubeconfig: "{{ capi_kubeconfig.path }}"
|
|
||||||
# TODO: move to git repo
|
|
||||||
@@ -1,44 +0,0 @@
|
|||||||
- block:
|
|
||||||
|
|
||||||
- name: Generate service account in workload cluster
|
|
||||||
kubernetes.core.k8s:
|
|
||||||
template: serviceaccount.j2
|
|
||||||
state: present
|
|
||||||
|
|
||||||
- name: Retrieve service account bearer token
|
|
||||||
kubernetes.core.k8s_info:
|
|
||||||
kind: ServiceAccount
|
|
||||||
name: "{{ _template.account.name }}"
|
|
||||||
namespace: "{{ _template.account.namespace }}"
|
|
||||||
register: workloadcluster_serviceaccount
|
|
||||||
|
|
||||||
- name: Retrieve service account bearer token
|
|
||||||
kubernetes.core.k8s_info:
|
|
||||||
kind: Secret
|
|
||||||
name: "{{ workloadcluster_serviceaccount.resources | json_query('[].secrets[].name') | first }}"
|
|
||||||
namespace: "{{ _template.account.namespace }}"
|
|
||||||
register: workloadcluster_bearertoken
|
|
||||||
|
|
||||||
- name: Register workload cluster in argo-cd
|
|
||||||
kubernetes.core.k8s:
|
|
||||||
template: cluster.j2
|
|
||||||
state: present
|
|
||||||
kubeconfig: "{{ kubeconfig.path }}"
|
|
||||||
vars:
|
|
||||||
_template:
|
|
||||||
cluster:
|
|
||||||
name: "{{ vapp['workloadcluster.name'] | lower }}"
|
|
||||||
secret: argocd-cluster-{{ vapp['workloadcluster.name'] | lower }}
|
|
||||||
url: https://{{ vapp['workloadcluster.vip'] }}:6443
|
|
||||||
token: "{{ workloadcluster_bearertoken.resources | json_query('[].data.token') }}"
|
|
||||||
|
|
||||||
vars:
|
|
||||||
_template:
|
|
||||||
account:
|
|
||||||
name: argocd-sa
|
|
||||||
namespace: default
|
|
||||||
clusterrolebinding:
|
|
||||||
name: argocd-crb
|
|
||||||
module_defaults:
|
|
||||||
group/k8s:
|
|
||||||
kubeconfig: "{{ capi_kubeconfig.path }}"
|
|
||||||
@@ -1,75 +0,0 @@
|
|||||||
- name: Gather hypervisor details
|
|
||||||
ansible.builtin.shell:
|
|
||||||
cmd: govc ls -L {{ item.moref }} | awk -F/ '{print ${{ item.part }}}'
|
|
||||||
environment:
|
|
||||||
GOVC_INSECURE: '1'
|
|
||||||
GOVC_URL: "{{ vapp['hv.fqdn'] }}"
|
|
||||||
GOVC_USERNAME: "{{ vapp['hv.username'] }}"
|
|
||||||
GOVC_PASSWORD: "{{ vapp['hv.password'] }}"
|
|
||||||
register: govc_inventory
|
|
||||||
loop:
|
|
||||||
- attribute: cluster
|
|
||||||
moref: >-
|
|
||||||
$(govc object.collect -json VirtualMachine:{{ moref_id }} | \
|
|
||||||
jq -r '.[] | select(.Name == "runtime").Val.Host | .Type + ":" + .Value')
|
|
||||||
part: (NF-1)
|
|
||||||
- attribute: datacenter
|
|
||||||
moref: VirtualMachine:{{ moref_id }}
|
|
||||||
part: 2
|
|
||||||
- attribute: datastore
|
|
||||||
moref: >-
|
|
||||||
$(govc object.collect -json VirtualMachine:{{ moref_id }} | \
|
|
||||||
jq -r '.[] | select(.Name == "datastore").Val.ManagedObjectReference | .[].Type + ":" + .[].Value')
|
|
||||||
part: NF
|
|
||||||
- attribute: folder
|
|
||||||
moref: >-
|
|
||||||
$(govc object.collect -json VirtualMachine:{{ moref_id }} | \
|
|
||||||
jq -r '.[] | select(.Name == "parent").Val | .Type + ":" + .Value')
|
|
||||||
part: 0
|
|
||||||
# - attribute: host
|
|
||||||
# moref: >-
|
|
||||||
# $(govc object.collect -json VirtualMachine:{{ moref_id }} | \
|
|
||||||
# jq -r '.[] | select(.Name == "runtime").Val.Host | .Type + ":" + .Value')
|
|
||||||
# part: NF
|
|
||||||
- attribute: network
|
|
||||||
moref: >-
|
|
||||||
$(govc object.collect -json VirtualMachine:{{ moref_id }} | \
|
|
||||||
jq -r '.[] | select(.Name == "network").Val.ManagedObjectReference | .[].Type + ":" + .[].Value')
|
|
||||||
part: NF
|
|
||||||
- attribute: resourcepool
|
|
||||||
moref: >-
|
|
||||||
$(govc object.collect -json VirtualMachine:{{ moref_id }} | \
|
|
||||||
jq -r '.[] | select(.Name == "resourcePool").Val | .Type + ":" + .Value')
|
|
||||||
part: 0
|
|
||||||
loop_control:
|
|
||||||
label: "{{ item.attribute }}"
|
|
||||||
|
|
||||||
- name: Retrieve hypervisor TLS thumbprint
|
|
||||||
ansible.builtin.shell:
|
|
||||||
cmd: openssl s_client -connect {{ vapp['hv.fqdn'] }}:443 < /dev/null 2>/dev/null | openssl x509 -fingerprint -noout -in /dev/stdin | awk -F'=' '{print $2}'
|
|
||||||
register: tls_thumbprint
|
|
||||||
|
|
||||||
- name: Store hypervisor details in dictionary
|
|
||||||
ansible.builtin.set_fact:
|
|
||||||
vcenter_info: "{{ vcenter_info | default({}) | combine({ item.item.attribute : item.stdout }) }}"
|
|
||||||
loop: "{{ govc_inventory.results }}"
|
|
||||||
loop_control:
|
|
||||||
label: "{{ item.item.attribute }}"
|
|
||||||
|
|
||||||
- name: Configure network protocol profile on hypervisor
|
|
||||||
ansible.builtin.shell:
|
|
||||||
cmd: >-
|
|
||||||
npp-prepper \
|
|
||||||
--server "{{ vapp['hv.fqdn'] }}" \
|
|
||||||
--username "{{ vapp['hv.username'] }}" \
|
|
||||||
--password "{{ vapp['hv.password'] }}" \
|
|
||||||
dc \
|
|
||||||
--name "{{ vcenter_info.datacenter }}" \
|
|
||||||
--portgroup "{{ vcenter_info.network }}" \
|
|
||||||
--startaddress {{ vapp['ippool.startip'] }} \
|
|
||||||
--endaddress {{ vapp['ippool.endip'] }} \
|
|
||||||
--netmask {{ (vapp['guestinfo.ipaddress'] ~ '/' ~ vapp['guestinfo.prefixlength']) | ansible.utils.ipaddr('netmask') }} \
|
|
||||||
{{ vapp['guestinfo.dnsserver'] | split(',') | map('trim') | map('regex_replace', '^', '--dnsserver ') | join(' ') }} \
|
|
||||||
--dnsdomain {{ vapp['metacluster.fqdn'] }} \
|
|
||||||
--gateway {{ vapp['guestinfo.gateway'] }} \
|
|
||||||
--force
|
|
||||||
@@ -1,5 +0,0 @@
|
|||||||
- import_tasks: hypervisor.yml
|
|
||||||
- import_tasks: registry.yml
|
|
||||||
- import_tasks: nodetemplates.yml
|
|
||||||
- import_tasks: clusterapi.yml
|
|
||||||
- import_tasks: gitops.yml
|
|
||||||
@@ -1,85 +0,0 @@
|
|||||||
- block:
|
|
||||||
|
|
||||||
- name: Check for existing templates on hypervisor
|
|
||||||
community.vmware.vmware_guest_info:
|
|
||||||
name: "{{ (item | basename | split('.'))[:-1] | join('.') }}"
|
|
||||||
register: existing_ova
|
|
||||||
loop: "{{ query('ansible.builtin.fileglob', '/opt/workloadcluster/node-templates/*.ova') | sort }}"
|
|
||||||
ignore_errors: yes
|
|
||||||
|
|
||||||
- name: Parse OVA files for network mappings
|
|
||||||
ansible.builtin.shell:
|
|
||||||
cmd: govc import.spec -json {{ item }}
|
|
||||||
environment:
|
|
||||||
GOVC_INSECURE: '1'
|
|
||||||
GOVC_URL: "{{ vapp['hv.fqdn'] }}"
|
|
||||||
GOVC_USERNAME: "{{ vapp['hv.username'] }}"
|
|
||||||
GOVC_PASSWORD: "{{ vapp['hv.password'] }}"
|
|
||||||
register: ova_spec
|
|
||||||
when: existing_ova.results[index] is failed
|
|
||||||
loop: "{{ query('ansible.builtin.fileglob', '/opt/workloadcluster/node-templates/*.ova') | sort }}"
|
|
||||||
loop_control:
|
|
||||||
index_var: index
|
|
||||||
|
|
||||||
- name: Deploy OVA templates on hypervisor
|
|
||||||
community.vmware.vmware_deploy_ovf:
|
|
||||||
cluster: "{{ vcenter_info.cluster }}"
|
|
||||||
datastore: "{{ vcenter_info.datastore }}"
|
|
||||||
folder: "{{ vcenter_info.folder }}"
|
|
||||||
name: "{{ (item | basename | split('.'))[:-1] | join('.') }}"
|
|
||||||
networks: "{u'{{ ova_spec.results[index].stdout | from_json | json_query('NetworkMapping[0].Name') }}':u'{{ vcenter_info.network }}'}"
|
|
||||||
allow_duplicates: no
|
|
||||||
power_on: false
|
|
||||||
ovf: "{{ item }}"
|
|
||||||
register: ova_deploy
|
|
||||||
when: existing_ova.results[index] is failed
|
|
||||||
loop: "{{ query('ansible.builtin.fileglob', '/opt/workloadcluster/node-templates/*.ova') | sort }}"
|
|
||||||
loop_control:
|
|
||||||
index_var: index
|
|
||||||
|
|
||||||
- name: Add vApp properties on deployed VM's
|
|
||||||
ansible.builtin.shell:
|
|
||||||
cmd: >-
|
|
||||||
npp-prepper \
|
|
||||||
--server "{{ vapp['hv.fqdn'] }}" \
|
|
||||||
--username "{{ vapp['hv.username'] }}" \
|
|
||||||
--password "{{ vapp['hv.password'] }}" \
|
|
||||||
vm \
|
|
||||||
--datacenter "{{ vcenter_info.datacenter }}" \
|
|
||||||
--portgroup "{{ vcenter_info.network }}" \
|
|
||||||
--name "{{ item.instance.hw_name }}"
|
|
||||||
when: existing_ova.results[index] is failed
|
|
||||||
loop: "{{ ova_deploy.results }}"
|
|
||||||
loop_control:
|
|
||||||
index_var: index
|
|
||||||
label: "{{ item.item }}"
|
|
||||||
|
|
||||||
- name: Create snapshot on deployed VM's
|
|
||||||
community.vmware.vmware_guest_snapshot:
|
|
||||||
folder: "{{ vcenter_info.folder }}"
|
|
||||||
name: "{{ item.instance.hw_name }}"
|
|
||||||
state: present
|
|
||||||
snapshot_name: "{{ ansible_date_time.iso8601_basic_short }}-base"
|
|
||||||
when: ova_deploy.results[index] is not skipped
|
|
||||||
loop: "{{ ova_deploy.results }}"
|
|
||||||
loop_control:
|
|
||||||
index_var: index
|
|
||||||
label: "{{ item.item }}"
|
|
||||||
|
|
||||||
- name: Mark deployed VM's as templates
|
|
||||||
community.vmware.vmware_guest:
|
|
||||||
name: "{{ item.instance.hw_name }}"
|
|
||||||
is_template: yes
|
|
||||||
when: ova_deploy.results[index] is not skipped
|
|
||||||
loop: "{{ ova_deploy.results }}"
|
|
||||||
loop_control:
|
|
||||||
index_var: index
|
|
||||||
label: "{{ item.item }}"
|
|
||||||
|
|
||||||
module_defaults:
|
|
||||||
group/vmware:
|
|
||||||
hostname: "{{ vapp['hv.fqdn'] }}"
|
|
||||||
validate_certs: no
|
|
||||||
username: "{{ vapp['hv.username'] }}"
|
|
||||||
password: "{{ vapp['hv.password'] }}"
|
|
||||||
datacenter: "{{ vcenter_info.datacenter }}"
|
|
||||||
@@ -1,40 +0,0 @@
|
|||||||
- block:
|
|
||||||
|
|
||||||
- name: Create dedicated kubeadm project within container registry
|
|
||||||
ansible.builtin.uri:
|
|
||||||
url: https://registry.{{ vapp['metacluster.fqdn'] }}/api/v2.0/projects
|
|
||||||
method: POST
|
|
||||||
headers:
|
|
||||||
Authorization: "Basic {{ ('admin:' ~ vapp['metacluster.password']) | b64encode }}"
|
|
||||||
body:
|
|
||||||
project_name: kubeadm
|
|
||||||
public: true
|
|
||||||
storage_limit: 0
|
|
||||||
metadata:
|
|
||||||
enable_content_trust: 'false'
|
|
||||||
enable_content_trust_cosign: 'false'
|
|
||||||
auto_scan: 'true'
|
|
||||||
severity: none
|
|
||||||
prevent_vul: 'false'
|
|
||||||
public: 'true'
|
|
||||||
reuse_sys_cve_allowlist: 'true'
|
|
||||||
|
|
||||||
- name: Lookup kubeadm container images
|
|
||||||
ansible.builtin.set_fact:
|
|
||||||
kubeadm_images: "{{ lookup('ansible.builtin.file', '/opt/metacluster/cluster-api/imagelist').splitlines() }}"
|
|
||||||
|
|
||||||
- name: Copy kubeadm container images to dedicated project
|
|
||||||
ansible.builtin.uri:
|
|
||||||
url: https://registry.{{ vapp['metacluster.fqdn'] }}/api/v2.0/projects/kubeadm/repositories/{{ ( item | regex_findall('([^:/]+)') )[-2] }}/artifacts?from=library/{{ item | replace('/', '%2F') | replace(':', '%3A') }}
|
|
||||||
method: POST
|
|
||||||
headers:
|
|
||||||
Authorization: "Basic {{ ('admin:' ~ vapp['metacluster.password']) | b64encode }}"
|
|
||||||
body:
|
|
||||||
from: "{{ item }}"
|
|
||||||
loop: "{{ kubeadm_images }}"
|
|
||||||
|
|
||||||
module_defaults:
|
|
||||||
ansible.builtin.uri:
|
|
||||||
validate_certs: no
|
|
||||||
status_code: [200, 201, 409]
|
|
||||||
body_format: json
|
|
||||||
@@ -1,28 +0,0 @@
|
|||||||
apiVersion: argoproj.io/v1alpha1
|
|
||||||
kind: ApplicationSet
|
|
||||||
metadata:
|
|
||||||
name: {{ _template.name }}
|
|
||||||
namespace: {{ _template.namespace }}
|
|
||||||
spec:
|
|
||||||
generators:
|
|
||||||
- git:
|
|
||||||
repoURL: ssh://git@gitea-ssh.gitea.svc.cluster.local/mc/GitOps.Config.git
|
|
||||||
revision: HEAD
|
|
||||||
directories:
|
|
||||||
- path: metacluster-applicationset/*
|
|
||||||
template:
|
|
||||||
metadata:
|
|
||||||
name: {% raw %}'{{ path.basename }}'{% endraw +%}
|
|
||||||
spec:
|
|
||||||
project: default
|
|
||||||
syncPolicy:
|
|
||||||
automated:
|
|
||||||
prune: true
|
|
||||||
selfHeal: true
|
|
||||||
source:
|
|
||||||
repoURL: ssh://git@gitea-ssh.gitea.svc.cluster.local/mc/GitOps.Config.git
|
|
||||||
targetRevision: HEAD
|
|
||||||
path: {% raw %}'{{ path }}'{% endraw +%}
|
|
||||||
destination:
|
|
||||||
server: https://kubernetes.default.svc
|
|
||||||
namespace: default
|
|
||||||
@@ -1,18 +0,0 @@
|
|||||||
apiVersion: v1
|
|
||||||
kind: Secret
|
|
||||||
metadata:
|
|
||||||
name: {{ _template.cluster.secret }}
|
|
||||||
namespace: argo-cd
|
|
||||||
labels:
|
|
||||||
argocd.argoproj.io/secret-type: cluster
|
|
||||||
type: Opaque
|
|
||||||
stringData:
|
|
||||||
name: {{ _template.cluster.name }}
|
|
||||||
server: {{ _template.cluster.url }}
|
|
||||||
config: |
|
|
||||||
{
|
|
||||||
"bearerToken": "{{ _template.cluster.token }}",
|
|
||||||
"tlsClientConfig": {
|
|
||||||
"insecure": true
|
|
||||||
}
|
|
||||||
}
|
|
||||||
@@ -1,42 +0,0 @@
|
|||||||
providers:
|
|
||||||
- name: "kubeadm"
|
|
||||||
url: "/opt/metacluster/cluster-api/bootstrap-kubeadm/{{ _template.version.base }}/bootstrap-components.yaml"
|
|
||||||
type: "BootstrapProvider"
|
|
||||||
- name: "cluster-api"
|
|
||||||
url: "/opt/metacluster/cluster-api/cluster-api/{{ _template.version.base }}/core-components.yaml"
|
|
||||||
type: "CoreProvider"
|
|
||||||
- name: "kubeadm"
|
|
||||||
url: "/opt/metacluster/cluster-api/control-plane-kubeadm/{{ _template.version.base }}/control-plane-components.yaml"
|
|
||||||
type: "ControlPlaneProvider"
|
|
||||||
- name: "vsphere"
|
|
||||||
url: "/opt/metacluster/cluster-api/infrastructure-vsphere/{{ _template.version.infrastructure_vsphere }}/infrastructure-components.yaml"
|
|
||||||
type: "InfrastructureProvider"
|
|
||||||
- name: "in-cluster"
|
|
||||||
url: "/opt/metacluster/cluster-api/ipam-in-cluster/{{ _template.version.ipam_incluster }}/ipam-components.yaml"
|
|
||||||
type: "IPAMProvider"
|
|
||||||
|
|
||||||
cert-manager:
|
|
||||||
url: "/opt/metacluster/cluster-api/cert-manager/{{ _template.version.cert_manager }}/cert-manager.yaml"
|
|
||||||
version: "{{ _template.version.cert_manager }}"
|
|
||||||
|
|
||||||
## -- Controller settings -- ##
|
|
||||||
VSPHERE_SERVER: "{{ _template.hv.fqdn }}"
|
|
||||||
VSPHERE_TLS_THUMBPRINT: "{{ _template.hv.tlsthumbprint }}"
|
|
||||||
VSPHERE_USERNAME: "{{ _template.hv.username }}"
|
|
||||||
VSPHERE_PASSWORD: "{{ _template.hv.password }}"
|
|
||||||
|
|
||||||
## -- Required workload cluster default settings -- ##
|
|
||||||
VSPHERE_DATACENTER: "{{ _template.hv.datacenter }}"
|
|
||||||
VSPHERE_DATASTORE: "{{ _template.hv.datastore }}"
|
|
||||||
VSPHERE_STORAGE_POLICY: ""
|
|
||||||
VSPHERE_NETWORK: "{{ _template.hv.network }}"
|
|
||||||
VSPHERE_RESOURCE_POOL: "{{ _template.hv.resourcepool }}"
|
|
||||||
VSPHERE_FOLDER: "{{ _template.hv.folder }}"
|
|
||||||
|
|
||||||
VSPHERE_TEMPLATE: "{{ _template.cluster.nodetemplate }}"
|
|
||||||
VSPHERE_SSH_AUTHORIZED_KEY: "{{ _template.cluster.publickey }}"
|
|
||||||
|
|
||||||
KUBERNETES_VERSION: "{{ _template.cluster.version }}"
|
|
||||||
CONTROL_PLANE_ENDPOINT_IP: "{{ _template.cluster.vip }}"
|
|
||||||
VIP_NETWORK_INTERFACE: ""
|
|
||||||
EXP_CLUSTER_RESOURCE_SET: "true"
|
|
||||||
@@ -1,14 +0,0 @@
|
|||||||
apiVersion: v1
|
|
||||||
kind: ConfigMap
|
|
||||||
metadata:
|
|
||||||
name: {{ _template.name }}
|
|
||||||
namespace: {{ _template.namespace }}
|
|
||||||
annotations:
|
|
||||||
{{ _template.annotations }}
|
|
||||||
labels:
|
|
||||||
{{ _template.labels }}
|
|
||||||
data:
|
|
||||||
{% for kv_pair in _template.data %}
|
|
||||||
"{{ kv_pair.key }}": |
|
|
||||||
{{ kv_pair.value | indent(width=4, first=True) }}
|
|
||||||
{% endfor %}
|
|
||||||
@@ -1,13 +0,0 @@
|
|||||||
apiVersion: v1
|
|
||||||
kind: Secret
|
|
||||||
metadata:
|
|
||||||
name: {{ _template.name }}-{{ _template.uid }}
|
|
||||||
namespace: {{ _template.namespace }}
|
|
||||||
labels:
|
|
||||||
argocd.argoproj.io/secret-type: repository
|
|
||||||
stringData:
|
|
||||||
url: ssh://git@gitea-ssh.gitea.svc.cluster.local/mc/GitOps.Config.git
|
|
||||||
name: {{ _template.name }}
|
|
||||||
insecure: 'true'
|
|
||||||
sshPrivateKey: |
|
|
||||||
{{ _template.privatekey }}
|
|
||||||
@@ -1,7 +0,0 @@
|
|||||||
apiVersion: traefik.containo.us/v1alpha1
|
|
||||||
kind: IngressRoute
|
|
||||||
metadata:
|
|
||||||
name: {{ _template.name }}
|
|
||||||
namespace: {{ _template.namespace }}
|
|
||||||
spec:
|
|
||||||
{{ _template.config }}
|
|
||||||
@@ -1,7 +0,0 @@
|
|||||||
apiVersion: traefik.containo.us/v1alpha1
|
|
||||||
kind: IngressRouteTCP
|
|
||||||
metadata:
|
|
||||||
name: {{ _template.name }}
|
|
||||||
namespace: {{ _template.namespace }}
|
|
||||||
spec:
|
|
||||||
{{ _template.config }}
|
|
||||||
@@ -1,104 +0,0 @@
|
|||||||
apiVersion: kustomize.config.k8s.io/v1beta1
|
|
||||||
kind: Kustomization
|
|
||||||
resources:
|
|
||||||
- cluster-template.yaml
|
|
||||||
|
|
||||||
patchesStrategicMerge:
|
|
||||||
- |-
|
|
||||||
apiVersion: controlplane.cluster.x-k8s.io/v1beta1
|
|
||||||
kind: KubeadmControlPlane
|
|
||||||
metadata:
|
|
||||||
name: '${CLUSTER_NAME}'
|
|
||||||
namespace: '${NAMESPACE}'
|
|
||||||
spec:
|
|
||||||
kubeadmConfigSpec:
|
|
||||||
clusterConfiguration:
|
|
||||||
imageRepository: registry.{{ _template.fqdn }}/kubeadm
|
|
||||||
- |-
|
|
||||||
apiVersion: bootstrap.cluster.x-k8s.io/v1beta1
|
|
||||||
kind: KubeadmConfigTemplate
|
|
||||||
metadata:
|
|
||||||
name: '${CLUSTER_NAME}-md-0'
|
|
||||||
namespace: '${NAMESPACE}'
|
|
||||||
spec:
|
|
||||||
template:
|
|
||||||
spec:
|
|
||||||
clusterConfiguration:
|
|
||||||
imageRepository: registry.{{ _template.fqdn }}/kubeadm
|
|
||||||
- |-
|
|
||||||
apiVersion: bootstrap.cluster.x-k8s.io/v1beta1
|
|
||||||
kind: KubeadmConfigTemplate
|
|
||||||
metadata:
|
|
||||||
name: '${CLUSTER_NAME}-md-0'
|
|
||||||
namespace: '${NAMESPACE}'
|
|
||||||
spec:
|
|
||||||
template:
|
|
||||||
spec:
|
|
||||||
files:
|
|
||||||
- encoding: base64
|
|
||||||
content: |
|
|
||||||
{{ _template.script.encoded }}
|
|
||||||
permissions: '0744'
|
|
||||||
owner: root:root
|
|
||||||
path: /root/network.sh
|
|
||||||
- content: |
|
|
||||||
network: {config: disabled}
|
|
||||||
owner: root:root
|
|
||||||
path: /etc/cloud/cloud.cfg.d/99-disable-network-config.cfg
|
|
||||||
- content: |
|
|
||||||
{{ _template.rootca | indent(width=14, first=False) | trim }}
|
|
||||||
owner: root:root
|
|
||||||
path: /usr/local/share/ca-certificates/root_ca.crt
|
|
||||||
|
|
||||||
patchesJson6902:
|
|
||||||
- target:
|
|
||||||
group: controlplane.cluster.x-k8s.io
|
|
||||||
version: v1beta1
|
|
||||||
kind: KubeadmControlPlane
|
|
||||||
name: .*
|
|
||||||
patch: |-
|
|
||||||
- op: add
|
|
||||||
path: /spec/kubeadmConfigSpec/files/-
|
|
||||||
value:
|
|
||||||
encoding: base64
|
|
||||||
content: |
|
|
||||||
{{ _template.script.encoded }}
|
|
||||||
owner: root:root
|
|
||||||
path: /root/network.sh
|
|
||||||
permissions: '0744'
|
|
||||||
- op: add
|
|
||||||
path: /spec/kubeadmConfigSpec/files/-
|
|
||||||
value:
|
|
||||||
content: |
|
|
||||||
network: {config: disabled}
|
|
||||||
owner: root:root
|
|
||||||
path: /etc/cloud/cloud.cfg.d/99-disable-network-config.cfg
|
|
||||||
- op: add
|
|
||||||
path: /spec/kubeadmConfigSpec/files/-
|
|
||||||
value:
|
|
||||||
content: |
|
|
||||||
{{ _template.rootca | indent(width=12, first=False) | trim }}
|
|
||||||
owner: root:root
|
|
||||||
path: /usr/local/share/ca-certificates/root_ca.crt
|
|
||||||
- target:
|
|
||||||
group: bootstrap.cluster.x-k8s.io
|
|
||||||
version: v1beta1
|
|
||||||
kind: KubeadmConfigTemplate
|
|
||||||
name: .*
|
|
||||||
patch: |-
|
|
||||||
{% for cmd in _template.runcmds %}
|
|
||||||
- op: add
|
|
||||||
path: /spec/template/spec/preKubeadmCommands/-
|
|
||||||
value: {{ cmd }}
|
|
||||||
{% endfor %}
|
|
||||||
- target:
|
|
||||||
group: controlplane.cluster.x-k8s.io
|
|
||||||
version: v1beta1
|
|
||||||
kind: KubeadmControlPlane
|
|
||||||
name: .*
|
|
||||||
patch: |-
|
|
||||||
{% for cmd in _template.runcmds %}
|
|
||||||
- op: add
|
|
||||||
path: /spec/kubeadmConfigSpec/preKubeadmCommands/-
|
|
||||||
value: {{ cmd }}
|
|
||||||
{% endfor %}
|
|
||||||
@@ -1,9 +0,0 @@
|
|||||||
apiVersion: v1
|
|
||||||
kind: Secret
|
|
||||||
metadata:
|
|
||||||
name: {{ _template.name }}
|
|
||||||
namespace: {{ _template.namespace }}
|
|
||||||
data:
|
|
||||||
{% for kv_pair in _template.data %}
|
|
||||||
"{{ kv_pair.key }}": {{ kv_pair.value }}
|
|
||||||
{% endfor %}
|
|
||||||
@@ -1,18 +0,0 @@
|
|||||||
apiVersion: v1
|
|
||||||
kind: ServiceAccount
|
|
||||||
metadata:
|
|
||||||
name: {{ _template.account.name }}
|
|
||||||
namespace: {{ _template.account.namespace }}
|
|
||||||
---
|
|
||||||
apiVersion: rbac.authorization.k8s.io/v1
|
|
||||||
kind: ClusterRoleBinding
|
|
||||||
metadata:
|
|
||||||
name: {{ _template.clusterrolebinding.name }}
|
|
||||||
subjects:
|
|
||||||
- kind: ServiceAccount
|
|
||||||
name: {{ _template.account.name }}
|
|
||||||
namespace: {{ _template.account.namespace }}
|
|
||||||
roleRef:
|
|
||||||
kind: ClusterRole
|
|
||||||
name: cluster-admin
|
|
||||||
apiGroup: rbac.authorization.k8s.io
|
|
||||||
@@ -1,4 +0,0 @@
|
|||||||
- name: Disable crontab job
|
|
||||||
ansible.builtin.cron:
|
|
||||||
name: firstboot
|
|
||||||
state: absent
|
|
||||||
@@ -1,12 +0,0 @@
|
|||||||
- import_tasks: service.yml
|
|
||||||
- import_tasks: cron.yml
|
|
||||||
|
|
||||||
- name: Cleanup tempfile
|
|
||||||
ansible.builtin.file:
|
|
||||||
path: "{{ kubeconfig.path }}"
|
|
||||||
state: absent
|
|
||||||
when: kubeconfig.path is defined
|
|
||||||
|
|
||||||
# - name: Reboot host
|
|
||||||
# ansible.builtin.shell:
|
|
||||||
# cmd: systemctl reboot
|
|
||||||
@@ -1,30 +0,0 @@
|
|||||||
- name: Create tarball compression service
|
|
||||||
ansible.builtin.template:
|
|
||||||
src: "{{ item.src }}"
|
|
||||||
dest: "{{ item.dest }}"
|
|
||||||
owner: root
|
|
||||||
group: root
|
|
||||||
mode: "{{ item.mode | default(omit) }}"
|
|
||||||
vars:
|
|
||||||
_template:
|
|
||||||
service:
|
|
||||||
name: compressTarballs
|
|
||||||
executable: /opt/firstboot/compresstarballs.sh
|
|
||||||
workingdir: /opt/metacluster/container-images/
|
|
||||||
loop:
|
|
||||||
- src: compresstarballs.j2
|
|
||||||
dest: "{{ _template.service.executable }}"
|
|
||||||
mode: o+x
|
|
||||||
- src: systemdunit.j2
|
|
||||||
dest: /etc/systemd/system/{{ _template.service.name }}.service
|
|
||||||
loop_control:
|
|
||||||
label: "{{ item.src }}"
|
|
||||||
|
|
||||||
- name: Enable/Start services
|
|
||||||
ansible.builtin.systemd:
|
|
||||||
name: "{{ item }}"
|
|
||||||
enabled: yes
|
|
||||||
state: started
|
|
||||||
loop:
|
|
||||||
- compressTarballs
|
|
||||||
- ttyConsoleMessage
|
|
||||||
@@ -1,24 +0,0 @@
|
|||||||
- name: Create volume group
|
|
||||||
community.general.lvg:
|
|
||||||
vg: longhorn_vg
|
|
||||||
pvs:
|
|
||||||
- /dev/sdb
|
|
||||||
pvresize: yes
|
|
||||||
|
|
||||||
- name: Create logical volume
|
|
||||||
community.general.lvol:
|
|
||||||
vg: longhorn_vg
|
|
||||||
lv: longhorn_lv
|
|
||||||
size: 100%VG
|
|
||||||
|
|
||||||
- name: Create filesystem
|
|
||||||
community.general.filesystem:
|
|
||||||
dev: /dev/mapper/longhorn_vg-longhorn_lv
|
|
||||||
fstype: ext4
|
|
||||||
|
|
||||||
- name: Mount dynamic disk
|
|
||||||
ansible.posix.mount:
|
|
||||||
path: /mnt/blockstorage
|
|
||||||
src: /dev/mapper/longhorn_vg-longhorn_lv
|
|
||||||
fstype: ext4
|
|
||||||
state: mounted
|
|
||||||
@@ -1,12 +0,0 @@
|
|||||||
- name: Import container images
|
|
||||||
ansible.builtin.command:
|
|
||||||
cmd: k3s ctr image import {{ item }} --digests
|
|
||||||
chdir: /opt/metacluster/container-images
|
|
||||||
register: import_result
|
|
||||||
loop: "{{ query('ansible.builtin.fileglob', '/opt/metacluster/container-images/*.tar') | sort }}"
|
|
||||||
loop_control:
|
|
||||||
label: "{{ item | basename }}"
|
|
||||||
# Probably should add a task before that ensures K3s node is fully initialized before starting imports; currently K3s goes away briefly during this loop
|
|
||||||
retries: "{{ playbook.retries }}"
|
|
||||||
delay: "{{ playbook.delays.short }}"
|
|
||||||
until: import_result is not failed
|
|
||||||
@@ -1,19 +0,0 @@
|
|||||||
- name: Set hostname
|
|
||||||
ansible.builtin.hostname:
|
|
||||||
name: "{{ vapp['guestinfo.hostname'] }}"
|
|
||||||
|
|
||||||
- name: Create netplan configuration file
|
|
||||||
ansible.builtin.template:
|
|
||||||
src: netplan.j2
|
|
||||||
dest: /etc/netplan/00-installer-config.yaml
|
|
||||||
vars:
|
|
||||||
_template:
|
|
||||||
macaddress: "{{ ansible_facts.default_ipv4.macaddress }}"
|
|
||||||
ipaddress: "{{ vapp['guestinfo.ipaddress'] }}"
|
|
||||||
prefixlength: "{{ vapp['guestinfo.prefixlength'] }}"
|
|
||||||
gateway: "{{ vapp['guestinfo.gateway'] }}"
|
|
||||||
dnsserver: "{{ vapp['guestinfo.dnsserver'] }}"
|
|
||||||
|
|
||||||
- name: Apply netplan configuration
|
|
||||||
ansible.builtin.shell:
|
|
||||||
cmd: /usr/sbin/netplan apply
|
|
||||||
@@ -1,13 +0,0 @@
|
|||||||
network:
|
|
||||||
version: 2
|
|
||||||
ethernets:
|
|
||||||
id0:
|
|
||||||
set-name: eth0
|
|
||||||
match:
|
|
||||||
macaddress: {{ _template.macaddress }}
|
|
||||||
addresses:
|
|
||||||
- {{ _template.ipaddress }}/{{ _template.prefixlength }}
|
|
||||||
gateway4: {{ _template.gateway }}
|
|
||||||
nameservers:
|
|
||||||
addresses:
|
|
||||||
- {{ _template.dnsserver }}
|
|
||||||
@@ -1 +0,0 @@
|
|||||||
- import_tasks: vcenter.yml
|
|
||||||
@@ -1,16 +0,0 @@
|
|||||||
- block:
|
|
||||||
|
|
||||||
- name: Check for vCenter connectivity
|
|
||||||
community.vmware.vmware_vcenter_settings_info:
|
|
||||||
schema: vsphere
|
|
||||||
register: vcenter_info
|
|
||||||
retries: "{{ playbook.retries }}"
|
|
||||||
delay: "{{ playbook.delays.short }}"
|
|
||||||
until: vcenter_info is not failed
|
|
||||||
|
|
||||||
module_defaults:
|
|
||||||
group/vmware:
|
|
||||||
hostname: "{{ vapp['hv.fqdn'] }}"
|
|
||||||
validate_certs: no
|
|
||||||
username: "{{ vapp['hv.username'] }}"
|
|
||||||
password: "{{ vapp['hv.password'] }}"
|
|
||||||
@@ -1,31 +0,0 @@
|
|||||||
- name: Create folder structure(s)
|
|
||||||
ansible.builtin.file:
|
|
||||||
path: "{{ item }}"
|
|
||||||
state: directory
|
|
||||||
loop:
|
|
||||||
- /opt/firstboot
|
|
||||||
|
|
||||||
- name: Create tty console message service
|
|
||||||
ansible.builtin.template:
|
|
||||||
src: "{{ item.src }}"
|
|
||||||
dest: "{{ item.dest }}"
|
|
||||||
owner: root
|
|
||||||
group: root
|
|
||||||
mode: "{{ item.mode | default(omit) }}"
|
|
||||||
vars:
|
|
||||||
_template:
|
|
||||||
service:
|
|
||||||
name: ttyConsoleMessage
|
|
||||||
executable: /opt/firstboot/tty.sh
|
|
||||||
workingdir: /tmp/
|
|
||||||
metacluster:
|
|
||||||
fqdn: "{{ vapp['metacluster.fqdn'] }}"
|
|
||||||
vip: "{{ vapp['metacluster.vip'] }}"
|
|
||||||
loop:
|
|
||||||
- src: tty.j2
|
|
||||||
dest: "{{ _template.service.executable }}"
|
|
||||||
mode: o+x
|
|
||||||
- src: systemdunit.j2
|
|
||||||
dest: /etc/systemd/system/{{ _template.service.name }}.service
|
|
||||||
loop_control:
|
|
||||||
label: "{{ item.src }}"
|
|
||||||
@@ -1,39 +0,0 @@
|
|||||||
- name: Set root password
|
|
||||||
ansible.builtin.user:
|
|
||||||
name: root
|
|
||||||
password: "{{ vapp['metacluster.password'] | password_hash('sha512', 65534 | random(seed=vapp['guestinfo.hostname']) | string) }}"
|
|
||||||
generate_ssh_key: yes
|
|
||||||
ssh_key_bits: 2048
|
|
||||||
ssh_key_file: .ssh/id_rsa
|
|
||||||
|
|
||||||
- name: Save root SSH publickey
|
|
||||||
ansible.builtin.lineinfile:
|
|
||||||
path: /root/.ssh/authorized_keys
|
|
||||||
line: "{{ vapp['guestinfo.rootsshkey'] }}"
|
|
||||||
|
|
||||||
- name: Disable SSH password authentication
|
|
||||||
ansible.builtin.lineinfile:
|
|
||||||
path: /etc/ssh/sshd_config
|
|
||||||
regex: "{{ item.regex }}"
|
|
||||||
line: "{{ item.line }}"
|
|
||||||
state: "{{ item.state }}"
|
|
||||||
loop:
|
|
||||||
- regex: '^#PasswordAuthentication'
|
|
||||||
line: 'PasswordAuthentication no'
|
|
||||||
state: present
|
|
||||||
- regex: '^PasswordAuthentication yes'
|
|
||||||
line: 'PasswordAuthentication yes'
|
|
||||||
state: absent
|
|
||||||
loop_control:
|
|
||||||
label: "{{ '[' ~ item.regex ~ '] ' ~ item.state }}"
|
|
||||||
|
|
||||||
- name: Create dedicated SSH keypair
|
|
||||||
community.crypto.openssh_keypair:
|
|
||||||
path: /root/.ssh/git_rsa_id
|
|
||||||
register: gitops_sshkey
|
|
||||||
|
|
||||||
- name: Delete 'ubuntu' user
|
|
||||||
ansible.builtin.user:
|
|
||||||
name: ubuntu
|
|
||||||
state: absent
|
|
||||||
remove: yes
|
|
||||||
@@ -1,38 +0,0 @@
|
|||||||
- name: Store current ovfEnvironment
|
|
||||||
ansible.builtin.shell:
|
|
||||||
cmd: /usr/bin/vmtoolsd --cmd "info-get guestinfo.ovfEnv"
|
|
||||||
register: ovfenv
|
|
||||||
|
|
||||||
- name: Parse XML for MoRef ID
|
|
||||||
community.general.xml:
|
|
||||||
xmlstring: "{{ ovfenv.stdout }}"
|
|
||||||
namespaces:
|
|
||||||
ns: http://schemas.dmtf.org/ovf/environment/1
|
|
||||||
ve: http://www.vmware.com/schema/ovfenv
|
|
||||||
xpath: /ns:Environment
|
|
||||||
content: attribute
|
|
||||||
register: environment_attribute
|
|
||||||
|
|
||||||
- name: Store MoRef ID
|
|
||||||
ansible.builtin.set_fact:
|
|
||||||
moref_id: "{{ ((environment_attribute.matches[0].values() | list)[0].values() | list)[1] }}"
|
|
||||||
|
|
||||||
- name: Parse XML for vApp properties
|
|
||||||
community.general.xml:
|
|
||||||
xmlstring: "{{ ovfenv.stdout }}"
|
|
||||||
namespaces:
|
|
||||||
ns: http://schemas.dmtf.org/ovf/environment/1
|
|
||||||
xpath: /ns:Environment/ns:PropertySection/ns:Property
|
|
||||||
content: attribute
|
|
||||||
register: property_section
|
|
||||||
|
|
||||||
- name: Assign vApp properties to dictionary
|
|
||||||
ansible.builtin.set_fact:
|
|
||||||
vapp: >-
|
|
||||||
{{ vapp | default({}) | combine({
|
|
||||||
((item.values() | list)[0].values() | list)[0]:
|
|
||||||
((item.values() | list)[0].values() | list)[1]})
|
|
||||||
}}
|
|
||||||
loop: "{{ property_section.matches }}"
|
|
||||||
loop_control:
|
|
||||||
label: "{{ ((item.values() | list)[0].values() | list)[0] }}"
|
|
||||||
@@ -1,10 +0,0 @@
|
|||||||
#!/bin/bash
|
|
||||||
|
|
||||||
# Change working directory
|
|
||||||
pushd {{ _template.service.workingdir }}
|
|
||||||
|
|
||||||
# Compress *.tar files
|
|
||||||
if tar -czf image-tarballs.tgz *.tar --remove-files; then
|
|
||||||
# Disable systemd unit
|
|
||||||
systemctl disable {{ _template.service.name }}
|
|
||||||
fi
|
|
||||||
@@ -1,8 +0,0 @@
|
|||||||
mirrors:
|
|
||||||
{% for entry in _template.data %}
|
|
||||||
{{ entry }}:
|
|
||||||
endpoint:
|
|
||||||
- https://registry.{{ _template.hv.fqdn }}
|
|
||||||
rewrite:
|
|
||||||
"(.*)": "library/{{ entry }}/$1"
|
|
||||||
{% endfor %}
|
|
||||||
@@ -1,9 +0,0 @@
|
|||||||
[Unit]
|
|
||||||
Description={{ _template.service.name }}
|
|
||||||
|
|
||||||
[Service]
|
|
||||||
ExecStart={{ _template.service.executable }}
|
|
||||||
Nice=10
|
|
||||||
|
|
||||||
[Install]
|
|
||||||
WantedBy=multi-user.target
|
|
||||||
@@ -1,47 +0,0 @@
|
|||||||
#!/bin/bash
|
|
||||||
|
|
||||||
export TERM=linux
|
|
||||||
|
|
||||||
BGRN='\033[1;92m'
|
|
||||||
BGRY='\033[1;30m'
|
|
||||||
BBLU='\033[1;34m'
|
|
||||||
BRED='\033[1;91m'
|
|
||||||
BWHI='\033[1;97m'
|
|
||||||
CBLA='\033[?16;0;30c' # Hide blinking cursor
|
|
||||||
DFLT='\033[0m' # Reset colour
|
|
||||||
LCLR='\033[K' # Clear to end of line
|
|
||||||
PRST='\033[0;0H' # Reset cursor position
|
|
||||||
|
|
||||||
# COMPONENTS=('ca' 'ingress' 'storage' 'registry' 'git' 'gitops')
|
|
||||||
COMPONENTS=('storage' 'registry' 'git' 'gitops')
|
|
||||||
FQDN='{{ _template.metacluster.fqdn }}'
|
|
||||||
IPADDRESS='{{ _template.metacluster.vip }}'
|
|
||||||
|
|
||||||
I=0
|
|
||||||
|
|
||||||
while /bin/true; do
|
|
||||||
if [[ $I -gt 59 ]]; then
|
|
||||||
clear > /dev/tty1
|
|
||||||
I=0
|
|
||||||
else
|
|
||||||
I=$(( $I + 1 ))
|
|
||||||
fi
|
|
||||||
|
|
||||||
echo -e "${PRST}" > /dev/tty1
|
|
||||||
echo -e "\n\n\t${DFLT}To manage this appliance, please connect to one of the following:${LCLR}\n" > /dev/tty1
|
|
||||||
|
|
||||||
for c in "${COMPONENTS[@]}"; do
|
|
||||||
STATUS=$(curl -ks "https://${c}.${FQDN}" -o /dev/null -w '%{http_code}')
|
|
||||||
|
|
||||||
if [[ "${STATUS}" -eq "200" ]]; then
|
|
||||||
echo -e "\t [${BGRN}+${DFLT}] ${BBLU}https://${c}.${FQDN}${DFLT}${LCLR}" > /dev/tty1
|
|
||||||
else
|
|
||||||
echo -e "\t [${BRED}-${DFLT}] ${BBLU}https://${c}.${FQDN}${DFLT}${LCLR}" > /dev/tty1
|
|
||||||
fi
|
|
||||||
done
|
|
||||||
|
|
||||||
echo -e "\n\t${BGRY}Note that your DNS zone ${DFLT}must have${BGRY} respective records defined,\n\teach pointing to: ${DFLT}${IPADDRESS}${LCLR}" > /dev/tty1
|
|
||||||
|
|
||||||
echo -e "${CBLA}" > /dev/tty1
|
|
||||||
sleep 1
|
|
||||||
done
|
|
||||||
@@ -1,6 +0,0 @@
|
|||||||
playbook:
|
|
||||||
retries: 5
|
|
||||||
delays:
|
|
||||||
long: 60
|
|
||||||
medium: 30
|
|
||||||
short: 10
|
|
||||||
10
ansible/roles/firstboot/files/ansible_payload/playbook.yml
Normal file
10
ansible/roles/firstboot/files/ansible_payload/playbook.yml
Normal file
@@ -0,0 +1,10 @@
|
|||||||
|
---
|
||||||
|
- hosts: 127.0.0.1
|
||||||
|
connection: local
|
||||||
|
gather_facts: false
|
||||||
|
# become: true
|
||||||
|
roles:
|
||||||
|
- vapp
|
||||||
|
- network
|
||||||
|
- users
|
||||||
|
- cleanup
|
||||||
@@ -0,0 +1,20 @@
|
|||||||
|
- name: Disable crontab job
|
||||||
|
ansible.builtin.cron:
|
||||||
|
name: firstboot
|
||||||
|
state: absent
|
||||||
|
- name: Restore extra tty
|
||||||
|
ansible.builtin.lineinfile:
|
||||||
|
path: /etc/systemd/logind.conf
|
||||||
|
regexp: "{{ item.regexp }}"
|
||||||
|
line: "{{ item.line }}"
|
||||||
|
loop:
|
||||||
|
- { regexp: '^NAutoVTs=', line: '#NAutoVTs=6'}
|
||||||
|
- { regexp: '^ReserveVT=', line: '#ReserveVT=6'}
|
||||||
|
- name: Unmask getty@tty1 service
|
||||||
|
ansible.builtin.systemd:
|
||||||
|
name: getty@tty1
|
||||||
|
enabled: yes
|
||||||
|
masked: no
|
||||||
|
- name: Reboot host
|
||||||
|
ansible.builtin.shell:
|
||||||
|
cmd: /usr/sbin/reboot now
|
||||||
@@ -0,0 +1,10 @@
|
|||||||
|
- name: Set hostname
|
||||||
|
ansible.builtin.hostname:
|
||||||
|
name: "{{ ovfproperties['guestinfo.hostname'] }}"
|
||||||
|
- name: Create netplan configuration file
|
||||||
|
ansible.builtin.template:
|
||||||
|
src: netplan.j2
|
||||||
|
dest: /etc/netplan/00-installer-config.yaml
|
||||||
|
- name: Apply netplan configuration
|
||||||
|
ansible.builtin.shell:
|
||||||
|
cmd: /usr/sbin/netplan apply
|
||||||
@@ -0,0 +1,10 @@
|
|||||||
|
network:
|
||||||
|
version: 2
|
||||||
|
ethernets:
|
||||||
|
ens192:
|
||||||
|
addresses:
|
||||||
|
- {{ ovfproperties['guestinfo.ipaddress'] }}/{{ ovfproperties['guestinfo.prefixlength'] }}
|
||||||
|
gateway4: {{ ovfproperties['guestinfo.gateway'] }}
|
||||||
|
nameservers:
|
||||||
|
addresses:
|
||||||
|
- {{ ovfproperties['guestinfo.dnsserver'] }}
|
||||||
@@ -0,0 +1,25 @@
|
|||||||
|
- name: Set root password
|
||||||
|
ansible.builtin.user:
|
||||||
|
name: root
|
||||||
|
password: "{{ ovfproperties['guestinfo.rootpw'] | password_hash('sha512', 65534 | random(seed=ovfproperties['guestinfo.hostname']) | string) }}"
|
||||||
|
generate_ssh_key: yes
|
||||||
|
ssh_key_bits: 2048
|
||||||
|
ssh_key_file: .ssh/id_rsa
|
||||||
|
- name: Save root SSH publickey
|
||||||
|
ansible.builtin.lineinfile:
|
||||||
|
path: /root/.ssh/authorized_keys
|
||||||
|
line: "{{ ovfproperties['guestinfo.rootsshkey'] }}"
|
||||||
|
- name: Disable SSH password authentication
|
||||||
|
ansible.builtin.lineinfile:
|
||||||
|
path: /etc/ssh/sshd_config
|
||||||
|
regex: "{{ item.regex }}"
|
||||||
|
line: "{{ item.line }}"
|
||||||
|
state: "{{ item.state }}"
|
||||||
|
loop:
|
||||||
|
- { regex: '^#PasswordAuthentication', line: 'PasswordAuthentication no', state: present}
|
||||||
|
- { regex: '^PasswordAuthentication yes', line: 'PasswordAuthentication yes', state: absent}
|
||||||
|
- name: Delete 'ubuntu' user
|
||||||
|
ansible.builtin.user:
|
||||||
|
name: ubuntu
|
||||||
|
state: absent
|
||||||
|
remove: yes
|
||||||
@@ -0,0 +1,21 @@
|
|||||||
|
- name: Store current ovfEnvironment
|
||||||
|
ansible.builtin.shell:
|
||||||
|
cmd: /usr/bin/vmtoolsd --cmd "info-get guestinfo.ovfEnv"
|
||||||
|
register: ovfenv
|
||||||
|
- name: Parse XML for vApp properties
|
||||||
|
community.general.xml:
|
||||||
|
xmlstring: "{{ ovfenv.stdout }}"
|
||||||
|
namespaces:
|
||||||
|
ns: http://schemas.dmtf.org/ovf/environment/1
|
||||||
|
xpath: /ns:Environment/ns:PropertySection/ns:Property
|
||||||
|
content: attribute
|
||||||
|
register: ovfenv
|
||||||
|
- name: Assign vApp properties to dictionary
|
||||||
|
ansible.builtin.set_fact:
|
||||||
|
ovfproperties: >-
|
||||||
|
{{ ovfproperties | default({}) |
|
||||||
|
combine({((item.values() | list)[0].values() | list)[0]:
|
||||||
|
((item.values() | list)[0].values() | list)[1]})
|
||||||
|
}}
|
||||||
|
loop: "{{ ovfenv.matches }}"
|
||||||
|
no_log: true
|
||||||
@@ -1,26 +0,0 @@
|
|||||||
---
|
|
||||||
- hosts: 127.0.0.1
|
|
||||||
connection: local
|
|
||||||
gather_facts: true
|
|
||||||
vars_files:
|
|
||||||
- defaults.yml
|
|
||||||
- metacluster.yml
|
|
||||||
# become: true
|
|
||||||
roles:
|
|
||||||
- vapp
|
|
||||||
- network
|
|
||||||
- preflight
|
|
||||||
- users
|
|
||||||
- disks
|
|
||||||
- metacluster
|
|
||||||
# - workloadcluster
|
|
||||||
- tty
|
|
||||||
- cleanup
|
|
||||||
handlers:
|
|
||||||
- name: Apply manifests
|
|
||||||
kubernetes.core.k8s:
|
|
||||||
src: "{{ item }}"
|
|
||||||
state: present
|
|
||||||
kubeconfig: "{{ kubeconfig.path }}"
|
|
||||||
loop: "{{ query('ansible.builtin.fileglob', '/var/lib/rancher/k3s/server/manifests/*.yaml') | sort }}"
|
|
||||||
ignore_errors: yes
|
|
||||||
@@ -1,50 +0,0 @@
|
|||||||
- block:
|
|
||||||
|
|
||||||
- name: Push images to registry
|
|
||||||
ansible.builtin.shell:
|
|
||||||
cmd: >-
|
|
||||||
skopeo copy \
|
|
||||||
--insecure-policy \
|
|
||||||
--dest-tls-verify=false \
|
|
||||||
--dest-creds admin:{{ vapp['metacluster.password'] }} \
|
|
||||||
docker-archive:./{{ item | basename }} \
|
|
||||||
docker://registry.{{ vapp['metacluster.fqdn'] }}/library/$( \
|
|
||||||
skopeo list-tags \
|
|
||||||
--insecure-policy \
|
|
||||||
docker-archive:./{{ item | basename }} | \
|
|
||||||
jq -r '.Tags[0]')
|
|
||||||
chdir: /opt/metacluster/container-images/
|
|
||||||
register: push_result
|
|
||||||
loop: "{{ query('ansible.builtin.fileglob', '/opt/metacluster/container-images/*.tar') | sort }}"
|
|
||||||
loop_control:
|
|
||||||
label: "{{ item | basename }}"
|
|
||||||
retries: "{{ playbook.retries }}"
|
|
||||||
delay: "{{ playbook.delays.short }}"
|
|
||||||
until: push_result is not failed
|
|
||||||
|
|
||||||
- name: Get all stored container images (=artifacts)
|
|
||||||
ansible.builtin.uri:
|
|
||||||
url: https://registry.{{ vapp['metacluster.fqdn'] }}/api/v2.0/search?q=library
|
|
||||||
method: GET
|
|
||||||
register: registry_artifacts
|
|
||||||
|
|
||||||
- name: Get source registries of all artifacts
|
|
||||||
ansible.builtin.set_fact:
|
|
||||||
source_registries: "{{ (source_registries | default([]) + [(item | split('/'))[1]]) | unique | sort }}"
|
|
||||||
loop: "{{ registry_artifacts.json.repository | json_query('[*].repository_name') }}"
|
|
||||||
|
|
||||||
- name: Configure K3s node for private registry
|
|
||||||
ansible.builtin.template:
|
|
||||||
dest: /etc/rancher/k3s/registries.yaml
|
|
||||||
src: registries.j2
|
|
||||||
vars:
|
|
||||||
_template:
|
|
||||||
data: "{{ source_registries }}"
|
|
||||||
hv:
|
|
||||||
fqdn: "{{ vapp['metacluster.fqdn'] }}"
|
|
||||||
|
|
||||||
module_defaults:
|
|
||||||
ansible.builtin.uri:
|
|
||||||
validate_certs: no
|
|
||||||
status_code: [200, 201, 401]
|
|
||||||
body_format: json
|
|
||||||
@@ -1,30 +0,0 @@
|
|||||||
- name: Configure fallback name resolution
|
|
||||||
ansible.builtin.lineinfile:
|
|
||||||
path: /etc/hosts
|
|
||||||
line: "{{ vapp['metacluster.vip'] }} {{ item ~ '.' ~ vapp['metacluster.fqdn'] }}"
|
|
||||||
state: present
|
|
||||||
loop:
|
|
||||||
# TODO: Make this list dynamic
|
|
||||||
- ca
|
|
||||||
- git
|
|
||||||
- gitops
|
|
||||||
- ingress
|
|
||||||
- registry
|
|
||||||
- storage
|
|
||||||
|
|
||||||
- name: Retrieve root CA certificate
|
|
||||||
ansible.builtin.uri:
|
|
||||||
url: https://ca.{{ vapp['metacluster.fqdn'] }}/roots
|
|
||||||
validate_certs: no
|
|
||||||
method: GET
|
|
||||||
status_code: [200, 201]
|
|
||||||
register: rootca_certificate
|
|
||||||
|
|
||||||
- name: Store root CA certificate
|
|
||||||
ansible.builtin.copy:
|
|
||||||
dest: /usr/local/share/ca-certificates/root_ca.crt
|
|
||||||
content: "{{ rootca_certificate.json.crts | list | join('\n') }}"
|
|
||||||
|
|
||||||
- name: Update certificate truststore
|
|
||||||
ansible.builtin.command:
|
|
||||||
cmd: update-ca-certificates
|
|
||||||
@@ -1,63 +0,0 @@
|
|||||||
- name: Store custom configuration files
|
|
||||||
ansible.builtin.copy:
|
|
||||||
dest: "{{ item.filename }}"
|
|
||||||
content: "{{ item.content }}"
|
|
||||||
loop:
|
|
||||||
- filename: /etc/rancher/k3s/config.yaml
|
|
||||||
content: |
|
|
||||||
kubelet-arg:
|
|
||||||
- "config=/etc/rancher/k3s/kubelet.config"
|
|
||||||
- filename: /etc/rancher/k3s/kubelet.config
|
|
||||||
content: |
|
|
||||||
apiVersion: kubelet.config.k8s.io/v1beta1
|
|
||||||
kind: KubeletConfiguration
|
|
||||||
|
|
||||||
shutdownGracePeriod: 180s
|
|
||||||
shtudownGracePeriodCriticalPods: 60s
|
|
||||||
loop_control:
|
|
||||||
label: "{{ item.filename }}"
|
|
||||||
|
|
||||||
- name: Gather service facts
|
|
||||||
ansible.builtin.service_facts:
|
|
||||||
# Module requires no attributes
|
|
||||||
|
|
||||||
- name: Install K3s
|
|
||||||
ansible.builtin.command:
|
|
||||||
cmd: ./install.sh
|
|
||||||
chdir: /opt/metacluster/k3s
|
|
||||||
environment:
|
|
||||||
INSTALL_K3S_SKIP_DOWNLOAD: 'true'
|
|
||||||
INSTALL_K3S_EXEC: "server --token {{ vapp['metacluster.token'] | trim }} --server https://{{ vapp['metacluster.vip'] }}:6443 --disable local-storage --config /etc/rancher/k3s/config.yaml"
|
|
||||||
when: ansible_facts.services['k3s.service'] is undefined
|
|
||||||
|
|
||||||
- name: Ensure API availability
|
|
||||||
ansible.builtin.uri:
|
|
||||||
url: https://{{ vapp['guestinfo.ipaddress'] }}:6443/livez?verbose
|
|
||||||
method: GET
|
|
||||||
validate_certs: no
|
|
||||||
status_code: [200, 401]
|
|
||||||
register: api_readycheck
|
|
||||||
until: api_readycheck.json.apiVersion is defined
|
|
||||||
retries: "{{ playbook.retries }}"
|
|
||||||
delay: "{{ playbook.delays.medium }}"
|
|
||||||
|
|
||||||
- name: Install kubectl tab-completion
|
|
||||||
ansible.builtin.shell:
|
|
||||||
cmd: kubectl completion bash | tee /etc/bash_completion.d/kubectl
|
|
||||||
|
|
||||||
- name: Initialize tempfile
|
|
||||||
ansible.builtin.tempfile:
|
|
||||||
state: file
|
|
||||||
register: kubeconfig
|
|
||||||
|
|
||||||
- name: Retrieve kubeconfig
|
|
||||||
ansible.builtin.command:
|
|
||||||
cmd: kubectl config view --raw
|
|
||||||
register: kubectl_config
|
|
||||||
|
|
||||||
- name: Store kubeconfig in tempfile
|
|
||||||
ansible.builtin.copy:
|
|
||||||
dest: "{{ kubeconfig.path }}"
|
|
||||||
content: "{{ kubectl_config.stdout }}"
|
|
||||||
mode: 0600
|
|
||||||
no_log: true
|
|
||||||
@@ -1,11 +0,0 @@
|
|||||||
- import_tasks: init.yml
|
|
||||||
- import_tasks: containerimages.yml
|
|
||||||
- import_tasks: k3s.yml
|
|
||||||
- import_tasks: assets.yml
|
|
||||||
- import_tasks: storage.yml
|
|
||||||
|
|
||||||
# - import_tasks: charts.yml
|
|
||||||
- import_tasks: registry.yml
|
|
||||||
# - import_tasks: certauthority.yml
|
|
||||||
# - import_tasks: git.yml
|
|
||||||
# - import_tasks: gitops.yml
|
|
||||||
@@ -1,28 +0,0 @@
|
|||||||
- block:
|
|
||||||
|
|
||||||
- name: Upgrade harbor chart
|
|
||||||
kubernetes.core.helm:
|
|
||||||
name: harbor
|
|
||||||
chart_ref: /opt/metacluster/helm-charts/harbor
|
|
||||||
release_namespace: harbor
|
|
||||||
create_namespace: yes
|
|
||||||
wait: no
|
|
||||||
kubeconfig: "{{ kubeconfig.path }}"
|
|
||||||
values: "{{ components.harbor.chart_values }}"
|
|
||||||
|
|
||||||
- name: Ensure harbor API availability
|
|
||||||
ansible.builtin.uri:
|
|
||||||
url: https://registry.{{ vapp['metacluster.fqdn'] }}/api/v2.0/health
|
|
||||||
method: GET
|
|
||||||
register: api_readycheck
|
|
||||||
until:
|
|
||||||
- api_readycheck.json.status is defined
|
|
||||||
- api_readycheck.json.status == 'healthy'
|
|
||||||
retries: "{{ playbook.retries }}"
|
|
||||||
delay: "{{ playbook.delays.long }}"
|
|
||||||
|
|
||||||
module_defaults:
|
|
||||||
ansible.builtin.uri:
|
|
||||||
validate_certs: no
|
|
||||||
status_code: [200, 201, 401]
|
|
||||||
body_format: json
|
|
||||||
@@ -1,53 +0,0 @@
|
|||||||
- block:
|
|
||||||
|
|
||||||
- name: Increase replicas for each volume
|
|
||||||
kubernetes.core.k8s:
|
|
||||||
api_version: longhorn.io/v1beta2
|
|
||||||
kind: volume
|
|
||||||
name: "{{ item.metadata.name }}"
|
|
||||||
namespace: longhorn-system
|
|
||||||
state: patched
|
|
||||||
definition: |
|
|
||||||
spec:
|
|
||||||
numberOfReplicas: {{ lookup('kubernetes.core.k8s', kind='node', kubeconfig=(kubeconfig.path)) | length | int }}
|
|
||||||
loop: "{{ lookup('kubernetes.core.k8s', api_version='longhorn.io/v1beta2', kind='volume', namespace='longhorn-system', kubeconfig=(kubeconfig.path)) }}"
|
|
||||||
loop_control:
|
|
||||||
label: "{{ item.metadata.name }}"
|
|
||||||
|
|
||||||
- name: Wait for replica rebuilds to complete
|
|
||||||
ansible.builtin.uri:
|
|
||||||
url: https://storage.{{ vapp['metacluster.fqdn'] }}/v1/volumes
|
|
||||||
method: GET
|
|
||||||
register: volume_details
|
|
||||||
until:
|
|
||||||
- (volume_details.json.data | json_query('[*].robustness') | unique | length) == 1
|
|
||||||
- (volume_details.json.data | json_query('[*].robustness') | first) == "healthy"
|
|
||||||
retries: "{{ playbook.retries }}"
|
|
||||||
delay: "{{ playbook.delays.medium }}"
|
|
||||||
|
|
||||||
- name: Install longhorn chart
|
|
||||||
kubernetes.core.helm:
|
|
||||||
name: longhorn
|
|
||||||
chart_ref: /opt/metacluster/helm-charts/longhorn
|
|
||||||
release_namespace: longhorn-system
|
|
||||||
create_namespace: yes
|
|
||||||
wait: no
|
|
||||||
values: "{{ components.longhorn.chart_values }}"
|
|
||||||
|
|
||||||
- name: Ensure longhorn API availability
|
|
||||||
ansible.builtin.uri:
|
|
||||||
url: https://storage.{{ vapp['metacluster.fqdn'] }}/v1
|
|
||||||
method: GET
|
|
||||||
register: api_readycheck
|
|
||||||
until:
|
|
||||||
- api_readycheck is not failed
|
|
||||||
retries: "{{ playbook.retries }}"
|
|
||||||
delay: "{{ playbook.delays.long }}"
|
|
||||||
|
|
||||||
module_defaults:
|
|
||||||
ansible.builtin.uri:
|
|
||||||
validate_certs: no
|
|
||||||
status_code: [200, 201]
|
|
||||||
body_format: json
|
|
||||||
group/k8s:
|
|
||||||
kubeconfig: "{{ kubeconfig.path }}"
|
|
||||||
@@ -1,2 +0,0 @@
|
|||||||
- import_tasks: vcenter.yml
|
|
||||||
- import_tasks: metacluster.yml
|
|
||||||
@@ -1,6 +0,0 @@
|
|||||||
- name: Check for metacluster connectivity
|
|
||||||
ansible.builtin.uri:
|
|
||||||
url: https://{{ vapp['metacluster.vip'] }}:6443/livez?verbose
|
|
||||||
method: GET
|
|
||||||
validate_certs: no
|
|
||||||
status_code: [200, 401]
|
|
||||||
@@ -2,7 +2,6 @@
|
|||||||
ansible.builtin.file:
|
ansible.builtin.file:
|
||||||
path: /opt/firstboot
|
path: /opt/firstboot
|
||||||
state: directory
|
state: directory
|
||||||
|
|
||||||
- name: Create firstboot script file
|
- name: Create firstboot script file
|
||||||
ansible.builtin.template:
|
ansible.builtin.template:
|
||||||
src: firstboot.j2
|
src: firstboot.j2
|
||||||
@@ -10,25 +9,18 @@
|
|||||||
owner: root
|
owner: root
|
||||||
group: root
|
group: root
|
||||||
mode: o+x
|
mode: o+x
|
||||||
|
|
||||||
- name: Create @reboot crontab job
|
- name: Create @reboot crontab job
|
||||||
ansible.builtin.cron:
|
ansible.builtin.cron:
|
||||||
name: firstboot
|
name: firstboot
|
||||||
special_time: reboot
|
special_time: reboot
|
||||||
job: "/opt/firstboot/firstboot.sh >/dev/tty1 2>&1"
|
job: "/opt/firstboot/firstboot.sh"
|
||||||
|
- name: Copy payload folder
|
||||||
- name: Copy payload folder (common)
|
|
||||||
ansible.builtin.copy:
|
ansible.builtin.copy:
|
||||||
src: ansible_payload/common/
|
src: ansible_payload/
|
||||||
dest: /opt/firstboot/ansible/
|
|
||||||
owner: root
|
|
||||||
group: root
|
|
||||||
mode: '0644'
|
|
||||||
|
|
||||||
- name: Copy payload folder (per appliancetype)
|
|
||||||
ansible.builtin.copy:
|
|
||||||
src: ansible_payload/{{ appliancetype }}/
|
|
||||||
dest: /opt/firstboot/ansible/
|
dest: /opt/firstboot/ansible/
|
||||||
owner: root
|
owner: root
|
||||||
group: root
|
group: root
|
||||||
mode: '0644'
|
mode: '0644'
|
||||||
|
- name: Install ansible-galaxy collection
|
||||||
|
ansible.builtin.shell:
|
||||||
|
cmd: ansible-galaxy collection install community.general
|
||||||
|
|||||||
@@ -1,10 +1,4 @@
|
|||||||
#!/bin/bash
|
#!/bin/bash
|
||||||
|
|
||||||
# Workaround for ansible output regression
|
|
||||||
export PYTHONUNBUFFERED=1
|
|
||||||
|
|
||||||
# Apply firstboot configuration w/ ansible
|
# Apply firstboot configuration w/ ansible
|
||||||
/usr/local/bin/ansible-playbook /opt/firstboot/ansible/playbook.yml | tee -a /var/log/firstboot.log > /dev/tty1 2>&1
|
/usr/local/bin/ansible-playbook /opt/firstboot/ansible/playbook.yml | tee -a /var/log/firstboot.log > /dev/tty1
|
||||||
|
|
||||||
# Cleanup console
|
|
||||||
clear > /dev/tty1
|
|
||||||
6
ansible/roles/os/tasks/ansible.yml
Normal file
6
ansible/roles/os/tasks/ansible.yml
Normal file
@@ -0,0 +1,6 @@
|
|||||||
|
- name: Install ansible (w/ dependencies)
|
||||||
|
ansible.builtin.pip:
|
||||||
|
name: "{{ item }}"
|
||||||
|
executable: pip3
|
||||||
|
state: latest
|
||||||
|
loop: "{{ pip_packages }}"
|
||||||
@@ -3,7 +3,6 @@
|
|||||||
name: cloud-init
|
name: cloud-init
|
||||||
state: absent
|
state: absent
|
||||||
purge: yes
|
purge: yes
|
||||||
|
|
||||||
- name: Delete cloud-init files
|
- name: Delete cloud-init files
|
||||||
ansible.builtin.file:
|
ansible.builtin.file:
|
||||||
path: "{{ item }}"
|
path: "{{ item }}"
|
||||||
|
|||||||
@@ -15,3 +15,6 @@
|
|||||||
|
|
||||||
- name: Install packages
|
- name: Install packages
|
||||||
import_tasks: packages.yml
|
import_tasks: packages.yml
|
||||||
|
|
||||||
|
- name: Install ansible
|
||||||
|
import_tasks: ansible.yml
|
||||||
|
|||||||
@@ -1,46 +1,14 @@
|
|||||||
- name: Configure 'needrestart' package
|
|
||||||
ansible.builtin.lineinfile:
|
|
||||||
path: /etc/needrestart/needrestart.conf
|
|
||||||
regexp: "{{ item.regexp }}"
|
|
||||||
line: "{{ item.line }}"
|
|
||||||
loop:
|
|
||||||
- regexp: "^#\\$nrconf\\{restart\\} = 'i';"
|
|
||||||
line: "$nrconf{restart} = 'a';"
|
|
||||||
- regexp: "^#\\$nrconf\\{kernelhints\\} = -1;"
|
|
||||||
line: "$nrconf{kernelhints} = -1;"
|
|
||||||
loop_control:
|
|
||||||
label: "{{ item.line }}"
|
|
||||||
|
|
||||||
- name: Install additional packages
|
- name: Install additional packages
|
||||||
ansible.builtin.apt:
|
ansible.builtin.apt:
|
||||||
pkg: "{{ packages.apt }}"
|
name: "{{ item }}"
|
||||||
state: latest
|
state: latest
|
||||||
update_cache: yes
|
update_cache: yes
|
||||||
install_recommends: no
|
loop: "{{ packages }}"
|
||||||
|
|
||||||
- name: Upgrade all packages
|
- name: Upgrade all packages
|
||||||
ansible.builtin.apt:
|
ansible.builtin.apt:
|
||||||
name: '*'
|
name: "*"
|
||||||
state: latest
|
state: latest
|
||||||
update_cache: yes
|
update_cache: yes
|
||||||
|
|
||||||
- name: Install additional python packages
|
|
||||||
ansible.builtin.pip:
|
|
||||||
name: "{{ item }}"
|
|
||||||
executable: pip3
|
|
||||||
state: latest
|
|
||||||
loop: "{{ packages.pip }}"
|
|
||||||
|
|
||||||
- name: Create folder
|
|
||||||
ansible.builtin.file:
|
|
||||||
path: /etc/ansible
|
|
||||||
state: directory
|
|
||||||
|
|
||||||
- name: Configure Ansible defaults
|
|
||||||
ansible.builtin.template:
|
|
||||||
src: ansible.j2
|
|
||||||
dest: /etc/ansible/ansible.cfg
|
|
||||||
|
|
||||||
- name: Cleanup
|
- name: Cleanup
|
||||||
ansible.builtin.apt:
|
ansible.builtin.apt:
|
||||||
autoremove: yes
|
autoremove: yes
|
||||||
|
|||||||
@@ -3,17 +3,14 @@
|
|||||||
name: snapd
|
name: snapd
|
||||||
state: absent
|
state: absent
|
||||||
purge: yes
|
purge: yes
|
||||||
|
|
||||||
- name: Delete leftover files
|
- name: Delete leftover files
|
||||||
ansible.builtin.file:
|
ansible.builtin.file:
|
||||||
path: /root/snap
|
path: /root/snap
|
||||||
state: absent
|
state: absent
|
||||||
|
|
||||||
- name: Hold snapd package
|
- name: Hold snapd package
|
||||||
ansible.builtin.dpkg_selections:
|
ansible.builtin.dpkg_selections:
|
||||||
name: snapd
|
name: snapd
|
||||||
selection: hold
|
selection: hold
|
||||||
|
|
||||||
- name: Reload systemd unit configurations
|
- name: Reload systemd unit configurations
|
||||||
ansible.builtin.systemd:
|
ansible.builtin.systemd:
|
||||||
daemon_reload: yes
|
daemon_reload: yes
|
||||||
|
|||||||
@@ -4,13 +4,8 @@
|
|||||||
regexp: "{{ item.regexp }}"
|
regexp: "{{ item.regexp }}"
|
||||||
line: "{{ item.line }}"
|
line: "{{ item.line }}"
|
||||||
loop:
|
loop:
|
||||||
- regexp: '^#NAutoVTs='
|
- { regexp: '^#NAutoVTs=', line: 'NAutoVTs=1'}
|
||||||
line: 'NAutoVTs=1'
|
- { regexp: '^#ReserveVT=', line: 'ReserveVT=11'}
|
||||||
- regexp: '^#ReserveVT='
|
|
||||||
line: 'ReserveVT=11'
|
|
||||||
loop_control:
|
|
||||||
label: "{{ item.line }}"
|
|
||||||
|
|
||||||
- name: Mask getty@tty1 service
|
- name: Mask getty@tty1 service
|
||||||
ansible.builtin.systemd:
|
ansible.builtin.systemd:
|
||||||
name: getty@tty1
|
name: getty@tty1
|
||||||
|
|||||||
@@ -1,2 +0,0 @@
|
|||||||
[defaults]
|
|
||||||
callbacks_enabled = ansible.posix.profile_tasks
|
|
||||||
@@ -1,13 +1,11 @@
|
|||||||
packages:
|
packages:
|
||||||
apt:
|
- jq
|
||||||
- jq
|
# (python3-*) Dependency for installation of Ansible
|
||||||
- python3-pip
|
- python3-pip
|
||||||
pip:
|
- python3-setuptools
|
||||||
# - ansible-core<2.14.0
|
- python3-wheel
|
||||||
- ansible-core
|
|
||||||
- jinja2
|
pip_packages:
|
||||||
- lxml
|
- pip
|
||||||
- markupsafe
|
- ansible-core
|
||||||
- pip
|
- lxml
|
||||||
- setuptools
|
|
||||||
- wheel
|
|
||||||
|
|||||||
@@ -1,266 +0,0 @@
|
|||||||
platform:
|
|
||||||
|
|
||||||
k3s:
|
|
||||||
version: v1.26.0+k3s1
|
|
||||||
|
|
||||||
gitops:
|
|
||||||
repository:
|
|
||||||
uri: https://code.spamasaurus.com/djpbessems/GitOps.MetaCluster.git
|
|
||||||
# revision: v0.1.0
|
|
||||||
revision: HEAD
|
|
||||||
|
|
||||||
packaged_components:
|
|
||||||
- name: traefik
|
|
||||||
namespace: kube-system
|
|
||||||
config: |2
|
|
||||||
additionalArguments:
|
|
||||||
- "--certificatesResolvers.stepca.acme.caserver=https://step-certificates.step-ca.svc.cluster.local/acme/acme/directory"
|
|
||||||
- "--certificatesResolvers.stepca.acme.email=admin"
|
|
||||||
- "--certificatesResolvers.stepca.acme.storage=/data/acme.json"
|
|
||||||
- "--certificatesResolvers.stepca.acme.tlsChallenge=true"
|
|
||||||
- "--certificatesresolvers.stepca.acme.certificatesduration=24"
|
|
||||||
deployment:
|
|
||||||
initContainers:
|
|
||||||
- name: volume-permissions
|
|
||||||
image: busybox:1
|
|
||||||
command: ["sh", "-c", "touch /data/acme.json && chmod -Rv 600 /data/* && chown 65532:65532 /data/acme.json"]
|
|
||||||
volumeMounts:
|
|
||||||
- name: data
|
|
||||||
mountPath: /data
|
|
||||||
globalArguments: []
|
|
||||||
ingressRoute:
|
|
||||||
dashboard:
|
|
||||||
enabled: false
|
|
||||||
persistence:
|
|
||||||
enabled: true
|
|
||||||
ports:
|
|
||||||
ssh:
|
|
||||||
port: 8022
|
|
||||||
protocol: TCP
|
|
||||||
web:
|
|
||||||
redirectTo: websecure
|
|
||||||
websecure:
|
|
||||||
tls:
|
|
||||||
certResolver: stepca
|
|
||||||
updateStrategy:
|
|
||||||
type: Recreate
|
|
||||||
rollingUpdate: null
|
|
||||||
|
|
||||||
helm_repositories:
|
|
||||||
- name: argo
|
|
||||||
url: https://argoproj.github.io/argo-helm
|
|
||||||
- name: gitea-charts
|
|
||||||
url: https://dl.gitea.io/charts/
|
|
||||||
- name: harbor
|
|
||||||
url: https://helm.goharbor.io
|
|
||||||
- name: jetstack
|
|
||||||
url: https://charts.jetstack.io
|
|
||||||
- name: longhorn
|
|
||||||
url: https://charts.longhorn.io
|
|
||||||
- name: smallstep
|
|
||||||
url: https://smallstep.github.io/helm-charts/
|
|
||||||
|
|
||||||
components:
|
|
||||||
|
|
||||||
argo-cd:
|
|
||||||
helm:
|
|
||||||
# version: 4.9.7 # (= ArgoCD v2.4.2)
|
|
||||||
version: 5.14.1 # (= ArgoCD v2.5.2)
|
|
||||||
chart: argo/argo-cd
|
|
||||||
parse_logic: helm template . | yq --no-doc eval '.. | .image? | select(.)' | sort -u | awk '!/ /'
|
|
||||||
chart_values: !unsafe |
|
|
||||||
configs:
|
|
||||||
secret:
|
|
||||||
argocdServerAdminPassword: "{{ vapp['metacluster.password'] | password_hash('bcrypt') }}"
|
|
||||||
server:
|
|
||||||
extraArgs:
|
|
||||||
- --insecure
|
|
||||||
ingress:
|
|
||||||
enabled: true
|
|
||||||
hosts:
|
|
||||||
- gitops.{{ vapp['metacluster.fqdn'] }}
|
|
||||||
|
|
||||||
cert-manager:
|
|
||||||
helm:
|
|
||||||
version: 1.10.1
|
|
||||||
chart: jetstack/cert-manager
|
|
||||||
parse_logic: helm template . | yq --no-doc eval '.. | .image? | select(.)' | sort -u | awk '!/ /'
|
|
||||||
# chart_values: !unsafe |
|
|
||||||
# installCRDs: true
|
|
||||||
|
|
||||||
clusterapi:
|
|
||||||
management:
|
|
||||||
version:
|
|
||||||
# Must match the version referenced at `dependencies.static_binaries[.filename==clusterctl].url`
|
|
||||||
base: v1.3.2
|
|
||||||
# Must match the version referenced at `components.cert-manager.helm.version`
|
|
||||||
cert_manager: v1.10.1
|
|
||||||
infrastructure_vsphere: v1.5.1
|
|
||||||
ipam_incluster: v0.1.0-alpha.1
|
|
||||||
workload:
|
|
||||||
version:
|
|
||||||
calico: v3.24.5
|
|
||||||
# k8s: v1.25.5
|
|
||||||
k8s: v1.23.5
|
|
||||||
node_template:
|
|
||||||
# Refer to `https://github.com/kubernetes-sigs/cluster-api-provider-vsphere/blob/v1.3.5/README.md#kubernetes-versions-with-published-ovas` for a list of supported node templates
|
|
||||||
# url: https://storage.googleapis.com/capv-templates/v1.25.5/ubuntu-2004-kube-v1.25.5.ova
|
|
||||||
url: https://storage.googleapis.com/capv-images/release/v1.23.5/ubuntu-2004-kube-v1.23.5.ova
|
|
||||||
|
|
||||||
gitea:
|
|
||||||
helm:
|
|
||||||
version: v6.0.3 # (= Gitea v1.17.3)
|
|
||||||
chart: gitea-charts/gitea
|
|
||||||
parse_logic: helm template . | yq --no-doc eval '.. | .image? | select(.)' | sort -u | sed '/:/!s/$/:latest/'
|
|
||||||
chart_values: !unsafe |
|
|
||||||
gitea:
|
|
||||||
admin:
|
|
||||||
username: administrator
|
|
||||||
password: "{{ vapp['metacluster.password'] }}"
|
|
||||||
email: admin@{{ vapp['metacluster.fqdn'] }}
|
|
||||||
config:
|
|
||||||
server:
|
|
||||||
OFFLINE_MODE: true
|
|
||||||
PROTOCOL: http
|
|
||||||
ROOT_URL: https://git.{{ vapp['metacluster.fqdn'] }}/
|
|
||||||
image:
|
|
||||||
pullPolicy: IfNotPresent
|
|
||||||
ingress:
|
|
||||||
enabled: true
|
|
||||||
hosts:
|
|
||||||
- host: git.{{ vapp['metacluster.fqdn'] }}
|
|
||||||
paths:
|
|
||||||
- path: /
|
|
||||||
pathType: Prefix
|
|
||||||
service:
|
|
||||||
ssh:
|
|
||||||
type: ClusterIP
|
|
||||||
port: 22
|
|
||||||
clusterIP:
|
|
||||||
|
|
||||||
harbor:
|
|
||||||
helm:
|
|
||||||
version: 1.10.2 # (= Harbor v2.6.2)
|
|
||||||
chart: harbor/harbor
|
|
||||||
parse_logic: helm template . | yq --no-doc eval '.. | .image? | select(.)' | sort -u | awk '!/ /'
|
|
||||||
chart_values: !unsafe |
|
|
||||||
expose:
|
|
||||||
ingress:
|
|
||||||
annotations: {}
|
|
||||||
hosts:
|
|
||||||
core: registry.{{ vapp['metacluster.fqdn'] }}
|
|
||||||
tls:
|
|
||||||
certSource: none
|
|
||||||
enabled: false
|
|
||||||
externalURL: https://registry.{{ vapp['metacluster.fqdn'] }}
|
|
||||||
harborAdminPassword: "{{ vapp['metacluster.password'] }}"
|
|
||||||
notary:
|
|
||||||
enabled: false
|
|
||||||
persistence:
|
|
||||||
persistentVolumeClaim:
|
|
||||||
registry:
|
|
||||||
size: 25Gi
|
|
||||||
|
|
||||||
kubevip:
|
|
||||||
# Must match the version referenced at `dependencies.container_images`
|
|
||||||
version: v0.5.8
|
|
||||||
|
|
||||||
longhorn:
|
|
||||||
helm:
|
|
||||||
version: 1.4.0
|
|
||||||
chart: longhorn/longhorn
|
|
||||||
parse_logic: cat values.yaml | yq eval '.. | select(has("repository")) | .repository + ":" + .tag'
|
|
||||||
chart_values: !unsafe |
|
|
||||||
defaultSettings:
|
|
||||||
allowNodeDrainWithLastHealthyReplica: true
|
|
||||||
defaultDataPath: /mnt/blockstorage
|
|
||||||
defaultReplicaCount: 1
|
|
||||||
ingress:
|
|
||||||
enabled: true
|
|
||||||
host: storage.{{ vapp['metacluster.fqdn'] }}
|
|
||||||
persistence:
|
|
||||||
defaultClassReplicaCount: 1
|
|
||||||
|
|
||||||
step-certificates:
|
|
||||||
helm:
|
|
||||||
# version: 1.18.2+20220324
|
|
||||||
version: 1.23.0
|
|
||||||
chart: smallstep/step-certificates
|
|
||||||
parse_logic: helm template . | yq --no-doc eval '.. | .image? | select(.)' | sed '/:/!s/$/:latest/' | sort -u
|
|
||||||
chart_values: !unsafe |
|
|
||||||
ca:
|
|
||||||
bootstrap:
|
|
||||||
postInitHook: |
|
|
||||||
echo '{{ vapp["metacluster.password"] }}' > ~/pwfile
|
|
||||||
step ca provisioner add acme \
|
|
||||||
--type ACME \
|
|
||||||
--password-file=~/pwfile \
|
|
||||||
--force-cn
|
|
||||||
rm ~/pwfile
|
|
||||||
dns: ca.{{ vapp['metacluster.fqdn'] }},step-certificates.step-ca.svc.cluster.local,127.0.0.1
|
|
||||||
password: "{{ vapp['metacluster.password'] }}"
|
|
||||||
provisioner:
|
|
||||||
name: admin
|
|
||||||
password: "{{ vapp['metacluster.password'] }}"
|
|
||||||
inject:
|
|
||||||
secrets:
|
|
||||||
ca_password: "{{ vapp['metacluster.password'] | b64encode }}"
|
|
||||||
provisioner_password: "{{ vapp['metacluster.password'] | b64encode }}"
|
|
||||||
service:
|
|
||||||
targetPort: 9000
|
|
||||||
|
|
||||||
dependencies:
|
|
||||||
|
|
||||||
ansible_galaxy_collections:
|
|
||||||
- ansible.posix
|
|
||||||
- ansible.utils
|
|
||||||
- community.crypto
|
|
||||||
- community.general
|
|
||||||
- community.vmware
|
|
||||||
- kubernetes.core
|
|
||||||
|
|
||||||
container_images:
|
|
||||||
# This should match the image tag referenced at `platform.packaged_components[.name==traefik].config`
|
|
||||||
- busybox:1
|
|
||||||
- ghcr.io/kube-vip/kube-vip:v0.5.8
|
|
||||||
# The following list is generated by running the following commands:
|
|
||||||
# $ clusterctl init -i vsphere:<version> [...]
|
|
||||||
# $ clusterctl generate cluster <name> [...] | yq eval '.data.data' | yq --no-doc eval '.. | .image? | select(.)' | sort -u
|
|
||||||
- gcr.io/cloud-provider-vsphere/cpi/release/manager:v1.18.1
|
|
||||||
- gcr.io/cloud-provider-vsphere/csi/release/driver:v2.1.0
|
|
||||||
- gcr.io/cloud-provider-vsphere/csi/release/syncer:v2.1.0
|
|
||||||
- quay.io/k8scsi/csi-attacher:v3.0.0
|
|
||||||
- quay.io/k8scsi/csi-node-driver-registrar:v2.0.1
|
|
||||||
- quay.io/k8scsi/csi-provisioner:v2.0.0
|
|
||||||
- quay.io/k8scsi/livenessprobe:v2.1.0
|
|
||||||
|
|
||||||
static_binaries:
|
|
||||||
- filename: clusterctl
|
|
||||||
url: https://github.com/kubernetes-sigs/cluster-api/releases/download/v1.3.2/clusterctl-linux-amd64
|
|
||||||
- filename: govc
|
|
||||||
url: https://github.com/vmware/govmomi/releases/download/v0.29.0/govc_Linux_x86_64.tar.gz
|
|
||||||
archive: compressed
|
|
||||||
- filename: helm
|
|
||||||
url: https://get.helm.sh/helm-v3.10.2-linux-amd64.tar.gz
|
|
||||||
archive: compressed
|
|
||||||
extra_opts: --strip-components=1
|
|
||||||
- filename: npp-prepper
|
|
||||||
url: https://code.spamasaurus.com/api/packages/djpbessems/generic/npp-prepper/v0.4.5/npp-prepper
|
|
||||||
- filename: skopeo
|
|
||||||
url: https://code.spamasaurus.com/api/packages/djpbessems/generic/skopeo/v1.11.0-dev/skopeo
|
|
||||||
- filename: step
|
|
||||||
url: https://dl.step.sm/gh-release/cli/gh-release-header/v0.23.0/step_linux_0.23.0_amd64.tar.gz
|
|
||||||
archive: compressed
|
|
||||||
extra_opts: --strip-components=2
|
|
||||||
- filename: yq
|
|
||||||
url: http://github.com/mikefarah/yq/releases/download/v4.30.5/yq_linux_amd64
|
|
||||||
|
|
||||||
packages:
|
|
||||||
apt:
|
|
||||||
- lvm2
|
|
||||||
pip:
|
|
||||||
- jmespath
|
|
||||||
- kubernetes
|
|
||||||
- netaddr
|
|
||||||
- passlib
|
|
||||||
- pyvmomi
|
|
||||||
@@ -1,47 +0,0 @@
|
|||||||
packer {
|
|
||||||
required_plugins {
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
build {
|
|
||||||
source "vsphere-iso.ubuntu" {
|
|
||||||
name = "bootstrap"
|
|
||||||
vm_name = "${var.vm_name}-bootstrap"
|
|
||||||
}
|
|
||||||
|
|
||||||
source "vsphere-iso.ubuntu" {
|
|
||||||
name = "upgrade"
|
|
||||||
vm_name = "${var.vm_name}-upgrade"
|
|
||||||
}
|
|
||||||
|
|
||||||
provisioner "ansible" {
|
|
||||||
pause_before = "2m30s"
|
|
||||||
|
|
||||||
playbook_file = "ansible/playbook.yml"
|
|
||||||
user = "ubuntu"
|
|
||||||
ansible_env_vars = [
|
|
||||||
"ANSIBLE_CONFIG=ansible/ansible.cfg",
|
|
||||||
"PYTHONUNBUFFERED=1"
|
|
||||||
]
|
|
||||||
use_proxy = "false"
|
|
||||||
extra_arguments = [
|
|
||||||
"--extra-vars", "appliancetype=${source.name}",
|
|
||||||
"--extra-vars", "ansible_ssh_pass=${var.ssh_password}"//,
|
|
||||||
// "--extra-vars", "repo_username=${var.repo_username}",
|
|
||||||
// "--extra-vars", "repo_password=${var.repo_password}"
|
|
||||||
]
|
|
||||||
}
|
|
||||||
|
|
||||||
post-processor "shell-local" {
|
|
||||||
inline = [
|
|
||||||
"pwsh -command \"& scripts/Update-OvfConfiguration.ps1 \\",
|
|
||||||
" -ApplianceType '${source.name}' \\",
|
|
||||||
" -OVFFile '/scratch/airgapped-k8s/${var.vm_name}-${source.name}.ovf' \"",
|
|
||||||
"pwsh -file scripts/Update-Manifest.ps1 \\",
|
|
||||||
" -ManifestFileName '/scratch/airgapped-k8s/${var.vm_name}-${source.name}.mf'",
|
|
||||||
"ovftool --acceptAllEulas --allowExtraConfig --overwrite \\",
|
|
||||||
" '/scratch/airgapped-k8s/${var.vm_name}-${source.name}.ovf' \\",
|
|
||||||
" /output/airgapped-k8s.${source.name}.ova"
|
|
||||||
]
|
|
||||||
}
|
|
||||||
}
|
|
||||||
@@ -1,5 +1,4 @@
|
|||||||
iso_url = "sn.itch.fyi/Repository/iso/Canonical/Ubuntu%20Server%2022.04/ubuntu-22.04.1-live-server-amd64.iso"
|
iso_url = "sn.itch.fyi/Repository/iso/Canonical/Ubuntu%20Server%2020.04/ubuntu-20.04.4-live-server-amd64.iso"
|
||||||
iso_checksum = "sha256:10F19C5B2B8D6DB711582E0E27F5116296C34FE4B313BA45F9B201A5007056CB"
|
iso_checksum = "sha256:28CCDB56450E643BAD03BB7BCF7507CE3D8D90E8BF09E38F6BD9AC298A98EAAD"
|
||||||
|
// iso_url = "sn.itch.fyi/Repository/iso/Canonical/Ubuntu%20Server%2020.04/ubuntu-20.04.2-live-server-amd64.iso"
|
||||||
// iso_url = "sn.itch.fyi/Repository/iso/Canonical/Ubuntu%20Server%2022.04/ubuntu-22.04-live-server-amd64.iso"
|
// iso_checksum = "sha256:D1F2BF834BBE9BB43FAF16F9BE992A6F3935E65BE0EDECE1DEE2AA6EB1767423"
|
||||||
// iso_checksum = "sha256:84AEAF7823C8C61BAA0AE862D0A06B03409394800000B3235854A6B38EB4856F"
|
|
||||||
|
|||||||
@@ -1,61 +0,0 @@
|
|||||||
source "vsphere-iso" "ubuntu" {
|
|
||||||
vcenter_server = var.vcenter_server
|
|
||||||
username = var.vsphere_username
|
|
||||||
password = var.vsphere_password
|
|
||||||
insecure_connection = "true"
|
|
||||||
|
|
||||||
datacenter = var.vsphere_datacenter
|
|
||||||
cluster = var.vsphere_cluster
|
|
||||||
host = var.vsphere_host
|
|
||||||
folder = var.vsphere_folder
|
|
||||||
datastore = var.vsphere_datastore
|
|
||||||
|
|
||||||
guest_os_type = "ubuntu64Guest"
|
|
||||||
|
|
||||||
boot_order = "disk,cdrom"
|
|
||||||
boot_command = [
|
|
||||||
"e<down><down><down><end>",
|
|
||||||
" autoinstall ds=nocloud;",
|
|
||||||
"<F10>"
|
|
||||||
]
|
|
||||||
boot_wait = "2s"
|
|
||||||
|
|
||||||
communicator = "ssh"
|
|
||||||
ssh_username = "ubuntu"
|
|
||||||
ssh_password = var.ssh_password
|
|
||||||
ssh_timeout = "20m"
|
|
||||||
ssh_handshake_attempts = "100"
|
|
||||||
ssh_pty = true
|
|
||||||
|
|
||||||
CPUs = 4
|
|
||||||
RAM = 8192
|
|
||||||
|
|
||||||
network_adapters {
|
|
||||||
network = var.vsphere_network
|
|
||||||
network_card = "vmxnet3"
|
|
||||||
}
|
|
||||||
storage {
|
|
||||||
disk_size = 76800
|
|
||||||
disk_thin_provisioned = true
|
|
||||||
}
|
|
||||||
disk_controller_type = ["pvscsi"]
|
|
||||||
usb_controller = ["xhci"]
|
|
||||||
|
|
||||||
cd_files = [
|
|
||||||
"packer/preseed/UbuntuServer22.04/user-data",
|
|
||||||
"packer/preseed/UbuntuServer22.04/meta-data"
|
|
||||||
]
|
|
||||||
cd_label = "cidata"
|
|
||||||
iso_url = local.iso_authenticatedurl
|
|
||||||
iso_checksum = var.iso_checksum
|
|
||||||
|
|
||||||
shutdown_command = "echo '${var.ssh_password}' | sudo -S shutdown -P now"
|
|
||||||
shutdown_timeout = "5m"
|
|
||||||
|
|
||||||
remove_cdrom = true
|
|
||||||
|
|
||||||
export {
|
|
||||||
images = false
|
|
||||||
output_directory = "/scratch/airgapped-k8s"
|
|
||||||
}
|
|
||||||
}
|
|
||||||
100
packer/ubuntuserver20.04.pkr.hcl
Normal file
100
packer/ubuntuserver20.04.pkr.hcl
Normal file
@@ -0,0 +1,100 @@
|
|||||||
|
packer {
|
||||||
|
required_plugins {
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
source "vsphere-iso" "ubuntuserver" {
|
||||||
|
vcenter_server = var.vcenter_server
|
||||||
|
username = var.vsphere_username
|
||||||
|
password = var.vsphere_password
|
||||||
|
insecure_connection = "true"
|
||||||
|
|
||||||
|
vm_name = "${var.vm_guestos}-${var.vm_name}"
|
||||||
|
datacenter = var.vsphere_datacenter
|
||||||
|
cluster = var.vsphere_cluster
|
||||||
|
host = var.vsphere_host
|
||||||
|
folder = var.vsphere_folder
|
||||||
|
datastore = var.vsphere_datastore
|
||||||
|
|
||||||
|
guest_os_type = "ubuntu64Guest"
|
||||||
|
|
||||||
|
boot_order = "disk,cdrom"
|
||||||
|
boot_command = [
|
||||||
|
"<enter><wait2><enter><wait><f6><esc><wait>",
|
||||||
|
" autoinstall<wait2> ds=nocloud;",
|
||||||
|
"<wait><enter>"
|
||||||
|
]
|
||||||
|
boot_wait = "2s"
|
||||||
|
|
||||||
|
communicator = "ssh"
|
||||||
|
ssh_username = "ubuntu"
|
||||||
|
ssh_password = var.ssh_password
|
||||||
|
ssh_timeout = "20m"
|
||||||
|
ssh_handshake_attempts = "100"
|
||||||
|
ssh_pty = true
|
||||||
|
|
||||||
|
CPUs = 2
|
||||||
|
RAM = 4096
|
||||||
|
|
||||||
|
network_adapters {
|
||||||
|
network = var.vsphere_network
|
||||||
|
network_card = "vmxnet3"
|
||||||
|
}
|
||||||
|
storage {
|
||||||
|
disk_size = 20480
|
||||||
|
disk_thin_provisioned = true
|
||||||
|
}
|
||||||
|
disk_controller_type = ["pvscsi"]
|
||||||
|
usb_controller = ["xhci"]
|
||||||
|
|
||||||
|
cd_files = [
|
||||||
|
"packer/preseed/UbuntuServer20.04/user-data",
|
||||||
|
"packer/preseed/UbuntuServer20.04/meta-data"
|
||||||
|
]
|
||||||
|
cd_label = "cidata"
|
||||||
|
iso_url = local.iso_authenticatedurl
|
||||||
|
iso_checksum = var.iso_checksum
|
||||||
|
|
||||||
|
shutdown_command = "echo '${var.ssh_password}' | sudo -S shutdown -P now"
|
||||||
|
shutdown_timeout = "5m"
|
||||||
|
|
||||||
|
export {
|
||||||
|
images = false
|
||||||
|
output_directory = "/scratch/ubuntuserver"
|
||||||
|
}
|
||||||
|
remove_cdrom = true
|
||||||
|
}
|
||||||
|
|
||||||
|
build {
|
||||||
|
sources = [
|
||||||
|
"source.vsphere-iso.ubuntuserver"
|
||||||
|
]
|
||||||
|
|
||||||
|
provisioner "ansible" {
|
||||||
|
only = ["vsphere-iso.ubuntuserver"]
|
||||||
|
|
||||||
|
playbook_file = "ansible/playbook.yml"
|
||||||
|
user = "ubuntu"
|
||||||
|
ansible_env_vars = [
|
||||||
|
"ANSIBLE_CONFIG=ansible/ansible.cfg"
|
||||||
|
]
|
||||||
|
use_proxy = "false"
|
||||||
|
extra_arguments = [
|
||||||
|
"--extra-vars", "ansible_ssh_pass=${var.ssh_password}"
|
||||||
|
]
|
||||||
|
}
|
||||||
|
|
||||||
|
post-processor "shell-local" {
|
||||||
|
only = ["vsphere-iso.ubuntuserver"]
|
||||||
|
inline = [
|
||||||
|
"pwsh -command \"& scripts/Update-OvfConfiguration.ps1 \\",
|
||||||
|
" -OVFFile '/scratch/ubuntuserver/${var.vm_guestos}-${var.vm_name}.ovf' \\",
|
||||||
|
" -Parameter @{'appliance.name'='${var.vm_guestos}';'appliance.version'='${var.vm_name}'}\"",
|
||||||
|
"pwsh -file scripts/Update-Manifest.ps1 \\",
|
||||||
|
" -ManifestFileName '/scratch/ubuntuserver/${var.vm_guestos}-${var.vm_name}.mf'",
|
||||||
|
"ovftool --acceptAllEulas --allowExtraConfig --overwrite \\",
|
||||||
|
" '/scratch/ubuntuserver/${var.vm_guestos}-${var.vm_name}.ovf' \\",
|
||||||
|
" /output/Ubuntu-Server-20.04.ova"
|
||||||
|
]
|
||||||
|
}
|
||||||
|
}
|
||||||
@@ -14,6 +14,7 @@ variable "vsphere_datastore" {}
|
|||||||
variable "vsphere_network" {}
|
variable "vsphere_network" {}
|
||||||
|
|
||||||
variable "vm_name" {}
|
variable "vm_name" {}
|
||||||
|
variable "vm_guestos" {}
|
||||||
variable "ssh_password" {
|
variable "ssh_password" {
|
||||||
sensitive = true
|
sensitive = true
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,197 +0,0 @@
|
|||||||
DeploymentConfigurations:
|
|
||||||
|
|
||||||
- Id: cp1w1
|
|
||||||
Label: 'Workload-cluster: 1 control-plane node/1 worker node'
|
|
||||||
Description: 1 control-plane node/1 worker node
|
|
||||||
|
|
||||||
- Id: cp1w2
|
|
||||||
Label: 'Workload-cluster: 1 control-plane node/2 worker nodes'
|
|
||||||
Description: 1 control-plane node/2 worker nodes
|
|
||||||
|
|
||||||
DynamicDisks:
|
|
||||||
|
|
||||||
- Description: Longhorn persistent storage
|
|
||||||
UnitSize: GB
|
|
||||||
Constraints:
|
|
||||||
Minimum: 100
|
|
||||||
Maximum: ''
|
|
||||||
PropertyCategory: 2
|
|
||||||
|
|
||||||
PropertyCategories:
|
|
||||||
|
|
||||||
- Name: 0) Deployment information
|
|
||||||
ProductProperties:
|
|
||||||
- Key: deployment.type
|
|
||||||
Type: string
|
|
||||||
Value:
|
|
||||||
- cp1w1
|
|
||||||
- cp1w2
|
|
||||||
UserConfigurable: false
|
|
||||||
|
|
||||||
- Name: 1) Meta-cluster
|
|
||||||
ProductProperties:
|
|
||||||
|
|
||||||
- Key: metacluster.fqdn
|
|
||||||
Type: string(1..)
|
|
||||||
Label: Meta-cluster FQDN*
|
|
||||||
Description: Respective subdomains will be available for each component (e.g. storage.example.org); this address should already be configured as a wildcard record within your DNS zone.
|
|
||||||
DefaultValue: meta.k8s.cluster
|
|
||||||
Configurations: '*'
|
|
||||||
UserConfigurable: true
|
|
||||||
|
|
||||||
- key: metacluster.vip
|
|
||||||
Type: ip
|
|
||||||
Label: Meta-cluster virtual IP*
|
|
||||||
Description: Meta-cluster control plane endpoint virtual IP
|
|
||||||
DefaultValue: ''
|
|
||||||
Configurations: '*'
|
|
||||||
UserConfigurable: true
|
|
||||||
|
|
||||||
- key: metacluster.token
|
|
||||||
Type: string(1..)
|
|
||||||
Label: K3s install token*
|
|
||||||
Description: Auto-generated; this value is used to join future new nodes to the metacluster after deployment
|
|
||||||
DefaultValue: '{{ metacluster.token }}'
|
|
||||||
Configurations: '*'
|
|
||||||
UserConfigurable: true
|
|
||||||
|
|
||||||
- Name: 2) Meta-cluster initial node
|
|
||||||
ProductProperties:
|
|
||||||
|
|
||||||
- Key: guestinfo.hostname
|
|
||||||
Type: string(1..15)
|
|
||||||
Label: Hostname*
|
|
||||||
Description: ''
|
|
||||||
DefaultValue: 'meta-{{ hostname.suffix }}'
|
|
||||||
Configurations: '*'
|
|
||||||
UserConfigurable: true
|
|
||||||
|
|
||||||
- Key: metacluster.password
|
|
||||||
Type: password(7..)
|
|
||||||
Label: Appliance password*
|
|
||||||
Description: 'Initial password for respective administrator accounts within each component'
|
|
||||||
DefaultValue: ''
|
|
||||||
Configurations: '*'
|
|
||||||
UserConfigurable: true
|
|
||||||
|
|
||||||
- Key: guestinfo.ipaddress
|
|
||||||
Type: ip
|
|
||||||
Label: IP Address*
|
|
||||||
Description: ''
|
|
||||||
DefaultValue: ''
|
|
||||||
Configurations: '*'
|
|
||||||
UserConfigurable: true
|
|
||||||
|
|
||||||
- Key: guestinfo.prefixlength
|
|
||||||
Type: int(8..32)
|
|
||||||
Label: Subnet prefix length*
|
|
||||||
Description: ''
|
|
||||||
DefaultValue: '24'
|
|
||||||
Configurations: '*'
|
|
||||||
UserConfigurable: true
|
|
||||||
|
|
||||||
- Key: guestinfo.dnsserver
|
|
||||||
Type: ip
|
|
||||||
Label: DNS server*
|
|
||||||
Description: ''
|
|
||||||
DefaultValue: ''
|
|
||||||
Configurations: '*'
|
|
||||||
UserConfigurable: true
|
|
||||||
|
|
||||||
- Key: guestinfo.gateway
|
|
||||||
Type: ip
|
|
||||||
Label: Gateway*
|
|
||||||
Description: ''
|
|
||||||
DefaultValue: ''
|
|
||||||
Configurations: '*'
|
|
||||||
UserConfigurable: true
|
|
||||||
|
|
||||||
# - Key: guestinfo.ntpserver
|
|
||||||
# Type: string(1..)
|
|
||||||
# Label: Time server*
|
|
||||||
# Description: A comma-separated list of timeservers
|
|
||||||
# DefaultValue: 0.pool.ntp.org,1.pool.ntp.org,2.pool.ntp.org
|
|
||||||
# Configurations: '*'
|
|
||||||
# UserConfigurable: true
|
|
||||||
|
|
||||||
- Name: 3) Workload-cluster
|
|
||||||
ProductProperties:
|
|
||||||
|
|
||||||
- Key: workloadcluster.name
|
|
||||||
Type: string(1..15)
|
|
||||||
Label: Workload-cluster name*
|
|
||||||
Description: ''
|
|
||||||
DefaultValue: 'workload-{{ hostname.suffix }}'
|
|
||||||
Configurations: '*'
|
|
||||||
UserConfigurable: true
|
|
||||||
|
|
||||||
- Key: workloadcluster.vip
|
|
||||||
Type: ip
|
|
||||||
Label: Workload-cluster virtual IP*
|
|
||||||
Description: Workload-cluster control plane endpoint virtual IP
|
|
||||||
DefaultValue: ''
|
|
||||||
Configurations: '*'
|
|
||||||
UserConfigurable: true
|
|
||||||
|
|
||||||
- Key: ippool.startip
|
|
||||||
Type: ip
|
|
||||||
Label: Workload-cluster IP-pool start IP*
|
|
||||||
Description: All nodes for the workload-cluster will be provisioned within this IP pool
|
|
||||||
DefaultValue: ''
|
|
||||||
Configurations: '*'
|
|
||||||
UserConfigurable: true
|
|
||||||
|
|
||||||
- Key: ippool.endip
|
|
||||||
Type: ip
|
|
||||||
Label: Workload-cluster IP-pool end IP*
|
|
||||||
Description: All nodes for the workload-cluster will be provisioned within this IP pool
|
|
||||||
DefaultValue: ''
|
|
||||||
Configurations: '*'
|
|
||||||
UserConfigurable: true
|
|
||||||
|
|
||||||
- Name: 4) Common
|
|
||||||
ProductProperties:
|
|
||||||
|
|
||||||
- Key: guestinfo.rootsshkey
|
|
||||||
Type: password(1..)
|
|
||||||
Label: SSH public key*
|
|
||||||
Description: Authentication for any node (meta-cluster *and* workloadcluster); this line should start with 'ssh-rsa AAAAB3N'
|
|
||||||
DefaultValue: ''
|
|
||||||
Configurations: '*'
|
|
||||||
UserConfigurable: true
|
|
||||||
|
|
||||||
- Name: 5) Hypervisor
|
|
||||||
ProductProperties:
|
|
||||||
|
|
||||||
- Key: hv.fqdn
|
|
||||||
Type: string(1..)
|
|
||||||
Label: vCenter FQDN/IP-address*
|
|
||||||
Description: The address of the vCenter instance which this bootstrap appliance will interact with for provisioning new VM's.
|
|
||||||
DefaultValue: ''
|
|
||||||
Configurations: '*'
|
|
||||||
UserConfigurable: true
|
|
||||||
|
|
||||||
- Key: hv.username
|
|
||||||
Type: string(1..)
|
|
||||||
Label: vCenter username*
|
|
||||||
Description: The username which this bootstrap appliance will authenticate with to the vCenter instance.
|
|
||||||
DefaultValue: 'administrator@vsphere.local'
|
|
||||||
Configurations: '*'
|
|
||||||
UserConfigurable: true
|
|
||||||
|
|
||||||
- Key: hv.password
|
|
||||||
Type: password(1..)
|
|
||||||
Label: vCenter password*
|
|
||||||
Description: The password which this bootstrap appliance will authenticate with to the vCenter instance.
|
|
||||||
DefaultValue: ''
|
|
||||||
Configurations: '*'
|
|
||||||
UserConfigurable: true
|
|
||||||
|
|
||||||
---
|
|
||||||
Variables:
|
|
||||||
- Name: hostname.suffix
|
|
||||||
Expression: |
|
|
||||||
(-join ((48..57) + (97..122) | Get-Random -Count 5 | % {[char]$_})).ToLower()
|
|
||||||
- Name: metacluster.token
|
|
||||||
Expression: |
|
|
||||||
(New-Guid).Guid -replace '-', ''
|
|
||||||
@@ -1,9 +1,6 @@
|
|||||||
#Requires -Modules 'powershell-yaml'
|
#Requires -Modules 'powershell-yaml'
|
||||||
[CmdletBinding()]
|
[CmdletBinding()]
|
||||||
Param(
|
Param(
|
||||||
[Parameter(Mandatory)]
|
|
||||||
[ValidateSet('Bootstrap', 'Upgrade')]
|
|
||||||
[string]$ApplianceType,
|
|
||||||
[Parameter(Mandatory)]
|
[Parameter(Mandatory)]
|
||||||
[ValidateScript({
|
[ValidateScript({
|
||||||
If (Test-Path($_)) {
|
If (Test-Path($_)) {
|
||||||
@@ -17,7 +14,7 @@ Param(
|
|||||||
)
|
)
|
||||||
|
|
||||||
$GetContentSplat = @{
|
$GetContentSplat = @{
|
||||||
Path = "$($PSScriptRoot)\$($MyInvocation.MyCommand)".Replace('.ps1', ".$($ApplianceType.ToLower()).yml")
|
Path = "$($PSScriptRoot)\$($MyInvocation.MyCommand)".Replace('.ps1', ".yml")
|
||||||
Raw = $True
|
Raw = $True
|
||||||
}
|
}
|
||||||
$RawContent = Get-Content @GetContentSplat
|
$RawContent = Get-Content @GetContentSplat
|
||||||
@@ -105,7 +102,7 @@ ForEach ($Disk in $OVFConfig.DynamicDisks) {
|
|||||||
$XML.SelectSingleNode("//ns:VirtualHardwareSection/ns:Item/rasd:InstanceID[.='$($HighestInstanceID)']", $NS).ParentNode
|
$XML.SelectSingleNode("//ns:VirtualHardwareSection/ns:Item/rasd:InstanceID[.='$($HighestInstanceID)']", $NS).ParentNode
|
||||||
)
|
)
|
||||||
|
|
||||||
$OVFConfig.PropertyCategories[@([int]$Disk.PropertyCategory, 0)[![boolean]$Disk.PropertyCategory]].ProductProperties += @{
|
$OVFConfig.PropertyCategories[0].ProductProperties += @{
|
||||||
Key = "vmconfig.disksize.$($DiskId)"
|
Key = "vmconfig.disksize.$($DiskId)"
|
||||||
Type = If ([boolean]$Disk.Constraints.Minimum -or [boolean]$Disk.Constraints.Maximum) {
|
Type = If ([boolean]$Disk.Constraints.Minimum -or [boolean]$Disk.Constraints.Maximum) {
|
||||||
"Int($($Disk.Constraints.Minimum)..$($Disk.Constraints.Maximum))"
|
"Int($($Disk.Constraints.Minimum)..$($Disk.Constraints.Maximum))"
|
||||||
|
|||||||
@@ -1,140 +0,0 @@
|
|||||||
DynamicDisks:
|
|
||||||
|
|
||||||
- Description: Longhorn persistent storage
|
|
||||||
UnitSize: GB
|
|
||||||
Constraints:
|
|
||||||
Minimum: 100
|
|
||||||
Maximum: ''
|
|
||||||
PropertyCategory: 1
|
|
||||||
|
|
||||||
PropertyCategories:
|
|
||||||
|
|
||||||
- Name: 1) Existing meta-cluster
|
|
||||||
ProductProperties:
|
|
||||||
|
|
||||||
- Key: metacluster.fqdn
|
|
||||||
Type: string(1..)
|
|
||||||
Label: Meta-cluster FQDN*
|
|
||||||
Description: The FQDN of the target meta-cluster which this appliance will perform an upgrade on.
|
|
||||||
DefaultValue: meta.k8s.cluster
|
|
||||||
Configurations: '*'
|
|
||||||
UserConfigurable: true
|
|
||||||
|
|
||||||
- key: metacluster.vip
|
|
||||||
Type: ip
|
|
||||||
Label: Meta-cluster virtual IP*
|
|
||||||
Description: Meta-cluster control plane endpoint virtual IP
|
|
||||||
DefaultValue: ''
|
|
||||||
Configurations: '*'
|
|
||||||
UserConfigurable: true
|
|
||||||
|
|
||||||
- Key: metacluster.password
|
|
||||||
Type: password(7..)
|
|
||||||
Label: Meta-cluster administrator password*
|
|
||||||
Description: 'Needed to authenticate with target meta-cluster'
|
|
||||||
DefaultValue: ''
|
|
||||||
Configurations: '*'
|
|
||||||
UserConfigurable: true
|
|
||||||
|
|
||||||
- key: metacluster.token
|
|
||||||
Type: string(1..)
|
|
||||||
Label: K3s install token*
|
|
||||||
Description: Must match the token originally used for the target meta-cluster
|
|
||||||
DefaultValue: ''
|
|
||||||
Configurations: '*'
|
|
||||||
UserConfigurable: true
|
|
||||||
|
|
||||||
- Name: 2) Add meta-cluster node
|
|
||||||
ProductProperties:
|
|
||||||
|
|
||||||
- Key: guestinfo.hostname
|
|
||||||
Type: string(1..15)
|
|
||||||
Label: Hostname*
|
|
||||||
Description: ''
|
|
||||||
DefaultValue: 'meta-{{ hostname.suffix }}'
|
|
||||||
Configurations: '*'
|
|
||||||
UserConfigurable: true
|
|
||||||
|
|
||||||
- Key: guestinfo.ipaddress
|
|
||||||
Type: ip
|
|
||||||
Label: IP Address*
|
|
||||||
Description: ''
|
|
||||||
DefaultValue: ''
|
|
||||||
Configurations: '*'
|
|
||||||
UserConfigurable: true
|
|
||||||
|
|
||||||
- Key: guestinfo.prefixlength
|
|
||||||
Type: int(8..32)
|
|
||||||
Label: Subnet prefix length*
|
|
||||||
Description: ''
|
|
||||||
DefaultValue: '24'
|
|
||||||
Configurations: '*'
|
|
||||||
UserConfigurable: true
|
|
||||||
|
|
||||||
- Key: guestinfo.dnsserver
|
|
||||||
Type: ip
|
|
||||||
Label: DNS server*
|
|
||||||
Description: ''
|
|
||||||
DefaultValue: ''
|
|
||||||
Configurations: '*'
|
|
||||||
UserConfigurable: true
|
|
||||||
|
|
||||||
- Key: guestinfo.gateway
|
|
||||||
Type: ip
|
|
||||||
Label: Gateway*
|
|
||||||
Description: ''
|
|
||||||
DefaultValue: ''
|
|
||||||
Configurations: '*'
|
|
||||||
UserConfigurable: true
|
|
||||||
|
|
||||||
# - Key: guestinfo.ntpserver
|
|
||||||
# Type: string(1..)
|
|
||||||
# Label: Time server*
|
|
||||||
# Description: A comma-separated list of timeservers
|
|
||||||
# DefaultValue: 0.pool.ntp.org,1.pool.ntp.org,2.pool.ntp.org
|
|
||||||
# Configurations: '*'
|
|
||||||
# UserConfigurable: true
|
|
||||||
|
|
||||||
- Name: 3) Common
|
|
||||||
ProductProperties:
|
|
||||||
|
|
||||||
- Key: guestinfo.rootsshkey
|
|
||||||
Type: password(1..)
|
|
||||||
Label: SSH public key*
|
|
||||||
Description: Authentication for this meta-cluster node; this line should start with 'ssh-rsa AAAAB3N'
|
|
||||||
DefaultValue: ''
|
|
||||||
Configurations: '*'
|
|
||||||
UserConfigurable: true
|
|
||||||
|
|
||||||
- Name: 4) Hypervisor
|
|
||||||
ProductProperties:
|
|
||||||
|
|
||||||
- Key: hv.fqdn
|
|
||||||
Type: string(1..)
|
|
||||||
Label: vCenter FQDN/IP-address*
|
|
||||||
Description: The address of the vCenter instance which this bootstrap appliance will interact with for provisioning new VM's.
|
|
||||||
DefaultValue: ''
|
|
||||||
Configurations: '*'
|
|
||||||
UserConfigurable: true
|
|
||||||
|
|
||||||
- Key: hv.username
|
|
||||||
Type: string(1..)
|
|
||||||
Label: vCenter username*
|
|
||||||
Description: The username which this bootstrap appliance will authenticate with to the vCenter instance.
|
|
||||||
DefaultValue: 'administrator@vsphere.local'
|
|
||||||
Configurations: '*'
|
|
||||||
UserConfigurable: true
|
|
||||||
|
|
||||||
- Key: hv.password
|
|
||||||
Type: password(1..)
|
|
||||||
Label: vCenter password*
|
|
||||||
Description: The password which this bootstrap appliance will authenticate with to the vCenter instance.
|
|
||||||
DefaultValue: ''
|
|
||||||
Configurations: '*'
|
|
||||||
UserConfigurable: true
|
|
||||||
|
|
||||||
---
|
|
||||||
Variables:
|
|
||||||
- Name: hostname.suffix
|
|
||||||
Expression: |
|
|
||||||
(-join ((48..57) + (97..122) | Get-Random -Count 5 | % {[char]$_})).ToLower()
|
|
||||||
99
scripts/Update-OvfConfiguration.yml
Normal file
99
scripts/Update-OvfConfiguration.yml
Normal file
@@ -0,0 +1,99 @@
|
|||||||
|
DeploymentConfigurations:
|
||||||
|
- Id: small
|
||||||
|
Label: 'Ubuntu Server 20.04 [SMALL: 1 vCPU/2GB RAM]'
|
||||||
|
Description: Ubuntu Server 20.04.x
|
||||||
|
Size:
|
||||||
|
CPU: 1
|
||||||
|
Memory: 2048
|
||||||
|
- Id: large
|
||||||
|
Label: 'Ubuntu Server 20.04 [LARGE: 4 vCPU/8GB RAM]'
|
||||||
|
Description: Ubuntu Server 20.04.x
|
||||||
|
Size:
|
||||||
|
CPU: 4
|
||||||
|
Memory: 8192
|
||||||
|
DynamicDisks: []
|
||||||
|
PropertyCategories:
|
||||||
|
# - Name: 0) Deployment information
|
||||||
|
# ProductProperties:
|
||||||
|
# - Key: deployment.type
|
||||||
|
# Type: string
|
||||||
|
# Value:
|
||||||
|
# - small
|
||||||
|
# - large
|
||||||
|
# UserConfigurable: false
|
||||||
|
- Name: 1) Operating System
|
||||||
|
ProductProperties:
|
||||||
|
- Key: guestinfo.hostname
|
||||||
|
Type: string(1..15)
|
||||||
|
Label: Hostname*
|
||||||
|
Description: '(max length: 15 characters)'
|
||||||
|
DefaultValue: ''
|
||||||
|
Configurations: '*'
|
||||||
|
UserConfigurable: true
|
||||||
|
- Key: guestinfo.rootpw
|
||||||
|
Type: password(7..)
|
||||||
|
Label: Local root password*
|
||||||
|
Description: ''
|
||||||
|
DefaultValue: ''
|
||||||
|
Configurations: '*'
|
||||||
|
UserConfigurable: true
|
||||||
|
- Key: guestinfo.rootsshkey
|
||||||
|
Type: password(1..)
|
||||||
|
Label: Local root SSH public key*
|
||||||
|
Description: This line should start with 'ssh-rsa AAAAB3N'
|
||||||
|
DefaultValue: ''
|
||||||
|
Configurations: '*'
|
||||||
|
UserConfigurable: true
|
||||||
|
- Key: guestinfo.ntpserver
|
||||||
|
Type: string(1..)
|
||||||
|
Label: Time server*
|
||||||
|
Description: A comma-separated list of timeservers
|
||||||
|
DefaultValue: 0.pool.ntp.org,1.pool.ntp.org,2.pool.ntp.org
|
||||||
|
Configurations: '*'
|
||||||
|
UserConfigurable: true
|
||||||
|
- Name: 2) Networking
|
||||||
|
ProductProperties:
|
||||||
|
- Key: guestinfo.ipaddress
|
||||||
|
Type: ip
|
||||||
|
Label: IP Address*
|
||||||
|
Description: ''
|
||||||
|
DefaultValue: ''
|
||||||
|
Configurations: '*'
|
||||||
|
UserConfigurable: true
|
||||||
|
- Key: guestinfo.prefixlength
|
||||||
|
Type: int(8..32)
|
||||||
|
Label: Subnet prefix length*
|
||||||
|
Description: ''
|
||||||
|
DefaultValue: '24'
|
||||||
|
Configurations: '*'
|
||||||
|
UserConfigurable: true
|
||||||
|
- Key: guestinfo.dnsserver
|
||||||
|
Type: ip
|
||||||
|
Label: DNS server*
|
||||||
|
Description: ''
|
||||||
|
DefaultValue: ''
|
||||||
|
Configurations: '*'
|
||||||
|
UserConfigurable: true
|
||||||
|
- Key: guestinfo.gateway
|
||||||
|
Type: ip
|
||||||
|
Label: Gateway*
|
||||||
|
Description: ''
|
||||||
|
DefaultValue: ''
|
||||||
|
Configurations: '*'
|
||||||
|
UserConfigurable: true
|
||||||
|
AdvancedOptions:
|
||||||
|
- Key: appliance.name
|
||||||
|
Value: "{{ appliance.name }}"
|
||||||
|
Required: false
|
||||||
|
- Key: appliance.version
|
||||||
|
Value: "{{ appliance.version }}"
|
||||||
|
Required: false
|
||||||
|
|
||||||
|
---
|
||||||
|
Variables:
|
||||||
|
- Name: appliance.name
|
||||||
|
Expression: |
|
||||||
|
$Parameter['appliance.name']
|
||||||
|
- Name: appliance.version
|
||||||
|
Expression: |
|
||||||
|
$Parameter['appliance.version']
|
||||||
Reference in New Issue
Block a user