Removing playbooks that will not be used for barebones setup

This commit is contained in:
shockrah 2024-02-11 17:05:15 -08:00
parent afd64caebd
commit 2cc9ab0815
16 changed files with 179 additions and 464 deletions

View File

@ -1,15 +0,0 @@
# Further changes to DNS config are going to vary so heavily that I'm not
# even going to bother putting down anything else for dns configuration
# At this point basically just change /etc/resolv.conf to change dns targets
# and then add stuff to the /etc/hosts file as required
---
- hosts: leftcoastlab
tasks:
- name: Install DNSMasq
become: yes
apt:
name: "{{ item }}"
update_cache: yes
loop:
- dnsmasq

View File

@ -1,31 +0,0 @@
# This playbook goes through the process of setting up a simple FTP server on
# the target host.
---
- hosts: leftcoastlab
become: yes
tasks:
- name: Include Vars
include_vars:
dir: ../vars
- name: Ensure Filebrowser Volume Exists
community.docker.docker_volume:
name: file-storage
state: present
- name: Setup Filebrowser Container
community.docker.docker_container:
name: filebrowser
image: filebrowser/filebrowser
restart_policy: always
recreate: yes
volumes:
- "file-storage:/srv"
ports:
- "{{FILEBROWSER_PORT}}:80"
log_driver: loki
log_options:
loki-url: "http://localhost:{{LOKI_PORT}}/loki/api/v1/push"

View File

@ -0,0 +1,25 @@
---
- hosts: lab
become: yes
vars:
lan_subnet: "192.168.1.0/24"
tasks:
- name: Enable Firewall rules for basic LAN access
community.general.ufw:
rule: allow
port: '{{ item }}'
proto: tcp
src: '{{ lan_subnet }}'
loop:
- 22
- 53
- 80
- 443
- name: Enable K3s Ports
community.general.ufw:
rule: allow
port: '{{ item }}'
proto: tcp
loop:
- 6443

View File

@ -1,23 +0,0 @@
# Setup a brand new system from the ground up.
---
- hosts: leftcoastlab
tasks:
- name: Import Global Vars
include_vars:
dir: ../vars
# System pre-reqs
- import_playbook: 'system-update.yml'
- import_playbook: 'install-docker.yml'
- import_playbook: 'dns.yml'
# Logging stack
- import_playbook: 'install-loki-logger.yml'
- import_playbook: 'grafana.yml'
# For host metrics
- import_playbook: 'vector.yml'
# Containerized services
- import_playbook: 'file-server.yml'
- import_playbook: 'git-stack.yml'
- import_playbook: 'subsonic.yml'
- import_playbook: 'nginx.yml'

View File

@ -1,33 +0,0 @@
---
- hosts: leftcoastlab
become: yes
vars:
GITEA_VOLUME: gitea-volume
tasks:
- name: Import Vars required for configuration
include_vars:
dir: ../vars
- name: Create Gitea Volume
community.docker.docker_volume:
name: "{{GITEA_VOLUME}}"
- name: Update/Install Git Tea Container Service
community.docker.docker_container:
name: gitea
image: gitea/gitea
pull: yes
restart_policy: always
recreate: yes
volumes:
- "{{GITEA_VOLUME}}:/data"
- /etc/timezone:/etc/timezone:ro
- /etc/localtime:/etc/localtime:ro
ports:
- "{{GITEA_PUB_PORT}}:3000"
- "{{GITEA_SSH_PORT}}:22"
log_driver: loki
log_options:
loki-url: "http://localhost:{{LOKI_PORT}}/loki/api/v1/push"

View File

@ -1,26 +0,0 @@
---
- hosts: leftcoastlab
become: yes
tasks:
- name: Include Vars
include_vars:
dir: ../vars
- name: Create Grafana Volume
community.docker.docker_volume:
name: grafana-volume
- name: Setup Grafana Container(latest)
community.docker.docker_container:
name: grafana
image: grafana/grafana-oss:latest
pull: yes
restart_policy: always
recreate: yes
timeout: 120
ports:
- "{{GRAFANA_PORT}}:3000"
volumes:
- "grafana-volume:/var/lib/grafana/"

View File

@ -1,29 +0,0 @@
---
- hosts: leftcoastlab
become: yes
tasks:
- name: Install docker dependencies
apt:
name: "{{item}}"
update_cache: yes
loop:
- apt-transport-https
- ca-certificates
- curl
- gnupg
- software-properties-common
- lsb-release
- python3-pip
- name: Install Docker
apt:
name: docker.io
update_cache: yes
- name: Install Docker Ansible Driver
pip:
name: docker

View File

@ -1,12 +0,0 @@
# This playbook installs the loki docker plugin which allows other docker based
# services to do they're regular logging actions but centralize them
---
- hosts: leftcoastlab
become: yes
tasks:
# --grant-all-permissions is the default for ansible
- name: Install Loki Docker Plugin
community.docker.docker_plugin:
plugin_name: grafana/loki-docker-driver:latest
state: enabled
alias: loki

View File

@ -0,0 +1,154 @@
---
# Auto install from the quickstart
# If airgapped, all K3s artifacts are already on the node.
- hosts:
tasks:
- name: Download K3s install script
ansible.builtin.get_url:
url: https://get.k3s.io/
timeout: 120
dest: /usr/local/bin/k3s-install.sh
owner: root
group: root
mode: 0755
- name: Download K3s binary
ansible.builtin.command:
cmd: /usr/local/bin/k3s-install.sh
changed_when: true
- name: Add K3s autocomplete to user bashrc
ansible.builtin.lineinfile:
path: "~{{ ansible_user }}/.bashrc"
regexp: '\.\s+<\(k3s completion bash\)'
line: ". <(k3s completion bash) # Added by k3s-ansible"
- name: Enable and check K3s service
ansible.builtin.systemd:
name: k3s
daemon_reload: true
state: started
enabled: true
- name: Pause to allow first server startup
when: (groups['server'] | length) > 1
ansible.builtin.pause:
seconds: 10
- name: Check whether kubectl is installed on control node
ansible.builtin.command: 'kubectl'
register: kubectl_installed
ignore_errors: true
delegate_to: 127.0.0.1
become: false
changed_when: false
- name: Apply K3S kubeconfig to control node
when: kubectl_installed.rc == 0
block:
- name: Copy kubeconfig to control node
ansible.builtin.fetch:
src: /etc/rancher/k3s/k3s.yaml
dest: "{{ kubeconfig }}"
flat: true
- name: Change server address in kubeconfig on control node
ansible.builtin.shell: |
KUBECONFIG={{ kubeconfig }} kubectl config set-cluster default --server=https://{{ api_endpoint }}:{{ api_port }}
delegate_to: 127.0.0.1
become: false
register: csa_result
changed_when:
- csa_result.rc == 0
- name: Setup kubeconfig k3s-ansible context on control node
when: kubeconfig != "~/.kube/config"
ansible.builtin.replace:
path: "{{ kubeconfig }}"
regexp: 'name: default'
replace: 'name: k3s-ansible'
delegate_to: 127.0.0.1
become: false
- name: Merge with any existing kubeconfig on control node
when: kubeconfig != "~/.kube/config"
ansible.builtin.shell: |
TFILE=$(mktemp)
KUBECONFIG={{ kubeconfig }} kubectl config set-context k3s-ansible --user=k3s-ansible --cluster=k3s-ansible
KUBECONFIG={{ kubeconfig }} kubectl config view --flatten > ${TFILE}
mv ${TFILE} {{ kubeconfig }}
delegate_to: 127.0.0.1
become: false
register: mv_result
changed_when:
- mv_result.rc == 0
- name: Start other server if any and verify status
when:
- (groups['server'] | length) > 1
- inventory_hostname != groups['server'][0]
block:
- name: Copy K3s service file [HA]
when: groups['server'] | length > 1
ansible.builtin.template:
src: "k3s-ha.service.j2"
dest: "{{ systemd_dir }}/k3s.service"
owner: root
group: root
mode: 0644
- name: Enable and check K3s service
ansible.builtin.systemd:
name: k3s
daemon_reload: true
state: started
enabled: true
- name: Verify that all server nodes joined
when: (groups['server'] | length) > 1
ansible.builtin.command:
cmd: >
k3s kubectl get nodes -l "node-role.kubernetes.io/control-plane=true" -o=jsonpath="{.items[*].metadata.name}"
register: nodes
until: nodes.rc == 0 and (nodes.stdout.split() | length) == (groups['server'] | length)
retries: 20
delay: 10
changed_when: false
- name: Setup kubectl for user
block:
- name: Create kubectl symlink
when: lookup('fileglob', '/usr/local/bin/kubectl', errors='warn') | length == 0
ansible.builtin.file:
src: /usr/local/bin/k3s
dest: /usr/local/bin/kubectl
state: link
- name: Create directory .kube
ansible.builtin.file:
path: ~{{ ansible_user }}/.kube
state: directory
owner: "{{ ansible_user }}"
mode: "u=rwx,g=rx,o="
- name: Copy config file to user home directory
ansible.builtin.copy:
src: /etc/rancher/k3s/k3s.yaml
dest: ~{{ ansible_user }}/.kube/config
remote_src: true
owner: "{{ ansible_user }}"
mode: "u=rw,g=,o="
- name: Configure default KUBECONFIG for user
ansible.builtin.lineinfile:
path: ~{{ ansible_user }}/.bashrc
regexp: 'export KUBECONFIG=~/.kube/config'
line: 'export KUBECONFIG=~/.kube/config # Added by k3s-ansible'
state: present
- name: Configure kubectl autocomplete
ansible.builtin.lineinfile:
path: ~{{ ansible_user }}/.bashrc
regexp: '\.\s+<\(kubectl completion bash\)'
line: ". <(kubectl completion bash) # Added by k3s-ansible"

View File

@ -1,63 +0,0 @@
# Sets up an instance of Kanboard from top to bottom
---
- hosts: leftcoastlab
become: yes
vars:
PRE: kanboard
DATA_VOLUME: "{{PRE}}-data-volume"
PLUGINS_VOLUME: "{{PRE}}-plugins-volume"
SSL_VOLUME: "{{PRE}}-ssl-volume"
tasks:
- name: Include Vars
include_vars:
dir: ../vars
- name: Ensure Data volume exists
community.docker.docker_volume:
name: "{{DATA_VOLUME}}"
- name: Ensure Plugins Volume exists
community.docker.docker_volume:
name: "{{PLUGINS_VOLUME}}"
- name: Ensure SSL Volume exists
community.docker.docker_volume:
name: "{{SSL_VOLUME}}"
- name: Ensure Docker service is active
community.docker.docker_container:
name: kanboard
image: kanboard/kanboard:latest
pull: yes
restart_policy: always
recreate: yes
ports:
- "{{ KANBOARD_PORT }}:80"
- "{{ KANBOARD_PORT_SSL }}:443"
volumes:
- "{{DATA_VOLUME}}:/var/www/app/data"
- "{{PLUGINS_VOLUME}}:/var/www/app/plugins"
- "{{SSL_VOLUME}}:/etc/nginx/ssl"
- name: Create Nginx Config for service
template:
dest: /etc/nginx/sites-available/todo.lablad
src: ../templates/proxy.conf.j2
notify:
- restart-nginx
vars:
SERVER_NAME: todo.lablad
PORT: KANBOARD_PORT
- name: Restart Nginx
service:
name: nginx
state: restarted

View File

@ -1,8 +0,0 @@
---
- hosts: leftcoastlab
become: yes
tasks:
- name: Copy landing page
copy:
src: "{{LANDING_PAGE_ROOT_PATH}}/public/"
dest: /var/www/html/

View File

@ -1,77 +0,0 @@
---
- hosts: leftcoastlab
become: yes
vars:
STATIC_USER: staticweb
tasks:
- name: Install Nginx Package
apt:
name: nginx
update_cache: yes
state: present
- name: Stop Apache2 Service
systemd:
name: apache2
state: stopped
enabled: no
- name: Remove Apache2
apt:
name: apache2
state: absent
- name: Ensure the default website is disabled
file:
path: /etc/nginx/sites-enabled/default
state: absent
- name: website user account
user:
name: "{{STATIC_USER}}"
comment: Static Website maintainer account
state: present
- name: Create Static web root directory
file:
path: /var/www
state: directory
recurse: yes
owner: "{{STATIC_USER}}"
group: "{{STATIC_USER}}"
- name: Copy Reverse proxy configs
copy:
src: "{{ item }}"
dest: /etc/nginx/sites-available/
loop:
- ../files/git.lablad
- ../files/music.lablad
- ../files/files.lablad
- ../files/metrics.lablad
- ../files/todo.lablad
- ../files/home.lablad
- ../files/vault.lablad
- name: Enable Sites
file:
src: /etc/nginx/sites-available/{{item}}
dest: /etc/nginx/sites-enabled/{{item}}
state: link
loop:
- git.lablad
- music.lablad
- files.lablad
- metrics.lablad
- vault.lablad
- todo.lablad
- home.lablad
notify:
- restart-nginx
handlers:
- import_tasks: ../handlers/nginx.yml

View File

@ -1,63 +0,0 @@
---
- hosts: leftcoastlab
vars:
PROM_URL: "https://github.com/prometheus/node_exporter/releases/download/v1.3.1/node_exporter-1.3.1.linux-amd64.tar.gz"
become: yes
tasks:
- name: Include Global Vars
include_vars:
dir: ../vars
- name: Download prometheus Node Explorer static binary tarball
get_url:
url: "{{ PROM_URL }}"
dest: /tmp/prom-node-exporer.tar.gz
- name: Untar node exporter tarball package
unarchive:
src: /tmp/prom-node-exporer.tar.gz
dest: /opt/prom-node-exporter
creates: /opt/prom-node-exporter
remote_src: yes
- name: Copy Unit file for node exporter
copy:
src: ../files/services/node_exporter.service
dest: /etc/systemd/system/node_exporter.service
- name: Install Node Exporter Service and Start
systemd:
daemon_reload: yes
name: node_exporter
state: restarted
enabled: yes
- name: Ensure Prom Docker Volume Exists
community.docker.docker_volume:
name: prometheus-volume
- name: Ensure Config directory exists
file:
path: /opt/prometheus
state: directory
- name: Copy Prometheus template to target directory
template:
dest: /opt/prometheus/prometheus.yml
src: ../templates/prometheus.yml.j2
- name: Setup Prometheus (Containerized) Service
community.docker.docker_container:
name: prometheus
image: prom/prometheus
pull: yes
restart_policy: always
recreate: yes
volumes:
- "prometheus-volume:/prometheus"
- "/opt/prometheus/prometheus.yml:/etc/prometheus/prometheus.yml"
ports:
- "{{PROM_PORT}}:{{PROM_PORT}}"

View File

@ -1,37 +0,0 @@
# This playbook sets up subsonic with/without a fresh to work with
---
- hosts: leftcoastlab
become: yes
vars:
SUBSONIC_DIR: /opt/subsonic/data
SUBSONIC_CONFIG_VOL: subsonic-config-volume
tasks:
- name: Include Vars
include_vars:
dir: ../vars
- name: Ensure Subsonic music directory exists
file:
path: "{{SUBSONIC_DIR}}"
state: directory
- name: Ensure Subsonic configuration volume exists
community.docker.docker_volume:
name: "{{SUBSONIC_CONFIG_VOL}}"
state: present
- name: Deploy Subsonic container
community.docker.docker_container:
name: subsonic
image: gzurowski/subsonic
restart_policy: always
recreate: yes
ports:
- "{{SUBSONIC_PORT}}:4040"
volumes:
- "{{SUBSONIC_DIR}}:/var/music"
log_driver: loki
log_options:
loki-url: "http://localhost:{{LOKI_PORT}}/loki/api/v1/push"

View File

@ -1,47 +0,0 @@
---
- hosts: leftcoastlab
tasks:
- name: Pull In Vars
include_vars:
dir: ../vars
- name: Create Vault Bind Root
become: yes
file:
path: /opt/vault
state: directory
- name: Ensure Vault Bind directories in place
become: yes
file:
path: "/opt/vault/{{item}}"
state: directory
loop:
- config
- logs
- file
- name: Create Vault Config
become: yes
template:
src: ../templates/vault-config.json.j2
dest: /opt/vault/config/vault.json
- name: Start up Vault Service
community.docker.docker_container:
name: vault
image: vault
ports:
- "{{VAULT_PORT}}:8200"
restart_policy: always
volumes:
- /opt/vault/logs:/vault/logs
- /opt/vault/file:/vault/file
- /opt/vault/config:/vault/config
env:
VAULT_ADDR: http://127.0.0.1:8200
capabilities:
- IPC_LOCK
entrypoint:
- vault server -config=/vault/config/vault.json