Compare commits
1 Commits
fix/ansibl
...
9b6f9b6656
| Author | SHA1 | Date | |
|---|---|---|---|
| 9b6f9b6656 |
@@ -17,7 +17,7 @@ resource kubernetes_pod admin {
|
|||||||
}
|
}
|
||||||
spec {
|
spec {
|
||||||
node_selector = {
|
node_selector = {
|
||||||
NodeType = var.admin_services.namespace
|
"vke.vultr.com/node-pool" = var.admin_services.namespace
|
||||||
}
|
}
|
||||||
container {
|
container {
|
||||||
image = each.value.image
|
image = each.value.image
|
||||||
|
|||||||
@@ -22,7 +22,7 @@ resource vultr_kubernetes_node_pools games {
|
|||||||
label = var.game_servers.namespace
|
label = var.game_servers.namespace
|
||||||
min_nodes = var.cluster.pools["games"].min
|
min_nodes = var.cluster.pools["games"].min
|
||||||
max_nodes = var.cluster.pools["games"].max
|
max_nodes = var.cluster.pools["games"].max
|
||||||
tag = var.admin_services.namespace
|
tag = var.game_servers.namespace
|
||||||
}
|
}
|
||||||
|
|
||||||
output k8s_config {
|
output k8s_config {
|
||||||
|
|||||||
@@ -29,4 +29,3 @@ resource vultr_firewall_rule admin-service-inbound {
|
|||||||
notes = each.value.port.notes
|
notes = each.value.port.notes
|
||||||
port = each.value.port.expose
|
port = each.value.port.expose
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -1,3 +0,0 @@
|
|||||||
static-web:
|
|
||||||
hosts:
|
|
||||||
shockrah.xyz:
|
|
||||||
@@ -1,9 +0,0 @@
|
|||||||
---
|
|
||||||
- name: Pre Pipeline Playbook for Static Hosts
|
|
||||||
hosts: static-web
|
|
||||||
remote_user: root
|
|
||||||
tasks:
|
|
||||||
- name: Import manual setup steps
|
|
||||||
ansible.builtin.import_role:
|
|
||||||
name: static
|
|
||||||
tasks_from: setup-webadmin.yaml
|
|
||||||
@@ -1,5 +0,0 @@
|
|||||||
# What is this
|
|
||||||
|
|
||||||
Here be the ansible based workflows that we use to keep things like the static
|
|
||||||
hosts properly setup with all the resources they need to properly host the
|
|
||||||
services we intended on hosting.
|
|
||||||
@@ -1,8 +0,0 @@
|
|||||||
---
|
|
||||||
- name: Static Host Maintenance and Setup
|
|
||||||
hosts: static-web
|
|
||||||
remote_user: webadmin
|
|
||||||
tasks:
|
|
||||||
- name: Import static host role
|
|
||||||
ansible.builtin.import_role:
|
|
||||||
name: static
|
|
||||||
@@ -1,5 +0,0 @@
|
|||||||
- name: Restart SSH
|
|
||||||
become: true
|
|
||||||
ansible.builtin.systemd:
|
|
||||||
name: sshd
|
|
||||||
state: restarted
|
|
||||||
@@ -1,40 +0,0 @@
|
|||||||
# Things that we definitely want to have are the following
|
|
||||||
# docker docker-compose python(latest) certbot
|
|
||||||
- name: Uhhh yea
|
|
||||||
become: true
|
|
||||||
block:
|
|
||||||
- name: Install base dependencies
|
|
||||||
ansible.builtin.apt:
|
|
||||||
update_cache: true
|
|
||||||
pkg:
|
|
||||||
- ca-certificates
|
|
||||||
- curl
|
|
||||||
- lsb-release
|
|
||||||
- name: Setup keyring directory
|
|
||||||
ansible.builtin.command:
|
|
||||||
cmd: "install -m 0755 -d {{ static_keyring_dir }}"
|
|
||||||
creates: "{{ static_keyring_dir }}"
|
|
||||||
- name: Download the docker GPG key
|
|
||||||
ansible.builtin.get_url:
|
|
||||||
url: "{{ static_docker_ubuntu }}/gpg"
|
|
||||||
dest: "{{ static_keyring_dir }}/docker.asc"
|
|
||||||
mode: "0644"
|
|
||||||
- name: Ensure docker.lst is present
|
|
||||||
vars:
|
|
||||||
key_path: "{{ static_keyring_dir }}/docker.asc"
|
|
||||||
repo: "{{ static_docker_ubuntu }}"
|
|
||||||
os_codename: jammy
|
|
||||||
ansible.builtin.template:
|
|
||||||
src: docker.list
|
|
||||||
dest: "{{ static_apt_sources_dir }}/docker.list"
|
|
||||||
mode: "0644"
|
|
||||||
- name: Install docker and python packages
|
|
||||||
ansible.builtin.apt:
|
|
||||||
update_cache: true
|
|
||||||
pkg:
|
|
||||||
- docker-ce
|
|
||||||
- docker-ce-cli
|
|
||||||
- containerd.io
|
|
||||||
- docker-buildx-plugin
|
|
||||||
- docker-compose-plugin
|
|
||||||
- python3
|
|
||||||
@@ -1,43 +0,0 @@
|
|||||||
- name: Ensure sudo is available
|
|
||||||
ansible.builtin.apt:
|
|
||||||
state: present
|
|
||||||
update_cache: true
|
|
||||||
pkg:
|
|
||||||
- sudo
|
|
||||||
- zsh
|
|
||||||
- name: Create webadmin user
|
|
||||||
ansible.builtin.user:
|
|
||||||
name: webadmin
|
|
||||||
state: present
|
|
||||||
shell: /bin/zsh
|
|
||||||
groups:
|
|
||||||
- nginx
|
|
||||||
append: true
|
|
||||||
- name: Copy webadmin public key
|
|
||||||
ansible.posix.authorized_key:
|
|
||||||
user: webadmin
|
|
||||||
state: present
|
|
||||||
key: "{{ lookup('file', 'files/webadmin.pem.pub') }}"
|
|
||||||
- name: Add webadmin to sudoers
|
|
||||||
ansible.builtin.copy:
|
|
||||||
dest: "/etc/sudoers.d/webadmin"
|
|
||||||
content: "webadmin ALL=(ALL) NOPASSWD: ALL"
|
|
||||||
mode: "0644"
|
|
||||||
owner: root
|
|
||||||
group: root
|
|
||||||
- name: Disable Password Authentication
|
|
||||||
ansible.builtin.lineinfile:
|
|
||||||
dest: /etc/ssh/sshd_config
|
|
||||||
line: PasswordAuthentication no
|
|
||||||
state: present
|
|
||||||
backup: true
|
|
||||||
notify:
|
|
||||||
- Restart SSH
|
|
||||||
- name: Disable root login
|
|
||||||
ansible.builtin.lineinfile:
|
|
||||||
dest: /etc/ssh/sshd_config
|
|
||||||
line: PermitRootLogin no
|
|
||||||
state: present
|
|
||||||
backup: true
|
|
||||||
notify:
|
|
||||||
- Restart SSH
|
|
||||||
@@ -1 +0,0 @@
|
|||||||
deb [arch=amd64 signed-by={{ key_path }}] {{ repo }} {{ os_codename }} stable
|
|
||||||
@@ -1,4 +0,0 @@
|
|||||||
static_keyring_dir: /etc/apt/keyrings
|
|
||||||
static_docker_ubuntu: https://download.docker.com/linux/ubuntu
|
|
||||||
static_apt_sources_dir: /etc/apt/sources.list.d
|
|
||||||
static_codename: jammy
|
|
||||||
45
runbooks/atlas-setup.md
Normal file
45
runbooks/atlas-setup.md
Normal file
@@ -0,0 +1,45 @@
|
|||||||
|
# What this covers
|
||||||
|
|
||||||
|
The creation of Atlas as it happened in order
|
||||||
|
|
||||||
|
## Commands Ran
|
||||||
|
|
||||||
|
Once the infra was provisioned and verified to be configured by Terraform correctly
|
||||||
|
we move on to the following
|
||||||
|
|
||||||
|
```sh
|
||||||
|
# Setup the machine to run docker
|
||||||
|
ansible-playbook -i hosts.ini atlas/init/system-deps.yml
|
||||||
|
|
||||||
|
# Second we copy over the contents of Alpha's mounted docker volumes
|
||||||
|
ansible-playbook -i hosts.ini atlas/init/perma-mount-drives.yml
|
||||||
|
|
||||||
|
# Next we copy over the data that we want to migrate ( if any )
|
||||||
|
ansible-playbook -i hosts.ini -e filebrowser=/path -e clippable=/path atlas/init/migrate-clips-files.yml
|
||||||
|
|
||||||
|
# Setup the services on the host that we want to run
|
||||||
|
ansible-playbook -i hosts.ini atlas/init/setup-containers.yml
|
||||||
|
|
||||||
|
# Next we put up the reverse proxy (nginx)
|
||||||
|
ansible-playbook -i hosts.ini atlas/init/setup-reverse-proxy.yml
|
||||||
|
|
||||||
|
# Finally we add TLS on top of nginx and we're done
|
||||||
|
ansible-playbook -i hosts.ini atlas/init/setup-certbot.yml
|
||||||
|
```
|
||||||
|
|
||||||
|
Maintenance should be straight forward for this machine as TLS is automatically
|
||||||
|
renewed every 3 months by a cron job. We can manually update the certs however
|
||||||
|
if we really want to. They also don't require anymore manual variable injection
|
||||||
|
like Alpha did as the only thing protected was `dev@shockrah.xyz` which is at
|
||||||
|
this point becoming semi-public. This means while it is associated with code
|
||||||
|
it is more of a _business e-mail_ so it can be placed in this repository with
|
||||||
|
very little concern.
|
||||||
|
|
||||||
|
System updates are now also to be fetched with a:
|
||||||
|
|
||||||
|
```sh
|
||||||
|
ansible-playbook -i hosts.ini atlas/maintain/analyze-system-deps.yml
|
||||||
|
```
|
||||||
|
|
||||||
|
Which performs purely read operations and does not affect the state of the
|
||||||
|
machine.
|
||||||
33
runbooks/new-drive.md
Normal file
33
runbooks/new-drive.md
Normal file
@@ -0,0 +1,33 @@
|
|||||||
|
# Mounting an attached drive
|
||||||
|
|
||||||
|
Assumptions:
|
||||||
|
|
||||||
|
* New drive is attached(in AWS) and detected in software
|
||||||
|
Ideally attachment is made through terraform
|
||||||
|
|
||||||
|
## Mounting Instructions (Step-by-Step)
|
||||||
|
|
||||||
|
1. Verify data does not have data: `sudo file -s /dev/xvdf`
|
||||||
|
|
||||||
|
Should return `data` if its ok. Other wise we're probably looking at the wrong
|
||||||
|
drive.
|
||||||
|
|
||||||
|
2. Create the filesystem on the new empty drive: `sudo mkfs -t ext4 /dev/xvdf`
|
||||||
|
|
||||||
|
3. Create mountpoint other wares to actaully use the drive
|
||||||
|
`sudo mkdir /mnt/example`.
|
||||||
|
|
||||||
|
Change _example_ to something that actually makes sense.
|
||||||
|
|
||||||
|
4. Add a new entry to /etc/fstab for automounting
|
||||||
|
|
||||||
|
`/dev/xvdf /newvolume ext4 defaults,nofail 0 0`
|
||||||
|
|
||||||
|
Tab delimited btw.
|
||||||
|
|
||||||
|
5. Mount all drives listed in `/etc/fstab` from before. `sudo mount -a`
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
Reference in New Issue
Block a user