• Stars
    star
    141
  • Rank 259,301 (Top 6 %)
  • Language HCL
  • License
    MIT License
  • Created over 3 years ago
  • Updated about 2 years ago

Reviews

There are no reviews yet. Be the first to send feedback to the community and the maintainers!

Repository Details

Low cost Kubernetes stack for startups, prototypes, and playgrounds on Hetzner Cloud.

k-andy

Zero friction Kubernetes stack on Hetzner Cloud

This terraform module will install a High Availability K3s Cluster with Embedded DB in a private network on Hetzner Cloud. The following resources are provisionised by default (20€/mo):

  • 3x Control-plane: CX11, 2GB RAM, 1VCPU, 20GB NVMe, 20TB Traffic.
  • 2x Worker: CX21, 4GB RAM, 2VCPU, 40GB NVMe, 20TB Traffic.
  • Network: Private network with one subnet.
  • Server and agent nodes are distributed across 3 Datacenters (nbg1, fsn1, hel1) for high availability.



Note: Are you looking for the next generation API Developer Platform? 🔎 Have a look at: WunderGraph Turn your services, databases and 3rd party APIs into a secure unified API in just a few minutes. 🪄


What is K3s?

K3s is a lightweight certified kubernetes distribution. It's packaged as single binary and comes with solid defaults for storage and networking but we replaced local-path-provisioner with hetzner CSI-driver and klipper load-balancer with hetzner Cloud Controller Manager. The default ingress controller (traefik) has been disabled.

Hetzner Cloud integration:

Auto-K3s-Upgrades

Enable the upgrade-controller (enable_upgrade_controller = true) and specify your target k3s version (upgrade_k3s_target_version). See here for possible versions.

Label the nodes you want to upgrade, e.g. kubectl label nodes core-control-plane-1 k3s-upgrade=true. The concurrency of the upgrade plan is set to 1, so you can also label them all at once. Agent nodes will be drained one by one during the upgrade.

You can label all control-plane nodes by using kubectl label nodes -l node-role.kubernetes.io/control-plane=true k3s-upgrade=true. All agent nodes can be labelled using kubectl label nodes -l !node-role.kubernetes.io/control-plane k3s-upgrade=true.

To remove the label from all nodes you can run kubectl label nodes --all k3s-upgrade-.

After a successful update you can also remove the upgrade controller and the plans again, setting enable_upgrade_controller to false.

Usage

See a more detailed example with walk-through in the example folder.

Inputs

Name Description Type Default Required
agent_groups Configuration of agent groups
map(object({
type = string
count = number
ip_offset = number
taints = list(string)
}))
{
"default": {
"count": 2,
"ip_offset": 33,
"taints": [],
"type": "cx21"
}
}
no
cluster_cidr Network CIDR to use for pod IPs string "10.42.0.0/16" no
control_plane_already_initialized Use this if you have to replace the first control plane and want the primary to join other already existing ones and not do an init anymore. You have to update control_plane_primary_index to something else too. bool false no
control_plane_primary_index Which of the servers should be the primary to connect to? If you change it from 1, also set control_plane_already_initialized to true. (1-indexed!) number 1 no
control_plane_server_count Number of control plane nodes number 3 no
control_plane_server_type Server type of control plane servers string "cx11" no
create_kubeconfig Create a local kubeconfig file to connect to the cluster bool true no
enable_upgrade_controller Install the rancher system-upgrade-controller bool false no
hcloud_csi_driver_version n/a string "v1.6.0" no
hcloud_token Token to authenticate against Hetzner Cloud any n/a yes
k3s_version K3s version string "v1.21.3+k3s1" no
kubeconfig_filename Specify the filename of the created kubeconfig file (defaults to kubeconfig-${var.name}.yaml any null no
name Cluster name (used in various places, don't use special chars) any n/a yes
network_cidr Network in which the cluster will be placed. Ignored if network_id is defined string "10.0.0.0/16" no
network_id If specified, no new network will be created. Make sure cluster_cidr and service_cidr don't collide with anything in the existing network. any null no
server_additional_packages Additional packages which will be installed on node creation list(string) [] no
server_locations Server locations in which servers will be distributed list(string)
[
"nbg1",
"fsn1",
"hel1"
]
no
service_cidr Network CIDR to use for services IPs string "10.43.0.0/16" no
ssh_private_key_location Use this private SSH key instead of generating a new one (Attention: Encrypted keys are not supported) string null no
subnet_cidr Subnet in which all nodes are placed string "10.0.1.0/24" no
upgrade_controller_image_tag The image tag of the upgrade controller (See https://github.com/rancher/system-upgrade-controller/releases) string "v0.8.0" no
upgrade_controller_kubectl_image_tag rancher/kubectl image tag string "v1.21.5" no
upgrade_k3s_target_version Target version of k3s (See https://github.com/k3s-io/k3s/releases) string null no
upgrade_node_additional_tolerations List of tolerations which upgrade jobs must have to run on every node (for control-plane and agents) list(map(any)) [] no

Outputs

Name Description
agents_public_ips The public IP addresses of the agent servers
cidr_block n/a
control_planes_public_ips The public IP addresses of the control plane servers
k3s_token Secret k3s authentication token
kubeconfig Structured kubeconfig data to supply to other providers
kubeconfig_file Kubeconfig file content with external IP address
network_id n/a
server_locations Array of hetzner server locations we deploy to
ssh_private_key Key to SSH into nodes
subnet_id n/a

Common Operations

Agent server replacement (common case)

If you need to cycle an agent, you can do that with a single node following this procedure. Replace the group name and number with the server you want to recreate!

Make sure you drain the nodes first.

kubectl drain that-agent
terraform taint 'module.my_cluster.module.agent_group["GROUP_NAME"].random_pet.agent_suffix[1]'
terraform apply

This will recreate the agent in that group on next apply.

Sophisticated agent server replacement

If you did some weird config change or recreate them by changing the base k3s version in the terraform configuration and terraform wants to replace all your agents at once you can do this. Replacing all by one is probably not a good idea.

Example for replacement of one agent (the first one of that group):

kubectl drain that-agent
terragrunt taint 'module.agent_group["GROUP_NAME"].random_pet.agent_suffix[0]'
terraform apply --target='module.agent_group["GROUP_NAME"].hcloud_server.agent["#0"]' --target='module.agent_group["GROUP_NAME"].hcloud_server_network.agent["#0"]' --target='module.agent_group["GROUP_NAME"].random_pet.agent_suffix[0]'

Control Plane server replacement

Control plane servers do not get recreated when the user-data for cloud-init changes. If you want to recreate one after you changed something which would change the cloud-init you need to taint them.

Primary server

If you for some reason need to replace the primary control plane, you'll need to tell it to join the others.

Set the variable control_plane_primary_index to one of the other control plane nodes (e.g. 2 or 3). Also set control_plane_already_initialized to true so it won't run a cluster-init again. This will make the primary connect to control-plane 2 or 3 after recreation.

Secondary servers

This is how you can replace the servers which didn't initialize the cluster.

terraform taint 'module.my_cluster.hcloud_server.control_plane["#1"]'
terraform apply

Auto-Upgrade

Prerequisite

Install the system-upgrade-controller in your cluster.

KUBECONFIG=kubeconfig.yaml kubectl apply -f ./upgrade/controller.yaml

Upgrade procedure

  1. Mark the nodes you want to upgrade (The script will mark all nodes).
KUBECONFIG=kubeconfig.yaml kubectl label --all node k3s-upgrade=true
  1. Run the plan for the servers.
KUBECONFIG=kubeconfig.yaml kubectl apply -f ./upgrade/server-plan.yaml

Warning: Wait for completion before you start upgrading your agents.

  1. Run the plan for the agents.
KUBECONFIG=kubeconfig.yaml kubectl apply -f ./upgrade/agent-plan.yaml

Backups

K3s will automatically backup your embedded etcd datastore every 12 hours to /var/lib/rancher/k3s/server/db/snapshots/. You can reset the cluster by pointing to a specific snapshot.

  1. Stop the master server.
sudo systemctl stop k3s
  1. Restore the master server with a snapshot
./k3s server \
  --cluster-reset \
  --cluster-reset-restore-path=<PATH-TO-SNAPSHOT>

Warning: This forget all peers and the server becomes the sole member of a new cluster. You have to manually rejoin all servers.

  1. Connect you with the different servers. Backup and delete /var/lib/rancher/k3s/server/db on each server.
sudo systemctl stop k3s
rm -rf /var/lib/rancher/k3s/server/db
sudo systemctl start k3s

This will rejoin the server one after another. After some time, all servers should be in sync again. Run kubectl get node to verify it.

Info: It exists no official tool to automate the procedure. In future, rancher might provide an operator to handle this (issue).

Debugging

Cloud init logs can be found on the remote machines in:

  • /var/log/cloud-init-output.log
  • /var/log/cloud-init.log
  • journalctl -u k3s.service -e last logs of the server
  • journalctl -u k3s-agent.service -e last logs of the agent

Credits

More Repositories

1

go-web

Modern Web Application with Golang
Go
249
star
2

profiling-nodejs

🌌 Collection of articles and tools to efficiently profile Node.js
148
star
3

next-localization

Lightweight Internationalization (i18n) library for Next.js 10+
JavaScript
123
star
4

FastGraph

Smart GraphQL CDN on Cloudflare edges with zero configuration
TypeScript
107
star
5

graphql-registry

GraphQL registry - single source of truth for registering and tracking your graph.
TypeScript
86
star
6

apollo-datasource-http

Optimized JSON HTTP Data Source for Apollo Server
TypeScript
73
star
7

k8s-gitops

The GitOps workflow to manage Kubernetes applications at any scale (without server components).
40
star
8

branch-comparer

Checkout multiple git branches, execute scripts and log the results
JavaScript
26
star
9

graphql-parser-bench

GraphQL parser comparison in different languages
Go
23
star
10

gulp-prestashop

Gulp prestashop build system
JavaScript
14
star
11

fay

Stateless, Fast and Reliable PDF rendering service.
Go
13
star
12

prettyhtml-vscode

Visual Studio Code extension for Prettyhtml https://github.com/Prettyhtml/prettyhtml
TypeScript
10
star
13

bootme

Configurable and extendable Task pipeline. Define hooks and revert your changes on failure.
JavaScript
9
star
14

judge-framework

Judge-Framework for decision making in large teams - http://judge-framework.starptech.de/
JavaScript
9
star
15

sveltejs-brunch

Compile Svelte components inside Brunch projects
JavaScript
6
star
16

ghconfig

Manage Github Workflows and Dependabot files as a fleet.
Go
5
star
17

golang-examples

The Syntax of Go and especially inheritance, the type system and the object-oriented characteristics are very differently as in beginner languages like Java and PHP.
Go
5
star
18

dart-nodejs

Compare https://www.dartlang.org/ and Node.js for server-side development
Dart
4
star
19

shapeleak

Find subsequent changes in Objects which can result in unoptimized code
JavaScript
3
star
20

zkmicro

Zookeeper as Centralized configuration management & Group membership and name services for a microservice landscape
JavaScript
2
star
21

peerCam

WIP - Video Sharing over WebTorrent
JavaScript
2
star
22

parcel-plugin-razor

Parcel plugin to handle Razor templates as entry points
JavaScript
2
star
23

shikaka

Opinionated UI component library bundler
JavaScript
2
star
24

nextjs-rkoa-505

TypeScript
1
star
25

starptech.com

My personal website
TypeScript
1
star
26

ViRest

ViRest creates a simple virtual JSON-REST interface to test your frontend against a fault-resilient backend for development purpose.
JavaScript
1
star
27

GraphQL

Experimental Graph grammar to get the most of the line.
Java
1
star