Merge branch 'master' into feat/calico
This commit is contained in:
@@ -1,21 +1,22 @@
|
||||
module "control_planes" {
|
||||
source = "./modules/host"
|
||||
|
||||
count = var.control_plane_count
|
||||
name = "${var.use_cluster_name_in_node_name ? "${var.cluster_name}-" : ""}control-plane"
|
||||
for_each = local.control_plane_nodepools
|
||||
|
||||
name = "${var.use_cluster_name_in_node_name ? "${var.cluster_name}-" : ""}${each.value.nodepool_name}"
|
||||
ssh_keys = [hcloud_ssh_key.k3s.id]
|
||||
public_key = var.public_key
|
||||
private_key = var.private_key
|
||||
additional_public_keys = var.additional_public_keys
|
||||
firewall_ids = [hcloud_firewall.k3s.id]
|
||||
placement_group_id = hcloud_placement_group.k3s.id
|
||||
location = var.location
|
||||
server_type = var.control_plane_server_type
|
||||
ipv4_subnet_id = hcloud_network_subnet.subnet[1].id
|
||||
placement_group_id = var.placement_group_disable ? 0 : element(hcloud_placement_group.control_plane.*.id, ceil(each.value.index / 10))
|
||||
location = each.value.location
|
||||
server_type = each.value.server_type
|
||||
ipv4_subnet_id = hcloud_network_subnet.control_plane[[for i, v in var.control_plane_nodepools : i if v.name == each.value.nodepool_name][0]].id
|
||||
|
||||
# We leave some room so 100 eventual Hetzner LBs that can be created perfectly safely
|
||||
# It leaves the subnet with 254 x 254 - 100 = 64416 IPs to use, so probably enough.
|
||||
private_ipv4 = cidrhost(local.network_ipv4_subnets[1], count.index + 101)
|
||||
private_ipv4 = cidrhost(hcloud_network_subnet.control_plane[[for i, v in var.control_plane_nodepools : i if v.name == each.value.nodepool_name][0]].ip_range, each.value.index + 101)
|
||||
|
||||
labels = {
|
||||
"provisioner" = "terraform",
|
||||
@@ -23,44 +24,45 @@ module "control_planes" {
|
||||
}
|
||||
|
||||
depends_on = [
|
||||
hcloud_network_subnet.subnet
|
||||
hcloud_network_subnet.control_plane
|
||||
]
|
||||
}
|
||||
|
||||
resource "null_resource" "control_planes" {
|
||||
count = var.control_plane_count
|
||||
for_each = local.control_plane_nodepools
|
||||
|
||||
triggers = {
|
||||
control_plane_id = module.control_planes[count.index].id
|
||||
control_plane_id = module.control_planes[each.key].id
|
||||
}
|
||||
|
||||
connection {
|
||||
user = "root"
|
||||
private_key = local.ssh_private_key
|
||||
agent_identity = local.ssh_identity
|
||||
host = module.control_planes[count.index].ipv4_address
|
||||
host = module.control_planes[each.key].ipv4_address
|
||||
}
|
||||
|
||||
# Generating k3s server config file
|
||||
provisioner "file" {
|
||||
content = yamlencode(merge({
|
||||
node-name = module.control_planes[count.index].name
|
||||
server = "https://${element(module.control_planes.*.private_ipv4_address, count.index > 0 ? 0 : 1)}:6443"
|
||||
node-name = module.control_planes[each.key].name
|
||||
server = length(module.control_planes) == 1 ? null : "https://${module.control_planes[each.key].private_ipv4_address == module.control_planes[keys(module.control_planes)[0]].private_ipv4_address ? module.control_planes[keys(module.control_planes)[1]].private_ipv4_address : module.control_planes[keys(module.control_planes)[0]].private_ipv4_address}:6443"
|
||||
token = random_password.k3s_token.result
|
||||
disable-cloud-controller = true
|
||||
disable = local.disable_extras
|
||||
flannel-iface = "eth1"
|
||||
kubelet-arg = ["cloud-provider=external", "volume-plugin-dir=/var/lib/kubelet/volumeplugins"]
|
||||
kube-controller-manager-arg = "flex-volume-plugin-dir=/var/lib/kubelet/volumeplugins"
|
||||
node-ip = module.control_planes[count.index].private_ipv4_address
|
||||
advertise-address = module.control_planes[count.index].private_ipv4_address
|
||||
node-taint = var.allow_scheduling_on_control_plane ? [] : ["node-role.kubernetes.io/master:NoSchedule"]
|
||||
node-label = var.automatically_upgrade_k3s ? ["k3s_upgrade=true"] : []
|
||||
node-ip = module.control_planes[each.key].private_ipv4_address
|
||||
advertise-address = module.control_planes[each.key].private_ipv4_address
|
||||
node-label = each.value.labels
|
||||
node-taint = each.value.taints
|
||||
disable-network-policy = var.cni_plugin == "calico" ? true : var.disable_network_policy
|
||||
},
|
||||
var.cni_plugin == "calico" ? {
|
||||
flannel-backend = "none"
|
||||
} : {}))
|
||||
|
||||
destination = "/tmp/config.yaml"
|
||||
}
|
||||
|
||||
@@ -87,6 +89,6 @@ resource "null_resource" "control_planes" {
|
||||
|
||||
depends_on = [
|
||||
null_resource.first_control_plane,
|
||||
hcloud_network_subnet.subnet
|
||||
hcloud_network_subnet.control_plane
|
||||
]
|
||||
}
|
||||
|
Reference in New Issue
Block a user