Skip to content

Commit

Permalink
Use global HTTP/TCP proxy load balancing for Ingress on GCP
Browse files Browse the repository at this point in the history
* Switch Ingress from regional network load balancers to global
HTTP/TCP Proxy load balancing
* Reduce cost by ~$19/month per cluster. Google bills the first 5
global and regional forwarding rules separately. Typhoon clusters now
use 3 global and 0 regional forwarding rules.
* Worker pools no longer include an extraneous load balancer. Remove
worker module's `ingress_static_ip` output.
* Add `ingress_static_ipv4` output variable
* Add `worker_instance_group` output to allow custom global load
balancing
* Deprecate `controllers_ipv4_public` module output
* Deprecate `ingress_static_ip` module output. Use `ingress_static_ipv4`
  • Loading branch information
dghubble committed Jun 23, 2018
1 parent 2eaf04c commit b517e83
Show file tree
Hide file tree
Showing 7 changed files with 175 additions and 77 deletions.
17 changes: 13 additions & 4 deletions CHANGES.md
Original file line number Diff line number Diff line change
Expand Up @@ -11,12 +11,10 @@ Notable changes between versions.

* Switch `kube-apiserver` port from 443 to 6443 ([#248](https://github.com/poseidon/typhoon/pull/248))
* Combine apiserver and ingress NLBs ([#249](https://github.com/poseidon/typhoon/pull/249))
* Simplify clusters to come with one NLB. Reduce cost by ~$18/month per cluster.
* Reduce cost by ~$18/month per cluster. Typhoon AWS clusters now use one network load balancer
* Users may keep using CNAME records to `ingress_dns_name` and the `nginx-ingress` addon for Ingress (up to a few million RPS)
* Users with heavy traffic (many million RPS) should create a separate NLB(s) for Ingress instead
* Listen for apiserver traffic on port 6443 and forward to controllers (with healthy apiserver)
* Listen for ingress traffic on ports 80/443 and forward to workers (with healthy ingress controller)
* Worker pools (advanced) no longer include an extraneous load balancer
* Worker pools no longer include an extraneous load balancer
* Disable detailed (paid) monitoring on worker nodes ([#251](https://github.com/poseidon/typhoon/pull/251))
* Favor Prometheus for cloud-agnostic metrics, aggregation, alerting, and visualization

Expand All @@ -31,6 +29,17 @@ Notable changes between versions.
* Switch `kube-apiserver` port from 443 to 6443 ([#248](https://github.com/poseidon/typhoon/pull/248))
* Update firewall rules and generated kubeconfig's

#### Google Cloud

* Use global HTTP and TCP proxy load balancing for Kubernetes Ingress
* Switch Ingress from regional network load balancers to global HTTP/TCP Proxy load balancing
* Reduce cost by ~$19/month per cluster. Google bills the first 5 global and regional forwarding rules separately. Typhoon clusters now use 3 global and 0 regional forwarding rules.
* Worker pools no longer include an extraneous load balancer. Remove worker module's `ingress_static_ip` output.
* Add `ingress_static_ipv4` output variable
* Add `worker_instance_group` output to allow custom global load balancing
* Deprecate `controllers_ipv4_public` module output
* Deprecate `ingress_static_ip` module output. Use `ingress_static_ipv4`

#### Addons

* Update CLUO from v0.6.0 to v0.7.0 ([#242](https://github.com/poseidon/typhoon/pull/242))
Expand Down
96 changes: 96 additions & 0 deletions google-cloud/container-linux/kubernetes/ingress.tf
Original file line number Diff line number Diff line change
@@ -0,0 +1,96 @@
# Static IPv4 address for the TCP Proxy Load Balancer
resource "google_compute_global_address" "ingress-ipv4" {
name = "${var.cluster_name}-ingress-ip"
ip_version = "IPV4"
}

# Forward IPv4 TCP traffic to the HTTP proxy load balancer
# Google Cloud does not allow TCP proxies for port 80. Must use HTTP proxy.
resource "google_compute_global_forwarding_rule" "ingress-http" {
name = "${var.cluster_name}-ingress-http"
ip_address = "${google_compute_global_address.ingress-ipv4.address}"
ip_protocol = "TCP"
port_range = "80"
target = "${google_compute_target_http_proxy.ingress-http.self_link}"
}

# Forward IPv4 TCP traffic to the TCP proxy load balancer
resource "google_compute_global_forwarding_rule" "ingress-https" {
name = "${var.cluster_name}-ingress-https"
ip_address = "${google_compute_global_address.ingress-ipv4.address}"
ip_protocol = "TCP"
port_range = "443"
target = "${google_compute_target_tcp_proxy.ingress-https.self_link}"
}

# HTTP proxy load balancer for ingress controllers
resource "google_compute_target_http_proxy" "ingress-http" {
name = "${var.cluster_name}-ingress-http"
description = "Distribute HTTP load across ${var.cluster_name} workers"
url_map = "${google_compute_url_map.ingress-http.self_link}"
}

# TCP proxy load balancer for ingress controllers
resource "google_compute_target_tcp_proxy" "ingress-https" {
name = "${var.cluster_name}-ingress-https"
description = "Distribute HTTPS load across ${var.cluster_name} workers"
backend_service = "${google_compute_backend_service.ingress-https.self_link}"
}

# HTTP URL Map (required)
resource "google_compute_url_map" "ingress-http" {
name = "${var.cluster_name}-ingress-http"
# Do not add host/path rules for applications here. Use Ingress resources.
default_service = "${google_compute_backend_service.ingress-http.self_link}"
}

# Backend service backed by managed instance group of workers
resource "google_compute_backend_service" "ingress-http" {
name = "${var.cluster_name}-ingress-http"
description = "${var.cluster_name} ingress service"

protocol = "HTTP"
port_name = "http"
session_affinity = "NONE"
timeout_sec = "60"

backend {
group = "${module.workers.instance_group}"
}

health_checks = ["${google_compute_health_check.ingress.self_link}"]
}

# Backend service backed by managed instance group of workers
resource "google_compute_backend_service" "ingress-https" {
name = "${var.cluster_name}-ingress-https"
description = "${var.cluster_name} ingress service"

protocol = "TCP"
port_name = "https"
session_affinity = "NONE"
timeout_sec = "60"

backend {
group = "${module.workers.instance_group}"
}

health_checks = ["${google_compute_health_check.ingress.self_link}"]
}

# Ingress HTTP Health Check
resource "google_compute_health_check" "ingress" {
name = "${var.cluster_name}-ingress-health"
description = "Health check for Ingress controller"

timeout_sec = 5
check_interval_sec = 5

healthy_threshold = 2
unhealthy_threshold = 4

http_health_check {
port = 10254
request_path = "/healthz"
}
}
14 changes: 14 additions & 0 deletions google-cloud/container-linux/kubernetes/network.tf
Original file line number Diff line number Diff line change
Expand Up @@ -161,3 +161,17 @@ resource "google_compute_firewall" "internal-kubelet-readonly" {
source_tags = ["${var.cluster_name}-controller", "${var.cluster_name}-worker"]
target_tags = ["${var.cluster_name}-controller", "${var.cluster_name}-worker"]
}

resource "google_compute_firewall" "google-health-checks" {
name = "${var.cluster_name}-google-health-checks"
network = "${google_compute_network.network.name}"

allow {
protocol = "tcp"
ports = [10254]
}

# https://cloud.google.com/compute/docs/load-balancing/tcp-ssl/tcp-proxy#health-checking
source_ranges = ["130.211.0.0/22", "35.191.0.0/16"]
target_tags = ["${var.cluster_name}-worker"]
}
26 changes: 22 additions & 4 deletions google-cloud/container-linux/kubernetes/outputs.tf
Original file line number Diff line number Diff line change
Expand Up @@ -3,12 +3,17 @@ output "controllers_ipv4_public" {
value = ["${google_compute_instance.controllers.*.network_interface.0.access_config.0.assigned_nat_ip}"]
}

output "ingress_static_ip" {
value = "${module.workers.ingress_static_ip}"
# Outputs for Kubernetes Ingress

output "ingress_static_ipv4" {
description = "Global IPv4 address for proxy load balancing to the nearest Ingress controller"
value = "${google_compute_global_address.ingress-ipv4.address}"
}

output "network_self_link" {
value = "${google_compute_network.network.self_link}"
# Deprecated, use ingress_static_ipv4
output "ingress_static_ip" {
description = "Global IPv4 address for proxy load balancing to the nearest Ingress controller"
value = "${google_compute_global_address.ingress-ipv4.address}"
}

# Outputs for worker pools
Expand All @@ -20,3 +25,16 @@ output "network_name" {
output "kubeconfig" {
value = "${module.bootkube.kubeconfig}"
}

# Outputs for custom firewalling

output "network_self_link" {
value = "${google_compute_network.network.self_link}"
}

# Outputs for custom load balancing

output "worker_instance_group" {
description = "Full URL of the worker managed instance group"
value = "${module.workers.instance_group}"
}
45 changes: 0 additions & 45 deletions google-cloud/container-linux/kubernetes/workers/ingress.tf

This file was deleted.

5 changes: 3 additions & 2 deletions google-cloud/container-linux/kubernetes/workers/outputs.tf
Original file line number Diff line number Diff line change
@@ -1,3 +1,4 @@
output "ingress_static_ip" {
value = "${google_compute_address.ingress-ip.address}"
output "instance_group" {
description = "Full URL of the worker managed instance group"
value = "${google_compute_region_instance_group_manager.workers.instance_group}"
}
49 changes: 27 additions & 22 deletions google-cloud/container-linux/kubernetes/workers/workers.tf
Original file line number Diff line number Diff line change
@@ -1,5 +1,4 @@
# Regional managed instance group maintains a homogeneous set of workers that
# span the zones in the region.
# Regional managed instance group of workers
resource "google_compute_region_instance_group_manager" "workers" {
name = "${var.name}-worker-group"
description = "Compute instance group of ${var.name} workers"
Expand All @@ -11,30 +10,18 @@ resource "google_compute_region_instance_group_manager" "workers" {

target_size = "${var.count}"

# target pool to which instances in the group should be added
target_pools = [
"${google_compute_target_pool.workers.self_link}",
]
}

# Worker Container Linux Config
data "template_file" "worker_config" {
template = "${file("${path.module}/cl/worker.yaml.tmpl")}"

vars = {
kubeconfig = "${indent(10, var.kubeconfig)}"
ssh_authorized_key = "${var.ssh_authorized_key}"
k8s_dns_service_ip = "${cidrhost(var.service_cidr, 10)}"
cluster_domain_suffix = "${var.cluster_domain_suffix}"
named_port {
name = "http"
port = "80"
}
}

data "ct_config" "worker_ign" {
content = "${data.template_file.worker_config.rendered}"
pretty_print = false
snippets = ["${var.clc_snippets}"]
named_port {
name = "https"
port = "443"
}
}

# Worker instance template
resource "google_compute_instance_template" "worker" {
name_prefix = "${var.name}-worker-"
description = "Worker Instance template"
Expand Down Expand Up @@ -76,3 +63,21 @@ resource "google_compute_instance_template" "worker" {
create_before_destroy = true
}
}

# Worker Container Linux Config
data "template_file" "worker_config" {
template = "${file("${path.module}/cl/worker.yaml.tmpl")}"

vars = {
kubeconfig = "${indent(10, var.kubeconfig)}"
ssh_authorized_key = "${var.ssh_authorized_key}"
k8s_dns_service_ip = "${cidrhost(var.service_cidr, 10)}"
cluster_domain_suffix = "${var.cluster_domain_suffix}"
}
}

data "ct_config" "worker_ign" {
content = "${data.template_file.worker_config.rendered}"
pretty_print = false
snippets = ["${var.clc_snippets}"]
}

0 comments on commit b517e83

Please sign in to comment.