-
Notifications
You must be signed in to change notification settings - Fork 0
/
locals.tf
122 lines (118 loc) · 5.74 KB
/
locals.tf
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
data "aws_ami" "al2gpu_ami" {
owners = ["amazon"]
most_recent = true
filter {
name = "name"
values = ["*amazon-eks-gpu-node-${var.kubernetesVersion}*"]
}
}
locals {
create_vpc = var.vpcId == null ? true : false
vpc_id = local.create_vpc ? module.vpc[0].vpc_id : var.vpcId
use_private_subnets_ids = length(var.private_subnet_ids) == 0 ? false : true
use_public_subnet_ids = length(var.public_subnet_ids) == 0 ? false : true
infrastructurename = var.infrastructurename
log_group_name = "/${module.eks.eks_cluster_id}/worker-fluentbit-logs"
account_id = data.aws_caller_identity.current.account_id
region = data.aws_region.current.name
license_server_role = "${local.infrastructurename}-license-server-role"
license_server_policy = "${local.infrastructurename}-license-server-policy"
license_server_bucket_name = "${local.infrastructurename}-license-server-bucket"
license_server = "${local.infrastructurename}-license-server"
license_server_instance_profile = "${local.infrastructurename}-license-server-instance-profile"
flowlogs_cloudwatch_loggroup = "/aws/vpc/${module.eks.eks_cluster_id}"
patch_manager_cloudwatch_loggroup_scan = "/aws/ssm/${module.eks.eks_cluster_id}/scan"
patch_manager_cloudwatch_loggroup_install = "/aws/ssm/${module.eks.eks_cluster_id}/install"
patchgroupid = "${var.infrastructurename}-patch-group"
s3_instance_buckets = flatten([for name, instance in module.simphera_instance : instance.s3_buckets])
license_server_bucket = var.licenseServer ? [aws_s3_bucket.license_server_bucket[0].bucket] : []
s3_buckets = concat(local.s3_instance_buckets, [aws_s3_bucket.bucket_logs.bucket], local.license_server_bucket)
private_subnets = local.create_vpc ? module.vpc[0].private_subnets : (local.use_private_subnets_ids ? var.private_subnet_ids : [for s in data.aws_subnet.private_subnet : s.id])
public_subnets = local.create_vpc ? module.vpc[0].public_subnets : (local.use_public_subnet_ids ? var.public_subnet_ids : [for s in data.aws_subnet.public_subnet : s.id])
# Using a one-line command for gpuPostUserData to avoid issues due to different line endings between Windows and Linux.
gpuPostUserData = "sudo yum -y erase nvidia-driver \nsudo yum -y install make gcc \nsudo yum -y update \nsudo yum -y install gcc kernel-devel-$(uname -r) \nsudo curl -fSsl -O https://us.download.nvidia.com/tesla/${var.gpuNvidiaDriverVersion}/NVIDIA-Linux-x86_64-${var.gpuNvidiaDriverVersion}.run \nsudo chmod +x NVIDIA-Linux-x86_64*.run \nsudo CC=/usr/bin/gcc10-cc ./NVIDIA-Linux-x86_64*.run -s --no-dkms --install-libglvnd \nsudo touch /etc/modprobe.d/nvidia.conf \necho \"options nvidia NVreg_EnableGpuFirmware=0\" | sudo tee --append /etc/modprobe.d/nvidia.conf \nsudo reboot"
default_managed_node_pools = {
"default" = {
node_group_name = "default"
instance_types = var.linuxNodeSize
subnet_ids = local.private_subnets
desired_size = var.linuxNodeCountMin
max_size = var.linuxNodeCountMax
min_size = var.linuxNodeCountMin
disk_size = var.linuxNodeDiskSize
},
"execnodes" = {
node_group_name = "execnodes"
instance_types = var.linuxExecutionNodeSize
subnet_ids = local.private_subnets
desired_size = var.linuxExecutionNodeCountMin
max_size = var.linuxExecutionNodeCountMax
min_size = var.linuxExecutionNodeCountMin
disk_size = var.linuxExecutionNodeDiskSize
k8s_labels = {
"purpose" = "execution"
}
k8s_taints = [
{
key = "purpose",
value = "execution",
"effect" = "NO_SCHEDULE"
}
]
}
}
gpu_node_pool = {
"gpuexecnodes" = {
node_group_name = "gpuexecnodes"
instance_types = var.gpuNodeSize
subnet_ids = local.private_subnets
desired_size = var.gpuNodeCountMin
max_size = var.gpuNodeCountMax
min_size = var.gpuNodeCountMin
disk_size = var.gpuNodeDiskSize
custom_ami_id = data.aws_ami.al2gpu_ami.image_id
create_launch_template = true
post_userdata = local.gpuPostUserData
k8s_labels = {
"purpose" = "gpu"
}
k8s_taints = [
{
key = "purpose",
value = "gpu",
"effect" = "NO_SCHEDULE"
}
]
}
}
ivsgpu_node_pool = {
"gpuivsnodes" = {
node_group_name = "gpuivsnodes"
instance_types = var.ivsGpuNodeSize
subnet_ids = local.private_subnets
desired_size = var.ivsGpuNodeCountMin
max_size = var.ivsGpuNodeCountMax
min_size = var.ivsGpuNodeCountMin
disk_size = var.ivsGpuNodeDiskSize
custom_ami_id = data.aws_ami.al2gpu_ami.image_id
create_launch_template = true
post_userdata = local.gpuPostUserData
k8s_labels = {
"product" = "ivs",
"purpose" = "gpu"
}
k8s_taints = [
{
key = "purpose",
value = "gpu",
"effect" = "NO_SCHEDULE"
},
{
key = "nvidia.com/gpu",
value = "",
"effect" = "NO_SCHEDULE"
}
]
}
}
}