-
Notifications
You must be signed in to change notification settings - Fork 69
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
<Arktos Mizar Integration>To able to add worker nodes into RP clusters in scale-out 1 X 1 enviroment #1230
<Arktos Mizar Integration>To able to add worker nodes into RP clusters in scale-out 1 X 1 enviroment #1230
Conversation
[APPROVALNOTIFIER] This PR is NOT APPROVED This pull-request has been approved by: The full list of commands accepted by this bot can be found here.
Needs approval from an approver in each of these files:
Approvers can indicate their approval by writing |
Question for set up instruction:
|
Question for set up instruction:
We can use same config folder for master and worker if it does not create confusion between master and worker.
|
This PR can be closed because it is replaced by larger PR 1382. |
What type of PR is this?
/kind documentation
/kind feature
What this PR does / why we need it:
When working on the project of Arktos Mizar Integration, Mizar team requests Arktos team to provide scale-out environment and the worker nodes can be added into RP cluster.
This PR enables to add the worker nodes into RP cluster in scale-out 1 TP X 1 RP environment and has been tested in the following 2 sets of scale out 1 TP X 1 RP (1 master node +1 worker node) environment, which is prerequisite of scale-out 2 TPs X 2 RPs environment. All nodes are AWS EC2 instance - t2.2xlarge running Ubuntu 18.04.
TP1: 172.31.3.192
RP1: 172.31.5.191
Worker node-1: 172.31.4.110
Worker node-2: 172.31.29.26
TP2: 172.31.5.56
RP1: 172.31.13.237
Worker node: 172.31.2.149
The script ./hack/arktos-worker-up.sh running on worker node enables kubelet and kube-proxy as well as the flannel is installed in process mode on worker node so that joining into RP cluster can be successful.
Also, the codes of this PR also works in scale-up environment ( master node + n x worker nodes) and have been tested in scale-up 1 + 2 environment below.
Master node : 172.31.22.85
Worker node-1: 172.31.29.128
Worker node-2: 172.31.24.185
Worker node-3: 172.21.5.205
Which issue(s) this PR fixes:
N/A
Special notes for your reviewer:
In scale-out 1 TP X 1 RP cluster, currently when 2nd worker node attempts to join RP cluster, in its flannel log /tmp/flanneld.log, you will see the following error:
E1203 06:00:52.242285 1298 route_network.go:115] Error adding route to 10.244.0.0/24 via 172.31.5.191 dev index 2: network is unreachable
I1203 06:00:52.242309 1298 route_network.go:86] Subnet added: 10.244.1.0/24 via 172.31.4.110
E1203 06:00:52.242497 1298 route_network.go:115] Error adding route to 10.244.1.0/24 via 172.31.4.110 dev index 2: network is unreachable
In scale-up cluster, currently when 3rd worker node attempts to join cluster, in its flannel log /tmp/flanneld.log, you will see the following error:
E1203 04:15:33.103839 18746 route_network.go:115] Error adding route to 10.244.0.0/24 via 172.31.22.85 dev index 2: network is unreachable
I1203 04:15:33.103858 18746 route_network.go:86] Subnet added: 10.244.1.0/24 via 172.31.29.128
E1203 04:15:33.103956 18746 route_network.go:115] Error adding route to 10.244.1.0/24 via 172.31.29.128 dev index 2: network is unreachable
I1203 04:15:33.103975 18746 route_network.go:86] Subnet added: 10.244.2.0/24 via 172.31.24.185
E1203 04:15:33.104110 18746 route_network.go:115] Error adding route to 10.244.2.0/24 via 172.31.24.185 dev index 2: network is unreachable
Need further investigate some limits of network on AWS EC2 instance type - t2.2xlarge.
Does this PR introduce a user-facing change?:
YES.