Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[libvirt] Add default router for worker #2614

Closed
wants to merge 1 commit into from

Conversation

jichenjc
Copy link
Contributor

@jichenjc jichenjc commented Nov 2, 2019

as console pod can't parse DNS something like
oauth-openshift.apps.test1.aa.testing so it failed to start
this PR added the definition directly into libvirt dnsmasq file

Related to #1007

so the virsh dumpxml will have following output

 <host ip='192.168.126.51'>
      <hostname>oauth-openshift.apps.test1.aa.testing</hostname>
      <hostname>console-openshift-console.apps.test1.aa.testing</hostname>
      <hostname>downloads-openshift-console.apps.test1.aa.testing</hostname>
      <hostname>alertmanager-main-openshift-monitoring.apps.test1.aa.testing</hostname>
      <hostname>grafana-openshift-monitoring.apps.test1.aa.testing</hostname>
      <hostname>prometheus-k8s-openshift-monitoring.apps.test1.aa.testing</hostname>
 </host>

The oc get co shows the console op is running after cluster startup

[core@test1-6dpnj-master-0 ~]$ oc get co
NAME                                       VERSION                         AVAILABLE   PROGRESSING   DEGRADED   SINCE
authentication                             4.3.0-0.okd-2019-10-29-180250   True        False         False      3m4s
cloud-credential                           4.3.0-0.okd-2019-10-29-180250   True        False         False      17m
cluster-autoscaler                         4.3.0-0.okd-2019-10-29-180250   True        False         False      8m48s
console                                    4.3.0-0.okd-2019-10-29-180250   True        False         False      2m18s
dns                                        4.3.0-0.okd-2019-10-29-180250   True        False         False      15m
image-registry                             4.3.0-0.okd-2019-10-29-180250   True        False         False      4m52s
ingress                                    4.3.0-0.okd-2019-10-29-180250   True        False         False      4m57s
insights                                   4.3.0-0.okd-2019-10-29-180250   True        False         False      16m
kube-apiserver                             4.3.0-0.okd-2019-10-29-180250   True        False         False      12m
kube-controller-manager                    4.3.0-0.okd-2019-10-29-180250   True        False         False      12m
kube-scheduler                             4.3.0-0.okd-2019-10-29-180250   True        False         False      13m
machine-api                                4.3.0-0.okd-2019-10-29-180250   True        False         False      15m
machine-config                             4.3.0-0.okd-2019-10-29-180250   True        False         False      14m
marketplace                                4.3.0-0.okd-2019-10-29-180250   True        False         False      9m14s
monitoring                                 4.3.0-0.okd-2019-10-29-180250   True        False         False      112s
network                                    4.3.0-0.okd-2019-10-29-180250   True        False         False      15m
node-tuning                                4.3.0-0.okd-2019-10-29-180250   True        False         False      9m17s
openshift-apiserver                        4.3.0-0.okd-2019-10-29-180250   True        False         False      7m3s
openshift-controller-manager               4.3.0-0.okd-2019-10-29-180250   True        False         False      7m2s
openshift-samples                          4.3.0-0.okd-2019-10-29-180250   True        False         False      5m44s
operator-lifecycle-manager                 4.3.0-0.okd-2019-10-29-180250   True        False         False      15m
operator-lifecycle-manager-catalog         4.3.0-0.okd-2019-10-29-180250   True        False         False      15m
operator-lifecycle-manager-packageserver   4.3.0-0.okd-2019-10-29-180250   True        False         False      10m
service-ca                                 4.3.0-0.okd-2019-10-29-180250   True        False         False      16m
service-catalog-apiserver                  4.3.0-0.okd-2019-10-29-180250   True        False         False      10m
service-catalog-controller-manager         4.3.0-0.okd-2019-10-29-180250   True        False         False      10m
storage                                    4.3.0-0.okd-2019-10-29-180250   True        False         False      10m

as console pod can't parse DNS something like
oauth-openshift.apps.test1.aa.testing so it failed to start
this PR added the definition directly into libvirt dnsmasq file

so the `virsh dumpxml` will have following output

 <host ip='192.168.126.51'>
      <hostname>oauth-openshift.apps.test1.aa.testing</hostname>
      <hostname>console-openshift-console.apps.test1.aa.testing</hostname>
      <hostname>downloads-openshift-console.apps.test1.aa.testing</hostname>
      <hostname>alertmanager-main-openshift-monitoring.apps.test1.aa.testing</hostname>
      <hostname>grafana-openshift-monitoring.apps.test1.aa.testing</hostname>
      <hostname>prometheus-k8s-openshift-monitoring.apps.test1.aa.testing</hostname>
 </host>

The `oc get co` shows the console op is running after cluster startup

[core@test1-6dpnj-master-0 ~]$ oc get co
NAME                                       VERSION                         AVAILABLE   PROGRESSING   DEGRADED   SINCE
authentication                             4.3.0-0.okd-2019-10-29-180250   True        False         False      3m4s
cloud-credential                           4.3.0-0.okd-2019-10-29-180250   True        False         False      17m
cluster-autoscaler                         4.3.0-0.okd-2019-10-29-180250   True        False         False      8m48s
console                                    4.3.0-0.okd-2019-10-29-180250   True        False         False      2m18s
@openshift-ci-robot openshift-ci-robot added the needs-ok-to-test Indicates a PR that requires an org member to verify it is safe to test. label Nov 2, 2019
@openshift-ci-robot
Copy link
Contributor

Hi @jichenjc. Thanks for your PR.

I'm waiting for a openshift member to verify that this patch is reasonable to test. If it is, they should reply with /ok-to-test on its own line. Until that is done, I will not automatically test new commits in this PR, but the usual testing commands by org members will still work. Regular contributors should join the org to skip this step.

Once the patch is verified, the new status will be reflected by the ok-to-test label.

I understand the commands that are listed here.

Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes/test-infra repository.

@openshift-ci-robot openshift-ci-robot added the size/M Denotes a PR that changes 30-99 lines, ignoring generated files. label Nov 2, 2019
@openshift-ci-robot
Copy link
Contributor

[APPROVALNOTIFIER] This PR is NOT APPROVED

This pull-request has been approved by: jichenjc
To complete the pull request process, please assign praveenkumar
You can assign the PR to them by writing /assign @praveenkumar in a comment when ready.

The full list of commands accepted by this bot can be found here.

Needs approval from an approver in each of these files:

Approvers can indicate their approval by writing /approve in a comment
Approvers can cancel approval by writing /approve cancel in a comment

@jichenjc
Copy link
Contributor Author

jichenjc commented Nov 4, 2019

@zeenix @jaypoulz @gyohuangxin
can you help to take a look? thanks

@oswee
Copy link

oswee commented Nov 4, 2019

Just comment. I had this issue, when i was trying to iPXE boot 3 Masters, 2 Infras and 3 Compute nodes on Libvirt via Ansible script. All at once. I have dedicated Bind server which handles *.apps.* but still had this issue.
When i disabled Compute node booting, everything went well because... some components was installed on Infra nodes.
Now when i try to boot Compute nodes, they do not join cluster automatically, but this is another story.
Spend on this HUGGGE amount of time. I wish there would be better documentation.

@jichenjc
Copy link
Contributor Author

jichenjc commented Nov 4, 2019

yes, my PR is for CI/CD for our platform (s390x IBM) ,as we are using libvirt as CI/CD infra so
we need make it automatically create a cluster with all op are running

I agree we need well document for this case.... some follow up PR will be added about this after additional talk with more experts here

@gyohuangxin
Copy link
Contributor

@zeenix @jaypoulz @gyohuangxin
can you help to take a look? thanks

Yes, the code works and make all apps run well automatically.

@zeenix @jaypoulz Could you pls have a look? Thanks.

@crawford
Copy link
Contributor

crawford commented Nov 5, 2019

This really doesn't belong in the installer as this couples more of the cluster internals to the installer (wrong direction). I'd would rather we did this as a one-off step in our CI as a stopgap.

cc @abhinavdahiya

/hold

@openshift-ci-robot openshift-ci-robot added the do-not-merge/hold Indicates that a PR should not merge because someone has issued a /hold command. label Nov 5, 2019
@jaypoulz
Copy link
Contributor

jaypoulz commented Nov 5, 2019

@gyohuangxin Thanks for your hard work to get this figured out. I will see if I can add a workaround in the CI job configuration to do this injection as part of the CI job, as per @crawford's guidance.

I think we'll draw more from #2600 in terms of implementation.

@abhinavdahiya
Copy link
Contributor

This really doesn't belong in the installer as this couples more of the cluster internals to the installer (wrong direction). I'd would rather we did this as a one-off step in our CI as a stopgap.

cc @abhinavdahiya

/hold

+1

this creates a coupling between installer and the routes created by operators, which are not restricted to this list.

@jichenjc
Copy link
Contributor Author

jichenjc commented Nov 6, 2019

thanks a lot for your comments :) I will close this PR

@jichenjc jichenjc closed this Nov 6, 2019
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
do-not-merge/hold Indicates that a PR should not merge because someone has issued a /hold command. needs-ok-to-test Indicates a PR that requires an org member to verify it is safe to test. size/M Denotes a PR that changes 30-99 lines, ignoring generated files.
Projects
None yet
Development

Successfully merging this pull request may close these issues.

7 participants