Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Unable to enable Azure CAPIProvider #483

Open
mantis-toboggan-md opened this issue Apr 16, 2024 · 4 comments
Open

Unable to enable Azure CAPIProvider #483

mantis-toboggan-md opened this issue Apr 16, 2024 · 4 comments
Assignees
Labels
area/azure kind/bug Something isn't working
Milestone

Comments

@mantis-toboggan-md
Copy link
Member

What steps did you take and what happened?

The Azure CAPIProvider fails to enable correctly in the 0.4.0 capi ui. The status changes from Provisioning to Ready but eventually becomes Unavailable. The capz-controller-manager pod is in CrashLoopBackoff with the error

“failed to get informer from cache” err=“failed to get API group resources: unable to retrieve the complete list of server APIs: bootstrap.cluster.x-k8s.io/v1beta1: the server could not find the requested resource” logger=“controller-runtime.source.EventHandler”

What did you expect to happen?

I would expect the capz-controller-manager pod to be running and azure capiprovider resource to be in Ready state

How to reproduce it?

No response

Rancher Turtles version

No response

Anything else you would like to add?

No response

Label(s) to be applied

/kind bug

@kkaempf kkaempf added kind/bug Something isn't working area/ui area/azure and removed area/ui labels Apr 17, 2024
@richardcase richardcase added this to the May release milestone May 3, 2024
@salasberryfin salasberryfin self-assigned this May 8, 2024
@salasberryfin
Copy link
Contributor

salasberryfin commented May 9, 2024

Hi @mantis-toboggan-md, thanks for reporting this. I was able to reproduce the issue with the following configuration:

  • Rancher v2.8.2
  • Rancher Turtles v0.7.0
  • Rancher Turtles UI v0.4.0

Looks like this may be related to missing resources bootstrap.cluster.x-k8s.io. This custom resource is generally available via Kubeadm but, since Turtles is using RKE2 for boostrap and control plane provisioning, CAPRKE2 is providing this resource instead.

For some reason CAPZ is not detecting the api resource via RKE2 but, if installing Kubeadm and the re-trying CAPZ installation, it applies the changes successfully.

I did apply this yaml file before installing CAPZ:

---
apiVersion: v1
kind: Namespace
metadata:
  name: capi-kubeadm-bootstrap-system
---
apiVersion: turtles-capi.cattle.io/v1alpha1
kind: CAPIProvider
metadata:
  name: kubeadm-bootstrap
  namespace: capi-kubeadm-bootstrap-system
spec:
  name: kubeadm
  type: bootstrap
  version: v1.4.6
  configSecret:
    name: variables
---
apiVersion: v1
kind: Namespace
metadata:
  name: capi-kubeadm-control-plane-system
---
apiVersion: turtles-capi.cattle.io/v1alpha1
kind: CAPIProvider
metadata:
  name: kubeadm-control-plane
  namespace: capi-kubeadm-control-plane-system
spec:
  name: kubeadm
  type: controlPlane
  version: v1.4.6
  configSecret:
    name: variables

And then the Azure provider was successfully installed via Rancher UI and the controller did not report any errors.

The custom resource that the logs report as missing should be available via the RKE2 provider so we need to investigate this a bit further to propose a solution.

@furkatgofurov7
Copy link
Contributor

furkatgofurov7 commented May 9, 2024

Could it be because there is no bootstrap.cluster.x-k8s.io/v1beta1 available in CAPRKE2, but only v1alpha1

@salasberryfin
Copy link
Contributor

Opened a new upstream issue kubernetes-sigs/cluster-api-provider-azure#4854 to track the fix on CAPZ. Once the community accepts this proposal, we'll submit the PR effectively removing the dependency on Kubeadm when enabling MachinePools.

@salasberryfin
Copy link
Contributor

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
area/azure kind/bug Something isn't working
Development

No branches or pull requests

6 participants