-
Notifications
You must be signed in to change notification settings - Fork 17
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Pepr Watch is not Responding to Changes after 90 mins #745
Comments
here is a temporary workaround that should force the watcher pod to reconnect. https://gist.github.com/cmwylie19/2c07e0e6f0962b8f18999488d1646a4a |
something with the AWS VPC CNI could have issues and is sometimes dropping policy. It could have something to do with it. Related Issue - aws/amazon-vpc-cni-k8s#2103 |
Fixed by #766. If you experience this, use the environment variable |
Some seemingly related issues:
Possible solution/workaround in istio/istio#8696 (comment) by creating a |
This is not technically resolved and setting a very low resync threshold like this is a really bad practice in production. Agree we either need to look at dropping Istio for this or looking at |
This should be solved based on c0d3aaa and the KFC informer pattern. All soak tests have passed |
Ran into this the other day using uds-core 0.26.1 (pepr v0.36.0) in our internal environments with |
Just ran through a UDS Core and bundling demo with Raytheon and ran into the problem where I had to restart the pepr-uds-core-watcher pod in order to get a successful deployment of mattermost, postgres operator, minio, etc. It looks like pepr blocks some of the secrets from being created, which then prevents mattermost from coming up. Killing the pepr watcher pod and then re-doing the deployment fixes the problem. This can happen regardless of whether the cluster/platform has been up for hours or just came up a couple minutes ago. Does not happen every time though. UDS Core 0.27.2 - k3d-core-demo:0.27.2 (also existed on 0.26.1) |
Just got it today, happened while deploying core into a cluster, by the time I got to Neuvector, Pepr was AWOL. Had to kill the watcher, and bounce all affected Neuvector pods. Core version 0.27.2. |
fixed in udici defenseunicorns/kubernetes-fluent-client#399, we have not heard any more complaints. Feel free to open it back up if it occurs again. |
Environment
Device and OS: UDS Core
App version:
Watch controller started responding after rolling. Only pod logs were health checks
Kubernetes distro being used:
Other: EKS Kubernetes 1.29
Steps to reproduce
Expected result
Upon disconnection, it reconnects
Actual Result
Visual Proof (screenshots, videos, text, etc)
Severity/Priority
Additional Context
Add any other context or screenshots about the technical debt here.
Well Known Bug:
kubernetes-client/csharp#533
kubernetes-client/javascript#596
The text was updated successfully, but these errors were encountered: