Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Logging Verbosity setting for kube* components #2165

Closed
ohadschn opened this issue Mar 7, 2021 · 12 comments
Closed

Logging Verbosity setting for kube* components #2165

ohadschn opened this issue Mar 7, 2021 · 12 comments
Assignees
Labels
feature-request Requested Features

Comments

@ohadschn
Copy link

ohadschn commented Mar 7, 2021

The following logs are available for cluster debugging:

Master

  • /var/log/kube-apiserver.log - API Server, responsible for serving the API
  • /var/log/kube-scheduler.log - Scheduler, responsible for making scheduling decisions
  • /var/log/kube-controller-manager.log - Controller that manages replication controllers

Worker Nodes

  • /var/log/kubelet.log - Kubelet, responsible for running containers on the node
  • /var/log/kube-proxy.log - Kube Proxy, responsible for service load balancing

It is my understanding that they all support a -v verbosity flag (klog convention), which can be immensely useful in test clusters. Unfortunately,

  • Master nodes cannot be interacted with, so the verbosity of the components they host is impossible to change.
  • Worker-level system pods can be hacked to enable this (Kured, DaemonSet) but there is no built-in, proper solution.

There is also the related matter of container logs settings (number of log files, max log file size, etc): #1062

@ghost ghost added the triage label Mar 7, 2021
@ghost
Copy link

ghost commented Mar 7, 2021

Hi ohadschn, AKS bot here 👋
Thank you for posting on the AKS Repo, I'll do my best to get a kind human from the AKS team to assist you.

I might be just a bot, but I'm told my suggestions are normally quite good, as such:

  1. If this case is urgent, please open a Support Request so that our 24/7 support team may help you faster.
  2. Please abide by the AKS repo Guidelines and Code of Conduct.
  3. If you're having an issue, could it be described on the AKS Troubleshooting guides or AKS Diagnostics?
  4. Make sure your subscribed to the AKS Release Notes to keep up to date with all that's new on AKS.
  5. Make sure there isn't a duplicate of this issue already reported. If there is, feel free to close this one and '+1' the existing issue.
  6. If you have a question, do take a look at our AKS FAQ. We place the most common ones there!

@ghost
Copy link

ghost commented Mar 10, 2021

Triage required from @Azure/aks-pm

@ghost
Copy link

ghost commented Mar 15, 2021

Action required from @Azure/aks-pm

@ghost ghost added the Needs Attention 👋 Issues needs attention/assignee/owner label Mar 15, 2021
@ghost
Copy link

ghost commented Mar 30, 2021

Issue needing attention of @Azure/aks-leads

@yuriwoof
Copy link

Same here.
It is necessary to get the more detailed log....

$ kubectl describe pvc gluster-pvc
Name:          gluster-pvc
Namespace:     default
StorageClass:  gluster-heketi-external
Status:        Pending
Volume:        
Labels:        <none>
Annotations:   volume.beta.kubernetes.io/storage-class: gluster-heketi-external
               volume.beta.kubernetes.io/storage-provisioner: kubernetes.io/glusterfs
Finalizers:    [kubernetes.io/pvc-protection]
Capacity:      
Access Modes:  
VolumeMode:    Filesystem
Used By:       <none>
Events:
  Type     Reason              Age                 From                         Message
  ----     ------              ----                ----                         -------
  Warning  ProvisioningFailed  51s (x205 over 9h)  persistentvolume-controller  Failed to provision volume with StorageClass "gluster-heketi-external": failed to create volume: failed to create volume: see kube-controller-manager.log for details

@palma21 palma21 added the feature-request Requested Features label Aug 30, 2021
@ghost ghost removed triage action-required Needs Attention 👋 Issues needs attention/assignee/owner labels Aug 30, 2021
@Azure Azure deleted a comment Aug 30, 2021
@Azure Azure deleted a comment Aug 30, 2021
@Azure Azure deleted a comment Aug 30, 2021
@Azure Azure deleted a comment Aug 30, 2021
@Azure Azure deleted a comment Aug 30, 2021
@Azure Azure deleted a comment Aug 30, 2021
@Azure Azure deleted a comment Aug 30, 2021
@Azure Azure deleted a comment Aug 30, 2021
@Azure Azure deleted a comment Aug 30, 2021
@pawelpabich
Copy link

We just hit a case that is pretty much impossible without access to detailed kube-scheduler logs. Having read only access to all logs would be amazing.

@pawelpabich
Copy link

@Kaarthis-zz
Copy link
Contributor

@yuri

Same here.
It is necessary to get the more detailed log....

$ kubectl describe pvc gluster-pvc
Name:          gluster-pvc
Namespace:     default
StorageClass:  gluster-heketi-external
Status:        Pending
Volume:        
Labels:        <none>
Annotations:   volume.beta.kubernetes.io/storage-class: gluster-heketi-external
               volume.beta.kubernetes.io/storage-provisioner: kubernetes.io/glusterfs
Finalizers:    [kubernetes.io/pvc-protection]
Capacity:      
Access Modes:  
VolumeMode:    Filesystem
Used By:       <none>
Events:
  Type     Reason              Age                 From                         Message
  ----     ------              ----                ----                         -------
  Warning  ProvisioningFailed  51s (x205 over 9h)  persistentvolume-controller  Failed to provision volume with StorageClass "gluster-heketi-external": failed to create volume: failed to create volume: see kube-controller-manager.log for details

Could you please elaborate what led you to this issue - 'ProvisioningFailed' , please describe the problem use case here and can you also let us know which log details are you most interested in - Any Support cases opened on this prior would be helpful for us .

@yuriwoof
Copy link

yuriwoof commented Sep 30, 2021

Hi @kaarthis you may mention another Yuri...
Anyway, in my scenario, I set up GluserFS cluster and create Stroage Class & PVC.
But I couldn't investigate the reason why ProvisioningFailed occur (it's limited in log level)

Any Support cases opened on this prior would be helpful for us .

Let me loop in the MS internal email thread.

@Kaarthis-zz
Copy link
Contributor

Got it @yuriwoof - which logs did you access for this and may i know if through the support case having increased verbosity helped you at all ,if so how ?

@Kaarthis-zz
Copy link
Contributor

@yuriwoof any luck with the MS Internal thread ? also happy to sync with you if you are available to discuss this.

@Kaarthis-zz Kaarthis-zz self-assigned this Oct 8, 2021
@Kaarthis-zz
Copy link
Contributor

We are evaluating this for long term impact especially balancing component CCP stability vs benefits of increased logging. Closing this for now as we have communicated with Octopus group in the thread.

@ghost ghost locked as resolved and limited conversation to collaborators Nov 26, 2021
@kaarthis kaarthis assigned kaarthis and unassigned Kaarthis-zz Jan 12, 2022
Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
feature-request Requested Features
Projects
None yet
Development

No branches or pull requests

6 participants