You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Originally posted by anirbansingharoy September 17, 2023
In our configuration, we have deployed vector in an AKS cluster as Daemonset. The agent collect logs using kubernetes source connect. There is transformation we do using remap. After that its piped to sink. In Sink, we used the kafka connector which published the logs to Azure Eventhub.
The set up was working perfectly until we got a big surge in amount of logs being generated. We found out during high load vector was shipping all events to Azure Eventhub which was throttling the requests.This was causing the vector pods to get into a state where it was getting OOM Killed.
We then read about buffering model and thought that it could be used to slow down the source when Azure Eventhub is unable to catchup the the pace of logs.
Initially we started with memory based buffering with a maximum events size of 2000. We expected that the behaviour will be when 2000 events are in the buffer, the source will be slowed down. However that didn't happen. It was still hammering azure eventhub with many requests and azure eventhub is rejecting those requests due to throttling.
The pod was continuously OOM Killed as well during this time.
Later we switch to disk based buffering to see if that helps. However it was still hammering the azure eventhub with many requests. The only way we were able to handle the situation at the moment is increasing the azure eventhub capacity.
So we would like to know, if we are doing any mistake in terms of buffer configuration due to which the source is not slowing down when sink is unable to catch up the with pace. Any other suggestion is also appreciated.
The text was updated successfully, but these errors were encountered:
Discussed in #18578
Originally posted by anirbansingharoy September 17, 2023
In our configuration, we have deployed vector in an AKS cluster as Daemonset. The agent collect logs using
kubernetes
source connect. There is transformation we do using remap. After that its piped to sink. In Sink, we used thekafka
connector which published the logs to Azure Eventhub.The set up was working perfectly until we got a big surge in amount of logs being generated. We found out during high load vector was shipping all events to Azure Eventhub which was throttling the requests.This was causing the vector pods to get into a state where it was getting OOM Killed.
We then read about buffering model and thought that it could be used to slow down the source when Azure Eventhub is unable to catchup the the pace of logs.
Initially we started with
memory
based buffering with a maximum events size of 2000. We expected that the behaviour will be when 2000 events are in the buffer, the source will be slowed down. However that didn't happen. It was still hammering azure eventhub with many requests and azure eventhub is rejecting those requests due to throttling.The configuration is as follows
The pod was continuously OOM Killed as well during this time.
Later we switch to disk based buffering to see if that helps. However it was still hammering the azure eventhub with many requests. The only way we were able to handle the situation at the moment is increasing the azure eventhub capacity.
So we would like to know, if we are doing any mistake in terms of buffer configuration due to which the source is not slowing down when sink is unable to catch up the with pace. Any other suggestion is also appreciated.
The text was updated successfully, but these errors were encountered: