Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Incorrect Metric Type for HPA Scaling #3286

Open
liaddrori1 opened this issue Aug 15, 2024 · 1 comment
Open

Incorrect Metric Type for HPA Scaling #3286

liaddrori1 opened this issue Aug 15, 2024 · 1 comment
Assignees
Labels
bug Something isn't working

Comments

@liaddrori1
Copy link

📚 The doc issue

In the kubernetes/autoscale.md file, the current implementation uses the ts_queue_latency_microseconds metric for scaling the Horizontal Pod Autoscaler (HPA). This metric is a counter, which only increases over time and does not decrease, leading to a potential issue where the HPA will continually scale up the number of pods without scaling them down when the load decreases.

Suggest a potential alternative/fix

To resolve this issue, it is recommended to use the rate of the counter metric over a time interval to enable both scaling up and down effectively.

  1. Use the Rate Function:

    • Utilize the rate function in Prometheus to calculate the rate of change of the ts_queue_latency_microseconds metric. This provides a per-second average rate of increase over a specified time window (e.g., 5 minutes).
  2. Modify the Prometheus Adapter Configuration:

    • Update the configuration to transform the counter metric into a rate-based metric. Here’s how the configuration should look:

      rules:
      - seriesQuery: 'ts_queue_latency_microseconds'
        resources:
          overrides:
            namespace:
              resource: namespace
            pod:
              resource: pod
        name:
          matches: "^(.*)_microseconds$"
          as: "${1}_per_second"
        metricsQuery: 'sum(rate(<<.Series>>{<<.LabelMatchers>>}[5m])) by (<<.GroupBy>>)'
  3. Modify the HPA Configuration:

    • Update the metrics section in the hpa.yaml file to use the rate of the metric:
      metrics:
        - type: Pods
          pods:
            metric:
              name: ts_queue_latency_per_second
            target:
              type: AverageValue
              averageValue: 1000000  # Set your desired threshold here
  4. Update Documentation:

    • Update the documentation in kubernetes/autoscale.md to reflect these changes and provide guidance on selecting appropriate target values based on the rate metric.

Why This Is Better

Using the rate of the counter metric allows the HPA to make scaling decisions based on the actual rate of change in queue latency rather than the cumulative value. This approach enables the HPA to scale pods up when the rate of incoming requests increases and scale down when the rate decreases, providing more responsive and efficient scaling behavior.

Example:

  • Current Configuration: If ts_queue_latency_microseconds is used directly, the HPA will see the metric as always increasing, causing continuous scaling up.
  • Proposed Configuration: By using sum(rate(ts_queue_latency_microseconds[5m])), the HPA can see the rate at which latency is increasing. For instance, if the rate increases to 7000 per second, the HPA will add pods. If the rate decreases to below the target value, it will scale down, allowing the system to adapt dynamically to load changes.

This improvement ensures better resource utilization and cost efficiency by aligning the number of pods with the actual workload.

@yardenhoch

@mreso
Copy link
Collaborator

mreso commented Aug 19, 2024

Thanks for flagging this @liaddrori1
@namannandan do you have bandwidth to look at this?

@mreso mreso added the bug Something isn't working label Aug 19, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working
Projects
None yet
Development

No branches or pull requests

3 participants