Skip to content

How will a partition data be processed when the data size of a partition exceeds the configuration size of spark.task.resource.gpu.amount? #5336

Answered by jlowe
YeahNew asked this question in General
Discussion options

You must be logged in to vote

Ultimately the number of tasks that can be allowed to run concurrently on the GPU is influenced by both configuration settings, and how many tasks will run concurrently at any point in time depends on what the tasks are trying to do at that moment. I'll try to explain in detail, so here's how those two configuration settings work in practice:

spark.task.resource.gpu.amount will direct Spark's scheduler to limit how many tasks are allowed to run concurrently on an executor, whether those tasks are actively using the GPU or not. For example, if spark.task.resource.gpu.amount=0.25 then the executor can run at most 4 tasks at a time, just like when spark.executor.cores=4. This is the maximum …

Replies: 7 comments

Comment options

You must be logged in to vote
0 replies
Comment options

You must be logged in to vote
0 replies
Comment options

You must be logged in to vote
0 replies
Answer selected by sameerz
Comment options

You must be logged in to vote
0 replies
Comment options

You must be logged in to vote
0 replies
Comment options

You must be logged in to vote
0 replies
Comment options

You must be logged in to vote
0 replies
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
documentation Improvements or additions to documentation question Further information is requested
3 participants
Converted from issue

This discussion was converted from issue #4617 on April 27, 2022 16:43.