-
Notifications
You must be signed in to change notification settings - Fork 30
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
changes to support db 13.3+ #716
changes to support db 13.3+ #716
Conversation
build |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Looks good. Minor comments.
dim = len(cluster_centers[0]) | ||
# inject unsupported expr (slice) that is essentially a noop | ||
df_for_scoring = df_for_scoring.select( | ||
F.slice(feature_col, 1, dim).alias(feature_col), output_col |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Interesting. Any intuition as to why slice can get the hanging resolved?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Yes. It is an unsupported expression in spark-rapids so falls back to cpu and injects columnartorow and rowtocolumnar transformations which do some batching that doesn't (but should after patching) happen otherwise.
|
||
6. Monitor progress periodically in case of a possible hang, to avoid incurring cloud costs in such cases. | ||
2. Monitor progress periodically in case of a possible hang, to avoid incurring cloud costs in such cases. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Do we keep item number 5, 6 or replace with 1, 2?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Interesting. Didn't intend to change the numbers, will revert.
"spark.sql.execution.arrow.pyspark.enabled": "true", | ||
"spark.sql.files.maxPartitionBytes": "2000000000000", | ||
"spark.databricks.delta.optimizeWrite.enabled": "false", | ||
"spark.rapids.sql.concurrentGpuTasks": "2" |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
While we keep this here, do we need to remove it from gpu_cluster_spec.sh (the one without GPU ETL)?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Yes. Will delete there. Good catch.
Signed-off-by: Erik Ordentlich <eordentlich@gmail.com>
Signed-off-by: Erik Ordentlich <eordentlich@gmail.com>
Signed-off-by: Erik Ordentlich <eordentlich@gmail.com>
40b9e84
to
9c85bd1
Compare
build |
Also adds db version as option to db benchmark script and a gpu only option (i.e. no spark rapids plugin).
Plus some other misc. updates (e.g. to logistic reg. notebooks, spark-rapids versions).
Includes a temporary patch till NVIDIA/spark-rapids#10770 is fixed.