-
Notifications
You must be signed in to change notification settings - Fork 244
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[BUG] udf_test udf_cudf_test failed require_minimum_pandas_version check in spark 320+ #4378
Comments
spark bumped up the minimum pandas version to 1.0.5 with apache/spark@3657703 That change went into 3.3.0, branch 3.2 does not have that change, so I'm curious why this started failing unless cudf changed their version but then I would expect it to fail other places. 3.2 shipped requiring pandas version 0.23.2 (https://github.com/apache/spark/blob/v3.2.0/python/pyspark/sql/pandas/utils.py#L23) Cudf seems to require: pandas>=1.0,<1.4.0dev0 which hasn't changed recently. From the jenkinsfile for this build it looks like we are using 11.0 and 11.2 cuda images, which I don't think are supported any more. I think we need to change to the 11.5. @pxLi @NvTimLiu @GaryShen2008 could you take a look ? |
I'll check this issue. |
Seems we're not importing the real pandas module when run cudf-udf tests, we have a directory in |
As Will update to |
Reason for the failure:
To fix:
|
close as #4419 merged |
Describe the bug
seems pandas lib in spark 320+ is incompatible w/ the one installed as cudf's dep
The text was updated successfully, but these errors were encountered: