From 3f31a67f07ae19326c08cac4e0b40fd4f91e5f30 Mon Sep 17 00:00:00 2001 From: Alessandro Bellina Date: Mon, 6 Dec 2021 13:48:25 -0600 Subject: [PATCH] Revert change in integration tests --- integration_tests/run_pyspark_from_build.sh | 3 +-- integration_tests/src/main/python/spark_session.py | 2 +- 2 files changed, 2 insertions(+), 3 deletions(-) diff --git a/integration_tests/run_pyspark_from_build.sh b/integration_tests/run_pyspark_from_build.sh index b4dd19831d9..4deb9d1b959 100755 --- a/integration_tests/run_pyspark_from_build.sh +++ b/integration_tests/run_pyspark_from_build.sh @@ -165,14 +165,13 @@ else export PYSP_TEST_spark_driver_extraClassPath="${ALL_JARS// /:}" export PYSP_TEST_spark_executor_extraClassPath="${ALL_JARS// /:}" - export PYSP_TEST_spark_driver_extraJavaOptions="-ea -Duser.timezone=UTC $COVERAGE_SUBMIT_FLAGS -Xrunjdwp:transport=dt_socket,server=y,suspend=n,address=5010" + export PYSP_TEST_spark_driver_extraJavaOptions="-ea -Duser.timezone=UTC $COVERAGE_SUBMIT_FLAGS" export PYSP_TEST_spark_executor_extraJavaOptions='-ea -Duser.timezone=UTC' export PYSP_TEST_spark_ui_showConsoleProgress='false' export PYSP_TEST_spark_sql_session_timeZone='UTC' export PYSP_TEST_spark_sql_shuffle_partitions='12' # prevent cluster shape to change export PYSP_TEST_spark_dynamicAllocation_enabled='false' - export PYSP_TEST_spark_rapids_sql_explain='ALL' # Extract Databricks version from deployed configs. This is set automatically on Databricks # notebooks but not when running Spark manually. diff --git a/integration_tests/src/main/python/spark_session.py b/integration_tests/src/main/python/spark_session.py index f4c9465ea5c..0a5e98b8139 100644 --- a/integration_tests/src/main/python/spark_session.py +++ b/integration_tests/src/main/python/spark_session.py @@ -96,7 +96,7 @@ def with_gpu_session(func, conf={}): if is_allowing_any_non_gpu(): copy['spark.rapids.sql.test.enabled'] = 'false' else: - copy['spark.rapids.sql.test.enabled'] = 'false' + copy['spark.rapids.sql.test.enabled'] = 'true' copy['spark.rapids.sql.test.allowedNonGpu'] = ','.join(get_non_gpu_allowed()) copy['spark.rapids.sql.test.validateExecsInGpuPlan'] = ','.join(get_validate_execs_in_gpu_plan())