diff --git a/common/utils/src/main/resources/error/error-classes.json b/common/utils/src/main/resources/error/error-classes.json index 06c47419fcb07..7012c66c8956e 100644 --- a/common/utils/src/main/resources/error/error-classes.json +++ b/common/utils/src/main/resources/error/error-classes.json @@ -409,6 +409,11 @@ "message" : [ "Error instantiating Spark Connect plugin: " ] + }, + "SESSION_NOT_SAME" : { + "message" : [ + "Both Datasets must belong to the same SparkSession." + ] } } }, diff --git a/connector/connect/client/jvm/src/main/scala/org/apache/spark/sql/Dataset.scala b/connector/connect/client/jvm/src/main/scala/org/apache/spark/sql/Dataset.scala index b0dd91293a0be..0f7b376955c96 100644 --- a/connector/connect/client/jvm/src/main/scala/org/apache/spark/sql/Dataset.scala +++ b/connector/connect/client/jvm/src/main/scala/org/apache/spark/sql/Dataset.scala @@ -1750,7 +1750,10 @@ class Dataset[T] private[sql] ( private def checkSameSparkSession(other: Dataset[_]): Unit = { if (this.sparkSession.sessionId != other.sparkSession.sessionId) { - throw new SparkException("Both Datasets must belong to the same SparkSession") + throw new SparkException( + errorClass = "CONNECT.SESSION_NOT_SAME", + messageParameters = Map.empty, + cause = null) } }