diff --git a/sql/core/src/main/scala/org/apache/spark/sql/SQLConf.scala b/sql/core/src/main/scala/org/apache/spark/sql/SQLConf.scala index 00a485d5f745e..5c858a3b6b648 100644 --- a/sql/core/src/main/scala/org/apache/spark/sql/SQLConf.scala +++ b/sql/core/src/main/scala/org/apache/spark/sql/SQLConf.scala @@ -67,7 +67,7 @@ private[spark] object SQLConf { val USE_SQL_SERIALIZER2 = "spark.sql.useSerializer2" - val USE_JSONRDD2 = "spark.sql.json.useJsonRDD2" + val USE_JACKSON_STREAMING_API = "spark.sql.json.useJacksonStreamingAPI" object Deprecated { val MAPRED_REDUCE_TASKS = "mapred.reduce.tasks" @@ -162,7 +162,7 @@ private[sql] class SQLConf extends Serializable { private[spark] def useSqlSerializer2: Boolean = getConf(USE_SQL_SERIALIZER2, "true").toBoolean - private[spark] def useJsonRDD2: Boolean = getConf(USE_JSONRDD2, "true").toBoolean + private[spark] def useJacksonStreamingAPI: Boolean = getConf(USE_JACKSON_STREAMING_API, "true").toBoolean /** * Upper bound on the sizes (in bytes) of the tables qualified for the auto conversion to diff --git a/sql/core/src/main/scala/org/apache/spark/sql/SQLContext.scala b/sql/core/src/main/scala/org/apache/spark/sql/SQLContext.scala index d99ee132f3699..dcb9226511c62 100644 --- a/sql/core/src/main/scala/org/apache/spark/sql/SQLContext.scala +++ b/sql/core/src/main/scala/org/apache/spark/sql/SQLContext.scala @@ -615,7 +615,7 @@ class SQLContext(@transient val sparkContext: SparkContext) */ @Experimental def jsonRDD(json: RDD[String], schema: StructType): DataFrame = { - if (conf.useJsonRDD2) { + if (conf.useJacksonStreamingAPI) { baseRelationToDataFrame(new JSONRelation(json, None, 1.0, Some(schema))(this)) } else { val columnNameOfCorruptJsonRecord = conf.columnNameOfCorruptRecord @@ -649,7 +649,7 @@ class SQLContext(@transient val sparkContext: SparkContext) */ @Experimental def jsonRDD(json: RDD[String], samplingRatio: Double): DataFrame = { - if (conf.useJsonRDD2) { + if (conf.useJacksonStreamingAPI) { baseRelationToDataFrame(new JSONRelation(json, None, samplingRatio, None)(this)) } else { val columnNameOfCorruptJsonRecord = conf.columnNameOfCorruptRecord diff --git a/sql/core/src/main/scala/org/apache/spark/sql/json/JSONRelation.scala b/sql/core/src/main/scala/org/apache/spark/sql/json/JSONRelation.scala index 355b295907628..b1d212ce20f51 100644 --- a/sql/core/src/main/scala/org/apache/spark/sql/json/JSONRelation.scala +++ b/sql/core/src/main/scala/org/apache/spark/sql/json/JSONRelation.scala @@ -125,12 +125,12 @@ private[sql] class JSONRelation( samplingRatio, userSpecifiedSchema)(sqlContext) - private val useJsonRDD2: Boolean = sqlContext.conf.useJsonRDD2 + private val useJacksonStreamingAPI: Boolean = sqlContext.conf.useJacksonStreamingAPI override val needConversion: Boolean = false override lazy val schema = userSpecifiedSchema.getOrElse { - if (useJsonRDD2) { + if (useJacksonStreamingAPI) { JsonRDD2.nullTypeToStringType( JsonRDD2.inferSchema( baseRDD, @@ -146,7 +146,7 @@ private[sql] class JSONRelation( } override def buildScan(): RDD[Row] = { - if (useJsonRDD2) { + if (useJacksonStreamingAPI) { JsonRDD2.jsonStringToRow( baseRDD, schema, @@ -160,7 +160,7 @@ private[sql] class JSONRelation( } override def buildScan(requiredColumns: Seq[Attribute], filters: Seq[Expression]): RDD[Row] = { - if (useJsonRDD2) { + if (useJacksonStreamingAPI) { JsonRDD2.jsonStringToRow( baseRDD, StructType.fromAttributes(requiredColumns),