From 2a802c6d3ce59210a04cf7c38d2f69ed7751128e Mon Sep 17 00:00:00 2001 From: madhukar Date: Thu, 23 Apr 2015 15:33:48 +0530 Subject: [PATCH] #5654 updated the doc according to comments --- sql/core/src/main/scala/org/apache/spark/sql/DataFrame.scala | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/sql/core/src/main/scala/org/apache/spark/sql/DataFrame.scala b/sql/core/src/main/scala/org/apache/spark/sql/DataFrame.scala index 951829ad2abb0..3c2fbcdef1e69 100644 --- a/sql/core/src/main/scala/org/apache/spark/sql/DataFrame.scala +++ b/sql/core/src/main/scala/org/apache/spark/sql/DataFrame.scala @@ -1086,8 +1086,8 @@ class DataFrame private[sql]( * an RDD out to a parquet file, and then register that file as a table. This "table" can then * be the target of an `insertInto`. * - * Also note that this doesn't create a hive table, but instead creates a Spark data source table. - * Here the metadata is persisted into Hive. But table will be not accessible from the hive. + * Also note that while this function can persist the table metadata into Hive's metastore, + * the table will NOT be accessible from Hive. * @group output */ @Experimental