Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[CH] Create external merge tree table failed after (#7170) #7437

Closed
baibaichen opened this issue Oct 8, 2024 · 0 comments · Fixed by #7438
Closed

[CH] Create external merge tree table failed after (#7170) #7437

baibaichen opened this issue Oct 8, 2024 · 0 comments · Fixed by #7438
Labels
bug Something isn't working triage

Comments

@baibaichen
Copy link
Contributor

baibaichen commented Oct 8, 2024

Backend

CH (ClickHouse)

Bug description

 10/08/2024 03:39:51 AM - INFO - CREATE EXTERNAL TABLE IF NOT EXISTS call_center
 (
     cc_call_center_sk INT,
     cc_call_center_id string,
     cc_rec_start_date DATE,
     cc_rec_end_date DATE,
     cc_closed_date_sk INT,
     cc_open_date_sk INT,
     cc_name string,
     cc_class string,
     cc_employees INT,
     cc_sq_ft INT,
     cc_hours string,
     cc_manager string,
     cc_mkt_id INT,
     cc_mkt_class string,
     cc_mkt_desc string,
     cc_market_manager string,
     cc_division INT,
     cc_division_name string,
     cc_company INT,
     cc_company_name string,
     cc_street_number string,
     cc_street_name string,
     cc_street_type string,
     cc_suite_number string,
     cc_city string,
     cc_county string,
     cc_state string,
     cc_zip string,
     cc_country string,
     cc_gmt_offset DECIMAL(5,2),
     cc_tax_percentage DECIMAL(5,2)
 )
 USING clickhouse
 TBLPROPERTIES (storage_policy='s3_main', delta.checkpointInterval=5)
 LOCATION 's3a://gluten-cicd/dataset/tpcds-sf100-mergetree-partition/call_center'
 10/08/2024 03:39:59 AM - ERROR - TExecuteStatementResp(status=TStatus(statusCode=3, infoMessages=['*org.apache.hive.service.cli.HiveSQLException:Error running query: [DELTA_CREATE_TABLE_WITH_DIFFERENT_PROPERTY] org.apache.spark.sql.delta.DeltaAnalysisException: [DELTA_CREATE_TABLE_WITH_DIFFERENT_PROPERTY] The specified properties do not match the existing properties at s3a://gluten-cicd/dataset/tpcds-sf100-mergetree-partition/call_center.

== Specified ==
delta.checkpointInterval=5
engine=mergetree
external=true
is_distribute=true
location=s3a://gluten-cicd/dataset/tpcds-sf100-mergetree-partition/call_center
metadata_path=_delta_log
owner=root
provider=clickhouse
sampling_key=
storage_db=tpcds
storage_policy=s3_main
storage_table=call_center

== Existing ==
delta.checkpointInterval=5
engine=mergetree
external=true
is_distribute=true
location=s3a://gluten-cicd/dataset/tpcds-sf100-mergetree-partition/call_center
metadata_path=_delta_log
owner=root
provider=clickhouse
sampling_key=
storage_policy=s3_main
:37:36', 'org.apache.spark.sql.hive.thriftserver.HiveThriftServerErrors$:runningQueryError:HiveThriftServerErrors.scala:43', 'org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation:org$apache$spark$sql$hive$thriftserver$SparkExecuteStatementOperation$$execute:SparkExecuteStatementOperation.scala:262', 'org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation:runInternal:SparkExecuteStatementOperation.scala:152', 'org.apache.hive.service.cli.operation.Operation:run:Operation.java:277', 'org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation:org$apache$spark$sql$hive$thriftserver$SparkOperation$$super$run:SparkExecuteStatementOperation.scala:41', 'org.apache.spark.sql.hive.thriftserver.SparkOperation:$anonfun$run$1:SparkOperation.scala:45', 'scala.runtime.java8.JFunction0$mcV$sp:apply:JFunction0$mcV$sp.java:23', 'org.apache.spark.sql.hive.thriftserver.SparkOperation:withLocalProperties:SparkOperation.scala:79', 'org.apache.spark.sql.hive.thriftserver.SparkOperation:withLocalProperties$:SparkOperation.scala:63', 'org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation:withLocalProperties:SparkExecuteStatementOperation.scala:41', 'org.apache.spark.sql.hive.thriftserver.SparkOperation:run:SparkOperation.scala:45', 'org.apache.spark.sql.hive.thriftserver.SparkOperation:run$:SparkOperation.scala:43', 'org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation:run:SparkExecuteStatementOperation.scala:41', 'org.apache.hive.service.cli.session.HiveSessionImpl:executeStatementInternal:HiveSessionImpl.java:484', 'org.apache.hive.service.cli.session.HiveSessionImpl:executeStatement:HiveSessionImpl.java:460', 'sun.reflect.NativeMethodAccessorImpl:invoke0:NativeMethodAccessorImpl.java:-2', 'sun.reflect.NativeMethodAccessorImpl:invoke:NativeMethodAccessorImpl.java:62', 'sun.reflect.DelegatingMethodAccessorImpl:invoke:DelegatingMethodAccessorImpl.java:43', 'java.lang.reflect.Method:invoke:Method.java:498', 'org.apache.hive.service.cli.session.HiveSessionProxy:invoke:HiveSessionProxy.java:71', 'org.apache.hive.service.cli.session.HiveSessionProxy:lambda$invoke$0:HiveSessionProxy.java:58', 'java.security.AccessController:doPrivileged:AccessController.java:-2', 'javax.security.auth.Subject:doAs:Subject.java:422', 'org.apache.hadoop.security.UserGroupInformation:doAs:UserGroupInformation.java:1878', 'org.apache.hive.service.cli.session.HiveSessionProxy:invoke:HiveSessionProxy.java:58', 'com.sun.proxy.$Proxy36:executeStatement::-1', 'org.apache.hive.service.cli.CLIService:executeStatement:CLIService.java:282', 'org.apache.hive.service.cli.thrift.ThriftCLIService:ExecuteStatement:ThriftCLIService.java:453', 'org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement:getResult:TCLIService.java:1557', 'org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement:getResult:TCLIService.java:1542', 'org.apache.thrift.ProcessFunction:process:ProcessFunction.java:38', 'org.apache.thrift.TBaseProcessor:process:TBaseProcessor.java:39', 'org.apache.hive.service.auth.TSetIpAddressProcessor:process:TSetIpAddressProcessor.java:52', 'org.apache.thrift.server.TThreadPoolServer$WorkerProcess:run:TThreadPoolServer.java:310', 'java.util.concurrent.ThreadPoolExecutor:runWorker:ThreadPoolExecutor.java:1149', 'java.util.concurrent.ThreadPoolExecutor$Worker:run:ThreadPoolExecutor.java:624', 'java.lang.Thread:run:Thread.java:750', '*org.apache.spark.sql.delta.DeltaAnalysisException:[DELTA_CREATE_TABLE_WITH_DIFFERENT_PROPERTY] The specified properties do not match the existing properties at s3a://gluten-cicd/dataset/tpcds-sf100-mergetree-partition/call_center.

== Specified ==
delta.checkpointInterval=5
engine=mergetree
external=true
is_distribute=true
location=s3a://gluten-cicd/dataset/tpcds-sf100-mergetree-partition/call_center
metadata_path=_delta_log
owner=root
provider=clickhouse
sampling_key=
storage_db=tpcds
storage_policy=s3_main
storage_table=call_center

== Existing ==
delta.checkpointInterval=5
engine=mergetree
external=true
is_distribute=true
location=s3a://gluten-cicd/dataset/tpcds-sf100-mergetree-partition/call_center
metadata_path=_delta_log
owner=root
provider=clickhouse
sampling_key=
storage_policy=s3_main
:93:57', 'org.apache.spark.sql.delta.DeltaErrorsBase:createTableWithDifferentPropertiesException:DeltaErrors.scala:1306', 'org.apache.spark.sql.delta.DeltaErrorsBase:createTableWithDifferentPropertiesException$:DeltaErrors.scala:1302', 'org.apache.spark.sql.delta.DeltaErrors$:createTableWithDifferentPropertiesException:DeltaErrors.scala:3382', 'org.apache.spark.sql.delta.commands.CreateDeltaTableCommand:verifyTableMetadata:CreateDeltaTableCommand.scala:494', 'org.apache.spark.sql.delta.commands.CreateDeltaTableCommand:createActionsForNewTableOrVerify$1:CreateDeltaTableCommand.scala:342', 'org.apache.spark.sql.delta.commands.CreateDeltaTableCommand:handleCreateTable:CreateDeltaTableCommand.scala:351', 'org.apache.spark.sql.delta.commands.CreateDeltaTableCommand:handleCommit:CreateDeltaTableCommand.scala:170', 'org.apache.spark.sql.delta.commands.CreateDeltaTableCommand:$anonfun$run$2:CreateDeltaTableCommand.scala:110', 'org.apache.spark.sql.delta.metering.DeltaLogging:recordFrameProfile:DeltaLogging.scala:168', 'org.apache.spark.sql.delta.metering.DeltaLogging:recordFrameProfile$:DeltaLogging.scala:166', 'org.apache.spark.sql.delta.commands.CreateDeltaTableCommand:recordFrameProfile:CreateDeltaTableCommand.scala:57', 'org.apache.spark.sql.delta.metering.DeltaLogging:$anonfun$recordDeltaOperationInternal$1:DeltaLogging.scala:136', 'com.databricks.spark.util.DatabricksLogging:recordOperation:DatabricksLogging.scala:128', 'com.databricks.spark.util.DatabricksLogging:recordOperation$:DatabricksLogging.scala:117', 'org.apache.spark.sql.delta.commands.CreateDeltaTableCommand:recordOperation:CreateDeltaTableCommand.scala:57', 'org.apache.spark.sql.delta.metering.DeltaLogging:recordDeltaOperationInternal:DeltaLogging.scala:135', 'org.apache.spark.sql.delta.metering.DeltaLogging:recordDeltaOperation:DeltaLogging.scala:125', 'org.apache.spark.sql.delta.metering.DeltaLogging:recordDeltaOperation$:DeltaLogging.scala:115', 'org.apache.spark.sql.delta.commands.CreateDeltaTableCommand:recordDeltaOperation:CreateDeltaTableCommand.scala:57', 'org.apache.spark.sql.delta.commands.CreateDeltaTableCommand:run:CreateDeltaTableCommand.scala:110', 'org.apache.spark.sql.execution.datasources.v2.clickhouse.ClickHouseSparkCatalog:org$apache$spark$sql$execution$datasources$v2$clickhouse$ClickHouseSparkCatalog$$createClickHouseTable:ClickHouseSparkCatalog.scala:203', 'org.apache.spark.sql.execution.datasources.v2.clickhouse.ClickHouseSparkCatalog:createTable:ClickHouseSparkCatalog.scala:95', 'org.apache.spark.sql.execution.datasources.v2.clickhouse.ClickHouseSparkCatalog:createTable:ClickHouseSparkCatalog.scala:79', 'org.apache.spark.sql.execution.datasources.v2.CreateTableExec:run:CreateTableExec.scala:44', 'org.apache.spark.sql.execution.datasources.v2.V2CommandExec:result$lzycompute:V2CommandExec.scala:43', 'org.apache.spark.sql.execution.datasources.v2.V2CommandExec:result:V2CommandExec.scala:43', 'org.apache.spark.sql.execution.datasources.v2.V2CommandExec:executeCollect:V2CommandExec.scala:49', 'org.apache.spark.sql.execution.QueryExecution$$anonfun$eagerlyExecuteCommands$1:$anonfun$applyOrElse$1:QueryExecution.scala:107', 'org.apache.spark.sql.execution.SQLExecution$:$anonfun$withNewExecutionId$6:SQLExecution.scala:125', 'org.apache.spark.sql.execution.SQLExecution$:withSQLConfPropagated:SQLExecution.scala:201', 'org.apache.spark.sql.execution.SQLExecution$:$anonfun$withNewExecutionId$1:SQLExecution.scala:108', 'org.apache.spark.sql.SparkSession:withActive:SparkSession.scala:900', 'org.apache.spark.sql.execution.SQLExecution$:withNewExecutionId:SQLExecution.scala:66', 'org.apache.spark.sql.execution.QueryExecution$$anonfun$eagerlyExecuteCommands$1:applyOrElse:QueryExecution.scala:107', 'org.apache.spark.sql.execution.QueryExecution$$anonfun$eagerlyExecuteCommands$1:applyOrElse:QueryExecution.scala:98', 'org.apache.spark.sql.catalyst.trees.TreeNode:$anonfun$transformDownWithPruning$1:TreeNode.scala:461', 'org.apache.spark.sql.catalyst.trees.CurrentOrigin$:withOrigin:origin.scala:76', 'org.apache.spark.sql.catalyst.trees.TreeNode:transformDownWithPruning:TreeNode.scala:461', 'org.apache.spark.sql.catalyst.plans.logical.LogicalPlan:org$apache$spark$sql$catalyst$plans$logical$AnalysisHelper$$super$transformDownWithPruning:LogicalPlan.scala:32', 'org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper:transformDownWithPruning:AnalysisHelper.scala:267', 'org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper:transformDownWithPruning$:AnalysisHelper.scala:263', 'org.apache.spark.sql.catalyst.plans.logical.LogicalPlan:transformDownWithPruning:LogicalPlan.scala:32', 'org.apache.spark.sql.catalyst.plans.logical.LogicalPlan:transformDownWithPruning:LogicalPlan.scala:32', 'org.apache.spark.sql.catalyst.trees.TreeNode:transformDown:TreeNode.scala:437', 'org.apache.spark.sql.execution.QueryExecution:eagerlyExecuteCommands:QueryExecution.scala:98', 'org.apache.spark.sql.execution.QueryExecution:commandExecuted$lzycompute:QueryExecution.scala:85', 'org.apache.spark.sql.execution.QueryExecution:commandExecuted:QueryExecution.scala:83', 'org.apache.spark.sql.Dataset:<init>:Dataset.scala:220', 'org.apache.spark.sql.Dataset$:$anonfun$ofRows$2:Dataset.scala:100', 'org.apache.spark.sql.SparkSession:withActive:SparkSession.scala:900', 'org.apache.spark.sql.Dataset$:ofRows:Dataset.scala:97', 'org.apache.spark.sql.SparkSession:$anonfun$sql$4:SparkSession.scala:691', 'org.apache.spark.sql.SparkSession:withActive:SparkSession.scala:900', 'org.apache.spark.sql.SparkSession:sql:SparkSession.scala:682', 'org.apache.spark.sql.SparkSession:sql:SparkSession.scala:713', 'org.apache.spark.sql.SparkSession:sql:SparkSession.scala:744', 'org.apache.spark.sql.SQLContext:sql:SQLContext.scala:651', 'org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation:org$apache$spark$sql$hive$thriftserver$SparkExecuteStatementOperation$$execute:SparkExecuteStatementOperation.scala:227'], sqlState='42KD7', errorCode=0, errorMessage='Error running query: [DELTA_CREATE_TABLE_WITH_DIFFERENT_PROPERTY] org.apache.spark.sql.delta.DeltaAnalysisException: [DELTA_CREATE_TABLE_WITH_DIFFERENT_PROPERTY] The specified properties do not match the existing properties at s3a://gluten-cicd/dataset/tpcds-sf100-mergetree-partition/call_center.

== Specified ==
delta.checkpointInterval=5
engine=mergetree
external=true
is_distribute=true
location=s3a://gluten-cicd/dataset/tpcds-sf100-mergetree-partition/call_center
metadata_path=_delta_log
owner=root
provider=clickhouse
sampling_key=
storage_db=tpcds
storage_policy=s3_main
storage_table=call_center

== Existing ==
delta.checkpointInterval=5
engine=mergetree
external=true
is_distribute=true
location=s3a://gluten-cicd/dataset/tpcds-sf100-mergetree-partition/call_center
metadata_path=_delta_log
owner=root
provider=clickhouse
sampling_key=
storage_policy=s3_main
'), operationHandle=None)
[2024-10-08T03:39:59.915Z] 10/08/2024 03:39:59 AM - INFO - drop table if exists catalog_page

Spark version

3.5

Spark configurations

No response

System information

No response

Relevant logs

No response

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working triage
Projects
None yet
Development

Successfully merging a pull request may close this issue.

1 participant