You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I had searched in the issues and found no similar issues.
What happened
Failed to synchronize data from ElasticSearch to Hive.
The error log is displayed in the error file.txt, and it contains the seatunnel's config information. error.txt
In addition, it has been verified in my environment: Es->Console and Mysql->Hive are executed normally.
I'm curious if this error is about ElasticsearchSink, it doesn't seem to be relevant to my configuration
24/03/14 16:39:31 INFO utils.ConfigBuilder: Loading config file from path: ./config/my/erdp_es_2_hive.config
24/03/14 16:39:31 INFO utils.ConfigShadeUtils: Load config shade spi: [base64]
24/03/14 16:39:31 INFO utils.ConfigBuilder: Parsed config file:
{
"env" : {
"parallelism" : 1,
"job.mode" : "BATCH"
},
"source" : [
{
"tls_verify_hostname" : false,
"password" : "",
"hosts" : [
"http://xxx:9200"
],
"query" : {
"match_all" : {}
},
"index" : "web_measurement",
"result_table_name" : "t1",
"source" : [
"timestamp",
"projectName",
"department",
"departmentId",
"user",
"module",
"page",
"object",
"behaviorType",
"token"
],
"plugin_name" : "Elasticsearch",
"username" : "elastic"
}
],
"transform" : [
{
"field_mapper" : {
"timestamp" : "timestamp",
"projectName" : "project_name",
"department" : "department",
"departmentId" : "department_id",
"user" : "user",
"page" : "page",
"object" : "object",
"behaviorType" : "behavior_type",
"token" : "token"
},
"source_table_name" : "t1",
"result_table_name" : "t2",
"plugin_name" : "FieldMapper"
}
],
"sink" : [
{
"source_table_name" : "t2",
"table_name" : "ods.ods_erdp_web_measurement",
"metastore_uri" : "thrift://xxx:9083",
"plugin_name" : "hive"
}
]
}
24/03/14 16:39:31 INFO conf.HiveConf: Found configuration file file:/etc/hive/conf.cloudera.hive/hive-site.xml
24/03/14 16:39:31 INFO spark.SparkContext: Running Spark version 2.4.0-cdh6.3.2
24/03/14 16:39:31 INFO logging.DriverLogger: Added a local log appender at: /tmp/spark-426fbceb-0f4f-4d34-b704-e0a0dd383a5f/__driver_logs__/driver.log
24/03/14 16:39:31 INFO spark.SparkContext: Submitted application: SeaTunnel
24/03/14 16:39:32 INFO spark.SecurityManager: Changing view acls to: bigdata
24/03/14 16:39:32 INFO spark.SecurityManager: Changing modify acls to: bigdata
24/03/14 16:39:32 INFO spark.SecurityManager: Changing view acls groups to:
24/03/14 16:39:32 INFO spark.SecurityManager: Changing modify acls groups to:
24/03/14 16:39:32 INFO spark.SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(bigdata); groups with view permissions: Set(); users with modify permissions: Set(bigdata); groups with modify permissions: Set()
24/03/14 16:39:32 INFO util.Utils: Successfully started service 'sparkDriver' on port 40655.
24/03/14 16:39:32 INFO spark.SparkEnv: Registering MapOutputTracker
24/03/14 16:39:32 INFO spark.SparkEnv: Registering BlockManagerMaster
24/03/14 16:39:32 INFO storage.BlockManagerMasterEndpoint: Using org.apache.spark.storage.DefaultTopologyMapper for getting topology information
24/03/14 16:39:32 INFO storage.BlockManagerMasterEndpoint: BlockManagerMasterEndpoint up
24/03/14 16:39:32 INFO storage.DiskBlockManager: Created local directory at /tmp/blockmgr-a32b1b1d-8595-4a68-a583-2c07a5dcd6e8
24/03/14 16:39:32 INFO memory.MemoryStore: MemoryStore started with capacity 366.3 MB
24/03/14 16:39:32 INFO spark.SparkEnv: Registering OutputCommitCoordinator
24/03/14 16:39:32 INFO util.log: Logging initialized @2311ms
24/03/14 16:39:32 INFO server.Server: jetty-9.3.z-SNAPSHOT, build timestamp: 2018-09-05T05:11:46+08:00, git hash: 3ce520221d0240229c862b122d2b06c12a625732
24/03/14 16:39:32 INFO server.Server: Started @2378ms
24/03/14 16:39:32 WARN util.Utils: Service 'SparkUI' could not bind on port 4040. Attempting port 4041.
24/03/14 16:39:32 INFO server.AbstractConnector: Started ServerConnector@5f18f9d2{HTTP/1.1,[http/1.1]}{0.0.0.0:4041}
24/03/14 16:39:32 INFO util.Utils: Successfully started service 'SparkUI' on port 4041.
24/03/14 16:39:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@4d178d55{/jobs,null,AVAILABLE,@Spark}
24/03/14 16:39:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@33364212{/jobs/json,null,AVAILABLE,@Spark}
24/03/14 16:39:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@2216effc{/jobs/job,null,AVAILABLE,@Spark}
24/03/14 16:39:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@6da9dc6{/jobs/job/json,null,AVAILABLE,@Spark}
24/03/14 16:39:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@7fd69dd{/stages,null,AVAILABLE,@Spark}
24/03/14 16:39:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@12010fd1{/stages/json,null,AVAILABLE,@Spark}
24/03/14 16:39:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@7c84195{/stages/stage,null,AVAILABLE,@Spark}
24/03/14 16:39:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@23940f86{/stages/stage/json,null,AVAILABLE,@Spark}
24/03/14 16:39:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@66153688{/stages/pool,null,AVAILABLE,@Spark}
24/03/14 16:39:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@455824ad{/stages/pool/json,null,AVAILABLE,@Spark}
24/03/14 16:39:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@7318daf8{/storage,null,AVAILABLE,@Spark}
24/03/14 16:39:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@70f31322{/storage/json,null,AVAILABLE,@Spark}
24/03/14 16:39:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@3f1ddac2{/storage/rdd,null,AVAILABLE,@Spark}
24/03/14 16:39:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@3be4fcc0{/storage/rdd/json,null,AVAILABLE,@Spark}
24/03/14 16:39:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@e1e2e5e{/environment,null,AVAILABLE,@Spark}
24/03/14 16:39:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@661c46bc{/environment/json,null,AVAILABLE,@Spark}
24/03/14 16:39:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@37864b77{/executors,null,AVAILABLE,@Spark}
24/03/14 16:39:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@2b98b3bb{/executors/json,null,AVAILABLE,@Spark}
24/03/14 16:39:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@540b0448{/executors/threadDump,null,AVAILABLE,@Spark}
24/03/14 16:39:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@50a691d3{/executors/threadDump/json,null,AVAILABLE,@Spark}
24/03/14 16:39:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@557eb543{/static,null,AVAILABLE,@Spark}
24/03/14 16:39:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@99407c2{/,null,AVAILABLE,@Spark}
24/03/14 16:39:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@6c796cc1{/api,null,AVAILABLE,@Spark}
24/03/14 16:39:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@128c502c{/jobs/job/kill,null,AVAILABLE,@Spark}
24/03/14 16:39:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@45667d98{/stages/stage/kill,null,AVAILABLE,@Spark}
24/03/14 16:39:32 INFO ui.SparkUI: Bound SparkUI to 0.0.0.0, and started at http://cdh1:4041
24/03/14 16:39:32 INFO spark.SparkContext: Added JAR file:///opt/module/seatunnel/apache-seatunnel-2.3.4/lib/seatunnel-transforms-v2.jar at spark://cdh1:40655/jars/seatunnel-transforms-v2.jar with timestamp 1710405572511
24/03/14 16:39:32 INFO spark.SparkContext: Added JAR file:///opt/module/seatunnel/apache-seatunnel-2.3.4/lib/seatunnel-hadoop3-3.1.4-uber.jar at spark://cdh1:40655/jars/seatunnel-hadoop3-3.1.4-uber.jar with timestamp 1710405572512
24/03/14 16:39:32 INFO spark.SparkContext: Added JAR file:///opt/module/seatunnel/apache-seatunnel-2.3.4/connectors/connector-elasticsearch-2.3.4.jar at spark://cdh1:40655/jars/connector-elasticsearch-2.3.4.jar with timestamp 1710405572512
24/03/14 16:39:32 INFO spark.SparkContext: Added JAR file:///opt/module/seatunnel/apache-seatunnel-2.3.4/connectors/connector-hive-2.3.4.jar at spark://cdh1:40655/jars/connector-hive-2.3.4.jar with timestamp 1710405572512
24/03/14 16:39:32 INFO spark.SparkContext: Added JAR file:/opt/module/seatunnel/apache-seatunnel-2.3.4/starter/seatunnel-spark-2-starter.jar at spark://cdh1:40655/jars/seatunnel-spark-2-starter.jar with timestamp 1710405572512
24/03/14 16:39:32 INFO executor.Executor: Starting executor ID driver on host localhost
24/03/14 16:39:32 INFO util.Utils: Successfully started service 'org.apache.spark.network.netty.NettyBlockTransferService' on port 43782.
24/03/14 16:39:32 INFO netty.NettyBlockTransferService: Server created on cdh1:43782
24/03/14 16:39:32 INFO storage.BlockManager: Using org.apache.spark.storage.RandomBlockReplicationPolicy for block replication policy
24/03/14 16:39:32 INFO storage.BlockManagerMaster: Registering BlockManager BlockManagerId(driver, cdh1, 43782, None)
24/03/14 16:39:32 INFO storage.BlockManagerMasterEndpoint: Registering block manager cdh1:43782 with 366.3 MB RAM, BlockManagerId(driver, cdh1, 43782, None)
24/03/14 16:39:32 INFO storage.BlockManagerMaster: Registered BlockManager BlockManagerId(driver, cdh1, 43782, None)
24/03/14 16:39:32 INFO storage.BlockManager: external shuffle service port = 7337
24/03/14 16:39:32 INFO storage.BlockManager: Initialized BlockManager: BlockManagerId(driver, cdh1, 43782, None)
24/03/14 16:39:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@4110765e{/metrics/json,null,AVAILABLE,@Spark}
24/03/14 16:39:33 INFO scheduler.EventLoggingListener: Logging events to hdfs://cdh1:8020/user/spark/applicationHistory/local-1710405572561
24/03/14 16:39:33 WARN lineage.LineageWriter: Lineage directory /data/var/log/spark/lineage doesn't exist or is not writable. Lineage for this application will be disabled.
24/03/14 16:39:33 INFO util.Utils: Extension com.cloudera.spark.lineage.NavigatorAppListener not being initialized.
24/03/14 16:39:33 INFO logging.DriverLogger$DfsAsyncWriter: Started driver log file sync to: /user/spark/driverLogs/local-1710405572561_driver.log
24/03/14 16:39:33 INFO discovery.AbstractPluginDiscovery: Load SeaTunnelSource Plugin from /opt/module/seatunnel/apache-seatunnel-2.3.4/connectors
24/03/14 16:39:33 INFO discovery.AbstractPluginDiscovery: Load Factory Plugin from /opt/module/seatunnel/apache-seatunnel-2.3.4/connectors
24/03/14 16:39:33 INFO discovery.AbstractPluginDiscovery: Discovery plugin jar for: PluginIdentifier{engineType='seatunnel', pluginType='source', pluginName='Elasticsearch'} at: file:/opt/module/seatunnel/apache-seatunnel-2.3.4/connectors/connector-elasticsearch-2.3.4.jar
24/03/14 16:39:33 INFO discovery.AbstractPluginDiscovery: Load plugin: PluginIdentifier{engineType='seatunnel', pluginType='source', pluginName='Elasticsearch'} from classpath
24/03/14 16:39:33 INFO client.EsRestClient: GET /web_measurement/_mappings respnse={"web_measurement":{"mappings":{"properties":{"_class":{"type":"text","fields":{"keyword":{"type":"keyword","ignore_above":256}}},"behaviorType":{"type":"text","fields":{"keyword":{"type":"keyword","ignore_above":256}}},"department":{"type":"text","fields":{"keyword":{"type":"keyword","ignore_above":256}}},"departmentId":{"type":"text","fields":{"keyword":{"type":"keyword","ignore_above":256}}},"event":{"type":"text","fields":{"keyword":{"type":"keyword","ignore_above":256}}},"extension":{"properties":{"accurate":{"type":"long"},"appID":{"type":"text","fields":{"keyword":{"type":"keyword","ignore_above":256}}},"eventType":{"type":"text","fields":{"keyword":{"type":"keyword","ignore_above":256}}},"height":{"type":"long"},"innerHTML":{"type":"text","fields":{"keyword":{"type":"keyword","ignore_above":256}}},"left":{"type":"long"},"outerHTML":{"type":"text","fields":{"keyword":{"type":"keyword","ignore_above":256}}},"pageHeight":{"type":"long"},"pageURL":{"type":"text","fields":{"keyword":{"type":"keyword","ignore_above":256}}},"paths":{"type":"text","fields":{"keyword":{"type":"keyword","ignore_above":256}}},"scrollTop":{"type":"long"},"startTime":{"type":"float"},"subType":{"type":"text","fields":{"keyword":{"type":"keyword","ignore_above":256}}},"target":{"type":"text","fields":{"keyword":{"type":"keyword","ignore_above":256}}},"token":{"type":"text","fields":{"keyword":{"type":"keyword","ignore_above":256}}},"top":{"type":"long"},"type":{"type":"text","fields":{"keyword":{"type":"keyword","ignore_above":256}}},"useId":{"type":"text","fields":{"keyword":{"type":"keyword","ignore_above":256}}},"uuid":{"type":"text","fields":{"keyword":{"type":"keyword","ignore_above":256}}},"viewport":{"properties":{"height":{"type":"long"},"width":{"type":"long"}}},"width":{"type":"long"}}},"id":{"type":"long"},"module":{"type":"text","fields":{"keyword":{"type":"keyword","ignore_above":256}}},"object":{"type":"text","fields":{"keyword":{"type":"keyword","ignore_above":256}}},"page":{"type":"text","fields":{"keyword":{"type":"keyword","ignore_above":256}}},"pageURL":{"type":"text","fields":{"keyword":{"type":"keyword","ignore_above":256}}},"projectName":{"type":"text","fields":{"keyword":{"type":"keyword","ignore_above":256}}},"referrer":{"type":"text","fields":{"keyword":{"type":"keyword","ignore_above":256}}},"timestamp":{"type":"long"},"token":{"type":"text","fields":{"keyword":{"type":"keyword","ignore_above":256}}},"user":{"type":"text","fields":{"keyword":{"type":"keyword","ignore_above":256}}}}}}}
24/03/14 16:39:33 INFO execution.SparkRuntimeEnvironment: register plugins :[file:/opt/module/seatunnel/apache-seatunnel-2.3.4/connectors/connector-elasticsearch-2.3.4.jar]
24/03/14 16:39:33 INFO discovery.AbstractPluginDiscovery: Load SeaTunnelTransform Plugin from /opt/module/seatunnel/apache-seatunnel-2.3.4/lib
24/03/14 16:39:33 INFO execution.SparkRuntimeEnvironment: register plugins :[]
24/03/14 16:39:33 INFO discovery.AbstractPluginDiscovery: Load Factory Plugin from /opt/module/seatunnel/apache-seatunnel-2.3.4/connectors
24/03/14 16:39:33 INFO discovery.AbstractPluginDiscovery: Load SeaTunnelSink Plugin from /opt/module/seatunnel/apache-seatunnel-2.3.4/connectors
24/03/14 16:39:33 INFO discovery.AbstractPluginDiscovery: Discovery plugin jar for: PluginIdentifier{engineType='seatunnel', pluginType='sink', pluginName='hive'} at: file:/opt/module/seatunnel/apache-seatunnel-2.3.4/connectors/connector-hive-2.3.4.jar
24/03/14 16:39:33 INFO discovery.AbstractPluginDiscovery: Load plugin: PluginIdentifier{engineType='seatunnel', pluginType='sink', pluginName='hive'} from classpath
24/03/14 16:39:33 INFO execution.SparkRuntimeEnvironment: register plugins :[file:/opt/module/seatunnel/apache-seatunnel-2.3.4/connectors/connector-hive-2.3.4.jar]
24/03/14 16:39:33 INFO internal.SharedState: loading hive config file: file:/etc/hive/conf.cloudera.hive/hive-site.xml
24/03/14 16:39:33 INFO internal.SharedState: spark.sql.warehouse.dir is not set, but hive.metastore.warehouse.dir is set. Setting spark.sql.warehouse.dir to the value of hive.metastore.warehouse.dir ('/user/hive/warehouse').
24/03/14 16:39:33 INFO internal.SharedState: Warehouse path is '/user/hive/warehouse'.
24/03/14 16:39:33 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@4776e209{/SQL,null,AVAILABLE,@Spark}
24/03/14 16:39:33 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@265a094b{/SQL/json,null,AVAILABLE,@Spark}
24/03/14 16:39:33 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@54c425b1{/SQL/execution,null,AVAILABLE,@Spark}
24/03/14 16:39:33 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@50b734c4{/SQL/execution/json,null,AVAILABLE,@Spark}
24/03/14 16:39:33 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@b14b60a{/static/sql,null,AVAILABLE,@Spark}
24/03/14 16:39:34 INFO state.StateStoreCoordinatorRef: Registered StateStoreCoordinator endpoint
24/03/14 16:39:34 WARN lineage.LineageWriter: Lineage directory /data/var/log/spark/lineage doesn't exist or is not writable. Lineage for this application will be disabled.
24/03/14 16:39:34 INFO util.Utils: Extension com.cloudera.spark.lineage.NavigatorQueryListener not being initialized.
24/03/14 16:39:35 INFO hive.HiveUtils: Initializing HiveMetastoreConnection version 2.1 using Spark classes.
24/03/14 16:39:35 INFO conf.HiveConf: Found configuration file file:/etc/hive/conf.cloudera.hive/hive-site.xml
24/03/14 16:39:35 INFO session.SessionState: Created HDFS directory: /tmp/hive/bigdata/0c77058d-a63d-4634-b205-89e6db8702bd
24/03/14 16:39:35 INFO session.SessionState: Created local directory: /tmp/bigdata/0c77058d-a63d-4634-b205-89e6db8702bd
24/03/14 16:39:35 INFO session.SessionState: Created HDFS directory: /tmp/hive/bigdata/0c77058d-a63d-4634-b205-89e6db8702bd/_tmp_space.db
24/03/14 16:39:35 INFO client.HiveClientImpl: Warehouse location for Hive client (version 2.1.1) is /user/hive/warehouse
24/03/14 16:39:36 INFO hive.metastore: HMS client filtering is enabled.
24/03/14 16:39:36 INFO hive.metastore: Trying to connect to metastore with URI thrift://cdh1:9083
24/03/14 16:39:36 INFO hive.metastore: Opened a connection to metastore, current connections: 1
24/03/14 16:39:36 INFO hive.metastore: Connected to metastore.
24/03/14 16:39:36 INFO discovery.AbstractPluginDiscovery: Load SeaTunnelSink Plugin from /opt/module/seatunnel/apache-seatunnel-2.3.4/connectors
Exception in thread "main" java.util.ServiceConfigurationError: org.apache.seatunnel.api.sink.SeaTunnelSink: Provider org.apache.seatunnel.connectors.seatunnel.elasticsearch.sink.ElasticsearchSink could not be instantiated
at java.util.ServiceLoader.fail(ServiceLoader.java:232)
at java.util.ServiceLoader.access$100(ServiceLoader.java:185)
at java.util.ServiceLoader$LazyIterator.nextService(ServiceLoader.java:384)
at java.util.ServiceLoader$LazyIterator.next(ServiceLoader.java:404)
at java.util.ServiceLoader$1.next(ServiceLoader.java:480)
at org.apache.seatunnel.plugin.discovery.AbstractPluginDiscovery.loadPluginInstance(AbstractPluginDiscovery.java:302)
at org.apache.seatunnel.plugin.discovery.AbstractPluginDiscovery.createOptionalPluginInstance(AbstractPluginDiscovery.java:183)
at org.apache.seatunnel.plugin.discovery.AbstractPluginDiscovery.createPluginInstance(AbstractPluginDiscovery.java:227)
at org.apache.seatunnel.plugin.discovery.AbstractPluginDiscovery.createPluginInstance(AbstractPluginDiscovery.java:171)
at org.apache.seatunnel.core.starter.spark.execution.SinkExecuteProcessor.fallbackCreateSink(SinkExecuteProcessor.java:176)
at org.apache.seatunnel.core.starter.spark.execution.SinkExecuteProcessor.execute(SinkExecuteProcessor.java:120)
at org.apache.seatunnel.core.starter.spark.execution.SparkExecution.execute(SparkExecution.java:71)
at org.apache.seatunnel.core.starter.spark.command.SparkTaskExecuteCommand.execute(SparkTaskExecuteCommand.java:60)
at org.apache.seatunnel.core.starter.SeaTunnel.run(SeaTunnel.java:40)
at org.apache.seatunnel.core.starter.spark.SeaTunnelSpark.main(SeaTunnelSpark.java:35)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.apache.spark.deploy.JavaMainApplication.start(SparkApplication.scala:52)
at org.apache.spark.deploy.SparkSubmit.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:851)
at org.apache.spark.deploy.SparkSubmit.doRunMain$1(SparkSubmit.scala:167)
at org.apache.spark.deploy.SparkSubmit.submit(SparkSubmit.scala:195)
at org.apache.spark.deploy.SparkSubmit.doSubmit(SparkSubmit.scala:86)
at org.apache.spark.deploy.SparkSubmit$$anon$2.doSubmit(SparkSubmit.scala:926)
at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:935)
at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
Caused by: java.lang.InstantiationException: org.apache.seatunnel.connectors.seatunnel.elasticsearch.sink.ElasticsearchSink
at java.lang.Class.newInstance(Class.java:427)
at java.util.ServiceLoader$LazyIterator.nextService(ServiceLoader.java:380)
... 24 more
Caused by: java.lang.NoSuchMethodException: org.apache.seatunnel.connectors.seatunnel.elasticsearch.sink.ElasticsearchSink.<init>()
at java.lang.Class.getConstructor0(Class.java:3082)
at java.lang.Class.newInstance(Class.java:412)
... 25 more
24/03/14 16:39:36 INFO spark.SparkContext: Invoking stop() from shutdown hook
24/03/14 16:39:36 INFO server.AbstractConnector: Stopped Spark@5f18f9d2{HTTP/1.1,[http/1.1]}{0.0.0.0:4041}
24/03/14 16:39:36 INFO ui.SparkUI: Stopped Spark web UI at http://cdh1:4041
24/03/14 16:39:36 INFO spark.MapOutputTrackerMasterEndpoint: MapOutputTrackerMasterEndpoint stopped!
24/03/14 16:39:36 INFO memory.MemoryStore: MemoryStore cleared
24/03/14 16:39:36 INFO storage.BlockManager: BlockManager stopped
24/03/14 16:39:36 INFO storage.BlockManagerMaster: BlockManagerMaster stopped
24/03/14 16:39:36 INFO scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint: OutputCommitCoordinator stopped!
24/03/14 16:39:36 INFO spark.SparkContext: Successfully stopped SparkContext
24/03/14 16:39:36 INFO util.ShutdownHookManager: Shutdown hook called
24/03/14 16:39:36 INFO util.ShutdownHookManager: Deleting directory /tmp/spark-eba84842-a375-450e-8aa2-604c616f818f
24/03/14 16:39:36 INFO util.ShutdownHookManager: Deleting directory /tmp/spark-426fbceb-0f4f-4d34-b704-e0a0dd383a5f
Search before asking
What happened
Failed to synchronize data from ElasticSearch to Hive.
The error log is displayed in the error file.txt, and it contains the seatunnel's config information.
error.txt
In addition, it has been verified in my environment: Es->Console and Mysql->Hive are executed normally.
I'm curious if this error is about ElasticsearchSink, it doesn't seem to be relevant to my configuration
SeaTunnel Version
2.3.4
SeaTunnel Config
Running Command
Error Exception
Zeta or Flink or Spark Version
spark version: 2.4.0-cdh6.3.2
Java or Scala Version
java version: 1.8.0_181
Screenshots
No response
Are you willing to submit PR?
Code of Conduct
The text was updated successfully, but these errors were encountered: