Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Bug] [ElasticSearch2Hive Fail] Failed to synchronize data from ElasticSearch to Hive. #6506

Closed
3 tasks done
jiangqingsong opened this issue Mar 14, 2024 · 1 comment · Fixed by #6551
Closed
3 tasks done
Labels

Comments

@jiangqingsong
Copy link

jiangqingsong commented Mar 14, 2024

Search before asking

  • I had searched in the issues and found no similar issues.

What happened

Failed to synchronize data from ElasticSearch to Hive.
The error log is displayed in the error file.txt, and it contains the seatunnel's config information.
error.txt

In addition, it has been verified in my environment: Es->Console and Mysql->Hive are executed normally.
I'm curious if this error is about ElasticsearchSink, it doesn't seem to be relevant to my configuration

SeaTunnel Version

2.3.4

SeaTunnel Config

en
irallelism = 1
  job.mode = "BATCH"
}

source {
        Elasticsearch {
        hosts = ["http://localhost:9200"]
        username = "elastic"
        password = ""
        tls_verify_hostname = false
        index = "person"
        source = ["name", "age"]
        query = {"match_all":{}}

    }
}

transform {}

sink {

  Hive {
    table_name = "test.person"
    metastore_uri = "thrift://xxx:9083"
  }

}

Running Command

./bin/start-seatunnel-spark-2-connector-v2.sh --master local[4] --deploy-mode client --config ./config/my/erdp_es_2_hive.config --plugin_name hive

Error Exception

24/03/14 16:39:31 INFO utils.ConfigBuilder: Loading config file from path: ./config/my/erdp_es_2_hive.config
24/03/14 16:39:31 INFO utils.ConfigShadeUtils: Load config shade spi: [base64]
24/03/14 16:39:31 INFO utils.ConfigBuilder: Parsed config file: 
{
    "env" : {
        "parallelism" : 1,
        "job.mode" : "BATCH"
    },
    "source" : [
        {
            "tls_verify_hostname" : false,
            "password" : "",
            "hosts" : [
                "http://xxx:9200"
            ],
            "query" : {
                "match_all" : {}
            },
            "index" : "web_measurement",
            "result_table_name" : "t1",
            "source" : [
                "timestamp",
                "projectName",
                "department",
                "departmentId",
                "user",
                "module",
                "page",
                "object",
                "behaviorType",
                "token"
            ],
            "plugin_name" : "Elasticsearch",
            "username" : "elastic"
        }
    ],
    "transform" : [
        {
            "field_mapper" : {
                "timestamp" : "timestamp",
                "projectName" : "project_name",
                "department" : "department",
                "departmentId" : "department_id",
                "user" : "user",
                "page" : "page",
                "object" : "object",
                "behaviorType" : "behavior_type",
                "token" : "token"
            },
            "source_table_name" : "t1",
            "result_table_name" : "t2",
            "plugin_name" : "FieldMapper"
        }
    ],
    "sink" : [
        {
            "source_table_name" : "t2",
            "table_name" : "ods.ods_erdp_web_measurement",
            "metastore_uri" : "thrift://xxx:9083",
            "plugin_name" : "hive"
        }
    ]
}

24/03/14 16:39:31 INFO conf.HiveConf: Found configuration file file:/etc/hive/conf.cloudera.hive/hive-site.xml
24/03/14 16:39:31 INFO spark.SparkContext: Running Spark version 2.4.0-cdh6.3.2
24/03/14 16:39:31 INFO logging.DriverLogger: Added a local log appender at: /tmp/spark-426fbceb-0f4f-4d34-b704-e0a0dd383a5f/__driver_logs__/driver.log
24/03/14 16:39:31 INFO spark.SparkContext: Submitted application: SeaTunnel
24/03/14 16:39:32 INFO spark.SecurityManager: Changing view acls to: bigdata
24/03/14 16:39:32 INFO spark.SecurityManager: Changing modify acls to: bigdata
24/03/14 16:39:32 INFO spark.SecurityManager: Changing view acls groups to: 
24/03/14 16:39:32 INFO spark.SecurityManager: Changing modify acls groups to: 
24/03/14 16:39:32 INFO spark.SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users  with view permissions: Set(bigdata); groups with view permissions: Set(); users  with modify permissions: Set(bigdata); groups with modify permissions: Set()
24/03/14 16:39:32 INFO util.Utils: Successfully started service 'sparkDriver' on port 40655.
24/03/14 16:39:32 INFO spark.SparkEnv: Registering MapOutputTracker
24/03/14 16:39:32 INFO spark.SparkEnv: Registering BlockManagerMaster
24/03/14 16:39:32 INFO storage.BlockManagerMasterEndpoint: Using org.apache.spark.storage.DefaultTopologyMapper for getting topology information
24/03/14 16:39:32 INFO storage.BlockManagerMasterEndpoint: BlockManagerMasterEndpoint up
24/03/14 16:39:32 INFO storage.DiskBlockManager: Created local directory at /tmp/blockmgr-a32b1b1d-8595-4a68-a583-2c07a5dcd6e8
24/03/14 16:39:32 INFO memory.MemoryStore: MemoryStore started with capacity 366.3 MB
24/03/14 16:39:32 INFO spark.SparkEnv: Registering OutputCommitCoordinator
24/03/14 16:39:32 INFO util.log: Logging initialized @2311ms
24/03/14 16:39:32 INFO server.Server: jetty-9.3.z-SNAPSHOT, build timestamp: 2018-09-05T05:11:46+08:00, git hash: 3ce520221d0240229c862b122d2b06c12a625732
24/03/14 16:39:32 INFO server.Server: Started @2378ms
24/03/14 16:39:32 WARN util.Utils: Service 'SparkUI' could not bind on port 4040. Attempting port 4041.
24/03/14 16:39:32 INFO server.AbstractConnector: Started ServerConnector@5f18f9d2{HTTP/1.1,[http/1.1]}{0.0.0.0:4041}
24/03/14 16:39:32 INFO util.Utils: Successfully started service 'SparkUI' on port 4041.
24/03/14 16:39:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@4d178d55{/jobs,null,AVAILABLE,@Spark}
24/03/14 16:39:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@33364212{/jobs/json,null,AVAILABLE,@Spark}
24/03/14 16:39:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@2216effc{/jobs/job,null,AVAILABLE,@Spark}
24/03/14 16:39:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@6da9dc6{/jobs/job/json,null,AVAILABLE,@Spark}
24/03/14 16:39:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@7fd69dd{/stages,null,AVAILABLE,@Spark}
24/03/14 16:39:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@12010fd1{/stages/json,null,AVAILABLE,@Spark}
24/03/14 16:39:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@7c84195{/stages/stage,null,AVAILABLE,@Spark}
24/03/14 16:39:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@23940f86{/stages/stage/json,null,AVAILABLE,@Spark}
24/03/14 16:39:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@66153688{/stages/pool,null,AVAILABLE,@Spark}
24/03/14 16:39:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@455824ad{/stages/pool/json,null,AVAILABLE,@Spark}
24/03/14 16:39:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@7318daf8{/storage,null,AVAILABLE,@Spark}
24/03/14 16:39:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@70f31322{/storage/json,null,AVAILABLE,@Spark}
24/03/14 16:39:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@3f1ddac2{/storage/rdd,null,AVAILABLE,@Spark}
24/03/14 16:39:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@3be4fcc0{/storage/rdd/json,null,AVAILABLE,@Spark}
24/03/14 16:39:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@e1e2e5e{/environment,null,AVAILABLE,@Spark}
24/03/14 16:39:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@661c46bc{/environment/json,null,AVAILABLE,@Spark}
24/03/14 16:39:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@37864b77{/executors,null,AVAILABLE,@Spark}
24/03/14 16:39:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@2b98b3bb{/executors/json,null,AVAILABLE,@Spark}
24/03/14 16:39:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@540b0448{/executors/threadDump,null,AVAILABLE,@Spark}
24/03/14 16:39:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@50a691d3{/executors/threadDump/json,null,AVAILABLE,@Spark}
24/03/14 16:39:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@557eb543{/static,null,AVAILABLE,@Spark}
24/03/14 16:39:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@99407c2{/,null,AVAILABLE,@Spark}
24/03/14 16:39:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@6c796cc1{/api,null,AVAILABLE,@Spark}
24/03/14 16:39:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@128c502c{/jobs/job/kill,null,AVAILABLE,@Spark}
24/03/14 16:39:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@45667d98{/stages/stage/kill,null,AVAILABLE,@Spark}
24/03/14 16:39:32 INFO ui.SparkUI: Bound SparkUI to 0.0.0.0, and started at http://cdh1:4041
24/03/14 16:39:32 INFO spark.SparkContext: Added JAR file:///opt/module/seatunnel/apache-seatunnel-2.3.4/lib/seatunnel-transforms-v2.jar at spark://cdh1:40655/jars/seatunnel-transforms-v2.jar with timestamp 1710405572511
24/03/14 16:39:32 INFO spark.SparkContext: Added JAR file:///opt/module/seatunnel/apache-seatunnel-2.3.4/lib/seatunnel-hadoop3-3.1.4-uber.jar at spark://cdh1:40655/jars/seatunnel-hadoop3-3.1.4-uber.jar with timestamp 1710405572512
24/03/14 16:39:32 INFO spark.SparkContext: Added JAR file:///opt/module/seatunnel/apache-seatunnel-2.3.4/connectors/connector-elasticsearch-2.3.4.jar at spark://cdh1:40655/jars/connector-elasticsearch-2.3.4.jar with timestamp 1710405572512
24/03/14 16:39:32 INFO spark.SparkContext: Added JAR file:///opt/module/seatunnel/apache-seatunnel-2.3.4/connectors/connector-hive-2.3.4.jar at spark://cdh1:40655/jars/connector-hive-2.3.4.jar with timestamp 1710405572512
24/03/14 16:39:32 INFO spark.SparkContext: Added JAR file:/opt/module/seatunnel/apache-seatunnel-2.3.4/starter/seatunnel-spark-2-starter.jar at spark://cdh1:40655/jars/seatunnel-spark-2-starter.jar with timestamp 1710405572512
24/03/14 16:39:32 INFO executor.Executor: Starting executor ID driver on host localhost
24/03/14 16:39:32 INFO util.Utils: Successfully started service 'org.apache.spark.network.netty.NettyBlockTransferService' on port 43782.
24/03/14 16:39:32 INFO netty.NettyBlockTransferService: Server created on cdh1:43782
24/03/14 16:39:32 INFO storage.BlockManager: Using org.apache.spark.storage.RandomBlockReplicationPolicy for block replication policy
24/03/14 16:39:32 INFO storage.BlockManagerMaster: Registering BlockManager BlockManagerId(driver, cdh1, 43782, None)
24/03/14 16:39:32 INFO storage.BlockManagerMasterEndpoint: Registering block manager cdh1:43782 with 366.3 MB RAM, BlockManagerId(driver, cdh1, 43782, None)
24/03/14 16:39:32 INFO storage.BlockManagerMaster: Registered BlockManager BlockManagerId(driver, cdh1, 43782, None)
24/03/14 16:39:32 INFO storage.BlockManager: external shuffle service port = 7337
24/03/14 16:39:32 INFO storage.BlockManager: Initialized BlockManager: BlockManagerId(driver, cdh1, 43782, None)
24/03/14 16:39:32 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@4110765e{/metrics/json,null,AVAILABLE,@Spark}
24/03/14 16:39:33 INFO scheduler.EventLoggingListener: Logging events to hdfs://cdh1:8020/user/spark/applicationHistory/local-1710405572561
24/03/14 16:39:33 WARN lineage.LineageWriter: Lineage directory /data/var/log/spark/lineage doesn't exist or is not writable. Lineage for this application will be disabled.
24/03/14 16:39:33 INFO util.Utils: Extension com.cloudera.spark.lineage.NavigatorAppListener not being initialized.
24/03/14 16:39:33 INFO logging.DriverLogger$DfsAsyncWriter: Started driver log file sync to: /user/spark/driverLogs/local-1710405572561_driver.log
24/03/14 16:39:33 INFO discovery.AbstractPluginDiscovery: Load SeaTunnelSource Plugin from /opt/module/seatunnel/apache-seatunnel-2.3.4/connectors
24/03/14 16:39:33 INFO discovery.AbstractPluginDiscovery: Load Factory Plugin from /opt/module/seatunnel/apache-seatunnel-2.3.4/connectors
24/03/14 16:39:33 INFO discovery.AbstractPluginDiscovery: Discovery plugin jar for: PluginIdentifier{engineType='seatunnel', pluginType='source', pluginName='Elasticsearch'} at: file:/opt/module/seatunnel/apache-seatunnel-2.3.4/connectors/connector-elasticsearch-2.3.4.jar
24/03/14 16:39:33 INFO discovery.AbstractPluginDiscovery: Load plugin: PluginIdentifier{engineType='seatunnel', pluginType='source', pluginName='Elasticsearch'} from classpath
24/03/14 16:39:33 INFO client.EsRestClient: GET /web_measurement/_mappings respnse={"web_measurement":{"mappings":{"properties":{"_class":{"type":"text","fields":{"keyword":{"type":"keyword","ignore_above":256}}},"behaviorType":{"type":"text","fields":{"keyword":{"type":"keyword","ignore_above":256}}},"department":{"type":"text","fields":{"keyword":{"type":"keyword","ignore_above":256}}},"departmentId":{"type":"text","fields":{"keyword":{"type":"keyword","ignore_above":256}}},"event":{"type":"text","fields":{"keyword":{"type":"keyword","ignore_above":256}}},"extension":{"properties":{"accurate":{"type":"long"},"appID":{"type":"text","fields":{"keyword":{"type":"keyword","ignore_above":256}}},"eventType":{"type":"text","fields":{"keyword":{"type":"keyword","ignore_above":256}}},"height":{"type":"long"},"innerHTML":{"type":"text","fields":{"keyword":{"type":"keyword","ignore_above":256}}},"left":{"type":"long"},"outerHTML":{"type":"text","fields":{"keyword":{"type":"keyword","ignore_above":256}}},"pageHeight":{"type":"long"},"pageURL":{"type":"text","fields":{"keyword":{"type":"keyword","ignore_above":256}}},"paths":{"type":"text","fields":{"keyword":{"type":"keyword","ignore_above":256}}},"scrollTop":{"type":"long"},"startTime":{"type":"float"},"subType":{"type":"text","fields":{"keyword":{"type":"keyword","ignore_above":256}}},"target":{"type":"text","fields":{"keyword":{"type":"keyword","ignore_above":256}}},"token":{"type":"text","fields":{"keyword":{"type":"keyword","ignore_above":256}}},"top":{"type":"long"},"type":{"type":"text","fields":{"keyword":{"type":"keyword","ignore_above":256}}},"useId":{"type":"text","fields":{"keyword":{"type":"keyword","ignore_above":256}}},"uuid":{"type":"text","fields":{"keyword":{"type":"keyword","ignore_above":256}}},"viewport":{"properties":{"height":{"type":"long"},"width":{"type":"long"}}},"width":{"type":"long"}}},"id":{"type":"long"},"module":{"type":"text","fields":{"keyword":{"type":"keyword","ignore_above":256}}},"object":{"type":"text","fields":{"keyword":{"type":"keyword","ignore_above":256}}},"page":{"type":"text","fields":{"keyword":{"type":"keyword","ignore_above":256}}},"pageURL":{"type":"text","fields":{"keyword":{"type":"keyword","ignore_above":256}}},"projectName":{"type":"text","fields":{"keyword":{"type":"keyword","ignore_above":256}}},"referrer":{"type":"text","fields":{"keyword":{"type":"keyword","ignore_above":256}}},"timestamp":{"type":"long"},"token":{"type":"text","fields":{"keyword":{"type":"keyword","ignore_above":256}}},"user":{"type":"text","fields":{"keyword":{"type":"keyword","ignore_above":256}}}}}}}
24/03/14 16:39:33 INFO execution.SparkRuntimeEnvironment: register plugins :[file:/opt/module/seatunnel/apache-seatunnel-2.3.4/connectors/connector-elasticsearch-2.3.4.jar]
24/03/14 16:39:33 INFO discovery.AbstractPluginDiscovery: Load SeaTunnelTransform Plugin from /opt/module/seatunnel/apache-seatunnel-2.3.4/lib
24/03/14 16:39:33 INFO execution.SparkRuntimeEnvironment: register plugins :[]
24/03/14 16:39:33 INFO discovery.AbstractPluginDiscovery: Load Factory Plugin from /opt/module/seatunnel/apache-seatunnel-2.3.4/connectors
24/03/14 16:39:33 INFO discovery.AbstractPluginDiscovery: Load SeaTunnelSink Plugin from /opt/module/seatunnel/apache-seatunnel-2.3.4/connectors
24/03/14 16:39:33 INFO discovery.AbstractPluginDiscovery: Discovery plugin jar for: PluginIdentifier{engineType='seatunnel', pluginType='sink', pluginName='hive'} at: file:/opt/module/seatunnel/apache-seatunnel-2.3.4/connectors/connector-hive-2.3.4.jar
24/03/14 16:39:33 INFO discovery.AbstractPluginDiscovery: Load plugin: PluginIdentifier{engineType='seatunnel', pluginType='sink', pluginName='hive'} from classpath
24/03/14 16:39:33 INFO execution.SparkRuntimeEnvironment: register plugins :[file:/opt/module/seatunnel/apache-seatunnel-2.3.4/connectors/connector-hive-2.3.4.jar]
24/03/14 16:39:33 INFO internal.SharedState: loading hive config file: file:/etc/hive/conf.cloudera.hive/hive-site.xml
24/03/14 16:39:33 INFO internal.SharedState: spark.sql.warehouse.dir is not set, but hive.metastore.warehouse.dir is set. Setting spark.sql.warehouse.dir to the value of hive.metastore.warehouse.dir ('/user/hive/warehouse').
24/03/14 16:39:33 INFO internal.SharedState: Warehouse path is '/user/hive/warehouse'.
24/03/14 16:39:33 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@4776e209{/SQL,null,AVAILABLE,@Spark}
24/03/14 16:39:33 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@265a094b{/SQL/json,null,AVAILABLE,@Spark}
24/03/14 16:39:33 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@54c425b1{/SQL/execution,null,AVAILABLE,@Spark}
24/03/14 16:39:33 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@50b734c4{/SQL/execution/json,null,AVAILABLE,@Spark}
24/03/14 16:39:33 INFO handler.ContextHandler: Started o.s.j.s.ServletContextHandler@b14b60a{/static/sql,null,AVAILABLE,@Spark}
24/03/14 16:39:34 INFO state.StateStoreCoordinatorRef: Registered StateStoreCoordinator endpoint
24/03/14 16:39:34 WARN lineage.LineageWriter: Lineage directory /data/var/log/spark/lineage doesn't exist or is not writable. Lineage for this application will be disabled.
24/03/14 16:39:34 INFO util.Utils: Extension com.cloudera.spark.lineage.NavigatorQueryListener not being initialized.
24/03/14 16:39:35 INFO hive.HiveUtils: Initializing HiveMetastoreConnection version 2.1 using Spark classes.
24/03/14 16:39:35 INFO conf.HiveConf: Found configuration file file:/etc/hive/conf.cloudera.hive/hive-site.xml
24/03/14 16:39:35 INFO session.SessionState: Created HDFS directory: /tmp/hive/bigdata/0c77058d-a63d-4634-b205-89e6db8702bd
24/03/14 16:39:35 INFO session.SessionState: Created local directory: /tmp/bigdata/0c77058d-a63d-4634-b205-89e6db8702bd
24/03/14 16:39:35 INFO session.SessionState: Created HDFS directory: /tmp/hive/bigdata/0c77058d-a63d-4634-b205-89e6db8702bd/_tmp_space.db
24/03/14 16:39:35 INFO client.HiveClientImpl: Warehouse location for Hive client (version 2.1.1) is /user/hive/warehouse
24/03/14 16:39:36 INFO hive.metastore: HMS client filtering is enabled.
24/03/14 16:39:36 INFO hive.metastore: Trying to connect to metastore with URI thrift://cdh1:9083
24/03/14 16:39:36 INFO hive.metastore: Opened a connection to metastore, current connections: 1
24/03/14 16:39:36 INFO hive.metastore: Connected to metastore.
24/03/14 16:39:36 INFO discovery.AbstractPluginDiscovery: Load SeaTunnelSink Plugin from /opt/module/seatunnel/apache-seatunnel-2.3.4/connectors
Exception in thread "main" java.util.ServiceConfigurationError: org.apache.seatunnel.api.sink.SeaTunnelSink: Provider org.apache.seatunnel.connectors.seatunnel.elasticsearch.sink.ElasticsearchSink could not be instantiated
        at java.util.ServiceLoader.fail(ServiceLoader.java:232)
        at java.util.ServiceLoader.access$100(ServiceLoader.java:185)
        at java.util.ServiceLoader$LazyIterator.nextService(ServiceLoader.java:384)
        at java.util.ServiceLoader$LazyIterator.next(ServiceLoader.java:404)
        at java.util.ServiceLoader$1.next(ServiceLoader.java:480)
        at org.apache.seatunnel.plugin.discovery.AbstractPluginDiscovery.loadPluginInstance(AbstractPluginDiscovery.java:302)
        at org.apache.seatunnel.plugin.discovery.AbstractPluginDiscovery.createOptionalPluginInstance(AbstractPluginDiscovery.java:183)
        at org.apache.seatunnel.plugin.discovery.AbstractPluginDiscovery.createPluginInstance(AbstractPluginDiscovery.java:227)
        at org.apache.seatunnel.plugin.discovery.AbstractPluginDiscovery.createPluginInstance(AbstractPluginDiscovery.java:171)
        at org.apache.seatunnel.core.starter.spark.execution.SinkExecuteProcessor.fallbackCreateSink(SinkExecuteProcessor.java:176)
        at org.apache.seatunnel.core.starter.spark.execution.SinkExecuteProcessor.execute(SinkExecuteProcessor.java:120)
        at org.apache.seatunnel.core.starter.spark.execution.SparkExecution.execute(SparkExecution.java:71)
        at org.apache.seatunnel.core.starter.spark.command.SparkTaskExecuteCommand.execute(SparkTaskExecuteCommand.java:60)
        at org.apache.seatunnel.core.starter.SeaTunnel.run(SeaTunnel.java:40)
        at org.apache.seatunnel.core.starter.spark.SeaTunnelSpark.main(SeaTunnelSpark.java:35)
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
        at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
        at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
        at java.lang.reflect.Method.invoke(Method.java:498)
        at org.apache.spark.deploy.JavaMainApplication.start(SparkApplication.scala:52)
        at org.apache.spark.deploy.SparkSubmit.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:851)
        at org.apache.spark.deploy.SparkSubmit.doRunMain$1(SparkSubmit.scala:167)
        at org.apache.spark.deploy.SparkSubmit.submit(SparkSubmit.scala:195)
        at org.apache.spark.deploy.SparkSubmit.doSubmit(SparkSubmit.scala:86)
        at org.apache.spark.deploy.SparkSubmit$$anon$2.doSubmit(SparkSubmit.scala:926)
        at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:935)
        at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
Caused by: java.lang.InstantiationException: org.apache.seatunnel.connectors.seatunnel.elasticsearch.sink.ElasticsearchSink
        at java.lang.Class.newInstance(Class.java:427)
        at java.util.ServiceLoader$LazyIterator.nextService(ServiceLoader.java:380)
        ... 24 more
Caused by: java.lang.NoSuchMethodException: org.apache.seatunnel.connectors.seatunnel.elasticsearch.sink.ElasticsearchSink.<init>()
        at java.lang.Class.getConstructor0(Class.java:3082)
        at java.lang.Class.newInstance(Class.java:412)
        ... 25 more
24/03/14 16:39:36 INFO spark.SparkContext: Invoking stop() from shutdown hook
24/03/14 16:39:36 INFO server.AbstractConnector: Stopped Spark@5f18f9d2{HTTP/1.1,[http/1.1]}{0.0.0.0:4041}
24/03/14 16:39:36 INFO ui.SparkUI: Stopped Spark web UI at http://cdh1:4041
24/03/14 16:39:36 INFO spark.MapOutputTrackerMasterEndpoint: MapOutputTrackerMasterEndpoint stopped!
24/03/14 16:39:36 INFO memory.MemoryStore: MemoryStore cleared
24/03/14 16:39:36 INFO storage.BlockManager: BlockManager stopped
24/03/14 16:39:36 INFO storage.BlockManagerMaster: BlockManagerMaster stopped
24/03/14 16:39:36 INFO scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint: OutputCommitCoordinator stopped!
24/03/14 16:39:36 INFO spark.SparkContext: Successfully stopped SparkContext
24/03/14 16:39:36 INFO util.ShutdownHookManager: Shutdown hook called
24/03/14 16:39:36 INFO util.ShutdownHookManager: Deleting directory /tmp/spark-eba84842-a375-450e-8aa2-604c616f818f
24/03/14 16:39:36 INFO util.ShutdownHookManager: Deleting directory /tmp/spark-426fbceb-0f4f-4d34-b704-e0a0dd383a5f

Zeta or Flink or Spark Version

spark version: 2.4.0-cdh6.3.2

Java or Scala Version

java version: 1.8.0_181

Screenshots

No response

Are you willing to submit PR?

  • Yes I am willing to submit a PR!

Code of Conduct

@LeonYoah
Copy link
Contributor

我也遇到了,2.3.4,s3file-hive报错:
1d690fb4a19d65acf4e406390533d49

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Projects
None yet
2 participants