Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[SUPPORT] flink 1.14 cdc 2.2 to hudi: org.apache.hudi.exception.HoodieRemoteException: 172.18.115.2:40827 failed to respond #5833

Closed
Chosen123Wang opened this issue Jun 10, 2022 · 15 comments
Labels
flink Issues related to flink priority:major degraded perf; unable to move forward; potential bugs

Comments

@Chosen123Wang
Copy link

Tips before filing an issue

  • Have you gone through our FAQs?

  • Join the mailing list to engage in conversations and get faster support at dev-subscribe@hudi.apache.org.

  • If you have triaged this as a bug, then file an issue directly.

Describe the problem you faced

A clear and concise description of the problem.

To Reproduce

Steps to reproduce the behavior:

  1. flink cdc to hudi

Expected behavior

A clear and concise description of what you expected to happen.

Environment Description

  • Hudi version : 0.11.0

  • Spark version :

  • Hive version :

  • Hadoop version : 3.2.2

  • Storage (HDFS/S3/GCS..) :HDFS 3.2.2

  • Running on Docker? (yes/no) :

Additional context

Add any other context about the problem here.

Stacktrace

Add the stacktrace of the error.

flink cdc to hudi

org.apache.hudi.exception.HoodieRemoteException: 172.18.115.2:40827 failed to respond
at org.apache.hudi.common.table.view.RemoteHoodieTableFileSystemView.refresh(RemoteHoodieTableFileSystemView.java:420)
at org.apache.hudi.common.table.view.RemoteHoodieTableFileSystemView.sync(RemoteHoodieTableFileSystemView.java:484)
at org.apache.hudi.common.table.view.PriorityBasedFileSystemView.sync(PriorityBasedFileSystemView.java:257)
at org.apache.hudi.sink.partitioner.profile.WriteProfile.reload(WriteProfile.java:259)
at org.apache.hudi.sink.partitioner.BucketAssigner.reload(BucketAssigner.java:211)
at org.apache.hudi.sink.partitioner.BucketAssignFunction.notifyCheckpointComplete(BucketAssignFunction.java:234)
at org.apache.flink.streaming.api.operators.AbstractUdfStreamOperator.notifyCheckpointComplete(AbstractUdfStreamOperator.java:126)
at org.apache.flink.streaming.runtime.tasks.StreamOperatorWrapper.notifyCheckpointComplete(StreamOperatorWrapper.java:99)
at org.apache.flink.streaming.runtime.tasks.RegularOperatorChain.notifyCheckpointComplete(RegularOperatorChain.java:152)
at org.apache.flink.streaming.runtime.tasks.SubtaskCheckpointCoordinatorImpl.notifyCheckpointComplete(SubtaskCheckpointCoordinatorImpl.java:348)
at org.apache.flink.streaming.runtime.tasks.StreamTask.notifyCheckpointComplete(StreamTask.java:1426)
at org.apache.flink.streaming.runtime.tasks.StreamTask.lambda$notifyCheckpointCompleteAsync$16(StreamTask.java:1374)
at org.apache.flink.streaming.runtime.tasks.StreamTask.lambda$notifyCheckpointOperation$18(StreamTask.java:1406)
at org.apache.flink.streaming.runtime.tasks.StreamTaskActionExecutor$1.runThrowing(StreamTaskActionExecutor.java:50)
at org.apache.flink.streaming.runtime.tasks.mailbox.Mail.run(Mail.java:90)
at org.apache.flink.streaming.runtime.tasks.mailbox.MailboxProcessor.processMailsWhenDefaultActionUnavailable(MailboxProcessor.java:338)
at org.apache.flink.streaming.runtime.tasks.mailbox.MailboxProcessor.processMail(MailboxProcessor.java:324)
at org.apache.flink.streaming.runtime.tasks.mailbox.MailboxProcessor.runMailboxLoop(MailboxProcessor.java:201)
at org.apache.flink.streaming.runtime.tasks.StreamTask.runMailboxLoop(StreamTask.java:809)
at org.apache.flink.streaming.runtime.tasks.StreamTask.invoke(StreamTask.java:761)
at org.apache.flink.runtime.taskmanager.Task.runWithSystemExitMonitoring(Task.java:958)
at org.apache.flink.runtime.taskmanager.Task.restoreAndInvoke(Task.java:937)
at org.apache.flink.runtime.taskmanager.Task.doRun(Task.java:766)
at org.apache.flink.runtime.taskmanager.Task.run(Task.java:575)
at java.lang.Thread.run(Thread.java:748)
Caused by: org.apache.http.NoHttpResponseException: 172.18.115.2:40827 failed to respond
at org.apache.http.impl.conn.DefaultHttpResponseParser.parseHead(DefaultHttpResponseParser.java:143)
at org.apache.http.impl.conn.DefaultHttpResponseParser.parseHead(DefaultHttpResponseParser.java:57)
at org.apache.http.impl.io.AbstractMessageParser.parse(AbstractMessageParser.java:261)
at org.apache.http.impl.DefaultBHttpClientConnection.receiveResponseHeader(DefaultBHttpClientConnection.java:165)
at org.apache.http.impl.conn.CPoolProxy.receiveResponseHeader(CPoolProxy.java:167)
at org.apache.http.protocol.HttpRequestExecutor.doReceiveResponse(HttpRequestExecutor.java:272)
at org.apache.http.protocol.HttpRequestExecutor.execute(HttpRequestExecutor.java:124)
at org.apache.http.impl.execchain.MainClientExec.execute(MainClientExec.java:271)
at org.apache.http.impl.execchain.ProtocolExec.execute(ProtocolExec.java:184)
at org.apache.http.impl.execchain.RetryExec.execute(RetryExec.java:88)
at org.apache.http.impl.execchain.RedirectExec.execute(RedirectExec.java:110)
at org.apache.http.impl.client.InternalHttpClient.doExecute(InternalHttpClient.java:184)
at org.apache.http.impl.client.CloseableHttpClient.execute(CloseableHttpClient.java:82)
at org.apache.http.impl.client.CloseableHttpClient.execute(CloseableHttpClient.java:107)
at org.apache.http.impl.client.CloseableHttpClient.execute(CloseableHttpClient.java:55)
at org.apache.http.client.fluent.Request.execute(Request.java:151)
at org.apache.hudi.common.table.view.RemoteHoodieTableFileSystemView.executeRequest(RemoteHoodieTableFileSystemView.java:176)
at org.apache.hudi.common.table.view.RemoteHoodieTableFileSystemView.refresh(RemoteHoodieTableFileSystemView.java:418)
... 24 more

@danny0405
Copy link
Contributor

Did you try the 0.10.1 already ?

@Chosen123Wang
Copy link
Author

Did you try the 0.10.1 already ?

not yet, just used 0.11.0

@danny0405
Copy link
Contributor

See the NetworkUtils the ip address seems can not work correctly for you local env.

@codope codope added the flink Issues related to flink label Jun 16, 2022
@codope
Copy link
Member

codope commented Jun 16, 2022

@Chosen123Wang did you check based on above suggestion by Danny?

@codope codope added the priority:major degraded perf; unable to move forward; potential bugs label Jun 17, 2022
@codope
Copy link
Member

codope commented Jun 17, 2022

@danny0405 @LinMingQiang Would #5884 be helpful here?

@LinMingQiang
Copy link
Contributor

@danny0405 @LinMingQiang Would #5884 be helpful here?

I'm afraid not, The current issue problem is that the network is unavailable and cannot be retried

@zhangty123321
Copy link

Excuse me, is there a solution to this problem? This problem occurs almost every time we use Hudi to read full data.

@danny0405
Copy link
Contributor

What do you mean by 'read full data' ?

@zhangty123321
Copy link

Hello, I mean reading the full amount of data

@danny0405
Copy link
Contributor

Hello, I mean reading the full amount of data

To read from the hudi table or write into it ?

@zhangty123321
Copy link

Read the data in full via flink cdc and write it to hudi

@zhangty123321
Copy link

Hello, I mean reading the full amount of data

To read from the hudi table or write into it ?
This error is almost a necessity, we tried to reduce the size of the table, and it still appeared

@danny0405
Copy link
Contributor

Read the data in full via flink cdc and write it to hudi

What kind of cdc source did you use ?

@nsivabalan
Copy link
Contributor

@zhangty123321 : can you give us steps to reproduce. full writer logs etc.

@xushiyan xushiyan moved this to 👤 User Action in Hudi Issue Support Oct 30, 2022
@danny0405
Copy link
Contributor

Try the new option hoodie.filesystem.view.remote.retry.enable which is introduced in release 0.12.1.

Repository owner moved this from 👤 User Action to ✅ Done in Hudi Issue Support Nov 8, 2022
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
flink Issues related to flink priority:major degraded perf; unable to move forward; potential bugs
Projects
Archived in project
Development

No branches or pull requests

6 participants