-
Notifications
You must be signed in to change notification settings - Fork 2.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[SUPPORT] [BUG] Duplicate fileID ??? from bucket ?? of partition found during the BucketStreamWriteFunction index bootstrap. #5330
Comments
cc @danny0405 |
see #5185 |
Thanks for the PR @wxplovecc , can you explain why the #5185 patch fixed the bug ? |
@Guanpx Did you use the code of latest master ? The shuffle strategy in Line 44 in b8e465f
Do you guys have any idea how the error happens ? |
I will try latest master tomorrow and debug that code ~ |
The mor table rollback was not delete the log files , If the job failed before first successful commit it will left some invalid |
@Guanpx have you tried the latest master and see if the fix solves the problem for you? |
@Guanpx : do you have any updates on this issue. if latest master works, feel free to close out the github issue. |
see #5763 |
closing the issue since the linked PR is landed. feel free to reopen or open a new issue if you are still facing issues. |
Seeing this issue with Flink 1.15 and Hudi 0.12.3:
Steps to reproduce and general setup the same as above. Looking for any suggestions, thank you. |
Did you write to an exiting table or a fresh new table then? |
This was a fresh table. Flink had restarted a few times due to a memory issue. It looks like this occured after that. |
Did you use the COW table or MOR? |
We are using COW. |
Describe the problem you faced
use flink1.13 ,bucket index , cow ,hudi-0.11.0(not latest)
To Reproduce
Steps to reproduce the behavior:
Environment Description
Hudi version : 0.11.0
Flink version : 1.13.2
Hadoop version : 3.0.0
Storage (HDFS/S3/GCS..) :HDFS
Running on Docker? (yes/no) : no
Additional context
that 0000007-f0da file should same as before bucket files, format : 0000007-b513-xxxx
Stacktrace
The text was updated successfully, but these errors were encountered: