-
Notifications
You must be signed in to change notification settings - Fork 2.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[SUPPORT] Multiple chaining of hudi tables via incremental source results in duplicate partition meta column #5189
Comments
|
@harsh1231 : Can you take a look at this issue |
@bvaradar can you please share some context on why can't we delete hudi/hudi-utilities/src/main/java/org/apache/hudi/utilities/sources/HoodieIncrSource.java Line 176 in b28f0d6
|
@harsh1231 : in the mean time (until @bvaradar responds), can you investigate as to why we are encountering duplicate issue. |
@nsivabalan @lowmmrfeeder load from previous table itself is failing |
@harsh1231 I am not sure of this, we aren't using columnname hudi/hudi-utilities/src/main/java/org/apache/hudi/utilities/sources/HoodieIncrSource.java Line 176 in b28f0d6
|
@lowmmrfeeder Got it , got distracted because of private package |
Ohh ya, we did little modification to HoodieIncrSource.java and created our own class. The modifications were related to reading whole source table when doing incremental read with deltastreamer for first time ( Its supported now in |
@harsh1231 Have you made any progress on reproducing the problem? |
@harsh1231 : can you spend some time on this. would be good to get to the bottom of this. |
I could able to reproduce this. will try to put in a fix by this weekend. |
Tips before filing an issue
Describe the problem you faced
From user:
I am trying to read a hoodie table and write to a hoodie table using delta streamer and I am getting this error:
Steps to reproduce:
stacktrace:
write configs:
Expected behavior
A clear and concise description of what you expected to happen.
Environment Description
Hudi version :
Spark version :
Hive version :
Hadoop version :
Storage (HDFS/S3/GCS..) :
Running on Docker? (yes/no) :
Additional context
Add any other context about the problem here.
Stacktrace
Add the stacktrace of the error.
The text was updated successfully, but these errors were encountered: