-
Notifications
You must be signed in to change notification settings - Fork 2.5k
Closed
Labels
area:ingestIngestion into HudiIngestion into Hudiarea:writerWrite client and core write operationsWrite client and core write operationspriority:mediumModerate impact; usability gapsModerate impact; usability gaps
Description
Describe the problem you faced
we accidentally configure wrong primary key in the spark write config, it cause duplicate data. wondering if there is a way to avoid it.
To Reproduce
change the primary config in write config and run the spark job.
Expected behavior
maybe should block the spark job to write data if the primary key config is different from the primary key in the original table.
Environment Description
-
Hudi version : 0.11.0
-
Spark version : 3.2.1
-
Storage (HDFS/S3/GCS..) : GCS
-
Running on Docker? (yes/no) : yes. on k8s.
Reactions are currently unavailable
Metadata
Metadata
Assignees
Labels
area:ingestIngestion into HudiIngestion into Hudiarea:writerWrite client and core write operationsWrite client and core write operationspriority:mediumModerate impact; usability gapsModerate impact; usability gaps
Type
Projects
Status
✅ Done