Aditya Goenka

Results 304 comments of Aditya Goenka

Sorry, Looks like you are using the AWS managed hudi. Can you try using emr-6.15.0 which has hudi 0.14.0

Great! Thanks lot @huliwuli. Closing out this issue then. Please reopen in case you have any concerns.

@Amar1404 With spark, Did you tried to give config along with write.df. - .option("parquet.compression.codec.zstd.level", "22")

@Amar1404 Were you able to check this, does this work?

@njalan What do you mean by "copied one partition file from same table". Are you referring copying the parquet files?

but partition directory only contains the parquet files AND log files (in case of MOR). Right? If you just copy partition files, how you are updating the .hoodie timeline?

@njalan No there is no way and we dont recommend also. Best way is to instead of moving use spark to write code and create another Hudi Table with partitions...

@Amar1404 Can you please try 0.14.1. This was fixed. I tried below code also to demonstrate - ``` DROP TABLE issue_11212; set hoodie.spark.sql.insert.into.operation=bulk_insert; CREATE TABLE issue_11212 ( ts BIGINT, uuid...

@Amar1404 With 0.12 we always used to delete records based on record key. That is the reason both of those records are getting filtered out. One way is to identify...

@Amar1404 Did the approach worked? Do you need any other help here?