Amarjeet Singh
Amarjeet Singh
I think so We need it partial Aware Scheduling here. Let say we are dependent on 3 dataset. Let say we update each dataset every hour. And we want one...
@ad1happy2go - Is there any other way to do it on hudi 0.12.3 like I am trying to use config hoodie.combine.before.delete setting it as false, or any other config
@ad1happy2go - Do you know any other way to delete duplicated record from the hudi table without rewriting whole table
@ad1happy2go - that approach worked thanks.
Hi @ad1happy2go - Please find below the configurations "hoodie.schema.on.read.enable": "true" "hoodie.cleaner.commits.retained": "3", "hoodie.datasource.write.reconcile.schema": "true", "hoodie.parquet.compression.codec": "zstd", "hoodie.delete.shuffle.parallelism": "200", "hoodie.parquet.max.file.size": "268435456", "hoodie.upsert.shuffle.parallelism": "200", "hoodie.datasource.hive_sync.support_timestamp": "true", "hoodie.datasource.write.keygenerator.class": "org.apache.hudi.keygen.CustomKeyGenerator", "hoodie.datasource.write.hive_style_partitioning": "true", "hoodie.insert.shuffle.parallelism": "200",...
@ad1happy2go - Were you able to recreate the issue? If you want we can have a meeting to create an issue.
hi @ad1happy2go - In my case the table is in long and changed to double.
Hi @ad1happy2go - I have found the issue is in Syncing of Table in Catalog, Since I am using GLue Catalog. But when I tried creating a table using the...
@ad1happy2go - Any updated on this.
@ad1happy2go - Any update on this?