alan_rodriguez
alan_rodriguez
Ok, let me have a try, and we have a spark job, needs to read base data & event data, then join them to generate new dataset. Base data are...
I disable metadata in reader side, so the reader job just treat them as normal parquet files and ignore reading metadata. Meanwhile, I disable metadata in writer side too, after...
[spark-hudi.log](https://github.com/apache/hudi/files/14753409/spark-hudi.log) I got more details for this error, actually this error occurs after this rollback error. But from log, I could not see any error before rollback occurs. Hope it...
Root cause is: rollback and async clean, two actions need to delete same file, then "dead lock" happens. Disable async cleaning works.
> @michael1991 Thanks for identifying the root cause. Do you have a fix in your mind. Created tracking jira for the same - https://issues.apache.org/jira/browse/HUDI-7560 > > Are you using spark...
As @xicm mentioned, lock is needed for async table service, close issue then.
I used Hudi0.14.1 on Dataproc2.1(Spark3.3.2 Hadoop3.3.6) to upsert Bloom indexed COW table with PartialUpdateAvroPayload, got same error on reading MDT bloomfilters partition hfiles. Missing some jars or not, how to...
Hi @ad1happy2go , glad to hear you again ~ Can you try column name with underscore, i'm not sure if enable urlencode for partition and partition column name with underscore...
@ad1happy2go Partitions are hours, for example, gs://bucket/tables/hudi/r_date=2024-06-17/r_hour=00. But problem only occurs on two partitions and underscore, we are using one partition column like yyyyMMddHH and it's going on well. Not...
在用ClashX Meta替代