Ranga Reddy

Results 87 comments of Ranga Reddy

Hi @Dakarun Implementing CDC on existing tables is not directly feasible. The recommended approach is to create a new table based on the existing data, then drop the original table...

Hi @Dakarun To migrate a non-CDC table to a CDC table, please follow these steps: 1. Create a new Hudi table with `hoodie.table.cdc.enabled=true`. 2. Copy (rewrite) the data from the...

Hi @njalan In order to get the primary key, pre-combined key or partition key we need to read the `/.hoodie/hoodie.properties` file. **Example:** ```properties hoodie.table.precombine.field=comb hoodie.table.partition.fields=par hoodie.table.recordkey.fields=id ``` Please let me...

Hi @RYiHui Have you tried reading the Hudi table data using the latest version of Trino?

Hi @stream2000 I am able to reproduce this issue while running the following spark code. Needs to test the Hudi test case. ```sh spark-shell \ --conf 'spark.serializer=org.apache.spark.serializer.KryoSerializer' \ --conf 'spark.sql.catalog.spark_catalog=org.apache.spark.sql.hudi.catalog.HoodieCatalog'...

**Spark with Hive** ```scala spark.sql( s""" |CREATE TABLE hive_test_null_partition ( | id INT, | boolean_field BOOLEAN, | float_field FLOAT, | byte_field BYTE, | short_field SHORT, | decimal_field DECIMAL(10, 5), |...

Created upstream jira for this issue - https://issues.apache.org/jira/browse/HUDI-8315

Hi @ligou525 I don't believe gzip compression is causing this issue; it's more likely that data corruption has occurred. Could you please share sample reproducible code to replicate this issue?

Hi @ligou525 It's great to know that Snappy compression solved your problem. Since I couldn't find any specific JIRA tickets about Gzip-related data loss, I'd suggest that if the application...

Hi @bithw1 I have confirmed that the **HoodieCleanMetadata** classes are present in both **Hudi 0.15.0** and **Hudi 1.0.0** jars. Could you please retry with following steps mentioned for 1.0.0 release....