SEZ
SEZ
想问,这个问题目前有解决思路了吗。 目前在实践mysql -flink cdc 增量数据同步过程中,因为一个库要同步40多张表,每个表一个cdc链接,发现都是独立的线程去执行binlog dump。这样导致binlog被重复读取,导致mysql生产压力大。目前有个两个疑惑: 1、单张表的cdc订阅是否也是从头拉取全库的binlog文件进行解析; 2、如果cdc统一只开一个任务订阅binlog ,可以不指定表结构,只把数据全部同步到kafka,再后端再去做解析,数据上能区分库和表?是否可行 
我这边也遇到了GTID 的问题,启动cdc任务,立刻报了个错,查阅了debezium文档 还没找到怎么指定gtid消费。。。有大佬可以指明一下吗。。。。 org.apache.kafka.connect.errors.ConnectException: The replication sender thread cannot start in AUTO_POSITION mode: this server has GTID_MODE = ON_PERMISSIVE instead of ON. Error code: 1236; SQLSTATE: HY000. at io.debezium.connector.mysql.AbstractReader.wrap(AbstractReader.java:230) at...
> @SEZ9 这个错误的意思是你的 server 没有开启 GTID。 GTID_MODE 需要设置为 ON 疑惑~~~ 我们测试环境一直都是OFF,今天产线发现是配置的ON_PERMISSIVE,然后就这样了。。
Is anyone working on this feature? We are preparing to make some contributions. We have been working with Apache Ranger and already have a version of the deployment guide, which...
> Hi @SEZ9, I no longer use Apache Ranger as a data security solution. As the workload I deal with is moving towards Trino for both Adhoc and [ETL](https://trino.io/blog/2022/05/05/tardigrade-launch.html), the...