datax
文章平均质量分 60
datax
冷漠;
我很懒,还没有添加简介
展开
-
DataX同步Hive数据丢失,源码修复
DataX同步Hive数据丢失,源码修复原创 2023-04-19 17:00:59 · 236 阅读 · 0 评论 -
ClickHouse写入常见问题: Too many parts (300). Merges are processing significantly slower than inserts.
ClickHouse exception, code: 1002, host: 10.129.170.80, port: 8123; Code: 252. DB::Exception: Too many parts (300). Merges are processing significantly slower than inserts. (TOO_MANY_PARTS) (version 22.3.2.1)原创 2023-02-02 17:56:36 · 5570 阅读 · 4 评论 -
Datax往ClickHouse数据库插入数据前执行truncate table遇到超时问题Read timed out
该问题是由于数据量过大(也有可能是网络问题),处理请求太久,导致 ClickHouse 连接超时,建议在 ClickHouse 链接字符串后面加上参数:`?socket_timeout=600000`原创 2023-01-11 10:52:14 · 2517 阅读 · 0 评论 -
限制Spark往HDFS写出数据时,生成_SUCCESS文件
限制Spark/SparkSQL往HDFS写出数据时,生成_SUCCESS文件原创 2022-10-28 15:48:22 · 2126 阅读 · 0 评论 -
datax同步数据到ClickHouse时同步时间特别长,原因:Too many partitions for single INSERT block (more than 100).
ClickHouse exception, code: 1002, host: 10.129.170.80, port: 8123; Code: 252. DB::Exception: Too many partitions for single INSERT block (more than 100). The limit is controlled by 'max_partitions_per_insert_block' setting.原创 2022-09-28 18:34:00 · 2364 阅读 · 0 评论