This is an automated email from the ASF dual-hosted git repository.
liaoxin pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/doris-website.git
The following commit(s) were added to refs/heads/master by this push:
new dc2932a4403 [doc](log analysis) update log-analysis example (#2404)
dc2932a4403 is described below
commit dc2932a4403bb430055641ab2b2498345fd2bc82
Author: hui lai <[email protected]>
AuthorDate: Fri May 23 11:14:28 2025 +0800
[doc](log analysis) update log-analysis example (#2404)
---
blog/log-analysis-elasticsearch-vs-apache-doris.md | 8 ++++----
docs/log-storage-analysis.md | 18 ++++++++----------
docs/observability/log.md | 18 ++++++++----------
.../current/log-storage-analysis.md | 18 ++++++++----------
.../current/observability/log-storage-analysis.md | 18 ++++++++----------
.../current/observability/log.md | 18 ++++++++----------
.../version-2.0/log-storage-analysis.md | 18 ++++++++----------
.../version-2.1/log-storage-analysis.md | 16 +++++++---------
.../version-2.1/observability/log-storage-analysis.md | 16 +++++++---------
.../version-2.1/observability/log.md | 16 +++++++---------
.../version-3.0/log-storage-analysis.md | 18 ++++++++----------
.../version-3.0/observability/log-storage-analysis.md | 16 +++++++---------
.../version-3.0/observability/log.md | 18 ++++++++----------
.../practical-guide/log-storage-analysis.md | 18 ++++++++----------
versioned_docs/version-2.1/log-storage-analysis.md | 16 +++++++---------
versioned_docs/version-2.1/observability/log.md | 18 ++++++++----------
.../practical-guide/log-storage-analysis.md | 18 ++++++++----------
versioned_docs/version-3.0/log-storage-analysis.md | 18 ++++++++----------
versioned_docs/version-3.0/observability/log.md | 18 ++++++++----------
.../practical-guide/log-storage-analysis.md | 18 ++++++++----------
20 files changed, 151 insertions(+), 189 deletions(-)
diff --git a/blog/log-analysis-elasticsearch-vs-apache-doris.md
b/blog/log-analysis-elasticsearch-vs-apache-doris.md
index a8c2bc026fc..890488c58c2 100644
--- a/blog/log-analysis-elasticsearch-vs-apache-doris.md
+++ b/blog/log-analysis-elasticsearch-vs-apache-doris.md
@@ -233,10 +233,10 @@ For JSON logs that are written into Kafka message queues,
create [Routine Load](
CREATE ROUTINE LOAD load_log_kafka ON log_db.log_table
COLUMNS(ts, clientip, request, status, size)
PROPERTIES (
- "max_batch_interval" = "10",
- "max_batch_rows" = "1000000",
- "max_batch_size" = "109715200",
- "strict_mode" = "false",
+ "max_batch_interval" = "60",
+ "max_batch_rows" = "20000000",
+ "max_batch_size" = "1073741824",
+ "load_to_single_tablet" = "true",
"format" = "json"
)
FROM KAFKA (
diff --git a/docs/log-storage-analysis.md b/docs/log-storage-analysis.md
index 258ebe170cd..332e70570b3 100644
--- a/docs/log-storage-analysis.md
+++ b/docs/log-storage-analysis.md
@@ -476,15 +476,13 @@ You can refer to the example below, where `property.*`
represents Librdkafka cli
```SQL
CREATE ROUTINE LOAD load_log_kafka ON log_db.log_table
COLUMNS(ts, clientip, request, status, size)
-PROPERTIES (
-"max_batch_interval" = "10",
-"max_batch_rows" = "1000000",
-"max_batch_size" = "109715200",
-"load_to_single_tablet" = "true",
-"timeout" = "600",
-"strict_mode" = "false",
-"format" = "json"
-)
+PROPERTIES (
+"max_batch_interval" = "60",
+"max_batch_rows" = "20000000",
+"max_batch_size" = "1073741824",
+"load_to_single_tablet" = "true",
+"format" = "json"
+)
FROM KAFKA (
"kafka_broker_list" = "host:port",
"kafka_topic" = "log__topic_",
@@ -498,7 +496,7 @@ FROM KAFKA (
<br />SHOW ROUTINE LOAD;
```
-For more information about Kafka, see [Routine
Load](./data-operate/import/import-way/routine-load-manual).
+For more information about Kafka, see [Routine
Load](./data-operate/import/import-way/routine-load-manual.md).
**Using customized programs to collect logs**
diff --git a/docs/observability/log.md b/docs/observability/log.md
index a244f95a67b..e38f35ee2b8 100644
--- a/docs/observability/log.md
+++ b/docs/observability/log.md
@@ -392,15 +392,13 @@ You can refer to the example below, where `property.*`
represents Librdkafka cli
```SQL
CREATE ROUTINE LOAD load_log_kafka ON log_db.log_table
COLUMNS(ts, clientip, request, status, size)
-PROPERTIES (
-"max_batch_interval" = "10",
-"max_batch_rows" = "1000000",
-"max_batch_size" = "109715200",
-"load_to_single_tablet" = "true",
-"timeout" = "600",
-"strict_mode" = "false",
-"format" = "json"
-)
+PROPERTIES (
+"max_batch_interval" = "60",
+"max_batch_rows" = "20000000",
+"max_batch_size" = "1073741824",
+"load_to_single_tablet" = "true",
+"format" = "json"
+)
FROM KAFKA (
"kafka_broker_list" = "host:port",
"kafka_topic" = "log__topic_",
@@ -414,7 +412,7 @@ FROM KAFKA (
<br />SHOW ROUTINE LOAD;
```
-For more information about Kafka, see [Routine
Load](../../data-operate/import/import-way/routine-load-manual.md).
+For more information about Kafka, see [Routine
Load](../data-operate/import/import-way/routine-load-manual.md).
**Using customized programs to collect logs**
diff --git
a/i18n/zh-CN/docusaurus-plugin-content-docs/current/log-storage-analysis.md
b/i18n/zh-CN/docusaurus-plugin-content-docs/current/log-storage-analysis.md
index 78e6bd25011..1db46f0d6c4 100644
--- a/i18n/zh-CN/docusaurus-plugin-content-docs/current/log-storage-analysis.md
+++ b/i18n/zh-CN/docusaurus-plugin-content-docs/current/log-storage-analysis.md
@@ -455,15 +455,13 @@ chmod +x filebeat-doris-1.0.0
-- 创建 routine load,从 kafka log__topic_将数据导入 log_table 表
CREATE ROUTINE LOAD load_log_kafka ON log_db.log_table
COLUMNS(ts, clientip, request, status, size)
-PROPERTIES (
-"max_batch_interval" = "10",
-"max_batch_rows" = "1000000",
-"max_batch_size" = "109715200",
-"load_to_single_tablet" = "true",
-"timeout" = "600",
-"strict_mode" = "false",
-"format" = "json"
-)
+PROPERTIES (
+"max_batch_interval" = "60",
+"max_batch_rows" = "20000000",
+"max_batch_size" = "1073741824",
+"load_to_single_tablet" = "true",
+"format" = "json"
+)
FROM KAFKA (
"kafka_broker_list" = "host:port",
"kafka_topic" = "log__topic_",
@@ -478,7 +476,7 @@ FROM KAFKA (
SHOW ROUTINE LOAD;
```
-更多关于 Kafka 配置和使用的说明,可参考 [Routine
Load](./data-operate/import/import-way/routine-load-manual)。
+更多关于 Kafka 配置和使用的说明,可参考 [Routine
Load](./data-operate/import/import-way/routine-load-manual.md)。
**使用自定义程序采集日志**
diff --git
a/i18n/zh-CN/docusaurus-plugin-content-docs/current/observability/log-storage-analysis.md
b/i18n/zh-CN/docusaurus-plugin-content-docs/current/observability/log-storage-analysis.md
index 14d41592983..7686bfb1ae7 100644
---
a/i18n/zh-CN/docusaurus-plugin-content-docs/current/observability/log-storage-analysis.md
+++
b/i18n/zh-CN/docusaurus-plugin-content-docs/current/observability/log-storage-analysis.md
@@ -374,15 +374,13 @@ chmod +x filebeat-doris-1.0.0
-- 创建 routine load,从 kafka log__topic_将数据导入 log_table 表
CREATE ROUTINE LOAD load_log_kafka ON log_db.log_table
COLUMNS(ts, clientip, request, status, size)
-PROPERTIES (
-"max_batch_interval" = "10",
-"max_batch_rows" = "1000000",
-"max_batch_size" = "109715200",
-"load_to_single_tablet" = "true",
-"timeout" = "600",
-"strict_mode" = "false",
-"format" = "json"
-)
+PROPERTIES (
+"max_batch_interval" = "60",
+"max_batch_rows" = "20000000",
+"max_batch_size" = "1073741824",
+"load_to_single_tablet" = "true",
+"format" = "json"
+)
FROM KAFKA (
"kafka_broker_list" = "host:port",
"kafka_topic" = "log__topic_",
@@ -397,7 +395,7 @@ FROM KAFKA (
SHOW ROUTINE LOAD;
```
-更多关于 Kafka 配置和使用的说明,可参考 [Routine
Load](./data-operate/import/import-way/routine-load-manual)。
+更多关于 Kafka 配置和使用的说明,可参考 [Routine
Load](../data-operate/import/import-way/routine-load-manual.md)。
**使用自定义程序采集日志**
diff --git
a/i18n/zh-CN/docusaurus-plugin-content-docs/current/observability/log.md
b/i18n/zh-CN/docusaurus-plugin-content-docs/current/observability/log.md
index efb467875b1..ce677e142c7 100644
--- a/i18n/zh-CN/docusaurus-plugin-content-docs/current/observability/log.md
+++ b/i18n/zh-CN/docusaurus-plugin-content-docs/current/observability/log.md
@@ -384,15 +384,13 @@ chmod +x filebeat-doris-7.17.5.4
-- 创建 routine load,从 kafka log__topic_将数据导入 log_table 表
CREATE ROUTINE LOAD load_log_kafka ON log_db.log_table
COLUMNS(ts, clientip, request, status, size)
-PROPERTIES (
-"max_batch_interval" = "10",
-"max_batch_rows" = "1000000",
-"max_batch_size" = "109715200",
-"load_to_single_tablet" = "true",
-"timeout" = "600",
-"strict_mode" = "false",
-"format" = "json"
-)
+PROPERTIES (
+"max_batch_interval" = "60",
+"max_batch_rows" = "20000000",
+"max_batch_size" = "1073741824",
+"load_to_single_tablet" = "true",
+"format" = "json"
+)
FROM KAFKA (
"kafka_broker_list" = "host:port",
"kafka_topic" = "log__topic_",
@@ -407,7 +405,7 @@ FROM KAFKA (
SHOW ROUTINE LOAD;
```
-更多关于 Kafka 配置和使用的说明,可参考 [Routine
Load](./data-operate/import/import-way/routine-load-manual)。
+更多关于 Kafka 配置和使用的说明,可参考 [Routine
Load](../data-operate/import/import-way/routine-load-manual.md)。
**使用自定义程序采集日志**
diff --git
a/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.0/log-storage-analysis.md
b/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.0/log-storage-analysis.md
index 80a0c89905a..87aef38492c 100644
---
a/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.0/log-storage-analysis.md
+++
b/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.0/log-storage-analysis.md
@@ -455,15 +455,13 @@ chmod +x filebeat-doris-1.0.0
-- 创建 routine load,从 kafka log__topic_将数据导入 log_table 表
CREATE ROUTINE LOAD load_log_kafka ON log_db.log_table
COLUMNS(ts, clientip, request, status, size)
-PROPERTIES (
-"max_batch_interval" = "10",
-"max_batch_rows" = "1000000",
-"max_batch_size" = "109715200",
-"load_to_single_tablet" = "true",
-"timeout" = "600",
-"strict_mode" = "false",
-"format" = "json"
-)
+PROPERTIES (
+"max_batch_interval" = "60",
+"max_batch_rows" = "20000000",
+"max_batch_size" = "1073741824",
+"load_to_single_tablet" = "true",
+"format" = "json"
+)
FROM KAFKA (
"kafka_broker_list" = "host:port",
"kafka_topic" = "log__topic_",
@@ -478,7 +476,7 @@ FROM KAFKA (
SHOW ROUTINE LOAD;
```
-更多关于 Kafka 配置和使用的说明,可参考 [Routine
Load](./data-operate/import/import-way/routine-load-manual)。
+更多关于 Kafka 配置和使用的说明,可参考 [Routine
Load](./data-operate/import/routine-load-manual.md)。
**使用自定义程序采集日志**
diff --git
a/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1/log-storage-analysis.md
b/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1/log-storage-analysis.md
index 14d41592983..e9b12f32ba5 100644
---
a/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1/log-storage-analysis.md
+++
b/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1/log-storage-analysis.md
@@ -374,14 +374,12 @@ chmod +x filebeat-doris-1.0.0
-- 创建 routine load,从 kafka log__topic_将数据导入 log_table 表
CREATE ROUTINE LOAD load_log_kafka ON log_db.log_table
COLUMNS(ts, clientip, request, status, size)
-PROPERTIES (
-"max_batch_interval" = "10",
-"max_batch_rows" = "1000000",
-"max_batch_size" = "109715200",
-"load_to_single_tablet" = "true",
-"timeout" = "600",
-"strict_mode" = "false",
-"format" = "json"
+PROPERTIES (
+"max_batch_interval" = "60",
+"max_batch_rows" = "20000000",
+"max_batch_size" = "1073741824",
+"load_to_single_tablet" = "true",
+"format" = "json"
)
FROM KAFKA (
"kafka_broker_list" = "host:port",
@@ -397,7 +395,7 @@ FROM KAFKA (
SHOW ROUTINE LOAD;
```
-更多关于 Kafka 配置和使用的说明,可参考 [Routine
Load](./data-operate/import/import-way/routine-load-manual)。
+更多关于 Kafka 配置和使用的说明,可参考 [Routine
Load](./data-operate/import/import-way/routine-load-manual.md)。
**使用自定义程序采集日志**
diff --git
a/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1/observability/log-storage-analysis.md
b/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1/observability/log-storage-analysis.md
index 14d41592983..a55010f09b0 100644
---
a/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1/observability/log-storage-analysis.md
+++
b/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1/observability/log-storage-analysis.md
@@ -374,14 +374,12 @@ chmod +x filebeat-doris-1.0.0
-- 创建 routine load,从 kafka log__topic_将数据导入 log_table 表
CREATE ROUTINE LOAD load_log_kafka ON log_db.log_table
COLUMNS(ts, clientip, request, status, size)
-PROPERTIES (
-"max_batch_interval" = "10",
-"max_batch_rows" = "1000000",
-"max_batch_size" = "109715200",
-"load_to_single_tablet" = "true",
-"timeout" = "600",
-"strict_mode" = "false",
-"format" = "json"
+PROPERTIES (
+"max_batch_interval" = "60",
+"max_batch_rows" = "20000000",
+"max_batch_size" = "1073741824",
+"load_to_single_tablet" = "true",
+"format" = "json"
)
FROM KAFKA (
"kafka_broker_list" = "host:port",
@@ -397,7 +395,7 @@ FROM KAFKA (
SHOW ROUTINE LOAD;
```
-更多关于 Kafka 配置和使用的说明,可参考 [Routine
Load](./data-operate/import/import-way/routine-load-manual)。
+更多关于 Kafka 配置和使用的说明,可参考 [Routine
Load](../data-operate/import/import-way/routine-load-manual.md)。
**使用自定义程序采集日志**
diff --git
a/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1/observability/log.md
b/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1/observability/log.md
index efb467875b1..4fd8cb9773e 100644
--- a/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1/observability/log.md
+++ b/i18n/zh-CN/docusaurus-plugin-content-docs/version-2.1/observability/log.md
@@ -384,14 +384,12 @@ chmod +x filebeat-doris-7.17.5.4
-- 创建 routine load,从 kafka log__topic_将数据导入 log_table 表
CREATE ROUTINE LOAD load_log_kafka ON log_db.log_table
COLUMNS(ts, clientip, request, status, size)
-PROPERTIES (
-"max_batch_interval" = "10",
-"max_batch_rows" = "1000000",
-"max_batch_size" = "109715200",
-"load_to_single_tablet" = "true",
-"timeout" = "600",
-"strict_mode" = "false",
-"format" = "json"
+PROPERTIES (
+"max_batch_interval" = "60",
+"max_batch_rows" = "20000000",
+"max_batch_size" = "1073741824",
+"load_to_single_tablet" = "true",
+"format" = "json"
)
FROM KAFKA (
"kafka_broker_list" = "host:port",
@@ -407,7 +405,7 @@ FROM KAFKA (
SHOW ROUTINE LOAD;
```
-更多关于 Kafka 配置和使用的说明,可参考 [Routine
Load](./data-operate/import/import-way/routine-load-manual)。
+更多关于 Kafka 配置和使用的说明,可参考 [Routine
Load](../data-operate/import/import-way/routine-load-manual.md)。
**使用自定义程序采集日志**
diff --git
a/i18n/zh-CN/docusaurus-plugin-content-docs/version-3.0/log-storage-analysis.md
b/i18n/zh-CN/docusaurus-plugin-content-docs/version-3.0/log-storage-analysis.md
index 14d41592983..29bb1556963 100644
---
a/i18n/zh-CN/docusaurus-plugin-content-docs/version-3.0/log-storage-analysis.md
+++
b/i18n/zh-CN/docusaurus-plugin-content-docs/version-3.0/log-storage-analysis.md
@@ -374,15 +374,13 @@ chmod +x filebeat-doris-1.0.0
-- 创建 routine load,从 kafka log__topic_将数据导入 log_table 表
CREATE ROUTINE LOAD load_log_kafka ON log_db.log_table
COLUMNS(ts, clientip, request, status, size)
-PROPERTIES (
-"max_batch_interval" = "10",
-"max_batch_rows" = "1000000",
-"max_batch_size" = "109715200",
-"load_to_single_tablet" = "true",
-"timeout" = "600",
-"strict_mode" = "false",
-"format" = "json"
-)
+PROPERTIES (
+"max_batch_interval" = "60",
+"max_batch_rows" = "20000000",
+"max_batch_size" = "1073741824",
+"load_to_single_tablet" = "true",
+"format" = "json"
+)
FROM KAFKA (
"kafka_broker_list" = "host:port",
"kafka_topic" = "log__topic_",
@@ -397,7 +395,7 @@ FROM KAFKA (
SHOW ROUTINE LOAD;
```
-更多关于 Kafka 配置和使用的说明,可参考 [Routine
Load](./data-operate/import/import-way/routine-load-manual)。
+更多关于 Kafka 配置和使用的说明,可参考 [Routine
Load](./data-operate/import/import-way/routine-load-manual.md)。
**使用自定义程序采集日志**
diff --git
a/i18n/zh-CN/docusaurus-plugin-content-docs/version-3.0/observability/log-storage-analysis.md
b/i18n/zh-CN/docusaurus-plugin-content-docs/version-3.0/observability/log-storage-analysis.md
index 14d41592983..a55010f09b0 100644
---
a/i18n/zh-CN/docusaurus-plugin-content-docs/version-3.0/observability/log-storage-analysis.md
+++
b/i18n/zh-CN/docusaurus-plugin-content-docs/version-3.0/observability/log-storage-analysis.md
@@ -374,14 +374,12 @@ chmod +x filebeat-doris-1.0.0
-- 创建 routine load,从 kafka log__topic_将数据导入 log_table 表
CREATE ROUTINE LOAD load_log_kafka ON log_db.log_table
COLUMNS(ts, clientip, request, status, size)
-PROPERTIES (
-"max_batch_interval" = "10",
-"max_batch_rows" = "1000000",
-"max_batch_size" = "109715200",
-"load_to_single_tablet" = "true",
-"timeout" = "600",
-"strict_mode" = "false",
-"format" = "json"
+PROPERTIES (
+"max_batch_interval" = "60",
+"max_batch_rows" = "20000000",
+"max_batch_size" = "1073741824",
+"load_to_single_tablet" = "true",
+"format" = "json"
)
FROM KAFKA (
"kafka_broker_list" = "host:port",
@@ -397,7 +395,7 @@ FROM KAFKA (
SHOW ROUTINE LOAD;
```
-更多关于 Kafka 配置和使用的说明,可参考 [Routine
Load](./data-operate/import/import-way/routine-load-manual)。
+更多关于 Kafka 配置和使用的说明,可参考 [Routine
Load](../data-operate/import/import-way/routine-load-manual.md)。
**使用自定义程序采集日志**
diff --git
a/i18n/zh-CN/docusaurus-plugin-content-docs/version-3.0/observability/log.md
b/i18n/zh-CN/docusaurus-plugin-content-docs/version-3.0/observability/log.md
index efb467875b1..ce677e142c7 100644
--- a/i18n/zh-CN/docusaurus-plugin-content-docs/version-3.0/observability/log.md
+++ b/i18n/zh-CN/docusaurus-plugin-content-docs/version-3.0/observability/log.md
@@ -384,15 +384,13 @@ chmod +x filebeat-doris-7.17.5.4
-- 创建 routine load,从 kafka log__topic_将数据导入 log_table 表
CREATE ROUTINE LOAD load_log_kafka ON log_db.log_table
COLUMNS(ts, clientip, request, status, size)
-PROPERTIES (
-"max_batch_interval" = "10",
-"max_batch_rows" = "1000000",
-"max_batch_size" = "109715200",
-"load_to_single_tablet" = "true",
-"timeout" = "600",
-"strict_mode" = "false",
-"format" = "json"
-)
+PROPERTIES (
+"max_batch_interval" = "60",
+"max_batch_rows" = "20000000",
+"max_batch_size" = "1073741824",
+"load_to_single_tablet" = "true",
+"format" = "json"
+)
FROM KAFKA (
"kafka_broker_list" = "host:port",
"kafka_topic" = "log__topic_",
@@ -407,7 +405,7 @@ FROM KAFKA (
SHOW ROUTINE LOAD;
```
-更多关于 Kafka 配置和使用的说明,可参考 [Routine
Load](./data-operate/import/import-way/routine-load-manual)。
+更多关于 Kafka 配置和使用的说明,可参考 [Routine
Load](../data-operate/import/import-way/routine-load-manual.md)。
**使用自定义程序采集日志**
diff --git a/versioned_docs/version-2.0/practical-guide/log-storage-analysis.md
b/versioned_docs/version-2.0/practical-guide/log-storage-analysis.md
index bddb46de306..7dc10c4fded 100644
--- a/versioned_docs/version-2.0/practical-guide/log-storage-analysis.md
+++ b/versioned_docs/version-2.0/practical-guide/log-storage-analysis.md
@@ -465,15 +465,13 @@ You can refer to the example below, where `property.*`
represents Librdkafka cli
```SQL
CREATE ROUTINE LOAD load_log_kafka ON log_db.log_table
COLUMNS(ts, clientip, request, status, size)
-PROPERTIES (
-"max_batch_interval" = "10",
-"max_batch_rows" = "1000000",
-"max_batch_size" = "109715200",
-"load_to_single_tablet" = "true",
-"timeout" = "600",
-"strict_mode" = "false",
-"format" = "json"
-)
+PROPERTIES (
+"max_batch_interval" = "60",
+"max_batch_rows" = "20000000",
+"max_batch_size" = "1073741824",
+"load_to_single_tablet" = "true",
+"format" = "json"
+)
FROM KAFKA (
"kafka_broker_list" = "host:port",
"kafka_topic" = "log__topic_",
@@ -487,7 +485,7 @@ FROM KAFKA (
<br />SHOW ROUTINE LOAD;
```
-For more information about Kafka, see [Routine
Load](../../data-operate/import/import-way/routine-load-manual.md).
+For more information about Kafka, see [Routine
Load](../data-operate/import/routine-load-manual.md).
**Using customized programs to collect logs**
diff --git a/versioned_docs/version-2.1/log-storage-analysis.md
b/versioned_docs/version-2.1/log-storage-analysis.md
index de063018b9d..f9c1a7af7f0 100644
--- a/versioned_docs/version-2.1/log-storage-analysis.md
+++ b/versioned_docs/version-2.1/log-storage-analysis.md
@@ -475,14 +475,12 @@ You can refer to the example below, where `property.*`
represents Librdkafka cli
```SQL
CREATE ROUTINE LOAD load_log_kafka ON log_db.log_table
COLUMNS(ts, clientip, request, status, size)
-PROPERTIES (
-"max_batch_interval" = "10",
-"max_batch_rows" = "1000000",
-"max_batch_size" = "109715200",
-"load_to_single_tablet" = "true",
-"timeout" = "600",
-"strict_mode" = "false",
-"format" = "json"
+PROPERTIES (
+"max_batch_interval" = "60",
+"max_batch_rows" = "20000000",
+"max_batch_size" = "1073741824",
+"load_to_single_tablet" = "true",
+"format" = "json"
)
FROM KAFKA (
"kafka_broker_list" = "host:port",
@@ -497,7 +495,7 @@ FROM KAFKA (
<br />SHOW ROUTINE LOAD;
```
-For more information about Kafka, see [Routine
Load](./data-operate/import/import-way/routine-load-manual).
+For more information about Kafka, see [Routine
Load](./data-operate/import/import-way/routine-load-manual.md).
**Using customized programs to collect logs**
diff --git a/versioned_docs/version-2.1/observability/log.md
b/versioned_docs/version-2.1/observability/log.md
index a244f95a67b..e38f35ee2b8 100644
--- a/versioned_docs/version-2.1/observability/log.md
+++ b/versioned_docs/version-2.1/observability/log.md
@@ -392,15 +392,13 @@ You can refer to the example below, where `property.*`
represents Librdkafka cli
```SQL
CREATE ROUTINE LOAD load_log_kafka ON log_db.log_table
COLUMNS(ts, clientip, request, status, size)
-PROPERTIES (
-"max_batch_interval" = "10",
-"max_batch_rows" = "1000000",
-"max_batch_size" = "109715200",
-"load_to_single_tablet" = "true",
-"timeout" = "600",
-"strict_mode" = "false",
-"format" = "json"
-)
+PROPERTIES (
+"max_batch_interval" = "60",
+"max_batch_rows" = "20000000",
+"max_batch_size" = "1073741824",
+"load_to_single_tablet" = "true",
+"format" = "json"
+)
FROM KAFKA (
"kafka_broker_list" = "host:port",
"kafka_topic" = "log__topic_",
@@ -414,7 +412,7 @@ FROM KAFKA (
<br />SHOW ROUTINE LOAD;
```
-For more information about Kafka, see [Routine
Load](../../data-operate/import/import-way/routine-load-manual.md).
+For more information about Kafka, see [Routine
Load](../data-operate/import/import-way/routine-load-manual.md).
**Using customized programs to collect logs**
diff --git a/versioned_docs/version-2.1/practical-guide/log-storage-analysis.md
b/versioned_docs/version-2.1/practical-guide/log-storage-analysis.md
index dddd69b9f4d..d48641b59ef 100644
--- a/versioned_docs/version-2.1/practical-guide/log-storage-analysis.md
+++ b/versioned_docs/version-2.1/practical-guide/log-storage-analysis.md
@@ -475,15 +475,13 @@ You can refer to the example below, where `property.*`
represents Librdkafka cli
```SQL
CREATE ROUTINE LOAD load_log_kafka ON log_db.log_table
COLUMNS(ts, clientip, request, status, size)
-PROPERTIES (
-"max_batch_interval" = "10",
-"max_batch_rows" = "1000000",
-"max_batch_size" = "109715200",
-"load_to_single_tablet" = "true",
-"timeout" = "600",
-"strict_mode" = "false",
-"format" = "json"
-)
+PROPERTIES (
+"max_batch_interval" = "60",
+"max_batch_rows" = "20000000",
+"max_batch_size" = "1073741824",
+"load_to_single_tablet" = "true",
+"format" = "json"
+)
FROM KAFKA (
"kafka_broker_list" = "host:port",
"kafka_topic" = "log__topic_",
@@ -497,7 +495,7 @@ FROM KAFKA (
<br />SHOW ROUTINE LOAD;
```
-For more information about Kafka, see [Routine
Load](../data-operate/import/import-way/routine-load-manual).
+For more information about Kafka, see [Routine
Load](../data-operate/import/import-way/routine-load-manual.md).
**Using customized programs to collect logs**
diff --git a/versioned_docs/version-3.0/log-storage-analysis.md
b/versioned_docs/version-3.0/log-storage-analysis.md
index 4c532b792f0..9ca0bcc4b1d 100644
--- a/versioned_docs/version-3.0/log-storage-analysis.md
+++ b/versioned_docs/version-3.0/log-storage-analysis.md
@@ -475,15 +475,13 @@ You can refer to the example below, where `property.*`
represents Librdkafka cli
```SQL
CREATE ROUTINE LOAD load_log_kafka ON log_db.log_table
COLUMNS(ts, clientip, request, status, size)
-PROPERTIES (
-"max_batch_interval" = "10",
-"max_batch_rows" = "1000000",
-"max_batch_size" = "109715200",
-"load_to_single_tablet" = "true",
-"timeout" = "600",
-"strict_mode" = "false",
-"format" = "json"
-)
+PROPERTIES (
+"max_batch_interval" = "60",
+"max_batch_rows" = "20000000",
+"max_batch_size" = "1073741824",
+"load_to_single_tablet" = "true",
+"format" = "json"
+)
FROM KAFKA (
"kafka_broker_list" = "host:port",
"kafka_topic" = "log__topic_",
@@ -497,7 +495,7 @@ FROM KAFKA (
<br />SHOW ROUTINE LOAD;
```
-For more information about Kafka, see [Routine
Load](./data-operate/import/import-way/routine-load-manual).
+For more information about Kafka, see [Routine
Load](./data-operate/import/import-way/routine-load-manual.md).
**Using customized programs to collect logs**
diff --git a/versioned_docs/version-3.0/observability/log.md
b/versioned_docs/version-3.0/observability/log.md
index a244f95a67b..e38f35ee2b8 100644
--- a/versioned_docs/version-3.0/observability/log.md
+++ b/versioned_docs/version-3.0/observability/log.md
@@ -392,15 +392,13 @@ You can refer to the example below, where `property.*`
represents Librdkafka cli
```SQL
CREATE ROUTINE LOAD load_log_kafka ON log_db.log_table
COLUMNS(ts, clientip, request, status, size)
-PROPERTIES (
-"max_batch_interval" = "10",
-"max_batch_rows" = "1000000",
-"max_batch_size" = "109715200",
-"load_to_single_tablet" = "true",
-"timeout" = "600",
-"strict_mode" = "false",
-"format" = "json"
-)
+PROPERTIES (
+"max_batch_interval" = "60",
+"max_batch_rows" = "20000000",
+"max_batch_size" = "1073741824",
+"load_to_single_tablet" = "true",
+"format" = "json"
+)
FROM KAFKA (
"kafka_broker_list" = "host:port",
"kafka_topic" = "log__topic_",
@@ -414,7 +412,7 @@ FROM KAFKA (
<br />SHOW ROUTINE LOAD;
```
-For more information about Kafka, see [Routine
Load](../../data-operate/import/import-way/routine-load-manual.md).
+For more information about Kafka, see [Routine
Load](../data-operate/import/import-way/routine-load-manual.md).
**Using customized programs to collect logs**
diff --git a/versioned_docs/version-3.0/practical-guide/log-storage-analysis.md
b/versioned_docs/version-3.0/practical-guide/log-storage-analysis.md
index 56c531a22fe..d6cb0a77b63 100644
--- a/versioned_docs/version-3.0/practical-guide/log-storage-analysis.md
+++ b/versioned_docs/version-3.0/practical-guide/log-storage-analysis.md
@@ -475,15 +475,13 @@ You can refer to the example below, where `property.*`
represents Librdkafka cli
```SQL
CREATE ROUTINE LOAD load_log_kafka ON log_db.log_table
COLUMNS(ts, clientip, request, status, size)
-PROPERTIES (
-"max_batch_interval" = "10",
-"max_batch_rows" = "1000000",
-"max_batch_size" = "109715200",
-"load_to_single_tablet" = "true",
-"timeout" = "600",
-"strict_mode" = "false",
-"format" = "json"
-)
+PROPERTIES (
+"max_batch_interval" = "60",
+"max_batch_rows" = "20000000",
+"max_batch_size" = "1073741824",
+"load_to_single_tablet" = "true",
+"format" = "json"
+)
FROM KAFKA (
"kafka_broker_list" = "host:port",
"kafka_topic" = "log__topic_",
@@ -497,7 +495,7 @@ FROM KAFKA (
<br />SHOW ROUTINE LOAD;
```
-For more information about Kafka, see [Routine
Load](../data-operate/import/import-way/routine-load-manual).
+For more information about Kafka, see [Routine
Load](../data-operate/import/import-way/routine-load-manual.md).
**Using customized programs to collect logs**
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]