This is an automated email from the ASF dual-hosted git repository.

yiguolei pushed a commit to branch branch-4.0
in repository https://gitbox.apache.org/repos/asf/doris.git


The following commit(s) were added to refs/heads/branch-4.0 by this push:
     new 8d494d1fecd branch-4.0: [fix](test) fix some routine load case 
unstable #61432 (#61609)
8d494d1fecd is described below

commit 8d494d1fecdca9ba4a4f19ada2437c67807044c6
Author: github-actions[bot] 
<41898282+github-actions[bot]@users.noreply.github.com>
AuthorDate: Mon Mar 23 13:59:49 2026 +0800

    branch-4.0: [fix](test) fix some routine load case unstable #61432 (#61609)
    
    Cherry-picked from #61432
    
    Co-authored-by: hui lai <[email protected]>
---
 be/src/cloud/cloud_rowset_builder.cpp                          |  2 ++
 be/src/runtime/routine_load/data_consumer.cpp                  | 10 +++++++---
 .../suites/load_p0/routine_load/test_black_list.groovy         |  4 ++--
 3 files changed, 11 insertions(+), 5 deletions(-)

diff --git a/be/src/cloud/cloud_rowset_builder.cpp 
b/be/src/cloud/cloud_rowset_builder.cpp
index 8ef15424a3d..ae082231076 100644
--- a/be/src/cloud/cloud_rowset_builder.cpp
+++ b/be/src/cloud/cloud_rowset_builder.cpp
@@ -96,6 +96,8 @@ Status CloudRowsetBuilder::init() {
 
 Status CloudRowsetBuilder::check_tablet_version_count() {
     int64_t version_count = 
cloud_tablet()->fetch_add_approximate_num_rowsets(0);
+    
DBUG_EXECUTE_IF("RowsetBuilder.check_tablet_version_count.too_many_version",
+                    { version_count = INT_MAX; });
     // TODO(plat1ko): load backoff algorithm
     int32_t max_version_config = cloud_tablet()->max_version_config();
     if (version_count > max_version_config) {
diff --git a/be/src/runtime/routine_load/data_consumer.cpp 
b/be/src/runtime/routine_load/data_consumer.cpp
index eecba26fc10..dad2b081060 100644
--- a/be/src/runtime/routine_load/data_consumer.cpp
+++ b/be/src/runtime/routine_load/data_consumer.cpp
@@ -420,9 +420,9 @@ Status KafkaDataConsumer::get_offsets_for_times(const 
std::vector<PIntegerPair>&
 Status KafkaDataConsumer::get_latest_offsets_for_partitions(
         const std::vector<int32_t>& partition_ids, std::vector<PIntegerPair>* 
offsets,
         int timeout) {
-    
DBUG_EXECUTE_IF("KafkaDataConsumer.get_latest_offsets_for_partitions.timeout", {
-        // sleep 60s
-        std::this_thread::sleep_for(std::chrono::seconds(60));
+    DBUG_EXECUTE_IF("KafkaDataConsumer.get_offsets_for_partitions.timeout", {
+        // sleep 61s
+        std::this_thread::sleep_for(std::chrono::seconds(61));
     });
     MonotonicStopWatch watch;
     watch.start();
@@ -456,6 +456,10 @@ Status 
KafkaDataConsumer::get_latest_offsets_for_partitions(
 Status KafkaDataConsumer::get_real_offsets_for_partitions(
         const std::vector<PIntegerPair>& offset_flags, 
std::vector<PIntegerPair>* offsets,
         int timeout) {
+    DBUG_EXECUTE_IF("KafkaDataConsumer.get_offsets_for_partitions.timeout", {
+        // sleep 61s
+        std::this_thread::sleep_for(std::chrono::seconds(61));
+    });
     MonotonicStopWatch watch;
     watch.start();
     for (const auto& entry : offset_flags) {
diff --git a/regression-test/suites/load_p0/routine_load/test_black_list.groovy 
b/regression-test/suites/load_p0/routine_load/test_black_list.groovy
index 29fc336492b..a68ced241d6 100644
--- a/regression-test/suites/load_p0/routine_load/test_black_list.groovy
+++ b/regression-test/suites/load_p0/routine_load/test_black_list.groovy
@@ -94,7 +94,7 @@ suite("test_black_list","nonConcurrent,p0") {
             PROPERTIES ("replication_allocation" = "tag.location.default: 1");
         """
 
-        def inject = 
"KafkaDataConsumer.get_latest_offsets_for_partitions.timeout"
+        def inject = "KafkaDataConsumer.get_offsets_for_partitions.timeout"
         try {
             GetDebugPoint().enableDebugPointForAllBEs(inject)
             sql """
@@ -116,7 +116,7 @@ suite("test_black_list","nonConcurrent,p0") {
                 log.info("routine load state: 
${state[0][8].toString()}".toString())
                 log.info("reason of state changed: 
${state[0][17].toString()}".toString())
                 log.info("other msg: ${state[0][19].toString()}".toString())
-                if (state[0][17].toString().contains("failed to get latest 
partition offset") || state[0][19].toString().contains("failed to get latest 
partition offset")) {
+                if (state[0][17].toString().contains("Failed to get real 
offsets of kafka topic") || state[0][19].toString().contains("Failed to get 
real offsets of kafka topic")) {
                     break
                 }
                 if (count >= 90) {


---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to