tomma-a commented on issue #10193:
URL: https://github.com/apache/seatunnel/issues/10193#issuecomment-3695580038

   > > > [@tomma-a](https://github.com/tomma-a) 
[@yzeng1618](https://github.com/yzeng1618) hi,Will it cause the checkpoint to 
be unable to restart, or will it just prevent the jobid from being obtained
   > > 
   > > 
   > > Hello [@Carl-Zhou-CN](https://github.com/Carl-Zhou-CN) , 
[@yzeng1618](https://github.com/yzeng1618)
   > > From my test, the job could not be restored from flink 
checkpoint/savepoint, The flink in a loop of trying to restore and fail
   > > For some jobs(depends on a specifi job) if flink 
jobmanager.execution.failover-strategy=Restart pipelined region , the job might 
start to run again, but I think it lost the interanal state , not restore from 
last run state.
   > > Really appreciate all your help here. Seatunnel is an awesome 
tool/framework , really appreciate all your contirubitions here!
   > > Tom
   > 
   > hi, [@tomma-a](https://github.com/tomma-a) could you provide more logs? It 
doesn't seem to be the main reason
   
   
   
   Hey @Carl-Zhou-CN ,
   
   I think even it's a 'WARN' message,  the problem could prevent the flink 
job(seatunnel job) to run normally as the the task could not be restored
   
   
   Seatunnel job   config
   `
   {
       "env" : {
           "parallelism" : 3,
           "job.mode" : "STREAMING",
           "checkpoint.interval" : 60000,
           "flink.execution.checkpointing.mode" : "EXACTLY_ONCE",
           "flink.pipeline.max-parallelism" : 64,
           "flink.execution.checkpointing.timeout" : 600000
       },
       "source" : [
           {
               "plugin_output" : "fake2",
               "topic" : "info",
               "consumer.group" : "testr",
               "bootstrap.servers" : 
"tom-cluster-kafka-bootstrap.kafka.svc.cluster.local:9092",
               "commit_on_checkpoint" : false,
               "format" : "json",
               "plugin_name" : "Kafka"
           }
       ],
       "sink" : [
           {
               "plugin_input" : "fake2",
               "topic" : "topic",
               "bootstrap.servers" : 
"tom-cluster1-kafka-bootstrap.kafka.svc.cluster.local:9092",
               "format" : "json",
               "kafka.request.timeout.ms" : 60000,
               "semantics" : "EXACTLY_ONCE",
               "plugin_name" : "Kafka"
           }
       ]
   }
   
   ....
   2025-12-29 06:07:17,485 INFO  
org.apache.seatunnel.core.starter.flink.execution.FlinkExecution [] - Flink 
Execution Plan: {
     "nodes" : [ {
       "id" : 1,
       "type" : "Source: Kafka-Source",
       "pact" : "Data Source",
       "contents" : "Source: Kafka-Source",
       "parallelism" : 3
     }, {
       "id" : 3,
       "type" : "Kafka-Sink: Writer",
       "pact" : "Operator",
       "contents" : "Kafka-Sink: Writer",
       "parallelism" : 3,
       "predecessors" : [ {
         "id" : 1,
         "ship_strategy" : "FORWARD",
         "side" : "second"
       } ]
     }, {
       "id" : 5,
       "type" : "Kafka-Sink: Committer",
       "pact" : "Operator",
       "contents" : "Kafka-Sink: Committer",
       "parallelism" : 3,
       "predecessors" : [ {
         "id" : 3,
         "ship_strategy" : "FORWARD",
         "side" : "second"
       } ]
     } ]
   }
   `
   
   After change the flink job yaml (I used flink upgradeMode: savepoint),  in 
k8s re-apply the flink yaml
   
   flink job is now in  a loop of :  created , initialized, failed.         
Never to be running of success!
   
   
   here is the full log:
   
   2025-12-29 06:07:31,116 INFO  
org.apache.flink.runtime.checkpoint.CheckpointCoordinator    [] - Reset the 
checkpoint ID of job bd6a33ade95d6a15ffaabc25c2fca8c2 to 5.
   2025-12-29 06:07:31,116 INFO  
org.apache.flink.runtime.checkpoint.CheckpointCoordinator    [] - Restoring job 
bd6a33ade95d6a15ffaabc25c2fca8c2 from Savepoint 4 @ 0 for 
bd6a33ade95d6a15ffaabc25c2fca8c2 located at 
oss://xxxxxxxxxxxxxxxxxxxx/testflink/tom/savepoints/savepoint-110c86-77fa991c785d.
   2025-12-29 06:07:31,193 INFO  
org.apache.flink.runtime.checkpoint.CheckpointCoordinator    [] - No master 
state to restore
   2025-12-29 06:07:31,195 INFO  
org.apache.flink.runtime.operators.coordination.RecreateOnResetOperatorCoordinator
 [] - Resetting coordinator to checkpoint.
   2025-12-29 06:07:31,199 INFO  
org.apache.flink.runtime.source.coordinator.SourceCoordinator [] - Closing 
SourceCoordinator for source Source: Kafka-Source.
   2025-12-29 06:07:31,199 INFO  
org.apache.flink.runtime.source.coordinator.SourceCoordinator [] - Source 
coordinator for source Source: Kafka-Source closed.
   2025-12-29 06:07:31,203 INFO  
org.apache.flink.runtime.source.coordinator.SourceCoordinator [] - Restoring 
SplitEnumerator of source Source: Kafka-Source from checkpoint.
   2025-12-29 06:07:31,218 WARN  
org.apache.seatunnel.translation.flink.source.FlinkSourceSplitEnumeratorContext 
[] - Get flink job id failed
   java.lang.IllegalStateException: Initialize flink job-id failed
        at 
org.apache.seatunnel.translation.flink.source.FlinkSourceSplitEnumeratorContext.getJobIdForV15(FlinkSourceSplitEnumeratorContext.java:152)
 
~[blob_p-ce35d9ba37fc821b91a3c1462ad9474638da52bc-02ed4dda7fda8af13c106a81eef8194f:2.3.12]
        at 
org.apache.seatunnel.translation.flink.source.FlinkSourceSplitEnumeratorContext.getFlinkJobId(FlinkSourceSplitEnumeratorContext.java:100)
 
~[blob_p-ce35d9ba37fc821b91a3c1462ad9474638da52bc-02ed4dda7fda8af13c106a81eef8194f:2.3.12]
        at 
org.apache.seatunnel.translation.flink.source.FlinkSourceSplitEnumeratorContext.<init>(FlinkSourceSplitEnumeratorContext.java:57)
 
~[blob_p-ce35d9ba37fc821b91a3c1462ad9474638da52bc-02ed4dda7fda8af13c106a81eef8194f:2.3.12]
        at 
org.apache.seatunnel.translation.flink.source.FlinkSource.restoreEnumerator(FlinkSource.java:116)
 
~[blob_p-ce35d9ba37fc821b91a3c1462ad9474638da52bc-02ed4dda7fda8af13c106a81eef8194f:2.3.12]
        at 
org.apache.seatunnel.translation.flink.source.FlinkSource.restoreEnumerator(FlinkSource.java:48)
 
~[blob_p-ce35d9ba37fc821b91a3c1462ad9474638da52bc-02ed4dda7fda8af13c106a81eef8194f:2.3.12]
        at 
org.apache.flink.runtime.source.coordinator.SourceCoordinator.resetToCheckpoint(SourceCoordinator.java:448)
 ~[flink-dist-1.18.0.jar:1.18.0]
        at 
org.apache.flink.runtime.operators.coordination.RecreateOnResetOperatorCoordinator$DeferrableCoordinator.resetAndStart(RecreateOnResetOperatorCoordinator.java:412)
 ~[flink-dist-1.18.0.jar:1.18.0]
        at 
org.apache.flink.runtime.operators.coordination.RecreateOnResetOperatorCoordinator.lambda$resetToCheckpoint$7(RecreateOnResetOperatorCoordinator.java:156)
 ~[flink-dist-1.18.0.jar:1.18.0]
        at java.util.concurrent.CompletableFuture.uniWhenComplete(Unknown 
Source) ~[?:?]
        at java.util.concurrent.CompletableFuture.uniWhenCompleteStage(Unknown 
Source) ~[?:?]
        at java.util.concurrent.CompletableFuture.whenComplete(Unknown Source) 
~[?:?]
        at 
org.apache.flink.runtime.operators.coordination.RecreateOnResetOperatorCoordinator.resetToCheckpoint(RecreateOnResetOperatorCoordinator.java:143)
 ~[flink-dist-1.18.0.jar:1.18.0]
        at 
org.apache.flink.runtime.operators.coordination.OperatorCoordinatorHolder.resetToCheckpoint(OperatorCoordinatorHolder.java:284)
 ~[flink-dist-1.18.0.jar:1.18.0]
        at 
org.apache.flink.runtime.checkpoint.CheckpointCoordinator.restoreStateToCoordinators(CheckpointCoordinator.java:2044)
 ~[flink-dist-1.18.0.jar:1.18.0]
        at 
org.apache.flink.runtime.checkpoint.CheckpointCoordinator.restoreLatestCheckpointedStateInternal(CheckpointCoordinator.java:1758)
 ~[flink-dist-1.18.0.jar:1.18.0]
        at 
org.apache.flink.runtime.checkpoint.CheckpointCoordinator.restoreSavepoint(CheckpointCoordinator.java:1872)
 ~[flink-dist-1.18.0.jar:1.18.0]
        at 
org.apache.flink.runtime.scheduler.DefaultExecutionGraphFactory.tryRestoreExecutionGraphFromSavepoint(DefaultExecutionGraphFactory.java:224)
 ~[flink-dist-1.18.0.jar:1.18.0]
        at 
org.apache.flink.runtime.scheduler.DefaultExecutionGraphFactory.createAndRestoreExecutionGraph(DefaultExecutionGraphFactory.java:199)
 ~[flink-dist-1.18.0.jar:1.18.0]
        at 
org.apache.flink.runtime.scheduler.SchedulerBase.createAndRestoreExecutionGraph(SchedulerBase.java:371)
 ~[flink-dist-1.18.0.jar:1.18.0]
        at 
org.apache.flink.runtime.scheduler.SchedulerBase.<init>(SchedulerBase.java:214) 
~[flink-dist-1.18.0.jar:1.18.0]
        at 
org.apache.flink.runtime.scheduler.DefaultScheduler.<init>(DefaultScheduler.java:140)
 ~[flink-dist-1.18.0.jar:1.18.0]
        at 
org.apache.flink.runtime.scheduler.DefaultSchedulerFactory.createInstance(DefaultSchedulerFactory.java:156)
 ~[flink-dist-1.18.0.jar:1.18.0]
        at 
org.apache.flink.runtime.jobmaster.DefaultSlotPoolServiceSchedulerFactory.createScheduler(DefaultSlotPoolServiceSchedulerFactory.java:122)
 ~[flink-dist-1.18.0.jar:1.18.0]
        at 
org.apache.flink.runtime.jobmaster.JobMaster.createScheduler(JobMaster.java:379)
 ~[flink-dist-1.18.0.jar:1.18.0]
        at 
org.apache.flink.runtime.jobmaster.JobMaster.<init>(JobMaster.java:356) 
~[flink-dist-1.18.0.jar:1.18.0]
        at 
org.apache.flink.runtime.jobmaster.factories.DefaultJobMasterServiceFactory.internalCreateJobMasterService(DefaultJobMasterServiceFactory.java:128)
 ~[flink-dist-1.18.0.jar:1.18.0]
        at 
org.apache.flink.runtime.jobmaster.factories.DefaultJobMasterServiceFactory.lambda$createJobMasterService$0(DefaultJobMasterServiceFactory.java:100)
 ~[flink-dist-1.18.0.jar:1.18.0]
        at 
org.apache.flink.util.function.FunctionUtils.lambda$uncheckedSupplier$4(FunctionUtils.java:112)
 ~[flink-dist-1.18.0.jar:1.18.0]
        at java.util.concurrent.CompletableFuture$AsyncSupply.run(Unknown 
Source) [?:?]
        at java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source) 
[?:?]
        at java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source) 
[?:?]
        at java.lang.Thread.run(Unknown Source) [?:?]
   Caused by: java.lang.NullPointerException: Cannot invoke "Object.getClass()" 
because "obj" is null
        at 
org.apache.seatunnel.translation.flink.source.FlinkSourceSplitEnumeratorContext.getJobIdForV15(FlinkSourceSplitEnumeratorContext.java:142)
 
~[blob_p-ce35d9ba37fc821b91a3c1462ad9474638da52bc-02ed4dda7fda8af13c106a81eef8194f:2.3.12]
        ... 31 more
   2025-12-29 06:07:31,309 INFO  
org.apache.kafka.clients.admin.AdminClientConfig             [] - 
AdminClientConfig values: 
        bootstrap.servers = 
[tom-cluster-kafka-bootstrap.kafka.svc.cluster.local:9092]
        client.dns.lookup = use_all_dns_ips
        client.id = seatunnel-enumerator-admin-client-1022909018
        connections.max.idle.ms = 300000
        default.api.timeout.ms = 60000
        metadata.max.age.ms = 300000
        metric.reporters = []
        metrics.num.samples = 2
        metrics.recording.level = INFO
        metrics.sample.window.ms = 30000
        receive.buffer.bytes = 65536
        reconnect.backoff.max.ms = 1000
        reconnect.backoff.ms = 50
        request.timeout.ms = 30000
        retries = 2147483647
        retry.backoff.ms = 100
        sasl.client.callback.handler.class = null
        sasl.jaas.config = null
        sasl.kerberos.kinit.cmd = /usr/bin/kinit
        sasl.kerberos.min.time.before.relogin = 60000
        sasl.kerberos.service.name = null
        sasl.kerberos.ticket.renew.jitter = 0.05
        sasl.kerberos.ticket.renew.window.factor = 0.8
        sasl.login.callback.handler.class = null
        sasl.login.class = null
        sasl.login.connect.timeout.ms = null
        sasl.login.read.timeout.ms = null
        sasl.login.refresh.buffer.seconds = 300
        sasl.login.refresh.min.period.seconds = 60
        sasl.login.refresh.window.factor = 0.8
        sasl.login.refresh.window.jitter = 0.05
        sasl.login.retry.backoff.max.ms = 10000
        sasl.login.retry.backoff.ms = 100
        sasl.mechanism = GSSAPI
        sasl.oauthbearer.clock.skew.seconds = 30
        sasl.oauthbearer.expected.audience = null
        sasl.oauthbearer.expected.issuer = null
        sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000
        sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000
        sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100
        sasl.oauthbearer.jwks.endpoint.url = null
        sasl.oauthbearer.scope.claim.name = scope
        sasl.oauthbearer.sub.claim.name = sub
        sasl.oauthbearer.token.endpoint.url = null
        security.protocol = PLAINTEXT
        security.providers = null
        send.buffer.bytes = 131072
        socket.connection.setup.timeout.max.ms = 30000
        socket.connection.setup.timeout.ms = 10000
        ssl.cipher.suites = null
        ssl.enabled.protocols = [TLSv1.2, TLSv1.3]
        ssl.endpoint.identification.algorithm = https
        ssl.engine.factory.class = null
        ssl.key.password = null
        ssl.keymanager.algorithm = SunX509
        ssl.keystore.certificate.chain = null
        ssl.keystore.key = null
        ssl.keystore.location = null
        ssl.keystore.password = null
        ssl.keystore.type = JKS
        ssl.protocol = TLSv1.3
        ssl.provider = null
        ssl.secure.random.implementation = null
        ssl.trustmanager.algorithm = PKIX
        ssl.truststore.certificates = null
        ssl.truststore.location = null
        ssl.truststore.password = null
        ssl.truststore.type = JKS
   
   2025-12-29 06:07:31,493 INFO  org.apache.kafka.common.utils.AppInfoParser    
              [] - Kafka version: 3.2.0
   2025-12-29 06:07:31,493 INFO  org.apache.kafka.common.utils.AppInfoParser    
              [] - Kafka commitId: 38103ffaa962ef50
   2025-12-29 06:07:31,493 INFO  org.apache.kafka.common.utils.AppInfoParser    
              [] - Kafka startTimeMs: 1766988451491
   2025-12-29 06:07:31,494 INFO  
org.apache.seatunnel.connectors.seatunnel.kafka.source.KafkaSourceSplitEnumerator
 [] - Task is being restored, forcing start mode to GROUP_OFFSETS for all topics
   2025-12-29 06:07:31,496 WARN  
org.apache.seatunnel.translation.flink.source.FlinkSourceSplitEnumeratorContext 
[] - Get flink job id failed
   java.lang.IllegalStateException: Initialize flink job-id failed
        at 
org.apache.seatunnel.translation.flink.source.FlinkSourceSplitEnumeratorContext.getJobIdForV15(FlinkSourceSplitEnumeratorContext.java:152)
 
~[blob_p-ce35d9ba37fc821b91a3c1462ad9474638da52bc-02ed4dda7fda8af13c106a81eef8194f:2.3.12]
        at 
org.apache.seatunnel.translation.flink.source.FlinkSourceSplitEnumeratorContext.getFlinkJobId(FlinkSourceSplitEnumeratorContext.java:100)
 
~[blob_p-ce35d9ba37fc821b91a3c1462ad9474638da52bc-02ed4dda7fda8af13c106a81eef8194f:2.3.12]
        at 
org.apache.seatunnel.translation.flink.source.FlinkSourceSplitEnumeratorContext.<init>(FlinkSourceSplitEnumeratorContext.java:57)
 
~[blob_p-ce35d9ba37fc821b91a3c1462ad9474638da52bc-02ed4dda7fda8af13c106a81eef8194f:2.3.12]
        at 
org.apache.seatunnel.translation.flink.source.FlinkSourceEnumerator.<init>(FlinkSourceEnumerator.java:69)
 
~[blob_p-ce35d9ba37fc821b91a3c1462ad9474638da52bc-02ed4dda7fda8af13c106a81eef8194f:2.3.12]
        at 
org.apache.seatunnel.translation.flink.source.FlinkSource.restoreEnumerator(FlinkSource.java:120)
 
~[blob_p-ce35d9ba37fc821b91a3c1462ad9474638da52bc-02ed4dda7fda8af13c106a81eef8194f:2.3.12]
        at 
org.apache.seatunnel.translation.flink.source.FlinkSource.restoreEnumerator(FlinkSource.java:48)
 
~[blob_p-ce35d9ba37fc821b91a3c1462ad9474638da52bc-02ed4dda7fda8af13c106a81eef8194f:2.3.12]
        at 
org.apache.flink.runtime.source.coordinator.SourceCoordinator.resetToCheckpoint(SourceCoordinator.java:448)
 ~[flink-dist-1.18.0.jar:1.18.0]
        at 
org.apache.flink.runtime.operators.coordination.RecreateOnResetOperatorCoordinator$DeferrableCoordinator.resetAndStart(RecreateOnResetOperatorCoordinator.java:412)
 ~[flink-dist-1.18.0.jar:1.18.0]
        at 
org.apache.flink.runtime.operators.coordination.RecreateOnResetOperatorCoordinator.lambda$resetToCheckpoint$7(RecreateOnResetOperatorCoordinator.java:156)
 ~[flink-dist-1.18.0.jar:1.18.0]
        at java.util.concurrent.CompletableFuture.uniWhenComplete(Unknown 
Source) ~[?:?]
        at java.util.concurrent.CompletableFuture.uniWhenCompleteStage(Unknown 
Source) ~[?:?]
        at java.util.concurrent.CompletableFuture.whenComplete(Unknown Source) 
~[?:?]
        at 
org.apache.flink.runtime.operators.coordination.RecreateOnResetOperatorCoordinator.resetToCheckpoint(RecreateOnResetOperatorCoordinator.java:143)
 ~[flink-dist-1.18.0.jar:1.18.0]
        at 
org.apache.flink.runtime.operators.coordination.OperatorCoordinatorHolder.resetToCheckpoint(OperatorCoordinatorHolder.java:284)
 ~[flink-dist-1.18.0.jar:1.18.0]
        at 
org.apache.flink.runtime.checkpoint.CheckpointCoordinator.restoreStateToCoordinators(CheckpointCoordinator.java:2044)
 ~[flink-dist-1.18.0.jar:1.18.0]
        at 
org.apache.flink.runtime.checkpoint.CheckpointCoordinator.restoreLatestCheckpointedStateInternal(CheckpointCoordinator.java:1758)
 ~[flink-dist-1.18.0.jar:1.18.0]
        at 
org.apache.flink.runtime.checkpoint.CheckpointCoordinator.restoreSavepoint(CheckpointCoordinator.java:1872)
 ~[flink-dist-1.18.0.jar:1.18.0]
        at 
org.apache.flink.runtime.scheduler.DefaultExecutionGraphFactory.tryRestoreExecutionGraphFromSavepoint(DefaultExecutionGraphFactory.java:224)
 ~[flink-dist-1.18.0.jar:1.18.0]
        at 
org.apache.flink.runtime.scheduler.DefaultExecutionGraphFactory.createAndRestoreExecutionGraph(DefaultExecutionGraphFactory.java:199)
 ~[flink-dist-1.18.0.jar:1.18.0]
        at 
org.apache.flink.runtime.scheduler.SchedulerBase.createAndRestoreExecutionGraph(SchedulerBase.java:371)
 ~[flink-dist-1.18.0.jar:1.18.0]
        at 
org.apache.flink.runtime.scheduler.SchedulerBase.<init>(SchedulerBase.java:214) 
~[flink-dist-1.18.0.jar:1.18.0]
        at 
org.apache.flink.runtime.scheduler.DefaultScheduler.<init>(DefaultScheduler.java:140)
 ~[flink-dist-1.18.0.jar:1.18.0]
        at 
org.apache.flink.runtime.scheduler.DefaultSchedulerFactory.createInstance(DefaultSchedulerFactory.java:156)
 ~[flink-dist-1.18.0.jar:1.18.0]
        at 
org.apache.flink.runtime.jobmaster.DefaultSlotPoolServiceSchedulerFactory.createScheduler(DefaultSlotPoolServiceSchedulerFactory.java:122)
 ~[flink-dist-1.18.0.jar:1.18.0]
        at 
org.apache.flink.runtime.jobmaster.JobMaster.createScheduler(JobMaster.java:379)
 ~[flink-dist-1.18.0.jar:1.18.0]
        at 
org.apache.flink.runtime.jobmaster.JobMaster.<init>(JobMaster.java:356) 
~[flink-dist-1.18.0.jar:1.18.0]
        at 
org.apache.flink.runtime.jobmaster.factories.DefaultJobMasterServiceFactory.internalCreateJobMasterService(DefaultJobMasterServiceFactory.java:128)
 ~[flink-dist-1.18.0.jar:1.18.0]
        at 
org.apache.flink.runtime.jobmaster.factories.DefaultJobMasterServiceFactory.lambda$createJobMasterService$0(DefaultJobMasterServiceFactory.java:100)
 ~[flink-dist-1.18.0.jar:1.18.0]
        at 
org.apache.flink.util.function.FunctionUtils.lambda$uncheckedSupplier$4(FunctionUtils.java:112)
 ~[flink-dist-1.18.0.jar:1.18.0]
        at java.util.concurrent.CompletableFuture$AsyncSupply.run(Unknown 
Source) [?:?]
        at java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source) 
[?:?]
        at java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source) 
[?:?]
        at java.lang.Thread.run(Unknown Source) [?:?]
   Caused by: java.lang.NullPointerException: Cannot invoke "Object.getClass()" 
because "obj" is null
        at 
org.apache.seatunnel.translation.flink.source.FlinkSourceSplitEnumeratorContext.getJobIdForV15(FlinkSourceSplitEnumeratorContext.java:142)
 
~[blob_p-ce35d9ba37fc821b91a3c1462ad9474638da52bc-02ed4dda7fda8af13c106a81eef8194f:2.3.12]
        ... 32 more
   2025-12-29 06:07:31,573 INFO  org.apache.flink.runtime.jobmaster.JobMast
   2025-12-29 06:07:31,573 INFO  org.apache.flink.runtime.jobmaster.JobMaster   
              [] - Using failover strategy 
org.apache.flink.runtime.executiongraph.failover.flip1.RestartPipelinedRegionFailoverStrategy@699c93b4
 for SeaTunnel (bd6a33ade95d6a15ffaabc25c2fca8c2).
   2025-12-29 06:07:31,589 INFO  
org.apache.flink.runtime.leaderretrieval.DefaultLeaderRetrievalService [] - 
Starting DefaultLeaderRetrievalService with 
KubernetesLeaderRetrievalDriver{configMapName='seatunnel-flink-streaming-example-3-cluster-config-map'}.
   2025-12-29 06:07:31,589 INFO  
org.apache.flink.kubernetes.kubeclient.resources.KubernetesConfigMapSharedInformer
 [] - Starting to watch for 
kafka/seatunnel-flink-streaming-example-3-cluster-config-map, watching 
id:715188fd-67ff-4c81-8dc5-2c8d5c2c0f99
   2025-12-29 06:07:31,589 INFO  org.apache.flink.runtime.jobmaster.JobMaster   
              [] - Starting execution of job 'SeaTunnel' 
(bd6a33ade95d6a15ffaabc25c2fca8c2) under job master id 
b51439a7fbdc7b39047daed657f64893.
   2025-12-29 06:07:31,593 INFO  
org.apache.flink.runtime.source.coordinator.SourceCoordinator [] - Starting 
split enumerator for source Source: Kafka-Source.
   2025-12-29 06:07:31,595 INFO  org.apache.flink.runtime.jobmaster.JobMaster   
              [] - Starting scheduling with scheduling strategy 
[org.apache.flink.runtime.scheduler.strategy.PipelinedRegionSchedulingStrategy]
   2025-12-29 06:07:31,595 INFO  
org.apache.flink.runtime.executiongraph.ExecutionGraph       [] - Job SeaTunnel 
(bd6a33ade95d6a15ffaabc25c2fca8c2) switched from state CREATED to RUNNING.
   2025-12-29 06:07:31,653 INFO  
org.apache.seatunnel.api.event.LoggingEventHandler           [] - log event: 
EnumeratorOpenEvent(createdTime=1766988451653, jobId=null, 
eventType=LIFECYCLE_ENUMERATOR_OPEN)
   2025-12-29 06:07:31,671 INFO  
org.apache.flink.runtime.executiongraph.ExecutionGraph       [] - Source: 
Kafka-Source -> Kafka-Sink: Writer -> Kafka-Sink: Committer (1/3) 
(d5762617aee8a970df5d8d9bedb25c17_cbc357ccb763df2852fee8c4fc7d55f2_0_0) 
switched from CREATED to SCHEDULED.
   2025-12-29 06:07:31,690 INFO  
org.apache.flink.runtime.executiongraph.ExecutionGraph       [] - Source: 
Kafka-Source -> Kafka-Sink: Writer -> Kafka-Sink: Committer (2/3) 
(d5762617aee8a970df5d8d9bedb25c17_cbc357ccb763df2852fee8c4fc7d55f2_1_0) 
switched from CREATED to SCHEDULED.
   2025-12-29 06:07:31,691 INFO  
org.apache.flink.runtime.executiongraph.ExecutionGraph       [] - Source: 
Kafka-Source -> Kafka-Sink: Writer -> Kafka-Sink: Committer (3/3) 
(d5762617aee8a970df5d8d9bedb25c17_cbc357ccb763df2852fee8c4fc7d55f2_2_0) 
switched from CREATED to SCHEDULED.
   2025-12-29 06:07:31,692 INFO  org.apache.flink.runtime.jobmaster.JobMaster   
              [] - Connecting to ResourceManager 
pekko.tcp://[email protected]:6123/user/rpc/resourcemanager_0(b51439a7fbdc7b39047daed657f64893)
   2025-12-29 06:07:31,772 INFO  org.apache.flink.runtime.jobmaster.JobMaster   
              [] - Resolved ResourceManager address, beginning registration
   2025-12-29 06:07:31,774 INFO  
org.apache.flink.runtime.leaderretrieval.DefaultLeaderRetrievalService [] - 
Starting DefaultLeaderRetrievalService with 
KubernetesLeaderRetrievalDriver{configMapName='seatunnel-flink-streaming-example-3-cluster-config-map'}.
   2025-12-29 06:07:31,774 INFO  
org.apache.flink.kubernetes.kubeclient.resources.KubernetesConfigMapSharedInformer
 [] - Starting to watch for 
kafka/seatunnel-flink-streaming-example-3-cluster-config-map, watching 
id:8f4ba652-bba7-469f-a0ce-dbb1973afa84
   2025-12-29 06:07:31,774 INFO  
org.apache.flink.runtime.resourcemanager.active.ActiveResourceManager [] - 
Registering job manager 
[email protected]://[email protected]:6123/user/rpc/jobmanager_2
 for job bd6a33ade95d6a15ffaabc25c2fca8c2.
   2025-12-29 06:07:31,786 INFO  
org.apache.flink.runtime.resourcemanager.active.ActiveResourceManager [] - 
Registered job manager 
[email protected]://[email protected]:6123/user/rpc/jobmanager_2
 for job bd6a33ade95d6a15ffaabc25c2fca8c2.
   2025-12-29 06:07:31,793 INFO  org.apache.flink.runtime.jobmaster.JobMaster   
              [] - JobManager successfully registered at ResourceManager, 
leader id: b51439a7fbdc7b39047daed657f64893.
   2025-12-29 06:07:31,794 INFO  
org.apache.flink.runtime.resourcemanager.slotmanager.FineGrainedSlotManager [] 
- Received resource requirements from job bd6a33ade95d6a15ffaabc25c2fca8c2: 
[ResourceRequirement{resourceProfile=ResourceProfile{UNKNOWN}, 
numberOfRequiredSlots=3}]
   2025-12-29 06:07:31,872 INFO  
org.apache.flink.runtime.resourcemanager.slotmanager.FineGrainedSlotManager [] 
- Matching resource requirements against available resources.
   Missing resources:
         Job bd6a33ade95d6a15ffaabc25c2fca8c2
                ResourceRequirement{resourceProfile=ResourceProfile{UNKNOWN}, 
numberOfRequiredSlots=3}
   Current resources:
        (none)
   2025-12-29 06:07:31,874 INFO  
org.apache.seatunnel.core.starter.flink.execution.FlinkExecution [] - Job 
finished, execution result: 
   
[!!!org.apache.seatunnel.translation.flink.metric.FlinkJobMetricsSummary@1ceda53f=>org.apache.flink.api.common.InvalidProgramException:Job
 was submitted in detached mode. Results of job execution, such as 
accumulators, runtime, etc. are not available. !!!]
   2025-12-29 06:07:31,971 INFO  
org.apache.flink.runtime.resourcemanager.active.ActiveResourceManager [] - need 
request 1 new workers, current worker number 0, declared worker number 1
   2025-12-29 06:07:31,972 INFO  
org.apache.flink.runtime.util.config.memory.ProcessMemoryUtils [] - The derived 
from fraction jvm overhead memory (92.444mb (96935027 bytes)) is less than its 
min value 192.000mb (201326592 bytes), min value will be used instead
   2025-12-29 06:07:31,972 INFO  
org.apache.flink.runtime.resourcemanager.active.ActiveResourceManager [] - 
Requesting new worker with resource spec WorkerResourceSpec {cpuCores=1.1, 
taskHeapSize=25.600mb (26843542 bytes), taskOffHeapSize=0 bytes, 
networkMemSize=64.000mb (67108864 bytes), managedMemSize=230.400mb (241591914 
bytes), numSlots=8}, current pending count: 1.
   2025-12-29 06:07:31,980 INFO  
org.apache.flink.runtime.externalresource.ExternalResourceUtils [] - Enabled 
external resources: []
   2025-12-29 06:07:32,301 INFO  org.apache.flink.configuration.Configuration   
              [] - Config uses fallback configuration key 
'kubernetes.service-account' instead of key 
'kubernetes.taskmanager.service-account'
   2025-12-29 06:07:32,307 INFO  
org.apache.flink.kubernetes.utils.KubernetesUtils            [] - The main 
container image pull policy configured in pod template will be overwritten to 
'Always' because of explicitly configured options.
   2025-12-29 06:07:32,376 INFO  
org.apache.flink.kubernetes.KubernetesResourceManagerDriver  [] - Creating new 
TaskManager pod with name seatunnel-flink-streaming-example-3-taskmanager-1-1 
and resource <1024,1.1>.
   2025-12-29 06:07:33,338 INFO  
org.apache.flink.kubernetes.KubernetesResourceManagerDriver  [] - Pod 
seatunnel-flink-streaming-example-3-taskmanager-1-1 is created.
   2025-12-29 06:07:33,382 INFO  
org.apache.flink.kubernetes.KubernetesResourceManagerDriver  [] - Received new 
TaskManager pod: seatunnel-flink-streaming-example-3-taskmanager-1-1
   2025-12-29 06:07:33,383 INFO  
org.apache.flink.runtime.resourcemanager.active.ActiveResourceManager [] - 
Requested worker seatunnel-flink-streaming-example-3-taskmanager-1-1 with 
resource spec WorkerResourceSpec {cpuCores=1.1, taskHeapSize=25.600mb (26843542 
bytes), taskOffHeapSize=0 bytes, networkMemSize=64.000mb (67108864 bytes), 
managedMemSize=230.400mb (241591914 bytes), numSlots=8}.
   2025-12-29 06:07:49,176 INFO  
org.apache.flink.runtime.checkpoint.CheckpointFailureManager [] - Failed to 
trigger checkpoint for job bd6a33ade95d6a15ffaabc25c2fca8c2 since Checkpoint 
triggering task Source: Kafka-Source -> Kafka-Sink: Writer -> Kafka-Sink: 
Committer (1/3) of job bd6a33ade95d6a15ffaabc25c2fca8c2 is not being executed 
at the moment. Aborting checkpoint. Failure reason: Not all required tasks are 
currently running..
   2025-12-29 06:08:04,588 INFO  
org.apache.flink.runtime.resourcemanager.active.ActiveResourceManager [] - 
Registering TaskManager with ResourceID 
seatunnel-flink-streaming-example-3-taskmanager-1-1 
(pekko.tcp://[email protected]:6122/user/rpc/taskmanager_0) at 
ResourceManager
   2025-12-29 06:08:04,682 INFO  
org.apache.flink.runtime.resourcemanager.slotmanager.FineGrainedSlotManager [] 
- Registering task executor seatunnel-flink-streaming-example-3-taskmanager-1-1 
under c0a7bcab446f42bcb2d1b76cc5d46e7f at the slot manager.
   2025-12-29 06:08:04,684 INFO  
org.apache.flink.runtime.resourcemanager.slotmanager.DefaultSlotStatusSyncer [] 
- Starting allocation of slot fc9290907d17d74942cf33818525ac0b from 
seatunnel-flink-streaming-example-3-taskmanager-1-1 for job 
bd6a33ade95d6a15ffaabc25c2fca8c2 with resource profile 
ResourceProfile{cpuCores=0.1375, taskHeapMemory=3.200mb (3355442 bytes), 
taskOffHeapMemory=0 bytes, managedMemory=28.800mb (30198989 bytes), 
networkMemory=8.000mb (8388608 bytes)}.
   2025-12-29 06:08:04,689 INFO  
org.apache.flink.runtime.resourcemanager.slotmanager.DefaultSlotStatusSyncer [] 
- Starting allocation of slot f461814e1852bddfcbc59b52742796d7 from 
seatunnel-flink-streaming-example-3-taskmanager-1-1 for job 
bd6a33ade95d6a15ffaabc25c2fca8c2 with resource profile 
ResourceProfile{cpuCores=0.1375, taskHeapMemory=3.200mb (3355442 bytes), 
taskOffHeapMemory=0 bytes, managedMemory=28.800mb (30198989 bytes), 
networkMemory=8.000mb (8388608 bytes)}.
   2025-12-29 06:08:04,690 INFO  
org.apache.flink.runtime.resourcemanager.slotmanager.DefaultSlotStatusSyncer [] 
- Starting allocation of slot 4e34c9a12c7cadb6b829d3ba71dd0c2e from 
seatunnel-flink-streaming-example-3-taskmanager-1-1 for job 
bd6a33ade95d6a15ffaabc25c2fca8c2 with resource profile 
ResourceProfile{cpuCores=0.1375, taskHeapMemory=3.200mb (3355442 bytes), 
taskOffHeapMemory=0 bytes, managedMemory=28.800mb (30198989 bytes), 
networkMemory=8.000mb (8388608 bytes)}.
   2025-12-29 06:08:04,691 INFO  
org.apache.flink.runtime.resourcemanager.active.ActiveResourceManager [] - 
Worker seatunnel-flink-streaming-example-3-taskmanager-1-1 is registered.
   2025-12-29 06:08:04,691 INFO  
org.apache.flink.runtime.resourcemanager.active.ActiveResourceManager [] - 
Worker seatunnel-flink-streaming-example-3-taskmanager-1-1 with resource spec 
WorkerResourceSpec {cpuCores=1.1, taskHeapSize=25.600mb (26843542 bytes), 
taskOffHeapSize=0 bytes, networkMemSize=64.000mb (67108864 bytes), 
managedMemSize=230.400mb (241591914 bytes), numSlots=8} was requested in 
current attempt. Current pending count after registering: 0.
   2025-12-29 06:08:04,986 INFO  
org.apache.flink.runtime.executiongraph.ExecutionGraph       [] - Source: 
Kafka-Source -> Kafka-Sink: Writer -> Kafka-Sink: Committer (1/3) 
(d5762617aee8a970df5d8d9bedb25c17_cbc357ccb763df2852fee8c4fc7d55f2_0_0) 
switched from SCHEDULED to DEPLOYING.
   2025-12-29 06:08:04,987 INFO  
org.apache.flink.runtime.executiongraph.ExecutionGraph       [] - Deploying 
Source: Kafka-Source -> Kafka-Sink: Writer -> Kafka-Sink: Committer (1/3) 
(attempt #0) with attempt id 
d5762617aee8a970df5d8d9bedb25c17_cbc357ccb763df2852fee8c4fc7d55f2_0_0 and 
vertex id cbc357ccb763df2852fee8c4fc7d55f2_0 to 
seatunnel-flink-streaming-example-3-taskmanager-1-1 @ 10.244.146.216 
(dataPort=41061) with allocation id fc9290907d17d74942cf33818525ac0b
   2025-12-29 06:08:05,008 INFO  
org.apache.flink.runtime.executiongraph.ExecutionGraph       [] - Source: 
Kafka-Source -> Kafka-Sink: Writer -> Kafka-Sink: Committer (2/3) 
(d5762617aee8a970df5d8d9bedb25c17_cbc357ccb763df2852fee8c4fc7d55f2_1_0) 
switched from SCHEDULED to DEPLOYING.
   2025-12-29 06:08:05,008 INFO  
org.apache.flink.runtime.executiongraph.ExecutionGraph       [] - Deploying 
Source: Kafka-Source -> Kafka-Sink: Writer -> Kafka-Sink: Committer (2/3) 
(attempt #0) with attempt id 
d5762617aee8a970df5d8d9bedb25c17_cbc357ccb763df2852fee8c4fc7d55f2_1_0 and 
vertex id cbc357ccb763df2852fee8c4fc7d55f2_1 to 
seatunnel-flink-streaming-example-3-taskmanager-1-1 @ 10.244.146.216 
(dataPort=41061) with allocation id f461814e1852bddfcbc59b52742796d7
   2025-12-29 06:08:05,008 INFO  
org.apache.flink.runtime.executiongraph.ExecutionGraph       [] - Source: 
Kafka-Source -> Kafka-Sink: Writer -> Kafka-Sink: Committer (3/3) 
(d5762617aee8a970df5d8d9bedb25c17_cbc357ccb763df2852fee8c4fc7d55f2_2_0) 
switched from SCHEDULED to DEPLOYING.
   2025-12-29 06:08:05,009 INFO  
org.apache.flink.runtime.executiongraph.ExecutionGraph       [] - Deploying 
Source: Kafka-Source -> Kafka-Sink: Writer -> Kafka-Sink: Committer (3/3) 
(attempt #0) with attempt id 
d5762617aee8a970df5d8d9bedb25c17_cbc357ccb763df2852fee8c4fc7d55f2_2_0 and 
vertex id cbc357ccb763df2852fee8c4fc7d55f2_2 to 
seatunnel-flink-streaming-example-3-taskmanager-1-1 @ 10.244.146.216 
(dataPort=41061) with allocation id 4e34c9a12c7cadb6b829d3ba71dd0c2e
   2025-12-29 06:08:20,478 INFO  
org.apache.flink.runtime.executiongraph.ExecutionGraph       [] - Source: 
Kafka-Source -> Kafka-Sink: Writer -> Kafka-Sink: Committer (3/3) 
(d5762617aee8a970df5d8d9bedb25c17_cbc357ccb763df2852fee8c4fc7d55f2_2_0) 
switched from DEPLOYING to INITIALIZING.
   2025-12-29 06:08:20,483 INFO  
org.apache.flink.runtime.executiongraph.ExecutionGraph       [] - Source: 
Kafka-Source -> Kafka-Sink: Writer -> Kafka-Sink: Committer (1/3) 
(d5762617aee8a970df5d8d9bedb25c17_cbc357ccb763df2852fee8c4fc7d55f2_0_0) 
switched from DEPLOYING to INITIALIZING.
   2025-12-29 06:08:20,484 INFO  
org.apache.flink.runtime.executiongraph.ExecutionGraph       [] - Source: 
Kafka-Source -> Kafka-Sink: Writer -> Kafka-Sink: Committer (2/3) 
(d5762617aee8a970df5d8d9bedb25c17_cbc357ccb763df2852fee8c4fc7d55f2_1_0) 
switched from DEPLOYING to INITIALIZING.
   2025-12-29 06:08:49,170 INFO  
org.apache.flink.runtime.checkpoint.CheckpointFailureManager [] - Failed to 
trigger checkpoint for job bd6a33ade95d6a15ffaabc25c2fca8c2 since Checkpoint 
triggering task Source: Kafka-Source -> Kafka-Sink: Writer -> Kafka-Sink: 
Committer (1/3) of job bd6a33ade95d6a15ffaabc25c2fca8c2 is not being executed 
at the moment. Aborting checkpoint. Failure reason: Not all required tasks are 
currently running..
   2025-12-29 06:09:49,170 INFO  
org.apache.flink.runtime.checkpoint.CheckpointFailureManager [] - Failed to 
trigger checkpoint for job bd6a33ade95d6a15ffaabc25c2fca8c2 since Checkpoint 
triggering task Source: Kafka-Source -> Kafka-Sink: Writer -> Kafka-Sink: 
Committer (1/3) of job bd6a33ade95d6a15ffaabc25c2fca8c2 is not being executed 
at the moment. Aborting checkpoint. Failure reason: Not all required tasks are 
currently running..
   2025-12-29 06:10:49,170 INFO  
org.apache.flink.runtime.checkpoint.CheckpointFailureManager [] - Failed to 
trigger checkpoint for job bd6a33ade95d6a15ffaabc25c2fca8c2 since Checkpoint 
triggering task Source: Kafka-Source -> Kafka-Sink: Writer -> Kafka-Sink: 
Committer (1/3) of job bd6a33ade95d6a15ffaabc25c2fca8c2 is not being executed 
at the moment. Aborting checkpoint. Failure reason: Not all required tasks are 
currently running..
   2025-12-29 06:11:49,170 INFO  
org.apache.flink.runtime.checkpoint.CheckpointFailureManager [] - Failed to 
trigger checkpoint for job bd6a33ade95d6a15ffaabc25c2fca8c2 since Checkpoint 
triggering task Source: Kafka-Source -> Kafka-Sink: Writer -> Kafka-Sink: 
Committer (1/3) of job bd6a33ade95d6a15ffaabc25c2fca8c2 is not being executed 
at the moment. Aborting checkpoint. Failure reason: Not all required tasks are 
currently running..
   2025-12-29 06:12:33,694 INFO  org.apache.kafka.clients.NetworkClient         
              [] - [AdminClient 
clientId=seatunnel-enumerator-admin-client-1022909018] Node -1 disconnected.
   2025-12-29 06:12:49,170 INFO  
org.apache.flink.runtime.checkpoint.CheckpointFailureManager [] - Failed to 
trigger checkpoint for job bd6a33ade95d6a15ffaabc25c2fca8c2 since Checkpoint 
triggering task Source: Kafka-Source -> Kafka-Sink: Writer -> Kafka-Sink: 
Committer (1/3) of job bd6a33ade95d6a15ffaabc25c2fca8c2 is not being executed 
at the moment. Aborting checkpoint. Failure reason: Not all required tasks are 
currently running..
   2025-12-29 06:13:49,170 INFO  
org.apache.flink.runtime.checkpoint.CheckpointFailureManager [] - Failed to 
trigger checkpoint for job bd6a33ade95d6a15ffaabc25c2fca8c2 since Checkpoint 
triggering task Source: Kafka-Source -> Kafka-Sink: Writer -> Kafka-Sink: 
Committer (1/3) of job bd6a33ade95d6a15ffaabc25c2fca8c2 is not being executed 
at the moment. Aborting checkpoint. Failure reason: Not all required tasks are 
currently running..
   2025-12-29 06:14:49,170 INFO  
org.apache.flink.runtime.checkpoint.CheckpointFailureManager [] - Failed to 
trigger checkpoint for job bd6a33ade95d6a15ffaabc25c2fca8c2 since Checkpoint 
triggering task Source: Kafka-Source -> Kafka-Sink: Writer -> Kafka-Sink: 
Committer (1/3) of job bd6a33ade95d6a15ffaabc25c2fca8c2 is not being executed 
at the moment. Aborting checkpoint. Failure reason: Not all required tasks are 
currently running..
   2025-12-29 06:15:49,170 INFO  
org.apache.flink.runtime.checkpoint.CheckpointFailureManager [] - Failed to 
trigger checkpoint for job bd6a33ade95d6a15ffaabc25c2fca8c2 since Checkpoint 
triggering task Source: Kafka-Source -> Kafka-Sink: Writer -> Kafka-Sink: 
Committer (1/3) of job bd6a33ade95d6a15ffaabc25c2fca8c2 is not being executed 
at the moment. Aborting checkpoint. Failure reason: Not all required tasks are 
currently running..
   2025-12-29 06:16:49,169 INFO  
org.apache.flink.runtime.checkpoint.CheckpointFailureManager [] - Failed to 
trigger checkpoint for job bd6a33ade95d6a15ffaabc25c2fca8c2 since Checkpoint 
triggering task Source: Kafka-Source -> Kafka-Sink: Writer -> Kafka-Sink: 
Committer (1/3) of job bd6a33ade95d6a15ffaabc25c2fca8c2 is not being executed 
at the moment. Aborting checkpoint. Failure reason: Not all required tasks are 
currently running..
   `


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]


Reply via email to