多谢!打开了DEBUG日志,仍然只有最后一个ERROR,不过之前有不少包含
kubernetes.client.dsl.internal.WatchConnectionManager  的日志,grep
了一部分,能看出些什么吗?

job-debug-0118.log:2021-01-19 02:12:25,551 DEBUG
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager [] -
WebSocket successfully opened
job-debug-0118.log:2021-01-19 02:12:25,646 DEBUG
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager [] -
Connecting websocket ...
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager@2553d42c
job-debug-0118.log:2021-01-19 02:12:25,647 DEBUG
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager [] -
WebSocket successfully opened
job-debug-0118.log:2021-01-19 02:12:30,128 DEBUG
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager [] -
Connecting websocket ...
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager@5a9fa83e
job-debug-0118.log:2021-01-19 02:12:30,176 DEBUG
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager [] -
WebSocket successfully opened
job-debug-0118.log:2021-01-19 02:12:39,028 DEBUG
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager [] - Force
closing the watch
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager@2553d42c
job-debug-0118.log:2021-01-19 02:12:39,028 DEBUG
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager [] -
Closing websocket
org.apache.flink.kubernetes.shaded.okhttp3.internal.ws.RealWebSocket@15b15029
job-debug-0118.log:2021-01-19 02:12:39,030 DEBUG
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager [] -
WebSocket close received. code: 1000, reason: 
job-debug-0118.log:2021-01-19 02:12:39,030 DEBUG
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager [] -
Ignoring onClose for already closed/closing websocket
job-debug-0118.log:2021-01-19 02:12:39,031 DEBUG
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager [] - Force
closing the watch
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager@2cdbe5a0
job-debug-0118.log:2021-01-19 02:12:39,031 DEBUG
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager [] -
Closing websocket
org.apache.flink.kubernetes.shaded.okhttp3.internal.ws.RealWebSocket@1e3f5396
job-debug-0118.log:2021-01-19 02:12:39,033 DEBUG
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager [] -
WebSocket close received. code: 1000, reason: 
job-debug-0118.log:2021-01-19 02:12:39,033 DEBUG
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager [] -
Ignoring onClose for already closed/closing websocket
job-debug-0118.log:2021-01-19 02:12:42,677 DEBUG
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager [] -
Connecting websocket ...
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager@210aab4b
job-debug-0118.log:2021-01-19 02:12:42,678 DEBUG
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager [] -
WebSocket successfully opened
job-debug-0118.log:2021-01-19 02:12:42,920 DEBUG
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager [] -
Connecting websocket ...
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager@278d8398
job-debug-0118.log:2021-01-19 02:12:42,921 DEBUG
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager [] -
WebSocket successfully opened
job-debug-0118.log:2021-01-19 02:12:45,130 DEBUG
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager [] -
Connecting websocket ...
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager@4b318628
job-debug-0118.log:2021-01-19 02:12:45,132 DEBUG
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager [] -
WebSocket successfully opened
job-debug-0118.log:2021-01-19 02:13:05,927 DEBUG
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager [] - Force
closing the watch
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager@278d8398
job-debug-0118.log:2021-01-19 02:13:05,927 DEBUG
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager [] -
Closing websocket
org.apache.flink.kubernetes.shaded.okhttp3.internal.ws.RealWebSocket@69d1ebd2
job-debug-0118.log:2021-01-19 02:13:05,930 DEBUG
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager [] -
WebSocket close received. code: 1000, reason: 
job-debug-0118.log:2021-01-19 02:13:05,930 DEBUG
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager [] -
Ignoring onClose for already closed/closing websocket
job-debug-0118.log:2021-01-19 02:13:05,940 DEBUG
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager [] - Force
closing the watch
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager@210aab4b
job-debug-0118.log:2021-01-19 02:13:05,940 DEBUG
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager [] -
Closing websocket
org.apache.flink.kubernetes.shaded.okhttp3.internal.ws.RealWebSocket@3db9d8d8
job-debug-0118.log:2021-01-19 02:13:05,942 DEBUG
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager [] -
WebSocket close received. code: 1000, reason: 
job-debug-0118.log:2021-01-19 02:13:05,942 DEBUG
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager [] -
Ignoring onClose for already closed/closing websocket
job-debug-0118.log:2021-01-19 02:13:08,378 DEBUG
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager [] -
Connecting websocket ...
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager@4dcf905
job-debug-0118.log:2021-01-19 02:13:08,381 DEBUG
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager [] -
WebSocket successfully opened
job-debug-0118.log:2021-01-19 02:13:08,471 DEBUG
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager [] -
Connecting websocket ...
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager@428ca061
job-debug-0118.log:2021-01-19 02:13:08,472 DEBUG
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager [] -
WebSocket successfully opened
job-debug-0118.log:2021-01-19 02:13:10,127 DEBUG
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager [] -
Connecting websocket ...
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager@46b49e58
job-debug-0118.log:2021-01-19 02:13:10,128 DEBUG
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager [] -
WebSocket successfully opened
job-debug-0118.log:2021-01-19 02:13:21,625 DEBUG
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager [] - Force
closing the watch
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager@428ca061
job-debug-0118.log:2021-01-19 02:13:21,625 DEBUG
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager [] -
Closing websocket
org.apache.flink.kubernetes.shaded.okhttp3.internal.ws.RealWebSocket@14e16427
job-debug-0118.log:2021-01-19 02:13:21,627 DEBUG
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager [] -
WebSocket close received. code: 1000, reason: 
job-debug-0118.log:2021-01-19 02:13:21,627 DEBUG
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager [] -
Ignoring onClose for already closed/closing websocket
job-debug-0118.log:2021-01-19 02:13:21,628 DEBUG
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager [] - Force
closing the watch
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager@4dcf905
job-debug-0118.log:2021-01-19 02:13:21,628 DEBUG
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager [] -
Closing websocket
org.apache.flink.kubernetes.shaded.okhttp3.internal.ws.RealWebSocket@11708e54
job-debug-0118.log:2021-01-19 02:13:21,630 DEBUG
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager [] -
WebSocket close received. code: 1000, reason: 
job-debug-0118.log:2021-01-19 02:13:21,630 DEBUG
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager [] -
Ignoring onClose for already closed/closing websocket
job-debug-0118.log:2021-01-19 02:13:25,680 DEBUG
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager [] -
Connecting websocket ...
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager@3ba4abd7
job-debug-0118.log:2021-01-19 02:13:25,681 DEBUG
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager [] -
WebSocket successfully opened
job-debug-0118.log:2021-01-19 02:13:25,908 DEBUG
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager [] -
Connecting websocket ...
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager@23fe4bdd
job-debug-0118.log:2021-01-19 02:13:25,909 DEBUG
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager [] -
WebSocket successfully opened
job-debug-0118.log:2021-01-19 02:13:30,128 DEBUG
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager [] -
Connecting websocket ...
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager@5cf8bd92
job-debug-0118.log:2021-01-19 02:13:30,175 DEBUG
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager [] -
WebSocket successfully opened
job-debug-0118.log:2021-01-19 02:13:46,104 DEBUG
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager [] -
WebSocket close received. code: 1000, reason: 
job-debug-0118.log:2021-01-19 02:13:46,105 DEBUG
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager [] -
Submitting reconnect task to the executor
job-debug-0118.log:2021-01-19 02:13:46,113 DEBUG
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager [] -
Scheduling reconnect task
job-debug-0118.log:2021-01-19 02:13:46,117 DEBUG
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager [] -
Current reconnect backoff is 1000 milliseconds (T0)
job-debug-0118.log:2021-01-19 02:13:47,117 DEBUG
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager [] -
Connecting websocket ...
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager@23f03575
job-debug-0118.log:2021-01-19 02:13:47,120 DEBUG
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager [] -
WebSocket successfully opened
job-debug-0118.log:     at
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager.closeEvent(WatchConnectionManager.java:367)
[flink-dist_2.11-1.12.1.jar:1.12.1]
job-debug-0118.log:     at
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager.access$700(WatchConnectionManager.java:50)
[flink-dist_2.11-1.12.1.jar:1.12.1]
job-debug-0118.log:     at
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager$1.onMessage(WatchConnectionManager.java:259)
[flink-dist_2.11-1.12.1.jar:1.12.1]
job-debug-0118.log:     at
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager.closeEvent(WatchConnectionManager.java:367)
[flink-dist_2.11-1.12.1.jar:1.12.1]
job-debug-0118.log:     at
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager.access$700(WatchConnectionManager.java:50)
[flink-dist_2.11-1.12.1.jar:1.12.1]
job-debug-0118.log:     at
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager$1.onMessage(WatchConnectionManager.java:259)
[flink-dist_2.11-1.12.1.jar:1.12.1]


最后的ERROR是这样的

2021-01-19 02:13:47,094 DEBUG
org.apache.flink.runtime.source.coordinator.SourceCoordinator [] - Handling
event from subtask 406 of source Source:
HiveSource-snmpprobe.p_snmp_ifXTable: RequestSplitEvent (host='172.0.37.8')
2021-01-19 02:13:47,094 INFO 
org.apache.flink.connector.file.src.impl.StaticFileSplitEnumerator [] -
Subtask 406 (on host '172.0.37.8') is requesting a file source split
2021-01-19 02:13:47,094 INFO 
org.apache.flink.connector.file.src.impl.StaticFileSplitEnumerator [] - No
more splits available for subtask 406
2021-01-19 02:13:47,097 INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph       [] - Source:
HiveSource-snmpprobe.p_snmp_ifXTable (318/458)
(710557b37a1e03f0f462ab5303842489) switched from RUNNING to FINISHED.
2021-01-19 02:13:47,097 DEBUG
org.apache.flink.runtime.executiongraph.ExecutionGraph       [] - Ignoring
transition of vertex Source: HiveSource-snmpprobe.p_snmp_ifXTable (318/458)
- execution #0 to FAILED while being FINISHED.
2021-01-19 02:13:47,097 DEBUG org.apache.flink.runtime.scheduler.SharedSlot     
          
[] - Remove logical slot (SlotRequestId{988c43b8a7b427ea962685f057438880})
for execution vertex (id 605b35e407e90cda15ad084365733fdd_317) from the
physical slot (SlotRequestId{37b03b71035c9d8c564bb7c299ee9b3d})
2021-01-19 02:13:47,097 DEBUG org.apache.flink.runtime.scheduler.SharedSlot     
          
[] - Release shared slot externally
(SlotRequestId{37b03b71035c9d8c564bb7c299ee9b3d})
2021-01-19 02:13:47,097 DEBUG
org.apache.flink.runtime.jobmaster.slotpool.SlotPoolImpl     [] - Releasing
slot [SlotRequestId{37b03b71035c9d8c564bb7c299ee9b3d}] because: Slot is
being returned from SlotSharingExecutionSlotAllocator.
2021-01-19 02:13:47,097 DEBUG org.apache.flink.runtime.scheduler.SharedSlot     
          
[] - Release shared slot (SlotRequestId{37b03b71035c9d8c564bb7c299ee9b3d})
2021-01-19 02:13:47,097 DEBUG
org.apache.flink.runtime.jobmaster.slotpool.SlotPoolImpl     [] - Fulfilling
pending slot request [SlotRequestId{bb5a48db898111288c811359cc2d7f51}] with
slot [385153f7c5efff54be584439258f7352]
2021-01-19 02:13:47,097 DEBUG org.apache.flink.runtime.scheduler.SharedSlot     
          
[] - Allocated logical slot
(SlotRequestId{78f370c05403ab3d703a8d89c19d23c8}) for execution vertex (id
605b35e407e90cda15ad084365733fdd_419) from the physical slot
(SlotRequestId{bb5a48db898111288c811359cc2d7f51})
2021-01-19 02:13:47,097 INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph       [] - Source:
HiveSource-snmpprobe.p_snmp_ifXTable (420/458)
(d04edd6e11b7cdc9e88c0ab6d756fed2) switched from SCHEDULED to DEPLOYING.
2021-01-19 02:13:47,097 INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph       [] - Deploying
Source: HiveSource-snmpprobe.p_snmp_ifXTable (420/458) (attempt #0) with
attempt id d04edd6e11b7cdc9e88c0ab6d756fed2 to 172.0.42.250:6122-5d505f @
172-0-42-250.flink-taskmanager-query-state.gem-flink.svc.cluster.local
(dataPort=40697) with allocation id 385153f7c5efff54be584439258f7352
2021-01-19 02:13:47,097 DEBUG
org.apache.flink.runtime.resourcemanager.slotmanager.SlotManagerImpl [] -
Cancel slot request 4da96bc97ef9b47ba7e408c78835d75a.
2021-01-19 02:13:47,100 INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph       [] - Source:
HiveSource-snmpprobe.p_snmp_ifXTable (413/458)
(2124587d1641d6cb05c05dfc742e8423) switched from DEPLOYING to RUNNING.
2021-01-19 02:13:47,100 INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph       [] - Source:
HiveSource-snmpprobe.p_snmp_ifXTable (414/458)
(4aa88cc1b61dc5b056ad59d373392c2f) switched from DEPLOYING to RUNNING.
2021-01-19 02:13:47,100 INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph       [] - Source:
HiveSource-snmpprobe.p_snmp_ifXTable (415/458)
(225d9a604e6b852f2ea6e87ebcf3107c) switched from DEPLOYING to RUNNING.
2021-01-19 02:13:47,112 INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph       [] - Source:
HiveSource-snmpprobe.p_snmp_ifXTable (417/458)
(c1a78898e76b3f1761cd5be1913dd24c) switched from DEPLOYING to RUNNING.
2021-01-19 02:13:47,112 INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph       [] - Source:
HiveSource-snmpprobe.p_snmp_ifXTable (418/458)
(6deb899dbd8bf373b349b980d1e78506) switched from DEPLOYING to RUNNING.
2021-01-19 02:13:47,113 INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph       [] - Source:
HiveSource-snmpprobe.p_snmp_ifXTable (416/458)
(5a99fa4a1d8bdbf93345a9b20ae1fa91) switched from DEPLOYING to RUNNING.
2021-01-19 02:13:47,117 INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph       [] - Source:
HiveSource-snmpprobe.p_snmp_ifXTable (406/458)
(d1f9edd1bfdd80eef6b32b8850020130) switched from RUNNING to FINISHED.
2021-01-19 02:13:47,117 DEBUG
org.apache.flink.runtime.executiongraph.ExecutionGraph       [] - Ignoring
transition of vertex Source: HiveSource-snmpprobe.p_snmp_ifXTable (406/458)
- execution #0 to FAILED while being FINISHED.
2021-01-19 02:13:47,117 DEBUG org.apache.flink.runtime.scheduler.SharedSlot     
          
[] - Remove logical slot (SlotRequestId{037efe676c5cec5fe6b549d3ebd5f72b})
for execution vertex (id 605b35e407e90cda15ad084365733fdd_405) from the
physical slot (SlotRequestId{a840e61c33cb3f250cfb54652c87aa64})
2021-01-19 02:13:47,117 DEBUG org.apache.flink.runtime.scheduler.SharedSlot     
          
[] - Release shared slot externally
(SlotRequestId{a840e61c33cb3f250cfb54652c87aa64})
2021-01-19 02:13:47,117 DEBUG
org.apache.flink.runtime.jobmaster.slotpool.SlotPoolImpl     [] - Releasing
slot [SlotRequestId{a840e61c33cb3f250cfb54652c87aa64}] because: Slot is
being returned from SlotSharingExecutionSlotAllocator.
2021-01-19 02:13:47,117 DEBUG org.apache.flink.runtime.scheduler.SharedSlot     
          
[] - Release shared slot (SlotRequestId{a840e61c33cb3f250cfb54652c87aa64})
2021-01-19 02:13:47,117 DEBUG
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager [] -
Connecting websocket ...
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager@23f03575
2021-01-19 02:13:47,117 DEBUG
org.apache.flink.runtime.jobmaster.slotpool.SlotPoolImpl     [] - Fulfilling
pending slot request [SlotRequestId{538f35a507cd0949bf547588eb436b49}] with
slot [ebf6ebbb9abe3a9e6ccb56e235b00b53]
2021-01-19 02:13:47,117 DEBUG org.apache.flink.runtime.scheduler.SharedSlot     
          
[] - Allocated logical slot
(SlotRequestId{a9bac8016853f9e86963b8ee11dea18f}) for execution vertex (id
605b35e407e90cda15ad084365733fdd_420) from the physical slot
(SlotRequestId{538f35a507cd0949bf547588eb436b49})
2021-01-19 02:13:47,117 INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph       [] - Source:
HiveSource-snmpprobe.p_snmp_ifXTable (421/458)
(706d012e7a572e1e5786536df9ab3bbb) switched from SCHEDULED to DEPLOYING.
2021-01-19 02:13:47,117 INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph       [] - Deploying
Source: HiveSource-snmpprobe.p_snmp_ifXTable (421/458) (attempt #0) with
attempt id 706d012e7a572e1e5786536df9ab3bbb to 172.0.37.8:6122-694869 @
172-0-37-8.flink-taskmanager-query-state.gem-flink.svc.cluster.local
(dataPort=32959) with allocation id ebf6ebbb9abe3a9e6ccb56e235b00b53
2021-01-19 02:13:47,117 INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph       [] - Source:
HiveSource-snmpprobe.p_snmp_ifXTable (407/458)
(396cb3fdd115a31d8575407fa9ee6e07) switched from RUNNING to FINISHED.
2021-01-19 02:13:47,117 DEBUG
org.apache.flink.runtime.executiongraph.ExecutionGraph       [] - Ignoring
transition of vertex Source: HiveSource-snmpprobe.p_snmp_ifXTable (407/458)
- execution #0 to FAILED while being FINISHED.
2021-01-19 02:13:47,117 DEBUG org.apache.flink.runtime.scheduler.SharedSlot     
          
[] - Remove logical slot (SlotRequestId{88e116a3dd0a40ef692734548aac9682})
for execution vertex (id 605b35e407e90cda15ad084365733fdd_406) from the
physical slot (SlotRequestId{9bb6a1762363d3996aded34c82abab54})
2021-01-19 02:13:47,117 DEBUG org.apache.flink.runtime.scheduler.SharedSlot     
          
[] - Release shared slot externally
(SlotRequestId{9bb6a1762363d3996aded34c82abab54})
2021-01-19 02:13:47,117 DEBUG
org.apache.flink.runtime.jobmaster.slotpool.SlotPoolImpl     [] - Releasing
slot [SlotRequestId{9bb6a1762363d3996aded34c82abab54}] because: Slot is
being returned from SlotSharingExecutionSlotAllocator.
2021-01-19 02:13:47,117 DEBUG org.apache.flink.runtime.scheduler.SharedSlot     
          
[] - Release shared slot (SlotRequestId{9bb6a1762363d3996aded34c82abab54})
2021-01-19 02:13:47,117 DEBUG
org.apache.flink.runtime.jobmaster.slotpool.SlotPoolImpl     [] - Fulfilling
pending slot request [SlotRequestId{88a7127cc4a86be0a962b9aa68d4feff}] with
slot [a30c9937af2c6de7ab471086cc9268f5]
2021-01-19 02:13:47,117 DEBUG org.apache.flink.runtime.scheduler.SharedSlot     
          
[] - Allocated logical slot
(SlotRequestId{da65200b5d50dcaaaff3f4373dd824c4}) for execution vertex (id
605b35e407e90cda15ad084365733fdd_421) from the physical slot
(SlotRequestId{88a7127cc4a86be0a962b9aa68d4feff})
2021-01-19 02:13:47,117 INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph       [] - Source:
HiveSource-snmpprobe.p_snmp_ifXTable (422/458)
(f12be47a0d11892e411d1afcb928b55a) switched from SCHEDULED to DEPLOYING.
2021-01-19 02:13:47,117 INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph       [] - Deploying
Source: HiveSource-snmpprobe.p_snmp_ifXTable (422/458) (attempt #0) with
attempt id f12be47a0d11892e411d1afcb928b55a to 172.0.37.8:6122-694869 @
172-0-37-8.flink-taskmanager-query-state.gem-flink.svc.cluster.local
(dataPort=32959) with allocation id a30c9937af2c6de7ab471086cc9268f5
2021-01-19 02:13:47,117 DEBUG
org.apache.flink.runtime.resourcemanager.slotmanager.SlotManagerImpl [] -
Cancel slot request 0a1cfa83b0d664615e0b9e1f938d7dee.
2021-01-19 02:13:47,117 DEBUG
org.apache.flink.runtime.resourcemanager.slotmanager.SlotManagerImpl [] -
Cancel slot request c1d63c4ffdf6e17212c4ca6be4071850.
2021-01-19 02:13:47,120 DEBUG
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager [] -
WebSocket successfully opened
2021-01-19 02:13:47,123 ERROR
org.apache.flink.runtime.resourcemanager.StandaloneResourceManager [] -
Fatal error occurred in ResourceManager.
org.apache.flink.runtime.leaderretrieval.LeaderRetrievalException: Error
while watching the ConfigMap
test-flink-etl-cb1c647ea7488765fd3e8cc1dc691e46-jobmanager-leader
        at
org.apache.flink.kubernetes.highavailability.KubernetesLeaderRetrievalDriver$ConfigMapCallbackHandlerImpl.handleFatalError(KubernetesLeaderRetrievalDriver.java:120)
[flink-dist_2.11-1.12.1.jar:1.12.1]
        at
org.apache.flink.kubernetes.kubeclient.resources.AbstractKubernetesWatcher.onClose(AbstractKubernetesWatcher.java:48)
[flink-dist_2.11-1.12.1.jar:1.12.1]
        at
io.fabric8.kubernetes.client.utils.WatcherToggle.onClose(WatcherToggle.java:56)
[flink-dist_2.11-1.12.1.jar:1.12.1]
        at
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager.closeEvent(WatchConnectionManager.java:367)
[flink-dist_2.11-1.12.1.jar:1.12.1]
        at
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager.access$700(WatchConnectionManager.java:50)
[flink-dist_2.11-1.12.1.jar:1.12.1]
        at
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager$1.onMessage(WatchConnectionManager.java:259)
[flink-dist_2.11-1.12.1.jar:1.12.1]
        at
org.apache.flink.kubernetes.shaded.okhttp3.internal.ws.RealWebSocket.onReadMessage(RealWebSocket.java:323)
[flink-dist_2.11-1.12.1.jar:1.12.1]
        at
org.apache.flink.kubernetes.shaded.okhttp3.internal.ws.WebSocketReader.readMessageFrame(WebSocketReader.java:219)
[flink-dist_2.11-1.12.1.jar:1.12.1]
        at
org.apache.flink.kubernetes.shaded.okhttp3.internal.ws.WebSocketReader.processNextFrame(WebSocketReader.java:105)
[flink-dist_2.11-1.12.1.jar:1.12.1]
        at
org.apache.flink.kubernetes.shaded.okhttp3.internal.ws.RealWebSocket.loopReader(RealWebSocket.java:274)
[flink-dist_2.11-1.12.1.jar:1.12.1]
        at
org.apache.flink.kubernetes.shaded.okhttp3.internal.ws.RealWebSocket$2.onResponse(RealWebSocket.java:214)
[flink-dist_2.11-1.12.1.jar:1.12.1]
        at
org.apache.flink.kubernetes.shaded.okhttp3.RealCall$AsyncCall.execute(RealCall.java:206)
[flink-dist_2.11-1.12.1.jar:1.12.1]
        at
org.apache.flink.kubernetes.shaded.okhttp3.internal.NamedRunnable.run(NamedRunnable.java:32)
[flink-dist_2.11-1.12.1.jar:1.12.1]
        at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
[?:1.8.0_275]
        at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
[?:1.8.0_275]
        at java.lang.Thread.run(Thread.java:748) [?:1.8.0_275]
2021-01-19 02:13:47,124 ERROR
org.apache.flink.runtime.entrypoint.ClusterEntrypoint        [] - Fatal
error occurred in the cluster entrypoint.
org.apache.flink.runtime.leaderretrieval.LeaderRetrievalException: Error
while watching the ConfigMap
test-flink-etl-cb1c647ea7488765fd3e8cc1dc691e46-jobmanager-leader
        at
org.apache.flink.kubernetes.highavailability.KubernetesLeaderRetrievalDriver$ConfigMapCallbackHandlerImpl.handleFatalError(KubernetesLeaderRetrievalDriver.java:120)
[flink-dist_2.11-1.12.1.jar:1.12.1]
        at
org.apache.flink.kubernetes.kubeclient.resources.AbstractKubernetesWatcher.onClose(AbstractKubernetesWatcher.java:48)
[flink-dist_2.11-1.12.1.jar:1.12.1]
        at
io.fabric8.kubernetes.client.utils.WatcherToggle.onClose(WatcherToggle.java:56)
[flink-dist_2.11-1.12.1.jar:1.12.1]
        at
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager.closeEvent(WatchConnectionManager.java:367)
[flink-dist_2.11-1.12.1.jar:1.12.1]
        at
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager.access$700(WatchConnectionManager.java:50)
[flink-dist_2.11-1.12.1.jar:1.12.1]
        at
io.fabric8.kubernetes.client.dsl.internal.WatchConnectionManager$1.onMessage(WatchConnectionManager.java:259)
[flink-dist_2.11-1.12.1.jar:1.12.1]
        at
org.apache.flink.kubernetes.shaded.okhttp3.internal.ws.RealWebSocket.onReadMessage(RealWebSocket.java:323)
[flink-dist_2.11-1.12.1.jar:1.12.1]
        at
org.apache.flink.kubernetes.shaded.okhttp3.internal.ws.WebSocketReader.readMessageFrame(WebSocketReader.java:219)
[flink-dist_2.11-1.12.1.jar:1.12.1]
        at
org.apache.flink.kubernetes.shaded.okhttp3.internal.ws.WebSocketReader.processNextFrame(WebSocketReader.java:105)
[flink-dist_2.11-1.12.1.jar:1.12.1]
        at
org.apache.flink.kubernetes.shaded.okhttp3.internal.ws.RealWebSocket.loopReader(RealWebSocket.java:274)
[flink-dist_2.11-1.12.1.jar:1.12.1]
        at
org.apache.flink.kubernetes.shaded.okhttp3.internal.ws.RealWebSocket$2.onResponse(RealWebSocket.java:214)
[flink-dist_2.11-1.12.1.jar:1.12.1]
        at
org.apache.flink.kubernetes.shaded.okhttp3.RealCall$AsyncCall.execute(RealCall.java:206)
[flink-dist_2.11-1.12.1.jar:1.12.1]
        at
org.apache.flink.kubernetes.shaded.okhttp3.internal.NamedRunnable.run(NamedRunnable.java:32)
[flink-dist_2.11-1.12.1.jar:1.12.1]
        at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
[?:1.8.0_275]
        at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
[?:1.8.0_275]
        at java.lang.Thread.run(Thread.java:748) [?:1.8.0_275]
2021-01-19 02:13:47,125 DEBUG
org.apache.flink.runtime.source.coordinator.SourceCoordinator [] - Handling
event from subtask 365 of source Source:
HiveSource-snmpprobe.p_snmp_ifXTable: ReaderRegistrationEvent[subtaskId =
365, location = 172.0.37.16)
2021-01-19 02:13:47,125 DEBUG
org.apache.flink.runtime.source.coordinator.SourceCoordinator [] - Handling
event from subtask 365 of source Source:
HiveSource-snmpprobe.p_snmp_ifXTable: RequestSplitEvent (host='172.0.37.16')
2021-01-19 02:13:47,125 INFO 
org.apache.flink.connector.file.src.impl.StaticFileSplitEnumerator [] -
Subtask 365 (on host '172.0.37.16') is requesting a file source split
2021-01-19 02:13:47,125 INFO 
org.apache.flink.connector.file.src.impl.StaticFileSplitEnumerator [] - No
more splits available for subtask 365
2021-01-19 02:13:47,125 DEBUG
org.apache.flink.runtime.source.coordinator.SourceCoordinator [] - Handling
event from subtask 379 of source Source:
HiveSource-snmpprobe.p_snmp_ifXTable: ReaderRegistrationEvent[subtaskId =
379, location = 172.0.37.16)
2021-01-19 02:13:47,125 DEBUG
org.apache.flink.runtime.source.coordinator.SourceCoordinator [] - Handling
event from subtask 379 of source Source:
HiveSource-snmpprobe.p_snmp_ifXTable: RequestSplitEvent (host='172.0.37.16')
2021-01-19 02:13:47,125 INFO 
org.apache.flink.connector.file.src.impl.StaticFileSplitEnumerator [] -
Subtask 379 (on host '172.0.37.16') is requesting a file source split
2021-01-19 02:13:47,125 INFO 
org.apache.flink.connector.file.src.impl.StaticFileSplitEnumerator [] - No
more splits available for subtask 379
2021-01-19 02:13:47,130 INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph       [] - Source:
HiveSource-snmpprobe.p_snmp_ifXTable (389/458)
(b0d8b877b1911ffca609f818693b68ad) switched from DEPLOYING to RUNNING.
2021-01-19 02:13:47,131 INFO  org.apache.flink.runtime.blob.BlobServer          
          
[] - Stopped BLOB server at 0.0.0.0:6124
2021-01-19 02:13:47,132 INFO 
org.apache.flink.runtime.executiongraph.ExecutionGraph       [] - Source:
HiveSource-snmpprobe.p_snmp_ifXTable (421/458)
(706d012e7a572e1e5786536df9ab3bbb) switched from DEPLOYING to RUNNING.





--
Sent from: http://apache-flink.147419.n8.nabble.com/

回复