Hi,
I couldn't make a savepoint for the following graph:
[image: image.png]
with stacktrace:
Caused by: java.util.concurrent.CompletionException:
java.util.concurrent.CompletionException:
org.apache.flink.runtime.checkpoint.CheckpointException: Failed to trigger
savepoint. Failure reason: Not all required tasks are currently running.
Here is my Snowflake source definition:
val batch = env.createInput(JDBCInputFormat.buildJDBCInputFormat
.setDrivername(options.driverName)
.setDBUrl(options.dbUrl)
.setUsername(options.username)
.setPassword(options.password)
.setQuery(query)
.setRowTypeInfo(getInputRowTypeInfo)
.setFetchSize(fetchSize)
.setParametersProvider(new
GenericParameterValuesProvider(buildQueryParams(parallelism)))
.finish, getReturnType)
where query is something like
select * from events where timestamp > t0 and timestamp < t1
My theory is that the snowflake_batch_source task changed to FINISHED once
it reads all the data. and then savepoint failed.
Is there any way to make a savepoint for such cases?
Thanks,
Fanbin