This is an automated email from the ASF dual-hosted git repository. sjwiesman pushed a commit to branch master in repository https://gitbox.apache.org/repos/asf/flink.git
commit 993289dcd12e9748492bccf84e622f6aa46a51dd Author: sjwiesman <sjwies...@gmail.com> AuthorDate: Mon Aug 16 10:45:07 2021 -0500 [hotfix][docs] replace references to MemoryStateBackend with HashMapStateBackend --- docs/content.zh/docs/libs/state_processor_api.md | 8 ++++---- docs/content/docs/libs/state_processor_api.md | 8 ++++---- 2 files changed, 8 insertions(+), 8 deletions(-) diff --git a/docs/content.zh/docs/libs/state_processor_api.md b/docs/content.zh/docs/libs/state_processor_api.md index 1db31e8..f7026d1 100644 --- a/docs/content.zh/docs/libs/state_processor_api.md +++ b/docs/content.zh/docs/libs/state_processor_api.md @@ -78,7 +78,7 @@ The compatibility guarantees for restoring state are identical to those when res ```java ExecutionEnvironment bEnv = ExecutionEnvironment.getExecutionEnvironment(); -ExistingSavepoint savepoint = Savepoint.load(bEnv, "hdfs://path/", new MemoryStateBackend()); +ExistingSavepoint savepoint = Savepoint.load(bEnv, "hdfs://path/", new HashMapStateBackend()); ``` @@ -302,7 +302,7 @@ class ClickReader extends WindowReaderFunction<Integer, ClickState, String, Time } ExecutionEnvironment batchEnv = ExecutionEnvironment.getExecutionEnvironment(); -ExistingSavepoint savepoint = Savepoint.load(batchEnv, "hdfs://checkpoint-dir", new MemoryStateBackend()); +ExistingSavepoint savepoint = Savepoint.load(batchEnv, "hdfs://checkpoint-dir", new HashMapStateBackend()); savepoint .window(TumblingEventTimeWindows.of(Time.minutes(1))) @@ -329,7 +329,7 @@ a savepoint for the Scala DataStream API please manually pass in all type inform int maxParallelism = 128; Savepoint - .create(new MemoryStateBackend(), maxParallelism) + .create(new HashMapStateBackend(), maxParallelism) .withOperator("uid1", transformation1) .withOperator("uid2", transformation2) .write(savepointPath); @@ -478,7 +478,7 @@ Besides creating a savepoint from scratch, you can base one off an existing save ```java Savepoint - .load(bEnv, new MemoryStateBackend(), oldPath) + .load(bEnv, new HashMapStateBackend(), oldPath) .withOperator("uid", transformation) .write(newPath); ``` diff --git a/docs/content/docs/libs/state_processor_api.md b/docs/content/docs/libs/state_processor_api.md index a6dd657..31e1a87 100644 --- a/docs/content/docs/libs/state_processor_api.md +++ b/docs/content/docs/libs/state_processor_api.md @@ -78,7 +78,7 @@ The compatibility guarantees for restoring state are identical to those when res ```java ExecutionEnvironment bEnv = ExecutionEnvironment.getExecutionEnvironment(); -ExistingSavepoint savepoint = Savepoint.load(bEnv, "hdfs://path/", new MemoryStateBackend()); +ExistingSavepoint savepoint = Savepoint.load(bEnv, "hdfs://path/", new HashMapStateBackend()); ``` @@ -302,7 +302,7 @@ class ClickReader extends WindowReaderFunction<Integer, ClickState, String, Time } ExecutionEnvironment batchEnv = ExecutionEnvironment.getExecutionEnvironment(); -ExistingSavepoint savepoint = Savepoint.load(batchEnv, "hdfs://checkpoint-dir", new MemoryStateBackend()); +ExistingSavepoint savepoint = Savepoint.load(batchEnv, "hdfs://checkpoint-dir", new HashMapStateBackend()); savepoint .window(TumblingEventTimeWindows.of(Time.minutes(1))) @@ -329,7 +329,7 @@ a savepoint for the Scala DataStream API please manually pass in all type inform int maxParallelism = 128; Savepoint - .create(new MemoryStateBackend(), maxParallelism) + .create(new HashMapStateBackend(), maxParallelism) .withOperator("uid1", transformation1) .withOperator("uid2", transformation2) .write(savepointPath); @@ -478,7 +478,7 @@ Besides creating a savepoint from scratch, you can base one off an existing save ```java Savepoint - .load(bEnv, new MemoryStateBackend(), oldPath) + .load(bEnv, new HashMapStateBackend(), oldPath) .withOperator("uid", transformation) .write(newPath); ```