[ 
https://issues.apache.org/jira/browse/FLINK-2111?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15093854#comment-15093854
 ] 

ASF GitHub Bot commented on FLINK-2111:
---------------------------------------

Github user tillrohrmann commented on a diff in the pull request:

    https://github.com/apache/flink/pull/750#discussion_r49451808
  
    --- Diff: 
flink-runtime/src/test/java/org/apache/flink/runtime/executiongraph/ExecutionGraphSignalsTest.java
 ---
    @@ -0,0 +1,223 @@
    +/*
    + * Licensed to the Apache Software Foundation (ASF) under one
    + * or more contributor license agreements.  See the NOTICE file
    + * distributed with this work for additional information
    + * regarding copyright ownership.  The ASF licenses this file
    + * to you under the Apache License, Version 2.0 (the
    + * "License"); you may not use this file except in compliance
    + * with the License.  You may obtain a copy of the License at
    + *
    + *     http://www.apache.org/licenses/LICENSE-2.0
    + *
    + * Unless required by applicable law or agreed to in writing, software
    + * distributed under the License is distributed on an "AS IS" BASIS,
    + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
    + * See the License for the specific language governing permissions and
    + * limitations under the License.
    + */
    +package org.apache.flink.runtime.executiongraph;
    +
    +import java.lang.reflect.Field;
    +import java.util.ArrayList;
    +import java.util.Arrays;
    +import java.util.List;
    +
    +import org.apache.flink.api.common.JobID;
    +import org.apache.flink.api.common.JobType;
    +import org.apache.flink.configuration.Configuration;
    +import org.apache.flink.runtime.StoppingException;
    +import org.apache.flink.runtime.akka.AkkaUtils;
    +import org.apache.flink.runtime.jobgraph.JobVertex;
    +import org.apache.flink.runtime.jobgraph.DistributionPattern;
    +import org.apache.flink.runtime.jobgraph.JobStatus;
    +import org.apache.flink.runtime.testingUtils.TestingUtils;
    +import org.junit.Assert;
    +import org.junit.Before;
    +import org.junit.Test;
    +import org.junit.runner.RunWith;
    +import org.powermock.core.classloader.annotations.PrepareForTest;
    +import org.powermock.modules.junit4.PowerMockRunner;
    +import org.powermock.api.mockito.PowerMockito;
    +
    +import scala.concurrent.duration.FiniteDuration;
    +import static org.mockito.Mockito.any;
    +import static org.mockito.Mockito.mock;
    +import static org.mockito.Mockito.same;
    +import static org.mockito.Mockito.times;
    +import static org.mockito.Mockito.verify;
    +import static org.mockito.Mockito.when;
    +
    +@RunWith(PowerMockRunner.class)
    +@PrepareForTest(ExecutionGraph.class)
    +public class ExecutionGraphSignalsTest {
    +   private ExecutionJobVertex[] mockEJV = new ExecutionJobVertex[5];
    +   private int[] dop = new int[] { 5, 7, 2, 11, 4 };
    +   private ExecutionVertex[][] mockEV = new 
ExecutionVertex[mockEJV.length][];
    +   private ExecutionGraph eg;
    +   private Field f;
    +
    +   @Before
    +   public void prepare() throws Exception {
    +           final JobID jobId = new JobID();
    +           final String jobName = "Test Job Sample Name";
    +           final Configuration cfg = new Configuration();
    +
    +
    +           assert (mockEJV.length == 5);
    +           JobVertex v1 = new JobVertex("vertex1");
    +           JobVertex v2 = new JobVertex("vertex2");
    +           JobVertex v3 = new JobVertex("vertex3");
    +           JobVertex v4 = new JobVertex("vertex4");
    +           JobVertex v5 = new JobVertex("vertex5");
    +
    +           for(int i = 0; i < mockEJV.length; ++i) {
    +                   mockEJV[i] = mock(ExecutionJobVertex.class);
    +
    +                   this.mockEV[i] = new ExecutionVertex[dop[i]];
    +                   for (int j = 0; j < dop[i]; ++j) {
    +                           this.mockEV[i][j] = mock(ExecutionVertex.class);
    +                   }
    +
    +                   when(mockEJV[i].getProducedDataSets()).thenReturn(new 
IntermediateResult[0]);
    +                   
when(mockEJV[i].getTaskVertices()).thenReturn(this.mockEV[i]);
    +           }
    +
    +           PowerMockito
    +                   .whenNew(ExecutionJobVertex.class)
    +                   .withArguments(any(ExecutionGraph.class), same(v1), 
any(Integer.class).intValue(),
    +                           any(FiniteDuration.class), 
any(Long.class).longValue()).thenReturn(mockEJV[0]);
    +           PowerMockito
    +                   .whenNew(ExecutionJobVertex.class)
    +                   .withArguments(any(ExecutionGraph.class), same(v2), 
any(Integer.class).intValue(),
    +                           any(FiniteDuration.class), 
any(Long.class).longValue()).thenReturn(mockEJV[1]);
    +           PowerMockito
    +                   .whenNew(ExecutionJobVertex.class)
    +                   .withArguments(any(ExecutionGraph.class), same(v3), 
any(Integer.class).intValue(),
    +                           any(FiniteDuration.class), 
any(Long.class).longValue()).thenReturn(mockEJV[2]);
    +           PowerMockito
    +                   .whenNew(ExecutionJobVertex.class)
    +                   .withArguments(any(ExecutionGraph.class), same(v4), 
any(Integer.class).intValue(),
    +                           any(FiniteDuration.class), 
any(Long.class).longValue()).thenReturn(mockEJV[3]);
    +           PowerMockito
    +                   .whenNew(ExecutionJobVertex.class)
    +                   .withArguments(any(ExecutionGraph.class), same(v5), 
any(Integer.class).intValue(),
    +                           any(FiniteDuration.class), 
any(Long.class).longValue()).thenReturn(mockEJV[4]);
    +
    +           v1.setParallelism(dop[0]);
    +           v2.setParallelism(dop[1]);
    +           v3.setParallelism(dop[2]);
    +           v4.setParallelism(dop[3]);
    +           v5.setParallelism(dop[4]);
    +
    +           v2.connectNewDataSetAsInput(v1, DistributionPattern.ALL_TO_ALL);
    +           mockNumberOfInputs(1,0);
    +           v4.connectNewDataSetAsInput(v2, DistributionPattern.ALL_TO_ALL);
    +           mockNumberOfInputs(3,1);
    +           v4.connectNewDataSetAsInput(v3, DistributionPattern.ALL_TO_ALL);
    +           mockNumberOfInputs(3,2);
    +           v5.connectNewDataSetAsInput(v4, DistributionPattern.ALL_TO_ALL);
    +           mockNumberOfInputs(4,3);
    +           v5.connectNewDataSetAsInput(v3, DistributionPattern.ALL_TO_ALL);
    +           mockNumberOfInputs(4,2);
    +
    +           List<JobVertex> ordered = new 
ArrayList<JobVertex>(Arrays.asList(v1, v2, v3, v4, v5));
    +
    +           eg = new ExecutionGraph(TestingUtils.defaultExecutionContext(), 
jobId, jobName,
    +                           JobType.BATCHING, cfg, 
AkkaUtils.getDefaultTimeout());
    +           eg.attachJobGraph(ordered);
    +
    +           f = eg.getClass().getDeclaredField("state");
    +           f.setAccessible(true);
    +   }
    +
    +   private void mockNumberOfInputs(int nodeIndex, int predecessorIndex) {
    +           for(int j = 0; j < dop[nodeIndex]; ++j) {
    +                   
when(mockEV[nodeIndex][j].getNumberOfInputs()).thenReturn(dop[predecessorIndex]);
    +           }
    +   }
    +
    +   @Test
    +   public void testCancel() throws Exception {
    +           Assert.assertEquals(JobStatus.CREATED, eg.getState());
    +           eg.cancel();
    +
    +           verifyCancel(1);
    +
    +           f.set(eg, JobStatus.RUNNING);
    +           eg.cancel();
    +
    +           verifyCancel(2);
    +           Assert.assertEquals(JobStatus.CANCELLING, eg.getState());
    +
    +           eg.cancel();
    +
    +           verifyCancel(2);
    +           Assert.assertEquals(JobStatus.CANCELLING, eg.getState());
    +
    +           f.set(eg, JobStatus.CANCELED);
    +           eg.cancel();
    +
    +           verifyCancel(2);
    +           Assert.assertEquals(JobStatus.CANCELED, eg.getState());
    +
    +           f.set(eg, JobStatus.FAILED);
    +           eg.cancel();
    +
    +           verifyCancel(2);
    +           Assert.assertEquals(JobStatus.FAILED, eg.getState());
    +
    +           f.set(eg, JobStatus.FAILING);
    +           eg.cancel();
    +
    +           verifyCancel(2);
    +           Assert.assertEquals(JobStatus.CANCELLING, eg.getState());
    +
    +           f.set(eg, JobStatus.FINISHED);
    +           eg.cancel();
    +
    +           verifyCancel(2);
    +           Assert.assertEquals(JobStatus.FINISHED, eg.getState());
    +
    +           f.set(eg, JobStatus.RESTARTING);
    +           eg.cancel();
    +
    +           verifyCancel(2);
    +           Assert.assertEquals(JobStatus.CANCELED, eg.getState());
    +   }
    +
    +   private void verifyCancel(int times) {
    +           for (int i = 0; i < mockEJV.length; ++i) {
    +                   verify(mockEJV[i], times(times)).cancel();
    +           }
    +
    +   }
    +
    +   @Test
    +   public void testStop() throws Exception {
    --- End diff --
    
    A short comment what we're testing here would be helpful


> Add "stop" signal to cleanly shutdown streaming jobs
> ----------------------------------------------------
>
>                 Key: FLINK-2111
>                 URL: https://issues.apache.org/jira/browse/FLINK-2111
>             Project: Flink
>          Issue Type: Improvement
>          Components: Distributed Runtime, JobManager, Local Runtime, 
> Streaming, TaskManager, Webfrontend
>            Reporter: Matthias J. Sax
>            Assignee: Matthias J. Sax
>            Priority: Minor
>
> Currently, streaming jobs can only be stopped using "cancel" command, what is 
> a "hard" stop with no clean shutdown.
> The new introduced "stop" signal, will only affect streaming source tasks 
> such that the sources can stop emitting data and shutdown cleanly, resulting 
> in a clean shutdown of the whole streaming job.
> This feature is a pre-requirment for 
> https://issues.apache.org/jira/browse/FLINK-1929



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to