[ https://issues.apache.org/jira/browse/FLINK-25940?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17489996#comment-17489996 ]
Yun Gao commented on FLINK-25940: --------------------------------- https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=31000&view=logs&j=9cada3cb-c1d3-5621-16da-0f718fb86602&t=c67e71ed-6451-5d26-8920-5a8cf9651901&l=24147 > pyflink/datastream/tests/test_data_stream.py::StreamingModeDataStreamTests::test_keyed_process_function_with_state > failed on AZP > -------------------------------------------------------------------------------------------------------------------------------- > > Key: FLINK-25940 > URL: https://issues.apache.org/jira/browse/FLINK-25940 > Project: Flink > Issue Type: Bug > Components: API / Python > Affects Versions: 1.15.0 > Reporter: Till Rohrmann > Assignee: Huang Xingbo > Priority: Critical > Labels: test-stability > > The test > {{pyflink/datastream/tests/test_data_stream.py::StreamingModeDataStreamTests::test_keyed_process_function_with_state}} > fails on AZP: > {code} > 2022-02-02T17:44:12.1898582Z Feb 02 17:44:12 > =================================== FAILURES > =================================== > 2022-02-02T17:44:12.1899860Z Feb 02 17:44:12 _____ > StreamingModeDataStreamTests.test_keyed_process_function_with_state ______ > 2022-02-02T17:44:12.1900493Z Feb 02 17:44:12 > 2022-02-02T17:44:12.1901218Z Feb 02 17:44:12 self = > <pyflink.datastream.tests.test_data_stream.StreamingModeDataStreamTests > testMethod=test_keyed_process_function_with_state> > 2022-02-02T17:44:12.1901948Z Feb 02 17:44:12 > 2022-02-02T17:44:12.1902745Z Feb 02 17:44:12 def > test_keyed_process_function_with_state(self): > 2022-02-02T17:44:12.1903722Z Feb 02 17:44:12 > self.env.get_config().set_auto_watermark_interval(2000) > 2022-02-02T17:44:12.1904473Z Feb 02 17:44:12 > self.env.set_stream_time_characteristic(TimeCharacteristic.EventTime) > 2022-02-02T17:44:12.1906780Z Feb 02 17:44:12 data_stream = > self.env.from_collection([(1, 'hi', '1603708211000'), > 2022-02-02T17:44:12.1908034Z Feb 02 17:44:12 > (2, 'hello', '1603708224000'), > 2022-02-02T17:44:12.1909166Z Feb 02 17:44:12 > (3, 'hi', '1603708226000'), > 2022-02-02T17:44:12.1910122Z Feb 02 17:44:12 > (4, 'hello', '1603708289000'), > 2022-02-02T17:44:12.1911099Z Feb 02 17:44:12 > (5, 'hi', '1603708291000'), > 2022-02-02T17:44:12.1912451Z Feb 02 17:44:12 > (6, 'hello', '1603708293000')], > 2022-02-02T17:44:12.1913456Z Feb 02 17:44:12 > type_info=Types.ROW([Types.INT(), Types.STRING(), > 2022-02-02T17:44:12.1914338Z Feb 02 17:44:12 > Types.STRING()])) > 2022-02-02T17:44:12.1914811Z Feb 02 17:44:12 > 2022-02-02T17:44:12.1915317Z Feb 02 17:44:12 class > MyTimestampAssigner(TimestampAssigner): > 2022-02-02T17:44:12.1915724Z Feb 02 17:44:12 > 2022-02-02T17:44:12.1916782Z Feb 02 17:44:12 def > extract_timestamp(self, value, record_timestamp) -> int: > 2022-02-02T17:44:12.1917621Z Feb 02 17:44:12 return > int(value[2]) > 2022-02-02T17:44:12.1918262Z Feb 02 17:44:12 > 2022-02-02T17:44:12.1918855Z Feb 02 17:44:12 class > MyProcessFunction(KeyedProcessFunction): > 2022-02-02T17:44:12.1919363Z Feb 02 17:44:12 > 2022-02-02T17:44:12.1919744Z Feb 02 17:44:12 def __init__(self): > 2022-02-02T17:44:12.1920143Z Feb 02 17:44:12 self.value_state > = None > 2022-02-02T17:44:12.1920648Z Feb 02 17:44:12 self.list_state > = None > 2022-02-02T17:44:12.1921298Z Feb 02 17:44:12 self.map_state = > None > 2022-02-02T17:44:12.1921864Z Feb 02 17:44:12 > 2022-02-02T17:44:12.1922479Z Feb 02 17:44:12 def open(self, > runtime_context: RuntimeContext): > 2022-02-02T17:44:12.1923907Z Feb 02 17:44:12 > value_state_descriptor = ValueStateDescriptor('value_state', Types.INT()) > 2022-02-02T17:44:12.1924922Z Feb 02 17:44:12 self.value_state > = runtime_context.get_state(value_state_descriptor) > 2022-02-02T17:44:12.1925741Z Feb 02 17:44:12 > list_state_descriptor = ListStateDescriptor('list_state', Types.INT()) > 2022-02-02T17:44:12.1926482Z Feb 02 17:44:12 self.list_state > = runtime_context.get_list_state(list_state_descriptor) > 2022-02-02T17:44:12.1927465Z Feb 02 17:44:12 > map_state_descriptor = MapStateDescriptor('map_state', Types.INT(), > Types.STRING()) > 2022-02-02T17:44:12.1927998Z Feb 02 17:44:12 state_ttl_config > = StateTtlConfig \ > 2022-02-02T17:44:12.1928444Z Feb 02 17:44:12 > .new_builder(Time.seconds(1)) \ > 2022-02-02T17:44:12.1928943Z Feb 02 17:44:12 > .set_update_type(StateTtlConfig.UpdateType.OnReadAndWrite) \ > 2022-02-02T17:44:12.1929462Z Feb 02 17:44:12 > .set_state_visibility( > 2022-02-02T17:44:12.1929939Z Feb 02 17:44:12 > StateTtlConfig.StateVisibility.ReturnExpiredIfNotCleanedUp) \ > 2022-02-02T17:44:12.1930601Z Feb 02 17:44:12 > .disable_cleanup_in_background() \ > 2022-02-02T17:44:12.1931032Z Feb 02 17:44:12 .build() > 2022-02-02T17:44:12.1931480Z Feb 02 17:44:12 > map_state_descriptor.enable_time_to_live(state_ttl_config) > 2022-02-02T17:44:12.1932018Z Feb 02 17:44:12 self.map_state = > runtime_context.get_map_state(map_state_descriptor) > 2022-02-02T17:44:12.1932610Z Feb 02 17:44:12 > 2022-02-02T17:44:12.1933172Z Feb 02 17:44:12 def > process_element(self, value, ctx): > 2022-02-02T17:44:12.1933623Z Feb 02 17:44:12 import time > 2022-02-02T17:44:12.1934007Z Feb 02 17:44:12 time.sleep(1) > 2022-02-02T17:44:12.1934419Z Feb 02 17:44:12 current_value = > self.value_state.value() > 2022-02-02T17:44:12.1934977Z Feb 02 17:44:12 > self.value_state.update(value[0]) > 2022-02-02T17:44:12.1935451Z Feb 02 17:44:12 current_list = > [_ for _ in self.list_state.get()] > 2022-02-02T17:44:12.1935921Z Feb 02 17:44:12 > self.list_state.add(value[0]) > 2022-02-02T17:44:12.1936401Z Feb 02 17:44:12 map_entries = > {k: v for k, v in self.map_state.items()} > 2022-02-02T17:44:12.1936862Z Feb 02 17:44:12 keys = > sorted(map_entries.keys()) > 2022-02-02T17:44:12.1937649Z Feb 02 17:44:12 > map_entries_string = [str(k) + ': ' + str(map_entries[k]) for k in keys] > 2022-02-02T17:44:12.1938404Z Feb 02 17:44:12 > map_entries_string = '{' + ', '.join(map_entries_string) + '}' > 2022-02-02T17:44:12.1938906Z Feb 02 17:44:12 > self.map_state.put(value[0], value[1]) > 2022-02-02T17:44:12.1939350Z Feb 02 17:44:12 current_key = > ctx.get_current_key() > 2022-02-02T17:44:12.1939889Z Feb 02 17:44:12 yield "current > key: {}, current value state: {}, current list state: {}, " \ > 2022-02-02T17:44:12.1940521Z Feb 02 17:44:12 "current > map state: {}, current value: {}".format(str(current_key), > 2022-02-02T17:44:12.1941111Z Feb 02 17:44:12 > str(current_value), > 2022-02-02T17:44:12.1941645Z Feb 02 17:44:12 > str(current_list), > 2022-02-02T17:44:12.1942254Z Feb 02 17:44:12 > map_entries_string, > 2022-02-02T17:44:12.1942796Z Feb 02 17:44:12 > str(value)) > 2022-02-02T17:44:12.1943369Z Feb 02 17:44:12 > 2022-02-02T17:44:12.1943761Z Feb 02 17:44:12 def on_timer(self, > timestamp, ctx): > 2022-02-02T17:44:12.1944178Z Feb 02 17:44:12 pass > 2022-02-02T17:44:12.1944503Z Feb 02 17:44:12 > 2022-02-02T17:44:12.1944898Z Feb 02 17:44:12 watermark_strategy = > WatermarkStrategy.for_monotonous_timestamps() \ > 2022-02-02T17:44:12.1945537Z Feb 02 17:44:12 > .with_timestamp_assigner(MyTimestampAssigner()) > 2022-02-02T17:44:12.1946018Z Feb 02 17:44:12 > data_stream.assign_timestamps_and_watermarks(watermark_strategy) \ > 2022-02-02T17:44:12.1946525Z Feb 02 17:44:12 .key_by(lambda x: > x[1], key_type=Types.STRING()) \ > 2022-02-02T17:44:12.1947019Z Feb 02 17:44:12 > .process(MyProcessFunction(), output_type=Types.STRING()) \ > 2022-02-02T17:44:12.1947465Z Feb 02 17:44:12 > .add_sink(self.test_sink) > 2022-02-02T17:44:12.1948146Z Feb 02 17:44:12 self.env.execute('test > time stamp assigner with keyed process function') > 2022-02-02T17:44:12.1948637Z Feb 02 17:44:12 results = > self.test_sink.get_results() > 2022-02-02T17:44:12.1949166Z Feb 02 17:44:12 expected = ["current > key: hi, current value state: None, current list state: [], " > 2022-02-02T17:44:12.1949957Z Feb 02 17:44:12 "current map > state: {}, current value: Row(f0=1, f1='hi', " > 2022-02-02T17:44:12.1950624Z Feb 02 17:44:12 > "f2='1603708211000')", > 2022-02-02T17:44:12.1951234Z Feb 02 17:44:12 "current > key: hello, current value state: None, " > 2022-02-02T17:44:12.1951822Z Feb 02 17:44:12 "current > list state: [], current map state: {}, current value: Row(f0=2," > 2022-02-02T17:44:12.1952596Z Feb 02 17:44:12 " > f1='hello', f2='1603708224000')", > 2022-02-02T17:44:12.1953292Z Feb 02 17:44:12 "current > key: hi, current value state: 1, current list state: [1], " > 2022-02-02T17:44:12.1954134Z Feb 02 17:44:12 "current map > state: {1: hi}, current value: Row(f0=3, f1='hi', " > 2022-02-02T17:44:12.1954799Z Feb 02 17:44:12 > "f2='1603708226000')", > 2022-02-02T17:44:12.1955331Z Feb 02 17:44:12 "current > key: hello, current value state: 2, current list state: [2], " > 2022-02-02T17:44:12.1956145Z Feb 02 17:44:12 "current map > state: {2: hello}, current value: Row(f0=4, f1='hello', " > 2022-02-02T17:44:12.1956826Z Feb 02 17:44:12 > "f2='1603708289000')", > 2022-02-02T17:44:12.1957362Z Feb 02 17:44:12 "current > key: hi, current value state: 3, current list state: [1, 3], " > 2022-02-02T17:44:12.1958156Z Feb 02 17:44:12 "current map > state: {1: hi, 3: hi}, current value: Row(f0=5, f1='hi', " > 2022-02-02T17:44:12.1958845Z Feb 02 17:44:12 > "f2='1603708291000')", > 2022-02-02T17:44:12.1959382Z Feb 02 17:44:12 "current > key: hello, current value state: 4, current list state: [2, 4]," > 2022-02-02T17:44:12.1960011Z Feb 02 17:44:12 " current > map state: {2: hello, 4: hello}, current value: Row(f0=6, " > 2022-02-02T17:44:12.1960715Z Feb 02 17:44:12 "f1='hello', > f2='1603708293000')"] > 2022-02-02T17:44:12.1961159Z Feb 02 17:44:12 > > self.assert_equals_sorted(expected, results) > 2022-02-02T17:44:12.1961533Z Feb 02 17:44:12 > 2022-02-02T17:44:12.1961906Z Feb 02 17:44:12 > pyflink/datastream/tests/test_data_stream.py:683: > 2022-02-02T17:44:12.1962464Z Feb 02 17:44:12 _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ > _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ > 2022-02-02T17:44:12.1963186Z Feb 02 17:44:12 > pyflink/datastream/tests/test_data_stream.py:62: in assert_equals_sorted > 2022-02-02T17:44:12.1963670Z Feb 02 17:44:12 self.assertEqual(expected, > actual) > 2022-02-02T17:44:12.1964685Z Feb 02 17:44:12 E AssertionError: Lists > differ: ["cur[719 chars]te: {1: hi, 3: hi}, current value: Row(f0=5, f[172 > chars]0')"] != ["cur[719 chars]te: {3: hi}, current value: Row(f0=5, > f1='hi',[165 chars]0')"] > 2022-02-02T17:44:12.1965369Z Feb 02 17:44:12 E > 2022-02-02T17:44:12.1965731Z Feb 02 17:44:12 E First differing element 4: > 2022-02-02T17:44:12.1966428Z Feb 02 17:44:12 E "curr[80 chars]te: {1: hi, > 3: hi}, current value: Row(f0=5, f[23 chars]00')" > 2022-02-02T17:44:12.1967192Z Feb 02 17:44:12 E "curr[80 chars]te: {3: hi}, > current value: Row(f0=5, f1='hi',[16 chars]00')" > 2022-02-02T17:44:12.1967860Z Feb 02 17:44:12 E > 2022-02-02T17:44:12.1968268Z Feb 02 17:44:12 E Diff is 1211 characters > long. Set self.maxDiff to None to see it. > 2022-02-02T17:44:12.1968783Z Feb 02 17:44:12 =============================== > warnings summary =============================== > 2022-02-02T17:44:12.1969374Z Feb 02 17:44:12 > pyflink/datastream/tests/test_stream_execution_environment.py::StreamExecutionEnvironmentTests::test_add_classpaths > 2022-02-02T17:44:12.1970541Z Feb 02 17:44:12 > /__w/1/s/flink-python/.tox/py38/lib/python3.8/site-packages/future/standard_library/__init__.py:65: > DeprecationWarning: the imp module is deprecated in favour of importlib; see > the module's documentation for alternative uses > 2022-02-02T17:44:12.1971219Z Feb 02 17:44:12 import imp > 2022-02-02T17:44:12.1971530Z Feb 02 17:44:12 > 2022-02-02T17:44:12.1972027Z Feb 02 17:44:12 > pyflink/datastream/tests/test_stream_execution_environment.py::StreamExecutionEnvironmentTests::test_add_classpaths > 2022-02-02T17:44:12.1973535Z Feb 02 17:44:12 > /__w/1/s/flink-python/.tox/py38/lib/python3.8/site-packages/apache_beam/typehints/typehints.py:693: > DeprecationWarning: Using or importing the ABCs from 'collections' instead > of from 'collections.abc' is deprecated since Python 3.3, and in 3.10 it will > stop working > 2022-02-02T17:44:12.1974526Z Feb 02 17:44:12 if not > isinstance(type_params, collections.Iterable): > 2022-02-02T17:44:12.1974930Z Feb 02 17:44:12 > 2022-02-02T17:44:12.1975407Z Feb 02 17:44:12 > pyflink/datastream/tests/test_stream_execution_environment.py::StreamExecutionEnvironmentTests::test_add_classpaths > 2022-02-02T17:44:12.1976680Z Feb 02 17:44:12 > /__w/1/s/flink-python/.tox/py38/lib/python3.8/site-packages/apache_beam/typehints/typehints.py:532: > DeprecationWarning: Using or importing the ABCs from 'collections' instead > of from 'collections.abc' is deprecated since Python 3.3, and in 3.10 it will > stop working > 2022-02-02T17:44:12.1977509Z Feb 02 17:44:12 if not > isinstance(type_params, (collections.Sequence, set)): > 2022-02-02T17:44:12.1977939Z Feb 02 17:44:12 > 2022-02-02T17:44:12.1978432Z Feb 02 17:44:12 > pyflink/datastream/tests/test_stream_execution_environment.py::StreamExecutionEnvironmentTests::test_add_python_archive > 2022-02-02T17:44:12.1979475Z Feb 02 17:44:12 > /__w/1/s/flink-python/.tox/py38/lib/python3.8/site-packages/_pytest/threadexception.py:75: > PytestUnhandledThreadExceptionWarning: Exception in thread > read_grpc_client_inputs > 2022-02-02T17:44:12.1980064Z Feb 02 17:44:12 > 2022-02-02T17:44:12.1980434Z Feb 02 17:44:12 Traceback (most recent call > last): > 2022-02-02T17:44:12.1981152Z Feb 02 17:44:12 File > "/__w/1/s/flink-python/dev/.conda/envs/3.8/lib/python3.8/threading.py", line > 932, in _bootstrap_inner > 2022-02-02T17:44:12.1981642Z Feb 02 17:44:12 self.run() > 2022-02-02T17:44:12.1982368Z Feb 02 17:44:12 File > "/__w/1/s/flink-python/dev/.conda/envs/3.8/lib/python3.8/threading.py", line > 870, in run > 2022-02-02T17:44:12.1982884Z Feb 02 17:44:12 self._target(*self._args, > **self._kwargs) > 2022-02-02T17:44:12.1983865Z Feb 02 17:44:12 File > "/__w/1/s/flink-python/.tox/py38/lib/python3.8/site-packages/apache_beam/runners/worker/data_plane.py", > line 598, in <lambda> > 2022-02-02T17:44:12.1984471Z Feb 02 17:44:12 target=lambda: > self._read_inputs(elements_iterator), > 2022-02-02T17:44:12.1985299Z Feb 02 17:44:12 File > "/__w/1/s/flink-python/.tox/py38/lib/python3.8/site-packages/apache_beam/runners/worker/data_plane.py", > line 581, in _read_inputs > 2022-02-02T17:44:12.1985881Z Feb 02 17:44:12 for elements in > elements_iterator: > 2022-02-02T17:44:12.1986760Z Feb 02 17:44:12 File > "/__w/1/s/flink-python/.tox/py38/lib/python3.8/site-packages/grpc/_channel.py", > line 426, in __next__ > 2022-02-02T17:44:12.1987262Z Feb 02 17:44:12 return self._next() > 2022-02-02T17:44:12.1987946Z Feb 02 17:44:12 File > "/__w/1/s/flink-python/.tox/py38/lib/python3.8/site-packages/grpc/_channel.py", > line 826, in _next > 2022-02-02T17:44:12.1988425Z Feb 02 17:44:12 raise self > 2022-02-02T17:44:12.1989023Z Feb 02 17:44:12 > grpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that > terminated with: > 2022-02-02T17:44:12.1990067Z Feb 02 17:44:12 status = > StatusCode.CANCELLED > 2022-02-02T17:44:12.1990653Z Feb 02 17:44:12 details = "Multiplexer > hanging up" > 2022-02-02T17:44:12.1991849Z Feb 02 17:44:12 debug_error_string = > "{"created":"@1643823819.576493566","description":"Error received from peer > ipv4:127.0.0.1:33091","file":"src/core/lib/surface/call.cc","file_line":1074,"grpc_message":"Multiplexer > hanging up","grpc_status":1}" > 2022-02-02T17:44:12.1993432Z Feb 02 17:44:12 > > 2022-02-02T17:44:12.1993889Z Feb 02 17:44:12 > 2022-02-02T17:44:12.1994521Z Feb 02 17:44:12 > warnings.warn(pytest.PytestUnhandledThreadExceptionWarning(msg)) > 2022-02-02T17:44:12.1995279Z Feb 02 17:44:12 > 2022-02-02T17:44:12.1996037Z Feb 02 17:44:12 > pyflink/datastream/tests/test_stream_execution_environment.py::StreamExecutionEnvironmentTests::test_add_python_file > 2022-02-02T17:44:12.1997435Z Feb 02 17:44:12 > /__w/1/s/flink-python/pyflink/table/table_environment.py:1997: > DeprecationWarning: Deprecated in 1.12. Use from_data_stream(DataStream, > *Expression) instead. > 2022-02-02T17:44:12.1998269Z Feb 02 17:44:12 warnings.warn( > 2022-02-02T17:44:12.1998594Z Feb 02 17:44:12 > 2022-02-02T17:44:12.1999075Z Feb 02 17:44:12 > pyflink/datastream/tests/test_stream_execution_environment.py::StreamExecutionEnvironmentTests::test_execute > 2022-02-02T17:44:12.2000008Z Feb 02 17:44:12 > /__w/1/s/flink-python/pyflink/table/table_environment.py:538: > DeprecationWarning: Deprecated in 1.10. Use create_table instead. > 2022-02-02T17:44:12.2000823Z Feb 02 17:44:12 warnings.warn("Deprecated in > 1.10. Use create_table instead.", DeprecationWarning) > 2022-02-02T17:44:12.2001522Z Feb 02 17:44:12 > 2022-02-02T17:44:12.2002614Z Feb 02 17:44:12 -- Docs: > https://docs.pytest.org/en/stable/warnings.html > 2022-02-02T17:44:12.2003603Z Feb 02 17:44:12 ============================= > slowest 20 durations ============================= > 2022-02-02T17:44:12.2004618Z Feb 02 17:44:12 10.16s call > pyflink/datastream/tests/test_connectors.py::ConnectorTests::test_stream_file_sink > 2022-02-02T17:44:12.2005726Z Feb 02 17:44:12 9.83s call > pyflink/datastream/tests/test_data_stream.py::BatchModeDataStreamTests::test_keyed_process_function_with_state > 2022-02-02T17:44:12.2006511Z Feb 02 17:44:12 8.79s call > pyflink/datastream/tests/test_data_stream.py::StreamingModeDataStreamTests::test_keyed_process_function_with_state > 2022-02-02T17:44:12.2007232Z Feb 02 17:44:12 6.78s call > pyflink/datastream/tests/test_stream_execution_environment.py::StreamExecutionEnvironmentTests::test_add_python_file > 2022-02-02T17:44:12.2007961Z Feb 02 17:44:12 5.52s call > pyflink/datastream/tests/test_data_stream.py::StreamingModeDataStreamTests::test_execute_and_collect > 2022-02-02T17:44:12.2009001Z Feb 02 17:44:12 5.44s call > pyflink/datastream/tests/test_data_stream.py::BatchModeDataStreamTests::test_execute_and_collect > 2022-02-02T17:44:12.2010033Z Feb 02 17:44:12 5.26s call > pyflink/datastream/tests/test_data_stream.py::BatchModeDataStreamTests::test_basic_co_operations_with_output_type > 2022-02-02T17:44:12.2011152Z Feb 02 17:44:12 5.25s call > pyflink/datastream/tests/test_data_stream.py::BatchModeDataStreamTests::test_basic_co_operations > 2022-02-02T17:44:12.2012377Z Feb 02 17:44:12 4.53s call > pyflink/datastream/tests/test_data_stream.py::BatchModeDataStreamTests::test_keyed_co_process > 2022-02-02T17:44:12.2013701Z Feb 02 17:44:12 4.35s call > pyflink/datastream/tests/test_stream_execution_environment.py::StreamExecutionEnvironmentTests::test_set_requirements_with_cached_directory > 2022-02-02T17:44:12.2014884Z Feb 02 17:44:12 4.32s call > pyflink/datastream/tests/test_data_stream.py::BatchModeDataStreamTests::test_reduce_with_state > 2022-02-02T17:44:12.2015900Z Feb 02 17:44:12 4.26s call > pyflink/datastream/tests/test_data_stream.py::BatchModeDataStreamTests::test_keyed_flat_map > 2022-02-02T17:44:12.2016970Z Feb 02 17:44:12 4.21s call > pyflink/datastream/tests/test_data_stream.py::BatchModeDataStreamTests::test_keyed_co_map > 2022-02-02T17:44:12.2018270Z Feb 02 17:44:12 4.06s call > pyflink/datastream/tests/test_data_stream.py::BatchModeDataStreamTests::test_keyed_filter > 2022-02-02T17:44:12.2019463Z Feb 02 17:44:12 3.90s call > pyflink/datastream/tests/test_stream_execution_environment.py::StreamExecutionEnvironmentTests::test_set_requirements_without_cached_directory > 2022-02-02T17:44:12.2020715Z Feb 02 17:44:12 3.90s call > pyflink/datastream/tests/test_data_stream.py::BatchModeDataStreamTests::test_aggregating_state > 2022-02-02T17:44:12.2021749Z Feb 02 17:44:12 3.87s call > pyflink/datastream/tests/test_data_stream.py::BatchModeDataStreamTests::test_keyed_co_flat_map > 2022-02-02T17:44:12.2022862Z Feb 02 17:44:12 3.84s call > pyflink/datastream/tests/test_data_stream.py::BatchModeDataStreamTests::test_multi_key_by > 2022-02-02T17:44:12.2024078Z Feb 02 17:44:12 3.83s call > pyflink/datastream/tests/test_data_stream.py::BatchModeDataStreamTests::test_time_window > 2022-02-02T17:44:12.2024912Z Feb 02 17:44:12 3.83s call > pyflink/datastream/tests/test_data_stream.py::BatchModeDataStreamTests::test_count_window > 2022-02-02T17:44:12.2025750Z Feb 02 17:44:12 =========================== > short test summary info ============================ > 2022-02-02T17:44:12.2026370Z Feb 02 17:44:12 FAILED > pyflink/datastream/tests/test_data_stream.py::StreamingModeDataStreamTests::test_keyed_process_function_with_state > 2022-02-02T17:44:12.2027008Z Feb 02 17:44:12 ======= 1 failed, 154 passed, 1 > skipped, 6 warnings in 235.76s (0:03:55) ======= > 2022-02-02T17:44:12.5428501Z Feb 02 17:44:12 test module > /__w/1/s/flink-python/pyflink/datastream failed > 2022-02-02T17:44:12.5431151Z Feb 02 17:44:12 ERROR: InvocationError for > command /bin/bash ./dev/integration_test.sh (exited with code 1) > 2022-02-02T17:44:12.5432097Z Feb 02 17:44:12 py38 finish: run-test after > 999.77 seconds > 2022-02-02T17:44:12.5436171Z Feb 02 17:44:12 py38 start: run-test-post > 2022-02-02T17:44:12.5437071Z Feb 02 17:44:12 py38 finish: run-test-post > after 0.00 seconds > 2022-02-02T17:44:12.5438162Z Feb 02 17:44:12 > ___________________________________ summary > ____________________________________ > 2022-02-02T17:44:12.5453873Z Feb 02 17:44:12 ERROR: py38: commands failed > 2022-02-02T17:44:12.5455066Z Feb 02 17:44:12 cleanup > /__w/1/s/flink-python/.tox/.tmp/package/1/apache-flink-1.15.dev0.zip > 2022-02-02T17:44:12.6013749Z Feb 02 17:44:12 ============tox checks... > [FAILED]============ > {code} > https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=30642&view=logs&j=9cada3cb-c1d3-5621-16da-0f718fb86602&t=c67e71ed-6451-5d26-8920-5a8cf9651901&l=24759 -- This message was sent by Atlassian Jira (v8.20.1#820001)