I'm not sure how apache-beam-testing permissions are managed; Kenn, could we grant read-access for contributors who need it for testing?
Here are two logs from the job that seem relevant: 2018-10-08 14:44:45.381 PDT Parsing unknown args: [u'--dataflowJobId=2018-10-08_14_41_03-9578125971484804239', u'--autoscalingAlgorithm=NONE', u'--direct_runner_use_stacked_bundle', u'--maxNumWorkers=0', u'--style=scrambled', u'--sleep_secs=20', u'--pipeline_type_check', u'--gcpTempLocation=gs://temp-storage-for-end-to-end-tests/temp-it/beamapp-jenkins-1008214058-522436.1539034858.522554', u'--numWorkers=1', u'--beam_plugins=apache_beam.io.filesystem.FileSystem', u'--beam_plugins=apache_beam.io.hadoopfilesystem.HadoopFileSystem', u'--beam_plugins=apache_beam.io.localfilesystem.LocalFileSystem', u'--beam_plugins=apache_beam.io.gcp.gcsfilesystem.GCSFileSystem', u'--beam_plugins=apache_beam.io.filesystem_test.TestingFileSystem', u'--beam_plugins=apache_beam.runners.interactive.display.pipeline_graph_renderer.PipelineGraphRenderer', u'--beam_plugins=apache_beam.runners.interactive.display.pipeline_graph_renderer.MuteRenderer', u'--beam_plugins=apache_beam.runners.interactive.display.pipeline_graph_renderer.TextRenderer', u'--beam_plugins=apache_beam.runners.interactive.display.pipeline_graph_renderer.PydotRenderer', u'--pipelineUrl=gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-1008214058-522436.1539034858.522554/pipeline.pb'] 2018-10-08 14:44:45.382 PDT Python sdk harness failed: Traceback (most recent call last): File "/usr/local/lib/python2.7/dist-packages/apache_beam/runners/worker/sdk_worker_main.py", line 133, in main sdk_pipeline_options.get_all_options(drop_default=True)) File "/usr/local/lib/python2.7/dist-packages/apache_beam/options/pipeline_options.py", line 227, in get_all_options action='append' if num_times > 1 else 'store') File "/usr/lib/python2.7/argparse.py", line 1308, in add_argument return self._add_action(action) File "/usr/lib/python2.7/argparse.py", line 1682, in _add_action self._optionals._add_action(action) File "/usr/lib/python2.7/argparse.py", line 1509, in _add_action action = super(_ArgumentGroup, self)._add_action(action) File "/usr/lib/python2.7/argparse.py", line 1322, in _add_action self._check_conflict(action) File "/usr/lib/python2.7/argparse.py", line 1460, in _check_conflict conflict_handler(action, confl_optionals) File "/usr/lib/python2.7/argparse.py", line 1467, in _handle_conflict_error raise ArgumentError(action, message % conflict_string) ArgumentError: argument --beam_plugins: conflicting option string(s): --beam_plugins On Wed, Oct 10, 2018 at 1:05 AM Maximilian Michels <m...@apache.org> wrote: > Would be great to provide access to Dataflow build logs. > > In the meantime, could someone with access send me the logs for the job > below? > > > https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-10-08_14_41_03-9578125971484804239?project=apache-beam-testing > > Thanks, > Max > > On 09.10.18 13:45, Maximilian Michels wrote: > > Hi, > > > > I'm debugging a test failure in Dataflow PostCommit. There are logs > > available which I can't access. Is it possible to be added to the > > apache-beam-testing project? > > > > Thanks, > > Max > > > > > > Example: > > ====================================================================== > > FAIL: test_streaming_with_attributes > > (apache_beam.io.gcp.pubsub_integration_test.PubSubIntegrationTest) > > ---------------------------------------------------------------------- > > Traceback (most recent call last): > > File > > > "/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python_Verify_PR/src/sdks/python/apache_beam/io/gcp/pubsub_integration_test.py", > > > line 175, in test_streaming_with_attributes > > self._test_streaming(with_attributes=True) > > File > > > "/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python_Verify_PR/src/sdks/python/apache_beam/io/gcp/pubsub_integration_test.py", > > > line 167, in _test_streaming > > timestamp_attribute=self.TIMESTAMP_ATTRIBUTE) > > File > > > "/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python_Verify_PR/src/sdks/python/apache_beam/io/gcp/pubsub_it_pipeline.py", > > > line 91, in run_pipeline > > result = p.run() > > File > > > "/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python_Verify_PR/src/sdks/python/apache_beam/pipeline.py", > > > line 416, in run > > return self.runner.run_pipeline(self) > > File > > > "/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python_Verify_PR/src/sdks/python/apache_beam/runners/dataflow/test_dataflow_runner.py", > > > line 65, in run_pipeline > > hc_assert_that(self.result, pickler.loads(on_success_matcher)) > > AssertionError: > > Expected: (Test pipeline expected terminated in state: RUNNING and > > Expected 2 messages.) > > but: Expected 2 messages. Got 0 messages. Diffs (item, count): > > Expected but not in actual: [(PubsubMessage(data001-seen, > > {'processed': 'IT'}), 1), (PubsubMessage(data002-seen, {'timestamp_out': > > '2018-07-11T02:02:50.149000Z', 'processed': 'IT'}), 1)] > > Unexpected: [] > > Stripped attributes: ['id', 'timestamp'] > > > > -------------------- >> begin captured stdout << --------------------- > > Found: > > > https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-10-08_14_41_03-9578125971484804239?project=apache-beam-testing. > > > > -- Got feedback? tinyurl.com/swegner-feedback