I'm not sure how apache-beam-testing permissions are managed; Kenn, could
we grant read-access for contributors who need it for testing?

Here are two logs from the job that seem relevant:

2018-10-08 14:44:45.381 PDT
Parsing unknown args:
[u'--dataflowJobId=2018-10-08_14_41_03-9578125971484804239',
u'--autoscalingAlgorithm=NONE', u'--direct_runner_use_stacked_bundle',
u'--maxNumWorkers=0', u'--style=scrambled', u'--sleep_secs=20',
u'--pipeline_type_check',
u'--gcpTempLocation=gs://temp-storage-for-end-to-end-tests/temp-it/beamapp-jenkins-1008214058-522436.1539034858.522554',
u'--numWorkers=1', u'--beam_plugins=apache_beam.io.filesystem.FileSystem',
u'--beam_plugins=apache_beam.io.hadoopfilesystem.HadoopFileSystem',
u'--beam_plugins=apache_beam.io.localfilesystem.LocalFileSystem',
u'--beam_plugins=apache_beam.io.gcp.gcsfilesystem.GCSFileSystem',
u'--beam_plugins=apache_beam.io.filesystem_test.TestingFileSystem',
u'--beam_plugins=apache_beam.runners.interactive.display.pipeline_graph_renderer.PipelineGraphRenderer',
u'--beam_plugins=apache_beam.runners.interactive.display.pipeline_graph_renderer.MuteRenderer',
u'--beam_plugins=apache_beam.runners.interactive.display.pipeline_graph_renderer.TextRenderer',
u'--beam_plugins=apache_beam.runners.interactive.display.pipeline_graph_renderer.PydotRenderer',
u'--pipelineUrl=gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-1008214058-522436.1539034858.522554/pipeline.pb']

2018-10-08 14:44:45.382 PDT
Python sdk harness failed: Traceback (most recent call last): File
"/usr/local/lib/python2.7/dist-packages/apache_beam/runners/worker/sdk_worker_main.py",
line 133, in main sdk_pipeline_options.get_all_options(drop_default=True))
File
"/usr/local/lib/python2.7/dist-packages/apache_beam/options/pipeline_options.py",
line 227, in get_all_options action='append' if num_times > 1 else 'store')
File "/usr/lib/python2.7/argparse.py", line 1308, in add_argument return
self._add_action(action) File "/usr/lib/python2.7/argparse.py", line 1682,
in _add_action self._optionals._add_action(action) File
"/usr/lib/python2.7/argparse.py", line 1509, in _add_action action =
super(_ArgumentGroup, self)._add_action(action) File
"/usr/lib/python2.7/argparse.py", line 1322, in _add_action
self._check_conflict(action) File "/usr/lib/python2.7/argparse.py", line
1460, in _check_conflict conflict_handler(action, confl_optionals) File
"/usr/lib/python2.7/argparse.py", line 1467, in _handle_conflict_error
raise ArgumentError(action, message % conflict_string) ArgumentError:
argument --beam_plugins: conflicting option string(s): --beam_plugins

On Wed, Oct 10, 2018 at 1:05 AM Maximilian Michels <m...@apache.org> wrote:

> Would be great to provide access to Dataflow build logs.
>
> In the meantime, could someone with access send me the logs for the job
> below?
>
>
> https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-10-08_14_41_03-9578125971484804239?project=apache-beam-testing
>
> Thanks,
> Max
>
> On 09.10.18 13:45, Maximilian Michels wrote:
> > Hi,
> >
> > I'm debugging a test failure in Dataflow PostCommit. There are logs
> > available which I can't access. Is it possible to be added to the
> > apache-beam-testing project?
> >
> > Thanks,
> > Max
> >
> >
> > Example:
> > ======================================================================
> > FAIL: test_streaming_with_attributes
> > (apache_beam.io.gcp.pubsub_integration_test.PubSubIntegrationTest)
> > ----------------------------------------------------------------------
> > Traceback (most recent call last):
> >    File
> >
> "/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python_Verify_PR/src/sdks/python/apache_beam/io/gcp/pubsub_integration_test.py",
>
> > line 175, in test_streaming_with_attributes
> >      self._test_streaming(with_attributes=True)
> >    File
> >
> "/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python_Verify_PR/src/sdks/python/apache_beam/io/gcp/pubsub_integration_test.py",
>
> > line 167, in _test_streaming
> >      timestamp_attribute=self.TIMESTAMP_ATTRIBUTE)
> >    File
> >
> "/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python_Verify_PR/src/sdks/python/apache_beam/io/gcp/pubsub_it_pipeline.py",
>
> > line 91, in run_pipeline
> >      result = p.run()
> >    File
> >
> "/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python_Verify_PR/src/sdks/python/apache_beam/pipeline.py",
>
> > line 416, in run
> >      return self.runner.run_pipeline(self)
> >    File
> >
> "/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python_Verify_PR/src/sdks/python/apache_beam/runners/dataflow/test_dataflow_runner.py",
>
> > line 65, in run_pipeline
> >      hc_assert_that(self.result, pickler.loads(on_success_matcher))
> > AssertionError:
> > Expected: (Test pipeline expected terminated in state: RUNNING and
> > Expected 2 messages.)
> >       but: Expected 2 messages. Got 0 messages. Diffs (item, count):
> >    Expected but not in actual: [(PubsubMessage(data001-seen,
> > {'processed': 'IT'}), 1), (PubsubMessage(data002-seen, {'timestamp_out':
> > '2018-07-11T02:02:50.149000Z', 'processed': 'IT'}), 1)]
> >    Unexpected: []
> >    Stripped attributes: ['id', 'timestamp']
> >
> > -------------------- >> begin captured stdout << ---------------------
> > Found:
> >
> https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-10-08_14_41_03-9578125971484804239?project=apache-beam-testing.
>
> >
>


-- 




Got feedback? tinyurl.com/swegner-feedback

Reply via email to