[jira] [Created] (BEAM-5613) Snapshot Python dependencies and add to Python_NightlySnapshot

2018-10-02 Thread Mark Liu (JIRA)
Mark Liu created BEAM-5613:
--

 Summary: Snapshot Python dependencies and add to 
Python_NightlySnapshot
 Key: BEAM-5613
 URL: https://issues.apache.org/jira/browse/BEAM-5613
 Project: Beam
  Issue Type: Improvement
  Components: dependencies
Reporter: Mark Liu
Assignee: Mark Liu


Python SDK depends on a list of libraries without specific versions. Unexpected 
or unaware version change could happen if there is a new release or range 
change in setup.py. This can be detected by dependency checking tools which 
require to track dependency list and their versions periodically. 

We could snapshot dependencies periodically using pip freeze and python 
setup.py egg_info and upload them to a public accessible location to benefit 
people who want to check dependency info or building related tools.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Resolved] (BEAM-5334) [beam_PerformanceTests_Python] Incompatible packages in dependencies

2018-09-21 Thread Mark Liu (JIRA)


 [ 
https://issues.apache.org/jira/browse/BEAM-5334?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mark Liu resolved BEAM-5334.

   Resolution: Fixed
Fix Version/s: Not applicable

> [beam_PerformanceTests_Python] Incompatible packages in dependencies
> 
>
> Key: BEAM-5334
> URL: https://issues.apache.org/jira/browse/BEAM-5334
> Project: Beam
>  Issue Type: Bug
>  Components: test-failures
>Reporter: Mikhail Gryzykhin
>Assignee: Mark Liu
>Priority: Major
> Fix For: Not applicable
>
>  Time Spent: 2h 50m
>  Remaining Estimate: 0h
>
> Failing job url: 
> https://builds.apache.org/job/beam_PerformanceTests_Python/1390/console
> Relevant log:
> 05:18:36 Obtaining 
> file:///home/jenkins/jenkins-slave/workspace/beam_PerformanceTests_Python/src/sdks/python
> 05:18:37 Complete output from command python setup.py egg_info:
> 05:18:37 
> /home/jenkins/jenkins-slave/workspace/beam_PerformanceTests_Python/env/.beam_env/local/lib/python2.7/site-packages/setuptools/dist.py:398:
>  UserWarning: Normalizing '2.8.0.dev' to '2.8.0.dev0'
> 05:18:37 normalized_version,
> 05:18:37 running egg_info
> 05:18:37 Requirement already satisfied: future==0.16.0 in 
> /home/jenkins/.local/lib/python2.7/site-packages (0.16.0)
> 05:18:37 gapic-google-cloud-pubsub-v1 0.15.4 has requirement 
> oauth2client<4.0dev,>=2.0.0, but you'll have oauth2client 4.1.2 which is 
> incompatible.
> 05:18:37 proto-google-cloud-datastore-v1 0.90.4 has requirement 
> oauth2client<4.0dev,>=2.0.0, but you'll have oauth2client 4.1.2 which is 
> incompatible.
> 05:18:37 proto-google-cloud-pubsub-v1 0.15.4 has requirement 
> oauth2client<4.0dev,>=2.0.0, but you'll have oauth2client 4.1.2 which is 
> incompatible.
> 05:18:37 googledatastore 7.0.1 has requirement httplib2<0.10,>=0.9.1, but 
> you'll have httplib2 0.11.3 which is incompatible.
> 05:18:37 googledatastore 7.0.1 has requirement oauth2client<4.0.0,>=2.0.1, 
> but you'll have oauth2client 4.1.2 which is incompatible.
> 05:18:37 You are using pip version 10.0.1, however version 18.0 is available.
> 05:18:37 You should consider upgrading via the 'pip install --upgrade pip' 
> command.
> 05:18:37 error: [Errno 2] No such file or directory



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (BEAM-5334) [beam_PerformanceTests_Python] Incompatible packages in dependencies

2018-09-21 Thread Mark Liu (JIRA)


[ 
https://issues.apache.org/jira/browse/BEAM-5334?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16623759#comment-16623759
 ] 

Mark Liu commented on BEAM-5334:


beam_PerformanceTests_Python is green again in 
https://builds.apache.org/view/A-D/view/Beam/job/beam_PerformanceTests_Python/1465/

> [beam_PerformanceTests_Python] Incompatible packages in dependencies
> 
>
> Key: BEAM-5334
> URL: https://issues.apache.org/jira/browse/BEAM-5334
> Project: Beam
>  Issue Type: Bug
>  Components: test-failures
>Reporter: Mikhail Gryzykhin
>Assignee: Mark Liu
>Priority: Major
> Fix For: Not applicable
>
>  Time Spent: 2h 50m
>  Remaining Estimate: 0h
>
> Failing job url: 
> https://builds.apache.org/job/beam_PerformanceTests_Python/1390/console
> Relevant log:
> 05:18:36 Obtaining 
> file:///home/jenkins/jenkins-slave/workspace/beam_PerformanceTests_Python/src/sdks/python
> 05:18:37 Complete output from command python setup.py egg_info:
> 05:18:37 
> /home/jenkins/jenkins-slave/workspace/beam_PerformanceTests_Python/env/.beam_env/local/lib/python2.7/site-packages/setuptools/dist.py:398:
>  UserWarning: Normalizing '2.8.0.dev' to '2.8.0.dev0'
> 05:18:37 normalized_version,
> 05:18:37 running egg_info
> 05:18:37 Requirement already satisfied: future==0.16.0 in 
> /home/jenkins/.local/lib/python2.7/site-packages (0.16.0)
> 05:18:37 gapic-google-cloud-pubsub-v1 0.15.4 has requirement 
> oauth2client<4.0dev,>=2.0.0, but you'll have oauth2client 4.1.2 which is 
> incompatible.
> 05:18:37 proto-google-cloud-datastore-v1 0.90.4 has requirement 
> oauth2client<4.0dev,>=2.0.0, but you'll have oauth2client 4.1.2 which is 
> incompatible.
> 05:18:37 proto-google-cloud-pubsub-v1 0.15.4 has requirement 
> oauth2client<4.0dev,>=2.0.0, but you'll have oauth2client 4.1.2 which is 
> incompatible.
> 05:18:37 googledatastore 7.0.1 has requirement httplib2<0.10,>=0.9.1, but 
> you'll have httplib2 0.11.3 which is incompatible.
> 05:18:37 googledatastore 7.0.1 has requirement oauth2client<4.0.0,>=2.0.1, 
> but you'll have oauth2client 4.1.2 which is incompatible.
> 05:18:37 You are using pip version 10.0.1, however version 18.0 is available.
> 05:18:37 You should consider upgrading via the 'pip install --upgrade pip' 
> command.
> 05:18:37 error: [Errno 2] No such file or directory



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Assigned] (BEAM-5334) [beam_PerformanceTests_Python] Incompatible packages in dependencies

2018-09-20 Thread Mark Liu (JIRA)


 [ 
https://issues.apache.org/jira/browse/BEAM-5334?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mark Liu reassigned BEAM-5334:
--

Assignee: Mark Liu

> [beam_PerformanceTests_Python] Incompatible packages in dependencies
> 
>
> Key: BEAM-5334
> URL: https://issues.apache.org/jira/browse/BEAM-5334
> Project: Beam
>  Issue Type: Bug
>  Components: test-failures
>Reporter: Mikhail Gryzykhin
>Assignee: Mark Liu
>Priority: Major
>  Time Spent: 2h 20m
>  Remaining Estimate: 0h
>
> Failing job url: 
> https://builds.apache.org/job/beam_PerformanceTests_Python/1390/console
> Relevant log:
> 05:18:36 Obtaining 
> file:///home/jenkins/jenkins-slave/workspace/beam_PerformanceTests_Python/src/sdks/python
> 05:18:37 Complete output from command python setup.py egg_info:
> 05:18:37 
> /home/jenkins/jenkins-slave/workspace/beam_PerformanceTests_Python/env/.beam_env/local/lib/python2.7/site-packages/setuptools/dist.py:398:
>  UserWarning: Normalizing '2.8.0.dev' to '2.8.0.dev0'
> 05:18:37 normalized_version,
> 05:18:37 running egg_info
> 05:18:37 Requirement already satisfied: future==0.16.0 in 
> /home/jenkins/.local/lib/python2.7/site-packages (0.16.0)
> 05:18:37 gapic-google-cloud-pubsub-v1 0.15.4 has requirement 
> oauth2client<4.0dev,>=2.0.0, but you'll have oauth2client 4.1.2 which is 
> incompatible.
> 05:18:37 proto-google-cloud-datastore-v1 0.90.4 has requirement 
> oauth2client<4.0dev,>=2.0.0, but you'll have oauth2client 4.1.2 which is 
> incompatible.
> 05:18:37 proto-google-cloud-pubsub-v1 0.15.4 has requirement 
> oauth2client<4.0dev,>=2.0.0, but you'll have oauth2client 4.1.2 which is 
> incompatible.
> 05:18:37 googledatastore 7.0.1 has requirement httplib2<0.10,>=0.9.1, but 
> you'll have httplib2 0.11.3 which is incompatible.
> 05:18:37 googledatastore 7.0.1 has requirement oauth2client<4.0.0,>=2.0.1, 
> but you'll have oauth2client 4.1.2 which is incompatible.
> 05:18:37 You are using pip version 10.0.1, however version 18.0 is available.
> 05:18:37 You should consider upgrading via the 'pip install --upgrade pip' 
> command.
> 05:18:37 error: [Errno 2] No such file or directory



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (BEAM-5334) [beam_PerformanceTests_Python] Incompatible packages in dependencies

2018-09-20 Thread Mark Liu (JIRA)


[ 
https://issues.apache.org/jira/browse/BEAM-5334?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16623075#comment-16623075
 ] 

Mark Liu commented on BEAM-5334:


[https://github.com/apache/beam/pull/6456] is out for review. It should fix 
this benchmark test.

> [beam_PerformanceTests_Python] Incompatible packages in dependencies
> 
>
> Key: BEAM-5334
> URL: https://issues.apache.org/jira/browse/BEAM-5334
> Project: Beam
>  Issue Type: Bug
>  Components: test-failures
>Reporter: Mikhail Gryzykhin
>Priority: Major
>  Time Spent: 2h 20m
>  Remaining Estimate: 0h
>
> Failing job url: 
> https://builds.apache.org/job/beam_PerformanceTests_Python/1390/console
> Relevant log:
> 05:18:36 Obtaining 
> file:///home/jenkins/jenkins-slave/workspace/beam_PerformanceTests_Python/src/sdks/python
> 05:18:37 Complete output from command python setup.py egg_info:
> 05:18:37 
> /home/jenkins/jenkins-slave/workspace/beam_PerformanceTests_Python/env/.beam_env/local/lib/python2.7/site-packages/setuptools/dist.py:398:
>  UserWarning: Normalizing '2.8.0.dev' to '2.8.0.dev0'
> 05:18:37 normalized_version,
> 05:18:37 running egg_info
> 05:18:37 Requirement already satisfied: future==0.16.0 in 
> /home/jenkins/.local/lib/python2.7/site-packages (0.16.0)
> 05:18:37 gapic-google-cloud-pubsub-v1 0.15.4 has requirement 
> oauth2client<4.0dev,>=2.0.0, but you'll have oauth2client 4.1.2 which is 
> incompatible.
> 05:18:37 proto-google-cloud-datastore-v1 0.90.4 has requirement 
> oauth2client<4.0dev,>=2.0.0, but you'll have oauth2client 4.1.2 which is 
> incompatible.
> 05:18:37 proto-google-cloud-pubsub-v1 0.15.4 has requirement 
> oauth2client<4.0dev,>=2.0.0, but you'll have oauth2client 4.1.2 which is 
> incompatible.
> 05:18:37 googledatastore 7.0.1 has requirement httplib2<0.10,>=0.9.1, but 
> you'll have httplib2 0.11.3 which is incompatible.
> 05:18:37 googledatastore 7.0.1 has requirement oauth2client<4.0.0,>=2.0.1, 
> but you'll have oauth2client 4.1.2 which is incompatible.
> 05:18:37 You are using pip version 10.0.1, however version 18.0 is available.
> 05:18:37 You should consider upgrading via the 'pip install --upgrade pip' 
> command.
> 05:18:37 error: [Errno 2] No such file or directory



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Comment Edited] (BEAM-5334) [beam_PerformanceTests_Python] Incompatible packages in dependencies

2018-09-17 Thread Mark Liu (JIRA)


[ 
https://issues.apache.org/jira/browse/BEAM-5334?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16617989#comment-16617989
 ] 

Mark Liu edited comment on BEAM-5334 at 9/17/18 8:07 PM:
-

https://github.com/apache/beam/pull/6392 enables Python benchmark to setup and 
run test with Gradle, which should be helpful to fix the failure. 
https://github.com/GoogleCloudPlatform/PerfKitBenchmarker/pull/1776 is the 
corresponding changes on Perfkit. Once that's in, we should be able to change 
Jenkins benchmark job to use Gradle for test setup and execution.


was (Author: markflyhigh):
https://github.com/apache/beam/pull/6392 enables Python benchmark to setup and 
run test with Gradle, which should be helpful to fix the failure. 
https://github.com/GoogleCloudPlatform/PerfKitBenchmarker/pull/1776 is the 
corresponding changes on Perfkit. Once that's in, we should be able to change 
Jenkins benchmark job to used Gradle for test setup and execution.

> [beam_PerformanceTests_Python] Incompatible packages in dependencies
> 
>
> Key: BEAM-5334
> URL: https://issues.apache.org/jira/browse/BEAM-5334
> Project: Beam
>  Issue Type: Bug
>  Components: test-failures
>Reporter: Mikhail Gryzykhin
>Priority: Major
>  Time Spent: 0.5h
>  Remaining Estimate: 0h
>
> Failing job url: 
> https://builds.apache.org/job/beam_PerformanceTests_Python/1390/console
> Relevant log:
> 05:18:36 Obtaining 
> file:///home/jenkins/jenkins-slave/workspace/beam_PerformanceTests_Python/src/sdks/python
> 05:18:37 Complete output from command python setup.py egg_info:
> 05:18:37 
> /home/jenkins/jenkins-slave/workspace/beam_PerformanceTests_Python/env/.beam_env/local/lib/python2.7/site-packages/setuptools/dist.py:398:
>  UserWarning: Normalizing '2.8.0.dev' to '2.8.0.dev0'
> 05:18:37 normalized_version,
> 05:18:37 running egg_info
> 05:18:37 Requirement already satisfied: future==0.16.0 in 
> /home/jenkins/.local/lib/python2.7/site-packages (0.16.0)
> 05:18:37 gapic-google-cloud-pubsub-v1 0.15.4 has requirement 
> oauth2client<4.0dev,>=2.0.0, but you'll have oauth2client 4.1.2 which is 
> incompatible.
> 05:18:37 proto-google-cloud-datastore-v1 0.90.4 has requirement 
> oauth2client<4.0dev,>=2.0.0, but you'll have oauth2client 4.1.2 which is 
> incompatible.
> 05:18:37 proto-google-cloud-pubsub-v1 0.15.4 has requirement 
> oauth2client<4.0dev,>=2.0.0, but you'll have oauth2client 4.1.2 which is 
> incompatible.
> 05:18:37 googledatastore 7.0.1 has requirement httplib2<0.10,>=0.9.1, but 
> you'll have httplib2 0.11.3 which is incompatible.
> 05:18:37 googledatastore 7.0.1 has requirement oauth2client<4.0.0,>=2.0.1, 
> but you'll have oauth2client 4.1.2 which is incompatible.
> 05:18:37 You are using pip version 10.0.1, however version 18.0 is available.
> 05:18:37 You should consider upgrading via the 'pip install --upgrade pip' 
> command.
> 05:18:37 error: [Errno 2] No such file or directory



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (BEAM-5334) [beam_PerformanceTests_Python] Incompatible packages in dependencies

2018-09-17 Thread Mark Liu (JIRA)


[ 
https://issues.apache.org/jira/browse/BEAM-5334?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16617989#comment-16617989
 ] 

Mark Liu commented on BEAM-5334:


https://github.com/apache/beam/pull/6392 enables Python benchmark to setup and 
run test with Gradle, which should be helpful to fix the failure. 
https://github.com/GoogleCloudPlatform/PerfKitBenchmarker/pull/1776 is the 
corresponding changes on Perfkit. Once that's in, we should be able to change 
Jenkins benchmark job to used Gradle for test setup and execution.

> [beam_PerformanceTests_Python] Incompatible packages in dependencies
> 
>
> Key: BEAM-5334
> URL: https://issues.apache.org/jira/browse/BEAM-5334
> Project: Beam
>  Issue Type: Bug
>  Components: test-failures
>Reporter: Mikhail Gryzykhin
>Priority: Major
>  Time Spent: 0.5h
>  Remaining Estimate: 0h
>
> Failing job url: 
> https://builds.apache.org/job/beam_PerformanceTests_Python/1390/console
> Relevant log:
> 05:18:36 Obtaining 
> file:///home/jenkins/jenkins-slave/workspace/beam_PerformanceTests_Python/src/sdks/python
> 05:18:37 Complete output from command python setup.py egg_info:
> 05:18:37 
> /home/jenkins/jenkins-slave/workspace/beam_PerformanceTests_Python/env/.beam_env/local/lib/python2.7/site-packages/setuptools/dist.py:398:
>  UserWarning: Normalizing '2.8.0.dev' to '2.8.0.dev0'
> 05:18:37 normalized_version,
> 05:18:37 running egg_info
> 05:18:37 Requirement already satisfied: future==0.16.0 in 
> /home/jenkins/.local/lib/python2.7/site-packages (0.16.0)
> 05:18:37 gapic-google-cloud-pubsub-v1 0.15.4 has requirement 
> oauth2client<4.0dev,>=2.0.0, but you'll have oauth2client 4.1.2 which is 
> incompatible.
> 05:18:37 proto-google-cloud-datastore-v1 0.90.4 has requirement 
> oauth2client<4.0dev,>=2.0.0, but you'll have oauth2client 4.1.2 which is 
> incompatible.
> 05:18:37 proto-google-cloud-pubsub-v1 0.15.4 has requirement 
> oauth2client<4.0dev,>=2.0.0, but you'll have oauth2client 4.1.2 which is 
> incompatible.
> 05:18:37 googledatastore 7.0.1 has requirement httplib2<0.10,>=0.9.1, but 
> you'll have httplib2 0.11.3 which is incompatible.
> 05:18:37 googledatastore 7.0.1 has requirement oauth2client<4.0.0,>=2.0.1, 
> but you'll have oauth2client 4.1.2 which is incompatible.
> 05:18:37 You are using pip version 10.0.1, however version 18.0 is available.
> 05:18:37 You should consider upgrading via the 'pip install --upgrade pip' 
> command.
> 05:18:37 error: [Errno 2] No such file or directory



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Comment Edited] (BEAM-5296) python beam_integration_benchmark(0) continuously failed.

2018-09-05 Thread Mark Liu (JIRA)


[ 
https://issues.apache.org/jira/browse/BEAM-5296?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16605115#comment-16605115
 ] 

Mark Liu edited comment on BEAM-5296 at 9/6/18 12:55 AM:
-

Investigations are made by Lukasz and summarized in 
https://github.com/apache/beam/pull/6321#issuecomment-418339689

If no further progress on the fix, we probably should consider upgrade fastavro 
on Jenkins nodes or rollback relevant 
[PR|https://github.com/apache/beam/pull/6321] since it affects many performance 
benchmarks.


was (Author: markflyhigh):
Investigations are made by Lukasz and summarized in 
https://github.com/apache/beam/pull/6321#issuecomment-418339689

If no further progress on the fix, we probably should consider upgrade fastavro 
on Jenkins nodes or rollback relevant PR 
https://github.com/apache/beam/pull/6321

> python beam_integration_benchmark(0) continuously failed.
> -
>
> Key: BEAM-5296
> URL: https://issues.apache.org/jira/browse/BEAM-5296
> Project: Beam
>  Issue Type: Bug
>  Components: test-failures
>Reporter: Boyuan Zhang
>Assignee: Mark Liu
>Priority: Major
>
> beam_integration_benchmark0) failed
> https://builds.apache.org/job/beam_PerformanceTests_Python/1377/



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Comment Edited] (BEAM-5296) python beam_integration_benchmark(0) continuously failed.

2018-09-05 Thread Mark Liu (JIRA)


[ 
https://issues.apache.org/jira/browse/BEAM-5296?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16605115#comment-16605115
 ] 

Mark Liu edited comment on BEAM-5296 at 9/6/18 12:54 AM:
-

Investigations are made by Lukasz and summarized in 
https://github.com/apache/beam/pull/6321#issuecomment-418339689

If no further progress on the fix, we probably should consider upgrade fastavro 
on Jenkins nodes or rollback relevant PR 
https://github.com/apache/beam/pull/6321


was (Author: markflyhigh):
Investigations are made by Lukasz and summarized in 
https://github.com/apache/beam/pull/6321#issuecomment-418339689

We probably should consider upgrade fastavro on Jenkins nodes or rollback 
relevant PR https://github.com/apache/beam/pull/6321

> python beam_integration_benchmark(0) continuously failed.
> -
>
> Key: BEAM-5296
> URL: https://issues.apache.org/jira/browse/BEAM-5296
> Project: Beam
>  Issue Type: Bug
>  Components: test-failures
>Reporter: Boyuan Zhang
>Assignee: Mark Liu
>Priority: Major
>
> beam_integration_benchmark0) failed
> https://builds.apache.org/job/beam_PerformanceTests_Python/1377/



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Comment Edited] (BEAM-5296) python beam_integration_benchmark(0) continuously failed.

2018-09-05 Thread Mark Liu (JIRA)


[ 
https://issues.apache.org/jira/browse/BEAM-5296?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16605115#comment-16605115
 ] 

Mark Liu edited comment on BEAM-5296 at 9/6/18 12:52 AM:
-

Investigations are made by Lukasz and summarized in 
https://github.com/apache/beam/pull/6321#issuecomment-418339689

We probably should consider upgrade fastavro on Jenkins nodes or rollback 
relevant PR https://github.com/apache/beam/pull/6321


was (Author: markflyhigh):
Investigations are made by Lukasz and summarized in 
https://github.com/apache/beam/pull/6321#issuecomment-418339689

> python beam_integration_benchmark(0) continuously failed.
> -
>
> Key: BEAM-5296
> URL: https://issues.apache.org/jira/browse/BEAM-5296
> Project: Beam
>  Issue Type: Bug
>  Components: test-failures
>Reporter: Boyuan Zhang
>Assignee: Mark Liu
>Priority: Major
>
> beam_integration_benchmark0) failed
> https://builds.apache.org/job/beam_PerformanceTests_Python/1377/



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (BEAM-5296) python beam_integration_benchmark(0) continuously failed.

2018-09-05 Thread Mark Liu (JIRA)


[ 
https://issues.apache.org/jira/browse/BEAM-5296?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16605115#comment-16605115
 ] 

Mark Liu commented on BEAM-5296:


Investigations are made by Lukasz and summarized in 
https://github.com/apache/beam/pull/6321#issuecomment-418339689

> python beam_integration_benchmark(0) continuously failed.
> -
>
> Key: BEAM-5296
> URL: https://issues.apache.org/jira/browse/BEAM-5296
> Project: Beam
>  Issue Type: Bug
>  Components: test-failures
>Reporter: Boyuan Zhang
>Assignee: Mark Liu
>Priority: Major
>
> beam_integration_benchmark0) failed
> https://builds.apache.org/job/beam_PerformanceTests_Python/1377/



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (BEAM-5329) Increase test coverage of Perfkit benchmark for Python case

2018-09-05 Thread Mark Liu (JIRA)
Mark Liu created BEAM-5329:
--

 Summary: Increase test coverage of Perfkit benchmark for Python 
case
 Key: BEAM-5329
 URL: https://issues.apache.org/jira/browse/BEAM-5329
 Project: Beam
  Issue Type: Test
  Components: testing
Reporter: Mark Liu
Assignee: Jason Kuster


Perfkit benchmark is defined in PerfKitBenchmarker repo (mostly 
[here|https://github.com/GoogleCloudPlatform/PerfKitBenchmarker/blob/master/perfkitbenchmarker/beam_benchmark_helper.py]
 and 
[here|https://github.com/GoogleCloudPlatform/PerfKitBenchmarker/blob/master/perfkitbenchmarker/linux_benchmarks/beam_integration_benchmark.py])
 but not enough test coverage to verify changes to the benchmark or it's utils. 
Breakage can be surfaced only when scheduled Beam Postcommit job runs which 
will directly affect performance measurement.

We want to add more tests starting from unit tests to cover missing branches 
and utils like Python benchmark.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Assigned] (BEAM-5302) Improve performance test documentation

2018-09-04 Thread Mark Liu (JIRA)


 [ 
https://issues.apache.org/jira/browse/BEAM-5302?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mark Liu reassigned BEAM-5302:
--

Assignee: (was: Jason Kuster)

> Improve performance test documentation
> --
>
> Key: BEAM-5302
> URL: https://issues.apache.org/jira/browse/BEAM-5302
> Project: Beam
>  Issue Type: Improvement
>  Components: testing
>Reporter: Mark Liu
>Priority: Major
>
> Current documentation for performance testing and benchmarks missing 
> following areas:
> How to write / use benchmark on Perfkit
> How to run benchmark locally or on Jenkins
> Benchmark summary / definition
> How to use performance metrics data and where to find
> Those documents can help new contributors to start and people who interested 
> in performance result to understand framework and look up results.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (BEAM-5302) Improve performance test documentation

2018-09-04 Thread Mark Liu (JIRA)
Mark Liu created BEAM-5302:
--

 Summary: Improve performance test documentation
 Key: BEAM-5302
 URL: https://issues.apache.org/jira/browse/BEAM-5302
 Project: Beam
  Issue Type: Improvement
  Components: testing
Reporter: Mark Liu
Assignee: Jason Kuster


Current documentation for performance testing and benchmarks missing following 
areas:

How to write / use benchmark on Perfkit
How to run benchmark locally or on Jenkins
Benchmark summary / definition
How to use performance metrics data and where to find

Those documents can help new contributors to start and people who interested in 
performance result to understand framework and look up results.




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (BEAM-5300) Add Python benchmark to beam_PerformanceTests_Analysis

2018-09-04 Thread Mark Liu (JIRA)
Mark Liu created BEAM-5300:
--

 Summary: Add Python benchmark to beam_PerformanceTests_Analysis
 Key: BEAM-5300
 URL: https://issues.apache.org/jira/browse/BEAM-5300
 Project: Beam
  Issue Type: Task
  Components: build-system
Reporter: Mark Liu
Assignee: Mark Liu


Benchmark: https://builds.apache.org/job/beam_PerformanceTests_Python/

We can add this benchmark to existing performance result analyze tool 
(beam_PerformanceTests_Analysis) to improve regression detection.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Comment Edited] (BEAM-5296) python beam_integration_benchmark(0) continuously failed.

2018-09-04 Thread Mark Liu (JIRA)


[ 
https://issues.apache.org/jira/browse/BEAM-5296?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16603465#comment-16603465
 ] 

Mark Liu edited comment on BEAM-5296 at 9/4/18 7:11 PM:


This flag is defined in BeamTestPlugin and seems it's failed to load due to 
fastavro:

{code:python}
/home/jenkins/.local/lib/python2.7/site-packages/nose/plugins/manager.py:395: 
RuntimeWarning: Unable to load plugin beam_test_plugin = 
test_config:BeamTestPlugin: (fastavro 0.19.7 
(/home/jenkins/.local/lib/python2.7/site-packages), 
Requirement.parse('fastavro==0.21.4'))
{code}

Tracing back to previous log and found fastavro 0.21.4 didn't install 
successfully:


{code:python}
23:18:23 Installing collected packages: fastavro, numpy, apache-beam
23:18:23   Found existing installation: fastavro 0.19.7
23:18:23 Not uninstalling fastavro at 
/home/jenkins/.local/lib/python2.7/site-packages, outside environment 
/home/jenkins/jenkins-slave/workspace/beam_PerformanceTests_Python/env/.beam_env
23:18:23 Can't uninstall 'fastavro'. No files were found to uninstall.
{code}

This test started to break since 
https://builds.apache.org/job/beam_PerformanceTests_Python/1366/ when fastavro 
was upgraded in 
https://github.com/apache/beam/commit/423a49d5135ef249a88de47a5713eabb07c952ae


was (Author: markflyhigh):
This flag is defined in BeamTestPlugin and seems it's failed to load due to 
fastavro:

{code:python}
/home/jenkins/.local/lib/python2.7/site-packages/nose/plugins/manager.py:395: 
RuntimeWarning: Unable to load plugin beam_test_plugin = 
test_config:BeamTestPlugin: (fastavro 0.19.7 
(/home/jenkins/.local/lib/python2.7/site-packages), 
Requirement.parse('fastavro==0.21.4'))
{code}

Tracing back to previous log and found fastavro 0.21.4 didn't install 
successfully:


{code:python}
18:31:08   Found existing installation: fastavro 0.19.7
18:31:08 Not uninstalling fastavro at 
/home/jenkins/.local/lib/python2.7/site-packages, outside environment 
/home/jenkins/jenkins-slave/workspace/beam_PerformanceTests_Python/env/.beam_env
{code}

This test started to break since 
https://builds.apache.org/job/beam_PerformanceTests_Python/1366/ when fastavro 
was upgraded in 
https://github.com/apache/beam/commit/423a49d5135ef249a88de47a5713eabb07c952ae

> python beam_integration_benchmark(0) continuously failed.
> -
>
> Key: BEAM-5296
> URL: https://issues.apache.org/jira/browse/BEAM-5296
> Project: Beam
>  Issue Type: Bug
>  Components: test-failures
>Reporter: Boyuan Zhang
>Assignee: Mark Liu
>Priority: Major
>
> beam_integration_benchmark0) failed
> https://builds.apache.org/job/beam_PerformanceTests_Python/1377/



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Comment Edited] (BEAM-5296) python beam_integration_benchmark(0) continuously failed.

2018-09-04 Thread Mark Liu (JIRA)


[ 
https://issues.apache.org/jira/browse/BEAM-5296?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16603465#comment-16603465
 ] 

Mark Liu edited comment on BEAM-5296 at 9/4/18 7:08 PM:


This flag is defined in BeamTestPlugin and seems it's failed to load due to 
fastavro:

{code:python}
/home/jenkins/.local/lib/python2.7/site-packages/nose/plugins/manager.py:395: 
RuntimeWarning: Unable to load plugin beam_test_plugin = 
test_config:BeamTestPlugin: (fastavro 0.19.7 
(/home/jenkins/.local/lib/python2.7/site-packages), 
Requirement.parse('fastavro==0.21.4'))
{code}

Tracing back to previous log and found fastavro 0.21.4 didn't install 
successfully:


{code:python}
18:31:08   Found existing installation: fastavro 0.19.7
18:31:08 Not uninstalling fastavro at 
/home/jenkins/.local/lib/python2.7/site-packages, outside environment 
/home/jenkins/jenkins-slave/workspace/beam_PerformanceTests_Python/env/.beam_env
{code}

This test started to break since 
https://builds.apache.org/job/beam_PerformanceTests_Python/1366/ when fastavro 
was upgraded in 
https://github.com/apache/beam/commit/423a49d5135ef249a88de47a5713eabb07c952ae


was (Author: markflyhigh):
This flag is defined in BeamTestPlugin and seems it's failed to load due to 
fastavro:

{code:python}
/home/jenkins/.local/lib/python2.7/site-packages/nose/plugins/manager.py:395: 
RuntimeWarning: Unable to load plugin beam_test_plugin = 
test_config:BeamTestPlugin: (fastavro 0.19.7 
(/home/jenkins/.local/lib/python2.7/site-packages), 
Requirement.parse('fastavro==0.21.4'))
{code}

Tracing back to previous log and found fastavro 0.21.4 didn't install 
successfully:


{code:python}
18:31:08   Found existing installation: fastavro 0.19.7
18:31:08 Not uninstalling fastavro at 
/home/jenkins/.local/lib/python2.7/site-packages, outside environment 
/home/jenkins/jenkins-slave/workspace/beam_PerformanceTests_Python/env/.beam_env
{code}

This test started to break since 
https://builds.apache.org/job/beam_PerformanceTests_Python/1366/ when fastavro 
is upgraded in 
https://github.com/apache/beam/commit/423a49d5135ef249a88de47a5713eabb07c952ae

> python beam_integration_benchmark(0) continuously failed.
> -
>
> Key: BEAM-5296
> URL: https://issues.apache.org/jira/browse/BEAM-5296
> Project: Beam
>  Issue Type: Bug
>  Components: test-failures
>Reporter: Boyuan Zhang
>Assignee: Mark Liu
>Priority: Major
>
> beam_integration_benchmark0) failed
> https://builds.apache.org/job/beam_PerformanceTests_Python/1377/



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (BEAM-5296) python beam_integration_benchmark(0) continuously failed.

2018-09-04 Thread Mark Liu (JIRA)


[ 
https://issues.apache.org/jira/browse/BEAM-5296?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16603465#comment-16603465
 ] 

Mark Liu commented on BEAM-5296:


This flag is defined in BeamTestPlugin and seems it's failed to load due to 
fastavro:

{code:python}
/home/jenkins/.local/lib/python2.7/site-packages/nose/plugins/manager.py:395: 
RuntimeWarning: Unable to load plugin beam_test_plugin = 
test_config:BeamTestPlugin: (fastavro 0.19.7 
(/home/jenkins/.local/lib/python2.7/site-packages), 
Requirement.parse('fastavro==0.21.4'))
{code}

Tracing back to previous log and found fastavro 0.21.4 didn't install 
successfully:


{code:python}
18:31:08   Found existing installation: fastavro 0.19.7
18:31:08 Not uninstalling fastavro at 
/home/jenkins/.local/lib/python2.7/site-packages, outside environment 
/home/jenkins/jenkins-slave/workspace/beam_PerformanceTests_Python/env/.beam_env
{code}

This test started to break since 
https://builds.apache.org/job/beam_PerformanceTests_Python/1366/ when fastavro 
is upgraded in 
https://github.com/apache/beam/commit/423a49d5135ef249a88de47a5713eabb07c952ae

> python beam_integration_benchmark(0) continuously failed.
> -
>
> Key: BEAM-5296
> URL: https://issues.apache.org/jira/browse/BEAM-5296
> Project: Beam
>  Issue Type: Bug
>  Components: test-failures
>Reporter: Boyuan Zhang
>Assignee: Mark Liu
>Priority: Major
>
> beam_integration_benchmark0) failed
> https://builds.apache.org/job/beam_PerformanceTests_Python/1377/



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (BEAM-5108) Improve Python test framework to prevent streaming pipeline leaks

2018-08-08 Thread Mark Liu (JIRA)


 [ 
https://issues.apache.org/jira/browse/BEAM-5108?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mark Liu updated BEAM-5108:
---
Summary: Improve Python test framework to prevent streaming pipeline leaks  
(was: Python test framework should prevent streaming pipeline leaks)

> Improve Python test framework to prevent streaming pipeline leaks
> -
>
> Key: BEAM-5108
> URL: https://issues.apache.org/jira/browse/BEAM-5108
> Project: Beam
>  Issue Type: Task
>  Components: testing
>Reporter: Mark Liu
>Priority: Major
>
> Recently, few Python streaming pipelines on Dataflow apache-beam-testing 
> project run for more than 5 days. This look like a leaking from Jenkins job 
> that runs e2e integration tests.
> Test framework has a pipeline resource clean up and applies to all 
> integration test, which is defined in 
> [TestDataflowRunner|https://github.com/apache/beam/blob/master/sdks/python/apache_beam/runners/dataflow/test_dataflow_runner.py#L67].
>  However, the cancellation may failed in a special case, like following (from 
> [this Jenkins 
> run|https://builds.apache.org/view/A-D/view/Beam/job/beam_PostCommit_Python_Verify/5636/consoleFull]):
> {quote}
> Workflow modification failed. Causes: (c53cc746f7bc7f49): Operation cancel 
> not allowed for job 2018-08-01_13_10_24-5019826606522054507. Job is not yet 
> ready for canceling. Please retry in a few minutes.
> {quote}
> Two possible approaches to improve test infra:
> 1. Add retry to the framework cancellation.
> 2. Instead of wait until pipeline in RUNNING state 
> ([here|https://github.com/apache/beam/blob/master/sdks/python/apache_beam/runners/dataflow/test_dataflow_runner.py#L57]),
>  we want to wait more to make sure worker pool starts successfully.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Closed] (BEAM-3809) beam_PerformanceTests_Python failing

2018-08-06 Thread Mark Liu (JIRA)


 [ 
https://issues.apache.org/jira/browse/BEAM-3809?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mark Liu closed BEAM-3809.
--

> beam_PerformanceTests_Python failing
> 
>
> Key: BEAM-3809
> URL: https://issues.apache.org/jira/browse/BEAM-3809
> Project: Beam
>  Issue Type: Bug
>  Components: testing
>Reporter: Ahmet Altay
>Assignee: Mark Liu
>Priority: Major
> Fix For: Not applicable
>
>  Time Spent: 0.5h
>  Remaining Estimate: 0h
>
> Looks like a misconfiguration based on the error.
>  
> Example test:
> [https://builds.apache.org/view/A-D/view/Beam/job/beam_PerformanceTests_Python/1000/console]
>  
> Error I see:
> error: option --test-pipeline-options not recognized
> 2018-03-08 18:48:08,202 9e009d45 MainThread beam_integration_benchmark(1/1) 
> ERRORError during benchmark beam_integration_benchmark
> Traceback (most recent call last):
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PerformanceTests_Python/PerfKitBenchmarker/perfkitbenchmarker/pkb.py",
>  line 622, in RunBenchmark
> DoRunPhase(spec, collector, detailed_timer)
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PerformanceTests_Python/PerfKitBenchmarker/perfkitbenchmarker/pkb.py",
>  line 525, in DoRunPhase
> samples = spec.BenchmarkRun(spec)
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PerformanceTests_Python/PerfKitBenchmarker/perfkitbenchmarker/linux_benchmarks/beam_integration_benchmark.py",
>  line 159, in Run
> job_type=job_type)
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PerformanceTests_Python/PerfKitBenchmarker/perfkitbenchmarker/providers/gcp/gcp_dpb_dataflow.py",
>  line 90, in SubmitJob
> assert retcode == 0, "Integration Test Failed."



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Resolved] (BEAM-3809) beam_PerformanceTests_Python failing

2018-08-06 Thread Mark Liu (JIRA)


 [ 
https://issues.apache.org/jira/browse/BEAM-3809?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mark Liu resolved BEAM-3809.

   Resolution: Fixed
Fix Version/s: Not applicable

> beam_PerformanceTests_Python failing
> 
>
> Key: BEAM-3809
> URL: https://issues.apache.org/jira/browse/BEAM-3809
> Project: Beam
>  Issue Type: Bug
>  Components: testing
>Reporter: Ahmet Altay
>Assignee: Mark Liu
>Priority: Major
> Fix For: Not applicable
>
>  Time Spent: 0.5h
>  Remaining Estimate: 0h
>
> Looks like a misconfiguration based on the error.
>  
> Example test:
> [https://builds.apache.org/view/A-D/view/Beam/job/beam_PerformanceTests_Python/1000/console]
>  
> Error I see:
> error: option --test-pipeline-options not recognized
> 2018-03-08 18:48:08,202 9e009d45 MainThread beam_integration_benchmark(1/1) 
> ERRORError during benchmark beam_integration_benchmark
> Traceback (most recent call last):
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PerformanceTests_Python/PerfKitBenchmarker/perfkitbenchmarker/pkb.py",
>  line 622, in RunBenchmark
> DoRunPhase(spec, collector, detailed_timer)
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PerformanceTests_Python/PerfKitBenchmarker/perfkitbenchmarker/pkb.py",
>  line 525, in DoRunPhase
> samples = spec.BenchmarkRun(spec)
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PerformanceTests_Python/PerfKitBenchmarker/perfkitbenchmarker/linux_benchmarks/beam_integration_benchmark.py",
>  line 159, in Run
> job_type=job_type)
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PerformanceTests_Python/PerfKitBenchmarker/perfkitbenchmarker/providers/gcp/gcp_dpb_dataflow.py",
>  line 90, in SubmitJob
> assert retcode == 0, "Integration Test Failed."



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Closed] (BEAM-3230) beam_PerformanceTests_Python is failing

2018-08-06 Thread Mark Liu (JIRA)


 [ 
https://issues.apache.org/jira/browse/BEAM-3230?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mark Liu closed BEAM-3230.
--

> beam_PerformanceTests_Python is failing
> ---
>
> Key: BEAM-3230
> URL: https://issues.apache.org/jira/browse/BEAM-3230
> Project: Beam
>  Issue Type: Bug
>  Components: sdk-py-core
>Reporter: Ankur Goenka
>Assignee: Mark Liu
>Priority: Critical
> Fix For: Not applicable
>
>  Time Spent: 40m
>  Remaining Estimate: 0h
>
> Jenkings beam_PerformanceTests_Python stage is failing for python builds.
> Here is the link to a failure console output 
> https://builds.apache.org/job/beam_PerformanceTests_Python/582/console



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (BEAM-3230) beam_PerformanceTests_Python is failing

2018-08-06 Thread Mark Liu (JIRA)


[ 
https://issues.apache.org/jira/browse/BEAM-3230?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16570810#comment-16570810
 ] 

Mark Liu commented on BEAM-3230:


Fix is merged. Latest schedule build passed: 
https://builds.apache.org/job/beam_PerformanceTests_Python/1272/

Will close this ticket.

> beam_PerformanceTests_Python is failing
> ---
>
> Key: BEAM-3230
> URL: https://issues.apache.org/jira/browse/BEAM-3230
> Project: Beam
>  Issue Type: Bug
>  Components: sdk-py-core
>Reporter: Ankur Goenka
>Assignee: Ankur Goenka
>Priority: Critical
> Fix For: Not applicable
>
>  Time Spent: 40m
>  Remaining Estimate: 0h
>
> Jenkings beam_PerformanceTests_Python stage is failing for python builds.
> Here is the link to a failure console output 
> https://builds.apache.org/job/beam_PerformanceTests_Python/582/console



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Resolved] (BEAM-3230) beam_PerformanceTests_Python is failing

2018-08-06 Thread Mark Liu (JIRA)


 [ 
https://issues.apache.org/jira/browse/BEAM-3230?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mark Liu resolved BEAM-3230.

   Resolution: Fixed
Fix Version/s: Not applicable

> beam_PerformanceTests_Python is failing
> ---
>
> Key: BEAM-3230
> URL: https://issues.apache.org/jira/browse/BEAM-3230
> Project: Beam
>  Issue Type: Bug
>  Components: sdk-py-core
>Reporter: Ankur Goenka
>Assignee: Mark Liu
>Priority: Critical
> Fix For: Not applicable
>
>  Time Spent: 40m
>  Remaining Estimate: 0h
>
> Jenkings beam_PerformanceTests_Python stage is failing for python builds.
> Here is the link to a failure console output 
> https://builds.apache.org/job/beam_PerformanceTests_Python/582/console



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Assigned] (BEAM-3230) beam_PerformanceTests_Python is failing

2018-08-06 Thread Mark Liu (JIRA)


 [ 
https://issues.apache.org/jira/browse/BEAM-3230?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mark Liu reassigned BEAM-3230:
--

Assignee: Mark Liu  (was: Ankur Goenka)

> beam_PerformanceTests_Python is failing
> ---
>
> Key: BEAM-3230
> URL: https://issues.apache.org/jira/browse/BEAM-3230
> Project: Beam
>  Issue Type: Bug
>  Components: sdk-py-core
>Reporter: Ankur Goenka
>Assignee: Mark Liu
>Priority: Critical
> Fix For: Not applicable
>
>  Time Spent: 40m
>  Remaining Estimate: 0h
>
> Jenkings beam_PerformanceTests_Python stage is failing for python builds.
> Here is the link to a failure console output 
> https://builds.apache.org/job/beam_PerformanceTests_Python/582/console



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (BEAM-3230) beam_PerformanceTests_Python is failing

2018-08-06 Thread Mark Liu (JIRA)


[ 
https://issues.apache.org/jira/browse/BEAM-3230?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16570809#comment-16570809
 ] 

Mark Liu commented on BEAM-3230:


https://github.com/apache/beam/pull/6142 will do the fix

> beam_PerformanceTests_Python is failing
> ---
>
> Key: BEAM-3230
> URL: https://issues.apache.org/jira/browse/BEAM-3230
> Project: Beam
>  Issue Type: Bug
>  Components: sdk-py-core
>Reporter: Ankur Goenka
>Assignee: Ankur Goenka
>Priority: Critical
>  Time Spent: 40m
>  Remaining Estimate: 0h
>
> Jenkings beam_PerformanceTests_Python stage is failing for python builds.
> Here is the link to a failure console output 
> https://builds.apache.org/job/beam_PerformanceTests_Python/582/console



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (BEAM-3230) beam_PerformanceTests_Python is failing

2018-08-03 Thread Mark Liu (JIRA)


[ 
https://issues.apache.org/jira/browse/BEAM-3230?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16568427#comment-16568427
 ] 

Mark Liu commented on BEAM-3230:


Works in https://github.com/apache/beam/pull/6074 is trying to fix this test. 
The current problem is in worker and pipeline can't finish. Welcome any help on 
that.

> beam_PerformanceTests_Python is failing
> ---
>
> Key: BEAM-3230
> URL: https://issues.apache.org/jira/browse/BEAM-3230
> Project: Beam
>  Issue Type: Bug
>  Components: sdk-py-core
>Reporter: Ankur Goenka
>Assignee: Ankur Goenka
>Priority: Critical
>  Time Spent: 40m
>  Remaining Estimate: 0h
>
> Jenkings beam_PerformanceTests_Python stage is failing for python builds.
> Here is the link to a failure console output 
> https://builds.apache.org/job/beam_PerformanceTests_Python/582/console



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Closed] (BEAM-4859) Run Python ValidatesRunner tests in streaming

2018-07-26 Thread Mark Liu (JIRA)


 [ 
https://issues.apache.org/jira/browse/BEAM-4859?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mark Liu closed BEAM-4859.
--

> Run Python ValidatesRunner tests in streaming
> -
>
> Key: BEAM-4859
> URL: https://issues.apache.org/jira/browse/BEAM-4859
> Project: Beam
>  Issue Type: Test
>  Components: testing
>Reporter: Mark Liu
>Assignee: Mark Liu
>Priority: Major
> Fix For: Not applicable
>
>  Time Spent: 4h 40m
>  Remaining Estimate: 0h
>
> Automated way to run existing Python ValidatesRunner tests in streaming. 
> Gradle should also be able to run them, so Jenkins automation will be much 
> simpler.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Resolved] (BEAM-4859) Run Python ValidatesRunner tests in streaming

2018-07-26 Thread Mark Liu (JIRA)


 [ 
https://issues.apache.org/jira/browse/BEAM-4859?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mark Liu resolved BEAM-4859.

   Resolution: Done
Fix Version/s: Not applicable

> Run Python ValidatesRunner tests in streaming
> -
>
> Key: BEAM-4859
> URL: https://issues.apache.org/jira/browse/BEAM-4859
> Project: Beam
>  Issue Type: Test
>  Components: testing
>Reporter: Mark Liu
>Assignee: Mark Liu
>Priority: Major
> Fix For: Not applicable
>
>  Time Spent: 4h 40m
>  Remaining Estimate: 0h
>
> Automated way to run existing Python ValidatesRunner tests in streaming. 
> Gradle should also be able to run them, so Jenkins automation will be much 
> simpler.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (BEAM-4859) Run Python ValidatesRunner tests in streaming

2018-07-26 Thread Mark Liu (JIRA)


[ 
https://issues.apache.org/jira/browse/BEAM-4859?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16558474#comment-16558474
 ] 

Mark Liu commented on BEAM-4859:


Python ValidatesRunner tests are enabled in streaming and added to same Jenkins 
suite beam_PostCommit_Py_VR_Dataflow.

Failed tests are tracked in https://issues.apache.org/jira/browse/BEAM-5025 and 
https://issues.apache.org/jira/browse/BEAM-3544

> Run Python ValidatesRunner tests in streaming
> -
>
> Key: BEAM-4859
> URL: https://issues.apache.org/jira/browse/BEAM-4859
> Project: Beam
>  Issue Type: Test
>  Components: testing
>Reporter: Mark Liu
>Assignee: Mark Liu
>Priority: Major
>  Time Spent: 4h 40m
>  Remaining Estimate: 0h
>
> Automated way to run existing Python ValidatesRunner tests in streaming. 
> Gradle should also be able to run them, so Jenkins automation will be much 
> simpler.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (BEAM-5025) Fix test_multi_valued_singleton_side_input on DataflowRunner in streaming

2018-07-25 Thread Mark Liu (JIRA)
Mark Liu created BEAM-5025:
--

 Summary: Fix test_multi_valued_singleton_side_input on 
DataflowRunner in streaming
 Key: BEAM-5025
 URL: https://issues.apache.org/jira/browse/BEAM-5025
 Project: Beam
  Issue Type: Bug
  Components: sdk-py-core, testing
Reporter: Mark Liu
Assignee: Ahmet Altay


test_multi_valued_singleton_side_input is a VR test and failed to run on 
DataflowRunner in streaming. 

Example failure link from pantheon:
https://pantheon.corp.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-07-25_16_16_40-5301315089771762711?project=apache-beam-testing

log:
  File 
"/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Py_VR_Dataflow_PR/src/sdks/python/apache_beam/pvalue.py",
 line 422, in _from_runtime_iterable
ValueError: PCollection of size 2 with more than one element accessed as a 
singleton view. First two elements encountered are "3", "4".

We should reenable this test in beam_PostCommit_Py_VR_Dataflow after fix it.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (BEAM-1124) Python ValidateRunner Test test_multi_valued_singleton_side_input Break Postcommit

2018-07-25 Thread Mark Liu (JIRA)


[ 
https://issues.apache.org/jira/browse/BEAM-1124?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16556453#comment-16556453
 ] 

Mark Liu commented on BEAM-1124:


The fix should work. This test is passing for a while.

Latest green Jenkins run for VR test:
https://builds.apache.org/view/A-D/view/Beam/job/beam_PostCommit_Py_VR_Dataflow/642/

> Python ValidateRunner Test test_multi_valued_singleton_side_input Break 
> Postcommit
> --
>
> Key: BEAM-1124
> URL: https://issues.apache.org/jira/browse/BEAM-1124
> Project: Beam
>  Issue Type: Bug
>  Components: runner-dataflow, sdk-py-core, testing
>Reporter: Mark Liu
>Assignee: Ahmet Altay
>Priority: Major
> Fix For: Not applicable
>
>
> Python test_multi_valued_singleton_side_input test, a ValidatesRunner test 
> that running on dataflow service, failed and broke 
> postcommit(https://builds.apache.org/view/Beam/job/beam_PostCommit_Python_Verify/853/).
> Here is the stack trace:
> {code}
> Traceback (most recent call last):
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python_Verify/sdks/python/apache_beam/dataflow_test.py",
>  line 186, in test_multi_valued_singleton_side_input
> pipeline.run()
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python_Verify/sdks/python/apache_beam/pipeline.py",
>  line 159, in run
> return self.runner.run(self)
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python_Verify/sdks/python/apache_beam/runners/dataflow_runner.py",
>  line 195, in run
> % getattr(self, 'last_error_msg', None), self.result)
> DataflowRuntimeException: Dataflow pipeline failed:
> (99aeafa7a8dffcc7): Traceback (most recent call last):
>   File 
> "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 
> 514, in do_work
> work_executor.execute()
>   File "dataflow_worker/executor.py", line 892, in 
> dataflow_worker.executor.MapTaskExecutor.execute 
> (dataflow_worker/executor.c:24008)
> op.start()
>   File "dataflow_worker/executor.py", line 456, in 
> dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:13870)
> def start(self):
>   File "dataflow_worker/executor.py", line 483, in 
> dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:13685)
> self.dofn_runner = common.DoFnRunner(
>   File "apache_beam/runners/common.py", line 89, in 
> apache_beam.runners.common.DoFnRunner.__init__ 
> (apache_beam/runners/common.c:3469)
> args, kwargs, [side_input[global_window]
>   File 
> "/usr/local/lib/python2.7/dist-packages/apache_beam/transforms/sideinputs.py",
>  line 192, in __getitem__
> _FilteringIterable(self._iterable, target_window), self._view_options)
>   File "/usr/local/lib/python2.7/dist-packages/apache_beam/pvalue.py", line 
> 279, in _from_runtime_iterable
> 'PCollection with more than one element accessed as '
> ValueError: PCollection with more than one element accessed as a singleton 
> view.
> {code}
> Worker logs in here:
> https://builds.apache.org/view/Beam/job/beam_PostCommit_Python_Verify/853/console
> In order to temporarily ignore this test in postcommit, we can comment out 
> annotation "@attr('ValidatesRunner')" of this test. Then it will only run as 
> a unit test (execute by DirectRunner), but not run as a ValidatesRunner test.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (BEAM-4859) Run Python ValidatesRunner tests in streaming

2018-07-24 Thread Mark Liu (JIRA)
Mark Liu created BEAM-4859:
--

 Summary: Run Python ValidatesRunner tests in streaming
 Key: BEAM-4859
 URL: https://issues.apache.org/jira/browse/BEAM-4859
 Project: Beam
  Issue Type: Test
  Components: testing
Reporter: Mark Liu
Assignee: Mark Liu


Automated way to run existing Python ValidatesRunner tests in streaming. Gradle 
should also be able to run them, so Jenkins automation will be much simpler.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (BEAM-4492) Update Python bigquery library to latest version

2018-07-19 Thread Mark Liu (JIRA)


[ 
https://issues.apache.org/jira/browse/BEAM-4492?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16549747#comment-16549747
 ] 

Mark Liu commented on BEAM-4492:


Following link shows places that use google-cloud-bigquery library:
https://github.com/apache/beam/search?l=Python=%22from+google.cloud+import+bigquery%22

> Update Python bigquery library to latest version
> 
>
> Key: BEAM-4492
> URL: https://issues.apache.org/jira/browse/BEAM-4492
> Project: Beam
>  Issue Type: Task
>  Components: testing
>Reporter: Mark Liu
>Assignee: Mark Liu
>Priority: Major
>
> Current google-cloud-bigquery is set to 0.25.0 in 
> https://github.com/apache/beam/blob/master/sdks/python/setup.py#L130
> However, the latest version is 1.2.0.
> According to comment in setup.py, this library is only used for testing, so 
> it should be easy to update. 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Closed] (BEAM-4512) Move DataflowRunner off of Maven build files

2018-07-16 Thread Mark Liu (JIRA)


 [ 
https://issues.apache.org/jira/browse/BEAM-4512?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mark Liu closed BEAM-4512.
--

> Move DataflowRunner off of Maven build files
> 
>
> Key: BEAM-4512
> URL: https://issues.apache.org/jira/browse/BEAM-4512
> Project: Beam
>  Issue Type: Sub-task
>  Components: runner-dataflow
>Reporter: Chamikara Jayalath
>Assignee: Mark Liu
>Priority: Major
> Fix For: 2.6.0
>
>
> Currently DataflowRunner (internally at Google) depends on Beam's Maven build 
> files. We have to move some internal build targets to use Gradle so that 
> Maven files can be deleted from Beam.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Resolved] (BEAM-4512) Move DataflowRunner off of Maven build files

2018-07-16 Thread Mark Liu (JIRA)


 [ 
https://issues.apache.org/jira/browse/BEAM-4512?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mark Liu resolved BEAM-4512.

   Resolution: Fixed
Fix Version/s: 2.6.0

> Move DataflowRunner off of Maven build files
> 
>
> Key: BEAM-4512
> URL: https://issues.apache.org/jira/browse/BEAM-4512
> Project: Beam
>  Issue Type: Sub-task
>  Components: runner-dataflow
>Reporter: Chamikara Jayalath
>Assignee: Mark Liu
>Priority: Major
> Fix For: 2.6.0
>
>
> Currently DataflowRunner (internally at Google) depends on Beam's Maven build 
> files. We have to move some internal build targets to use Gradle so that 
> Maven files can be deleted from Beam.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Comment Edited] (BEAM-4512) Move DataflowRunner off of Maven build files

2018-07-09 Thread Mark Liu (JIRA)


[ 
https://issues.apache.org/jira/browse/BEAM-4512?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16537751#comment-16537751
 ] 

Mark Liu edited comment on BEAM-4512 at 7/9/18 10:57 PM:
-

The work inside Google that moves project build dependency from Maven pom to 
Gradle is done and no longer a blocker to delete pom in Beam.


was (Author: markflyhigh):
The work inside Google that moves project build dependency from Maven pom to 
Gradle is done and no longer a blocker to delete pom in Beam.

> Move DataflowRunner off of Maven build files
> 
>
> Key: BEAM-4512
> URL: https://issues.apache.org/jira/browse/BEAM-4512
> Project: Beam
>  Issue Type: Sub-task
>  Components: runner-dataflow
>Reporter: Chamikara Jayalath
>Assignee: Mark Liu
>Priority: Major
>
> Currently DataflowRunner (internally at Google) depends on Beam's Maven build 
> files. We have to move some internal build targets to use Gradle so that 
> Maven files can be deleted from Beam.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (BEAM-4512) Move DataflowRunner off of Maven build files

2018-07-09 Thread Mark Liu (JIRA)


[ 
https://issues.apache.org/jira/browse/BEAM-4512?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16537751#comment-16537751
 ] 

Mark Liu commented on BEAM-4512:


The work inside Google that moves project build dependency from Maven pom to 
Gradle is done and no longer a blocker to delete pom in Beam.

> Move DataflowRunner off of Maven build files
> 
>
> Key: BEAM-4512
> URL: https://issues.apache.org/jira/browse/BEAM-4512
> Project: Beam
>  Issue Type: Sub-task
>  Components: runner-dataflow
>Reporter: Chamikara Jayalath
>Assignee: Mark Liu
>Priority: Major
>
> Currently DataflowRunner (internally at Google) depends on Beam's Maven build 
> files. We have to move some internal build targets to use Gradle so that 
> Maven files can be deleted from Beam.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Closed] (BEAM-4743) jars are overwritten during release with -Ppublishing

2018-07-09 Thread Mark Liu (JIRA)


 [ 
https://issues.apache.org/jira/browse/BEAM-4743?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mark Liu closed BEAM-4743.
--
   Resolution: Duplicate
Fix Version/s: Not applicable

duplicated with BEAM-4744

> jars are overwritten during release with -Ppublishing
> -
>
> Key: BEAM-4743
> URL: https://issues.apache.org/jira/browse/BEAM-4743
> Project: Beam
>  Issue Type: Bug
>  Components: build-system
>Reporter: Mark Liu
>Assignee: Alan Myrvold
>Priority: Major
> Fix For: Not applicable
>
>
> +cc [~pabloem] [~boyuanz]
> Few jars (*-shaded.jar and *-tests.jar) are missing if I run
> `./gradlew -Ppublishing :publishToMavenLocal`.
> However, I can get them through assemble and shadowJar. For example, 
> `./gradlew :beam-sdks-java-core:assemble :beam-sdks-java-core:shadowJar`
> It could be the reason why we can't run publishing in parallel. 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (BEAM-4744) Jars are overwritten during release with -Ppublishing

2018-07-09 Thread Mark Liu (JIRA)
Mark Liu created BEAM-4744:
--

 Summary: Jars are overwritten during release with -Ppublishing
 Key: BEAM-4744
 URL: https://issues.apache.org/jira/browse/BEAM-4744
 Project: Beam
  Issue Type: Bug
  Components: build-system
Reporter: Mark Liu
Assignee: Alan Myrvold


+cc [~pabloem] [~boyuanz] [~lukecwik]

Few jars (*-shaded.jar and *-tests.jar) are missing if I run `./gradlew 
-Ppublishing :publishToMavenLocal`. 

However, I can get them through assemble and shadowJar. For example, `./gradlew 
:beam-sdks-java-core:assemble :beam-sdks-java-core:shadowJar`.

This could be the reason why we can't run parallel build during release.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (BEAM-4743) jars are overwritten during release with -Ppublishing

2018-07-09 Thread Mark Liu (JIRA)
Mark Liu created BEAM-4743:
--

 Summary: jars are overwritten during release with -Ppublishing
 Key: BEAM-4743
 URL: https://issues.apache.org/jira/browse/BEAM-4743
 Project: Beam
  Issue Type: Bug
  Components: build-system
Reporter: Mark Liu
Assignee: Alan Myrvold


+cc [~pabloem] [~boyuanz]

Few jars (*-shaded.jar and *-tests.jar) are missing if I run
`./gradlew -Ppublishing :publishToMavenLocal`.

However, I can get them through assemble and shadowJar. For example, `./gradlew 
:beam-sdks-java-core:assemble :beam-sdks-java-core:shadowJar`

It could be the reason why we can't run publishing in parallel. 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (BEAM-4512) Move DataflowRunner off of Maven build files

2018-06-22 Thread Mark Liu (JIRA)


[ 
https://issues.apache.org/jira/browse/BEAM-4512?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16520655#comment-16520655
 ] 

Mark Liu commented on BEAM-4512:


We made a good progress in recent days:
- migrate build_rules script to buildSrc plugin. merged 
[PR5683|https://github.com/apache/beam/pull/5683] and 
[PR5730|https://github.com/apache/beam/pull/5730].
- allow people specify repositores for build scripts by providing extra 
property "extraMavenUrls". merged 
[PR5728|https://github.com/apache/beam/pull/5728].

> Move DataflowRunner off of Maven build files
> 
>
> Key: BEAM-4512
> URL: https://issues.apache.org/jira/browse/BEAM-4512
> Project: Beam
>  Issue Type: Sub-task
>  Components: runner-dataflow
>Reporter: Chamikara Jayalath
>Assignee: Mark Liu
>Priority: Major
>
> Currently DataflowRunner (internally at Google) depends on Beam's Maven build 
> files. We have to move some internal build targets to use Gradle so that 
> Maven files can be deleted from Beam.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (BEAM-4492) Update Python bigquery library to latest version

2018-06-05 Thread Mark Liu (JIRA)
Mark Liu created BEAM-4492:
--

 Summary: Update Python bigquery library to latest version
 Key: BEAM-4492
 URL: https://issues.apache.org/jira/browse/BEAM-4492
 Project: Beam
  Issue Type: Task
  Components: testing
Reporter: Mark Liu
Assignee: Mark Liu


Current google-cloud-bigquery is set to 0.25.0 in 
https://github.com/apache/beam/blob/master/sdks/python/setup.py#L130

However, the latest version is 1.2.0.

According to comment in setup.py, this library is only used for testing, so it 
should be easy to update. 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (BEAM-3944) Convert beam_PerformanceTests_Python to use Gradle

2018-05-09 Thread Mark Liu (JIRA)

[ 
https://issues.apache.org/jira/browse/BEAM-3944?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16469294#comment-16469294
 ] 

Mark Liu commented on BEAM-3944:


https://issues.apache.org/jira/browse/BEAM-3809 is tracking the Python 
performance test failure. The failure is not very obvious, need more 
investigation on Perfkit side. However, I may not be able to spend more time on 
that. Disable the test probably a better solution.

> Convert beam_PerformanceTests_Python to use Gradle
> --
>
> Key: BEAM-3944
> URL: https://issues.apache.org/jira/browse/BEAM-3944
> Project: Beam
>  Issue Type: Sub-task
>  Components: testing
>Reporter: Chamikara Jayalath
>Priority: Major
>




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Closed] (BEAM-4092) Enable Python ValidatesRunner test against TestDataflowRunner in streaming

2018-04-16 Thread Mark Liu (JIRA)

 [ 
https://issues.apache.org/jira/browse/BEAM-4092?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mark Liu closed BEAM-4092.
--
   Resolution: Duplicate
Fix Version/s: Not applicable

> Enable Python ValidatesRunner test against TestDataflowRunner in streaming
> --
>
> Key: BEAM-4092
> URL: https://issues.apache.org/jira/browse/BEAM-4092
> Project: Beam
>  Issue Type: New Feature
>  Components: sdk-py-core, testing
>Reporter: Mark Liu
>Assignee: Mark Liu
>Priority: Major
> Fix For: Not applicable
>
>




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (BEAM-4093) Support Python ValidatesRunner test against TestDataflowRunner in streaming

2018-04-16 Thread Mark Liu (JIRA)
Mark Liu created BEAM-4093:
--

 Summary: Support Python ValidatesRunner test against 
TestDataflowRunner in streaming
 Key: BEAM-4093
 URL: https://issues.apache.org/jira/browse/BEAM-4093
 Project: Beam
  Issue Type: New Feature
  Components: sdk-py-core, testing
Reporter: Mark Liu
Assignee: Mark Liu






--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (BEAM-4092) Enable Python ValidatesRunner test against TestDataflowRunner in streaming

2018-04-16 Thread Mark Liu (JIRA)
Mark Liu created BEAM-4092:
--

 Summary: Enable Python ValidatesRunner test against 
TestDataflowRunner in streaming
 Key: BEAM-4092
 URL: https://issues.apache.org/jira/browse/BEAM-4092
 Project: Beam
  Issue Type: New Feature
  Components: sdk-py-core, testing
Reporter: Mark Liu
Assignee: Mark Liu






--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Resolved] (BEAM-3946) Python SDK tests are failing if no GOOGLE_APPLICATION_CREDENTIALS was set

2018-04-04 Thread Mark Liu (JIRA)

 [ 
https://issues.apache.org/jira/browse/BEAM-3946?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mark Liu resolved BEAM-3946.

   Resolution: Fixed
Fix Version/s: Not applicable

> Python SDK tests are failing if no GOOGLE_APPLICATION_CREDENTIALS was set
> -
>
> Key: BEAM-3946
> URL: https://issues.apache.org/jira/browse/BEAM-3946
> Project: Beam
>  Issue Type: Bug
>  Components: examples-python
>Reporter: Alexey Romanenko
>Assignee: Mark Liu
>Priority: Major
> Fix For: Not applicable
>
>  Time Spent: 20m
>  Remaining Estimate: 0h
>
> Running locally mvn clean install fails on following Apache Beam :: SDKs :: 
> Python tests:
> {{ERROR: test_message_matcher_mismatch 
> (apache_beam.io.gcp.tests.pubsub_matcher_test.PubSubMatcherTest)}}
>  {{ERROR: test_message_matcher_success 
> (apache_beam.io.gcp.tests.pubsub_matcher_test.PubSubMatcherTest)}}
>  {{ERROR: test_message_metcher_timeout 
> (apache_beam.io.gcp.tests.pubsub_matcher_test.PubSubMatcherTest)}}
>  
> with an error:
> DefaultCredentialsError: Could not automatically determine credentials. 
> Please set GOOGLE_APPLICATION_CREDENTIALS or
>  explicitly create credential and re-run the application. For more
>  information, please see
>  
> [https://developers.google.com/accounts/docs/application-default-credentials].
>   >> begin captured logging << 
>  google.auth.transport._http_client: DEBUG: Making request: GET 
> [http://169.254.169.254|http://169.254.169.254/]
>  google.auth.compute_engine._metadata: INFO: Compute Engine Metadata server 
> unavailable.
>  - >> end captured logging << -
>  
> It looks like it's a regression and it was caused by this commit: 
> [301853647f2c726c04c5bdb02cab6ff6b39f09d0|https://github.com/apache/beam/commit/301853647f2c726c04c5bdb02cab6ff6b39f09d0]



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (BEAM-3946) Python SDK tests are failing if no GOOGLE_APPLICATION_CREDENTIALS was set

2018-04-04 Thread Mark Liu (JIRA)

[ 
https://issues.apache.org/jira/browse/BEAM-3946?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16426204#comment-16426204
 ] 

Mark Liu commented on BEAM-3946:


https://github.com/apache/beam/pull/5021 should fix this problem. 

> Python SDK tests are failing if no GOOGLE_APPLICATION_CREDENTIALS was set
> -
>
> Key: BEAM-3946
> URL: https://issues.apache.org/jira/browse/BEAM-3946
> Project: Beam
>  Issue Type: Bug
>  Components: examples-python
>Reporter: Alexey Romanenko
>Assignee: Mark Liu
>Priority: Major
>  Time Spent: 20m
>  Remaining Estimate: 0h
>
> Running locally mvn clean install fails on following Apache Beam :: SDKs :: 
> Python tests:
> {{ERROR: test_message_matcher_mismatch 
> (apache_beam.io.gcp.tests.pubsub_matcher_test.PubSubMatcherTest)}}
>  {{ERROR: test_message_matcher_success 
> (apache_beam.io.gcp.tests.pubsub_matcher_test.PubSubMatcherTest)}}
>  {{ERROR: test_message_metcher_timeout 
> (apache_beam.io.gcp.tests.pubsub_matcher_test.PubSubMatcherTest)}}
>  
> with an error:
> DefaultCredentialsError: Could not automatically determine credentials. 
> Please set GOOGLE_APPLICATION_CREDENTIALS or
>  explicitly create credential and re-run the application. For more
>  information, please see
>  
> [https://developers.google.com/accounts/docs/application-default-credentials].
>   >> begin captured logging << 
>  google.auth.transport._http_client: DEBUG: Making request: GET 
> [http://169.254.169.254|http://169.254.169.254/]
>  google.auth.compute_engine._metadata: INFO: Compute Engine Metadata server 
> unavailable.
>  - >> end captured logging << -
>  
> It looks like it's a regression and it was caused by this commit: 
> [301853647f2c726c04c5bdb02cab6ff6b39f09d0|https://github.com/apache/beam/commit/301853647f2c726c04c5bdb02cab6ff6b39f09d0]



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Resolved] (BEAM-2861) test_delete_bq_table_succeeds fails with GOOGLE_APPLICATION_CREDENTIALS

2018-04-04 Thread Mark Liu (JIRA)

 [ 
https://issues.apache.org/jira/browse/BEAM-2861?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mark Liu resolved BEAM-2861.

   Resolution: Fixed
Fix Version/s: Not applicable

> test_delete_bq_table_succeeds fails with GOOGLE_APPLICATION_CREDENTIALS
> ---
>
> Key: BEAM-2861
> URL: https://issues.apache.org/jira/browse/BEAM-2861
> Project: Beam
>  Issue Type: Bug
>  Components: sdk-py-core
>Reporter: Ahmet Altay
>Assignee: Mark Liu
>Priority: Major
> Fix For: Not applicable
>
>
> This is a variation of the https://issues.apache.org/jira/browse/BEAM-2101
> The tests are not skipped if the GCP libraries are installed. But the tests 
> also require GCP authentication. We should probably also skip the tests if 
> GCP is installed but user is not authenticated.
> cc: [~pei...@gmail.com]



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Closed] (BEAM-2861) test_delete_bq_table_succeeds fails with GOOGLE_APPLICATION_CREDENTIALS

2018-04-04 Thread Mark Liu (JIRA)

 [ 
https://issues.apache.org/jira/browse/BEAM-2861?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mark Liu closed BEAM-2861.
--

> test_delete_bq_table_succeeds fails with GOOGLE_APPLICATION_CREDENTIALS
> ---
>
> Key: BEAM-2861
> URL: https://issues.apache.org/jira/browse/BEAM-2861
> Project: Beam
>  Issue Type: Bug
>  Components: sdk-py-core
>Reporter: Ahmet Altay
>Assignee: Mark Liu
>Priority: Major
> Fix For: Not applicable
>
>
> This is a variation of the https://issues.apache.org/jira/browse/BEAM-2101
> The tests are not skipped if the GCP libraries are installed. But the tests 
> also require GCP authentication. We should probably also skip the tests if 
> GCP is installed but user is not authenticated.
> cc: [~pei...@gmail.com]



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Resolved] (BEAM-1033) BigQueryMatcher is flaky

2018-04-04 Thread Mark Liu (JIRA)

 [ 
https://issues.apache.org/jira/browse/BEAM-1033?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mark Liu resolved BEAM-1033.

   Resolution: Fixed
Fix Version/s: Not applicable

> BigQueryMatcher is flaky
> 
>
> Key: BEAM-1033
> URL: https://issues.apache.org/jira/browse/BEAM-1033
> Project: Beam
>  Issue Type: Bug
>  Components: testing
>Reporter: Pei He
>Assignee: Mark Liu
>Priority: Major
> Fix For: Not applicable
>
>
> Jenkins link:
> https://builds.apache.org/job/beam_PreCommit_MavenVerify/5145/console
> Running org.apache.beam.examples.WindowedWordCountIT
> Tests run: 2, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 304.282 sec 
> <<< FAILURE! - in org.apache.beam.examples.WindowedWordCountIT
> testWindowedWordCountInBatch(org.apache.beam.examples.WindowedWordCountIT)  
> Time elapsed: 304.282 sec  <<< FAILURE!
> java.lang.AssertionError: 
> Expected: Expected checksum is (cd5b52939257e12428a9fa085c32a84dd209b180)
>  but: Invalid BigQuery response: 
> {"jobComplete":false,"jobReference":{"jobId":"job_0STNX_OD83tQOzo6MvmqXCrk61U","projectId":"apache-beam-testing"},"kind":"bigquery#queryResponse"}
>   at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:20)
>   at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:8)
>   at 
> org.apache.beam.runners.dataflow.testing.TestDataflowRunner.run(TestDataflowRunner.java:164)
>   at 
> org.apache.beam.runners.dataflow.testing.TestDataflowRunner.run(TestDataflowRunner.java:93)
>   at 
> org.apache.beam.runners.dataflow.testing.TestDataflowRunner.run(TestDataflowRunner.java:61)
>   at org.apache.beam.sdk.Pipeline.run(Pipeline.java:179)
>   at 
> org.apache.beam.examples.WindowedWordCount.main(WindowedWordCount.java:224)
>   at 
> org.apache.beam.examples.WindowedWordCountIT.testWindowedWordCountPipeline(WindowedWordCountIT.java:88)
>   at 
> org.apache.beam.examples.WindowedWordCountIT.testWindowedWordCountInBatch(WindowedWordCountIT.java:59)
>   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>   at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>   at java.lang.reflect.Method.invoke(Method.java:498)
>   at 
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47)
>   at 
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
>   at 
> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44)
>   at 
> org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
>   at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:271)
>   at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:70)
>   at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
>   at org.junit.runners.ParentRunner$3.run(ParentRunner.java:238)
>   at 
> org.apache.maven.surefire.junitcore.pc.Scheduler$1.run(Scheduler.java:393)
>   at 
> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
>   at java.util.concurrent.FutureTask.run(FutureTask.java:266)
>   at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
>   at java.lang.Thread.run(Thread.java:745)
> Results :
> Failed tests: 
>   
> WindowedWordCountIT.testWindowedWordCountInBatch:59->testWindowedWordCountPipeline:88
>  
> Expected: Expected checksum is (cd5b52939257e12428a9fa085c32a84dd209b180)
>  but: Invalid BigQuery response: 
> {"jobComplete":false,"jobReference":{"jobId":"job_0STNX_OD83tQOzo6MvmqXCrk61U","projectId":"apache-beam-testing"},"kind":"bigquery#queryResponse"}
> Tests run: 3, Failures: 1, Errors: 0, Skipped: 0



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Closed] (BEAM-1033) BigQueryMatcher is flaky

2018-04-04 Thread Mark Liu (JIRA)

 [ 
https://issues.apache.org/jira/browse/BEAM-1033?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mark Liu closed BEAM-1033.
--

> BigQueryMatcher is flaky
> 
>
> Key: BEAM-1033
> URL: https://issues.apache.org/jira/browse/BEAM-1033
> Project: Beam
>  Issue Type: Bug
>  Components: testing
>Reporter: Pei He
>Assignee: Mark Liu
>Priority: Major
> Fix For: Not applicable
>
>
> Jenkins link:
> https://builds.apache.org/job/beam_PreCommit_MavenVerify/5145/console
> Running org.apache.beam.examples.WindowedWordCountIT
> Tests run: 2, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 304.282 sec 
> <<< FAILURE! - in org.apache.beam.examples.WindowedWordCountIT
> testWindowedWordCountInBatch(org.apache.beam.examples.WindowedWordCountIT)  
> Time elapsed: 304.282 sec  <<< FAILURE!
> java.lang.AssertionError: 
> Expected: Expected checksum is (cd5b52939257e12428a9fa085c32a84dd209b180)
>  but: Invalid BigQuery response: 
> {"jobComplete":false,"jobReference":{"jobId":"job_0STNX_OD83tQOzo6MvmqXCrk61U","projectId":"apache-beam-testing"},"kind":"bigquery#queryResponse"}
>   at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:20)
>   at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:8)
>   at 
> org.apache.beam.runners.dataflow.testing.TestDataflowRunner.run(TestDataflowRunner.java:164)
>   at 
> org.apache.beam.runners.dataflow.testing.TestDataflowRunner.run(TestDataflowRunner.java:93)
>   at 
> org.apache.beam.runners.dataflow.testing.TestDataflowRunner.run(TestDataflowRunner.java:61)
>   at org.apache.beam.sdk.Pipeline.run(Pipeline.java:179)
>   at 
> org.apache.beam.examples.WindowedWordCount.main(WindowedWordCount.java:224)
>   at 
> org.apache.beam.examples.WindowedWordCountIT.testWindowedWordCountPipeline(WindowedWordCountIT.java:88)
>   at 
> org.apache.beam.examples.WindowedWordCountIT.testWindowedWordCountInBatch(WindowedWordCountIT.java:59)
>   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>   at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>   at java.lang.reflect.Method.invoke(Method.java:498)
>   at 
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47)
>   at 
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
>   at 
> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44)
>   at 
> org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
>   at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:271)
>   at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:70)
>   at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
>   at org.junit.runners.ParentRunner$3.run(ParentRunner.java:238)
>   at 
> org.apache.maven.surefire.junitcore.pc.Scheduler$1.run(Scheduler.java:393)
>   at 
> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
>   at java.util.concurrent.FutureTask.run(FutureTask.java:266)
>   at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
>   at java.lang.Thread.run(Thread.java:745)
> Results :
> Failed tests: 
>   
> WindowedWordCountIT.testWindowedWordCountInBatch:59->testWindowedWordCountPipeline:88
>  
> Expected: Expected checksum is (cd5b52939257e12428a9fa085c32a84dd209b180)
>  but: Invalid BigQuery response: 
> {"jobComplete":false,"jobReference":{"jobId":"job_0STNX_OD83tQOzo6MvmqXCrk61U","projectId":"apache-beam-testing"},"kind":"bigquery#queryResponse"}
> Tests run: 3, Failures: 1, Errors: 0, Skipped: 0



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Resolved] (BEAM-1583) Separate GCP test required packages from general GCP dependencies

2018-04-04 Thread Mark Liu (JIRA)

 [ 
https://issues.apache.org/jira/browse/BEAM-1583?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mark Liu resolved BEAM-1583.

   Resolution: Done
Fix Version/s: Not applicable

> Separate GCP test required packages from general GCP dependencies
> -
>
> Key: BEAM-1583
> URL: https://issues.apache.org/jira/browse/BEAM-1583
> Project: Beam
>  Issue Type: Task
>  Components: sdk-py-core
>Reporter: Mark Liu
>Assignee: Mark Liu
>Priority: Major
> Fix For: Not applicable
>
>
> This issue comes from discussion under:
> https://github.com/apache/beam/pull/2064#discussion_r103755653
> If more GCP dependencies introduced for test only purpose, thinking to move 
> them to a separate group.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (BEAM-1583) Separate GCP test required packages from general GCP dependencies

2018-04-04 Thread Mark Liu (JIRA)

[ 
https://issues.apache.org/jira/browse/BEAM-1583?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16426032#comment-16426032
 ] 

Mark Liu commented on BEAM-1583:


Currently, there is a separate dependency group GCP_REQUIREMENTS in setup.py. 
People can choose to install GCP related dependencies separately.

I think it's a good situation for this JIRA. Will close it and feel free to 
reopen if we need improvement.

> Separate GCP test required packages from general GCP dependencies
> -
>
> Key: BEAM-1583
> URL: https://issues.apache.org/jira/browse/BEAM-1583
> Project: Beam
>  Issue Type: Task
>  Components: sdk-py-core
>Reporter: Mark Liu
>Assignee: Mark Liu
>Priority: Major
> Fix For: Not applicable
>
>
> This issue comes from discussion under:
> https://github.com/apache/beam/pull/2064#discussion_r103755653
> If more GCP dependencies introduced for test only purpose, thinking to move 
> them to a separate group.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Closed] (BEAM-1583) Separate GCP test required packages from general GCP dependencies

2018-04-04 Thread Mark Liu (JIRA)

 [ 
https://issues.apache.org/jira/browse/BEAM-1583?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mark Liu closed BEAM-1583.
--

> Separate GCP test required packages from general GCP dependencies
> -
>
> Key: BEAM-1583
> URL: https://issues.apache.org/jira/browse/BEAM-1583
> Project: Beam
>  Issue Type: Task
>  Components: sdk-py-core
>Reporter: Mark Liu
>Assignee: Mark Liu
>Priority: Major
> Fix For: Not applicable
>
>
> This issue comes from discussion under:
> https://github.com/apache/beam/pull/2064#discussion_r103755653
> If more GCP dependencies introduced for test only purpose, thinking to move 
> them to a separate group.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (BEAM-3946) Python SDK tests are failing if no GOOGLE_APPLICATION_CREDENTIALS was set

2018-04-04 Thread Mark Liu (JIRA)

[ 
https://issues.apache.org/jira/browse/BEAM-3946?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16425826#comment-16425826
 ] 

Mark Liu commented on BEAM-3946:


Sorry for late response. 

I think the cause is that pubsub client is initialized in 
pubsub_matcher_test.py, which requires GOOGLE_APPLICATION_CREDENTIALS. The 
right way to do is fully mock pubsub in unit test and I can have a quick fix 
for it.

> Python SDK tests are failing if no GOOGLE_APPLICATION_CREDENTIALS was set
> -
>
> Key: BEAM-3946
> URL: https://issues.apache.org/jira/browse/BEAM-3946
> Project: Beam
>  Issue Type: Bug
>  Components: examples-python
>Reporter: Alexey Romanenko
>Assignee: Mark Liu
>Priority: Major
>
> Running locally mvn clean install fails on following Apache Beam :: SDKs :: 
> Python tests:
> {{ERROR: test_message_matcher_mismatch 
> (apache_beam.io.gcp.tests.pubsub_matcher_test.PubSubMatcherTest)}}
>  {{ERROR: test_message_matcher_success 
> (apache_beam.io.gcp.tests.pubsub_matcher_test.PubSubMatcherTest)}}
>  {{ERROR: test_message_metcher_timeout 
> (apache_beam.io.gcp.tests.pubsub_matcher_test.PubSubMatcherTest)}}
>  
> with an error:
> DefaultCredentialsError: Could not automatically determine credentials. 
> Please set GOOGLE_APPLICATION_CREDENTIALS or
>  explicitly create credential and re-run the application. For more
>  information, please see
>  
> [https://developers.google.com/accounts/docs/application-default-credentials].
>   >> begin captured logging << 
>  google.auth.transport._http_client: DEBUG: Making request: GET 
> [http://169.254.169.254|http://169.254.169.254/]
>  google.auth.compute_engine._metadata: INFO: Compute Engine Metadata server 
> unavailable.
>  - >> end captured logging << -
>  
> It looks like it's a regression and it was caused by this commit: 
> [301853647f2c726c04c5bdb02cab6ff6b39f09d0|https://github.com/apache/beam/commit/301853647f2c726c04c5bdb02cab6ff6b39f09d0]



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Assigned] (BEAM-3922) beam_PostCommit_Python_Verify is broken

2018-03-23 Thread Mark Liu (JIRA)

 [ 
https://issues.apache.org/jira/browse/BEAM-3922?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mark Liu reassigned BEAM-3922:
--

Assignee: Robert Bradshaw  (was: Jason Kuster)

> beam_PostCommit_Python_Verify is broken
> ---
>
> Key: BEAM-3922
> URL: https://issues.apache.org/jira/browse/BEAM-3922
> Project: Beam
>  Issue Type: Bug
>  Components: testing
>Reporter: Mark Liu
>Assignee: Robert Bradshaw
>Priority: Major
>
> Jenkins 
>  
> [beam_PostCommit_Python_Verify|https://builds.apache.org/job/beam_PostCommit_Python_Verify/]
>  is broken since Mar 21. 
> From the [console 
> log|https://builds.apache.org/job/beam_PostCommit_Python_Verify/4490/consoleFull]:
> {code}
> ==
> ERROR: test_wordcount_it (apache_beam.examples.wordcount_it_test.WordCountIT)
> --
> Traceback (most recent call last):
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python_Verify/src/sdks/python/.eggs/nose-1.3.7-py2.7.egg/nose/plugins/multiprocess.py",
>  line 812, in run
> test(orig)
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python_Verify/src/sdks/python/.eggs/nose-1.3.7-py2.7.egg/nose/case.py",
>  line 45, in __call__
> return self.run(*arg, **kwarg)
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python_Verify/src/sdks/python/.eggs/nose-1.3.7-py2.7.egg/nose/case.py",
>  line 133, in run
> self.runTest(result)
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python_Verify/src/sdks/python/.eggs/nose-1.3.7-py2.7.egg/nose/case.py",
>  line 151, in runTest
> test(result)
>   File "/usr/lib/python2.7/unittest/case.py", line 395, in __call__
> return self.run(*args, **kwds)
>   File "/usr/lib/python2.7/unittest/case.py", line 331, in run
> testMethod()
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python_Verify/src/sdks/python/apache_beam/examples/wordcount_it_test.py",
>  line 66, in test_wordcount_it
> wordcount.run(test_pipeline.get_full_options_as_args(**extra_opts))
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python_Verify/src/sdks/python/apache_beam/examples/wordcount.py",
>  line 115, in run
> result = p.run()
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python_Verify/src/sdks/python/apache_beam/pipeline.py",
>  line 389, in run
> self.to_runner_api(), self.runner, self._options).run(False)
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python_Verify/src/sdks/python/apache_beam/pipeline.py",
>  line 402, in run
> return self.runner.run_pipeline(self)
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python_Verify/src/sdks/python/apache_beam/runners/dataflow/test_dataflow_runner.py",
>  line 57, in run_pipeline
> self.result.wait_until_finish()
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python_Verify/src/sdks/python/apache_beam/runners/dataflow/dataflow_runner.py",
>  line 1071, in wait_until_finish
> time.sleep(5.0)
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python_Verify/src/sdks/python/.eggs/nose-1.3.7-py2.7.egg/nose/plugins/multiprocess.py",
>  line 276, in signalhandler
> raise TimedOutException()
> TimedOutException: 'test_wordcount_it 
> (apache_beam.examples.wordcount_it_test.WordCountIT)'
> {code}
> Looks like wordcount pipeline didn't finish after 900s (set from command 
> --process-timeout=900) and test failed in timeout. Generally this test should 
> finish in 10min, so probably something wrong in the pipeline.
> One failure pipeline link (found in console log):
> https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-03-23_08_28_06-8460792149394878073?project=apache-beam-testing



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Assigned] (BEAM-3922) beam_PostCommit_Python_Verify is broken

2018-03-23 Thread Mark Liu (JIRA)

 [ 
https://issues.apache.org/jira/browse/BEAM-3922?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mark Liu reassigned BEAM-3922:
--

Assignee: Jason Kuster  (was: Anton Kedin)

> beam_PostCommit_Python_Verify is broken
> ---
>
> Key: BEAM-3922
> URL: https://issues.apache.org/jira/browse/BEAM-3922
> Project: Beam
>  Issue Type: Bug
>  Components: testing
>Reporter: Mark Liu
>Assignee: Jason Kuster
>Priority: Major
>
> Jenkins 
>  
> [beam_PostCommit_Python_Verify|https://builds.apache.org/job/beam_PostCommit_Python_Verify/]
>  is broken since Mar 21. 
> From the [console 
> log|https://builds.apache.org/job/beam_PostCommit_Python_Verify/4490/consoleFull]:
> {code}
> ==
> ERROR: test_wordcount_it (apache_beam.examples.wordcount_it_test.WordCountIT)
> --
> Traceback (most recent call last):
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python_Verify/src/sdks/python/.eggs/nose-1.3.7-py2.7.egg/nose/plugins/multiprocess.py",
>  line 812, in run
> test(orig)
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python_Verify/src/sdks/python/.eggs/nose-1.3.7-py2.7.egg/nose/case.py",
>  line 45, in __call__
> return self.run(*arg, **kwarg)
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python_Verify/src/sdks/python/.eggs/nose-1.3.7-py2.7.egg/nose/case.py",
>  line 133, in run
> self.runTest(result)
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python_Verify/src/sdks/python/.eggs/nose-1.3.7-py2.7.egg/nose/case.py",
>  line 151, in runTest
> test(result)
>   File "/usr/lib/python2.7/unittest/case.py", line 395, in __call__
> return self.run(*args, **kwds)
>   File "/usr/lib/python2.7/unittest/case.py", line 331, in run
> testMethod()
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python_Verify/src/sdks/python/apache_beam/examples/wordcount_it_test.py",
>  line 66, in test_wordcount_it
> wordcount.run(test_pipeline.get_full_options_as_args(**extra_opts))
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python_Verify/src/sdks/python/apache_beam/examples/wordcount.py",
>  line 115, in run
> result = p.run()
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python_Verify/src/sdks/python/apache_beam/pipeline.py",
>  line 389, in run
> self.to_runner_api(), self.runner, self._options).run(False)
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python_Verify/src/sdks/python/apache_beam/pipeline.py",
>  line 402, in run
> return self.runner.run_pipeline(self)
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python_Verify/src/sdks/python/apache_beam/runners/dataflow/test_dataflow_runner.py",
>  line 57, in run_pipeline
> self.result.wait_until_finish()
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python_Verify/src/sdks/python/apache_beam/runners/dataflow/dataflow_runner.py",
>  line 1071, in wait_until_finish
> time.sleep(5.0)
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python_Verify/src/sdks/python/.eggs/nose-1.3.7-py2.7.egg/nose/plugins/multiprocess.py",
>  line 276, in signalhandler
> raise TimedOutException()
> TimedOutException: 'test_wordcount_it 
> (apache_beam.examples.wordcount_it_test.WordCountIT)'
> {code}
> Looks like wordcount pipeline didn't finish after 900s (set from command 
> --process-timeout=900) and test failed in timeout. Generally this test should 
> finish in 10min, so probably something wrong in the pipeline.
> One failure pipeline link (found in console log):
> https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-03-23_08_28_06-8460792149394878073?project=apache-beam-testing



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (BEAM-3922) beam_PostCommit_Python_Verify is broken

2018-03-23 Thread Mark Liu (JIRA)
Mark Liu created BEAM-3922:
--

 Summary: beam_PostCommit_Python_Verify is broken
 Key: BEAM-3922
 URL: https://issues.apache.org/jira/browse/BEAM-3922
 Project: Beam
  Issue Type: Bug
  Components: testing
Reporter: Mark Liu
Assignee: Anton Kedin


Jenkins 
 
[beam_PostCommit_Python_Verify|https://builds.apache.org/job/beam_PostCommit_Python_Verify/]
 is broken since Mar 21. 

>From the [console 
>log|https://builds.apache.org/job/beam_PostCommit_Python_Verify/4490/consoleFull]:

{code}
==
ERROR: test_wordcount_it (apache_beam.examples.wordcount_it_test.WordCountIT)
--
Traceback (most recent call last):
  File 
"/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python_Verify/src/sdks/python/.eggs/nose-1.3.7-py2.7.egg/nose/plugins/multiprocess.py",
 line 812, in run
test(orig)
  File 
"/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python_Verify/src/sdks/python/.eggs/nose-1.3.7-py2.7.egg/nose/case.py",
 line 45, in __call__
return self.run(*arg, **kwarg)
  File 
"/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python_Verify/src/sdks/python/.eggs/nose-1.3.7-py2.7.egg/nose/case.py",
 line 133, in run
self.runTest(result)
  File 
"/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python_Verify/src/sdks/python/.eggs/nose-1.3.7-py2.7.egg/nose/case.py",
 line 151, in runTest
test(result)
  File "/usr/lib/python2.7/unittest/case.py", line 395, in __call__
return self.run(*args, **kwds)
  File "/usr/lib/python2.7/unittest/case.py", line 331, in run
testMethod()
  File 
"/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python_Verify/src/sdks/python/apache_beam/examples/wordcount_it_test.py",
 line 66, in test_wordcount_it
wordcount.run(test_pipeline.get_full_options_as_args(**extra_opts))
  File 
"/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python_Verify/src/sdks/python/apache_beam/examples/wordcount.py",
 line 115, in run
result = p.run()
  File 
"/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python_Verify/src/sdks/python/apache_beam/pipeline.py",
 line 389, in run
self.to_runner_api(), self.runner, self._options).run(False)
  File 
"/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python_Verify/src/sdks/python/apache_beam/pipeline.py",
 line 402, in run
return self.runner.run_pipeline(self)
  File 
"/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python_Verify/src/sdks/python/apache_beam/runners/dataflow/test_dataflow_runner.py",
 line 57, in run_pipeline
self.result.wait_until_finish()
  File 
"/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python_Verify/src/sdks/python/apache_beam/runners/dataflow/dataflow_runner.py",
 line 1071, in wait_until_finish
time.sleep(5.0)
  File 
"/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python_Verify/src/sdks/python/.eggs/nose-1.3.7-py2.7.egg/nose/plugins/multiprocess.py",
 line 276, in signalhandler
raise TimedOutException()
TimedOutException: 'test_wordcount_it 
(apache_beam.examples.wordcount_it_test.WordCountIT)'
{code}

Looks like wordcount pipeline didn't finish after 900s (set from command 
--process-timeout=900) and test failed in timeout. Generally this test should 
finish in 10min, so probably something wrong in the pipeline.

One failure pipeline link (found in console log):
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-03-23_08_28_06-8460792149394878073?project=apache-beam-testing



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (BEAM-3861) Build test infra for end-to-end streaming test in Python SDK

2018-03-15 Thread Mark Liu (JIRA)
Mark Liu created BEAM-3861:
--

 Summary: Build test infra for end-to-end streaming test in Python 
SDK
 Key: BEAM-3861
 URL: https://issues.apache.org/jira/browse/BEAM-3861
 Project: Beam
  Issue Type: Task
  Components: testing
Reporter: Mark Liu
Assignee: Mark Liu






--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Resolved] (BEAM-1584) Support clean-up step in integration test

2018-03-15 Thread Mark Liu (JIRA)

 [ 
https://issues.apache.org/jira/browse/BEAM-1584?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mark Liu resolved BEAM-1584.

   Resolution: Fixed
Fix Version/s: Not applicable

> Support clean-up step in integration test
> -
>
> Key: BEAM-1584
> URL: https://issues.apache.org/jira/browse/BEAM-1584
> Project: Beam
>  Issue Type: Task
>  Components: testing
>Reporter: Mark Liu
>Assignee: Mark Liu
>Priority: Major
> Fix For: Not applicable
>
>
> Idea comes from: 
> https://github.com/apache/beam/pull/2064/files/628fafed098ac5550356a201c6ccdcdcc2e9604e
> Integration tests in all sdks should be able to do clean-up at the end of 
> each run.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Assigned] (BEAM-2339) Jenkins cross JDK version test on Windows

2018-03-15 Thread Mark Liu (JIRA)

 [ 
https://issues.apache.org/jira/browse/BEAM-2339?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mark Liu reassigned BEAM-2339:
--

Assignee: Davor Bonaci  (was: Mark Liu)

> Jenkins cross JDK version test on Windows
> -
>
> Key: BEAM-2339
> URL: https://issues.apache.org/jira/browse/BEAM-2339
> Project: Beam
>  Issue Type: Task
>  Components: build-system, testing
>Reporter: Mark Liu
>Assignee: Davor Bonaci
>Priority: Major
>
> We can set os variant to choose windows for Jenkins test, which can be 
> combined with JDK version test. So that we can have cross OS / cross JDK 
> version test. 
> This discussion came from 
> https://github.com/apache/beam/pull/3184#pullrequestreview-39303400



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Closed] (BEAM-3841) Python TestDataflowRunner should oeverride run_pipeline

2018-03-15 Thread Mark Liu (JIRA)

 [ 
https://issues.apache.org/jira/browse/BEAM-3841?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mark Liu closed BEAM-3841.
--

> Python TestDataflowRunner should oeverride run_pipeline
> ---
>
> Key: BEAM-3841
> URL: https://issues.apache.org/jira/browse/BEAM-3841
> Project: Beam
>  Issue Type: Bug
>  Components: testing
>Reporter: Mark Liu
>Assignee: Mark Liu
>Priority: Major
> Fix For: Not applicable
>
>  Time Spent: 0.5h
>  Remaining Estimate: 0h
>
> [TestDataflowRunner|https://github.com/apache/beam/blob/master/sdks/python/apache_beam/runners/dataflow/test_dataflow_runner.py]
>  is inherited from 
> [DataflowRunner|https://github.com/apache/beam/blob/master/sdks/python/apache_beam/runners/dataflow/dataflow_runner.py].
>  Basically, it wraps DataflowRunner.run_pipeline and provide more test 
> actions. 
> However DataflowRunner.run renamed to run_pipeline in [this 
> commit|https://github.com/apache/beam/commit/8cf222d3db1188aff5432af548961fc670f97635],
>  but run function in TestDataflowRunner didn't change.
> We should change it accordingly.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Resolved] (BEAM-3841) Python TestDataflowRunner should oeverride run_pipeline

2018-03-15 Thread Mark Liu (JIRA)

 [ 
https://issues.apache.org/jira/browse/BEAM-3841?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mark Liu resolved BEAM-3841.

   Resolution: Fixed
Fix Version/s: Not applicable

> Python TestDataflowRunner should oeverride run_pipeline
> ---
>
> Key: BEAM-3841
> URL: https://issues.apache.org/jira/browse/BEAM-3841
> Project: Beam
>  Issue Type: Bug
>  Components: testing
>Reporter: Mark Liu
>Assignee: Mark Liu
>Priority: Major
> Fix For: Not applicable
>
>  Time Spent: 0.5h
>  Remaining Estimate: 0h
>
> [TestDataflowRunner|https://github.com/apache/beam/blob/master/sdks/python/apache_beam/runners/dataflow/test_dataflow_runner.py]
>  is inherited from 
> [DataflowRunner|https://github.com/apache/beam/blob/master/sdks/python/apache_beam/runners/dataflow/dataflow_runner.py].
>  Basically, it wraps DataflowRunner.run_pipeline and provide more test 
> actions. 
> However DataflowRunner.run renamed to run_pipeline in [this 
> commit|https://github.com/apache/beam/commit/8cf222d3db1188aff5432af548961fc670f97635],
>  but run function in TestDataflowRunner didn't change.
> We should change it accordingly.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (BEAM-3841) Python TestDataflowRunner should oeverride run_pipeline

2018-03-15 Thread Mark Liu (JIRA)

[ 
https://issues.apache.org/jira/browse/BEAM-3841?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16401182#comment-16401182
 ] 

Mark Liu commented on BEAM-3841:


Fix PR [4856|https://github.com/apache/beam/pull/4856] is merged. 

> Python TestDataflowRunner should oeverride run_pipeline
> ---
>
> Key: BEAM-3841
> URL: https://issues.apache.org/jira/browse/BEAM-3841
> Project: Beam
>  Issue Type: Bug
>  Components: testing
>Reporter: Mark Liu
>Assignee: Mark Liu
>Priority: Major
>  Time Spent: 0.5h
>  Remaining Estimate: 0h
>
> [TestDataflowRunner|https://github.com/apache/beam/blob/master/sdks/python/apache_beam/runners/dataflow/test_dataflow_runner.py]
>  is inherited from 
> [DataflowRunner|https://github.com/apache/beam/blob/master/sdks/python/apache_beam/runners/dataflow/dataflow_runner.py].
>  Basically, it wraps DataflowRunner.run_pipeline and provide more test 
> actions. 
> However DataflowRunner.run renamed to run_pipeline in [this 
> commit|https://github.com/apache/beam/commit/8cf222d3db1188aff5432af548961fc670f97635],
>  but run function in TestDataflowRunner didn't change.
> We should change it accordingly.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (BEAM-3841) Python TestDataflowRunner should oeverride run_pipeline

2018-03-12 Thread Mark Liu (JIRA)
Mark Liu created BEAM-3841:
--

 Summary: Python TestDataflowRunner should oeverride run_pipeline
 Key: BEAM-3841
 URL: https://issues.apache.org/jira/browse/BEAM-3841
 Project: Beam
  Issue Type: Test
  Components: testing
Reporter: Mark Liu
Assignee: Mark Liu


[TestDataflowRunner|https://github.com/apache/beam/blob/master/sdks/python/apache_beam/runners/dataflow/test_dataflow_runner.py]
 is inherited from 
[DataflowRunner|https://github.com/apache/beam/blob/master/sdks/python/apache_beam/runners/dataflow/dataflow_runner.py].
 Basically, it wraps DataflowRunner.run_pipeline and provide more test actions. 

However DataflowRunner.run renamed to run_pipeline in [this 
commit|https://github.com/apache/beam/commit/8cf222d3db1188aff5432af548961fc670f97635],
 but run function in TestDataflowRunner didn't change.

We should change it accordingly.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (BEAM-3841) Python TestDataflowRunner should oeverride run_pipeline

2018-03-12 Thread Mark Liu (JIRA)

 [ 
https://issues.apache.org/jira/browse/BEAM-3841?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mark Liu updated BEAM-3841:
---
Issue Type: Bug  (was: Test)

> Python TestDataflowRunner should oeverride run_pipeline
> ---
>
> Key: BEAM-3841
> URL: https://issues.apache.org/jira/browse/BEAM-3841
> Project: Beam
>  Issue Type: Bug
>  Components: testing
>Reporter: Mark Liu
>Assignee: Mark Liu
>Priority: Major
>
> [TestDataflowRunner|https://github.com/apache/beam/blob/master/sdks/python/apache_beam/runners/dataflow/test_dataflow_runner.py]
>  is inherited from 
> [DataflowRunner|https://github.com/apache/beam/blob/master/sdks/python/apache_beam/runners/dataflow/dataflow_runner.py].
>  Basically, it wraps DataflowRunner.run_pipeline and provide more test 
> actions. 
> However DataflowRunner.run renamed to run_pipeline in [this 
> commit|https://github.com/apache/beam/commit/8cf222d3db1188aff5432af548961fc670f97635],
>  but run function in TestDataflowRunner didn't change.
> We should change it accordingly.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (BEAM-3810) Add new matcher to check output file content for WordCountIT and WindowedWordCountIT

2018-03-08 Thread Mark Liu (JIRA)
Mark Liu created BEAM-3810:
--

 Summary: Add new matcher to check output file content for 
WordCountIT and WindowedWordCountIT
 Key: BEAM-3810
 URL: https://issues.apache.org/jira/browse/BEAM-3810
 Project: Beam
  Issue Type: Test
  Components: testing
Reporter: Mark Liu
Assignee: Mark Liu


Test use case like 
[https://github.com/apache/beam/pull/4755#discussion_r173254182] shows that 
verify lines of output for integration test makes the validation and debugging 
easier than simple verify checksum. We'd like to add a new matcher (like 
FileContentMatcher) to very lines given expected files.

This matcher can be applied to ITs that already use FileChecksumMatcher (etc. 
WordCountIT and WindowedWordCountIT).



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Comment Edited] (BEAM-3809) beam_PerformanceTests_Python failing

2018-03-08 Thread Mark Liu (JIRA)

[ 
https://issues.apache.org/jira/browse/BEAM-3809?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16391749#comment-16391749
 ] 

Mark Liu edited comment on BEAM-3809 at 3/8/18 7:11 PM:


The command pkb use to execute pipeline is:
{code}
python setup.py nosetests --tests=apache_beam.examples.wordcount_it_test 
--attr=IT --test-pipeline-options="--project=apache-beam-testing" 
"--staging_location=gs://temp-storage-for-end-to-end-tests/staging-it" 
"--temp_location=gs://temp-storage-for-end-to-end-tests/temp-it" 
"--output=gs://temp-storage-for-end-to-end-tests/py-it-cloud/output" 
--runner=TestDataflowRunner 
--sdk_location=/home/jenkins/jenkins-slave/workspace/beam_PerformanceTests_Python/runs/9e009d45/beam/sdks/python/target/apache-beam-2.5.0.dev0.tar.gz
{code}
which looks good to me. Probably need more investigate on pkb side.

 


was (Author: markflyhigh):
The command pkb use to execute pipeline is:
python setup.py nosetests --tests=apache_beam.examples.wordcount_it_test 
--attr=IT --test-pipeline-options="--project=apache-beam-testing" 
"--staging_location=gs://temp-storage-for-end-to-end-tests/staging-it" 
"--temp_location=gs://temp-storage-for-end-to-end-tests/temp-it" 
"--output=gs://temp-storage-for-end-to-end-tests/py-it-cloud/output" 
--runner=TestDataflowRunner 
--sdk_location=/home/jenkins/jenkins-slave/workspace/beam_PerformanceTests_Python/runs/9e009d45/beam/sdks/python/target/apache-beam-2.5.0.dev0.tar.gz
which looks good to me. Probably need more investigate on pkb side.

 

> beam_PerformanceTests_Python failing
> 
>
> Key: BEAM-3809
> URL: https://issues.apache.org/jira/browse/BEAM-3809
> Project: Beam
>  Issue Type: Bug
>  Components: testing
>Reporter: Ahmet Altay
>Assignee: Mark Liu
>Priority: Major
>
> Looks like a misconfiguration based on the error.
>  
> Example test:
> [https://builds.apache.org/view/A-D/view/Beam/job/beam_PerformanceTests_Python/1000/console]
>  
> Error I see:
> error: option --test-pipeline-options not recognized
> 2018-03-08 18:48:08,202 9e009d45 MainThread beam_integration_benchmark(1/1) 
> ERRORError during benchmark beam_integration_benchmark
> Traceback (most recent call last):
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PerformanceTests_Python/PerfKitBenchmarker/perfkitbenchmarker/pkb.py",
>  line 622, in RunBenchmark
> DoRunPhase(spec, collector, detailed_timer)
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PerformanceTests_Python/PerfKitBenchmarker/perfkitbenchmarker/pkb.py",
>  line 525, in DoRunPhase
> samples = spec.BenchmarkRun(spec)
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PerformanceTests_Python/PerfKitBenchmarker/perfkitbenchmarker/linux_benchmarks/beam_integration_benchmark.py",
>  line 159, in Run
> job_type=job_type)
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PerformanceTests_Python/PerfKitBenchmarker/perfkitbenchmarker/providers/gcp/gcp_dpb_dataflow.py",
>  line 90, in SubmitJob
> assert retcode == 0, "Integration Test Failed."



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (BEAM-3809) beam_PerformanceTests_Python failing

2018-03-08 Thread Mark Liu (JIRA)

[ 
https://issues.apache.org/jira/browse/BEAM-3809?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16391749#comment-16391749
 ] 

Mark Liu commented on BEAM-3809:


The command pkb use to execute pipeline is:
python setup.py nosetests --tests=apache_beam.examples.wordcount_it_test 
--attr=IT --test-pipeline-options="--project=apache-beam-testing" 
"--staging_location=gs://temp-storage-for-end-to-end-tests/staging-it" 
"--temp_location=gs://temp-storage-for-end-to-end-tests/temp-it" 
"--output=gs://temp-storage-for-end-to-end-tests/py-it-cloud/output" 
--runner=TestDataflowRunner 
--sdk_location=/home/jenkins/jenkins-slave/workspace/beam_PerformanceTests_Python/runs/9e009d45/beam/sdks/python/target/apache-beam-2.5.0.dev0.tar.gz
which looks good to me. Probably need more investigate on pkb side.

 

> beam_PerformanceTests_Python failing
> 
>
> Key: BEAM-3809
> URL: https://issues.apache.org/jira/browse/BEAM-3809
> Project: Beam
>  Issue Type: Bug
>  Components: testing
>Reporter: Ahmet Altay
>Assignee: Mark Liu
>Priority: Major
>
> Looks like a misconfiguration based on the error.
>  
> Example test:
> [https://builds.apache.org/view/A-D/view/Beam/job/beam_PerformanceTests_Python/1000/console]
>  
> Error I see:
> error: option --test-pipeline-options not recognized
> 2018-03-08 18:48:08,202 9e009d45 MainThread beam_integration_benchmark(1/1) 
> ERRORError during benchmark beam_integration_benchmark
> Traceback (most recent call last):
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PerformanceTests_Python/PerfKitBenchmarker/perfkitbenchmarker/pkb.py",
>  line 622, in RunBenchmark
> DoRunPhase(spec, collector, detailed_timer)
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PerformanceTests_Python/PerfKitBenchmarker/perfkitbenchmarker/pkb.py",
>  line 525, in DoRunPhase
> samples = spec.BenchmarkRun(spec)
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PerformanceTests_Python/PerfKitBenchmarker/perfkitbenchmarker/linux_benchmarks/beam_integration_benchmark.py",
>  line 159, in Run
> job_type=job_type)
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PerformanceTests_Python/PerfKitBenchmarker/perfkitbenchmarker/providers/gcp/gcp_dpb_dataflow.py",
>  line 90, in SubmitJob
> assert retcode == 0, "Integration Test Failed."



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (BEAM-3750) Integration test will fail when use TestPipeline.convertToArgs in some special cases

2018-02-26 Thread Mark Liu (JIRA)
Mark Liu created BEAM-3750:
--

 Summary: Integration test will fail when use 
TestPipeline.convertToArgs in some special cases
 Key: BEAM-3750
 URL: https://issues.apache.org/jira/browse/BEAM-3750
 Project: Beam
  Issue Type: Bug
  Components: testing
Reporter: Mark Liu
Assignee: Mark Liu


ITs(Integration Tests) are using {{TestPipeline.convertToArgs}} to convert 
well-defined options to list of pipeline-recognizable arguments and pass them 
to pipeline main function. A example use case is 
{{WordCountIT}}([https://github.com/apache/beam/blob/master/examples/java/src/test/java/org/apache/beam/examples/WordCountIT.java#L70]).

However, incorrect format will be generated by this function if the option that 
we want to convert has attribute type {{List}}.

Ideally, we want to get rid of this complicated conversion step in testing 
which can avoid more problematic cases, but also make integration test better 
understandable.

Some discussions happened in [https://github.com/apache/beam/pull/4346.] One 
possible approach is to extract part of pipeline code so that integration test 
can reference directly. 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Commented] (BEAM-3364) Use multiprocess nose plugin to improve unit test run time

2017-12-15 Thread Mark Liu (JIRA)

[ 
https://issues.apache.org/jira/browse/BEAM-3364?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16293413#comment-16293413
 ] 

Mark Liu commented on BEAM-3364:


Integration tests in PostCommit are enabled with miltiprocess. AFAIK, the 
machine that host jenkins instance has 4 cores, so I think set processes=4 
should be fine.

> Use multiprocess nose plugin to improve unit test run time
> --
>
> Key: BEAM-3364
> URL: https://issues.apache.org/jira/browse/BEAM-3364
> Project: Beam
>  Issue Type: Improvement
>  Components: sdk-py-core, testing
>Reporter: Udi Meiri
>Priority: Minor
>
> The multiprocess plugin is installed by default. You can try it out by 
> running:
> nosetests --processes 8 --process-timeout 60
> The running time is reduced from ~3m30s to ~30s on my workstation, but with 
> some tests reporting errors.
> Errors can be avoided by disabling multiprocess for certain classes by 
> setting this global in the test file:
> {code}
> _multiprocess_ = False  # Disable nosetests multiprocess plugin.
> {code}
> The problem is that failures are flaky, so repeated runs are required to 
> verify we're not adding any new flakes by enabling multiprocess. 
> Alternatively, we could try fixing the tests to make them pass.
> Setting up multiprocess runs is done by adding these lines to the bottom of 
> setup.cfg:
> {code}
> # Run tests in multiple processes for shorter overall running time.
> processes=8
> # This timeout only applies to multiprocess runs. The default is 10 seconds.
> process-timeout=60
> {code}



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (BEAM-3068) Document Jenkins ghprb commands

2017-10-17 Thread Mark Liu (JIRA)

 [ 
https://issues.apache.org/jira/browse/BEAM-3068?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mark Liu updated BEAM-3068:
---
Component/s: (was: build-system)
 testing

> Document Jenkins ghprb commands
> ---
>
> Key: BEAM-3068
> URL: https://issues.apache.org/jira/browse/BEAM-3068
> Project: Beam
>  Issue Type: Task
>  Components: testing
>Reporter: Mark Liu
>
> Summarize current ghprb (github pull request builder plugin) commands for 
> people to easily find and use instead of to check each groovy file.
> commands includes:
> "retest this please",
> command to run specific Jenkins build (defined under 
> .test-infra/jenkins/job_beam_*.groovy).



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Created] (BEAM-3068) Document Jenkins ghprb commands

2017-10-17 Thread Mark Liu (JIRA)
Mark Liu created BEAM-3068:
--

 Summary: Document Jenkins ghprb commands
 Key: BEAM-3068
 URL: https://issues.apache.org/jira/browse/BEAM-3068
 Project: Beam
  Issue Type: Task
  Components: build-system
Reporter: Mark Liu


Summarize current ghprb (github pull request builder plugin) commands for 
people to easily find and use instead of to check each groovy file.

commands includes:
"retest this please",
command to run specific Jenkins build (defined under 
.test-infra/jenkins/job_beam_*.groovy).



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (BEAM-3059) Timeout option for individual integration tests

2017-10-16 Thread Mark Liu (JIRA)

[ 
https://issues.apache.org/jira/browse/BEAM-3059?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16206677#comment-16206677
 ] 

Mark Liu commented on BEAM-3059:


Some explanation: we want to have a timeout option for each individual 
integration test, so that test will fail if it exceeds the time limit. 

This will have the following benefits:
1. build fail fast if one test stall or running too long. 
2. improve debugging information in the situation when Jenkins build stall and 
timeout.

> Timeout option for individual integration tests
> ---
>
> Key: BEAM-3059
> URL: https://issues.apache.org/jira/browse/BEAM-3059
> Project: Beam
>  Issue Type: Wish
>  Components: sdk-py-core
>Reporter: Ahmet Altay
>Priority: Minor
>  Labels: starter
>
> cc: [~markflyhigh]



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Resolved] (BEAM-2999) Split validatesrunner tests from Python postcommit

2017-10-05 Thread Mark Liu (JIRA)

 [ 
https://issues.apache.org/jira/browse/BEAM-2999?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mark Liu resolved BEAM-2999.

   Resolution: Done
Fix Version/s: Not applicable

> Split validatesrunner tests from Python postcommit
> --
>
> Key: BEAM-2999
> URL: https://issues.apache.org/jira/browse/BEAM-2999
> Project: Beam
>  Issue Type: Task
>  Components: testing
>Reporter: Mark Liu
>Assignee: Mark Liu
> Fix For: Not applicable
>
>
> The only Python Postcommit Jenkins build includes too many tests which makes 
> the build (and test) time over 1 hour. Also it became hard to found error in 
> long console logs if build failed.
> We can separate validatesrunner tests which currently take ~20mins out from 
> the Postcommit build to a separate Jenkins branch. This will shorten the 
> total build time of Postcommit.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (BEAM-2999) Split validatesrunner tests from Python postcommit

2017-10-05 Thread Mark Liu (JIRA)

 [ 
https://issues.apache.org/jira/browse/BEAM-2999?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mark Liu updated BEAM-2999:
---
Issue Type: Task  (was: Bug)

> Split validatesrunner tests from Python postcommit
> --
>
> Key: BEAM-2999
> URL: https://issues.apache.org/jira/browse/BEAM-2999
> Project: Beam
>  Issue Type: Task
>  Components: testing
>Reporter: Mark Liu
>Assignee: Mark Liu
> Fix For: Not applicable
>
>
> The only Python Postcommit Jenkins build includes too many tests which makes 
> the build (and test) time over 1 hour. Also it became hard to found error in 
> long console logs if build failed.
> We can separate validatesrunner tests which currently take ~20mins out from 
> the Postcommit build to a separate Jenkins branch. This will shorten the 
> total build time of Postcommit.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Closed] (BEAM-2999) Split validatesrunner tests from Python postcommit

2017-10-05 Thread Mark Liu (JIRA)

 [ 
https://issues.apache.org/jira/browse/BEAM-2999?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mark Liu closed BEAM-2999.
--

> Split validatesrunner tests from Python postcommit
> --
>
> Key: BEAM-2999
> URL: https://issues.apache.org/jira/browse/BEAM-2999
> Project: Beam
>  Issue Type: Task
>  Components: testing
>Reporter: Mark Liu
>Assignee: Mark Liu
> Fix For: Not applicable
>
>
> The only Python Postcommit Jenkins build includes too many tests which makes 
> the build (and test) time over 1 hour. Also it became hard to found error in 
> long console logs if build failed.
> We can separate validatesrunner tests which currently take ~20mins out from 
> the Postcommit build to a separate Jenkins branch. This will shorten the 
> total build time of Postcommit.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Created] (BEAM-2999) Split validatesrunner tests from Python postcommit

2017-09-27 Thread Mark Liu (JIRA)
Mark Liu created BEAM-2999:
--

 Summary: Split validatesrunner tests from Python postcommit
 Key: BEAM-2999
 URL: https://issues.apache.org/jira/browse/BEAM-2999
 Project: Beam
  Issue Type: Bug
  Components: testing
Reporter: Mark Liu
Assignee: Mark Liu


The only Python Postcommit Jenkins build includes too many tests which makes 
the build (and test) time over 1 hour. Also it became hard to found error in 
long console logs if build failed.

We can separate validatesrunner tests which currently take ~20mins out from the 
Postcommit build to a separate Jenkins branch. This will shorten the total 
build time of Postcommit.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (BEAM-2813) error: option --test-pipeline-options not recognized

2017-08-28 Thread Mark Liu (JIRA)

[ 
https://issues.apache.org/jira/browse/BEAM-2813?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16144523#comment-16144523
 ] 

Mark Liu commented on BEAM-2813:


Not sure if pinning a dependency will work since both pyasn1 and pyasn1-modules 
are not directly required by Beam. Also looks like Nose needs to check each 
installed packages and it's dependencies' version before load a customized 
plugin.

However we can use tools like 
[pipdeptree|https://pypi.python.org/pypi/pipdeptree] to show and verify 
dependency tree separately. 
[pipdeptree|https://pypi.python.org/pypi/pipdeptree] can print potential 
conflicts to console. 

> error: option --test-pipeline-options not recognized
> 
>
> Key: BEAM-2813
> URL: https://issues.apache.org/jira/browse/BEAM-2813
> Project: Beam
>  Issue Type: Bug
>  Components: sdk-py
>Reporter: Ahmet Altay
>Assignee: Mark Liu
>
> Python post commits 3004 to 3008 (all 5) failed with this error, but somehow 
> fixed in 3009. Mark do you know what might be causing this?
> https://builds.apache.org/view/A-D/view/Beam/job/beam_PostCommit_Python_Verify/3004/
> https://builds.apache.org/view/A-D/view/Beam/job/beam_PostCommit_Python_Verify/3009/
> The error is:
> # Run ValidatesRunner tests on Google Cloud Dataflow service
> echo ">>> RUNNING DATAFLOW RUNNER VALIDATESRUNNER TESTS"
> >>> RUNNING DATAFLOW RUNNER VALIDATESRUNNER TESTS
> python setup.py nosetests \
>   --attr ValidatesRunner \
>   --nocapture \
>   --processes=4 \
>   --process-timeout=900 \
>   --test-pipeline-options=" \
> --runner=TestDataflowRunner \
> --project=$PROJECT \
> --staging_location=$GCS_LOCATION/staging-validatesrunner-test \
> --temp_location=$GCS_LOCATION/temp-validatesrunner-test \
> --sdk_location=$SDK_LOCATION \
> --requirements_file=postcommit_requirements.txt \
> --num_workers=1"
> /home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python_Verify/sdks/python/local/lib/python2.7/site-packages/setuptools/dist.py:341:
>  UserWarning: Normalizing '2.2.0.dev' to '2.2.0.dev0'
>   normalized_version,
> /home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python_Verify/sdks/python/.eggs/nose-1.3.7-py2.7.egg/nose/plugins/manager.py:395:
>  RuntimeWarning: Unable to load plugin beam_test_plugin = 
> test_config:BeamTestPlugin: (pyasn1 0.3.3 
> (/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python_Verify/sdks/python/lib/python2.7/site-packages),
>  Requirement.parse('pyasn1==0.3.2'), set(['pyasn1-modules']))
>   RuntimeWarning)
> usage: setup.py [global_opts] cmd1 [cmd1_opts] [cmd2 [cmd2_opts] ...]
>or: setup.py --help [cmd1 cmd2 ...]
>or: setup.py --help-commands
>or: setup.py cmd --help
> error: option --test-pipeline-options not recognized



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Comment Edited] (BEAM-2813) error: option --test-pipeline-options not recognized

2017-08-28 Thread Mark Liu (JIRA)

[ 
https://issues.apache.org/jira/browse/BEAM-2813?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16144500#comment-16144500
 ] 

Mark Liu edited comment on BEAM-2813 at 8/28/17 10:58 PM:
--

I think this is a version mismatch in dependencies. This line may related to 
the actual error:
{code}
/nose/plugins/manager.py:395: RuntimeWarning: Unable to load plugin 
beam_test_plugin = test_config:BeamTestPlugin: (pyasn1 0.3.3 
(/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python_Verify/sdks/python/lib/python2.7/site-packages),
 Requirement.parse('pyasn1==0.3.2'), set(['pyasn1-modules']))
RuntimeWarning)
{code}

Both pyasn1 and pyasn1-modules are indirect dependencies and used in many GCP 
packages. Two days ago, pyasn1 was updated to 0.3.3. However, 
pyasn1-modules(0.0.11) only requires pyasn1 0.3.2. I guess when nose upload 
customized plugin, it also examined the existing packages and found versions 
conflict.

Yesterday, pyasn1-modules was updated to 
[0.1.1|https://pypi.python.org/pypi/pyasn1-modules] which fix this problem by 
pointing to latest pyasn1 (0.3.3). So Jenkins started passing.


was (Author: markflyhigh):
I think this is a version mismatch in dependencies. This line may related to 
the actual error:
{code}
/nose/plugins/manager.py:395: RuntimeWarning: Unable to load plugin 
beam_test_plugin = test_config:BeamTestPlugin: (pyasn1 0.3.3 
(/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python_Verify/sdks/python/lib/python2.7/site-packages),
 Requirement.parse('pyasn1==0.3.2'), set(['pyasn1-modules']))
RuntimeWarning)
{code}

Both pyasn1 and pyasn1-modules are indirect dependencies for many GCP packages. 
Two days ago, pyasn1 was updated to 0.3.3. However, pyasn1-modules(0.0.11) only 
requires pyasn1 0.3.2. I guess when nose upload customized plugin, it also 
examined the existing packages and found versions conflict.

Yesterday, pyasn1-modules was updated to 
[0.1.1|https://pypi.python.org/pypi/pyasn1-modules] which fix this problem by 
pointing to latest pyasn1 (0.3.3). So Jenkins started passing.

> error: option --test-pipeline-options not recognized
> 
>
> Key: BEAM-2813
> URL: https://issues.apache.org/jira/browse/BEAM-2813
> Project: Beam
>  Issue Type: Bug
>  Components: sdk-py
>Reporter: Ahmet Altay
>Assignee: Mark Liu
>
> Python post commits 3004 to 3008 (all 5) failed with this error, but somehow 
> fixed in 3009. Mark do you know what might be causing this?
> https://builds.apache.org/view/A-D/view/Beam/job/beam_PostCommit_Python_Verify/3004/
> https://builds.apache.org/view/A-D/view/Beam/job/beam_PostCommit_Python_Verify/3009/
> The error is:
> # Run ValidatesRunner tests on Google Cloud Dataflow service
> echo ">>> RUNNING DATAFLOW RUNNER VALIDATESRUNNER TESTS"
> >>> RUNNING DATAFLOW RUNNER VALIDATESRUNNER TESTS
> python setup.py nosetests \
>   --attr ValidatesRunner \
>   --nocapture \
>   --processes=4 \
>   --process-timeout=900 \
>   --test-pipeline-options=" \
> --runner=TestDataflowRunner \
> --project=$PROJECT \
> --staging_location=$GCS_LOCATION/staging-validatesrunner-test \
> --temp_location=$GCS_LOCATION/temp-validatesrunner-test \
> --sdk_location=$SDK_LOCATION \
> --requirements_file=postcommit_requirements.txt \
> --num_workers=1"
> /home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python_Verify/sdks/python/local/lib/python2.7/site-packages/setuptools/dist.py:341:
>  UserWarning: Normalizing '2.2.0.dev' to '2.2.0.dev0'
>   normalized_version,
> /home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python_Verify/sdks/python/.eggs/nose-1.3.7-py2.7.egg/nose/plugins/manager.py:395:
>  RuntimeWarning: Unable to load plugin beam_test_plugin = 
> test_config:BeamTestPlugin: (pyasn1 0.3.3 
> (/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python_Verify/sdks/python/lib/python2.7/site-packages),
>  Requirement.parse('pyasn1==0.3.2'), set(['pyasn1-modules']))
>   RuntimeWarning)
> usage: setup.py [global_opts] cmd1 [cmd1_opts] [cmd2 [cmd2_opts] ...]
>or: setup.py --help [cmd1 cmd2 ...]
>or: setup.py --help-commands
>or: setup.py cmd --help
> error: option --test-pipeline-options not recognized



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (BEAM-2813) error: option --test-pipeline-options not recognized

2017-08-28 Thread Mark Liu (JIRA)

[ 
https://issues.apache.org/jira/browse/BEAM-2813?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16144500#comment-16144500
 ] 

Mark Liu commented on BEAM-2813:


I think this is a version mismatch in dependencies. This line may related to 
the actual error:
{code}
/nose/plugins/manager.py:395: RuntimeWarning: Unable to load plugin 
beam_test_plugin = test_config:BeamTestPlugin: (pyasn1 0.3.3 
(/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python_Verify/sdks/python/lib/python2.7/site-packages),
 Requirement.parse('pyasn1==0.3.2'), set(['pyasn1-modules']))
RuntimeWarning)
{code}

Both pyasn1 and pyasn1-modules are indirect dependencies for many GCP packages. 
Two days ago, pyasn1 was updated to 0.3.3. However, pyasn1-modules(0.0.11) only 
requires pyasn1 0.3.2. I guess when nose upload customized plugin, it also 
examined the existing packages and found versions conflict.

Yesterday, pyasn1-modules was updated to 
[0.1.1|https://pypi.python.org/pypi/pyasn1-modules] which fix this problem by 
pointing to latest pyasn1 (0.3.3). So Jenkins started passing.

> error: option --test-pipeline-options not recognized
> 
>
> Key: BEAM-2813
> URL: https://issues.apache.org/jira/browse/BEAM-2813
> Project: Beam
>  Issue Type: Bug
>  Components: sdk-py
>Reporter: Ahmet Altay
>Assignee: Mark Liu
>
> Python post commits 3004 to 3008 (all 5) failed with this error, but somehow 
> fixed in 3009. Mark do you know what might be causing this?
> https://builds.apache.org/view/A-D/view/Beam/job/beam_PostCommit_Python_Verify/3004/
> https://builds.apache.org/view/A-D/view/Beam/job/beam_PostCommit_Python_Verify/3009/
> The error is:
> # Run ValidatesRunner tests on Google Cloud Dataflow service
> echo ">>> RUNNING DATAFLOW RUNNER VALIDATESRUNNER TESTS"
> >>> RUNNING DATAFLOW RUNNER VALIDATESRUNNER TESTS
> python setup.py nosetests \
>   --attr ValidatesRunner \
>   --nocapture \
>   --processes=4 \
>   --process-timeout=900 \
>   --test-pipeline-options=" \
> --runner=TestDataflowRunner \
> --project=$PROJECT \
> --staging_location=$GCS_LOCATION/staging-validatesrunner-test \
> --temp_location=$GCS_LOCATION/temp-validatesrunner-test \
> --sdk_location=$SDK_LOCATION \
> --requirements_file=postcommit_requirements.txt \
> --num_workers=1"
> /home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python_Verify/sdks/python/local/lib/python2.7/site-packages/setuptools/dist.py:341:
>  UserWarning: Normalizing '2.2.0.dev' to '2.2.0.dev0'
>   normalized_version,
> /home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python_Verify/sdks/python/.eggs/nose-1.3.7-py2.7.egg/nose/plugins/manager.py:395:
>  RuntimeWarning: Unable to load plugin beam_test_plugin = 
> test_config:BeamTestPlugin: (pyasn1 0.3.3 
> (/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python_Verify/sdks/python/lib/python2.7/site-packages),
>  Requirement.parse('pyasn1==0.3.2'), set(['pyasn1-modules']))
>   RuntimeWarning)
> usage: setup.py [global_opts] cmd1 [cmd1_opts] [cmd2 [cmd2_opts] ...]
>or: setup.py --help [cmd1 cmd2 ...]
>or: setup.py --help-commands
>or: setup.py cmd --help
> error: option --test-pipeline-options not recognized



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (BEAM-2755) Wordcount it test fail

2017-08-11 Thread Mark Liu (JIRA)

[ 
https://issues.apache.org/jira/browse/BEAM-2755?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16124110#comment-16124110
 ] 

Mark Liu commented on BEAM-2755:


This build run against my PR (https://github.com/apache/beam/pull/3688), not 
master branch. The error is fixed in the PR and that PR already merged.

We can mark this JIRA as fixed.

> Wordcount it test fail
> --
>
> Key: BEAM-2755
> URL: https://issues.apache.org/jira/browse/BEAM-2755
> Project: Beam
>  Issue Type: Bug
>  Components: sdk-py
>Reporter: Ahmet Altay
>Assignee: Mark Liu
>
> https://builds.apache.org/view/A-D/view/Beam/job/beam_PostCommit_Python_Verify/2886/console
>  failed with the error below. It might be a flake, or an error in test_utils:
> Delete operation failed with exceptions {'*': IOError(OSError(2, 'No such 
> file or directory'),), '-': IOError(OSError(2, 'No such file or 
> directory'),), '/': IOError(OSError(13, 'Permission denied'),), '1': 
> IOError(OSError(2, 'No such file or directory'),), '0': IOError(OSError(2, 
> 'No such file or directory'),), '3': IOError(OSError(2, 'No such file or 
> directory'),), '2': IOError(OSError(2, 'No such file or directory'),), '5': 
> IOError(OSError(2, 'No such file or directory'),), ':': IOError(OSError(2, 
> 'No such file or directory'),), 'a': IOError(OSError(2, 'No such file or 
> directory'),), 'c': IOError(OSError(2, 'No such file or directory'),), 'e': 
> IOError(OSError(2, 'No such file or directory'),), 'd': IOError(OSError(2, 
> 'No such file or directory'),), 'g': IOError(OSError(2, 'No such file or 
> directory'),), 'f': IOError(OSError(2, 'No such file or directory'),), 'i': 
> IOError(OSError(2, 'No such file or directory'),), 'm': IOError(OSError(2, 
> 'No such file or directory'),), 'l': IOError(OSError(2, 'No such file or 
> directory'),), 'o': IOError(OSError(2, 'No such file or directory'),), 'n': 
> IOError(OSError(2, 'No such file or directory'),), 'p': IOError(OSError(2, 
> 'No such file or directory'),), 's': IOError(OSError(2, 'No such file or 
> directory'),), 'r': IOError(OSError(2, 'No such file or directory'),), 'u': 
> IOError(OSError(2, 'No such file or directory'),), 't': IOError(OSError(2, 
> 'No such file or directory'),), 'y': IOError(OSError(2, 'No such file or 
> directory'),)}
>  Traceback for above exception (most recent call last):
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python_Verify/sdks/python/apache_beam/utils/retry.py",
>  line 175, in wrapper
> return fun(*args, **kwargs)
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python_Verify/sdks/python/apache_beam/testing/test_utils.py",
>  line 91, in delete_files
> FileSystems.delete(file_paths)
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python_Verify/sdks/python/apache_beam/io/filesystems.py",
>  line 224, in delete
> return filesystem.delete(paths)
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python_Verify/sdks/python/apache_beam/io/localfilesystem.py",
>  line 267, in delete
> raise BeamIOError("Delete operation failed", exceptions)
> Mark, could you take a look at this?



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (BEAM-2762) Coverage report for Python code

2017-08-10 Thread Mark Liu (JIRA)

 [ 
https://issues.apache.org/jira/browse/BEAM-2762?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mark Liu updated BEAM-2762:
---
Description: 
It's good to have code coverage in Python SDK to show the test coverage. Java 
is using jacoco-maven-plugin to generate coverage report and coveralls service 
to manage/display data. 

Python have similar tool called coverage.py for report generation and 
coveralls-python to send report to api of coveralls service. 

It's nice to have one place (like coveralls service) to manage and show data 
from different SDKs together/separately. However, there are still some problems 
due to the fact of Beam CI system and multi languages in Beam. The coveralls 
service doesn't have a good way to collect data separately from different 
projects. But the postcommit builds are separated by sdks. 

As the first step, I think it's good to have the python coverage report 
printout in the build to give people a idea of the current coverage status at 
lease.

  was:
It's good to have code coverage in Python SDK to show the test coverage. Java 
is using jacoco-maven-plugin to generate coverage report and coveralls service 
to manage/display data. 

Python have similar tool called coverage.py for report generation and 
coveralls-python to send report to api of coveralls service. 

It's nice to have one place (like coveralls service) to manage and show data 
from different SDKs together/separately. However, there are still some problems 
due to the fact of Beam CI system and multi languages in Beam. The coveralls 
service doesn't have a good way to collect data separately from different 
projects. But the postcommit build are running in separate sdks. 

As the first step, I think it's good to have the python coverage report 
printout in the build to give people a idea of the current coverage status at 
lease.


> Coverage report for Python code
> ---
>
> Key: BEAM-2762
> URL: https://issues.apache.org/jira/browse/BEAM-2762
> Project: Beam
>  Issue Type: Task
>  Components: sdk-py, testing
>Reporter: Mark Liu
>Assignee: Mark Liu
>
> It's good to have code coverage in Python SDK to show the test coverage. Java 
> is using jacoco-maven-plugin to generate coverage report and coveralls 
> service to manage/display data. 
> Python have similar tool called coverage.py for report generation and 
> coveralls-python to send report to api of coveralls service. 
> It's nice to have one place (like coveralls service) to manage and show data 
> from different SDKs together/separately. However, there are still some 
> problems due to the fact of Beam CI system and multi languages in Beam. The 
> coveralls service doesn't have a good way to collect data separately from 
> different projects. But the postcommit builds are separated by sdks. 
> As the first step, I think it's good to have the python coverage report 
> printout in the build to give people a idea of the current coverage status at 
> lease.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (BEAM-2762) Coverage report for Python code

2017-08-10 Thread Mark Liu (JIRA)

 [ 
https://issues.apache.org/jira/browse/BEAM-2762?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mark Liu updated BEAM-2762:
---
Description: 
It's good to have code coverage in Python SDK to show the test coverage. Java 
is using jacoco-maven-plugin to generate coverage report and coveralls service 
to manage/display data. 

Python have similar tool called coverage.py for report generation and 
coveralls-python to send report to api of coveralls service. 

It's nice to have one place (like coveralls service) to manage and show data 
from different SDKs together/separately. However, there are still some problems 
due to the fact of Beam CI system and multi languages in Beam. The coveralls 
service doesn't have a good way to collect data separately from different 
projects. But the postcommit build are running in separate sdks. 

As the first step, I think it's good to have the python coverage report 
printout in the build to give people a idea of the current coverage status at 
lease.

  was:
It's good to have code coverage in Python SDK to show the test coverage. Java 
is using jacoco-maven-plugin to generate coverage report and coveralls service 
to manage/display data. 

Python have similar tool called coverage.py for report generation and 
coveralls-python to send report to api of coveralls service. 

It's nice to have one place (like coveralls service) to manage and show data 
from different SDKs together/separately. However, there are still some problems 
due to the fact of Beam CI system and multi languages. The coveralls service 
doesn't have a good way to collect data separately from different projects. But 
the postcommit build are running in separate sdks. 

As the first step, I think it's good to have the python coverage report 
printout in the build to give people a idea of the current coverage status at 
lease.


> Coverage report for Python code
> ---
>
> Key: BEAM-2762
> URL: https://issues.apache.org/jira/browse/BEAM-2762
> Project: Beam
>  Issue Type: Task
>  Components: sdk-py, testing
>Reporter: Mark Liu
>Assignee: Mark Liu
>
> It's good to have code coverage in Python SDK to show the test coverage. Java 
> is using jacoco-maven-plugin to generate coverage report and coveralls 
> service to manage/display data. 
> Python have similar tool called coverage.py for report generation and 
> coveralls-python to send report to api of coveralls service. 
> It's nice to have one place (like coveralls service) to manage and show data 
> from different SDKs together/separately. However, there are still some 
> problems due to the fact of Beam CI system and multi languages in Beam. The 
> coveralls service doesn't have a good way to collect data separately from 
> different projects. But the postcommit build are running in separate sdks. 
> As the first step, I think it's good to have the python coverage report 
> printout in the build to give people a idea of the current coverage status at 
> lease.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Created] (BEAM-2762) Coverage report for Python code

2017-08-10 Thread Mark Liu (JIRA)
Mark Liu created BEAM-2762:
--

 Summary: Coverage report for Python code
 Key: BEAM-2762
 URL: https://issues.apache.org/jira/browse/BEAM-2762
 Project: Beam
  Issue Type: Task
  Components: sdk-py, testing
Reporter: Mark Liu
Assignee: Mark Liu


It's good to have code coverage in Python SDK to show the test coverage. Java 
is using jacoco-maven-plugin to generate coverage report and coveralls service 
to manage/display data. 

Python have similar tool called coverage.py for report generation and 
coveralls-python to send report to api of coveralls service. 

It's nice to have one place (like coveralls service) to manage and show data 
from different SDKs together/separately. However, there are still some problems 
due to the fact of Beam CI system and multi languages. The coveralls service 
doesn't have a good way to collect data separately from different projects. But 
the postcommit build are running in separate sdks. 

As the first step, I think it's good to have the python coverage report 
printout in the build to give people a idea of the current coverage status at 
lease.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (BEAM-2648) beam_PerformanceTests_Python failing since 2017-07-17

2017-07-25 Thread Mark Liu (JIRA)

[ 
https://issues.apache.org/jira/browse/BEAM-2648?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16101035#comment-16101035
 ] 

Mark Liu commented on BEAM-2648:


PR is merged. Recent Jenkins run passed.
https://builds.apache.org/view/Beam/job/beam_PerformanceTests_Python/142/

> beam_PerformanceTests_Python failing since 2017-07-17
> -
>
> Key: BEAM-2648
> URL: https://issues.apache.org/jira/browse/BEAM-2648
> Project: Beam
>  Issue Type: Bug
>  Components: sdk-py
>Reporter: Kenneth Knowles
>Assignee: Mark Liu
>
> See 
> https://builds.apache.org/blue/organizations/jenkins/beam_PerformanceTests_Python/activity



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (BEAM-2648) beam_PerformanceTests_Python failing since 2017-07-17

2017-07-21 Thread Mark Liu (JIRA)

[ 
https://issues.apache.org/jira/browse/BEAM-2648?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16096694#comment-16096694
 ] 

Mark Liu commented on BEAM-2648:


fix PR: https://github.com/GoogleCloudPlatform/PerfKitBenchmarker/pull/1449 is 
created.

> beam_PerformanceTests_Python failing since 2017-07-17
> -
>
> Key: BEAM-2648
> URL: https://issues.apache.org/jira/browse/BEAM-2648
> Project: Beam
>  Issue Type: Bug
>  Components: sdk-py
>Reporter: Kenneth Knowles
>Assignee: Mark Liu
>
> See 
> https://builds.apache.org/blue/organizations/jenkins/beam_PerformanceTests_Python/activity



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Updated] (BEAM-2633) Update contribution guide to include information on running wordcount with modified sdk from head

2017-07-20 Thread Mark Liu (JIRA)

 [ 
https://issues.apache.org/jira/browse/BEAM-2633?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mark Liu updated BEAM-2633:
---
Description: 
[Quickstart-Python - Execute a pipeline locally - 
DataflowRunner|https://beam.apache.org/get-started/quickstart-py/#execute-a-pipeline-locally]
 is missing sdk artifacts portion. 

Need to add one step before pipeline execution to build python artifacts (or 
download from Apache release source). Also need to update execution command 
with "--sdk_location=". 

Otherwise, developers is getting following errors:
{code}
  Could not find a version that satisfies the requirement apache-beam==2.1.0 
(from versions: 0.6.0, 2.0.0)
No matching distribution found for apache-beam==2.1.0
Traceback (most recent call last):
  File 
"/System/Library/Frameworks/Python.framework/Versions/2.7/lib/python2.7/runpy.py",
 line 162, in _run_module_as_main
"__main__", fname, loader, pkg_name)
  File 
"/System/Library/Frameworks/Python.framework/Versions/2.7/lib/python2.7/runpy.py",
 line 72, in _run_code
exec code in run_globals
  File 
"/Users/markliu/Downloads/apache-beam-2.1.0/apache_beam/examples/wordcount.py", 
line 126, in 
run()
  File 
"/Users/markliu/Downloads/apache-beam-2.1.0/apache_beam/examples/wordcount.py", 
line 105, in run
result = p.run()
  File "apache_beam/pipeline.py", line 328, in run
return self.runner.run(self)
  File "apache_beam/runners/dataflow/dataflow_runner.py", line 283, in run
self.dataflow_client.create_job(self.job), self)
  File "apache_beam/utils/retry.py", line 168, in wrapper
return fun(*args, **kwargs)
  File "apache_beam/runners/dataflow/internal/apiclient.py", line 423, in 
create_job
self.create_job_description(job)
  File "apache_beam/runners/dataflow/internal/apiclient.py", line 446, in 
create_job_description
job.options, file_copy=self._gcs_file_copy)
  File "apache_beam/runners/dataflow/internal/dependency.py", line 399, in 
stage_job_resources
_stage_beam_sdk_tarball(sdk_remote_location, staged_path, temp_dir)
  File "apache_beam/runners/dataflow/internal/dependency.py", line 484, in 
_stage_beam_sdk_tarball
_dependency_file_copy(_download_pypi_sdk_package(temp_dir), staged_path)
  File "apache_beam/runners/dataflow/internal/dependency.py", line 580, in 
_download_pypi_sdk_package
processes.check_call(cmd_args)
  File "apache_beam/utils/processes.py", line 44, in check_call
return subprocess.check_call(*args, **kwargs)
  File 
"/System/Library/Frameworks/Python.framework/Versions/2.7/lib/python2.7/subprocess.py",
 line 540, in check_call
raise CalledProcessError(retcode, cmd)
subprocess.CalledProcessError: Command 
'['/Users/markliu/tmp/tmp-env/bin/python', '-m', 'pip', 'install', 
'--download', '/var/folders/vh/2rbdqyz53m905z_t_yhnvrt000bwz2/T/tmpBf92vB', 
'apache-beam==2.1.0', '--no-binary', ':all:', '--no-deps']' returned non-zero 
exit status 1
{code}

  was:
Developing guide in [Quickstart-Python - Execute a pipeline locally - 
DataflowRunner|https://beam.apache.org/get-started/quickstart-py/#execute-a-pipeline-locally]
 is missing sdk artifacts portion. 

Need to add one step before pipeline execution to build python artifacts (or 
download from Apache release source). Also need to update execution command 
with "--sdk_location=". 

Otherwise, developers is getting following errors:
{code}
  Could not find a version that satisfies the requirement apache-beam==2.1.0 
(from versions: 0.6.0, 2.0.0)
No matching distribution found for apache-beam==2.1.0
Traceback (most recent call last):
  File 
"/System/Library/Frameworks/Python.framework/Versions/2.7/lib/python2.7/runpy.py",
 line 162, in _run_module_as_main
"__main__", fname, loader, pkg_name)
  File 
"/System/Library/Frameworks/Python.framework/Versions/2.7/lib/python2.7/runpy.py",
 line 72, in _run_code
exec code in run_globals
  File 
"/Users/markliu/Downloads/apache-beam-2.1.0/apache_beam/examples/wordcount.py", 
line 126, in 
run()
  File 
"/Users/markliu/Downloads/apache-beam-2.1.0/apache_beam/examples/wordcount.py", 
line 105, in run
result = p.run()
  File "apache_beam/pipeline.py", line 328, in run
return self.runner.run(self)
  File "apache_beam/runners/dataflow/dataflow_runner.py", line 283, in run
self.dataflow_client.create_job(self.job), self)
  File "apache_beam/utils/retry.py", line 168, in wrapper
return fun(*args, **kwargs)
  File "apache_beam/runners/dataflow/internal/apiclient.py", line 423, in 
create_job
self.create_job_description(job)
  File "apache_beam/runners/dataflow/internal/apiclient.py", line 446, in 
create_job_description
job.options, file_copy=self._gcs_file_copy)
  File "apache_beam/runners/dataflow/internal/dependency.py", line 399, in 
stage_job_resources
_stage_beam_sdk_tarball(sdk_remote_location, staged_path, temp_dir)
  File "apache_beam/runners/dataflow/internal/dependency.py", line 484, in 

[jira] [Created] (BEAM-2649) Add documents for performance test

2017-07-20 Thread Mark Liu (JIRA)
Mark Liu created BEAM-2649:
--

 Summary: Add documents for performance test
 Key: BEAM-2649
 URL: https://issues.apache.org/jira/browse/BEAM-2649
 Project: Beam
  Issue Type: Task
  Components: website
Reporter: Mark Liu
Assignee: Mark Liu


Testing document need to be updated to include performance tests for Java and 
Python, as we already have three Java tests (DataflowRunner, JDBC and Spark) 
and one Python test running in Jenkins postcommit 
(https://github.com/apache/beam/tree/master/.test-infra/jenkins).



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Created] (BEAM-2633) Missing --sdk_location in Quickstart-Python - run wordcount.py with DataflowRunner

2017-07-18 Thread Mark Liu (JIRA)
Mark Liu created BEAM-2633:
--

 Summary: Missing --sdk_location in Quickstart-Python - run 
wordcount.py with DataflowRunner
 Key: BEAM-2633
 URL: https://issues.apache.org/jira/browse/BEAM-2633
 Project: Beam
  Issue Type: Bug
  Components: website
Reporter: Mark Liu
Assignee: Ahmet Altay


Developing guide in [Quickstart-Python - Execute a pipeline locally - 
DataflowRunner|https://beam.apache.org/get-started/quickstart-py/#execute-a-pipeline-locally]
 is missing sdk artifacts portion. 

Need to add one step before pipeline execution to build python artifacts (or 
download from Apache release source). Also need to update execution command 
with "--sdk_location=". 

Otherwise, developers is getting following errors:
{code}
  Could not find a version that satisfies the requirement apache-beam==2.1.0 
(from versions: 0.6.0, 2.0.0)
No matching distribution found for apache-beam==2.1.0
Traceback (most recent call last):
  File 
"/System/Library/Frameworks/Python.framework/Versions/2.7/lib/python2.7/runpy.py",
 line 162, in _run_module_as_main
"__main__", fname, loader, pkg_name)
  File 
"/System/Library/Frameworks/Python.framework/Versions/2.7/lib/python2.7/runpy.py",
 line 72, in _run_code
exec code in run_globals
  File 
"/Users/markliu/Downloads/apache-beam-2.1.0/apache_beam/examples/wordcount.py", 
line 126, in 
run()
  File 
"/Users/markliu/Downloads/apache-beam-2.1.0/apache_beam/examples/wordcount.py", 
line 105, in run
result = p.run()
  File "apache_beam/pipeline.py", line 328, in run
return self.runner.run(self)
  File "apache_beam/runners/dataflow/dataflow_runner.py", line 283, in run
self.dataflow_client.create_job(self.job), self)
  File "apache_beam/utils/retry.py", line 168, in wrapper
return fun(*args, **kwargs)
  File "apache_beam/runners/dataflow/internal/apiclient.py", line 423, in 
create_job
self.create_job_description(job)
  File "apache_beam/runners/dataflow/internal/apiclient.py", line 446, in 
create_job_description
job.options, file_copy=self._gcs_file_copy)
  File "apache_beam/runners/dataflow/internal/dependency.py", line 399, in 
stage_job_resources
_stage_beam_sdk_tarball(sdk_remote_location, staged_path, temp_dir)
  File "apache_beam/runners/dataflow/internal/dependency.py", line 484, in 
_stage_beam_sdk_tarball
_dependency_file_copy(_download_pypi_sdk_package(temp_dir), staged_path)
  File "apache_beam/runners/dataflow/internal/dependency.py", line 580, in 
_download_pypi_sdk_package
processes.check_call(cmd_args)
  File "apache_beam/utils/processes.py", line 44, in check_call
return subprocess.check_call(*args, **kwargs)
  File 
"/System/Library/Frameworks/Python.framework/Versions/2.7/lib/python2.7/subprocess.py",
 line 540, in check_call
raise CalledProcessError(retcode, cmd)
subprocess.CalledProcessError: Command 
'['/Users/markliu/tmp/tmp-env/bin/python', '-m', 'pip', 'install', 
'--download', '/var/folders/vh/2rbdqyz53m905z_t_yhnvrt000bwz2/T/tmpBf92vB', 
'apache-beam==2.1.0', '--no-binary', ':all:', '--no-deps']' returned non-zero 
exit status 1
{code}



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (BEAM-2542) Python performance test failed in Beam repo init

2017-07-06 Thread Mark Liu (JIRA)

[ 
https://issues.apache.org/jira/browse/BEAM-2542?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16077350#comment-16077350
 ] 

Mark Liu commented on BEAM-2542:


Build back to stable: 
https://builds.apache.org/view/Beam/job/beam_PerformanceTests_Python/67

Will mark this JIRA as resolved.

> Python performance test failed in Beam repo init
> 
>
> Key: BEAM-2542
> URL: https://issues.apache.org/jira/browse/BEAM-2542
> Project: Beam
>  Issue Type: Bug
>  Components: testing
>Reporter: Mark Liu
>Assignee: Mark Liu
> Fix For: Not applicable
>
>
> Jenkins link: 
> https://builds.apache.org/view/Beam/job/beam_PerformanceTests_Python/39
> Errors:
> {code}
> 2017-06-29 00:00:32,173 b368d0d5 MainThread beam_integration_benchmark(1/1) 
> ERRORError during benchmark beam_integration_benchmark
> Traceback (most recent call last):
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PerformanceTests_Python/PerfKitBenchmarker/perfkitbenchmarker/pkb.py",
>  line 555, in RunBenchmark
> DoPreparePhase(spec, detailed_timer)
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PerformanceTests_Python/PerfKitBenchmarker/perfkitbenchmarker/pkb.py",
>  line 450, in DoPreparePhase
> spec.BenchmarkPrepare(spec)
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PerformanceTests_Python/PerfKitBenchmarker/perfkitbenchmarker/linux_benchmarks/beam_integration_benchmark.py",
>  line 91, in Prepare
> beam_benchmark_helper.InitializeBeamRepo(benchmark_spec)
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PerformanceTests_Python/PerfKitBenchmarker/perfkitbenchmarker/beam_benchmark_helper.py",
>  line 136, in InitializeBeamRepo
> mvn_command)
> TypeError: AddRunnerProfileMvnArgument() takes exactly 3 arguments (2 given)
> {code}
> Need to investigate on benchmark code in Perfkit.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Resolved] (BEAM-2542) Python performance test failed in Beam repo init

2017-07-06 Thread Mark Liu (JIRA)

 [ 
https://issues.apache.org/jira/browse/BEAM-2542?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mark Liu resolved BEAM-2542.

   Resolution: Fixed
Fix Version/s: Not applicable

> Python performance test failed in Beam repo init
> 
>
> Key: BEAM-2542
> URL: https://issues.apache.org/jira/browse/BEAM-2542
> Project: Beam
>  Issue Type: Bug
>  Components: testing
>Reporter: Mark Liu
>Assignee: Mark Liu
> Fix For: Not applicable
>
>
> Jenkins link: 
> https://builds.apache.org/view/Beam/job/beam_PerformanceTests_Python/39
> Errors:
> {code}
> 2017-06-29 00:00:32,173 b368d0d5 MainThread beam_integration_benchmark(1/1) 
> ERRORError during benchmark beam_integration_benchmark
> Traceback (most recent call last):
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PerformanceTests_Python/PerfKitBenchmarker/perfkitbenchmarker/pkb.py",
>  line 555, in RunBenchmark
> DoPreparePhase(spec, detailed_timer)
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PerformanceTests_Python/PerfKitBenchmarker/perfkitbenchmarker/pkb.py",
>  line 450, in DoPreparePhase
> spec.BenchmarkPrepare(spec)
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PerformanceTests_Python/PerfKitBenchmarker/perfkitbenchmarker/linux_benchmarks/beam_integration_benchmark.py",
>  line 91, in Prepare
> beam_benchmark_helper.InitializeBeamRepo(benchmark_spec)
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PerformanceTests_Python/PerfKitBenchmarker/perfkitbenchmarker/beam_benchmark_helper.py",
>  line 136, in InitializeBeamRepo
> mvn_command)
> TypeError: AddRunnerProfileMvnArgument() takes exactly 3 arguments (2 given)
> {code}
> Need to investigate on benchmark code in Perfkit.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Closed] (BEAM-2542) Python performance test failed in Beam repo init

2017-07-06 Thread Mark Liu (JIRA)

 [ 
https://issues.apache.org/jira/browse/BEAM-2542?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mark Liu closed BEAM-2542.
--

> Python performance test failed in Beam repo init
> 
>
> Key: BEAM-2542
> URL: https://issues.apache.org/jira/browse/BEAM-2542
> Project: Beam
>  Issue Type: Bug
>  Components: testing
>Reporter: Mark Liu
>Assignee: Mark Liu
> Fix For: Not applicable
>
>
> Jenkins link: 
> https://builds.apache.org/view/Beam/job/beam_PerformanceTests_Python/39
> Errors:
> {code}
> 2017-06-29 00:00:32,173 b368d0d5 MainThread beam_integration_benchmark(1/1) 
> ERRORError during benchmark beam_integration_benchmark
> Traceback (most recent call last):
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PerformanceTests_Python/PerfKitBenchmarker/perfkitbenchmarker/pkb.py",
>  line 555, in RunBenchmark
> DoPreparePhase(spec, detailed_timer)
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PerformanceTests_Python/PerfKitBenchmarker/perfkitbenchmarker/pkb.py",
>  line 450, in DoPreparePhase
> spec.BenchmarkPrepare(spec)
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PerformanceTests_Python/PerfKitBenchmarker/perfkitbenchmarker/linux_benchmarks/beam_integration_benchmark.py",
>  line 91, in Prepare
> beam_benchmark_helper.InitializeBeamRepo(benchmark_spec)
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PerformanceTests_Python/PerfKitBenchmarker/perfkitbenchmarker/beam_benchmark_helper.py",
>  line 136, in InitializeBeamRepo
> mvn_command)
> TypeError: AddRunnerProfileMvnArgument() takes exactly 3 arguments (2 given)
> {code}
> Need to investigate on benchmark code in Perfkit.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Comment Edited] (BEAM-2542) Python performance test failed in Beam repo init

2017-07-05 Thread Mark Liu (JIRA)

[ 
https://issues.apache.org/jira/browse/BEAM-2542?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16075654#comment-16075654
 ] 

Mark Liu edited comment on BEAM-2542 at 7/6/17 12:00 AM:
-

it's merged. Waiting for next scheduled run to verify the fix.
https://builds.apache.org/view/Beam/job/beam_PerformanceTests_Python/


was (Author: markflyhigh):
it's merged. Waiting for next scheduled run to verify the fix.

> Python performance test failed in Beam repo init
> 
>
> Key: BEAM-2542
> URL: https://issues.apache.org/jira/browse/BEAM-2542
> Project: Beam
>  Issue Type: Bug
>  Components: testing
>Reporter: Mark Liu
>Assignee: Mark Liu
>
> Jenkins link: 
> https://builds.apache.org/view/Beam/job/beam_PerformanceTests_Python/39
> Errors:
> {code}
> 2017-06-29 00:00:32,173 b368d0d5 MainThread beam_integration_benchmark(1/1) 
> ERRORError during benchmark beam_integration_benchmark
> Traceback (most recent call last):
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PerformanceTests_Python/PerfKitBenchmarker/perfkitbenchmarker/pkb.py",
>  line 555, in RunBenchmark
> DoPreparePhase(spec, detailed_timer)
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PerformanceTests_Python/PerfKitBenchmarker/perfkitbenchmarker/pkb.py",
>  line 450, in DoPreparePhase
> spec.BenchmarkPrepare(spec)
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PerformanceTests_Python/PerfKitBenchmarker/perfkitbenchmarker/linux_benchmarks/beam_integration_benchmark.py",
>  line 91, in Prepare
> beam_benchmark_helper.InitializeBeamRepo(benchmark_spec)
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PerformanceTests_Python/PerfKitBenchmarker/perfkitbenchmarker/beam_benchmark_helper.py",
>  line 136, in InitializeBeamRepo
> mvn_command)
> TypeError: AddRunnerProfileMvnArgument() takes exactly 3 arguments (2 given)
> {code}
> Need to investigate on benchmark code in Perfkit.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (BEAM-2542) Python performance test failed in Beam repo init

2017-07-05 Thread Mark Liu (JIRA)

[ 
https://issues.apache.org/jira/browse/BEAM-2542?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16075654#comment-16075654
 ] 

Mark Liu commented on BEAM-2542:


it's merged. Waiting for next scheduled run to verify the fix.

> Python performance test failed in Beam repo init
> 
>
> Key: BEAM-2542
> URL: https://issues.apache.org/jira/browse/BEAM-2542
> Project: Beam
>  Issue Type: Bug
>  Components: testing
>Reporter: Mark Liu
>Assignee: Mark Liu
>
> Jenkins link: 
> https://builds.apache.org/view/Beam/job/beam_PerformanceTests_Python/39
> Errors:
> {code}
> 2017-06-29 00:00:32,173 b368d0d5 MainThread beam_integration_benchmark(1/1) 
> ERRORError during benchmark beam_integration_benchmark
> Traceback (most recent call last):
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PerformanceTests_Python/PerfKitBenchmarker/perfkitbenchmarker/pkb.py",
>  line 555, in RunBenchmark
> DoPreparePhase(spec, detailed_timer)
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PerformanceTests_Python/PerfKitBenchmarker/perfkitbenchmarker/pkb.py",
>  line 450, in DoPreparePhase
> spec.BenchmarkPrepare(spec)
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PerformanceTests_Python/PerfKitBenchmarker/perfkitbenchmarker/linux_benchmarks/beam_integration_benchmark.py",
>  line 91, in Prepare
> beam_benchmark_helper.InitializeBeamRepo(benchmark_spec)
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PerformanceTests_Python/PerfKitBenchmarker/perfkitbenchmarker/beam_benchmark_helper.py",
>  line 136, in InitializeBeamRepo
> mvn_command)
> TypeError: AddRunnerProfileMvnArgument() takes exactly 3 arguments (2 given)
> {code}
> Need to investigate on benchmark code in Perfkit.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Commented] (BEAM-2542) Python performance test failed in Beam repo init

2017-07-05 Thread Mark Liu (JIRA)

[ 
https://issues.apache.org/jira/browse/BEAM-2542?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16075476#comment-16075476
 ] 

Mark Liu commented on BEAM-2542:


PR: https://github.com/GoogleCloudPlatform/PerfKitBenchmarker/pull/1427 is in 
review.

> Python performance test failed in Beam repo init
> 
>
> Key: BEAM-2542
> URL: https://issues.apache.org/jira/browse/BEAM-2542
> Project: Beam
>  Issue Type: Bug
>  Components: testing
>Reporter: Mark Liu
>Assignee: Mark Liu
>
> Jenkins link: 
> https://builds.apache.org/view/Beam/job/beam_PerformanceTests_Python/39
> Errors:
> {code}
> 2017-06-29 00:00:32,173 b368d0d5 MainThread beam_integration_benchmark(1/1) 
> ERRORError during benchmark beam_integration_benchmark
> Traceback (most recent call last):
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PerformanceTests_Python/PerfKitBenchmarker/perfkitbenchmarker/pkb.py",
>  line 555, in RunBenchmark
> DoPreparePhase(spec, detailed_timer)
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PerformanceTests_Python/PerfKitBenchmarker/perfkitbenchmarker/pkb.py",
>  line 450, in DoPreparePhase
> spec.BenchmarkPrepare(spec)
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PerformanceTests_Python/PerfKitBenchmarker/perfkitbenchmarker/linux_benchmarks/beam_integration_benchmark.py",
>  line 91, in Prepare
> beam_benchmark_helper.InitializeBeamRepo(benchmark_spec)
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PerformanceTests_Python/PerfKitBenchmarker/perfkitbenchmarker/beam_benchmark_helper.py",
>  line 136, in InitializeBeamRepo
> mvn_command)
> TypeError: AddRunnerProfileMvnArgument() takes exactly 3 arguments (2 given)
> {code}
> Need to investigate on benchmark code in Perfkit.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Closed] (BEAM-2475) Add Jenkins Suite for Python Performance Test

2017-07-05 Thread Mark Liu (JIRA)

 [ 
https://issues.apache.org/jira/browse/BEAM-2475?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mark Liu closed BEAM-2475.
--

> Add Jenkins Suite for Python Performance Test
> -
>
> Key: BEAM-2475
> URL: https://issues.apache.org/jira/browse/BEAM-2475
> Project: Beam
>  Issue Type: Task
>  Components: testing
>Reporter: Mark Liu
>Assignee: Mark Liu
> Fix For: Not applicable
>
>
> Create a Jenkins suite that is scheduled to run Python performance test using 
> Perfkit. This task will focus on running Perfkit beam_integration_benchmark 
> using DataflowRunner. Once other runners are supported in Python SDK, we 
> should add more suites on Jenkins.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Resolved] (BEAM-2475) Add Jenkins Suite for Python Performance Test

2017-07-05 Thread Mark Liu (JIRA)

 [ 
https://issues.apache.org/jira/browse/BEAM-2475?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mark Liu resolved BEAM-2475.

   Resolution: Done
Fix Version/s: Not applicable

> Add Jenkins Suite for Python Performance Test
> -
>
> Key: BEAM-2475
> URL: https://issues.apache.org/jira/browse/BEAM-2475
> Project: Beam
>  Issue Type: Task
>  Components: testing
>Reporter: Mark Liu
>Assignee: Mark Liu
> Fix For: Not applicable
>
>
> Create a Jenkins suite that is scheduled to run Python performance test using 
> Perfkit. This task will focus on running Perfkit beam_integration_benchmark 
> using DataflowRunner. Once other runners are supported in Python SDK, we 
> should add more suites on Jenkins.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


  1   2   3   >