[jira] [Work logged] (BEAM-9510) Dependencies in base_image_requirements.txt are not compatible with each other

2020-03-16 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-9510?focusedWorklogId=404480&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-404480
 ]

ASF GitHub Bot logged work on BEAM-9510:


Author: ASF GitHub Bot
Created on: 17/Mar/20 06:52
Start Date: 17/Mar/20 06:52
Worklog Time Spent: 10m 
  Work Description: tvalentyn commented on issue #11142: [BEAM-9510] Fixing 
version incompatibilities in base_image_requirements.txt
URL: https://github.com/apache/beam/pull/11142#issuecomment-599906766
 
 
   While Jenkins does not show test result on the PR, the tests are actually 
running: https://builds.apache.org/job/beam_PostCommit_Py_ValCont_PR/122/. 
   Let's check the console logs that there are no conflicts before the merge.
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 404480)
Time Spent: 1h 20m  (was: 1h 10m)

> Dependencies in base_image_requirements.txt are not compatible with each other
> --
>
> Key: BEAM-9510
> URL: https://issues.apache.org/jira/browse/BEAM-9510
> Project: Beam
>  Issue Type: Bug
>  Components: sdk-py-harness
>Reporter: David Yan
>Assignee: Hannah Jiang
>Priority: Major
>  Time Spent: 1h 20m
>  Remaining Estimate: 0h
>
> [https://github.com/apache/beam/blob/master/sdks/python/container/base_image_requirements.txt#L56]
> says it requires google-cloud-bigquery==1.24.0, google-cloud-core==1.0.2, 
> google-cloud-bigtable==0.32.1, grpc-1.22.0 and tensorflow-2.1.0
> But they are incompatible with each other:
> ERROR: google-cloud-bigquery 1.24.0 has requirement 
> google-cloud-core<2.0dev,>=1.1.0, but you'll have google-cloud-core 1.0.2 
> which is incompatible.
> ERROR: google-cloud-bigtable 0.32.1 has requirement 
> google-cloud-core<0.30dev,>=0.29.0, but you'll have google-cloud-core 1.0.2 
> which is incompatible.
> ERROR: tensorboard 2.1.1 has requirement grpcio>=1.24.3, but you'll have 
> grpcio 1.22.0 which is incompatible.
> ERROR: tensorflow 2.1.0 has requirement scipy==1.4.1; python_version >= "3", 
> but you'll have scipy 1.2.2 which is incompatible.
>  
>  
>  
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (BEAM-9510) Dependencies in base_image_requirements.txt are not compatible with each other

2020-03-16 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-9510?focusedWorklogId=404477&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-404477
 ]

ASF GitHub Bot logged work on BEAM-9510:


Author: ASF GitHub Bot
Created on: 17/Mar/20 06:29
Start Date: 17/Mar/20 06:29
Worklog Time Spent: 10m 
  Work Description: tvalentyn commented on issue #11142: [BEAM-9510] Fixing 
version incompatibilities in base_image_requirements.txt
URL: https://github.com/apache/beam/pull/11142#issuecomment-599900258
 
 
   LGTM
   A few things out of scope of this PR:
   1) This PR fixes the incompatibility in dependencies, but the list of 
dependencies is in base_image_requirements.txt is still not exhaustive. So, 
container builds may not be repeatable over time, and there is a risk that we 
will introduce incompatible dependencies down the road and won't notice it.
   2) We can consider switching the base image to `python:slim-stretch` instead 
of `python:stretch` to reduce the size of the images: 
https://github.com/apache/beam/blob/master/sdks/python/container/Dockerfile#L20
   cc: @Hannah-Jiang 
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 404477)
Time Spent: 1h 10m  (was: 1h)

> Dependencies in base_image_requirements.txt are not compatible with each other
> --
>
> Key: BEAM-9510
> URL: https://issues.apache.org/jira/browse/BEAM-9510
> Project: Beam
>  Issue Type: Bug
>  Components: sdk-py-harness
>Reporter: David Yan
>Assignee: Hannah Jiang
>Priority: Major
>  Time Spent: 1h 10m
>  Remaining Estimate: 0h
>
> [https://github.com/apache/beam/blob/master/sdks/python/container/base_image_requirements.txt#L56]
> says it requires google-cloud-bigquery==1.24.0, google-cloud-core==1.0.2, 
> google-cloud-bigtable==0.32.1, grpc-1.22.0 and tensorflow-2.1.0
> But they are incompatible with each other:
> ERROR: google-cloud-bigquery 1.24.0 has requirement 
> google-cloud-core<2.0dev,>=1.1.0, but you'll have google-cloud-core 1.0.2 
> which is incompatible.
> ERROR: google-cloud-bigtable 0.32.1 has requirement 
> google-cloud-core<0.30dev,>=0.29.0, but you'll have google-cloud-core 1.0.2 
> which is incompatible.
> ERROR: tensorboard 2.1.1 has requirement grpcio>=1.24.3, but you'll have 
> grpcio 1.22.0 which is incompatible.
> ERROR: tensorflow 2.1.0 has requirement scipy==1.4.1; python_version >= "3", 
> but you'll have scipy 1.2.2 which is incompatible.
>  
>  
>  
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (BEAM-9510) Dependencies in base_image_requirements.txt are not compatible with each other

2020-03-16 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-9510?focusedWorklogId=404471&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-404471
 ]

ASF GitHub Bot logged work on BEAM-9510:


Author: ASF GitHub Bot
Created on: 17/Mar/20 06:14
Start Date: 17/Mar/20 06:14
Worklog Time Spent: 10m 
  Work Description: tvalentyn commented on issue #11142: [BEAM-9510] Fixing 
version incompatibilities in base_image_requirements.txt
URL: https://github.com/apache/beam/pull/11142#issuecomment-599896374
 
 
   Thanks, @davidyan74! Added a small change to reflect that `typing` should 
only be installed on Python 2 (this was recently changed in Beam master).
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 404471)
Time Spent: 50m  (was: 40m)

> Dependencies in base_image_requirements.txt are not compatible with each other
> --
>
> Key: BEAM-9510
> URL: https://issues.apache.org/jira/browse/BEAM-9510
> Project: Beam
>  Issue Type: Bug
>  Components: sdk-py-harness
>Reporter: David Yan
>Assignee: Hannah Jiang
>Priority: Major
>  Time Spent: 50m
>  Remaining Estimate: 0h
>
> [https://github.com/apache/beam/blob/master/sdks/python/container/base_image_requirements.txt#L56]
> says it requires google-cloud-bigquery==1.24.0, google-cloud-core==1.0.2, 
> google-cloud-bigtable==0.32.1, grpc-1.22.0 and tensorflow-2.1.0
> But they are incompatible with each other:
> ERROR: google-cloud-bigquery 1.24.0 has requirement 
> google-cloud-core<2.0dev,>=1.1.0, but you'll have google-cloud-core 1.0.2 
> which is incompatible.
> ERROR: google-cloud-bigtable 0.32.1 has requirement 
> google-cloud-core<0.30dev,>=0.29.0, but you'll have google-cloud-core 1.0.2 
> which is incompatible.
> ERROR: tensorboard 2.1.1 has requirement grpcio>=1.24.3, but you'll have 
> grpcio 1.22.0 which is incompatible.
> ERROR: tensorflow 2.1.0 has requirement scipy==1.4.1; python_version >= "3", 
> but you'll have scipy 1.2.2 which is incompatible.
>  
>  
>  
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (BEAM-9510) Dependencies in base_image_requirements.txt are not compatible with each other

2020-03-16 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-9510?focusedWorklogId=404472&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-404472
 ]

ASF GitHub Bot logged work on BEAM-9510:


Author: ASF GitHub Bot
Created on: 17/Mar/20 06:14
Start Date: 17/Mar/20 06:14
Worklog Time Spent: 10m 
  Work Description: tvalentyn commented on issue #11142: [BEAM-9510] Fixing 
version incompatibilities in base_image_requirements.txt
URL: https://github.com/apache/beam/pull/11142#issuecomment-599896422
 
 
   Run Python Dataflow ValidatesContainer
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 404472)
Time Spent: 1h  (was: 50m)

> Dependencies in base_image_requirements.txt are not compatible with each other
> --
>
> Key: BEAM-9510
> URL: https://issues.apache.org/jira/browse/BEAM-9510
> Project: Beam
>  Issue Type: Bug
>  Components: sdk-py-harness
>Reporter: David Yan
>Assignee: Hannah Jiang
>Priority: Major
>  Time Spent: 1h
>  Remaining Estimate: 0h
>
> [https://github.com/apache/beam/blob/master/sdks/python/container/base_image_requirements.txt#L56]
> says it requires google-cloud-bigquery==1.24.0, google-cloud-core==1.0.2, 
> google-cloud-bigtable==0.32.1, grpc-1.22.0 and tensorflow-2.1.0
> But they are incompatible with each other:
> ERROR: google-cloud-bigquery 1.24.0 has requirement 
> google-cloud-core<2.0dev,>=1.1.0, but you'll have google-cloud-core 1.0.2 
> which is incompatible.
> ERROR: google-cloud-bigtable 0.32.1 has requirement 
> google-cloud-core<0.30dev,>=0.29.0, but you'll have google-cloud-core 1.0.2 
> which is incompatible.
> ERROR: tensorboard 2.1.1 has requirement grpcio>=1.24.3, but you'll have 
> grpcio 1.22.0 which is incompatible.
> ERROR: tensorflow 2.1.0 has requirement scipy==1.4.1; python_version >= "3", 
> but you'll have scipy 1.2.2 which is incompatible.
>  
>  
>  
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (BEAM-9136) Add LICENSES and NOTICES to docker images

2020-03-16 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-9136?focusedWorklogId=40&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-40
 ]

ASF GitHub Bot logged work on BEAM-9136:


Author: ASF GitHub Bot
Created on: 17/Mar/20 04:26
Start Date: 17/Mar/20 04:26
Worklog Time Spent: 10m 
  Work Description: Hannah-Jiang commented on issue #11067: [BEAM-9136]Add 
licenses for dependencies
URL: https://github.com/apache/beam/pull/11067#issuecomment-599871035
 
 
   > In general, it's preferable to avoid checking in script-created files, 
preferring instead to create them on demand at build time. (There are 
exceptions, e.g. bootstrapping and/or difficult to obtain build tools, but I 
don't think that applies here.)
   > 
   > As another principle, it's preferable, where possible, to enforce 
correctness of the repository state before things get committed/merged (e.g. at 
pre-commit time), rather than have a separate, asynchronous, external process 
to try to fix things up after the fact.
   > 
   > Perhaps it's worth expanding the design doc on how we could do this?
   
   Yep, I will add it to the doc and send out for review when it is ready. 
Thanks for explanation!
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 40)
Time Spent: 2.5h  (was: 2h 20m)

> Add LICENSES and NOTICES to docker images
> -
>
> Key: BEAM-9136
> URL: https://issues.apache.org/jira/browse/BEAM-9136
> Project: Beam
>  Issue Type: Task
>  Components: build-system
>Reporter: Hannah Jiang
>Assignee: Hannah Jiang
>Priority: Major
>  Time Spent: 2.5h
>  Remaining Estimate: 0h
>
> Scan dependencies and add licenses and notices of the dependencies to SDK 
> docker images.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (BEAM-9494) Remove workaround for BQ transform for Dataflow

2020-03-16 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-9494?focusedWorklogId=404409&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-404409
 ]

ASF GitHub Bot logged work on BEAM-9494:


Author: ASF GitHub Bot
Created on: 17/Mar/20 02:34
Start Date: 17/Mar/20 02:34
Worklog Time Spent: 10m 
  Work Description: pabloem commented on issue #5: 
[cherry-pick][release-2.20.0][BEAM-9494] Reifying outputs from BQ file writing
URL: https://github.com/apache/beam/pull/5#issuecomment-599846684
 
 
   This is no longer needed. Sorry and thanks @amaliujia
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 404409)
Time Spent: 2.5h  (was: 2h 20m)

> Remove workaround for BQ transform for Dataflow
> ---
>
> Key: BEAM-9494
> URL: https://issues.apache.org/jira/browse/BEAM-9494
> Project: Beam
>  Issue Type: Bug
>  Components: io-py-gcp
>Reporter: Luke Cwik
>Assignee: Pablo Estrada
>Priority: Minor
> Fix For: 2.20.0
>
>  Time Spent: 2.5h
>  Remaining Estimate: 0h
>
> Dataflow incorrectly uses the Flatten input PCollection coder when it 
> performs an optimization instead of the output PCollection coder which can 
> lead to issues if these coders differ.
>  
> The workaround was introduced in [https://github.com/apache/beam/pull/11103]



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (BEAM-9494) Remove workaround for BQ transform for Dataflow

2020-03-16 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-9494?focusedWorklogId=404408&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-404408
 ]

ASF GitHub Bot logged work on BEAM-9494:


Author: ASF GitHub Bot
Created on: 17/Mar/20 02:34
Start Date: 17/Mar/20 02:34
Worklog Time Spent: 10m 
  Work Description: pabloem commented on pull request #5: 
[cherry-pick][release-2.20.0][BEAM-9494] Reifying outputs from BQ file writing
URL: https://github.com/apache/beam/pull/5
 
 
   
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 404408)
Time Spent: 2h 20m  (was: 2h 10m)

> Remove workaround for BQ transform for Dataflow
> ---
>
> Key: BEAM-9494
> URL: https://issues.apache.org/jira/browse/BEAM-9494
> Project: Beam
>  Issue Type: Bug
>  Components: io-py-gcp
>Reporter: Luke Cwik
>Assignee: Pablo Estrada
>Priority: Minor
> Fix For: 2.20.0
>
>  Time Spent: 2h 20m
>  Remaining Estimate: 0h
>
> Dataflow incorrectly uses the Flatten input PCollection coder when it 
> performs an optimization instead of the output PCollection coder which can 
> lead to issues if these coders differ.
>  
> The workaround was introduced in [https://github.com/apache/beam/pull/11103]



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (BEAM-9444) Shall we use GCP Libraries BOM to specify Google-related library versions?

2020-03-16 Thread Tomo Suzuki (Jira)


[ 
https://issues.apache.org/jira/browse/BEAM-9444?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17060584#comment-17060584
 ] 

Tomo Suzuki commented on BEAM-9444:
---

To run the build in my Mac:


{noformat}
suztomo-macbookpro44% ./gradlew -p sdks/java check -x 
extensions:sql:zetasql:check -x harness:test -x io:jdbc:test  -x io:kafka:test 
-x io:solr:test
...
BUILD SUCCESSFUL in 1h 15m 30s
551 actionable tasks: 44 executed, 507 up-to-date
{noformat}


> Shall we use GCP Libraries BOM to specify Google-related library versions?
> --
>
> Key: BEAM-9444
> URL: https://issues.apache.org/jira/browse/BEAM-9444
> Project: Beam
>  Issue Type: Task
>  Components: build-system
>Reporter: Tomo Suzuki
>Assignee: Tomo Suzuki
>Priority: Major
> Attachments: Screen Shot 2020-03-13 at 13.33.01.png
>
>  Time Spent: 1h 20m
>  Remaining Estimate: 0h
>
> Shall we use GCP Libraries BOM to specify Google-related library versions?
>   
>  I've been working on Beam's dependency upgrades in the past few months. I 
> think it's time to consider a long-term solution to keep the libraries 
> up-to-date with small maintenance effort. To achieve that, I propose Beam to 
> use GCP Libraries BOM to set the Google-related library versions, rather than 
> trying to make changes in each of ~30 Google libraries.
>   
> h1. Background
> A BOM is pom.xml that provides dependencyManagement to importing projects.
>   
>  GCP Libraries BOM is a BOM that includes many Google Cloud related libraries 
> + gRPC + protobuf. We (Google Cloud Java Diamond Dependency team) maintain 
> the BOM so that the set of the libraries are compatible with each other.
>   
> h1. Implementation
> Notes for obstacles.
> h2. BeamModulePlugin's "force" does not take BOM into account (thus fails)
> {{forcedModules}} via version resolution strategy is playing bad. This causes
> {noformat}
> A problem occurred evaluating project ':sdks:java:extensions:sql'. 
> Could not resolve all dependencies for configuration 
> ':sdks:java:extensions:sql:fmppTemplates'.
> Invalid format: 'com.google.cloud:google-cloud-core'. Group, name and version 
> cannot be empty. Correct example: 'org.gradle:gradle-core:1.0'{noformat}
> !Screen Shot 2020-03-13 at 13.33.01.png|width=489,height=287! 
>   
> h2.  :sdks:java:maven-archetypes:examples needs the version of 
> google-http-client
> The task requires the version for the library:
> {code}
> 'google-http-client.version': 
> dependencies.create(project.library.java.google_http_client).getVersion(),
> {code}
> This would generate NullPointerException. Running gradlew without the 
> subproject:
>  
> {code:java}
> ./gradlew -p sdks/java check -x :sdks:java:maven-archetypes:examples:check
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (BEAM-9444) Shall we use GCP Libraries BOM to specify Google-related library versions?

2020-03-16 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-9444?focusedWorklogId=404403&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-404403
 ]

ASF GitHub Bot logged work on BEAM-9444:


Author: ASF GitHub Bot
Created on: 17/Mar/20 02:09
Start Date: 17/Mar/20 02:09
Worklog Time Spent: 10m 
  Work Description: suztomo commented on pull request #11125: [BEAM-9444] 
(Old) Use Google Cloud Libraries BOM to set GCP library dependencies
URL: https://github.com/apache/beam/pull/11125
 
 
   
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 404403)
Time Spent: 1h 20m  (was: 1h 10m)

> Shall we use GCP Libraries BOM to specify Google-related library versions?
> --
>
> Key: BEAM-9444
> URL: https://issues.apache.org/jira/browse/BEAM-9444
> Project: Beam
>  Issue Type: Task
>  Components: build-system
>Reporter: Tomo Suzuki
>Assignee: Tomo Suzuki
>Priority: Major
> Attachments: Screen Shot 2020-03-13 at 13.33.01.png
>
>  Time Spent: 1h 20m
>  Remaining Estimate: 0h
>
> Shall we use GCP Libraries BOM to specify Google-related library versions?
>   
>  I've been working on Beam's dependency upgrades in the past few months. I 
> think it's time to consider a long-term solution to keep the libraries 
> up-to-date with small maintenance effort. To achieve that, I propose Beam to 
> use GCP Libraries BOM to set the Google-related library versions, rather than 
> trying to make changes in each of ~30 Google libraries.
>   
> h1. Background
> A BOM is pom.xml that provides dependencyManagement to importing projects.
>   
>  GCP Libraries BOM is a BOM that includes many Google Cloud related libraries 
> + gRPC + protobuf. We (Google Cloud Java Diamond Dependency team) maintain 
> the BOM so that the set of the libraries are compatible with each other.
>   
> h1. Implementation
> Notes for obstacles.
> h2. BeamModulePlugin's "force" does not take BOM into account (thus fails)
> {{forcedModules}} via version resolution strategy is playing bad. This causes
> {noformat}
> A problem occurred evaluating project ':sdks:java:extensions:sql'. 
> Could not resolve all dependencies for configuration 
> ':sdks:java:extensions:sql:fmppTemplates'.
> Invalid format: 'com.google.cloud:google-cloud-core'. Group, name and version 
> cannot be empty. Correct example: 'org.gradle:gradle-core:1.0'{noformat}
> !Screen Shot 2020-03-13 at 13.33.01.png|width=489,height=287! 
>   
> h2.  :sdks:java:maven-archetypes:examples needs the version of 
> google-http-client
> The task requires the version for the library:
> {code}
> 'google-http-client.version': 
> dependencies.create(project.library.java.google_http_client).getVersion(),
> {code}
> This would generate NullPointerException. Running gradlew without the 
> subproject:
>  
> {code:java}
> ./gradlew -p sdks/java check -x :sdks:java:maven-archetypes:examples:check
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (BEAM-7923) Interactive Beam

2020-03-16 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-7923?focusedWorklogId=404401&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-404401
 ]

ASF GitHub Bot logged work on BEAM-7923:


Author: ASF GitHub Bot
Created on: 17/Mar/20 02:02
Start Date: 17/Mar/20 02:02
Worklog Time Spent: 10m 
  Work Description: aaltay commented on issue #11141: [BEAM-7923] Include 
side effects in p.run
URL: https://github.com/apache/beam/pull/11141#issuecomment-599839271
 
 
   Could you resolve the conflict?
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 404401)
Time Spent: 2h 20m  (was: 2h 10m)

> Interactive Beam
> 
>
> Key: BEAM-7923
> URL: https://issues.apache.org/jira/browse/BEAM-7923
> Project: Beam
>  Issue Type: New Feature
>  Components: runner-py-interactive
>Reporter: Ning Kang
>Assignee: Ning Kang
>Priority: Major
>  Time Spent: 2h 20m
>  Remaining Estimate: 0h
>
> This is the top level ticket for all efforts leveraging [interactive 
> Beam|[https://github.com/apache/beam/tree/master/sdks/python/apache_beam/runners/interactive]]
> As the development goes, blocking tickets will be added to this one.
>  
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (BEAM-3301) Go SplittableDoFn support

2020-03-16 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-3301?focusedWorklogId=404390&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-404390
 ]

ASF GitHub Bot logged work on BEAM-3301:


Author: ASF GitHub Bot
Created on: 17/Mar/20 01:31
Start Date: 17/Mar/20 01:31
Worklog Time Spent: 10m 
  Work Description: youngoli commented on issue #11144: [BEAM-3301] Perform 
SDF validation (missing RestrictionTrackers).
URL: https://github.com/apache/beam/pull/11144#issuecomment-599832332
 
 
   Btw, only major thing I was unsure of. Should I be hiding this functionality 
behind an experiment flag of some kind? This initial one doesn't do so, and I 
don't expect people would randomly use it before it's ready because there's no 
user-facing documentation indicating its usage in the "beam" package, but it's 
still technically possible to use it right now.
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 404390)
Time Spent: 5.5h  (was: 5h 20m)

> Go SplittableDoFn support
> -
>
> Key: BEAM-3301
> URL: https://issues.apache.org/jira/browse/BEAM-3301
> Project: Beam
>  Issue Type: Improvement
>  Components: sdk-go
>Reporter: Henning Rohde
>Assignee: Daniel Oliveira
>Priority: Major
>  Time Spent: 5.5h
>  Remaining Estimate: 0h
>
> SDFs will be the only way to add streaming and liquid sharded IO for Go.
> Design doc: https://s.apache.org/splittable-do-fn



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (BEAM-3301) Go SplittableDoFn support

2020-03-16 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-3301?focusedWorklogId=404388&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-404388
 ]

ASF GitHub Bot logged work on BEAM-3301:


Author: ASF GitHub Bot
Created on: 17/Mar/20 01:29
Start Date: 17/Mar/20 01:29
Worklog Time Spent: 10m 
  Work Description: youngoli commented on pull request #11144: [BEAM-3301] 
Perform SDF validation (missing RestrictionTrackers).
URL: https://github.com/apache/beam/pull/11144
 
 
   Added SDF detection and method validation when creating DoFns. This
   adds all the validation except for anything to do with restriction
   trackers, which will be added in a later change.
   
   
   
   Thank you for your contribution! Follow this checklist to help us 
incorporate your contribution quickly and easily:
   
- [x] [**Choose 
reviewer(s)**](https://beam.apache.org/contribute/#make-your-change) and 
mention them in a comment (`R: @username`).
- [x] Format the pull request title like `[BEAM-XXX] Fixes bug in 
ApproximateQuantiles`, where you replace `BEAM-XXX` with the appropriate JIRA 
issue, if applicable. This will automatically link the pull request to the 
issue.
- [ ] Update `CHANGES.md` with noteworthy changes.
- [x] If this contribution is large, please file an Apache [Individual 
Contributor License Agreement](https://www.apache.org/licenses/icla.pdf).
   
   See the [Contributor Guide](https://beam.apache.org/contribute) for more 
tips on [how to make review process 
smoother](https://beam.apache.org/contribute/#make-reviewers-job-easier).
   
   Post-Commit Tests Status (on master branch)
   

   
   Lang | SDK | Apex | Dataflow | Flink | Gearpump | Samza | Spark
   --- | --- | --- | --- | --- | --- | --- | ---
   Go | [![Build 
Status](https://builds.apache.org/job/beam_PostCommit_Go/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Go/lastCompletedBuild/)
 | --- | --- | [![Build 
Status](https://builds.apache.org/job/beam_PostCommit_Go_VR_Flink/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Go_VR_Flink/lastCompletedBuild/)
 | --- | --- | [![Build 
Status](https://builds.apache.org/job/beam_PostCommit_Go_VR_Spark/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Go_VR_Spark/lastCompletedBuild/)
   Java | [![Build 
Status](https://builds.apache.org/job/beam_PostCommit_Java/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Java/lastCompletedBuild/)
 | [![Build 
Status](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Apex/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Apex/lastCompletedBuild/)
 | [![Build 
Status](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Dataflow/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Dataflow/lastCompletedBuild/)[![Build
 
Status](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Dataflow_Java11/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Dataflow_Java11/lastCompletedBuild/)
 | [![Build 
Status](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Flink/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Flink/lastCompletedBuild/)[![Build
 
Status](https://builds.apache.org/job/beam_PostCommit_Java_PVR_Flink_Batch/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Java_PVR_Flink_Batch/lastCompletedBuild/)[![Build
 
Status](https://builds.apache.org/job/beam_PostCommit_Java_PVR_Flink_Streaming/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Java_PVR_Flink_Streaming/lastCompletedBuild/)
 | [![Build 
Status](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Gearpump/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Gearpump/lastCompletedBuild/)
 | [![Build 
Status](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Samza/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Samza/lastCompletedBuild/)
 | [![Build 
Status](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/lastCompletedBuild/)[![Build
 
Status](https://builds.apache.org/job/beam_PostCommit_Java_PVR_Spark_Batch/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Java_PVR_Spark_Batch/lastCompletedBuild/)[![Build
 
Status](https://builds.apache.org/job/beam_PostCommit_

[jira] [Work logged] (BEAM-3301) Go SplittableDoFn support

2020-03-16 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-3301?focusedWorklogId=404389&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-404389
 ]

ASF GitHub Bot logged work on BEAM-3301:


Author: ASF GitHub Bot
Created on: 17/Mar/20 01:29
Start Date: 17/Mar/20 01:29
Worklog Time Spent: 10m 
  Work Description: youngoli commented on issue #11144: [BEAM-3301] Perform 
SDF validation (missing RestrictionTrackers).
URL: https://github.com/apache/beam/pull/11144#issuecomment-599831761
 
 
   R: @lostluck 
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 404389)
Time Spent: 5h 20m  (was: 5h 10m)

> Go SplittableDoFn support
> -
>
> Key: BEAM-3301
> URL: https://issues.apache.org/jira/browse/BEAM-3301
> Project: Beam
>  Issue Type: Improvement
>  Components: sdk-go
>Reporter: Henning Rohde
>Assignee: Daniel Oliveira
>Priority: Major
>  Time Spent: 5h 20m
>  Remaining Estimate: 0h
>
> SDFs will be the only way to add streaming and liquid sharded IO for Go.
> Design doc: https://s.apache.org/splittable-do-fn



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (BEAM-3301) Go SplittableDoFn support

2020-03-16 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-3301?focusedWorklogId=404382&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-404382
 ]

ASF GitHub Bot logged work on BEAM-3301:


Author: ASF GitHub Bot
Created on: 17/Mar/20 01:09
Start Date: 17/Mar/20 01:09
Worklog Time Spent: 10m 
  Work Description: youngoli commented on pull request #10991: [BEAM-3301] 
Refactor DoFn validation & allow specifying main inputs.
URL: https://github.com/apache/beam/pull/10991
 
 
   
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 404382)
Time Spent: 5h  (was: 4h 50m)

> Go SplittableDoFn support
> -
>
> Key: BEAM-3301
> URL: https://issues.apache.org/jira/browse/BEAM-3301
> Project: Beam
>  Issue Type: Improvement
>  Components: sdk-go
>Reporter: Henning Rohde
>Assignee: Daniel Oliveira
>Priority: Major
>  Time Spent: 5h
>  Remaining Estimate: 0h
>
> SDFs will be the only way to add streaming and liquid sharded IO for Go.
> Design doc: https://s.apache.org/splittable-do-fn



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (BEAM-9498) RowJson exception for unsupported types should list the relevant fields

2020-03-16 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-9498?focusedWorklogId=404380&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-404380
 ]

ASF GitHub Bot logged work on BEAM-9498:


Author: ASF GitHub Bot
Created on: 17/Mar/20 01:06
Start Date: 17/Mar/20 01:06
Worklog Time Spent: 10m 
  Work Description: apilloud commented on pull request #9: [BEAM-9498] 
Include descriptor and type of unsupported fields in RowJson exception
URL: https://github.com/apache/beam/pull/9#discussion_r393390344
 
 

 ##
 File path: sdks/java/core/src/main/java/org/apache/beam/sdk/util/RowJson.java
 ##
 @@ -84,41 +85,76 @@
   private static final ImmutableSet SUPPORTED_TYPES =
   ImmutableSet.of(BYTE, INT16, INT32, INT64, FLOAT, DOUBLE, BOOLEAN, 
STRING, DECIMAL);
 
+  /**
+   * Throws {@link UnsupportedRowJsonException} if {@code schema} contains an 
unsupported field
+   * type.
+   */
   public static void verifySchemaSupported(Schema schema) {
-schema.getFields().forEach(RowJson::verifyFieldTypeSupported);
+ImmutableList unsupportedFields = 
findUnsupportedFields(schema);
+if (!unsupportedFields.isEmpty()) {
+  throw new UnsupportedRowJsonException(
+  String.format(
+  "Field type%s %s not supported when converting between JSON and 
Rows. Supported types are: %s",
+  unsupportedFields.size() > 1 ? "s" : "",
+  unsupportedFields.toString(),
+  SUPPORTED_TYPES.toString()));
+}
+  }
+
+  private static class UnsupportedField {
+final String descriptor;
+final TypeName typeName;
+
+UnsupportedField(String descriptor, TypeName typeName) {
+  this.descriptor = descriptor;
+  this.typeName = typeName;
+}
+
+@Override
+public String toString() {
+  return this.descriptor + "=" + this.typeName;
+}
+  }
+
+  private static ImmutableList findUnsupportedFields(Schema 
schema) {
+return schema.getFields().stream()
+.flatMap((field) -> findUnsupportedFields(field).stream())
+.collect(toImmutableList());
   }
 
-  static void verifyFieldTypeSupported(Field field) {
+  private static ImmutableList findUnsupportedFields(Field 
field) {
 FieldType fieldType = field.getType();
 
 Review comment:
   nit: drop the variable, it doesn't add value.
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 404380)
Time Spent: 20m  (was: 10m)

> RowJson exception for unsupported types should list the relevant fields
> ---
>
> Key: BEAM-9498
> URL: https://issues.apache.org/jira/browse/BEAM-9498
> Project: Beam
>  Issue Type: Improvement
>  Components: dsl-sql
>Reporter: Brian Hulette
>Assignee: Brian Hulette
>Priority: Major
>  Time Spent: 20m
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (BEAM-9510) Dependencies in base_image_requirements.txt are not compatible with each other

2020-03-16 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-9510?focusedWorklogId=404381&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-404381
 ]

ASF GitHub Bot logged work on BEAM-9510:


Author: ASF GitHub Bot
Created on: 17/Mar/20 01:06
Start Date: 17/Mar/20 01:06
Worklog Time Spent: 10m 
  Work Description: tvalentyn commented on issue #11142: [BEAM-9510] Fixing 
version incompatibilities in base_image_requirements.txt
URL: https://github.com/apache/beam/pull/11142#issuecomment-599826365
 
 
   Run Python Dataflow ValidatesContainer
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 404381)
Time Spent: 40m  (was: 0.5h)

> Dependencies in base_image_requirements.txt are not compatible with each other
> --
>
> Key: BEAM-9510
> URL: https://issues.apache.org/jira/browse/BEAM-9510
> Project: Beam
>  Issue Type: Bug
>  Components: sdk-py-harness
>Reporter: David Yan
>Assignee: Hannah Jiang
>Priority: Major
>  Time Spent: 40m
>  Remaining Estimate: 0h
>
> [https://github.com/apache/beam/blob/master/sdks/python/container/base_image_requirements.txt#L56]
> says it requires google-cloud-bigquery==1.24.0, google-cloud-core==1.0.2, 
> google-cloud-bigtable==0.32.1, grpc-1.22.0 and tensorflow-2.1.0
> But they are incompatible with each other:
> ERROR: google-cloud-bigquery 1.24.0 has requirement 
> google-cloud-core<2.0dev,>=1.1.0, but you'll have google-cloud-core 1.0.2 
> which is incompatible.
> ERROR: google-cloud-bigtable 0.32.1 has requirement 
> google-cloud-core<0.30dev,>=0.29.0, but you'll have google-cloud-core 1.0.2 
> which is incompatible.
> ERROR: tensorboard 2.1.1 has requirement grpcio>=1.24.3, but you'll have 
> grpcio 1.22.0 which is incompatible.
> ERROR: tensorflow 2.1.0 has requirement scipy==1.4.1; python_version >= "3", 
> but you'll have scipy 1.2.2 which is incompatible.
>  
>  
>  
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (BEAM-9498) RowJson exception for unsupported types should list the relevant fields

2020-03-16 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-9498?focusedWorklogId=404379&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-404379
 ]

ASF GitHub Bot logged work on BEAM-9498:


Author: ASF GitHub Bot
Created on: 17/Mar/20 01:06
Start Date: 17/Mar/20 01:06
Worklog Time Spent: 10m 
  Work Description: apilloud commented on pull request #9: [BEAM-9498] 
Include descriptor and type of unsupported fields in RowJson exception
URL: https://github.com/apache/beam/pull/9#discussion_r393391148
 
 

 ##
 File path: sdks/java/core/src/main/java/org/apache/beam/sdk/util/RowJson.java
 ##
 @@ -84,41 +85,76 @@
   private static final ImmutableSet SUPPORTED_TYPES =
   ImmutableSet.of(BYTE, INT16, INT32, INT64, FLOAT, DOUBLE, BOOLEAN, 
STRING, DECIMAL);
 
+  /**
+   * Throws {@link UnsupportedRowJsonException} if {@code schema} contains an 
unsupported field
+   * type.
+   */
   public static void verifySchemaSupported(Schema schema) {
-schema.getFields().forEach(RowJson::verifyFieldTypeSupported);
+ImmutableList unsupportedFields = 
findUnsupportedFields(schema);
+if (!unsupportedFields.isEmpty()) {
+  throw new UnsupportedRowJsonException(
+  String.format(
+  "Field type%s %s not supported when converting between JSON and 
Rows. Supported types are: %s",
+  unsupportedFields.size() > 1 ? "s" : "",
+  unsupportedFields.toString(),
+  SUPPORTED_TYPES.toString()));
+}
+  }
+
+  private static class UnsupportedField {
+final String descriptor;
+final TypeName typeName;
+
+UnsupportedField(String descriptor, TypeName typeName) {
+  this.descriptor = descriptor;
+  this.typeName = typeName;
+}
+
+@Override
+public String toString() {
+  return this.descriptor + "=" + this.typeName;
+}
+  }
+
+  private static ImmutableList findUnsupportedFields(Schema 
schema) {
+return schema.getFields().stream()
+.flatMap((field) -> findUnsupportedFields(field).stream())
+.collect(toImmutableList());
   }
 
-  static void verifyFieldTypeSupported(Field field) {
+  private static ImmutableList findUnsupportedFields(Field 
field) {
 FieldType fieldType = field.getType();
-verifyFieldTypeSupported(fieldType);
+return findUnsupportedFields(fieldType, field.getName());
   }
 
-  static void verifyFieldTypeSupported(FieldType fieldType) {
+  private static ImmutableList findUnsupportedFields(
+  FieldType fieldType, String fieldName) {
 TypeName fieldTypeName = fieldType.getTypeName();
 
 if (fieldTypeName.isCompositeType()) {
   Schema rowFieldSchema = fieldType.getRowSchema();
 
 Review comment:
   nit: drop the variable, it doesn't add value.
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 404379)
Time Spent: 20m  (was: 10m)

> RowJson exception for unsupported types should list the relevant fields
> ---
>
> Key: BEAM-9498
> URL: https://issues.apache.org/jira/browse/BEAM-9498
> Project: Beam
>  Issue Type: Improvement
>  Components: dsl-sql
>Reporter: Brian Hulette
>Assignee: Brian Hulette
>Priority: Major
>  Time Spent: 20m
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (BEAM-9507) Beam dependency check failing

2020-03-16 Thread Ahmet Altay (Jira)


[ 
https://issues.apache.org/jira/browse/BEAM-9507?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17060556#comment-17060556
 ] 

Ahmet Altay commented on BEAM-9507:
---

cc: [~yifanzou]

> Beam dependency check failing
> -
>
> Key: BEAM-9507
> URL: https://issues.apache.org/jira/browse/BEAM-9507
> Project: Beam
>  Issue Type: Bug
>  Components: testing
>Reporter: Michał Walenia
>Priority: Major
>
> Here are the logs:
> [https://builds.apache.org/job/beam_Dependency_Check/257/console]
>  
>from grpc_tools import protoc*13:04:25* ImportError: No module 
> named 'grpc_tools'*13:04:25* *13:04:25* During handling of the above 
> exception, another exception occurred:*13:04:25* *13:04:25* Traceback 
> (most recent call last):*13:04:25*   File 
> "/usr/lib/python3.5/multiprocessing/process.py", line 249, in 
> _bootstrap*13:04:25* self.run()*13:04:25*   File 
> "/usr/lib/python3.5/multiprocessing/process.py", line 93, in run*13:04:25*
>  self._target(*self._args, **self._kwargs)*13:04:25*   File 
> "/home/jenkins/jenkins-slave/workspace/beam_Dependency_Check/src/sdks/python/gen_protos.py",
>  line 378, in _install_grpcio_tools_and_generate_proto_files*13:04:25*
>  generate_proto_files(force=force)*13:04:25*   File 
> "/home/jenkins/jenkins-slave/workspace/beam_Dependency_Check/src/sdks/python/gen_protos.py",
>  line 315, in generate_proto_files*13:04:25* protoc_gen_mypy = 
> _find_protoc_gen_mypy()*13:04:25*   File 
> "/home/jenkins/jenkins-slave/workspace/beam_Dependency_Check/src/sdks/python/gen_protos.py",
>  line 233, in _find_protoc_gen_mypy*13:04:25* (fname, ', 
> '.join(search_paths)))*13:04:25* RuntimeError: Could not find 
> protoc-gen-mypy in 
> /home/jenkins/jenkins-slave/workspace/beam_Dependency_Check/src/sdks/python/sdks/python/bin,
>  
> /home/jenkins/jenkins-slave/workspace/beam_Dependency_Check/src/sdks/python/sdks/python/bin,
>  /home/jenkins/tools/java/latest1.8/bin, /usr/local/sbin, /usr/local/bin, 
> /usr/sbin, /usr/bin, /sbin, /bin, /usr/games, /usr/local/games*13:04:25* 
> Traceback (most recent call last):*13:04:25*   File 
> "/home/jenkins/jenkins-slave/workspace/beam_Dependency_Check/src/sdks/python/gen_protos.py",
>  line 292, in generate_proto_files*13:04:25* from grpc_tools import 
> protoc*13:04:25* ImportError: No module named 'grpc_tools'*13:04:25* 
> *13:04:25* During handling of the above exception, another exception 
> occurred:*13:04:25* *13:04:25* Traceback (most recent call 
> last):*13:04:25*   File "", line 1, in *13:04:25*   
> File 
> "/home/jenkins/jenkins-slave/workspace/beam_Dependency_Check/src/sdks/python/setup.py",
>  line 315, in *13:04:25* 'mypy': 
> generate_protos_first(mypy),*13:04:25*   File 
> "/home/jenkins/jenkins-slave/workspace/beam_Dependency_Check/src/sdks/python/sdks/python/lib/python3.5/site-packages/setuptools/__init__.py",
>  line 144, in setup*13:04:25* return 
> distutils.core.setup(**attrs)*13:04:25*   File 
> "/usr/lib/python3.5/distutils/core.py", line 148, in setup*13:04:25* 
> dist.run_commands()*13:04:25*   File 
> "/usr/lib/python3.5/distutils/dist.py", line 955, in run_commands*13:04:25*   
>   self.run_command(cmd)*13:04:25*   File 
> "/usr/lib/python3.5/distutils/dist.py", line 974, in run_command*13:04:25*
>  cmd_obj.run()*13:04:25*   File 
> "/home/jenkins/jenkins-slave/workspace/beam_Dependency_Check/src/sdks/python/setup.py",
>  line 239, in run*13:04:25* 
> gen_protos.generate_proto_files()*13:04:25*   File 
> "/home/jenkins/jenkins-slave/workspace/beam_Dependency_Check/src/sdks/python/gen_protos.py",
>  line 310, in generate_proto_files*13:04:25* raise ValueError("Proto 
> generation failed (see log for details).")*13:04:25* ValueError: Proto 
> generation failed (see log for details).*13:04:25* 
> *13:04:25* ERROR: Command errored out 
> with exit status 1: python setup.py egg_info Check the logs for full command 
> output.*13:04:25* *13:04:25* >
>  *Task :sdks:python:dependencyUpdates*
>  FAILED*13:04:25* *13:04:25* FAILURE: Build failed with an 
> exception.*13:04:25* *13:04:25* * Where:*13:04:25* Build file 
> '/home/jenkins/jenkins-slave/workspace/beam_Dependency_Check/src/sdks/python/build.gradle'
>  line: 94*13:04:25* *13:04:25* * What went wrong:*13:04:25* Execution failed 
> for task ':sdks:python:dependencyUpdates'.*13:04:25* > Process 'command 'sh'' 
> finished with non-zero exit value 1



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (BEAM-9526) Bug in top.LargestPerKey, disappearing elements.

2020-03-16 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-9526?focusedWorklogId=404378&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-404378
 ]

ASF GitHub Bot logged work on BEAM-9526:


Author: ASF GitHub Bot
Created on: 17/Mar/20 01:05
Start Date: 17/Mar/20 01:05
Worklog Time Spent: 10m 
  Work Description: youngoli commented on issue #11143: [BEAM-9526] Add 
missing unmarshalling in top.LargestPerKey.
URL: https://github.com/apache/beam/pull/11143#issuecomment-599826047
 
 
   R: @lostluck 
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 404378)
Time Spent: 20m  (was: 10m)

> Bug in top.LargestPerKey, disappearing elements.
> 
>
> Key: BEAM-9526
> URL: https://issues.apache.org/jira/browse/BEAM-9526
> Project: Beam
>  Issue Type: Bug
>  Components: sdk-go
>Reporter: Daniel Oliveira
>Assignee: Daniel Oliveira
>Priority: Minor
>  Time Spent: 20m
>  Remaining Estimate: 0h
>
> There's a bug in top.LargestPerKey that appears when running it on a runner 
> that has combiner lifting. Since it's similar to a GBK, we should be 
> inputting a K, V and receiving a K, []V as output. When the runner has 
> combiner lifting, the []V always contains an empty slice.
>  
> I already found the fix for this, just logging this to attach it to the PR.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (BEAM-9526) Bug in top.LargestPerKey, disappearing elements.

2020-03-16 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-9526?focusedWorklogId=404377&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-404377
 ]

ASF GitHub Bot logged work on BEAM-9526:


Author: ASF GitHub Bot
Created on: 17/Mar/20 01:03
Start Date: 17/Mar/20 01:03
Worklog Time Spent: 10m 
  Work Description: youngoli commented on pull request #11143: [BEAM-9526] 
Add missing unmarshalling in top.LargestPerKey.
URL: https://github.com/apache/beam/pull/11143
 
 
   ExtractOutput wasn't unmarshalling the accumulator first, which would
   cause the values to get dropped if the accumulator was marshalled when
   ExtractOutput was called. This fixes that and modifies the unit test
   slightly to catch it if it stops working in the future.
   
   
   
   Thank you for your contribution! Follow this checklist to help us 
incorporate your contribution quickly and easily:
   
- [x] [**Choose 
reviewer(s)**](https://beam.apache.org/contribute/#make-your-change) and 
mention them in a comment (`R: @username`).
- [x] Format the pull request title like `[BEAM-XXX] Fixes bug in 
ApproximateQuantiles`, where you replace `BEAM-XXX` with the appropriate JIRA 
issue, if applicable. This will automatically link the pull request to the 
issue.
- [ ] Update `CHANGES.md` with noteworthy changes.
- [ ] If this contribution is large, please file an Apache [Individual 
Contributor License Agreement](https://www.apache.org/licenses/icla.pdf).
   
   See the [Contributor Guide](https://beam.apache.org/contribute) for more 
tips on [how to make review process 
smoother](https://beam.apache.org/contribute/#make-reviewers-job-easier).
   
   Post-Commit Tests Status (on master branch)
   

   
   Lang | SDK | Apex | Dataflow | Flink | Gearpump | Samza | Spark
   --- | --- | --- | --- | --- | --- | --- | ---
   Go | [![Build 
Status](https://builds.apache.org/job/beam_PostCommit_Go/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Go/lastCompletedBuild/)
 | --- | --- | [![Build 
Status](https://builds.apache.org/job/beam_PostCommit_Go_VR_Flink/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Go_VR_Flink/lastCompletedBuild/)
 | --- | --- | [![Build 
Status](https://builds.apache.org/job/beam_PostCommit_Go_VR_Spark/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Go_VR_Spark/lastCompletedBuild/)
   Java | [![Build 
Status](https://builds.apache.org/job/beam_PostCommit_Java/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Java/lastCompletedBuild/)
 | [![Build 
Status](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Apex/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Apex/lastCompletedBuild/)
 | [![Build 
Status](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Dataflow/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Dataflow/lastCompletedBuild/)[![Build
 
Status](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Dataflow_Java11/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Dataflow_Java11/lastCompletedBuild/)
 | [![Build 
Status](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Flink/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Flink/lastCompletedBuild/)[![Build
 
Status](https://builds.apache.org/job/beam_PostCommit_Java_PVR_Flink_Batch/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Java_PVR_Flink_Batch/lastCompletedBuild/)[![Build
 
Status](https://builds.apache.org/job/beam_PostCommit_Java_PVR_Flink_Streaming/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Java_PVR_Flink_Streaming/lastCompletedBuild/)
 | [![Build 
Status](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Gearpump/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Gearpump/lastCompletedBuild/)
 | [![Build 
Status](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Samza/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Samza/lastCompletedBuild/)
 | [![Build 
Status](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/lastCompletedBuild/)[![Build
 
Status](https://builds.apache.org/job/beam_PostCommit_Java_PVR_Spark_Batch/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Java_PVR_Spark_Batch/lastCo

[jira] [Work logged] (BEAM-9510) Dependencies in base_image_requirements.txt are not compatible with each other

2020-03-16 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-9510?focusedWorklogId=404374&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-404374
 ]

ASF GitHub Bot logged work on BEAM-9510:


Author: ASF GitHub Bot
Created on: 17/Mar/20 01:00
Start Date: 17/Mar/20 01:00
Worklog Time Spent: 10m 
  Work Description: tvalentyn commented on issue #11142: [BEAM-9510] Fixing 
version incompatibilities in base_image_requirements.txt
URL: https://github.com/apache/beam/pull/11142#issuecomment-599825064
 
 
   Run Dataflow Python ValidatesContainer
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 404374)
Time Spent: 0.5h  (was: 20m)

> Dependencies in base_image_requirements.txt are not compatible with each other
> --
>
> Key: BEAM-9510
> URL: https://issues.apache.org/jira/browse/BEAM-9510
> Project: Beam
>  Issue Type: Bug
>  Components: sdk-py-harness
>Reporter: David Yan
>Assignee: Hannah Jiang
>Priority: Major
>  Time Spent: 0.5h
>  Remaining Estimate: 0h
>
> [https://github.com/apache/beam/blob/master/sdks/python/container/base_image_requirements.txt#L56]
> says it requires google-cloud-bigquery==1.24.0, google-cloud-core==1.0.2, 
> google-cloud-bigtable==0.32.1, grpc-1.22.0 and tensorflow-2.1.0
> But they are incompatible with each other:
> ERROR: google-cloud-bigquery 1.24.0 has requirement 
> google-cloud-core<2.0dev,>=1.1.0, but you'll have google-cloud-core 1.0.2 
> which is incompatible.
> ERROR: google-cloud-bigtable 0.32.1 has requirement 
> google-cloud-core<0.30dev,>=0.29.0, but you'll have google-cloud-core 1.0.2 
> which is incompatible.
> ERROR: tensorboard 2.1.1 has requirement grpcio>=1.24.3, but you'll have 
> grpcio 1.22.0 which is incompatible.
> ERROR: tensorflow 2.1.0 has requirement scipy==1.4.1; python_version >= "3", 
> but you'll have scipy 1.2.2 which is incompatible.
>  
>  
>  
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (BEAM-9510) Dependencies in base_image_requirements.txt are not compatible with each other

2020-03-16 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-9510?focusedWorklogId=404371&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-404371
 ]

ASF GitHub Bot logged work on BEAM-9510:


Author: ASF GitHub Bot
Created on: 17/Mar/20 00:49
Start Date: 17/Mar/20 00:49
Worklog Time Spent: 10m 
  Work Description: davidyan74 commented on issue #11142: [BEAM-9510] 
Fixing version incompatibilities in base_image_requirements.txt
URL: https://github.com/apache/beam/pull/11142#issuecomment-599822693
 
 
   R: @aaltay 
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 404371)
Time Spent: 20m  (was: 10m)

> Dependencies in base_image_requirements.txt are not compatible with each other
> --
>
> Key: BEAM-9510
> URL: https://issues.apache.org/jira/browse/BEAM-9510
> Project: Beam
>  Issue Type: Bug
>  Components: sdk-py-harness
>Reporter: David Yan
>Assignee: Hannah Jiang
>Priority: Major
>  Time Spent: 20m
>  Remaining Estimate: 0h
>
> [https://github.com/apache/beam/blob/master/sdks/python/container/base_image_requirements.txt#L56]
> says it requires google-cloud-bigquery==1.24.0, google-cloud-core==1.0.2, 
> google-cloud-bigtable==0.32.1, grpc-1.22.0 and tensorflow-2.1.0
> But they are incompatible with each other:
> ERROR: google-cloud-bigquery 1.24.0 has requirement 
> google-cloud-core<2.0dev,>=1.1.0, but you'll have google-cloud-core 1.0.2 
> which is incompatible.
> ERROR: google-cloud-bigtable 0.32.1 has requirement 
> google-cloud-core<0.30dev,>=0.29.0, but you'll have google-cloud-core 1.0.2 
> which is incompatible.
> ERROR: tensorboard 2.1.1 has requirement grpcio>=1.24.3, but you'll have 
> grpcio 1.22.0 which is incompatible.
> ERROR: tensorflow 2.1.0 has requirement scipy==1.4.1; python_version >= "3", 
> but you'll have scipy 1.2.2 which is incompatible.
>  
>  
>  
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (BEAM-9510) Dependencies in base_image_requirements.txt are not compatible with each other

2020-03-16 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-9510?focusedWorklogId=404370&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-404370
 ]

ASF GitHub Bot logged work on BEAM-9510:


Author: ASF GitHub Bot
Created on: 17/Mar/20 00:46
Start Date: 17/Mar/20 00:46
Worklog Time Spent: 10m 
  Work Description: davidyan74 commented on pull request #11142: 
[BEAM-9510] Fixing version incompatibilities in base_image_requirements.txt
URL: https://github.com/apache/beam/pull/11142
 
 
   Tried this out and it doesn't have any compatibility error messages any more 
for python3.
   
   R: @tvalentyn 
   
   **Please** add a meaningful description for your change here
   
   
   
   Thank you for your contribution! Follow this checklist to help us 
incorporate your contribution quickly and easily:
   
- [ ] [**Choose 
reviewer(s)**](https://beam.apache.org/contribute/#make-your-change) and 
mention them in a comment (`R: @username`).
- [ ] Format the pull request title like `[BEAM-XXX] Fixes bug in 
ApproximateQuantiles`, where you replace `BEAM-XXX` with the appropriate JIRA 
issue, if applicable. This will automatically link the pull request to the 
issue.
- [ ] Update `CHANGES.md` with noteworthy changes.
- [ ] If this contribution is large, please file an Apache [Individual 
Contributor License Agreement](https://www.apache.org/licenses/icla.pdf).
   
   See the [Contributor Guide](https://beam.apache.org/contribute) for more 
tips on [how to make review process 
smoother](https://beam.apache.org/contribute/#make-reviewers-job-easier).
   
   Post-Commit Tests Status (on master branch)
   

   
   Lang | SDK | Apex | Dataflow | Flink | Gearpump | Samza | Spark
   --- | --- | --- | --- | --- | --- | --- | ---
   Go | [![Build 
Status](https://builds.apache.org/job/beam_PostCommit_Go/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Go/lastCompletedBuild/)
 | --- | --- | [![Build 
Status](https://builds.apache.org/job/beam_PostCommit_Go_VR_Flink/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Go_VR_Flink/lastCompletedBuild/)
 | --- | --- | [![Build 
Status](https://builds.apache.org/job/beam_PostCommit_Go_VR_Spark/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Go_VR_Spark/lastCompletedBuild/)
   Java | [![Build 
Status](https://builds.apache.org/job/beam_PostCommit_Java/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Java/lastCompletedBuild/)
 | [![Build 
Status](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Apex/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Apex/lastCompletedBuild/)
 | [![Build 
Status](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Dataflow/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Dataflow/lastCompletedBuild/)[![Build
 
Status](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Dataflow_Java11/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Dataflow_Java11/lastCompletedBuild/)
 | [![Build 
Status](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Flink/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Flink/lastCompletedBuild/)[![Build
 
Status](https://builds.apache.org/job/beam_PostCommit_Java_PVR_Flink_Batch/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Java_PVR_Flink_Batch/lastCompletedBuild/)[![Build
 
Status](https://builds.apache.org/job/beam_PostCommit_Java_PVR_Flink_Streaming/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Java_PVR_Flink_Streaming/lastCompletedBuild/)
 | [![Build 
Status](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Gearpump/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Gearpump/lastCompletedBuild/)
 | [![Build 
Status](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Samza/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Samza/lastCompletedBuild/)
 | [![Build 
Status](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/lastCompletedBuild/)[![Build
 
Status](https://builds.apache.org/job/beam_PostCommit_Java_PVR_Spark_Batch/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Java_PVR_Spark_Batch/lastCompletedBuild/)[![Build
 
Status](https://builds.apache.org/job/beam_PostComm

[jira] [Work logged] (BEAM-8057) Support NAN, INF, and -INF

2020-03-16 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-8057?focusedWorklogId=404369&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-404369
 ]

ASF GitHub Bot logged work on BEAM-8057:


Author: ASF GitHub Bot
Created on: 17/Mar/20 00:45
Start Date: 17/Mar/20 00:45
Worklog Time Spent: 10m 
  Work Description: apilloud commented on pull request #11131: [BEAM-8057] 
Reject Infinite or NaN literals at parse time
URL: https://github.com/apache/beam/pull/11131
 
 
   
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 404369)
Time Spent: 0.5h  (was: 20m)

> Support NAN, INF, and -INF
> --
>
> Key: BEAM-8057
> URL: https://issues.apache.org/jira/browse/BEAM-8057
> Project: Beam
>  Issue Type: Sub-task
>  Components: dsl-sql-zetasql
>Reporter: Rui Wang
>Priority: Major
>  Time Spent: 0.5h
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (BEAM-8070) Support empty array literal

2020-03-16 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-8070?focusedWorklogId=404368&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-404368
 ]

ASF GitHub Bot logged work on BEAM-8070:


Author: ASF GitHub Bot
Created on: 17/Mar/20 00:43
Start Date: 17/Mar/20 00:43
Worklog Time Spent: 10m 
  Work Description: apilloud commented on pull request #11130: [BEAM-8070] 
Preserve type for empty array
URL: https://github.com/apache/beam/pull/11130
 
 
   
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 404368)
Time Spent: 1h 20m  (was: 1h 10m)

> Support empty array literal
> ---
>
> Key: BEAM-8070
> URL: https://issues.apache.org/jira/browse/BEAM-8070
> Project: Beam
>  Issue Type: Sub-task
>  Components: dsl-sql-zetasql
>Reporter: Rui Wang
>Assignee: Andrew Pilloud
>Priority: Major
>  Labels: zetasql-compliance
>  Time Spent: 1h 20m
>  Remaining Estimate: 0h
>
> Currently BeamSQL throws an IndexOutOfBoundsException when given a query with 
> an empty array literal. This happens because Calcite attempts to infer the 
> element types [1,2] from an empty element list.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (BEAM-9526) Bug in top.LargestPerKey, disappearing elements.

2020-03-16 Thread Daniel Oliveira (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-9526?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Daniel Oliveira updated BEAM-9526:
--
Status: Open  (was: Triage Needed)

> Bug in top.LargestPerKey, disappearing elements.
> 
>
> Key: BEAM-9526
> URL: https://issues.apache.org/jira/browse/BEAM-9526
> Project: Beam
>  Issue Type: Bug
>  Components: sdk-go
>Reporter: Daniel Oliveira
>Assignee: Daniel Oliveira
>Priority: Minor
>
> There's a bug in top.LargestPerKey that appears when running it on a runner 
> that has combiner lifting. Since it's similar to a GBK, we should be 
> inputting a K, V and receiving a K, []V as output. When the runner has 
> combiner lifting, the []V always contains an empty slice.
>  
> I already found the fix for this, just logging this to attach it to the PR.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work started] (BEAM-9526) Bug in top.LargestPerKey, disappearing elements.

2020-03-16 Thread Daniel Oliveira (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-9526?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Work on BEAM-9526 started by Daniel Oliveira.
-
> Bug in top.LargestPerKey, disappearing elements.
> 
>
> Key: BEAM-9526
> URL: https://issues.apache.org/jira/browse/BEAM-9526
> Project: Beam
>  Issue Type: Bug
>  Components: sdk-go
>Reporter: Daniel Oliveira
>Assignee: Daniel Oliveira
>Priority: Minor
>
> There's a bug in top.LargestPerKey that appears when running it on a runner 
> that has combiner lifting. Since it's similar to a GBK, we should be 
> inputting a K, V and receiving a K, []V as output. When the runner has 
> combiner lifting, the []V always contains an empty slice.
>  
> I already found the fix for this, just logging this to attach it to the PR.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (BEAM-9526) Bug in top.LargestPerKey, disappearing elements.

2020-03-16 Thread Daniel Oliveira (Jira)
Daniel Oliveira created BEAM-9526:
-

 Summary: Bug in top.LargestPerKey, disappearing elements.
 Key: BEAM-9526
 URL: https://issues.apache.org/jira/browse/BEAM-9526
 Project: Beam
  Issue Type: Bug
  Components: sdk-go
Reporter: Daniel Oliveira
Assignee: Daniel Oliveira


There's a bug in top.LargestPerKey that appears when running it on a runner 
that has combiner lifting. Since it's similar to a GBK, we should be inputting 
a K, V and receiving a K, []V as output. When the runner has combiner lifting, 
the []V always contains an empty slice.

 

I already found the fix for this, just logging this to attach it to the PR.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (BEAM-3301) Go SplittableDoFn support

2020-03-16 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-3301?focusedWorklogId=404362&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-404362
 ]

ASF GitHub Bot logged work on BEAM-3301:


Author: ASF GitHub Bot
Created on: 17/Mar/20 00:29
Start Date: 17/Mar/20 00:29
Worklog Time Spent: 10m 
  Work Description: lostluck commented on pull request #10991: [BEAM-3301] 
Refactor DoFn validation & allow specifying main inputs.
URL: https://github.com/apache/beam/pull/10991#discussion_r393382209
 
 

 ##
 File path: sdks/go/pkg/beam/core/graph/fn.go
 ##
 @@ -239,52 +279,50 @@ func AsDoFn(fn *Fn) (*DoFn, error) {
return nil, addContext(err, fn)
}
 
-   // Start validating DoFn. First, check that ProcessElement has a main 
input.
+   // Validate ProcessElement has correct number of main inputs (as 
indicated by
+   // numMainIn), and that main inputs are before side inputs.
processFn := fn.methods[processElementName]
-   pos, num, ok := processFn.Inputs()
-   if ok {
-   first := processFn.Param[pos].Kind
-   if first != funcx.FnValue {
-   err := errors.New("side input parameters must follow 
main input parameter")
-   err = errors.SetTopLevelMsgf(err,
-   "Method %v of DoFns should always have a main 
input before side inputs, "+
-   "but it has side inputs (as Iters or 
ReIters) first in DoFn %v.",
-   processElementName, fn.Name())
-   err = errors.WithContextf(err, "method %v", 
processElementName)
-   return nil, addContext(err, fn)
-   }
+   if err := validateMainInputs(fn, processFn, processElementName, 
numMainIn); err != nil {
+   return nil, addContext(err, fn)
+   }
+
+   // If numMainIn is unknown, we can try inferring it from the second 
input in ProcessElement.
+   // If there is none, or it's not a FnValue type, then we can safely 
infer that there's only
+   // one main input.
+   pos, num, _ := processFn.Inputs()
+   if numMainIn == MainUnknown && (num == 1 || processFn.Param[pos+1].Kind 
!= funcx.FnValue) {
+   numMainIn = MainSingle
}
 
// If the ProcessElement function includes side inputs or emit 
functions those must also be
 
 Review comment:
   At most relaxed we'd be able to either not require them at all if none are 
used, or isolate them by their types. All instances of a given side input or 
emit with the same type would need to be listed at once, since otherwise we 
have no way to distinguish them except by position.  Permitting Nothing to be 
set would be the most convenient, or permitting only the Side Inputs and not 
requireing the Emits.
   
   For now though, it's better to be more strict now and relax later, since the 
inverse is impossible, and such variety is harder to maintain if unnecessary.
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 404362)
Time Spent: 4h 50m  (was: 4h 40m)

> Go SplittableDoFn support
> -
>
> Key: BEAM-3301
> URL: https://issues.apache.org/jira/browse/BEAM-3301
> Project: Beam
>  Issue Type: Improvement
>  Components: sdk-go
>Reporter: Henning Rohde
>Assignee: Daniel Oliveira
>Priority: Major
>  Time Spent: 4h 50m
>  Remaining Estimate: 0h
>
> SDFs will be the only way to add streaming and liquid sharded IO for Go.
> Design doc: https://s.apache.org/splittable-do-fn



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (BEAM-8335) Add streaming support to Interactive Beam

2020-03-16 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-8335?focusedWorklogId=404347&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-404347
 ]

ASF GitHub Bot logged work on BEAM-8335:


Author: ASF GitHub Bot
Created on: 17/Mar/20 00:14
Start Date: 17/Mar/20 00:14
Worklog Time Spent: 10m 
  Work Description: pabloem commented on issue #11138: [BEAM-8335] Refactor 
IPythonLogHandler
URL: https://github.com/apache/beam/pull/11138#issuecomment-599814751
 
 
   LGTM
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 404347)
Time Spent: 113h 10m  (was: 113h)

> Add streaming support to Interactive Beam
> -
>
> Key: BEAM-8335
> URL: https://issues.apache.org/jira/browse/BEAM-8335
> Project: Beam
>  Issue Type: Improvement
>  Components: runner-py-interactive
>Reporter: Sam Rohde
>Assignee: Sam Rohde
>Priority: Major
>  Time Spent: 113h 10m
>  Remaining Estimate: 0h
>
> This issue tracks the work items to introduce streaming support to the 
> Interactive Beam experience. This will allow users to:
>  * Write and run a streaming job in IPython
>  * Automatically cache records from unbounded sources
>  * Add a replay experience that replays all cached records to simulate the 
> original pipeline execution
>  * Add controls to play/pause/stop/step individual elements from the cached 
> records
>  * Add ability to inspect/visualize unbounded PCollections



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (BEAM-8335) Add streaming support to Interactive Beam

2020-03-16 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-8335?focusedWorklogId=404346&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-404346
 ]

ASF GitHub Bot logged work on BEAM-8335:


Author: ASF GitHub Bot
Created on: 17/Mar/20 00:14
Start Date: 17/Mar/20 00:14
Worklog Time Spent: 10m 
  Work Description: pabloem commented on pull request #11138: [BEAM-8335] 
Refactor IPythonLogHandler
URL: https://github.com/apache/beam/pull/11138
 
 
   
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 404346)
Time Spent: 113h  (was: 112h 50m)

> Add streaming support to Interactive Beam
> -
>
> Key: BEAM-8335
> URL: https://issues.apache.org/jira/browse/BEAM-8335
> Project: Beam
>  Issue Type: Improvement
>  Components: runner-py-interactive
>Reporter: Sam Rohde
>Assignee: Sam Rohde
>Priority: Major
>  Time Spent: 113h
>  Remaining Estimate: 0h
>
> This issue tracks the work items to introduce streaming support to the 
> Interactive Beam experience. This will allow users to:
>  * Write and run a streaming job in IPython
>  * Automatically cache records from unbounded sources
>  * Add a replay experience that replays all cached records to simulate the 
> original pipeline execution
>  * Add controls to play/pause/stop/step individual elements from the cached 
> records
>  * Add ability to inspect/visualize unbounded PCollections



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (BEAM-9421) AI Platform pipeline patterns

2020-03-16 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-9421?focusedWorklogId=404344&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-404344
 ]

ASF GitHub Bot logged work on BEAM-9421:


Author: ASF GitHub Bot
Created on: 17/Mar/20 00:12
Start Date: 17/Mar/20 00:12
Worklog Time Spent: 10m 
  Work Description: limingxi commented on pull request #11075: [BEAM-9421] 
Website section that describes getting predictions using AI Platform Prediciton
URL: https://github.com/apache/beam/pull/11075#discussion_r393378386
 
 

 ##
 File path: website/src/documentation/patterns/ai-platform.md
 ##
 @@ -0,0 +1,87 @@
+---
+layout: section
+title: "AI Platform integration patterns"
+section_menu: section-menu/documentation.html
+permalink: /documentation/patterns/ai-platform/
+---
+
+
+# AI Platform integration patterns
+
+This page describes common patterns in pipelines with Google AI Platform 
transforms.
+
+
+  Adapt for:
+  
+Java SDK
+Python SDK
+  
+
+
+## Getting predictions
+
+This section shows how to use a cloud-hosted machine learning model to make 
predictions about new data using Google Cloud AI Platform Prediction within 
Beam's pipeline.
+ 
+[tfx_bsl](https://github.com/tensorflow/tfx-bsl) is a library that provides 
`RunInference` Beam's PTransform. `RunInference` is a PTransform able to 
perform two types of inference. One of them can use a service endpoint. When 
using a service endpoint, the transform takes a PCollection of type 
`tf.train.Example` and, for each element, sends a request to Google Cloud AI 
Platform Prediction service. The transform produces a PCollection of type 
`PredictLog` which contains predictions.
+
+Before getting started, deploy a machine learning model to the cloud. The 
cloud service manages the infrastructure needed to handle prediction requests 
in both efficient and scalable way. Only Tensorflow models are supported. For 
more information, see [Exporting a SavedModel for 
prediction](https://cloud.google.com/ai-platform/prediction/docs/exporting-savedmodel-for-prediction).
+
+Once a machine learning model is deployed, prepare a list of instances to get 
predictions for. 
+
+Here is an example of a pipeline that reads input instances from the file, 
converts JSON objects to `tf.train.Example` objects and sends data to the 
service. The content of a file can look like this:
+
+```
+{"input": "the quick brown"}
+{"input": "la bruja le"}
+``` 
+
+The example creates `tf.train.BytesList` instances, thus it expects byte-like 
strings as input, but other data types, like `tf.train.FloatList` and 
`tf.train.Int64List`, are also supported by the transform. To send binary data, 
make sure that the name of an input ends in `_bytes`.
 
 Review comment:
   I think it would be good to list all possible input formats and output 
formats here or somewhere else for reference. And for the last sentence, do you 
mean that we need to change l74 to something like: 
   feature={name+'_bytes', value} for sending binary data to endpoint?
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 404344)
Time Spent: 3h 50m  (was: 3h 40m)

> AI Platform pipeline patterns
> -
>
> Key: BEAM-9421
> URL: https://issues.apache.org/jira/browse/BEAM-9421
> Project: Beam
>  Issue Type: Sub-task
>  Components: website
>Reporter: Kamil Wasilewski
>Assignee: Kamil Wasilewski
>Priority: Major
>  Labels: pipeline-patterns
>  Time Spent: 3h 50m
>  Remaining Estimate: 0h
>
> New pipeline patterns should be contributed to the Beam's website in order to 
> demonstrate how newly implemented Google Cloud AI PTransforms can be used in 
> pipelines.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (BEAM-9421) AI Platform pipeline patterns

2020-03-16 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-9421?focusedWorklogId=404327&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-404327
 ]

ASF GitHub Bot logged work on BEAM-9421:


Author: ASF GitHub Bot
Created on: 16/Mar/20 23:28
Start Date: 16/Mar/20 23:28
Worklog Time Spent: 10m 
  Work Description: wenchenglu commented on issue #11075: [BEAM-9421] 
Website section that describes getting predictions using AI Platform Prediciton
URL: https://github.com/apache/beam/pull/11075#issuecomment-599803048
 
 
   Thanks for addressing my comments. These changes look good to me. I will ask 
another engineer from AI Platform prediction to do another pass.
   
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 404327)
Time Spent: 3h 40m  (was: 3.5h)

> AI Platform pipeline patterns
> -
>
> Key: BEAM-9421
> URL: https://issues.apache.org/jira/browse/BEAM-9421
> Project: Beam
>  Issue Type: Sub-task
>  Components: website
>Reporter: Kamil Wasilewski
>Assignee: Kamil Wasilewski
>Priority: Major
>  Labels: pipeline-patterns
>  Time Spent: 3h 40m
>  Remaining Estimate: 0h
>
> New pipeline patterns should be contributed to the Beam's website in order to 
> demonstrate how newly implemented Google Cloud AI PTransforms can be used in 
> pipelines.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (BEAM-9421) AI Platform pipeline patterns

2020-03-16 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-9421?focusedWorklogId=404326&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-404326
 ]

ASF GitHub Bot logged work on BEAM-9421:


Author: ASF GitHub Bot
Created on: 16/Mar/20 23:26
Start Date: 16/Mar/20 23:26
Worklog Time Spent: 10m 
  Work Description: wenchenglu commented on pull request #11075: 
[BEAM-9421] Website section that describes getting predictions using AI 
Platform Prediciton
URL: https://github.com/apache/beam/pull/11075#discussion_r393365839
 
 

 ##
 File path: website/src/documentation/patterns/ai-platform.md
 ##
 @@ -0,0 +1,87 @@
+---
+layout: section
+title: "AI Platform integration patterns"
+section_menu: section-menu/documentation.html
+permalink: /documentation/patterns/ai-platform/
+---
+
+
+# AI Platform integration patterns
+
+This page describes common patterns in pipelines with Google AI Platform 
transforms.
+
+
+  Adapt for:
+  
+Java SDK
+Python SDK
+  
+
+
+## Getting predictions
+
+This section shows how to use a cloud-hosted machine learning model to make 
predictions about new data using Google Cloud AI Platform Prediction within 
Beam's pipeline.
+ 
+[tfx_bsl](https://github.com/tensorflow/tfx-bsl) is a library that provides 
`RunInference` Beam's PTransform. `RunInference` is a PTransform able to 
perform two types of inference. One of them can use a service endpoint. When 
using a service endpoint, the transform takes a PCollection of type 
`tf.train.Example` and, for each element, sends a request to Google Cloud AI 
Platform Prediction service. The transform produces a PCollection of type 
`PredictLog` which contains predictions.
+
+Before getting started, deploy a machine learning model to the cloud. The 
cloud service manages the infrastructure needed to handle prediction requests 
in both efficient and scalable way. Only Tensorflow models are supported. For 
more information, see [Exporting a SavedModel for 
prediction](https://cloud.google.com/ai-platform/prediction/docs/exporting-savedmodel-for-prediction).
 
 Review comment:
   LGTM
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 404326)
Time Spent: 3.5h  (was: 3h 20m)

> AI Platform pipeline patterns
> -
>
> Key: BEAM-9421
> URL: https://issues.apache.org/jira/browse/BEAM-9421
> Project: Beam
>  Issue Type: Sub-task
>  Components: website
>Reporter: Kamil Wasilewski
>Assignee: Kamil Wasilewski
>Priority: Major
>  Labels: pipeline-patterns
>  Time Spent: 3.5h
>  Remaining Estimate: 0h
>
> New pipeline patterns should be contributed to the Beam's website in order to 
> demonstrate how newly implemented Google Cloud AI PTransforms can be used in 
> pipelines.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (BEAM-9421) AI Platform pipeline patterns

2020-03-16 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-9421?focusedWorklogId=404325&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-404325
 ]

ASF GitHub Bot logged work on BEAM-9421:


Author: ASF GitHub Bot
Created on: 16/Mar/20 23:23
Start Date: 16/Mar/20 23:23
Worklog Time Spent: 10m 
  Work Description: aaltay commented on pull request #11075: [BEAM-9421] 
Website section that describes getting predictions using AI Platform Prediciton
URL: https://github.com/apache/beam/pull/11075#discussion_r393365169
 
 

 ##
 File path: website/src/documentation/patterns/ai-platform.md
 ##
 @@ -0,0 +1,87 @@
+---
+layout: section
+title: "AI Platform integration patterns"
+section_menu: section-menu/documentation.html
+permalink: /documentation/patterns/ai-platform/
+---
+
+
+# AI Platform integration patterns
+
+This page describes common patterns in pipelines with Google AI Platform 
transforms.
+
+
+  Adapt for:
+  
+Java SDK
+Python SDK
+  
+
+
+## Getting predictions
+
+This section shows how to use a cloud-hosted machine learning model to make 
predictions about new data using Google Cloud AI Platform Prediction within 
Beam's pipeline.
+ 
+[tfx_bsl](https://github.com/tensorflow/tfx-bsl) is a library that provides 
`RunInference` Beam's PTransform. `RunInference` is a PTransform able to 
perform two types of inference. One of them can use a service endpoint. When 
using a service endpoint, the transform takes a PCollection of type 
`tf.train.Example` and, for each element, sends a request to Google Cloud AI 
Platform Prediction service. The transform produces a PCollection of type 
`PredictLog` which contains predictions.
+
+Before getting started, deploy a machine learning model to the cloud. The 
cloud service manages the infrastructure needed to handle prediction requests 
in both efficient and scalable way. Only Tensorflow models are supported. For 
more information, see [Exporting a SavedModel for 
prediction](https://cloud.google.com/ai-platform/prediction/docs/exporting-savedmodel-for-prediction).
 
 Review comment:
   Beam does not have hooks for job start and job finish. One way to emulate 
those would be using an airflow wrapper to run some setup and cleanup tasks 
before and after a task. I do not believe we have a recommendation here but an 
example pattern might be useful (@rezarokni -- in case this will be a good 
pattern to add to the backlog.)
   
   For this version, it is probably fine to assume that the life cycle of the 
model will be managed by the user separately. For a next phase, we can try to 
build a graph similar to IOs (like BQ IO) to execute a certain setup and clean 
task in a single worker. (Although we cannot guarantee that it will not be 
retried)
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 404325)
Time Spent: 3h 20m  (was: 3h 10m)

> AI Platform pipeline patterns
> -
>
> Key: BEAM-9421
> URL: https://issues.apache.org/jira/browse/BEAM-9421
> Project: Beam
>  Issue Type: Sub-task
>  Components: website
>Reporter: Kamil Wasilewski
>Assignee: Kamil Wasilewski
>Priority: Major
>  Labels: pipeline-patterns
>  Time Spent: 3h 20m
>  Remaining Estimate: 0h
>
> New pipeline patterns should be contributed to the Beam's website in order to 
> demonstrate how newly implemented Google Cloud AI PTransforms can be used in 
> pipelines.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (BEAM-9485) Dataflow Silently drops Non implemented transform in fnapi mode.

2020-03-16 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-9485?focusedWorklogId=404318&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-404318
 ]

ASF GitHub Bot logged work on BEAM-9485:


Author: ASF GitHub Bot
Created on: 16/Mar/20 23:08
Start Date: 16/Mar/20 23:08
Worklog Time Spent: 10m 
  Work Description: angoenka commented on issue #1: 
[cherry-pick][BEAM-9485] Raise error when transform urn is not implemented
URL: https://github.com/apache/beam/pull/1#issuecomment-599797732
 
 
   already merged in 2.20 
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 404318)
Time Spent: 3h 10m  (was: 3h)

> Dataflow Silently drops Non implemented transform in fnapi mode.
> 
>
> Key: BEAM-9485
> URL: https://issues.apache.org/jira/browse/BEAM-9485
> Project: Beam
>  Issue Type: Bug
>  Components: runner-dataflow, sdk-py-core
>Reporter: Ankur Goenka
>Assignee: Ankur Goenka
>Priority: Major
> Fix For: 2.20.0
>
>  Time Spent: 3h 10m
>  Remaining Estimate: 0h
>
> We need to raise an error here 
> https://github.com/apache/beam/blob/02cb8d807314a38542c9894b19483e4333d8223b/sdks/python/apache_beam/runners/dataflow/dataflow_runner.py#L857



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (BEAM-9485) Dataflow Silently drops Non implemented transform in fnapi mode.

2020-03-16 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-9485?focusedWorklogId=404319&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-404319
 ]

ASF GitHub Bot logged work on BEAM-9485:


Author: ASF GitHub Bot
Created on: 16/Mar/20 23:08
Start Date: 16/Mar/20 23:08
Worklog Time Spent: 10m 
  Work Description: angoenka commented on pull request #1: 
[cherry-pick][BEAM-9485] Raise error when transform urn is not implemented
URL: https://github.com/apache/beam/pull/1
 
 
   
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 404319)
Time Spent: 3h 20m  (was: 3h 10m)

> Dataflow Silently drops Non implemented transform in fnapi mode.
> 
>
> Key: BEAM-9485
> URL: https://issues.apache.org/jira/browse/BEAM-9485
> Project: Beam
>  Issue Type: Bug
>  Components: runner-dataflow, sdk-py-core
>Reporter: Ankur Goenka
>Assignee: Ankur Goenka
>Priority: Major
> Fix For: 2.20.0
>
>  Time Spent: 3h 20m
>  Remaining Estimate: 0h
>
> We need to raise an error here 
> https://github.com/apache/beam/blob/02cb8d807314a38542c9894b19483e4333d8223b/sdks/python/apache_beam/runners/dataflow/dataflow_runner.py#L857



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (BEAM-9465) Reshuffle should trigger repeatedly

2020-03-16 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-9465?focusedWorklogId=404321&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-404321
 ]

ASF GitHub Bot logged work on BEAM-9465:


Author: ASF GitHub Bot
Created on: 16/Mar/20 23:08
Start Date: 16/Mar/20 23:08
Worklog Time Spent: 10m 
  Work Description: angoenka commented on pull request #11071: 
[cherry-pick][BEAM-9465] Fire repeatedly in reshuffle
URL: https://github.com/apache/beam/pull/11071
 
 
   
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 404321)
Time Spent: 3h 50m  (was: 3h 40m)

> Reshuffle should trigger repeatedly
> ---
>
> Key: BEAM-9465
> URL: https://issues.apache.org/jira/browse/BEAM-9465
> Project: Beam
>  Issue Type: Bug
>  Components: sdk-py-core
>Reporter: Ankur Goenka
>Assignee: Ankur Goenka
>Priority: Major
> Fix For: 2.20.0
>
>  Time Spent: 3h 50m
>  Remaining Estimate: 0h
>
> [https://github.com/apache/beam/blob/403a08f8b95d13e5381a22c1c032ad22c8848650/sdks/python/apache_beam/transforms/trigger.py#L516|https://www.google.com/url?q=https://github.com/apache/beam/blob/403a08f8b95d13e5381a22c1c032ad22c8848650/sdks/python/apache_beam/transforms/trigger.py%23L516&sa=D]
> should fire repeatedly 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (BEAM-9465) Reshuffle should trigger repeatedly

2020-03-16 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-9465?focusedWorklogId=404320&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-404320
 ]

ASF GitHub Bot logged work on BEAM-9465:


Author: ASF GitHub Bot
Created on: 16/Mar/20 23:08
Start Date: 16/Mar/20 23:08
Worklog Time Spent: 10m 
  Work Description: angoenka commented on issue #11071: 
[cherry-pick][BEAM-9465] Fire repeatedly in reshuffle
URL: https://github.com/apache/beam/pull/11071#issuecomment-599797804
 
 
   already merged in 2.20 
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 404320)
Time Spent: 3h 40m  (was: 3.5h)

> Reshuffle should trigger repeatedly
> ---
>
> Key: BEAM-9465
> URL: https://issues.apache.org/jira/browse/BEAM-9465
> Project: Beam
>  Issue Type: Bug
>  Components: sdk-py-core
>Reporter: Ankur Goenka
>Assignee: Ankur Goenka
>Priority: Major
> Fix For: 2.20.0
>
>  Time Spent: 3h 40m
>  Remaining Estimate: 0h
>
> [https://github.com/apache/beam/blob/403a08f8b95d13e5381a22c1c032ad22c8848650/sdks/python/apache_beam/transforms/trigger.py#L516|https://www.google.com/url?q=https://github.com/apache/beam/blob/403a08f8b95d13e5381a22c1c032ad22c8848650/sdks/python/apache_beam/transforms/trigger.py%23L516&sa=D]
> should fire repeatedly 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (BEAM-9485) Dataflow Silently drops Non implemented transform in fnapi mode.

2020-03-16 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-9485?focusedWorklogId=404317&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-404317
 ]

ASF GitHub Bot logged work on BEAM-9485:


Author: ASF GitHub Bot
Created on: 16/Mar/20 23:04
Start Date: 16/Mar/20 23:04
Worklog Time Spent: 10m 
  Work Description: angoenka commented on issue #1: 
[cherry-pick][BEAM-9485] Raise error when transform urn is not implemented
URL: https://github.com/apache/beam/pull/1#issuecomment-599796689
 
 
   Successful Python precommit 
https://builds.apache.org/job/beam_PreCommit_Python_Commit/11740/
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 404317)
Time Spent: 3h  (was: 2h 50m)

> Dataflow Silently drops Non implemented transform in fnapi mode.
> 
>
> Key: BEAM-9485
> URL: https://issues.apache.org/jira/browse/BEAM-9485
> Project: Beam
>  Issue Type: Bug
>  Components: runner-dataflow, sdk-py-core
>Reporter: Ankur Goenka
>Assignee: Ankur Goenka
>Priority: Major
> Fix For: 2.20.0
>
>  Time Spent: 3h
>  Remaining Estimate: 0h
>
> We need to raise an error here 
> https://github.com/apache/beam/blob/02cb8d807314a38542c9894b19483e4333d8223b/sdks/python/apache_beam/runners/dataflow/dataflow_runner.py#L857



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (BEAM-9496) Add a Dataframe API for Python

2020-03-16 Thread Jira


[ 
https://issues.apache.org/jira/browse/BEAM-9496?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17060524#comment-17060524
 ] 

Ismaël Mejía commented on BEAM-9496:


Awesome (y)!

> Add a Dataframe API for Python
> --
>
> Key: BEAM-9496
> URL: https://issues.apache.org/jira/browse/BEAM-9496
> Project: Beam
>  Issue Type: New Feature
>  Components: sdk-py-core
>Reporter: Robert Bradshaw
>Assignee: Robert Bradshaw
>Priority: Major
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> This is an umbrella bug for the dataframes work.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (BEAM-9496) Add a Dataframe API for Python

2020-03-16 Thread Robert Bradshaw (Jira)


[ 
https://issues.apache.org/jira/browse/BEAM-9496?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17060523#comment-17060523
 ] 

Robert Bradshaw commented on BEAM-9496:
---

Sorry. The doc should be publicly available now. 

> Add a Dataframe API for Python
> --
>
> Key: BEAM-9496
> URL: https://issues.apache.org/jira/browse/BEAM-9496
> Project: Beam
>  Issue Type: New Feature
>  Components: sdk-py-core
>Reporter: Robert Bradshaw
>Assignee: Robert Bradshaw
>Priority: Major
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> This is an umbrella bug for the dataframes work.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (BEAM-9525) KafkaIO attempting to make connection when uploading via Dataflow Runner

2020-03-16 Thread Ross Rothenstine (Jira)
Ross Rothenstine created BEAM-9525:
--

 Summary: KafkaIO attempting to make connection when uploading via 
Dataflow Runner
 Key: BEAM-9525
 URL: https://issues.apache.org/jira/browse/BEAM-9525
 Project: Beam
  Issue Type: Bug
  Components: io-java-kafka
Reporter: Ross Rothenstine


Given our Kafka is running in GCP, within a VPC that is not accessible outside 
of the VPC.

When uploading a pipeline to GCS using the Dataflow Runner via our CI/CD 
worker, it attempts to make a connection to the broker. Our broker isn't 
available outside the VPC, so this process fails. We can remedy this by running 
a pod from within the VPC to upload, but it causes an awkward step within our 
CI/CD process.

If there is a more elegant solution to this, such as KafkaIO taking in a 
ValueProvider for bootstrap servers.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (BEAM-7923) Interactive Beam

2020-03-16 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-7923?focusedWorklogId=404314&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-404314
 ]

ASF GitHub Bot logged work on BEAM-7923:


Author: ASF GitHub Bot
Created on: 16/Mar/20 22:51
Start Date: 16/Mar/20 22:51
Worklog Time Spent: 10m 
  Work Description: KevinGG commented on pull request #11141: [BEAM-7923] 
Include side effects in p.run
URL: https://github.com/apache/beam/pull/11141#discussion_r393352768
 
 

 ##
 File path: sdks/python/apache_beam/runners/interactive/pipeline_instrument.py
 ##
 @@ -418,10 +420,16 @@ def visit_transform(self, transform_node):
   tuple(ie.current_env().options.capturable_sources)):
   unbounded_source_pcolls.update(transform_node.outputs.values())
 cacheable_inputs.update(self._pin._cacheable_inputs(transform_node))
+ins, outs = self._pin._all_inputs_outputs(transform_node)
+all_inputs.update(ins)
+all_outputs.update(outs)
 
 v = InstrumentVisitor(self)
 self._pipeline.visit(v)
 
+# Every output PCollection that is never used as an input PCollection is
+# considered as a side effect of the pipeline run and should be included.
+self._extended_targets.update(all_outputs.difference(all_inputs))
 
 Review comment:
   It's not necessary. The intended behavior is not ambiguous: When the user 
uses `show`, `head`, `collect` APIs, these PCollections are excluded completely 
as the user explicitly wishes. And when the user invokes `p.run()`, all 
transforms in the pipeline should be executed as expected.
   
   This change is only to make sure that the prune logic doesn't affect the 
above intended behavior.
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 404314)
Time Spent: 2h 10m  (was: 2h)

> Interactive Beam
> 
>
> Key: BEAM-7923
> URL: https://issues.apache.org/jira/browse/BEAM-7923
> Project: Beam
>  Issue Type: New Feature
>  Components: runner-py-interactive
>Reporter: Ning Kang
>Assignee: Ning Kang
>Priority: Major
>  Time Spent: 2h 10m
>  Remaining Estimate: 0h
>
> This is the top level ticket for all efforts leveraging [interactive 
> Beam|[https://github.com/apache/beam/tree/master/sdks/python/apache_beam/runners/interactive]]
> As the development goes, blocking tickets will be added to this one.
>  
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Resolved] (BEAM-9485) Dataflow Silently drops Non implemented transform in fnapi mode.

2020-03-16 Thread Rui Wang (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-9485?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Rui Wang resolved BEAM-9485.

Resolution: Fixed

> Dataflow Silently drops Non implemented transform in fnapi mode.
> 
>
> Key: BEAM-9485
> URL: https://issues.apache.org/jira/browse/BEAM-9485
> Project: Beam
>  Issue Type: Bug
>  Components: runner-dataflow, sdk-py-core
>Reporter: Ankur Goenka
>Assignee: Ankur Goenka
>Priority: Major
> Fix For: 2.20.0
>
>  Time Spent: 2h 50m
>  Remaining Estimate: 0h
>
> We need to raise an error here 
> https://github.com/apache/beam/blob/02cb8d807314a38542c9894b19483e4333d8223b/sdks/python/apache_beam/runners/dataflow/dataflow_runner.py#L857



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Resolved] (BEAM-9465) Reshuffle should trigger repeatedly

2020-03-16 Thread Rui Wang (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-9465?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Rui Wang resolved BEAM-9465.

Resolution: Fixed

> Reshuffle should trigger repeatedly
> ---
>
> Key: BEAM-9465
> URL: https://issues.apache.org/jira/browse/BEAM-9465
> Project: Beam
>  Issue Type: Bug
>  Components: sdk-py-core
>Reporter: Ankur Goenka
>Assignee: Ankur Goenka
>Priority: Major
> Fix For: 2.20.0
>
>  Time Spent: 3.5h
>  Remaining Estimate: 0h
>
> [https://github.com/apache/beam/blob/403a08f8b95d13e5381a22c1c032ad22c8848650/sdks/python/apache_beam/transforms/trigger.py#L516|https://www.google.com/url?q=https://github.com/apache/beam/blob/403a08f8b95d13e5381a22c1c032ad22c8848650/sdks/python/apache_beam/transforms/trigger.py%23L516&sa=D]
> should fire repeatedly 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (BEAM-9503) SyntaxError in process worker startup

2020-03-16 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-9503?focusedWorklogId=404310&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-404310
 ]

ASF GitHub Bot logged work on BEAM-9503:


Author: ASF GitHub Bot
Created on: 16/Mar/20 22:45
Start Date: 16/Mar/20 22:45
Worklog Time Spent: 10m 
  Work Description: ibzib commented on issue #11123: [BEAM-9503] Insert 
missing comma in process worker script.
URL: https://github.com/apache/beam/pull/11123#issuecomment-599791452
 
 
   Run Python 3.7 PostCommit
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 404310)
Time Spent: 4h 20m  (was: 4h 10m)

> SyntaxError in process worker startup
> -
>
> Key: BEAM-9503
> URL: https://issues.apache.org/jira/browse/BEAM-9503
> Project: Beam
>  Issue Type: Bug
>  Components: sdk-py-harness
>Reporter: Kyle Weaver
>Assignee: Kyle Weaver
>Priority: Major
> Fix For: 2.20.0
>
>  Time Spent: 4h 20m
>  Remaining Estimate: 0h
>
> WARNING:apache_beam.runners.worker.worker_pool_main:Starting worker with 
> command ['python', '-c', 'from apache_beam.runners.worker.sdk_worker import 
> SdkHarness; 
> SdkHarness("localhost:57103",worker_id="1-1",state_cache_size=0data_buffer_time_limit_ms=0).run()']
> Note that 'state_cache_size=0data_buffer_time_limit_ms=0' is all mashed 
> together. Looks like we're missing a comma: 
> https://github.com/apache/beam/blob/feefaca793d8358d5386d0725863c03e4e37b5b1/sdks/python/apache_beam/runners/worker/worker_pool_main.py#L116



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (BEAM-7923) Interactive Beam

2020-03-16 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-7923?focusedWorklogId=404311&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-404311
 ]

ASF GitHub Bot logged work on BEAM-7923:


Author: ASF GitHub Bot
Created on: 16/Mar/20 22:45
Start Date: 16/Mar/20 22:45
Worklog Time Spent: 10m 
  Work Description: aaltay commented on pull request #11141: [BEAM-7923] 
Include side effects in p.run
URL: https://github.com/apache/beam/pull/11141#discussion_r393351076
 
 

 ##
 File path: sdks/python/apache_beam/runners/interactive/pipeline_instrument.py
 ##
 @@ -418,10 +420,16 @@ def visit_transform(self, transform_node):
   tuple(ie.current_env().options.capturable_sources)):
   unbounded_source_pcolls.update(transform_node.outputs.values())
 cacheable_inputs.update(self._pin._cacheable_inputs(transform_node))
+ins, outs = self._pin._all_inputs_outputs(transform_node)
+all_inputs.update(ins)
+all_outputs.update(outs)
 
 v = InstrumentVisitor(self)
 self._pipeline.visit(v)
 
+# Every output PCollection that is never used as an input PCollection is
+# considered as a side effect of the pipeline run and should be included.
+self._extended_targets.update(all_outputs.difference(all_inputs))
 
 Review comment:
   Do you want to track, mark side effects differently? Does users want to 
specifically track these pcollections?
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 404311)
Time Spent: 2h  (was: 1h 50m)

> Interactive Beam
> 
>
> Key: BEAM-7923
> URL: https://issues.apache.org/jira/browse/BEAM-7923
> Project: Beam
>  Issue Type: New Feature
>  Components: runner-py-interactive
>Reporter: Ning Kang
>Assignee: Ning Kang
>Priority: Major
>  Time Spent: 2h
>  Remaining Estimate: 0h
>
> This is the top level ticket for all efforts leveraging [interactive 
> Beam|[https://github.com/apache/beam/tree/master/sdks/python/apache_beam/runners/interactive]]
> As the development goes, blocking tickets will be added to this one.
>  
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (BEAM-9503) SyntaxError in process worker startup

2020-03-16 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-9503?focusedWorklogId=404308&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-404308
 ]

ASF GitHub Bot logged work on BEAM-9503:


Author: ASF GitHub Bot
Created on: 16/Mar/20 22:45
Start Date: 16/Mar/20 22:45
Worklog Time Spent: 10m 
  Work Description: ibzib commented on issue #11123: [BEAM-9503] Insert 
missing comma in process worker script.
URL: https://github.com/apache/beam/pull/11123#issuecomment-599791417
 
 
   Run Python 3.5 PostCommit
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 404308)
Time Spent: 4h 10m  (was: 4h)

> SyntaxError in process worker startup
> -
>
> Key: BEAM-9503
> URL: https://issues.apache.org/jira/browse/BEAM-9503
> Project: Beam
>  Issue Type: Bug
>  Components: sdk-py-harness
>Reporter: Kyle Weaver
>Assignee: Kyle Weaver
>Priority: Major
> Fix For: 2.20.0
>
>  Time Spent: 4h 10m
>  Remaining Estimate: 0h
>
> WARNING:apache_beam.runners.worker.worker_pool_main:Starting worker with 
> command ['python', '-c', 'from apache_beam.runners.worker.sdk_worker import 
> SdkHarness; 
> SdkHarness("localhost:57103",worker_id="1-1",state_cache_size=0data_buffer_time_limit_ms=0).run()']
> Note that 'state_cache_size=0data_buffer_time_limit_ms=0' is all mashed 
> together. Looks like we're missing a comma: 
> https://github.com/apache/beam/blob/feefaca793d8358d5386d0725863c03e4e37b5b1/sdks/python/apache_beam/runners/worker/worker_pool_main.py#L116



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (BEAM-7923) Interactive Beam

2020-03-16 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-7923?focusedWorklogId=404305&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-404305
 ]

ASF GitHub Bot logged work on BEAM-7923:


Author: ASF GitHub Bot
Created on: 16/Mar/20 22:35
Start Date: 16/Mar/20 22:35
Worklog Time Spent: 10m 
  Work Description: KevinGG commented on issue #11141: [BEAM-7923] Include 
side effects in p.run
URL: https://github.com/apache/beam/pull/11141#issuecomment-599788607
 
 
   Formatted with yapf.
   Lint passed locally.
   
   R: @aaltay 
   R: @davidyan74 
   R: @rohdesamuel 
   
   PTAL, thx!
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 404305)
Time Spent: 1h 50m  (was: 1h 40m)

> Interactive Beam
> 
>
> Key: BEAM-7923
> URL: https://issues.apache.org/jira/browse/BEAM-7923
> Project: Beam
>  Issue Type: New Feature
>  Components: runner-py-interactive
>Reporter: Ning Kang
>Assignee: Ning Kang
>Priority: Major
>  Time Spent: 1h 50m
>  Remaining Estimate: 0h
>
> This is the top level ticket for all efforts leveraging [interactive 
> Beam|[https://github.com/apache/beam/tree/master/sdks/python/apache_beam/runners/interactive]]
> As the development goes, blocking tickets will be added to this one.
>  
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (BEAM-7923) Interactive Beam

2020-03-16 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-7923?focusedWorklogId=404304&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-404304
 ]

ASF GitHub Bot logged work on BEAM-7923:


Author: ASF GitHub Bot
Created on: 16/Mar/20 22:33
Start Date: 16/Mar/20 22:33
Worklog Time Spent: 10m 
  Work Description: KevinGG commented on pull request #11141: [BEAM-7923] 
Include side effects in p.run
URL: https://github.com/apache/beam/pull/11141#discussion_r393347005
 
 

 ##
 File path: sdks/python/apache_beam/runners/interactive/utils_test.py
 ##
 @@ -56,7 +56,7 @@ def test_parse_windowedvalue_with_window_info(self):
 [['a', 2, int(1e6), els[0].windows, els[0].pane_info],
  ['b', 3, int(1e6), els[1].windows, els[1].pane_info]],
 columns=[0, 1, 'event_time', 'windows', 'pane_info'])
-pd.testing.assert_frame_equal(actual_df, expected_df)
+pd.testing.assert_frame_equal(actual_df, expected_df, check_like=True)
 
 Review comment:
   This test was flaky because the dataframe columns can be built in arbitrary 
orders. This option makes sure it doesn't take column positioning into 
consideration since we only care about the equivalence of data.
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 404304)
Time Spent: 1h 40m  (was: 1.5h)

> Interactive Beam
> 
>
> Key: BEAM-7923
> URL: https://issues.apache.org/jira/browse/BEAM-7923
> Project: Beam
>  Issue Type: New Feature
>  Components: runner-py-interactive
>Reporter: Ning Kang
>Assignee: Ning Kang
>Priority: Major
>  Time Spent: 1h 40m
>  Remaining Estimate: 0h
>
> This is the top level ticket for all efforts leveraging [interactive 
> Beam|[https://github.com/apache/beam/tree/master/sdks/python/apache_beam/runners/interactive]]
> As the development goes, blocking tickets will be added to this one.
>  
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (BEAM-9503) SyntaxError in process worker startup

2020-03-16 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-9503?focusedWorklogId=404301&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-404301
 ]

ASF GitHub Bot logged work on BEAM-9503:


Author: ASF GitHub Bot
Created on: 16/Mar/20 22:29
Start Date: 16/Mar/20 22:29
Worklog Time Spent: 10m 
  Work Description: ibzib commented on issue #11124: 
[cherry-pick][release-2.20.0][BEAM-9503] Insert missing comma in process worker 
script.
URL: https://github.com/apache/beam/pull/11124#issuecomment-599786492
 
 
   Run Python PreCommit
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 404301)
Time Spent: 4h  (was: 3h 50m)

> SyntaxError in process worker startup
> -
>
> Key: BEAM-9503
> URL: https://issues.apache.org/jira/browse/BEAM-9503
> Project: Beam
>  Issue Type: Bug
>  Components: sdk-py-harness
>Reporter: Kyle Weaver
>Assignee: Kyle Weaver
>Priority: Major
> Fix For: 2.20.0
>
>  Time Spent: 4h
>  Remaining Estimate: 0h
>
> WARNING:apache_beam.runners.worker.worker_pool_main:Starting worker with 
> command ['python', '-c', 'from apache_beam.runners.worker.sdk_worker import 
> SdkHarness; 
> SdkHarness("localhost:57103",worker_id="1-1",state_cache_size=0data_buffer_time_limit_ms=0).run()']
> Note that 'state_cache_size=0data_buffer_time_limit_ms=0' is all mashed 
> together. Looks like we're missing a comma: 
> https://github.com/apache/beam/blob/feefaca793d8358d5386d0725863c03e4e37b5b1/sdks/python/apache_beam/runners/worker/worker_pool_main.py#L116



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (BEAM-7923) Interactive Beam

2020-03-16 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-7923?focusedWorklogId=404302&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-404302
 ]

ASF GitHub Bot logged work on BEAM-7923:


Author: ASF GitHub Bot
Created on: 16/Mar/20 22:29
Start Date: 16/Mar/20 22:29
Worklog Time Spent: 10m 
  Work Description: KevinGG commented on pull request #11141: [BEAM-7923] 
Include side effects in p.run
URL: https://github.com/apache/beam/pull/11141
 
 
   1. PCollections never used as inputs and not watched, such as sinks without 
being assigned
   to variables will be pruned before `p.run()`. The change makes sure that
   these side effect PCollections are now considered as extended targets
   and will be executed on `p.run()`.
   2. Note the change will not affect `show`, `head` and `collect` because
   they have an additional pipeline fragment logic that already prunes
   everything unrelated before the instrumenting and the prune logic inside
   instrumenting.
   
   **Please** add a meaningful description for your change here
   
   
   
   Thank you for your contribution! Follow this checklist to help us 
incorporate your contribution quickly and easily:
   
- [ ] [**Choose 
reviewer(s)**](https://beam.apache.org/contribute/#make-your-change) and 
mention them in a comment (`R: @username`).
- [ ] Format the pull request title like `[BEAM-XXX] Fixes bug in 
ApproximateQuantiles`, where you replace `BEAM-XXX` with the appropriate JIRA 
issue, if applicable. This will automatically link the pull request to the 
issue.
- [ ] Update `CHANGES.md` with noteworthy changes.
- [ ] If this contribution is large, please file an Apache [Individual 
Contributor License Agreement](https://www.apache.org/licenses/icla.pdf).
   
   See the [Contributor Guide](https://beam.apache.org/contribute) for more 
tips on [how to make review process 
smoother](https://beam.apache.org/contribute/#make-reviewers-job-easier).
   
   Post-Commit Tests Status (on master branch)
   

   
   Lang | SDK | Apex | Dataflow | Flink | Gearpump | Samza | Spark
   --- | --- | --- | --- | --- | --- | --- | ---
   Go | [![Build 
Status](https://builds.apache.org/job/beam_PostCommit_Go/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Go/lastCompletedBuild/)
 | --- | --- | [![Build 
Status](https://builds.apache.org/job/beam_PostCommit_Go_VR_Flink/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Go_VR_Flink/lastCompletedBuild/)
 | --- | --- | [![Build 
Status](https://builds.apache.org/job/beam_PostCommit_Go_VR_Spark/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Go_VR_Spark/lastCompletedBuild/)
   Java | [![Build 
Status](https://builds.apache.org/job/beam_PostCommit_Java/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Java/lastCompletedBuild/)
 | [![Build 
Status](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Apex/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Apex/lastCompletedBuild/)
 | [![Build 
Status](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Dataflow/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Dataflow/lastCompletedBuild/)[![Build
 
Status](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Dataflow_Java11/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Dataflow_Java11/lastCompletedBuild/)
 | [![Build 
Status](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Flink/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Flink/lastCompletedBuild/)[![Build
 
Status](https://builds.apache.org/job/beam_PostCommit_Java_PVR_Flink_Batch/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Java_PVR_Flink_Batch/lastCompletedBuild/)[![Build
 
Status](https://builds.apache.org/job/beam_PostCommit_Java_PVR_Flink_Streaming/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Java_PVR_Flink_Streaming/lastCompletedBuild/)
 | [![Build 
Status](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Gearpump/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Gearpump/lastCompletedBuild/)
 | [![Build 
Status](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Samza/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Samza/lastCompletedBuild/)
 | [![Build 
Status](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/lastCompletedBuild/badg

[jira] [Work logged] (BEAM-9511) ArrayScanToUncollectConverter: ResolvedParameter cannot be cast to ResolvedLiteral

2020-03-16 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-9511?focusedWorklogId=404299&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-404299
 ]

ASF GitHub Bot logged work on BEAM-9511:


Author: ASF GitHub Bot
Created on: 16/Mar/20 22:24
Start Date: 16/Mar/20 22:24
Worklog Time Spent: 10m 
  Work Description: robinyqiu commented on issue #11133: 
[BEAM-9511][BEAM-9515][BEAM-9516] Uncollect takes arbitrary expressions
URL: https://github.com/apache/beam/pull/11133#issuecomment-599784806
 
 
   Would be great to add the query in 
https://issues.apache.org/jira/browse/BEAM-9511 as a unit test.
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 404299)
Remaining Estimate: 0h
Time Spent: 10m

> ArrayScanToUncollectConverter: ResolvedParameter cannot be cast to 
> ResolvedLiteral
> --
>
> Key: BEAM-9511
> URL: https://issues.apache.org/jira/browse/BEAM-9511
> Project: Beam
>  Issue Type: Bug
>  Components: dsl-sql-zetasql
>Reporter: Andrew Pilloud
>Assignee: Andrew Pilloud
>Priority: Major
>  Labels: zetasql-compliance
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
>  
> {code:java}
> Mar 16, 2020 12:57:42 PM 
> cloud.dataflow.sql.ExecuteQueryServiceServer$SqlComplianceServiceImpl 
> executeQuery
> INFO: Processing Sql statement: SELECT elem FROM UNNEST(@string_arr) AS elem
> Mar 16, 2020 12:57:42 PM 
> com.google.zetasql.io.grpc.internal.SerializingExecutor run
> SEVERE: Exception while executing runnable 
> com.google.zetasql.io.grpc.internal.ServerImpl$JumpToApplicationThreadServerStreamListener$1HalfClosed@2f5b49be
> java.lang.ClassCastException: 
> com.google.zetasql.resolvedast.ResolvedNodes$ResolvedParameter cannot be cast 
> to com.google.zetasql.resolvedast.ResolvedNodes$ResolvedLiteral
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.ArrayScanToUncollectConverter.convert(ArrayScanToUncollectConverter.java:45)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.ArrayScanToUncollectConverter.convert(ArrayScanToUncollectConverter.java:31)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.QueryStatementConverter.convertNode(QueryStatementConverter.java:97)
>   at 
> java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:193)
>   at java.util.Collections$2.tryAdvance(Collections.java:4717)
>   at java.util.Collections$2.forEachRemaining(Collections.java:4725)
>   at java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:481)
>   at 
> java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:471)
>   at 
> java.util.stream.ReduceOps$ReduceOp.evaluateSequential(ReduceOps.java:708)
>   at java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:234)
>   at 
> java.util.stream.ReferencePipeline.collect(ReferencePipeline.java:499)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.QueryStatementConverter.convertNode(QueryStatementConverter.java:96)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.QueryStatementConverter.convert(QueryStatementConverter.java:84)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.QueryStatementConverter.convertRootQuery(QueryStatementConverter.java:51)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLPlannerImpl.rel(ZetaSQLPlannerImpl.java:160)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLQueryPlanner.convertToBeamRelInternal(ZetaSQLQueryPlanner.java:131)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLQueryPlanner.convertToBeamRel(ZetaSQLQueryPlanner.java:115)
>   at 
> cloud.dataflow.sql.ExecuteQueryServiceServer$SqlComplianceServiceImpl.executeQuery(ExecuteQueryServiceServer.java:242)
>   at 
> com.google.zetasql.testing.SqlComplianceServiceGrpc$MethodHandlers.invoke(SqlComplianceServiceGrpc.java:423)
>   at 
> com.google.zetasql.io.grpc.stub.ServerCalls$UnaryServerCallHandler$UnaryServerCallListener.onHalfClose(ServerCalls.java:171)
>   at 
> com.google.zetasql.io.grpc.internal.ServerCallImpl$ServerStreamListenerImpl.halfClosed(ServerCallImpl.java:283)
>   at 
> com.google.zetasql.io.grpc.internal.ServerImpl$JumpToApplicationThreadServerStreamListener$1HalfClosed.runInContext(ServerImpl.java:711)
>   at 
> com.google.zetasql.io.grpc.internal.ContextRunnable.

[jira] [Work logged] (BEAM-8866) portableWordCount Flink/Spark - flaky post commits

2020-03-16 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-8866?focusedWorklogId=404297&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-404297
 ]

ASF GitHub Bot logged work on BEAM-8866:


Author: ASF GitHub Bot
Created on: 16/Mar/20 22:23
Start Date: 16/Mar/20 22:23
Worklog Time Spent: 10m 
  Work Description: ibzib commented on issue #11140: [BEAM-8866] Use unique 
temp dir for output of portable word count tests.
URL: https://github.com/apache/beam/pull/11140#issuecomment-599784705
 
 
   Run Python 3.7 PostCommit
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 404297)
Time Spent: 20m  (was: 10m)

> portableWordCount Flink/Spark - flaky post commits
> --
>
> Key: BEAM-8866
> URL: https://issues.apache.org/jira/browse/BEAM-8866
> Project: Beam
>  Issue Type: Sub-task
>  Components: runner-flink
>Reporter: Ahmet Altay
>Assignee: Kyle Weaver
>Priority: Critical
>  Labels: portability-flink
>  Time Spent: 20m
>  Remaining Estimate: 0h
>
> Logs: 
> [https://scans.gradle.com/s/rkdiftvzvr7cy/console-log?task=:sdks:python:test-suites:portable:py36:portableWordCountFlinkRunnerStreaming]
> Error:
> ..
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python36/src/sdks/python/apache_beam/io/localfilesystem.py",
>  line 335, in delete   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python36/src/sdks/python/apache_beam/io/localfilesystem.py",
>  line 335, in delete     raise BeamIOError("Delete operation failed", 
> exceptions) apache_beam.io.filesystem.BeamIOError: Delete operation failed 
> with exceptions \{'/tmp/py-wordcount-direct-1-of-2': OSError('No 
> files found to delete under: /tmp/py-wordcount-direct-1-of-2',), 
> '/tmp/py-wordcount-direct-0-of-2': OSError('No files found to delete 
> under: /tmp/py-wordcount-direct-0-of-2',)} During handling of the 
> above exception, another exception occurred:
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (BEAM-9510) Dependencies in base_image_requirements.txt are not compatible with each other

2020-03-16 Thread Valentyn Tymofieiev (Jira)


[ 
https://issues.apache.org/jira/browse/BEAM-9510?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17060503#comment-17060503
 ] 

Valentyn Tymofieiev commented on BEAM-9510:
---

Related: https://issues.apache.org/jira/browse/BEAM-8551

> Dependencies in base_image_requirements.txt are not compatible with each other
> --
>
> Key: BEAM-9510
> URL: https://issues.apache.org/jira/browse/BEAM-9510
> Project: Beam
>  Issue Type: Bug
>  Components: sdk-py-harness
>Reporter: David Yan
>Assignee: Hannah Jiang
>Priority: Major
>
> [https://github.com/apache/beam/blob/master/sdks/python/container/base_image_requirements.txt#L56]
> says it requires google-cloud-bigquery==1.24.0, google-cloud-core==1.0.2, 
> google-cloud-bigtable==0.32.1, grpc-1.22.0 and tensorflow-2.1.0
> But they are incompatible with each other:
> ERROR: google-cloud-bigquery 1.24.0 has requirement 
> google-cloud-core<2.0dev,>=1.1.0, but you'll have google-cloud-core 1.0.2 
> which is incompatible.
> ERROR: google-cloud-bigtable 0.32.1 has requirement 
> google-cloud-core<0.30dev,>=0.29.0, but you'll have google-cloud-core 1.0.2 
> which is incompatible.
> ERROR: tensorboard 2.1.1 has requirement grpcio>=1.24.3, but you'll have 
> grpcio 1.22.0 which is incompatible.
> ERROR: tensorflow 2.1.0 has requirement scipy==1.4.1; python_version >= "3", 
> but you'll have scipy 1.2.2 which is incompatible.
>  
>  
>  
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (BEAM-8551) Beam Python containers should include all Beam SDK dependencies, and do not have conflicting dependencies

2020-03-16 Thread Valentyn Tymofieiev (Jira)


[ 
https://issues.apache.org/jira/browse/BEAM-8551?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17060501#comment-17060501
 ] 

Valentyn Tymofieiev commented on BEAM-8551:
---

cc: [~hannahjiang]

> Beam Python containers should include all Beam SDK dependencies, and do not 
> have conflicting dependencies
> -
>
> Key: BEAM-8551
> URL: https://issues.apache.org/jira/browse/BEAM-8551
> Project: Beam
>  Issue Type: Bug
>  Components: sdk-py-core
>Reporter: Valentyn Tymofieiev
>Priority: Major
>
> Checks could be introduced during container creation, and be enforced by 
> ValidatesContainer test suites. We could:
> - Check pip output or status code for incompatible dependency errors.
> - Remove internet access when installing apache-beam in the container, to 
> makes sure all dependencies are installed.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (BEAM-9524) ib.show() spins forever when cells are re-executed

2020-03-16 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-9524?focusedWorklogId=404290&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-404290
 ]

ASF GitHub Bot logged work on BEAM-9524:


Author: ASF GitHub Bot
Created on: 16/Mar/20 22:12
Start Date: 16/Mar/20 22:12
Worklog Time Spent: 10m 
  Work Description: rohdesamuel commented on issue #11128: [BEAM-9524] Fix 
for ib.show() executing indefinitely
URL: https://github.com/apache/beam/pull/11128#issuecomment-599781031
 
 
   R: @pabloem 
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 404290)
Remaining Estimate: 0h
Time Spent: 10m

> ib.show() spins forever when cells are re-executed
> --
>
> Key: BEAM-9524
> URL: https://issues.apache.org/jira/browse/BEAM-9524
> Project: Beam
>  Issue Type: Bug
>  Components: runner-py-interactive
>Reporter: Sam Rohde
>Priority: Major
>  Time Spent: 10m
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (BEAM-8070) Support empty array literal

2020-03-16 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-8070?focusedWorklogId=404278&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-404278
 ]

ASF GitHub Bot logged work on BEAM-8070:


Author: ASF GitHub Bot
Created on: 16/Mar/20 22:02
Start Date: 16/Mar/20 22:02
Worklog Time Spent: 10m 
  Work Description: robinyqiu commented on issue #11130: [BEAM-8070] 
Preserve type for empty array
URL: https://github.com/apache/beam/pull/11130#issuecomment-599777966
 
 
   LGTM.
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 404278)
Time Spent: 1h 10m  (was: 1h)

> Support empty array literal
> ---
>
> Key: BEAM-8070
> URL: https://issues.apache.org/jira/browse/BEAM-8070
> Project: Beam
>  Issue Type: Sub-task
>  Components: dsl-sql-zetasql
>Reporter: Rui Wang
>Assignee: Andrew Pilloud
>Priority: Major
>  Labels: zetasql-compliance
>  Time Spent: 1h 10m
>  Remaining Estimate: 0h
>
> Currently BeamSQL throws an IndexOutOfBoundsException when given a query with 
> an empty array literal. This happens because Calcite attempts to infer the 
> element types [1,2] from an empty element list.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (BEAM-8070) Support empty array literal

2020-03-16 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-8070?focusedWorklogId=404277&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-404277
 ]

ASF GitHub Bot logged work on BEAM-8070:


Author: ASF GitHub Bot
Created on: 16/Mar/20 22:01
Start Date: 16/Mar/20 22:01
Worklog Time Spent: 10m 
  Work Description: apilloud commented on pull request #11130: [BEAM-8070] 
Preserve type for empty array
URL: https://github.com/apache/beam/pull/11130#discussion_r393334075
 
 

 ##
 File path: 
sdks/java/extensions/sql/zetasql/src/test/java/org/apache/beam/sdk/extensions/sql/zetasql/ZetaSQLDialectSpecTest.java
 ##
 @@ -658,7 +658,6 @@ public void testIfNullNegative() {
   }
 
   @Test
-  @Ignore("Throws IndexOutOfBoundsException")
   public void testConstructEmptyArrayLiteral() {
 
 Review comment:
   Done.
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 404277)
Time Spent: 1h  (was: 50m)

> Support empty array literal
> ---
>
> Key: BEAM-8070
> URL: https://issues.apache.org/jira/browse/BEAM-8070
> Project: Beam
>  Issue Type: Sub-task
>  Components: dsl-sql-zetasql
>Reporter: Rui Wang
>Assignee: Andrew Pilloud
>Priority: Major
>  Labels: zetasql-compliance
>  Time Spent: 1h
>  Remaining Estimate: 0h
>
> Currently BeamSQL throws an IndexOutOfBoundsException when given a query with 
> an empty array literal. This happens because Calcite attempts to infer the 
> element types [1,2] from an empty element list.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (BEAM-9524) ib.show() spins forever when cells are re-executed

2020-03-16 Thread Sam Rohde (Jira)
Sam Rohde created BEAM-9524:
---

 Summary: ib.show() spins forever when cells are re-executed
 Key: BEAM-9524
 URL: https://issues.apache.org/jira/browse/BEAM-9524
 Project: Beam
  Issue Type: Bug
  Components: runner-py-interactive
Reporter: Sam Rohde






--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (BEAM-9496) Add a Dataframe API for Python

2020-03-16 Thread Jira


[ 
https://issues.apache.org/jira/browse/BEAM-9496?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17060491#comment-17060491
 ] 

Ismaël Mejía commented on BEAM-9496:


Great, seems the doc is not public yet.
cc: [~chadrik] who I suppose could be interested on this and the typings for it 
:)

> Add a Dataframe API for Python
> --
>
> Key: BEAM-9496
> URL: https://issues.apache.org/jira/browse/BEAM-9496
> Project: Beam
>  Issue Type: New Feature
>  Components: sdk-py-core
>Reporter: Robert Bradshaw
>Assignee: Robert Bradshaw
>Priority: Major
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> This is an umbrella bug for the dataframes work.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (BEAM-8070) Support empty array literal

2020-03-16 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-8070?focusedWorklogId=404275&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-404275
 ]

ASF GitHub Bot logged work on BEAM-8070:


Author: ASF GitHub Bot
Created on: 16/Mar/20 21:54
Start Date: 16/Mar/20 21:54
Worklog Time Spent: 10m 
  Work Description: robinyqiu commented on pull request #11130: [BEAM-8070] 
Preserve type for empty array
URL: https://github.com/apache/beam/pull/11130#discussion_r393330866
 
 

 ##
 File path: 
sdks/java/extensions/sql/zetasql/src/test/java/org/apache/beam/sdk/extensions/sql/zetasql/ZetaSQLDialectSpecTest.java
 ##
 @@ -658,7 +658,6 @@ public void testIfNullNegative() {
   }
 
   @Test
-  @Ignore("Throws IndexOutOfBoundsException")
   public void testConstructEmptyArrayLiteral() {
 
 Review comment:
   Since a similar new test is added, can we rename this one to 
`testEmptyArrayParameter`? That makes more sense to me.
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 404275)
Time Spent: 50m  (was: 40m)

> Support empty array literal
> ---
>
> Key: BEAM-8070
> URL: https://issues.apache.org/jira/browse/BEAM-8070
> Project: Beam
>  Issue Type: Sub-task
>  Components: dsl-sql-zetasql
>Reporter: Rui Wang
>Assignee: Andrew Pilloud
>Priority: Major
>  Labels: zetasql-compliance
>  Time Spent: 50m
>  Remaining Estimate: 0h
>
> Currently BeamSQL throws an IndexOutOfBoundsException when given a query with 
> an empty array literal. This happens because Calcite attempts to infer the 
> element types [1,2] from an empty element list.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (BEAM-8070) Support empty array literal

2020-03-16 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-8070?focusedWorklogId=404269&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-404269
 ]

ASF GitHub Bot logged work on BEAM-8070:


Author: ASF GitHub Bot
Created on: 16/Mar/20 21:35
Start Date: 16/Mar/20 21:35
Worklog Time Spent: 10m 
  Work Description: apilloud commented on issue #11130: [BEAM-8070] 
Preserve type for empty array
URL: https://github.com/apache/beam/pull/11130#issuecomment-599768814
 
 
   Done.
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 404269)
Time Spent: 40m  (was: 0.5h)

> Support empty array literal
> ---
>
> Key: BEAM-8070
> URL: https://issues.apache.org/jira/browse/BEAM-8070
> Project: Beam
>  Issue Type: Sub-task
>  Components: dsl-sql-zetasql
>Reporter: Rui Wang
>Assignee: Andrew Pilloud
>Priority: Major
>  Labels: zetasql-compliance
>  Time Spent: 40m
>  Remaining Estimate: 0h
>
> Currently BeamSQL throws an IndexOutOfBoundsException when given a query with 
> an empty array literal. This happens because Calcite attempts to infer the 
> element types [1,2] from an empty element list.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (BEAM-2572) Implement an S3 filesystem for Python SDK

2020-03-16 Thread Badrul Chowdhury (Jira)


[ 
https://issues.apache.org/jira/browse/BEAM-2572?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17060475#comment-17060475
 ] 

Badrul Chowdhury commented on BEAM-2572:


[~pabloem], the table for built-in IO transforms needs to be updated: 
[https://beam.apache.org/documentation/io/built-in/]

> Implement an S3 filesystem for Python SDK
> -
>
> Key: BEAM-2572
> URL: https://issues.apache.org/jira/browse/BEAM-2572
> Project: Beam
>  Issue Type: Task
>  Components: sdk-py-core
>Reporter: Dmitry Demeshchuk
>Priority: Minor
>  Labels: GSoC2019, gsoc, gsoc2019, mentor, outreachy19dec
> Fix For: 2.19.0
>
>  Time Spent: 5h 10m
>  Remaining Estimate: 0h
>
> There are two paths worth exploring, to my understanding:
> 1. Sticking to the HDFS-based approach (like it's done in Java).
> 2. Using boto/boto3 for accessing S3 through its common API endpoints.
> I personally prefer the second approach, for a few reasons:
> 1. In real life, HDFS and S3 have different consistency guarantees, therefore 
> their behaviors may contradict each other in some edge cases (say, we write 
> something to S3, but it's not immediately accessible for reading from another 
> end).
> 2. There are other AWS-based sources and sinks we may want to create in the 
> future: DynamoDB, Kinesis, SQS, etc.
> 3. boto3 already provides somewhat good logic for basic things like 
> reattempting.
> Whatever path we choose, there's another problem related to this: we 
> currently cannot pass any global settings (say, pipeline options, or just an 
> arbitrary kwarg) to a filesystem. Because of that, we'd have to setup the 
> runner nodes to have AWS keys set up in the environment, which is not trivial 
> to achieve and doesn't look too clean either (I'd rather see one single place 
> for configuring the runner options).
> Also, it's worth mentioning that I already have a janky S3 filesystem 
> implementation that only supports DirectRunner at the moment (because of the 
> previous paragraph). I'm perfectly fine finishing it myself, with some 
> guidance from the maintainers.
> Where should I move on from here, and whose input should I be looking for?
> Thanks!



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (BEAM-9468) Add Google Cloud Healthcare API IO Connectors

2020-03-16 Thread Jira


 [ 
https://issues.apache.org/jira/browse/BEAM-9468?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ismaël Mejía updated BEAM-9468:
---
Status: Open  (was: Triage Needed)

> Add Google Cloud Healthcare API IO Connectors
> -
>
> Key: BEAM-9468
> URL: https://issues.apache.org/jira/browse/BEAM-9468
> Project: Beam
>  Issue Type: New Feature
>  Components: io-java-gcp
>Reporter: Jacob Ferriero
>Assignee: Jacob Ferriero
>Priority: Minor
>  Time Spent: 1h 20m
>  Remaining Estimate: 0h
>
> Add IO Transforms for the HL7v2, FHIR and DICOM stores in the [Google Cloud 
> Healthcare API|https://cloud.google.com/healthcare/docs/]
> HL7v2IO
> FHIRIO
> DICOM 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (BEAM-9506) _CustomBigQuerySource value provider parameter gcs_location is wrongly evaluated

2020-03-16 Thread Jira


 [ 
https://issues.apache.org/jira/browse/BEAM-9506?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ismaël Mejía updated BEAM-9506:
---
Status: Open  (was: Triage Needed)

> _CustomBigQuerySource value provider parameter gcs_location is wrongly 
> evaluated
> 
>
> Key: BEAM-9506
> URL: https://issues.apache.org/jira/browse/BEAM-9506
> Project: Beam
>  Issue Type: Bug
>  Components: io-py-gcp
>Affects Versions: 2.19.0
>Reporter: Elias Djurfeldt
>Assignee: Kamil Wasilewski
>Priority: Minor
>
> The `gcs_location` parameter is wrongly evaluated at pipeline construction 
> time, resulting in calling value_provider.get() from a non-runtime context 
> when using a value provider for the `gcs_location`. 
> See discussion at 
> [https://github.com/apache/beam/pull/11040#issuecomment-597872563]
> The code in question is at: 
> [https://github.com/apache/beam/blob/master/sdks/python/apache_beam/io/gcp/bigquery.py#L1575]



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Resolved] (BEAM-9472) Remove excessive logging in python fn_api_runner

2020-03-16 Thread Jira


 [ 
https://issues.apache.org/jira/browse/BEAM-9472?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ismaël Mejía resolved BEAM-9472.

Resolution: Fixed

> Remove excessive logging in python fn_api_runner
> 
>
> Key: BEAM-9472
> URL: https://issues.apache.org/jira/browse/BEAM-9472
> Project: Beam
>  Issue Type: Task
>  Components: sdk-py-harness
>Reporter: Rui Wang
>Assignee: Robert Bradshaw
>Priority: Major
> Fix For: 2.20.0
>
>  Time Spent: 50m
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (BEAM-9510) Dependencies in base_image_requirements.txt are not compatible with each other

2020-03-16 Thread Jira


 [ 
https://issues.apache.org/jira/browse/BEAM-9510?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ismaël Mejía updated BEAM-9510:
---
Status: Open  (was: Triage Needed)

> Dependencies in base_image_requirements.txt are not compatible with each other
> --
>
> Key: BEAM-9510
> URL: https://issues.apache.org/jira/browse/BEAM-9510
> Project: Beam
>  Issue Type: Bug
>  Components: sdk-py-harness
>Reporter: David Yan
>Assignee: Hannah Jiang
>Priority: Major
>
> [https://github.com/apache/beam/blob/master/sdks/python/container/base_image_requirements.txt#L56]
> says it requires google-cloud-bigquery==1.24.0, google-cloud-core==1.0.2, 
> google-cloud-bigtable==0.32.1, grpc-1.22.0 and tensorflow-2.1.0
> But they are incompatible with each other:
> ERROR: google-cloud-bigquery 1.24.0 has requirement 
> google-cloud-core<2.0dev,>=1.1.0, but you'll have google-cloud-core 1.0.2 
> which is incompatible.
> ERROR: google-cloud-bigtable 0.32.1 has requirement 
> google-cloud-core<0.30dev,>=0.29.0, but you'll have google-cloud-core 1.0.2 
> which is incompatible.
> ERROR: tensorboard 2.1.1 has requirement grpcio>=1.24.3, but you'll have 
> grpcio 1.22.0 which is incompatible.
> ERROR: tensorflow 2.1.0 has requirement scipy==1.4.1; python_version >= "3", 
> but you'll have scipy 1.2.2 which is incompatible.
>  
>  
>  
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (BEAM-8866) portableWordCount Flink/Spark - flaky post commits

2020-03-16 Thread Kyle Weaver (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-8866?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kyle Weaver updated BEAM-8866:
--
Labels: portability-flink  (was: )

> portableWordCount Flink/Spark - flaky post commits
> --
>
> Key: BEAM-8866
> URL: https://issues.apache.org/jira/browse/BEAM-8866
> Project: Beam
>  Issue Type: Sub-task
>  Components: runner-flink
>Reporter: Ahmet Altay
>Assignee: Kyle Weaver
>Priority: Critical
>  Labels: portability-flink
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> Logs: 
> [https://scans.gradle.com/s/rkdiftvzvr7cy/console-log?task=:sdks:python:test-suites:portable:py36:portableWordCountFlinkRunnerStreaming]
> Error:
> ..
>   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python36/src/sdks/python/apache_beam/io/localfilesystem.py",
>  line 335, in delete   File 
> "/home/jenkins/jenkins-slave/workspace/beam_PostCommit_Python36/src/sdks/python/apache_beam/io/localfilesystem.py",
>  line 335, in delete     raise BeamIOError("Delete operation failed", 
> exceptions) apache_beam.io.filesystem.BeamIOError: Delete operation failed 
> with exceptions \{'/tmp/py-wordcount-direct-1-of-2': OSError('No 
> files found to delete under: /tmp/py-wordcount-direct-1-of-2',), 
> '/tmp/py-wordcount-direct-0-of-2': OSError('No files found to delete 
> under: /tmp/py-wordcount-direct-0-of-2',)} During handling of the 
> above exception, another exception occurred:
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (BEAM-9472) Remove excessive logging in python fn_api_runner

2020-03-16 Thread Jira


 [ 
https://issues.apache.org/jira/browse/BEAM-9472?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ismaël Mejía updated BEAM-9472:
---
Status: Open  (was: Triage Needed)

> Remove excessive logging in python fn_api_runner
> 
>
> Key: BEAM-9472
> URL: https://issues.apache.org/jira/browse/BEAM-9472
> Project: Beam
>  Issue Type: Task
>  Components: sdk-py-harness
>Reporter: Rui Wang
>Assignee: Robert Bradshaw
>Priority: Major
> Fix For: 2.20.0
>
>  Time Spent: 50m
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (BEAM-8070) Support empty array literal

2020-03-16 Thread Andrew Pilloud (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-8070?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrew Pilloud updated BEAM-8070:
-
Labels: zetasql-compliance  (was: )

> Support empty array literal
> ---
>
> Key: BEAM-8070
> URL: https://issues.apache.org/jira/browse/BEAM-8070
> Project: Beam
>  Issue Type: Sub-task
>  Components: dsl-sql-zetasql
>Reporter: Rui Wang
>Assignee: Andrew Pilloud
>Priority: Major
>  Labels: zetasql-compliance
>  Time Spent: 0.5h
>  Remaining Estimate: 0h
>
> Currently BeamSQL throws an IndexOutOfBoundsException when given a query with 
> an empty array literal. This happens because Calcite attempts to infer the 
> element types [1,2] from an empty element list.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (BEAM-8866) portableWordCount Flink/Spark - flaky post commits

2020-03-16 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-8866?focusedWorklogId=404257&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-404257
 ]

ASF GitHub Bot logged work on BEAM-8866:


Author: ASF GitHub Bot
Created on: 16/Mar/20 21:11
Start Date: 16/Mar/20 21:11
Worklog Time Spent: 10m 
  Work Description: ibzib commented on pull request #11140: [BEAM-8866] Use 
unique temp dir for output of portable word count tests.
URL: https://github.com/apache/beam/pull/11140
 
 
   **Please** add a meaningful description for your change here
   
   
   
   Thank you for your contribution! Follow this checklist to help us 
incorporate your contribution quickly and easily:
   
- [ ] [**Choose 
reviewer(s)**](https://beam.apache.org/contribute/#make-your-change) and 
mention them in a comment (`R: @username`).
- [ ] Format the pull request title like `[BEAM-XXX] Fixes bug in 
ApproximateQuantiles`, where you replace `BEAM-XXX` with the appropriate JIRA 
issue, if applicable. This will automatically link the pull request to the 
issue.
- [ ] Update `CHANGES.md` with noteworthy changes.
- [ ] If this contribution is large, please file an Apache [Individual 
Contributor License Agreement](https://www.apache.org/licenses/icla.pdf).
   
   See the [Contributor Guide](https://beam.apache.org/contribute) for more 
tips on [how to make review process 
smoother](https://beam.apache.org/contribute/#make-reviewers-job-easier).
   
   Post-Commit Tests Status (on master branch)
   

   
   Lang | SDK | Apex | Dataflow | Flink | Gearpump | Samza | Spark
   --- | --- | --- | --- | --- | --- | --- | ---
   Go | [![Build 
Status](https://builds.apache.org/job/beam_PostCommit_Go/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Go/lastCompletedBuild/)
 | --- | --- | [![Build 
Status](https://builds.apache.org/job/beam_PostCommit_Go_VR_Flink/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Go_VR_Flink/lastCompletedBuild/)
 | --- | --- | [![Build 
Status](https://builds.apache.org/job/beam_PostCommit_Go_VR_Spark/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Go_VR_Spark/lastCompletedBuild/)
   Java | [![Build 
Status](https://builds.apache.org/job/beam_PostCommit_Java/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Java/lastCompletedBuild/)
 | [![Build 
Status](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Apex/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Apex/lastCompletedBuild/)
 | [![Build 
Status](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Dataflow/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Dataflow/lastCompletedBuild/)[![Build
 
Status](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Dataflow_Java11/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Dataflow_Java11/lastCompletedBuild/)
 | [![Build 
Status](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Flink/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Flink/lastCompletedBuild/)[![Build
 
Status](https://builds.apache.org/job/beam_PostCommit_Java_PVR_Flink_Batch/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Java_PVR_Flink_Batch/lastCompletedBuild/)[![Build
 
Status](https://builds.apache.org/job/beam_PostCommit_Java_PVR_Flink_Streaming/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Java_PVR_Flink_Streaming/lastCompletedBuild/)
 | [![Build 
Status](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Gearpump/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Gearpump/lastCompletedBuild/)
 | [![Build 
Status](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Samza/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Samza/lastCompletedBuild/)
 | [![Build 
Status](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/lastCompletedBuild/)[![Build
 
Status](https://builds.apache.org/job/beam_PostCommit_Java_PVR_Spark_Batch/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Java_PVR_Spark_Batch/lastCompletedBuild/)[![Build
 
Status](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_SparkStructuredStreaming/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_J

[jira] [Work logged] (BEAM-9509) Subprocess job server treats missing local file as remote URL

2020-03-16 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-9509?focusedWorklogId=404256&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-404256
 ]

ASF GitHub Bot logged work on BEAM-9509:


Author: ASF GitHub Bot
Created on: 16/Mar/20 21:06
Start Date: 16/Mar/20 21:06
Worklog Time Spent: 10m 
  Work Description: ibzib commented on pull request #11139: [BEAM-9509] 
Improve error message for bad job server URL.
URL: https://github.com/apache/beam/pull/11139
 
 
   **Please** add a meaningful description for your change here
   
   
   
   Thank you for your contribution! Follow this checklist to help us 
incorporate your contribution quickly and easily:
   
- [ ] [**Choose 
reviewer(s)**](https://beam.apache.org/contribute/#make-your-change) and 
mention them in a comment (`R: @username`).
- [ ] Format the pull request title like `[BEAM-XXX] Fixes bug in 
ApproximateQuantiles`, where you replace `BEAM-XXX` with the appropriate JIRA 
issue, if applicable. This will automatically link the pull request to the 
issue.
- [ ] Update `CHANGES.md` with noteworthy changes.
- [ ] If this contribution is large, please file an Apache [Individual 
Contributor License Agreement](https://www.apache.org/licenses/icla.pdf).
   
   See the [Contributor Guide](https://beam.apache.org/contribute) for more 
tips on [how to make review process 
smoother](https://beam.apache.org/contribute/#make-reviewers-job-easier).
   
   Post-Commit Tests Status (on master branch)
   

   
   Lang | SDK | Apex | Dataflow | Flink | Gearpump | Samza | Spark
   --- | --- | --- | --- | --- | --- | --- | ---
   Go | [![Build 
Status](https://builds.apache.org/job/beam_PostCommit_Go/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Go/lastCompletedBuild/)
 | --- | --- | [![Build 
Status](https://builds.apache.org/job/beam_PostCommit_Go_VR_Flink/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Go_VR_Flink/lastCompletedBuild/)
 | --- | --- | [![Build 
Status](https://builds.apache.org/job/beam_PostCommit_Go_VR_Spark/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Go_VR_Spark/lastCompletedBuild/)
   Java | [![Build 
Status](https://builds.apache.org/job/beam_PostCommit_Java/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Java/lastCompletedBuild/)
 | [![Build 
Status](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Apex/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Apex/lastCompletedBuild/)
 | [![Build 
Status](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Dataflow/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Dataflow/lastCompletedBuild/)[![Build
 
Status](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Dataflow_Java11/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Dataflow_Java11/lastCompletedBuild/)
 | [![Build 
Status](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Flink/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Flink/lastCompletedBuild/)[![Build
 
Status](https://builds.apache.org/job/beam_PostCommit_Java_PVR_Flink_Batch/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Java_PVR_Flink_Batch/lastCompletedBuild/)[![Build
 
Status](https://builds.apache.org/job/beam_PostCommit_Java_PVR_Flink_Streaming/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Java_PVR_Flink_Streaming/lastCompletedBuild/)
 | [![Build 
Status](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Gearpump/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Gearpump/lastCompletedBuild/)
 | [![Build 
Status](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Samza/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Samza/lastCompletedBuild/)
 | [![Build 
Status](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark/lastCompletedBuild/)[![Build
 
Status](https://builds.apache.org/job/beam_PostCommit_Java_PVR_Spark_Batch/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Java_PVR_Spark_Batch/lastCompletedBuild/)[![Build
 
Status](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_SparkStructuredStreaming/lastCompletedBuild/badge/icon)](https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRu

[jira] [Assigned] (BEAM-9511) ArrayScanToUncollectConverter: ResolvedParameter cannot be cast to ResolvedLiteral

2020-03-16 Thread Andrew Pilloud (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-9511?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrew Pilloud reassigned BEAM-9511:


Assignee: Andrew Pilloud

> ArrayScanToUncollectConverter: ResolvedParameter cannot be cast to 
> ResolvedLiteral
> --
>
> Key: BEAM-9511
> URL: https://issues.apache.org/jira/browse/BEAM-9511
> Project: Beam
>  Issue Type: Bug
>  Components: dsl-sql-zetasql
>Reporter: Andrew Pilloud
>Assignee: Andrew Pilloud
>Priority: Major
>  Labels: zetasql-compliance
>
>  
> {code:java}
> Mar 16, 2020 12:57:42 PM 
> cloud.dataflow.sql.ExecuteQueryServiceServer$SqlComplianceServiceImpl 
> executeQuery
> INFO: Processing Sql statement: SELECT elem FROM UNNEST(@string_arr) AS elem
> Mar 16, 2020 12:57:42 PM 
> com.google.zetasql.io.grpc.internal.SerializingExecutor run
> SEVERE: Exception while executing runnable 
> com.google.zetasql.io.grpc.internal.ServerImpl$JumpToApplicationThreadServerStreamListener$1HalfClosed@2f5b49be
> java.lang.ClassCastException: 
> com.google.zetasql.resolvedast.ResolvedNodes$ResolvedParameter cannot be cast 
> to com.google.zetasql.resolvedast.ResolvedNodes$ResolvedLiteral
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.ArrayScanToUncollectConverter.convert(ArrayScanToUncollectConverter.java:45)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.ArrayScanToUncollectConverter.convert(ArrayScanToUncollectConverter.java:31)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.QueryStatementConverter.convertNode(QueryStatementConverter.java:97)
>   at 
> java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:193)
>   at java.util.Collections$2.tryAdvance(Collections.java:4717)
>   at java.util.Collections$2.forEachRemaining(Collections.java:4725)
>   at java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:481)
>   at 
> java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:471)
>   at 
> java.util.stream.ReduceOps$ReduceOp.evaluateSequential(ReduceOps.java:708)
>   at java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:234)
>   at 
> java.util.stream.ReferencePipeline.collect(ReferencePipeline.java:499)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.QueryStatementConverter.convertNode(QueryStatementConverter.java:96)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.QueryStatementConverter.convert(QueryStatementConverter.java:84)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.QueryStatementConverter.convertRootQuery(QueryStatementConverter.java:51)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLPlannerImpl.rel(ZetaSQLPlannerImpl.java:160)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLQueryPlanner.convertToBeamRelInternal(ZetaSQLQueryPlanner.java:131)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLQueryPlanner.convertToBeamRel(ZetaSQLQueryPlanner.java:115)
>   at 
> cloud.dataflow.sql.ExecuteQueryServiceServer$SqlComplianceServiceImpl.executeQuery(ExecuteQueryServiceServer.java:242)
>   at 
> com.google.zetasql.testing.SqlComplianceServiceGrpc$MethodHandlers.invoke(SqlComplianceServiceGrpc.java:423)
>   at 
> com.google.zetasql.io.grpc.stub.ServerCalls$UnaryServerCallHandler$UnaryServerCallListener.onHalfClose(ServerCalls.java:171)
>   at 
> com.google.zetasql.io.grpc.internal.ServerCallImpl$ServerStreamListenerImpl.halfClosed(ServerCallImpl.java:283)
>   at 
> com.google.zetasql.io.grpc.internal.ServerImpl$JumpToApplicationThreadServerStreamListener$1HalfClosed.runInContext(ServerImpl.java:711)
>   at 
> com.google.zetasql.io.grpc.internal.ContextRunnable.run(ContextRunnable.java:37)
>   at 
> com.google.zetasql.io.grpc.internal.SerializingExecutor.run(SerializingExecutor.java:123)
>   at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
>   at java.lang.Thread.run(Thread.java:748) {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Assigned] (BEAM-9516) ArrayScanToUncollectConverter: ResolvedSubqueryExpr cannot be cast to ResolvedLiteral

2020-03-16 Thread Andrew Pilloud (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-9516?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrew Pilloud reassigned BEAM-9516:


Assignee: Andrew Pilloud

> ArrayScanToUncollectConverter: ResolvedSubqueryExpr cannot be cast to 
> ResolvedLiteral
> -
>
> Key: BEAM-9516
> URL: https://issues.apache.org/jira/browse/BEAM-9516
> Project: Beam
>  Issue Type: Bug
>  Components: dsl-sql-zetasql
>Reporter: Andrew Pilloud
>Assignee: Andrew Pilloud
>Priority: Major
>  Labels: zetasql-compliance
>
> {code:java}
> Mar 16, 2020 12:58:49 PM 
> com.google.zetasql.io.grpc.internal.SerializingExecutor run
> SEVERE: Exception while executing runnable 
> com.google.zetasql.io.grpc.internal.ServerImpl$JumpToApplicationThreadServerStreamListener$1HalfClosed@7c1d4ea6
> java.lang.ClassCastException: 
> com.google.zetasql.resolvedast.ResolvedNodes$ResolvedSubqueryExpr cannot be 
> cast to com.google.zetasql.resolvedast.ResolvedNodes$ResolvedLiteral
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.ArrayScanToUncollectConverter.convert(ArrayScanToUncollectConverter.java:45)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.ArrayScanToUncollectConverter.convert(ArrayScanToUncollectConverter.java:31)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.QueryStatementConverter.convertNode(QueryStatementConverter.java:97)
>   at 
> java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:193)
>   at java.util.Collections$2.tryAdvance(Collections.java:4717)
>   at java.util.Collections$2.forEachRemaining(Collections.java:4725)
>   at java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:481)
>   at 
> java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:471)
>   at 
> java.util.stream.ReduceOps$ReduceOp.evaluateSequential(ReduceOps.java:708)
>   at java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:234)
>   at 
> java.util.stream.ReferencePipeline.collect(ReferencePipeline.java:499)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.QueryStatementConverter.convertNode(QueryStatementConverter.java:96)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.QueryStatementConverter.convert(QueryStatementConverter.java:84)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.QueryStatementConverter.convertRootQuery(QueryStatementConverter.java:51)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLPlannerImpl.rel(ZetaSQLPlannerImpl.java:160)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLQueryPlanner.convertToBeamRelInternal(ZetaSQLQueryPlanner.java:131)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLQueryPlanner.convertToBeamRel(ZetaSQLQueryPlanner.java:115)
>   at 
> cloud.dataflow.sql.ExecuteQueryServiceServer$SqlComplianceServiceImpl.executeQuery(ExecuteQueryServiceServer.java:242)
>   at 
> com.google.zetasql.testing.SqlComplianceServiceGrpc$MethodHandlers.invoke(SqlComplianceServiceGrpc.java:423)
>   at 
> com.google.zetasql.io.grpc.stub.ServerCalls$UnaryServerCallHandler$UnaryServerCallListener.onHalfClose(ServerCalls.java:171)
>   at 
> com.google.zetasql.io.grpc.internal.ServerCallImpl$ServerStreamListenerImpl.halfClosed(ServerCallImpl.java:283)
>   at 
> com.google.zetasql.io.grpc.internal.ServerImpl$JumpToApplicationThreadServerStreamListener$1HalfClosed.runInContext(ServerImpl.java:711)
>   at 
> com.google.zetasql.io.grpc.internal.ContextRunnable.run(ContextRunnable.java:37)
>   at 
> com.google.zetasql.io.grpc.internal.SerializingExecutor.run(SerializingExecutor.java:123)
>   at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
>   at java.lang.Thread.run(Thread.java:748)
>  {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Work logged] (BEAM-9503) SyntaxError in process worker startup

2020-03-16 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-9503?focusedWorklogId=404253&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-404253
 ]

ASF GitHub Bot logged work on BEAM-9503:


Author: ASF GitHub Bot
Created on: 16/Mar/20 20:57
Start Date: 16/Mar/20 20:57
Worklog Time Spent: 10m 
  Work Description: amaliujia commented on issue #11124: 
[cherry-pick][release-2.20.0][BEAM-9503] Insert missing comma in process worker 
script.
URL: https://github.com/apache/beam/pull/11124#issuecomment-599754982
 
 
   Run Python PreCommit
 

This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
---

Worklog Id: (was: 404253)
Time Spent: 3h 50m  (was: 3h 40m)

> SyntaxError in process worker startup
> -
>
> Key: BEAM-9503
> URL: https://issues.apache.org/jira/browse/BEAM-9503
> Project: Beam
>  Issue Type: Bug
>  Components: sdk-py-harness
>Reporter: Kyle Weaver
>Assignee: Kyle Weaver
>Priority: Major
> Fix For: 2.20.0
>
>  Time Spent: 3h 50m
>  Remaining Estimate: 0h
>
> WARNING:apache_beam.runners.worker.worker_pool_main:Starting worker with 
> command ['python', '-c', 'from apache_beam.runners.worker.sdk_worker import 
> SdkHarness; 
> SdkHarness("localhost:57103",worker_id="1-1",state_cache_size=0data_buffer_time_limit_ms=0).run()']
> Note that 'state_cache_size=0data_buffer_time_limit_ms=0' is all mashed 
> together. Looks like we're missing a comma: 
> https://github.com/apache/beam/blob/feefaca793d8358d5386d0725863c03e4e37b5b1/sdks/python/apache_beam/runners/worker/worker_pool_main.py#L116



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Assigned] (BEAM-9515) ResolvedSubqueryExpr can not be cast to ResolvedLiteral

2020-03-16 Thread Andrew Pilloud (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-9515?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrew Pilloud reassigned BEAM-9515:


Assignee: Andrew Pilloud

> ResolvedSubqueryExpr can not be cast to ResolvedLiteral
> ---
>
> Key: BEAM-9515
> URL: https://issues.apache.org/jira/browse/BEAM-9515
> Project: Beam
>  Issue Type: Bug
>  Components: dsl-sql-zetasql
>Reporter: Andrew Pilloud
>Assignee: Andrew Pilloud
>Priority: Major
>  Labels: zetasql-compliance
>
> {code:java}
> Mar 16, 2020 1:00:02 PM 
> cloud.dataflow.sql.ExecuteQueryServiceServer$SqlComplianceServiceImpl 
> executeQuery
> INFO: Processing Sql statement: SELECT * FROM UNNEST(ARRAY(
>   SELECT bool_val FROM AllTypesTable t
>   ORDER BY bool_val ASC
> )) x WITH OFFSET POS
> Mar 16, 2020 1:00:02 PM 
> com.google.zetasql.io.grpc.internal.SerializingExecutor run
> SEVERE: Exception while executing runnable 
> com.google.zetasql.io.grpc.internal.ServerImpl$JumpToApplicationThreadServerStreamListener$1HalfClosed@7b42f9e7
> java.lang.ClassCastException: 
> com.google.zetasql.resolvedast.ResolvedNodes$ResolvedSubqueryExpr cannot be 
> cast to com.google.zetasql.resolvedast.ResolvedNodes$ResolvedLiteral
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.ArrayScanToUncollectConverter.convert(ArrayScanToUncollectConverter.java:45)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.ArrayScanToUncollectConverter.convert(ArrayScanToUncollectConverter.java:31)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.QueryStatementConverter.convertNode(QueryStatementConverter.java:97)
>   at 
> java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:193)
>   at java.util.Collections$2.tryAdvance(Collections.java:4717)
>   at java.util.Collections$2.forEachRemaining(Collections.java:4725)
>   at java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:481)
>   at 
> java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:471)
>   at 
> java.util.stream.ReduceOps$ReduceOp.evaluateSequential(ReduceOps.java:708)
>   at java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:234)
>   at 
> java.util.stream.ReferencePipeline.collect(ReferencePipeline.java:499)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.QueryStatementConverter.convertNode(QueryStatementConverter.java:96)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.QueryStatementConverter.convert(QueryStatementConverter.java:84)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.QueryStatementConverter.convertRootQuery(QueryStatementConverter.java:51)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLPlannerImpl.rel(ZetaSQLPlannerImpl.java:160)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLQueryPlanner.convertToBeamRelInternal(ZetaSQLQueryPlanner.java:131)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLQueryPlanner.convertToBeamRel(ZetaSQLQueryPlanner.java:115)
>   at 
> cloud.dataflow.sql.ExecuteQueryServiceServer$SqlComplianceServiceImpl.executeQuery(ExecuteQueryServiceServer.java:242)
>   at 
> com.google.zetasql.testing.SqlComplianceServiceGrpc$MethodHandlers.invoke(SqlComplianceServiceGrpc.java:423)
>   at 
> com.google.zetasql.io.grpc.stub.ServerCalls$UnaryServerCallHandler$UnaryServerCallListener.onHalfClose(ServerCalls.java:171)
>   at 
> com.google.zetasql.io.grpc.internal.ServerCallImpl$ServerStreamListenerImpl.halfClosed(ServerCallImpl.java:283)
>   at 
> com.google.zetasql.io.grpc.internal.ServerImpl$JumpToApplicationThreadServerStreamListener$1HalfClosed.runInContext(ServerImpl.java:711)
>   at 
> com.google.zetasql.io.grpc.internal.ContextRunnable.run(ContextRunnable.java:37)
>   at 
> com.google.zetasql.io.grpc.internal.SerializingExecutor.run(SerializingExecutor.java:123)
>   at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
>   at java.lang.Thread.run(Thread.java:748)
>  {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Assigned] (BEAM-9513) NullPointerException in convertRexNodeFromResolvedExprWithRefScan

2020-03-16 Thread Andrew Pilloud (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-9513?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrew Pilloud reassigned BEAM-9513:


Assignee: Andrew Pilloud

> NullPointerException in convertRexNodeFromResolvedExprWithRefScan
> -
>
> Key: BEAM-9513
> URL: https://issues.apache.org/jira/browse/BEAM-9513
> Project: Beam
>  Issue Type: Bug
>  Components: dsl-sql-zetasql
>Reporter: Andrew Pilloud
>Assignee: Andrew Pilloud
>Priority: Major
>  Labels: zetasql-compliance
>
> {code:java}
> Mar 16, 2020 12:58:26 PM 
> cloud.dataflow.sql.ExecuteQueryServiceServer$SqlComplianceServiceImpl 
> executeQuery
> INFO: Processing Sql statement: WITH
>   R AS (SELECT 2 a),
>   S AS (SELECT a-1 x, a+1 y FROM R)
> SELECT a, x, y FROM R, S
> Mar 16, 2020 12:58:26 PM 
> com.google.zetasql.io.grpc.internal.SerializingExecutor run
> SEVERE: Exception while executing runnable 
> com.google.zetasql.io.grpc.internal.ServerImpl$JumpToApplicationThreadServerStreamListener$1HalfClosed@5341cc1c
> java.lang.NullPointerException
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.ExpressionConverter.convertRexNodeFromResolvedExprWithRefScan(ExpressionConverter.java:374)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.JoinScanWithRefConverter.convert(JoinScanWithRefConverter.java:63)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.JoinScanWithRefConverter.convert(JoinScanWithRefConverter.java:35)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.QueryStatementConverter.convertNode(QueryStatementConverter.java:97)
>   at 
> java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:193)
>   at java.util.Collections$2.tryAdvance(Collections.java:4717)
>   at java.util.Collections$2.forEachRemaining(Collections.java:4725)
>   at java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:481)
>   at 
> java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:471)
>   at 
> java.util.stream.ReduceOps$ReduceOp.evaluateSequential(ReduceOps.java:708)
>   at java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:234)
>   at 
> java.util.stream.ReferencePipeline.collect(ReferencePipeline.java:499)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.QueryStatementConverter.convertNode(QueryStatementConverter.java:96)
>   at 
> java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:193)
>   at java.util.Collections$2.tryAdvance(Collections.java:4717)
>   at java.util.Collections$2.forEachRemaining(Collections.java:4725)
>   at java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:481)
>   at 
> java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:471)
>   at 
> java.util.stream.ReduceOps$ReduceOp.evaluateSequential(ReduceOps.java:708)
>   at java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:234)
>   at 
> java.util.stream.ReferencePipeline.collect(ReferencePipeline.java:499)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.QueryStatementConverter.convertNode(QueryStatementConverter.java:96)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.QueryStatementConverter.convert(QueryStatementConverter.java:84)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.QueryStatementConverter.convertRootQuery(QueryStatementConverter.java:51)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLPlannerImpl.rel(ZetaSQLPlannerImpl.java:160)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLQueryPlanner.convertToBeamRelInternal(ZetaSQLQueryPlanner.java:131)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLQueryPlanner.convertToBeamRel(ZetaSQLQueryPlanner.java:115)
>   at 
> cloud.dataflow.sql.ExecuteQueryServiceServer$SqlComplianceServiceImpl.executeQuery(ExecuteQueryServiceServer.java:242)
>   at 
> com.google.zetasql.testing.SqlComplianceServiceGrpc$MethodHandlers.invoke(SqlComplianceServiceGrpc.java:423)
>   at 
> com.google.zetasql.io.grpc.stub.ServerCalls$UnaryServerCallHandler$UnaryServerCallListener.onHalfClose(ServerCalls.java:171)
>   at 
> com.google.zetasql.io.grpc.internal.ServerCallImpl$ServerStreamListenerImpl.halfClosed(ServerCallImpl.java:283)
>   at 
> com.google.zetasql.io.grpc.internal.ServerImpl$JumpToApplicationThreadServerStreamListener$1HalfClosed.runInContext(ServerImpl.java:711)
>   at 
> com.google.zetasql.io.grpc.internal.ContextRunnable.run(ContextRunnable.java:37)
>   at 
> com.google.zetasql.io.grpc.internal.SerializingExecutor.run(SerializingExecutor.java:123)
>   at 
> java.util.concurrent.ThreadPoolExecutor.runWorker

[jira] [Assigned] (BEAM-9521) NullPointerException in convertRexNodeFromResolvedExpr

2020-03-16 Thread Andrew Pilloud (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-9521?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrew Pilloud reassigned BEAM-9521:


Assignee: Andrew Pilloud

> NullPointerException in convertRexNodeFromResolvedExpr
> --
>
> Key: BEAM-9521
> URL: https://issues.apache.org/jira/browse/BEAM-9521
> Project: Beam
>  Issue Type: Bug
>  Components: dsl-sql-zetasql
>Reporter: Andrew Pilloud
>Assignee: Andrew Pilloud
>Priority: Major
>  Labels: zetasql-compliance
>
> {code:java}
> Mar 16, 2020 12:57:10 PM 
> cloud.dataflow.sql.ExecuteQueryServiceServer$SqlComplianceServiceImpl 
> executeQuery
> INFO: Processing Sql statement: SELECT (SELECT AS STRUCT
> (SELECT AS STRUCT l.float_val lf, r.float_val rf),
> l.float_val = r.float_val),
>(SELECT AS STRUCT
> (SELECT AS STRUCT l.double_val ld, r.double_val rd),
> l.double_val = r.double_val)
> FROM TestTableZero l, TestTableZero r
> WHERE l.row_id > r.row_id
> Mar 16, 2020 12:57:10 PM 
> com.google.zetasql.io.grpc.internal.SerializingExecutor run
> SEVERE: Exception while executing runnable 
> com.google.zetasql.io.grpc.internal.ServerImpl$JumpToApplicationThreadServerStreamListener$1HalfClosed@710f8c0d
> java.lang.NullPointerException
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.ExpressionConverter.convertRexNodeFromResolvedExpr(ExpressionConverter.java:301)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.JoinScanConverter.convert(JoinScanConverter.java:83)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.JoinScanConverter.convert(JoinScanConverter.java:36)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.QueryStatementConverter.convertNode(QueryStatementConverter.java:97)
>   at 
> java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:193)
>   at java.util.Collections$2.tryAdvance(Collections.java:4717)
>   at java.util.Collections$2.forEachRemaining(Collections.java:4725)
>   at java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:481)
>   at 
> java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:471)
>   at 
> java.util.stream.ReduceOps$ReduceOp.evaluateSequential(ReduceOps.java:708)
>   at java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:234)
>   at 
> java.util.stream.ReferencePipeline.collect(ReferencePipeline.java:499)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.QueryStatementConverter.convertNode(QueryStatementConverter.java:96)
>   at 
> java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:193)
>   at java.util.Collections$2.tryAdvance(Collections.java:4717)
>   at java.util.Collections$2.forEachRemaining(Collections.java:4725)
>   at java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:481)
>   at 
> java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:471)
>   at 
> java.util.stream.ReduceOps$ReduceOp.evaluateSequential(ReduceOps.java:708)
>   at java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:234)
>   at 
> java.util.stream.ReferencePipeline.collect(ReferencePipeline.java:499)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.QueryStatementConverter.convertNode(QueryStatementConverter.java:96)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.QueryStatementConverter.convert(QueryStatementConverter.java:84)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.QueryStatementConverter.convertRootQuery(QueryStatementConverter.java:51)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLPlannerImpl.rel(ZetaSQLPlannerImpl.java:160)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLQueryPlanner.convertToBeamRelInternal(ZetaSQLQueryPlanner.java:131)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLQueryPlanner.convertToBeamRel(ZetaSQLQueryPlanner.java:115)
>   at 
> cloud.dataflow.sql.ExecuteQueryServiceServer$SqlComplianceServiceImpl.executeQuery(ExecuteQueryServiceServer.java:242)
>   at 
> com.google.zetasql.testing.SqlComplianceServiceGrpc$MethodHandlers.invoke(SqlComplianceServiceGrpc.java:423)
>   at 
> com.google.zetasql.io.grpc.stub.ServerCalls$UnaryServerCallHandler$UnaryServerCallListener.onHalfClose(ServerCalls.java:171)
>   at 
> com.google.zetasql.io.grpc.internal.ServerCallImpl$ServerStreamListenerImpl.halfClosed(ServerCallImpl.java:283)
>   at 
> com.google.zetasql.io.grpc.internal.ServerImpl$JumpToApplicationThreadServerStreamListener$1HalfClosed.runInContext(ServerImpl.java:711)
>   at 
> com.google.zetasql.io.grpc.internal.ContextRunnable.run(ContextRunnable.java:37)

[jira] [Updated] (BEAM-9511) ArrayScanToUncollectConverter: ResolvedParameter cannot be cast to ResolvedLiteral

2020-03-16 Thread Andrew Pilloud (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-9511?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrew Pilloud updated BEAM-9511:
-
Status: Open  (was: Triage Needed)

> ArrayScanToUncollectConverter: ResolvedParameter cannot be cast to 
> ResolvedLiteral
> --
>
> Key: BEAM-9511
> URL: https://issues.apache.org/jira/browse/BEAM-9511
> Project: Beam
>  Issue Type: Bug
>  Components: dsl-sql-zetasql
>Reporter: Andrew Pilloud
>Priority: Major
>  Labels: zetasql-compliance
>
>  
> {code:java}
> Mar 16, 2020 12:57:42 PM 
> cloud.dataflow.sql.ExecuteQueryServiceServer$SqlComplianceServiceImpl 
> executeQuery
> INFO: Processing Sql statement: SELECT elem FROM UNNEST(@string_arr) AS elem
> Mar 16, 2020 12:57:42 PM 
> com.google.zetasql.io.grpc.internal.SerializingExecutor run
> SEVERE: Exception while executing runnable 
> com.google.zetasql.io.grpc.internal.ServerImpl$JumpToApplicationThreadServerStreamListener$1HalfClosed@2f5b49be
> java.lang.ClassCastException: 
> com.google.zetasql.resolvedast.ResolvedNodes$ResolvedParameter cannot be cast 
> to com.google.zetasql.resolvedast.ResolvedNodes$ResolvedLiteral
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.ArrayScanToUncollectConverter.convert(ArrayScanToUncollectConverter.java:45)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.ArrayScanToUncollectConverter.convert(ArrayScanToUncollectConverter.java:31)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.QueryStatementConverter.convertNode(QueryStatementConverter.java:97)
>   at 
> java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:193)
>   at java.util.Collections$2.tryAdvance(Collections.java:4717)
>   at java.util.Collections$2.forEachRemaining(Collections.java:4725)
>   at java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:481)
>   at 
> java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:471)
>   at 
> java.util.stream.ReduceOps$ReduceOp.evaluateSequential(ReduceOps.java:708)
>   at java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:234)
>   at 
> java.util.stream.ReferencePipeline.collect(ReferencePipeline.java:499)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.QueryStatementConverter.convertNode(QueryStatementConverter.java:96)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.QueryStatementConverter.convert(QueryStatementConverter.java:84)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.QueryStatementConverter.convertRootQuery(QueryStatementConverter.java:51)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLPlannerImpl.rel(ZetaSQLPlannerImpl.java:160)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLQueryPlanner.convertToBeamRelInternal(ZetaSQLQueryPlanner.java:131)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLQueryPlanner.convertToBeamRel(ZetaSQLQueryPlanner.java:115)
>   at 
> cloud.dataflow.sql.ExecuteQueryServiceServer$SqlComplianceServiceImpl.executeQuery(ExecuteQueryServiceServer.java:242)
>   at 
> com.google.zetasql.testing.SqlComplianceServiceGrpc$MethodHandlers.invoke(SqlComplianceServiceGrpc.java:423)
>   at 
> com.google.zetasql.io.grpc.stub.ServerCalls$UnaryServerCallHandler$UnaryServerCallListener.onHalfClose(ServerCalls.java:171)
>   at 
> com.google.zetasql.io.grpc.internal.ServerCallImpl$ServerStreamListenerImpl.halfClosed(ServerCallImpl.java:283)
>   at 
> com.google.zetasql.io.grpc.internal.ServerImpl$JumpToApplicationThreadServerStreamListener$1HalfClosed.runInContext(ServerImpl.java:711)
>   at 
> com.google.zetasql.io.grpc.internal.ContextRunnable.run(ContextRunnable.java:37)
>   at 
> com.google.zetasql.io.grpc.internal.SerializingExecutor.run(SerializingExecutor.java:123)
>   at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
>   at java.lang.Thread.run(Thread.java:748) {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (BEAM-9513) NullPointerException in convertRexNodeFromResolvedExprWithRefScan

2020-03-16 Thread Andrew Pilloud (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-9513?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrew Pilloud updated BEAM-9513:
-
Status: Open  (was: Triage Needed)

> NullPointerException in convertRexNodeFromResolvedExprWithRefScan
> -
>
> Key: BEAM-9513
> URL: https://issues.apache.org/jira/browse/BEAM-9513
> Project: Beam
>  Issue Type: Bug
>  Components: dsl-sql-zetasql
>Reporter: Andrew Pilloud
>Priority: Major
>  Labels: zetasql-compliance
>
> {code:java}
> Mar 16, 2020 12:58:26 PM 
> cloud.dataflow.sql.ExecuteQueryServiceServer$SqlComplianceServiceImpl 
> executeQuery
> INFO: Processing Sql statement: WITH
>   R AS (SELECT 2 a),
>   S AS (SELECT a-1 x, a+1 y FROM R)
> SELECT a, x, y FROM R, S
> Mar 16, 2020 12:58:26 PM 
> com.google.zetasql.io.grpc.internal.SerializingExecutor run
> SEVERE: Exception while executing runnable 
> com.google.zetasql.io.grpc.internal.ServerImpl$JumpToApplicationThreadServerStreamListener$1HalfClosed@5341cc1c
> java.lang.NullPointerException
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.ExpressionConverter.convertRexNodeFromResolvedExprWithRefScan(ExpressionConverter.java:374)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.JoinScanWithRefConverter.convert(JoinScanWithRefConverter.java:63)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.JoinScanWithRefConverter.convert(JoinScanWithRefConverter.java:35)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.QueryStatementConverter.convertNode(QueryStatementConverter.java:97)
>   at 
> java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:193)
>   at java.util.Collections$2.tryAdvance(Collections.java:4717)
>   at java.util.Collections$2.forEachRemaining(Collections.java:4725)
>   at java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:481)
>   at 
> java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:471)
>   at 
> java.util.stream.ReduceOps$ReduceOp.evaluateSequential(ReduceOps.java:708)
>   at java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:234)
>   at 
> java.util.stream.ReferencePipeline.collect(ReferencePipeline.java:499)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.QueryStatementConverter.convertNode(QueryStatementConverter.java:96)
>   at 
> java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:193)
>   at java.util.Collections$2.tryAdvance(Collections.java:4717)
>   at java.util.Collections$2.forEachRemaining(Collections.java:4725)
>   at java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:481)
>   at 
> java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:471)
>   at 
> java.util.stream.ReduceOps$ReduceOp.evaluateSequential(ReduceOps.java:708)
>   at java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:234)
>   at 
> java.util.stream.ReferencePipeline.collect(ReferencePipeline.java:499)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.QueryStatementConverter.convertNode(QueryStatementConverter.java:96)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.QueryStatementConverter.convert(QueryStatementConverter.java:84)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.QueryStatementConverter.convertRootQuery(QueryStatementConverter.java:51)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLPlannerImpl.rel(ZetaSQLPlannerImpl.java:160)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLQueryPlanner.convertToBeamRelInternal(ZetaSQLQueryPlanner.java:131)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLQueryPlanner.convertToBeamRel(ZetaSQLQueryPlanner.java:115)
>   at 
> cloud.dataflow.sql.ExecuteQueryServiceServer$SqlComplianceServiceImpl.executeQuery(ExecuteQueryServiceServer.java:242)
>   at 
> com.google.zetasql.testing.SqlComplianceServiceGrpc$MethodHandlers.invoke(SqlComplianceServiceGrpc.java:423)
>   at 
> com.google.zetasql.io.grpc.stub.ServerCalls$UnaryServerCallHandler$UnaryServerCallListener.onHalfClose(ServerCalls.java:171)
>   at 
> com.google.zetasql.io.grpc.internal.ServerCallImpl$ServerStreamListenerImpl.halfClosed(ServerCallImpl.java:283)
>   at 
> com.google.zetasql.io.grpc.internal.ServerImpl$JumpToApplicationThreadServerStreamListener$1HalfClosed.runInContext(ServerImpl.java:711)
>   at 
> com.google.zetasql.io.grpc.internal.ContextRunnable.run(ContextRunnable.java:37)
>   at 
> com.google.zetasql.io.grpc.internal.SerializingExecutor.run(SerializingExecutor.java:123)
>   at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
>   

[jira] [Updated] (BEAM-9516) ArrayScanToUncollectConverter: ResolvedSubqueryExpr cannot be cast to ResolvedLiteral

2020-03-16 Thread Andrew Pilloud (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-9516?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrew Pilloud updated BEAM-9516:
-
Status: Open  (was: Triage Needed)

> ArrayScanToUncollectConverter: ResolvedSubqueryExpr cannot be cast to 
> ResolvedLiteral
> -
>
> Key: BEAM-9516
> URL: https://issues.apache.org/jira/browse/BEAM-9516
> Project: Beam
>  Issue Type: Bug
>  Components: dsl-sql-zetasql
>Reporter: Andrew Pilloud
>Priority: Major
>  Labels: zetasql-compliance
>
> {code:java}
> Mar 16, 2020 12:58:49 PM 
> com.google.zetasql.io.grpc.internal.SerializingExecutor run
> SEVERE: Exception while executing runnable 
> com.google.zetasql.io.grpc.internal.ServerImpl$JumpToApplicationThreadServerStreamListener$1HalfClosed@7c1d4ea6
> java.lang.ClassCastException: 
> com.google.zetasql.resolvedast.ResolvedNodes$ResolvedSubqueryExpr cannot be 
> cast to com.google.zetasql.resolvedast.ResolvedNodes$ResolvedLiteral
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.ArrayScanToUncollectConverter.convert(ArrayScanToUncollectConverter.java:45)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.ArrayScanToUncollectConverter.convert(ArrayScanToUncollectConverter.java:31)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.QueryStatementConverter.convertNode(QueryStatementConverter.java:97)
>   at 
> java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:193)
>   at java.util.Collections$2.tryAdvance(Collections.java:4717)
>   at java.util.Collections$2.forEachRemaining(Collections.java:4725)
>   at java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:481)
>   at 
> java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:471)
>   at 
> java.util.stream.ReduceOps$ReduceOp.evaluateSequential(ReduceOps.java:708)
>   at java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:234)
>   at 
> java.util.stream.ReferencePipeline.collect(ReferencePipeline.java:499)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.QueryStatementConverter.convertNode(QueryStatementConverter.java:96)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.QueryStatementConverter.convert(QueryStatementConverter.java:84)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.QueryStatementConverter.convertRootQuery(QueryStatementConverter.java:51)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLPlannerImpl.rel(ZetaSQLPlannerImpl.java:160)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLQueryPlanner.convertToBeamRelInternal(ZetaSQLQueryPlanner.java:131)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLQueryPlanner.convertToBeamRel(ZetaSQLQueryPlanner.java:115)
>   at 
> cloud.dataflow.sql.ExecuteQueryServiceServer$SqlComplianceServiceImpl.executeQuery(ExecuteQueryServiceServer.java:242)
>   at 
> com.google.zetasql.testing.SqlComplianceServiceGrpc$MethodHandlers.invoke(SqlComplianceServiceGrpc.java:423)
>   at 
> com.google.zetasql.io.grpc.stub.ServerCalls$UnaryServerCallHandler$UnaryServerCallListener.onHalfClose(ServerCalls.java:171)
>   at 
> com.google.zetasql.io.grpc.internal.ServerCallImpl$ServerStreamListenerImpl.halfClosed(ServerCallImpl.java:283)
>   at 
> com.google.zetasql.io.grpc.internal.ServerImpl$JumpToApplicationThreadServerStreamListener$1HalfClosed.runInContext(ServerImpl.java:711)
>   at 
> com.google.zetasql.io.grpc.internal.ContextRunnable.run(ContextRunnable.java:37)
>   at 
> com.google.zetasql.io.grpc.internal.SerializingExecutor.run(SerializingExecutor.java:123)
>   at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
>   at java.lang.Thread.run(Thread.java:748)
>  {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (BEAM-9522) BeamJoinRel.extractJoinRexNode RexLiteral cannot be cast to RexCall

2020-03-16 Thread Andrew Pilloud (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-9522?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrew Pilloud updated BEAM-9522:
-
Status: Open  (was: Triage Needed)

> BeamJoinRel.extractJoinRexNode RexLiteral cannot be cast to RexCall
> ---
>
> Key: BEAM-9522
> URL: https://issues.apache.org/jira/browse/BEAM-9522
> Project: Beam
>  Issue Type: Bug
>  Components: dsl-sql-zetasql
>Reporter: Andrew Pilloud
>Priority: Minor
>  Labels: zetasql-compliance
>
> {code:java}
> Mar 16, 2020 1:26:41 PM 
> cloud.dataflow.sql.ExecuteQueryServiceServer$SqlComplianceServiceImpl 
> executeQuery
> INFO: Processing Sql statement: SELECT r.*, s.* FROM
> (SELECT x, x FROM (SELECT 2 x)) r FULL JOIN (SELECT 3) s ON false
> Mar 16, 2020 1:26:41 PM 
> com.google.zetasql.io.grpc.internal.SerializingExecutor run
> SEVERE: Exception while executing runnable 
> com.google.zetasql.io.grpc.internal.ServerImpl$JumpToApplicationThreadServerStreamListener$1HalfClosed@16f4bbdb
> java.lang.ClassCastException: 
> org.apache.beam.vendor.calcite.v1_20_0.org.apache.calcite.rex.RexLiteral 
> cannot be cast to 
> org.apache.beam.vendor.calcite.v1_20_0.org.apache.calcite.rex.RexCall
>   at 
> org.apache.beam.sdk.extensions.sql.impl.rel.BeamJoinRel.extractJoinRexNodes(BeamJoinRel.java:304)
>   at 
> org.apache.beam.sdk.extensions.sql.impl.rel.BeamJoinRel$ExtractJoinKeys.expand(BeamJoinRel.java:194)
>   at 
> org.apache.beam.sdk.extensions.sql.impl.rel.BeamJoinRel$ExtractJoinKeys.expand(BeamJoinRel.java:177)
>   at org.apache.beam.sdk.Pipeline.applyInternal(Pipeline.java:542)
>   at org.apache.beam.sdk.Pipeline.applyTransform(Pipeline.java:476)
>   at 
> org.apache.beam.sdk.values.PCollectionList.apply(PCollectionList.java:178)
>   at 
> org.apache.beam.sdk.extensions.sql.impl.rel.BeamCoGBKJoinRel$StandardJoin.expand(BeamCoGBKJoinRel.java:93)
>   at 
> org.apache.beam.sdk.extensions.sql.impl.rel.BeamCoGBKJoinRel$StandardJoin.expand(BeamCoGBKJoinRel.java:86)
>   at org.apache.beam.sdk.Pipeline.applyInternal(Pipeline.java:542)
>   at org.apache.beam.sdk.Pipeline.applyTransform(Pipeline.java:493)
>   at 
> org.apache.beam.sdk.extensions.sql.impl.rel.BeamSqlRelUtils.toPCollection(BeamSqlRelUtils.java:69)
>   at 
> org.apache.beam.sdk.extensions.sql.impl.rel.BeamSqlRelUtils.lambda$buildPCollectionList$0(BeamSqlRelUtils.java:50)
>   at 
> java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:193)
>   at java.util.Iterator.forEachRemaining(Iterator.java:116)
>   at 
> java.util.Spliterators$IteratorSpliterator.forEachRemaining(Spliterators.java:1801)
>   at java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:481)
>   at 
> java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:471)
>   at 
> java.util.stream.ReduceOps$ReduceOp.evaluateSequential(ReduceOps.java:708)
>   at java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:234)
>   at 
> java.util.stream.ReferencePipeline.collect(ReferencePipeline.java:499)
>   at 
> org.apache.beam.sdk.extensions.sql.impl.rel.BeamSqlRelUtils.buildPCollectionList(BeamSqlRelUtils.java:51)
>   at 
> org.apache.beam.sdk.extensions.sql.impl.rel.BeamSqlRelUtils.toPCollection(BeamSqlRelUtils.java:67)
>   at 
> org.apache.beam.sdk.extensions.sql.impl.rel.BeamSqlRelUtils.toPCollection(BeamSqlRelUtils.java:39)
>   at 
> org.apache.beam.sdk.extensions.sql.impl.rel.BeamEnumerableConverter.runCollector(BeamEnumerableConverter.java:199)
>   at 
> org.apache.beam.sdk.extensions.sql.impl.rel.BeamEnumerableConverter.collectRows(BeamEnumerableConverter.java:218)
>   at 
> org.apache.beam.sdk.extensions.sql.impl.rel.BeamEnumerableConverter.toRowList(BeamEnumerableConverter.java:150)
>   at 
> org.apache.beam.sdk.extensions.sql.impl.rel.BeamEnumerableConverter.toRowList(BeamEnumerableConverter.java:127)
>   at 
> cloud.dataflow.sql.ExecuteQueryServiceServer$SqlComplianceServiceImpl.executeQuery(ExecuteQueryServiceServer.java:244)
>   at 
> com.google.zetasql.testing.SqlComplianceServiceGrpc$MethodHandlers.invoke(SqlComplianceServiceGrpc.java:423)
>   at 
> com.google.zetasql.io.grpc.stub.ServerCalls$UnaryServerCallHandler$UnaryServerCallListener.onHalfClose(ServerCalls.java:171)
>   at 
> com.google.zetasql.io.grpc.internal.ServerCallImpl$ServerStreamListenerImpl.halfClosed(ServerCallImpl.java:283)
>   at 
> com.google.zetasql.io.grpc.internal.ServerImpl$JumpToApplicationThreadServerStreamListener$1HalfClosed.runInContext(ServerImpl.java:711)
>   at 
> com.google.zetasql.io.grpc.internal.ContextRunnable.run(ContextRunnable.java:37)
>   at 
> com.google.zetasql.io.grpc.internal.SerializingExecutor.run(SerializingExecutor.java:

[jira] [Updated] (BEAM-9518) BeamZetaSqlCalcRel$CalcFn.processElement Unsupported Beam fieldType: LOGICAL_TYPE

2020-03-16 Thread Andrew Pilloud (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-9518?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrew Pilloud updated BEAM-9518:
-
Status: Open  (was: Triage Needed)

> BeamZetaSqlCalcRel$CalcFn.processElement Unsupported Beam fieldType: 
> LOGICAL_TYPE
> -
>
> Key: BEAM-9518
> URL: https://issues.apache.org/jira/browse/BEAM-9518
> Project: Beam
>  Issue Type: Bug
>  Components: dsl-sql-zetasql
>Reporter: Andrew Pilloud
>Priority: Minor
>  Labels: zetasql-compliance
>
> Mar 16, 2020 1:23:15 PM 
> com.google.zetasql.io.grpc.internal.SerializingExecutor run
> SEVERE: Exception while executing runnable 
> com.google.zetasql.io.grpc.internal.ServerImpl$JumpToApplicationThreadServerStreamListener$1HalfClosed@61d0fcb4
> org.apache.beam.sdk.Pipeline$PipelineExecutionException: 
> java.lang.UnsupportedOperationException: Unsupported Beam fieldType: 
> LOGICAL_TYPE
>  at 
> org.apache.beam.runners.direct.DirectRunner$DirectPipelineResult.waitUntilFinish(DirectRunner.java:348)
>  at 
> org.apache.beam.runners.direct.DirectRunner$DirectPipelineResult.waitUntilFinish(DirectRunner.java:318)
>  at org.apache.beam.runners.direct.DirectRunner.run(DirectRunner.java:213)
>  at org.apache.beam.runners.direct.DirectRunner.run(DirectRunner.java:67)
>  at org.apache.beam.sdk.Pipeline.run(Pipeline.java:317)
>  at org.apache.beam.sdk.Pipeline.run(Pipeline.java:303)
>  at 
> org.apache.beam.sdk.extensions.sql.impl.rel.BeamEnumerableConverter.runCollector(BeamEnumerableConverter.java:201)
>  at 
> org.apache.beam.sdk.extensions.sql.impl.rel.BeamEnumerableConverter.collectRows(BeamEnumerableConverter.java:218)
>  at 
> org.apache.beam.sdk.extensions.sql.impl.rel.BeamEnumerableConverter.toRowList(BeamEnumerableConverter.java:150)
>  at 
> org.apache.beam.sdk.extensions.sql.impl.rel.BeamEnumerableConverter.toRowList(BeamEnumerableConverter.java:127)
>  at 
> cloud.dataflow.sql.ExecuteQueryServiceServer$SqlComplianceServiceImpl.executeQuery(ExecuteQueryServiceServer.java:244)
>  at 
> com.google.zetasql.testing.SqlComplianceServiceGrpc$MethodHandlers.invoke(SqlComplianceServiceGrpc.java:423)
>  at 
> com.google.zetasql.io.grpc.stub.ServerCalls$UnaryServerCallHandler$UnaryServerCallListener.onHalfClose(ServerCalls.java:171)
>  at 
> com.google.zetasql.io.grpc.internal.ServerCallImpl$ServerStreamListenerImpl.halfClosed(ServerCallImpl.java:283)
>  at 
> com.google.zetasql.io.grpc.internal.ServerImpl$JumpToApplicationThreadServerStreamListener$1HalfClosed.runInContext(ServerImpl.java:711)
>  at 
> com.google.zetasql.io.grpc.internal.ContextRunnable.run(ContextRunnable.java:37)
>  at 
> com.google.zetasql.io.grpc.internal.SerializingExecutor.run(SerializingExecutor.java:123)
>  at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
>  at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
>  at java.lang.Thread.run(Thread.java:748)
> Caused by: java.lang.UnsupportedOperationException: Unsupported Beam 
> fieldType: LOGICAL_TYPE
>  at 
> org.apache.beam.sdk.extensions.sql.zetasql.ZetaSqlUtils.zetaSqlValueToJavaObject(ZetaSqlUtils.java:219)
>  at 
> org.apache.beam.sdk.extensions.sql.zetasql.BeamZetaSqlCalcRel$CalcFn.processElement(BeamZetaSqlCalcRel.java:193)



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (BEAM-9519) BeamZetaSqlCalcRel$CalcFn.processElement field is not nullable in Array field $arrayx

2020-03-16 Thread Andrew Pilloud (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-9519?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrew Pilloud updated BEAM-9519:
-
Status: Open  (was: Triage Needed)

> BeamZetaSqlCalcRel$CalcFn.processElement field is not nullable in Array field 
> $arrayx
> -
>
> Key: BEAM-9519
> URL: https://issues.apache.org/jira/browse/BEAM-9519
> Project: Beam
>  Issue Type: Bug
>  Components: dsl-sql-zetasql
>Reporter: Andrew Pilloud
>Priority: Minor
>  Labels: zetasql-compliance
>
> {code:java}
> Mar 16, 2020 1:27:47 PM 
> cloud.dataflow.sql.ExecuteQueryServiceServer$SqlComplianceServiceImpl 
> executeQuery
> INFO: Processing Sql statement: SELECT x FROM UNNEST([3, NULL]) x;
> Mar 16, 2020 1:27:48 PM 
> com.google.zetasql.io.grpc.internal.SerializingExecutor run
> SEVERE: Exception while executing runnable 
> com.google.zetasql.io.grpc.internal.ServerImpl$JumpToApplicationThreadServerStreamListener$1HalfClosed@39086547
> org.apache.beam.sdk.Pipeline$PipelineExecutionException: 
> java.lang.IllegalArgumentException: FieldType{typeName=INT64, nullable=false, 
> logicalType=null, collectionElementType=null, mapKeyType=null, 
> mapValueType=null, rowSchema=null, metadata={}} is not nullable in Array 
> field $arrayx
>   at 
> org.apache.beam.runners.direct.DirectRunner$DirectPipelineResult.waitUntilFinish(DirectRunner.java:348)
>   at 
> org.apache.beam.runners.direct.DirectRunner$DirectPipelineResult.waitUntilFinish(DirectRunner.java:318)
>   at 
> org.apache.beam.runners.direct.DirectRunner.run(DirectRunner.java:213)
>   at org.apache.beam.runners.direct.DirectRunner.run(DirectRunner.java:67)
>   at org.apache.beam.sdk.Pipeline.run(Pipeline.java:317)
>   at org.apache.beam.sdk.Pipeline.run(Pipeline.java:303)
>   at 
> org.apache.beam.sdk.extensions.sql.impl.rel.BeamEnumerableConverter.runCollector(BeamEnumerableConverter.java:201)
>   at 
> org.apache.beam.sdk.extensions.sql.impl.rel.BeamEnumerableConverter.collectRows(BeamEnumerableConverter.java:218)
>   at 
> org.apache.beam.sdk.extensions.sql.impl.rel.BeamEnumerableConverter.toRowList(BeamEnumerableConverter.java:150)
>   at 
> org.apache.beam.sdk.extensions.sql.impl.rel.BeamEnumerableConverter.toRowList(BeamEnumerableConverter.java:127)
>   at 
> cloud.dataflow.sql.ExecuteQueryServiceServer$SqlComplianceServiceImpl.executeQuery(ExecuteQueryServiceServer.java:244)
>   at 
> com.google.zetasql.testing.SqlComplianceServiceGrpc$MethodHandlers.invoke(SqlComplianceServiceGrpc.java:423)
>   at 
> com.google.zetasql.io.grpc.stub.ServerCalls$UnaryServerCallHandler$UnaryServerCallListener.onHalfClose(ServerCalls.java:171)
>   at 
> com.google.zetasql.io.grpc.internal.ServerCallImpl$ServerStreamListenerImpl.halfClosed(ServerCallImpl.java:283)
>   at 
> com.google.zetasql.io.grpc.internal.ServerImpl$JumpToApplicationThreadServerStreamListener$1HalfClosed.runInContext(ServerImpl.java:711)
>   at 
> com.google.zetasql.io.grpc.internal.ContextRunnable.run(ContextRunnable.java:37)
>   at 
> com.google.zetasql.io.grpc.internal.SerializingExecutor.run(SerializingExecutor.java:123)
>   at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
>   at java.lang.Thread.run(Thread.java:748)
> Caused by: java.lang.IllegalArgumentException: FieldType{typeName=INT64, 
> nullable=false, logicalType=null, collectionElementType=null, 
> mapKeyType=null, mapValueType=null, rowSchema=null, metadata={}} is not 
> nullable in Array field $arrayx
>   at org.apache.beam.sdk.values.Row$Builder.verifyArray(Row.java:682)
>   at org.apache.beam.sdk.values.Row$Builder.verify(Row.java:647)
>   at org.apache.beam.sdk.values.Row$Builder.verify(Row.java:639)
>   at org.apache.beam.sdk.values.Row$Builder.build(Row.java:844)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.BeamZetaSqlCalcRel$CalcFn.processElement(BeamZetaSqlCalcRel.java:195)
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (BEAM-9512) Anonymous structs have name collision in schema

2020-03-16 Thread Andrew Pilloud (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-9512?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrew Pilloud updated BEAM-9512:
-
Status: Open  (was: Triage Needed)

> Anonymous structs have name collision in schema
> ---
>
> Key: BEAM-9512
> URL: https://issues.apache.org/jira/browse/BEAM-9512
> Project: Beam
>  Issue Type: Bug
>  Components: dsl-sql-zetasql
>Reporter: Andrew Pilloud
>Priority: Major
>  Labels: zetasql-compliance
>
> {code:java}
> Mar 16, 2020 12:57:42 PM 
> cloud.dataflow.sql.ExecuteQueryServiceServer$SqlComplianceServiceImpl 
> executeQuery
> INFO: Processing Sql statement: SELECT STRUCT(ARRAY INT64>>[(11, 12), (21, 22)])
> Mar 16, 2020 12:57:42 PM 
> com.google.zetasql.io.grpc.internal.SerializingExecutor run
> SEVERE: Exception while executing runnable 
> com.google.zetasql.io.grpc.internal.ServerImpl$JumpToApplicationThreadServerStreamListener$1HalfClosed@c73b08a
> java.lang.IllegalArgumentException: Duplicate field  added to schema
>   at org.apache.beam.sdk.schemas.Schema.(Schema.java:228)
>   at org.apache.beam.sdk.schemas.Schema.fromFields(Schema.java:966)
>   at 
> java.util.stream.ReferencePipeline.collect(ReferencePipeline.java:503)
>   at 
> org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils.toSchema(CalciteUtils.java:194)
>   at 
> org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils.toFieldType(CalciteUtils.java:251)
>   at 
> org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils.toFieldType(CalciteUtils.java:246)
>   at 
> org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils.toField(CalciteUtils.java:239)
>   at 
> org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils.toField(CalciteUtils.java:235)
>   at 
> java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:193)
>   at java.util.Iterator.forEachRemaining(Iterator.java:116)
>   at 
> java.util.Spliterators$IteratorSpliterator.forEachRemaining(Spliterators.java:1801)
>   at java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:481)
>   at 
> java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:471)
>   at 
> java.util.stream.ReduceOps$ReduceOp.evaluateSequential(ReduceOps.java:708)
>   at java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:234)
>   at 
> java.util.stream.ReferencePipeline.collect(ReferencePipeline.java:499)
>   at 
> org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils.toSchema(CalciteUtils.java:194)
>   at 
> org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils.toFieldType(CalciteUtils.java:251)
>   at 
> org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils.toField(CalciteUtils.java:239)
>   at 
> org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils.toField(CalciteUtils.java:235)
>   at 
> java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:193)
>   at java.util.Iterator.forEachRemaining(Iterator.java:116)
>   at 
> java.util.Spliterators$IteratorSpliterator.forEachRemaining(Spliterators.java:1801)
>   at java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:481)
>   at 
> java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:471)
>   at 
> java.util.stream.ReduceOps$ReduceOp.evaluateSequential(ReduceOps.java:708)
>   at java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:234)
>   at 
> java.util.stream.ReferencePipeline.collect(ReferencePipeline.java:499)
>   at 
> org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils.toSchema(CalciteUtils.java:194)
>   at 
> cloud.dataflow.sql.ExecuteQueryServiceServer$SqlComplianceServiceImpl.executeQuery(ExecuteQueryServiceServer.java:243)
>   at 
> com.google.zetasql.testing.SqlComplianceServiceGrpc$MethodHandlers.invoke(SqlComplianceServiceGrpc.java:423)
>   at 
> com.google.zetasql.io.grpc.stub.ServerCalls$UnaryServerCallHandler$UnaryServerCallListener.onHalfClose(ServerCalls.java:171)
>   at 
> com.google.zetasql.io.grpc.internal.ServerCallImpl$ServerStreamListenerImpl.halfClosed(ServerCallImpl.java:283)
>   at 
> com.google.zetasql.io.grpc.internal.ServerImpl$JumpToApplicationThreadServerStreamListener$1HalfClosed.runInContext(ServerImpl.java:711)
>   at 
> com.google.zetasql.io.grpc.internal.ContextRunnable.run(ContextRunnable.java:37)
>   at 
> com.google.zetasql.io.grpc.internal.SerializingExecutor.run(SerializingExecutor.java:123)
>   at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
>   at java.lang.Thread.run(Thread.java:748)
>  {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (BEAM-9520) Forbidden IOException when reading from InputStream

2020-03-16 Thread Andrew Pilloud (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-9520?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrew Pilloud updated BEAM-9520:
-
Status: Open  (was: Triage Needed)

> Forbidden IOException when reading from InputStream
> ---
>
> Key: BEAM-9520
> URL: https://issues.apache.org/jira/browse/BEAM-9520
> Project: Beam
>  Issue Type: Bug
>  Components: dsl-sql-zetasql
>Reporter: Andrew Pilloud
>Priority: Minor
>  Labels: zetasql-compliance
>
> {code:java}
> Mar 16, 2020 1:26:17 PM 
> cloud.dataflow.sql.ExecuteQueryServiceServer$SqlComplianceServiceImpl 
> executeQuery
> INFO: Processing Sql statement: (SELECT 1 UNION ALL SELECT NULL UNION ALL 
> SELECT 1 UNION ALL SELECT NULL)
> INTERSECT DISTINCT
> (SELECT 1 UNION ALL SELECT NULL UNION ALL SELECT NULL)
> Mar 16, 2020 1:26:18 PM 
> com.google.zetasql.io.grpc.internal.SerializingExecutor run
> SEVERE: Exception while executing runnable 
> com.google.zetasql.io.grpc.internal.ServerImpl$JumpToApplicationThreadServerStreamListener$1HalfClosed@7f643e84
> org.apache.beam.sdk.Pipeline$PipelineExecutionException: 
> java.lang.IllegalArgumentException: Forbidden IOException when reading from 
> InputStream
>   at 
> org.apache.beam.runners.direct.DirectRunner$DirectPipelineResult.waitUntilFinish(DirectRunner.java:348)
>   at 
> org.apache.beam.runners.direct.DirectRunner$DirectPipelineResult.waitUntilFinish(DirectRunner.java:318)
>   at 
> org.apache.beam.runners.direct.DirectRunner.run(DirectRunner.java:213)
>   at org.apache.beam.runners.direct.DirectRunner.run(DirectRunner.java:67)
>   at org.apache.beam.sdk.Pipeline.run(Pipeline.java:317)
>   at org.apache.beam.sdk.Pipeline.run(Pipeline.java:303)
>   at 
> org.apache.beam.sdk.extensions.sql.impl.rel.BeamEnumerableConverter.runCollector(BeamEnumerableConverter.java:201)
>   at 
> org.apache.beam.sdk.extensions.sql.impl.rel.BeamEnumerableConverter.collectRows(BeamEnumerableConverter.java:218)
>   at 
> org.apache.beam.sdk.extensions.sql.impl.rel.BeamEnumerableConverter.toRowList(BeamEnumerableConverter.java:150)
>   at 
> org.apache.beam.sdk.extensions.sql.impl.rel.BeamEnumerableConverter.toRowList(BeamEnumerableConverter.java:127)
>   at 
> cloud.dataflow.sql.ExecuteQueryServiceServer$SqlComplianceServiceImpl.executeQuery(ExecuteQueryServiceServer.java:244)
>   at 
> com.google.zetasql.testing.SqlComplianceServiceGrpc$MethodHandlers.invoke(SqlComplianceServiceGrpc.java:423)
>   at 
> com.google.zetasql.io.grpc.stub.ServerCalls$UnaryServerCallHandler$UnaryServerCallListener.onHalfClose(ServerCalls.java:171)
>   at 
> com.google.zetasql.io.grpc.internal.ServerCallImpl$ServerStreamListenerImpl.halfClosed(ServerCallImpl.java:283)
>   at 
> com.google.zetasql.io.grpc.internal.ServerImpl$JumpToApplicationThreadServerStreamListener$1HalfClosed.runInContext(ServerImpl.java:711)
>   at 
> com.google.zetasql.io.grpc.internal.ContextRunnable.run(ContextRunnable.java:37)
>   at 
> com.google.zetasql.io.grpc.internal.SerializingExecutor.run(SerializingExecutor.java:123)
>   at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
>   at java.lang.Thread.run(Thread.java:748)
> Caused by: java.lang.IllegalArgumentException: Forbidden IOException when 
> reading from InputStream
>   at 
> org.apache.beam.sdk.util.CoderUtils.decodeFromSafeStream(CoderUtils.java:118)
>   at 
> org.apache.beam.sdk.util.CoderUtils.decodeFromByteArray(CoderUtils.java:98)
>   at 
> org.apache.beam.sdk.util.CoderUtils.decodeFromByteArray(CoderUtils.java:92)
>   at org.apache.beam.sdk.util.CoderUtils.clone(CoderUtils.java:141)
>   at 
> org.apache.beam.sdk.util.MutationDetectors$CodedValueMutationDetector.(MutationDetectors.java:115)
>   at 
> org.apache.beam.sdk.util.MutationDetectors.forValueWithCoder(MutationDetectors.java:46)
>   at 
> org.apache.beam.runners.direct.ImmutabilityCheckingBundleFactory$ImmutabilityEnforcingBundle.add(ImmutabilityCheckingBundleFactory.java:112)
>   at 
> org.apache.beam.runners.direct.ParDoEvaluator$BundleOutputManager.output(ParDoEvaluator.java:299)
>   at 
> org.apache.beam.repackaged.direct_java.runners.core.SimpleDoFnRunner.outputWindowedValue(SimpleDoFnRunner.java:258)
>   at 
> org.apache.beam.repackaged.direct_java.runners.core.SimpleDoFnRunner.access$800(SimpleDoFnRunner.java:78)
>   at 
> org.apache.beam.repackaged.direct_java.runners.core.SimpleDoFnRunner$DoFnProcessContext.output(SimpleDoFnRunner.java:409)
>   at 
> org.apache.beam.repackaged.direct_java.runners.core.SimpleDoFnRunner$DoFnProcessContext.output(SimpleDoFnRunner.java:397)
>   at 
> org.apache.

[jira] [Updated] (BEAM-9523) GROUP BY DOUBLE fails at runtime instead of planning time

2020-03-16 Thread Andrew Pilloud (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-9523?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrew Pilloud updated BEAM-9523:
-
Status: Open  (was: Triage Needed)

> GROUP BY DOUBLE fails at runtime instead of planning time
> -
>
> Key: BEAM-9523
> URL: https://issues.apache.org/jira/browse/BEAM-9523
> Project: Beam
>  Issue Type: Bug
>  Components: dsl-sql-zetasql
>Reporter: Andrew Pilloud
>Priority: Minor
>  Labels: zetasql-compliance
>
> This should fail earlier.
> {code:java}
> Mar 16, 2020 1:27:32 PM 
> cloud.dataflow.sql.ExecuteQueryServiceServer$SqlComplianceServiceImpl 
> executeQuery
> INFO: Processing Sql statement: SELECT COUNT(a) FROM (
> SELECT a FROM (SELECT 1.2 a UNION ALL SELECT 2.3 UNION ALL SELECT 3.4) LIMIT 
> 1)
> Mar 16, 2020 1:27:33 PM 
> com.google.zetasql.io.grpc.internal.SerializingExecutor run
> SEVERE: Exception while executing runnable 
> com.google.zetasql.io.grpc.internal.ServerImpl$JumpToApplicationThreadServerStreamListener$1HalfClosed@7946abe5
> java.lang.IllegalStateException: the keyCoder of a GroupByKey must be 
> deterministic
>   at org.apache.beam.sdk.transforms.GroupByKey.expand(GroupByKey.java:234)
>   at org.apache.beam.sdk.transforms.GroupByKey.expand(GroupByKey.java:110)
>   at org.apache.beam.sdk.Pipeline.applyInternal(Pipeline.java:542)
>   at org.apache.beam.sdk.Pipeline.applyTransform(Pipeline.java:493)
>   at org.apache.beam.sdk.values.PCollection.apply(PCollection.java:368)
>   at 
> org.apache.beam.sdk.transforms.join.CoGroupByKey.expand(CoGroupByKey.java:118)
>   at 
> org.apache.beam.sdk.transforms.join.CoGroupByKey.expand(CoGroupByKey.java:71)
>   at org.apache.beam.sdk.Pipeline.applyInternal(Pipeline.java:542)
>   at org.apache.beam.sdk.Pipeline.applyTransform(Pipeline.java:476)
>   at 
> org.apache.beam.sdk.transforms.join.KeyedPCollectionTuple.apply(KeyedPCollectionTuple.java:108)
>   at 
> org.apache.beam.sdk.extensions.sql.impl.rel.BeamSetOperatorRelBase.expand(BeamSetOperatorRelBase.java:96)
>   at 
> org.apache.beam.sdk.extensions.sql.impl.rel.BeamSetOperatorRelBase.expand(BeamSetOperatorRelBase.java:41)
>   at org.apache.beam.sdk.Pipeline.applyInternal(Pipeline.java:542)
>   at org.apache.beam.sdk.Pipeline.applyTransform(Pipeline.java:493)
>   at 
> org.apache.beam.sdk.extensions.sql.impl.rel.BeamSqlRelUtils.toPCollection(BeamSqlRelUtils.java:69)
>   at 
> org.apache.beam.sdk.extensions.sql.impl.rel.BeamSqlRelUtils.lambda$buildPCollectionList$0(BeamSqlRelUtils.java:50)
>   at 
> java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:193)
>   at java.util.Iterator.forEachRemaining(Iterator.java:116)
>   at 
> java.util.Spliterators$IteratorSpliterator.forEachRemaining(Spliterators.java:1801)
>   at java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:481)
>   at 
> java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:471)
>   at 
> java.util.stream.ReduceOps$ReduceOp.evaluateSequential(ReduceOps.java:708)
>   at java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:234)
>   at 
> java.util.stream.ReferencePipeline.collect(ReferencePipeline.java:499)
>   at 
> org.apache.beam.sdk.extensions.sql.impl.rel.BeamSqlRelUtils.buildPCollectionList(BeamSqlRelUtils.java:51)
>   at 
> org.apache.beam.sdk.extensions.sql.impl.rel.BeamSqlRelUtils.toPCollection(BeamSqlRelUtils.java:67)
>   at 
> org.apache.beam.sdk.extensions.sql.impl.rel.BeamSqlRelUtils.lambda$buildPCollectionList$0(BeamSqlRelUtils.java:50)
>   at 
> java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:193)
>   at java.util.Iterator.forEachRemaining(Iterator.java:116)
>   at 
> java.util.Spliterators$IteratorSpliterator.forEachRemaining(Spliterators.java:1801)
>   at java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:481)
>   at 
> java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:471)
>   at 
> java.util.stream.ReduceOps$ReduceOp.evaluateSequential(ReduceOps.java:708)
>   at java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:234)
>   at 
> java.util.stream.ReferencePipeline.collect(ReferencePipeline.java:499)
>   at 
> org.apache.beam.sdk.extensions.sql.impl.rel.BeamSqlRelUtils.buildPCollectionList(BeamSqlRelUtils.java:51)
>   at 
> org.apache.beam.sdk.extensions.sql.impl.rel.BeamSqlRelUtils.toPCollection(BeamSqlRelUtils.java:67)
>   at 
> org.apache.beam.sdk.extensions.sql.impl.rel.BeamSqlRelUtils.lambda$buildPCollectionList$0(BeamSqlRelUtils.java:50)
>   at 
> java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:193)
>   at java.util.Iterator.forEachRemaining(Iterator.java:116)
>

[jira] [Updated] (BEAM-9517) BeamZetaSqlCalcRel$CalcFn.processElement Field is not nullable

2020-03-16 Thread Andrew Pilloud (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-9517?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrew Pilloud updated BEAM-9517:
-
Status: Open  (was: Triage Needed)

> BeamZetaSqlCalcRel$CalcFn.processElement Field is not nullable
> --
>
> Key: BEAM-9517
> URL: https://issues.apache.org/jira/browse/BEAM-9517
> Project: Beam
>  Issue Type: Bug
>  Components: dsl-sql-zetasql
>Reporter: Andrew Pilloud
>Priority: Minor
>  Labels: zetasql-compliance
>
> {code:java}
> Mar 16, 2020 1:24:42 PM 
> cloud.dataflow.sql.ExecuteQueryServiceServer$SqlComplianceServiceImpl 
> executeQuery
> INFO: Processing Sql statement: SELECT timestamp_diff(@p0, @p1, MICROSECOND) 
> AS ColA
> Mar 16, 2020 1:24:43 PM 
> com.google.zetasql.io.grpc.internal.SerializingExecutor run
> SEVERE: Exception while executing runnable 
> com.google.zetasql.io.grpc.internal.ServerImpl$JumpToApplicationThreadServerStreamListener$1HalfClosed@20ca06
> org.apache.beam.sdk.Pipeline$PipelineExecutionException: 
> java.lang.IllegalArgumentException: Field ColA is not nullable
>   at 
> org.apache.beam.runners.direct.DirectRunner$DirectPipelineResult.waitUntilFinish(DirectRunner.java:348)
>   at 
> org.apache.beam.runners.direct.DirectRunner$DirectPipelineResult.waitUntilFinish(DirectRunner.java:318)
>   at 
> org.apache.beam.runners.direct.DirectRunner.run(DirectRunner.java:213)
>   at org.apache.beam.runners.direct.DirectRunner.run(DirectRunner.java:67)
>   at org.apache.beam.sdk.Pipeline.run(Pipeline.java:317)
>   at org.apache.beam.sdk.Pipeline.run(Pipeline.java:303)
>   at 
> org.apache.beam.sdk.extensions.sql.impl.rel.BeamEnumerableConverter.runCollector(BeamEnumerableConverter.java:201)
>   at 
> org.apache.beam.sdk.extensions.sql.impl.rel.BeamEnumerableConverter.collectRows(BeamEnumerableConverter.java:218)
>   at 
> org.apache.beam.sdk.extensions.sql.impl.rel.BeamEnumerableConverter.toRowList(BeamEnumerableConverter.java:150)
>   at 
> org.apache.beam.sdk.extensions.sql.impl.rel.BeamEnumerableConverter.toRowList(BeamEnumerableConverter.java:127)
>   at 
> cloud.dataflow.sql.ExecuteQueryServiceServer$SqlComplianceServiceImpl.executeQuery(ExecuteQueryServiceServer.java:244)
>   at 
> com.google.zetasql.testing.SqlComplianceServiceGrpc$MethodHandlers.invoke(SqlComplianceServiceGrpc.java:423)
>   at 
> com.google.zetasql.io.grpc.stub.ServerCalls$UnaryServerCallHandler$UnaryServerCallListener.onHalfClose(ServerCalls.java:171)
>   at 
> com.google.zetasql.io.grpc.internal.ServerCallImpl$ServerStreamListenerImpl.halfClosed(ServerCallImpl.java:283)
>   at 
> com.google.zetasql.io.grpc.internal.ServerImpl$JumpToApplicationThreadServerStreamListener$1HalfClosed.runInContext(ServerImpl.java:711)
>   at 
> com.google.zetasql.io.grpc.internal.ContextRunnable.run(ContextRunnable.java:37)
>   at 
> com.google.zetasql.io.grpc.internal.SerializingExecutor.run(SerializingExecutor.java:123)
>   at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
>   at java.lang.Thread.run(Thread.java:748)
> Caused by: java.lang.IllegalArgumentException: Field ColA is not nullable
>   at org.apache.beam.sdk.values.Row$Builder.verify(Row.java:635)
>   at org.apache.beam.sdk.values.Row$Builder.build(Row.java:844)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.BeamZetaSqlCalcRel$CalcFn.processElement(BeamZetaSqlCalcRel.java:195)
>  {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (BEAM-9514) AssertionError type mismatch from SUM

2020-03-16 Thread Andrew Pilloud (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-9514?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrew Pilloud updated BEAM-9514:
-
Status: Open  (was: Triage Needed)

> AssertionError type mismatch from SUM
> -
>
> Key: BEAM-9514
> URL: https://issues.apache.org/jira/browse/BEAM-9514
> Project: Beam
>  Issue Type: Bug
>  Components: dsl-sql-zetasql
>Reporter: Andrew Pilloud
>Priority: Major
>  Labels: zetasql-compliance
>
> {code:java}
> Mar 16, 2020 12:59:49 PM 
> cloud.dataflow.sql.ExecuteQueryServiceServer$SqlComplianceServiceImpl 
> executeQuery
> INFO: Processing Sql statement: select sum(distinct_4) from TableDistincts
> group by distinct_2
> having false
> Exception in thread "pool-1-thread-1" java.lang.AssertionError: Type mismatch:
> rowtype of new rel:
> RecordType(BIGINT distinct_2, BIGINT $col1) NOT NULL
> rowtype of set:
> RecordType(BIGINT distinct_2, BIGINT NOT NULL $col1) NOT NULL
>   at 
> org.apache.beam.vendor.calcite.v1_20_0.org.apache.calcite.util.Litmus$1.fail(Litmus.java:31)
>   at 
> org.apache.beam.vendor.calcite.v1_20_0.org.apache.calcite.plan.RelOptUtil.equal(RelOptUtil.java:1984)
>   at 
> org.apache.beam.vendor.calcite.v1_20_0.org.apache.calcite.plan.volcano.RelSubset.add(RelSubset.java:284)
>   at 
> org.apache.beam.vendor.calcite.v1_20_0.org.apache.calcite.plan.volcano.RelSet.add(RelSet.java:148)
>   at 
> org.apache.beam.vendor.calcite.v1_20_0.org.apache.calcite.plan.volcano.VolcanoPlanner.addRelToSet(VolcanoPlanner.java:1806)
>   at 
> org.apache.beam.vendor.calcite.v1_20_0.org.apache.calcite.plan.volcano.VolcanoPlanner.reregister(VolcanoPlanner.java:1480)
>   at 
> org.apache.beam.vendor.calcite.v1_20_0.org.apache.calcite.plan.volcano.RelSet.mergeWith(RelSet.java:331)
>   at 
> org.apache.beam.vendor.calcite.v1_20_0.org.apache.calcite.plan.volcano.VolcanoPlanner.merge(VolcanoPlanner.java:1571)
>   at 
> org.apache.beam.vendor.calcite.v1_20_0.org.apache.calcite.plan.volcano.VolcanoPlanner.ensureRegistered(VolcanoPlanner.java:863)
>   at 
> org.apache.beam.vendor.calcite.v1_20_0.org.apache.calcite.plan.volcano.VolcanoPlanner.ensureRegistered(VolcanoPlanner.java:1927)
>   at 
> org.apache.beam.vendor.calcite.v1_20_0.org.apache.calcite.plan.volcano.VolcanoRuleCall.transformTo(VolcanoRuleCall.java:129)
>   at 
> org.apache.beam.vendor.calcite.v1_20_0.org.apache.calcite.plan.RelOptRuleCall.transformTo(RelOptRuleCall.java:236)
>   at 
> org.apache.beam.vendor.calcite.v1_20_0.org.apache.calcite.rel.rules.AggregateRemoveRule.onMatch(AggregateRemoveRule.java:126)
>   at 
> org.apache.beam.vendor.calcite.v1_20_0.org.apache.calcite.plan.volcano.VolcanoRuleCall.onMatch(VolcanoRuleCall.java:208)
>   at 
> org.apache.beam.vendor.calcite.v1_20_0.org.apache.calcite.plan.volcano.VolcanoPlanner.findBestExp(VolcanoPlanner.java:631)
>   at 
> org.apache.beam.vendor.calcite.v1_20_0.org.apache.calcite.tools.Programs$RuleSetProgram.run(Programs.java:328)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLPlannerImpl.transform(ZetaSQLPlannerImpl.java:180)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLQueryPlanner.convertToBeamRelInternal(ZetaSQLQueryPlanner.java:150)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLQueryPlanner.convertToBeamRel(ZetaSQLQueryPlanner.java:115)
>   at 
> cloud.dataflow.sql.ExecuteQueryServiceServer$SqlComplianceServiceImpl.executeQuery(ExecuteQueryServiceServer.java:242)
>   at 
> com.google.zetasql.testing.SqlComplianceServiceGrpc$MethodHandlers.invoke(SqlComplianceServiceGrpc.java:423)
>   at 
> com.google.zetasql.io.grpc.stub.ServerCalls$UnaryServerCallHandler$UnaryServerCallListener.onHalfClose(ServerCalls.java:171)
>   at 
> com.google.zetasql.io.grpc.internal.ServerCallImpl$ServerStreamListenerImpl.halfClosed(ServerCallImpl.java:283)
>   at 
> com.google.zetasql.io.grpc.internal.ServerImpl$JumpToApplicationThreadServerStreamListener$1HalfClosed.runInContext(ServerImpl.java:711)
>   at 
> com.google.zetasql.io.grpc.internal.ContextRunnable.run(ContextRunnable.java:37)
>   at 
> com.google.zetasql.io.grpc.internal.SerializingExecutor.run(SerializingExecutor.java:123)
>   at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
>   at java.lang.Thread.run(Thread.java:748) {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (BEAM-9521) NullPointerException in convertRexNodeFromResolvedExpr

2020-03-16 Thread Andrew Pilloud (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-9521?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrew Pilloud updated BEAM-9521:
-
Status: Open  (was: Triage Needed)

> NullPointerException in convertRexNodeFromResolvedExpr
> --
>
> Key: BEAM-9521
> URL: https://issues.apache.org/jira/browse/BEAM-9521
> Project: Beam
>  Issue Type: Bug
>  Components: dsl-sql-zetasql
>Reporter: Andrew Pilloud
>Priority: Major
>  Labels: zetasql-compliance
>
> {code:java}
> Mar 16, 2020 12:57:10 PM 
> cloud.dataflow.sql.ExecuteQueryServiceServer$SqlComplianceServiceImpl 
> executeQuery
> INFO: Processing Sql statement: SELECT (SELECT AS STRUCT
> (SELECT AS STRUCT l.float_val lf, r.float_val rf),
> l.float_val = r.float_val),
>(SELECT AS STRUCT
> (SELECT AS STRUCT l.double_val ld, r.double_val rd),
> l.double_val = r.double_val)
> FROM TestTableZero l, TestTableZero r
> WHERE l.row_id > r.row_id
> Mar 16, 2020 12:57:10 PM 
> com.google.zetasql.io.grpc.internal.SerializingExecutor run
> SEVERE: Exception while executing runnable 
> com.google.zetasql.io.grpc.internal.ServerImpl$JumpToApplicationThreadServerStreamListener$1HalfClosed@710f8c0d
> java.lang.NullPointerException
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.ExpressionConverter.convertRexNodeFromResolvedExpr(ExpressionConverter.java:301)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.JoinScanConverter.convert(JoinScanConverter.java:83)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.JoinScanConverter.convert(JoinScanConverter.java:36)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.QueryStatementConverter.convertNode(QueryStatementConverter.java:97)
>   at 
> java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:193)
>   at java.util.Collections$2.tryAdvance(Collections.java:4717)
>   at java.util.Collections$2.forEachRemaining(Collections.java:4725)
>   at java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:481)
>   at 
> java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:471)
>   at 
> java.util.stream.ReduceOps$ReduceOp.evaluateSequential(ReduceOps.java:708)
>   at java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:234)
>   at 
> java.util.stream.ReferencePipeline.collect(ReferencePipeline.java:499)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.QueryStatementConverter.convertNode(QueryStatementConverter.java:96)
>   at 
> java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:193)
>   at java.util.Collections$2.tryAdvance(Collections.java:4717)
>   at java.util.Collections$2.forEachRemaining(Collections.java:4725)
>   at java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:481)
>   at 
> java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:471)
>   at 
> java.util.stream.ReduceOps$ReduceOp.evaluateSequential(ReduceOps.java:708)
>   at java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:234)
>   at 
> java.util.stream.ReferencePipeline.collect(ReferencePipeline.java:499)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.QueryStatementConverter.convertNode(QueryStatementConverter.java:96)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.QueryStatementConverter.convert(QueryStatementConverter.java:84)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.QueryStatementConverter.convertRootQuery(QueryStatementConverter.java:51)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLPlannerImpl.rel(ZetaSQLPlannerImpl.java:160)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLQueryPlanner.convertToBeamRelInternal(ZetaSQLQueryPlanner.java:131)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLQueryPlanner.convertToBeamRel(ZetaSQLQueryPlanner.java:115)
>   at 
> cloud.dataflow.sql.ExecuteQueryServiceServer$SqlComplianceServiceImpl.executeQuery(ExecuteQueryServiceServer.java:242)
>   at 
> com.google.zetasql.testing.SqlComplianceServiceGrpc$MethodHandlers.invoke(SqlComplianceServiceGrpc.java:423)
>   at 
> com.google.zetasql.io.grpc.stub.ServerCalls$UnaryServerCallHandler$UnaryServerCallListener.onHalfClose(ServerCalls.java:171)
>   at 
> com.google.zetasql.io.grpc.internal.ServerCallImpl$ServerStreamListenerImpl.halfClosed(ServerCallImpl.java:283)
>   at 
> com.google.zetasql.io.grpc.internal.ServerImpl$JumpToApplicationThreadServerStreamListener$1HalfClosed.runInContext(ServerImpl.java:711)
>   at 
> com.google.zetasql.io.grpc.internal.ContextRunnable.run(ContextRunnable.java:37)
>   at 
> com.google.zetasql.io

[jira] [Updated] (BEAM-9515) ResolvedSubqueryExpr can not be cast to ResolvedLiteral

2020-03-16 Thread Andrew Pilloud (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-9515?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrew Pilloud updated BEAM-9515:
-
Status: Open  (was: Triage Needed)

> ResolvedSubqueryExpr can not be cast to ResolvedLiteral
> ---
>
> Key: BEAM-9515
> URL: https://issues.apache.org/jira/browse/BEAM-9515
> Project: Beam
>  Issue Type: Bug
>  Components: dsl-sql-zetasql
>Reporter: Andrew Pilloud
>Priority: Major
>  Labels: zetasql-compliance
>
> {code:java}
> Mar 16, 2020 1:00:02 PM 
> cloud.dataflow.sql.ExecuteQueryServiceServer$SqlComplianceServiceImpl 
> executeQuery
> INFO: Processing Sql statement: SELECT * FROM UNNEST(ARRAY(
>   SELECT bool_val FROM AllTypesTable t
>   ORDER BY bool_val ASC
> )) x WITH OFFSET POS
> Mar 16, 2020 1:00:02 PM 
> com.google.zetasql.io.grpc.internal.SerializingExecutor run
> SEVERE: Exception while executing runnable 
> com.google.zetasql.io.grpc.internal.ServerImpl$JumpToApplicationThreadServerStreamListener$1HalfClosed@7b42f9e7
> java.lang.ClassCastException: 
> com.google.zetasql.resolvedast.ResolvedNodes$ResolvedSubqueryExpr cannot be 
> cast to com.google.zetasql.resolvedast.ResolvedNodes$ResolvedLiteral
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.ArrayScanToUncollectConverter.convert(ArrayScanToUncollectConverter.java:45)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.ArrayScanToUncollectConverter.convert(ArrayScanToUncollectConverter.java:31)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.QueryStatementConverter.convertNode(QueryStatementConverter.java:97)
>   at 
> java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:193)
>   at java.util.Collections$2.tryAdvance(Collections.java:4717)
>   at java.util.Collections$2.forEachRemaining(Collections.java:4725)
>   at java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:481)
>   at 
> java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:471)
>   at 
> java.util.stream.ReduceOps$ReduceOp.evaluateSequential(ReduceOps.java:708)
>   at java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:234)
>   at 
> java.util.stream.ReferencePipeline.collect(ReferencePipeline.java:499)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.QueryStatementConverter.convertNode(QueryStatementConverter.java:96)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.QueryStatementConverter.convert(QueryStatementConverter.java:84)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.translation.QueryStatementConverter.convertRootQuery(QueryStatementConverter.java:51)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLPlannerImpl.rel(ZetaSQLPlannerImpl.java:160)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLQueryPlanner.convertToBeamRelInternal(ZetaSQLQueryPlanner.java:131)
>   at 
> org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLQueryPlanner.convertToBeamRel(ZetaSQLQueryPlanner.java:115)
>   at 
> cloud.dataflow.sql.ExecuteQueryServiceServer$SqlComplianceServiceImpl.executeQuery(ExecuteQueryServiceServer.java:242)
>   at 
> com.google.zetasql.testing.SqlComplianceServiceGrpc$MethodHandlers.invoke(SqlComplianceServiceGrpc.java:423)
>   at 
> com.google.zetasql.io.grpc.stub.ServerCalls$UnaryServerCallHandler$UnaryServerCallListener.onHalfClose(ServerCalls.java:171)
>   at 
> com.google.zetasql.io.grpc.internal.ServerCallImpl$ServerStreamListenerImpl.halfClosed(ServerCallImpl.java:283)
>   at 
> com.google.zetasql.io.grpc.internal.ServerImpl$JumpToApplicationThreadServerStreamListener$1HalfClosed.runInContext(ServerImpl.java:711)
>   at 
> com.google.zetasql.io.grpc.internal.ContextRunnable.run(ContextRunnable.java:37)
>   at 
> com.google.zetasql.io.grpc.internal.SerializingExecutor.run(SerializingExecutor.java:123)
>   at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
>   at java.lang.Thread.run(Thread.java:748)
>  {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (BEAM-9523) GROUP BY DOUBLE fails at runtime instead of planning time

2020-03-16 Thread Andrew Pilloud (Jira)
Andrew Pilloud created BEAM-9523:


 Summary: GROUP BY DOUBLE fails at runtime instead of planning time
 Key: BEAM-9523
 URL: https://issues.apache.org/jira/browse/BEAM-9523
 Project: Beam
  Issue Type: Bug
  Components: dsl-sql-zetasql
Reporter: Andrew Pilloud


This should fail earlier.
{code:java}
Mar 16, 2020 1:27:32 PM 
cloud.dataflow.sql.ExecuteQueryServiceServer$SqlComplianceServiceImpl 
executeQuery
INFO: Processing Sql statement: SELECT COUNT(a) FROM (
SELECT a FROM (SELECT 1.2 a UNION ALL SELECT 2.3 UNION ALL SELECT 3.4) LIMIT 1)
Mar 16, 2020 1:27:33 PM com.google.zetasql.io.grpc.internal.SerializingExecutor 
run
SEVERE: Exception while executing runnable 
com.google.zetasql.io.grpc.internal.ServerImpl$JumpToApplicationThreadServerStreamListener$1HalfClosed@7946abe5
java.lang.IllegalStateException: the keyCoder of a GroupByKey must be 
deterministic
at org.apache.beam.sdk.transforms.GroupByKey.expand(GroupByKey.java:234)
at org.apache.beam.sdk.transforms.GroupByKey.expand(GroupByKey.java:110)
at org.apache.beam.sdk.Pipeline.applyInternal(Pipeline.java:542)
at org.apache.beam.sdk.Pipeline.applyTransform(Pipeline.java:493)
at org.apache.beam.sdk.values.PCollection.apply(PCollection.java:368)
at 
org.apache.beam.sdk.transforms.join.CoGroupByKey.expand(CoGroupByKey.java:118)
at 
org.apache.beam.sdk.transforms.join.CoGroupByKey.expand(CoGroupByKey.java:71)
at org.apache.beam.sdk.Pipeline.applyInternal(Pipeline.java:542)
at org.apache.beam.sdk.Pipeline.applyTransform(Pipeline.java:476)
at 
org.apache.beam.sdk.transforms.join.KeyedPCollectionTuple.apply(KeyedPCollectionTuple.java:108)
at 
org.apache.beam.sdk.extensions.sql.impl.rel.BeamSetOperatorRelBase.expand(BeamSetOperatorRelBase.java:96)
at 
org.apache.beam.sdk.extensions.sql.impl.rel.BeamSetOperatorRelBase.expand(BeamSetOperatorRelBase.java:41)
at org.apache.beam.sdk.Pipeline.applyInternal(Pipeline.java:542)
at org.apache.beam.sdk.Pipeline.applyTransform(Pipeline.java:493)
at 
org.apache.beam.sdk.extensions.sql.impl.rel.BeamSqlRelUtils.toPCollection(BeamSqlRelUtils.java:69)
at 
org.apache.beam.sdk.extensions.sql.impl.rel.BeamSqlRelUtils.lambda$buildPCollectionList$0(BeamSqlRelUtils.java:50)
at 
java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:193)
at java.util.Iterator.forEachRemaining(Iterator.java:116)
at 
java.util.Spliterators$IteratorSpliterator.forEachRemaining(Spliterators.java:1801)
at java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:481)
at 
java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:471)
at 
java.util.stream.ReduceOps$ReduceOp.evaluateSequential(ReduceOps.java:708)
at java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:234)
at 
java.util.stream.ReferencePipeline.collect(ReferencePipeline.java:499)
at 
org.apache.beam.sdk.extensions.sql.impl.rel.BeamSqlRelUtils.buildPCollectionList(BeamSqlRelUtils.java:51)
at 
org.apache.beam.sdk.extensions.sql.impl.rel.BeamSqlRelUtils.toPCollection(BeamSqlRelUtils.java:67)
at 
org.apache.beam.sdk.extensions.sql.impl.rel.BeamSqlRelUtils.lambda$buildPCollectionList$0(BeamSqlRelUtils.java:50)
at 
java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:193)
at java.util.Iterator.forEachRemaining(Iterator.java:116)
at 
java.util.Spliterators$IteratorSpliterator.forEachRemaining(Spliterators.java:1801)
at java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:481)
at 
java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:471)
at 
java.util.stream.ReduceOps$ReduceOp.evaluateSequential(ReduceOps.java:708)
at java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:234)
at 
java.util.stream.ReferencePipeline.collect(ReferencePipeline.java:499)
at 
org.apache.beam.sdk.extensions.sql.impl.rel.BeamSqlRelUtils.buildPCollectionList(BeamSqlRelUtils.java:51)
at 
org.apache.beam.sdk.extensions.sql.impl.rel.BeamSqlRelUtils.toPCollection(BeamSqlRelUtils.java:67)
at 
org.apache.beam.sdk.extensions.sql.impl.rel.BeamSqlRelUtils.lambda$buildPCollectionList$0(BeamSqlRelUtils.java:50)
at 
java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:193)
at java.util.Iterator.forEachRemaining(Iterator.java:116)
at 
java.util.Spliterators$IteratorSpliterator.forEachRemaining(Spliterators.java:1801)
at java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:481)
at 
java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:471)
at 
java.util.stream.ReduceOps$ReduceOp.evaluateSequential(ReduceOps.java:708)

[jira] [Created] (BEAM-9522) BeamJoinRel.extractJoinRexNode RexLiteral cannot be cast to RexCall

2020-03-16 Thread Andrew Pilloud (Jira)
Andrew Pilloud created BEAM-9522:


 Summary: BeamJoinRel.extractJoinRexNode RexLiteral cannot be cast 
to RexCall
 Key: BEAM-9522
 URL: https://issues.apache.org/jira/browse/BEAM-9522
 Project: Beam
  Issue Type: Bug
  Components: dsl-sql-zetasql
Reporter: Andrew Pilloud


{code:java}
Mar 16, 2020 1:26:41 PM 
cloud.dataflow.sql.ExecuteQueryServiceServer$SqlComplianceServiceImpl 
executeQuery
INFO: Processing Sql statement: SELECT r.*, s.* FROM
(SELECT x, x FROM (SELECT 2 x)) r FULL JOIN (SELECT 3) s ON false
Mar 16, 2020 1:26:41 PM com.google.zetasql.io.grpc.internal.SerializingExecutor 
run
SEVERE: Exception while executing runnable 
com.google.zetasql.io.grpc.internal.ServerImpl$JumpToApplicationThreadServerStreamListener$1HalfClosed@16f4bbdb
java.lang.ClassCastException: 
org.apache.beam.vendor.calcite.v1_20_0.org.apache.calcite.rex.RexLiteral cannot 
be cast to org.apache.beam.vendor.calcite.v1_20_0.org.apache.calcite.rex.RexCall
at 
org.apache.beam.sdk.extensions.sql.impl.rel.BeamJoinRel.extractJoinRexNodes(BeamJoinRel.java:304)
at 
org.apache.beam.sdk.extensions.sql.impl.rel.BeamJoinRel$ExtractJoinKeys.expand(BeamJoinRel.java:194)
at 
org.apache.beam.sdk.extensions.sql.impl.rel.BeamJoinRel$ExtractJoinKeys.expand(BeamJoinRel.java:177)
at org.apache.beam.sdk.Pipeline.applyInternal(Pipeline.java:542)
at org.apache.beam.sdk.Pipeline.applyTransform(Pipeline.java:476)
at 
org.apache.beam.sdk.values.PCollectionList.apply(PCollectionList.java:178)
at 
org.apache.beam.sdk.extensions.sql.impl.rel.BeamCoGBKJoinRel$StandardJoin.expand(BeamCoGBKJoinRel.java:93)
at 
org.apache.beam.sdk.extensions.sql.impl.rel.BeamCoGBKJoinRel$StandardJoin.expand(BeamCoGBKJoinRel.java:86)
at org.apache.beam.sdk.Pipeline.applyInternal(Pipeline.java:542)
at org.apache.beam.sdk.Pipeline.applyTransform(Pipeline.java:493)
at 
org.apache.beam.sdk.extensions.sql.impl.rel.BeamSqlRelUtils.toPCollection(BeamSqlRelUtils.java:69)
at 
org.apache.beam.sdk.extensions.sql.impl.rel.BeamSqlRelUtils.lambda$buildPCollectionList$0(BeamSqlRelUtils.java:50)
at 
java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:193)
at java.util.Iterator.forEachRemaining(Iterator.java:116)
at 
java.util.Spliterators$IteratorSpliterator.forEachRemaining(Spliterators.java:1801)
at java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:481)
at 
java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:471)
at 
java.util.stream.ReduceOps$ReduceOp.evaluateSequential(ReduceOps.java:708)
at java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:234)
at 
java.util.stream.ReferencePipeline.collect(ReferencePipeline.java:499)
at 
org.apache.beam.sdk.extensions.sql.impl.rel.BeamSqlRelUtils.buildPCollectionList(BeamSqlRelUtils.java:51)
at 
org.apache.beam.sdk.extensions.sql.impl.rel.BeamSqlRelUtils.toPCollection(BeamSqlRelUtils.java:67)
at 
org.apache.beam.sdk.extensions.sql.impl.rel.BeamSqlRelUtils.toPCollection(BeamSqlRelUtils.java:39)
at 
org.apache.beam.sdk.extensions.sql.impl.rel.BeamEnumerableConverter.runCollector(BeamEnumerableConverter.java:199)
at 
org.apache.beam.sdk.extensions.sql.impl.rel.BeamEnumerableConverter.collectRows(BeamEnumerableConverter.java:218)
at 
org.apache.beam.sdk.extensions.sql.impl.rel.BeamEnumerableConverter.toRowList(BeamEnumerableConverter.java:150)
at 
org.apache.beam.sdk.extensions.sql.impl.rel.BeamEnumerableConverter.toRowList(BeamEnumerableConverter.java:127)
at 
cloud.dataflow.sql.ExecuteQueryServiceServer$SqlComplianceServiceImpl.executeQuery(ExecuteQueryServiceServer.java:244)
at 
com.google.zetasql.testing.SqlComplianceServiceGrpc$MethodHandlers.invoke(SqlComplianceServiceGrpc.java:423)
at 
com.google.zetasql.io.grpc.stub.ServerCalls$UnaryServerCallHandler$UnaryServerCallListener.onHalfClose(ServerCalls.java:171)
at 
com.google.zetasql.io.grpc.internal.ServerCallImpl$ServerStreamListenerImpl.halfClosed(ServerCallImpl.java:283)
at 
com.google.zetasql.io.grpc.internal.ServerImpl$JumpToApplicationThreadServerStreamListener$1HalfClosed.runInContext(ServerImpl.java:711)
at 
com.google.zetasql.io.grpc.internal.ContextRunnable.run(ContextRunnable.java:37)
at 
com.google.zetasql.io.grpc.internal.SerializingExecutor.run(SerializingExecutor.java:123)
at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
 {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (BEAM-9521) NullPointerException in convertRexNodeFromResolvedExpr

2020-03-16 Thread Andrew Pilloud (Jira)
Andrew Pilloud created BEAM-9521:


 Summary: NullPointerException in convertRexNodeFromResolvedExpr
 Key: BEAM-9521
 URL: https://issues.apache.org/jira/browse/BEAM-9521
 Project: Beam
  Issue Type: Bug
  Components: dsl-sql-zetasql
Reporter: Andrew Pilloud


{code:java}
Mar 16, 2020 12:57:10 PM 
cloud.dataflow.sql.ExecuteQueryServiceServer$SqlComplianceServiceImpl 
executeQuery
INFO: Processing Sql statement: SELECT (SELECT AS STRUCT
(SELECT AS STRUCT l.float_val lf, r.float_val rf),
l.float_val = r.float_val),
   (SELECT AS STRUCT
(SELECT AS STRUCT l.double_val ld, r.double_val rd),
l.double_val = r.double_val)
FROM TestTableZero l, TestTableZero r
WHERE l.row_id > r.row_id
Mar 16, 2020 12:57:10 PM 
com.google.zetasql.io.grpc.internal.SerializingExecutor run
SEVERE: Exception while executing runnable 
com.google.zetasql.io.grpc.internal.ServerImpl$JumpToApplicationThreadServerStreamListener$1HalfClosed@710f8c0d
java.lang.NullPointerException
at 
org.apache.beam.sdk.extensions.sql.zetasql.translation.ExpressionConverter.convertRexNodeFromResolvedExpr(ExpressionConverter.java:301)
at 
org.apache.beam.sdk.extensions.sql.zetasql.translation.JoinScanConverter.convert(JoinScanConverter.java:83)
at 
org.apache.beam.sdk.extensions.sql.zetasql.translation.JoinScanConverter.convert(JoinScanConverter.java:36)
at 
org.apache.beam.sdk.extensions.sql.zetasql.translation.QueryStatementConverter.convertNode(QueryStatementConverter.java:97)
at 
java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:193)
at java.util.Collections$2.tryAdvance(Collections.java:4717)
at java.util.Collections$2.forEachRemaining(Collections.java:4725)
at java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:481)
at 
java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:471)
at 
java.util.stream.ReduceOps$ReduceOp.evaluateSequential(ReduceOps.java:708)
at java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:234)
at 
java.util.stream.ReferencePipeline.collect(ReferencePipeline.java:499)
at 
org.apache.beam.sdk.extensions.sql.zetasql.translation.QueryStatementConverter.convertNode(QueryStatementConverter.java:96)
at 
java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:193)
at java.util.Collections$2.tryAdvance(Collections.java:4717)
at java.util.Collections$2.forEachRemaining(Collections.java:4725)
at java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:481)
at 
java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:471)
at 
java.util.stream.ReduceOps$ReduceOp.evaluateSequential(ReduceOps.java:708)
at java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:234)
at 
java.util.stream.ReferencePipeline.collect(ReferencePipeline.java:499)
at 
org.apache.beam.sdk.extensions.sql.zetasql.translation.QueryStatementConverter.convertNode(QueryStatementConverter.java:96)
at 
org.apache.beam.sdk.extensions.sql.zetasql.translation.QueryStatementConverter.convert(QueryStatementConverter.java:84)
at 
org.apache.beam.sdk.extensions.sql.zetasql.translation.QueryStatementConverter.convertRootQuery(QueryStatementConverter.java:51)
at 
org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLPlannerImpl.rel(ZetaSQLPlannerImpl.java:160)
at 
org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLQueryPlanner.convertToBeamRelInternal(ZetaSQLQueryPlanner.java:131)
at 
org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLQueryPlanner.convertToBeamRel(ZetaSQLQueryPlanner.java:115)
at 
cloud.dataflow.sql.ExecuteQueryServiceServer$SqlComplianceServiceImpl.executeQuery(ExecuteQueryServiceServer.java:242)
at 
com.google.zetasql.testing.SqlComplianceServiceGrpc$MethodHandlers.invoke(SqlComplianceServiceGrpc.java:423)
at 
com.google.zetasql.io.grpc.stub.ServerCalls$UnaryServerCallHandler$UnaryServerCallListener.onHalfClose(ServerCalls.java:171)
at 
com.google.zetasql.io.grpc.internal.ServerCallImpl$ServerStreamListenerImpl.halfClosed(ServerCallImpl.java:283)
at 
com.google.zetasql.io.grpc.internal.ServerImpl$JumpToApplicationThreadServerStreamListener$1HalfClosed.runInContext(ServerImpl.java:711)
at 
com.google.zetasql.io.grpc.internal.ContextRunnable.run(ContextRunnable.java:37)
at 
com.google.zetasql.io.grpc.internal.SerializingExecutor.run(SerializingExecutor.java:123)
at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
 {code}



--
This message was sent by Atlassian Jira
(v8.3.4#80

[jira] [Created] (BEAM-9520) Forbidden IOException when reading from InputStream

2020-03-16 Thread Andrew Pilloud (Jira)
Andrew Pilloud created BEAM-9520:


 Summary: Forbidden IOException when reading from InputStream
 Key: BEAM-9520
 URL: https://issues.apache.org/jira/browse/BEAM-9520
 Project: Beam
  Issue Type: Bug
  Components: dsl-sql-zetasql
Reporter: Andrew Pilloud


{code:java}
Mar 16, 2020 1:26:17 PM 
cloud.dataflow.sql.ExecuteQueryServiceServer$SqlComplianceServiceImpl 
executeQuery
INFO: Processing Sql statement: (SELECT 1 UNION ALL SELECT NULL UNION ALL 
SELECT 1 UNION ALL SELECT NULL)
INTERSECT DISTINCT
(SELECT 1 UNION ALL SELECT NULL UNION ALL SELECT NULL)
Mar 16, 2020 1:26:18 PM com.google.zetasql.io.grpc.internal.SerializingExecutor 
run
SEVERE: Exception while executing runnable 
com.google.zetasql.io.grpc.internal.ServerImpl$JumpToApplicationThreadServerStreamListener$1HalfClosed@7f643e84
org.apache.beam.sdk.Pipeline$PipelineExecutionException: 
java.lang.IllegalArgumentException: Forbidden IOException when reading from 
InputStream
at 
org.apache.beam.runners.direct.DirectRunner$DirectPipelineResult.waitUntilFinish(DirectRunner.java:348)
at 
org.apache.beam.runners.direct.DirectRunner$DirectPipelineResult.waitUntilFinish(DirectRunner.java:318)
at 
org.apache.beam.runners.direct.DirectRunner.run(DirectRunner.java:213)
at org.apache.beam.runners.direct.DirectRunner.run(DirectRunner.java:67)
at org.apache.beam.sdk.Pipeline.run(Pipeline.java:317)
at org.apache.beam.sdk.Pipeline.run(Pipeline.java:303)
at 
org.apache.beam.sdk.extensions.sql.impl.rel.BeamEnumerableConverter.runCollector(BeamEnumerableConverter.java:201)
at 
org.apache.beam.sdk.extensions.sql.impl.rel.BeamEnumerableConverter.collectRows(BeamEnumerableConverter.java:218)
at 
org.apache.beam.sdk.extensions.sql.impl.rel.BeamEnumerableConverter.toRowList(BeamEnumerableConverter.java:150)
at 
org.apache.beam.sdk.extensions.sql.impl.rel.BeamEnumerableConverter.toRowList(BeamEnumerableConverter.java:127)
at 
cloud.dataflow.sql.ExecuteQueryServiceServer$SqlComplianceServiceImpl.executeQuery(ExecuteQueryServiceServer.java:244)
at 
com.google.zetasql.testing.SqlComplianceServiceGrpc$MethodHandlers.invoke(SqlComplianceServiceGrpc.java:423)
at 
com.google.zetasql.io.grpc.stub.ServerCalls$UnaryServerCallHandler$UnaryServerCallListener.onHalfClose(ServerCalls.java:171)
at 
com.google.zetasql.io.grpc.internal.ServerCallImpl$ServerStreamListenerImpl.halfClosed(ServerCallImpl.java:283)
at 
com.google.zetasql.io.grpc.internal.ServerImpl$JumpToApplicationThreadServerStreamListener$1HalfClosed.runInContext(ServerImpl.java:711)
at 
com.google.zetasql.io.grpc.internal.ContextRunnable.run(ContextRunnable.java:37)
at 
com.google.zetasql.io.grpc.internal.SerializingExecutor.run(SerializingExecutor.java:123)
at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
Caused by: java.lang.IllegalArgumentException: Forbidden IOException when 
reading from InputStream
at 
org.apache.beam.sdk.util.CoderUtils.decodeFromSafeStream(CoderUtils.java:118)
at 
org.apache.beam.sdk.util.CoderUtils.decodeFromByteArray(CoderUtils.java:98)
at 
org.apache.beam.sdk.util.CoderUtils.decodeFromByteArray(CoderUtils.java:92)
at org.apache.beam.sdk.util.CoderUtils.clone(CoderUtils.java:141)
at 
org.apache.beam.sdk.util.MutationDetectors$CodedValueMutationDetector.(MutationDetectors.java:115)
at 
org.apache.beam.sdk.util.MutationDetectors.forValueWithCoder(MutationDetectors.java:46)
at 
org.apache.beam.runners.direct.ImmutabilityCheckingBundleFactory$ImmutabilityEnforcingBundle.add(ImmutabilityCheckingBundleFactory.java:112)
at 
org.apache.beam.runners.direct.ParDoEvaluator$BundleOutputManager.output(ParDoEvaluator.java:299)
at 
org.apache.beam.repackaged.direct_java.runners.core.SimpleDoFnRunner.outputWindowedValue(SimpleDoFnRunner.java:258)
at 
org.apache.beam.repackaged.direct_java.runners.core.SimpleDoFnRunner.access$800(SimpleDoFnRunner.java:78)
at 
org.apache.beam.repackaged.direct_java.runners.core.SimpleDoFnRunner$DoFnProcessContext.output(SimpleDoFnRunner.java:409)
at 
org.apache.beam.repackaged.direct_java.runners.core.SimpleDoFnRunner$DoFnProcessContext.output(SimpleDoFnRunner.java:397)
at 
org.apache.beam.sdk.transforms.join.CoGroupByKey$ConstructUnionTableFn.processElement(CoGroupByKey.java:175)
Caused by: java.io.EOFException: reached end of stream after reading 0 bytes; 1 
bytes expected
at 
org.apache.beam.vendor.guava.v26_0_jre.com.google.common.io.ByteStreams.readFully(ByteStreams.java:780)
at 
org.apache.beam.vendor.guava.v26_0_jre.com.go

[jira] [Created] (BEAM-9519) BeamZetaSqlCalcRel$CalcFn.processElement field is not nullable in Array field $arrayx

2020-03-16 Thread Andrew Pilloud (Jira)
Andrew Pilloud created BEAM-9519:


 Summary: BeamZetaSqlCalcRel$CalcFn.processElement field is not 
nullable in Array field $arrayx
 Key: BEAM-9519
 URL: https://issues.apache.org/jira/browse/BEAM-9519
 Project: Beam
  Issue Type: Bug
  Components: dsl-sql-zetasql
Reporter: Andrew Pilloud


{code:java}
Mar 16, 2020 1:27:47 PM 
cloud.dataflow.sql.ExecuteQueryServiceServer$SqlComplianceServiceImpl 
executeQuery
INFO: Processing Sql statement: SELECT x FROM UNNEST([3, NULL]) x;
Mar 16, 2020 1:27:48 PM com.google.zetasql.io.grpc.internal.SerializingExecutor 
run
SEVERE: Exception while executing runnable 
com.google.zetasql.io.grpc.internal.ServerImpl$JumpToApplicationThreadServerStreamListener$1HalfClosed@39086547
org.apache.beam.sdk.Pipeline$PipelineExecutionException: 
java.lang.IllegalArgumentException: FieldType{typeName=INT64, nullable=false, 
logicalType=null, collectionElementType=null, mapKeyType=null, 
mapValueType=null, rowSchema=null, metadata={}} is not nullable in Array field 
$arrayx
at 
org.apache.beam.runners.direct.DirectRunner$DirectPipelineResult.waitUntilFinish(DirectRunner.java:348)
at 
org.apache.beam.runners.direct.DirectRunner$DirectPipelineResult.waitUntilFinish(DirectRunner.java:318)
at 
org.apache.beam.runners.direct.DirectRunner.run(DirectRunner.java:213)
at org.apache.beam.runners.direct.DirectRunner.run(DirectRunner.java:67)
at org.apache.beam.sdk.Pipeline.run(Pipeline.java:317)
at org.apache.beam.sdk.Pipeline.run(Pipeline.java:303)
at 
org.apache.beam.sdk.extensions.sql.impl.rel.BeamEnumerableConverter.runCollector(BeamEnumerableConverter.java:201)
at 
org.apache.beam.sdk.extensions.sql.impl.rel.BeamEnumerableConverter.collectRows(BeamEnumerableConverter.java:218)
at 
org.apache.beam.sdk.extensions.sql.impl.rel.BeamEnumerableConverter.toRowList(BeamEnumerableConverter.java:150)
at 
org.apache.beam.sdk.extensions.sql.impl.rel.BeamEnumerableConverter.toRowList(BeamEnumerableConverter.java:127)
at 
cloud.dataflow.sql.ExecuteQueryServiceServer$SqlComplianceServiceImpl.executeQuery(ExecuteQueryServiceServer.java:244)
at 
com.google.zetasql.testing.SqlComplianceServiceGrpc$MethodHandlers.invoke(SqlComplianceServiceGrpc.java:423)
at 
com.google.zetasql.io.grpc.stub.ServerCalls$UnaryServerCallHandler$UnaryServerCallListener.onHalfClose(ServerCalls.java:171)
at 
com.google.zetasql.io.grpc.internal.ServerCallImpl$ServerStreamListenerImpl.halfClosed(ServerCallImpl.java:283)
at 
com.google.zetasql.io.grpc.internal.ServerImpl$JumpToApplicationThreadServerStreamListener$1HalfClosed.runInContext(ServerImpl.java:711)
at 
com.google.zetasql.io.grpc.internal.ContextRunnable.run(ContextRunnable.java:37)
at 
com.google.zetasql.io.grpc.internal.SerializingExecutor.run(SerializingExecutor.java:123)
at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
Caused by: java.lang.IllegalArgumentException: FieldType{typeName=INT64, 
nullable=false, logicalType=null, collectionElementType=null, mapKeyType=null, 
mapValueType=null, rowSchema=null, metadata={}} is not nullable in Array field 
$arrayx
at org.apache.beam.sdk.values.Row$Builder.verifyArray(Row.java:682)
at org.apache.beam.sdk.values.Row$Builder.verify(Row.java:647)
at org.apache.beam.sdk.values.Row$Builder.verify(Row.java:639)
at org.apache.beam.sdk.values.Row$Builder.build(Row.java:844)
at 
org.apache.beam.sdk.extensions.sql.zetasql.BeamZetaSqlCalcRel$CalcFn.processElement(BeamZetaSqlCalcRel.java:195)
{code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Assigned] (BEAM-9510) Dependencies in base_image_requirements.txt are not compatible with each other

2020-03-16 Thread Ahmet Altay (Jira)


 [ 
https://issues.apache.org/jira/browse/BEAM-9510?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ahmet Altay reassigned BEAM-9510:
-

Assignee: Hannah Jiang

> Dependencies in base_image_requirements.txt are not compatible with each other
> --
>
> Key: BEAM-9510
> URL: https://issues.apache.org/jira/browse/BEAM-9510
> Project: Beam
>  Issue Type: Bug
>  Components: sdk-py-harness
>Reporter: David Yan
>Assignee: Hannah Jiang
>Priority: Major
>
> [https://github.com/apache/beam/blob/master/sdks/python/container/base_image_requirements.txt#L56]
> says it requires google-cloud-bigquery==1.24.0, google-cloud-core==1.0.2, 
> google-cloud-bigtable==0.32.1, grpc-1.22.0 and tensorflow-2.1.0
> But they are incompatible with each other:
> ERROR: google-cloud-bigquery 1.24.0 has requirement 
> google-cloud-core<2.0dev,>=1.1.0, but you'll have google-cloud-core 1.0.2 
> which is incompatible.
> ERROR: google-cloud-bigtable 0.32.1 has requirement 
> google-cloud-core<0.30dev,>=0.29.0, but you'll have google-cloud-core 1.0.2 
> which is incompatible.
> ERROR: tensorboard 2.1.1 has requirement grpcio>=1.24.3, but you'll have 
> grpcio 1.22.0 which is incompatible.
> ERROR: tensorflow 2.1.0 has requirement scipy==1.4.1; python_version >= "3", 
> but you'll have scipy 1.2.2 which is incompatible.
>  
>  
>  
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (BEAM-9510) Dependencies in base_image_requirements.txt are not compatible with each other

2020-03-16 Thread Ahmet Altay (Jira)


[ 
https://issues.apache.org/jira/browse/BEAM-9510?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17060457#comment-17060457
 ] 

Ahmet Altay commented on BEAM-9510:
---

This file looks out of date. We can update it to match 
(https://cloud.google.com/dataflow/docs/concepts/sdk-worker-dependencies#sdk-for-python)

> Dependencies in base_image_requirements.txt are not compatible with each other
> --
>
> Key: BEAM-9510
> URL: https://issues.apache.org/jira/browse/BEAM-9510
> Project: Beam
>  Issue Type: Bug
>  Components: sdk-py-harness
>Reporter: David Yan
>Priority: Major
>
> [https://github.com/apache/beam/blob/master/sdks/python/container/base_image_requirements.txt#L56]
> says it requires google-cloud-bigquery==1.24.0, google-cloud-core==1.0.2, 
> google-cloud-bigtable==0.32.1, grpc-1.22.0 and tensorflow-2.1.0
> But they are incompatible with each other:
> ERROR: google-cloud-bigquery 1.24.0 has requirement 
> google-cloud-core<2.0dev,>=1.1.0, but you'll have google-cloud-core 1.0.2 
> which is incompatible.
> ERROR: google-cloud-bigtable 0.32.1 has requirement 
> google-cloud-core<0.30dev,>=0.29.0, but you'll have google-cloud-core 1.0.2 
> which is incompatible.
> ERROR: tensorboard 2.1.1 has requirement grpcio>=1.24.3, but you'll have 
> grpcio 1.22.0 which is incompatible.
> ERROR: tensorflow 2.1.0 has requirement scipy==1.4.1; python_version >= "3", 
> but you'll have scipy 1.2.2 which is incompatible.
>  
>  
>  
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


  1   2   3   >