[ 
https://issues.apache.org/jira/browse/BEAM-6102?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16693892#comment-16693892
 ] 

Chamikara Jayalath commented on BEAM-6102:
------------------------------------------

Is this a release blocker ?

> Wordcount broken on master
> --------------------------
>
>                 Key: BEAM-6102
>                 URL: https://issues.apache.org/jira/browse/BEAM-6102
>             Project: Beam
>          Issue Type: Bug
>          Components: sdk-java-core
>            Reporter: Ankur Goenka
>            Assignee: Kenneth Knowles
>            Priority: Major
>
> The wordcount is broken on the master.
> Its failing with serialization error mentioned below
>  
> {code:java}
> 11:08:59 AM: Executing task 'WordCount.main()'... Parallel execution is an 
> incubating feature. > Task :buildSrc:compileJava NO-SOURCE > Task 
> :buildSrc:compileGroovy UP-TO-DATE > Task :buildSrc:processResources 
> NO-SOURCE > Task :buildSrc:classes UP-TO-DATE > Task :buildSrc:jar UP-TO-DATE 
> > Task :buildSrc:assemble UP-TO-DATE > Task :buildSrc:spotlessGroovy 
> UP-TO-DATE > Task :buildSrc:spotlessGroovyCheck UP-TO-DATE > Task 
> :buildSrc:spotlessGroovyGradle UP-TO-DATE > Task 
> :buildSrc:spotlessGroovyGradleCheck UP-TO-DATE > Task :buildSrc:spotlessCheck 
> UP-TO-DATE > Task :buildSrc:compileTestJava NO-SOURCE > Task 
> :buildSrc:compileTestGroovy NO-SOURCE > Task :buildSrc:processTestResources 
> NO-SOURCE > Task :buildSrc:testClasses UP-TO-DATE > Task :buildSrc:test 
> NO-SOURCE > Task :buildSrc:check UP-TO-DATE > Task :buildSrc:build UP-TO-DATE 
> Parallel execution with configuration on demand is an incubating feature. > 
> Configure project :beam-model-pipeline applyPortabilityNature with default 
> configuration for project beam-model-pipeline > Configure project 
> :beam-model-job-management applyPortabilityNature with default configuration 
> for project beam-model-job-management > Configure project 
> :beam-model-fn-execution applyPortabilityNature with default configuration 
> for project beam-model-fn-execution > Task 
> :beam-examples-java:processResources NO-SOURCE > Task 
> :beam-sdks-java-core:generateAvroProtocol NO-SOURCE > Task 
> :beam-sdks-java-extensions-google-cloud-platform-core:processResources 
> NO-SOURCE > Task :beam-sdks-java-io-google-cloud-platform:processResources 
> NO-SOURCE > Task :beam-vendor-grpc-v1_13_1:compileJava NO-SOURCE > Task 
> :beam-runners-core-construction-java:processResources NO-SOURCE > Task 
> :beam-sdks-java-extensions-protobuf:extractProto UP-TO-DATE > Task 
> :beam-model-pipeline:extractProto UP-TO-DATE > Task 
> :beam-model-job-management:processResources UP-TO-DATE > Task 
> :beam-runners-core-java:processResources NO-SOURCE > Task 
> :beam-sdks-java-fn-execution:processResources NO-SOURCE > Task 
> :beam-sdks-java-harness:processResources NO-SOURCE > Task 
> :beam-vendor-sdks-java-extensions-protobuf:processResources NO-SOURCE > Task 
> :beam-sdks-java-extensions-protobuf:processResources NO-SOURCE > Task 
> :beam-sdks-java-core:generateAvroJava NO-SOURCE > Task 
> :beam-vendor-grpc-v1_13_1:processResources NO-SOURCE > Task 
> :beam-model-fn-execution:processResources UP-TO-DATE > Task 
> :beam-model-pipeline:processResources UP-TO-DATE > Task 
> :beam-runners-local-java-core:processResources NO-SOURCE > Task 
> :beam-runners-java-fn-execution:processResources NO-SOURCE > Task 
> :beam-runners-direct-java:processResources NO-SOURCE > Task 
> :beam-vendor-grpc-v1_13_1:classes UP-TO-DATE > Task 
> :beam-sdks-java-core:processResources UP-TO-DATE > Task 
> :beam-runners-google-cloud-dataflow-java:processResources UP-TO-DATE > Task 
> :beam-vendor-grpc-v1_13_1:shadowJar UP-TO-DATE > Task 
> :beam-model-fn-execution:extractIncludeProto UP-TO-DATE > Task 
> :beam-model-pipeline:extractIncludeProto UP-TO-DATE > Task 
> :beam-model-job-management:extractIncludeProto UP-TO-DATE > Task 
> :beam-model-pipeline:generateProto UP-TO-DATE > Task 
> :beam-model-pipeline:compileJava UP-TO-DATE > Task 
> :beam-model-pipeline:classes UP-TO-DATE > Task :beam-model-pipeline:shadowJar 
> UP-TO-DATE > Task :beam-model-pipeline:jar UP-TO-DATE > Task 
> :beam-model-job-management:extractProto UP-TO-DATE > Task 
> :beam-model-fn-execution:extractProto UP-TO-DATE > Task 
> :beam-model-fn-execution:generateProto UP-TO-DATE > Task 
> :beam-model-job-management:generateProto UP-TO-DATE > Task 
> :beam-model-job-management:compileJava UP-TO-DATE > Task 
> :beam-model-job-management:classes UP-TO-DATE > Task 
> :beam-model-fn-execution:compileJava UP-TO-DATE > Task 
> :beam-model-fn-execution:classes UP-TO-DATE > Task 
> :beam-model-job-management:shadowJar UP-TO-DATE > Task 
> :beam-sdks-java-core:compileJava UP-TO-DATE > Task 
> :beam-sdks-java-core:classes UP-TO-DATE > Task 
> :beam-model-fn-execution:shadowJar UP-TO-DATE > Task 
> :beam-sdks-java-core:shadowJar UP-TO-DATE > Task 
> :beam-sdks-java-extensions-protobuf:extractIncludeProto UP-TO-DATE > Task 
> :beam-sdks-java-extensions-protobuf:generateProto NO-SOURCE > Task 
> :beam-sdks-java-fn-execution:compileJava UP-TO-DATE > Task 
> :beam-sdks-java-fn-execution:classes UP-TO-DATE > Task 
> :beam-runners-local-java-core:compileJava UP-TO-DATE > Task 
> :beam-runners-local-java-core:classes UP-TO-DATE > Task 
> :beam-runners-core-construction-java:compileJava UP-TO-DATE > Task 
> :beam-sdks-java-extensions-google-cloud-platform-core:compileJava UP-TO-DATE 
> > Task :beam-sdks-java-extensions-google-cloud-platform-core:classes 
> UP-TO-DATE > Task :beam-sdks-java-extensions-protobuf:compileJava UP-TO-DATE 
> > Task :beam-sdks-java-extensions-protobuf:classes UP-TO-DATE > Task 
> :beam-vendor-sdks-java-extensions-protobuf:compileJava UP-TO-DATE > Task 
> :beam-vendor-sdks-java-extensions-protobuf:classes UP-TO-DATE > Task 
> :beam-runners-core-construction-java:classes UP-TO-DATE > Task 
> :beam-runners-local-java-core:shadowJar UP-TO-DATE > Task 
> :beam-sdks-java-fn-execution:shadowJar UP-TO-DATE > Task 
> :beam-sdks-java-extensions-protobuf:shadowJar UP-TO-DATE > Task 
> :beam-sdks-java-extensions-google-cloud-platform-core:shadowJar UP-TO-DATE > 
> Task :beam-vendor-sdks-java-extensions-protobuf:shadowJar UP-TO-DATE > Task 
> :beam-runners-core-construction-java:shadowJar UP-TO-DATE > Task 
> :beam-runners-core-java:compileJava UP-TO-DATE > Task 
> :beam-runners-core-java:classes UP-TO-DATE > Task 
> :beam-runners-core-java:shadowJar UP-TO-DATE > Task 
> :beam-sdks-java-io-google-cloud-platform:compileJava UP-TO-DATE > Task 
> :beam-sdks-java-io-google-cloud-platform:classes UP-TO-DATE > Task 
> :beam-sdks-java-harness:compileJava UP-TO-DATE > Task 
> :beam-sdks-java-harness:classes UP-TO-DATE > Task :beam-sdks-java-harness:jar 
> UP-TO-DATE > Task :beam-sdks-java-io-google-cloud-platform:shadowJar 
> UP-TO-DATE > Task :beam-sdks-java-harness:shadowJar UP-TO-DATE > Task 
> :beam-examples-java:compileJava UP-TO-DATE > Task :beam-examples-java:classes 
> UP-TO-DATE > Task :beam-runners-java-fn-execution:compileJava UP-TO-DATE > 
> Task :beam-runners-java-fn-execution:classes UP-TO-DATE > Task 
> :beam-runners-java-fn-execution:shadowJar UP-TO-DATE > Task 
> :beam-runners-google-cloud-dataflow-java:compileJava UP-TO-DATE > Task 
> :beam-runners-google-cloud-dataflow-java:classes UP-TO-DATE > Task 
> :beam-runners-direct-java:compileJava UP-TO-DATE > Task 
> :beam-runners-direct-java:classes UP-TO-DATE > Task 
> :beam-runners-google-cloud-dataflow-java:shadowJar UP-TO-DATE > Task 
> :beam-runners-direct-java:shadowJar UP-TO-DATE > Task 
> :beam-examples-java:WordCount.main() Nov 20, 2018 11:09:02 AM 
> org.apache.beam.sdk.extensions.gcp.options.GcpOptions$DefaultProjectFactory 
> create INFO: Inferred default GCP project 'google.com:clouddfe' from gcloud. 
> If this is the incorrect project, please cancel this Pipeline and specify the 
> command-line argument --project. Nov 20, 2018 11:09:02 AM 
> com.google.auth.oauth2.DefaultCredentialsProvider 
> warnAboutProblematicCredentials WARNING: Your application has authenticated 
> using end user credentials from Google Cloud SDK. We recommend that most 
> server applications use service accounts instead. If your application 
> continues to use end user credentials from Cloud SDK, you might receive a 
> "quota exceeded" or "API not enabled" error. For more information about 
> service accounts, see https://cloud.google.com/docs/authentication/. Nov 20, 
> 2018 11:09:03 AM 
> org.apache.beam.sdk.extensions.gcp.options.GcpOptions$GcpTempLocationFactory 
> tryCreateDefaultBucket INFO: No tempLocation specified, attempting to use 
> default bucket: dataflow-staging-us-central1-927334603519 Nov 20, 2018 
> 11:09:03 AM 
> org.apache.beam.sdk.util.RetryHttpRequestInitializer$LoggingHttpBackOffHandler
>  handleResponse WARNING: Request failed with code 409, performed 0 retries 
> due to IOExceptions, performed 0 retries due to unsuccessful status codes, 
> HTTP framework says request can be retried, (caller responsible for 
> retrying): 
> https://www.googleapis.com/storage/v1/b?predefinedAcl=projectPrivate&predefinedDefaultObjectAcl=projectPrivate&project=google.com:clouddfe
>  Nov 20, 2018 11:09:03 AM 
> org.apache.beam.runners.dataflow.options.DataflowPipelineOptions$StagingLocationFactory
>  create INFO: No stagingLocation provided, falling back to gcpTempLocation 
> Nov 20, 2018 11:09:04 AM org.apache.beam.runners.dataflow.DataflowRunner 
> fromOptions INFO: PipelineOptions.filesToStage was not specified. Defaulting 
> to files from the classpath: will stage 120 files. Enable logging at DEBUG 
> level to see which files will be staged. Nov 20, 2018 11:09:04 AM 
> org.apache.beam.runners.dataflow.DataflowRunner run INFO: Executing pipeline 
> on the Dataflow Service, which will have billing implications related to 
> Google Compute Engine usage and other Google Cloud Services. Nov 20, 2018 
> 11:09:04 AM org.apache.beam.runners.dataflow.util.PackageUtil 
> stageClasspathElements INFO: Uploading 120 files from 
> PipelineOptions.filesToStage to staging location to prepare for execution. 
> Nov 20, 2018 11:09:04 AM org.apache.beam.runners.dataflow.util.PackageUtil 
> stageClasspathElements WARNING: Skipping non-existent file to stage 
> /usr/local/google/home/goenka/d/work/tmp/tmp_beam/beam/examples/java/build/resources/main.
>  Nov 20, 2018 11:09:05 AM org.apache.beam.runners.dataflow.util.PackageUtil 
> tryStagePackage INFO: Uploading 
> /usr/local/google/home/goenka/.gradle/caches/modules-2/files-2.1/com.google.http-client/google-http-client/1.27.0/52d6076f58edfedb55526c0cac4ba155c6dc645b/google-http-client-1.27.0.jar
>  to 
> gs://dataflow-staging-us-central1-927334603519/temp/staging/google-http-client-1.27.0-XiDeJfgpNiiCZxI1a4Q27Q.jar
>  Nov 20, 2018 11:09:05 AM org.apache.beam.runners.dataflow.util.PackageUtil 
> tryStagePackage INFO: Uploading 
> /usr/local/google/home/goenka/.gradle/caches/modules-2/files-2.1/com.google.oauth-client/google-oauth-client-java6/1.27.0/90570eedf1ae6188ee5028c11ec423fe52336373/google-oauth-client-java6-1.27.0.jar
>  to 
> gs://dataflow-staging-us-central1-927334603519/temp/staging/google-oauth-client-java6-1.27.0-VWEWXu2J1auJGvb5eLfbhg.jar
>  Nov 20, 2018 11:09:05 AM org.apache.beam.runners.dataflow.util.PackageUtil 
> tryStagePackage INFO: Uploading 
> /usr/local/google/home/goenka/.gradle/caches/modules-2/files-2.1/com.google.apis/google-api-services-pubsub/v1-rev20181105-1.27.0/31f49e03ec44c72845a6a0572ccba74d7aee1f57/google-api-services-pubsub-v1-rev20181105-1.27.0.jar
>  to 
> gs://dataflow-staging-us-central1-927334603519/temp/staging/google-api-services-pubsub-v1-rev20181105-1.27.0-_3JmrTY281LX4AcOiP7LkQ.jar
>  Nov 20, 2018 11:09:05 AM org.apache.beam.runners.dataflow.util.PackageUtil 
> tryStagePackage INFO: Uploading 
> /usr/local/google/home/goenka/.gradle/caches/modules-2/files-2.1/com.google.apis/google-api-services-dataflow/v1b3-rev20181107-1.27.0/d0b3764765916d65fd75a52ae1a8cf13a150450a/google-api-services-dataflow-v1b3-rev20181107-1.27.0.jar
>  to 
> gs://dataflow-staging-us-central1-927334603519/temp/staging/google-api-services-dataflow-v1b3-rev20181107-1.27.0-ZCpCwXOULzPs_S2gBBH7TA.jar
>  Nov 20, 2018 11:09:05 AM org.apache.beam.runners.dataflow.util.PackageUtil 
> tryStagePackage INFO: Uploading 
> /usr/local/google/home/goenka/.gradle/caches/modules-2/files-2.1/com.google.oauth-client/google-oauth-client/1.27.0/1979e3b499f1ef49959005ca59101791361abca9/google-oauth-client-1.27.0.jar
>  to 
> gs://dataflow-staging-us-central1-927334603519/temp/staging/google-oauth-client-1.27.0-waNrxu589XHIvY73DhrxNA.jar
>  Nov 20, 2018 11:09:05 AM org.apache.beam.runners.dataflow.util.PackageUtil 
> tryStagePackage INFO: Uploading 
> /usr/local/google/home/goenka/.gradle/caches/modules-2/files-2.1/com.google.apis/google-api-services-clouddebugger/v2-rev20180801-1.27.0/2866c6cfca03ef7bbf687ba3b62fa843377e664b/google-api-services-clouddebugger-v2-rev20180801-1.27.0.jar
>  to 
> gs://dataflow-staging-us-central1-927334603519/temp/staging/google-api-services-clouddebugger-v2-rev20180801-1.27.0-xrIcl7yUGCktYEb9_xtvjQ.jar
>  Nov 20, 2018 11:09:05 AM org.apache.beam.runners.dataflow.util.PackageUtil 
> tryStagePackage INFO: Uploading 
> /usr/local/google/home/goenka/.gradle/caches/modules-2/files-2.1/com.google.api-client/google-api-client-jackson2/1.27.0/1e1a91b684fc2e6e97d2369c7d8d1d9efc115543/google-api-client-jackson2-1.27.0.jar
>  to 
> gs://dataflow-staging-us-central1-927334603519/temp/staging/google-api-client-jackson2-1.27.0-M222JGX1Wh3HoNtcvwCaEw.jar
>  Nov 20, 2018 11:09:05 AM org.apache.beam.runners.dataflow.util.PackageUtil 
> tryStagePackage INFO: Uploading 
> /usr/local/google/home/goenka/.gradle/caches/modules-2/files-2.1/com.google.api-client/google-api-client-java6/1.27.0/2c06247935819b429424797d9844aa33955f4fb0/google-api-client-java6-1.27.0.jar
>  to 
> gs://dataflow-staging-us-central1-927334603519/temp/staging/google-api-client-java6-1.27.0-zPlGOao_rbTkn6baqjfy2w.jar
>  Nov 20, 2018 11:09:05 AM org.apache.beam.runners.dataflow.util.PackageUtil 
> tryStagePackage INFO: Uploading 
> /usr/local/google/home/goenka/d/work/tmp/tmp_beam/beam/sdks/java/io/google-cloud-platform/build/libs/beam-sdks-java-io-google-cloud-platform-2.9.0-SNAPSHOT.jar
>  to 
> gs://dataflow-staging-us-central1-927334603519/temp/staging/beam-sdks-java-io-google-cloud-platform-2.9.0-SNAPSHOT-xFFgKvdT2GS2td3iSJaJqQ.jar
>  Nov 20, 2018 11:09:05 AM org.apache.beam.runners.dataflow.util.PackageUtil 
> tryStagePackage INFO: Uploading 
> /usr/local/google/home/goenka/.gradle/caches/modules-2/files-2.1/com.google.apis/google-api-services-bigquery/v2-rev20181104-1.27.0/d539fb4b7ac318cadc344579fa6f80ae6cb8070e/google-api-services-bigquery-v2-rev20181104-1.27.0.jar
>  to 
> gs://dataflow-staging-us-central1-927334603519/temp/staging/google-api-services-bigquery-v2-rev20181104-1.27.0-oUKA0dpY-zmhgM8DIAq8Zg.jar
>  Nov 20, 2018 11:09:05 AM org.apache.beam.runners.dataflow.util.PackageUtil 
> tryStagePackage INFO: Uploading 
> /usr/local/google/home/goenka/.gradle/caches/modules-2/files-2.1/com.google.http-client/google-http-client-protobuf/1.27.0/b81859b8a284ea416d74644a70524c4d6be3e9a/google-http-client-protobuf-1.27.0.jar
>  to 
> gs://dataflow-staging-us-central1-927334603519/temp/staging/google-http-client-protobuf-1.27.0-1E0HsNaSwDtqOduUs-NssA.jar
>  Nov 20, 2018 11:09:05 AM org.apache.beam.runners.dataflow.util.PackageUtil 
> tryStagePackage INFO: Uploading 
> /usr/local/google/home/goenka/.gradle/caches/modules-2/files-2.1/com.google.http-client/google-http-client-jackson/1.27.0/2b5eff5b6a343e716db21673743576db9a53e5ae/google-http-client-jackson-1.27.0.jar
>  to 
> gs://dataflow-staging-us-central1-927334603519/temp/staging/google-http-client-jackson-1.27.0-XQrBFTlCGWMqVu9vUPkMcw.jar
>  Nov 20, 2018 11:09:05 AM org.apache.beam.runners.dataflow.util.PackageUtil 
> tryStagePackage INFO: Uploading 
> /usr/local/google/home/goenka/.gradle/caches/modules-2/files-2.1/com.google.api-client/google-api-client/1.27.0/7d498c11db92fb72521784212c953612a42e50db/google-api-client-1.27.0.jar
>  to 
> gs://dataflow-staging-us-central1-927334603519/temp/staging/google-api-client-1.27.0-UjH-dlvVLqF5D_4Trh8foQ.jar
>  Nov 20, 2018 11:09:05 AM org.apache.beam.runners.dataflow.util.PackageUtil 
> tryStagePackage INFO: Uploading 
> /usr/local/google/home/goenka/.gradle/caches/modules-2/files-2.1/com.google.apis/google-api-services-storage/v1-rev20181013-1.27.0/10593f99e06d962017b0663348ee1fed0f8528/google-api-services-storage-v1-rev20181013-1.27.0.jar
>  to 
> gs://dataflow-staging-us-central1-927334603519/temp/staging/google-api-services-storage-v1-rev20181013-1.27.0-Y65RPabM4FA03rQX5lUvKg.jar
>  Nov 20, 2018 11:09:05 AM org.apache.beam.runners.dataflow.util.PackageUtil 
> tryStagePackage INFO: Uploading 
> /usr/local/google/home/goenka/d/work/tmp/tmp_beam/beam/model/job-management/build/libs/beam-model-job-management-2.9.0-SNAPSHOT.jar
>  to 
> gs://dataflow-staging-us-central1-927334603519/temp/staging/beam-model-job-management-2.9.0-SNAPSHOT-PRRkG0x941fKfrBa33f-Qg.jar
>  Nov 20, 2018 11:09:05 AM org.apache.beam.runners.dataflow.util.PackageUtil 
> tryStagePackage INFO: Uploading 
> /usr/local/google/home/goenka/.gradle/caches/modules-2/files-2.1/com.google.apis/google-api-services-cloudresourcemanager/v1-rev20181015-1.27.0/354796285db3b2fce9089147db3fe1f2a8cca444/google-api-services-cloudresourcemanager-v1-rev20181015-1.27.0.jar
>  to 
> gs://dataflow-staging-us-central1-927334603519/temp/staging/google-api-services-cloudresourcemanager-v1-rev20181015-1.27.0-SQfpsiUEwOlTbzfnWC4EzA.jar
>  Nov 20, 2018 11:09:05 AM org.apache.beam.runners.dataflow.util.PackageUtil 
> tryStagePackage INFO: Uploading 
> /usr/local/google/home/goenka/d/work/tmp/tmp_beam/beam/model/pipeline/build/libs/beam-model-pipeline-2.9.0-SNAPSHOT.jar
>  to 
> gs://dataflow-staging-us-central1-927334603519/temp/staging/beam-model-pipeline-2.9.0-SNAPSHOT-XD9dbaFWaiUJnjDcCAkrOg.jar
>  Nov 20, 2018 11:09:05 AM org.apache.beam.runners.dataflow.util.PackageUtil 
> tryStagePackage INFO: Uploading 
> /usr/local/google/home/goenka/.gradle/caches/modules-2/files-2.1/org.apache.httpcomponents/httpcore/4.4.9/a86ce739e5a7175b4b234c290a00a5fdb80957a0/httpcore-4.4.9.jar
>  to 
> gs://dataflow-staging-us-central1-927334603519/temp/staging/httpcore-4.4.9-uJRVUHg5wJ1hGWYd79IWag.jar
>  Nov 20, 2018 11:09:05 AM org.apache.beam.runners.dataflow.util.PackageUtil 
> tryStagePackage INFO: Uploading 
> /usr/local/google/home/goenka/d/work/tmp/tmp_beam/beam/sdks/java/extensions/protobuf/build/libs/beam-sdks-java-extensions-protobuf-2.9.0-SNAPSHOT.jar
>  to 
> gs://dataflow-staging-us-central1-927334603519/temp/staging/beam-sdks-java-extensions-protobuf-2.9.0-SNAPSHOT-4q_iSCAMQwCGDowjT6QSKw.jar
>  Nov 20, 2018 11:09:05 AM org.apache.beam.runners.dataflow.util.PackageUtil 
> tryStagePackage INFO: Uploading 
> /usr/local/google/home/goenka/d/work/tmp/tmp_beam/beam/sdks/java/extensions/google-cloud-platform-core/build/libs/beam-sdks-java-extensions-google-cloud-platform-core-2.9.0-SNAPSHOT.jar
>  to 
> gs://dataflow-staging-us-central1-927334603519/temp/staging/beam-sdks-java-extensions-google-cloud-platform-core-2.9.0-SNAPSHOT-qVkigYLlAbU1ujEdaTQZ0Q.jar
>  Nov 20, 2018 11:09:05 AM org.apache.beam.runners.dataflow.util.PackageUtil 
> tryStagePackage INFO: Uploading 
> /usr/local/google/home/goenka/.gradle/caches/modules-2/files-2.1/org.apache.httpcomponents/httpclient/4.5.5/1603dfd56ebcd583ccdf337b6c3984ac55d89e58/httpclient-4.5.5.jar
>  to 
> gs://dataflow-staging-us-central1-927334603519/temp/staging/httpclient-4.5.5-l-flsTVHa30lpasx4epJIg.jar
>  Nov 20, 2018 11:09:05 AM org.apache.beam.runners.dataflow.util.PackageUtil 
> tryStagePackage INFO: Uploading 
> /usr/local/google/home/goenka/d/work/tmp/tmp_beam/beam/runners/core-construction-java/build/libs/beam-runners-core-construction-java-2.9.0-SNAPSHOT.jar
>  to 
> gs://dataflow-staging-us-central1-927334603519/temp/staging/beam-runners-core-construction-java-2.9.0-SNAPSHOT-XbHTR4DoxxMKF5aSfjqWYQ.jar
>  Nov 20, 2018 11:09:05 AM org.apache.beam.runners.dataflow.util.PackageUtil 
> tryStagePackage INFO: Uploading 
> /usr/local/google/home/goenka/.gradle/caches/modules-2/files-2.1/com.google.http-client/google-http-client-jackson2/1.27.0/25ee9240648b5dfd01a3385937bcf332b6f222a7/google-http-client-jackson2-1.27.0.jar
>  to 
> gs://dataflow-staging-us-central1-927334603519/temp/staging/google-http-client-jackson2-1.27.0-s26dP1SAvqbYDq_HiJ9Bbw.jar
>  Nov 20, 2018 11:09:05 AM org.apache.beam.runners.dataflow.util.PackageUtil 
> tryStagePackage INFO: Uploading 
> /usr/local/google/home/goenka/d/work/tmp/tmp_beam/beam/runners/direct-java/build/libs/beam-runners-direct-java-2.9.0-SNAPSHOT.jar
>  to 
> gs://dataflow-staging-us-central1-927334603519/temp/staging/beam-runners-direct-java-2.9.0-SNAPSHOT-jn9fS-cJcB_-dTRLAU-meQ.jar
>  Nov 20, 2018 11:09:05 AM org.apache.beam.runners.dataflow.util.PackageUtil 
> tryStagePackage INFO: Uploading 
> /usr/local/google/home/goenka/d/work/tmp/tmp_beam/beam/runners/google-cloud-dataflow-java/build/libs/beam-runners-google-cloud-dataflow-java-2.9.0-SNAPSHOT.jar
>  to 
> gs://dataflow-staging-us-central1-927334603519/temp/staging/beam-runners-google-cloud-dataflow-java-2.9.0-SNAPSHOT-oQdbiEtnM6C2iUC5axtRjA.jar
>  Nov 20, 2018 11:09:05 AM org.apache.beam.runners.dataflow.util.PackageUtil 
> tryStagePackage INFO: Uploading 
> /usr/local/google/home/goenka/d/work/tmp/tmp_beam/beam/examples/java/build/classes/java/main
>  to 
> gs://dataflow-staging-us-central1-927334603519/temp/staging/main-cxzukVaCwjdrlMI-NQ8Jew.jar
>  Nov 20, 2018 11:09:05 AM org.apache.beam.runners.dataflow.util.PackageUtil 
> tryStagePackage INFO: Uploading 
> /usr/local/google/home/goenka/d/work/tmp/tmp_beam/beam/vendor/grpc-v1_13_1/build/libs/beam-vendor-grpc-v1_13_1-2.9.0-SNAPSHOT.jar
>  to 
> gs://dataflow-staging-us-central1-927334603519/temp/staging/beam-vendor-grpc-v1_13_1-2.9.0-SNAPSHOT-E8CXf40HkGDxZZU040A-bA.jar
>  Nov 20, 2018 11:09:06 AM org.apache.beam.runners.dataflow.util.PackageUtil 
> tryStagePackage INFO: Uploading 
> /usr/local/google/home/goenka/d/work/tmp/tmp_beam/beam/sdks/java/core/build/libs/beam-sdks-java-core-2.9.0-SNAPSHOT.jar
>  to 
> gs://dataflow-staging-us-central1-927334603519/temp/staging/beam-sdks-java-core-2.9.0-SNAPSHOT-Aoyvg0doZsYvr7jINSJAKA.jar
>  Nov 20, 2018 11:09:21 AM org.apache.beam.runners.dataflow.util.PackageUtil 
> stageClasspathElements INFO: Staging files complete: 91 files cached, 28 
> files newly uploaded Nov 20, 2018 11:09:21 AM 
> org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator 
> addStep INFO: Adding ReadLines/Read as step s1 Nov 20, 2018 11:09:21 AM 
> org.apache.beam.sdk.io.FileBasedSource getEstimatedSizeBytes INFO: 
> Filepattern gs://apache-beam-samples/shakespeare/kinglear.txt matched 1 files 
> with total size 157283 Nov 20, 2018 11:09:21 AM 
> org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator 
> addStep INFO: Adding WordCount.CountWords/ParDo(ExtractWords) as step s2 Nov 
> 20, 2018 11:09:21 AM 
> org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator 
> addStep INFO: Adding WordCount.CountWords/Count.PerElement/Init/Map as step 
> s3 Nov 20, 2018 11:09:21 AM 
> org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator 
> addStep INFO: Adding 
> WordCount.CountWords/Count.PerElement/Combine.perKey(Count)/GroupByKey as 
> step s4 Nov 20, 2018 11:09:21 AM 
> org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator 
> addStep INFO: Adding 
> WordCount.CountWords/Count.PerElement/Combine.perKey(Count)/Combine.GroupedValues
>  as step s5 Nov 20, 2018 11:09:21 AM 
> org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator 
> addStep INFO: Adding MapElements/Map as step s6 Nov 20, 2018 11:09:21 AM 
> org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator 
> addStep INFO: Adding WriteCounts/WriteFiles/RewindowIntoGlobal/Window.Assign 
> as step s7 Nov 20, 2018 11:09:21 AM 
> org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator 
> addStep INFO: Adding 
> WriteCounts/WriteFiles/WriteUnshardedBundlesToTempFiles/WriteUnshardedBundles 
> as step s8 Nov 20, 2018 11:09:21 AM 
> org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator 
> addStep INFO: Adding 
> WriteCounts/WriteFiles/WriteUnshardedBundlesToTempFiles/GroupUnwritten as 
> step s9 Nov 20, 2018 11:09:21 AM 
> org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator 
> addStep INFO: Adding 
> WriteCounts/WriteFiles/WriteUnshardedBundlesToTempFiles/WriteUnwritten as 
> step s10 Nov 20, 2018 11:09:21 AM 
> org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator 
> addStep INFO: Adding 
> WriteCounts/WriteFiles/WriteUnshardedBundlesToTempFiles/DropShardNum as step 
> s11 Nov 20, 2018 11:09:21 AM 
> org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator 
> addStep INFO: Adding 
> WriteCounts/WriteFiles/WriteUnshardedBundlesToTempFiles/Flatten.PCollections 
> as step s12 Nov 20, 2018 11:09:21 AM 
> org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator 
> addStep INFO: Adding 
> WriteCounts/WriteFiles/GatherTempFileResults/View.AsList/ParDo(ToIsmRecordForGlobalWindow)
>  as step s13 Nov 20, 2018 11:09:21 AM 
> org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator 
> addStep INFO: Adding 
> WriteCounts/WriteFiles/GatherTempFileResults/View.AsList/CreateDataflowView 
> as step s14 Nov 20, 2018 11:09:21 AM 
> org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator 
> addStep INFO: Adding 
> WriteCounts/WriteFiles/GatherTempFileResults/Reify.ReifyViewInGlobalWindow/Create.Values/Read(CreateSource)
>  as step s15 Nov 20, 2018 11:09:21 AM 
> org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator 
> addStep INFO: Adding 
> WriteCounts/WriteFiles/GatherTempFileResults/Reify.ReifyViewInGlobalWindow/Reify.ReifyView/ParDo(Anonymous)
>  as step s16 Nov 20, 2018 11:09:21 AM 
> org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator 
> addStep INFO: Adding 
> WriteCounts/WriteFiles/GatherTempFileResults/Reify.ReifyViewInGlobalWindow/Values/Values/Map
>  as step s17 Nov 20, 2018 11:09:21 AM 
> org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator 
> addStep INFO: Adding WriteCounts/WriteFiles/FinalizeTempFileBundles/Finalize 
> as step s18 Nov 20, 2018 11:09:21 AM 
> org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator 
> addStep INFO: Adding 
> WriteCounts/WriteFiles/FinalizeTempFileBundles/Reshuffle.ViaRandomKey/Pair 
> with random key as step s19 Nov 20, 2018 11:09:21 AM 
> org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator 
> addStep INFO: Adding 
> WriteCounts/WriteFiles/FinalizeTempFileBundles/Reshuffle.ViaRandomKey/Reshuffle/Window.Into()/Window.Assign
>  as step s20 Nov 20, 2018 11:09:21 AM 
> org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator 
> addStep INFO: Adding 
> WriteCounts/WriteFiles/FinalizeTempFileBundles/Reshuffle.ViaRandomKey/Reshuffle/GroupByKey
>  as step s21 Nov 20, 2018 11:09:21 AM 
> org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator 
> addStep INFO: Adding 
> WriteCounts/WriteFiles/FinalizeTempFileBundles/Reshuffle.ViaRandomKey/Reshuffle/ExpandIterable
>  as step s22 Nov 20, 2018 11:09:21 AM 
> org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator 
> addStep INFO: Adding 
> WriteCounts/WriteFiles/FinalizeTempFileBundles/Reshuffle.ViaRandomKey/Values/Values/Map
>  as step s23 Nov 20, 2018 11:09:21 AM 
> org.apache.beam.runners.dataflow.DataflowRunner run INFO: Staging pipeline 
> description to gs://dataflow-staging-us-central1-927334603519/temp/staging/ 
> Nov 20, 2018 11:09:22 AM org.apache.beam.runners.dataflow.util.PackageUtil 
> tryStagePackage INFO: Uploading <63064 bytes, hash 1eMyLHEGQDSnbL_rtt2mOQ> to 
> gs://dataflow-staging-us-central1-927334603519/temp/staging/pipeline-1eMyLHEGQDSnbL_rtt2mOQ.pb
>  Dataflow SDK version: 2.9.0-SNAPSHOT Nov 20, 2018 11:09:24 AM 
> org.apache.beam.runners.dataflow.DataflowRunner run Submitted job: 
> 2018-11-20_11_09_23-3437480453113408186 INFO: To access the Dataflow 
> monitoring console, please navigate to 
> https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-11-20_11_09_23-3437480453113408186?project=google.com%3Aclouddfe
>  Nov 20, 2018 11:09:24 AM org.apache.beam.runners.dataflow.DataflowRunner run 
> INFO: To cancel the job using the 'gcloud' tool, run: > gcloud dataflow jobs 
> --project=google.com:clouddfe cancel --region=us-central1 
> 2018-11-20_11_09_23-3437480453113408186 Nov 20, 2018 11:09:34 AM 
> org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process 
> INFO: 2018-11-20T19:09:23.285Z: Autoscaling is enabled for job 
> 2018-11-20_11_09_23-3437480453113408186. The number of workers will be 
> between 1 and 1000. Nov 20, 2018 11:09:34 AM 
> org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process 
> INFO: 2018-11-20T19:09:23.346Z: Autoscaling was automatically enabled for job 
> 2018-11-20_11_09_23-3437480453113408186. Nov 20, 2018 11:09:34 AM 
> org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process 
> INFO: 2018-11-20T19:09:26.020Z: Checking permissions granted to controller 
> Service Account. Nov 20, 2018 11:09:34 AM 
> org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process 
> INFO: 2018-11-20T19:09:30.430Z: Worker configuration: n1-standard-1 in 
> us-central1-a. Nov 20, 2018 11:09:34 AM 
> org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process 
> INFO: 2018-11-20T19:09:31.335Z: Expanding CoGroupByKey operations into 
> optimizable parts. Nov 20, 2018 11:09:34 AM 
> org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process 
> INFO: 2018-11-20T19:09:31.555Z: Expanding GroupByKey operations into 
> optimizable parts. Nov 20, 2018 11:09:34 AM 
> org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process 
> INFO: 2018-11-20T19:09:31.614Z: Lifting ValueCombiningMappingFns into 
> MergeBucketsMappingFns Nov 20, 2018 11:09:34 AM 
> org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process 
> INFO: 2018-11-20T19:09:31.886Z: Fusing adjacent ParDo, Read, Write, and 
> Flatten operations Nov 20, 2018 11:09:34 AM 
> org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process 
> INFO: 2018-11-20T19:09:31.941Z: Fusing consumer 
> WriteCounts/WriteFiles/GatherTempFileResults/Reify.ReifyViewInGlobalWindow/Reify.ReifyView/ParDo(Anonymous)
>  into 
> WriteCounts/WriteFiles/GatherTempFileResults/Reify.ReifyViewInGlobalWindow/Create.Values/Read(CreateSource)
>  Nov 20, 2018 11:09:34 AM 
> org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process 
> INFO: 2018-11-20T19:09:31.994Z: Fusing consumer 
> WriteCounts/WriteFiles/FinalizeTempFileBundles/Reshuffle.ViaRandomKey/Reshuffle/Window.Into()/Window.Assign
>  into 
> WriteCounts/WriteFiles/FinalizeTempFileBundles/Reshuffle.ViaRandomKey/Pair 
> with random key Nov 20, 2018 11:09:34 AM 
> org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process 
> INFO: 2018-11-20T19:09:32.044Z: Fusing consumer 
> WriteCounts/WriteFiles/FinalizeTempFileBundles/Reshuffle.ViaRandomKey/Reshuffle/ExpandIterable
>  into 
> WriteCounts/WriteFiles/FinalizeTempFileBundles/Reshuffle.ViaRandomKey/Reshuffle/GroupByKey/GroupByWindow
>  Nov 20, 2018 11:09:34 AM 
> org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process 
> INFO: 2018-11-20T19:09:32.099Z: Fusing consumer 
> WriteCounts/WriteFiles/FinalizeTempFileBundles/Reshuffle.ViaRandomKey/Values/Values/Map
>  into 
> WriteCounts/WriteFiles/FinalizeTempFileBundles/Reshuffle.ViaRandomKey/Reshuffle/ExpandIterable
>  Nov 20, 2018 11:09:34 AM 
> org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process 
> INFO: 2018-11-20T19:09:32.152Z: Fusing consumer 
> WriteCounts/WriteFiles/FinalizeTempFileBundles/Reshuffle.ViaRandomKey/Reshuffle/GroupByKey/Reify
>  into 
> WriteCounts/WriteFiles/FinalizeTempFileBundles/Reshuffle.ViaRandomKey/Reshuffle/Window.Into()/Window.Assign
>  Nov 20, 2018 11:09:34 AM 
> org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process 
> INFO: 2018-11-20T19:09:32.196Z: Fusing consumer 
> WriteCounts/WriteFiles/FinalizeTempFileBundles/Reshuffle.ViaRandomKey/Reshuffle/GroupByKey/Write
>  into 
> WriteCounts/WriteFiles/FinalizeTempFileBundles/Reshuffle.ViaRandomKey/Reshuffle/GroupByKey/Reify
>  Nov 20, 2018 11:09:34 AM 
> org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process 
> INFO: 2018-11-20T19:09:32.244Z: Fusing consumer 
> WriteCounts/WriteFiles/FinalizeTempFileBundles/Reshuffle.ViaRandomKey/Reshuffle/GroupByKey/GroupByWindow
>  into 
> WriteCounts/WriteFiles/FinalizeTempFileBundles/Reshuffle.ViaRandomKey/Reshuffle/GroupByKey/Read
>  Nov 20, 2018 11:09:34 AM 
> org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process 
> INFO: 2018-11-20T19:09:32.312Z: Fusing consumer 
> WriteCounts/WriteFiles/GatherTempFileResults/Reify.ReifyViewInGlobalWindow/Values/Values/Map
>  into 
> WriteCounts/WriteFiles/GatherTempFileResults/Reify.ReifyViewInGlobalWindow/Reify.ReifyView/ParDo(Anonymous)
>  Nov 20, 2018 11:09:34 AM 
> org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process 
> INFO: 2018-11-20T19:09:32.364Z: Fusing consumer 
> WriteCounts/WriteFiles/FinalizeTempFileBundles/Finalize into 
> WriteCounts/WriteFiles/GatherTempFileResults/Reify.ReifyViewInGlobalWindow/Values/Values/Map
>  Nov 20, 2018 11:09:34 AM 
> org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process 
> INFO: 2018-11-20T19:09:32.410Z: Fusing consumer 
> WriteCounts/WriteFiles/FinalizeTempFileBundles/Reshuffle.ViaRandomKey/Pair 
> with random key into WriteCounts/WriteFiles/FinalizeTempFileBundles/Finalize 
> Nov 20, 2018 11:09:34 AM 
> org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process 
> INFO: 2018-11-20T19:09:32.459Z: Unzipping flatten s12 for input 
> s11.org.apache.beam.sdk.values.PCollection.<init>:402#20ff67585e33a8f6 Nov 
> 20, 2018 11:09:34 AM 
> org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process 
> INFO: 2018-11-20T19:09:32.491Z: Fusing unzipped copy of 
> WriteCounts/WriteFiles/GatherTempFileResults/View.AsList/ParDo(ToIsmRecordForGlobalWindow),
>  through flatten 
> WriteCounts/WriteFiles/WriteUnshardedBundlesToTempFiles/Flatten.PCollections, 
> into producer 
> WriteCounts/WriteFiles/WriteUnshardedBundlesToTempFiles/DropShardNum Nov 20, 
> 2018 11:09:34 AM 
> org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process 
> INFO: 2018-11-20T19:09:32.536Z: Fusing consumer 
> WriteCounts/WriteFiles/GatherTempFileResults/View.AsList/ParDo(ToIsmRecordForGlobalWindow)
>  into 
> WriteCounts/WriteFiles/WriteUnshardedBundlesToTempFiles/WriteUnshardedBundles 
> Nov 20, 2018 11:09:34 AM 
> org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process 
> INFO: 2018-11-20T19:09:32.575Z: Fusing consumer 
> WriteCounts/WriteFiles/WriteUnshardedBundlesToTempFiles/DropShardNum into 
> WriteCounts/WriteFiles/WriteUnshardedBundlesToTempFiles/WriteUnwritten Nov 
> 20, 2018 11:09:34 AM 
> org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process 
> INFO: 2018-11-20T19:09:32.627Z: Fusing consumer 
> WordCount.CountWords/Count.PerElement/Combine.perKey(Count)/Combine.GroupedValues
>  into 
> WordCount.CountWords/Count.PerElement/Combine.perKey(Count)/GroupByKey/Read 
> Nov 20, 2018 11:09:34 AM 
> org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process 
> INFO: 2018-11-20T19:09:32.686Z: Fusing consumer 
> WriteCounts/WriteFiles/WriteUnshardedBundlesToTempFiles/GroupUnwritten/Write 
> into 
> WriteCounts/WriteFiles/WriteUnshardedBundlesToTempFiles/GroupUnwritten/Reify 
> Nov 20, 2018 11:09:34 AM 
> org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process 
> INFO: 2018-11-20T19:09:32.726Z: Fusing consumer 
> WriteCounts/WriteFiles/RewindowIntoGlobal/Window.Assign into MapElements/Map 
> Nov 20, 2018 11:09:34 AM 
> org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process 
> INFO: 2018-11-20T19:09:32.781Z: Fusing consumer 
> WordCount.CountWords/Count.PerElement/Combine.perKey(Count)/Combine.GroupedValues/Extract
>  into 
> WordCount.CountWords/Count.PerElement/Combine.perKey(Count)/Combine.GroupedValues
>  Nov 20, 2018 11:09:34 AM 
> org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process 
> INFO: 2018-11-20T19:09:32.838Z: Fusing consumer 
> WriteCounts/WriteFiles/WriteUnshardedBundlesToTempFiles/WriteUnwritten into 
> WriteCounts/WriteFiles/WriteUnshardedBundlesToTempFiles/GroupUnwritten/GroupByWindow
>  Nov 20, 2018 11:09:34 AM 
> org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process 
> INFO: 2018-11-20T19:09:32.884Z: Fusing consumer 
> WordCount.CountWords/Count.PerElement/Combine.perKey(Count)/GroupByKey+WordCount.CountWords/Count.PerElement/Combine.perKey(Count)/Combine.GroupedValues/Partial
>  into WordCount.CountWords/Count.PerElement/Init/Map Nov 20, 2018 11:09:34 AM 
> org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process 
> INFO: 2018-11-20T19:09:32.928Z: Fusing consumer 
> WordCount.CountWords/Count.PerElement/Combine.perKey(Count)/GroupByKey/Write 
> into 
> WordCount.CountWords/Count.PerElement/Combine.perKey(Count)/GroupByKey/Reify 
> Nov 20, 2018 11:09:34 AM 
> org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process 
> INFO: 2018-11-20T19:09:32.975Z: Fusing consumer 
> WriteCounts/WriteFiles/WriteUnshardedBundlesToTempFiles/WriteUnshardedBundles 
> into WriteCounts/WriteFiles/RewindowIntoGlobal/Window.Assign Nov 20, 2018 
> 11:09:34 AM 
> org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process 
> INFO: 2018-11-20T19:09:33.029Z: Fusing consumer 
> WordCount.CountWords/Count.PerElement/Init/Map into 
> WordCount.CountWords/ParDo(ExtractWords) Nov 20, 2018 11:09:34 AM 
> org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process 
> INFO: 2018-11-20T19:09:33.074Z: Fusing consumer 
> WriteCounts/WriteFiles/WriteUnshardedBundlesToTempFiles/GroupUnwritten/GroupByWindow
>  into 
> WriteCounts/WriteFiles/WriteUnshardedBundlesToTempFiles/GroupUnwritten/Read 
> Nov 20, 2018 11:09:34 AM 
> org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process 
> INFO: 2018-11-20T19:09:33.112Z: Fusing consumer MapElements/Map into 
> WordCount.CountWords/Count.PerElement/Combine.perKey(Count)/Combine.GroupedValues/Extract
>  Nov 20, 2018 11:09:34 AM 
> org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process 
> INFO: 2018-11-20T19:09:33.152Z: Fusing consumer 
> WordCount.CountWords/Count.PerElement/Combine.perKey(Count)/GroupByKey/Reify 
> into 
> WordCount.CountWords/Count.PerElement/Combine.perKey(Count)/GroupByKey+WordCount.CountWords/Count.PerElement/Combine.perKey(Count)/Combine.GroupedValues/Partial
>  Nov 20, 2018 11:09:34 AM 
> org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process 
> INFO: 2018-11-20T19:09:33.190Z: Fusing consumer 
> WriteCounts/WriteFiles/WriteUnshardedBundlesToTempFiles/GroupUnwritten/Reify 
> into 
> WriteCounts/WriteFiles/WriteUnshardedBundlesToTempFiles/WriteUnshardedBundles 
> Nov 20, 2018 11:09:34 AM 
> org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process 
> INFO: 2018-11-20T19:09:33.245Z: Fusing consumer 
> WordCount.CountWords/ParDo(ExtractWords) into ReadLines/Read Nov 20, 2018 
> 11:09:34 AM 
> org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process 
> INFO: 2018-11-20T19:09:33.764Z: Executing operation 
> WriteCounts/WriteFiles/FinalizeTempFileBundles/Reshuffle.ViaRandomKey/Reshuffle/GroupByKey/Create
>  Nov 20, 2018 11:09:34 AM 
> org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process 
> INFO: 2018-11-20T19:09:33.810Z: Executing operation 
> WordCount.CountWords/Count.PerElement/Combine.perKey(Count)/GroupByKey/Create 
> Nov 20, 2018 11:09:34 AM 
> org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process 
> INFO: 2018-11-20T19:09:33.850Z: Executing operation 
> WriteCounts/WriteFiles/WriteUnshardedBundlesToTempFiles/GroupUnwritten/Create 
> Nov 20, 2018 11:09:34 AM 
> org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process 
> INFO: 2018-11-20T19:09:33.889Z: Starting 1 workers in us-central1-a... Nov 
> 20, 2018 11:09:37 AM 
> org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process 
> INFO: 2018-11-20T19:09:34.261Z: Executing operation 
> ReadLines/Read+WordCount.CountWords/ParDo(ExtractWords)+WordCount.CountWords/Count.PerElement/Init/Map+WordCount.CountWords/Count.PerElement/Combine.perKey(Count)/GroupByKey+WordCount.CountWords/Count.PerElement/Combine.perKey(Count)/Combine.GroupedValues/Partial+WordCount.CountWords/Count.PerElement/Combine.perKey(Count)/GroupByKey/Reify+WordCount.CountWords/Count.PerElement/Combine.perKey(Count)/GroupByKey/Write
>  Nov 20, 2018 11:09:44 AM 
> org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process 
> INFO: 2018-11-20T19:09:43.287Z: Autoscaling: Raised the number of workers to 
> 0 based on the rate of progress in the currently running step(s). Nov 20, 
> 2018 11:10:16 AM 
> org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process 
> INFO: 2018-11-20T19:10:15.798Z: Autoscaling: Raised the number of workers to 
> 1 based on the rate of progress in the currently running step(s). Nov 20, 
> 2018 11:10:16 AM 
> org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process 
> INFO: 2018-11-20T19:10:15.859Z: Autoscaling: Would further reduce the number 
> of workers but reached the minimum number allowed for the job. Nov 20, 2018 
> 11:10:22 AM 
> org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process 
> INFO: 2018-11-20T19:10:21.327Z: Workers have started successfully. Nov 20, 
> 2018 11:10:22 AM 
> org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process 
> INFO: 2018-11-20T19:10:21.511Z: Workers have started successfully. Nov 20, 
> 2018 11:10:35 AM 
> org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process 
> SEVERE: 2018-11-20T19:10:32.138Z: java.lang.RuntimeException: 
> org.apache.beam.runners.dataflow.worker.repackaged.com.google.common.util.concurrent.UncheckedExecutionException:
>  java.lang.IllegalArgumentException: unable to deserialize Serialized 
> DoFnInfo at 
> org.apache.beam.runners.dataflow.worker.IntrinsicMapTaskExecutorFactory$1.typedApply(IntrinsicMapTaskExecutorFactory.java:193)
>  at 
> org.apache.beam.runners.dataflow.worker.IntrinsicMapTaskExecutorFactory$1.typedApply(IntrinsicMapTaskExecutorFactory.java:164)
>  at 
> org.apache.beam.runners.dataflow.worker.graph.Networks$TypeSafeNodeFunction.apply(Networks.java:63)
>  at 
> org.apache.beam.runners.dataflow.worker.graph.Networks$TypeSafeNodeFunction.apply(Networks.java:50)
>  at 
> org.apache.beam.runners.dataflow.worker.graph.Networks.replaceDirectedNetworkNodes(Networks.java:87)
>  at 
> org.apache.beam.runners.dataflow.worker.IntrinsicMapTaskExecutorFactory.create(IntrinsicMapTaskExecutorFactory.java:124)
>  at 
> org.apache.beam.runners.dataflow.worker.BatchDataflowWorker.doWork(BatchDataflowWorker.java:337)
>  at 
> org.apache.beam.runners.dataflow.worker.BatchDataflowWorker.getAndPerformWork(BatchDataflowWorker.java:291)
>  at 
> org.apache.beam.runners.dataflow.worker.DataflowBatchWorkerHarness$WorkerThread.doWork(DataflowBatchWorkerHarness.java:135)
>  at 
> org.apache.beam.runners.dataflow.worker.DataflowBatchWorkerHarness$WorkerThread.call(DataflowBatchWorkerHarness.java:115)
>  at 
> org.apache.beam.runners.dataflow.worker.DataflowBatchWorkerHarness$WorkerThread.call(DataflowBatchWorkerHarness.java:102)
>  at java.util.concurrent.FutureTask.run(FutureTask.java:266) at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
>  at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
>  at java.lang.Thread.run(Thread.java:745) Caused by: 
> org.apache.beam.runners.dataflow.worker.repackaged.com.google.common.util.concurrent.UncheckedExecutionException:
>  java.lang.IllegalArgumentException: unable to deserialize Serialized 
> DoFnInfo at 
> org.apache.beam.runners.dataflow.worker.repackaged.com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2214)
>  at 
> org.apache.beam.runners.dataflow.worker.repackaged.com.google.common.cache.LocalCache.get(LocalCache.java:4053)
>  at 
> org.apache.beam.runners.dataflow.worker.repackaged.com.google.common.cache.LocalCache$LocalManualCache.get(LocalCache.java:4899)
>  at 
> org.apache.beam.runners.dataflow.worker.UserParDoFnFactory.create(UserParDoFnFactory.java:91)
>  at 
> org.apache.beam.runners.dataflow.worker.DefaultParDoFnFactory.create(DefaultParDoFnFactory.java:75)
>  at 
> org.apache.beam.runners.dataflow.worker.IntrinsicMapTaskExecutorFactory.createParDoOperation(IntrinsicMapTaskExecutorFactory.java:263)
>  at 
> org.apache.beam.runners.dataflow.worker.IntrinsicMapTaskExecutorFactory.access$000(IntrinsicMapTaskExecutorFactory.java:85)
>  at 
> org.apache.beam.runners.dataflow.worker.IntrinsicMapTaskExecutorFactory$1.typedApply(IntrinsicMapTaskExecutorFactory.java:182)
>  ... 14 more Caused by: java.lang.IllegalArgumentException: unable to 
> deserialize Serialized DoFnInfo at 
> org.apache.beam.sdk.util.SerializableUtils.deserializeFromByteArray(SerializableUtils.java:74)
>  at 
> org.apache.beam.runners.dataflow.worker.UserParDoFnFactory$UserDoFnExtractor.getDoFnInfo(UserParDoFnFactory.java:62)
>  at 
> org.apache.beam.runners.dataflow.worker.UserParDoFnFactory.lambda$create$0(UserParDoFnFactory.java:93)
>  at 
> org.apache.beam.runners.dataflow.worker.repackaged.com.google.common.cache.LocalCache$LocalManualCache$1.load(LocalCache.java:4904)
>  at 
> org.apache.beam.runners.dataflow.worker.repackaged.com.google.common.cache.LocalCache$LoadingValueReference.loadFuture(LocalCache.java:3628)
>  at 
> org.apache.beam.runners.dataflow.worker.repackaged.com.google.common.cache.LocalCache$Segment.loadSync(LocalCache.java:2336)
>  at 
> org.apache.beam.runners.dataflow.worker.repackaged.com.google.common.cache.LocalCache$Segment.lockedGetOrLoad(LocalCache.java:2295)
>  at 
> org.apache.beam.runners.dataflow.worker.repackaged.com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2208)
>  ... 21 more Caused by: java.io.InvalidClassException: 
> org.apache.beam.sdk.values.WindowingStrategy; local class incompatible: 
> stream classdesc serialVersionUID = -6607512772692666907, local class 
> serialVersionUID = -3616600070988263902 at 
> java.io.ObjectStreamClass.initNonProxy(ObjectStreamClass.java:616) at 
> java.io.ObjectInputStream.readNonProxyDesc(ObjectInputStream.java:1630) at 
> java.io.ObjectInputStream.readClassDesc(ObjectInputStream.java:1521) at 
> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1781) at 
> java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1353) at 
> java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:2018) at 
> java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1942) at 
> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1808) at 
> java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1353) at 
> java.io.ObjectInputStream.readObject(ObjectInputStream.java:373) at 
> org.apache.beam.sdk.util.SerializableUtils.deserializeFromByteArray(SerializableUtils.java:71)
>  ... 28 more Nov 20, 2018 11:10:39 AM 
> org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process 
> SEVERE: 2018-11-20T19:10:36.934Z: java.lang.RuntimeException: 
> org.apache.beam.runners.dataflow.worker.repackaged.com.google.common.util.concurrent.UncheckedExecutionException:
>  java.lang.IllegalArgumentException: unable to deserialize Serialized 
> DoFnInfo at 
> org.apache.beam.runners.dataflow.worker.IntrinsicMapTaskExecutorFactory$1.typedApply(IntrinsicMapTaskExecutorFactory.java:193)
>  at 
> org.apache.beam.runners.dataflow.worker.IntrinsicMapTaskExecutorFactory$1.typedApply(IntrinsicMapTaskExecutorFactory.java:164)
>  at 
> org.apache.beam.runners.dataflow.worker.graph.Networks$TypeSafeNodeFunction.apply(Networks.java:63)
>  at 
> org.apache.beam.runners.dataflow.worker.graph.Networks$TypeSafeNodeFunction.apply(Networks.java:50)
>  at 
> org.apache.beam.runners.dataflow.worker.graph.Networks.replaceDirectedNetworkNodes(Networks.java:87)
>  at 
> org.apache.beam.runners.dataflow.worker.IntrinsicMapTaskExecutorFactory.create(IntrinsicMapTaskExecutorFactory.java:124)
>  at 
> org.apache.beam.runners.dataflow.worker.BatchDataflowWorker.doWork(BatchDataflowWorker.java:337)
>  at 
> org.apache.beam.runners.dataflow.worker.BatchDataflowWorker.getAndPerformWork(BatchDataflowWorker.java:291)
>  at 
> org.apache.beam.runners.dataflow.worker.DataflowBatchWorkerHarness$WorkerThread.doWork(DataflowBatchWorkerHarness.java:135)
>  at 
> org.apache.beam.runners.dataflow.worker.DataflowBatchWorkerHarness$WorkerThread.call(DataflowBatchWorkerHarness.java:115)
>  at 
> org.apache.beam.runners.dataflow.worker.DataflowBatchWorkerHarness$WorkerThread.call(DataflowBatchWorkerHarness.java:102)
>  at java.util.concurrent.FutureTask.run(FutureTask.java:266) at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
>  at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
>  at java.lang.Thread.run(Thread.java:745) Caused by: 
> org.apache.beam.runners.dataflow.worker.repackaged.com.google.common.util.concurrent.UncheckedExecutionException:
>  java.lang.IllegalArgumentException: unable to deserialize Serialized 
> DoFnInfo at 
> org.apache.beam.runners.dataflow.worker.repackaged.com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2214)
>  at 
> org.apache.beam.runners.dataflow.worker.repackaged.com.google.common.cache.LocalCache.get(LocalCache.java:4053)
>  at 
> org.apache.beam.runners.dataflow.worker.repackaged.com.google.common.cache.LocalCache$LocalManualCache.get(LocalCache.java:4899)
>  at 
> org.apache.beam.runners.dataflow.worker.UserParDoFnFactory.create(UserParDoFnFactory.java:91)
>  at 
> org.apache.beam.runners.dataflow.worker.DefaultParDoFnFactory.create(DefaultParDoFnFactory.java:75)
>  at 
> org.apache.beam.runners.dataflow.worker.IntrinsicMapTaskExecutorFactory.createParDoOperation(IntrinsicMapTaskExecutorFactory.java:263)
>  at 
> org.apache.beam.runners.dataflow.worker.IntrinsicMapTaskExecutorFactory.access$000(IntrinsicMapTaskExecutorFactory.java:85)
>  at 
> org.apache.beam.runners.dataflow.worker.IntrinsicMapTaskExecutorFactory$1.typedApply(IntrinsicMapTaskExecutorFactory.java:182)
>  ... 14 more Caused by: java.lang.IllegalArgumentException: unable to 
> deserialize Serialized DoFnInfo at 
> org.apache.beam.sdk.util.SerializableUtils.deserializeFromByteArray(SerializableUtils.java:74)
>  at 
> org.apache.beam.runners.dataflow.worker.UserParDoFnFactory$UserDoFnExtractor.getDoFnInfo(UserParDoFnFactory.java:62)
>  at 
> org.apache.beam.runners.dataflow.worker.UserParDoFnFactory.lambda$create$0(UserParDoFnFactory.java:93)
>  at 
> org.apache.beam.runners.dataflow.worker.repackaged.com.google.common.cache.LocalCache$LocalManualCache$1.load(LocalCache.java:4904)
>  at 
> org.apache.beam.runners.dataflow.worker.repackaged.com.google.common.cache.LocalCache$LoadingValueReference.loadFuture(LocalCache.java:3628)
>  at 
> org.apache.beam.runners.dataflow.worker.repackaged.com.google.common.cache.LocalCache$Segment.loadSync(LocalCache.java:2336)
>  at 
> org.apache.beam.runners.dataflow.worker.repackaged.com.google.common.cache.LocalCache$Segment.lockedGetOrLoad(LocalCache.java:2295)
>  at 
> org.apache.beam.runners.dataflow.worker.repackaged.com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2208)
>  ... 21 more Caused by: java.io.InvalidClassException: 
> org.apache.beam.sdk.values.WindowingStrategy; local class incompatible: 
> stream classdesc serialVersionUID = -6607512772692666907, local class 
> serialVersionUID = -3616600070988263902 at 
> java.io.ObjectStreamClass.initNonProxy(ObjectStreamClass.java:616) at 
> java.io.ObjectInputStream.readNonProxyDesc(ObjectInputStream.java:1630) at 
> java.io.ObjectInputStream.readClassDesc(ObjectInputStream.java:1521) at 
> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1781) at 
> java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1353) at 
> java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:2018) at 
> java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1942) at 
> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1808) at 
> java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1353) at 
> java.io.ObjectInputStream.readObject(ObjectInputStream.java:373) at 
> org.apache.beam.sdk.util.SerializableUtils.deserializeFromByteArray(SerializableUtils.java:71)
>  ... 28 more Nov 20, 2018 11:10:45 AM 
> org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process 
> SEVERE: 2018-11-20T19:10:44.927Z: java.lang.RuntimeException: 
> org.apache.beam.runners.dataflow.worker.repackaged.com.google.common.util.concurrent.UncheckedExecutionException:
>  java.lang.IllegalArgumentException: unable to deserialize Serialized 
> DoFnInfo at 
> org.apache.beam.runners.dataflow.worker.IntrinsicMapTaskExecutorFactory$1.typedApply(IntrinsicMapTaskExecutorFactory.java:193)
>  at 
> org.apache.beam.runners.dataflow.worker.IntrinsicMapTaskExecutorFactory$1.typedApply(IntrinsicMapTaskExecutorFactory.java:164)
>  at 
> org.apache.beam.runners.dataflow.worker.graph.Networks$TypeSafeNodeFunction.apply(Networks.java:63)
>  at 
> org.apache.beam.runners.dataflow.worker.graph.Networks$TypeSafeNodeFunction.apply(Networks.java:50)
>  at 
> org.apache.beam.runners.dataflow.worker.graph.Networks.replaceDirectedNetworkNodes(Networks.java:87)
>  at 
> org.apache.beam.runners.dataflow.worker.IntrinsicMapTaskExecutorFactory.create(IntrinsicMapTaskExecutorFactory.java:124)
>  at 
> org.apache.beam.runners.dataflow.worker.BatchDataflowWorker.doWork(BatchDataflowWorker.java:337)
>  at 
> org.apache.beam.runners.dataflow.worker.BatchDataflowWorker.getAndPerformWork(BatchDataflowWorker.java:291)
>  at 
> org.apache.beam.runners.dataflow.worker.DataflowBatchWorkerHarness$WorkerThread.doWork(DataflowBatchWorkerHarness.java:135)
>  at 
> org.apache.beam.runners.dataflow.worker.DataflowBatchWorkerHarness$WorkerThread.call(DataflowBatchWorkerHarness.java:115)
>  at 
> org.apache.beam.runners.dataflow.worker.DataflowBatchWorkerHarness$WorkerThread.call(DataflowBatchWorkerHarness.java:102)
>  at java.util.concurrent.FutureTask.run(FutureTask.java:266) at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
>  at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
>  at java.lang.Thread.run(Thread.java:745) Caused by: 
> org.apache.beam.runners.dataflow.worker.repackaged.com.google.common.util.concurrent.UncheckedExecutionException:
>  java.lang.IllegalArgumentException: unable to deserialize Serialized 
> DoFnInfo at 
> org.apache.beam.runners.dataflow.worker.repackaged.com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2214)
>  at 
> org.apache.beam.runners.dataflow.worker.repackaged.com.google.common.cache.LocalCache.get(LocalCache.java:4053)
>  at 
> org.apache.beam.runners.dataflow.worker.repackaged.com.google.common.cache.LocalCache$LocalManualCache.get(LocalCache.java:4899)
>  at 
> org.apache.beam.runners.dataflow.worker.UserParDoFnFactory.create(UserParDoFnFactory.java:91)
>  at 
> org.apache.beam.runners.dataflow.worker.DefaultParDoFnFactory.create(DefaultParDoFnFactory.java:75)
>  at 
> org.apache.beam.runners.dataflow.worker.IntrinsicMapTaskExecutorFactory.createParDoOperation(IntrinsicMapTaskExecutorFactory.java:263)
>  at 
> org.apache.beam.runners.dataflow.worker.IntrinsicMapTaskExecutorFactory.access$000(IntrinsicMapTaskExecutorFactory.java:85)
>  at 
> org.apache.beam.runners.dataflow.worker.IntrinsicMapTaskExecutorFactory$1.typedApply(IntrinsicMapTaskExecutorFactory.java:182)
>  ... 14 more Caused by: java.lang.IllegalArgumentException: unable to 
> deserialize Serialized DoFnInfo at 
> org.apache.beam.sdk.util.SerializableUtils.deserializeFromByteArray(SerializableUtils.java:74)
>  at 
> org.apache.beam.runners.dataflow.worker.UserParDoFnFactory$UserDoFnExtractor.getDoFnInfo(UserParDoFnFactory.java:62)
>  at 
> org.apache.beam.runners.dataflow.worker.UserParDoFnFactory.lambda$create$0(UserParDoFnFactory.java:93)
>  at 
> org.apache.beam.runners.dataflow.worker.repackaged.com.google.common.cache.LocalCache$LocalManualCache$1.load(LocalCache.java:4904)
>  at 
> org.apache.beam.runners.dataflow.worker.repackaged.com.google.common.cache.LocalCache$LoadingValueReference.loadFuture(LocalCache.java:3628)
>  at 
> org.apache.beam.runners.dataflow.worker.repackaged.com.google.common.cache.LocalCache$Segment.loadSync(LocalCache.java:2336)
>  at 
> org.apache.beam.runners.dataflow.worker.repackaged.com.google.common.cache.LocalCache$Segment.lockedGetOrLoad(LocalCache.java:2295)
>  at 
> org.apache.beam.runners.dataflow.worker.repackaged.com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2208)
>  ... 21 more Caused by: java.io.InvalidClassException: 
> org.apache.beam.sdk.values.WindowingStrategy; local class incompatible: 
> stream classdesc serialVersionUID = -6607512772692666907, local class 
> serialVersionUID = -3616600070988263902 at 
> java.io.ObjectStreamClass.initNonProxy(ObjectStreamClass.java:616) at 
> java.io.ObjectInputStream.readNonProxyDesc(ObjectInputStream.java:1630) at 
> java.io.ObjectInputStream.readClassDesc(ObjectInputStream.java:1521) at 
> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1781) at 
> java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1353) at 
> java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:2018) at 
> java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1942) at 
> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1808) at 
> java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1353) at 
> java.io.ObjectInputStream.readObject(ObjectInputStream.java:373) at 
> org.apache.beam.sdk.util.SerializableUtils.deserializeFromByteArray(SerializableUtils.java:71)
>  ... 28 more Nov 20, 2018 11:11:01 AM 
> org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process 
> SEVERE: 2018-11-20T19:11:00.729Z: java.lang.RuntimeException: 
> org.apache.beam.runners.dataflow.worker.repackaged.com.google.common.util.concurrent.UncheckedExecutionException:
>  java.lang.IllegalArgumentException: unable to deserialize Serialized 
> DoFnInfo at 
> org.apache.beam.runners.dataflow.worker.IntrinsicMapTaskExecutorFactory$1.typedApply(IntrinsicMapTaskExecutorFactory.java:193)
>  at 
> org.apache.beam.runners.dataflow.worker.IntrinsicMapTaskExecutorFactory$1.typedApply(IntrinsicMapTaskExecutorFactory.java:164)
>  at 
> org.apache.beam.runners.dataflow.worker.graph.Networks$TypeSafeNodeFunction.apply(Networks.java:63)
>  at 
> org.apache.beam.runners.dataflow.worker.graph.Networks$TypeSafeNodeFunction.apply(Networks.java:50)
>  at 
> org.apache.beam.runners.dataflow.worker.graph.Networks.replaceDirectedNetworkNodes(Networks.java:87)
>  at 
> org.apache.beam.runners.dataflow.worker.IntrinsicMapTaskExecutorFactory.create(IntrinsicMapTaskExecutorFactory.java:124)
>  at 
> org.apache.beam.runners.dataflow.worker.BatchDataflowWorker.doWork(BatchDataflowWorker.java:337)
>  at 
> org.apache.beam.runners.dataflow.worker.BatchDataflowWorker.getAndPerformWork(BatchDataflowWorker.java:291)
>  at 
> org.apache.beam.runners.dataflow.worker.DataflowBatchWorkerHarness$WorkerThread.doWork(DataflowBatchWorkerHarness.java:135)
>  at 
> org.apache.beam.runners.dataflow.worker.DataflowBatchWorkerHarness$WorkerThread.call(DataflowBatchWorkerHarness.java:115)
>  at 
> org.apache.beam.runners.dataflow.worker.DataflowBatchWorkerHarness$WorkerThread.call(DataflowBatchWorkerHarness.java:102)
>  at java.util.concurrent.FutureTask.run(FutureTask.java:266) at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
>  at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
>  at java.lang.Thread.run(Thread.java:745) Caused by: 
> org.apache.beam.runners.dataflow.worker.repackaged.com.google.common.util.concurrent.UncheckedExecutionException:
>  java.lang.IllegalArgumentException: unable to deserialize Serialized 
> DoFnInfo at 
> org.apache.beam.runners.dataflow.worker.repackaged.com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2214)
>  at 
> org.apache.beam.runners.dataflow.worker.repackaged.com.google.common.cache.LocalCache.get(LocalCache.java:4053)
>  at 
> org.apache.beam.runners.dataflow.worker.repackaged.com.google.common.cache.LocalCache$LocalManualCache.get(LocalCache.java:4899)
>  at 
> org.apache.beam.runners.dataflow.worker.UserParDoFnFactory.create(UserParDoFnFactory.java:91)
>  at 
> org.apache.beam.runners.dataflow.worker.DefaultParDoFnFactory.create(DefaultParDoFnFactory.java:75)
>  at 
> org.apache.beam.runners.dataflow.worker.IntrinsicMapTaskExecutorFactory.createParDoOperation(IntrinsicMapTaskExecutorFactory.java:263)
>  at 
> org.apache.beam.runners.dataflow.worker.IntrinsicMapTaskExecutorFactory.access$000(IntrinsicMapTaskExecutorFactory.java:85)
>  at 
> org.apache.beam.runners.dataflow.worker.IntrinsicMapTaskExecutorFactory$1.typedApply(IntrinsicMapTaskExecutorFactory.java:182)
>  ... 14 more Caused by: java.lang.IllegalArgumentException: unable to 
> deserialize Serialized DoFnInfo at 
> org.apache.beam.sdk.util.SerializableUtils.deserializeFromByteArray(SerializableUtils.java:74)
>  at 
> org.apache.beam.runners.dataflow.worker.UserParDoFnFactory$UserDoFnExtractor.getDoFnInfo(UserParDoFnFactory.java:62)
>  at 
> org.apache.beam.runners.dataflow.worker.UserParDoFnFactory.lambda$create$0(UserParDoFnFactory.java:93)
>  at 
> org.apache.beam.runners.dataflow.worker.repackaged.com.google.common.cache.LocalCache$LocalManualCache$1.load(LocalCache.java:4904)
>  at 
> org.apache.beam.runners.dataflow.worker.repackaged.com.google.common.cache.LocalCache$LoadingValueReference.loadFuture(LocalCache.java:3628)
>  at 
> org.apache.beam.runners.dataflow.worker.repackaged.com.google.common.cache.LocalCache$Segment.loadSync(LocalCache.java:2336)
>  at 
> org.apache.beam.runners.dataflow.worker.repackaged.com.google.common.cache.LocalCache$Segment.lockedGetOrLoad(LocalCache.java:2295)
>  at 
> org.apache.beam.runners.dataflow.worker.repackaged.com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2208)
>  ... 21 more Caused by: java.io.InvalidClassException: 
> org.apache.beam.sdk.values.WindowingStrategy; local class incompatible: 
> stream classdesc serialVersionUID = -6607512772692666907, local class 
> serialVersionUID = -3616600070988263902 at 
> java.io.ObjectStreamClass.initNonProxy(ObjectStreamClass.java:616) at 
> java.io.ObjectInputStream.readNonProxyDesc(ObjectInputStream.java:1630) at 
> java.io.ObjectInputStream.readClassDesc(ObjectInputStream.java:1521) at 
> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1781) at 
> java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1353) at 
> java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:2018) at 
> java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1942) at 
> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1808) at 
> java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1353) at 
> java.io.ObjectInputStream.readObject(ObjectInputStream.java:373) at 
> org.apache.beam.sdk.util.SerializableUtils.deserializeFromByteArray(SerializableUtils.java:71)
>  ... 28 more Nov 20, 2018 11:11:01 AM 
> org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process 
> SEVERE: 2018-11-20T19:11:00.885Z: Workflow failed. Causes: 
> S02:ReadLines/Read+WordCount.CountWords/ParDo(ExtractWords)+WordCount.CountWords/Count.PerElement/Init/Map+WordCount.CountWords/Count.PerElement/Combine.perKey(Count)/GroupByKey+WordCount.CountWords/Count.PerElement/Combine.perKey(Count)/Combine.GroupedValues/Partial+WordCount.CountWords/Count.PerElement/Combine.perKey(Count)/GroupByKey/Reify+WordCount.CountWords/Count.PerElement/Combine.perKey(Count)/GroupByKey/Write
>  failed., A work item was attempted 4 times without success. Each time the 
> worker eventually lost contact with the service. The work item was attempted 
> on: wordcount-goenka-11201909-11201109-n7q9-harness-6q3n, 
> wordcount-goenka-11201909-11201109-n7q9-harness-6q3n, 
> wordcount-goenka-11201909-11201109-n7q9-harness-6q3n, 
> wordcount-goenka-11201909-11201109-n7q9-harness-6q3n Nov 20, 2018 11:11:02 AM 
> org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process 
> INFO: 2018-11-20T19:11:01.059Z: Cleaning up. Nov 20, 2018 11:11:02 AM 
> org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process 
> INFO: 2018-11-20T19:11:01.473Z: Stopping worker pool...
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

Reply via email to