Re: Where to specify trust.jks

2023-05-20 Thread Yu Watanabe
://github.com/yuwtennis/apache-beam-pipeline-apps/blob/main/java/Dockerfile Thanks, Yu Watanabe On Fri, May 19, 2023 at 12:28 AM Pablo Estrada via user wrote: > > Hi Utkarsh, > you can pass a path in GCS (or a filesystem), and the workers should be able > to download it onto themselves

Re: How to run expansion service using go sdk in local development environment ?

2022-08-29 Thread Yu Watanabe
Hello Danny. Ah . I see . Thank you for your advice. Thanks, Yu Watanabe On Mon, Aug 29, 2022 at 9:26 AM Danny McCormick via user wrote: > > Hey Yu, as the error you posted suggests, the Go direct runner which you're > using in your local development environment doesn'

How to run expansion service using go sdk in local development environment ?

2022-08-28 Thread Yu Watanabe
.github.com/yuwtennis/dec3bf3bfc0c4fa54d9d3565c98d008e Thanks, Yu -- Yu Watanabe linkedin: www.linkedin.com/in/yuwatanabe1/ twitter: twitter.com/yuwtennis

Re: How to register as external cross language transform ?

2022-08-17 Thread Yu Watanabe
: > > > > On Wed, Aug 17, 2022 at 3:05 AM Yu Watanabe wrote: >> >> Hello. >> >> I am trying to write code for cross language transform for >> ElasticsearchIO but having trouble with it. >> I would appreciate it if I could get help. >> >> A

How to register as external cross language transform ?

2022-08-17 Thread Yu Watanabe
egards, Yu Watanabe -- Yu Watanabe linkedin: www.linkedin.com/in/yuwatanabe1/ twitter: twitter.com/yuwtennis

Re: Any guideline for building golang connector ?

2022-07-08 Thread Yu Watanabe
Hello Danny. Thank you for the details. I appreciate your message. I am a newbie around building io . So I will look into the links and first build my knowledge. Thanks, Yu Watanabe On Fri, Jul 8, 2022 at 8:28 PM Danny McCormick via user < user@beam.apache.org> wrote: > Hey Yu

Any guideline for building golang connector ?

2022-07-08 Thread Yu Watanabe
Hello . Is there any guideline for building a go sdk connector ? I was reviewing the document but I could not find one for golang. https://beam.apache.org/documentation/io/developing-io-overview/ I was thinking of building one for elasticsearch. Thanks, Yu Watanabe -- Yu Watanabe linkedin

Re: Is FileIO for Azure Blob Storage GA?

2021-12-21 Thread Yu Watanabe
olidays for a little while. > > [1] > https://beam.apache.org/releases/javadoc/2.34.0/org/apache/beam/sdk/io/azure/blobstore/package-summary.html > > On Tue, Dec 21, 2021 at 4:47 AM Yu Watanabe wrote: > >> Hello . >> >> Looks like unit tests are passed in these PRs

Re: Is FileIO for Azure Blob Storage GA?

2021-12-21 Thread Yu Watanabe
/documentation/io/built-in/ On Tue, Dec 21, 2021 at 10:50 AM 池上有希乃 wrote: > Hi > > I think I will use Apache Beam for my business. > So there is a question. > Whether FileIO for Azure Blob Storage is GA phase or not? > > Thanks, > Yukino Ikegami > -- Yu Watanabe li

Re: java.io.InvalidClassException with Spark 3.1.2

2021-08-22 Thread Yu Watanabe
he code, if you would like to contribute. help me *star if if it > is useful for you. thank you > > On Monday, August 16, 2021, 12:37:46 AM GMT+8, Yu Watanabe < > yu.w.ten...@gmail.com> wrote: > > > Hello . > > I would like to ask question for spark runner. > &g

Re: java.io.InvalidClassException with Spark 3.1.2

2021-08-19 Thread Yu Watanabe
it. [2] > > [1] > https://github.com/apache/beam/pull/14897/commits/b6fca2bb79d9e7a69044b477460445456720ec58 > [2] https://issues.apache.org/jira/browse/BEAM-12762 > > > On Sun, Aug 15, 2021 at 9:37 AM Yu Watanabe wrote: > >> Hello . >> >> I would like to ask

java.io.InvalidClassException with Spark 3.1.2

2021-08-15 Thread Yu Watanabe
/spark-2.4.8-bin-hadoop2.7.tgz Would there be any setting which you need to be aware of for spark 3.1.2 ? Thanks, Yu Watanabe -- Yu Watanabe linkedin: www.linkedin.com/in/yuwatanabe1/ twitter: twitter.com/yuwtennis

Re: Submit Python Beam on Spark Dataproc

2021-08-15 Thread Yu Watanabe
> *SSH tunnel to the master node:* > gcloud compute ssh \ > --project \ > --zone \ > -- -NL 7077:localhost:7077 > > - > > Thanks, > Mahan > > On Tue, Aug 10, 2021 at 3:53 PM Yu Watanabe wrote: > >> Hello . >> >> Would

Re: Submit Python Beam on Spark Dataproc

2021-08-10 Thread Yu Watanabe
;s the environment_type? Can we use DOCKER? Then what's the SDK > Harness Configuration? > 4- Should we run the job-server outside of the Dataproc cluster or should > we run it in the master node? > > Thanks, > Mahan > -- Yu Watanabe linkedin: www.linkedin.com/in/yuwatanabe1/ twitter: twitter.com/yuwtennis

Re: Running Beam on Flink

2020-02-16 Thread Yu Watanabe
esolving with 127.0.0.1 and see if it works ? Thanks, Yu Watanabe On Mon, Feb 10, 2020 at 3:29 AM Xander Song wrote: > > Hi Jincheng, > > Thanks for your help. Yes, I am using Mac. Your suggestion allowed me to > submit the job on port 8099. However, I am now encountering

Re: Assertion error when using kafka module in python

2020-01-25 Thread Yu Watanabe
hu, Jan 9, 2020 at 7:17 AM Chamikara Jayalath wrote: > > Hmm, seems like a Java (external) ParDo is being forwarded to Python SDK for > execution somehow. +Maximilian Michels might know more. > > On Sun, Jan 5, 2020 at 2:57 AM Yu Watanabe wrote: >> >> Hello. >&

Assertion error when using kafka module in python

2020-01-05 Thread Yu Watanabe
- I have my platform built upon docker engine and used below combination of modules. 1. apache-beam: 2.16.0 2. flink: 1.8 3. python-sdk(37): compiled using release-2.16.0 4. jobserver: compiled using release-2.16.0 5. kafka:

Re: Worker pool dies with error: context deadline exceeded

2020-01-05 Thread Yu Watanabe
== $ gsutil cat gs://${PROJECT_ID}/sample.txt-0-of-1 Hello World. Apache beam ==== Thanks, Yu Watanabe On Fri, Jan 3, 2020 at 5:58 AM Kyle Weaver wrote: > This is the root cause: > > > python-sdk_

Re: Beam on Flink with Python SDK and using GCS as artifacts directory

2019-12-31 Thread Yu Watanabe
Matthew > Just to verify that the preferred python version is python3? Harness container supports both python 2 and 3. https://beam.apache.org/documentation/runtime/environments/ In my opinion, considering that python2's EOL is Jan 1, 2020 , python 3 would be the choice. Thanks, Yu

Worker pool dies with error: context deadline exceeded

2019-12-30 Thread Yu Watanabe
oyment/blob/master/flink-session-cluster/docker/docker-compose.yml My pipeline code https://github.com/yuwtennis/beam-deployment/blob/master/flink-session-cluster/docker/samples/src/sample.py Would there be any settings I need to use for starting up sdk container ? Best Regards, Yu Watanabe -- Yu Watanabe yu.w.ten...@gmail.com

Re: Protocol message had invalid UTF-8

2019-12-30 Thread Yu Watanabe
Kyle. Thank you for the reply. Error had disappeared after creating sdk and job-server with correct apache beam branch (2.16.0 in this case). Thanks, Yu Watanabe On Tue, Dec 31, 2019 at 2:30 AM Kyle Weaver wrote: > This error can happen when the job server and sdk versions are mismatc

Protocol message had invalid UTF-8

2019-12-30 Thread Yu Watanabe
st", "${JOB_HOST}", "--job-port", "${JOB_PORT}" ] ========== I appreciate if I could get some help. Thanks, Yu Watanabe -- Yu Watanabe yu.w.ten...@gmail.com

Re: How to import external module inside ParDo using Apache Flink ?

2019-10-07 Thread Yu Watanabe
7, 2019 at 4:49 PM Yu Watanabe wrote: > Thank you for the comment. > > I finally got this working. I would like to share my experience for people > whom are beginner with portable runner. > What I done was below items when calling functions and classes from > external pa

Re: How to import external module inside ParDo using Apache Flink ?

2019-10-07 Thread Yu Watanabe
0', 'elasticsearch>=7.0.0,<8.0.0', 'urllib3', 'boto3' ] setuptools.setup( author = 'Yu Watanabe', author_email = 'AUTHOR_

Re: How to reference manifest from apache flink worker node ?

2019-09-27 Thread Yu Watanabe
eature > complete, but we're still smoothing out some of the ease-of-use issues. And > feedback like this really helps, so thanks! > > On Tue, Sep 24, 2019 at 8:10 PM Yu Watanabe wrote: > >> I needed little adjustment with the pipeline option to work it out. >> >> P

How to import external module inside ParDo using Apache Flink ?

2019-09-25 Thread Yu Watanabe
;--environment_type=DOCKER", "--experiments=beam_fn_api", "--job_endpoint=localhost:8099" ]) options.view_as(SetupOptions).save_main_session = True options.view_as(SetupOptions).setup_file = '/home/admin/quality-validation/bin/setup.py' = Is it possible to solve dependency in ParDo linked to external module when using Apache Flink? Thanks, Yu Watanabe -- Yu Watanabe Weekend Freelancer who loves to challenge building data platform yu.w.ten...@gmail.com [image: LinkedIn icon] <https://www.linkedin.com/in/yuwatanabe1> [image: Twitter icon] <https://twitter.com/yuwtennis>

Re: How do you call user defined ParDo class from the pipeline in Portable Runner (Apache Flink) ?

2019-09-25 Thread Yu Watanabe
Actually there was a good example in the latest wordcount.py in master repo. https://github.com/apache/beam/blob/master/sdks/python/apache_beam/examples/wordcount.py On Thu, Sep 26, 2019 at 12:00 PM Yu Watanabe wrote: > Thank you for the help. > > I have chosen to remove the super()

Re: How do you call user defined ParDo class from the pipeline in Portable Runner (Apache Flink) ?

2019-09-25 Thread Yu Watanabe
reference to super in your dofn should help. > > On Wed, Sep 25, 2019 at 5:13 PM Yu Watanabe wrote: > >> Thank you for the reply. >> >> " save_main_session" did not work, however, situation had changed. >> >>

Re: How do you call user defined ParDo class from the pipeline in Portable Runner (Apache Flink) ?

2019-09-25 Thread Yu Watanabe
Pre Frame Line postFrm[f], # Post Frame Line info['pre'][0][1],# Pre Last Modified Time info['post'][0][1]) # Post Last Modified Time yield (fra

How do you call user defined ParDo class from the pipeline in Portable Runner (Apache Flink) ?

2019-09-25 Thread Yu Watanabe
======= This is not a problem when running pipeline using DirectRunner. May I ask , how should I import class for ParDo when running on Flink ? Thanks, Yu Watanabe -- Yu Watanabe Weekend Freelancer who loves to challenge building data platform yu.w.ten...@gmail.com [image: LinkedIn icon] <https://www.linkedin.com/in/yuwatanabe1> [image: Twitter icon] <https://twitter.com/yuwtennis>

Re: How to reference manifest from apache flink worker node ?

2019-09-24 Thread Yu Watanabe
s-dir /home/ec2-user --job-port 8099 > > 3. Submit your pipeline > options = PipelineOptions([ > "--runner=PortableRunner", > "--environment_config= > asia.gcr.io/creationline001/beam/python3:latest&qu

Re: How to reference manifest from apache flink worker node ?

2019-09-24 Thread Yu Watanabe
"--experiments=beam_fn_api" > ]) > > On Mon, Sep 23, 2019 at 8:20 PM Yu Watanabe wrote: > >> Kyle. >> >> Thank you for the assistance. >> >> Looks like "--artifacts-dir" is not parsed. Below is the DEBUG

Re: How to reference manifest from apache flink worker node ?

2019-09-23 Thread Yu Watanabe
27;-jar', '/home/admin/.apache_beam/cache/beam-runners-flink-1.8-job-server-2.15.0.jar', '--flink-master-url', 'ip-172-31-12-113.ap-northeast-1.compute.internal:8081', '--artifacts-dir', '/home/admin/artifacts18unmki3', '--job-port', 46767, '--artifact-port', 0, 

How to reference manifest from apache flink worker node ?

2019-09-23 Thread Yu Watanabe
act directory from worker node? I appreciate if I could get some advice . Best Regards, Yu Watanabe -- Yu Watanabe Weekend Freelancer who loves to challenge building data platform yu.w.ten...@gmail.com [image: LinkedIn icon] <https://www.linkedin.com/in/yuwatanabe1> [image: Twitter icon] &

Re: Submitting job to AWS EMR fails with error=2, No such file or directory

2019-09-20 Thread Yu Watanabe
t;docker pull" if the container is already pulled ? Thanks, Yu Watanabe On Fri, Sep 20, 2019 at 7:48 PM Benjamin Tan wrote: > Seems like some file is missing. Why not manually docker pull the image > (remember to adjust the container location to omit to registry) locally > f

Re: Submitting job to AWS EMR fails with error=2, No such file or directory

2019-09-20 Thread Yu Watanabe
un(Task.java:711) at java.lang.Thread.run(Thread.java:748) ... ====== Thanks, Yu Watanabe On Thu, Sep 19, 2019 at 3:47 AM Ankur Goenka wrote: > Adding to the previous suggestions. > You can also add "--retain_docker_container" to your pipeline option an

Re: Submitting job to AWS EMR fails with error=2, No such file or directory

2019-09-20 Thread Yu Watanabe
==-- java.util.concurrent.ExecutionException: java.io.FileNotFoundException: /tmp/artifactskjsmgv8p/job_ea4673de-71d0-4dd9-b683-fe8c52464666/MANIFEST (No such file or directory) ===-- Thanks, Y

Re: apache beam 2.16.0 ?

2019-09-18 Thread Yu Watanabe
16.0 is ongoing and it will be released > once blockers are solved. > > > -Rui > > On Wed, Sep 18, 2019 at 9:34 PM Yu Watanabe wrote: > >> Hello. >> >> I would like to use 2.16.0 to diagnose container problem, however, looks >> like the job-server is not

apache beam 2.16.0 ?

2019-09-18 Thread Yu Watanabe
-flink-1.8-job-server-2.16.0.jar: HTTP Error 404: Not Found Checked maven repo and indeed there is no job-server 2.16.0 yet. https://mvnrepository.com/artifact/org.apache.beam/beam-runners-flink Will 2.16.0 released soon ? Thanks, Yu Watanabe -- Yu Watanabe Weekend Freelancer who loves to

Re: Submitting job to AWS EMR fails with error=2, No such file or directory

2019-09-18 Thread Yu Watanabe
ot;:\"/opt/apache/beam/boot\"} > > On 2019/09/18 10:40:42, Yu Watanabe wrote: > > Hello. > > > > I am trying to run FlinkRunner (2.15.0) on AWS EC2 instance and submit > job > > to

Re: Submitting job to AWS EMR fails with error=2, No such file or directory

2019-09-18 Thread Yu Watanabe
nner.py#L138 config = json.loads(portable_options.environment_config) May I ask what command I need to set as shell script for each task managers ? Best Regards, Yu Watanabe On Wed, Sep 18, 2019 at 9:39 PM Benjamin Tan wrote: > Seems like docker is not installed. Maybe run with PROC

Submitting job to AWS EMR fails with error=2, No such file or directory

2019-09-18 Thread Yu Watanabe
--flink-master-url ip-172-31-1-84.ap-northeast-1.compute.internal:43581 --artifacts-dir /tmp/artifactskj47j8yn --job-port 48205 --artifact-port 0 --expansion-port 0 admin@ip-172-31-9-89:/opt/flink$ ====- Would there be any other setting I need to

Re: How to use google container registry for FlinkRunner ?

2019-09-17 Thread Yu Watanabe
]) p = beam.Pipeline(options=options) ======= Thanks, Yu Watanabe On Tue, Sep 17, 2019 at 11:21 AM Benjamin Tan wrote: > Something like this: > > options = PipelineOptions(["--runner=PortableRunner&quo

How to use google container registry for FlinkRunner ?

2019-09-16 Thread Yu Watanabe
======= Perhaps is there any environment variable to specify which image to use ? Best Regards, Yu Watanabe -- Yu Watanabe Weekend Freelancer who loves to challenge building data platform yu.w.ten...@gmail.com [image: LinkedIn icon] <https://www.linkedin.com/in/yuwatanabe1> [image: Twitter icon] <https://twitter.com/yuwtennis>

Re: Flink Runner logging FAILED_TO_UNCOMPRESS

2019-09-14 Thread Yu Watanabe
nstead of key 'rest.address' [flink-runner-job-invoker] INFO org.apache.flink.runtime.rest.RestClient - Rest client endpoint started. [flink-runner-job-invoker] INFO org.apache.flink.client.program.rest.RestClusterClient - Submitting job 4e055a8878dda3f564a7b7c84d48510d (detached: false). Thanks, Yu Watan

Flink Runner logging FAILED_TO_UNCOMPRESS

2019-09-14 Thread Yu Watanabe
with beam.Pipeline(options=options) as p: (p | beam.Create(["Hello World"])) === Would there be any other settings should I look for ? Thanks, Yu Watanabe -- Yu Watanabe Weekend Freelancer who loves to c

Re: How do you write portable runner pipeline on separate python code ?

2019-09-14 Thread Yu Watanabe
DFS/S3/GCS/... > * Mounting an external directory into the container so that any "local" > writes appear outside the container > * Using a non-docker environment such as external or process. > > On Thu, Sep 12, 2019 at 4:51 AM Yu Watanabe wrote: > >> Hello. >> >>

Re: How do you write portable runner pipeline on separate python code ?

2019-09-14 Thread Yu Watanabe
localhost:8099 (this is the default address of the JobService). For example: ======== Thanks, Yu Watanabe On Fri, Sep 13, 2019 at 5:09 AM Kyle Weaver wrote: > +dev I think we should probably point new users of > the porta

Re: How to buffer events using spark portable runner ?

2019-09-12 Thread Yu Watanabe
Lukasaz Thank you for the reply. I will try apache flink. Thanks, Yu On Sun, Sep 8, 2019 at 11:59 PM Lukasz Cwik wrote: > Try using Apache Flink. > > On Sun, Sep 8, 2019 at 6:23 AM Yu Watanabe wrote: > >> Hello . >> >> I would like to ask question related to

How do you write portable runner pipeline on separate python code ?

2019-09-12 Thread Yu Watanabe
l_endpoint=localhost:36079'. stderr: Unable to find image ' ywatanabe-docker-apache.bintray.io/beam/python3:latest' locallydocker: Error response from daemon: unknown: Subject ywatanabe was not found.See 'docker run --help'. ----------- A

How to buffer events using spark portable runner ?

2019-09-08 Thread Yu Watanabe
://beam.apache.org/documentation/runners/capability-matrix/#cap-summary-when Is there anyway in portable runner that you can do similar processing as *timely processing* ? This is my first time using portable runner and I appreciate if I can get help with this. Best Regards, Yu Watanabe -- Yu

Which memory profiler is used for python 3.7.3 ?

2019-07-24 Thread Yu Watanabe
ich memory profiler do users of apache beam use for python 3.7.3 ? https://pypi.org/project/memory-profiler/ Thanks, Yu Watanabe -- Yu Watanabe Weekend Freelancer who loves to challenge building data platform yu.w.ten...@gmail.com [image: LinkedIn icon] <https://www.linkedin.com/in/yuwatan

Re: Any way to profile speed for each transforms ?

2019-07-24 Thread Yu Watanabe
t; For a raw dump you could do something like: >> >> p = beam.Pipeline(...) >> p | beam.Read... >> results = p.run() >> results.wait_until_finish() >> import pprint >> pprint.pprint(results._metrics_by_stage) >> >> >> >> >

Any way to profile speed for each transforms ?

2019-07-24 Thread Yu Watanabe
speed for each transform. Best Regards, Yu Watanabe -- Yu Watanabe Weekend Freelancer who loves to challenge building data platform yu.w.ten...@gmail.com [image: LinkedIn icon] <https://www.linkedin.com/in/yuwatanabe1> [image: Twitter icon] <https://twitter.com/yuwtennis>