Thank You Xintong, I will look for these updates in the near future.

Regards,
Vijay

On Wed, May 26, 2021 at 6:40 PM Xintong Song <tonysong...@gmail.com> wrote:

> Hi Vijay,
>
> Currently, Flink only supports shipping files from the local machine where
> job is submitted.
>
> There are tickets [1][2][3] tracking the efforts that shipping files from
> remote paths, e.g., http, hdfs, etc. Once the efforts are done, adding s3
> as an additional supported schema should be straightforward.
>
> Unfortunately, these efforts are still in progress, and are more or less
> staled recently.
>
> Thank you~
>
> Xintong Song
>
>
> [1] https://issues.apache.org/jira/browse/FLINK-20681
> [2] https://issues.apache.org/jira/browse/FLINK-20811
> [3] https://issues.apache.org/jira/browse/FLINK-20867
>
> On Thu, May 27, 2021 at 12:23 AM Vijayendra Yadav <contact....@gmail.com>
> wrote:
>
>> Hi Pohl,
>>
>> I tried to ship my property file. Example: *-yarn.ship-files
>> s3://applib/xx/xx/1.0-SNAPSHOT/application.properties  \*
>>
>>
>> *Error:*
>>
>> 6:21:37.163 [main] ERROR org.apache.flink.client.cli.CliFrontend -
>> Invalid command line arguments.
>> org.apache.flink.client.cli.CliArgsException: Could not build the program
>> from JAR file: JAR file does not exist: -yarn.ship-files
>>         at
>> org.apache.flink.client.cli.CliFrontend.getPackagedProgram(CliFrontend.java:244)
>> ~[flink-dist_2.11-1.11.0.jar:1.11.0]
>>         at
>> org.apache.flink.client.cli.CliFrontend.run(CliFrontend.java:223)
>> ~[flink-dist_2.11-1.11.0.jar:1.11.0]
>>         at
>> org.apache.flink.client.cli.CliFrontend.parseParameters(CliFrontend.java:916)
>> ~[flink-dist_2.11-1.11.0.jar:1.11.0]
>>         at
>> org.apache.flink.client.cli.CliFrontend.lambda$main$10(CliFrontend.java:992)
>> ~[flink-dist_2.11-1.11.0.jar:1.11.0]
>>         at java.security.AccessController.doPrivileged(Native Method)
>> ~[?:1.8.0_292]
>>         at javax.security.auth.Subject.doAs(Subject.java:422)
>> [?:1.8.0_292]
>>         at
>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1893)
>> [hadoop-common-2.10.0-amzn-0.jar:?]
>>         at
>> org.apache.flink.runtime.security.contexts.HadoopSecurityContext.runSecured(HadoopSecurityContext.java:41)
>> [flink-dist_2.11-1.11.0.jar:1.11.0]
>>         at
>> org.apache.flink.client.cli.CliFrontend.main(CliFrontend.java:992)
>> [flink-dist_2.11-1.11.0.jar:1.11.0]
>> Caused by: java.io.FileNotFoundException: JAR file does not exist:
>> -yarn.ship-files
>>         at
>> org.apache.flink.client.cli.CliFrontend.getJarFile(CliFrontend.java:740)
>> ~[flink-dist_2.11-1.11.0.jar:1.11.0]
>>         at
>> org.apache.flink.client.cli.CliFrontend.buildProgram(CliFrontend.java:717)
>> ~[flink-dist_2.11-1.11.0.jar:1.11.0]
>>         at
>> org.apache.flink.client.cli.CliFrontend.getPackagedProgram(CliFrontend.java:242)
>> ~[flink-dist_2.11-1.11.0.jar:1.11.0]
>>         ... 8 more
>> Could not build the program from JAR file: JAR file does not exist:
>> -yarn.ship-files
>>
>>
>> *Thanks,*
>>
>> *Vijay*
>>
>> On Tue, May 25, 2021 at 11:58 PM Matthias Pohl <matth...@ververica.com>
>> wrote:
>>
>>> Hi Vijay,
>>> have you tried yarn-ship-files [1] or yarn-ship-archives [2]? Maybe,
>>> that's what you're looking for...
>>>
>>> Best,
>>> Matthias
>>>
>>> [1]
>>> https://ci.apache.org/projects/flink/flink-docs-release-1.13/docs/deployment/config/#yarn-ship-files
>>> [2]
>>> https://ci.apache.org/projects/flink/flink-docs-release-1.13/docs/deployment/config/#yarn-ship-archives
>>>
>>> On Tue, May 25, 2021 at 5:56 PM Vijayendra Yadav <contact....@gmail.com>
>>> wrote:
>>>
>>>> Hi Piotr,
>>>>
>>>> I have been doing the same process as you mentioned so far, now I am
>>>> migrating the deployment process using AWS CDK and AWS Step Functions, kind
>>>> of like the CICD process.
>>>> I added a download step of jar and configs (1, 2, 3 and 4) from S3
>>>> using command-runner.jar (AWS Step); it loaded that into one of the Master
>>>> nodes (out of 3). In the next step when I launched Flink Job it would not
>>>> find build because Job is launched in some other yarn node.
>>>>
>>>> I was hoping just like *Apache spark *where whatever files we provide
>>>> in *--file*s are shipped to yarn (s3 to yarn workfirectory), Flink
>>>> should also have a solution.
>>>>
>>>> Thanks,
>>>> Vijay
>>>>
>>>>
>>>> On Tue, May 25, 2021 at 12:50 AM Piotr Nowojski <pnowoj...@apache.org>
>>>> wrote:
>>>>
>>>>> Hi Vijay,
>>>>>
>>>>> I'm not sure if I understand your question correctly. You have jar and
>>>>> configs (1, 2, 3 and 4) on S3 and you want to start a Flink job using
>>>>> those? Can you simply download those things (whole directory containing
>>>>> those) to the machine that will be starting the Flink job?
>>>>>
>>>>> Best, Piotrek
>>>>>
>>>>> wt., 25 maj 2021 o 07:50 Vijayendra Yadav <contact....@gmail.com>
>>>>> napisaƂ(a):
>>>>>
>>>>>> Hi Team,
>>>>>>
>>>>>> I am trying to find a way to ship files from aws s3 for a flink
>>>>>> streaming job, I am running on AWS EMR. What i need to ship are 
>>>>>> following:
>>>>>> 1) application jar
>>>>>> 2) application property file
>>>>>> 3) custom flink-conf.yaml
>>>>>> 4) log4j application specific
>>>>>>
>>>>>> Please let me know options.
>>>>>>
>>>>>> Thanks,
>>>>>> Vijay
>>>>>
>>>>>

Reply via email to