Well I think I will build a REST API over Livy in order to import/export data 
into HDFS.

 

Thanks all!

 

Joaquín Silva | Pentagon Security & AKAINIX

Av. Kennedy 4.700, Piso 10, Of. 1002, Edificio New Century, Vitacura | Código 
Postal (ZIP Code) 7561127

Cel: (56-9) 6304 2498 

 

From: Vivek Suvarna [mailto:vikk...@gmail.com] 
Sent: martes, 25 de julio de 2017 1:19
To: user@livy.incubator.apache.org
Subject: Re: Input file as an argument og a Spark code

 

I had a similar requirement. 

I used webhdfs to first copy the file across to hdfs before starting the spark 
job via Livy. 



Sent from my iPhone


On 25 Jul 2017, at 9:39 AM, Saisai Shao <sai.sai.s...@gmail.com> wrote:

        I think you have to make this csv file accessible from Spark cluster, 
putting to HDFS is one possible solution. 

         

        On Tue, Jul 25, 2017 at 1:26 AM, Joaquín Silva <joaq...@akainix.com> 
wrote:

        Hello,

         

        I'm building a BASH program (using Curl)  that should run a Spark code 
remotely using Livy. But one of the code argument  is a CSV file, how can I 
make that spark reads this file?. The file is going to be in client side, not 
in the Spark cluster machines.

         

        Regards,

         

        Joaquín Silva

         

         

Reply via email to