Yes, I already tested with spark-shell and pyspark , with the same result.

Can't I use Linux filesystem to read CSV, such as file:///data/file.csv. My
understanding is that the job is sent and is interpreted in the worker,
isn't it?

Thanks.

El El mar, 9 may 2017 a las 20:23, Jongyoul Lee <jongy...@gmail.com>
escribió:

> Could you test if it works with spark-shell?
>
> On Sun, May 7, 2017 at 5:22 PM, Sofiane Cherchalli <sofian...@gmail.com>
> wrote:
>
>> Hi,
>>
>> I have a standalone cluster, one master and one worker, running in
>> separate nodes. Zeppelin is running is in a separate node too in client
>> mode.
>>
>> When I run a notebook that reads a CSV file located in the worker
>> node with Spark-CSV package, Zeppelin tries to read the CSV locally and
>> fails because the CVS is in the worker node and not in Zeppelin node.
>>
>> Is this the expected behavior?
>>
>> Thanks.
>>
>
>
>
> --
> 이종열, Jongyoul Lee, 李宗烈
> http://madeng.net
>

Reply via email to