Maybe your application is overriding the master variable when it creates its
SparkContext. I see you are still passing “yarn-client” as an argument later to
it in your command.
> On Jun 17, 2018, at 11:53 AM, Raymond Xie wrote:
>
> Thank you Subhash.
>
> Here is the new command:
>
Thank you Subhash.
Here is the new command:
spark-submit --master local[*] --class retail_db.GetRevenuePerOrder --conf
spark.ui.port=12678 spark2practice_2.11-0.1.jar yarn-client
/public/retail_db/order_items /home/rxie/output/revenueperorder
Still seeing the same issue here.
2018-06-17 11:51:25
Hi Raymond,
If you set your master to local[*] instead of yarn-client, it should run on
your local machine.
Thanks,
Subhash
Sent from my iPhone
> On Jun 17, 2018, at 2:32 PM, Raymond Xie wrote:
>
> Hello,
>
> I am wondering how can I run spark job in my environment which is a single
>
Hello,
I am wondering how can I run spark job in my environment which is a single
Ubuntu host with no hadoop installed? if I run my job like below, I will
end up with infinite loop at the end. Thank you very much.
rxie@ubuntu:~/data$ spark-submit --class retail_db.GetRevenuePerOrder
--conf