Thank you for your quick reply again, Jeff. Yes i know the difference of the execution environment between "%sh" and ">spark-submit". And my question was "how to execute spark-submit as shell interpreter". That mean, i am searching how to execute a binary program from a note of zeppelin. This time it has been limited on Spark.
Seems like Zeppelin have several procedure to execute Spark shell, like spark.pyspark, spark.sql.... So how to do "spark-submit" was my wondering. I am sorry for bothering Your time, but at the same time, i am appreciated if You get my wondering clearly, and show me some tips. -Keiji 2017-10-03 18:30 GMT+09:00 Jeff Zhang <zjf...@gmail.com>: > %sh is shell interpreter, you can run spark-submit just as you run it in > shell terminal. > > 小野圭二 <onoke...@gmail.com>于2017年10月3日周二 下午4:58写道: > >> Thank you for your reply, Jeff >> >> "%sh" ? >> "sh" seems like request something execution code. >> I tried "%sh", then >> >> %sh <program name with full path> >> %sh bash: <program name>: no permission >> >> I made binary file from .py to .pyc, but the answer was as same. >> I am sorry seems like doubting you, but Is "%sh" the resolution? >> >> -Keiji >> >> 2017-10-03 17:35 GMT+09:00 Jianfeng (Jeff) Zhang <jzh...@hortonworks.com> >> : >> >>> >>> I am surprised why would you use %spark-submit, there’s no document >>> about %spark-submit. If you want to use spark-submit in zeppelin, then >>> you could use %sh >>> >>> >>> Best Regard, >>> Jeff Zhang >>> >>> >>> From: 小野圭二 <onoke...@gmail.com> >>> Reply-To: "users@zeppelin.apache.org" <users@zeppelin.apache.org> >>> Date: Tuesday, October 3, 2017 at 12:49 PM >>> To: "users@zeppelin.apache.org" <users@zeppelin.apache.org> >>> Subject: How to execute spark-submit on Note >>> >>> Hi all, >>> >>> I searched this topic on the archive of ml, but still could not find out >>> the solution clearly. >>> So i have tried to post this again(maybe). >>> >>> I am using ver 0.8.0, and have installed spark 2.2 on the other path, >>> just for checking my test program. >>> Then i wrote a quite simple sample python code to check the how to. >>> >>> 1. the code works fine on a note in Zeppelin >>> 2. the same code but added the initialize code for SparkContext in it >>> works fine on the Spark by using 'spark-submit'. >>> 3. tried to execute "2" from a note in Zeppelin with the following >>> script. >>> yes, "spark" interpreter has been implemented in the note. >>> then on the note, >>> %spark-submit <program name with full path> >>> -> interpreter not found error >>> 4.I have arranged 'SPARK_SUBMIT_OPTIONS' in zeppelin-env.sh order by the >>> doc >>> ex. export SPARK_SUBMIT_OPTIONS='--packages >>> com.databricks:spark-csv_2.10:1.2.0' >>> 5. then running.... >>> %spark-submit <program name with full path> >>> -> interpreter not found error (as same as "3") >>> >>> How can i use spark-submit from a note? >>> Any advice thanks. >>> >>> -Keiji >>> >> >>