Re: PySpark RDD method errors

2015-07-20 Thread moon soo Lee
VA_OPTS[@]}" >> >> cp conf/zeppelin-env.sh.template conf/zeppelin-env.sh >> cat <> conf/zeppelin-env.sh >> export MASTER=yarn-client >> export HADOOP_CONF_DIR=$HADOOP_CONF_DIR >> export ZEPPELIN_SPARK_USEHIVECONTEXT=false >> export ZEPPELIN_JAVA_OPTS

Re: PySpark RDD method errors

2015-07-17 Thread moon soo Lee
ent > export HADOOP_CONF_DIR=$HADOOP_CONF_DIR > export ZEPPELIN_SPARK_USEHIVECONTEXT=false > export ZEPPELIN_JAVA_OPTS="${ZEPPELIN_JAVA_OPTS[@]}" > EOF > > > Thank you so much for helping > > -Chad > > From: moon soo Lee > Reply-To: "users@zeppelin.incubator.apache

Re: PySpark RDD method errors

2015-07-13 Thread Chad Timmins
pelin.incubator.apache.org>" mailto:users@zeppelin.incubator.apache.org>> Date: Monday, July 13, 2015 at 12:25 PM To: "users@zeppelin.incubator.apache.org<mailto:users@zeppelin.incubator.apache.org>" mailto:users@zeppelin.incubator.apache.org>> Subject: Re: Py

Re: PySpark RDD method errors

2015-07-13 Thread moon soo Lee
> > From: moon soo Lee > Reply-To: "users@zeppelin.incubator.apache.org" < > users@zeppelin.incubator.apache.org> > Date: Sunday, July 12, 2015 at 8:59 AM > To: "users@zeppelin.incubator.apache.org" < > users@zeppelin.incubator.apache.org> > Su

Re: PySpark RDD method errors

2015-07-13 Thread Chad Timmins
@zeppelin.incubator.apache.org<mailto:users@zeppelin.incubator.apache.org>" mailto:users@zeppelin.incubator.apache.org>> Subject: Re: PySpark RDD method errors Hi, Thanks for sharing the problem. I have tried with AWS EMR and i could make all the code works without error. I've set

Re: PySpark RDD method errors

2015-07-12 Thread moon soo Lee
Hi, Thanks for sharing the problem. I have tried with AWS EMR and i could make all the code works without error. I've set export HADOOP_CONF_DIR=/home/hadoop/conf export SPARK_HOME=/home/hadoop/spark export ZEPPELIN_PORT=9090 with 'yarn-client' for master property. export SPARK_HOME is not corr

PySpark RDD method errors

2015-07-09 Thread Chad Timmins
Hi, When I run the filter() method on an RDD object and then try to print its results using collect(), I get a Py4JJavaError. It is not only filter but other methods that cause similar errors and I cannot figure out what is causing this. PySpark from the command line works fine, but it does n