[jira] [Commented] (SPARK-7909) spark-ec2 and associated tools not py3 ready

2015-07-08 Thread Shivaram Venkataraman (JIRA)

[ 
https://issues.apache.org/jira/browse/SPARK-7909?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14619799#comment-14619799
 ] 

Shivaram Venkataraman commented on SPARK-7909:
--

I think we merged all the Python3 patches that were open in mesos/spark-ec2 -- 
[~meawoppl] [~davies] can we mark this as fixed now ?

> spark-ec2 and associated tools not py3 ready
> 
>
> Key: SPARK-7909
> URL: https://issues.apache.org/jira/browse/SPARK-7909
> Project: Spark
>  Issue Type: Improvement
>  Components: EC2
> Environment: ec2 python3
>Reporter: Matthew Goodman
>Priority: Blocker
>
> At present there is not a possible permutation of tools that supports Python3 
> on both the launching computer and running cluster.  There are a couple 
> problems involved:
>  - There is no prebuilt spark binary with python3 support.
>  - spark-ec2/spark/init.sh contains inline py3 unfriendly print statements
>  - Config files for cluster processes don't seem to make it to all nodes in a 
> working format.
> I have fixes for some of this, but the config and running context debugging 
> remains elusive to me.  



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-7909) spark-ec2 and associated tools not py3 ready

2015-07-08 Thread Sean Owen (JIRA)

[ 
https://issues.apache.org/jira/browse/SPARK-7909?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14619764#comment-14619764
 ] 

Sean Owen commented on SPARK-7909:
--

In parallel, there's a suggestion to move the EC2 support out of apache/spark 
entirely into its own repo, where it can rejoin the other EC2 Spark support 
from mesos/spark-ec2. Would that make this any easier, if it could evolve 
separately and as one unit?

> spark-ec2 and associated tools not py3 ready
> 
>
> Key: SPARK-7909
> URL: https://issues.apache.org/jira/browse/SPARK-7909
> Project: Spark
>  Issue Type: Improvement
>  Components: EC2
> Environment: ec2 python3
>Reporter: Matthew Goodman
>Priority: Blocker
>
> At present there is not a possible permutation of tools that supports Python3 
> on both the launching computer and running cluster.  There are a couple 
> problems involved:
>  - There is no prebuilt spark binary with python3 support.
>  - spark-ec2/spark/init.sh contains inline py3 unfriendly print statements
>  - Config files for cluster processes don't seem to make it to all nodes in a 
> working format.
> I have fixes for some of this, but the config and running context debugging 
> remains elusive to me.  



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-7909) spark-ec2 and associated tools not py3 ready

2015-06-01 Thread Shivaram Venkataraman (JIRA)

[ 
https://issues.apache.org/jira/browse/SPARK-7909?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14567735#comment-14567735
 ] 

Shivaram Venkataraman commented on SPARK-7909:
--

Yeah feel free to open a PR for the `print` fixes.

> spark-ec2 and associated tools not py3 ready
> 
>
> Key: SPARK-7909
> URL: https://issues.apache.org/jira/browse/SPARK-7909
> Project: Spark
>  Issue Type: Improvement
>  Components: EC2
> Environment: ec2 python3
>Reporter: Matthew Goodman
>
> At present there is not a possible permutation of tools that supports Python3 
> on both the launching computer and running cluster.  There are a couple 
> problems involved:
>  - There is no prebuilt spark binary with python3 support.
>  - spark-ec2/spark/init.sh contains inline py3 unfriendly print statements
>  - Config files for cluster processes don't seem to make it to all nodes in a 
> working format.
> I have fixes for some of this, but the config and running context debugging 
> remains elusive to me.  



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-7909) spark-ec2 and associated tools not py3 ready

2015-06-01 Thread Matthew Goodman (JIRA)

[ 
https://issues.apache.org/jira/browse/SPARK-7909?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14567685#comment-14567685
 ] 

Matthew Goodman commented on SPARK-7909:


Awesome, thanks for all the help on this.  There is one (possibly unrelated) 
issue remains, which is that httpd seems to fail to startup, giving the 
following traceback:

{code:title=HTTPD Failure Traceback|borderStyle=solid}
Starting httpd: httpd: Syntax error on line 154 of /etc/httpd/conf/httpd.conf: 
Cannot load /etc/httpd/modules/mod_authz_core.so into server: 
/etc/httpd/modules/mod_authz_core.so: cannot open shared object file: No such 
file or directory
{code}

Should I send in a PR [for this 
change|https://github.com/3Scan/spark-ec2/commit/3416dd07c492b0cddcc98c4fa83f9e4284ed8fc9]?
  

> spark-ec2 and associated tools not py3 ready
> 
>
> Key: SPARK-7909
> URL: https://issues.apache.org/jira/browse/SPARK-7909
> Project: Spark
>  Issue Type: Improvement
>  Components: EC2
> Environment: ec2 python3
>Reporter: Matthew Goodman
>
> At present there is not a possible permutation of tools that supports Python3 
> on both the launching computer and running cluster.  There are a couple 
> problems involved:
>  - There is no prebuilt spark binary with python3 support.
>  - spark-ec2/spark/init.sh contains inline py3 unfriendly print statements
>  - Config files for cluster processes don't seem to make it to all nodes in a 
> working format.
> I have fixes for some of this, but the config and running context debugging 
> remains elusive to me.  



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-7909) spark-ec2 and associated tools not py3 ready

2015-05-28 Thread Shivaram Venkataraman (JIRA)

[ 
https://issues.apache.org/jira/browse/SPARK-7909?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14564168#comment-14564168
 ] 

Shivaram Venkataraman commented on SPARK-7909:
--

The packages will get to S3 once the 1.4 release is finalized. We are still 
testing / voting on release candidates and you can follow these on the Spark 
developer mailing list.  BTW I also have a change open at spark-ec2 for 
substituting the Spark version based on pattern 
https://github.com/mesos/spark-ec2/pull/116/files#diff-1d040c3294246f2b59643d63868fc2ad,
 so that should take care of picking up the binary once its released.

However, feel free to send out PRs for the other python3 print fixes you had to 
make in init.sh etc. 

> spark-ec2 and associated tools not py3 ready
> 
>
> Key: SPARK-7909
> URL: https://issues.apache.org/jira/browse/SPARK-7909
> Project: Spark
>  Issue Type: Improvement
>  Components: EC2
> Environment: ec2 python3
>Reporter: Matthew Goodman
>
> At present there is not a possible permutation of tools that supports Python3 
> on both the launching computer and running cluster.  There are a couple 
> problems involved:
>  - There is no prebuilt spark binary with python3 support.
>  - spark-ec2/spark/init.sh contains inline py3 unfriendly print statements
>  - Config files for cluster processes don't seem to make it to all nodes in a 
> working format.
> I have fixes for some of this, but the config and running context debugging 
> remains elusive to me.  



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-7909) spark-ec2 and associated tools not py3 ready

2015-05-28 Thread Matthew Goodman (JIRA)

[ 
https://issues.apache.org/jira/browse/SPARK-7909?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14564166#comment-14564166
 ] 

Matthew Goodman commented on SPARK-7909:


Using the prebuilt binaries from the links provided yields a working cluster.  
Is there a timeline for when the spark 1.4.0 binaries make the s3 bucket?  I 
can add the link to the spark/init.sh script, but it will bounce until the 
binary is actually place in the bucket.

In either case I suspect the naming convention will be similar, so would a PR 
for the changes outlined above be a good step at this stage?

> spark-ec2 and associated tools not py3 ready
> 
>
> Key: SPARK-7909
> URL: https://issues.apache.org/jira/browse/SPARK-7909
> Project: Spark
>  Issue Type: Improvement
>  Components: EC2
> Environment: ec2 python3
>Reporter: Matthew Goodman
>
> At present there is not a possible permutation of tools that supports Python3 
> on both the launching computer and running cluster.  There are a couple 
> problems involved:
>  - There is no prebuilt spark binary with python3 support.
>  - spark-ec2/spark/init.sh contains inline py3 unfriendly print statements
>  - Config files for cluster processes don't seem to make it to all nodes in a 
> working format.
> I have fixes for some of this, but the config and running context debugging 
> remains elusive to me.  



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-7909) spark-ec2 and associated tools not py3 ready

2015-05-28 Thread Shivaram Venkataraman (JIRA)

[ 
https://issues.apache.org/jira/browse/SPARK-7909?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14563634#comment-14563634
 ] 

Shivaram Venkataraman commented on SPARK-7909:
--

Yeah I think that error comes from trying to use Hadoop2 while the cluster is 
configured for Hadoop1. I think the pre-built binaries shouldn't have that 
problem

> spark-ec2 and associated tools not py3 ready
> 
>
> Key: SPARK-7909
> URL: https://issues.apache.org/jira/browse/SPARK-7909
> Project: Spark
>  Issue Type: Improvement
>  Components: EC2
> Environment: ec2 python3
>Reporter: Matthew Goodman
>
> At present there is not a possible permutation of tools that supports Python3 
> on both the launching computer and running cluster.  There are a couple 
> problems involved:
>  - There is no prebuilt spark binary with python3 support.
>  - spark-ec2/spark/init.sh contains inline py3 unfriendly print statements
>  - Config files for cluster processes don't seem to make it to all nodes in a 
> working format.
> I have fixes for some of this, but the config and running context debugging 
> remains elusive to me.  



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-7909) spark-ec2 and associated tools not py3 ready

2015-05-28 Thread Matthew Goodman (JIRA)

[ 
https://issues.apache.org/jira/browse/SPARK-7909?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14563609#comment-14563609
 ] 

Matthew Goodman commented on SPARK-7909:


There are 11 folders in /root/spark/work/app-20150528200603-/, all with the 
same traceback below different only in time of error
{code:title=Spark worker Traceback|borderStyle=solid}
15/05/28 20:06:04 INFO executor.CoarseGrainedExecutorBackend: Registered signal 
handlers for [TERM, HUP, INT]
Exception in thread "main" java.lang.ExceptionInInitializerError
at 
org.apache.spark.executor.CoarseGrainedExecutorBackend$.run(CoarseGrainedExecutorBackend.scala:146)
at 
org.apache.spark.executor.CoarseGrainedExecutorBackend$.main(CoarseGrainedExecutorBackend.scala:245)
at 
org.apache.spark.executor.CoarseGrainedExecutorBackend.main(CoarseGrainedExecutorBackend.scala)
Caused by: java.lang.RuntimeException: 
java.lang.reflect.InvocationTargetException
at 
org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:131)
at org.apache.hadoop.security.Groups.(Groups.java:55)
at 
org.apache.hadoop.security.Groups.getUserToGroupsMappingService(Groups.java:182)
at 
org.apache.hadoop.security.UserGroupInformation.initialize(UserGroupInformation.java:235)
at 
org.apache.hadoop.security.UserGroupInformation.setConfiguration(UserGroupInformation.java:249)
at 
org.apache.spark.deploy.SparkHadoopUtil.(SparkHadoopUtil.scala:50)
at 
org.apache.spark.deploy.SparkHadoopUtil$.(SparkHadoopUtil.scala:353)
at 
org.apache.spark.deploy.SparkHadoopUtil$.(SparkHadoopUtil.scala)
... 3 more
Caused by: java.lang.reflect.InvocationTargetException
at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
at 
sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57)
at 
sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
at java.lang.reflect.Constructor.newInstance(Constructor.java:526)
at 
org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:129)
... 10 more
Caused by: java.lang.UnsatisfiedLinkError: 
org.apache.hadoop.security.JniBasedUnixGroupsMapping.anchorNative()V
at 
org.apache.hadoop.security.JniBasedUnixGroupsMapping.anchorNative(Native Method)
at 
org.apache.hadoop.security.JniBasedUnixGroupsMapping.(JniBasedUnixGroupsMapping.java:49)
at 
org.apache.hadoop.security.JniBasedUnixGroupsMappingWithFallback.(JniBasedUnixGroupsMappingWithFallback.java:38)
... 15 more
{code}

My launch script is as follows:
{code:title=Spark Launch Call|borderStyle=solid}
bash spark-ec2 --spark-version=ab62d73ddb973c25de043e8e9ade7800adf244e8 
--spark-ec2-git-repo=https://github.com/3scan/spark-ec2 
--spark-ec2-git-branch=branch-1.4 --key-pair=blahblahblah 
--identity-file=blahblahblah.pem --region us-west-2 --user-data 
/home/meawoppl/repos/3scan-analysis/spark/linux-bootstrap.sh login test-cluster
{code}

I am going to try the prebuilt spark next.  I suspect this is surrounding the 
compiled/checked out version that I am running?  Not sure.


> spark-ec2 and associated tools not py3 ready
> 
>
> Key: SPARK-7909
> URL: https://issues.apache.org/jira/browse/SPARK-7909
> Project: Spark
>  Issue Type: Improvement
>  Components: EC2
> Environment: ec2 python3
>Reporter: Matthew Goodman
>
> At present there is not a possible permutation of tools that supports Python3 
> on both the launching computer and running cluster.  There are a couple 
> problems involved:
>  - There is no prebuilt spark binary with python3 support.
>  - spark-ec2/spark/init.sh contains inline py3 unfriendly print statements
>  - Config files for cluster processes don't seem to make it to all nodes in a 
> working format.
> I have fixes for some of this, but the config and running context debugging 
> remains elusive to me.  



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-7909) spark-ec2 and associated tools not py3 ready

2015-05-28 Thread Shivaram Venkataraman (JIRA)

[ 
https://issues.apache.org/jira/browse/SPARK-7909?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14563581#comment-14563581
 ] 

Shivaram Venkataraman commented on SPARK-7909:
--

You can find the 1.4 RC2 binaries at 
http://people.apache.org/~pwendell/spark-releases/spark-1.4.0-rc2-bin/  

Also the worker logs are at `/root/spark/work///stderr` on 
the slave machines 

> spark-ec2 and associated tools not py3 ready
> 
>
> Key: SPARK-7909
> URL: https://issues.apache.org/jira/browse/SPARK-7909
> Project: Spark
>  Issue Type: Improvement
>  Components: EC2
> Environment: ec2 python3
>Reporter: Matthew Goodman
>
> At present there is not a possible permutation of tools that supports Python3 
> on both the launching computer and running cluster.  There are a couple 
> problems involved:
>  - There is no prebuilt spark binary with python3 support.
>  - spark-ec2/spark/init.sh contains inline py3 unfriendly print statements
>  - Config files for cluster processes don't seem to make it to all nodes in a 
> working format.
> I have fixes for some of this, but the config and running context debugging 
> remains elusive to me.  



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-7909) spark-ec2 and associated tools not py3 ready

2015-05-28 Thread Matthew Goodman (JIRA)

[ 
https://issues.apache.org/jira/browse/SPARK-7909?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14563488#comment-14563488
 ] 

Matthew Goodman commented on SPARK-7909:


Thanks for the leads.  I had some major problems with the JIRA yesterday, was I 
alone in that?

[~shivaram] Where are the worker logs stored?

[~davies] I didn't see any of the RC binaries for spark in the EC2 in the s3 
bucket:
https://github.com/mesos/spark-ec2/blob/branch-1.4/spark/init.sh
http://s3.amazonaws.com/spark-related-packages/

Is there somewhere else they are stored?  

I recall seeing the RC branches in github, but that appears to have just 
dissapeared, and the branch-1.4 no longer has a RC.  Is the release happening 
today or some such? 

The first thing I did was [add the 1.3.1| 
https://github.com/3Scan/spark-ec2/commit/08d210dc8d44c07383e46fcd303c8f0c20828bcf]
 build that I found there, and try that.  That didn't feature any of the py3 
support, and crashes fast and early.  Next I had it build from a checkout of 
the current master, which is where the above errors come from. To get a source 
build to work, I had to make [this change | 
https://github.com/3Scan/spark-ec2/commit/3416dd07c492b0cddcc98c4fa83f9e4284ed8fc9],
 and at least one other . . . to be determined following sorting out the above 
trace.  

> spark-ec2 and associated tools not py3 ready
> 
>
> Key: SPARK-7909
> URL: https://issues.apache.org/jira/browse/SPARK-7909
> Project: Spark
>  Issue Type: Improvement
>  Components: EC2
> Environment: ec2 python3
>Reporter: Matthew Goodman
>
> At present there is not a possible permutation of tools that supports Python3 
> on both the launching computer and running cluster.  There are a couple 
> problems involved:
>  - There is no prebuilt spark binary with python3 support.
>  - spark-ec2/spark/init.sh contains inline py3 unfriendly print statements
>  - Config files for cluster processes don't seem to make it to all nodes in a 
> working format.
> I have fixes for some of this, but the config and running context debugging 
> remains elusive to me.  



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-7909) spark-ec2 and associated tools not py3 ready

2015-05-28 Thread Davies Liu (JIRA)

[ 
https://issues.apache.org/jira/browse/SPARK-7909?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14563352#comment-14563352
 ] 

Davies Liu commented on SPARK-7909:
---

[~meawoppl] It's true that some tools don't work with Python3, but PySpark 1.4 
should work with Python3 (it's tested with Python3 for every PR), you can test 
with the 1.4-RC2 package.

The loggings didn't say something wrong with Python, It said the executors are 
keep dying. Could you check the logging of executors?

> spark-ec2 and associated tools not py3 ready
> 
>
> Key: SPARK-7909
> URL: https://issues.apache.org/jira/browse/SPARK-7909
> Project: Spark
>  Issue Type: Improvement
>  Components: EC2
> Environment: ec2 python3
>Reporter: Matthew Goodman
>
> At present there is not a possible permutation of tools that supports Python3 
> on both the launching computer and running cluster.  There are a couple 
> problems involved:
>  - There is no prebuilt spark binary with python3 support.
>  - spark-ec2/spark/init.sh contains inline py3 unfriendly print statements
>  - Config files for cluster processes don't seem to make it to all nodes in a 
> working format.
> I have fixes for some of this, but the config and running context debugging 
> remains elusive to me.  



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-7909) spark-ec2 and associated tools not py3 ready

2015-05-27 Thread Shivaram Venkataraman (JIRA)

[ 
https://issues.apache.org/jira/browse/SPARK-7909?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14562161#comment-14562161
 ] 

Shivaram Venkataraman commented on SPARK-7909:
--

Could you check the worker logs to figure out why the executors were dying ? 
Also regarding the points described in the JIRA

1. Regd. pre-built Spark binary, I think Spark 1.4 will be Python3 friendly, 
but [~davies] might be able to confirm
2. I am not sure how print statements in shell script can interfere with 
Python3 (cc [~nchammas] who knows more than me about Python 3)

> spark-ec2 and associated tools not py3 ready
> 
>
> Key: SPARK-7909
> URL: https://issues.apache.org/jira/browse/SPARK-7909
> Project: Spark
>  Issue Type: Improvement
>  Components: EC2
> Environment: ec2 python3
>Reporter: Matthew Goodman
>
> At present there is not a possible permutation of tools that supports Python3 
> on both the launching computer and running cluster.  There are a couple 
> problems involved:
>  - There is no prebuilt spark binary with python3 support.
>  - spark-ec2/spark/init.sh contains inline py3 unfriendly print statements
>  - Config files for cluster processes don't seem to make it to all nodes in a 
> working format.
> I have fixes for some of this, but the config and running context debugging 
> remains elusive to me.  



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-7909) spark-ec2 and associated tools not py3 ready

2015-05-27 Thread Matthew Goodman (JIRA)

[ 
https://issues.apache.org/jira/browse/SPARK-7909?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14562121#comment-14562121
 ] 

Matthew Goodman commented on SPARK-7909:


I have almost everything working, but I am getting hung-up on getting the 
pyspark binary to launch things correctly.  When logging into ec2:

{code:title=PySpark Output Loop|borderStyle=solid}
root@ip-172-31-6-84 ~]$ ./spark/bin/pyspark
Python 3.4.3 |Continuum Analytics, Inc.| (default, Mar  6 2015, 12:03:53) 
[GCC 4.4.7 20120313 (Red Hat 4.4.7-1)] on linux
Type "help", "copyright", "credits" or "license" for more information.
Using Spark's default log4j profile: org/apache/spark/log4j-defaults.properties
15/05/28 00:58:28 INFO SparkContext: Running Spark version 1.4.0-SNAPSHOT
15/05/28 00:58:28 WARN NativeCodeLoader: Unable to load native-hadoop library 
for your platform... using builtin-java classes where applicable
15/05/28 00:58:28 INFO SecurityManager: Changing view acls to: root
15/05/28 00:58:28 INFO SecurityManager: Changing modify acls to: root
15/05/28 00:58:28 INFO SecurityManager: SecurityManager: authentication 
disabled; ui acls disabled; users with view permissions: Set(root); users with 
modify permissions: Set(root)
15/05/28 00:58:29 INFO Slf4jLogger: Slf4jLogger started
15/05/28 00:58:29 INFO Remoting: Starting remoting
15/05/28 00:58:29 INFO Remoting: Remoting started; listening on addresses 
:[akka.tcp://sparkDriver@172.31.6.84:59125]
15/05/28 00:58:30 INFO Utils: Successfully started service 'sparkDriver' on 
port 59125.
15/05/28 00:58:30 INFO SparkEnv: Registering MapOutputTracker
15/05/28 00:58:30 INFO SparkEnv: Registering BlockManagerMaster
15/05/28 00:58:30 INFO DiskBlockManager: Created local directory at 
/mnt/spark/spark-985d5a6c-150e-40ad-875f-351733a40276/blockmgr-e36c9174-ff48-42e1-bbd0-c2b0649ab751
15/05/28 00:58:30 INFO DiskBlockManager: Created local directory at 
/mnt2/spark/spark-fb2a7e42-2998-4ad2-be5f-d25472727d57/blockmgr-660e39f5-3561-4bec-a042-7cab1ea8cf54
15/05/28 00:58:30 INFO MemoryStore: MemoryStore started with capacity 265.4 MB
15/05/28 00:58:30 INFO HttpFileServer: HTTP File server directory is 
/mnt/spark/spark-985d5a6c-150e-40ad-875f-351733a40276/httpd-40a7ad26-25d9-482b-bc7f-68d9f126b32d
15/05/28 00:58:30 INFO HttpServer: Starting HTTP Server
15/05/28 00:58:30 INFO Server: jetty-8.y.z-SNAPSHOT
15/05/28 00:58:30 INFO AbstractConnector: Started SocketConnector@0.0.0.0:46564
15/05/28 00:58:30 INFO Utils: Successfully started service 'HTTP file server' 
on port 46564.
15/05/28 00:58:30 INFO SparkEnv: Registering OutputCommitCoordinator
15/05/28 00:58:30 INFO Server: jetty-8.y.z-SNAPSHOT
15/05/28 00:58:30 INFO AbstractConnector: Started 
SelectChannelConnector@0.0.0.0:4040
15/05/28 00:58:30 INFO Utils: Successfully started service 'SparkUI' on port 
4040.
15/05/28 00:58:30 INFO SparkUI: Started SparkUI at 
http://ec2-52-24-65-198.us-west-2.compute.amazonaws.com:4040
15/05/28 00:58:30 INFO AppClient$ClientActor: Connecting to master 
akka.tcp://sparkmas...@ec2-52-24-65-198.us-west-2.compute.amazonaws.com:7077/user/Master...
15/05/28 00:58:31 INFO SparkDeploySchedulerBackend: Connected to Spark cluster 
with app ID app-20150528005831-0005
15/05/28 00:58:31 INFO AppClient$ClientActor: Executor added: 
app-20150528005831-0005/0 on worker-20150527230803-172.31.13.150-50730 
(172.31.13.150:50730) with 2 cores
15/05/28 00:58:31 INFO SparkDeploySchedulerBackend: Granted executor ID 
app-20150528005831-0005/0 on hostPort 172.31.13.150:50730 with 2 cores, 6.0 GB 
RAM
15/05/28 00:58:31 INFO AppClient$ClientActor: Executor updated: 
app-20150528005831-0005/0 is now LOADING
15/05/28 00:58:31 INFO AppClient$ClientActor: Executor updated: 
app-20150528005831-0005/0 is now RUNNING
15/05/28 00:58:31 INFO Utils: Successfully started service 
'org.apache.spark.network.netty.NettyBlockTransferService' on port 34430.
15/05/28 00:58:31 INFO NettyBlockTransferService: Server created on 34430
15/05/28 00:58:31 INFO BlockManagerMaster: Trying to register BlockManager
15/05/28 00:58:31 INFO BlockManagerMasterEndpoint: Registering block manager 
172.31.6.84:34430 with 265.4 MB RAM, BlockManagerId(driver, 172.31.6.84, 34430)
15/05/28 00:58:31 INFO BlockManagerMaster: Registered BlockManager
15/05/28 00:58:31 INFO SparkDeploySchedulerBackend: SchedulerBackend is ready 
for scheduling beginning after reached minRegisteredResourcesRatio: 0.0
Welcome to
    __
 / __/__  ___ _/ /__
_\ \/ _ \/ _ `/ __/  '_/
   /__ / .__/\_,_/_/ /_/\_\   version 1.4.0-SNAPSHOT
  /_/

Using Python version 3.4.3 (default, Mar  6 2015 12:03:53)
SparkContext available as sc, SQLContext available as sqlContext.
>>> 15/05/28 00:58:32 INFO AppClient$ClientActor: Executor updated: 
>>> app-20150528005831-0005/0 is now EXITED (Command exited with code 1)
15/05/28 00:58:32 INFO SparkDeploySchedulerBackend: Executor 
app-2015052800