Re: issue Running Spark Job on Yarn Cluster

2015-03-04 Thread sachin Singh
-- If you reply to this email, your message will be added to the discussion below: http://apache-spark-user-list.1001560.n3.nabble.com/issue-Running-Spark-Job-on-Yarn-Cluster-tp21697p21909.html To unsubscribe from issue Running Spark Job on Yarn Cluster, click here http://apache-spark-user-list

Re: Spark Monitoring UI for Hadoop Yarn Cluster

2015-03-04 Thread Srini Karri
is, is it possible to achieve the same monitoring UI experience with Yarn Cluster like Viewing workers, running/completed job stages in the Web UI. Currently, if we go to our Yarn Resource manager UI, we are able to see the Spark Jobs and it's logs. But it is not as rich as Spark Standalone master UI

Re: Spark Monitoring UI for Hadoop Yarn Cluster

2015-03-04 Thread Srini Karri
Yes. I do see files, actually I missed copying the other settings: spark.master spark:// skarri-lt05.redmond.corp.microsoft.com:7077 spark.eventLog.enabled true spark.rdd.compress true spark.storage.memoryFraction 1 spark.core.connection.ack.wait.timeout 6000

Re: Spark Monitoring UI for Hadoop Yarn Cluster

2015-03-04 Thread Marcelo Vanzin
On Wed, Mar 4, 2015 at 10:08 AM, Srini Karri skarri@gmail.com wrote: spark.executor.extraClassPath D:\\Apache\\spark-1.2.1-bin-hadoop2\\spark-1.2.1-bin-hadoop2.4\\bin\\classes spark.eventLog.dir D:/Apache/spark-1.2.1-bin-hadoop2/spark-1.2.1-bin-hadoop2.4/bin/tmp/spark-events

Re: Spark Monitoring UI for Hadoop Yarn Cluster

2015-03-04 Thread Srini Karri
Hi Marcelo, I found the problem from http://mail-archives.apache.org/mod_mbox/spark-user/201409.mbox/%3cCAL+LEBfzzjugOoB2iFFdz_=9TQsH=DaiKY=cvydfydg3ac5...@mail.gmail.com%3e this link. The problem is the application I am running, is not generating APPLICATION_COMPLETE file. If I add this file

Re: issue Running Spark Job on Yarn Cluster

2015-03-04 Thread roni
-user-list.1001560.n3.nabble.com/issue-Running-Spark-Job-on-Yarn-Cluster-tp21697p21909.html To unsubscribe from issue Running Spark Job on Yarn Cluster, click here. NAML http://apache-spark-user-list.1001560.n3.nabble.com/template/NamlServlet.jtp?macro=macro_viewerid=instant_html%21nabble

Issue with yarn cluster - hangs in accepted state.

2015-03-03 Thread abhi
I am trying to run below java class with yarn cluster, but it hangs in accepted state . i don't see any error . Below is the class and command . Any help is appreciated . Thanks, Abhi bin/spark-submit --class com.mycompany.app.SimpleApp --master yarn-cluster /home/hduser/my-app-1.0.jar

Re: Spark Monitoring UI for Hadoop Yarn Cluster

2015-03-03 Thread Todd Nist
, I have tried Standalone Spark Installation on Windows, I am able to submit the logs, able to see the history of events. My question is, is it possible to achieve the same monitoring UI experience with Yarn Cluster like Viewing workers, running/completed job stages in the Web UI. Currently

Re: Spark Monitoring UI for Hadoop Yarn Cluster

2015-03-03 Thread Marcelo Vanzin
is the context, I have tried Standalone Spark Installation on Windows, I am able to submit the logs, able to see the history of events. My question is, is it possible to achieve the same monitoring UI experience with Yarn Cluster like Viewing workers, running/completed job stages in the Web UI. Currently

Re: Issue with yarn cluster - hangs in accepted state.

2015-03-03 Thread Zhan Zhang
Do you have enough resource in your cluster? You can check your resource manager to see the usage. Thanks. Zhan Zhang On Mar 3, 2015, at 8:51 AM, abhi abhishek...@gmail.commailto:abhishek...@gmail.com wrote: I am trying to run below java class with yarn cluster, but it hangs in accepted

Re: Issue with yarn cluster - hangs in accepted state.

2015-03-03 Thread Tobias Pfeiffer
Hi, On Wed, Mar 4, 2015 at 6:20 AM, Zhan Zhang zzh...@hortonworks.com wrote: Do you have enough resource in your cluster? You can check your resource manager to see the usage. Yep, I can confirm that this is a very annoying issue. If there is not enough memory or VCPUs available, your app

Spark Monitoring UI for Hadoop Yarn Cluster

2015-03-03 Thread Srini Karri
Cluster like Viewing workers, running/completed job stages in the Web UI. Currently, if we go to our Yarn Resource manager UI, we are able to see the Spark Jobs and it's logs. But it is not as rich as Spark Standalone master UI. Is this limitation for hadoop yarn cluster or is there any way we can

What is best way to run spark job in yarn-cluster mode from java program(servlet container) and NOT using spark-submit command.

2015-02-25 Thread kshekhram
a situation where I am running web application in Jetty using Spring boot.My web application receives a REST web service request based on that It needs to trigger spark calculation job in Yarn cluster. Since my job can take longer to run and can access data from HDFS, so I want to run the spark job in yarn

Re: issue Running Spark Job on Yarn Cluster

2015-02-19 Thread Harshvardhan Chauhan
Is this the full stack trace ? On Wed, Feb 18, 2015 at 2:39 AM, sachin Singh sachin.sha...@gmail.com wrote: Hi, I want to run my spark Job in Hadoop yarn Cluster mode, I am using below command - spark-submit --master yarn-cluster --driver-memory 1g --executor-memory 1g --executor-cores 1

Re: issue Running Spark Job on Yarn Cluster

2015-02-19 Thread Sachin Singh
Yes. On 19 Feb 2015 23:40, Harshvardhan Chauhan ha...@gumgum.com wrote: Is this the full stack trace ? On Wed, Feb 18, 2015 at 2:39 AM, sachin Singh sachin.sha...@gmail.com wrote: Hi, I want to run my spark Job in Hadoop yarn Cluster mode, I am using below command - spark-submit --master

Re: issue Running Spark Job on Yarn Cluster

2015-02-19 Thread Marcelo Vanzin
You'll need to look at your application's logs. You can use yarn logs --applicationId [id] to see them. On Wed, Feb 18, 2015 at 2:39 AM, sachin Singh sachin.sha...@gmail.com wrote: Hi, I want to run my spark Job in Hadoop yarn Cluster mode, I am using below command - spark-submit --master

issue Running Spark Job on Yarn Cluster

2015-02-18 Thread sachin Singh
Hi, I want to run my spark Job in Hadoop yarn Cluster mode, I am using below command - spark-submit --master yarn-cluster --driver-memory 1g --executor-memory 1g --executor-cores 1 --class com.dc.analysis.jobs.AggregationJob sparkanalitic.jar param1 param2 param3 I am getting error as under

RE: Spark sql failed in yarn-cluster mode when connecting to non-default hive database

2015-02-17 Thread Andrew Lee
started in yarn-cluster mode. ./bin/spark-submit --verbose --queue research --driver-java-options -XX:MaxPermSize=8192M --files /etc/hive/hive-site.xml --driver-class-path /etc/hive/hive-site.xml --master yarn --deploy-mode cluster The problem here is that --files only look for the local

Re: How to broadcast a variable read from a file in yarn-cluster mode?

2015-02-12 Thread Tathagata Das
yarn-client, even with both a SparkContext and StreamingContext. It looks to me that in yarn-cluster mode it's grabbing resources for the StreamingContext but not for the SparkContext. Any ideas? Jon 15/02/10 12:06:16 INFO MemoryStore: MemoryStore started with capacity 1177.8 MB. 15/02/10

Re: Resource allocation in yarn-cluster mode

2015-02-10 Thread Sandy Ryza
toth.zsolt@gmail.com: Hi, I'm using Spark in yarn-cluster mode and submit the jobs programmatically from the client in Java. I ran into a few issues when tried to set the resource allocation properties. 1. It looks like setting spark.executor.memory, spark.executor.cores

Re: How to broadcast a variable read from a file in yarn-cluster mode?

2015-02-10 Thread Sandy Ryza
(sbStreamingTv) does work successfully using yarn-client, even with both a SparkContext and StreamingContext. It looks to me that in yarn-cluster mode it's grabbing resources for the StreamingContext but not for the SparkContext. Any ideas? Jon 15/02/10 12:06:16 INFO MemoryStore

Re: Resource allocation in yarn-cluster mode

2015-02-10 Thread Zsolt Tóth
One more question: Is there reason why Spark throws an error when requesting too much memory instead of capping it to the maximum value (as YARN would do by default)? Thanks! 2015-02-10 17:32 GMT+01:00 Zsolt Tóth toth.zsolt@gmail.com: Hi, I'm using Spark in yarn-cluster mode and submit

Re: How to broadcast a variable read from a file in yarn-cluster mode?

2015-02-10 Thread Jon Gregg
Spark doesn't get beyond that point in the code.* Also, this job (sbStreamingTv) does work successfully using yarn-client, even with both a SparkContext and StreamingContext. It looks to me that in yarn-cluster mode it's grabbing resources for the StreamingContext but not for the SparkContext. Any

Resource allocation in yarn-cluster mode

2015-02-10 Thread Zsolt Tóth
Hi, I'm using Spark in yarn-cluster mode and submit the jobs programmatically from the client in Java. I ran into a few issues when tried to set the resource allocation properties. 1. It looks like setting spark.executor.memory, spark.executor.cores and spark.executor.instances have no effect

Re: How to broadcast a variable read from a file in yarn-cluster mode?

2015-02-10 Thread Jon Gregg
(sbStreamingTv) does work successfully using yarn-client, even with both a SparkContext and StreamingContext. It looks to me that in yarn-cluster mode it's grabbing resources for the StreamingContext but not for the SparkContext. Any ideas? Jon 15/02/10 12:06:16 INFO MemoryStore

Re: How to broadcast a variable read from a file in yarn-cluster mode?

2015-02-10 Thread Sandy Ryza
it: val badIPs = fromFile(edgeDir + badfullIPs.csv) val badIPsLines = badIPs.getLines val badIpSet = badIPsLines.toSet val badIPsBC = sc.broadcast(badIpSet) badIPs.close How can I accomplish this in yarn-cluster mode? Jon -- View this message in context

Re: How to broadcast a variable read from a file in yarn-cluster mode?

2015-02-10 Thread Jon Gregg
that in yarn-cluster mode it's grabbing resources for the StreamingContext but not for the SparkContext. Any ideas? Jon 15/02/10 12:06:16 INFO MemoryStore: MemoryStore started with capacity 1177.8 MB. 15/02/10 12:06:16 INFO ConnectionManager: Bound socket to port 30129 with id

Re: How to broadcast a variable read from a file in yarn-cluster mode?

2015-02-06 Thread Jon Gregg
it, and then broadcast it: val badIPs = fromFile(edgeDir + badfullIPs.csv) val badIPsLines = badIPs.getLines val badIpSet = badIPsLines.toSet val badIPsBC = sc.broadcast(badIpSet) badIPs.close How can I accomplish this in yarn-cluster mode? Jon -- View this message

Re: How to broadcast a variable read from a file in yarn-cluster mode?

2015-02-06 Thread Sandy Ryza
= badIPs.getLines val badIpSet = badIPsLines.toSet val badIPsBC = sc.broadcast(badIpSet) badIPs.close How can I accomplish this in yarn-cluster mode? Jon -- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/How-to-broadcast-a-variable-read-from

How to broadcast a variable read from a file in yarn-cluster mode?

2015-02-05 Thread YaoPau
val badIPsBC = sc.broadcast(badIpSet) badIPs.close How can I accomplish this in yarn-cluster mode? Jon -- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/How-to-broadcast-a-variable-read-from-a-file-in-yarn-cluster-mode-tp21524.html Sent from

Re: Spark sql failed in yarn-cluster mode when connecting to non-default hive database

2015-02-05 Thread Cheng Lian
. through hive cli, I don't see any problem. but for spark on yarn-cluster mode, I am not able to switch to a database other than the default one, for Yarn-client mode, it works fine. Thanks! Jenny On Tue, Aug 12, 2014 at 12:53 PM, Yin Huai huaiyin@gmail.com mailto:huaiyin@gmail.com wrote

Re: How to broadcast a variable read from a file in yarn-cluster mode?

2015-02-05 Thread Sandy Ryza
this in yarn-cluster mode? Jon -- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/How-to-broadcast-a-variable-read-from-a-file-in-yarn-cluster-mode-tp21524.html Sent from the Apache Spark User List mailing list archive at Nabble.com

Re: Spark Job running on localhost on yarn cluster

2015-02-05 Thread kundan kumar
to the standalone mode. However, as compared to the standalone mode, spark on yarn runs very slow. I am running it as $SPARK_HOME/bin/spark-submit --class EDDApp --master yarn-cluster --num-executors 10 --executor-memory 14g target/scala-2.10/edd-application_2.10-1.0.jar hdfs://hm41:9000/user/hduser

Re: Spark Job running on localhost on yarn cluster

2015-02-05 Thread Kostas Sakellis
with more memory? For your setup the memory calculation is:1 executorMemoryGB * 1.07 = 14GB = 14GB/1.07 ~ 13GB. Your command args should be something like: --master yarn-cluster --num-executors 5 --num-executor-cores 7 --executor-memory 13g As for the UI, where did you see 7.2GB? can you send a screen

Re: Spark Job running on localhost on yarn cluster

2015-02-04 Thread Michael Albert
1) Parameters like --num-executors should come before the jar.  That is, you want something like$SPARK_HOME --num-executors 3 --driver-memory 6g --executor-memory 7g \--master yarn-cluster  --class EDDApp target/scala-2.10/eddjar \outputPath That is, *your* parameters come after the jar

Re: Spark Job running on localhost on yarn cluster

2015-02-04 Thread Felix C
Is YARN_CONF_DIR set? --- Original Message --- From: Aniket Bhatnagar aniket.bhatna...@gmail.com Sent: February 4, 2015 6:16 AM To: kundan kumar iitr.kun...@gmail.com, spark users user@spark.apache.org Subject: Re: Spark Job running on localhost on yarn cluster Have you set master in SparkConf

Spark Job running on localhost on yarn cluster

2015-02-04 Thread kundan kumar
Hi, I am trying to execute my code on a yarn cluster The command which I am using is $SPARK_HOME/bin/spark-submit --class EDDApp target/scala-2.10/edd-application_2.10-1.0.jar --master yarn-cluster --num-executors 3 --driver-memory 6g --executor-memory 7g outpuPath But, I can see

[spark-jobserver] Submit Job in yarn-cluster mode (?)

2015-01-14 Thread Pietro Gentile
Hi all, I'm able to submit spark jobs through spark-jobserver. But this allows to use spark only in yarn-client mode. I want to use spark also in yarn-cluster mode but jobserver does not allow it, like says in the README file https://github.com/spark-jobserver/spark-jobserver. Could you tell

Problem getting Spark running on a Yarn cluster

2015-01-06 Thread Sharon Rapoport
Hello, We have hadoop 2.6.0 and Yarn set up on ec2. Trying to get spark 1.1.1 running on the Yarn cluster. I have of course googled around and found that this problem is solved for most after removing the line including 127.0.1.1 from /etc/hosts. This hasn’t seemed to solve this for me

Re: Problem getting Spark running on a Yarn cluster

2015-01-06 Thread Akhil Das
to get spark 1.1.1 running on the Yarn cluster. I have of course googled around and found that this problem is solved for most after removing the line including 127.0.1.1 from /etc/hosts. This hasn’t seemed to solve this for me. Anyone has an idea where else might 127.0.1.1 be hiding in some conf

RE: Spark sql failed in yarn-cluster mode when connecting to non-default hive database

2014-12-29 Thread Andrew Lee
(javax.jdo.option.ConnectionPassword,xxx) hiveContext.setConf(javax.jdo.option.ConnectionURL,jdbc:mysql://localhost:3306/hive?createDatabaseIfNotExist=true) From: huaiyin@gmail.com Date: Wed, 13 Aug 2014 16:56:13 -0400 Subject: Re: Spark sql failed in yarn-cluster mode when

RE: Spark sql failed in yarn-cluster mode when connecting to non-default hive database

2014-12-29 Thread Andrew Lee
with -Phadoop-provided, and other common libraries that are required. From: alee...@hotmail.com To: user@spark.apache.org CC: lian.cs@gmail.com; linlin200...@gmail.com; huaiyin@gmail.com Subject: RE: Spark sql failed in yarn-cluster mode when connecting to non-default hive database Date

Spark UI port issue when deploying Spark driver on YARN in yarn-cluster mode on EMR

2014-12-23 Thread Roberto Coluccio
Hello folks, I'm trying to deploy a Spark driver on Amazon EMR in yarn-cluster mode expecting to be able to access the Spark UI from the spark-master-ip:4040 address (default port). The problem here is that the Spark UI port is always defined randomly at runtime, although I also tried to specify

Re: Can we specify driver running on a specific machine of the cluster on yarn-cluster mode?

2014-12-19 Thread Sean Owen
That's not true in yarn-cluster mode, where the driver runs in a container that YARN creates, which may not be on the machine that runs spark-submit. As far as I know, however, you can't control where YARN allocates that, and shouldn't need to. You can probably query YARN to find where it did

Can we specify driver running on a specific machine of the cluster on yarn-cluster mode?

2014-12-18 Thread LinQili
Hi all,On yarn-cluster mode, can we let the driver running on a specific machine that we choose in cluster ? Or, even the machine not in the cluster?

create table in yarn-cluster mode vs yarn-client mode

2014-12-18 Thread Chirag Aggarwal
Hi, I have a simple app, where I am trying to create a table. I am able to create the table on running app in yarn-client mode, but not with yarn-cluster mode. Is this some known issue? Has this already been fixed? Please note that I am using spark-1.1 over hadoop-2.4.0 App: - import

Re: custom spark app name in yarn-cluster mode

2014-12-15 Thread Tomer Benyamini
Thanks Sandy, passing --name works fine :) Tomer On Fri, Dec 12, 2014 at 9:35 AM, Sandy Ryza sandy.r...@cloudera.com wrote: Hi Tomer, In yarn-cluster mode, the application has already been submitted to YARN by the time the SparkContext is created, so it's too late to set the app name

HTTP 500 Error for SparkUI in YARN Cluster mode

2014-12-14 Thread Benyi Wang
I got this error when I click Track URL: ApplicationMaster when I run a spark job in YARN cluster mode. I found this jira https://issues.apache.org/jira/browse/YARN-800, but I could not get this problem fixed. I'm running CDH 5.1.0 with Both HDFS and RM HA enabled. Does anybody has the similar

custom spark app name in yarn-cluster mode

2014-12-11 Thread Tomer Benyamini
Hi, I'm trying to set a custom spark app name when running a java spark app in yarn-cluster mode. SparkConf sparkConf = new SparkConf(); sparkConf.setMaster(System.getProperty(spark.master)); sparkConf.setAppName(myCustomName); sparkConf.set(spark.logConf, true); JavaSparkContext sc

Re: custom spark app name in yarn-cluster mode

2014-12-11 Thread Tomer Benyamini
On Thu, Dec 11, 2014 at 8:27 PM, Tomer Benyamini tomer@gmail.com wrote: Hi, I'm trying to set a custom spark app name when running a java spark app in yarn-cluster mode. SparkConf sparkConf = new SparkConf(); sparkConf.setMaster(System.getProperty(spark.master

Re: custom spark app name in yarn-cluster mode

2014-12-11 Thread Sandy Ryza
Hi Tomer, In yarn-cluster mode, the application has already been submitted to YARN by the time the SparkContext is created, so it's too late to set the app name there. I believe giving it with the --name property to spark-submit should work. -Sandy On Thu, Dec 11, 2014 at 10:28 AM, Tomer

Issue on [SPARK-3877][YARN]: Return code of the spark-submit in yarn-cluster mode

2014-12-05 Thread LinQili
Hi, all: According to https://github.com/apache/spark/pull/2732, When a spark job fails or exits nonzero in yarn-cluster mode, the spark-submit will get the corresponding return code of the spark job. But I tried in spark-1.1.1 yarn cluster, spark-submit return zero anyway. Here is my spark

RE: Issue on [SPARK-3877][YARN]: Return code of the spark-submit in yarn-cluster mode

2014-12-05 Thread LinQili
I tried in spark client mode, spark-submit can get the correct return code from spark job. But in yarn-cluster mode, It failed. From: lin_q...@outlook.com To: u...@spark.incubator.apache.org Subject: Issue on [SPARK-3877][YARN]: Return code of the spark-submit in yarn-cluster mode Date: Fri, 5

RE: Issue on [SPARK-3877][YARN]: Return code of the spark-submit in yarn-cluster mode

2014-12-05 Thread LinQili
-submit cannot get the second return code 100. What's the difference between these two `exit`? I was so confused. From: lin_q...@outlook.com To: u...@spark.incubator.apache.org Subject: RE: Issue on [SPARK-3877][YARN]: Return code of the spark-submit in yarn-cluster mode Date: Fri, 5 Dec 2014 17

Re: Issue on [SPARK-3877][YARN]: Return code of the spark-submit in yarn-cluster mode

2014-12-05 Thread Shixiong Zhu
. -- From: lin_q...@outlook.com To: u...@spark.incubator.apache.org Subject: RE: Issue on [SPARK-3877][YARN]: Return code of the spark-submit in yarn-cluster mode Date: Fri, 5 Dec 2014 17:11:39 +0800 I tried in spark client mode, spark-submit can get the correct return code

Re: Issue on [SPARK-3877][YARN]: Return code of the spark-submit in yarn-cluster mode

2014-12-05 Thread Shixiong Zhu
, spark-submit returned 1 for both two cases. That’s expected. In the yarn-cluster mode, the driver runs in the ApplicationMaster. The exit code of driver is also the exit code of ApplicationMaster. However, for now, Spark cannot get the exit code of ApplicationMaster from Yarn, because Yarn does

Spark yarn cluster Application Master not running yarn container

2014-11-25 Thread firemonk9
I am running a 3 node(32 core, 60gb) Yarn cluster for Spark jobs. 1) Below are my Yarn memory settings yarn.nodemanager.resource.memory-mb = 52224 yarn.scheduler.minimum-allocation-mb = 40960 yarn.scheduler.maximum-allocation-mb = 52224 Apache Spark Memory Settings export SPARK_EXECUTOR_MEMORY

Submit Spark driver on Yarn Cluster in client mode

2014-11-24 Thread Naveen Kumar Pokala
Hi, I want to submit my spark program from my machine on a YARN Cluster in yarn client mode. How to specify al l the required details through SPARK submitter. Please provide me some details. -Naveen.

Re: Submit Spark driver on Yarn Cluster in client mode

2014-11-24 Thread Akhil Das
You can export the hadoop configurations dir (export HADOOP_CONF_DIR=XXX) in the environment and then submit it like: ./bin/spark-submit \ --class org.apache.spark.examples.SparkPi \ --master yarn-cluster \ # can also be `yarn-client` for client mode --executor-memory 20G \ --num

Re: Submit Spark driver on Yarn Cluster in client mode

2014-11-24 Thread Naveen Kumar Pokala
, 2014 4:08 PM To: Naveen Kumar Pokala Cc: user@spark.apache.orgmailto:user@spark.apache.org Subject: Re: Submit Spark driver on Yarn Cluster in client mode You can export the hadoop configurations dir (export HADOOP_CONF_DIR=XXX) in the environment and then submit it like: ./bin/spark-submit

Re: Submit Spark driver on Yarn Cluster in client mode

2014-11-24 Thread Akhil Das
:* Akhil Das [mailto:ak...@sigmoidanalytics.com ak...@sigmoidanalytics.com] *Sent:* Monday, November 24, 2014 4:08 PM *To:* Naveen Kumar Pokala *Cc:* user@spark.apache.org *Subject:* Re: Submit Spark driver on Yarn Cluster in client mode You can export the hadoop configurations dir (export

How to get applicationId for yarn mode(both yarn-client and yarn-cluster mode)

2014-11-21 Thread Earthson
Is there any way to get the yarn application_id inside the program? -- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/How-to-get-applicationId-for-yarn-mode-both-yarn-client-and-yarn-cluster-mode-tp19462.html Sent from the Apache Spark User List mailing

Re: How to get applicationId for yarn mode(both yarn-client and yarn-cluster mode)

2014-11-21 Thread Earthson
-applicationId-for-yarn-mode-both-yarn-client-and-yarn-cluster-mode-tp19462p19466.html Sent from the Apache Spark User List mailing list archive at Nabble.com. - To unsubscribe, e-mail: user-unsubscr...@spark.apache.org For additional

Building a hash table from a csv file using yarn-cluster, and giving it to each executor

2014-11-13 Thread YaoPau
) val ipLines = badIpSource.getLines() val set = new HashSet[String]() val badIpSet = set ++ ipLines badIpSource.close() def isGoodIp(ip: String): Boolean = !badIpSet.contains(ip) But when I try this using --master yarn-cluster I get Exception in thread Thread-4

Re: Building a hash table from a csv file using yarn-cluster, and giving it to each executor

2014-11-13 Thread aappddeevv
If the file is not present on each node, it may not find it. -- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/Building-a-hash-table-from-a-csv-file-using-yarn-cluster-and-giving-it-to-each-executor-tp18850p18877.html Sent from the Apache Spark User List

Re: Spark 1.0.0 on yarn cluster problem

2014-10-23 Thread firemonk9
-- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/Spark-1-0-0-on-yarn-cluster-problem-tp7560p17175.html Sent from the Apache Spark User List mailing list archive at Nabble.com. - To unsubscribe, e

Re: Spark 1.0.0 on yarn cluster problem

2014-10-23 Thread Andrew Or
. Thank you -- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/Spark-1-0-0-on-yarn-cluster-problem-tp7560p17175.html Sent from the Apache Spark User List mailing list archive at Nabble.com

Re: how to set log level of spark executor on YARN(using yarn-cluster mode)

2014-10-20 Thread eric wong
copy of log4j.properties like:* *log4j.rootCategory=DEBUG, console* *2. upload when using spark-submit script:* *./bin/spark-submit --class edu.bjut.spark.SparkPageRank --master yarn-cluster --num-executors 5 --executor-memory 2g --executor-cores 1 /data/hadoopspark/MySparkTest.jar

Re: Application failure in yarn-cluster mode

2014-10-16 Thread Christophe Préaud
anybody confirm if it is a bug, or a (configuration?) problem on my side? Thanks, Christophe. On 10/10/2014 18:24, Christophe Préaud wrote: Hi, After updating from spark-1.0.0 to spark-1.1.0, my spark applications failed most of the time (but not always) in yarn-cluster mode (but not in yarn-client

Re: system.out.println with --master yarn-cluster

2014-10-15 Thread vishnu86
. Regards, Vishnu -- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/system-out-println-with-master-yarn-cluster-tp16370p16473.html Sent from the Apache Spark User List mailing list archive at Nabble.com

how to set log level of spark executor on YARN(using yarn-cluster mode)

2014-10-15 Thread eric wong
Hi, I want to check the DEBUG log of spark executor on YARN(using yarn-cluster mode), but 1. yarn daemonlog setlevel DEBUG YarnChild.class 2. set log4j.properties in spark/conf folder on client node. no means above works. So how could i set the log level of spark executor* on YARN container

Re: how to set log level of spark executor on YARN(using yarn-cluster mode)

2014-10-15 Thread Marcelo Vanzin
, Oct 15, 2014 at 5:58 PM, eric wong win19...@gmail.com wrote: Hi, I want to check the DEBUG log of spark executor on YARN(using yarn-cluster mode), but 1. yarn daemonlog setlevel DEBUG YarnChild.class 2. set log4j.properties in spark/conf folder on client node. no means above works. So

Re: Application failure in yarn-cluster mode

2014-10-14 Thread Christophe Préaud
spark applications failed most of the time (but not always) in yarn-cluster mode (but not in yarn-client mode). Here is my configuration: * spark-1.1.0 * hadoop-2.2.0 And the hadoop.tmp.dir definition in the hadoop core-site.xml file (each directory is on its own partition, on different

Issue on running spark application in Yarn-cluster mode

2014-10-13 Thread vishnu86
When I execute the following in yarn-client mode its working fine and giving the result properly, but when i try to run in Yarn-cluster mode i am getting error spark-submit --class org.apache.spark.examples.SparkPi --master yarn-client /home/rug885/spark/examples/lib/spark-examples_2.10-1.0.0

Application failure in yarn-cluster mode

2014-10-10 Thread Christophe Préaud
Hi, After updating from spark-1.0.0 to spark-1.1.0, my spark applications failed most of the time (but not always) in yarn-cluster mode (but not in yarn-client mode). Here is my configuration: * spark-1.1.0 * hadoop-2.2.0 And the hadoop.tmp.dir definition in the hadoop core-site.xml

Re: executors not created yarn-cluster mode

2014-10-08 Thread Andrew Or
anything, -Andrew 2014-10-08 12:00 GMT-07:00 jamborta jambo...@gmail.com: Hi all, I have a setup that works fine in yarn-client mode, but when I change that to yarn-cluster, the executors don't get created, apart from the driver (it seems that it does not even appear in yarn's resource manager

Re: executors not created yarn-cluster mode

2014-10-08 Thread jamborta
that is the executors cannot connect back to the driver (in my case I am not sure if they are even started). I could not find a way to debug, as the log files don't have any error in them. thanks -- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/executors-not-created-yarn

Using addFile with pipe on a yarn cluster

2014-09-29 Thread esamanas
Hi, I've been using pyspark with my YARN cluster with success. The work I'm doing involves using the RDD's pipe command to send data through a binary I've made. I can do this easily in pyspark like so (assuming 'sc' is already defined): sc.addFile(./dumb_prog) t= sc.parallelize(range(10

SPARK 1.1.0 on yarn-cluster and external JARs

2014-09-25 Thread rzykov
(MongoDB, algebird and so on)? Thanks in advance -- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/SPARK-1-1-0-on-yarn-cluster-and-external-JARs-tp15136.html Sent from the Apache Spark User List mailing list archive at Nabble.com

Re: SPARK 1.1.0 on yarn-cluster and external JARs

2014-09-25 Thread Egor Pahomov
spark jar in a HDFS folder and set up the variable SPARK_JAR. What is the best way to do that for other external jars (MongoDB, algebird and so on)? Thanks in advance -- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/SPARK-1-1-0-on-yarn-cluster

Re: SPARK 1.1.0 on yarn-cluster and external JARs

2014-09-25 Thread Marcelo Vanzin
.nabble.com/SPARK-1-1-0-on-yarn-cluster-and-external-JARs-tp15136.html Sent from the Apache Spark User List mailing list archive at Nabble.com. - To unsubscribe, e-mail: user-unsubscr...@spark.apache.org For additional commands

Error launching spark application from Windows to Linux YARN Cluster - Could not find or load main class org.apache.spark.deploy.yarn.ExecutorLauncher

2014-09-23 Thread dxrodri
: http://apache-spark-user-list.1001560.n3.nabble.com/Error-launching-spark-application-from-Windows-to-Linux-YARN-Cluster-Could-not-find-or-load-main-clar-tp14888.html Sent from the Apache Spark User List mailing list archive at Nabble.com

Re: FetchFailed when collect at YARN cluster

2014-08-23 Thread Victor Tso-Guillen
lines of short strings (about 10 characters each line) from a YARN cluster with 400 nodes: *14/08/22 11:43:27 WARN scheduler.TaskSetManager: Lost task 205.0 in stage 0.0 (TID 1228, aaa.xxx.com): FetchFailed(BlockManagerId(220, aaa.xxx.com, 37899, 0), shuffleId=0, mapId=420, reduceId=205) 14/08/22

Re: FetchFailed when collect at YARN cluster

2014-08-23 Thread Jiayu Zhou
I saw your post. What are the operations you did? Are you trying to collect data from driver? Did you try the akka configurations? -- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/FetchFailed-when-collect-at-YARN-cluster-tp12670p12703.html Sent from

Re: FetchFailed when collect at YARN cluster

2014-08-23 Thread Victor Tso-Guillen
are the operations you did? Are you trying to collect data from driver? Did you try the akka configurations? -- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/FetchFailed-when-collect-at-YARN-cluster-tp12670p12703.html Sent from the Apache Spark User List mailing list

FetchFailed when collect at YARN cluster

2014-08-22 Thread Jiayu Zhou
Hi, I am having this FetchFailed issue when the driver is about to collect about 2.5M lines of short strings (about 10 characters each line) from a YARN cluster with 400 nodes: *14/08/22 11:43:27 WARN scheduler.TaskSetManager: Lost task 205.0 in stage 0.0 (TID 1228, aaa.xxx.com): FetchFailed

Re: spark on yarn cluster can't launch

2014-08-16 Thread Sandy Ryza
) at org.apache.spark.Logging$class.logInfo(Logging.scala:58) However, when I removed --deploy-mode cluster \ Exception disappear. I think with the deploy-mode cluster is running in yarn cluster mode, if not, the default will be run in yarn client mode. But why did yarn cluster get Exception? Thanks

Re: spark on yarn cluster can't launch

2014-08-16 Thread Sandy Ryza
--deploy-mode cluster \ Exception disappear. I think with the deploy-mode cluster is running in yarn cluster mode, if not, the default will be run in yarn client mode. But why did yarn cluster get Exception? Thanks -- cente...@gmail.com|齐忠

spark on yarn cluster can't launch

2014-08-15 Thread centerqi hu
) However, when I removed --deploy-mode cluster \ Exception disappear. I think with the deploy-mode cluster is running in yarn cluster mode, if not, the default will be run in yarn client mode. But why did yarn cluster get Exception? Thanks -- cente...@gmail.com|齐忠

Re: spark on yarn cluster can't launch

2014-08-15 Thread Andrew Or
$class.logInfo(Logging.scala:58) However, when I removed --deploy-mode cluster \ Exception disappear. I think with the deploy-mode cluster is running in yarn cluster mode, if not, the default will be run in yarn client mode. But why did yarn cluster get Exception? Thanks -- cente

Re: Spark sql failed in yarn-cluster mode when connecting to non-default hive database

2014-08-13 Thread Yin Huai
I think the problem is that when you are using yarn-cluster mode, because the Spark driver runs inside the application master, the hive-conf is not accessible by the driver. Can you try to set those confs by using hiveContext.set(...)? Or, maybe you can copy hive-site.xml to spark/conf in the node

Re: Spark sql failed in yarn-cluster mode when connecting to non-default hive database

2014-08-12 Thread Yin Huai
...@gmail.com wrote: you can reproduce this issue with the following steps (assuming you have Yarn cluster + Hive 12): 1) using hive shell, create a database, e.g: create database ttt 2) write a simple spark sql program import org.apache.spark.{SparkConf, SparkContext} import

Re: Spark sql failed in yarn-cluster mode when connecting to non-default hive database

2014-08-12 Thread Jenny Zhao
Hi Yin, hive-site.xml was copied to spark/conf and the same as the one under $HIVE_HOME/conf. through hive cli, I don't see any problem. but for spark on yarn-cluster mode, I am not able to switch to a database other than the default one, for Yarn-client mode, it works fine. Thanks! Jenny

Re: Spark sql failed in yarn-cluster mode when connecting to non-default hive database

2014-08-11 Thread Cheng Lian
Since you were using hql(...), it’s probably not related to JDBC driver. But I failed to reproduce this issue locally with a single node pseudo distributed YARN cluster. Would you mind to elaborate more about steps to reproduce this bug? Thanks ​ On Sun, Aug 10, 2014 at 9:36 PM, Cheng Lian

Re: Spark sql failed in yarn-cluster mode when connecting to non-default hive database

2014-08-11 Thread Yin Huai
reproduce this issue with the following steps (assuming you have Yarn cluster + Hive 12): 1) using hive shell, create a database, e.g: create database ttt 2) write a simple spark sql program import org.apache.spark.{SparkConf, SparkContext} import org.apache.spark.sql._ import

Re: Spark sql failed in yarn-cluster mode when connecting to non-default hive database

2014-08-11 Thread Jenny Zhao
can reproduce this issue with the following steps (assuming you have Yarn cluster + Hive 12): 1) using hive shell, create a database, e.g: create database ttt 2) write a simple spark sql program import org.apache.spark.{SparkConf, SparkContext} import org.apache.spark.sql._ import

Re: Spark sql failed in yarn-cluster mode when connecting to non-default hive database

2014-08-10 Thread Cheng Lian
Hi Jenny, does this issue only happen when running Spark SQL with YARN in your environment? On Sat, Aug 9, 2014 at 3:56 AM, Jenny Zhao linlin200...@gmail.com wrote: Hi, I am able to run my hql query on yarn cluster mode when connecting to the default hive metastore defined in hive-site.xml

Spark sql failed in yarn-cluster mode when connecting to non-default hive database

2014-08-08 Thread Jenny Zhao
Hi, I am able to run my hql query on yarn cluster mode when connecting to the default hive metastore defined in hive-site.xml. however, if I want to switch to a different database, like: hql(use other-database) it only works in yarn client mode, but failed on yarn-cluster mode

Re: How to specify the job to run on the specific nodes(machines) in the hadoop yarn cluster?

2014-07-30 Thread Haiyang Fu
It's really a good question !I'm also working on it On Wed, Jul 30, 2014 at 11:45 AM, adu dujinh...@hzduozhun.com wrote: Hi all, RT. I want to run a job on specific two nodes in the cluster? How to configure the yarn? Dose yarn queue help? Thanks

How to specify the job to run on the specific nodes(machines) in the hadoop yarn cluster?

2014-07-29 Thread adu
Hi all, RT. I want to run a job on specific two nodes in the cluster? How to configure the yarn? Dose yarn queue help? Thanks

<    1   2   3   4   5   >