spark with label nodes in yarn

2015-12-15 Thread 张志强(旺轩)
Hi all, Has anyone tried label based scheduling via spark on yarn? I've tried that, it didn't work, spark 1.4.1 + apache hadoop 2.6.0 Any feedbacks are welcome. Thanks Allen

Re: spark with label nodes in yarn

2015-12-15 Thread Ted Yu
SPARK-6470 was integrated to 1.5.0 release. Please use 1.5.0 or newer release. SPARK-7173 adds support for setting "spark.yarn.am.nodeLabelExpression" Cheers On Tue, Dec 15, 2015 at 1:55 AM, 张志强(旺轩) wrote: > Hi Ted,

Re: spark with label nodes in yarn

2015-12-15 Thread Saisai Shao
SPARK-6470 only supports node label expression for executors. SPARK-7173 supports node label expression for AM (will be in 1.6). If you want to schedule your whole application through label expression, you have to configure both am and executor label expression. If you only want to schedule

Re: [VOTE] Release Apache Spark 1.6.0 (RC2)

2015-12-15 Thread Michael Armbrust
This vote is canceled due to the issue with the incorrect version. This issue will be fixed by https://github.com/apache/spark/pull/10317 We can wait a little bit for a fix to https://issues.apache.org/jira/browse/SPARK-12345. However if it looks like there is not an easy fix coming soon, I'm

Re: BIRCH clustering algorithm

2015-12-15 Thread Joseph Bradley
Hi Dzeno, I'm not familiar with the algorithm myself, but if you have an important use case for it, you could open a JIRA to discuss it. However, if it is a less common algorithm, I'd recommend first submitting it as a Spark package (but publicizing the package on the user list). If it gains

java.lang.NoSuchMethodError while saving a random forest model Spark version 1.5

2015-12-15 Thread Rachana Srivastava
I have recently upgraded spark version but when I try to run save a random forest model using model save command I am getting nosuchmethoderror. My code works fine with 1.3x version. model.save(sc.sc(), "modelsavedir"); ERROR:

​Spark 1.6 - H​ive remote metastore not working

2015-12-15 Thread syepes
​Hello, I am testing out the 1.6 branch (#08aa3b4) and I have just noticed that spark-shell "HiveContext" is no longer able to connect to my remote metastore. Using the same build options and configuration files with 1.5 (#0fdf554) it works. Does anyone know if there have been any mayor changes

Re: [VOTE] Release Apache Spark 1.6.0 (RC2)

2015-12-15 Thread Iulian Dragoș
Thanks for the heads up. On Tue, Dec 15, 2015 at 11:40 PM, Michael Armbrust wrote: > This vote is canceled due to the issue with the incorrect version. This > issue will be fixed by https://github.com/apache/spark/pull/10317 > > We can wait a little bit for a fix to >

Re: [VOTE] Release Apache Spark 1.6.0 (RC2)

2015-12-15 Thread Iulian Dragoș
-1 (non-binding) Cluster mode on Mesos is broken (regression compared to 1.5.2). It seems to be related to the way SPARK_HOME is handled. In the driver logs I see: I1215 15:00:39.411212 28032 exec.cpp:134] Version: 0.25.0 I1215 15:00:39.413512 28037 exec.cpp:208] Executor registered on slave

Re: SparkML algos limitations question.

2015-12-15 Thread Joseph Bradley
Hi Eugene, The maxDepth parameter exists because the implementation uses Integer node IDs which correspond to positions in the binary tree. This simplified the implementation. I'd like to eventually modify it to avoid depending on tree node IDs, but that is not yet on the roadmap. There is not

Re: spark with label nodes in yarn

2015-12-15 Thread Saisai Shao
Yes, of course, capacity scheduler also needs to be configured. On Wed, Dec 16, 2015 at 10:41 AM, 张志强(旺轩) wrote: > one more question , do I have to configure label for my capacity > scheduler? is this mandatory? > > > > *发件人:* AllenZ [mailto:zzq98...@alibaba-inc.com] >

Re: spark with label nodes in yarn

2015-12-15 Thread 张志强(旺轩)
one more question , do I have to configure label for my capacity scheduler? is this mandatory? 发件人: AllenZ [mailto:zzq98...@alibaba-inc.com] 发送时间: 2015年12月16日 9:21 收件人: 'Ted Yu' 抄送: 'Saisai Shao'; 'dev' 主题: Re: spark with label nodes in yarn Oops... I do use spark 1.5.0 and apache

security testing on spark ?

2015-12-15 Thread Judy Nash
Hi all, Does anyone know of any effort from the community on security testing spark clusters. I.e. Static source code analysis to find security flaws Penetration testing to identify ways to compromise spark cluster Fuzzing to crash spark Thanks, Judy

Re: spark with label nodes in yarn

2015-12-15 Thread 张志强(旺轩)
Oops... I do use spark 1.5.0 and apache hadoop 2.6.0 (spark 1.4.1 + apache hadoop 2.6.0 is a typo), sorry Thanks, Allen 发件人: Ted Yu [mailto:yuzhih...@gmail.com] 发送时间: 2015年12月15日 22:59 收件人: 张志强(旺轩) 抄送: Saisai Shao; dev 主题: Re: spark with label nodes in yarn Please upgrade to

Re: spark with label nodes in yarn

2015-12-15 Thread Ted Yu
Please upgrade to Spark 1.5.x 1.4.1 didn't support node label feature. Cheers On Tue, Dec 15, 2015 at 2:20 AM, 张志强(旺轩) wrote: > Hi SaiSai, > > > > OK, it make sense to me , what I need is just to schedule the executors, > AND I leave one nodemanager at least with no