Re: Clustering text data with MLlib

2014-12-30 Thread xhudik
Kmeans really needs to have identified number of clusters in advance. There are multiple algorithms (XMeans, ART,...) which do not need this information. Unfortunately, none of them is implemented in MLLib for the moment (you can give a hand and help community). Anyway, it seems to me you will

Re: building spark1.2 meet error

2014-12-30 Thread xhudik
Hi, well, spark 1.2 was prepared for scala 2.10. If you want stable and fully functional tool I'd compile it this default compiler. *I was able to compile Spar 1.2 by Java 7 and scala 2.10 seamlessly.* I also tried Java8 and scala 2.11 (no -Dscala.usejavacp=true), but I failed for some other

Re: Mllib native netlib-java/OpenBLAS

2014-12-30 Thread xhudik
I'm half-way there follow 1. compiled and installed open blas library 2. ln -s libopenblas_sandybridgep-r0.2.13.so /usr/lib/libblas.so.3 3. compiled and built spark: mvn -Pnetlib-lgpl -DskipTests clean compile package So far so fine. Then I run into problems by testing the solution:

Re: building spark1.2 meet error

2014-12-31 Thread xhudik
Hi J_soft, for me it is working, I didn't put -Dscala-2.10 -X parameters. I got only one warning, since I don't have hadoop 2.5 it didn't activate this profile: /larix@kovral:~/sources/spark-1.2.0 mvn -Pyarn -Phadoop-2.5 -Dhadoop.version=2.5.0 -DskipTests clean package Found 0 infos

MLLIB and Openblas library in non-default dir

2015-01-02 Thread xhudik
Hi I have compiled OpenBlas library into nonstandard directory and I want to inform Spark app about it via: -Dcom.github.fommil.netlib.NativeSystemBLAS.natives=/usr/local/lib/libopenblas.so which is a standard option in netlib-java (https://github.com/fommil/netlib-java) I tried 2 ways: 1. via

Re: building spark1.2 meet error

2015-01-04 Thread xhudik
Hi J_soft mvn do not provide tar packages by default. You got many jar files - each project has its own jar (e.g. mllib has mllib/target/spark-mllib_2.10-1.2.0.jar). However, if you want one big tar package with all dependencies - look here: https://github.com/apache/spark/tree/master/assembly

Re: Problem with building spark-1.2.0

2015-01-04 Thread xhudik
The error you provided says that build was unsuccessful. If you write what you did (what command you used), whole error trace - someone might be able to help you ... -- View this message in context: