PYSPARK_PYTHON doesn't work in spark worker

2016-03-07 Thread guoqing0...@yahoo.com.hk
4j-0.8.2.1-src.zip:/data/hadoop_local/usercache/hdptest/filecache/38/spark-assembly-1.4.1-hadoop2.5.2.jar:pyspark.zip:py4j-0.8.2.1-src.zip It seem that PYSPARK_PYTHON doesn't work in spark worker , can someone please help me to solve it ? Thanks~ guoqing0...@yahoo.com.hk

Table is modified by DataFrameWriter

2015-09-16 Thread guoqing0...@yahoo.com.hk
Hi all, I found the table structure was modified when use DataFrameWriter.jdbc to save the content of DataFrame , sqlContext.sql("select '2015-09-17',count(1) from test").write.jdbc(url,test,properties) table structure before saving: app_key text t_amount bigint(20) saved: _c0 text _c1

Re: Re: Table is modified by DataFrameWriter

2015-09-16 Thread guoqing0...@yahoo.com.hk
name was modified. From: Josh Rosen Date: 2015-09-17 11:42 To: guoqing0...@yahoo.com.hk CC: Ted Yu; user Subject: Re: Re: Table is modified by DataFrameWriter What are your JDBC properties configured to? Do you have overwrite mode enabled? On Wed, Sep 16, 2015 at 7:39 PM, guoqing0

Re: Re: Table is modified by DataFrameWriter

2015-09-16 Thread guoqing0...@yahoo.com.hk
Spark-1.4.1 From: Ted Yu Date: 2015-09-17 10:29 To: guoqing0...@yahoo.com.hk CC: user Subject: Re: Table is modified by DataFrameWriter Can you tell us which release you were using ? Thanks On Sep 16, 2015, at 7:11 PM, "guoqing0...@yahoo.com.hk" <guoqing0...@yahoo.com.hk> w

java.util.NoSuchElementException: key not found

2015-09-11 Thread guoqing0...@yahoo.com.hk
(RDD.scala:262) at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:297) guoqing0...@yahoo.com.hk

Insert operation in Dataframe

2015-08-07 Thread guoqing0...@yahoo.com.hk
Hi all , Is the Dataframe support the insert operation , like sqlContext.sql(insert into table1 xxx select xxx from table2) ? guoqing0...@yahoo.com.hk

streaming issue

2015-07-27 Thread guoqing0...@yahoo.com.hk
Hi, I got a error when running spark streaming as below . java.lang.reflect.InvocationTargetException at sun.reflect.GeneratedMethodAccessor13.invoke(Unknown Source) at

Spark build with Hive

2015-05-20 Thread guoqing0...@yahoo.com.hk
2.4.X with Hive 12 support mvn -Pyarn -Phadoop-2.4 -Dhadoop.version=2.4.0 -Phive -Phive-0.12.0 -Phive-thriftserver -DskipTests clean package guoqing0...@yahoo.com.hk

Re: RE: Spark build with Hive

2015-05-20 Thread guoqing0...@yahoo.com.hk
Thanks very much , Which version will be support In the upcome 1.4 ? I hope it will be support more versions. guoqing0...@yahoo.com.hk From: Cheng, Hao Date: 2015-05-21 11:20 To: Ted Yu; guoqing0...@yahoo.com.hk CC: user Subject: RE: Spark build with Hive Yes, ONLY support 0.12.0 and 0.13.1

Hive on Spark VS Spark SQL

2015-05-19 Thread guoqing0...@yahoo.com.hk
Hive on Spark and SparkSQL which should be better , and what are the key characteristics and the advantages and the disadvantages between ? guoqing0...@yahoo.com.hk

Worker Core in Spark

2015-05-12 Thread guoqing0...@yahoo.com.hk
Assume that i had several mathines with 8cores , 1 core per work with 8 workers , 8 cores per work with 1 work , which one is better ?

Re: Re: implicit function in SparkStreaming

2015-04-29 Thread guoqing0...@yahoo.com.hk
Thank you for your pointers , it`s very helpful to me , in this scenario how can i use the implicit def in the enclosing class ? From: Tathagata Das Date: 2015-04-30 07:00 To: guoqing0...@yahoo.com.hk CC: user Subject: Re: implicit function in SparkStreaming I believe that the implicit def

Re: Re: implicit function in SparkStreaming

2015-04-29 Thread guoqing0...@yahoo.com.hk
Appreciate for your help , it works . i`m curious why the enclosing class cannot serialized , is it need to extends java.io.Serializable ? if object never serialized how it works in the task .whether there`s any association with the spark.closure.serializer . guoqing0...@yahoo.com.hk From

Is the Spark-1.3.1 support build with scala 2.8 ?

2015-04-23 Thread guoqing0...@yahoo.com.hk
Is the Spark-1.3.1 support build with scala 2.8 ? Wether it can integrated with kafka_2.8.0-0.8.0 If build with scala 2.10 . Thanks.

Re: Re: HiveContext setConf seems not stable

2015-04-23 Thread guoqing0...@yahoo.com.hk
Hi all , My understanding for this problem is SQLConf will be overwrite by the hiveconfig in initialization phase when setConf(key: String, value: String) being called in the first time as below code snippets , so it is correctly in later. I`m not sure whether it is right , any point are

Re: Re: Is the Spark-1.3.1 support build with scala 2.8 ?

2015-04-23 Thread guoqing0...@yahoo.com.hk
Thank you very much for your suggestion. Regards, From: madhu phatak Date: 2015-04-24 13:06 To: guoqing0...@yahoo.com.hk CC: user Subject: Re: Is the Spark-1.3.1 support build with scala 2.8 ? Hi, AFAIK it's only build with 2.10 and 2.11. You should integrate kafka_2.10.0-0.8.0 to make

Re: Re: problem with spark thrift server

2015-04-23 Thread guoqing0...@yahoo.com.hk
:40 To: guoqing0...@yahoo.com.hk CC: user Subject: Re: problem with spark thrift server Hi What do you mean disable the driver? what are you trying to achieve. Thanks Arush On Thu, Apr 23, 2015 at 12:29 PM, guoqing0...@yahoo.com.hk guoqing0...@yahoo.com.hk wrote: Hi , I have a question about