Re: unsubsribe

2018-10-30 Thread Anu B Nair
I have already send minimum 10 times! Today also I have send one! On Tue, Oct 30, 2018 at 3:51 PM Biplob Biswas wrote: > You need to send the email to user-unsubscr...@spark.apache.org and not > to the usergroup. > > Thanks & Regards > Biplob Biswas > > > On Tue, Oc

Re: unsubsribe

2018-10-30 Thread Anu B Nair
I am sending this Unsubscribe mail for last few months! It never happens! If anyone can help us to unsubscribe it wil be really helpful! On Tue, Oct 30, 2018 at 3:27 PM Mohan Palavancha wrote: > >

Unsubscribe

2018-09-06 Thread Anu B Nair
Hi, I have tried all possible way to unsubscripted from this group. Can anyone help? -- Anu

Unsubscribe

2018-04-18 Thread Anu B Nair

Unsubscribe

2018-02-25 Thread Anu B Nair

Unsubscribe

2018-01-18 Thread Anu B Nair

unsubscribe

2018-01-16 Thread Anu B Nair

Java heap space OutOfMemoryError in pyspark spark-submit (spark version:2.2)

2018-01-04 Thread Anu B Nair
yOf(Arrays.java:3230) at java.io.ByteArrayOutputStream.grow(ByteArrayOutputStream.java:113) at java.io.ByteArrayOutputStream.ensureCapacity(ByteArrayOutputStream.java:93* Please let me know how to resolve this ? -- Anu

Fwd: [pyspark][MLlib] Getting WARN FPGrowth: Input data is not cached for cached data

2017-12-21 Thread Anu B Nair
int(result)* Understood that it is a warning, but just wanted to know in detail -- Anu r z h k p z y x w v u t s s x o n r x z y m t s q e z x z y r q t p from pyspark.mllib.fpm import FPGrowth import pyspark from pyspark.context import SparkContext from pyspark.sql.session import Spark

Re: SparkSQL Timestamp query failure

2015-03-30 Thread anu
Hi Alessandro Could you specify which query were you able to run successfully? 1. sqlContext.sql(SELECT * FROM Logs as l where l.timestamp = '2012-10-08 16:10:36' ).collect OR 2. sqlContext.sql(SELECT * FROM Logs as l where cast(l.timestamp as string) = '2012-10-08 16:10:36.0').collect I am

Measuer Bytes READ and Peak Memory Usage for Query

2015-03-20 Thread anu
Hi All I would like to measure Bytes Read and Peak Memory Usage for a Spark SQL Query. Please clarify if Bytes Read = aggregate size of all RDDs ?? All my RDDs are in memory and 0B spill to disk. And I am clueless how to measure Peak Memory Usage. -- View this message in context:

Hive on Spark with Spark as a service on CDH5.2

2015-03-17 Thread anu
*I am not clear if spark sql supports HIve on Spark when spark is run as a service in CDH 5.2? * Can someone please clarify this. If this is possible, how what configuration changes have I to make to import hive context in spark shell as well as to be able to do a spark-submit for the job to be

Transform a Schema RDD to another Schema RDD with a different schema

2015-03-17 Thread anu
I have a schema RDD with thw following Schema : scala mainRDD.printSchema root |-- COL1: integer (nullable = false) |-- COL2: integer (nullable = false) |-- COL3: string (nullable = true) |-- COL4: double (nullable = false) |-- COL5: string (nullable = true) Now, I transform the mainRDD

Iterate over contents of schemaRDD loaded from parquet file to extract timestamp

2015-03-16 Thread anu
Spark Version - 1.1.0 Scala - 2.10.4 I have loaded following type data from a parquet file, stored in a schemaRDD [7654321,2015-01-01 00:00:00.007,0.49,THU] Since, in spark version 1.1.0, parquet format doesn't support saving timestamp valuues, I have saved the timestamp data as string. Can you

Optimizing SQL Query

2015-03-06 Thread anu
I have a query that's like: Could you help in providing me pointers as to how to start to optimize it w.r.t. spark sql: sqlContext.sql( SELECT dw.DAY_OF_WEEK, dw.HOUR, avg(dw.SDP_USAGE) AS AVG_SDP_USAGE FROM ( SELECT sdp.WID, DAY_OF_WEEK, HOUR, SUM(INTERVAL_VALUE) AS SDP_USAGE

Re: SparkSQL Timestamp query failure

2015-03-02 Thread anu
Thank you Alessandro :) On Tue, Mar 3, 2015 at 10:03 AM, whitebread [via Apache Spark User List] ml-node+s1001560n2188...@n3.nabble.com wrote: Anu, 1) I defined my class Header as it follows: case class Header(timestamp: java.sql.Timestamp, c_ip: String, cs_username: String, s_ip: String

Re: SparkSQL Timestamp query failure

2015-03-02 Thread anu
Can you please post how did you overcome this issue. -- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/SparkSQL-Timestamp-query-failure-tp19502p21868.html Sent from the Apache Spark User List mailing list archive at Nabble.com.

Facing error: java.lang.ArrayIndexOutOfBoundsException while executing SparkSQL join query

2015-02-27 Thread anu
I have three tables with the following schema: case class *date_d*(WID: Int, CALENDAR_DATE: java.sql.Timestamp, DATE_STRING: String, DAY_OF_WEEK: String, DAY_OF_MONTH: Int, DAY_OF_YEAR: Int, END_OF_MONTH_FLAG: String, YEARWEEK: Int, CALENDAR_MONTH: String, MONTH_NUM: Int, YEARMONTH: Int, QUARTER:

Facing error while extending scala class with Product interface to overcome limit of 22 fields in spark-shell

2015-02-24 Thread anu
My issue is posted here on stack-overflow. What am I doing wrong here? http://stackoverflow.com/questions/28689186/facing-error-while-extending-scala-class-with-product-interface-to-overcome-limi -- View this message in context: