how to decide broadcast join data size

2018-07-14 Thread Selvam Raman
= 8gb, executor-core=4 Memory: 8gb(0.4% per internal) - 4.8gb for actual computation and storage. lets consider i have not done any persist in this case i could utilize 4.8gb per executor. IS IT POSSIBLE FOR ME TO USE 400MB file for BROADCAST JOIN? -- Selvam Raman "லஞ்சம் தவிர்த்து நெ

Sequence file to Image in spark

2018-04-28 Thread Selvam Raman
WARN TaskSetManager:66 - Lost task 0.0 in stage 8.0 (TID 14 , localhost, executor driver): java.lang.IllegalArgumentException: image == null ! at javax.imageio.ImageTypeSpecifier.createFromRenderedImage(Unknown Sour ce) -- Selvam Raman "லஞ்சம் தவிர்த்து நெஞ்சம் நிமிர்த்து"

Spark Higher order function

2018-03-05 Thread Selvam Raman
Dear All, i read about higher order function in databricks blog. https://docs.databricks.com/spark/latest/spark-sql/higher-order-functions-lambda-functions.html does higher order functionality available in our spark(open source)? -- Selvam Raman "லஞ்சம் தவிர்த்து நெஞ்சம் நிமிர்த்து"

Re: Spark EMR executor-core vs Vcores

2018-02-26 Thread Selvam Raman
All used cores aren't getting reported correctly in EMR, and YARN itself > has no control over it, so whatever you put in `spark.executor.cores` will > be used, > but in the ResourceManager you will only see 1 vcore used per nodemanager. > > On Mon, Feb 26, 2018 at 5:20 AM, Selvam

Re: Spark EMR executor-core vs Vcores

2018-02-26 Thread Selvam Raman
; enough memory. > > Use see 5 executor because 4 for the job and one for the application > master. > > serr the used menory and the total memory. > > On Mon, Feb 26, 2018 at 12:20 PM, Selvam Raman <sel...@gmail.com> wrote: > >> Hi, >> >> spark version

Re: Spark EMR executor-core vs Vcores

2018-02-26 Thread Selvam Raman
: 2500.054 BogoMIPS: 5000.10 Hypervisor vendor: Xen Virtualization type: full L1d cache: 32K L1i cache: 32K L2 cache: 256K L3 cache: 25600K NUMA node0 CPU(s): 0-7 On Mon, Feb 26, 2018 at 10:20 AM, Selvam Raman <sel...@gmail.

Spark EMR executor-core vs Vcores

2018-02-26 Thread Selvam Raman
e 20g+10%overhead ram(22gb), 10 core(number of threads), 1 Vcore(cpu). please correct me if my understand is wrong. how can i utilize number of vcore in EMR effectively. Will Vcore boost performance? -- Selvam Raman "லஞ்சம் தவிர்த்து நெஞ்சம் நிமிர்த்து"

Spark-Solr -- unresolved dependencies

2018-02-23 Thread Selvam Raman
.scala:294) at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:158) at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:124) at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) -- Selvam Raman "லஞ்சம் தவிர்த்து நெஞ்சம் நிமிர்த்து"

Re: pyspark+spacy throwing pickling exception

2018-02-15 Thread Selvam Raman
left out the exception. On one hand I’m also not sure how well > spacy serializes, so to debug this I would start off by moving the nlp = > inside of my function and see if it still fails. > > On Thu, Feb 15, 2018 at 9:08 PM Selvam Raman <sel...@gmail.com> wrote: > >> imp

pyspark+spacy throwing pickling exception

2018-02-15 Thread Selvam Raman
. def f(x) : print(x) description = xmlData.filter(col("dcterms:description").isNotNull()).select(col("dcterms:description").alias("desc")) description.rdd.flatMap(lambda row: getPhrases(row.desc)).foreach(f) when i am trying to access getphrases i am getting below

Re: Pyspark UDF/map fucntion throws pickling exception

2018-02-15 Thread Selvam Raman
hon3.6/site-packages/pyspark/rdd.py", line 906, in fold vals = self.mapPartitions(func).collect() File "/Library/Frameworks/Python.framework/Versions/3.6/lib/python3.6/site-packages/pyspark/rdd.py", line 809, in collect port = self.ctx._jvm.PythonRDD.collectAndServe(self._jr

Pyspark UDF/map fucntion throws pickling exception

2018-02-15 Thread Selvam Raman
ns/3.6/lib/python3.6/pickle.py", line 476, in save f(self, obj) # Call unbound method with explicit self File "/Library/Frameworks/Python.framework/Versions/3.6/lib/python3.6/pickle.py", line 751, in save_tuple save(element) File "/Library/Frameworks/Python.framework/Versions/3.6/lib/python3.6/pickle.py", line 476, in save f(self, obj) # Call unbound method with explicit self File "/Library/Frameworks/Python.framework/Versions/3.6/lib/python3.6/site-packages/pyspark/cloudpickle.py", line 368, in save_builtin_function return self.save_function(obj) File "/Library/Frameworks/Python.framework/Versions/3.6/lib/python3.6/site-packages/pyspark/cloudpickle.py", line 247, in save_function if islambda(obj) or obj.__code__.co_filename == '' or themodule is None: AttributeError: 'builtin_function_or_method' object has no attribute '__code__' please help me. -- Selvam Raman "லஞ்சம் தவிர்த்து நெஞ்சம் நிமிர்த்து"

Re: How to convert Array of Json rows into Dataset of specific columns in Spark 2.2.0?

2018-01-10 Thread Selvam Raman
> val empInfoSchema = ArrayType(employeeSchema) > > empInfoSchema.json > > val empInfoStrDF = Seq((emp_info)).toDF("emp_info_str") > empInfoStrDF.printSchema > empInfoStrDF.show(false) > > val empInfoDF = empInfoStrDF.select(from_json('emp_info_str, > empInfoSchema).as("emp_info")) > empInfoDF.printSchema > > empInfoDF.select(struct("*")).show(false) > > empInfoDF.select("emp_info.name", "emp_info.address", > "emp_info.docs").show(false) > > empInfoDF.select(explode('emp_info.getItem("name"))).show > > > > > -- > Sent from: http://apache-spark-user-list.1001560.n3.nabble.com/ > > - > To unsubscribe e-mail: user-unsubscr...@spark.apache.org > > -- Selvam Raman "லஞ்சம் தவிர்த்து நெஞ்சம் நிமிர்த்து"

how to get Cache size from storage

2017-09-05 Thread Selvam Raman
can i get those details. -- Selvam Raman "லஞ்சம் தவிர்த்து நெஞ்சம் நிமிர்த்து"

convert ps to jpg file

2017-05-26 Thread Selvam Raman
frequently i got yan OOM and disk full issue. Could you please share your thoughts? -- Selvam Raman "லஞ்சம் தவிர்த்து நெஞ்சம் நிமிர்த்து"

how to create List in pyspark

2017-04-24 Thread Selvam Raman
) How can i achieve the same df while i am reading from source? doc = spark.read.text("/Users/rs/Desktop/nohup.out") how can i create array type with "sentences" column from doc(dataframe) The below one creates more than one column. rdd.map(lambda rdd: rdd[0]).map(lambda row:row.split(" ")) -- Selvam Raman "லஞ்சம் தவிர்த்து நெஞ்சம் நிமிர்த்து"

Re: Spark Mlib - java.lang.OutOfMemoryError: Java heap space

2017-04-24 Thread Selvam Raman
# -XX:OnOutOfMemoryError="kill -9 %p" # Executing /bin/sh -c "kill -9 15090"... Killed Node-45.dev contains 8.9GB free while it throws out of memory. Can anyone please help me to understand the issue? On Mon, Apr 24, 2017 at 11:22 AM, Selvam Raman <sel...@gmail.com> wrote: > Hi, &g

Spark Mlib - java.lang.OutOfMemoryError: Java heap space

2017-04-24 Thread Selvam Raman
ecutors 4 --executor-cores 2 --executor-memory 20g Word2VecExample.py -- Selvam Raman "லஞ்சம் தவிர்த்து நெஞ்சம் நிமிர்த்து"

Pyspark - pickle.PicklingError: Can't pickle

2017-04-03 Thread Selvam Raman
ne 681, in _batch_setitems save(v) File "/Library/Frameworks/Python.framework/Versions/2.7/lib/python2.7/pickle.py", line 317, in save self.save_global(obj, rv) File "/Users/rs/Downloads/spark-2.0.1-bin-hadoop2.7/python/pyspark/cloudpickle.py", line 390, in save_

Update DF record with delta data in spark

2017-04-02 Thread Selvam Raman
Test2 2 1 Test3 3 2 Current approach: 1) Delete row in table1 where table1.composite key = table2.composite key. 2) Union all table and table2 to get updated result. is this right approach?. is there any other way to achieve it?​ -- Selvam Raman "லஞ்சம் தவிர்த்து நெஞ்சம் நிமிர்த்து"

Convert Dataframe to Dataset in pyspark

2017-04-01 Thread Selvam Raman
In Scala, val ds = sqlContext.read.text("/home/spark/1.6/lines").as[String] what is the equivalent code in pyspark? -- Selvam Raman "லஞ்சம் தவிர்த்து நெஞ்சம் நிமிர்த்து"

Re: how to read object field within json file

2017-03-25 Thread Selvam Raman
ption easily. >> >> Now you need write your own UDF, maybe can do what you want. >> >> Yong >> >> -- >> *From:* Selvam Raman <sel...@gmail.com> >> *Sent:* Thursday, March 23, 2017 5:03 PM >> *To:* user >> *Subject:* how to read object field w

how to read object field within json file

2017-03-23 Thread Selvam Raman
9":{} } I am having bzip json files like above format. some json row contains two objects within source(like F1 and F2), sometime five(F1,F2,F3,F4,F5),etc. So the final schema will contains combination of all objects for the source field. Now, every row will contain n number of objects but only some

is it possible to read .mdb file in spark

2017-01-25 Thread Selvam Raman
-- Selvam Raman "லஞ்சம் தவிர்த்து நெஞ்சம் நிமிர்த்து"

Writing into parquet throws Array out of bounds exception

2016-12-21 Thread Selvam Raman
there is one column which is longblob, if i convert to unbase64. I face this problem. i could able to write to parquet without conversion. So is there some limit for bytes per line?. Please give me your suggestion. -- Selvam Raman "லஞ்சம் தவிர்த்து நெஞ்சம் நிமிர்த்து"

Reading xls and xlsx files

2016-12-19 Thread Selvam Raman
Hi, Is there a way to read xls and xlsx files using spark?. is there any hadoop inputformat available to read xls and xlsx files which could be used in spark? -- Selvam Raman "லஞ்சம் தவிர்த்து நெஞ்சம் நிமிர்த்து"

Re: Spark dump in slave Node EMR

2016-12-16 Thread Selvam Raman
t 12:30 PM, Selvam Raman <sel...@gmail.com> wrote: > Hi, > > how can i take heap dump in EMR slave node to analyze. > > I have one master and two slave. > > if i enter jps command in Master, i could see sparksubmit with pid. > > But i could not see anything

Spark dump in slave Node EMR

2016-12-16 Thread Selvam Raman
Hi, how can i take heap dump in EMR slave node to analyze. I have one master and two slave. if i enter jps command in Master, i could see sparksubmit with pid. But i could not see anything in slave node. how can i take heap dump for spark job. -- Selvam Raman "லஞ்சம் தவிர்த்து நெ

Re: Spark Batch checkpoint

2016-12-16 Thread Selvam Raman
nUID = 1L; @Override public void call(Iterator row) throws Exception { while(row.hasNext()) { //Process data and insert into No-Sql DB } } }); } } Now where can i apply rdd.checkpoint(). Thanks, selvam On Thu, Dec 15, 2016 at 10:44 PM, Selvam Raman <sel...@gmail.com> w

Re: Spark Batch checkpoint

2016-12-15 Thread Selvam Raman
;. > This will store checkpoints on that directory that I called checkpoint. > > > Thank You, > > Irving Duran > > On Thu, Dec 15, 2016 at 10:33 AM, Selvam Raman <sel...@gmail.com> wrote: > >> Hi, >> >> is there any provision in spark batch for chec

Spark Batch checkpoint

2016-12-15 Thread Selvam Raman
or is there way for checkpoint provision. Checkpoint,what i am expecting is start from 71 partition to till end. Please give me your suggestions. -- Selvam Raman "லஞ்சம் தவிர்த்து நெஞ்சம் நிமிர்த்து"

Re: Spark Job not exited and shows running

2016-12-01 Thread Selvam Raman
times (and wanting to spare the resources > of our submitting machines) we have now switched to use yarn cluster mode > by default. This seems to resolve the problem. > > Hope this helps, > > Daniel > > On 29 Nov 2016 11:20 p.m., "Selvam Raman" <sel...@gmail.com> w

Spark Job not exited and shows running

2016-11-29 Thread Selvam Raman
. Spark version:2.0( AWS EMR). -- Selvam Raman "லஞ்சம் தவிர்த்து நெஞ்சம் நிமிர்த்து"

Java Collections.emptyList inserted as null object in cassandra

2016-11-29 Thread Selvam Raman
Filed Type in cassandra : List I am trying to insert Collections.emptyList() from spark to cassandra list field. In cassandra it stores as null object. How can i avoid null values here. -- Selvam Raman "லஞ்சம் தவிர்த்து நெஞ்சம் நிமிர்த்து"

Re: Dataframe broadcast join hint not working

2016-11-26 Thread Selvam Raman
584 bytes) compared to data it contains. b has just 85 rows and > around 4964 bytes. > Help is very much appreciated!! > > Thanks > Swapnil > > > -- Selvam Raman "லஞ்சம் தவிர்த்து நெஞ்சம் நிமிர்த்து"

How Spark determines Parquet partition size

2016-11-08 Thread Selvam Raman
"file:///Users/rs/parti").rdd.partitions.length res4: Int = 5 ​so how does parquet partitioning the data in spark?​ -- Selvam Raman "லஞ்சம் தவிர்த்து நெஞ்சம் நிமிர்த்து"

Spark Sql - "broadcast-exchange-1" java.lang.OutOfMemoryError: Java heap space

2016-10-25 Thread Selvam Raman
l data from the table. Why it is reading all the data from table and doing sort merge join for 3 or 4 tables. Why it is not applying any filtering value. Though i have given large memory for executor it is still throws the same error. when spark sql do the joining how it is utilizing memory and cores. Any guidelines would be greatly welcome. -- Selvam Raman "லஞ்சம் தவிர்த்து நெஞ்சம் நிமிர்த்து"

Re: Spark Sql 2.0 throws null pointer exception

2016-10-24 Thread Selvam Raman
have faced the problem earlier. Thanks, Selvam R​ On Mon, Oct 24, 2016 at 10:23 AM, Selvam Raman <sel...@gmail.com> wrote: > Hi All, > > Please help me. > > I have 10 (tables data) parquet file in s3. > > I am reading and storing as Dataset then registered as temp table.

Spark Sql 2.0 throws null pointer exception

2016-10-24 Thread Selvam Raman
a:1142) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) at java.lang.Thread.run(Thread.java:745)​ -- Selvam Raman "லஞ்சம் தவிர்த்து நெஞ்சம் நிமிர்த்து"

Re: PostgresSql queries vs spark sql

2016-10-23 Thread Selvam Raman
I found it. We can use pivot which is similar to cross tab In postgres. Thank you. On Oct 17, 2016 10:00 PM, "Selvam Raman" <sel...@gmail.com> wrote: > Hi, > > Please share me some idea if you work on this earlier. > How can i develop postgres CROSSTAB function in

Spark SQL parallelize

2016-10-20 Thread Selvam Raman
Hi, I am having 40+ structured data stored in s3 bucket as parquet file . I am going to use 20 table in the use case. There s a Main table which drive the whole flow. Main table contains 1k record. My use case is for every record in the main table process the rest of table( join group by

PostgresSql queries vs spark sql

2016-10-17 Thread Selvam Raman
--+++ test1| val2 | val3 | test2| val6 | val7 | -- Selvam Raman "லஞ்சம் தவிர்த்து நெஞ்சம் நிமிர்த்து"

Re: Spark-Sql 2.0 nullpointerException

2016-10-12 Thread Selvam Raman
What i am trying to achieve is Trigger query to get number(i.e.,1,2,3,...n) for every number i have to trigger another 3 queries. Thanks, selvam R On Wed, Oct 12, 2016 at 4:10 PM, Selvam Raman <sel...@gmail.com> wrote: > Hi , > > I am reading parquet file and creating temp t

Spark-Sql 2.0 nullpointerException

2016-10-12 Thread Selvam Raman
ang.Thread.run(Thread.java:745) 16/10/12 15:59:53 INFO SparkContext: Invoking stop() from shutdown hook ​ ​Please let me know if i am missing anything. Thank you for the help.​ -- Selvam Raman "லஞ்சம் தவிர்த்து நெஞ்சம் நிமிர்த்து"

Re: Spark S3

2016-10-10 Thread Selvam Raman
I mentioned parquet as input format. On Oct 10, 2016 11:06 PM, "ayan guha" <guha.a...@gmail.com> wrote: > It really depends on the input format used. > On 11 Oct 2016 08:46, "Selvam Raman" <sel...@gmail.com> wrote: > >> Hi, >> >> How spar

Spark S3

2016-10-10 Thread Selvam Raman
RDD , then we can look at the partitions.size or length to check how many partition for a file. But how this will be accomplished in terms of S3 bucket.​ -- Selvam Raman "லஞ்சம் தவிர்த்து நெஞ்சம் நிமிர்த்து"

Re: Spark CSV skip lines

2016-09-10 Thread Selvam Raman
aFrame is not yet supported. > > There is an issue open[2]. I hope this is helpful. > > Thanks. > > [1] https://github.com/apache/spark/blob/27209252f09ff73c58e60c6df8aaba > 73b308088c/sql/core/src/main/scala/org/apache/spark/sql/ > DataFrameReader.scala#L369 > [2] https://

Spark CSV skip lines

2016-09-10 Thread Selvam Raman
. -- Selvam Raman "லஞ்சம் தவிர்த்து நெஞ்சம் நிமிர்த்து"

Spark Checkpoint for JDBC/ODBC

2016-09-06 Thread Selvam Raman
Hi, Need your input to take decision. We have an n-number of databases(ie oracle, MySQL,etc). I want to read a data from the sources but how it is maintaining fault tolerance in source side. if source side system went down. how the spark system reads the data. -- Selvam Raman "ல

Cassandra timestamp to spark Date field

2016-09-05 Thread Selvam Raman
ues)),schema) in schema fields I have mentioned timestamp as *StructField*("shipped_datetime", *DateType*), when I try to show the result, it throws java.util.Date can not convert to java.sql.Date. so how can I solve the issue. First I have converted cassandrascanrdd to -- Selvam

Re: spark cassandra issue

2016-09-04 Thread Selvam Raman
; LinkedIn * >>> https://www.linkedin.com/profile/view?id=AAEWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw >>> <https://www.linkedin.com/profile/view?id=AAEWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw>* >>> >>> >>> >>> http://talebzadehmich.wordpress.c

Re: spark cassandra issue

2016-09-04 Thread Selvam Raman
own risk. Any and all responsibility for any > loss, damage or destruction of data or any other property which may arise > from relying on this email's technical content is explicitly disclaimed. > The author will in no case be liable for any monetary damages arising from > such loss, damag

Re: spark cassandra issue

2016-09-04 Thread Selvam Raman
its very urgent. please help me guys. On Sun, Sep 4, 2016 at 8:05 PM, Selvam Raman <sel...@gmail.com> wrote: > Please help me to solve the issue. > > spark-shell --packages com.datastax.spark:spark-cassandra-connector_2.10:1.3.0 > --conf spark.cassandra.connection.host

spark cassandra issue

2016-09-04 Thread Selvam Raman
ndra.DefaultSource.createRelation(DefaultSource.scala:56) at org.apache.spark.sql.execution.datasources.ResolvedDataSource$.apply(ResolvedDataSource.scala:125) a -- Selvam Raman "லஞ்சம் தவிர்த்து நெஞ்சம் நிமிர்த்து"

Need a help in row repetation

2016-09-03 Thread Selvam Raman
please give me any suggestion in terms of dataframe. -- Selvam Raman "லஞ்சம் தவிர்த்து நெஞ்சம் நிமிர்த்து"

Re: Insert non-null values from dataframe

2016-08-29 Thread Selvam Raman
from being read as null though, it > will only skip writing tombstones. > > On Thu, Aug 25, 2016, 1:23 PM Selvam Raman <sel...@gmail.com> wrote: > >> Hi , >> >> Dataframe: >> colA colB colC colD colE >> 1 2 3 4 5 >> 1 2 3 null null >> 1 null

Insert non-null values from dataframe

2016-08-25 Thread Selvam Raman
) Record 2:(1,2,3) Record 3:(1,5) Record 4:(3,4,5) -- Selvam Raman "லஞ்சம் தவிர்த்து நெஞ்சம் நிமிர்த்து"

Re: Windows operation orderBy desc

2016-08-24 Thread Selvam Raman
m.invalid> > wrote: > >> Hi, >> >> in the following Window spec I want orderBy ("") to be displayed >> in descending order please >> >> val W = Window.partitionBy("col1").orderBy("col2") >> >> If I Do >> >

Extract year from string format of date

2016-08-17 Thread Selvam Raman
() *function but it gives only null values for the string same for *to_date()* function. -- Selvam Raman "லஞ்சம் தவிர்த்து நெஞ்சம் நிமிர்த்து"

Re: Data frame Performance

2016-08-16 Thread Selvam Raman
own risk. Any and all responsibility for any > loss, damage or destruction of data or any other property which may arise > from relying on this email's technical content is explicitly disclaimed. > The author will in no case be liable for any monetary damages arising from > such lo

Data frame Performance

2016-08-16 Thread Selvam Raman
qlContext.sql("select site,valudf(collect_set(requests)) as test from sel group by site").first -- Selvam Raman "லஞ்சம் தவிர்த்து நெஞ்சம் நிமிர்த்து"

Re: Get distinct column data from grouped data

2016-08-09 Thread Selvam Raman
selvam R On Tue, Aug 9, 2016 at 4:19 PM, Selvam Raman <sel...@gmail.com> wrote: > Example: > > sel1 test > sel1 test > sel1 ok > sel2 ok > sel2 test > > > expected result: > > sel1, [test,ok] > sel2,[test,ok] > > How to achieve the above re

Get distinct column data from grouped data

2016-08-09 Thread Selvam Raman
Example: sel1 test sel1 test sel1 ok sel2 ok sel2 test expected result: sel1, [test,ok] sel2,[test,ok] How to achieve the above result using spark dataframe. please suggest me. -- Selvam Raman "லஞ்சம் தவிர்த்து நெஞ்சம் நிமிர்த்து"

SparkSession for RDBMS

2016-08-03 Thread Selvam Raman
. -- Selvam Raman "லஞ்சம் தவிர்த்து நெஞ்சம் நிமிர்த்து"

Sqoop On Spark

2016-08-01 Thread Selvam Raman
Hi Team, how can i use spark as execution engine in sqoop2. i see the patch(S QOOP-1532 <https://issues.apache.org/jira/browse/SQOOP-1532>) but it shows in progess. so can not we use sqoop on spark. Please help me if you have an any idea. -- Selvam Raman "லஞ்சம் தவிர்த்து நெஞ்சம் நிமிர்த்து"

Skew data

2016-06-16 Thread Selvam Raman
Hi, What is skew data. I read that if the data was skewed while joining it would take long time to finish the job.(99 percent finished in seconds where 1 percent of task taking minutes to hour). How to handle skewed data in spark. Thanks, Selvam R +91-97877-87724

Windows Rstudio to Linux spakR

2016-06-01 Thread Selvam Raman
Hi , How to connect to sparkR (which is available in Linux env) using Rstudio(Windows env). Please help me. -- Selvam Raman "லஞ்சம் தவிர்த்து நெஞ்சம் நிமிர்த்து"

Re: Release Announcement: XGBoost4J - Portable Distributed XGBoost in Spark, Flink and Dataflow

2016-05-25 Thread Selvam Raman
XGBoost4J could integrate with spark from 1.6 version. Currently I am using spark 1.5.2. Can I use XGBoost instead of XGBoost4j. Will both provides same result. Thanks, Selvam R +91-97877-87724 On Mar 15, 2016 9:23 PM, "Nan Zhu" wrote: > Dear Spark Users and

next on empty iterator though i used hasNext

2016-04-25 Thread Selvam Raman
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214) ... 3 more -- Selvam Raman "லஞ்சம் தவிர்த்து நெஞ்சம் நிமிர்த்து"

HiveContext in spark

2016-04-12 Thread Selvam Raman
Ql.scala:1217) ​ -- Selvam Raman "லஞ்சம் தவிர்த்து நெஞ்சம் நிமிர்த்து"

Re: Null pointer exception when using com.databricks.spark.csv

2016-03-30 Thread Selvam Raman
ny file (eg. textFile()) work as well? > > I think this is related with this thread, > http://apache-spark-user-list.1001560.n3.nabble.com/Error-while-running-example-scala-application-using-spark-submit-td10056.html > . > > > 2016-03-30 12:44 GMT+09:00 Selvam Raman <sel..

Null pointer exception when using com.databricks.spark.csv

2016-03-29 Thread Selvam Raman
] -- Selvam Raman "லஞ்சம் தவிர்த்து நெஞ்சம் நிமிர்த்து"