Just wanted to add more info.. I was using SparkSQL reading in the
tab-delimited raw data files converting the timestamp to Date format:
sc.textFile("rawdata/*").map(_.split("\t")).map(p => Point(df.format(new
Date( p(0).trim.toLong*1000L )), p(1), p(2).trim.toInt ,p(3).trim.toInt,
p(4).trim.toInt ,p(5)))
Then I go about registering it as table and when I run simple query like
select count(*) from
, I get the ArrayIndexOutOfBoundsException.
I bumped up the SPARK_DRIVER_MEMORY to 8g but still didn't help:
export SPARK_DRIVER_MEMORY=8g
Let me know if I'm missing any steps.. thanks!
--
View this message in context:
http://apache-spark-user-list.1001560.n3.nabble.com/spark-shell-running-into-ArrayIndexOutOfBoundsException-tp10480p10520.html
Sent from the Apache Spark User List mailing list archive at Nabble.com.