Re: spark sql writing in avro

2015-03-13 Thread M. Dale
to it not being implemented (the job quits and says non implemented method or something along those lines). I will try going the spark shell and passing in the jar built from github since I haven't tried that quite yet. On Thu, Mar 12, 2015 at 6:44 PM, M. Dale medal...@yahoo.com mailto:medal

Re: spark sql writing in avro

2015-03-12 Thread M. Dale
Short answer: if you downloaded spark-avro from the repo.maven.apache.org repo you might be using an old version (pre-November 14, 2014) - see timestamps at http://repo.maven.apache.org/maven2/com/databricks/spark-avro_2.10/0.1/ Lots of changes at https://github.com/databricks/spark-avro since

Re: IncompatibleClassChangeError

2015-03-05 Thread M. Dale
In Hadoop 1.x TaskAttemptContext is a class (for example, https://hadoop.apache.org/docs/r1.2.1/api/org/apache/hadoop/mapred/TaskAttemptContext.html) In Hadoop 2.x TaskAttemptContext is an interface (https://hadoop.apache.org/docs/r2.4.0/api/org/apache/hadoop/mapreduce/TaskAttemptContext.html)

Re: “mapreduce.job.user.classpath.first” for Spark

2015-02-03 Thread M. Dale
Try spark.yarn.user.classpath.first (see https://issues.apache.org/jira/browse/SPARK-2996 - only works for YARN). Also thread at http://apache-spark-user-list.1001560.n3.nabble.com/netty-on-classpath-when-using-spark-submit-td18030.html. HTH, Markus On 02/03/2015 11:20 PM, Corey Nolet wrote:

Re: java.io.NotSerializableException: org.apache.avro.mapred.AvroKey using spark with avro

2014-12-18 Thread M. Dale
I did not encounter this with my Avro records using Spark 1.10 (see https://github.com/medale/spark-mail/blob/master/analytics/src/main/scala/com/uebercomputing/analytics/basic/UniqueSenderCounter.scala). I do use the default Java serialization but all the fields in my Avro object are

Re: netty on classpath when using spark-submit

2014-11-04 Thread M. Dale
Tobias, From http://spark.apache.org/docs/latest/configuration.html it seems that there is an experimental property: spark.files.userClassPathFirst Whether to give user-added jars precedence over Spark's own jars when loading classes in Executors. This feature can be used to mitigate