The value you're specifying for io.serializations below is incorrect:

<property>
<name>io.serializations</name>
<value>org.apache.avro.mapred.AvroSerialization,
avro.serialization.key.reader.schema,
avro.serialization.value.reader.schema,
avro.serialization.key.writer.schema,avro.serialization.value.writer.schema
</value>
</property>

If the goal is to include org.apache.avro.mapred.AvroSerialization,
then it should look more like:

<property>
  <name>io.serializations</name>
  
<value>org.apache.hadoop.io.serializer.WritableSerialization,org.apache.hadoop.io.serializer.avro.AvroSpecificSerialization,org.apache.hadoop.io.serializer.avro.AvroReflectSerialization,org.apache.avro.mapred.AvroSerialization</value>
</property>

That is, it must be an extension of the default values, and not a
replacement of them.

On Wed, Mar 13, 2013 at 4:05 AM, M, Paul <pa...@iqt.org> wrote:
> Hello,
>
> I am trying to run an M/R job with Avro serialization via Oozie.  I've made
> some progress in the workflow.xml, however I am still running into the
> following error.  Any thoughts?  I believe it may have to do with the
> io.serializations property below.   FYI, I am using CDH 4.2.0 mr1.
>
> 2013-03-12 15:24:32,334 INFO org.apache.hadoop.mapred.TaskInProgress: Error
> from attempt_201303111118_0080_m_000000_3: java.lang.NullPointerException
> at org.apache.hadoop.mapred.MapTask.getSplitDetails(MapTask.java:356)
> at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:389)
> at org.apache.hadoop.mapred.MapTask.run(MapTask.java:333)
> at org.apache.hadoop.mapred.Child$4.run(Child.java:268)
> at java.security.AccessController.doPrivileged(Native Method)
> at javax.security.auth.Subject.doAs(Subject.java:396)
> at
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1407)
> at org.apache.hadoop.mapred.Child.main(Child.java:262)
>
>
> <action name="mr-node">
> <map-reduce>
> <job-tracker>${jobTracker}</job-tracker>
> <name-node>${nameNode}</name-node>
> <prepare>
> <delete path="${nameNode}/user/${wf:user()}/${outputDir}" />
> </prepare>
> <configuration>
> <property>
> <name>mapred.job.queue.name</name>
> <value>${queueName}</value>
> </property>
>
> <property>
> <name>mapreduce.reduce.class</name>
> <value>org.apache.avro.mapred.HadoopReducer</value>
> </property>
> <property>
> <name>mapreduce.map.class</name>
> <value>org.apache.avro.mapred.HadoopMapper</value>
> </property>
>
>
> <property>
> <name>avro.reducer</name>
> <value>org.my.project.mapreduce.CombineAvroRecordsByHourReducer
> </value>
> </property>
>
> <property>
> <name>avro.mapper</name>
> <value>org.my.project.mapreduce.ParseMetadataAsTextIntoAvroMapper
> </value>
> </property>
>
>
> <property>
> <name>mapreduce.inputformat.class</name>
> <value>org.my.project.mapreduce.NonSplitableInputFormat</value>
> </property>
>
> <!-- Key Value Mapper -->
> <property>
> <name>avro.output.schema</name>
> <value>{"type":"record","name":"Pair","namespace":"org.apache.avro.mapred","fields":..."}]}
> </value>
> </property>
> <property>
> <name>mapred.mapoutput.key.class</name>
> <value>org.apache.avro.mapred.AvroKey</value>
> </property>
> <property>
> <name>mapred.mapoutput.value.class</name>
> <value>org.apache.avro.mapred.AvroValue</value>
> </property>
>
>
> <property>
> <name>avro.schema.output.key</name>
> <value>{"type":"record","name":"DataRecord","namespace":...]}]}
> </value>
> </property>
>
> <property>
> <name>mapreduce.outputformat.class</name>
> <value>org.apache.hadoop.mapreduce.lib.output.TextOutputFormat
> </value>
> </property>
>
> <property>
> <name>mapred.output.key.comparator.class</name>
> <value>org.apache.avro.mapred.AvroKeyComparator</value>
> </property>
>
> <property>
> <name>io.serializations</name>
> <value>org.apache.avro.mapred.AvroSerialization,
> avro.serialization.key.reader.schema,
> avro.serialization.value.reader.schema,
> avro.serialization.key.writer.schema,avro.serialization.value.writer.schema
> </value>
> </property>
>
> <property>
> <name>mapred.map.tasks</name>
> <value>1</value>
> </property>
>
>
>
> <!--Input/Output -->
> <property>
> <name>mapred.input.dir</name>
> <value>/user/${wf:user()}/input/</value>
> </property>
> <property>
> <name>mapred.output.dir</name>
> <value>/user/${wf:user()}/${outputDir}</value>
> </property>
> </configuration>
> </map-reduce>



-- 
Harsh J

Reply via email to