Hi,

Glad to see you sorted out the problem. Out of curiosity what version of
jackson were you using and what was pulling it in? Can you share you maven
pom/gradle build?


On Sun, Jul 6, 2014 at 10:27 PM, Brian Thomas <brianjthoma...@gmail.com>
wrote:

> I figured it out, dependency issue in my classpath.  Maven was pulling
> down a very old version of the jackson jar.  I added the following line to
> my dependencies and the error went away:
>
> compile 'org.codehaus.jackson:jackson-mapper-asl:1.9.13'
>
>
> On Friday, July 4, 2014 3:22:30 PM UTC-4, Brian Thomas wrote:
>>
>>  I am trying to test querying elasticsearch using Apache Spark using
>> elasticsearch-hadoop.  I am just trying to do a query to the elasticsearch
>> server and return the count of results.
>>
>> Below is my test class using the Java API:
>>
>> import org.apache.hadoop.conf.Configuration;
>> import org.apache.hadoop.io.MapWritable;
>> import org.apache.hadoop.io.Text;
>> import org.apache.spark.SparkConf;
>> import org.apache.spark.api.java.JavaPairRDD;
>> import org.apache.spark.api.java.JavaSparkContext;
>> import org.apache.spark.serializer.KryoSerializer;
>> import org.elasticsearch.hadoop.mr.EsInputFormat;
>>
>> import scala.Tuple2;
>>
>> public class ElasticsearchSparkQuery{
>>
>>     public static int query(String masterUrl, String
>> elasticsearchHostPort) {
>>         SparkConf sparkConfig = new SparkConf().setAppName("
>> ESQuery").setMaster(masterUrl);
>>         sparkConfig.set("spark.serializer",
>> KryoSerializer.class.getName());
>>         JavaSparkContext sparkContext = new JavaSparkContext(sparkConfig);
>>
>>         Configuration conf = new Configuration();
>>         conf.setBoolean("mapred.map.tasks.speculative.execution", false);
>>         conf.setBoolean("mapred.reduce.tasks.speculative.execution",
>> false);
>>         conf.set("es.nodes", elasticsearchHostPort);
>>         conf.set("es.resource", "media/docs");
>>         conf.set("es.query", "?q=*");
>>
>>         JavaPairRDD<Text, MapWritable> esRDD =
>> sparkContext.newAPIHadoopRDD(conf, EsInputFormat.class, Text.class,
>>                 MapWritable.class);
>>         return (int) esRDD.count();
>>     }
>> }
>>
>>
>> When I try to run this I get the following error:
>>
>>
>> 4/07/04 14:58:07 INFO executor.Executor: Running task ID 0
>> 14/07/04 14:58:07 INFO storage.BlockManager: Found block broadcast_0
>> locally
>> 14/07/04 14:58:07 INFO rdd.NewHadoopRDD: Input split: ShardInputSplit
>> [node=[5UATWUzmTUuNzhmGxXWy_w/S'byll|10.45.71.152:9200],shard=0]
>> 14/07/04 14:58:07 WARN mr.EsInputFormat: Cannot determine task id...
>> 14/07/04 14:58:07 ERROR executor.Executor: Exception in task ID 0
>> java.lang.NoSuchFieldError: ALLOW_UNQUOTED_FIELD_NAMES
>>     at org.elasticsearch.hadoop.serialization.json.
>> JacksonJsonParser.<clinit>(JacksonJsonParser.java:38)
>>     at org.elasticsearch.hadoop.serialization.ScrollReader.
>> read(ScrollReader.java:75)
>>     at org.elasticsearch.hadoop.rest.RestRepository.scroll(
>> RestRepository.java:267)
>>     at org.elasticsearch.hadoop.rest.ScrollQuery.hasNext(
>> ScrollQuery.java:75)
>>     at org.elasticsearch.hadoop.mr.EsInputFormat$ShardRecordReader.next(
>> EsInputFormat.java:319)
>>     at org.elasticsearch.hadoop.mr.EsInputFormat$ShardRecordReader.
>> nextKeyValue(EsInputFormat.java:255)
>>     at org.apache.spark.rdd.NewHadoopRDD$$anon$1.hasNext(
>> NewHadoopRDD.scala:122)
>>     at org.apache.spark.InterruptibleIterator.hasNext(
>> InterruptibleIterator.scala:39)
>>     at org.apache.spark.util.Utils$.getIteratorSize(Utils.scala:1014)
>>     at org.apache.spark.rdd.RDD$$anonfun$count$1.apply(RDD.scala:847)
>>     at org.apache.spark.rdd.RDD$$anonfun$count$1.apply(RDD.scala:847)
>>     at org.apache.spark.SparkContext$$anonfun$runJob$4.apply(
>> SparkContext.scala:1080)
>>     at org.apache.spark.SparkContext$$anonfun$runJob$4.apply(
>> SparkContext.scala:1080)
>>     at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.
>> scala:111)
>>     at org.apache.spark.scheduler.Task.run(Task.scala:51)
>>     at org.apache.spark.executor.Executor$TaskRunner.run(
>> Executor.scala:187)
>>     at java.util.concurrent.ThreadPoolExecutor.runWorker(
>> ThreadPoolExecutor.java:1145)
>>     at java.util.concurrent.ThreadPoolExecutor$Worker.run(
>> ThreadPoolExecutor.java:615)
>>     at java.lang.Thread.run(Thread.java:745)
>>
>> Has anyone run into this issue with the JacksonJsonParser?
>>
>>  --
> You received this message because you are subscribed to the Google Groups
> "elasticsearch" group.
> To unsubscribe from this group and stop receiving emails from it, send an
> email to elasticsearch+unsubscr...@googlegroups.com.
> To view this discussion on the web visit
> https://groups.google.com/d/msgid/elasticsearch/9c2b2f2e-5196-4a72-bfbc-4cd0fda9edf0%40googlegroups.com
> <https://groups.google.com/d/msgid/elasticsearch/9c2b2f2e-5196-4a72-bfbc-4cd0fda9edf0%40googlegroups.com?utm_medium=email&utm_source=footer>
> .
>
> For more options, visit https://groups.google.com/d/optout.
>

-- 
You received this message because you are subscribed to the Google Groups 
"elasticsearch" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to elasticsearch+unsubscr...@googlegroups.com.
To view this discussion on the web visit 
https://groups.google.com/d/msgid/elasticsearch/CAJogdmd0fR795J7p3BRMOgN2U%2BTq3hYr9bzCKabmk9q8VOT0VQ%40mail.gmail.com.
For more options, visit https://groups.google.com/d/optout.

Reply via email to