Hello,
I am using nutch together with solr.
Here an extract from hadoop.log
2009-05-03 14:41:32,290 INFO solr.SolrIndexer - SolrIndexer: starting
2009-05-03 14:41:32,290 INFO indexer.IndexerMapReduce -
IndexerMapReduce: crawldb: Z:/dev/searching/nutch-1.0/crawl_test/crawldb
2009-05-03 14:41:32,290 INFO indexer.IndexerMapReduce -
IndexerMapReduce: linkdb: Z:/dev/searching/nutch-1.0/crawl_test/linkdb
2009-05-03 14:41:32,290 INFO indexer.IndexerMapReduce -
IndexerMapReduces: adding segment:
Z:/dev/searching/nutch-1.0/crawl_test/segments/*
2009-05-03 14:41:34,040 WARN util.NativeCodeLoader - Unable to load
native-hadoop library for your platform... using builtin-java classes
where applicable
2009-05-03 14:41:34,040 INFO plugin.PluginRepository - Plugins: looking
in: Z:\dev\searching\nutch-1.0\plugins
2009-05-03 14:41:34,134 INFO plugin.PluginRepository - Plugin
Auto-activation mode: [true]
2009-05-03 14:41:34,134 INFO plugin.PluginRepository - Registered Plugins:
2009-05-03 14:41:34,134 INFO plugin.PluginRepository - the nutch
core extension points (nutch-extensionpoints)
...
2009-05-03 14:43:28,118 INFO indexer.IndexingFilters - Adding
org.apache.nutch.indexer.anchor.AnchorIndexingFilter
2009-05-03 14:43:37,853 FATAL solr.SolrIndexer - SolrIndexer:
java.io.IOException: Job failed!
at org.apache.hadoop.mapred.JobClient.runJob(JobClient.java:1232)
at
org.apache.nutch.indexer.solr.SolrIndexer.indexSolr(SolrIndexer.java:73)
Could anyone help me to figure out what is going wrong or where to find
more information why the solr indexer job fails.
Once this exception occurs it will happen with each call to the solr
indexer.
The only workaround I found was to remove all nutch data and recrawl. :((
- Ron