On Wed, Dec 3, 2008 at 8:55 PM, brainstorm <[EMAIL PROTECTED]> wrote: > Using nutch 0.9 (hadoop 0.17.1): > > [EMAIL PROTECTED] working]$ bin/nutch readlinkdb > /home/hadoop/crawl-20081201/crawldb -dump crawled_urls.txt > LinkDb dump: starting > LinkDb db: /home/hadoop/crawl-urls-20081201/crawldb ^^^^^^^^^^^^^^^^^^^^^^^^^^^^
It seems you are providing a crawldb as argument. You should pass the linkdb. > java.io.IOException: Type mismatch in value from map: expected > org.apache.nutch.crawl.Inlinks, recieved > org.apache.nutch.crawl.CrawlDatum > at > org.apache.hadoop.mapred.MapTask$MapOutputBuffer.collect(MapTask.java:427) > at > org.apache.hadoop.mapred.lib.IdentityMapper.map(IdentityMapper.java:37) > at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:47) > at org.apache.hadoop.mapred.MapTask.run(MapTask.java:219) > at > org.apache.hadoop.mapred.TaskTracker$Child.main(TaskTracker.java:2124) > > LinkDbReader: java.io.IOException: Job failed! > at org.apache.hadoop.mapred.JobClient.runJob(JobClient.java:1062) > at > org.apache.nutch.crawl.LinkDbReader.processDumpJob(LinkDbReader.java:110) > at org.apache.nutch.crawl.LinkDbReader.run(LinkDbReader.java:127) > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65) > at org.apache.nutch.crawl.LinkDbReader.main(LinkDbReader.java:114) > > This is the first time I use readlinkdb and the rest of the crawling > process is working ok, I've looked up JIRA and there's no related bug. > > I've also tried latest trunk nutch but DFS is not working for me: > > [EMAIL PROTECTED] trunk]$ bin/hadoop dfs -ls > > Exception in thread "main" java.lang.RuntimeException: > java.lang.ClassNotFoundException: > org.apache.hadoop.hdfs.DistributedFileSystem > at > org.apache.hadoop.conf.Configuration.getClass(Configuration.java:648) > at > org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:1334) > at org.apache.hadoop.fs.FileSystem.access$300(FileSystem.java:56) > at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:1351) > at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:213) > at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:118) > at org.apache.hadoop.fs.FsShell.init(FsShell.java:88) > at org.apache.hadoop.fs.FsShell.run(FsShell.java:1698) > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65) > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:79) > at org.apache.hadoop.fs.FsShell.main(FsShell.java:1847) > Caused by: java.lang.ClassNotFoundException: > org.apache.hadoop.hdfs.DistributedFileSystem > at java.net.URLClassLoader$1.run(URLClassLoader.java:200) > at java.security.AccessController.doPrivileged(Native Method) > at java.net.URLClassLoader.findClass(URLClassLoader.java:188) > at java.lang.ClassLoader.loadClass(ClassLoader.java:307) > at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:301) > at java.lang.ClassLoader.loadClass(ClassLoader.java:252) > at java.lang.ClassLoader.loadClassInternal(ClassLoader.java:320) > at java.lang.Class.forName0(Native Method) > at java.lang.Class.forName(Class.java:247) > at > org.apache.hadoop.conf.Configuration.getClassByName(Configuration.java:628) > at > org.apache.hadoop.conf.Configuration.getClass(Configuration.java:646) > ... 10 more > > Should I file both bugs on JIRA ? > This I am not sure, but did you try ant clean; ant? It may be a version mismatch. -- Doğacan Güney