I GOT IT TO WORK!!! (on windopws) =) =) =)

Thank you everyone :)

Sas

-----Original Message-----
From: Markus Jelsma [mailto:markus.jel...@openindex.io] 
Sent: Wednesday, July 13, 2016 10:44 AM
To: user@nutch.apache.org
Subject: RE: Running into an Issue

Jamal, i really have no idea but it smells like a Windows related problem again.

See:
http://stackoverflow.com/questions/27201505/hadoop-exception-in-thread-main-java-lang-nullpointerexception
http://stackoverflow.com/questions/30379441/mapreduce-development-inside-eclipse-on-windows

Markus
 
 
-----Original message-----
> From:Jamal, Sarfaraz <sarfaraz.ja...@verizonwireless.com.INVALID>
> Sent: Wednesday 13th July 2016 16:09
> To: user@nutch.apache.org
> Subject: RE: Running into an Issue
> 
> Hi Markus,
> 
> So I am now running this on a machine that I have administrative 
> access to And I am getting a different error message:
> 
> Do you (or anyone) have any ideas?
> 
> $ ./nutch inject ../TestCrawl/crawldb ../url [I tried from the bin 
> folder and the root nutch folder[
> 
> Injector: starting at 2016-07-13 10:02:54
> Injector: crawlDb: ../TestCrawl/crawldb
> Injector: urlDir: ../url
> Injector: Converting injected urls to crawl db entries.
> Injector: java.lang.NullPointerException
>         at java.lang.ProcessBuilder.start(Unknown Source)
>         at org.apache.hadoop.util.Shell.runCommand(Shell.java:445)
>         at org.apache.hadoop.util.Shell.run(Shell.java:418)
>         at 
>org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:          
>                      
>650)
>         at org.apache.hadoop.util.Shell.execCommand(Shell.java:739)
>         at org.apache.hadoop.util.Shell.execCommand(Shell.java:722)
>         at 
>org.apache.hadoop.fs.RawLocalFileSystem.setPermission(RawLocalFileSys          
>                      
>tem.java:633)
>         at 
>org.apache.hadoop.fs.FilterFileSystem.setPermission(FilterFileSystem.          
>                      
>java:467)
>         at 
>org.apache.hadoop.fs.ChecksumFileSystem.create(ChecksumFileSystem.jav          
>                      
>a:456)
>         at 
>org.apache.hadoop.fs.ChecksumFileSystem.create(ChecksumFileSystem.jav          
>                      
>a:424)
>         at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:906)
>         at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:887)
>         at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:849)
>         at 
>org.apache.hadoop.fs.FileSystem.createNewFile(FileSystem.java:1149)
>         at 
>org.apache.nutch.util.LockUtil.createLockFile(LockUtil.java:58)
>         at org.apache.nutch.crawl.Injector.inject(Injector.java:357)
>         at org.apache.nutch.crawl.Injector.run(Injector.java:467)
>         at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
>         at org.apache.nutch.crawl.Injector.main(Injector.java:441)
> 
>  
> 
> -----Original Message-----
> From: Markus Jelsma [mailto:markus.jel...@openindex.io]
> Sent: Tuesday, July 12, 2016 4:52 AM
> To: user@nutch.apache.org
> Subject: RE: Running into an Issue
> 
> Hi , there are some Windows API calls in there that i will never understand. 
> Are there some kinds of symlinks you are working with or whatever they are 
> called in Windows? There must be something with Nutch/Hadoop getting access 
> to your disk. Check permissions, disk space and whatever you can think of.
> 
> M.
>  
> -----Original message-----
> > From:Jamal, Sarfaraz <sarfaraz.ja...@verizonwireless.com.INVALID>
> > Sent: Monday 11th July 2016 22:46
> > To: Nutch help <user@nutch.apache.org>
> > Subject: Running into an Issue
> > 
> > So I feel I have made some progress on Nutch
> > 
> > However I am now getting another error which I am having difficulty 
> > navigating through:
> > 
> > bin/nutch inject TestCrawl/crawldb url
> > 
> > produces this below
> > 
> > Do you have to run Cygwin under Admin for it to work?
> > 
> > Injector: Converting injected urls to crawl db entries.
> > Exception in thread "main" java.lang.UnsatisfiedLinkError: 
> >org.apache.hadoop.io.nativeio.NativeIO$Windows.access0(Ljava/lang/Str
> >ing;I)Z
> >         at 
> >org.apache.hadoop.io.nativeio.NativeIO$Windows.access0(Native Method)
> >         at 
> >org.apache.hadoop.io.nativeio.NativeIO$Windows.access(NativeIO.java:5
> >70)
> >         at org.apache.hadoop.fs.FileUtil.canRead(FileUtil.java:977)
> >         at 
> >org.apache.hadoop.util.DiskChecker.checkAccessByFileMethods(DiskCheck
> >er.java:173)
> >         at 
> >org.apache.hadoop.util.DiskChecker.checkDirAccess(DiskChecker.java:16
> >0)
> >         at 
> >org.apache.hadoop.util.DiskChecker.checkDir(DiskChecker.java:94)
> >         at 
> >org.apache.hadoop.fs.LocalDirAllocator$AllocatorPerContext.confChange
> >d(LocalDirAllocator.java:285)
> >         at 
> >org.apache.hadoop.fs.LocalDirAllocator$AllocatorPerContext.getLocalPa
> >thForWrite(LocalDirAllocator.java:344)
> >         at 
> >org.apache.hadoop.fs.LocalDirAllocator.getLocalPathForWrite(LocalDirA
> >llocator.java:150)
> >         at 
> >org.apache.hadoop.fs.LocalDirAllocator.getLocalPathForWrite(LocalDirA
> >llocator.java:131)
> >         at 
> >org.apache.hadoop.fs.LocalDirAllocator.getLocalPathForWrite(LocalDirA
> >llocator.java:115)
> >         at 
> >org.apache.hadoop.mapred.LocalDistributedCacheManager.setup(LocalDist
> >ributedCacheManager.java:131)
> >         at 
> >org.apache.hadoop.mapred.LocalJobRunner$Job.<init>(LocalJobRunner.jav
> >a:163)
> >         at 
> >org.apache.hadoop.mapred.LocalJobRunner.submitJob(LocalJobRunner.java
> >:731)
> >         at 
> >org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitt
> >er.java:432)
> >         at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1285)
> >         at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1282)
> >         at java.security.AccessController.doPrivileged(Native 
> >Method)
> >         at javax.security.auth.Subject.doAs(Unknown Source)
> >         at 
> >org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInforma
> >tion.java:1548)
> >         at org.apache.hadoop.mapreduce.Job.submit(Job.java:1282)
> >         at 
> >org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:1303)
> >         at org.apache.nutch.crawl.Injector.inject(Injector.java:376)
> >         at org.apache.nutch.crawl.Injector.run(Injector.java:467)
> >         at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
> >         at org.apache.nutch.crawl.Injector.main(Injector.java:441)
> > 
> 

Reply via email to