Thanks.

I do my crawl using the Intranet Recrawl script available in the wiki.
I have put these statements in a loop iterating 10 times.

 1. bin/nutch generate crawl/crawldb crawl/segments -topN 1000
 2. seg1=`ls -d crawl/segments/* | tail -1`
 3. bin/nutch fetch $seg1 -threads 50
 4. bin/nutch updatedb crawl/crawldb $seg1

So, to fetch without parsing I need to modify the statement 3 to:-

bin/nutch fetch $seg1 -threads 50 -noParsing.

Now where do I put this statement:-

bin/nutch parse $seg1

In between statement 3 and statement 4?

On 5/31/07, Vishal Shah <[EMAIL PROTECTED]> wrote:
> Hi Manoharam,
>
>   You can use the parse command to parse a segment after it is fetched with
> -noParsing option. The result will be equivalent to running fetch without
> the noparsing option.
>
>    In your nutch installation directory, try the command bin/nutch. It will
> give you the usage for the parse command.
>
> Regards,
>
> -vishal.
>
> -----Original Message-----
> From: Manoharam Reddy [mailto:[EMAIL PROTECTED]
> Sent: Thursday, May 31, 2007 11:24 AM
> To: [EMAIL PROTECTED]
> Subject: Re: OutOfMemoryError - Why should the while(1) loop stop?
>
> If I run fetcher in non-parsing mode how can I later parse the pages
> so that ultimately when a user searches in the Nutch search engine, he
> can see the content of PDF files, etc as summary? Please help or point
> me to proper articles or wiki where I can learn this.
>
> On 5/30/07, Doğacan Güney <[EMAIL PROTECTED]> wrote:
> > On 5/30/07, Manoharam Reddy <[EMAIL PROTECTED]> wrote:
> > > Time and again I get this error and as a result the segment remains
> > > incomplete. This wastes one iteration of the for() loop in which I am
> > > doing generate, fetch and update.
> > >
> > > Can someone please tell me what are the measures I can take to avoid
> > > this error? And isn't it possible to make some code changes so that
> > > the whole fetch doesn't have to stop suddenly when this error occurs.
> > > Can't we do something in the code so that, the fetch still continues
> > > like in case of SocketException, in which case the fetch while(1) loop
> > > continues.
> > >
> > > If it is not possible, please tell me how can I prevent this error
> > > from happening?
> >
> > Are you also parsing during fetch? If you are, I would suggest running
> > Fetcher in non-parsing mode.
> >
> > >
> > > ----- ERROR -----
> > >
> > > fetch of http://telephony/register.asp failed with:
> > > java.lang.OutOfMemoryError: Java heap space
> > > java.lang.NullPointerException
> > > at
> org.apache.hadoop.fs.FSDataInputStream$Buffer.getPos(FSDataInputStream.java:
> 87)
> > > at
> org.apache.hadoop.fs.FSDataInputStream.getPos(FSDataInputStream.java:125)
> > > ......
> > > at org.apache.nutch.fetcher.Fetcher$FetcherThread.run(Fetcher.java:115)
> > > fetcher caught:java.lang.NullPointerException
> > > java.lang.NullPointerException
> > > at
> org.apache.hadoop.fs.FSDataInputStream$Buffer.getPos(FSDataInputStream.java:
> 87)
> > > at
> org.apache.hadoop.fs.FSDataInputStream.getPos(FSDataInputStream.java:125)
> > > .......
> > > at org.apache.nutch.fetcher.Fetcher$FetcherThread.run(Fetcher.java:115)
> > > fetcher caught:java.lang.NullPointerException
> > > Fetcher: java.io.IOException: Job failed!
> > >   at org.apache.hadoop.mapred.JobClient.runJob(JobClient.java:604)
> > >   at org.apache.nutch.fetcher.Fetcher.fetch(Fetcher.java:470)
> > >   at org.apache.nutch.fetcher.Fetcher.run(Fetcher.java:505)
> > >   at org.apache.hadoop.util.ToolBase.doMain(ToolBase.java:189)
> > >   at org.apache.nutch.fetcher.Fetcher.main(Fetcher.java:477)
> > >
> >
> >
> > --
> > Doğacan Güney
> >
>
>
-------------------------------------------------------------------------
This SF.net email is sponsored by DB2 Express
Download DB2 Express C - the FREE version of DB2 express and take
control of your XML. No limits. Just data. Click to get it now.
http://sourceforge.net/powerbar/db2/
_______________________________________________
Nutch-general mailing list
Nutch-general@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/nutch-general

Reply via email to