Hi Jesse,
thanks for spotting this. I fixed the problem in trunk, see
https://issues.apache.org/jira/browse/NUTCH-793
--
Sami Siren
Jesse Hires wrote:
I am seeing the following and am able to find any notes anywhere on it.
org.apache.jasper.JasperException: Unable to compile class for
I am seeing the following and am able to find any notes anywhere on it.
org.apache.jasper.JasperException: Unable to compile class for JSP:
An error occurred at line: 207 in the jsp file: /search.jsp
query.getParams cannot be resolved or is not a field
204: // position this is good,
with the Hadoop configuration or machine setup.
it is not a Nutch issue as such so I'll mark this as invalid.
Running Nutch On linux having whoami exception?
---
Key: NUTCH-778
URL: https://issues.apache.org/jira/browse
Running Nutch On linux having whoami exception?
---
Key: NUTCH-778
URL: https://issues.apache.org/jira/browse/NUTCH-778
Project: Nutch
Issue Type: Bug
Affects Versions: 1.0.0
)
Please tell us: version of nutch
any changes
different configurations (different then crawl-urlfilter - adding your
domain).
Thanks,
Bartosz
Frank McCown pisze:
Adding cygwin to my PATH solved my problem with whoami. But now I'm
getting an exception when running the crawler:
Injector
to my PATH solved my problem with whoami. But now I'm
getting an exception when running the crawler:
Injector: Converting injected urls to crawl db entries.
Exception in thread main java.io.IOException: Job failed!
at org.apache.hadoop.mapred.JobClient.runJob(JobClient.java:1232
any changes
different configurations (different then crawl-urlfilter - adding your
domain).
Thanks,
Bartosz
Frank McCown pisze:
Adding cygwin to my PATH solved my problem with whoami. But now I'm
getting an exception when running the crawler:
Injector: Converting injected urls to crawl
fmccown pisze:
You must run Nutch's crawler using cygwin on Windows since cygwin has the
whoami program. If you run it from Eclipse on Windows, it can't use
cygwin's whoami program and will fail with the exceptions you saw. This is
an unfortunately design decision in Hadoop which makes
I've been running 0.9 in Eclipse on Windows for some time, and I was
successful in running the NutchBean from version 1.0 in Eclipse, but
the crawler gave me the same exception as it gave this individual.
Maybe there's something else I'm overlooking, but I followed the
Tutorial at
http
0.9 in Eclipse on Windows for some time, and I was
successful in running the NutchBean from version 1.0 in Eclipse, but
the crawler gave me the same exception as it gave this individual.
Maybe there's something else I'm overlooking, but I followed the
Tutorial at
http://wiki.apache.org/nutch
-Original Message-
From: Bartosz Gadzimski [mailto:bartek...@o2.pl]
Sent: Friday, April 10, 2009 5:06 AM
To: nutch-dev@lucene.apache.org
Subject: Re: login failed exception
Hello,
I am not sure if it's the case but you should try to add whoami to your
windows box.
for example
without running commandline
cygwin? Is there a way to turn off this security in Hadoop?
Thanks,
Sanjoy
-Original Message-
From: Bartosz Gadzimski [mailto:bartek...@o2.pl]
Sent: Friday, April 10, 2009 5:06 AM
To: nutch-dev@lucene.apache.org
Subject: Re: login failed exception
Hello,
I am
exception
Hello,
So now you have to install cygwin and be sure that you add it to PATH
it's in http://wiki.apache.org/nutch/RunNutchInEclipse0.9
After this you should be able to run bash command from command prompt
(Menu Start RUN cmd.exe)
Then you'r done - everything will be working.
I
changes
different configurations (different then crawl-urlfilter - adding your
domain).
Thanks,
Bartosz
Frank McCown pisze:
Adding cygwin to my PATH solved my problem with whoami. But now I'm
getting an exception when running the crawler:
Injector: Converting injected urls to crawl db entries
and notes with a
JobStatus.FAILED. As Bartosz suggests these are logged as WARNs in hadoop.log.
Sanjoy
-Original Message-
From: Frank McCown [mailto:fmcc...@harding.edu]
Sent: Friday, April 10, 2009 2:29 PM
To: nutch-dev@lucene.apache.org
Subject: Re: login failed exception
Adding cygwin to my
failed exception
Hello Frank,
Please look into hadoop.log and let maybe there is something more.
About your error - you must give us more specific configuration of your
nutch.
Default nutch installation is working with no problems (I'v never
changed src/plugin path)
Please tell us: version
9.0 not work in Eclipse on Windows.
--
View this message in context:
http://www.nabble.com/login-failed-exception-tp21539952p22979522.html
Sent from the Nutch - Dev mailing list archive at Nabble.com.
Hi,
I'm using Nutch 0.9 to crawl part of my intranet, and am getting the
following when attempting to parse ppt files:
2009-03-11 16:30:47,000 ERROR mspowerpoint.ContentReaderListener -
extractClientTextBoxes
java.lang.ArrayIndexOutOfBoundsException: -55133188
at
Hi Pradeep,
I was able to run the nutch 0.9 stable version but its crawling only very
few urls. So i tried with the latest build from the nightly build. I am
getting this login failed exception. I still not able to resolve it...
Vimal Varghese
Pradeep Pujari pradeep.puj...@macys.com
20
Hi,
I have configured the latest nutch from the nightly build in eclipse.
I am getting this following error.
Exception in thread main java.io.IOException: Failed to get the current
user's information.
at org.apache.hadoop.mapred.JobClient.getUGI(JobClient.java:717
Pages with no ContentType cause a Null Pointer exception
Key: NUTCH-531
URL: https://issues.apache.org/jira/browse/NUTCH-531
Project: Nutch
Issue Type: Bug
Components
/JavaDemoApplication
However, I'm getting an exception while creating an instance of the
configuration class:
Configuration nutchConf = NutchConfiguration.create();
the exception is:
Exception occurred in target VM: Could not initialize class
org.apache.hadoop.conf.Configuration
://nmgroup.tsinghua.edu.cn/cn/people_asso.htm
- Original Message -
From: Doun [EMAIL PROTECTED]
To: nutch-dev@lucene.apache.org
Sent: Friday, November 28, 2008 9:52 AM
Subject: Exception in NutchConfiguration class using java servlet
Hi,
I'm pretty newbie in dealing with nutch. I've done the crawling
to Hadoop.
TaskRunner io pipes are not setDaemon(true), cleanup and exception errors are
incomplete, max heap too small
Key: NUTCH-152
URL: https
TaskRunner io pipes are not setDaemon(true), cleanup and exception errors are
incomplete, max heap too small
Key: NUTCH-152
URL: https://issues.apache.org
a couple of months back, and am not using nutch
anymore or following the mailing lists.
Cheers,
Carl.
Pages with no ContentType cause a Null Pointer exception
Key: NUTCH-531
URL: https
with the integration of the framwerok
Tika.
I guess we could close it. Isn't it?
Pages with no ContentType cause a Null Pointer exception
Key: NUTCH-531
URL: https://issues.apache.org/jira/browse/NUTCH-531
Lately I've been getting this error while running Fetcher2:
java.io.EOFException at
java.io.DataInputStream.readFully(DataInputStream.java:178) at
java.io.DataInputStream.readFully(DataInputStream.java:152) at
org.apache.hadoop.io.SequenceFile$Reader.init(SequenceFile.java:1383) at
Null Pointer Exception on getSummary when segment no longer exists
--
Key: NUTCH-542
URL: https://issues.apache.org/jira/browse/NUTCH-542
Project: Nutch
Issue Type: Bug
be enough.
I am attaching a draft patch with the suggested change.
Pages with no ContentType cause a Null Pointer exception
Key: NUTCH-531
URL: https://issues.apache.org/jira/browse/NUTCH-531
Pages with no ContentType cause a Null Pointer exception
Key: NUTCH-531
URL: https://issues.apache.org/jira/browse/NUTCH-531
Project: Nutch
Issue Type: Bug
Components
via Tomcat / nutch-0.9-dev.war raises exception
-
Key: NUTCH-394
URL: http://issues.apache.org/jira/browse/NUTCH-394
Project: Nutch
Issue Type: Bug
Components: searcher, web
via Tomcat / nutch-0.9-dev.war raises exception
-
Key: NUTCH-394
URL: http://issues.apache.org/jira/browse/NUTCH-394
Project: Nutch
Issue Type: Bug
Components: searcher, web
Searching via Tomcat / nutch-0.9-dev.war raises exception
-
Key: NUTCH-394
URL: http://issues.apache.org/jira/browse/NUTCH-394
Project: Nutch
Issue Type: Bug
Components
change to add some junit test along with this ?
http: proxy exception list:
---
Key: NUTCH-208
URL: http://issues.apache.org/jira/browse/NUTCH-208
Project: Nutch
Issue Type: New Feature
Components: fetcher
[ http://issues.apache.org/jira/browse/NUTCH-208?page=all ]
Renaud Richardet updated NUTCH-208:
---
Attachment: proxy_exception_list-0.8.diff
I updated the patch to 0.8 and corrected small typo (if
(!.equals(input[i].trim())){ ). The proxy exception
[ http://issues.apache.org/jira/browse/NUTCH-287?page=all ]
Stefan Groschupf closed NUTCH-287:
--
Resolution: Won't Fix
http://www.mail-archive.com/nutch-user%40lucene.apache.org/msg04696.html
Exception when searching with sort
Exception when searching with sort
--
Key: NUTCH-287
URL: http://issues.apache.org/jira/browse/NUTCH-287
Project: Nutch
Type: Bug
Components: searcher
Versions: 0.8-dev
Reporter: Stefan Neufeind
Priority: Critical
PROTECTED]
Sent: Thursday, April 27, 2006 12:48 AM
To: nutch-dev@lucene.apache.org
Subject: Re: exception
Importance: High
This is a Hadoop DFS error. It could mean that you don't have any
datanodes running, or that all your datanodes are full. Or, it could be
a bug in dfs. You might try a recent
[EMAIL PROTECTED] wrote:
We updated hadoop from trunk branch. But now we get new errors:
Oops. Looks like I introduced a bug yesterday. Let me fix it...
Sorry,
Doug
12:48 AM
To: nutch-dev@lucene.apache.org
Subject: Re: exception
Importance: High
This is a Hadoop DFS error. It could mean that you don't have any
datanodes running, or that all your datanodes are full. Or, it could be
a bug in dfs. You might try a recent nightly build of Hadoop to see
Message-
From: Doug Cutting [mailto:[EMAIL PROTECTED] Sent: Thursday, April
27, 2006 12:48 AM
To: nutch-dev@lucene.apache.org
Subject: Re: exception
Importance: High
This is a Hadoop DFS error. It could mean that you don't have any
datanodes running, or that all your datanodes are full
This is a Hadoop DFS error. It could mean that you don't have any
datanodes running, or that all your datanodes are full. Or, it could be
a bug in dfs. You might try a recent nightly build of Hadoop to see if
it works any better.
Doug
Anton Potehin wrote:
What means error of following
Hi Folks,
I updated to the latest SVN revision (385691) today, and I am now seeing a
Null Pointer exception in the AnalyzerFactory.java class. It seems that in
some cases, the method:
private Extension getExtension(String lang) { Extension extension =
(Extension) this.conf.getObject(lang
I updated to the latest SVN revision (385691) today, and I am now seeing
a
Null Pointer exception in the AnalyzerFactory.java class.
Fixed (r385702). Thanks Chris.
NOTE: not sure if returning null is the right thing to do here, but hey,
at
least it made my crawl finish
[ http://issues.apache.org/jira/browse/NUTCH-91?page=all ]
Piotr Kosiorowski closed NUTCH-91:
--
Fix Version: 0.7.2-dev
0.8-dev
Resolution: Fixed
Commited with small extension. Thanks.
empty encoding causes exception
to execute such a class.
Exception in thread main java.lang.NoClassDefFoundError: invertlink
-
Key: NUTCH-222
URL: http://issues.apache.org/jira/browse/NUTCH-222
Project: Nutch
Type: Bug
That was a typo. Same thing happens with invertlinks.
$ bin/nutch invertlinks taxcrawl/db/ -dir
taxcrawl/segments/20060302033503/
run java in C:\Program Files\Java\jdk1.5.0_04
Exception in thread main java.lang.NoClassDefFoundError: invertlinks
-Original Message-
From: Stefan Groschupf
in 0.8
Exception in thread main java.lang.NoClassDefFoundError: invertlink
-
Key: NUTCH-222
URL: http://issues.apache.org/jira/browse/NUTCH-222
Project: Nutch
Type: Bug
Components: fetcher
0.7
script to running a nutch 0.8 command.
Stefan
Am 04.03.2006 um 17:24 schrieb Richard Braman:
That was a typo. Same thing happens with invertlinks.
$ bin/nutch invertlinks taxcrawl/db/ -dir
taxcrawl/segments/20060302033503/
run java in C:\Program Files\Java\jdk1.5.0_04
Exception in thread
@lucene.apache.org
Subject: Re: [jira] Closed: (NUTCH-222) Exception in thread main
java.lang.NoClassDefFoundError: invertlink
Hi Richard,
I can not reproduce the problem: tom:~/Documents/workspace/nutch-trunk
joa$ bin/nutch invertlinks
Usage: linkdb (-dir segmentsDir | segment1 segment2
can just index your segment.
Stefan
-Original Message-
From: Stefan Groschupf [mailto:[EMAIL PROTECTED]
Sent: Saturday, March 04, 2006 11:47 AM
To: nutch-dev@lucene.apache.org
Subject: Re: [jira] Closed: (NUTCH-222) Exception in thread main
java.lang.NoClassDefFoundError: invertlink
Hi
http: proxy exception list:
Key: NUTCH-208
URL: http://issues.apache.org/jira/browse/NUTCH-208
Project: Nutch
Type: New Feature
Components: fetcher
Versions: 0.8-dev
Reporter: Matthias Günter
Priority: Minor
I
[ http://issues.apache.org/jira/browse/NUTCH-208?page=all ]
Matthias Günter updated NUTCH-208:
--
Attachment: patch.txt
A preliminary patch!!
http: proxy exception list:
---
Key: NUTCH-208
URL: http
[ http://issues.apache.org/jira/browse/NUTCH-208?page=all ]
Matthias Günter updated NUTCH-208:
--
Attachment: patch.txt
A preliminary patch!!
http: proxy exception list:
---
Key: NUTCH-208
URL: http
,
I found the reason of that exception!
If you look into my crawl.log carefully then you notice these lines:
060104 213608 Parsing
[http://220.000.000.001/otd_04_Detailed_Design_Document.doc] with
[EMAIL PROTECTED]
060104 213609 Unable to successfully parse content
http://220.000.000.001
with:
java.lang.ClassCastException: java.util.ArrayList
-Matt Zytaruk
Could you please add a call to printStackTrace() in that catch{}
statement, so that we know where the exception is thrown?
--
Best regards,
Andrzej Bialecki
with:
java.lang.ClassCastException: java.util.ArrayList
-Matt Zytaruk
Could you please add a call to printStackTrace() in that catch{}
statement, so that we know where the exception is thrown?
Matt Zytaruk wrote:
Here you go.
java.lang.ClassCastException: java.util.ArrayList
at org.apache.nutch.parse.ParseData.write(ParseData.java:122)
at org.apache.nutch.parse.ParseImpl.write(ParseImpl.java:51)
at
Hi,
I attached the patch. Please test.
--
Best regards,
Andrzej Bialecki
___. ___ ___ ___ _ _ __
[__ || __|__/|__||\/| Information Retrieval, Semantic Web
___|||__|| \| || | Embedded Unix, System Integration
http://www.sigram.com Contact: info at
So will this throw an exception on older segments? or will it just not
get the correct metadata? I have a lot of older segments I still need to
use.
Thanks for your help.
-Matt Zytaruk
Andrzej Bialecki wrote:
Matt Zytaruk wrote:
Here you go.
java.lang.ClassCastException
and exception errors are
incomplete, max heap too small
Key: NUTCH-152
URL: http://issues.apache.org/jira/browse/NUTCH-152
Project: Nutch
Type: Bug
Lukas Vlcek wrote:
How can I learn that?
What I do is running regular one-step command [/bin/nutch crawl]
In that case your nutch-default.xml / nutch-site.xml decides, there is a
boolean option there. If you didn't change this, then it defaults to
true (i.e. your fetcher is parsing the
Hi,
I found the reason of that exception!
If you look into my crawl.log carefully then you notice these lines:
060104 213608 Parsing
[http://220.000.000.001/otd_04_Detailed_Design_Document.doc] with
[EMAIL PROTECTED]
060104 213609 Unable to successfully parse content
http://220.000.000.001
Yes it was fixed. just update your code from trunk.
On Wed, 2006-01-04 at 08:51 +0100, Andrzej Bialecki wrote:
Lukas Vlcek wrote:
Hi,
I am trying to use the latest nutch-trunk version but I am facing
unexpected Job failed! exception. It seems that all crawling work
has been already done
:
Lukas Vlcek wrote:
Hi,
I am trying to use the latest nutch-trunk version but I am facing
unexpected Job failed! exception. It seems that all crawling work
has been already done but some threads are hunged which results into
exception after some timeout.
This was fixed
Fixed in the copy i run as i've been able to get my
100k pages indexed without getting that error.
-byron
--- Andrzej Bialecki [EMAIL PROTECTED] wrote:
Lukas Vlcek wrote:
Hi,
I am trying to use the latest nutch-trunk version
but I am facing
unexpected Job failed! exception. It seems
,
I am trying to use the latest nutch-trunk version
but I am facing
unexpected Job failed! exception. It seems that
all crawling work
has been already done but some threads are hunged
which results into
exception after some timeout.
This was fixed (or should be fixed
I gave it a next try this night and I still have troubles.
This is the very end of my log (full version is attached) and you can
see another nasty exception:
...
060104 213644 map 100%
060104 213645 Optimizing index.
java.lang.NullPointerException: value cannot be null
Lukas Vlcek wrote:
I gave it a next try this night and I still have troubles.
This is the very end of my log (full version is attached) and you can
see another nasty exception:
Do you use the Fetcher in parsing or non-parsing mode, i.e. do you run a
ParseSegment as a separate step
Hi,
I am trying to use the latest nutch-trunk version but I am facing
unexpected Job failed! exception. It seems that all crawling work
has been already done but some threads are hunged which results into
exception after some timeout.
I am not sure whether this is a real nutch issue or just mine
mail-lists
On 1/4/06, Lukas Vlcek [EMAIL PROTECTED] wrote:
Hi,
I am trying to use the latest nutch-trunk version but I am facing
unexpected Job failed! exception. It seems that all crawling work
has been already done but some threads are hunged which results into
exception after some
Lukas Vlcek wrote:
Hi,
I am trying to use the latest nutch-trunk version but I am facing
unexpected Job failed! exception. It seems that all crawling work
has been already done but some threads are hunged which results into
exception after some timeout.
This was fixed (or should be fixed
as a source change comment.
TaskRunner io pipes are not setDaemon(true), cleanup and exception errors are
incomplete, max heap too small
Key: NUTCH-152
URL: http
[ http://issues.apache.org/jira/browse/NUTCH-89?page=all ]
Piotr Kosiorowski closed NUTCH-89:
--
Fix Version: 0.8-dev
0.7
Resolution: Fixed
Applied in trunk and 0.7 branch. Thanks.
parse-rss null pointer exception
MapFile.Writer throws directory exists exception if run multiple times in the
same JVM or server JVM.
-
Key: NUTCH-96
URL: http://issues.apache.org/jira/browse/NUTCH-96
a folder in NDFS. But
when we try to put a local file to the NDFS through
the NDFS client, we get an exception “Could not obtain
new output block for file”. Sometimes the exception is
“NullPointer at java.net.Socket”. Could you please
share your thoughts on why such an exception could
occur.
Thanks
77 matches
Mail list logo