gt;
> >>g.
> >>
> >>
> >>Dan Morrill wrote:
> >>
> >>
> >>
> >>>If I remember it correctly, google as been sued and won a number of
> times on
> >>>this issue, you can cache, you can search others web sites, groc
t on the web, it is considered public in
that regard.
I am not a lawyer, check grocklaw.
r/d
-Original Message-
From: TDLN [mailto:[EMAIL PROTECTED]
Sent: Thursday, March 30, 2006 3:34 AM
To: nutch-user@lucene.apache.org
Subject: Re: Legal issues
Google's and Yahoo's Terms of
ng passwords,
> >and honor robots.txt and they post it on the web, it is considered public in
> >that regard.
> >
> >I am not a lawyer, check grocklaw.
> >
> >r/d
> >
> >-Original Message-
> >From: TDLN [mailto:[EMAIL PROTECTED]
> &g
hat regard.
I am not a lawyer, check grocklaw.
r/d
-Original Message-
From: TDLN [mailto:[EMAIL PROTECTED]
Sent: Thursday, March 30, 2006 3:34 AM
To: nutch-user@lucene.apache.org
Subject: Re: Legal issues
Google's and Yahoo's Terms of Service provide interesting reading regarding
suc
issues
Google's and Yahoo's Terms of Service provide interesting reading regarding
such legal issues.
http://www.google.com/terms_of_service.html
http://docs.yahoo.com/info/terms/
Rgrds, Thomas
On 3/30/06, gekkokid <[EMAIL PROTECTED]> wrote:
>
> Shouldn't be a problem if yo
Google's and Yahoo's Terms of Service provide interesting reading regarding
such legal issues.
http://www.google.com/terms_of_service.html
http://docs.yahoo.com/info/terms/
Rgrds, Thomas
On 3/30/06, gekkokid <[EMAIL PROTECTED]> wrote:
>
> Shouldn't be a problem if yo
Shouldn't be a problem if your honouring the robots.txt
Legal issues could be Stealing Copyrighted Material? thats if your
reproducing it but if your analysing the content and links and keeping to
the robots.txt rules I doubt your have a problem unless its crawling every
10 mi
What are say the legal issues of crawling a site like reddit, digg or
slashdot. Assuming that you are just collecting links that users post
through that service and then you are regathering those links. I
can't see an issue there.
The other extreme would be crawling google and requeryi