No regular expressions needed. You can limit URLs based on query patterns
already. See the bad_querystr attribute:
<http://www.htdig.org/attrs.html#bad_querystr>

--
-Geoff Hutchison
Williams Students Online
http://wso.williams.edu/

On Thu, 11 Jan 2001, Richard Bethany wrote:

> Geoff,
> 
> I'm the SysAdmin for our web servers and I'm working with Chuck (who does
> the development work) on this problem.  Here's the "nuts & bolts" of the
> problem.  Our entire web server is set up with a menuing system being run
> through PHP3.  This menuing system basically allows local documents/links to
> be reached via a URL off of the PHP3 file.  In other words, if I try to
> access a particular page it will be accessed as
> http://ourweb.com/DEPT/index.php3?i=1&e=3&p=2:3:4:.
> 
> In this scenario the only relevant piece of info is the "i" value; the
> remainder of the info simply describes which portions of the menu should be
> displayed.  What ends up happening is that, for a page with eight(8) main
> menu items, 40,320 (8*7*6*5*4*3*2*1) different "hits" show up in htDig for
> each link!!  I essentially need to exclude any URL where "p" has more than
> one value (i.e. - &p=1: is okay, &p=1:2: is not).
> 
> I've looked through the mailing list archives and found a great deal of
> discussion on the topic of regular expressions with exclusions and also some
> talk of stripping parts of the URL, but I've seen nothing to indicate that
> any of this has actually been implemented.  Do you know if there is any
> implementation of this?  If not, I saw a reply to a different problem from
> Gilles indicating that the URL::normalizePath() function would be the best
> place to start hacking so I guess I'll try that.
> 
> Thanks for your time!!!
> Richard


------------------------------------
To unsubscribe from the htdig mailing list, send a message to
[EMAIL PROTECTED]
You will receive a message to confirm this.
List archives:  <http://www.htdig.org/mail/menu.html>
FAQ:            <http://www.htdig.org/FAQ.html>

Reply via email to