No regular expressions needed. You can limit URLs based on query patterns
already. See the bad_querystr attribute:
<http://www.htdig.org/attrs.html#bad_querystr>
--
-Geoff Hutchison
Williams Students Online
http://wso.williams.edu/
On Thu, 11 Jan 2001, Richard Bethany wrote:
> Geoff,
>
> I'm the SysAdmin for our web servers and I'm working with Chuck (who does
> the development work) on this problem. Here's the "nuts & bolts" of the
> problem. Our entire web server is set up with a menuing system being run
> through PHP3. This menuing system basically allows local documents/links to
> be reached via a URL off of the PHP3 file. In other words, if I try to
> access a particular page it will be accessed as
> http://ourweb.com/DEPT/index.php3?i=1&e=3&p=2:3:4:.
>
> In this scenario the only relevant piece of info is the "i" value; the
> remainder of the info simply describes which portions of the menu should be
> displayed. What ends up happening is that, for a page with eight(8) main
> menu items, 40,320 (8*7*6*5*4*3*2*1) different "hits" show up in htDig for
> each link!! I essentially need to exclude any URL where "p" has more than
> one value (i.e. - &p=1: is okay, &p=1:2: is not).
>
> I've looked through the mailing list archives and found a great deal of
> discussion on the topic of regular expressions with exclusions and also some
> talk of stripping parts of the URL, but I've seen nothing to indicate that
> any of this has actually been implemented. Do you know if there is any
> implementation of this? If not, I saw a reply to a different problem from
> Gilles indicating that the URL::normalizePath() function would be the best
> place to start hacking so I guess I'll try that.
>
> Thanks for your time!!!
> Richard
------------------------------------
To unsubscribe from the htdig mailing list, send a message to
[EMAIL PROTECTED]
You will receive a message to confirm this.
List archives: <http://www.htdig.org/mail/menu.html>
FAQ: <http://www.htdig.org/FAQ.html>