RE: [htdig] htdig

Subject: RE: [htdig] htdig
From: Geoff Hutchison (
Date: Thu Jan 11 2001 - 08:34:56 PST

No regular expressions needed. You can limit URLs based on query patterns
already. See the bad_querystr attribute:

-Geoff Hutchison
Williams Students Online

On Thu, 11 Jan 2001, Richard Bethany wrote:

> Geoff, > > I'm the SysAdmin for our web servers and I'm working with Chuck (who does > the development work) on this problem. Here's the "nuts & bolts" of the > problem. Our entire web server is set up with a menuing system being run > through PHP3. This menuing system basically allows local documents/links to > be reached via a URL off of the PHP3 file. In other words, if I try to > access a particular page it will be accessed as > > > In this scenario the only relevant piece of info is the "i" value; the > remainder of the info simply describes which portions of the menu should be > displayed. What ends up happening is that, for a page with eight(8) main > menu items, 40,320 (8*7*6*5*4*3*2*1) different "hits" show up in htDig for > each link!! I essentially need to exclude any URL where "p" has more than > one value (i.e. - &p=1: is okay, &p=1:2: is not). > > I've looked through the mailing list archives and found a great deal of > discussion on the topic of regular expressions with exclusions and also some > talk of stripping parts of the URL, but I've seen nothing to indicate that > any of this has actually been implemented. Do you know if there is any > implementation of this? If not, I saw a reply to a different problem from > Gilles indicating that the URL::normalizePath() function would be the best > place to start hacking so I guess I'll try that. > > Thanks for your time!!! > Richard

------------------------------------ To unsubscribe from the htdig mailing list, send a message to You will receive a message to confirm this. List archives: <> FAQ: <>

This archive was generated by hypermail 2b28 : Thu Jan 11 2001 - 08:49:07 PST