S. Hayles (email@example.com)
Tue, 6 Jul 1999 16:42:51 +0100 (BST)
This is with ht://Dig 3.1.2 under IRIX 6.5
I used a script to build a list of all files on our server that were not
externally accessible - so they could be excluded from the externally
accessible index. It ended up ~250Kb with ~5000 entries. I wasn't too
surprised that it didn't seem to work.
On a quick examination, the only limit in this area I could find was the
buffer length in Configuration::Read - but increasing this didn't seem to
help. I tried using robots.txt to restrict indexing, and once max_doc_size
was adjusted this worked fine - but it seems an unwieldy solution.
Since they both appear to use StringMatch for comparisons I would have
expected exclude_urls to work if robots.txt works. Has anyone else had
problems with exclude_urls?
Steven Hayles - Computer Systems Developer, firstname.lastname@example.org
Learning Technology Group, Computer Centre, University of
Leicester, University Rd, Leicester, LE1 7RH
Fax (0/+44)116 2522198 WWW <URL:http://www.le.ac.uk/home/sh23>
To unsubscribe from the htdig3-dev mailing list, send a message to
email@example.com containing the single word "unsubscribe" in
the SUBJECT of the message.
This archive was generated by hypermail 2.0b3 on Tue Jul 06 1999 - 07:57:29 PDT