ht://Dig Copyright © 1995-2002 The ht://Dig Group
Please see the file COPYING for license information.


htdig [options]


Htdig retrieves HTML documents using the HTTP protocol and gathers information from these documents which can later be used to search these documents. This program can be referred to as the search robot.


Use alternate work files. Tells htdig to append .work to database files, causing a second copy of the database to be built. This allows the original files to be used by htsearch during the indexing run. When used without the "-i" flag for an update dig, htdig will use any existing .work files for the databases to update.
-c configfile
Use the specified configfile file instead of the default.
-h maxhops
Restrict the dig to documents that are at most maxhops links away from the starting document. This only works if -i is also given.
Initial. Do not use any old databases. This is accomplished by first erasing the databases.
-m [url_file]
Minimal. Only index the URLs in the file provided and no others. The url_file can be a "-", causing htdig to read the URLs from the STDIN.
Print statistics about the dig after completion.
Create an ASCII version of the document database. This database is easy to parse with other programs so that information can be extracted from it for purposes other than searching. One could gather some interesting statistics from this database.

Each line in the file starts with the document id followed by a list of \tfieldname:value. The fields always appear in the order listed below:

aState (0 = normal, 1 = not found, 2 = not indexed, 3 = obsolete)
mLast modification time as reported by the server
sSize in bytes
hMeta description
lTime of last retrieval
LCount of the links in the document (outgoing links)
bCount of the links to the document (incoming links or backlinks)
cHopCount of this document
gSignature of the document used for duplicate-detection
eE-mail address to use for a notification message from htnotify
nDate to send out a notification e-mail message
SSubject for a notification e-mail message
dThe text of links pointing to this document. (e.g. <a href="docURL">description</a>)
AAnchors in the document (i.e. <A NAME=...)
-u username:password
Tells htdig to send the supplied username and password with each HTTP request. The credentials will be encoded using the 'Basic' authentication scheme. There HAS to be a colon (:) between the username and password.
Verbose mode. This increases the verbosity of the program. Using more than 2 is probably only useful for debugging purposes. The default verbose mode (using only one -v) gives a nice progress report while digging. This progress report can be a bit cryptic, so here is a brief explanation. A line is shown for each URL, with 3 numbers before the URL and some symbols after the URL. The first number is the number of documents parsed so far, the second is the DocID for this document, and the third is the hop count of the document (number of hops from one of the start_url documents). After the URL, it shows a "*" for a link in the document that it already visited, a "+" for a new link it just queued, and a "-" for a link it rejected for any of a number of reasons. To find out what those reasons are, you need to run htdig with at least 3 -v options, i.e. -vvv. If there are no "*", "+" or "-" symbols after the URL, it doesn't mean the document was not parsed or was empty, but only that no links to other documents were found within it. With more verbose output, these symbols will get interspersed in several lines of debugging output.
url_file (at end of arguments, after options)
Get the list URLs to start indexing from the file provided. This will override the default start_url. The url_file can be a "-", causing htdig to read the URLs from the STDIN.


The default configuration file.

See Also

htmerge, htsearch, Configuration file format, and A Standard for Robot Exclusion.

Last modified: $Date: 2002/02/08 21:33:19 $
SourceForge Logo