Download Options

(** Not implemented in interface **)
--bind-address=ADDRESS
	 When making client TCP/IP connections, `bind()' to
	 ADDRESS on the local machine.  ADDRESS may be speci-
	 fied as a hostname or IP address.  This option can be
	 useful if your machine is bound to multiple IPs.

Number of retries
--tries=number
	 Set number of retries to number.  Specify 0 or inf for
	 infinite retrying.

Concatenate files to
--output-document=file
	 The documents will not be written to the appropriate
	 files, but all will be concatenated together and writ-
	 ten to file.  If file already exists, it will be over-
	 written.  If the file is -, the documents will be
	 written to standard output.  Including this option
	 automatically sets the number of tries to 1.

Do not overwrite existing files
--no-clobber
	 If a file is downloaded more than once in the same
	 directory, Wget's behavior depends on a few options,
	 including -nc.  In certain cases, the local file will
	 be clobbered, or overwritten, upon repeated download.
	 In other cases it will be preserved.

	 When running Wget without -N, -nc, or -r, downloading
	 the same file in the same directory will result in the
	 original copy of file being preserved and the second
	 copy being named file.1.  If that file is downloaded
	 yet again, the third copy will be named file.2, and so
	 on.  When -nc is specified, this behavior is sup-
	 pressed, and Wget will refuse to download newer copies
	 of file.  Therefore, ```no-clobber''' is actually a
	 misnomer in this mode---it's not clobbering that's
	 prevented (as the numeric suffixes were already pre-
	 venting clobbering), but rather the multiple version
	 saving that's prevented.

	 When running Wget with -r, but without -N or -nc, re-
	 downloading a file will result in the new copy simply
	 overwriting the old.  Adding -nc will prevent this
	 behavior, instead causing the original version to be
	 preserved and any newer copies on the server to be
	 ignored.

	 When running Wget with -N, with or without -r, the
	 decision as to whether or not to download a newer copy
	 of a file depends on the local and remote timestamp
	 and size of the file.  -nc may not be specified at the
	 same time as -N.

	 Note that when -nc is specified, files with the suf-
	 fixes .html or (yuck) .htm will be loaded from the
	 local disk and parsed as if they had been retrieved
	 from the Web.

Resume getting partially downloaded files
--continue
	 Continue getting a partially-downloaded file.  This is
	 useful when you want to finish up a download started
	 by a previous instance of Wget, or by another program.
	 For instance:

					 wget -c ftp://sunsite.doc.ic.ac.uk/ls-lR.Z

	 If there is a file named ls-lR.Z in the current direc-
	 tory, Wget will assume that it is the first portion of
	 the remote file, and will ask the server to continue
	 the retrieval from an offset equal to the length of
	 the local file.

	 Note that you don't need to specify this option if you
	 just want the current invocation of Wget to retry
	 downloading a file should the connection be lost mid-
	 way through.  This is the default behavior.  -c only
	 affects resumption of downloads started prior to this
	 invocation of Wget, and whose local files are still
	 sitting around.

	 Without -c, the previous example would just download
	 the remote file to ls-lR.Z.1, leaving the truncated
	 ls-lR.Z file alone.

	 Beginning with Wget 1.7, if you use -c on a non-empty
	 file, and it turns out that the server does not sup-
	 port continued downloading, Wget will refuse to start
	 the download from scratch, which would effectively
	 ruin existing contents.  If you really want the down-
	 load to start from scratch, remove the file.

	 Also beginning with Wget 1.7, if you use -c on a file
	 which is of equal size as the one on the server, Wget
	 will refuse to download the file and print an explana-
	 tory message.  The same happens when the file is
	 smaller on the server than locally (presumably because
	 it was changed on the server since your last download
	 attempt)---because ``continuing'' is not meaningful,
	 no download occurs.

	 On the other side of the coin, while using -c, any
	 file that's bigger on the server than locally will be
	 considered an incomplete download and only
	 `(length(remote) - length(local))' bytes will be down-
	 loaded and tacked onto the end of the local file.
	 This behavior can be desirable in certain cases---for
	 instance, you can use wget -c to download just the new
	 portion that's been appended to a data collection or
	 log file.

	 However, if the file is bigger on the server because
	 it's been changed, as opposed to just appended to,
	 you'll end up with a garbled file.  Wget has no way of
	 verifying that the local file is really a valid prefix
	 of the remote file.  You need to be especially careful
	 of this when using -c in conjunction with -r, since
	 every file will be considered as an "incomplete down-
	 load" candidate.

	 Another instance where you'll get a garbled file if
	 you try to use -c is if you have a lame HTTP proxy
	 that inserts a ``transfer interrupted'' string into
	 the local file.  In the future a ``rollback'' option
	 may be added to deal with this case.

	 Note that -c only works with FTP servers and with HTTP
	 servers that support the `Range' header.

(** Not implemented in interface **)
--dot-style=style
	 Set the retrieval style to style.  Wget traces the
	 retrieval of each document by printing dots on the
	 screen, each dot representing a fixed amount of
	 retrieved data.  Any number of dots may be separated
	 in a cluster, to make counting easier.  This option

	 allows you to choose one of the pre-defined styles,
	 determining the number of bytes represented by a dot,
	 the number of dots in a cluster, and the number of
	 dots on the line.

	 With the `default' style each dot represents 1K, there
	 are ten dots in a cluster and 50 dots in a line.  The
	 `binary' style has a more ``computer''-like orienta-
	 tion---8K dots, 16-dots clusters and 48 dots per line
	 (which makes for 384K lines).  The `mega' style is
	 suitable for downloading very large files---each dot
	 represents 64K retrieved, there are eight dots in a
	 cluster, and 48 dots on each line (so each line con-
	 tains 3M).  The `micro' style is exactly the reverse;
	 it is suitable for downloading small files, with
	 128-byte dots, 8 dots per cluster, and 48 dots (6K)
	 per line.

Only retrieve files newer than local files
--timestamping
	 Turn on time-stamping.

Print responses/headers from FTP/HTTP servers
--server-response
	 Print the headers sent by HTTP servers and responses
	 sent by FTP servers.

Check for files but do not download (spider)
--spider
	 When invoked with this option, Wget will behave as a
	 Web spider, which means that it will not download the
	 pages, just check that they are there.  You can use it
	 to check your bookmarks, e.g. with:

					 wget --spider --force-html -i bookmarks.html

	 This feature needs much more work for Wget to get
	 close to the functionality of real WWW spiders.

-T seconds

(????)
--timeout=seconds
	 Set the read timeout to seconds seconds.  Whenever a
	 network read is issued, the file descriptor is checked
	 for a timeout, which could otherwise leave a pending
	 connection (uninterrupted read).  The default timeout
	 is 900 seconds (fifteen minutes).  Setting timeout to
	 0 will disable checking for timeouts.

	 Please do not lower the default timeout value with
	 this option unless you know what you are doing.

Pause between retrievals
--wait=seconds
	 Wait the specified number of seconds between the
	 retrievals.  Use of this option is recommended, as it
	 lightens the server load by making the requests less
	 frequent.  Instead of in seconds, the time can be
	 specified in minutes using the `m' suffix, in hours
	 using `h' suffix, or in days using `d' suffix.

	 Specifying a large value for this option is useful if
	 the network or the destination host is down, so that
	 Wget can wait long enough to reasonably expect the
	 network error to be fixed before the retry.

Wait between failed retrievals
--waitretry=seconds
	 If you don't want Wget to wait between every
	 retrieval, but only between retries of failed down-
	 loads, you can use this option.  Wget will use linear
	 backoff, waiting 1 second after the first failure on a
	 given file, then waiting 2 seconds after the second
	 failure on that file, up to the maximum number of sec-
	 onds you specify.  Therefore, a value of 10 will actu-
	 ally make Wget wait up to (1 + 2 + ... + 10) = 55 sec-
	 onds per file.

	 Note that this option is turned on by default in the
	 global wgetrc file.

Turn proxy support on
--proxy=on/off
	 Turn proxy support on or off.  The proxy is on by
	 default if the appropriate environmental variable is
	 defined.

Limit download to
--quota=quota
	 Specify download quota for automatic retrievals.  The
	 value can be specified in bytes (default), kilobytes
	 (with k suffix), or megabytes (with m suffix).

	 Note that quota will never affect downloading a single
	 file.  So if you specify wget -Q10k
	 ftp://wuarchive.wustl.edu/ls-lR.gz, all of the ls-
	 lR.gz will be downloaded.  The same goes even when
	 several URLs are specified on the command-line.  How-
	 ever, quota is respected when retrieving either recur-
	 sively, or from an input file.  Thus you may safely
	 type wget -Q2m -i sites---download will be aborted
	 when the quota is exceeded.

	 Setting quota to 0 or to inf unlimits the download

Directory Options


Do not create directories for downloads 
--no-directories
	 Do not create a hierarchy of directories when retriev-
	 ing recursively.  With this option turned on, all
	 files will get saved to the current directory, without
	 clobbering (if a name shows up more than once, the
	 filenames will get extensions .n).

(?? Is this case handled in code ??)
--force-directories
	 The opposite of -nd---create a hierarchy of directo-
	 ries, even if one would not have been created other-
	 wise.  E.g. wget -x http://fly.srk.fer.hr/robots.txt
	 will save the downloaded file to
	 fly.srk.fer.hr/robots.txt.

Do not prefix directories with host name
--no-host-directories
	 Disable generation of host-prefixed directories.  By
	 default, invoking Wget with -r http://fly.srk.fer.hr/
	 will create a structure of directories beginning with
	 fly.srk.fer.hr/.  This option disables such behavior.

Remove N directory components
--cut-dirs=number
	 Ignore number directory components.  This is useful
	 for getting a fine-grained control over the directory
	 where recursive retrieval will be saved.

	 Take, for example, the directory at
	 ftp://ftp.xemacs.org/pub/xemacs/.  If you retrieve it
	 with -r, it will be saved locally under
	 ftp.xemacs.org/pub/xemacs/.  While the -nH option can
	 remove the ftp.xemacs.org/ part, you are still stuck
	 with pub/xemacs.  This is where --cut-dirs comes in
	 handy; it makes Wget not ``see'' number remote direc-
	 tory components.  Here are several examples of how
	 --cut-dirs option works.

					 No options        -> ftp.xemacs.org/pub/xemacs/
					 -nH               -> pub/xemacs/
					 -nH --cut-dirs=1  -> xemacs/
					 -nH --cut-dirs=2  -> .

					 --cut-dirs=1      -> ftp.xemacs.org/xemacs/
					 ...

	 If you just want to get rid of the directory struc-
	 ture, this option is similar to a combination of -nd
	 and -P.  However, unlike -nd, --cut-dirs does not lose
	 with subdirectories---for instance, with -nH --cut-
	 dirs=1, a beta/ subdirectory will be placed to
	 xemacs/beta, as one would expect.

Output directory
--directory-prefix=prefix
	 Set directory prefix to prefix.  The directory prefix
	 is the directory where all other files and subdirecto-
	 ries will be saved to, i.e. the top of the retrieval
	 tree.  The default is . (the current directory).