Download Options (** Not implemented in interface **) --bind-address=ADDRESS When making client TCP/IP connections, `bind()' to ADDRESS on the local machine. ADDRESS may be speci- fied as a hostname or IP address. This option can be useful if your machine is bound to multiple IPs. Number of retries --tries=number Set number of retries to number. Specify 0 or inf for infinite retrying. Concatenate files to --output-document=file The documents will not be written to the appropriate files, but all will be concatenated together and writ- ten to file. If file already exists, it will be over- written. If the file is -, the documents will be written to standard output. Including this option automatically sets the number of tries to 1. Do not overwrite existing files --no-clobber If a file is downloaded more than once in the same directory, Wget's behavior depends on a few options, including -nc. In certain cases, the local file will be clobbered, or overwritten, upon repeated download. In other cases it will be preserved. When running Wget without -N, -nc, or -r, downloading the same file in the same directory will result in the original copy of file being preserved and the second copy being named file.1. If that file is downloaded yet again, the third copy will be named file.2, and so on. When -nc is specified, this behavior is sup- pressed, and Wget will refuse to download newer copies of file. Therefore, ```no-clobber''' is actually a misnomer in this mode---it's not clobbering that's prevented (as the numeric suffixes were already pre- venting clobbering), but rather the multiple version saving that's prevented. When running Wget with -r, but without -N or -nc, re- downloading a file will result in the new copy simply overwriting the old. Adding -nc will prevent this behavior, instead causing the original version to be preserved and any newer copies on the server to be ignored. When running Wget with -N, with or without -r, the decision as to whether or not to download a newer copy of a file depends on the local and remote timestamp and size of the file. -nc may not be specified at the same time as -N. Note that when -nc is specified, files with the suf- fixes .html or (yuck) .htm will be loaded from the local disk and parsed as if they had been retrieved from the Web. Resume getting partially downloaded files --continue Continue getting a partially-downloaded file. This is useful when you want to finish up a download started by a previous instance of Wget, or by another program. For instance: wget -c ftp://sunsite.doc.ic.ac.uk/ls-lR.Z If there is a file named ls-lR.Z in the current direc- tory, Wget will assume that it is the first portion of the remote file, and will ask the server to continue the retrieval from an offset equal to the length of the local file. Note that you don't need to specify this option if you just want the current invocation of Wget to retry downloading a file should the connection be lost mid- way through. This is the default behavior. -c only affects resumption of downloads started prior to this invocation of Wget, and whose local files are still sitting around. Without -c, the previous example would just download the remote file to ls-lR.Z.1, leaving the truncated ls-lR.Z file alone. Beginning with Wget 1.7, if you use -c on a non-empty file, and it turns out that the server does not sup- port continued downloading, Wget will refuse to start the download from scratch, which would effectively ruin existing contents. If you really want the down- load to start from scratch, remove the file. Also beginning with Wget 1.7, if you use -c on a file which is of equal size as the one on the server, Wget will refuse to download the file and print an explana- tory message. The same happens when the file is smaller on the server than locally (presumably because it was changed on the server since your last download attempt)---because ``continuing'' is not meaningful, no download occurs. On the other side of the coin, while using -c, any file that's bigger on the server than locally will be considered an incomplete download and only `(length(remote) - length(local))' bytes will be down- loaded and tacked onto the end of the local file. This behavior can be desirable in certain cases---for instance, you can use wget -c to download just the new portion that's been appended to a data collection or log file. However, if the file is bigger on the server because it's been changed, as opposed to just appended to, you'll end up with a garbled file. Wget has no way of verifying that the local file is really a valid prefix of the remote file. You need to be especially careful of this when using -c in conjunction with -r, since every file will be considered as an "incomplete down- load" candidate. Another instance where you'll get a garbled file if you try to use -c is if you have a lame HTTP proxy that inserts a ``transfer interrupted'' string into the local file. In the future a ``rollback'' option may be added to deal with this case. Note that -c only works with FTP servers and with HTTP servers that support the `Range' header. (** Not implemented in interface **) --dot-style=style Set the retrieval style to style. Wget traces the retrieval of each document by printing dots on the screen, each dot representing a fixed amount of retrieved data. Any number of dots may be separated in a cluster, to make counting easier. This option allows you to choose one of the pre-defined styles, determining the number of bytes represented by a dot, the number of dots in a cluster, and the number of dots on the line. With the `default' style each dot represents 1K, there are ten dots in a cluster and 50 dots in a line. The `binary' style has a more ``computer''-like orienta- tion---8K dots, 16-dots clusters and 48 dots per line (which makes for 384K lines). The `mega' style is suitable for downloading very large files---each dot represents 64K retrieved, there are eight dots in a cluster, and 48 dots on each line (so each line con- tains 3M). The `micro' style is exactly the reverse; it is suitable for downloading small files, with 128-byte dots, 8 dots per cluster, and 48 dots (6K) per line. Only retrieve files newer than local files --timestamping Turn on time-stamping. Print responses/headers from FTP/HTTP servers --server-response Print the headers sent by HTTP servers and responses sent by FTP servers. Check for files but do not download (spider) --spider When invoked with this option, Wget will behave as a Web spider, which means that it will not download the pages, just check that they are there. You can use it to check your bookmarks, e.g. with: wget --spider --force-html -i bookmarks.html This feature needs much more work for Wget to get close to the functionality of real WWW spiders. -T seconds (????) --timeout=seconds Set the read timeout to seconds seconds. Whenever a network read is issued, the file descriptor is checked for a timeout, which could otherwise leave a pending connection (uninterrupted read). The default timeout is 900 seconds (fifteen minutes). Setting timeout to 0 will disable checking for timeouts. Please do not lower the default timeout value with this option unless you know what you are doing. Pause between retrievals --wait=seconds Wait the specified number of seconds between the retrievals. Use of this option is recommended, as it lightens the server load by making the requests less frequent. Instead of in seconds, the time can be specified in minutes using the `m' suffix, in hours using `h' suffix, or in days using `d' suffix. Specifying a large value for this option is useful if the network or the destination host is down, so that Wget can wait long enough to reasonably expect the network error to be fixed before the retry. Wait between failed retrievals --waitretry=seconds If you don't want Wget to wait between every retrieval, but only between retries of failed down- loads, you can use this option. Wget will use linear backoff, waiting 1 second after the first failure on a given file, then waiting 2 seconds after the second failure on that file, up to the maximum number of sec- onds you specify. Therefore, a value of 10 will actu- ally make Wget wait up to (1 + 2 + ... + 10) = 55 sec- onds per file. Note that this option is turned on by default in the global wgetrc file. Turn proxy support on --proxy=on/off Turn proxy support on or off. The proxy is on by default if the appropriate environmental variable is defined. Limit download to --quota=quota Specify download quota for automatic retrievals. The value can be specified in bytes (default), kilobytes (with k suffix), or megabytes (with m suffix). Note that quota will never affect downloading a single file. So if you specify wget -Q10k ftp://wuarchive.wustl.edu/ls-lR.gz, all of the ls- lR.gz will be downloaded. The same goes even when several URLs are specified on the command-line. How- ever, quota is respected when retrieving either recur- sively, or from an input file. Thus you may safely type wget -Q2m -i sites---download will be aborted when the quota is exceeded. Setting quota to 0 or to inf unlimits the download Directory Options Do not create directories for downloads --no-directories Do not create a hierarchy of directories when retriev- ing recursively. With this option turned on, all files will get saved to the current directory, without clobbering (if a name shows up more than once, the filenames will get extensions .n). (?? Is this case handled in code ??) --force-directories The opposite of -nd---create a hierarchy of directo- ries, even if one would not have been created other- wise. E.g. wget -x http://fly.srk.fer.hr/robots.txt will save the downloaded file to fly.srk.fer.hr/robots.txt. Do not prefix directories with host name --no-host-directories Disable generation of host-prefixed directories. By default, invoking Wget with -r http://fly.srk.fer.hr/ will create a structure of directories beginning with fly.srk.fer.hr/. This option disables such behavior. Remove N directory components --cut-dirs=number Ignore number directory components. This is useful for getting a fine-grained control over the directory where recursive retrieval will be saved. Take, for example, the directory at ftp://ftp.xemacs.org/pub/xemacs/. If you retrieve it with -r, it will be saved locally under ftp.xemacs.org/pub/xemacs/. While the -nH option can remove the ftp.xemacs.org/ part, you are still stuck with pub/xemacs. This is where --cut-dirs comes in handy; it makes Wget not ``see'' number remote direc- tory components. Here are several examples of how --cut-dirs option works. No options -> ftp.xemacs.org/pub/xemacs/ -nH -> pub/xemacs/ -nH --cut-dirs=1 -> xemacs/ -nH --cut-dirs=2 -> . --cut-dirs=1 -> ftp.xemacs.org/xemacs/ ... If you just want to get rid of the directory struc- ture, this option is similar to a combination of -nd and -P. However, unlike -nd, --cut-dirs does not lose with subdirectories---for instance, with -nH --cut- dirs=1, a beta/ subdirectory will be placed to xemacs/beta, as one would expect. Output directory --directory-prefix=prefix Set directory prefix to prefix. The directory prefix is the directory where all other files and subdirecto- ries will be saved to, i.e. the top of the retrieval tree. The default is . (the current directory).