Available command options

The table that follows list of all of the available command options:

Command Option1 Description1
-A --accept=<LIST> Download only files with the extensions or patterns specified as the <LIST> variable. Separate multiple entries with a whitespace.
-B --base=<URL> Prepend any relative links in an input file with a specific URL (i.e., as defined for the <URL> variable).
-D --domains=<LIST> Follow only domains specified as the <LIST> variable. Separate multiple entries with a whitespace.
-F --force-html Regard an input file to be HTML.
-g --glob=<on/off> Enable or disable globbing to allow or disallow special wildcard characters.
-G --ignore-tags=<LIST> Follow all of the default HTML tags that are normally followed during recursion (see the -r -- recursive option below) except those specified as the <LIST> variable. Separate multiple entries with a whitespace.
-h --help Display help information
-H --span-<hosts> Allow recursion to move to other hosts, by providing the URLs as the<hosts> variable, separating multiple entries with a whitespace. This command must be used with the -D --domains=<UST> option.
-i --input-file=<FILE> Get the list of URLs to download from a named file (input the complete path and file name for this file as the <FILE> variable).
-I --include-directories=<LIST> Follow only directories specified as the <LIST> variable. Separate multiple entries with a whitespace.
-k --convert-links Change absolute hyperlinks to relative.
-l --level=<#> Limit recursion depth to a specific number of levels, by setting the <#> variable to the desired number. The maximum recursion depth is 10. Recursion limits apply to directory depth and following links.
-L --relative Follow only relative links.
-m --mirror Enable options necessary to perform mirroring.
-nd --no-directories When performing recursive downloads, do not recreate the site’s directory hierarchy structure; instead, copy all files to the working directory.
-nH --no-host-directories Do not include a hostname directory in the hierarchy.
-N --timestamping Only include files if they are newer than any files that were previously downloaded.
-o --output-file=<FILE> Send operation information to a specified file instead of to the standard output. Include the desired path and filename for this file as the <FILE> variable.
-O ---output-document=<FILE> Do not download files, but concatenate their contents and write them to a specific file. Include the desired path and filename for this file as the <FILE> variable.
-p --page-requisites In addition to the specified HTML page, also download any other associated files that may be required to properly display the page (for example, image files, CSS files, etc.)
-P -- directory-prefix=<PREFIX> Download all files and subdirectories to a directory called <PREFIX>.
-q --quiet Do not display the operation’s step-by-step execution.
-Q --quota=<#> Set a byte-count limit as the <#> variable for downloading multiple files recursively, or from an input file (suffix with k for kilobytes or m for megabytes. For example, 20m for 20 megabytes).
-r --recursive Include this option to perform the download recursively (for example, links within the targeted page will be followed and their associated content/pages will also be downloaded.)

See Considerations when using the Site Snapshot Tool for recursion considerations.

Depth limits:
  • Default recursion depth: If omitted, the default recursion depth is set to 10.
  • Maximum recursion depth: 10

By default, if you use recursion the following tags/attributes will be followed:

a/href

applet/code

area/href

bgsound/src

body/background

embed/href

embed/src

fig/src

frame/src

iframe/src

img/href

img/lowsrc

img/src

input/src

layer/src

link/href

overlay/src

script/src

table/background

td/background

th/background

base/href

-R --reject=<LIST> Download all files except those with the extensions or patterns specified as the <LIST> variable. Separate multiple entries with a whitespace.
-S --server-response Display sent HTTP server headers and FTP server responses.
-t --tries=<#> Perform a specified <#> of attempts to download each URL (20 is the default; and 0 - zero - can be set to make unlimited retries).
-T --timeout=<#> Do not allow DNS lookups, connection attempts, and/or read idle times to exceed a specific <#> of seconds.
-v --verbose Display the operation’s execution step by step (this is implied when using the sst command).
-w --wait=<#> At the end of a file retrieval, wait a specified <#> of seconds before retrieving the next file.
-x --force-directories Re-create the directory hierarchy, regardless of whether one normally would be created.
-X --exclude-directories=<LIST> Follow all directories except those specified as the <LIST> variable. Separate multiple entries with a whitespace.
-z --convert-absolute Change relative hyperlinks to absolute.
--exclude-domains=<LIST> Follow all domains except those specified as the <LIST> variable. Separate multiple entries with a whitespace.
--follow-ftp Do not ignore FTP links within HTML pages.
--follow-tags=<LIST> Follow only a subset of default HTML tags that when recursing, by specifying the desired tags as the <LIST> variable (please see the -r --recursive option above).
--header=<STRING> Specify a desired <STRING> to be included with HTTP requests’ headers.
--http-user=<USER> Specify a username as the <USER> variable for access to the HTTP server (if applicable).
--http-passwd=<PASS> Set the <PASS> variable to the password associated with the username defined via the — http-user=<USER> command for access to the HTTP server (if applicable).
--ignore-robots Do not honor the robot.txt file or the robots metatag.
--limit-rate=<RATE> Do not download faster than <RATE> (suffix with k for kilobytes/second or m for megabytes/second. For example, 1m for a one megabyte maximum rate).
--load-cookies=<FILE> Prior to the first download, load the cookies contained in a specific file (i.e., by setting the complete path and filename to the desired file as the <FILE> variable). The cookie file must follow a specific format:

domain

ignore

path

secure

expires

name

--no-clobber Do not download a file if it already exists in the working directory.
--no-http-keep-alive Disable the persistent connection feature.
--no-parent When using recursion (-r --recursive), never ascend to the starting point’s parent directory.
--passive-ftp Use passive mode to require the client to start communications with the server.
--random-wait If a file fails to download, wait either OxWAIT, IxWAIT, and 2xWAIT, determined randomly, before reattempting the download (“WAIT” is the <SECONDS> variable set with the option, -waitretry=<SECONDS>).
--retr-symlinks Ignore symbolic links when performing recursive download (-r --recursive), and download the link targets instead, unless the target is a directory.
--save-cookies=<FILE> Before quitting the session, save all of the valid cookies to a specific file (i.e., by setting the complete path and filename to the desired file as the <FILE> variable). The cookie file follows a specific format:

domain

ignore

path

secure

expires

name

--spider Check for the presence of files without actually downloading them.
--upload-quota=<QUOTA> Override the default 50,000 file limit for this operation only, and set the new limit to <QUOTA> (for example, --upload-quota=100000).
--waitretry=<SECONDS> If a file fails to download, reattempt after one (1) second; if it again fails to download, wait two seconds and try again, and so on until <SECONDS> between attempts is reached and then stop.
1 Values revealed as <value> represent variables that need to be input. The “< >” are not required (and SHOULD NOT BE INCLUDED) in the input.