[dan@dan ~]$ wget --help
GNU Wget 1.9+cvs-stable (Red Hat modified), un recuperador por red no interactivo.
Modo de empleo: wget [OPCIÓN]... [URL]...
Los argumentos obligatorios para las opciones largas son también obligatorios
para las opciones cortas.
Inicio:
-V, --version muestra la versión de wget y termina.
-h, --help muestra esta ayuda.
-b, --background pasa a segundo plano al iniciar.
-e, --execute=ORDEN ejecuta una orden como las de `.wgetrc'.
Logging and input file:
-o, --output-file=FILE log messages to FILE.
-a, --append-output=FILE append messages to FILE.
-d, --debug print debug output.
-q, --quiet quiet (no output).
-v, --verbose be verbose (this is the default).
-nv, --non-verbose turn off verboseness, without being quiet.
-i, --input-file=FILE download URLs found in FILE.
-F, --force-html treat input file as HTML.
-B, --base=URL prepends URL to relative links in -F -i file.
Download:
-t, --tries=NUMBER set number of retries to NUMBER (0 unlimits).
--retry-connrefused retry even if connection is refused.
-O --output-document=FILE write documents to FILE.
-nc, --no-clobber don't clobber existing files or use .# suffixes.
-c, --continue resume getting a partially-downloaded file.
--progress=TYPE select progress gauge type.
-N, --timestamping don't re-retrieve files unless newer than local.
-S, --server-response print server response.
--spider don't download anything.
-T, --timeout=SECONDS set all timeout values to SECONDS.
--dns-timeout=SECS set the DNS lookup timeout to SECS.
--connect-timeout=SECS set the connect timeout to SECS.
--read-timeout=SECS set the read timeout to SECS.
-w, --wait=SECONDS wait SECONDS between retrievals.
--waitretry=SECONDS wait 1...SECONDS between retries of a retrieval.
--random-wait wait from 0...2*WAIT secs between retrievals.
-Y, --proxy=on/off turn proxy on or off.
-Q, --quota=NUMBER set retrieval quota to NUMBER.
--bind-address=ADDRESS bind to ADDRESS (hostname or IP) on local host.
--limit-rate=RATE limit download rate to RATE.
--dns-cache=off disable caching DNS lookups.
--restrict-file-names=OS restrict chars in file names to ones OS allows.
Directories:
-nd, --no-directories don't create directories.
-x, --force-directories force creation of directories.
-nH, --no-host-directories don't create host directories.
-P, --directory-prefix=PREFIX save files to PREFIX/...
--cut-dirs=NUMBER ignore NUMBER remote directory components.
HTTP options:
--http-user=USER set http user to USER.
--http-passwd=PASS set http password to PASS.
-C, --cache=on/off (dis)allow server-cached data (normally allowed).
-E, --html-extension save all text/html documents with .html extension.
--ignore-length ignore `Content-Length' header field.
--header=STRING insert STRING among the headers.
--proxy-user=USER set USER as proxy username.
--proxy-passwd=PASS set PASS as proxy password.
--referer=URL include `Referer: URL' header in HTTP request.
-s, --save-headers save the HTTP headers to file.
-U, --user-agent=AGENT identify as AGENT instead of Wget/VERSION.
--no-http-keep-alive disable HTTP keep-alive (persistent connections).
--cookies=off don't use cookies.
--load-cookies=FILE load cookies from FILE before session.
--save-cookies=FILE save cookies to FILE after session.
--post-data=STRING use the POST method; send STRING as the data.
--post-file=FILE use the POST method; send contents of FILE.
HTTPS (SSL) options:
--sslcertfile=FILE optional client certificate.
--sslcertkey=KEYFILE optional keyfile for this certificate.
--egd-file=FILE file name of the EGD socket.
--sslcadir=DIR dir where hash list of CA's are stored.
--sslcafile=FILE file with bundle of CA's
--sslcerttype=0/1 Client-Cert type 0=PEM (default) / 1=ASN1 (DER)
--sslcheckcert=0/1 Check the server cert agenst given CA
--sslprotocol=0-3 choose SSL protocol; 0=automatic,
1=SSLv2 2=SSLv3 3=TLSv1
FTP options:
-nr, --dont-remove-listing don't remove `.listing' files.
-g, --glob=on/off turn file name globbing on or off.
--passive-ftp use the "passive" transfer mode (default).
--active-ftp use the "active" transfer mode.
--retr-symlinks when recursing, get linked-to files (not dirs).
Recursive retrieval:
-r, --recursive recursive download.
-l, --level=NUMBER maximum recursion depth (inf or 0 for infinite).
--delete-after delete files locally after downloading them.
-k, --convert-links convert non-relative links to relative.
-K, --backup-converted before converting file X, back up as X.orig.
-m, --mirror shortcut option equivalent to -r -N -l inf -nr.
-p, --page-requisites get all images, etc. needed to display HTML page.
--strict-comments turn on strict (SGML) handling of HTML comments.
Aceptar/rechazar recursivo:
-A, --accept=LISTA lista, separada por comas, de extensiones
aceptadas.
-R, --reject=LISTA lista, separada por comas, de extensiones
rechazadas.
-D, --domains=LISTA lista, separada por comas, de dominios
aceptados.
--exclude-domains=LISTA lista de dominios rechazados, separados
por comas.
--follow-ftp sigue enlaces a FTP desde documentos HTML.
--follow-tags=LISTA lista de etiquetas HTML a las que se sigue.
-G, --ignore-tags=LISTA lista de etiquetas HTML que se ignorarán.
-H, --span-hosts ir a anfitriones distintos en la recursión.
-L, --relative sigue sólo enlaces relativos.
-I, --include-directories=LISTA lista de directorios permitidos.
-X, --exclude-directories=LISTA lista de directorios excluidos.
-np, --no-parent no asciende al directorio padre.
Envíe información sobre errores y sugerencias a <
[email protected]>.
[dan@dan ~]$