‑h |
‑‑help |
Print this help message and exit | |
‑‑version |
Print program version and exit | ||
‑i |
‑‑input‑file |
FILE |
Download URLs found in FILE (- for stdin). More than one --input-file can be specified |
‑d |
‑‑destination |
PATH |
Target location for file downloads |
‑D |
‑‑directory |
PATH |
Exact location for file downloads |
‑f |
‑‑filename |
FORMAT |
Filename format string for downloaded files (/O for "original" filenames) |
‑‑proxy |
URL |
Use the specified proxy | |
‑‑source‑address |
IP |
Client-side IP address to bind to | |
‑‑user‑agent |
UA |
User-Agent request header | |
‑‑clear‑cache |
MODULE |
Delete cached login sessions, cookies, etc. for MODULE (ALL to delete everything) | |
‑‑cookies |
FILE |
File to load additional cookies from | |
‑‑cookies‑from‑browser |
BROWSER |
Name of the browser to load cookies from, with optional keyring name prefixed with + , profile prefixed with : , and container prefixed with :: (none for no container) |
‑q |
‑‑quiet |
Activate quiet mode | |
‑v |
‑‑verbose |
Print various debugging information | |
‑g |
‑‑get‑urls |
Print URLs instead of downloading | |
‑G |
‑‑resolve‑urls |
Print URLs instead of downloading; resolve intermediary URLs | |
‑j |
‑‑dump‑json |
Print JSON information | |
‑s |
‑‑simulate |
Simulate data extraction; do not download anything | |
‑E |
‑‑extractor‑info |
Print extractor defaults and settings | |
‑K |
‑‑list‑keywords |
Print a list of available keywords and example values for the given URLs | |
‑‑list‑modules |
Print a list of available extractor modules | ||
‑‑list‑extractors |
Print a list of extractor classes with description, (sub)category and example URL | ||
‑‑write‑log |
FILE |
Write logging output to FILE | |
‑‑write‑unsupported |
FILE |
Write URLs, which get emitted by other extractors but cannot be handled, to FILE | |
‑‑write‑pages |
Write downloaded intermediary pages to files in the current directory to debug problems |
‑r |
‑‑limit‑rate |
RATE |
Maximum download rate (e.g. 500k or 2.5M) |
‑R |
‑‑retries |
N |
Maximum number of retries for failed HTTP requests or -1 for infinite retries (default: 4) |
‑‑http‑timeout |
SECONDS |
Timeout for HTTP connections (default: 30.0) | |
‑‑sleep |
SECONDS |
Number of seconds to wait before each download. This can be either a constant value or a range (e.g. 2.7 or 2.0-3.5) | |
‑‑sleep‑request |
SECONDS |
Number of seconds to wait between HTTP requests during data extraction | |
‑‑sleep‑extractor |
SECONDS |
Number of seconds to wait before starting data extraction for an input URL | |
‑‑filesize‑min |
SIZE |
Do not download files smaller than SIZE (e.g. 500k or 2.5M) | |
‑‑filesize‑max |
SIZE |
Do not download files larger than SIZE (e.g. 500k or 2.5M) | |
‑‑chunk‑size |
SIZE |
Size of in-memory data chunks (default: 32k) | |
‑‑no‑part |
Do not use .part files | ||
‑‑no‑skip |
Do not skip downloads; overwrite existing files | ||
‑‑no‑mtime |
Do not set file modification times according to Last-Modified HTTP response headers | ||
‑‑no‑download |
Do not download any files | ||
‑‑no‑postprocessors |
Do not run any post processors | ||
‑‑no‑check‑certificate |
Disable HTTPS certificate validation |
‑c |
‑‑config |
FILE |
Additional configuration files |
‑o |
‑‑option |
OPT |
Additional <key>=<value> option values |
‑‑ignore‑config |
Do not read default configuration files |
‑u |
‑‑username |
USER |
Username to login with |
‑p |
‑‑password |
PASS |
Password belonging to the given username |
‑‑netrc |
Enable .netrc authentication data |
‑‑download‑archive |
FILE |
Record all downloaded or skipped files in FILE and skip downloading any file already in it | |
‑A |
‑‑abort |
N |
Stop current extractor run after N consecutive file downloads were skipped |
‑T |
‑‑terminate |
N |
Stop current and parent extractor runs after N consecutive file downloads were skipped |
‑‑range |
RANGE |
Index range(s) specifying which files to download. These can be either a constant value, range, or slice (e.g. 5 , 8-20 , or 1:24:3 ) |
|
‑‑chapter‑range |
RANGE |
Like --range , but applies to manga chapters and other delegated URLs |
|
‑‑filter |
EXPR |
Python expression controlling which files to download. Files for which the expression evaluates to False are ignored. Available keys are the filename-specific ones listed by -K . Example: --filter "image_width >= 1000 and rating in ('s', 'q')" |
|
‑‑chapter‑filter |
EXPR |
Like --filter , but applies to manga chapters and other delegated URLs |
‑‑zip |
Store downloaded files in a ZIP archive | ||
‑‑ugoira‑conv |
Convert Pixiv Ugoira to WebM (requires FFmpeg) | ||
‑‑ugoira‑conv‑lossless |
Convert Pixiv Ugoira to WebM in VP9 lossless mode | ||
‑‑ugoira‑conv‑copy |
Convert Pixiv Ugoira to MKV without re-encoding any frames | ||
‑‑write‑metadata |
Write metadata to separate JSON files | ||
‑‑write‑info‑json |
Write gallery metadata to a info.json file | ||
‑‑write‑tags |
Write image tags to separate text files | ||
‑‑mtime‑from‑date |
Set file modification times according to date metadata |
||
‑‑exec |
CMD |
Execute CMD for each downloaded file. Example: --exec "convert {} {}.png && rm {}" |
|
‑‑exec‑after |
CMD |
Execute CMD after all files were downloaded successfully. Example: --exec-after "cd {} && convert * ../doc.pdf" |
|
‑P |
‑‑postprocessor |
NAME |
Activate the specified post processor |