Skip to content

Latest commit

 

History

History
134 lines (124 loc) · 7.91 KB

options.md

File metadata and controls

134 lines (124 loc) · 7.91 KB

Command-Line Options

General Options:

-h, --help                  Print this help message and exit
--version                   Print program version and exit
-i, --input-file FILE       Download URLs found in FILE ('-' for stdin).
                            More than one --input-file can be specified
-f, --filename FORMAT       Filename format string for downloaded files
                            ('/O' for "original" filenames)
-d, --destination PATH      Target location for file downloads
-D, --directory PATH        Exact location for file downloads
-X, --extractors PATH       Load external extractors from PATH
--proxy URL                 Use the specified proxy
--source-address IP         Client-side IP address to bind to
--user-agent UA             User-Agent request header
--clear-cache MODULE        Delete cached login sessions, cookies, etc. for
                            MODULE (ALL to delete everything)

Output Options:

-q, --quiet                 Activate quiet mode
-v, --verbose               Print various debugging information
-g, --get-urls              Print URLs instead of downloading
-G, --resolve-urls          Print URLs instead of downloading; resolve
                            intermediary URLs
-j, --dump-json             Print JSON information
-s, --simulate              Simulate data extraction; do not download
                            anything
-E, --extractor-info        Print extractor defaults and settings
-K, --list-keywords         Print a list of available keywords and example
                            values for the given URLs
--list-modules              Print a list of available extractor modules
--list-extractors           Print a list of extractor classes with
                            description, (sub)category and example URL
--write-log FILE            Write logging output to FILE
--write-unsupported FILE    Write URLs, which get emitted by other
                            extractors but cannot be handled, to FILE
--write-pages               Write downloaded intermediary pages to files in
                            the current directory to debug problems

Downloader Options:

-r, --limit-rate RATE       Maximum download rate (e.g. 500k or 2.5M)
-R, --retries N             Maximum number of retries for failed HTTP
                            requests or -1 for infinite retries (default: 4)
--http-timeout SECONDS      Timeout for HTTP connections (default: 30.0)
--sleep SECONDS             Number of seconds to wait before each download.
                            This can be either a constant value or a range
                            (e.g. 2.7 or 2.0-3.5)
--sleep-request SECONDS     Number of seconds to wait between HTTP requests
                            during data extraction
--sleep-extractor SECONDS   Number of seconds to wait before starting data
                            extraction for an input URL
--filesize-min SIZE         Do not download files smaller than SIZE (e.g.
                            500k or 2.5M)
--filesize-max SIZE         Do not download files larger than SIZE (e.g.
                            500k or 2.5M)
--chunk-size SIZE           Size of in-memory data chunks (default: 32k)
--no-part                   Do not use .part files
--no-skip                   Do not skip downloads; overwrite existing files
--no-mtime                  Do not set file modification times according to
                            Last-Modified HTTP response headers
--no-download               Do not download any files
--no-postprocessors         Do not run any post processors
--no-check-certificate      Disable HTTPS certificate validation

Configuration Options:

-o, --option KEY=VALUE      Additional options. Example: -o browser=firefox
-c, --config FILE           Additional configuration files
--config-yaml FILE          Additional configuration files in YAML format
--config-toml FILE          Additional configuration files in TOML format
--config-create             Create a basic configuration file
--config-ignore             Do not read default configuration files

Authentication Options:

-u, --username USER         Username to login with
-p, --password PASS         Password belonging to the given username
--netrc                     Enable .netrc authentication data

Cookie Options:

-C, --cookies FILE          File to load additional cookies from
--cookies-export FILE       Export session cookies to FILE
--cookies-from-browser BROWSER[/DOMAIN][+KEYRING][:PROFILE][::CONTAINER]
                            Name of the browser to load cookies from, with
                            optional domain prefixed with '/', keyring name
                            prefixed with '+', profile prefixed with ':',
                            and container prefixed with '::' ('none' for no
                            container)

Selection Options:

--download-archive FILE     Record all downloaded or skipped files in FILE
                            and skip downloading any file already in it
-A, --abort N               Stop current extractor run after N consecutive
                            file downloads were skipped
-T, --terminate N           Stop current and parent extractor runs after N
                            consecutive file downloads were skipped
--range RANGE               Index range(s) specifying which files to
                            download. These can be either a constant value,
                            range, or slice (e.g. '5', '8-20', or '1:24:3')
--chapter-range RANGE       Like '--range', but applies to manga chapters
                            and other delegated URLs
--filter EXPR               Python expression controlling which files to
                            download. Files for which the expression
                            evaluates to False are ignored. Available keys
                            are the filename-specific ones listed by '-K'.
                            Example: --filter "image_width >= 1000 and
                            rating in ('s', 'q')"
--chapter-filter EXPR       Like '--filter', but applies to manga chapters
                            and other delegated URLs

Post-processing Options:

--zip                       Store downloaded files in a ZIP archive
--ugoira-conv               Convert Pixiv Ugoira to WebM (requires FFmpeg)
--ugoira-conv-lossless      Convert Pixiv Ugoira to WebM in VP9 lossless
                            mode
--ugoira-conv-copy          Convert Pixiv Ugoira to MKV without re-encoding
                            any frames
--write-metadata            Write metadata to separate JSON files
--write-info-json           Write gallery metadata to a info.json file
--write-tags                Write image tags to separate text files
--mtime-from-date           Set file modification times according to 'date'
                            metadata
--exec CMD                  Execute CMD for each downloaded file. Supported
                            replacement fields are {} or {_path},
                            {_directory}, {_filename}. Example: --exec
                            "convert {} {}.png && rm {}"
--exec-after CMD            Execute CMD after all files were downloaded
                            successfully. Example: --exec-after "cd
                            {_directory} && convert * ../doc.pdf"
-P, --postprocessor NAME    Activate the specified post processor
-O, --postprocessor-option OPT
                            Additional '<key>=<value>' post processor
                            options