# wikiget [![Made with Python](https://img.shields.io/badge/made_with-python-3776AB?logo=python)][python] [![GitHub Actions](https://img.shields.io/github/actions/workflow/status/clpo13/wikiget/python.yml?logo=github)][action] [![Codecov coverage](https://img.shields.io/codecov/c/gh/clpo13/wikiget)][codecov] [![PyPI version](https://img.shields.io/pypi/v/wikiget)][pypi] [![PyPI license](https://img.shields.io/pypi/l/wikiget)][license] Something like [wget] for downloading a file from MediaWiki sites (like Wikipedia or Wikimedia Commons) using only the file name or the URL of its description page. ## Installation Requires Python 3.7+ and pip. Install the latest version with: ```bash pip install wikiget ``` For the latest features, at the risk of bugs and undocumented behavior, you can install the development version directly from [GitHub]: ```bash pip install https://github.com/clpo13/wikiget/archive/refs/heads/master.zip ``` Alternatively, if you have [Homebrew] installed: ```bash brew tap clpo13/clpo13 brew install wikiget ``` ## Usage `wikiget [-h] [-V] [-q | -v] [-f] [-s SITE] [-P PATH] [-u USERNAME] [-p PASSWORD] [-o OUTPUT | -a] [-l LOGFILE] [-j THREADS] FILE` The only required parameter is `FILE`, which is the file you want to download. It can either be the name of the file on the wiki, including the namespace prefix, or a link to the file description page. If `FILE` is in the form `File:Example.jpg` or `Image:Example.jpg`, it will be fetched from the default site, which is "commons.wikimedia.org". If it's the fully-qualified URL of a file description page, like `https://en.wikipedia.org/wiki/File:Example.jpg`, the file is fetched from the site in the URL, in this case "en.wikipedia.org". Note: full URLs may contain characters your shell interprets differently, so you can either escape those characters with a backslash `\` or surround the entire URL with single `'` or double `"` quotes. Use of a fully-qualified URL like this may require setting the `--path` flag (see next paragraph). The site can also be specified with the `--site` flag, though this will not have any effect if the full URL is given. Non-Wikimedia sites should work, but you may need to specify the wiki's script path with `--path` (where `index.php` and `api.php` live; on Wikimedia sites it's `/w/`, but other sites may use `/` or something else entirely). Private wikis (those requiring login even for read access) are also supported with the use of the `--username` and `--password` flags. More detailed information, such as the site used and full URL of the file, can be displayed with `-v` or `--verbose`. Use `-vv` to display even more detail, mainly debugging information or API messages. `-q` can be used to silence warnings. A logfile can be specified with `-l` or `--logfile`. If this option is present, the logfile will contain the same information as `-v` along with timestamps. New log entries will be appended to an existing logfile. By default, the program won't overwrite existing files with the same name as the target, but this can be forced with `-f` or `--force`. Additionally, the file can be downloaded to a different name with `-o`. Files can be batch downloaded with the `-a` or `--batch` flag. In this mode, `FILE` will be treated as an input file containing multiple files to download, one filename or URL per line. Blank lines and lines starting with "#" are ignored. If an error is encountered, execution stops immediately and the offending filename is printed. For large batches, the process can be sped up by downloading files in parallel. The number of parallel downloads can be set with `-j`. For instance, with `-a -j4`, wikiget will download four files at once. Without `-j` or with `-j` by itself without a number, wikiget will download the files one at a time. ### Example usage ```bash wikiget File:Example.jpg wikiget --site en.wikipedia.org File:Example.jpg wikiget https://en.wikipedia.org/wiki/File:Example.jpg -o test.jpg ``` ## Future plans - optional machine-readable (JSON) log output - batch download by (Commons) category or user uploads - maybe: download Wikipedia articles, in plain text, wikitext, or other formats ## Contributing Pull requests, bug reports, or feature requests are more than welcome. See [CONTRIBUTING] for more info. ## License Copyright (C) 2018-2023 Cody Logan and contributors This program is free software: you can redistribute it and/or modify it under the terms of the GNU General Public License as published by the Free Software Foundation, either version 3 of the License, or (at your option) any later version. This program is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more details. You should have received a copy of the GNU General Public License along with this program. If not, see . [wget]: https://www.gnu.org/software/wget/ [github]: https://github.com/clpo13/wikiget/ [homebrew]: https://brew.sh/ [python]: https://www.python.org/ [action]: https://github.com/clpo13/wikiget/actions/workflows/python.yml [codecov]: https://app.codecov.io/gh/clpo13/wikiget/ [pypi]: https://pypi.org/project/wikiget/ [license]: https://github.com/clpo13/wikiget/blob/master/LICENSE [contributing]: CONTRIBUTING.md