How to use Wget: Install, Commands and Examples (Mac & Windows)

Share this post

Wget is free command-line tool that you can use to download files from the internet.

In this wget tutorial, we will learn how to install and how to use wget.

Subscribe to my Newsletter

What is Wget and What Does it Do?

WGET is a free tool to crawl websites and download files via the command line.

Subscribe to my Newsletter

  • It lets you download files from the internet via FTP, HTTP or HTTPS (web pages, pdf, xml sitemaps, etc.).
  • It provides recursive downloads, which means that Wget downloads the requested document, then the documents linked from that document, and then the next, etc.
  • It follows the links and directory structure.
  • It lets you overwrite the links with the correct domain, helping you create mirrors of websites.

Install Wget

Check if Wget is installed

Open Terminal and type:

$ wget -V

If it is installed, it will return the version.

You Might Also Like  Python Script Automation Using Task Scheduler (Windows)

If not, follow the next steps to download wget on either Mac or Windows.

Download Wget on Mac

The recommended method to install wget on Mac is with Homebrew.

First, install Homebrew.

$ ruby -e "$(curl -fsSL"

Then, install wget.

$ brew install wget

Download Wget on Windows

To install and configure wget for Windows:

  1. Download wget for Windows and install the package.
  2. Copy the wget.exe file into your C:\Windows\System32 folder.
  3. Open the command prompt (cmd.exe) and run wget to see if it is installed.

Here is a quick video showing you how to download wget on windows 10.

Wget Basics

Let’s look at the wget syntax, view the basic commands structure and understand the most important options.

Wget Syntax

Wget has two arguments: [OPTION] and [URL] .

wget [OPTION]... [URL]...
  • [OPTION] tells what to do with the [URL] argument provided after. It has a short and a long-form (ex: -V and --version are doing the same thing).
  • [URL] is the file or the directory you wish to download.
  • You can call many OPTIONS or URLs at once.

View WGET commands

To view available wget commands, use wget -h.

Extract Web pages with Wget Commands

Download a single file

$ wget

Download a File to a Specific Output Directory

Here replace <YOUR-PATH> by the output directory location where you want to save the file.

$ wget ‐P <YOUR-PATH>

Rename Downloaded File

To output the file with a different name:

$ wget -O <YOUR-FILENAME.html>

Define User Agent

Identify yourself. Define your user-agent.

$ wget --user-agent=Chrome

Extract as Google bot

$ wget --user-agent="Mozilla/5.0 (Linux; Android 6.0.1; Nexus 5X Build/MMB29P) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/86.0.4240.198 Mobile Safari/537.36 (compatible; Googlebot/2.1; +"

Extract Robots.txt only When it Changes

Let’s extract robots.txt only if the latest version in the server is more recent than the local copy.

You Might Also Like  Get Started With Django (Python Tutorial)

First time that you extract use -S to keep a timestamps of the file.

$ wget -S

Later, to check if the robots.txt file has changed, and download it if it has.

$ wget -N

Convert Links on a Page

Convert the links in the HTML so they still work in your local version. (ex: to localhost:8000/path)

$ wget --convert-links

Mirror a Single Webpage

To mirror a single web page so that it can work on your local.

$ wget -E -H -k -K -p --convert-links

Extract Multiple URLs

Add all urls in a urls.txt file.
$ wget -i urls.txt

Limit Speed

To be a good citizen of the web, it is important not to crawl too fast by using --wait and --limit-rate.

  • --wait=1: Wait 1 second between extractions.
  • --limit-rate=10K: Limit the download speed (bytes per second)

Extract Entire Site (Proceed with Caution)

Recursive mode extract a page, and follows the links on the pages to extract them as well.

This is extracting your entire site and can put extra load on your server. Be sure that you know what you do or that you involve the devs.

$ wget --recursive --page-requisites --adjust-extension --span-hosts --wait=1 --limit-rate=10K --convert-links --restrict-file-names=windows --no-clobber --domains --no-parent
  • --recursive: Follow links in the document. The maximum depth is 5.
  • --page-requisites: Get all assets (CSS/JS/images)
  • --adjust-extension: Save files with .html at the end.
  • --span-hosts: Include necessary assets from offsite as well.
  • --wait=1: Wait 1 second between extractions.
  • --limit-rate=10K: Limit the download speed (bytes per second)
  • --convert-links: Convert the links in the HTML so they still work in your local version.
  • --restrict-file-names=windows: Modify filenames to work in Windows.
  • --no-clobber: Overwrite existing files.
  • --domains Do not follow links outside this domain.
  • --no-parent: Do not ever ascend to the parent directory when retrieving recursively
  • --level: Specify the depth of crawling. inf is used for infinite.
You Might Also Like  Mirror a Webpage on NGrok with Python and Wget

(Extra) Run Spider Mode

$ wget --spider -r -o wget.log

Wget VS Curl

Wget’s strength compared to curl is its ability to download recursively. This means that it will download a document, then follow the links and then download those documents as well.

Use Wget With Python

Wget is strictly command line, but there is a package that you can import the wget package that mimics wget.

import wget
url = ''
filename =

Debug Wget Command Not Found

If you get the -bash: wget: command not found error on Mac, Linux or Windows, it means that the wget GNU is either not installed or does not work properly.

Go back and make sure that you installed wget properly.


This is it.

You now know how to install and use Wget in your command-line.

Enjoyed This Post?