HACKINTOSH.ORG | Macintosh discussion forums

Macintosh News => iPhone/iPod/iPad News => Topic started by: HCK on September 11, 2018, 04:05:20 pm



Title: Use wget to archive websites and online docs on your Mac
Post by: HCK on September 11, 2018, 04:05:20 pm
Use wget to archive websites and online docs on your Mac

Need to bulk save online resources? You can with wget!

Do you need to download a large quantity of online documentation for your work or university studies but have limited internet access? Or perhaps do you simply want to be able to locally store web documents so you can parse them with desktop tools? On macOS, you can easily archive any freely accessible online URL (or an entire subdomain if you have the disk capacity!) with free and open source software (FOSS) in one simple terminal command. Here's how!

The wget command
Options galore
Using wget
Getting wget
Final comments
The wget command

The wget command is network downloader that can follow and archive HTTP, HTTPS, and FTP protocols. It's designated as a "non-interactive" command because you can initiate the program and leave it to do its work without any other user interaction. The wget manual explains it this way:


  Wget can follow links in HTML, XHTML, and CSS pages, to create local versions of remote web sites, fully r...

Source: Use wget to archive websites and online docs on your Mac (http://feedproxy.google.com/~r/TheIphoneBlog/~3/WemUahbjiqw/how-archive-websites-online-documents-macos-wget)