![]() ![]() Protocols used for data exchange have a lot of metadata embedded in the packets computers send to communicate. If you're using wget to archive a site, then the options -no-cookies -page-requisites -convert-links are also useful to ensure that every page is fresh, complete, and that the site copy is more or less self-contained. Depending on how old the website is, that could mean you're getting a lot more content than you realize. This option is the same as running -recursive -level inf -timestamping -no-remove-listing, which means it's infinitely recursive, so you're getting everything on the domain you specify. You can download an entire site, including its directory structure, using the -mirror option. ![]() Assuming you know the location and filename pattern of the files you want to download, you can use Bash syntax to specify the start and end points between a range of integers to represent a sequence of filenames: $ wget http: // /file_. ![]() If it's not one big file but several files that you need to download, wget can help you with that. $ wget -continue https: // /linux-distro.iso Download a sequence of files That means the next time you download a 4 GB Linux distribution ISO you don't ever have to go back to the start when something goes wrong. With the -continue ( -c for short), wget can determine where the download left off and continue the file transfer. If you're downloading a very large file, you might find that you have to interrupt the download. You can use the -output-document option ( -O for short) to name your download whatever you want: $ wget http: // -output-document foo.html Continue a partial download $ wget http: // -output-document - | head -n4 ![]()
0 Comments
Leave a Reply. |
AuthorWrite something about yourself. No need to be fancy, just an overview. ArchivesCategories |