简体   繁体   中英

How can I get wget to download all the pdf files from this website?

The website is: https://dgriffinchess.wordpress.com/

I already downloaded the entire website, but I'd also like to have the pdf files, and yes, I've tried this , this and this answer, and unless wget saves the pdf files other than the main site folder(the one at the Home directory), I don't see them downloading at all..(I don't wait until the command finishes, I just wait for a few minutes and see that no pdf file has been downloaded yet, and considering that there is one almost on every webpage, I conclude that the pdf's aren't getting downloaded) I don't really care if I have to re-download the entire website again, it's not that big to begin with, what matters most to me are the.pdf files, which doesn't seem to download in any way..

Many thanks in advance

The PDF files are stored on another domain, dgriffinchess.files.wordpress.com .

To completely download this website along with the PDF files, you need to authorize the domain name where the PDF files are stored using --span-hosts and --domains=domain_a,domain_b :

wget --recursive --page-requisites --convert-links --span-hosts --domains=dgriffinchess.wordpress.com,dgriffinchess.files.wordpress.com https://dgriffinchess.wordpress.com/

The technical post webpages of this site follow the CC BY-SA 4.0 protocol. If you need to reprint, please indicate the site URL or the original address.Any question please contact:yoyou2525@163.com.

 
粤ICP备18138465号  © 2020-2024 STACKOOM.COM