[英]scraping: download files from url
使用BeautifulSoup
Python版本。
try:
# Python 3.x
from urllib.request import urlopen, urlretrieve, quote
from urllib.parse import urljoin
except ImportError:
# Python 2.x
from urllib import urlopen, urlretrieve, quote
from urlparse import urljoin
from bs4 import BeautifulSoup
url = 'http://oilandgas.ky.gov/Pages/ProductionReports.aspx'
u = urlopen(url)
try:
html = u.read().decode('utf-8')
finally:
u.close()
soup = BeautifulSoup(html)
for link in soup.select('div[webpartid] a'):
href = link.get('href')
if href.startswith('javascript:'):
continue
filename = href.rsplit('/', 1)[-1]
href = urljoin(url, quote(href))
try:
urlretrieve(href, filename)
except:
print('failed to download')
這對我有用:
getIt = function(what,when){
url=paste0("http://oilandgas.ky.gov/Production%20Reports%20Library/",
when,"%20-%20",what,
"%20Production.xls")
destfile=paste0("/tmp/",what,when,".xls")
download.file(url,destfile)
}
例如:
> getIt("gas",2006)
trying URL 'http://oilandgas.ky.gov/Production%20Reports%20Library/2006%20-%20gas%20Production.xls'
Content type 'application/vnd.ms-excel' length 3490304 bytes (3.3 Mb)
opened URL
==================================================
downloaded 3.3 Mb
除了第一個:
> getIt("oil",2010)
trying URL 'http://oilandgas.ky.gov/Production%20Reports%20Library/2010%20-%20oil%20Production.xls'
Error in download.file(url, destfile) :
cannot open URL 'http://oilandgas.ky.gov/Production%20Reports%20Library/2010%20-%20oil%20Production.xls'
In addition: Warning message:
In download.file(url, destfile) :
cannot open: HTTP status was '404 NOT FOUND'
雖然我可以獲得2010年的天然氣數據:
> getIt("gas",2010)
trying URL 'http://oilandgas.ky.gov/Production%20Reports%20Library/2010%20-%20gas%20Production.xls'
Content type 'application/vnd.ms-excel' length 4177408 bytes (4.0 Mb)
opened URL
==================================================
downloaded 4.0 Mb
所以看起來他們改變了那個鏈接的系統。 您可以通過鏈接獲取該數據,然后在cruddy Sharepoint HTML中查找下載鏈接。
這就是為什么我們討厭Sharepoint,小孩。
聲明:本站的技術帖子網頁,遵循CC BY-SA 4.0協議,如果您需要轉載,請注明本站網址或者原文地址。任何問題請咨詢:yoyou2525@163.com.