这是源码,20页的图片链接存放在links中用循环写入本地,链接能够正确得出,但写入本地部分会报错
import requests
import urllib.request
from time import sleep
from bs4 import BeautifulSoup
# url = 'https://weheartit.com/inspirations/taylorswift?page=1&before=275543698'
urls = ['https://weheartit.com/inspirations/taylorswift?page={}&before=275543698'.format(str(i)) for i in range(1,21)]
headers = {
'User-Agent':'Mozilla/5.0 (Windows NT 6.1; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/59.0.3071.115 Safari/537.36'
}
proxies = {"http": "http://122.72.32.73:80"}
folder_path = r'E:\文档\Python\Taylor Swift'
links = []
i = 0
def get_image(url):
data = requests.get(url, headers=headers)
sleep(2)
soup = BeautifulSoup(data.text,'lxml')
imgs = soup.select('img[width="300"]')
# 判断链接后缀并替换为.webp
for img in imgs:
link = img.get('src')
if link[-4:] =='.png':
img_link = link.replace('png', 'webp')
else:
img_link = link.replace('jpg','webp')
links.append(img_link)
for link in links:
urllib.request.urlretrieve(link, folder_path + '\\' + str(i) + link[-4:])
global i
i += 1
print('完成第%d张' % i)
for url in urls:
get_image(url)
报错内容:urllib.error.URLError: <urlopen error [SSL: CERTIFICATE_VERIFY_FAILED] certificate verify failed (_ssl.c:720)>
你该import ssl
https://stackoverflow.com/que...