学了这一节感觉难度不大,开始上不去网发现需要翻~wall,好吧。还有就是下载图片的命名发现切片少了命名会重就会覆盖,下载了半天发现的。
我的成果
![](https://img.haomeiwen.com/i1709416/d9896b99857583cc.jpg)
我的代码
from bs4 import BeautifulSoup
import requests
import urllib.request
urls=['http://weheartit.com/inspirations/taylorswift?page={}&before='.format(i) for i in range(1,21)]
for url in urls:
wb_data=requests.get(url)
soup=BeautifulSoup(wb_data.text,'lxml')
imgs=soup.select('#main-container > div > div > div > div > div > a > img')
download_links=[]
for img in imgs:
download_links.append(img.get('src'))
file_path='/Users/mac/Desktop/taylor/'
for item,i in zip(download_links,range(1,len(download_links))):
urllib.request.urlretrieve(item,file_path+str(i)+item[-8:])
print('done')
总结
- 发现命名重了会覆盖,真是的,加了个变量让他增长肯定就不重了
- 貌似挂了vpn就不用写代理了
网友评论