标签:加载 bin like google details first x86_64 mat x86
# -*- coding: utf-8 -*- import time import os import urllib from bs4 import BeautifulSoup def SaveImage(link,InputData,count): try: time.sleep(0.2) urllib.request.urlretrieve(link,‘./‘+InputData+‘/‘+str(count)+‘.jpg‘) except Exception : time.sleep(1) print("产生未知错误,放弃保存") else: print("图+1,已有" + str(count) + "张图") def FindLink(PageNum, InputData, word): for i in range(PageNum): print(i) try: url = ‘http://cn.bing.com/images/async?q={0}&first={1}&count=35&relp=35&lostate=r&mmasync=1&dgState=x*175_y*848_h*199_c*1_i*106_r*0‘ # 定义请求头 agent = { ‘User-Agent‘: "Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/31.0.165063 Safari/537.36 AppEngine-Google."} page1 = urllib.request.Request(url.format(InputData, i * 35 + 1), headers=agent) page = urllib.request.urlopen(page1) #使用beautifulSoup进行解析网页 soup = BeautifulSoup(page.read(), ‘html.parser‘) # 创建文件夹 if not os.path.exists("./" + word): os.mkdir(‘./‘ + word) for StepOne in soup.select(‘.mimg‘): link = StepOne.attrs[‘src‘] count = len(os.listdir(‘./‘ + word)) + 1 SaveImage(link, word, count) except: print(‘URL OPENING ERROR !‘) if __name__==‘__main__‘: #输入需要加载的页数,每页35幅图像 PageNum = 100 #输入需要搜索的关键字from bs4 import BeautifulSoup word=‘金发‘ #UTF-8编码 InputData=urllib.parse.quote(word) print(InputData) FindLink(PageNum,InputData,word)
原文链接:https://blog.csdn.net/qq_37482202/article/details/82352870
标签:加载 bin like google details first x86_64 mat x86
原文地址:https://www.cnblogs.com/ya-cpp/p/12011808.html