from bs4 import BeautifulSoup
import requests
import pymysql
import warnings
# import pinyin
# from pinyin import PinYin
from pypinyin import pinyin, lazy_pinyin
import pypinyin
warnings.filterwarnings("ignore")
conn = pymysql.connect(host=‘localhost‘, user=‘root‘, passwd=‘root‘, db=‘test2‘, port=3306, charset=‘utf8‘)
cursor = conn.cursor()
def get_temperature(url,city):
headers = {
‘User-Agent‘: ‘Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/63.0.3239.132 Safari/537.36‘} # 设置头文件信息
response = requests.get(url, headers=headers).content # 提交requests get 请求
soup = BeautifulSoup(response, "lxml") # 用Beautifulsoup 进行解析
conmid2 = soup.findAll(‘div‘, class_=‘wdetail‘)
# conmid2 = conmid.findAll(‘div‘, class_=‘wdetail‘)
for info in conmid2:
tr_list = info.find_all(‘tr‘)[1:] # 使用切片取到第三个tr标签
for index, tr in enumerate(tr_list): # enumerate可以返回元素的位置及内容
td_list = tr.find_all(‘td‘)
# if index == 0:
date = td_list[0].text.strip().replace("\n", "") # 取每个标签的text信息,并使用replace()函数将换行符删除
weather = td_list[1].text.strip().replace("\n", "").split("/")[0].strip()
temperature = td_list[2].text.strip().replace("\n", "").split("/")[0].strip()
wind = td_list[3].text.strip().replace("\n", "").split("/")[0].strip()
# else:
# city_name = td_list[0].text.replace(‘\n‘, ‘‘)
# weather = td_list[4].text.replace(‘\n‘, ‘‘)
# wind = td_list[5].text.replace(‘\n‘, ‘‘)
# max = td_list[3].text.replace(‘\n‘, ‘‘)
# min = td_list[6].text.replace(‘\n‘, ‘‘)
print(city,date, weather, wind, temperature)
cursor.execute(‘insert into weather(city, date, weather, wind, temp) values(%s, %s, %s, %s, %s)‘
, (city, date, weather, wind, temperature ))
if __name__==‘__main__‘:
# citys1= ["成都市","广元市","绵阳市","德阳市","南充市","广安市","遂宁市","内江市","乐山市","自贡市","泸州市","宜宾市","攀枝花市","巴中市","达州市","资阳市","眉山市","雅安市","崇州市","邛崃市","都江堰市","彭州市","江油市","什邡市","广汉市","绵竹市","阆中市","华蓥市","峨眉山市","万源市","简阳市","西昌市","康定市","马尔康市","隆昌市"]
# citys1= ["郑州市","开封市","洛阳市","平顶山市","安阳市","鹤壁市","新乡市","焦作市","濮阳市","许昌市","漯河市","三门峡市","南阳市","商丘市","周口市","驻马店市","信阳市","荥阳市","新郑市","登封市","新密市","偃师市","孟州市","沁阳市","卫辉市","辉县市","林州市","禹州市","长葛市","舞钢市","义马市","灵宝市","项城市","巩义市","邓州市","永城市","汝州市","济源市"]
# citys1= ["呼和浩特市","包头市","乌海市","赤峰市","通辽市","鄂尔多斯市","呼伦贝尔市","巴彦淖尔市","乌兰察布市","霍林郭勒市","满洲里市","牙克石市","扎兰屯市","额尔古纳市","根河市","丰镇市","乌兰浩特市","阿尔山市","二连浩特市","锡林浩特市"]
# citys1= ["沈阳市","大连市","鞍山市","抚顺市","本溪市","丹东市","锦州市","营口市","阜新市","辽阳市","盘锦市","铁岭市","朝阳市","葫芦岛市","新民市","瓦房店市","庄河市","海城市","东港市","凤城市","凌海市","北镇市","盖州市","大石桥市","灯塔市","调兵山市","开原市","北票市","凌源市","兴城市"]
# citys1= ["葫芦岛市","新民市","瓦房店市","庄河市","海城市","东港市","凤城市","凌海市","北镇市","盖州市","大石桥市","灯塔市","调兵山市","开原市","北票市","凌源市","兴城市"]
citys1= ["开原市","北票市","凌源市","兴城市"]
for city in citys1:
city1 = ‘‘.join(lazy_pinyin(city[:-1]))
print(city1)
urls = [‘http://www.tianqihoubao.com/lishi/‘+city1+‘/month/201801.html‘,
‘http://www.tianqihoubao.com/lishi/‘+city1+‘/month/201802.html‘,
‘http://www.tianqihoubao.com/lishi/‘+city1+‘/month/201803.html‘,
‘http://www.tianqihoubao.com/lishi/‘+city1+‘/month/201804.html‘,
‘http://www.tianqihoubao.com/lishi/‘+city1+‘/month/201805.html‘,
‘http://www.tianqihoubao.com/lishi/‘+city1+‘/month/201806.html‘,
‘http://www.tianqihoubao.com/lishi/‘+city1+‘/month/201807.html‘]
for url in urls:
get_temperature(url, city)
conn.commit()