码迷,mamicode.com
首页 > 其他好文 > 详细

requests库和BeautifulSoup4库爬取新闻列表

时间:2017-09-28 16:48:04      阅读:167      评论:0      收藏:0      [点我收藏+]

标签:url   split   tle   sele   准备   image   gen   com   enc   

画图显示:

 

import jieba
from wordcloud import WordCloud
import matplotlib.pyplot as plt

txt = open("zuihou.txt","r",encoding=utf-8).read()
wordlist = jieba.lcut(txt)

 

wl_split=" ".join(wordlist)
mywc = WordCloud().generate(wl_split)
plt.imshow(mywc)
plt.axis("off")
plt.show()

结果:

技术分享

 

 用requests库和BeautifulSoup4库,爬取校园新闻列表的时间、标题、链接、来源、详细内容

爬虫,网页信息

import requests
from bs4 import BeautifulSoup

gzccurl = http://news.gzcc.cn/html/xiaoyuanxinwen/
res = requests.get(gzccurl)
res.encoding = utf-8

soup = BeautifulSoup(res.text,html.parser)
for news in soup.select(li):
    if len(news.select(.news-list-title))>0:
        title = news.select(.news-list-title)[0].text
        url = news.select(a)[0][href]
        print(title,url)

结果:

技术分享

 加上时间:

for news in soup.select(li):
    if len(news.select(.news-list-title))>0:
        title = news.select(.news-list-title)[0].text
        url = news.select(a)[0][href]
        time = news.select(.news-list-info)[0].contents[0].text
        
        print(time,title,url)

效果:

技术分享

  • 将其中的时间str转换成datetime类型。
  • 将取得详细内容的代码包装成函数。
import requests
from bs4 import BeautifulSoup
from datetime import datetime

gzccurl = http://news.gzcc.cn/html/xiaoyuanxinwen/
res = requests.get(gzccurl)
res.encoding = utf-8

soup = BeautifulSoup(res.text,html.parser)
def getdetail(url):
    resd = requests.get(url)
    resd.encoding= utf-8
    soupd = BeautifulSoup(resd.text,html.parser)
    return(soupd.select(.show-content)[0].text)

for news in soup.select(li):
    if len(news.select(.news-list-title))>0:
        title = news.select(.news-list-title)[0].text
        url = news.select(a)[0][href]
        time = news.select(.news-list-info)[0].contents[0].text
        dt = datetime.strptime(time,%Y-%m-%d)
        source = news.select(.news-list-info)[0].contents[1].text
        detail = getdetail(url)
        print(dt,title,url,source,detail)

结果:

技术分享

  • 选一个自己感兴趣的主题,做类似的操作,为后面“爬取网络数据并进行文本分析”做准备。
import requests
from bs4 import BeautifulSoup
from datetime import datetime

gzccurl = http://www.lbldy.com/tag/gqdy/
res = requests.get(gzccurl)
res.encoding = utf-8

soup = BeautifulSoup(res.text,html.parser)
def getdetail(url):
    resd = requests.get(url)
    resd.encoding= utf-8
    soupd = BeautifulSoup(resd.text,html.parser)
    return(soupd.select(.show-content)[0].text)

for news in soup.select(h4):
    
        print(news)

结果:

技术分享

 

requests库和BeautifulSoup4库爬取新闻列表

标签:url   split   tle   sele   准备   image   gen   com   enc   

原文地址:http://www.cnblogs.com/ruijin-chen/p/7605982.html

(0)
(0)
   
举报
评论 一句话评论(0
登录后才能评论!
© 2014 mamicode.com 版权所有  联系我们:gaon5@hotmail.com
迷上了代码!