• 微信公众号:美女很有趣。 工作之余,放松一下,关注即送10G+美女照片!

web爬虫01-单线程爬虫

开发技术 开发技术 3小时前 2次浏览

单线程爬虫

目标网站:https://movie.douban.com/top250 豆瓣top250

代码:

import codecs
import requests
from bs4 import BeautifulSoup
import time

DOWNLOAD_URL = 'https://movie.douban.com/top250'


# 时间装饰器
def timer(func):
    def inner(*args, **kw):
        t1 = time.time()
        func(*args, **kw)
        t2 = time.time()
        print("-------一共花费时间:{}秒".format(t2-t1))
        return t2 - t1
    return inner


# 下载html
def download_page(url):
    return requests.get(url, headers={
        'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/61.0.3163.100 Safari/537.36'
    }).content


# 提取需要信息
def parse_html(html):
    soup = BeautifulSoup(html, 'html.parser')
    # 电影列表
    movie_list_soup = soup.find('ol', attrs={'class': 'grid_view'})

    movie_name_list = []

    for movie_li in movie_list_soup.find_all('li'):
        movie_name = movie_li.find('span', attrs={'class': 'title'}).get_text()
        movie_info = movie_li.find('div', attrs={'class': 'bd'}).find('p').get_text()
        movie_star = movie_li.find('span', attrs={'class': 'rating_num'}).get_text()

        movie_name_list.append(movie_name)
        movie_name_list.append(movie_info)
        movie_name_list.append(movie_star)

    # 下一页链接
    next_page = soup.find('span', attrs={'class': 'next'}).find('a')
    if next_page:
        return movie_name_list, DOWNLOAD_URL + next_page['href']
    return movie_name_list, None


@timer
def main():
    url = DOWNLOAD_URL
    with codecs.open('movies', 'wb', encoding='utf-8') as f:
        while True:
            if url == None:
                break
            html = download_page(url)
            movies, url = parse_html(html)
            f.write(u'{movies}n'.format(movies='n'.join(movies)))


if __name__ == '__main__':
    main()
#参考链接:https://www.jianshu.com/p/8a460be5a26e
花费时间:

web爬虫01-单线程爬虫


程序员灯塔
转载请注明原文链接:web爬虫01-单线程爬虫
喜欢 (0)