美文网首页
BeautifulSoup4小试牛刀

BeautifulSoup4小试牛刀

作者: Snow_DZG | 来源:发表于2016-08-03 14:44 被阅读49次

    python真好玩,python真有趣,python真该死!尝试爬取起点中文网的小说,因为他家的小说网页比较复杂,有一点挑战性!

    问题(慢慢解决):
    1.代码杂乱,可以适当精简重构
    2.可以采用多线程的方式去爬,效率可以更高
    3.适当的采用各种包各种方式,让网站不认为自己是爬虫

    不多说直接上代码:

    需要注意:
    1."C:/Users/DZG/Desktop/" 这个路径换成自己本机的
    2.爬取的url,可以复制一个去看下 这个url是哪里的url,-~

    #! usr/bin/env python3
    # -*- coding:utf-8 -*-
    # author DZG 2016-08-03
    
    #问题(慢慢解决):
    #1.代码杂乱,可以适当精简重构
    #2.可以采用多线程的方式去爬,效率可以更高
    #3.适当的采用各种包各种方式,让网站不认为自己是爬虫
    #4.目前只是爬取免费,后期应该可以加入 模拟VIP登录的功能
    #需要注意:
    #1."C:/Users/DZG/Desktop/" 这个路径换成自己本机的
    #2.爬取的url,可以复制一个去看下 这个url是哪里的url,^-^~
    
    import urllib.request
    import chardet
    import time
    import random
    from bs4 import BeautifulSoup as bs
    
    ISOTIMEFORMAT='%Y-%m-%d %X'
    
    #通用装饰器1
    def writeBook(func):
        def swr(*args,**kw):
            f = open('C:/Users/DZG/Desktop/' + args[-1] + '.md', 'a', encoding='utf-8');
            f.write('\n\n')
            result = func(*args, **kw)
            f.write('>')
            return result;
        return swr;
    
    #创建文件并定义名称
    def createBook(bookName):
        f=open('C:/Users/DZG/Desktop/'+ bookName +'.md','a',encoding='utf-8');
    
    #写卷目录
    def createTitle(title,bookName):
        f=open('C:/Users/DZG/Desktop/'+bookName+'.md','a',encoding='utf-8');
        f.write('\n\n### ' + title)
    
    #写每卷中对应的章节目录
    def writeZhangjie(menuContentBs, bookName,i):
        menuZhangJie = menuContentBs.find_all("div", "box_cont")[1:];
        menuZhangJieS = bs(str(menuZhangJie[i]), 'html.parser');
        menuZhangJieSpan = menuZhangJieS.find_all('span')
        menuZhangJieHref = menuZhangJieS.find_all('a')
        menuZhangJieHrefList = getHref(menuZhangJieHref)
        writeZhangjieDetail(menuZhangJieSpan, bookName,menuZhangJieHrefList)
    
    def writeZhangjie2(menuContentBs, bookName,i):
        menuZhangJie = menuContentBs.find_all("div", "box_cont")[0:];
        menuZhangJieS = bs(str(menuZhangJie[i]), 'html.parser');
        menuZhangJieSpan = menuZhangJieS.find_all('span')
        menuZhangJieHref = menuZhangJieS.find_all('a')
        menuZhangJieHrefList = getHref(menuZhangJieHref)
        writeZhangjieDetail(menuZhangJieSpan, bookName,menuZhangJieHrefList)
    
    #写章节细节和每章内容的方法(还有点问题)
    def writeZhangjieDetail(menuZhangJieSpan, bookName,menuZhangJieHrefList):
        j = 0
        for spanValue in menuZhangJieSpan:
            timeout = random.choice(range(80, 180)) / 100
            print('延时:', timeout)
    
            spanValue = bs(str(spanValue), 'html.parser').get_text()
    
            print(j, spanValue, '--执行到此章节--', time.strftime(ISOTIMEFORMAT, time.localtime())) #打印已经写到哪个章节
            f = open('C:/Users/DZG/Desktop/' + bookName + '.md', 'a', encoding='utf-8');
            f.write('\n\n---')
            f.write('\n\n#### ' + spanValue + '\n\n')
            try:
                chapterCode = urllib.request.urlopen(menuZhangJieHrefList[j]).read()
                result = chardet.detect(chapterCode)  # 检验读取的页面的编码方式
                if (result['confidence'] > 0.5):  # 如果概率大于0.5 即采取这种编码
                    chapterCode = chapterCode.decode(result['encoding'])
                chapterSoup = bs(chapterCode, 'html.parser')  # 使用BS读取解析网页代码
                chapterResult = chapterSoup.find(id='chaptercontent')  # 找到id='chaptercontent'的节点
    
                chapterResultChilds = chapterResult.children
                storySrc = bs(str(list(chapterResultChilds)[3]), "html.parser")
                fileurl = storySrc.contents[0].attrs['src']
    
                fileArticle1 = urllib.request.urlopen(fileurl).read().decode('gb2312', errors='replace')
                fileArticle2 = bs(fileArticle1, 'html.parser');
                fileArticle = str(fileArticle2).replace("document.write('  ", "")\
                    .replace("<p>  ", "  \n\n").replace("</p>", "").replace("');","")\
                    .replace('<a href="http://www.qidian.com">起点'
                             '中文网 www.qidian.com 欢迎广大书友光临阅读,最新、最快、最火的连载作品'
                             '尽在起点原创!</a><a>手机用户请到m.qidian.com阅读。</a>',"---来自DZG")\
                    .replace('<ahref=http: www.qidian.com="">起点'
                             '中文网www.qidian.com欢迎广大书友光临阅读,最新、最快、最火的连载作品'
                             '尽在起点原创!</ahref=http:><a>手机用户请到m.qidian.com阅读。</a>',"---来自DZG")\
                    .replace("起点中文网www.qidian.com欢迎广大书友光临阅读,最新、最快、最火的连载作品"
                             "尽在起点原创!</a><a>手机用户请到m.qidian.com阅读。</a></a>","---来自DZG")\
                    .replace("http://www.qidian.com","http://www.jianshu.com/users/6ca20b30c14c");
                f.write(fileArticle)
            except:
                f.write('章节丢失')
                print('章节丢失')
            j = j + 1
            time.sleep(timeout)
    
    #获取各个章节的URL
    def getHref(menuZhangJieHref):
        menuZhangJieHrefList = []
        for index in menuZhangJieHref:
            menuZhangJieHrefList.append(index.get('href'));
        return menuZhangJieHrefList;
    
    def write(menuContentBs,menu,n):
        i = 0
        for index in menu:
            MuLu = bs(str(index), 'html.parser');
            MLb = MuLu.find('b');
            MuLua = bs(str(MLb), 'html.parser');
            MLa = MuLua.find('a');
            forTitle = bs(str(MLa), 'html.parser');
            muluTitle = MuLua.get_text()
            bookName = forTitle.get_text()
            createBook(bookName)
            createTitle(muluTitle, bookName)
            # MD文件和卷标题写好,就要写入正式的章节目录和文本内容
            if n==0:
                writeZhangjie2(menuContentBs, bookName, i)
            else:
                writeZhangjie(menuContentBs, bookName, i)
            print(i + 1, muluTitle, '--执行到此目录', time.strftime(ISOTIMEFORMAT, time.localtime()))  # 打印已经写入的目录
            i = i + 1;
            # time.sleep(1)
    
    def main():
        #url = "http://read.qidian.com/BookReader/uMyvtX3MUFw1.aspx"  # 日在火影
        #url = "http://read.qidian.com/BookReader/7FnsrjmItYk1.aspx"  # 大时代1958
        url = "http://read.qidian.com/BookReader/nxkadQuAGmYJiWg6PYdjVg2.aspx"  # 绿茵圣骑士
    
        webStory = urllib.request.urlopen(url).read().decode('utf8',errors='replace');          #获取整个网页
    
        print('开始:',time.strftime( ISOTIMEFORMAT, time.localtime() ))
        soup = bs(webStory,'html.parser')                       #解析整个网页
        menuContent = soup.find_all(id="content");              #获取解析好的网页上id位content的元素
        #menuContentStr = ','.join(str(v) for v in menuContent)  #将 menuContent转换为str 如果有多个
        menuContentBs = bs(str(menuContent),'html.parser');       #解析转换好的menuContentStr为bs对象
    
        menu = menuContentBs.find_all("div","box_title")[0:1]
        if('正文卷' in bs(str(menu),'html.parser').get_text()):
            menu = menuContentBs.find_all("div", "box_title")[0:];  # 从解析好的bs对象中获取是div且class为box_title的元素
            write(menuContentBs,menu, 0)
        else:
            menu = menuContentBs.find_all("div", "box_title")[1:];  # 从解析好的bs对象中获取是div且class为box_title的元素
            write(menuContentBs,menu, 1)
        #soup2 = bs(str(menu),'html.parser');                       #转换为bs对象
    
        print('结束:',time.strftime( ISOTIMEFORMAT, time.localtime() ))
    
    #主函数
    if __name__ == '__main__':
        main();
    
    

    下面是爬的样品:
    http://www.jianshu.com/p/a7ae79bd721a

    相关文章

      网友评论

          本文标题:BeautifulSoup4小试牛刀

          本文链接:https://www.haomeiwen.com/subject/yibqsttx.html