Python爬虫之爬取小说(二)
合并章节生成一个TXT文件
Description
改进原因:
原本每个章节生成一个TXT文件,不方便阅读。
过程中遇到的问题:
1、文件名排序出错,使用os.listdir后排序有问题
解决方法:将文件名改为统一形式,按照前两位数字sort排序,在debug中观察出错点。
2、将合并TXT的代码加入原本爬虫代码中,文件夹为同一文件夹,注意文件夹路径调整。
My solution
# 爬取小说每个章节的内容并放进同一个txt文件内
# 《一千八百昼》,作者:蟹总
# 网站:https://www.35xs.com/book/276274/
# charset = utf-8
import requests
import os
import time
from bs4 import BeautifulSoup
import os
headers = {
'User-Agent':'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/63.0.3239.84 Safari/537.36'
}
# 获取小说各个章节的名字和url
def getChapterInfo(novel_url):
chapter_html = requests.get(novel_url, headers=headers).text
soup = BeautifulSoup(chapter_html,'lxml')
soup2 = soup.find('ul',{'class':'mulu_list'})
chapter_list = soup2.find_all('li') # 先找到'li'
chapter_all_dict = {}
for each in chapter_list:
import re
chapter_each = {}
chapter_each['title'] = each.find('a').get_text()
chapter_each['chapter_url'] = 'http://www.35xs.com'+each.find('a')['href']
chapter_num = int(re.findall('\d+',each.get_text())[0])
chapter_all_dict[chapter_num] = chapter_each # 记录到所有的章节的字典中保存,返回字典
return chapter_all_dict
# 获取小说章节内容,写入文本
def getChapterContent(each_chapter_dict):
content_html = requests.get(each_chapter_dict['chapter_url'],headers=headers).text
soup = BeautifulSoup(content_html,'lxml')
content_tag = soup.find('div',{'class':'bookreadercontent'})
p_tag = content_tag.find_all('p')
print('正在保存的章节-->'+ each_chapter_dict['title'])
for each in p_tag: # 正文内容
paragraph = each.get_text().strip() # 去掉首尾字符
with open(each_chapter_dict['title']+r'.txt','a',encoding = 'utf8')as f:
f.write(''+ paragraph + '\n\n')
f.close()
if __name__ == '__main__':
start = time.clock() # 程序运行起始时间
novel_url = 'https://www.35xs.com/book/276274/'
novel_info = getChapterInfo(novel_url)
dir_name = 'test'
if not os.path.exists(dir_name):
os.mkdir(dir_name)
os.chdir(dir_name)
for each in novel_info:
getChapterContent(novel_info[each])
end = time.clock() # 记录程序结束时间
print('保存结束,共保存 %d 章,花费时间:%f s'%(len(novel_info),(end-start)))
# 获取目标文件夹的路径
meragefiledir = os.getcwd() # 当前文件夹中的test文件夹
# 获取当前文件夹中的文件名称列表
filenames = os.listdir(meragefiledir)
# 打开当前目录下的result.txt文件,如果没有则创建
def fn(filenames):
x = []
for i in range(0,len(filenames)):
if filenames[i][1].isdigit():
i+=1
else:
os.rename(os.path.join(meragefiledir,filenames[i]),os.path.join(meragefiledir,'0'+filenames[i]))
fn(filenames)
filenames2 = os.listdir(meragefiledir)
filenames2.sort()
filenames2.sort(key= lambda x:int(x[0:2])) # 文件名按数字大小排序
file = open('result.txt','w')
# 向文件中写入字符
# 先遍历文件名
for filename in filenames2:
filepath = meragefiledir + '/'
filepath = filepath + filename
# 遍历单个文件,读取行数
for line in open(filepath):
file.writelines(line)
file.write('\n')
# 关闭文件
file.close()