396 lines
18 KiB
Python
396 lines
18 KiB
Python
# 声明:本代码仅供学习和研究目的使用。使用者应遵守以下原则:
|
||
# 1. 不得用于任何商业用途。
|
||
# 2. 使用时应遵守目标平台的使用条款和robots.txt规则。
|
||
# 3. 不得进行大规模爬取或对平台造成运营干扰。
|
||
# 4. 应合理控制请求频率,避免给目标平台带来不必要的负担。
|
||
# 5. 不得用于任何非法或不当的用途。
|
||
#
|
||
# 详细许可条款请参阅项目根目录下的LICENSE文件。
|
||
# 使用本代码即表示您同意遵守上述原则和LICENSE中的所有条款。
|
||
|
||
|
||
# -*- coding: utf-8 -*-
|
||
import html
|
||
import json
|
||
import re
|
||
from typing import Dict, List, Tuple
|
||
from urllib.parse import parse_qs, unquote
|
||
|
||
from parsel import Selector
|
||
|
||
from constant import baidu_tieba as const
|
||
from model.m_baidu_tieba import TiebaComment, TiebaCreator, TiebaNote
|
||
from tools import utils
|
||
|
||
GENDER_MALE = "sex_male"
|
||
GENDER_FEMALE = "sex_female"
|
||
|
||
|
||
class TieBaExtractor:
|
||
def __init__(self):
|
||
pass
|
||
|
||
@staticmethod
|
||
def extract_search_note_list(page_content: str) -> List[TiebaNote]:
|
||
"""
|
||
提取贴吧帖子列表,这里提取的关键词搜索结果页的数据,还缺少帖子的回复数和回复页等数据
|
||
Args:
|
||
page_content: 页面内容的HTML字符串
|
||
|
||
Returns:
|
||
包含帖子信息的字典列表
|
||
"""
|
||
xpath_selector = "//div[@class='s_post']"
|
||
post_list = Selector(text=page_content).xpath(xpath_selector)
|
||
result: List[TiebaNote] = []
|
||
for post in post_list:
|
||
tieba_note = TiebaNote(note_id=post.xpath(".//span[@class='p_title']/a/@data-tid").get(default='').strip(),
|
||
title=post.xpath(".//span[@class='p_title']/a/text()").get(default='').strip(),
|
||
desc=post.xpath(".//div[@class='p_content']/text()").get(default='').strip(),
|
||
note_url=const.TIEBA_URL + post.xpath(".//span[@class='p_title']/a/@href").get(
|
||
default=''),
|
||
user_nickname=post.xpath(".//a[starts-with(@href, '/home/main')]/font/text()").get(
|
||
default='').strip(), user_link=const.TIEBA_URL + post.xpath(
|
||
".//a[starts-with(@href, '/home/main')]/@href").get(default=''),
|
||
tieba_name=post.xpath(".//a[@class='p_forum']/font/text()").get(default='').strip(),
|
||
tieba_link=const.TIEBA_URL + post.xpath(".//a[@class='p_forum']/@href").get(
|
||
default=''),
|
||
publish_time=post.xpath(".//font[@class='p_green p_date']/text()").get(
|
||
default='').strip(), )
|
||
result.append(tieba_note)
|
||
return result
|
||
|
||
def extract_tieba_note_list(self, page_content: str) -> List[TiebaNote]:
|
||
"""
|
||
提取贴吧帖子列表
|
||
Args:
|
||
page_content:
|
||
|
||
Returns:
|
||
|
||
"""
|
||
page_content = page_content.replace('<!--', "")
|
||
content_selector = Selector(text=page_content)
|
||
xpath_selector = "//ul[@id='thread_list']/li"
|
||
post_list = content_selector.xpath(xpath_selector)
|
||
result: List[TiebaNote] = []
|
||
for post_selector in post_list:
|
||
post_field_value: Dict = self.extract_data_field_value(post_selector)
|
||
if not post_field_value:
|
||
continue
|
||
note_id = str(post_field_value.get("id"))
|
||
tieba_note = TiebaNote(note_id=note_id,
|
||
title=post_selector.xpath(".//a[@class='j_th_tit ']/text()").get(default='').strip(),
|
||
desc=post_selector.xpath(
|
||
".//div[@class='threadlist_abs threadlist_abs_onlyline ']/text()").get(
|
||
default='').strip(), note_url=const.TIEBA_URL + f"/p/{note_id}",
|
||
user_link=const.TIEBA_URL + post_selector.xpath(
|
||
".//a[@class='frs-author-name j_user_card ']/@href").get(default='').strip(),
|
||
user_nickname=post_field_value.get("authoer_nickname") or post_field_value.get(
|
||
"author_name"),
|
||
tieba_name=content_selector.xpath("//a[@class='card_title_fname']/text()").get(
|
||
default='').strip(), tieba_link=const.TIEBA_URL + content_selector.xpath(
|
||
"//a[@class='card_title_fname']/@href").get(default=''),
|
||
total_replay_num=post_field_value.get("reply_num", 0))
|
||
result.append(tieba_note)
|
||
return result
|
||
|
||
def extract_note_detail(self, page_content: str) -> TiebaNote:
|
||
"""
|
||
提取贴吧帖子详情
|
||
Args:
|
||
page_content:
|
||
|
||
Returns:
|
||
|
||
"""
|
||
content_selector = Selector(text=page_content)
|
||
first_floor_selector = content_selector.xpath("//div[@class='p_postlist'][1]")
|
||
only_view_author_link = content_selector.xpath("//*[@id='lzonly_cntn']/@href").get(default='').strip()
|
||
note_id = only_view_author_link.split("?")[0].split("/")[-1]
|
||
# 帖子回复数、回复页数
|
||
thread_num_infos = content_selector.xpath(
|
||
"//div[@id='thread_theme_5']//li[@class='l_reply_num']//span[@class='red']")
|
||
# IP地理位置、发表时间
|
||
other_info_content = content_selector.xpath(".//div[@class='post-tail-wrap']").get(default="").strip()
|
||
ip_location, publish_time = self.extract_ip_and_pub_time(other_info_content)
|
||
note = TiebaNote(note_id=note_id, title=content_selector.xpath("//title/text()").get(default='').strip(),
|
||
desc=content_selector.xpath("//meta[@name='description']/@content").get(default='').strip(),
|
||
note_url=const.TIEBA_URL + f"/p/{note_id}",
|
||
user_link=const.TIEBA_URL + first_floor_selector.xpath(
|
||
".//a[@class='p_author_face ']/@href").get(default='').strip(),
|
||
user_nickname=first_floor_selector.xpath(
|
||
".//a[@class='p_author_name j_user_card']/text()").get(default='').strip(),
|
||
user_avatar=first_floor_selector.xpath(".//a[@class='p_author_face ']/img/@src").get(
|
||
default='').strip(),
|
||
tieba_name=content_selector.xpath("//a[@class='card_title_fname']/text()").get(
|
||
default='').strip(), tieba_link=const.TIEBA_URL + content_selector.xpath(
|
||
"//a[@class='card_title_fname']/@href").get(default=''), ip_location=ip_location,
|
||
publish_time=publish_time,
|
||
total_replay_num=thread_num_infos[0].xpath("./text()").get(default='').strip(),
|
||
total_replay_page=thread_num_infos[1].xpath("./text()").get(default='').strip(), )
|
||
note.title = note.title.replace(f"【{note.tieba_name}】_百度贴吧", "")
|
||
return note
|
||
|
||
def extract_tieba_note_parment_comments(self, page_content: str, note_id: str) -> List[TiebaComment]:
|
||
"""
|
||
提取贴吧帖子一级评论
|
||
Args:
|
||
page_content:
|
||
note_id:
|
||
|
||
Returns:
|
||
|
||
"""
|
||
xpath_selector = "//div[@class='l_post l_post_bright j_l_post clearfix ']"
|
||
comment_list = Selector(text=page_content).xpath(xpath_selector)
|
||
result: List[TiebaComment] = []
|
||
for comment_selector in comment_list:
|
||
comment_field_value: Dict = self.extract_data_field_value(comment_selector)
|
||
if not comment_field_value:
|
||
continue
|
||
tieba_name = comment_selector.xpath("//a[@class='card_title_fname']/text()").get(default='').strip()
|
||
other_info_content = comment_selector.xpath(".//div[@class='post-tail-wrap']").get(default="").strip()
|
||
ip_location, publish_time = self.extract_ip_and_pub_time(other_info_content)
|
||
tieba_comment = TiebaComment(comment_id=str(comment_field_value.get("content").get("post_id")),
|
||
sub_comment_count=comment_field_value.get("content").get("comment_num"),
|
||
content=utils.extract_text_from_html(
|
||
comment_field_value.get("content").get("content")),
|
||
note_url=const.TIEBA_URL + f"/p/{note_id}",
|
||
user_link=const.TIEBA_URL + comment_selector.xpath(
|
||
".//a[@class='p_author_face ']/@href").get(default='').strip(),
|
||
user_nickname=comment_selector.xpath(
|
||
".//a[@class='p_author_name j_user_card']/text()").get(default='').strip(),
|
||
user_avatar=comment_selector.xpath(
|
||
".//a[@class='p_author_face ']/img/@src").get(default='').strip(),
|
||
tieba_id=str(comment_field_value.get("content").get("forum_id", "")),
|
||
tieba_name=tieba_name, tieba_link=f"https://tieba.baidu.com/f?kw={tieba_name}",
|
||
ip_location=ip_location, publish_time=publish_time, note_id=note_id, )
|
||
result.append(tieba_comment)
|
||
return result
|
||
|
||
def extract_tieba_note_sub_comments(self, page_content: str, parent_comment: TiebaComment) -> List[TiebaComment]:
|
||
"""
|
||
提取贴吧帖子二级评论
|
||
Args:
|
||
page_content:
|
||
parent_comment:
|
||
|
||
Returns:
|
||
|
||
"""
|
||
selector = Selector(page_content)
|
||
comments = []
|
||
comment_ele_list = selector.xpath("//li[@class='lzl_single_post j_lzl_s_p first_no_border']")
|
||
comment_ele_list.extend(selector.xpath("//li[@class='lzl_single_post j_lzl_s_p ']"))
|
||
for comment_ele in comment_ele_list:
|
||
comment_value = self.extract_data_field_value(comment_ele)
|
||
if not comment_value:
|
||
continue
|
||
comment_user_a_selector = comment_ele.xpath("./a[@class='j_user_card lzl_p_p']")[0]
|
||
content = utils.extract_text_from_html(
|
||
comment_ele.xpath(".//span[@class='lzl_content_main']").get(default=""))
|
||
comment = TiebaComment(
|
||
comment_id=str(comment_value.get("spid")), content=content,
|
||
user_link=comment_user_a_selector.xpath("./@href").get(default=""),
|
||
user_nickname=comment_value.get("showname"),
|
||
user_avatar=comment_user_a_selector.xpath("./img/@src").get(default=""),
|
||
publish_time=comment_ele.xpath(".//span[@class='lzl_time']/text()").get(default="").strip(),
|
||
parent_comment_id=parent_comment.comment_id,
|
||
note_id=parent_comment.note_id, note_url=parent_comment.note_url,
|
||
tieba_id=parent_comment.tieba_id, tieba_name=parent_comment.tieba_name,
|
||
tieba_link=parent_comment.tieba_link)
|
||
comments.append(comment)
|
||
|
||
return comments
|
||
|
||
def extract_creator_info(self, html_content: str) -> TiebaCreator:
|
||
"""
|
||
提取贴吧创作者信息
|
||
Args:
|
||
html_content:
|
||
|
||
Returns:
|
||
|
||
"""
|
||
selector = Selector(text=html_content)
|
||
user_link_selector = selector.xpath("//p[@class='space']/a")
|
||
user_link: str = user_link_selector.xpath("./@href").get(default='')
|
||
user_link_params: Dict = parse_qs(unquote(user_link.split("?")[-1]))
|
||
user_name = user_link_params.get("un")[0] if user_link_params.get("un") else ""
|
||
user_id = user_link_params.get("id")[0] if user_link_params.get("id") else ""
|
||
userinfo_userdata_selector = selector.xpath("//div[@class='userinfo_userdata']")
|
||
follow_fans_selector = selector.xpath("//span[@class='concern_num']")
|
||
follows, fans = 0, 0
|
||
if len(follow_fans_selector) == 2:
|
||
follows, fans = self.extract_follow_and_fans(follow_fans_selector)
|
||
user_content = userinfo_userdata_selector.get(default='')
|
||
return TiebaCreator(user_id=user_id, user_name=user_name,
|
||
nickname=selector.xpath(".//span[@class='userinfo_username ']/text()").get(
|
||
default='').strip(),
|
||
avatar=selector.xpath(".//div[@class='userinfo_left_head']//img/@src").get(
|
||
default='').strip(),
|
||
gender=self.extract_gender(user_content),
|
||
ip_location=self.extract_ip(user_content),
|
||
follows=follows,
|
||
fans=fans,
|
||
registration_duration=self.extract_registration_duration(user_content)
|
||
)
|
||
|
||
def extract_ip_and_pub_time(self, html_content: str) -> Tuple[str, str]:
|
||
"""
|
||
提取IP位置和发布时间
|
||
Args:
|
||
html_content:
|
||
|
||
Returns:
|
||
|
||
"""
|
||
pattern_pub_time = re.compile(r'<span class="tail-info">(\d{4}-\d{2}-\d{2} \d{2}:\d{2})</span>')
|
||
time_match = pattern_pub_time.search(html_content)
|
||
pub_time = time_match.group(1) if time_match else ""
|
||
return self.extract_ip(html_content), pub_time
|
||
|
||
@staticmethod
|
||
def extract_ip(html_content: str) -> str:
|
||
"""
|
||
提取IP
|
||
Args:
|
||
html_content:
|
||
|
||
Returns:
|
||
|
||
"""
|
||
pattern_ip = re.compile(r'IP属地:(\S+)</span>')
|
||
ip_match = pattern_ip.search(html_content)
|
||
ip = ip_match.group(1) if ip_match else ""
|
||
return ip
|
||
|
||
@staticmethod
|
||
def extract_gender(html_content: str) -> str:
|
||
"""
|
||
提取性别
|
||
Args:
|
||
html_content:
|
||
|
||
Returns:
|
||
|
||
"""
|
||
if GENDER_MALE in html_content:
|
||
return '男'
|
||
elif GENDER_FEMALE in html_content:
|
||
return '女'
|
||
return '未知'
|
||
|
||
@staticmethod
|
||
def extract_follow_and_fans(selectors: List[Selector]) -> Tuple[str, str]:
|
||
"""
|
||
提取关注数和粉丝数
|
||
Args:
|
||
selectors:
|
||
|
||
Returns:
|
||
|
||
"""
|
||
pattern = re.compile(r'<span class="concern_num">\(<a[^>]*>(\d+)</a>\)</span>')
|
||
follow_match = pattern.findall(selectors[0].get())
|
||
fans_match = pattern.findall(selectors[1].get())
|
||
follows = follow_match[0] if follow_match else 0
|
||
fans = fans_match[0] if fans_match else 0
|
||
return follows, fans
|
||
|
||
@staticmethod
|
||
def extract_registration_duration(html_content: str) -> str:
|
||
"""
|
||
"<span>吧龄:1.9年</span>"
|
||
Returns: 1.9年
|
||
|
||
"""
|
||
pattern = re.compile(r'<span>吧龄:(\S+)</span>')
|
||
match = pattern.search(html_content)
|
||
return match.group(1) if match else ""
|
||
|
||
@staticmethod
|
||
def extract_data_field_value(selector: Selector) -> Dict:
|
||
"""
|
||
提取data-field的值
|
||
Args:
|
||
selector:
|
||
|
||
Returns:
|
||
|
||
"""
|
||
data_field_value = selector.xpath("./@data-field").get(default='').strip()
|
||
if not data_field_value or data_field_value == "{}":
|
||
return {}
|
||
try:
|
||
# 先使用 html.unescape 处理转义字符 再json.loads 将 JSON 字符串转换为 Python 字典
|
||
unescaped_json_str = html.unescape(data_field_value)
|
||
data_field_dict_value = json.loads(unescaped_json_str)
|
||
except Exception as ex:
|
||
print(f"extract_data_field_value,错误信息:{ex}, 尝试使用其他方式解析")
|
||
data_field_dict_value = {}
|
||
return data_field_dict_value
|
||
|
||
|
||
def test_extract_search_note_list():
|
||
with open("test_data/search_keyword_notes.html", "r", encoding="utf-8") as f:
|
||
content = f.read()
|
||
extractor = TieBaExtractor()
|
||
result = extractor.extract_search_note_list(content)
|
||
print(result)
|
||
|
||
|
||
def test_extract_note_detail():
|
||
with open("test_data/note_detail.html", "r", encoding="utf-8") as f:
|
||
content = f.read()
|
||
extractor = TieBaExtractor()
|
||
result = extractor.extract_note_detail(content)
|
||
print(result.model_dump())
|
||
|
||
|
||
def test_extract_tieba_note_parment_comments():
|
||
with open("test_data/note_comments.html", "r", encoding="utf-8") as f:
|
||
content = f.read()
|
||
extractor = TieBaExtractor()
|
||
result = extractor.extract_tieba_note_parment_comments(content, "123456")
|
||
print(result)
|
||
|
||
|
||
def test_extract_tieba_note_sub_comments():
|
||
with open("test_data/note_sub_comments.html", "r", encoding="utf-8") as f:
|
||
content = f.read()
|
||
extractor = TieBaExtractor()
|
||
fake_parment_comment = TiebaComment(comment_id="123456", content="content", user_link="user_link",
|
||
user_nickname="user_nickname", user_avatar="user_avatar",
|
||
publish_time="publish_time", parent_comment_id="parent_comment_id",
|
||
note_id="note_id", note_url="note_url", tieba_id="tieba_id",
|
||
tieba_name="tieba_name", )
|
||
result = extractor.extract_tieba_note_sub_comments(content, fake_parment_comment)
|
||
print(result)
|
||
|
||
|
||
def test_extract_tieba_note_list():
|
||
with open("test_data/tieba_note_list.html", "r", encoding="utf-8") as f:
|
||
content = f.read()
|
||
extractor = TieBaExtractor()
|
||
result = extractor.extract_tieba_note_list(content)
|
||
print(result)
|
||
pass
|
||
|
||
|
||
def test_extract_creator_info():
|
||
with open("test_data/creator_info.html", "r", encoding="utf-8") as f:
|
||
content = f.read()
|
||
extractor = TieBaExtractor()
|
||
result = extractor.extract_creator_info(content)
|
||
print(result.model_dump_json())
|
||
|
||
|
||
if __name__ == '__main__':
|
||
# test_extract_search_note_list()
|
||
# test_extract_note_detail()
|
||
# test_extract_tieba_note_parment_comments()
|
||
# test_extract_tieba_note_list()
|
||
test_extract_creator_info()
|