mirror of
https://gitee.com/houhuan/TrendRadar.git
synced 2025-12-21 14:27:15 +08:00
v1.4.1
This commit is contained in:
parent
39b9c9a243
commit
0aafa8c3e2
299
main.py
299
main.py
@ -3,6 +3,7 @@
|
||||
import json
|
||||
import time
|
||||
import random
|
||||
import re
|
||||
from datetime import datetime
|
||||
import webbrowser
|
||||
from typing import Dict, List, Tuple, Optional, Union
|
||||
@ -13,7 +14,7 @@ import requests
|
||||
import pytz
|
||||
|
||||
CONFIG = {
|
||||
"VERSION": "1.3.0",
|
||||
"VERSION": "1.4.1",
|
||||
"VERSION_CHECK_URL": "https://raw.githubusercontent.com/sansan0/TrendRadar/refs/heads/master/version",
|
||||
"SHOW_VERSION_UPDATE": True, # 控制显示版本更新提示,改成 False 将不接受新版本提示
|
||||
"FEISHU_MESSAGE_SEPARATOR": "━━━━━━━━━━━━━━━━━━━", # feishu消息分割线
|
||||
@ -24,6 +25,10 @@ CONFIG = {
|
||||
"DEFAULT_PROXY": "http://127.0.0.1:10086",
|
||||
"ENABLE_CRAWLER": True, # 是否启用爬取新闻功能,False时直接停止程序
|
||||
"ENABLE_NOTIFICATION": True, # 是否启用通知功能,False时不发送手机通知
|
||||
"FOCUS_NEW_ONLY": False, # 是否只关注新增新闻,True时只统计和推送新增的新闻(增量推送)
|
||||
# FOCUS_NEW_ONLY 增量推送开关:避免重复推送相同内容,只在有新内容时才发通知
|
||||
# 优点:1.减少重复推送噪音 2.专注最新动态 3.避免通知疲劳
|
||||
# 适用场景:1.高频监控(≤30分钟间隔) 2.实时热点追踪 3.只关心新话题而非持续热度
|
||||
"MESSAGE_BATCH_SIZE": 4000, # 消息分批大小(字节)
|
||||
"BATCH_SEND_INTERVAL": 1, # 批次发送间隔(秒)
|
||||
# 飞书机器人的 webhook URL
|
||||
@ -266,6 +271,35 @@ class DataFetcher:
|
||||
class DataProcessor:
|
||||
"""数据处理器"""
|
||||
|
||||
@staticmethod
|
||||
def clean_title(title: str) -> str:
|
||||
"""清理标题中的特殊字符"""
|
||||
if not isinstance(title, str):
|
||||
title = str(title)
|
||||
|
||||
# 移除或替换常见的特殊字符
|
||||
cleaned_title = title.replace("\n", " ").replace( # 换行符替换为空格
|
||||
"\r", " "
|
||||
) # 回车符替换为空格
|
||||
|
||||
cleaned_title = re.sub(r"\s+", " ", cleaned_title)
|
||||
|
||||
cleaned_title = cleaned_title.strip()
|
||||
|
||||
return cleaned_title
|
||||
|
||||
@staticmethod
|
||||
def is_first_crawl_today() -> bool:
|
||||
"""检测是否是当天第一次爬取"""
|
||||
date_folder = TimeHelper.format_date_folder()
|
||||
txt_dir = Path("output") / date_folder / "txt"
|
||||
|
||||
if not txt_dir.exists():
|
||||
return True
|
||||
|
||||
files = sorted([f for f in txt_dir.iterdir() if f.suffix == ".txt"])
|
||||
return len(files) <= 1 # 0个文件或1个文件都算第一次
|
||||
|
||||
@staticmethod
|
||||
def detect_latest_new_titles(id_to_alias: Dict) -> Dict:
|
||||
"""检测当日最新批次的新增标题"""
|
||||
@ -365,7 +399,7 @@ class DataProcessor:
|
||||
if url_part.endswith("]"):
|
||||
url = url_part[:-1]
|
||||
|
||||
title = title_part.strip()
|
||||
title = DataProcessor.clean_title(title_part.strip())
|
||||
ranks = [rank] if rank is not None else [1]
|
||||
|
||||
titles_by_source[source_name][title] = {
|
||||
@ -394,6 +428,7 @@ class DataProcessor:
|
||||
# 按排名排序标题
|
||||
sorted_titles = []
|
||||
for title, info in title_data.items():
|
||||
cleaned_title = DataProcessor.clean_title(title)
|
||||
if isinstance(info, dict):
|
||||
ranks = info.get("ranks", [])
|
||||
url = info.get("url", "")
|
||||
@ -404,12 +439,12 @@ class DataProcessor:
|
||||
mobile_url = ""
|
||||
|
||||
rank = ranks[0] if ranks else 1
|
||||
sorted_titles.append((rank, title, url, mobile_url))
|
||||
sorted_titles.append((rank, cleaned_title, url, mobile_url))
|
||||
|
||||
sorted_titles.sort(key=lambda x: x[0])
|
||||
|
||||
for rank, title, url, mobile_url in sorted_titles:
|
||||
line = f"{rank}. {title}"
|
||||
for rank, cleaned_title, url, mobile_url in sorted_titles:
|
||||
line = f"{rank}. {cleaned_title}"
|
||||
|
||||
if url:
|
||||
line += f" [URL:{url}]"
|
||||
@ -760,11 +795,45 @@ class StatisticsCalculator:
|
||||
title_info: Optional[Dict] = None,
|
||||
rank_threshold: int = CONFIG["RANK_THRESHOLD"],
|
||||
new_titles: Optional[Dict] = None,
|
||||
focus_new_only: bool = False,
|
||||
) -> Tuple[List[Dict], int]:
|
||||
"""统计词频,支持必须词、频率词、过滤词,并标记新增标题"""
|
||||
|
||||
# 检测是否是当天第一次爬取
|
||||
is_first_today = DataProcessor.is_first_crawl_today()
|
||||
|
||||
# 确定处理的数据源和新增标记逻辑
|
||||
if focus_new_only:
|
||||
if is_first_today:
|
||||
# 新增模式 + 当天第一次:处理所有新闻,都标记为新增
|
||||
results_to_process = results
|
||||
all_news_are_new = True
|
||||
total_input_news = sum(len(titles) for titles in results.values())
|
||||
print(
|
||||
f"新增模式:当天第一次爬取,处理 {total_input_news} 条新闻(所有匹配的新闻都视为新增)"
|
||||
)
|
||||
else:
|
||||
# 新增模式 + 当天非第一次:只处理新增的新闻
|
||||
results_to_process = new_titles if new_titles else {}
|
||||
all_news_are_new = True # 处理的都是新增新闻
|
||||
if new_titles:
|
||||
total_new_count = sum(len(titles) for titles in new_titles.values())
|
||||
print(
|
||||
f"新增模式:检测到 {total_new_count} 条新增新闻,开始进行频率词匹配..."
|
||||
)
|
||||
else:
|
||||
print("新增模式:未检测到新增新闻")
|
||||
else:
|
||||
# 正常模式:处理所有新闻
|
||||
results_to_process = results
|
||||
all_news_are_new = False
|
||||
total_input_news = sum(len(titles) for titles in results.values())
|
||||
print(f"正常模式:处理 {total_input_news} 条新闻")
|
||||
|
||||
word_stats = {}
|
||||
total_titles = 0
|
||||
processed_titles = {}
|
||||
matched_new_count = 0
|
||||
|
||||
if title_info is None:
|
||||
title_info = {}
|
||||
@ -775,7 +844,7 @@ class StatisticsCalculator:
|
||||
group_key = group["group_key"]
|
||||
word_stats[group_key] = {"count": 0, "titles": {}}
|
||||
|
||||
for source_id, titles_data in results.items():
|
||||
for source_id, titles_data in results_to_process.items():
|
||||
total_titles += len(titles_data)
|
||||
|
||||
if source_id not in processed_titles:
|
||||
@ -786,11 +855,17 @@ class StatisticsCalculator:
|
||||
continue
|
||||
|
||||
# 使用统一的匹配逻辑
|
||||
if not StatisticsCalculator._matches_word_groups(
|
||||
matches_frequency_words = StatisticsCalculator._matches_word_groups(
|
||||
title, word_groups, filter_words
|
||||
):
|
||||
)
|
||||
|
||||
if not matches_frequency_words:
|
||||
continue
|
||||
|
||||
# 如果是新增模式,统计匹配的新增新闻数量
|
||||
if focus_new_only and all_news_are_new:
|
||||
matched_new_count += 1
|
||||
|
||||
source_ranks = title_data.get("ranks", [])
|
||||
source_url = title_data.get("url", "")
|
||||
source_mobile_url = title_data.get("mobileUrl", "")
|
||||
@ -853,9 +928,13 @@ class StatisticsCalculator:
|
||||
|
||||
source_alias = id_to_alias.get(source_id, source_id)
|
||||
|
||||
# 修复is_new判断逻辑,添加容错处理
|
||||
# 判断是否为新增
|
||||
is_new = False
|
||||
if new_titles and source_id in new_titles:
|
||||
if all_news_are_new:
|
||||
# 新增模式下所有处理的新闻都是新增,或者当天第一次的所有新闻都是新增
|
||||
is_new = True
|
||||
elif new_titles and source_id in new_titles:
|
||||
# 正常模式下,检查是否在新增列表中
|
||||
new_titles_for_source = new_titles[source_id]
|
||||
if title in new_titles_for_source:
|
||||
is_new = True
|
||||
@ -888,6 +967,21 @@ class StatisticsCalculator:
|
||||
processed_titles[source_id][title] = True
|
||||
break
|
||||
|
||||
if focus_new_only and not is_first_today:
|
||||
if new_titles:
|
||||
total_new_count = sum(len(titles) for titles in new_titles.values())
|
||||
print(
|
||||
f"新增模式:{total_new_count} 条新增新闻中,有 {matched_new_count} 条匹配频率词"
|
||||
)
|
||||
if matched_new_count == 0:
|
||||
print("新增模式:没有新增新闻匹配频率词,将不会发送通知")
|
||||
else:
|
||||
print("新增模式:未检测到新增新闻")
|
||||
elif focus_new_only and is_first_today:
|
||||
print(
|
||||
f"新增模式:当天第一次爬取,{matched_new_count} 条新闻匹配频率词并将推送"
|
||||
)
|
||||
|
||||
stats = []
|
||||
for group_key, data in word_stats.items():
|
||||
all_titles = []
|
||||
@ -1041,17 +1135,54 @@ class ReportGenerator:
|
||||
failed_ids: Optional[List] = None,
|
||||
new_titles: Optional[Dict] = None,
|
||||
id_to_alias: Optional[Dict] = None,
|
||||
hide_new_section: bool = False,
|
||||
) -> Dict:
|
||||
"""准备报告数据"""
|
||||
filtered_new_titles = {}
|
||||
if new_titles and id_to_alias:
|
||||
word_groups, filter_words = DataProcessor.load_frequency_words()
|
||||
for source_id, titles_data in new_titles.items():
|
||||
filtered_titles = ReportGenerator._apply_frequency_filter(
|
||||
titles_data, word_groups, filter_words
|
||||
)
|
||||
if filtered_titles:
|
||||
filtered_new_titles[source_id] = filtered_titles
|
||||
processed_new_titles = []
|
||||
|
||||
# 只有在非隐藏模式下才处理新增新闻部分
|
||||
if not hide_new_section:
|
||||
filtered_new_titles = {}
|
||||
if new_titles and id_to_alias:
|
||||
word_groups, filter_words = DataProcessor.load_frequency_words()
|
||||
for source_id, titles_data in new_titles.items():
|
||||
filtered_titles = ReportGenerator._apply_frequency_filter(
|
||||
titles_data, word_groups, filter_words
|
||||
)
|
||||
if filtered_titles:
|
||||
filtered_new_titles[source_id] = filtered_titles
|
||||
|
||||
if filtered_new_titles and id_to_alias:
|
||||
for source_id, titles_data in filtered_new_titles.items():
|
||||
source_alias = id_to_alias.get(source_id, source_id)
|
||||
source_titles = []
|
||||
|
||||
for title, title_data in titles_data.items():
|
||||
url, mobile_url, ranks = (
|
||||
ReportGenerator._extract_title_data_fields(title_data)
|
||||
)
|
||||
|
||||
processed_title = {
|
||||
"title": title,
|
||||
"source_alias": source_alias,
|
||||
"time_display": "",
|
||||
"count": 1,
|
||||
"ranks": ranks,
|
||||
"rank_threshold": CONFIG["RANK_THRESHOLD"],
|
||||
"url": url,
|
||||
"mobile_url": mobile_url,
|
||||
"is_new": True,
|
||||
}
|
||||
source_titles.append(processed_title)
|
||||
|
||||
if source_titles:
|
||||
processed_new_titles.append(
|
||||
{
|
||||
"source_id": source_id,
|
||||
"source_alias": source_alias,
|
||||
"titles": source_titles,
|
||||
}
|
||||
)
|
||||
|
||||
processed_stats = []
|
||||
for stat in stats:
|
||||
@ -1082,39 +1213,6 @@ class ReportGenerator:
|
||||
}
|
||||
)
|
||||
|
||||
processed_new_titles = []
|
||||
if filtered_new_titles and id_to_alias:
|
||||
for source_id, titles_data in filtered_new_titles.items():
|
||||
source_alias = id_to_alias.get(source_id, source_id)
|
||||
source_titles = []
|
||||
|
||||
for title, title_data in titles_data.items():
|
||||
url, mobile_url, ranks = ReportGenerator._extract_title_data_fields(
|
||||
title_data
|
||||
)
|
||||
|
||||
processed_title = {
|
||||
"title": title,
|
||||
"source_alias": source_alias,
|
||||
"time_display": "",
|
||||
"count": 1,
|
||||
"ranks": ranks,
|
||||
"rank_threshold": CONFIG["RANK_THRESHOLD"],
|
||||
"url": url,
|
||||
"mobile_url": mobile_url,
|
||||
"is_new": True,
|
||||
}
|
||||
source_titles.append(processed_title)
|
||||
|
||||
if source_titles:
|
||||
processed_new_titles.append(
|
||||
{
|
||||
"source_id": source_id,
|
||||
"source_alias": source_alias,
|
||||
"titles": source_titles,
|
||||
}
|
||||
)
|
||||
|
||||
return {
|
||||
"stats": processed_stats,
|
||||
"new_titles": processed_new_titles,
|
||||
@ -1170,7 +1268,9 @@ class ReportGenerator:
|
||||
)
|
||||
|
||||
link_url = title_data["mobile_url"] or title_data["url"]
|
||||
escaped_title = ReportGenerator._html_escape(title_data["title"])
|
||||
|
||||
cleaned_title = DataProcessor.clean_title(title_data["title"])
|
||||
escaped_title = ReportGenerator._html_escape(cleaned_title)
|
||||
escaped_source_alias = ReportGenerator._html_escape(title_data["source_alias"])
|
||||
|
||||
if link_url:
|
||||
@ -1357,10 +1457,13 @@ class ReportGenerator:
|
||||
)
|
||||
|
||||
link_url = title_data["mobile_url"] or title_data["url"]
|
||||
|
||||
cleaned_title = DataProcessor.clean_title(title_data["title"])
|
||||
|
||||
if link_url:
|
||||
formatted_title = f"[{title_data['title']}]({link_url})"
|
||||
formatted_title = f"[{cleaned_title}]({link_url})"
|
||||
else:
|
||||
formatted_title = title_data["title"]
|
||||
formatted_title = cleaned_title
|
||||
|
||||
title_prefix = "🆕 " if title_data["is_new"] else ""
|
||||
|
||||
@ -1386,10 +1489,13 @@ class ReportGenerator:
|
||||
)
|
||||
|
||||
link_url = title_data["mobile_url"] or title_data["url"]
|
||||
|
||||
cleaned_title = DataProcessor.clean_title(title_data["title"])
|
||||
|
||||
if link_url:
|
||||
formatted_title = f"[{title_data['title']}]({link_url})"
|
||||
formatted_title = f"[{cleaned_title}]({link_url})"
|
||||
else:
|
||||
formatted_title = title_data["title"]
|
||||
formatted_title = cleaned_title
|
||||
|
||||
title_prefix = "🆕 " if title_data["is_new"] else ""
|
||||
|
||||
@ -1415,10 +1521,13 @@ class ReportGenerator:
|
||||
)
|
||||
|
||||
link_url = title_data["mobile_url"] or title_data["url"]
|
||||
|
||||
cleaned_title = DataProcessor.clean_title(title_data["title"])
|
||||
|
||||
if link_url:
|
||||
formatted_title = f"[{title_data['title']}]({link_url})"
|
||||
formatted_title = f"[{cleaned_title}]({link_url})"
|
||||
else:
|
||||
formatted_title = title_data["title"]
|
||||
formatted_title = cleaned_title
|
||||
|
||||
title_prefix = "🆕 " if title_data["is_new"] else ""
|
||||
|
||||
@ -1444,10 +1553,13 @@ class ReportGenerator:
|
||||
)
|
||||
|
||||
link_url = title_data["mobile_url"] or title_data["url"]
|
||||
|
||||
cleaned_title = DataProcessor.clean_title(title_data["title"])
|
||||
|
||||
if link_url:
|
||||
formatted_title = f'<a href="{link_url}">{ReportGenerator._html_escape(title_data["title"])}</a>'
|
||||
formatted_title = f'<a href="{link_url}">{ReportGenerator._html_escape(cleaned_title)}</a>'
|
||||
else:
|
||||
formatted_title = title_data["title"]
|
||||
formatted_title = cleaned_title
|
||||
|
||||
title_prefix = "🆕 " if title_data["is_new"] else ""
|
||||
|
||||
@ -1988,13 +2100,14 @@ class ReportGenerator:
|
||||
id_to_alias: Optional[Dict] = None,
|
||||
update_info: Optional[Dict] = None,
|
||||
proxy_url: Optional[str] = None,
|
||||
hide_new_section: bool = False,
|
||||
) -> Dict[str, bool]:
|
||||
"""发送数据到多个webhook平台"""
|
||||
results = {}
|
||||
|
||||
# 数据处理层
|
||||
report_data = ReportGenerator._prepare_report_data(
|
||||
stats, failed_ids, new_titles, id_to_alias
|
||||
stats, failed_ids, new_titles, id_to_alias, hide_new_section
|
||||
)
|
||||
|
||||
# 获取环境变量中的webhook配置
|
||||
@ -2336,6 +2449,21 @@ class NewsAnalyzer:
|
||||
except Exception as e:
|
||||
print(f"版本检查出错: {e}")
|
||||
|
||||
def _has_valid_content(
|
||||
self, stats: List[Dict], new_titles: Optional[Dict] = None
|
||||
) -> bool:
|
||||
"""检查是否有有效的新闻内容"""
|
||||
if CONFIG["FOCUS_NEW_ONLY"]:
|
||||
# 新增模式下,只要stats有内容就说明有新增的匹配新闻
|
||||
return any(stat["count"] > 0 for stat in stats)
|
||||
else:
|
||||
# 正常模式下,检查是否有匹配的频率词新闻或新增新闻
|
||||
has_matched_news = any(stat["count"] > 0 for stat in stats)
|
||||
has_new_news = bool(
|
||||
new_titles and any(len(titles) > 0 for titles in new_titles.values())
|
||||
)
|
||||
return has_matched_news or has_new_news
|
||||
|
||||
def generate_daily_summary(self) -> Optional[str]:
|
||||
"""生成当日统计报告"""
|
||||
print("生成当日统计报告...")
|
||||
@ -2349,11 +2477,7 @@ class NewsAnalyzer:
|
||||
total_titles = sum(len(titles) for titles in all_results.values())
|
||||
print(f"读取到 {total_titles} 个标题")
|
||||
|
||||
latest_new_titles = DataProcessor.detect_latest_new_titles(id_to_alias)
|
||||
if latest_new_titles:
|
||||
total_new_count = sum(len(titles) for titles in latest_new_titles.values())
|
||||
print(f"检测到 {total_new_count} 条最新新增新闻")
|
||||
|
||||
new_titles = DataProcessor.detect_latest_new_titles(id_to_alias)
|
||||
word_groups, filter_words = DataProcessor.load_frequency_words()
|
||||
|
||||
stats, total_titles = StatisticsCalculator.count_word_frequency(
|
||||
@ -2363,14 +2487,15 @@ class NewsAnalyzer:
|
||||
id_to_alias,
|
||||
title_info,
|
||||
self.rank_threshold,
|
||||
latest_new_titles,
|
||||
new_titles,
|
||||
focus_new_only=CONFIG["FOCUS_NEW_ONLY"],
|
||||
)
|
||||
|
||||
html_file = ReportGenerator.generate_html_report(
|
||||
stats,
|
||||
total_titles,
|
||||
is_daily=True,
|
||||
new_titles=latest_new_titles,
|
||||
new_titles=new_titles,
|
||||
id_to_alias=id_to_alias,
|
||||
)
|
||||
print(f"当日HTML统计报告已生成: {html_file}")
|
||||
@ -2392,20 +2517,33 @@ class NewsAnalyzer:
|
||||
CONFIG["ENABLE_NOTIFICATION"]
|
||||
and has_webhook
|
||||
and self.report_type in ["daily", "both"]
|
||||
and self._has_valid_content(stats, new_titles)
|
||||
):
|
||||
hide_new_section = CONFIG["FOCUS_NEW_ONLY"]
|
||||
|
||||
ReportGenerator.send_to_webhooks(
|
||||
stats,
|
||||
[],
|
||||
"当日汇总",
|
||||
latest_new_titles,
|
||||
new_titles,
|
||||
id_to_alias,
|
||||
self.update_info,
|
||||
self.proxy_url,
|
||||
hide_new_section=hide_new_section,
|
||||
)
|
||||
elif CONFIG["ENABLE_NOTIFICATION"] and not has_webhook:
|
||||
print("⚠️ 警告:通知功能已启用但未配置webhook URL,将跳过通知发送")
|
||||
elif not CONFIG["ENABLE_NOTIFICATION"]:
|
||||
print("跳过当日汇总通知:通知功能已禁用")
|
||||
elif (
|
||||
CONFIG["ENABLE_NOTIFICATION"]
|
||||
and has_webhook
|
||||
and not self._has_valid_content(stats, new_titles)
|
||||
):
|
||||
if CONFIG["FOCUS_NEW_ONLY"]:
|
||||
print("跳过当日汇总通知:新增模式下未检测到匹配的新增新闻")
|
||||
else:
|
||||
print("跳过当日汇总通知:未匹配到有效的新闻内容")
|
||||
|
||||
return html_file
|
||||
|
||||
@ -2441,6 +2579,11 @@ class NewsAnalyzer:
|
||||
|
||||
print(f"报告类型: {self.report_type}")
|
||||
|
||||
if CONFIG["FOCUS_NEW_ONLY"]:
|
||||
print("运行模式: 新增检测模式(只关注新增新闻)")
|
||||
else:
|
||||
print("运行模式: 正常模式(频率词统计 + 新增检测)")
|
||||
|
||||
ids = [
|
||||
("toutiao", "今日头条"),
|
||||
("baidu", "百度热搜"),
|
||||
@ -2496,14 +2639,18 @@ class NewsAnalyzer:
|
||||
title_info,
|
||||
self.rank_threshold,
|
||||
new_titles,
|
||||
focus_new_only=CONFIG["FOCUS_NEW_ONLY"],
|
||||
)
|
||||
|
||||
# 只有启用通知且配置了webhook时才发送通知
|
||||
# 只有启用通知且配置了webhook且有有效内容时才发送通知
|
||||
if (
|
||||
CONFIG["ENABLE_NOTIFICATION"]
|
||||
and has_webhook
|
||||
and self.report_type in ["current", "both"]
|
||||
and self._has_valid_content(stats, new_titles)
|
||||
):
|
||||
hide_new_section = CONFIG["FOCUS_NEW_ONLY"]
|
||||
|
||||
ReportGenerator.send_to_webhooks(
|
||||
stats,
|
||||
failed_ids,
|
||||
@ -2512,11 +2659,21 @@ class NewsAnalyzer:
|
||||
id_to_alias,
|
||||
self.update_info,
|
||||
self.proxy_url,
|
||||
hide_new_section=hide_new_section,
|
||||
)
|
||||
elif CONFIG["ENABLE_NOTIFICATION"] and not has_webhook:
|
||||
print("⚠️ 警告:通知功能已启用但未配置webhook URL,将跳过通知发送")
|
||||
elif not CONFIG["ENABLE_NOTIFICATION"]:
|
||||
print("跳过单次爬取通知:通知功能已禁用")
|
||||
elif (
|
||||
CONFIG["ENABLE_NOTIFICATION"]
|
||||
and has_webhook
|
||||
and not self._has_valid_content(stats, new_titles)
|
||||
):
|
||||
if CONFIG["FOCUS_NEW_ONLY"]:
|
||||
print("跳过单次爬取通知:新增模式下未检测到匹配的新增新闻")
|
||||
else:
|
||||
print("跳过单次爬取通知:未匹配到有效的新闻内容")
|
||||
|
||||
html_file = ReportGenerator.generate_html_report(
|
||||
stats, total_titles, failed_ids, False, new_titles, id_to_alias
|
||||
|
||||
Loading…
Reference in New Issue
Block a user