Browse Source

Merge branch 'bugfix/20250728-bugfix' of Server/LongArticleTaskServer into master

luojunhui 1 month ago
parent
commit
6679130e1f

+ 1 - 1
app_config.toml

@@ -3,4 +3,4 @@ bind = "0.0.0.0:6060"
 workers = 6
 keep_alive_timeout = 120  # 保持连接的最大秒数,根据需要调整
 graceful_timeout = 30    # 重启或停止之前等待当前工作完成的时间
-loglevel = "warning"  # 日志级别
+loglevel = "warning"  # 日志级别

+ 3 - 0
applications/tasks/crawler_tasks/crawler_toutiao.py

@@ -85,12 +85,14 @@ class CrawlerToutiao(CrawlerPipeline, CrawlerToutiaoConst):
             key="toutiao_blogger_cookie", output_type="string"
         )
         while has_more:
+            print(account_id, max_cursor)
             response = await get_toutiao_account_info_list(
                 account_id=account_id,
                 cookie=cookie,
                 media_type=media_type,
                 max_behot_time=current_cursor,
             )
+            print(response)
             if not response:
                 break
 
@@ -118,6 +120,7 @@ class CrawlerToutiao(CrawlerPipeline, CrawlerToutiaoConst):
                     raise Exception(f"unknown media type: {media_type}")
 
             crawler_info_list_bar = tqdm(info_list, desc=bar_description)
+            print(json.dumps(info_list, ensure_ascii=False, indent=4))
             for info in crawler_info_list_bar:
                 try:
                     crawler_info_list_bar.set_postfix({"id": info["id"]})

+ 3 - 0
applications/tasks/data_recycle_tasks/recycle_daily_publish_articles.py

@@ -216,6 +216,9 @@ class CheckDailyPublishArticlesTask(RecycleDailyPublishArticlesTask):
                 fail_list.append(second_task)
 
         if fail_list:
+            now = datetime.datetime.now()
+            if now.hour < 20:
+                return
             columns = [
                 feishu_robot.create_feishu_columns_sheet(
                     sheet_type="plain_text",