Jelajahi Sumber

feat:添加日志

zhaohaipeng 1 bulan lalu
induk
melakukan
8d3c05088c

+ 9 - 9
enums/automation_job.py

@@ -2,15 +2,15 @@ from enum import Enum
 
 
 class AutomationJobCronInfo(Enum):
-    account_top = ("account_top", "top", 9)
-    account_extend_top = ("account_extend_top", "top", 9)
-    channel_topic_top = ("channel_topic_top", "top", 9)
-    channel_topic_extend_top = ("channel_topic_extend_top", "top", 9)
-    channel_image_search_video_top = ("channel_image_search_video_top", "top", 9)
-    channel_image_search_topic_top = ("channel_image_search_topic_top", "top", 9)
-    channel_image_search_topic_extend_top = ("channel_image_search_topic_extend_top", "top", 9)
-    channel_image_search_video_all_cate_top = ("channel_image_search_video_all_cate_top", "top", 9)
-    video_decode_accurate_text_top = ("video_decode_accurate_text_top", "top", 9)
+    account_top = ("account", "top", 9)
+    account_extend_top = ("account_extend", "top", 9)
+    channel_topic_top = ("channel_topic", "top", 9)
+    channel_topic_extend_top = ("channel_topic_extend", "top", 9)
+    channel_image_search_video_top = ("channel_image_search_video", "top", 9)
+    channel_image_search_topic_top = ("channel_image_search_topic", "top", 9)
+    channel_image_search_topic_extend_top = ("channel_image_search_topic_extend", "top", 9)
+    channel_image_search_video_all_cate_top = ("channel_image_search_video_all_cate", "top", 9)
+    video_decode_accurate_text_top = ("video_decode_accurate_text", "top", 9)
 
     def __init__(self, crawler_mode, video_source, min_hour):
         self.crawler_mode = crawler_mode

+ 0 - 3
monitor/automation_provide_job_monitor.py

@@ -68,9 +68,6 @@ def job_run_state(start_ts: int, end_ts: int):
     all_crawler_mode = list(dict.fromkeys([cron_info.crawler_mode for cron_info in AutomationJobCronInfo]))
     collapsible_limit = 5
     crawler_mode_group = [all_crawler_mode[i:i + collapsible_limit] for i in range(0, len(all_crawler_mode), collapsible_limit)]
-    print(f"all_crawler_mode: {all_crawler_mode}")
-    print(f"crawler_mode_group: {crawler_mode_group}")
-    print(f"log_data: {log_data}")
     for crawler_mode_partition in crawler_mode_group:
         elements = []
         for crawler_mode in crawler_mode_partition:

+ 2 - 2
util/feishu_inform_util.py

@@ -14,8 +14,8 @@ def send_card_msg_to_feishu(webhook, card_json):
         "card": card_json
     }
     print(f"推送飞书消息webhook地址 - {webhook},  消息内容: {json.dumps(payload_message, indent=4, ensure_ascii=False)}")
-    # response = requests.request('POST', url=webhook, headers=headers, data=json.dumps(payload_message))
-    # print(f"推送飞书消息返回结果: {response.text}")
+    response = requests.request('POST', url=webhook, headers=headers, data=json.dumps(payload_message))
+    print(f"推送飞书消息返回结果: {response.text}")
 
 
 def timestamp_format(timestamp: str) -> str: