瀏覽代碼

feat:修改脚本

zhaohaipeng 1 周之前
父節點
當前提交
8dd7151cb4

+ 0 - 0
aliyun/__init__.py


+ 1 - 1
client/CrawlerClient.py

@@ -65,7 +65,7 @@ class CrawlerClient:
             "publish_time": publish_time,
             "duration": duration_val,
             "cursor": cursor if cursor else "",
-            "accountId": account_id
+            "account_id": account_id
         }
 
         # 发送POST请求并处理响应

+ 0 - 0
monitor/aliyun_resource_monitor.py


+ 29 - 0
script/automation_provide_job_repair.py

@@ -0,0 +1,29 @@
+import json
+
+from client.AIGCClient import AIGCClient
+
+aigc_client = AIGCClient(token="8bf14f27fc3a486788f3383452422d72", base_url="https://aigc-api.aiddit.com")
+
+log_list = []
+with open("/Users/zhao/Downloads/cb08b033-8ab9-48e9-baf2-1ec5b031e0e5.json", 'r') as f:
+    line = f.readline()
+    while line:
+        log_list.append(json.loads(line))
+        line = f.readline()
+
+for log in log_list:
+    crawler_plan_id = log.get("crawlerPlanId")
+    video_id = log.get("videoId")
+    if crawler_plan_id == '20260122023138114334284':
+        continue
+    error_msg, crawler_plan_info = aigc_client.get_content_crawler_plan_by_id(crawler_plan_id)
+    if error_msg:
+        print(f"获取 {crawler_plan_id} 的爬取信息异常")
+        continue
+    update_use_date = crawler_plan_info.get("updateData")
+    crawler_plan_name = update_use_date['name']
+    if "视频vid" not in crawler_plan_name:
+        continue
+    new_crawler_plan_name = crawler_plan_name.replace("{视频vid}", video_id)
+    update_use_date['name'] = new_crawler_plan_name
+    # aigc_client.crawler_plan_save(update_use_date)

+ 6 - 6
script/dou_yin_keywords_search.py

@@ -26,6 +26,12 @@ def log_info_print_title():
 
 
 def log_info_print(log_json: dict[str, Any], account_id: Optional[int] = None):
+    if 'ext' in log_json and isinstance(log_json['ext'], dict):
+        log_json['ext'] = json.dumps(log_json['ext'], ensure_ascii=False)
+
+    if 'modelValueConfig' and isinstance(log_json['modelValueConfig'], dict):
+        log_json['modelValueConfig'] = json.dumps(log_json['modelValueConfig'], ensure_ascii=False)
+
     video_id = log_json["videoId"]
     keywords = log_json['keywords']
     crawler_plan_id = log_json.get("crawlerPlanId", "")
@@ -142,8 +148,6 @@ def keywords_not_login_comprehensive_sort(keywords: str, log_json: dict[str, Any
     """
     account_id = 0
     log_json = keywords_search_and_filter(keywords=keywords, sort_type="综合排序", account_id=account_id, log_json=log_json, filters=filters)
-    log_json['ext'] = json.dumps(log_json['ext'], ensure_ascii=False)
-    log_json['modelValueConfig'] = json.dumps(log_json['modelValueConfig'], ensure_ascii=False)
     log_info_print(log_json, account_id=account_id)
 
 
@@ -153,8 +157,6 @@ def keywords_login_comprehensive_sort(keywords: str, log_json: dict[str, Any], f
     """
     account_id = 771431186
     log_json = keywords_search_and_filter(keywords=keywords, sort_type="综合排序", account_id=account_id, log_json=log_json, filters=filters)
-    log_json['ext'] = json.dumps(log_json['ext'], ensure_ascii=False)
-    log_json['modelValueConfig'] = json.dumps(log_json['modelValueConfig'], ensure_ascii=False)
     log_info_print(log_json, account_id=account_id)
 
 
@@ -164,8 +166,6 @@ def keywords_login_like_sort(keywords: str, log_json: dict[str, Any], filters: L
     """
     account_id = 771431186
     log_json = keywords_search_and_filter(keywords=keywords, sort_type="最多点赞", account_id=account_id, log_json=log_json, filters=filters)
-    log_json['ext'] = json.dumps(log_json['ext'], ensure_ascii=False)
-    log_json['modelValueConfig'] = json.dumps(log_json['modelValueConfig'], ensure_ascii=False)
     log_info_print(log_json, account_id=account_id)