Ver código fonte

update feishu_api

wangkun 3 anos atrás
pai
commit
4d1e65319d

Diferenças do arquivo suprimidas por serem muito extensas
+ 0 - 0
chlsfiles/charles202205071008.chlsj


+ 40 - 10
main/common.py

@@ -13,6 +13,7 @@ import os
 import time
 import requests
 import urllib3
+from loguru import logger
 
 proxies = {"http": None, "https": None}
 
@@ -27,6 +28,7 @@ class Common:
     # 明天 <class 'str'>  2022-04-15
     tomorrow = (date.today() + timedelta(days=1)).strftime("%Y-%m-%d")
 
+    # 使用 logging 模块生成日志
     @staticmethod
     def crawler_log():
         """
@@ -49,6 +51,33 @@ class Common:
 
         return crawler_logger
 
+    # 使用 logger 模块生成日志
+    @staticmethod
+    def logger():
+        """
+        使用 logger 模块生成日志
+        """
+        # 日志路径
+        log_dir = r"./logs/"
+        log_path = os.getcwd() + os.sep + log_dir
+        if not os.path.isdir(log_path):
+            os.makedirs(log_path)
+
+        # 日志文件名
+        log_name = time.strftime("%Y-%m-%d", time.localtime(time.time())) + '.log'
+
+        # 日志不打印到控制台
+        logger.remove(handler_id=None)
+
+        # rotation="500 MB",实现每 500MB 存储一个文件
+        # rotation="12:00",实现每天 12:00 创建一个文件
+        # rotation="1 week",每周创建一个文件
+        # retention="10 days",每隔10天之后就会清理旧的日志
+        # 初始化日志
+        logger.add(log_dir + log_name, level="INFO", rotation='00:00')
+
+        return logger
+
     @classmethod
     def del_logs(cls):
         """
@@ -68,7 +97,7 @@ class Common:
         else:
             for file in all_logs[:len(all_logs) - 7]:
                 os.remove(log_dir + file)
-        cls.crawler_log().info("清除冗余日志成功")
+        cls.logger().info("清除冗余日志成功")
 
     # 删除 charles 缓存文件,只保留最近的两个文件
     @classmethod
@@ -77,6 +106,7 @@ class Common:
         all_file = sorted(os.listdir(r"./chlsfiles/"))
         for file in all_file[0:-2]:
             os.remove(r"./chlsfiles/" + file)
+        cls.logger().info("删除 charles 缓存文件成功")
 
     @classmethod
     def download_method(cls, text, d_name, d_url):
@@ -105,9 +135,9 @@ class Common:
                 with open(video_dir + video_name, "wb") as f:
                     for chunk in response.iter_content(chunk_size=10240):
                         f.write(chunk)
-                cls.crawler_log().info("==========视频下载完成==========")
+                cls.logger().info("==========视频下载完成==========")
             except Exception as e:
-                cls.crawler_log().error("视频下载失败:{}".format(e))
+                cls.logger().exception("视频下载失败:{}", e)
 
         # 下载封面
         elif text == "cover":
@@ -122,9 +152,9 @@ class Common:
             try:
                 with open(video_dir + cover_name, "wb") as f:
                     f.write(response.content)
-                cls.crawler_log().info("==========封面下载完成==========")
+                cls.logger().info("==========封面下载完成==========")
             except Exception as e:
-                cls.crawler_log().error("封面下载失败:{}".format(e))
+                cls.logger().exception("封面下载失败:{}", e)
 
     @staticmethod
     def read_txt(t_name):
@@ -142,7 +172,7 @@ class Common:
         charles_file_dir = r"./chlsfiles/"
 
         if int(len(os.listdir(charles_file_dir))) == 1:
-            Common.crawler_log().info("未找到chlsfile文件,等待60s")
+            Common.logger().info("未找到chlsfile文件,等待60s")
             time.sleep(60)
         else:
             try:
@@ -208,11 +238,11 @@ class Common:
                             else:
                                 return sessions
                 else:
-                    cls.crawler_log().info("未找到 session,10s后重新获取")
+                    cls.logger().info("未找到 session,10s后重新获取")
                     time.sleep(10)
                     cls.get_session()
             except Exception as e:
-                cls.crawler_log().error("获取 session 异常,30s后重试:{}".format(e))
+                cls.logger().exception("获取 session 异常,30s后重试:{}", e)
                 time.sleep(30)
                 cls.get_session()
 
@@ -222,7 +252,7 @@ class Common:
         count = 0
         for count, line in enumerate(open(videoid_path, "rb").readlines()):
             count += 1
-        cls.crawler_log().info('累计下载视频数: {}\n'.format(count))
+        cls.logger().info('累计下载视频数: {}\n', count)
 
     @classmethod
     def kanyikan_today_download_count(cls):
@@ -251,7 +281,7 @@ class Common:
             name = os.path.splitext(file)[0]
             if name == cls.yesterday + "_kanyikan_videoid":
                 os.remove(yesterday_kanyikan_videoid_txt_dir + file)
-        Common.crawler_log().info("删除快手昨天下载统计文件成功")
+        Common.logger().info("删除快手昨天下载统计文件成功")
 
 
 if __name__ == "__main__":

+ 27 - 27
main/download_play.py

@@ -42,31 +42,31 @@ class DownloadPlay:
         测试环境:env == dev
         正式环境:env == prod
         """
-        if len(Feishu.get_values_batch("Y8N3Vl")) == 1:
+        if len(Feishu.get_values_batch("SdCHOM")) == 1:
             pass
         else:
-            for i in range(len(Feishu.get_values_batch("Y8N3Vl"))):
+            for i in range(len(Feishu.get_values_batch("SdCHOM"))):
                 try:
-                    download_video_id = Feishu.get_values_batch("Y8N3Vl")[i+1][1]
-                    download_video_play_cnt = Feishu.get_values_batch("Y8N3Vl")[i+1][2]
-                    download_video_title = Feishu.get_values_batch("Y8N3Vl")[i+1][3]
-                    download_video_duration = Feishu.get_values_batch("Y8N3Vl")[i+1][4]
-                    download_video_comment_cnt = Feishu.get_values_batch("Y8N3Vl")[i+1][5]
-                    download_video_like_cnt = Feishu.get_values_batch("Y8N3Vl")[i+1][6]
-                    download_video_share_cnt = Feishu.get_values_batch("Y8N3Vl")[i+1][7]
-                    download_video_resolution = Feishu.get_values_batch("Y8N3Vl")[i+1][8]
+                    download_video_id = Feishu.get_values_batch("SdCHOM")[i+1][1]
+                    download_video_play_cnt = Feishu.get_values_batch("SdCHOM")[i+1][2]
+                    download_video_title = Feishu.get_values_batch("SdCHOM")[i+1][3]
+                    download_video_duration = Feishu.get_values_batch("SdCHOM")[i+1][4]
+                    download_video_comment_cnt = Feishu.get_values_batch("SdCHOM")[i+1][5]
+                    download_video_like_cnt = Feishu.get_values_batch("SdCHOM")[i+1][6]
+                    download_video_share_cnt = Feishu.get_values_batch("SdCHOM")[i+1][7]
+                    download_video_resolution = Feishu.get_values_batch("SdCHOM")[i+1][8]
                     download_video_width = download_video_resolution.split("*")[0]
                     download_video_height = download_video_resolution.split("*")[-1]
-                    download_video_send_time = Feishu.get_values_batch("Y8N3Vl")[i+1][9]
-                    download_user_name = Feishu.get_values_batch("Y8N3Vl")[i+1][10]
-                    download_head_url = Feishu.get_values_batch("Y8N3Vl")[i+1][11]
-                    download_cover_url = Feishu.get_values_batch("Y8N3Vl")[i+1][12]
-                    download_video_url = Feishu.get_values_batch("Y8N3Vl")[i+1][13]
-                    download_video_session = Feishu.get_values_batch("Y8N3Vl")[i+1][14]
+                    download_video_send_time = Feishu.get_values_batch("SdCHOM")[i+1][9]
+                    download_user_name = Feishu.get_values_batch("SdCHOM")[i+1][10]
+                    download_head_url = Feishu.get_values_batch("SdCHOM")[i+1][11]
+                    download_cover_url = Feishu.get_values_batch("SdCHOM")[i+1][12]
+                    download_video_url = Feishu.get_values_batch("SdCHOM")[i+1][13]
+                    download_video_session = Feishu.get_values_batch("SdCHOM")[i+1][14]
 
                     if cls.play_rule(download_video_width, download_video_height,
                                      download_video_duration, download_video_play_cnt) is True:
-                        Common.crawler_log().info("开始下载视频:{}".format(download_video_title))
+                        Common.logger().info("开始下载视频:{}", download_video_title)
 
                         # 下载封面
                         Common.download_method(text="cover", d_name=download_video_title, d_url=download_cover_url)
@@ -90,32 +90,32 @@ class DownloadPlay:
                                       str(download_video_session))
 
                         # 上传视频
-                        Common.crawler_log().info("开始上传视频:{}".format(download_video_title))
+                        Common.logger().info("开始上传视频:{}", download_video_title)
                         Publish.upload_and_publish(env, "play")
 
                         # 保存视频 ID 到云文档:https://w42nne6hzg.feishu.cn/sheets/shtcngRPoDYAi24x52j2nDuHMih?sheet=20ce0c
-                        Common.crawler_log().info("保存视频ID至云文档:{}".format(download_video_title))
+                        Common.logger().info("保存视频ID至云文档:{}", download_video_title)
                         # 看一看+ ,视频ID工作表,插入首行
                         Feishu.insert_columns("20ce0c")
                         # 看一看+ ,视频ID工作表,首行写入数据
                         Feishu.update_values("20ce0c", download_video_id, "", "", "",
                                              "", "", "", "", "", "", "", "", "", "", "")
 
-                        # 从云文档删除该视频信息:https://w42nne6hzg.feishu.cn/sheets/shtcngRPoDYAi24x52j2nDuHMih?sheet=Y8N3Vl
-                        Common.crawler_log().info("从云文档删除该视频信息:{}".format(download_video_title))
+                        # 从云文档删除该视频信息:https://w42nne6hzg.feishu.cn/sheets/shtcngRPoDYAi24x52j2nDuHMih?sheet=SdCHOM
+                        Common.logger().info("从云文档删除该视频信息:{}", download_video_title)
                         # 删除行或列,可选 ROWS、COLUMNS
-                        Feishu.dimension_range("Y8N3Vl", "ROWS", i + 2, i + 2)
+                        Feishu.dimension_range("SdCHOM", "ROWS", i + 2, i + 2)
 
                     else:
-                        # 从云文档删除该视频信息:https://w42nne6hzg.feishu.cn/sheets/shtcngRPoDYAi24x52j2nDuHMih?sheet=Y8N3Vl
-                        Common.crawler_log().info("该视频不满足下载规则,删除在云文档中的信息:{}".format(download_video_title))
+                        # 从云文档删除该视频信息:https://w42nne6hzg.feishu.cn/sheets/shtcngRPoDYAi24x52j2nDuHMih?sheet=SdCHOM
+                        Common.logger().info("该视频不满足下载规则,删除在云文档中的信息:{}", download_video_title)
                         # 删除行或列,可选 ROWS、COLUMNS
-                        Feishu.dimension_range("Y8N3Vl", "ROWS", i + 2, i + 2)
+                        Feishu.dimension_range("SdCHOM", "ROWS", i + 2, i + 2)
 
                 except Exception as e:
-                    Common.crawler_log().error("视频 info 异常,删除该视频信息".format(e))
+                    Common.logger().error("视频 info 异常,删除该视频信息", e)
                     # 删除行或列,可选 ROWS、COLUMNS
-                    Feishu.dimension_range("Y8N3Vl", "ROWS", i + 2, i + 2)
+                    Feishu.dimension_range("SdCHOM", "ROWS", i + 2, i + 2)
 
             cls.download_play_video("prod")
 

+ 28 - 33
main/download_sendtime.py

@@ -9,7 +9,6 @@
 """
 import json
 import os
-import random
 import sys
 import time
 import requests
@@ -52,15 +51,15 @@ class DownloadSendtime:
         :param env: 测试环境:dev;正式环境:prod
         :return: 下载并上传视频
         """
-        if len(Feishu.get_values_batch("Y8N3Vl")) == 1:
+        if len(Feishu.get_values_batch("SdCHOM")) == 1:
             pass
         else:
-            for i in range(len(Feishu.get_values_batch("Y8N3Vl"))):
+            for i in range(len(Feishu.get_values_batch("SdCHOM"))):
                 try:
                     sendtime_session = Common.get_session()
-                    Common.crawler_log().info("获取视频info时,session:{}".format(sendtime_session))
-                    download_video_id = Feishu.get_values_batch("Y8N3Vl")[i+1][1]
-                    download_video_title = Feishu.get_values_batch("Y8N3Vl")[i+1][3]
+                    Common.logger().info("获取视频info时,session:{}", sendtime_session)
+                    download_video_id = Feishu.get_values_batch("SdCHOM")[i+1][1]
+                    download_video_title = Feishu.get_values_batch("SdCHOM")[i+1][3]
                     url = "https://search.weixin.qq.com/cgi-bin/recwxa/recwxagetonevideoinfo?"
                     param = {
                         "session": sendtime_session,
@@ -78,9 +77,9 @@ class DownloadSendtime:
                     r = requests.get(url=url, params=param, proxies=proxies, verify=False)
                     response = json.loads(r.content.decode("utf8"))
                     if "data" not in response:
-                        Common.crawler_log().error("获取视频info时错误,删除该视频:{}".format(download_video_title))
+                        Common.logger().error("获取视频info时错误,删除该视频:{}", download_video_title)
                         # 删除行或列,可选 ROWS、COLUMNS
-                        Feishu.dimension_range("Y8N3Vl", "ROWS", i + 2, i + 2)
+                        Feishu.dimension_range("SdCHOM", "ROWS", i + 2, i + 2)
                     else:
                         data = response["data"]
                         v_duration = data["duration"]
@@ -98,17 +97,13 @@ class DownloadSendtime:
                         if "items" not in data["play_info"]:
                             if len(data["play_info"]) > 2:
                                 download_url_up = data["play_info"][2]["play_url"]
-                                # Common.crawler_log().info('视频下载地址:{}'.format(download_url_up))
                             else:
                                 download_url_up = data["play_info"][0]["play_url"]
-                                # Common.crawler_log().info('视频下载地址:{}'.format(download_url_up))
                         else:
                             if len(data["play_info"]["items"]) > 2:
                                 download_url_up = data["play_info"]["items"][2]["play_url"]
-                                # Common.crawler_log().info('视频下载地址:{}'.format(download_url_up))
                             else:
                                 download_url_up = data["play_info"]["items"][0]["play_url"]
-                                # Common.crawler_log().info('视频下载地址:{}'.format(download_url_up))
 
                         # 判断基本规则
                         if cls.send_time_rule(v_width, v_height, v_duration, v_play_cnt_sendtime) is True \
@@ -120,8 +115,8 @@ class DownloadSendtime:
                             # 满足下载条件:当前时间 - 发布时间 <= 3天,播放量大于1万
                             if int(time.time()) - int(v_send_date) <= 604800:
                                 if int(v_play_cnt_sendtime) >= 10000:
-                                    Common.crawler_log().info("该视频:{}".format(
-                                        download_video_title) + " " + "在7天内的播放量{}>=10000".format(v_play_cnt_sendtime))
+                                    Common.logger().info("该视频:{} ,在7天内的播放量{}>=10000",
+                                                         download_video_title, v_play_cnt_sendtime)
 
                                     # 下载封面
                                     Common.download_method("cover", download_video_title, v_video_cover)
@@ -146,46 +141,46 @@ class DownloadSendtime:
                                                    str(sendtime_session))
 
                                     # 上传该视频
-                                    Common.crawler_log().info("开始上传视频:{}".format(download_video_title))
+                                    Common.logger().info("开始上传视频:{}", download_video_title)
                                     Publish.upload_and_publish(env, "send_time")
 
                                     # 保存视频 ID 到云文档:
                                     # https://w42nne6hzg.feishu.cn/sheets/shtcngRPoDYAi24x52j2nDuHMih?sheet=20ce0c
-                                    Common.crawler_log().info("保存视频ID至云文档:{}".format(download_video_title))
+                                    Common.logger().info("保存视频ID至云文档:{}", download_video_title)
                                     # 看一看+ ,视频ID工作表,插入首行
                                     Feishu.insert_columns("20ce0c")
                                     # 看一看+ ,视频ID工作表,首行写入数据
                                     Feishu.update_values("20ce0c", download_video_id, "", "", "",
                                                          "", "", "", "", "", "", "", "", "", "", "")
 
-                                    # 从云文档删除该视频信息:https://w42nne6hzg.feishu.cn/sheets/shtcngRPoDYAi24x52j2nDuHMih?sheet=Y8N3Vl
-                                    Common.crawler_log().info("从云文档删除该视频信息:{}".format(download_video_title))
+                                    # 从云文档删除该视频信息:https://w42nne6hzg.feishu.cn/sheets/shtcngRPoDYAi24x52j2nDuHMih?sheet=SdCHOM
+                                    Common.logger().info("从云文档删除该视频信息:{}", download_video_title)
                                     # 删除行或列,可选 ROWS、COLUMNS
-                                    Feishu.dimension_range("Y8N3Vl", "ROWS", i + 2, i + 2)
+                                    Feishu.dimension_range("SdCHOM", "ROWS", i + 2, i + 2)
 
                                 else:
-                                    # 从云文档删除该视频信息:https://w42nne6hzg.feishu.cn/sheets/shtcngRPoDYAi24x52j2nDuHMih?sheet=Y8N3Vl
-                                    Common.crawler_log().info("该视频7天播放量:{}<10000".format(
-                                        int(v_play_cnt_sendtime)) + ";" + "不满足下载规则:{}".format(download_video_title))
+                                    # 从云文档删除该视频信息:https://w42nne6hzg.feishu.cn/sheets/shtcngRPoDYAi24x52j2nDuHMih?sheet=SdCHOM
+                                    Common.logger().info("该视频7天播放量:{}<10000 ;不满足下载规则:{}",
+                                                         int(v_play_cnt_sendtime), download_video_title)
                                     # 删除行或列,可选 ROWS、COLUMNS
-                                    Feishu.dimension_range("Y8N3Vl", "ROWS", i + 2, i + 2)
+                                    Feishu.dimension_range("SdCHOM", "ROWS", i + 2, i + 2)
                             else:
-                                # 从云文档删除该视频信息:https://w42nne6hzg.feishu.cn/sheets/shtcngRPoDYAi24x52j2nDuHMih?sheet=Y8N3Vl
-                                Common.crawler_log().info("视频发布时间大于7天:{}天".format(
-                                    int((int(time.time()) - int(v_send_date)) / 86400))
-                                                          + ";" + "标题:{}".format(download_video_title))
+                                # 从云文档删除该视频信息:https://w42nne6hzg.feishu.cn/sheets/shtcngRPoDYAi24x52j2nDuHMih?sheet=SdCHOM
+                                Common.logger().info("视频发布时间大于7天:{}天;标题:{}",
+                                                     int((int(time.time()) - int(v_send_date)) / 86400),
+                                                     download_video_title)
                                 # 删除行或列,可选 ROWS、COLUMNS
-                                Feishu.dimension_range("Y8N3Vl", "ROWS", i + 2, i + 2)
+                                Feishu.dimension_range("SdCHOM", "ROWS", i + 2, i + 2)
                         else:
-                            # 从云文档删除该视频信息:https://w42nne6hzg.feishu.cn/sheets/shtcngRPoDYAi24x52j2nDuHMih?sheet=Y8N3Vl
-                            Common.crawler_log().info("不满足下载规则:{}".format(download_video_title))
+                            # 从云文档删除该视频信息:https://w42nne6hzg.feishu.cn/sheets/shtcngRPoDYAi24x52j2nDuHMih?sheet=SdCHOM
+                            Common.logger().info("不满足下载规则:{}", download_video_title)
                             # 删除行或列,可选 ROWS、COLUMNS
-                            Feishu.dimension_range("Y8N3Vl", "ROWS", i + 2, i + 2)
+                            Feishu.dimension_range("SdCHOM", "ROWS", i + 2, i + 2)
 
                 except Exception as e:
-                    Common.crawler_log().error("获取视频info异常:{},删除该视频".format(e))
+                    Common.logger().error("获取视频info异常:{},删除该视频", e)
                     # 删除行或列,可选 ROWS、COLUMNS
-                    Feishu.dimension_range("Y8N3Vl", "ROWS", i + 2, i + 2)
+                    Feishu.dimension_range("SdCHOM", "ROWS", i + 2, i + 2)
 
             cls.download_sendtime_video("prod")
 

+ 40 - 40
main/download_up.py

@@ -54,19 +54,19 @@ class DownloadUp:
             1 更新视频 ID 到 https://w42nne6hzg.feishu.cn/sheets/shtcngRPoDYAi24x52j2nDuHMih?sheet=20ce0c
             2 视频信息写入文件 "./videos/{d_title}/info.txt"
         4.上传完成:
-            1 删除该视频在 https://w42nne6hzg.feishu.cn/sheets/shtcngRPoDYAi24x52j2nDuHMih?sheet=Y8N3Vl 中的信息
+            1 删除该视频在 https://w42nne6hzg.feishu.cn/sheets/shtcngRPoDYAi24x52j2nDuHMih?sheet=SdCHOM 中的信息
         """
-        if len(Feishu.get_values_batch("Y8N3Vl")) == 1:
+        if len(Feishu.get_values_batch("SdCHOM")) == 1:
             pass
         else:
-            for i in range(len(Feishu.get_values_batch("Y8N3Vl"))):
+            for i in range(len(Feishu.get_values_batch("SdCHOM"))):
                 try:
                     video_info_session = Common.get_session()
-                    Common.crawler_log().info("获取视频info时,session:{}".format(video_info_session))
-                    download_time = Feishu.get_values_batch("Y8N3Vl")[i+1][0]  # 第一次获取该视频的时间
-                    download_video_id = Feishu.get_values_batch("Y8N3Vl")[i+1][1]  # 外网视频 ID
-                    download_video_play_cnt = Feishu.get_values_batch("Y8N3Vl")[i+1][2]  # 播放量
-                    download_video_title = Feishu.get_values_batch("Y8N3Vl")[i+1][3]
+                    Common.logger().info("获取视频info时,session:{}", video_info_session)
+                    download_time = Feishu.get_values_batch("SdCHOM")[i+1][0]  # 第一次获取该视频的时间
+                    download_video_id = Feishu.get_values_batch("SdCHOM")[i+1][1]  # 外网视频 ID
+                    download_video_play_cnt = Feishu.get_values_batch("SdCHOM")[i+1][2]  # 播放量
+                    download_video_title = Feishu.get_values_batch("SdCHOM")[i+1][3]
 
                     url = "https://search.weixin.qq.com/cgi-bin/recwxa/recwxagetonevideoinfo?"
                     param = {
@@ -85,9 +85,9 @@ class DownloadUp:
                     r = requests.get(url=url, params=param, proxies=proxies, verify=False)
                     response = json.loads(r.content.decode("utf8"))
                     if "data" not in response:
-                        Common.crawler_log().error("获取视频info时错误,删除该视频:{}".format(download_video_title))
+                        Common.logger().error("获取视频info时错误,删除该视频:{}", download_video_title)
                         # 删除行或列,可选 ROWS、COLUMNS
-                        Feishu.dimension_range("Y8N3Vl", "ROWS", i + 2, i + 2)
+                        Feishu.dimension_range("SdCHOM", "ROWS", i + 2, i + 2)
                     else:
                         data = response["data"]
                         v_duration = data["duration"]
@@ -121,14 +121,14 @@ class DownloadUp:
                                 and v_send_date != "" and v_username != "" and v_user_cover != "" \
                                 and v_video_cover != "" and download_url_up != "":
                             if int(time.time()) - int(download_time) < 3600:
-                                Common.crawler_log().info("距上次获取该视频时间:{}分钟".format(
-                                    int((int(int(time.time()) - int(download_time))) / 60))
-                                                          + ";{}".format(download_video_title))
+                                Common.logger().info("距上次获取该视频时间:{}分钟;{}",
+                                                     int((int(int(time.time()) - int(download_time))) / 60),
+                                                     download_video_title)
                             elif 7200 >= int(time.time()) - int(download_time) >= 3600:
                                 if int(v_play_cnt_up) - int(download_video_play_cnt) >= 1000:
-                                    Common.crawler_log().info("该视频:{}".format(
-                                        download_video_title) + " " + "在1小时内的播放量{}>=1000".format(
-                                        int(v_play_cnt_up) - int(download_video_play_cnt)))
+                                    Common.logger().info("该视频:{}在1小时内的播放量{}>=1000",
+                                                         download_video_title,
+                                                         int(v_play_cnt_up) - int(download_video_play_cnt))
 
                                     # 下载封面
                                     Common.download_method("cover", download_video_title, v_video_cover)
@@ -153,39 +153,39 @@ class DownloadUp:
                                                    str(video_info_session))
     
                                     # 上传该视频
-                                    Common.crawler_log().info("开始上传视频:{}".format(download_video_title))
+                                    Common.logger().info("开始上传视频:{}", download_video_title)
                                     Publish.upload_and_publish(env, "up")
 
                                     # 保存视频 ID 到云文档:
                                     # https://w42nne6hzg.feishu.cn/sheets/shtcngRPoDYAi24x52j2nDuHMih?sheet=20ce0c
-                                    Common.crawler_log().info("保存视频ID至云文档:{}".format(download_video_title))
+                                    Common.logger().info("保存视频ID至云文档:{}", download_video_title)
                                     # 看一看+ ,视频ID工作表,插入首行
                                     Feishu.insert_columns("20ce0c")
                                     # 看一看+ ,视频ID工作表,首行写入数据
                                     Feishu.update_values("20ce0c", download_video_id, "", "", "",
                                                          "", "", "", "", "", "", "", "", "", "", "")
 
-                                    # 从云文档删除该视频信息:https://w42nne6hzg.feishu.cn/sheets/shtcngRPoDYAi24x52j2nDuHMih?sheet=Y8N3Vl
-                                    Common.crawler_log().info("从云文档删除该视频信息:{}".format(download_video_title))
+                                    # 从云文档删除该视频信息:https://w42nne6hzg.feishu.cn/sheets/shtcngRPoDYAi24x52j2nDuHMih?sheet=SdCHOM
+                                    Common.logger().info("从云文档删除该视频信息:{}", download_video_title)
                                     # 删除行或列,可选 ROWS、COLUMNS
-                                    Feishu.dimension_range("Y8N3Vl", "ROWS", i + 2, i + 2)
+                                    Feishu.dimension_range("SdCHOM", "ROWS", i + 2, i + 2)
                                 else:
                                     # 删除之前保存的该视频信息,并把现在的信息保存进去
-                                    Common.crawler_log().info("该视频1小时内的播放量:{}<1000".format(
-                                        int(v_play_cnt_up) - int(download_video_play_cnt)
-                                    ) + ";" + "更新该视频在kanyikan_feeds.txt中的信息:{}".format(download_video_title))
-                                    # 从云文档删除该视频信息:https://w42nne6hzg.feishu.cn/sheets/shtcngRPoDYAi24x52j2nDuHMih?sheet=Y8N3Vl
-                                    Common.crawler_log().info("从云文档删除该视频信息:{}".format(download_video_title))
+                                    Common.logger().info("该视频1小时内的播放量:{}<1000;更新该视频信息:{}",
+                                                         int(v_play_cnt_up) - int(download_video_play_cnt),
+                                                         download_video_title)
+                                    # 从云文档删除该视频信息:https://w42nne6hzg.feishu.cn/sheets/shtcngRPoDYAi24x52j2nDuHMih?sheet=SdCHOM
+                                    Common.logger().info("从云文档删除该视频信息:{}", download_video_title)
                                     # 删除行或列,可选 ROWS、COLUMNS
-                                    Feishu.dimension_range("Y8N3Vl", "ROWS", i + 2, i + 2)
+                                    Feishu.dimension_range("SdCHOM", "ROWS", i + 2, i + 2)
 
                                     # 看一看+工作表,插入首行
-                                    print(Feishu.insert_columns("Y8N3Vl"))
+                                    print(Feishu.insert_columns("SdCHOM"))
 
                                     # 获取当前时间
                                     download_up_time = int(time.time())
                                     # 看一看云文档,工作表 kanyikan_feeds_1 中写入数据
-                                    Feishu.update_values("Y8N3Vl",
+                                    Feishu.update_values("SdCHOM",
                                                          a1=str(download_up_time),
                                                          b1=str(download_video_id),
                                                          c1=str(v_play_cnt_up),
@@ -202,22 +202,22 @@ class DownloadUp:
                                                          n1=str(download_url_up),
                                                          o1=str(video_info_session))
                             elif int(time.time()) - int(download_time) > 7200:
-                                Common.crawler_log().info("距上次获取该视频时间:""{}分钟。超过2小时,删除该视频"
-                                                          .format(int((int(time.time()) - int(download_time)) / 60))
-                                                          + ";" + "标题:{}".format(download_video_title))
+                                Common.logger().info("距上次获取该视频时间:""{}分钟。超过2小时,删除该视频;标题:{}",
+                                                     int((int(time.time()) - int(download_time)) / 60),
+                                                     download_video_title)
                                 # 删除行或列,可选 ROWS、COLUMNS
-                                Feishu.dimension_range("Y8N3Vl", "ROWS", i + 2, i + 2)
+                                Feishu.dimension_range("SdCHOM", "ROWS", i + 2, i + 2)
                         else:
-                            Common.crawler_log().info("不满足下载规则:{}".format(download_video_title))
-                            # 从云文档删除该视频信息:https://w42nne6hzg.feishu.cn/sheets/shtcngRPoDYAi24x52j2nDuHMih?sheet=Y8N3Vl
-                            Common.crawler_log().info("从云文档删除该视频信息:{}".format(download_video_title))
+                            Common.logger().info("不满足下载规则:{}", download_video_title)
+                            # 从云文档删除该视频信息:https://w42nne6hzg.feishu.cn/sheets/shtcngRPoDYAi24x52j2nDuHMih?sheet=SdCHOM
+                            Common.logger().info("从云文档删除该视频信息:{}", download_video_title)
                             # 删除行或列,可选 ROWS、COLUMNS
-                            Feishu.dimension_range("Y8N3Vl", "ROWS", i + 2, i + 2)
+                            Feishu.dimension_range("SdCHOM", "ROWS", i + 2, i + 2)
                 except Exception as e:
-                    # 从云文档删除该视频信息:https://w42nne6hzg.feishu.cn/sheets/shtcngRPoDYAi24x52j2nDuHMih?sheet=Y8N3Vl
-                    Common.crawler_log().error("获取视频info异常:{},删除该视频".format(e))
+                    # 从云文档删除该视频信息:https://w42nne6hzg.feishu.cn/sheets/shtcngRPoDYAi24x52j2nDuHMih?sheet=SdCHOM
+                    Common.logger().exception("获取视频info异常:{},删除该视频", e)
                     # 删除行或列,可选 ROWS、COLUMNS
-                    Feishu.dimension_range("Y8N3Vl", "ROWS", i + 2, i + 2)
+                    Feishu.dimension_range("SdCHOM", "ROWS", i + 2, i + 2)
 
 
 if __name__ == "__main__":

+ 17 - 17
main/feishu_lib.py

@@ -5,6 +5,8 @@ import json
 import requests
 import urllib3
 
+from main.common import Common
+
 proxies = {"http": None, "https": None}
 
 
@@ -117,7 +119,7 @@ class Feishu:
         urllib3.disable_warnings()
         r = requests.post(url=url, headers=headers, json=body, proxies=proxies, verify=False)
 
-        return r.json()["msg"]
+        Common.logger().info(r.json()["msg"])
 
     # 看一看+工作表,首行写入数据
     @classmethod
@@ -161,7 +163,7 @@ class Feishu:
         urllib3.disable_warnings()
         r = requests.post(url=url, headers=headers, json=body, proxies=proxies, verify=False)
 
-        return r.json()["msg"]
+        Common.logger().info(r.json()["msg"])
 
     # 读取单元格数据
     @classmethod
@@ -222,7 +224,7 @@ class Feishu:
         urllib3.disable_warnings()
         r = requests.delete(url=url, headers=headers, json=body, proxies=proxies, verify=False)
 
-        return r.json()["msg"]
+        Common.logger().info(r.json()["msg"])
 
 
 if __name__ == "__main__":
@@ -234,28 +236,26 @@ if __name__ == "__main__":
     # feishu.get_metainfo()
 
     # 读取工作表中所有数据
-    print(feishu.get_values_batch("Y8N3Vl"))
-    print(len(feishu.get_values_batch("SdCHOM")))
-    for i in range(len(feishu.get_values_batch("Y8N3Vl"))):
-        videoid = feishu.get_values_batch("Y8N3Vl")[i][1]
-        if videoid == "b3":
-            # 删除行或列,可选 ROWS、COLUMNS
-            feishu.dimension_range("Y8N3Vl", "ROWS", i+1, i+1)
-            print(videoid)
-
+    # print(feishu.get_values_batch("SdCHOM"))
+    # print(len(feishu.get_values_batch("SdCHOM")))
+    # for i in range(len(feishu.get_values_batch("SdCHOM"))):
+    #     videoid = feishu.get_values_batch("SdCHOM")[i][1]
+    #     if videoid == "b3":
+    #         # 删除行或列,可选 ROWS、COLUMNS
+    #         feishu.dimension_range("SdCHOM", "ROWS", i+1, i+1)
+    #         print(videoid)
 
-    #
     # # 看一看+工作表,插入首行
-    # print(feishu.insert_columns("Y8N3Vl"))
+    # print(feishu.insert_columns("SdCHOM"))
     #
     # # 看一看+工作表,首行写入数据
-    # print(feishu.update_values("Y8N3Vl", "a1", "b1", "c1", "d1", "e1", "f1", "g1",
+    # print(feishu.update_values("SdCHOM", "a1", "b1", "c1", "d1", "e1", "f1", "g1",
     #                            "h1", "i1", "j1", "k1", "l1", "m1", "n1", "o1"))
 
     # # 查询单元格内容
-    # print(feishu.get_range_value("Y8N3Vl", "B8:C8"))
+    # print(feishu.get_range_value("SdCHOM", "B8:C8"))
     #
     # # 删除行或列,可选 ROWS、COLUMNS
-    # feishu.dimension_range("Y8N3Vl", "ROWS")
+    # feishu.dimension_range("SdCHOM", "ROWS")
 
     pass

+ 36 - 36
main/get_feeds.py

@@ -24,13 +24,13 @@ def get_feeds():
     """
     1.从看一看+小程序首页推荐,获取视频列表
     2.先在 https://w42nne6hzg.feishu.cn/sheets/shtcngRPoDYAi24x52j2nDuHMih?sheet=20ce0c 中去重
-    3.再从 https://w42nne6hzg.feishu.cn/sheets/shtcngRPoDYAi24x52j2nDuHMih?sheet=Y8N3Vl 中去重
-    4.添加视频信息至 https://w42nne6hzg.feishu.cn/sheets/shtcngRPoDYAi24x52j2nDuHMih?sheet=Y8N3Vl
+    3.再从 https://w42nne6hzg.feishu.cn/sheets/shtcngRPoDYAi24x52j2nDuHMih?sheet=SdCHOM 中去重
+    4.添加视频信息至 https://w42nne6hzg.feishu.cn/sheets/shtcngRPoDYAi24x52j2nDuHMih?sheet=SdCHOM
     """
     host = "https://search.weixin.qq.com"
     url = '/cgi-bin/recwxa/recwxavideolist?'
     video_list_session = Common.get_session()
-    Common.crawler_log().info("获取视频list时,session:{}".format(video_list_session))
+    Common.logger().info("获取视频list时,session:{}", video_list_session)
     header = {
         "Connection": "keep-alive",
         "content-type": "application/json",
@@ -60,12 +60,12 @@ def get_feeds():
         response = json.loads(r.content.decode("utf8"))
 
         if "data" not in response:
-            Common.crawler_log().info("获取视频list时,session过期,随机睡眠 31-50 秒")
+            Common.logger().info("获取视频list时,session过期,随机睡眠 31-50 秒")
             # 如果返回空信息,则随机睡眠 31-40 秒
             time.sleep(random.randint(31, 40))
             get_feeds()
         elif "items" not in response["data"]:
-            Common.crawler_log().info("获取视频list时,返回空信息,随机睡眠 1-3 分钟")
+            Common.logger().info("获取视频list时,返回空信息,随机睡眠 1-3 分钟")
             # 如果返回空信息,则随机睡眠 1-3 分钟
             time.sleep(random.randint(60, 180))
             get_feeds()
@@ -74,11 +74,11 @@ def get_feeds():
             for i in range(len(items)):
                 # 如果该视频没有视频信息,则忽略
                 if "videoInfo" not in items[i]:
-                    Common.crawler_log().info("无视频信息")
+                    Common.logger().info("无视频信息")
                 else:
                     # 获取视频ID
                     video_id = items[i]["videoId"]
-                    Common.crawler_log().info('视频ID:{}'.format(video_id))
+                    Common.logger().info('视频ID:{}', video_id)
                 
                     # 获取视频标题
                     video_title = items[i]["title"].strip().replace("\n", "")\
@@ -87,48 +87,48 @@ def get_feeds():
                         .replace("?", "").replace('"', "").replace("<", "")\
                         .replace(">", "").replace("|", "").replace(" ", "")\
                         .replace("&NBSP", "").replace(".", "。").replace(" ", "")
-                    Common.crawler_log().info('视频标题:{}'.format(video_title))
+                    Common.logger().info('视频标题:{}', video_title)
                 
                     # 获取视频播放次数
                     video_play_cnt = items[i]["playCount"]
-                    Common.crawler_log().info('视频播放次数:{}'.format(video_play_cnt))
+                    Common.logger().info('视频播放次数:{}', video_play_cnt)
                 
                     # 获取视频点赞数
                     video_liked_cnt = items[i]["liked_cnt"]
-                    Common.crawler_log().info('视频点赞数:{}'.format(video_liked_cnt))
+                    Common.logger().info('视频点赞数:{}', video_liked_cnt)
                 
                     # 获取视频时长
                     video_duration = items[i]["mediaDuration"]
-                    Common.crawler_log().info('视频时长:{}秒'.format(video_duration))
+                    Common.logger().info('视频时长:{}秒', video_duration)
                 
                     # 获取视频评论数
                     video_comment_cnt = items[i]["comment_cnt"]
-                    Common.crawler_log().info('视频评论数:{}'.format(video_comment_cnt))
+                    Common.logger().info('视频评论数:{}', video_comment_cnt)
                 
                     # 获取视频分享数
                     video_shared_cnt = items[i]["shared_cnt"]
-                    Common.crawler_log().info('视频分享数:{}'.format(video_shared_cnt))
+                    Common.logger().info('视频分享数:{}', video_shared_cnt)
                 
                     # 获取视频发布时间
                     video_send_date = items[i]["date"]
-                    Common.crawler_log().info('视频发布时间:{}'.format(
-                        time.strftime("%Y-%m-%d %H:%M:%S", time.localtime(video_send_date))))
+                    Common.logger().info("视频发布时间:{}",
+                                         time.strftime("%Y-%m-%d %H:%M:%S", time.localtime(video_send_date)))
                 
                     # 获取视频用户名
                     video_user = items[i]["source"].strip().replace("\n", "")
-                    Common.crawler_log().info('视频用户名:{}'.format(video_user))
+                    Common.logger().info('视频用户名:{}', video_user)
                 
                     # 获取视频宽高
                     if "short_video_info" not in items[i]:
                         video_width = "0"
                         video_height = "0"
                         video_resolution = str(video_width) + "*" + str(video_height)
-                        Common.crawler_log().info("无分辨率:{}".format(video_resolution))
+                        Common.logger().info("无分辨率:{}", video_resolution)
                     elif len(items[i]["short_video_info"]) == 0:
                         video_width = "0"
                         video_height = "0"
                         video_resolution = str(video_width) + "*" + str(video_height)
-                        Common.crawler_log().info("无分辨率:{}".format(video_resolution))
+                        Common.logger().info("无分辨率:{}", video_resolution)
 
                     else:
                         # 视频宽
@@ -136,34 +136,34 @@ def get_feeds():
                         # 视频高
                         video_height = items[i]["short_video_info"]["height"]
                         video_resolution = str(video_width) + "*" + str(video_height)
-                        Common.crawler_log().info('视频宽高:{}'.format(video_resolution))
+                        Common.logger().info('视频宽高:{}', video_resolution)
                 
                     # 获取视频用户头像
                     video_user_cover = items[i]["bizIcon"]
-                    Common.crawler_log().info('视频用户头像:{}'.format(video_user_cover))
+                    Common.logger().info('视频用户头像:{}', video_user_cover)
                 
                     # 获取视频封面
                     if "smartCoverUrl" in items[i]:
                         video_cover = items[i]["smartCoverUrl"]
-                        Common.crawler_log().info('视频封面:{}'.format(video_cover))
+                        Common.logger().info('视频封面:{}', video_cover)
                     else:
                         video_cover = items[i]["thumbUrl"]
-                        Common.crawler_log().info('视频封面:{}'.format(video_cover))
+                        Common.logger().info('视频封面:{}', video_cover)
                 
                     # 获取播放地址
                     if "mpInfo" in items[i]["videoInfo"]["videoCdnInfo"].keys():
                         if len(items[i]["videoInfo"]["videoCdnInfo"]["mpInfo"]["urlInfo"]) > 2:
                             url = items[i]["videoInfo"]["videoCdnInfo"]["mpInfo"]["urlInfo"][2]["url"]
-                            Common.crawler_log().info('视频播放地址:{}'.format(url))
+                            Common.logger().info('视频播放地址:{}', url)
                         else:
                             url = items[i]["videoInfo"]["videoCdnInfo"]["mpInfo"]["urlInfo"][0]["url"]
-                            Common.crawler_log().info('视频播放地址:{}'.format(url))
+                            Common.logger().info('视频播放地址:{}', url)
                     elif "ctnInfo" in items[i]["videoInfo"]["videoCdnInfo"]:
                         url = items[i]["videoInfo"]["videoCdnInfo"]["ctnInfo"]["urlInfo"][0]["url"]
-                        Common.crawler_log().info('视频播放地址:{}'.format(url))
+                        Common.logger().info('视频播放地址:{}', url)
                     else:
                         url = items[i]["videoInfo"]["videoCdnInfo"]["urlInfo"][0]["url"]
-                        Common.crawler_log().info('视频播放地址:{}'.format(url))
+                        Common.logger().info('视频播放地址:{}', url)
 
                     # 过滤无效视频
                     if video_id == "" \
@@ -178,27 +178,27 @@ def get_feeds():
                             or video_user_cover == "" \
                             or video_cover == "" \
                             or url == "":
-                        Common.crawler_log().info("无效视频")
+                        Common.logger().info("无效视频")
                     else:
                         # 从 云文档 去重:https://w42nne6hzg.feishu.cn/sheets/shtcngRPoDYAi24x52j2nDuHMih?sheet=20ce0c
                         if video_id in [j for i in Feishu.get_values_batch("20ce0c") for j in i]:
-                            Common.crawler_log().info("该视频已下载:{}".format(video_title))
+                            Common.logger().info("该视频已下载:{}", video_title)
                         else:
-                            Common.crawler_log().info("该视频未下载:{}".format(video_title))
+                            Common.logger().info("该视频未下载:{}", video_title)
 
-                            # 从 云文档 去重:https://w42nne6hzg.feishu.cn/sheets/shtcngRPoDYAi24x52j2nDuHMih?sheet=Y8N3Vl
-                            if video_id in [j for i in Feishu.get_values_batch("Y8N3Vl") for j in i]:
-                                Common.crawler_log().info("该视频已在kanyikan_feeds_1中:{}".format(video_title))
+                            # 从 云文档 去重:https://w42nne6hzg.feishu.cn/sheets/shtcngRPoDYAi24x52j2nDuHMih?sheet=SdCHOM
+                            if video_id in [j for i in Feishu.get_values_batch("SdCHOM") for j in i]:
+                                Common.logger().info("该视频已在kanyikan_feeds_1中:{}", video_title)
                             else:
-                                Common.crawler_log().info("添加该视频信息至kanyikan_feeds_1:{}".format(video_title))
+                                Common.logger().info("添加该视频信息至kanyikan_feeds_1:{}", video_title)
 
                                 # 看一看+工作表,插入首行
-                                print(Feishu.insert_columns("Y8N3Vl"))
+                                print(Feishu.insert_columns("SdCHOM"))
 
                                 # 获取当前时间
                                 get_feeds_time = int(time.time())
                                 # 看一看云文档,工作表 kanyikan_feeds_1 中写入数据
-                                Feishu.update_values("Y8N3Vl",
+                                Feishu.update_values("SdCHOM",
                                                      a1=str(get_feeds_time),
                                                      b1=str(video_id),
                                                      c1=str(video_play_cnt),
@@ -215,7 +215,7 @@ def get_feeds():
                                                      n1=str(url),
                                                      o1=str(video_list_session))
     except Exception as e:
-        Common.crawler_log().error("获取视频 list 时异常:{}".format(e))
+        Common.logger().exception("获取视频 list 时异常:{}", e)
 
 
 if __name__ == "__main__":

+ 21 - 26
main/publish.py

@@ -36,13 +36,12 @@ class Publish:
         versionCode  版本 默认1
         :return:
         """
-        # Common.crawler_log().info('publish request data: {}'.format(request_data))
         result = cls.request_post('https://videotest.yishihui.com/longvideoapi/crawler/video/send', request_data)
-        Common.crawler_log().info('publish result: {}'.format(result))
+        Common.logger().info('publish result: {}', result)
         if result['code'] != 0:
-            Common.crawler_log().error('pushlish failure msg = {}'.format(result['msg']))
+            Common.logger().error('pushlish failure msg = {}', result['msg'])
         else:
-            Common.crawler_log().info('publish success video_id = : {}'.format(request_data['crawlerSrcId']))
+            Common.logger().info('publish success video_id = : {}', request_data['crawlerSrcId'])
 
     @classmethod
     def publish_video_prod(cls, request_data):
@@ -61,13 +60,12 @@ class Publish:
         versionCode  版本 默认1
         :return:
         """
-        # Common.crawler_log().info('publish request data: {}'.format(request_data))
         result = cls.request_post('https://longvideoapi.piaoquantv.com/longvideoapi/crawler/video/send', request_data)
-        Common.crawler_log().info('publish result: {}'.format(result))
+        Common.logger().info('publish result: {}', result)
         if result['code'] != 0:
-            Common.crawler_log().error('pushlish failure msg = {}'.format(result['msg']))
+            Common.logger().error('pushlish failure msg = {}', result['msg'])
         else:
-            Common.crawler_log().info('publish success video_id = : {}'.format(request_data['crawlerSrcId']))
+            Common.logger().info('publish success video_id = : {}', request_data['crawlerSrcId'])
 
     @classmethod
     def request_post(cls, request_url, request_data):
@@ -120,23 +118,20 @@ class Publish:
 
     @classmethod
     def put_file(cls, oss_file, local_file):
-        # Common.crawler_log().info("put oss file = {}, local file = {}".format(oss_file, local_file))
         cls.bucket.put_object_from_file(oss_file, local_file)
-        Common.crawler_log().info("put oss file = {}, local file = {} success".format(oss_file, local_file))
+        Common.logger().info("put oss file = {}, local file = {} success", oss_file, local_file)
 
     # 清除本地文件
     @classmethod
     def remove_local_file(cls, local_file):
-        # Common.crawler_log().info("remove local file = {}".format(local_file))
         os.remove(local_file)
-        Common.crawler_log().info("remove local file = {} success".format(local_file))
+        Common.logger().info("remove local file = {} success", local_file)
 
     # 清除本地文件夹
     @classmethod
     def remove_local_file_dir(cls, local_file):
-        # Common.crawler_log().info("remove local file dir = {}".format(local_file))
         os.rmdir(local_file)
-        Common.crawler_log().info("remove local file dir = {} success".format(local_file))
+        Common.logger().info("remove local file dir = {} success", local_file)
 
     local_file_path = '.\\videos'
     video_file = 'video'
@@ -161,7 +156,7 @@ class Publish:
         :param env: 测试环境:dev,正式环境:prod
         :param job: 上升榜:up,播放量:play, send_time:发布时间榜
         """
-        Common.crawler_log().info("upload_and_publish starting...")
+        Common.logger().info("upload_and_publish starting...")
         today = time.strftime("%Y%m%d", time.localtime())
         # videos 目录下的所有视频文件夹
         files = os.listdir(cls.local_file_path)
@@ -171,7 +166,7 @@ class Publish:
                 fi_d = os.path.join(cls.local_file_path, f)
                 # 确认为视频文件夹
                 if os.path.isdir(fi_d):
-                    Common.crawler_log().info('dir = {}'.format(fi_d))
+                    Common.logger().info('dir = {}', fi_d)
                     # 列出所有视频文件夹
                     dir_files = os.listdir(fi_d)
                     data = {'appType': '888888', 'crawlerSrcCode': 'KANYIKAN', 'viewStatus': '1', 'versionCode': '1'}
@@ -195,7 +190,7 @@ class Publish:
                     for fi in dir_files:
                         # 视频文件夹下的所有文件路径
                         fi_path = fi_d + '\\' + fi
-                        Common.crawler_log().info('dir fi_path = {}'.format(fi_path))
+                        Common.logger().info('dir fi_path = {}', fi_path)
                         # 读取 info.txt,赋值给 data
                         if cls.info_file in fi:
                             f = open(fi_path, "r", encoding="UTF-8")
@@ -204,7 +199,7 @@ class Publish:
                                 line = f.readline()
                                 line = line.replace('\n', '')
                                 if line is not None and len(line) != 0 and not line.isspace():
-                                    Common.crawler_log().info("line = {}".format(line))
+                                    Common.logger().info("line = {}", line)
                                     if i == 0:
                                         data['crawlerSrcId'] = line
                                     elif i == 1:
@@ -214,7 +209,7 @@ class Publish:
                                     elif i == 8:
                                         data['crawlerSrcPublishTimestamp'] = line
                                 else:
-                                    Common.crawler_log().warning("{} line is None".format(fi_path))
+                                    Common.logger().warning("{} line is None", fi_path)
                             f.close()
                             # remove info.txt
                             cls.remove_local_file(fi_path)
@@ -222,7 +217,7 @@ class Publish:
                     dir_files = os.listdir(fi_d)
                     for fi in dir_files:
                         fi_path = fi_d + '\\' + fi
-                        Common.crawler_log().info('dir fi_path = {}'.format(fi_path))
+                        Common.logger().info('dir fi_path = {}', fi_path)
                         # 上传oss
                         if cls.video_file in fi:
                             global oss_video_file
@@ -230,20 +225,20 @@ class Publish:
                                 oss_video_file = cls.oss_file_path_video.format("dev", today, data['crawlerSrcId'])
                             elif env == "prod":
                                 oss_video_file = cls.oss_file_path_video.format("prod", today, data['crawlerSrcId'])
-                            Common.crawler_log().info("oss_video_file = {}".format(oss_video_file))
+                            Common.logger().info("oss_video_file = {}", oss_video_file)
                             cls.put_file(oss_video_file, fi_path)
                             data['videoPath'] = oss_video_file
-                            Common.crawler_log().info("videoPath = {}".format(oss_video_file))
+                            Common.logger().info("videoPath = {}", oss_video_file)
                         elif cls.image_file in fi:
                             global oss_image_file
                             if env == "dev":
                                 oss_image_file = cls.oss_file_path_image.format("env", today, data['crawlerSrcId'])
                             elif env == "prod":
                                 oss_image_file = cls.oss_file_path_image.format("prod", today, data['crawlerSrcId'])
-                            Common.crawler_log().info("oss_image_file = {}".format(oss_image_file))
+                            Common.logger().info("oss_image_file = {}", oss_image_file)
                             cls.put_file(oss_image_file, fi_path)
                             data['coverImgPath'] = oss_image_file
-                            Common.crawler_log().info("coverImgPath = {}".format(oss_image_file))
+                            Common.logger().info("coverImgPath = {}", oss_image_file)
                         # 全部remove
                         cls.remove_local_file(fi_path)
 
@@ -255,6 +250,6 @@ class Publish:
                     cls.remove_local_file_dir(fi_d)
 
                 else:
-                    Common.crawler_log().error('file not a dir = {}'.format(fi_d))
+                    Common.logger().error('file not a dir = {}', fi_d)
             except Exception as e:
-                Common.crawler_log().exception('upload_and_publish error', e)
+                Common.logger().exception('upload_and_publish error', e)

+ 7 - 7
main/run.py

@@ -25,7 +25,7 @@ class Main:
             play_now = datetime.datetime.now()
             # 指定时间结束抓取视频
             if play_now.hour == 0:
-                Common.crawler_log().info("结束抓取播放量视频\n")
+                Common.logger().info("结束抓取播放量视频\n")
                 time.sleep(3)
                 break
             else:
@@ -49,7 +49,7 @@ class Main:
             up_now = datetime.datetime.now()
             # 指定时间结束抓取视频
             if up_now.hour > 19:
-                Common.crawler_log().info("结束抓取上升榜视频\n")
+                Common.logger().info("结束抓取上升榜视频\n")
                 time.sleep(3)
                 break
             else:
@@ -73,7 +73,7 @@ class Main:
             sendtime_now = datetime.datetime.now()
             # 指定时间结束抓取视频
             if sendtime_now.hour > 9:
-                Common.crawler_log().info("结束抓取3天榜单视频\n")
+                Common.logger().info("结束抓取3天榜单视频\n")
                 time.sleep(3)
                 break
             else:
@@ -99,19 +99,19 @@ class Main:
             while True:
                 main_now = datetime.datetime.now()
                 if 9 >= main_now.hour >= 0:
-                    Common.crawler_log().info("开始抓取3天榜单视频\n")
+                    Common.logger().info("开始抓取3天榜单视频\n")
                     time.sleep(1)
                     cls.download_sendtime_job()
                 elif 19 >= main_now.hour >= 10:
-                    Common.crawler_log().info("开始抓取上升榜视频\n")
+                    Common.logger().info("开始抓取上升榜视频\n")
                     time.sleep(1)
                     cls.download_up_job()
                 elif 24 >= main_now.hour >= 20:
-                    Common.crawler_log().info("开始抓取播放量视频\n")
+                    Common.logger().info("开始抓取播放量视频\n")
                     time.sleep(1)
                     cls.download_play_job()
                 else:
-                    Common.crawler_log().info("结束今天抓取及上传任务\n")
+                    Common.logger().info("结束今天抓取及上传任务\n")
                     time.sleep(60)
                     break
 

Alguns arquivos não foram mostrados porque muitos arquivos mudaram nesse diff