zhangyong vor 2 Monaten
Ursprung
Commit
3a6a509c1b
4 geänderte Dateien mit 22 neuen und 21 gelöschten Zeilen
  1. 0 1
      utils/__init__.py
  2. 19 19
      utils/feishu_utils.py
  3. 1 1
      workers/consumption_work.py
  4. 2 0
      workers/select_work.py

+ 0 - 1
utils/__init__.py

@@ -1 +0,0 @@
-from .common_log import Common

+ 19 - 19
utils/feishu_utils.py

@@ -7,9 +7,9 @@ import os
 import sys
 import requests
 import urllib3
+from loguru import logger
 
 sys.path.append(os.getcwd())
-from utils import Common
 
 proxies = {"http": None, "https": None}
 
@@ -46,7 +46,7 @@ class Feishu:
             tenant_access_token = response.json()["tenant_access_token"]
             return tenant_access_token
         except Exception as e:
-            Common.logger("feishu").error("获取飞书 api token 异常:{}", e)
+            logger.error("获取飞书 api token 异常:{}", e)
 
     # 获取表格元数据
     @classmethod
@@ -72,7 +72,7 @@ class Feishu:
             response = json.loads(r.content.decode("utf8"))
             return response
         except Exception as e:
-            Common.logger("feishu").error("获取表格元数据异常:{}", e)
+            logger.error("获取表格元数据异常:{}", e)
 
     # 读取工作表中所有数据
     @classmethod
@@ -102,7 +102,7 @@ class Feishu:
             values = response["data"]["valueRanges"][0]["values"]
             return values
         except Exception as e:
-            Common.logger("feishu").error("读取工作表所有数据异常:{}", e)
+            logger.error("读取工作表所有数据异常:{}", e)
 
     # 工作表,插入行或列
     @classmethod
@@ -135,9 +135,9 @@ class Feishu:
 
             urllib3.disable_warnings()
             r = requests.post(url=insert_columns_url, headers=headers, json=body, proxies=proxies, verify=False)
-            Common.logger("feishu").info("插入行或列:{}", r.json()["msg"])
+            logger.info("插入行或列:{}", r.json()["msg"])
         except Exception as e:
-            Common.logger("feishu").error("插入行或列异常:{}", e)
+            logger.error("插入行或列异常:{}", e)
 
     # 写入数据
     @classmethod
@@ -167,9 +167,9 @@ class Feishu:
             }
             urllib3.disable_warnings()
             r = requests.post(url=update_values_url, headers=headers, json=body, proxies=proxies, verify=False)
-            Common.logger("feishu").info("写入数据:{}", r.json()["msg"])
+            logger.info("写入数据:{}", r.json()["msg"])
         except Exception as e:
-            Common.logger("feishu").error("写入数据异常:{}", e)
+            logger.error("写入数据异常:{}", e)
 
     # 合并单元格
     @classmethod
@@ -195,9 +195,9 @@ class Feishu:
             }
             urllib3.disable_warnings()
             r = requests.post(url=merge_cells_url, headers=headers, json=body, proxies=proxies, verify=False)
-            Common.logger("feishu").info("合并单元格:{}", r.json()["msg"])
+            logger.info("合并单元格:{}", r.json()["msg"])
         except Exception as e:
-            Common.logger("feishu").error("合并单元格异常:{}", e)
+            logger.error("合并单元格异常:{}", e)
 
     # 读取单元格数据
     @classmethod
@@ -231,7 +231,7 @@ class Feishu:
             # print(r.text)
             return r.json()["data"]["valueRange"]["values"][0]
         except Exception as e:
-            Common.logger("feishu").error("读取单元格数据异常:{}", e)
+            logger.error("读取单元格数据异常:{}", e)
     # 获取表内容
     @classmethod
     def get_sheet_content(cls, crawler, sheet_id):
@@ -246,7 +246,7 @@ class Feishu:
                         content_list.append(y)
             return content_list
         except Exception as e:
-            Common.logger("feishu").error(f'get_sheet_content:{e}\n')
+            logger.error(f'get_sheet_content:{e}\n')
 
     # 删除行或列,可选 ROWS、COLUMNS
     @classmethod
@@ -278,9 +278,9 @@ class Feishu:
             }
             urllib3.disable_warnings()
             r = requests.delete(url=dimension_range_url, headers=headers, json=body, proxies=proxies, verify=False)
-            Common.logger("feishu").info("删除视频数据:{}", r.json()["msg"])
+            logger.info("删除视频数据:{}", r.json()["msg"])
         except Exception as e:
-            Common.logger("feishu").error("删除视频数据异常:{}", e)
+            logger.error("删除视频数据异常:{}", e)
 
     # 获取用户 ID
     @classmethod
@@ -310,7 +310,7 @@ class Feishu:
 
             return open_id
         except Exception as e:
-            Common.logger("feishu").error(f"get_userid异常:{e}\n")
+            logger.error(f"get_userid异常:{e}\n")
 
     # 飞书机器人
     @classmethod
@@ -368,9 +368,9 @@ class Feishu:
             })
             urllib3.disable_warnings()
             r = requests.post(url, headers=headers, data=data, verify=False, proxies=proxies)
-            Common.logger("feishu").info(f'触发机器人消息:{r.status_code}, {text}')
+            logger.info(f'触发机器人消息:{r.status_code}, {text}')
         except Exception as e:
-            Common.logger("feishu").error(f"bot异常:{e}\n")
+            logger.error(f"bot异常:{e}\n")
 
     # 飞书机器人-改造计划完成通知
     @classmethod
@@ -401,9 +401,9 @@ class Feishu:
             })
             urllib3.disable_warnings()
             r = requests.post(url, headers=headers, data=data, verify=False, proxies=proxies)
-            Common.logger("feishu").info(f'触发机器人消息:{r.status_code}, {text}')
+            logger.info(f'触发机器人消息:{r.status_code}, {text}')
         except Exception as e:
-            Common.logger("feishu").error(f"bot异常:{e}\n")
+            logger.error(f"bot异常:{e}\n")
 
 
 if __name__ == "__main__":

+ 1 - 1
workers/consumption_work.py

@@ -6,13 +6,13 @@ import orjson
 from apscheduler.schedulers.asyncio import AsyncIOScheduler
 from apscheduler.triggers.interval import IntervalTrigger
 from loguru import logger
+sys.path.append('/app')
 
 from utils.aliyun_log import AliyunLogger
 from utils.google_ai_studio import GoogleAI
 from utils.piaoquan import PQ
 from utils.redis import RedisHelper, content_video_data
 
-sys.path.append('/app')
 
 
 

+ 2 - 0
workers/select_work.py

@@ -1,8 +1,10 @@
 import datetime
+import sys
 import time
 
 import schedule
 from loguru import logger
+sys.path.append('/app')
 
 from utils.odps_data import OdpsDataCount
 from utils.redis import RedisHelper