common.py 4.6 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137
  1. # -*- coding: utf-8 -*-
  2. # @Author: wangkun
  3. # @Time: 2022/5/27
  4. from datetime import date, timedelta
  5. from loguru import logger
  6. import datetime
  7. import os
  8. import time
  9. import requests
  10. import urllib3
  11. proxies = {"http": None, "https": None}
  12. # proxies = {"http": "127.0.0.1:19180", "https": "127.0.0.1:19180"}
  13. class Common:
  14. # 统一获取当前时间 <class 'datetime.datetime'> 2022-04-14 20:13:51.244472
  15. now = datetime.datetime.now()
  16. # 昨天 <class 'str'> 2022-04-13
  17. yesterday = (date.today() + timedelta(days=-1)).strftime("%Y-%m-%d")
  18. # 今天 <class 'datetime.date'> 2022-04-14
  19. today = date.today()
  20. # 明天 <class 'str'> 2022-04-15
  21. tomorrow = (date.today() + timedelta(days=1)).strftime("%Y-%m-%d")
  22. # 使用 logger 模块生成日志
  23. @staticmethod
  24. def logger():
  25. """
  26. 使用 logger 模块生成日志
  27. """
  28. # 日志路径
  29. log_dir = "./logs/"
  30. log_path = os.getcwd() + os.sep + log_dir
  31. if not os.path.isdir(log_path):
  32. os.makedirs(log_path)
  33. # 日志文件名
  34. log_name = time.strftime("%Y-%m-%d", time.localtime(time.time())) + '.log'
  35. # 日志不打印到控制台
  36. logger.remove(handler_id=None)
  37. # rotation="500 MB",实现每 500MB 存储一个文件
  38. # rotation="12:00",实现每天 12:00 创建一个文件
  39. # rotation="1 week",每周创建一个文件
  40. # retention="10 days",每隔10天之后就会清理旧的日志
  41. # 初始化日志
  42. logger.add(log_dir + log_name, level="INFO", rotation='00:00')
  43. return logger
  44. # 清除日志,保留最近 7 个文件
  45. @classmethod
  46. def del_logs(cls, d_dir):
  47. """
  48. 清除冗余日志文件
  49. :d_dir: 需要删除的 log 地址
  50. :return: 保留最近 7 个日志
  51. """
  52. if d_dir == "logs":
  53. logs_dir = "./logs/"
  54. all_files = sorted(os.listdir(logs_dir))
  55. all_logs = []
  56. for log in all_files:
  57. name = os.path.splitext(log)[-1]
  58. if name == ".log":
  59. all_logs.append(log)
  60. if len(all_logs) <= 7:
  61. pass
  62. else:
  63. for file in all_logs[:len(all_logs) - 7]:
  64. os.remove(logs_dir + file)
  65. cls.logger().info("清除冗余日志成功")
  66. # 封装下载视频或封面的方法
  67. @classmethod
  68. def download_method(cls, log_path, text, d_name, d_url):
  69. """
  70. 下载封面:text == "cover" ; 下载视频:text == "video"
  71. 需要下载的视频标题:d_title
  72. 视频封面,或视频播放地址:d_url
  73. 下载保存路径:"./files/{d_title}/"
  74. """
  75. # 首先创建一个保存该视频相关信息的文件夹
  76. video_dir = "./videos/" + d_name + "/"
  77. # video_dir = "./videos/"
  78. if not os.path.exists(video_dir):
  79. os.mkdir(video_dir)
  80. # 下载视频
  81. if text == "video":
  82. # 需要下载的视频地址
  83. video_url = d_url
  84. # 视频名
  85. video_name = "video.mp4"
  86. # # 视频名
  87. # video_name = d_name + ".mp4"
  88. # 下载视频
  89. urllib3.disable_warnings()
  90. response = requests.get(video_url, stream=True, proxies=proxies, verify=False)
  91. try:
  92. with open(video_dir + video_name, "wb") as f:
  93. for chunk in response.iter_content(chunk_size=10240):
  94. f.write(chunk)
  95. if log_path == "logs":
  96. cls.logger().info("==========视频下载完成==========")
  97. except Exception as e:
  98. if log_path == "logs":
  99. cls.logger().exception("视频下载失败:{}", e)
  100. # 下载封面
  101. elif text == "cover":
  102. # 需要下载的封面地址
  103. cover_url = d_url
  104. # 封面名
  105. cover_name = "image.jpg"
  106. # # 封面名
  107. # cover_name = d_name + ".jpg"
  108. # 下载封面
  109. urllib3.disable_warnings()
  110. response = requests.get(cover_url, proxies=proxies, verify=False)
  111. try:
  112. with open(video_dir + cover_name, "wb") as f:
  113. f.write(response.content)
  114. if log_path == "logs":
  115. cls.logger().info("==========封面下载完成==========")
  116. except Exception as e:
  117. if log_path == "logs":
  118. cls.logger().exception("封面下载失败:{}", e)
  119. if __name__ == "__main__":
  120. common = Common()