history_task.py 7.7 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229
  1. """
  2. @author: luojunhui
  3. """
  4. import json
  5. import time
  6. import asyncio
  7. from applications.config import Config
  8. from applications.log import logging
  9. from applications.functions.pqFunctions import publish_to_pq
  10. from applications.functions.common import shuffle_list
  11. class historyContentIdTask(object):
  12. """
  13. 处理已经匹配过小程序的文章
  14. """
  15. def __init__(self, mysql_client):
  16. """
  17. :param mysql_client:
  18. """
  19. self.mysql_client = mysql_client
  20. self.config = Config()
  21. self.article_match_video_table = self.config.article_match_video_table
  22. self.article_text_table = self.config.article_text_table
  23. self.article_crawler_video_table = self.config.article_crawler_video_table
  24. self.gh_id_dict = json.loads(self.config.get_config_value("testAccountLevel2"))
  25. self.history_coroutines = self.config.get_config_value("historyArticleCoroutines")
  26. async def get_tasks(self):
  27. """
  28. 获取任务
  29. :return:
  30. """
  31. select_sql1 = f"""
  32. SELECT
  33. ART.trace_id,
  34. ART.content_id,
  35. ART.flow_pool_level,
  36. ART.gh_id,
  37. ART.process_times
  38. FROM {self.article_match_video_table} ART
  39. JOIN (
  40. select content_id, count(1) as cnt
  41. from {self.article_crawler_video_table}
  42. where download_status = 2
  43. group by content_id
  44. ) VID on ART.content_id = VID.content_id and VID.cnt >= 3
  45. WHERE ART.content_status = 0 and ART.process_times <= 3
  46. ORDER BY request_timestamp
  47. LIMIT {self.history_coroutines};
  48. """
  49. tasks = await self.mysql_client.async_select(sql=select_sql1)
  50. task_obj_list = [
  51. {
  52. "trace_id": item[0],
  53. "content_id": item[1],
  54. "flow_pool_level": item[2],
  55. "gh_id": item[3],
  56. "process_times": item[4]
  57. } for item in tasks
  58. ]
  59. logging(
  60. code="9001",
  61. info="本次任务获取到 {} 条视频".format(len(task_obj_list)),
  62. data=task_obj_list
  63. )
  64. return task_obj_list
  65. async def get_video_list(self, content_id):
  66. """
  67. content_id
  68. :return:
  69. """
  70. sql = f"""
  71. SELECT platform, play_count, like_count, video_oss_path, cover_oss_path, user_id
  72. FROM {self.article_crawler_video_table}
  73. WHERE content_id = '{content_id}' and download_status = 2;
  74. """
  75. res_tuple = await self.mysql_client.async_select(sql)
  76. if len(res_tuple) >= 3:
  77. return [
  78. {
  79. "platform": i[0],
  80. "play_count": i[1],
  81. "like_count": i[2],
  82. "video_oss_path": i[3],
  83. "cover_oss_path": i[4],
  84. "uid": i[5]
  85. }
  86. for i in res_tuple]
  87. else:
  88. return []
  89. async def get_kimi_title(self, content_id):
  90. """
  91. 获取 kimiTitle
  92. :param content_id:
  93. :return:
  94. """
  95. select_sql = f"""
  96. select kimi_title from {self.article_text_table} where content_id = '{content_id}';
  97. """
  98. res_tuple = await self.mysql_client.async_select(select_sql)
  99. if res_tuple:
  100. return res_tuple[0][0]
  101. else:
  102. return False
  103. async def publish_videos_to_pq(self, trace_id, flow_pool_level, kimi_title, gh_id, download_videos, process_times):
  104. """
  105. 发布至 pq
  106. :param process_times:
  107. :param trace_id:
  108. :param download_videos: 已下载的视频---> list [{}, {}, {}.... ]
  109. :param gh_id: 公众号 id ---> str
  110. :param kimi_title: kimi 标题 ---> str
  111. :param flow_pool_level: 流量池层级 ---> str
  112. :return:
  113. """
  114. published_status = 4
  115. match flow_pool_level:
  116. case "autoArticlePoolLevel4":
  117. # 冷启层, 全量做
  118. video_list = shuffle_list(download_videos)[:3]
  119. case "autoArticlePoolLevel3":
  120. # 次条,只针对具体账号做
  121. if self.gh_id_dict.get(gh_id):
  122. video_list = shuffle_list(download_videos)[:3]
  123. else:
  124. video_list = download_videos[:3]
  125. case "autoArticlePoolLevel2":
  126. video_list = []
  127. case "autoArticlePoolLevel1":
  128. # 头条,先不做
  129. video_list = download_videos[:3]
  130. case _:
  131. print("未传流量池信息")
  132. video_list = download_videos[:3]
  133. L = []
  134. for video_obj in video_list:
  135. params = {
  136. "videoPath": video_obj['video_oss_path'],
  137. "uid": video_obj['uid'],
  138. "title": kimi_title
  139. }
  140. response = await publish_to_pq(params)
  141. time.sleep(2)
  142. obj = {
  143. "uid": video_obj['uid'],
  144. "source": video_obj['platform'],
  145. "kimiTitle": kimi_title,
  146. "videoId": response['data']['id'],
  147. "videoCover": response['data']['shareImgPath'],
  148. "videoPath": response['data']['videoPath'],
  149. "videoOss": video_obj['video_oss_path']
  150. }
  151. L.append(obj)
  152. update_sql = f"""
  153. UPDATE {self.article_match_video_table}
  154. SET content_status = %s, response = %s, process_times = %s
  155. WHERE trace_id = %s;
  156. """
  157. await self.mysql_client.async_insert(
  158. sql=update_sql,
  159. params=(published_status, json.dumps(L, ensure_ascii=False), process_times + 1, trace_id)
  160. )
  161. logging(
  162. code="9002",
  163. info="已经从历史文章更新",
  164. trace_id=trace_id
  165. )
  166. async def process_task(self, params):
  167. """
  168. 异步执行
  169. :param params:
  170. :return:
  171. """
  172. content_id = params['content_id']
  173. trace_id = params['trace_id']
  174. flow_pool_level = params['flow_pool_level']
  175. gh_id = params['gh_id']
  176. process_times = params['process_times']
  177. download_videos = await self.get_video_list(content_id=content_id)
  178. if download_videos:
  179. # 把状态修改为 4
  180. """
  181. todo: 加上状态锁,防止多个进程同时处理一个视频, 通过 update_time && content_id来判断
  182. """
  183. # update_sql = f"""
  184. # UPDATE {self.article_crawler_video_table}
  185. # SET content_status = %s
  186. # WHERE trace_id = %s;
  187. # """
  188. # await self.mysql_client.asyncInsert(
  189. # sql=update_sql,
  190. # params=(4, trace_id)
  191. # )
  192. kimi_title = await self.get_kimi_title(content_id)
  193. await self.publish_videos_to_pq(
  194. flow_pool_level=flow_pool_level,
  195. kimi_title=kimi_title,
  196. gh_id=gh_id,
  197. trace_id=trace_id,
  198. download_videos=download_videos,
  199. process_times=process_times
  200. )
  201. else:
  202. pass
  203. async def deal(self):
  204. """
  205. 处理
  206. :return:
  207. """
  208. task_list = await self.get_tasks()
  209. logging(
  210. code="5002",
  211. info="History content_task Task Got {} this time".format(len(task_list)),
  212. function="History Contents Task"
  213. )
  214. if task_list:
  215. tasks = [self.process_task(params) for params in task_list]
  216. await asyncio.gather(*tasks)
  217. else:
  218. print("暂时未获得历史已存在文章")