gzh_recommend.py 23 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455
  1. # -*- coding: utf-8 -*-
  2. # @Author: wangkun
  3. # @Time: 2022/8/1
  4. # import time
  5. # import base64
  6. import json
  7. import os
  8. import time
  9. # import urllib.parse
  10. import requests
  11. import urllib3
  12. from crawler_gzh.main.common import Common
  13. from crawler_gzh.main.feishu_lib import Feishu
  14. from crawler_gzh.main.publish import Publish
  15. proxies = {"http": None, "https": None}
  16. class Recommend:
  17. # 获取 token,保存至飞书云文档
  18. @classmethod
  19. def get_token(cls, log_type):
  20. # charles 抓包文件保存目录
  21. charles_file_dir = "./crawler-kanyikan-recommend/chlsfiles/"
  22. # charles_file_dir = "../chlsfiles/"
  23. if int(len(os.listdir(charles_file_dir))) == 1:
  24. Common.logger(log_type).info("未找到chlsfile文件,等待60s")
  25. time.sleep(60)
  26. else:
  27. try:
  28. # 目标文件夹下所有文件
  29. all_file = sorted(os.listdir(charles_file_dir))
  30. # 获取到目标文件
  31. old_file = all_file[-1]
  32. # 分离文件名与扩展名
  33. new_file = os.path.splitext(old_file)
  34. # 重命名文件后缀
  35. os.rename(os.path.join(charles_file_dir, old_file),
  36. os.path.join(charles_file_dir, new_file[0] + ".txt"))
  37. with open(charles_file_dir + new_file[0] + ".txt", encoding='utf-8-sig', errors='ignore') as f:
  38. contents = json.load(f, strict=False)
  39. for content in contents:
  40. if "mp.weixin.qq.com" in content['host']:
  41. if content["path"] == r"/mp/getappmsgext":
  42. # query
  43. query = content["query"]
  44. Feishu.update_values("recommend", "gzh", "VzrN7E", "B9:B9", [[query]])
  45. # body
  46. headers = content["request"]["header"]["headers"]
  47. body = content["request"]["body"]["text"]
  48. # time.sleep(1)
  49. Feishu.update_values("recommend", "gzh", "VzrN7E", "B8:B8", [[body]])
  50. # title / vid
  51. title = content["request"]["body"]["text"].split("title=")[-1].split("&ct=")[0]
  52. vid = content["request"]["body"]["text"].split("vid=")[-1].split("&is_pay_subscribe")[0]
  53. # time.sleep(1)
  54. Feishu.update_values("recommend", "gzh", "VzrN7E", "B1:B1", [[title]])
  55. # time.sleep(1)
  56. Feishu.update_values("recommend", "gzh", "VzrN7E", "B2:B2", [[vid]])
  57. for h in headers:
  58. if h["name"] == "cookie" and "pass_ticket" in h["value"]:
  59. pass_ticket = h["value"].split("pass_ticket=")[-1]
  60. # print(f"pass_ticket:{pass_ticket}")
  61. Feishu.update_values("recommend", "gzh", "VzrN7E", "B5:B5", [[pass_ticket]])
  62. if h["name"] == "referer":
  63. referer = h["value"]
  64. # print(f"__biz:{referer}")
  65. Feishu.update_values("recommend", "gzh", "VzrN7E", "B7:B7", [[referer]])
  66. if h["name"] == "referer":
  67. __biz = h["value"].split("__biz=")[-1].split("&mid=")[0]
  68. # print(f"__biz:{__biz}")
  69. Feishu.update_values("recommend", "gzh", "VzrN7E", "B3:B3", [[__biz]])
  70. if h["name"] == "cookie" and "appmsg_token" in h["value"]:
  71. appmsg_token = h["value"].split("appmsg_token=")[-1]
  72. # print(f"appmsg_token:{appmsg_token}")
  73. Feishu.update_values("recommend", "gzh", "VzrN7E", "B4:B4", [[appmsg_token]])
  74. if h["name"] == "cookie" and "wap_sid2" in h["value"]:
  75. wap_sid2 = h["value"].split("wap_sid2=")[-1]
  76. # print(f"wap_sid2:{wap_sid2}")
  77. Feishu.update_values("recommend", "gzh", "VzrN7E", "B6:B6", [[wap_sid2]])
  78. except Exception as e:
  79. Common.logger(log_type).error("获取session异常,30s后重试:{}", e)
  80. time.sleep(30)
  81. cls.get_token(log_type)
  82. # 获取推荐列表
  83. @classmethod
  84. def get_recommend(cls, log_type):
  85. try:
  86. token_sheet = Feishu.get_values_batch("recommend", "gzh", "VzrN7E")
  87. if token_sheet is None:
  88. Common.logger(log_type).info("未获取到token等信息,30s后重试")
  89. time.sleep(30)
  90. cls.get_recommend(log_type)
  91. else:
  92. # __biz = token_sheet[2][1]
  93. appmsg_token = token_sheet[3][1]
  94. pass_ticket = token_sheet[4][1]
  95. wap_sid2 = token_sheet[5][1]
  96. referer = token_sheet[6][1]
  97. body = token_sheet[7][1]
  98. query = token_sheet[8][1]
  99. url = "https://mp.weixin.qq.com/mp/getappmsgext?"
  100. headers = {
  101. # "content-type": "application/x-www-form-urlencoded; charset=UTF-8",
  102. "content-type": 'text/plain',
  103. "accept": "*/*",
  104. "x-requested-with": "XMLHttpRequest",
  105. "accept-language": "zh-cn",
  106. "accept-encoding": "gzip, deflate, br",
  107. "origin": "https://mp.weixin.qq.com",
  108. "user-agent": "Mozilla/5.0 (iPhone; CPU iPhone OS 14_7_1 like Mac OS X) AppleWebKit/605.1.15 "
  109. "(KHTML, like Gecko) Mobile/15E148 MicroMessenger/8.0.26(0x18001a29)"
  110. " NetType/WIFI Language/zh_CN",
  111. "referer": referer
  112. }
  113. cookies = {
  114. "appmsg_token": appmsg_token,
  115. "devicetype": "iOS14.7.1",
  116. "lang": "zh_CN",
  117. "pass_ticket": pass_ticket,
  118. "rewardsn": "",
  119. "version": "18001a29",
  120. "wap_sid2": wap_sid2,
  121. "wxtokenkey": "777",
  122. "wxuin": "2010747860"
  123. }
  124. urllib3.disable_warnings()
  125. response = requests.post(url=url, headers=headers, cookies=cookies, params=query, data=body,
  126. verify=False, proxies=proxies)
  127. if "related_tag_video" not in response.json():
  128. Common.logger(log_type).warning("response:{}\n", response.text)
  129. elif len(response.json()["related_tag_video"]) == 0:
  130. Common.logger(log_type).warning("response:{}\n", response.text)
  131. time.sleep(10)
  132. cls.get_recommend(log_type)
  133. else:
  134. feeds = response.json()["related_tag_video"]
  135. for m in range(len(feeds)):
  136. # video_title
  137. if "title" not in feeds[m]:
  138. video_title = 0
  139. else:
  140. video_title = feeds[m]["title"]
  141. # video_title = base64.b64decode(video_title).decode("utf-8")
  142. # video_id
  143. if "vid" not in feeds[m]:
  144. video_id = 0
  145. else:
  146. video_id = feeds[m]["vid"]
  147. # play_cnt
  148. if "read_num" not in feeds[m]:
  149. play_cnt = 0
  150. else:
  151. play_cnt = feeds[m]["read_num"]
  152. # like_cnt
  153. if "like_num" not in feeds[m]:
  154. like_cnt = 0
  155. else:
  156. like_cnt = feeds[m]["like_num"]
  157. # duration
  158. if "duration" not in feeds[m]:
  159. duration = 0
  160. else:
  161. duration = feeds[m]["duration"]
  162. # video_width / video_height
  163. if "videoWidth" not in feeds[m] or "videoHeight" not in feeds[m]:
  164. video_width = 0
  165. video_height = 0
  166. else:
  167. video_width = feeds[m]["videoWidth"]
  168. video_height = feeds[m]["videoHeight"]
  169. # send_time
  170. if "pubTime" not in feeds[m]:
  171. send_time = 0
  172. else:
  173. send_time = feeds[m]["pubTime"]
  174. # user_name
  175. if "srcDisplayName" not in feeds[m]:
  176. user_name = 0
  177. else:
  178. user_name = feeds[m]["srcDisplayName"]
  179. # user_name = base64.b64decode(user_name).decode("utf-8")
  180. # user_id
  181. if "srcUserName" not in feeds[m]:
  182. user_id = 0
  183. else:
  184. user_id = feeds[m]["srcUserName"]
  185. # head_url
  186. if "head_img_url" not in feeds[m]:
  187. head_url = 0
  188. else:
  189. head_url = feeds[m]["head_img_url"]
  190. # cover_url
  191. if "cover" not in feeds[m]:
  192. cover_url = 0
  193. else:
  194. cover_url = feeds[m]["cover"]
  195. # video_url
  196. if "url" not in feeds[m]:
  197. video_url = 0
  198. else:
  199. video_url = feeds[m]["url"]
  200. # 下载链接
  201. download_url = cls.get_url(log_type, video_url)
  202. Common.logger(log_type).info("video_title:{}", video_title)
  203. Common.logger(log_type).info("video_id:{}", video_id)
  204. Common.logger(log_type).info("play_cnt:{}", play_cnt)
  205. Common.logger(log_type).info("like_cnt:{}", like_cnt)
  206. Common.logger(log_type).info("duration:{}", duration)
  207. Common.logger(log_type).info("video_width:{}", video_width)
  208. Common.logger(log_type).info("video_height:{}", video_height)
  209. Common.logger(log_type).info("send_time:{}", send_time)
  210. Common.logger(log_type).info("user_name:{}", user_name)
  211. Common.logger(log_type).info("user_id:{}", user_id)
  212. Common.logger(log_type).info("head_url:{}", head_url)
  213. Common.logger(log_type).info("cover_url:{}", cover_url)
  214. Common.logger(log_type).info("video_url:{}", video_url)
  215. Common.logger(log_type).info("download_url:{}", download_url)
  216. if video_id == 0 or video_title == 0 or duration == 0 or video_url == 0:
  217. Common.logger(log_type).info("无效视频\n")
  218. elif str(video_id) in [x for y in Feishu.get_values_batch(log_type, "gzh", "fCs3BT") for x in
  219. y]:
  220. Common.logger(log_type).info("该视频已下载\n")
  221. elif str(video_id) in [x for y in Feishu.get_values_batch(log_type, "gzh", "zWKFGb") for x in
  222. y]:
  223. Common.logger(log_type).info("该视频已在feeds中\n")
  224. else:
  225. Feishu.insert_columns(log_type, "gzh", "zWKFGb", "ROWS", 1, 2)
  226. get_feeds_time = int(time.time())
  227. values = [[time.strftime("%Y/%m/%d %H:%M:%S", time.localtime(get_feeds_time)),
  228. "推荐榜",
  229. video_title,
  230. str(video_id),
  231. play_cnt,
  232. like_cnt,
  233. duration,
  234. str(video_width) + "*" + str(video_height),
  235. time.strftime("%Y/%m/%d %H:%M:%S", time.localtime(send_time)),
  236. user_name,
  237. user_id,
  238. head_url,
  239. cover_url,
  240. video_url,
  241. download_url
  242. ]]
  243. time.sleep(1)
  244. Feishu.update_values(log_type, "gzh", "zWKFGb", "D2:T2", values)
  245. Common.logger(log_type).info("添加至recommend_feeds成功\n")
  246. except Exception as e:
  247. Common.logger(log_type).error("get_recommend异常:{}", e)
  248. # 获取视频下载链接
  249. @classmethod
  250. def get_url(cls, log_type, url):
  251. try:
  252. payload = {}
  253. headers = {
  254. 'Cookie': 'rewardsn=; wxtokenkey=777'
  255. }
  256. urllib3.disable_warnings()
  257. response = requests.get(url=url, headers=headers, data=payload, verify=False, proxies=proxies)
  258. response_list = response.text.splitlines()
  259. video_url_list = []
  260. for m in response_list:
  261. if "mpvideo.qpic.cn" in m:
  262. video_url = m.split("url: '")[1].split("',")[0].replace(r"\x26amp;", "&")
  263. video_url_list.append(video_url)
  264. video_url = video_url_list[0]
  265. return video_url
  266. except Exception as e:
  267. Common.logger(log_type).error("get_url异常:{}", e)
  268. # 下载/上传
  269. @classmethod
  270. def download_publish(cls, log_type, env):
  271. try:
  272. recommend_feeds_sheet = Feishu.get_values_batch(log_type, "gzh", "zWKFGb")
  273. for i in range(1, len(recommend_feeds_sheet)):
  274. download_video_title = recommend_feeds_sheet[i][5]
  275. download_video_id = recommend_feeds_sheet[i][6]
  276. download_video_play_cnt = recommend_feeds_sheet[i][7]
  277. download_video_like_cnt = recommend_feeds_sheet[i][8]
  278. download_video_duration = recommend_feeds_sheet[i][9]
  279. download_width_height = recommend_feeds_sheet[i][10]
  280. download_video_send_time = recommend_feeds_sheet[i][11]
  281. download_user_name = recommend_feeds_sheet[i][12]
  282. download_user_id = recommend_feeds_sheet[i][13]
  283. download_head_url = recommend_feeds_sheet[i][14]
  284. download_cover_url = recommend_feeds_sheet[i][15]
  285. download_video_url = recommend_feeds_sheet[i][17]
  286. download_video_comment_cnt = 0
  287. download_video_share_cnt = 0
  288. Common.logger(log_type).info("正在判断第{}行", i + 1)
  289. Common.logger(log_type).info("download_video_title:{}", download_video_title)
  290. Common.logger(log_type).info("download_video_id:{}", download_video_id)
  291. Common.logger(log_type).info("download_video_play_cnt:{}", download_video_play_cnt)
  292. Common.logger(log_type).info("download_video_duration:{}", download_video_duration)
  293. Common.logger(log_type).info("download_video_send_time:{}", download_video_send_time)
  294. Common.logger(log_type).info("download_video_url:{}\n", download_video_url)
  295. # Common.logger(log_type).info("download_video_like_cnt:{}", download_video_like_cnt)
  296. # Common.logger(log_type).info("download_width_height:{}", download_width_height)
  297. # Common.logger(log_type).info("download_user_name:{}", download_user_name)
  298. # Common.logger(log_type).info("download_user_id:{}", download_user_id)
  299. # Common.logger(log_type).info("download_head_url:{}", download_head_url)
  300. # Common.logger(log_type).info("download_cover_url:{}", download_cover_url)
  301. # 过滤空行
  302. if download_video_id is None or download_video_title is None or download_video_play_cnt is None:
  303. Common.logger(log_type).warning("空行,略过\n")
  304. # # 过滤敏感词
  305. # elif any(word if word in download_video_title else False for word in
  306. # cls.sensitive_words(log_type)) is True:
  307. # Feishu.dimension_range(log_type, "music_album", "69UxPo", "ROWS", i + 1, i + 1)
  308. # Common.logger(log_type).info("视频已中敏感词,删除成功\n")
  309. # return
  310. # # 下载规则
  311. # elif cls.download_rule(download_video_share_cnt, download_video_play_cnt) is False:
  312. # Feishu.dimension_range(log_type, "music_album", "69UxPo", "ROWS", i + 1, i + 1)
  313. # Common.logger(log_type).info("不满足下载规则,删除成功\n")
  314. # return
  315. # 时长小于 60s,删除
  316. elif int(download_video_duration) < 60:
  317. Feishu.dimension_range(log_type, "gzh", "zWKFGb", "ROWS", i + 1, i + 1)
  318. Common.logger(log_type).info("时长{}<60,删除成功\n", download_video_duration)
  319. return
  320. # 已下载视频表去重
  321. elif str(download_video_id) in [n for m in Feishu.get_values_batch(log_type, "gzh", "fCs3BT")
  322. for n in m]:
  323. Feishu.dimension_range(log_type, "gzh", "zWKFGb", "ROWS", i + 1, i + 1)
  324. Common.logger(log_type).info("该视频在公众号中已下载,删除成功\n")
  325. return
  326. # 看一看已下载表去重
  327. elif str(download_video_id) in [n for m in Feishu.get_values_batch(log_type, "kanyikan", "20ce0c")
  328. for n in m]:
  329. Feishu.dimension_range(log_type, "gzh", "zWKFGb", "ROWS", i + 1, i + 1)
  330. Common.logger(log_type).info("该视频在看一看中已下载,删除成功\n")
  331. return
  332. else:
  333. # 下载封面
  334. Common.download_method(log_type=log_type, text="cover",
  335. d_name=str(download_video_title), d_url=str(download_cover_url))
  336. # 下载视频
  337. Common.download_method(log_type=log_type, text="video",
  338. d_name=str(download_video_title), d_url=str(download_video_url))
  339. # 保存视频信息至 "./videos/{download_video_title}/info.txt"
  340. with open("./crawler_gzh/videos/" + download_video_title + "/" + "info.txt",
  341. "a", encoding="UTF-8") as f_a:
  342. f_a.write(str(download_video_id) + "\n" +
  343. str(download_video_title) + "\n" +
  344. str(download_video_duration) + "\n" +
  345. str(download_video_play_cnt) + "\n" +
  346. str(download_video_comment_cnt) + "\n" +
  347. str(download_video_like_cnt) + "\n" +
  348. str(download_video_share_cnt) + "\n" +
  349. str(download_width_height) + "\n" +
  350. str(int(time.mktime(
  351. time.strptime(download_video_send_time, "%Y/%m/%d %H:%M:%S")))) + "\n" +
  352. str(download_user_name) + "\n" +
  353. str(download_head_url) + "\n" +
  354. str(download_video_url) + "\n" +
  355. str(download_cover_url) + "\n" +
  356. "gzh")
  357. Common.logger(log_type).info("==========视频信息已保存至info.txt==========")
  358. # 上传视频
  359. Common.logger(log_type).info("开始上传视频:{}".format(download_video_title))
  360. our_video_id = Publish.upload_and_publish(log_type, env, "play")
  361. our_video_link = "https://admin.piaoquantv.com/cms/post-detail/" + str(our_video_id) + "/info"
  362. Common.logger(log_type).info("视频上传完成:{}", download_video_title)
  363. # 保存视频 ID 到云文档
  364. Common.logger(log_type).info("保存视频ID至云文档:{}", download_video_title)
  365. # 视频ID工作表,插入首行
  366. Feishu.insert_columns(log_type, "gzh", "fCs3BT", "ROWS", 1, 2)
  367. # 视频ID工作表,首行写入数据
  368. upload_time = int(time.time())
  369. values = [[time.strftime("%Y/%m/%d %H:%M:%S", time.localtime(upload_time)),
  370. "推荐榜",
  371. str(download_video_title),
  372. str(download_video_id),
  373. our_video_link,
  374. download_video_play_cnt,
  375. download_video_like_cnt,
  376. download_video_duration,
  377. str(download_width_height),
  378. str(download_video_send_time),
  379. str(download_user_name),
  380. str(download_user_id),
  381. str(download_head_url),
  382. str(download_cover_url),
  383. str(download_video_url)]]
  384. time.sleep(1)
  385. Feishu.update_values(log_type, "gzh", "fCs3BT", "D2:W2", values)
  386. # 删除行或列,可选 ROWS、COLUMNS
  387. Feishu.dimension_range(log_type, "gzh", "zWKFGb", "ROWS", i + 1, i + 1)
  388. Common.logger(log_type).info("视频:{},下载/上传成功\n", download_video_title)
  389. return
  390. except Exception as e:
  391. Common.logger(log_type).error("download_publish异常:{}", e)
  392. # 执行下载/上传
  393. @classmethod
  394. def run_download_publish(cls, log_type, env):
  395. try:
  396. while True:
  397. recommend_feeds_sheet = Feishu.get_values_batch(log_type, "gzh", "zWKFGb")
  398. if len(recommend_feeds_sheet) == 1:
  399. Common.logger(log_type).info("下载/上传完成")
  400. break
  401. else:
  402. cls.download_publish(log_type, env)
  403. except Exception as e:
  404. Common.logger(log_type).error("run_download_publish异常:{}", e)
  405. if __name__ == "__main__":
  406. Recommend.get_recommend("recommend")
  407. # Recommend.download_publish("recommend")
  408. # Recommend.run_download_publish("recommend", "dev")
  409. # token = Recommend.get_token("recommend")
  410. # print(token)