run_douyin_author_scheduling.py 1.9 KB

12345678910111213141516171819202122232425262728293031323334353637383940414243
  1. # -*- coding: utf-8 -*-
  2. # @Author: wangkun
  3. # @Time: 2023/5/26
  4. import argparse
  5. import os
  6. import sys
  7. sys.path.append(os.getcwd())
  8. from common.common import Common
  9. from common.public import task_fun
  10. from common.scheduling_db import MysqlHelper
  11. from douyin.douyin_author.douyin_author_scheduling import DouyinauthorScheduling
  12. def main(log_type, crawler, task, env):
  13. task_dict = task_fun(task)['task_dict']
  14. rule_dict = task_fun(task)['rule_dict']
  15. task_id = task_dict['task_id']
  16. select_user_sql = f"""select * from crawler_user_v3 where task_id={task_id}"""
  17. user_list = MysqlHelper.get_values(log_type, crawler, select_user_sql, env, action="")
  18. Common.logger(log_type, crawler).info(f"调度任务:{task_dict}")
  19. Common.logger(log_type, crawler).info(f"抓取规则:{rule_dict}")
  20. # Common.logger(log_type, crawler).info(f"用户列表:{user_list}\n")
  21. Common.logger(log_type, crawler).info(f'开始抓取 {task_dict["task_name"]}\n')
  22. DouyinauthorScheduling.get_author_videos(log_type=log_type,
  23. crawler=crawler,
  24. rule_dict=rule_dict,
  25. user_list=user_list,
  26. env=env)
  27. Common.del_logs(log_type, crawler)
  28. Common.logger(log_type, crawler).info('抓取完一轮\n')
  29. if __name__ == "__main__":
  30. parser = argparse.ArgumentParser() ## 新建参数解释器对象
  31. parser.add_argument('--log_type', default='recommend') ## 添加参数,注明参数类型
  32. parser.add_argument('--crawler', default='kuaishou') ## 添加参数
  33. parser.add_argument('--task') ## 添加参数
  34. parser.add_argument('--env', default='prod') ## 添加参数
  35. args = parser.parse_args() ### 参数赋值,也可以通过终端赋值
  36. main(log_type=args.log_type,
  37. crawler=args.crawler,
  38. task=args.task,
  39. env=args.env)