|
@@ -10,10 +10,10 @@ sys.path.append(os.getcwd())
|
|
from common.public import task_fun_mq, get_consumer, ack_message
|
|
from common.public import task_fun_mq, get_consumer, ack_message
|
|
from common.common import Common
|
|
from common.common import Common
|
|
from common.scheduling_db import MysqlHelper
|
|
from common.scheduling_db import MysqlHelper
|
|
-from gongzhonghao.gongzhonghao_author.gongzhonghao_author import GongzhonghaoAuthor
|
|
|
|
|
|
+from gongzhonghao.gongzhonghao_author.gongzhonghao_author_lock import GongzhonghaoAuthor
|
|
|
|
|
|
|
|
|
|
-def get_author_videos(log_type, crawler, token_index, task_dict, rule_dict, user_list, env):
|
|
|
|
|
|
+def get_author_videos(log_type, crawler, task_dict, rule_dict, user_list, env):
|
|
Common.logger(log_type, crawler).info(f'开始抓取:{task_dict["taskName"]}\n')
|
|
Common.logger(log_type, crawler).info(f'开始抓取:{task_dict["taskName"]}\n')
|
|
Common.logging(log_type, crawler, env, f'开始抓取:{task_dict["taskName"]}\n')
|
|
Common.logging(log_type, crawler, env, f'开始抓取:{task_dict["taskName"]}\n')
|
|
Common.logger(log_type, crawler).info(f"user_list:{user_list}")
|
|
Common.logger(log_type, crawler).info(f"user_list:{user_list}")
|
|
@@ -21,11 +21,10 @@ def get_author_videos(log_type, crawler, token_index, task_dict, rule_dict, user
|
|
GongzhonghaoAuthor.get_all_videos(log_type=log_type,
|
|
GongzhonghaoAuthor.get_all_videos(log_type=log_type,
|
|
crawler=crawler,
|
|
crawler=crawler,
|
|
task_dict=task_dict,
|
|
task_dict=task_dict,
|
|
- token_index=token_index,
|
|
|
|
rule_dict=rule_dict,
|
|
rule_dict=rule_dict,
|
|
user_list=user_list,
|
|
user_list=user_list,
|
|
env=env)
|
|
env=env)
|
|
- Common.del_logs(log_type, crawler)
|
|
|
|
|
|
+ # Common.del_logs(log_type, crawler)
|
|
Common.logger(log_type, crawler).info('抓取一轮结束\n')
|
|
Common.logger(log_type, crawler).info('抓取一轮结束\n')
|
|
Common.logging(log_type, crawler, env, '抓取一轮结束\n')
|
|
Common.logging(log_type, crawler, env, '抓取一轮结束\n')
|
|
|
|
|
|
@@ -103,7 +102,7 @@ def main(log_type, crawler, topic_name, group_id, env):
|
|
start = i * chunk_size
|
|
start = i * chunk_size
|
|
end = min((i + 1) * chunk_size, user_num + 1)
|
|
end = min((i + 1) * chunk_size, user_num + 1)
|
|
process = Process(target=get_author_videos, args=(
|
|
process = Process(target=get_author_videos, args=(
|
|
- f"{log_type}{i + 1}", crawler, i + 1, task_dict, rule_dict, user_list[start:end], env))
|
|
|
|
|
|
+ f"{log_type}{i + 1}", crawler, task_dict, rule_dict, user_list[start:end], env))
|
|
process.start()
|
|
process.start()
|
|
processes.append(process)
|
|
processes.append(process)
|
|
|
|
|
|
@@ -135,4 +134,4 @@ if __name__ == "__main__":
|
|
crawler=args.crawler,
|
|
crawler=args.crawler,
|
|
topic_name=args.topic_name,
|
|
topic_name=args.topic_name,
|
|
group_id=args.group_id,
|
|
group_id=args.group_id,
|
|
- env=args.env)
|
|
|
|
|
|
+ env=args.env)
|