|
@@ -60,25 +60,28 @@ class Scheduling:
|
|
|
|
|
|
@classmethod
|
|
@classmethod
|
|
def get_redis(cls, log_type, crawler, env):
|
|
def get_redis(cls, log_type, crawler, env):
|
|
- while True:
|
|
|
|
- if env == 'hk':
|
|
|
|
- task_key = 'crawler_config_task_queue:hk'
|
|
|
|
- elif env == 'prod':
|
|
|
|
- task_key = 'crawler_config_task_queue:aliyun'
|
|
|
|
- else:
|
|
|
|
- task_key = 'crawler_config_task_queue:dev'
|
|
|
|
|
|
+ if env == 'hk':
|
|
|
|
+ task_key = 'crawler_config_task_queue:hk'
|
|
|
|
+ elif env == 'prod':
|
|
|
|
+ task_key = 'crawler_config_task_queue:aliyun'
|
|
|
|
+ else:
|
|
|
|
+ task_key = 'crawler_config_task_queue:dev'
|
|
|
|
|
|
- redis_data = RedisHelper.redis_pop(env, task_key)
|
|
|
|
- if redis_data is None or len(redis_data) == 0:
|
|
|
|
- Common.logger(log_type, crawler).info("Redis为空,等待1秒")
|
|
|
|
- time.sleep(1)
|
|
|
|
- else:
|
|
|
|
- task = eval(str(redis_data, encoding="utf8"))
|
|
|
|
- return task
|
|
|
|
|
|
+ redis_data = RedisHelper.redis_pop(env, task_key)
|
|
|
|
+ if redis_data is None or len(redis_data) == 0:
|
|
|
|
+ # Common.logger(log_type, crawler).info("Redis为空,程序退出")
|
|
|
|
+ # time.sleep(1)
|
|
|
|
+ return
|
|
|
|
+ else:
|
|
|
|
+ task = eval(str(redis_data, encoding="utf8"))
|
|
|
|
+ return task
|
|
|
|
|
|
@classmethod
|
|
@classmethod
|
|
def scheduling_task(cls, log_type, crawler, env):
|
|
def scheduling_task(cls, log_type, crawler, env):
|
|
task = cls.get_redis(log_type, crawler, env)
|
|
task = cls.get_redis(log_type, crawler, env)
|
|
|
|
+ if not task:
|
|
|
|
+ Common.logger(log_type, crawler).info("Redis为空,程序退出")
|
|
|
|
+ return
|
|
Common.logger(log_type, crawler).info(f"task: {task}")
|
|
Common.logger(log_type, crawler).info(f"task: {task}")
|
|
Common.logger(log_type, crawler).info(f"已获取调度任务:{task}")
|
|
Common.logger(log_type, crawler).info(f"已获取调度任务:{task}")
|
|
task_id = task['task_id']
|
|
task_id = task['task_id']
|