Quellcode durchsuchen

提交main函数线下测试版本(无阿里云日志)
PEP8格式化代码

罗俊辉 vor 1 Jahr
Ursprung
Commit
089b8e76c6
2 geänderte Dateien mit 35 neuen und 32 gelöschten Zeilen
  1. 19 17
      app/main.py
  2. 16 15
      app/main_dev_version.py

+ 19 - 17
app/main.py

@@ -25,13 +25,12 @@ async def run(task_id, mode, platform):
     """
     # 创建一个aliyun日志对象
     logger = AliyunLogger(platform=platform, mode=mode)
-    logger.logging(
-        code=1003,
-        message="{}: 开始一轮抓取".format(platform)
-    )
+    logger.logging(code=1003, message="{}: 开始一轮抓取".format(platform))
     # 创建并一个子进程
     await asyncio.create_subprocess_shell(
-        "python3 scheduler/run_spider_online.py --task_id {} --mode {} --platform {}".format(task_id, mode, platform)
+        "python3 scheduler/run_spider_online.py --task_id {} --mode {} --platform {}".format(
+            task_id, mode, platform
+        )
     )
     logger.logging(code=5002, message="successfully run spider")
 
@@ -41,10 +40,10 @@ async def consume_single_message(spider):
     消费单个消息,若消费成功则启动爬虫新协程;
     :param spider: 爬虫类
     """
-    topic = spider['topic']
-    group = spider['group']
-    platform = spider['platform']
-    mode = spider['mode']
+    topic = spider["topic"]
+    group = spider["group"]
+    platform = spider["platform"]
+    mode = spider["mode"]
     logger = AliyunLogger(platform=platform, mode=mode)
     consumer = get_consumer(topic, group)
     try:
@@ -53,16 +52,19 @@ async def consume_single_message(spider):
             # 在这里消费消息,做一些数据处理分析
             for single_message in messages:
                 ack_message(
-                    mode=mode,
-                    platform=platform,
-                    recv_msgs=messages,
-                    consumer=consumer
+                    mode=mode, platform=platform, recv_msgs=messages, consumer=consumer
+                )
+                logger.logging(
+                    code=5000,
+                    message="successfully consumed message",
+                    data=single_message.message_body,
                 )
-                logger.logging(code=5000, message="successfully consumed message", data=single_message.message_body)
                 message_body = single_message.message_body
-                task_id = json.loads(message_body)['id']
+                task_id = json.loads(message_body)["id"]
                 # 创建爬虫task
-                await asyncio.create_task(run(task_id, spider['mode'], spider['platform']))
+                await asyncio.create_task(
+                    run(task_id, spider["mode"], spider["platform"])
+                )
                 logger.logging(code=5001, message="successfully created task")
         else:
             logger.logging(code=5003, message="Messages Queue is Empty")
@@ -91,6 +93,6 @@ async def main():
         await asyncio.sleep(60)  # 每分钟接收一次MQ,
 
 
-if __name__ == '__main__':
+if __name__ == "__main__":
     # 运行主事件循环
     asyncio.run(main())

+ 16 - 15
app/main_dev_version.py

@@ -29,7 +29,9 @@ async def run(task_id, mode, platform):
     print(message)
     # 创建并一个子进程
     await asyncio.create_subprocess_shell(
-        "python3 scheduler/run_spider_online.py --task_id {} --mode {} --platform {}".format(task_id, mode, platform)
+        "python3 scheduler/run_spider_online.py --task_id {} --mode {} --platform {}".format(
+            task_id, mode, platform
+        )
     )
     print("successfully run spider")
 
@@ -39,10 +41,10 @@ async def consume_single_message(spider):
     消费单个消息,若消费成功则启动爬虫新协程;
     :param spider: 爬虫类
     """
-    topic = spider['topic']
-    group = spider['group']
-    platform = spider['platform']
-    mode = spider['mode']
+    topic = spider["topic"]
+    group = spider["group"]
+    platform = spider["platform"]
+    mode = spider["mode"]
     consumer = get_consumer(topic, group)
     try:
         messages = consumer.consume_message(wait_seconds=10, batch_size=1)
@@ -50,21 +52,20 @@ async def consume_single_message(spider):
             # 在这里消费消息,做一些数据处理分析
             for single_message in messages:
                 ack_message(
-                    mode=mode,
-                    platform=platform,
-                    recv_msgs=messages,
-                    consumer=consumer
+                    mode=mode, platform=platform, recv_msgs=messages, consumer=consumer
                 )
-                message="successfully consumed message"
+                message = "successfully consumed message"
                 print(message)
                 message_body = single_message.message_body
-                task_id = json.loads(message_body)['id']
+                task_id = json.loads(message_body)["id"]
                 # 创建爬虫task
-                await asyncio.create_task(run(task_id, spider['mode'], spider['platform']))
-                message="successfully created task"
+                await asyncio.create_task(
+                    run(task_id, spider["mode"], spider["platform"])
+                )
+                message = "successfully created task"
                 print(message)
         else:
-            message="Messages Queue is Empty"
+            message = "Messages Queue is Empty"
             print(message)
 
     except MQExceptionBase as err:
@@ -91,6 +92,6 @@ async def main():
         # await asyncio.sleep(60)  # 每分钟接收一次MQ,
 
 
-if __name__ == '__main__':
+if __name__ == "__main__":
     # 运行主事件循环
     asyncio.run(main())