Sfoglia il codice sorgente

update 二期开发

lierqiang 2 anni fa
parent
commit
0b713a9cc2
1 ha cambiato i file con 20 aggiunte e 20 eliminazioni
  1. 20 20
      server/conf_task.py

+ 20 - 20
server/conf_task.py

@@ -14,28 +14,28 @@ from conf.config import get_config
 app = Flask(__name__)
 app.config['JSON_AS_ASCII'] = False
 # mysql实例
-MysqlHelper = MysqlHelper()
+mysql_con = MysqlHelper()
 conf = get_config()
 
 
 @app.route("/v1/crawler/task/getcategory", methods=["GET"])
 def getCategory():
     sql = f'select id, content_category from crawler_content_category'
-    result = MysqlHelper.get_values(sql)
+    result = mysql_con.get_values(sql)
     return jsonify({'code': 200, 'data': result})
 
 
 @app.route("/v1/crawler/task/getmodename", methods=["GET"])
 def getModeName():
     sql = f'select id, mode_name from crawler_mode'
-    result = MysqlHelper.get_values(sql)
+    result = mysql_con.get_values(sql)
     return jsonify({'code': 200, 'data': result})
 
 
 @app.route("/v1/crawler/task/getmodeboard", methods=["GET"])
 def getModeBoard():
     sql = f'select id, mode_board from crawler_board'
-    result = MysqlHelper.get_values(sql)
+    result = mysql_con.get_values(sql)
     return jsonify({'code': 200, 'data': result})
 
 
@@ -44,7 +44,7 @@ def getMediaInfo():
     data = request.args.to_dict()
     task_id = data['task_id']
     sql = f'select * from crawler_author_map where task_id={task_id}'
-    result = MysqlHelper.get_values(sql)
+    result = mysql_con.get_values(sql)
     task_user_info = []
     for task_info in result:
         media_id = task_info['media_id']
@@ -77,11 +77,11 @@ def getTaskUserInfo():
             values += f'{k}="{v}" and '
 
     sql = f"select task_id from crawler_author_map where {values[:-4]}"  # [:-1]是为了去掉末尾的逗号
-    res = MysqlHelper.get_values(sql)
+    res = mysql_con.get_values(sql)
     task_id = res['task_id']
 
     sql = f'select task_name, source, task_type, create_task_user, insert_time, update_task_user, update_time from crawler_task where task_id={task_id} '
-    task_info = MysqlHelper.get_values(sql)
+    task_info = mysql_con.get_values(sql)
 
     return jsonify({'code': 200, 'data': task_info})
 
@@ -95,7 +95,7 @@ def getSource():
         # # 对参数进行操作
         sql = 'select source, task_type, spider_name, machine, source_desc, task_type_desc, spider_name_desc from crawler_source'
 
-        result = MysqlHelper.get_values(sql)
+        result = mysql_con.get_values(sql)
         if not result:
             return jsonify({'code': '200', 'result': [], 'message': '没有更多数据'})
         source_list = list()
@@ -135,7 +135,7 @@ def get_repeat_list():
             s_sql = f"""select spider_link from crawler_author_map where spider_link={spider_link}"""
         else:
             s_sql = f"""select spider_link from crawler_author_map where spider_link='{spider_link}'"""
-        result = MysqlHelper.get_values(s_sql)
+        result = mysql_con.get_values(s_sql)
         if result:
             repeat_list.append(spider_link)
     if repeat_list:
@@ -182,7 +182,7 @@ def insertTask():
         keys = ','.join(data.keys())
         values = ','.join(['%s'] * len(data))
         sql = 'insert into {table}({keys}) VALUES({values})'.format(table='crawler_task', keys=keys, values=values)
-        task_id = MysqlHelper.insert_values(sql, tuple(data.values()))
+        task_id = mysql_con.insert_values(sql, tuple(data.values()))
         if task_id:
             success_list, fail_list = create_uid(user_data, task_id)
             return jsonify(
@@ -200,7 +200,7 @@ def getAllTask():
         start_count = (page * offset) - offset
         end_count = page * offset
         sql = f"""select task_name, source, task_type, create_task_user, insert_time, update_task_user, update_time  from crawler_task order by update_time desc limit {start_count}, {end_count} """
-        result = MysqlHelper.get_values(sql)
+        result = mysql_con.get_values(sql)
         if not result:
             return jsonify({'code': '200', 'result': [], 'message': '没有更多任务'})
         task_list = list()
@@ -217,7 +217,7 @@ def getAllTask():
             task_list.append(task_data)
 
         t_sql = f"""select count(*) from crawler_task"""
-        t_res = MysqlHelper.get_values(t_sql)
+        t_res = mysql_con.get_values(t_sql)
         total = t_res[0]['count(*)']
     except Exception as e:
         return jsonify({"code": "400", 'message': "任务列表获取失败"})
@@ -231,7 +231,7 @@ def getOneTask():
         get_data = request.args.to_dict()
         task_id = get_data['task_id']
         sql = f'select * from crawler_task where task_id={task_id}'
-        result = MysqlHelper.get_values(sql)
+        result = mysql_con.get_values(sql)
 
         if not result:
             return jsonify({'code': '400', 'result': [], 'message': 'no data'})
@@ -256,7 +256,7 @@ def updateTask():
                 values += f'{k}="{v}",'
 
         sql = f'update crawler_task set {values[:-1]} where task_id={task_id}'
-        result = MysqlHelper.update_values(sql)
+        result = mysql_con.update_values(sql)
         if result:
             return jsonify({'code': 200, 'message': 'task update success'})
         else:
@@ -280,11 +280,11 @@ def create_uid(task, task_id):
     content_category_id = task['content_category_id']
 
     mn_sql = f'select * from crawler_mode_name where id={mode_name_id}'
-    mode_name_list = MysqlHelper.get_values(mn_sql)
+    mode_name_list = mysql_con.get_values(mn_sql)
     mb_sql = f'select * from crawler_mode_board where id={mode_board_id}'
-    mode_board_list = MysqlHelper.get_values(mb_sql)
+    mode_board_list = mysql_con.get_values(mb_sql)
     cc_sql = f'select * from crawler_content_category where id={content_category_id}'
-    content_category_list = MysqlHelper.get_values(cc_sql)
+    content_category_list = mysql_con.get_values(cc_sql)
     task['mode_name_str'] = mode_name_list[0]['mode_name']
     task['mode_board_str'] = mode_board_list[0]['mode_board']
     task['content_category_str'] = content_category_list[0]['content_category']
@@ -346,7 +346,7 @@ def create_uid(task, task_id):
         values = ','.join(['%s'] * len(data))
         table = 'crawler_author_map'
         sql = f"""insert into {table}({keys}) VALUES({values})"""
-        MysqlHelper.insert_values(sql, tuple(data.values()))
+        mysql_con.insert_values(sql, tuple(data.values()))
         uer_info = dict(
             outer_id=author_url,
             uid=media_id
@@ -370,7 +370,7 @@ def createUser():
     for author_url in spider_link:
         try:
             f_sql = f"""select spider_link from crawler_author_map where spider_link="{author_url}" """
-            result = MysqlHelper.get_values(f_sql)
+            result = mysql_con.get_values(f_sql)
             if result:
                 success_list.append(author_url)
                 continue
@@ -411,7 +411,7 @@ def createUser():
                 values = ','.join(['%s'] * len(data))
                 table = 'crawler_author_map'
                 sql = f"""insert into {table}({keys}) VALUES({values})"""
-                result = MysqlHelper.insert_values(sql, tuple(data.values()))
+                result = mysql_con.insert_values(sql, tuple(data.values()))
                 if not result:
                     fail_list.append(author_url)
                 else: