|
@@ -14,28 +14,28 @@ from conf.config import get_config
|
|
|
app = Flask(__name__)
|
|
|
app.config['JSON_AS_ASCII'] = False
|
|
|
# mysql实例
|
|
|
-MysqlHelper = MysqlHelper()
|
|
|
+mysql_con = MysqlHelper()
|
|
|
conf = get_config()
|
|
|
|
|
|
|
|
|
@app.route("/v1/crawler/task/getcategory", methods=["GET"])
|
|
|
def getCategory():
|
|
|
sql = f'select id, content_category from crawler_content_category'
|
|
|
- result = MysqlHelper.get_values(sql)
|
|
|
+ result = mysql_con.get_values(sql)
|
|
|
return jsonify({'code': 200, 'data': result})
|
|
|
|
|
|
|
|
|
@app.route("/v1/crawler/task/getmodename", methods=["GET"])
|
|
|
def getModeName():
|
|
|
sql = f'select id, mode_name from crawler_mode'
|
|
|
- result = MysqlHelper.get_values(sql)
|
|
|
+ result = mysql_con.get_values(sql)
|
|
|
return jsonify({'code': 200, 'data': result})
|
|
|
|
|
|
|
|
|
@app.route("/v1/crawler/task/getmodeboard", methods=["GET"])
|
|
|
def getModeBoard():
|
|
|
sql = f'select id, mode_board from crawler_board'
|
|
|
- result = MysqlHelper.get_values(sql)
|
|
|
+ result = mysql_con.get_values(sql)
|
|
|
return jsonify({'code': 200, 'data': result})
|
|
|
|
|
|
|
|
@@ -44,7 +44,7 @@ def getMediaInfo():
|
|
|
data = request.args.to_dict()
|
|
|
task_id = data['task_id']
|
|
|
sql = f'select * from crawler_author_map where task_id={task_id}'
|
|
|
- result = MysqlHelper.get_values(sql)
|
|
|
+ result = mysql_con.get_values(sql)
|
|
|
task_user_info = []
|
|
|
for task_info in result:
|
|
|
media_id = task_info['media_id']
|
|
@@ -77,11 +77,11 @@ def getTaskUserInfo():
|
|
|
values += f'{k}="{v}" and '
|
|
|
|
|
|
sql = f"select task_id from crawler_author_map where {values[:-4]}" # [:-1]是为了去掉末尾的逗号
|
|
|
- res = MysqlHelper.get_values(sql)
|
|
|
+ res = mysql_con.get_values(sql)
|
|
|
task_id = res['task_id']
|
|
|
|
|
|
sql = f'select task_name, source, task_type, create_task_user, insert_time, update_task_user, update_time from crawler_task where task_id={task_id} '
|
|
|
- task_info = MysqlHelper.get_values(sql)
|
|
|
+ task_info = mysql_con.get_values(sql)
|
|
|
|
|
|
return jsonify({'code': 200, 'data': task_info})
|
|
|
|
|
@@ -95,7 +95,7 @@ def getSource():
|
|
|
# # 对参数进行操作
|
|
|
sql = 'select source, task_type, spider_name, machine, source_desc, task_type_desc, spider_name_desc from crawler_source'
|
|
|
|
|
|
- result = MysqlHelper.get_values(sql)
|
|
|
+ result = mysql_con.get_values(sql)
|
|
|
if not result:
|
|
|
return jsonify({'code': '200', 'result': [], 'message': '没有更多数据'})
|
|
|
source_list = list()
|
|
@@ -135,7 +135,7 @@ def get_repeat_list():
|
|
|
s_sql = f"""select spider_link from crawler_author_map where spider_link={spider_link}"""
|
|
|
else:
|
|
|
s_sql = f"""select spider_link from crawler_author_map where spider_link='{spider_link}'"""
|
|
|
- result = MysqlHelper.get_values(s_sql)
|
|
|
+ result = mysql_con.get_values(s_sql)
|
|
|
if result:
|
|
|
repeat_list.append(spider_link)
|
|
|
if repeat_list:
|
|
@@ -182,7 +182,7 @@ def insertTask():
|
|
|
keys = ','.join(data.keys())
|
|
|
values = ','.join(['%s'] * len(data))
|
|
|
sql = 'insert into {table}({keys}) VALUES({values})'.format(table='crawler_task', keys=keys, values=values)
|
|
|
- task_id = MysqlHelper.insert_values(sql, tuple(data.values()))
|
|
|
+ task_id = mysql_con.insert_values(sql, tuple(data.values()))
|
|
|
if task_id:
|
|
|
success_list, fail_list = create_uid(user_data, task_id)
|
|
|
return jsonify(
|
|
@@ -200,7 +200,7 @@ def getAllTask():
|
|
|
start_count = (page * offset) - offset
|
|
|
end_count = page * offset
|
|
|
sql = f"""select task_name, source, task_type, create_task_user, insert_time, update_task_user, update_time from crawler_task order by update_time desc limit {start_count}, {end_count} """
|
|
|
- result = MysqlHelper.get_values(sql)
|
|
|
+ result = mysql_con.get_values(sql)
|
|
|
if not result:
|
|
|
return jsonify({'code': '200', 'result': [], 'message': '没有更多任务'})
|
|
|
task_list = list()
|
|
@@ -217,7 +217,7 @@ def getAllTask():
|
|
|
task_list.append(task_data)
|
|
|
|
|
|
t_sql = f"""select count(*) from crawler_task"""
|
|
|
- t_res = MysqlHelper.get_values(t_sql)
|
|
|
+ t_res = mysql_con.get_values(t_sql)
|
|
|
total = t_res[0]['count(*)']
|
|
|
except Exception as e:
|
|
|
return jsonify({"code": "400", 'message': "任务列表获取失败"})
|
|
@@ -231,7 +231,7 @@ def getOneTask():
|
|
|
get_data = request.args.to_dict()
|
|
|
task_id = get_data['task_id']
|
|
|
sql = f'select * from crawler_task where task_id={task_id}'
|
|
|
- result = MysqlHelper.get_values(sql)
|
|
|
+ result = mysql_con.get_values(sql)
|
|
|
|
|
|
if not result:
|
|
|
return jsonify({'code': '400', 'result': [], 'message': 'no data'})
|
|
@@ -256,7 +256,7 @@ def updateTask():
|
|
|
values += f'{k}="{v}",'
|
|
|
|
|
|
sql = f'update crawler_task set {values[:-1]} where task_id={task_id}'
|
|
|
- result = MysqlHelper.update_values(sql)
|
|
|
+ result = mysql_con.update_values(sql)
|
|
|
if result:
|
|
|
return jsonify({'code': 200, 'message': 'task update success'})
|
|
|
else:
|
|
@@ -280,11 +280,11 @@ def create_uid(task, task_id):
|
|
|
content_category_id = task['content_category_id']
|
|
|
|
|
|
mn_sql = f'select * from crawler_mode_name where id={mode_name_id}'
|
|
|
- mode_name_list = MysqlHelper.get_values(mn_sql)
|
|
|
+ mode_name_list = mysql_con.get_values(mn_sql)
|
|
|
mb_sql = f'select * from crawler_mode_board where id={mode_board_id}'
|
|
|
- mode_board_list = MysqlHelper.get_values(mb_sql)
|
|
|
+ mode_board_list = mysql_con.get_values(mb_sql)
|
|
|
cc_sql = f'select * from crawler_content_category where id={content_category_id}'
|
|
|
- content_category_list = MysqlHelper.get_values(cc_sql)
|
|
|
+ content_category_list = mysql_con.get_values(cc_sql)
|
|
|
task['mode_name_str'] = mode_name_list[0]['mode_name']
|
|
|
task['mode_board_str'] = mode_board_list[0]['mode_board']
|
|
|
task['content_category_str'] = content_category_list[0]['content_category']
|
|
@@ -346,7 +346,7 @@ def create_uid(task, task_id):
|
|
|
values = ','.join(['%s'] * len(data))
|
|
|
table = 'crawler_author_map'
|
|
|
sql = f"""insert into {table}({keys}) VALUES({values})"""
|
|
|
- MysqlHelper.insert_values(sql, tuple(data.values()))
|
|
|
+ mysql_con.insert_values(sql, tuple(data.values()))
|
|
|
uer_info = dict(
|
|
|
outer_id=author_url,
|
|
|
uid=media_id
|
|
@@ -370,7 +370,7 @@ def createUser():
|
|
|
for author_url in spider_link:
|
|
|
try:
|
|
|
f_sql = f"""select spider_link from crawler_author_map where spider_link="{author_url}" """
|
|
|
- result = MysqlHelper.get_values(f_sql)
|
|
|
+ result = mysql_con.get_values(f_sql)
|
|
|
if result:
|
|
|
success_list.append(author_url)
|
|
|
continue
|
|
@@ -411,7 +411,7 @@ def createUser():
|
|
|
values = ','.join(['%s'] * len(data))
|
|
|
table = 'crawler_author_map'
|
|
|
sql = f"""insert into {table}({keys}) VALUES({values})"""
|
|
|
- result = MysqlHelper.insert_values(sql, tuple(data.values()))
|
|
|
+ result = mysql_con.insert_values(sql, tuple(data.values()))
|
|
|
if not result:
|
|
|
fail_list.append(author_url)
|
|
|
else:
|