|  | @@ -21,6 +21,46 @@ mysql_con = MysqlHelper()
 | 
	
		
			
				|  |  |  conf = get_config()
 | 
	
		
			
				|  |  |  
 | 
	
		
			
				|  |  |  
 | 
	
		
			
				|  |  | +@app.route("/v1/crawler/task/addlink", methods=["POST"])
 | 
	
		
			
				|  |  | +def addSpiderLink():
 | 
	
		
			
				|  |  | +    try:
 | 
	
		
			
				|  |  | +        data = request.json
 | 
	
		
			
				|  |  | +        spider_link = data['spider_link']
 | 
	
		
			
				|  |  | +        task_id = data['task_id']
 | 
	
		
			
				|  |  | +
 | 
	
		
			
				|  |  | +        sql = f'select * from crawler_author_map where spider_link={spider_link}'
 | 
	
		
			
				|  |  | +        result = mysql_con.get_values(sql)
 | 
	
		
			
				|  |  | +        if result:
 | 
	
		
			
				|  |  | +            old_task_id = result[0]['task_id']
 | 
	
		
			
				|  |  | +            if task_id == old_task_id:
 | 
	
		
			
				|  |  | +                up_sql = f'update crawler_author_map set is_del=1 where spider_link="{spider_link}"'
 | 
	
		
			
				|  |  | +            else:
 | 
	
		
			
				|  |  | +                up_sql = f'update crawler_author_map set task_id={task_id} where spider_link="{spider_link}"'
 | 
	
		
			
				|  |  | +            res = mysql_con.update_values(up_sql)
 | 
	
		
			
				|  |  | +
 | 
	
		
			
				|  |  | +            task_sql = f'select spider_link from crawler_task where task_id ={task_id}'
 | 
	
		
			
				|  |  | +            task = mysql_con.get_values(task_sql)
 | 
	
		
			
				|  |  | +            spider_links = eval(task[0]['spider_link'])
 | 
	
		
			
				|  |  | +            spider_links.append(spider_link)
 | 
	
		
			
				|  |  | +            u_sql = f'update crawler_task set spider_link="{spider_links}" where task_id={task_id}'
 | 
	
		
			
				|  |  | +            mysql_con.update_values(u_sql)
 | 
	
		
			
				|  |  | +            return jsonify({'code': 200, 'message': '抓取名单增加成功', 'del_link': spider_link})
 | 
	
		
			
				|  |  | +        else:
 | 
	
		
			
				|  |  | +            sql = f'select * from crawler_task where task_id={task_id}'
 | 
	
		
			
				|  |  | +            result = mysql_con.get_values(sql)
 | 
	
		
			
				|  |  | +            success_list, fail_list = create_uid(result[0], task_id)
 | 
	
		
			
				|  |  | +
 | 
	
		
			
				|  |  | +            spider_links = eval(result[0]['spider_link'])
 | 
	
		
			
				|  |  | +            spider_links.append(spider_link)
 | 
	
		
			
				|  |  | +            u_sql = f'update crawler_task set spider_link="{spider_links}" where task_id={task_id}'
 | 
	
		
			
				|  |  | +            mysql_con.update_values(u_sql)
 | 
	
		
			
				|  |  | +
 | 
	
		
			
				|  |  | +            return jsonify({'code': 200, 'message': '抓取名单增加成功', 'add_link': success_list})
 | 
	
		
			
				|  |  | +    except Exception as e:
 | 
	
		
			
				|  |  | +        return jsonify(
 | 
	
		
			
				|  |  | +            {'code': 400, 'message': '抓取名单删除失败', 'spider_link': spider_link})
 | 
	
		
			
				|  |  | +
 | 
	
		
			
				|  |  | +
 | 
	
		
			
				|  |  |  @app.route("/v1/crawler/task/dellink", methods=["POST"])
 | 
	
		
			
				|  |  |  def delSpiderLink():
 | 
	
		
			
				|  |  |      data = request.json
 | 
	
	
		
			
				|  | @@ -29,18 +69,18 @@ def delSpiderLink():
 | 
	
		
			
				|  |  |      up_sql = f'update crawler_author_map set is_del=0 where spider_link="{spider_link}"'
 | 
	
		
			
				|  |  |      MysqlHelper.update_values(up_sql)
 | 
	
		
			
				|  |  |  
 | 
	
		
			
				|  |  | -    sql = f'select spider_link from crawler_task where task_id ={task_id}'
 | 
	
		
			
				|  |  | +    sql = f'select * from crawler_task where task_id ={task_id}'
 | 
	
		
			
				|  |  |      task = mysql_con.get_values(sql)
 | 
	
		
			
				|  |  |      spider_links = eval(task[0]['spider_link'])
 | 
	
		
			
				|  |  | -    for link in spider_links:
 | 
	
		
			
				|  |  | -        spider_links.remove(link)
 | 
	
		
			
				|  |  | +    spider_links.remove(spider_link)
 | 
	
		
			
				|  |  | +
 | 
	
		
			
				|  |  |      u_sql = f'update crawler_task set spider_link="{spider_links}" where task_id={task_id}'
 | 
	
		
			
				|  |  |      mysql_con.update_values(u_sql)
 | 
	
		
			
				|  |  |      if spider_link:
 | 
	
		
			
				|  |  |          return jsonify({'code': 200, 'message': '抓取名单删除成功', 'del_link': spider_link})
 | 
	
		
			
				|  |  |      else:
 | 
	
		
			
				|  |  |          return jsonify(
 | 
	
		
			
				|  |  | -            {'code': 400, 'message': '抓取名单删除失败', 'del_link': spider_link, 'spider_link': spider_links})
 | 
	
		
			
				|  |  | +            {'code': 400, 'message': '抓取名单删除失败', 'del_link': spider_link})
 | 
	
		
			
				|  |  |  
 | 
	
		
			
				|  |  |  
 | 
	
		
			
				|  |  |  @app.route("/v1/crawler/task/getcategory", methods=["GET"])
 | 
	
	
		
			
				|  | @@ -125,6 +165,7 @@ def getSource():
 | 
	
		
			
				|  |  |  
 | 
	
		
			
				|  |  |      return jsonify({'code': '200', 'result': result})
 | 
	
		
			
				|  |  |  
 | 
	
		
			
				|  |  | +
 | 
	
		
			
				|  |  |  @app.route("/v1/crawler/source/getasktype", methods=["GET"])
 | 
	
		
			
				|  |  |  def getTaskType():
 | 
	
		
			
				|  |  |      try:
 | 
	
	
		
			
				|  | @@ -157,7 +198,6 @@ def getTaskType():
 | 
	
		
			
				|  |  |      return jsonify({'code': '200', 'result': source_dict})
 | 
	
		
			
				|  |  |  
 | 
	
		
			
				|  |  |  
 | 
	
		
			
				|  |  | -
 | 
	
		
			
				|  |  |  @app.route("/v1/crawler/task/checkrepeat", methods=["POST"])
 | 
	
		
			
				|  |  |  def get_repeat_list():
 | 
	
		
			
				|  |  |      data = request.json
 |