123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481 |
- #! /usr/bin/env python
- # -*- coding: utf-8 -*-
- # vim:fenc=utf-8
- import re
- import signal
- import sys
- import time
- import random
- from typing import Dict, List, Tuple, Any, Optional
- import logging
- from datetime import datetime, timedelta
- import traceback
- import apscheduler.triggers.cron
- from apscheduler.schedulers.background import BackgroundScheduler
- import chat_service
- import configs
- import logging_service
- from configs import apollo_config
- from logging_service import logger
- from chat_service import CozeChat, ChatServiceType
- from dialogue_manager import DialogueManager, DialogueState, DialogueStateCache
- from response_type_detector import ResponseTypeDetector
- from user_manager import UserManager, LocalUserManager, MySQLUserManager, MySQLUserRelationManager, UserRelationManager, \
- LocalUserRelationManager
- from openai import OpenAI
- from message_queue_backend import MessageQueueBackend, MemoryQueueBackend, AliyunRocketMQQueueBackend
- from user_profile_extractor import UserProfileExtractor
- import threading
- from message import MessageType, Message, MessageChannel
- class AgentService:
- def __init__(
- self,
- receive_backend: MessageQueueBackend,
- send_backend: MessageQueueBackend,
- human_backend: MessageQueueBackend,
- user_manager: UserManager,
- user_relation_manager: UserRelationManager,
- chat_service_type: ChatServiceType = ChatServiceType.OPENAI_COMPATIBLE
- ):
- self.receive_queue = receive_backend
- self.send_queue = send_backend
- self.human_queue = human_backend
- # 核心服务模块
- self.agent_state_cache = DialogueStateCache()
- self.user_manager = user_manager
- self.user_relation_manager = user_relation_manager
- self.user_profile_extractor = UserProfileExtractor()
- self.response_type_detector = ResponseTypeDetector()
- self.agent_registry: Dict[str, DialogueManager] = {}
- self.config = configs.get()
- chat_config = self.config['chat_api']['openai_compatible']
- self.text_model_name = chat_config['text_model']
- self.multimodal_model_name = chat_config['multimodal_model']
- self.text_model_client = chat_service.OpenAICompatible.create_client(self.text_model_name)
- self.multimodal_model_client = chat_service.OpenAICompatible.create_client(self.multimodal_model_name)
- coze_config = configs.get()['chat_api']['coze']
- coze_oauth_app = CozeChat.get_oauth_app(
- coze_config['oauth_client_id'], coze_config['private_key_path'], str(coze_config['public_key_id']),
- account_id=coze_config.get('account_id', None)
- )
- self.coze_client = CozeChat(
- base_url=chat_service.COZE_CN_BASE_URL,
- auth_app=coze_oauth_app
- )
- self.chat_service_type = chat_service_type
- # 定时任务调度器
- self.scheduler = None
- self.scheduler_mode = self.config.get('system', {}).get('scheduler_mode', 'local')
- self.scheduler_queue = None
- self.msg_scheduler_thread = None
- self.limit_initiative_conversation_rate = True
- self.running = False
- self.process_thread = None
- self._sigint_cnt = 0
- def setup_initiative_conversations(self, schedule_params: Optional[Dict] = None):
- if not schedule_params:
- schedule_params = {'hour': '8,16,20'}
- self.scheduler.add_job(
- self._check_initiative_conversations,
- apscheduler.triggers.cron.CronTrigger(**schedule_params)
- )
- def setup_scheduler(self):
- self.scheduler = BackgroundScheduler()
- if self.scheduler_mode == 'mq':
- logging.info("setup event message scheduler with MQ")
- mq_conf = self.config['mq']
- topic = mq_conf['scheduler_topic']
- self.scheduler_queue = AliyunRocketMQQueueBackend(
- mq_conf['endpoints'],
- mq_conf['instance_id'],
- topic,
- has_consumer=True, has_producer=True,
- group_id=mq_conf['scheduler_group'],
- topic_type='DELAY'
- )
- self.msg_scheduler_thread = threading.Thread(target=self.process_scheduler_events)
- self.msg_scheduler_thread.start()
- self.scheduler.start()
- def process_scheduler_events(self):
- while self.running:
- msg = self.scheduler_queue.consume()
- if msg:
- try:
- self.process_scheduler_event(msg)
- self.scheduler_queue.ack(msg)
- except Exception as e:
- logger.error("Error processing scheduler event: {}".format(e))
- time.sleep(1)
- logger.info("Scheduler event processing thread exit")
- def process_scheduler_event(self, msg: Message):
- if msg.type == MessageType.AGGREGATION_TRIGGER:
- # 延迟触发的消息,需放入接收队列以驱动Agent运转
- self.receive_queue.produce(msg)
- else:
- logger.warning(f"Unknown message type: {msg.type}")
- def _get_agent_instance(self, staff_id: str, user_id: str) -> DialogueManager:
- """获取Agent实例"""
- agent_key = 'agent_{}_{}'.format(staff_id, user_id)
- if agent_key not in self.agent_registry:
- self.agent_registry[agent_key] = DialogueManager(
- staff_id, user_id, self.user_manager, self.agent_state_cache)
- return self.agent_registry[agent_key]
- def process_messages(self):
- """持续处理接收队列消息"""
- while self.running:
- message = self.receive_queue.consume()
- if message:
- try:
- self.process_single_message(message)
- self.receive_queue.ack(message)
- except Exception as e:
- logger.error("Error processing message: {}".format(e))
- traceback.print_exc()
- time.sleep(1)
- logger.info("Message processing thread exit")
- def start(self, blocking=False):
- self.running = True
- self.process_thread = threading.Thread(target=service.process_messages)
- self.process_thread.start()
- self.setup_scheduler()
- # 只有企微场景需要主动发起
- if not self.config['debug_flags'].get('disable_active_conversation', False):
- schedule_param = self.config['agent_behavior'].get('active_conversation_schedule_param', None)
- self.setup_initiative_conversations(schedule_param)
- signal.signal(signal.SIGINT, self._handle_sigint)
- if blocking:
- self.process_thread.join()
- def shutdown(self, sync=True):
- if not self.running:
- raise Exception("Service is not running")
- self.running = False
- self.scheduler.shutdown()
- if sync:
- self.process_thread.join()
- self.receive_queue.shutdown()
- self.send_queue.shutdown()
- if self.msg_scheduler_thread:
- self.msg_scheduler_thread.join()
- self.scheduler_queue.shutdown()
- def _handle_sigint(self, signum, frame):
- self._sigint_cnt += 1
- if self._sigint_cnt == 1:
- logger.warning("Try to shutdown gracefully...")
- self.shutdown(sync=True)
- else:
- logger.warning("Forcing exit")
- sys.exit(0)
- def _update_user_profile(self, user_id, user_profile, recent_dialogue: List[Dict]):
- profile_to_update = self.user_profile_extractor.extract_profile_info(user_profile, recent_dialogue)
- if not profile_to_update:
- logger.debug("user_id: {}, no profile info extracted".format(user_id))
- return
- logger.warning("update user profile: {}".format(profile_to_update))
- if profile_to_update.get('interaction_frequency', None) == 'stopped':
- # 和企微日常push联动,减少对用户的干扰
- if self.user_relation_manager.stop_user_daily_push(user_id):
- logger.warning(f"user[{user_id}]: daily push set to be stopped")
- merged_profile = self.user_profile_extractor.merge_profile_info(user_profile, profile_to_update)
- self.user_manager.save_user_profile(user_id, merged_profile)
- return merged_profile
- def _schedule_aggregation_trigger(self, staff_id: str, user_id: str, delay_sec: int):
- logger.debug("user: {}, schedule trigger message after {} seconds".format(user_id, delay_sec))
- message_ts = int((time.time() + delay_sec) * 1000)
- msg = Message.build(MessageType.AGGREGATION_TRIGGER, MessageChannel.SYSTEM, user_id, staff_id, None, message_ts)
- # 系统消息使用特定的msgId,无实际意义
- msg.msgId = -MessageType.AGGREGATION_TRIGGER.value
- if self.scheduler_mode == 'mq':
- self.scheduler_queue.produce(msg)
- else:
- self.scheduler.add_job(lambda: self.receive_queue.produce(msg),
- 'date',
- run_date=datetime.now() + timedelta(seconds=delay_sec))
- def process_single_message(self, message: Message):
- user_id = message.sender
- staff_id = message.receiver
- # 获取用户信息和Agent实例
- user_profile = self.user_manager.get_user_profile(user_id)
- agent = self._get_agent_instance(staff_id, user_id)
- # 更新对话状态
- logger.debug("process message: {}".format(message))
- need_response, message_text = agent.update_state(message)
- logger.debug("user: {}, next state: {}".format(user_id, agent.current_state))
- # 根据状态路由消息
- try:
- if agent.is_in_human_intervention():
- self._route_to_human_intervention(user_id, message)
- elif agent.current_state == DialogueState.MESSAGE_AGGREGATING:
- if message.type != MessageType.AGGREGATION_TRIGGER:
- # 产生一个触发器,但是不能由触发器递归产生
- logger.debug("user: {}, waiting next message for aggregation".format(user_id))
- self._schedule_aggregation_trigger(staff_id, user_id, agent.message_aggregation_sec)
- elif need_response:
- # 先更新用户画像再处理回复
- self._update_user_profile(user_id, user_profile, agent.dialogue_history[-10:])
- resp = self._get_chat_response(user_id, agent, message_text)
- if resp:
- recent_dialogue = agent.dialogue_history[-10:]
- if len(recent_dialogue) < 2 or staff_id not in ('1688855931724582', '1688854492669990'):
- message_type = MessageType.TEXT
- else:
- message_type = self.response_type_detector.detect_type(
- recent_dialogue[:-1], recent_dialogue[-1], enable_random=True)
- self._send_response(staff_id, user_id, resp, message_type)
- else:
- logger.debug(f"staff[{staff_id}], user[{user_id}]: do not need response")
- # 当前消息处理成功,commit并持久化agent状态
- agent.persist_state()
- except Exception as e:
- agent.rollback_state()
- raise e
- def _send_response(self, staff_id, user_id, response, message_type: MessageType, skip_check=False):
- logger.warning(f"staff[{staff_id}] user[{user_id}]: response[{message_type}] {response}")
- current_ts = int(time.time() * 1000)
- user_tags = self.user_relation_manager.get_user_tags(user_id)
- white_list_tags = set(apollo_config.get_json_value("agent_response_whitelist_tags"))
- hit_white_list_tags = len(set(user_tags).intersection(white_list_tags)) > 0
- # FIXME(zhoutian)
- # 测试期间临时逻辑,只发送特定的账号或特定用户
- staff_white_lists = set(apollo_config.get_json_value("agent_response_whitelist_staffs"))
- if not (staff_id in staff_white_lists or hit_white_list_tags or skip_check):
- logger.warning(f"staff[{staff_id}] user[{user_id}]: skip reply")
- return None
- self.send_queue.produce(
- Message.build(message_type, MessageChannel.CORP_WECHAT,
- staff_id, user_id, response, current_ts)
- )
- def _route_to_human_intervention(self, user_id: str, origin_message: Message):
- """路由到人工干预"""
- self.human_queue.produce(Message.build(
- MessageType.TEXT,
- origin_message.channel,
- origin_message.sender,
- origin_message.receiver,
- "用户对话需人工介入,用户名:{}".format(user_id),
- int(time.time() * 1000)
- ))
- def _check_initiative_conversations(self):
- logger.info("start to check initiative conversations")
- if not DialogueManager.is_time_suitable_for_active_conversation():
- logger.info("time is not suitable for active conversation")
- return
- white_list_tags = set(apollo_config.get_json_value('agent_initiate_whitelist_tags'))
- first_initiate_tags = set(apollo_config.get_json_value('agent_first_initiate_whitelist_tags', []))
- # 合并白名单,减少配置成本
- white_list_tags.update(first_initiate_tags)
- voice_tags = set(apollo_config.get_json_value('agent_initiate_by_voice_tags'))
- """定时检查主动发起对话"""
- for staff_user in self.user_relation_manager.list_staff_users():
- staff_id = staff_user['staff_id']
- user_id = staff_user['user_id']
- agent = self._get_agent_instance(staff_id, user_id)
- should_initiate = agent.should_initiate_conversation()
- user_tags = self.user_relation_manager.get_user_tags(user_id)
- if configs.get_env() != 'dev' and not white_list_tags.intersection(user_tags):
- should_initiate = False
- if should_initiate:
- logger.warning(f"user[{user_id}], tags{user_tags}: initiate conversation")
- # FIXME:虽然需要主动唤起的用户同时发来消息的概率很低,但仍可能会有并发冲突 需要并入事件驱动框架
- agent.do_state_change(DialogueState.GREETING)
- try:
- if agent.previous_state == DialogueState.INITIALIZED or first_initiate_tags.intersection(user_tags):
- # 完全无交互历史的用户才使用此策略,但新用户接入即会产生“我已添加了你”的消息将Agent初始化
- # 因此存量用户无法使用该状态做实验
- # TODO:增加基于对话历史的判断、策略去重;如果对话间隔过长需要使用长期记忆检索;在无长期记忆时,可采用用户添加时间来判断
- resp = self._generate_active_greeting_message(agent, user_tags)
- else:
- resp = self._get_chat_response(user_id, agent, None)
- if resp:
- if set(user_tags).intersection(voice_tags):
- message_type = MessageType.VOICE
- else:
- message_type = MessageType.TEXT
- self._send_response(staff_id, user_id, resp, message_type, skip_check=True)
- if self.limit_initiative_conversation_rate:
- time.sleep(random.randint(10,20))
- agent.persist_state()
- except Exception as e:
- # FIXME:虽然需要主动唤起的用户同时发来消息的概率很低,但仍可能会有并发冲突
- agent.rollback_state()
- logger.error("Error in active greeting: {}".format(e))
- else:
- logger.debug(f"user[{user_id}], do not initiate conversation")
- def _generate_active_greeting_message(self, agent: DialogueManager, user_tags: List[str]=None):
- chat_config = agent.build_active_greeting_config(user_tags)
- chat_response = self._call_chat_api(chat_config, ChatServiceType.OPENAI_COMPATIBLE)
- chat_response = self.sanitize_response(chat_response)
- if response := agent.generate_response(chat_response):
- return response
- else:
- logger.warning(f"staff[{agent.staff_id}] user[{agent.user_id}]: no response generated")
- return None
- def _get_chat_response(self, user_id: str, agent: DialogueManager,
- user_message: Optional[str]):
- """处理LLM响应"""
- chat_config = agent.build_chat_configuration(user_message, self.chat_service_type)
- logger.debug(chat_config)
- chat_response = self._call_chat_api(chat_config, self.chat_service_type)
- chat_response = self.sanitize_response(chat_response)
- if response := agent.generate_response(chat_response):
- return response
- else:
- logger.warning(f"staff[{agent.staff_id}] user[{user_id}]: no response generated")
- return None
- def _call_chat_api(self, chat_config: Dict, chat_service_type: ChatServiceType) -> str:
- if configs.get().get('debug_flags', {}).get('disable_llm_api_call', False):
- return 'LLM模拟回复 {}'.format(datetime.now().strftime("%Y-%m-%d %H:%M:%S"))
- if chat_service_type == ChatServiceType.OPENAI_COMPATIBLE:
- # 指定了LLM模型则优先使用指定模型
- if chat_config.get('model_name', None):
- llm_client = chat_service.OpenAICompatible.create_client(chat_config['model_name'])
- chat_completion = llm_client.chat.completions.create(
- messages=chat_config['messages'],
- model=chat_config['model_name'],
- )
- elif chat_config.get('use_multimodal_model', False):
- chat_completion = self.multimodal_model_client.chat.completions.create(
- messages=chat_config['messages'],
- model=self.multimodal_model_name,
- )
- else:
- chat_completion = self.text_model_client.chat.completions.create(
- messages=chat_config['messages'],
- model=self.text_model_name,
- )
- response = chat_completion.choices[0].message.content
- elif chat_service_type == ChatServiceType.COZE_CHAT:
- bot_user_id = 'qywx_{}'.format(chat_config['user_id'])
- response = self.coze_client.create(
- chat_config['bot_id'], bot_user_id, chat_config['messages'],
- chat_config['custom_variables']
- )
- else:
- raise Exception('Unsupported chat service type: {}'.format(self.chat_service_type))
- return response
- @staticmethod
- def sanitize_response(response: str):
- pattern = r'\[?\d{4}-\d{2}-\d{2} \d{2}:\d{2}:\d{2}\]?'
- response = re.sub(pattern, '', response)
- response = response.strip()
- return response
- if __name__ == "__main__":
- config = configs.get()
- logging_service.setup_root_logger()
- logger.warning("current env: {}".format(configs.get_env()))
- scheduler_logger = logging.getLogger('apscheduler')
- scheduler_logger.setLevel(logging.WARNING)
- use_aliyun_mq = config['debug_flags']['use_aliyun_mq']
- # 初始化不同队列的后端
- if use_aliyun_mq:
- receive_queue = AliyunRocketMQQueueBackend(
- config['mq']['endpoints'],
- config['mq']['instance_id'],
- config['mq']['receive_topic'],
- has_consumer=True, has_producer=True,
- group_id=config['mq']['receive_group'],
- topic_type='FIFO'
- )
- send_queue = AliyunRocketMQQueueBackend(
- config['mq']['endpoints'],
- config['mq']['instance_id'],
- config['mq']['send_topic'],
- has_consumer=False, has_producer=True,
- topic_type='FIFO'
- )
- else:
- receive_queue = MemoryQueueBackend()
- send_queue = MemoryQueueBackend()
- human_queue = MemoryQueueBackend()
- # 初始化用户管理服务
- # FIXME(zhoutian): 如果不使用MySQL,此数据库配置非必须
- user_db_config = config['storage']['user']
- staff_db_config = config['storage']['staff']
- wecom_db_config = config['storage']['user_relation']
- if config['debug_flags'].get('use_local_user_storage', False):
- user_manager = LocalUserManager()
- user_relation_manager = LocalUserRelationManager()
- else:
- user_manager = MySQLUserManager(user_db_config['mysql'], user_db_config['table'], staff_db_config['table'])
- user_relation_manager = MySQLUserRelationManager(
- user_db_config['mysql'], wecom_db_config['mysql'],
- config['storage']['staff']['table'],
- user_db_config['table'],
- wecom_db_config['table']['staff'],
- wecom_db_config['table']['relation'],
- wecom_db_config['table']['user']
- )
- # 创建Agent服务
- service = AgentService(
- receive_backend=receive_queue,
- send_backend=send_queue,
- human_backend=human_queue,
- user_manager=user_manager,
- user_relation_manager=user_relation_manager,
- chat_service_type=ChatServiceType.COZE_CHAT
- )
- if not config['debug_flags'].get('console_input', False):
- service.start(blocking=True)
- sys.exit(0)
- else:
- service.start()
- message_id = 0
- while service.running:
- print("Input next message: ")
- text = sys.stdin.readline().strip()
- if not text:
- continue
- message_id += 1
- sender = '7881301263964433'
- receiver = '1688854492669990'
- if text == MessageType.AGGREGATION_TRIGGER.name:
- message = Message.build(MessageType.AGGREGATION_TRIGGER, MessageChannel.CORP_WECHAT,
- sender, receiver, None, int(time.time() * 1000))
- else:
- message = Message.build(MessageType.TEXT, MessageChannel.CORP_WECHAT,
- sender,receiver, text, int(time.time() * 1000)
- )
- message.msgId = message_id
- receive_queue.produce(message)
- time.sleep(0.1)
|