""" 压缩后缓存功能测试 通过降低压缩阈值来快速触发压缩,测试压缩后的缓存行为 """ import argparse import os import sys import asyncio from pathlib import Path # 添加项目根目录到 Python 路径 sys.path.insert(0, str(Path(__file__).parent.parent.parent)) from dotenv import load_dotenv load_dotenv() from agent.llm.prompts import SimplePrompt from agent.core.runner import AgentRunner, RunConfig from agent.trace import FileSystemTraceStore, Trace, Message from agent.trace.compaction import CompressionConfig from agent.llm import create_openrouter_llm_call async def main(): # 路径配置 base_dir = Path(__file__).parent prompt_path = base_dir / "test.prompt" output_dir = base_dir / "output" output_dir.mkdir(exist_ok=True) print("=" * 60) print("压缩后缓存功能测试") print("=" * 60) print() # 加载 prompt print("1. 加载 prompt 配置...") prompt = SimplePrompt(prompt_path) # 构建消息 print("2. 构建任务消息...") messages = prompt.build_messages() # 创建 Agent Runner with 低压缩阈值 print("3. 创建 Agent Runner...") print(f" - 模型: {prompt.config.get('model', 'sonnet-4.6')}") print(f" - 压缩阈值: 10,000 tokens (降低以快速触发)") store = FileSystemTraceStore(base_path=".trace") # 创建自定义压缩配置 compression_config = CompressionConfig( max_tokens=10000, # 降低到10K以快速触发压缩 threshold_ratio=0.5, keep_recent_messages=10 ) runner = AgentRunner( trace_store=store, llm_call=create_openrouter_llm_call(model=f"anthropic/claude-{prompt.config.get('model', 'sonnet-4.6')}"), skills_dir=None, debug=True, compression_config=compression_config # 使用自定义压缩配置 ) print(f"4. 启动新 Agent 模式...") print() current_trace_id = None compression_detected = False try: initial_messages = messages config = RunConfig( model=f"anthropic/claude-{prompt.config.get('model', 'sonnet-4.6')}", temperature=float(prompt.config.get('temperature', 0.3)), max_iterations=100, name="压缩缓存测试", ) print("▶️ 开始执行...") print() async for item in runner.run(messages=initial_messages, config=config): # 处理 Trace 对象 if isinstance(item, Trace): current_trace_id = item.trace_id if item.status == "running": print(f"[Trace] 开始: {item.trace_id[:8]}...") elif item.status == "completed": print(f"\n[Trace] ✅ 完成") print(f" - Total messages: {item.total_messages}") print(f" - Total tokens: {item.total_tokens:,}") print(f" - Cache creation: {item.total_cache_creation_tokens:,}") print(f" - Cache read: {item.total_cache_read_tokens:,}") if item.total_prompt_tokens > 0: print(f" - Cache hit rate: {item.total_cache_read_tokens / item.total_prompt_tokens * 100:.1f}%") print(f" - Total cost: ${item.total_cost:.4f}") elif item.status == "failed": print(f"\n[Trace] ❌ 失败: {item.error_message}") # 处理 Message 对象 elif isinstance(item, Message): if item.role == "assistant": content = item.content if isinstance(content, dict): tool_calls = content.get("tool_calls") if tool_calls: print(f"[{item.sequence}] Tool calls: {len(tool_calls)}") # 检测压缩消息 if item.role == "user" and isinstance(item.content, str): if "对话历史摘要" in item.content or "自动压缩" in item.content: if not compression_detected: compression_detected = True print(f"\n{'='*60}") print(f"🔄 检测到压缩发生在 sequence {item.sequence}") print(f"{'='*60}\n") except KeyboardInterrupt: print("\n\n用户中断 (Ctrl+C)") if current_trace_id: await runner.stop(current_trace_id) # 分析缓存情况 if current_trace_id: print() print("=" * 60) print("缓存分析") print("=" * 60) trace = await store.get_trace(current_trace_id) if trace: print(f"\nTrace ID: {current_trace_id}") print(f"总消息数: {trace.total_messages}") print(f"总 tokens: {trace.total_tokens:,}") print(f"Prompt tokens: {trace.total_prompt_tokens:,}") print(f"Cache creation: {trace.total_cache_creation_tokens:,} ({trace.total_cache_creation_tokens / trace.total_prompt_tokens * 100:.1f}%)") print(f"Cache read: {trace.total_cache_read_tokens:,} ({trace.total_cache_read_tokens / trace.total_prompt_tokens * 100:.1f}%)") print(f"总成本: ${trace.total_cost:.4f}") if compression_detected: print(f"\n✅ 压缩已触发") print(f" - 压缩后缓存机制应该只保留系统prompt缓存") print(f" - 新的message缓存点会在压缩后重新创建") else: print(f"\n⚠️ 未检测到压缩") print() print(f"Trace 目录: .trace/{current_trace_id}") if __name__ == "__main__": asyncio.run(main())