| |
| """ |
| Code Implementation MCP Server |
| |
| This MCP server provides core functions needed for paper code reproduction: |
| 1. File read/write operations |
| 2. Code execution and testing |
| 3. Code search and analysis |
| 4. Iterative improvement support |
| |
| Usage: |
| python tools/code_implementation_server.py |
| """ |
|
|
| import os |
| import subprocess |
| import json |
| import sys |
| import io |
| from pathlib import Path |
| import re |
| from typing import Dict, Any, List |
| import tempfile |
| import shutil |
| import logging |
| from datetime import datetime |
|
|
| |
| if sys.stdout.encoding != "utf-8": |
| try: |
| if hasattr(sys.stdout, "reconfigure"): |
| sys.stdout.reconfigure(encoding="utf-8") |
| sys.stderr.reconfigure(encoding="utf-8") |
| else: |
| sys.stdout = io.TextIOWrapper(sys.stdout.detach(), encoding="utf-8") |
| sys.stderr = io.TextIOWrapper(sys.stderr.detach(), encoding="utf-8") |
| except Exception as e: |
| print(f"Warning: Could not set UTF-8 encoding: {e}") |
|
|
| |
| from mcp.server.fastmcp import FastMCP |
|
|
| |
| logging.basicConfig(level=logging.INFO) |
| logger = logging.getLogger(__name__) |
|
|
| |
| mcp = FastMCP("code-implementation-server") |
|
|
| |
| WORKSPACE_DIR = None |
| OPERATION_HISTORY = [] |
| CURRENT_FILES = {} |
|
|
|
|
| def initialize_workspace(workspace_dir: str = None): |
| """ |
| Initialize workspace |
| |
| By default, the workspace will be set by the workflow via the set_workspace tool to: |
| {plan_file_parent}/generate_code |
| |
| Args: |
| workspace_dir: Optional workspace directory path |
| """ |
| global WORKSPACE_DIR |
| if workspace_dir is None: |
| |
| |
| WORKSPACE_DIR = Path.cwd() / "generate_code" |
| |
| |
| else: |
| WORKSPACE_DIR = Path(workspace_dir).resolve() |
| |
| WORKSPACE_DIR.mkdir(parents=True, exist_ok=True) |
| logger.info(f"Workspace initialized: {WORKSPACE_DIR}") |
|
|
|
|
| def ensure_workspace_exists(): |
| """Ensure workspace directory exists""" |
| global WORKSPACE_DIR |
| if WORKSPACE_DIR is None: |
| initialize_workspace() |
|
|
| |
| if not WORKSPACE_DIR.exists(): |
| WORKSPACE_DIR.mkdir(parents=True, exist_ok=True) |
| logger.info(f"Workspace directory created: {WORKSPACE_DIR}") |
|
|
|
|
| def validate_path(path: str) -> Path: |
| """Validate if path is within workspace""" |
| if WORKSPACE_DIR is None: |
| initialize_workspace() |
|
|
| full_path = (WORKSPACE_DIR / path).resolve() |
| if not str(full_path).startswith(str(WORKSPACE_DIR)): |
| raise ValueError(f"Path {path} is outside workspace scope") |
| return full_path |
|
|
|
|
| def log_operation(action: str, details: Dict[str, Any]): |
| """Log operation history""" |
| OPERATION_HISTORY.append( |
| {"timestamp": datetime.now().isoformat(), "action": action, "details": details} |
| ) |
|
|
|
|
| |
|
|
|
|
| @mcp.tool() |
| async def read_file( |
| file_path: str, start_line: int = None, end_line: int = None |
| ) -> str: |
| """ |
| Read file content, supports specifying line number range |
| |
| Args: |
| file_path: File path, relative to workspace |
| start_line: Starting line number (1-based, optional) |
| end_line: Ending line number (1-based, optional) |
| |
| Returns: |
| JSON string of file content or error message |
| """ |
| try: |
| full_path = validate_path(file_path) |
|
|
| if not full_path.exists(): |
| result = {"status": "error", "message": f"File does not exist: {file_path}"} |
| log_operation( |
| "read_file_error", {"file_path": file_path, "error": "file_not_found"} |
| ) |
| return json.dumps(result, ensure_ascii=False, indent=2) |
|
|
| with open(full_path, "r", encoding="utf-8") as f: |
| lines = f.readlines() |
|
|
| |
| if start_line is not None or end_line is not None: |
| start_idx = (start_line - 1) if start_line else 0 |
| end_idx = end_line if end_line else len(lines) |
| lines = lines[start_idx:end_idx] |
|
|
| content = "".join(lines) |
|
|
| result = { |
| "status": "success", |
| "content": content, |
| "file_path": file_path, |
| "total_lines": len(lines), |
| "size_bytes": len(content.encode("utf-8")), |
| } |
|
|
| log_operation( |
| "read_file", |
| { |
| "file_path": file_path, |
| "start_line": start_line, |
| "end_line": end_line, |
| "lines_read": len(lines), |
| }, |
| ) |
|
|
| return json.dumps(result, ensure_ascii=False, indent=2) |
|
|
| except Exception as e: |
| result = { |
| "status": "error", |
| "message": f"Failed to read file: {str(e)}", |
| "file_path": file_path, |
| } |
| log_operation("read_file_error", {"file_path": file_path, "error": str(e)}) |
| return json.dumps(result, ensure_ascii=False, indent=2) |
|
|
|
|
| @mcp.tool() |
| async def read_multiple_files(file_requests: str, max_files: int = 5) -> str: |
| """ |
| Read multiple files in a single operation (for batch reading) |
| |
| Args: |
| file_requests: JSON string with file requests, e.g., |
| '{"file1.py": {}, "file2.py": {"start_line": 1, "end_line": 10}}' |
| or simple array: '["file1.py", "file2.py"]' |
| max_files: Maximum number of files to read in one operation (default: 5) |
| |
| Returns: |
| JSON string of operation results for all files |
| """ |
| try: |
| |
| try: |
| requests_data = json.loads(file_requests) |
| except json.JSONDecodeError as e: |
| return json.dumps( |
| { |
| "status": "error", |
| "message": f"Invalid JSON format for file_requests: {str(e)}", |
| "operation_type": "multi_file", |
| "timestamp": datetime.now().isoformat(), |
| }, |
| ensure_ascii=False, |
| indent=2, |
| ) |
|
|
| |
| if isinstance(requests_data, list): |
| |
| normalized_requests = {file_path: {} for file_path in requests_data} |
| elif isinstance(requests_data, dict): |
| normalized_requests = requests_data |
| else: |
| return json.dumps( |
| { |
| "status": "error", |
| "message": "file_requests must be a JSON object or array", |
| "operation_type": "multi_file", |
| "timestamp": datetime.now().isoformat(), |
| }, |
| ensure_ascii=False, |
| indent=2, |
| ) |
|
|
| |
| if len(normalized_requests) == 0: |
| return json.dumps( |
| { |
| "status": "error", |
| "message": "No files provided for reading", |
| "operation_type": "multi_file", |
| "timestamp": datetime.now().isoformat(), |
| }, |
| ensure_ascii=False, |
| indent=2, |
| ) |
|
|
| if len(normalized_requests) > max_files: |
| return json.dumps( |
| { |
| "status": "error", |
| "message": f"Too many files provided ({len(normalized_requests)}), maximum is {max_files}", |
| "operation_type": "multi_file", |
| "timestamp": datetime.now().isoformat(), |
| }, |
| ensure_ascii=False, |
| indent=2, |
| ) |
|
|
| |
| results = { |
| "status": "success", |
| "message": f"Successfully processed {len(normalized_requests)} files", |
| "operation_type": "multi_file", |
| "timestamp": datetime.now().isoformat(), |
| "files_processed": len(normalized_requests), |
| "files": {}, |
| "summary": { |
| "successful": 0, |
| "failed": 0, |
| "total_size_bytes": 0, |
| "total_lines": 0, |
| "files_not_found": 0, |
| }, |
| } |
|
|
| |
| for file_path, options in normalized_requests.items(): |
| try: |
| full_path = validate_path(file_path) |
| start_line = options.get("start_line") |
| end_line = options.get("end_line") |
|
|
| if not full_path.exists(): |
| results["files"][file_path] = { |
| "status": "error", |
| "message": f"File does not exist: {file_path}", |
| "file_path": file_path, |
| "content": "", |
| "total_lines": 0, |
| "size_bytes": 0, |
| "start_line": start_line, |
| "end_line": end_line, |
| } |
| results["summary"]["failed"] += 1 |
| results["summary"]["files_not_found"] += 1 |
| continue |
|
|
| with open(full_path, "r", encoding="utf-8") as f: |
| lines = f.readlines() |
|
|
| |
| original_line_count = len(lines) |
| if start_line is not None or end_line is not None: |
| start_idx = (start_line - 1) if start_line else 0 |
| end_idx = end_line if end_line else len(lines) |
| lines = lines[start_idx:end_idx] |
|
|
| content = "".join(lines) |
| size_bytes = len(content.encode("utf-8")) |
| lines_count = len(lines) |
|
|
| |
| results["files"][file_path] = { |
| "status": "success", |
| "message": f"File read successfully: {file_path}", |
| "file_path": file_path, |
| "content": content, |
| "total_lines": lines_count, |
| "original_total_lines": original_line_count, |
| "size_bytes": size_bytes, |
| "start_line": start_line, |
| "end_line": end_line, |
| "line_range_applied": start_line is not None |
| or end_line is not None, |
| } |
|
|
| |
| results["summary"]["successful"] += 1 |
| results["summary"]["total_size_bytes"] += size_bytes |
| results["summary"]["total_lines"] += lines_count |
|
|
| |
| log_operation( |
| "read_file_multi", |
| { |
| "file_path": file_path, |
| "start_line": start_line, |
| "end_line": end_line, |
| "lines_read": lines_count, |
| "size_bytes": size_bytes, |
| "batch_operation": True, |
| }, |
| ) |
|
|
| except Exception as file_error: |
| |
| results["files"][file_path] = { |
| "status": "error", |
| "message": f"Failed to read file: {str(file_error)}", |
| "file_path": file_path, |
| "content": "", |
| "total_lines": 0, |
| "size_bytes": 0, |
| "start_line": options.get("start_line"), |
| "end_line": options.get("end_line"), |
| } |
|
|
| results["summary"]["failed"] += 1 |
|
|
| |
| log_operation( |
| "read_file_multi_error", |
| { |
| "file_path": file_path, |
| "error": str(file_error), |
| "batch_operation": True, |
| }, |
| ) |
|
|
| |
| if results["summary"]["failed"] > 0: |
| if results["summary"]["successful"] > 0: |
| results["status"] = "partial_success" |
| results["message"] = ( |
| f"Read {results['summary']['successful']} files successfully, {results['summary']['failed']} failed" |
| ) |
| else: |
| results["status"] = "failed" |
| results["message"] = ( |
| f"All {results['summary']['failed']} files failed to read" |
| ) |
|
|
| |
| log_operation( |
| "read_multiple_files", |
| { |
| "files_count": len(normalized_requests), |
| "successful": results["summary"]["successful"], |
| "failed": results["summary"]["failed"], |
| "total_size_bytes": results["summary"]["total_size_bytes"], |
| "status": results["status"], |
| }, |
| ) |
|
|
| return json.dumps(results, ensure_ascii=False, indent=2) |
|
|
| except Exception as e: |
| result = { |
| "status": "error", |
| "message": f"Failed to read multiple files: {str(e)}", |
| "operation_type": "multi_file", |
| "timestamp": datetime.now().isoformat(), |
| "files_processed": 0, |
| } |
| log_operation("read_multiple_files_error", {"error": str(e)}) |
| return json.dumps(result, ensure_ascii=False, indent=2) |
|
|
|
|
| @mcp.tool() |
| async def write_file( |
| file_path: str, content: str, create_dirs: bool = True, create_backup: bool = False |
| ) -> str: |
| """ |
| Write content to file |
| |
| Args: |
| file_path: File path, relative to workspace |
| content: Content to write to file |
| create_dirs: Whether to create directories if they don't exist |
| create_backup: Whether to create backup file if file already exists |
| |
| Returns: |
| JSON string of operation result |
| """ |
| try: |
| full_path = validate_path(file_path) |
|
|
| |
| if create_dirs: |
| full_path.parent.mkdir(parents=True, exist_ok=True) |
|
|
| |
| backup_created = False |
| if full_path.exists() and create_backup: |
| backup_path = full_path.with_suffix(full_path.suffix + ".backup") |
| shutil.copy2(full_path, backup_path) |
| backup_created = True |
|
|
| |
| with open(full_path, "w", encoding="utf-8") as f: |
| f.write(content) |
|
|
| |
| CURRENT_FILES[file_path] = { |
| "last_modified": datetime.now().isoformat(), |
| "size_bytes": len(content.encode("utf-8")), |
| "lines": len(content.split("\n")), |
| } |
|
|
| result = { |
| "status": "success", |
| "message": f"File written successfully: {file_path}", |
| "file_path": file_path, |
| "size_bytes": len(content.encode("utf-8")), |
| "lines_written": len(content.split("\n")), |
| "backup_created": backup_created, |
| } |
|
|
| log_operation( |
| "write_file", |
| { |
| "file_path": file_path, |
| "size_bytes": len(content.encode("utf-8")), |
| "lines": len(content.split("\n")), |
| "backup_created": backup_created, |
| }, |
| ) |
|
|
| return json.dumps(result, ensure_ascii=False, indent=2) |
|
|
| except Exception as e: |
| result = { |
| "status": "error", |
| "message": f"Failed to write file: {str(e)}", |
| "file_path": file_path, |
| } |
| log_operation("write_file_error", {"file_path": file_path, "error": str(e)}) |
| return json.dumps(result, ensure_ascii=False, indent=2) |
|
|
|
|
| @mcp.tool() |
| async def write_multiple_files( |
| file_implementations: str, |
| create_dirs: bool = True, |
| create_backup: bool = False, |
| max_files: int = 5, |
| ) -> str: |
| """ |
| Write multiple files in a single operation (for batch implementation) |
| |
| Args: |
| file_implementations: JSON string mapping file paths to content, e.g., |
| '{"file1.py": "content1", "file2.py": "content2"}' |
| create_dirs: Whether to create directories if they don't exist |
| create_backup: Whether to create backup files if they already exist |
| max_files: Maximum number of files to write in one operation (default: 5) |
| |
| Returns: |
| JSON string of operation results for all files |
| """ |
| try: |
| |
| try: |
| files_dict = json.loads(file_implementations) |
| except json.JSONDecodeError as e: |
| return json.dumps( |
| { |
| "status": "error", |
| "message": f"Invalid JSON format for file_implementations: {str(e)}", |
| "operation_type": "multi_file", |
| "timestamp": datetime.now().isoformat(), |
| }, |
| ensure_ascii=False, |
| indent=2, |
| ) |
|
|
| |
| if not isinstance(files_dict, dict): |
| return json.dumps( |
| { |
| "status": "error", |
| "message": "file_implementations must be a JSON object mapping file paths to content", |
| "operation_type": "multi_file", |
| "timestamp": datetime.now().isoformat(), |
| }, |
| ensure_ascii=False, |
| indent=2, |
| ) |
|
|
| if len(files_dict) == 0: |
| return json.dumps( |
| { |
| "status": "error", |
| "message": "No files provided for writing", |
| "operation_type": "multi_file", |
| "timestamp": datetime.now().isoformat(), |
| }, |
| ensure_ascii=False, |
| indent=2, |
| ) |
|
|
| if len(files_dict) > max_files: |
| return json.dumps( |
| { |
| "status": "error", |
| "message": f"Too many files provided ({len(files_dict)}), maximum is {max_files}", |
| "operation_type": "multi_file", |
| "timestamp": datetime.now().isoformat(), |
| }, |
| ensure_ascii=False, |
| indent=2, |
| ) |
|
|
| |
| results = { |
| "status": "success", |
| "message": f"Successfully processed {len(files_dict)} files", |
| "operation_type": "multi_file", |
| "timestamp": datetime.now().isoformat(), |
| "files_processed": len(files_dict), |
| "files": {}, |
| "summary": { |
| "successful": 0, |
| "failed": 0, |
| "total_size_bytes": 0, |
| "total_lines": 0, |
| "backups_created": 0, |
| }, |
| } |
|
|
| |
| for file_path, content in files_dict.items(): |
| try: |
| full_path = validate_path(file_path) |
|
|
| |
| if create_dirs: |
| full_path.parent.mkdir(parents=True, exist_ok=True) |
|
|
| |
| backup_created = False |
| if full_path.exists() and create_backup: |
| backup_path = full_path.with_suffix(full_path.suffix + ".backup") |
| shutil.copy2(full_path, backup_path) |
| backup_created = True |
| results["summary"]["backups_created"] += 1 |
|
|
| |
| with open(full_path, "w", encoding="utf-8") as f: |
| f.write(content) |
|
|
| |
| size_bytes = len(content.encode("utf-8")) |
| lines_count = len(content.split("\n")) |
|
|
| |
| CURRENT_FILES[file_path] = { |
| "last_modified": datetime.now().isoformat(), |
| "size_bytes": size_bytes, |
| "lines": lines_count, |
| } |
|
|
| |
| results["files"][file_path] = { |
| "status": "success", |
| "message": f"File written successfully: {file_path}", |
| "size_bytes": size_bytes, |
| "lines_written": lines_count, |
| "backup_created": backup_created, |
| } |
|
|
| |
| results["summary"]["successful"] += 1 |
| results["summary"]["total_size_bytes"] += size_bytes |
| results["summary"]["total_lines"] += lines_count |
|
|
| |
| log_operation( |
| "write_file_multi", |
| { |
| "file_path": file_path, |
| "size_bytes": size_bytes, |
| "lines": lines_count, |
| "backup_created": backup_created, |
| "batch_operation": True, |
| }, |
| ) |
|
|
| except Exception as file_error: |
| |
| results["files"][file_path] = { |
| "status": "error", |
| "message": f"Failed to write file: {str(file_error)}", |
| "size_bytes": 0, |
| "lines_written": 0, |
| "backup_created": False, |
| } |
|
|
| results["summary"]["failed"] += 1 |
|
|
| |
| log_operation( |
| "write_file_multi_error", |
| { |
| "file_path": file_path, |
| "error": str(file_error), |
| "batch_operation": True, |
| }, |
| ) |
|
|
| |
| if results["summary"]["failed"] > 0: |
| if results["summary"]["successful"] > 0: |
| results["status"] = "partial_success" |
| results["message"] = ( |
| f"Processed {results['summary']['successful']} files successfully, {results['summary']['failed']} failed" |
| ) |
| else: |
| results["status"] = "failed" |
| results["message"] = ( |
| f"All {results['summary']['failed']} files failed to write" |
| ) |
|
|
| |
| log_operation( |
| "write_multiple_files", |
| { |
| "files_count": len(files_dict), |
| "successful": results["summary"]["successful"], |
| "failed": results["summary"]["failed"], |
| "total_size_bytes": results["summary"]["total_size_bytes"], |
| "status": results["status"], |
| }, |
| ) |
|
|
| return json.dumps(results, ensure_ascii=False, indent=2) |
|
|
| except Exception as e: |
| result = { |
| "status": "error", |
| "message": f"Failed to write multiple files: {str(e)}", |
| "operation_type": "multi_file", |
| "timestamp": datetime.now().isoformat(), |
| "files_processed": 0, |
| } |
| log_operation("write_multiple_files_error", {"error": str(e)}) |
| return json.dumps(result, ensure_ascii=False, indent=2) |
|
|
|
|
| |
|
|
|
|
| @mcp.tool() |
| async def execute_python(code: str, timeout: int = 30) -> str: |
| """ |
| Execute Python code and return output |
| |
| Args: |
| code: Python code to execute |
| timeout: Timeout in seconds |
| |
| Returns: |
| JSON string of execution result |
| """ |
| try: |
| |
| with tempfile.NamedTemporaryFile( |
| mode="w", suffix=".py", delete=False, encoding="utf-8" |
| ) as f: |
| f.write(code) |
| temp_file = f.name |
|
|
| try: |
| |
| ensure_workspace_exists() |
|
|
| |
| result = subprocess.run( |
| [sys.executable, temp_file], |
| cwd=WORKSPACE_DIR, |
| capture_output=True, |
| text=True, |
| timeout=timeout, |
| encoding="utf-8", |
| ) |
|
|
| execution_result = { |
| "status": "success" if result.returncode == 0 else "error", |
| "return_code": result.returncode, |
| "stdout": result.stdout, |
| "stderr": result.stderr, |
| "timeout": timeout, |
| } |
|
|
| if result.returncode != 0: |
| execution_result["message"] = "Python code execution failed" |
| else: |
| execution_result["message"] = "Python code execution successful" |
|
|
| log_operation( |
| "execute_python", |
| { |
| "return_code": result.returncode, |
| "stdout_length": len(result.stdout), |
| "stderr_length": len(result.stderr), |
| }, |
| ) |
|
|
| return json.dumps(execution_result, ensure_ascii=False, indent=2) |
|
|
| finally: |
| |
| os.unlink(temp_file) |
|
|
| except subprocess.TimeoutExpired: |
| result = { |
| "status": "error", |
| "message": f"Python code execution timeout ({timeout}秒)", |
| "timeout": timeout, |
| } |
| log_operation("execute_python_timeout", {"timeout": timeout}) |
| return json.dumps(result, ensure_ascii=False, indent=2) |
|
|
| except Exception as e: |
| result = { |
| "status": "error", |
| "message": f"Python code execution failed: {str(e)}", |
| } |
| log_operation("execute_python_error", {"error": str(e)}) |
| return json.dumps(result, ensure_ascii=False, indent=2) |
|
|
|
|
| @mcp.tool() |
| async def execute_bash(command: str, timeout: int = 30) -> str: |
| """ |
| Execute bash command |
| |
| Args: |
| command: Bash command to execute |
| timeout: Timeout in seconds |
| |
| Returns: |
| JSON string of execution result |
| """ |
| try: |
| |
| dangerous_commands = ["rm -rf", "sudo", "chmod 777", "mkfs", "dd if="] |
| if any(dangerous in command.lower() for dangerous in dangerous_commands): |
| result = { |
| "status": "error", |
| "message": f"Dangerous command execution prohibited: {command}", |
| } |
| log_operation( |
| "execute_bash_blocked", |
| {"command": command, "reason": "dangerous_command"}, |
| ) |
| return json.dumps(result, ensure_ascii=False, indent=2) |
|
|
| |
| ensure_workspace_exists() |
|
|
| |
| result = subprocess.run( |
| command, |
| shell=True, |
| cwd=WORKSPACE_DIR, |
| capture_output=True, |
| text=True, |
| timeout=timeout, |
| encoding="utf-8", |
| ) |
|
|
| execution_result = { |
| "status": "success" if result.returncode == 0 else "error", |
| "return_code": result.returncode, |
| "stdout": result.stdout, |
| "stderr": result.stderr, |
| "command": command, |
| "timeout": timeout, |
| } |
|
|
| if result.returncode != 0: |
| execution_result["message"] = "Bash command execution failed" |
| else: |
| execution_result["message"] = "Bash command execution successful" |
|
|
| log_operation( |
| "execute_bash", |
| { |
| "command": command, |
| "return_code": result.returncode, |
| "stdout_length": len(result.stdout), |
| "stderr_length": len(result.stderr), |
| }, |
| ) |
|
|
| return json.dumps(execution_result, ensure_ascii=False, indent=2) |
|
|
| except subprocess.TimeoutExpired: |
| result = { |
| "status": "error", |
| "message": f"Bash command execution timeout ({timeout} seconds)", |
| "command": command, |
| "timeout": timeout, |
| } |
| log_operation("execute_bash_timeout", {"command": command, "timeout": timeout}) |
| return json.dumps(result, ensure_ascii=False, indent=2) |
|
|
| except Exception as e: |
| result = { |
| "status": "error", |
| "message": f"Failed to execute bash command: {str(e)}", |
| "command": command, |
| } |
| log_operation("execute_bash_error", {"command": command, "error": str(e)}) |
| return json.dumps(result, ensure_ascii=False, indent=2) |
|
|
|
|
| @mcp.tool() |
| async def read_code_mem(file_paths: List[str]) -> str: |
| """ |
| Check if file summaries exist in implement_code_summary.md for multiple files |
| |
| Args: |
| file_paths: List of file paths to check for summary information in implement_code_summary.md |
| |
| Returns: |
| Summary information for all requested files if available |
| """ |
| try: |
| if not file_paths or not isinstance(file_paths, list): |
| result = { |
| "status": "error", |
| "message": "file_paths parameter is required and must be a list", |
| } |
| log_operation( |
| "read_code_mem_error", {"error": "missing_or_invalid_file_paths"} |
| ) |
| return json.dumps(result, ensure_ascii=False, indent=2) |
|
|
| |
| unique_file_paths = list(dict.fromkeys(file_paths)) |
|
|
| |
| ensure_workspace_exists() |
|
|
| |
| current_path = Path(WORKSPACE_DIR) |
| summary_file_path = current_path.parent / "implement_code_summary.md" |
|
|
| if not summary_file_path.exists(): |
| result = { |
| "status": "no_summary", |
| "file_paths": unique_file_paths, |
| "message": "No summary file found.", |
| "results": [], |
| } |
| log_operation( |
| "read_code_mem", |
| {"file_paths": unique_file_paths, "status": "no_summary_file"}, |
| ) |
| return json.dumps(result, ensure_ascii=False, indent=2) |
|
|
| |
| with open(summary_file_path, "r", encoding="utf-8") as f: |
| summary_content = f.read() |
|
|
| if not summary_content.strip(): |
| result = { |
| "status": "no_summary", |
| "file_paths": unique_file_paths, |
| "message": "Summary file is empty.", |
| "results": [], |
| } |
| log_operation( |
| "read_code_mem", |
| {"file_paths": unique_file_paths, "status": "empty_summary"}, |
| ) |
| return json.dumps(result, ensure_ascii=False, indent=2) |
|
|
| |
| results = [] |
| summaries_found = 0 |
|
|
| for file_path in unique_file_paths: |
| |
| file_section = _extract_file_section_from_summary( |
| summary_content, file_path |
| ) |
|
|
| if file_section: |
| file_result = { |
| "file_path": file_path, |
| "status": "summary_found", |
| "summary_content": file_section, |
| "message": f"Summary information found for {file_path}", |
| } |
| summaries_found += 1 |
| else: |
| file_result = { |
| "file_path": file_path, |
| "status": "no_summary", |
| "summary_content": None, |
| "message": f"No summary found for {file_path}", |
| } |
|
|
| results.append(file_result) |
|
|
| |
| if summaries_found == len(unique_file_paths): |
| overall_status = "all_summaries_found" |
| elif summaries_found > 0: |
| overall_status = "partial_summaries_found" |
| else: |
| overall_status = "no_summaries_found" |
|
|
| result = { |
| "status": overall_status, |
| "file_paths": unique_file_paths, |
| "total_requested": len(unique_file_paths), |
| "summaries_found": summaries_found, |
| "message": f"Found summaries for {summaries_found}/{len(unique_file_paths)} files", |
| "results": results, |
| } |
|
|
| log_operation( |
| "read_code_mem", |
| { |
| "file_paths": unique_file_paths, |
| "status": overall_status, |
| "total_requested": len(unique_file_paths), |
| "summaries_found": summaries_found, |
| }, |
| ) |
|
|
| return json.dumps(result, ensure_ascii=False, indent=2) |
|
|
| except Exception as e: |
| result = { |
| "status": "error", |
| "message": f"Failed to check code memory: {str(e)}", |
| "file_paths": file_paths |
| if isinstance(file_paths, list) |
| else [str(file_paths)], |
| "results": [], |
| } |
| log_operation( |
| "read_code_mem_error", {"file_paths": file_paths, "error": str(e)} |
| ) |
| return json.dumps(result, ensure_ascii=False, indent=2) |
|
|
|
|
| def _extract_file_section_from_summary( |
| summary_content: str, target_file_path: str |
| ) -> str: |
| """ |
| Extract the specific section for a file from the summary content |
| |
| Args: |
| summary_content: Full summary content |
| target_file_path: Path of the target file |
| |
| Returns: |
| File-specific section or None if not found |
| """ |
| import re |
|
|
| |
| normalized_target = _normalize_file_path(target_file_path) |
|
|
| |
| section_pattern = r"={80}\s*\n## IMPLEMENTATION File ([^;]+); ROUND \d+\s*\n={80}(.*?)(?=\n={80}|\Z)" |
|
|
| matches = re.findall(section_pattern, summary_content, re.DOTALL) |
|
|
| for file_path_in_summary, section_content in matches: |
| file_path_in_summary = file_path_in_summary.strip() |
| section_content = section_content.strip() |
|
|
| |
| normalized_summary_path = _normalize_file_path(file_path_in_summary) |
|
|
| |
| if _paths_match( |
| normalized_target, |
| normalized_summary_path, |
| target_file_path, |
| file_path_in_summary, |
| ): |
| |
| file_section = f"""================================================================================ |
| ## IMPLEMENTATION File {file_path_in_summary}; ROUND [X] |
| ================================================================================ |
| |
| {section_content} |
| |
| --- |
| *Extracted from implement_code_summary.md*""" |
| return file_section |
|
|
| |
| return _extract_file_section_alternative(summary_content, target_file_path) |
|
|
|
|
| def _normalize_file_path(file_path: str) -> str: |
| """Normalize file path for comparison""" |
| |
| normalized = file_path.strip("/").lower() |
| |
| normalized = normalized.replace("\\", "/") |
|
|
| |
| common_prefixes = ["src/", "./src/", "./", "core/", "lib/", "main/"] |
| for prefix in common_prefixes: |
| if normalized.startswith(prefix): |
| normalized = normalized[len(prefix) :] |
| break |
|
|
| return normalized |
|
|
|
|
| def _paths_match( |
| normalized_target: str, |
| normalized_summary: str, |
| original_target: str, |
| original_summary: str, |
| ) -> bool: |
| """Check if two file paths match using multiple strategies""" |
|
|
| |
| if normalized_target == normalized_summary: |
| return True |
|
|
| |
| target_basename = os.path.basename(original_target) |
| summary_basename = os.path.basename(original_summary) |
| if target_basename == summary_basename and len(target_basename) > 4: |
| return True |
|
|
| |
| target_suffix = _remove_common_prefixes(normalized_target) |
| summary_suffix = _remove_common_prefixes(normalized_summary) |
| if target_suffix == summary_suffix: |
| return True |
|
|
| |
| if normalized_target.endswith(normalized_summary) or normalized_summary.endswith( |
| normalized_target |
| ): |
| return True |
|
|
| |
| if len(normalized_target) > 10 and normalized_target in normalized_summary: |
| return True |
| if len(normalized_summary) > 10 and normalized_summary in normalized_target: |
| return True |
|
|
| return False |
|
|
|
|
| def _remove_common_prefixes(file_path: str) -> str: |
| """Remove common prefixes from file path""" |
| prefixes_to_remove = ["src/", "core/", "./", "lib/", "main/"] |
| path = file_path |
|
|
| for prefix in prefixes_to_remove: |
| if path.startswith(prefix): |
| path = path[len(prefix) :] |
|
|
| return path |
|
|
|
|
| def _extract_file_section_alternative( |
| summary_content: str, target_file_path: str |
| ) -> str: |
| """Alternative method to extract file section using simpler pattern matching""" |
|
|
| |
| target_basename = os.path.basename(target_file_path) |
|
|
| |
| sections = summary_content.split("=" * 80) |
|
|
| for i, section in enumerate(sections): |
| if "## IMPLEMENTATION File" in section: |
| |
| lines = section.strip().split("\n") |
| for line in lines: |
| if "## IMPLEMENTATION File" in line: |
| |
| try: |
| file_part = line.split("File ")[1].split("; ROUND")[0].strip() |
|
|
| |
| if ( |
| _normalize_file_path(target_file_path) |
| == _normalize_file_path(file_part) |
| or target_basename == os.path.basename(file_part) |
| or target_file_path in file_part |
| or file_part.endswith(target_file_path) |
| ): |
| |
| if i + 1 < len(sections): |
| content_section = sections[i + 1].strip() |
| return f"""================================================================================ |
| ## IMPLEMENTATION File {file_part} |
| ================================================================================ |
| |
| {content_section} |
| |
| --- |
| *Extracted from implement_code_summary.md using alternative method*""" |
| except (IndexError, AttributeError): |
| continue |
|
|
| return None |
|
|
|
|
| |
|
|
|
|
| @mcp.tool() |
| async def search_code( |
| pattern: str, |
| file_pattern: str = "*.json", |
| use_regex: bool = False, |
| search_directory: str = None, |
| ) -> str: |
| """ |
| Search patterns in code files |
| |
| Args: |
| pattern: Search pattern |
| file_pattern: File pattern (e.g., '*.py') |
| use_regex: Whether to use regular expressions |
| search_directory: Specify search directory (optional, uses WORKSPACE_DIR if not specified) |
| |
| Returns: |
| JSON string of search results |
| """ |
| try: |
| |
| if search_directory: |
| |
| if os.path.isabs(search_directory): |
| search_path = Path(search_directory) |
| else: |
| |
| search_path = Path.cwd() / search_directory |
| else: |
| |
| ensure_workspace_exists() |
| search_path = WORKSPACE_DIR |
|
|
| |
| if not search_path.exists(): |
| result = { |
| "status": "error", |
| "message": f"Search directory不存在: {search_path}", |
| "pattern": pattern, |
| } |
| return json.dumps(result, ensure_ascii=False, indent=2) |
|
|
| import glob |
|
|
| |
| file_paths = glob.glob(str(search_path / "**" / file_pattern), recursive=True) |
|
|
| matches = [] |
| total_files_searched = 0 |
|
|
| for file_path in file_paths: |
| try: |
| with open(file_path, "r", encoding="utf-8") as f: |
| lines = f.readlines() |
|
|
| total_files_searched += 1 |
| relative_path = os.path.relpath(file_path, search_path) |
|
|
| for line_num, line in enumerate(lines, 1): |
| if use_regex: |
| if re.search(pattern, line): |
| matches.append( |
| { |
| "file": relative_path, |
| "line_number": line_num, |
| "line_content": line.strip(), |
| "match_type": "regex", |
| } |
| ) |
| else: |
| if pattern.lower() in line.lower(): |
| matches.append( |
| { |
| "file": relative_path, |
| "line_number": line_num, |
| "line_content": line.strip(), |
| "match_type": "substring", |
| } |
| ) |
|
|
| except Exception as e: |
| logger.warning(f"Error searching file {file_path}: {e}") |
| continue |
|
|
| result = { |
| "status": "success", |
| "pattern": pattern, |
| "file_pattern": file_pattern, |
| "use_regex": use_regex, |
| "search_directory": str(search_path), |
| "total_matches": len(matches), |
| "total_files_searched": total_files_searched, |
| "matches": matches[:50], |
| } |
|
|
| if len(matches) > 50: |
| result["note"] = f"显示前50个匹配,总共找到{len(matches)}个匹配" |
|
|
| log_operation( |
| "search_code", |
| { |
| "pattern": pattern, |
| "file_pattern": file_pattern, |
| "use_regex": use_regex, |
| "search_directory": str(search_path), |
| "total_matches": len(matches), |
| "files_searched": total_files_searched, |
| }, |
| ) |
|
|
| return json.dumps(result, ensure_ascii=False, indent=2) |
|
|
| except Exception as e: |
| result = { |
| "status": "error", |
| "message": f"Code search failed: {str(e)}", |
| "pattern": pattern, |
| } |
| log_operation("search_code_error", {"pattern": pattern, "error": str(e)}) |
| return json.dumps(result, ensure_ascii=False, indent=2) |
|
|
|
|
| |
|
|
|
|
| @mcp.tool() |
| async def get_file_structure(directory: str = ".", max_depth: int = 5) -> str: |
| """ |
| Get directory file structure |
| |
| Args: |
| directory: Directory path, relative to workspace |
| max_depth: 最大遍历深度 |
| |
| Returns: |
| JSON string of file structure |
| """ |
| try: |
| ensure_workspace_exists() |
|
|
| if directory == ".": |
| target_dir = WORKSPACE_DIR |
| else: |
| target_dir = validate_path(directory) |
|
|
| if not target_dir.exists(): |
| result = { |
| "status": "error", |
| "message": f"Directory does not exist: {directory}", |
| } |
| return json.dumps(result, ensure_ascii=False, indent=2) |
|
|
| def scan_directory(path: Path, current_depth: int = 0) -> Dict[str, Any]: |
| """Recursively scan directory""" |
| if current_depth >= max_depth: |
| return {"type": "directory", "name": path.name, "truncated": True} |
|
|
| items = [] |
| try: |
| for item in sorted(path.iterdir()): |
| relative_path = os.path.relpath(item, WORKSPACE_DIR) |
|
|
| if item.is_file(): |
| file_info = { |
| "type": "file", |
| "name": item.name, |
| "path": relative_path, |
| "size_bytes": item.stat().st_size, |
| "extension": item.suffix, |
| } |
| items.append(file_info) |
| elif item.is_dir() and not item.name.startswith("."): |
| dir_info = scan_directory(item, current_depth + 1) |
| dir_info["path"] = relative_path |
| items.append(dir_info) |
| except PermissionError: |
| pass |
|
|
| return { |
| "type": "directory", |
| "name": path.name, |
| "items": items, |
| "item_count": len(items), |
| } |
|
|
| structure = scan_directory(target_dir) |
|
|
| |
| def count_items(node): |
| if node["type"] == "file": |
| return {"files": 1, "directories": 0} |
| else: |
| counts = {"files": 0, "directories": 1} |
| for item in node.get("items", []): |
| item_counts = count_items(item) |
| counts["files"] += item_counts["files"] |
| counts["directories"] += item_counts["directories"] |
| return counts |
|
|
| counts = count_items(structure) |
|
|
| result = { |
| "status": "success", |
| "directory": directory, |
| "max_depth": max_depth, |
| "structure": structure, |
| "summary": { |
| "total_files": counts["files"], |
| "total_directories": counts["directories"] |
| - 1, |
| }, |
| } |
|
|
| log_operation( |
| "get_file_structure", |
| { |
| "directory": directory, |
| "max_depth": max_depth, |
| "total_files": counts["files"], |
| "total_directories": counts["directories"] - 1, |
| }, |
| ) |
|
|
| return json.dumps(result, ensure_ascii=False, indent=2) |
|
|
| except Exception as e: |
| result = { |
| "status": "error", |
| "message": f"Failed to get file structure: {str(e)}", |
| "directory": directory, |
| } |
| log_operation( |
| "get_file_structure_error", {"directory": directory, "error": str(e)} |
| ) |
| return json.dumps(result, ensure_ascii=False, indent=2) |
|
|
|
|
| |
|
|
|
|
| @mcp.tool() |
| async def set_workspace(workspace_path: str) -> str: |
| """ |
| Set workspace directory |
| |
| Called by workflow to set workspace to: {plan_file_parent}/generate_code |
| This ensures all file operations are executed relative to the correct project directory |
| |
| Args: |
| workspace_path: Workspace path (Usually {plan_file_parent}/generate_code) |
| |
| Returns: |
| JSON string of operation result |
| """ |
| try: |
| global WORKSPACE_DIR |
| new_workspace = Path(workspace_path).resolve() |
|
|
| |
| new_workspace.mkdir(parents=True, exist_ok=True) |
|
|
| old_workspace = WORKSPACE_DIR |
| WORKSPACE_DIR = new_workspace |
|
|
| logger.info(f"New Workspace: {WORKSPACE_DIR}") |
|
|
| result = { |
| "status": "success", |
| "message": f"Workspace setup successful: {workspace_path}", |
| "new_workspace": str(WORKSPACE_DIR), |
| } |
|
|
| log_operation( |
| "set_workspace", |
| { |
| "old_workspace": str(old_workspace) if old_workspace else None, |
| "new_workspace": str(WORKSPACE_DIR), |
| "workspace_alignment": "plan_file_parent/generate_code", |
| }, |
| ) |
|
|
| return json.dumps(result, ensure_ascii=False, indent=2) |
|
|
| except Exception as e: |
| result = { |
| "status": "error", |
| "message": f"Failed to set workspace: {str(e)}", |
| "workspace_path": workspace_path, |
| } |
| log_operation( |
| "set_workspace_error", {"workspace_path": workspace_path, "error": str(e)} |
| ) |
| return json.dumps(result, ensure_ascii=False, indent=2) |
|
|
|
|
| @mcp.tool() |
| async def get_operation_history(last_n: int = 10) -> str: |
| """ |
| Get operation history |
| |
| Args: |
| last_n: Return the last N operations |
| |
| Returns: |
| JSON string of operation history |
| """ |
| try: |
| recent_history = ( |
| OPERATION_HISTORY[-last_n:] if last_n > 0 else OPERATION_HISTORY |
| ) |
|
|
| result = { |
| "status": "success", |
| "total_operations": len(OPERATION_HISTORY), |
| "returned_operations": len(recent_history), |
| "workspace": str(WORKSPACE_DIR) if WORKSPACE_DIR else None, |
| "history": recent_history, |
| } |
|
|
| return json.dumps(result, ensure_ascii=False, indent=2) |
|
|
| except Exception as e: |
| result = { |
| "status": "error", |
| "message": f"Failed to get operation history: {str(e)}", |
| } |
| return json.dumps(result, ensure_ascii=False, indent=2) |
|
|
|
|
| |
|
|
|
|
| def main(): |
| """Start MCP server""" |
| print("🚀 Code Implementation MCP Server") |
| print( |
| "📝 Paper Code Implementation Tool Server / Paper Code Implementation Tool Server" |
| ) |
| print("") |
| print("Available tools / Available tools:") |
| |
| print( |
| " • read_code_mem - Read code summary from implement_code_summary.md / Read code summary from implement_code_summary.md" |
| ) |
| print(" • write_file - Write file contents / Write file contents") |
| print(" • execute_python - Execute Python code / Execute Python code") |
| print(" • execute_bash - Execute bash command / Execute bash commands") |
| print(" • search_code - Search code patterns / Search code patterns") |
| print(" • get_file_structure - Get file structure / Get file structure") |
| print(" • set_workspace - Set workspace / Set workspace") |
| print(" • get_operation_history - Get operation history / Get operation history") |
| print("") |
| print("🔧 Server starting...") |
|
|
| |
| initialize_workspace() |
|
|
| |
| mcp.run() |
|
|
|
|
| if __name__ == "__main__": |
| main() |
|
|