EdgeBox
A fully-featured, GUI-powered local LLM Agent sandbox with complete MCP protocol support. Features both CLI and full desktop environment, enabling AI agents to operate browsers, terminal, and other desktop applications just like humans. Based on E2B oss code.
claude mcp add --transport stdio bigppwong-edgebox node src/server.js \ --env EDGEBOX_CONFIG="path/to/config.json (optional)" \ --env EDGEBOX_LOG_LEVEL="info (optional)"
How to use
EdgeBox exposes its capabilities through the MCP (Model Context Protocol) API, providing both a full GUI desktop environment and a command-line shell for AI agents. This server lets an LLM agent perform real-world desktop interactions, including programmatically controlling applications, executing code in isolated sandboxes, and managing files, all through MCP tools. When GUI tools are enabled, agents can interact with a local desktop via a VNC session, launch and use applications like a browser and code editor, and visually perceive the desktop state via screenshots. The MCP surface is organized into Core Tools (CLI mode, always available) for code execution, shell access, and filesystem operations, plus Desktop Tools (GUI mode) for mouse, keyboard, window, and visual control tasks. Agents connect using MCP-compatible clients and can manage multiple isolated sessions using the x-session-id header.
How to install
Prerequisites
- A supported runtime for EdgeBox (as the MCP server, typically Node.js for this build).
- Git and network access to clone the repository and install dependencies.
- Optional: Docker if you prefer containerized execution.
Installation steps
-
Clone the repository: git clone https://github.com/BIGPPWONG/EdgeBox.git cd EdgeBox
-
Install dependencies (Node.js project): npm install
-
Configure runtime options (optional):
- Create or edit a config.json to adjust paths, ports, and environment variables.
- Example: { "port": 8080, "guiEnabled": true, "logLevel": "info" }
-
Run the MCP server: npm run start
-
Verify the MCP endpoint is reachable (e.g., http://localhost:8080/mcp): curl http://localhost:8080/mcp/health
If you prefer Docker, you can adapt the following pattern after building an image named edgebox:
- Docker (alternative): docker run -p 8080:8080 edgebox:latest
Prerequisites recap
- Node.js (or Docker) installed on the host
- Network access for dependencies and assets
- Basic familiarity with MCP clients and session headers
Additional notes
Tips and notes:
- GUI Tools require the GUI subsystem to be available; enable guiEnabled in your config if you want desktop interactions via VNC.
- Each MCP session is isolated; use the x-session-id header to manage multiple concurrent sandboxes.
- For security, run EdgeBox on a trusted machine and consider network isolation if exposing MCP endpoints externally.
- If you encounter permission issues with desktop automation, ensure the host user has appropriate rights to simulate input and capture screenshots.
- Environment variables like EDGEBOX_CONFIG and EDGEBOX_LOG_LEVEL can help customize behavior without editing code.
- When debugging, check logs at the configured log level to trace MCP tool invocations and session management.
Related MCP Servers
obsidian
Obsidian Knowledge-Management MCP (Model Context Protocol) server that enables AI agents and development tools to interact with an Obsidian vault. It provides a comprehensive suite of tools for reading, writing, searching, and managing notes, tags, and frontmatter, acting as a bridge to the Obsidian Local REST API plugin.
Security-Detections
MCP to help Defenders Detection Engineer Harder and Smarter
furi
CLI & API for MCP management
mentor
A Model Context Protocol server providing LLM Agents a second opinion via AI-powered Deepseek-Reasoning R1 mentorship capabilities, including code review, design critique, writing feedback, and idea brainstorming through the Deepseek API.
perplexity
A Perplexity API MCP server that unlocks Perplexity's search-augmented AI capabilities for LLM agents. Features robust error handling, secure input validation, and transparent reasoning with the showThinking parameter.
toolkit
A Model Context Protocol server providing LLM Agents with system utilities and tools, including IP geolocation, network diagnostics, system monitoring, cryptographic operations, and QR code generation.