vllora
Debug your AI agents
claude mcp add --transport stdio vllora-vllora docker run -i vllora/vllora
How to use
vLLora is a lightweight, real-time debugging and observability tool for AI agents that integrates with MCP servers. It traces and analyzes interactions between your AI agents and their providers (such as OpenAI-compatible APIs) in real time, giving you visibility into prompts, responses, tool calls, and agent workflows. The MCP capability lets you connect external tooling and dashboards to the server via HTTP and Server-Sent Events, enabling seamless data streaming and integration with your existing debugging pipelines. To start, run the MCP server (the recommended approach here is via Docker for portability). Once running, navigate to the UI at the configured ports to configure provider API keys and inspect traces as they occur. You can test tracing by sending a chat completion request to the server’s HTTP endpoint and watching the live trace feed appear in the UI.
How to install
Prerequisites:
- Docker installed on your machine (recommended for simplicity and portability).
- Optionally, ensure you have internet access to pull the vLLora image from Docker Hub.
Install and run:
-
Pull and run the vLLora MCP server via Docker:
docker run -d --name vllora-mcp -p 9090:9090 -p 9091:9091 vllora/vllora
-
Verify the server is running:
docker ps | grep vllora
-
Open the UI and configure API keys:
- Visit http://localhost:9091 to configure your OpenAI-compatible provider API keys.
-
Optional: If you prefer building locally, you can clone the repository and build the Rust project (requires Rust toolchain):
git clone https://github.com/vllora/vllora.git cd vllora cargo build --release ./target/release/vlora
Then access http://localhost:9090 (API) and http://localhost:9091 (UI).
Additional notes
Tips and considerations:
- The UI at /9091 is used to configure provider API keys and view traces. Ensure your OpenAI-compatible API keys are set up there for tracing to begin.
- By default, vLLora exposes HTTP endpoints for MCP-style interactions. You can integrate with other MCP-compatible tooling via HTTP/SSE streams.
- If you run behind a firewall or reverse proxy, ensure that ports 9090 (API) and 9091 (UI) are accessible from your development environment.
- The Docker image vllora/vllora may be updated over time; pin to a specific tag if version stability is required, e.g., docker run vllora/vllora:1.0.0.
- Review the docs at the official site for updated MCP integration guidance and any platform-specific notes.
Related MCP Servers
LibreChat
Enhanced ChatGPT Clone: Features Agents, MCP, DeepSeek, Anthropic, AWS, OpenAI, Responses API, Azure, Groq, o1, GPT-5, Mistral, OpenRouter, Vertex AI, Gemini, Artifacts, AI model switching, message search, Code Interpreter, langchain, DALL-E-3, OpenAPI Actions, Functions, Secure Multi-User Auth, Presets, open-source for self-hosting. Active.
repomix
📦 Repomix is a powerful tool that packs your entire repository into a single, AI-friendly file. Perfect for when you need to feed your codebase to Large Language Models (LLMs) or other AI tools like Claude, ChatGPT, DeepSeek, Perplexity, Gemini, Gemma, Llama, Grok, and more.
sre
The SmythOS Runtime Environment (SRE) is an open-source, cloud-native runtime for agentic AI. Secure, modular, and production-ready, it lets developers build, run, and manage intelligent agents across local, cloud, and edge environments.
weam
Web app for teams of 20+ members. In-built connections to major LLMs via API. Share chats, prompts, and agents in team or private folders. Modern, fully responsive stack (Next.js, Node.js). Deploy your own vibe-coded AI apps, agents, or workflows—or use ready-made solutions from the library.
openapi
OpenAPI definitions, converters and LLM function calling schema composer.
coplay-unity-plugin
Unity plugin for Coplay