flowllm
FlowLLM: Simplifying LLM-based HTTP/MCP Service Development
claude mcp add --transport stdio flowllm-ai-flowllm python -m flowllm \ --env FLOW_LLM_API_KEY="your-api-key-here" \ --env FLOW_LLM_BASE_URL="https://api.example.com/v1" \ --env FLOW_EMBEDDING_API_KEY="your-embedding-api-key-here" \ --env FLOW_EMBEDDING_BASE_URL="https://embed.example.com/v1"
How to use
FlowLLM provides HTTP and MCP (Model Context Protocol) services built around LLMs, embeddings, and vector stores. It enables you to expose LLM-powered capabilities as MCP tools that MCP-compatible clients can discover and invoke, alongside a REST API for general usage. FlowLLM auto-generates services so you can run a single MCP-enabled endpoint and test flows via a minimal setup. Once started, you can register new Ops, compose them into flows, and serve them through MCP tools or HTTP endpoints. The included token counting support helps you estimate usage, and the architecture supports both synchronous JSON HTTP responses and streaming responses where appropriate.
How to install
Prerequisites:
- Python 3.10+
- Git
- Internet access to install dependencies
Install from PyPI (recommended):
pip install flowllm
Install from source (editable mode):
git clone https://github.com/flowllm-ai/flowllm.git
cd flowllm
pip install -e .
Configure environment variables (examples):
# API keys and base URLs
export FLOW_LLM_API_KEY=your-api-key-here
export FLOW_LLM_BASE_URL=https://api.example.com/v1
export FLOW_EMBEDDING_API_KEY=your-embedding-api-key-here
export FLOW_EMBEDDING_BASE_URL=https://embed.example.com/v1
Run the MCP server:
# Run via the module entry point (as configured in mcp_config)
python -m flowllm
Optional: for development, install dependencies and run tests if provided in the repo.
Additional notes
Tips and considerations:
- Ensure your API keys and base URLs are correctly configured in the environment to enable LLM and embedding services.
- The MCP service will register as tools for MCP-compatible clients; you can discover and invoke tools defined in your FlowLLM configuration.
- If you plan heavy concurrent usage, consider configuring token_count and backends (base, OpenAI, HF) as described in the project docs to get accurate usage metrics.
- For debugging, FlowLLM supports a CMD/service mode to run a single Op from the command line; this is useful for quick tests of a specific operation.
- Check the project docs for guidance on YAML configuration of flows, input/output schemas, and how to compose Ops with serial (>>) and parallel (|) operators.
Related MCP Servers
PPTAgent
An Agentic Framework for Reflective PowerPoint Generation
AgentChat
AgentChat 是一个基于 LLM 的智能体交流平台,内置默认 Agent 并支持用户自定义 Agent。通过多轮对话和任务协作,Agent 可以理解并协助完成复杂任务。项目集成 LangChain、Function Call、MCP 协议、RAG、Memory、Milvus 和 ElasticSearch 等技术,实现高效的知识检索与工具调用,使用 FastAPI 构建高性能后端服务。
mcp-aktools
📈 提供股票、加密货币的数据查询和分析功能MCP服务器
python -client
支持查询主流agent框架技术文档的MCP server(支持stdio和sse两种传输协议), 支持 langchain、llama-index、autogen、agno、openai-agents-sdk、mcp-doc、camel-ai 和 crew-ai
ultimate_mcp_server
Comprehensive MCP server exposing dozens of capabilities to AI agents: multi-provider LLM delegation, browser automation, document processing, vector ops, and cognitive memory systems
awsome_kali_MCPServers
awsome kali MCPServers is a set of MCP servers tailored for Kali Linux