mcp -funasr
MCPServer is a Python-based server that leverages Alibaba's FunASR library to provide speech processing services through the FastMCP framework.
claude mcp add --transport stdio radial-hks-mcp-server-funasr uvx radial-hks-mcp-server-funasr
How to use
This MCP server exposes FunASR-powered speech processing capabilities via the FastMCP framework. It supports audio validation, asynchronous speech transcription, and voice activity detection (VAD), with multi-model support and dynamic loading of ASR and VAD models. You can validate audio files to ensure they are readable and correctly formatted, start asynchronous transcription tasks with optional per-task model selection and generation parameters, and query for task status or retrieve results. The server is designed to load models on demand and can switch between different ASR and VAD models as needed for different requests.
How to install
Prerequisites:
- Python 3.8+
- pip
- Clone the repository or navigate to the MCPServer directory that contains this README and the server code.
- Create a virtual environment and activate it:
python -m venv .venv source .venv/bin/activate # On Windows: .venv\Scripts\activate - Install dependencies:
This installs FastMCP, FunASR, and their dependencies. If you need a specific PyTorch setup (e.g., CUDA-enabled), install PyTorch manually prior to running the server as per the official PyTorch instructions.pip install -r requirements.txt - Run the server:
The server will start and be accessible at http://0.0.0.0:9000. On first run, FunASR will download default ASR and VAD models, which may take some time.uvicorn main:app --host 0.0.0.0 --port 9000
Additional notes
Tips and notes:
- Environment variable: MODELSCOPE_API_TOKEN is optional but may be required if you access private models or face rate limits. Set it in your environment if needed, e.g., export MODELSCOPE_API_TOKEN="YOUR_TOKEN_HERE".
- The server supports per-request model selection for transcription (model_name) and per-request generation parameters (model_generate_kwargs). Models can be loaded or switched dynamically.
- If you encounter large model downloads, ensure the network allows access to ModelScope and related model repositories.
- The default ASR/VAD models are downloaded on the first run. You can explicitly configure which models to load by adjusting server configuration or environment variables if supported by your deployment.
- For GPU-enabled environments, ensure PyTorch with CUDA is installed prior to running the server for optimal performance.
Related MCP Servers
nautex
MCP server for guiding Coding Agents via end-to-end requirements to implementation plan pipeline
mcp-yfinance
Real-time stock API with Python, MCP server example, yfinance stock analysis dashboard
pfsense
pfSense MCP Server enables security administrators to manage their pfSense firewalls using natural language through AI assistants like Claude Desktop. Simply ask "Show me blocked IPs" or "Run a PCI compliance check" instead of navigating complex interfaces. Supports REST/XML-RPC/SSH connections, and includes built-in complian
cloudwatch-logs
MCP server from serkanh/cloudwatch-logs-mcp
servicenow-api
ServiceNow MCP Server and API Wrapper
the -company
TheMCPCompany: Creating General-purpose Agents with Task-specific Tools