Skill_Seekers
Convert documentation websites, GitHub repositories, and PDFs into Claude AI skills with automatic conflict detection
claude mcp add --transport stdio yusufkaraaslan-skill_seekers python -m skill_seekers \ --env PYTHONUNBUFFERED="1"
How to use
Skill Seekers acts as the data preprocessing layer for AI systems. It ingests documentation, GitHub repositories, PDFs, and videos, then analyzes and structures the content into reusable knowledge assets and SKILL.md files that can power AI skills across platforms. As an MCP server, you can run Skill Seekers locally and expose its capabilities to your MCP workflow, enabling other agents or tooling in your pipeline to request packaged outputs (Claude, Gemini, OpenAI, LangChain documents, etc.). The toolset focuses on extracting, organizing, and exporting data in multiple target formats so you can feed downstream AI systems without re-scraping sources.
Once running, you can leverage its common commands to create and package knowledge assets. For example, you can create a knowledge asset from a source, then package it for multiple targets (Claude, LangChain, LlamaIndex, etc.). This enables end-to-end preparation from raw sources to deployment-ready outputs that integrate with your RAG pipelines, IDEs, or AI coding assistants.
How to install
Prerequisites:
- Python 3.10+ installed on your system
- Access to install Python packages (pip)
Installation and setup:
- Create a virtual environment (optional but recommended):
python -m venv venv
source venv/bin/activate # on macOS/Linux
venv\Scripts\activate # on Windows
- Install Skill Seekers from PyPI:
pip install skill-seekers
- Run the MCP server for Skill Seekers:
python -m skill_seekers
- (Optional) Create a dedicated config file or environment variables if you run multiple MCP servers in one environment.
# Example: run with a custom port or settings via environment variables as needed
export SKILL_SEEKERS_CONFIG=/path/to/config.json
Additional notes
Tips and caveats:
- The Skill Seekers package focuses on converting diverse sources into structured outputs suitable for multiple AI targets (Claude, Gemini, OpenAI, LangChain, LlamaIndex, etc.).
- If running behind a corporate proxy, ensure HTTP(S)_PROXY environment variables are set for pip and Python network access.
- When packaging, you can export to several targets in one go to speed up integration workflows (see CLI docs in the project for exact commands).
- For best results, pin Python dependencies in a virtual environment to avoid compatibility issues across projects.
- If you encounter performance issues with large sources, consider preprocessing in chunks and using the packaging steps to produce smaller, incremental outputs.
Related MCP Servers
claude-codex-settings
My personal Claude Code and OpenAI Codex setup with battle-tested skills, commands, hooks, agents and MCP servers that I use daily.
mcp-n8n-workflow-builder
AI-powered n8n workflow automation through natural language. MCP server enabling Claude AI & Cursor IDE to create, manage, and monitor workflows via Model Context Protocol. Multi-instance support, 17 tools, comprehensive docs. Build workflows conversationally without manual JSON editing.
mcp-batchit
🚀 MCP aggregator for batching multiple tool calls into a single request. Reduces overhead, saves tokens, and simplifies complex operations in AI agent workflows.
mcp-mail
📧 MCP Mail Tool - AI-powered email management tool | 基于 MCP 的智能邮件管理工具
review-flow
Automated AI code reviews powered — webhook-driven, real-time dashboard, MCP integration, smart queue with deduplication, multi-agent audits, and iterative follow-up reviews for GitLab MRs and GitHub PRs
mcp-ragex
MCP server for intelligent code search: semantic (RAG), symbolic (tree-sitter), and regex (ripgrep) search modes. Built for Claude Code and AI coding assistants.