langgraph-multiagent
Scannednpx machina-cli add skill BjornMelin/dev-skills/langgraph-multiagent --openclawLangGraph Multi-Agent
What this skill does
Build, review, and modernize production-grade multi-agent systems with LangGraph/LangChain while staying version-accurate by default: always resolve the current APIs from docs + installed versions, and fall back to opensrc/ source snapshots for under-the-hood edge cases.
Default operating rules (do these every time)
- Treat all LangGraph/LangChain APIs as versioned; never “code from memory” for any API surface that might have changed.
- Establish ground truth before coding:
- Determine installed/pinned versions (repo lockfiles +
importlib.metadata). - Query official docs via
langchain-docs.SearchDocsByLangChain. - Use Context7 for canonical API references and examples.
- Determine installed/pinned versions (repo lockfiles +
- When docs are ambiguous or behavior is subtle, inspect dependency internals via
opensrc/(read-only):- Run
npx opensrc pypi:<package>@<version> --modify=false - Check
opensrc/sources.jsonand cite exactopensrc/...paths + versions in your writeup.
- Run
Use references/research_playbook.md when you need a rigorous doc-sweep, including llms.txt-driven crawling of LangGraph docs.
Workflow decision tree
- Build something new → follow Build workflow.
- Audit/deprecations/migration → follow Audit & migrate workflow.
- Latency/cost/scale issues → follow Performance & scale workflow.
Quick start (repo-aware)
- Generate a deprecations + framework audit report:
- Run
python scripts/audit_repo_agents.py --root .(from the skill folder) - Optional:
python scripts/audit_repo_agents.py --root . --json agent_audit.json - Optional:
python scripts/generate_migration_plan.py --audit-json agent_audit.json --out migration_plan.md
- Run
- If you need “latest docs” for a specific area, start with:
langchain-docs.SearchDocsByLangChainusing queries fromreferences/docs_index.md
- If you hit a behavior edge case, snapshot internals:
- Run
python scripts/opensrc_snapshot.py --packages langgraph langchain langchain-core(from the skill folder)
- Run
- If you need a doc sitemap for LangGraph (for agentic RAG or doc crawling):
- Run
python scripts/fetch_llms_txt_urls.py --print --unique(from the skill folder)
- Run
- If you need to build an offline docs cache (bounded crawl):
- Use seeds in
references/doc_crawl_targets.md - Run
python scripts/crawl_docs.py --llms-txt https://langchain-ai.github.io/langgraph/llms.txt --allow-prefixes https://langchain-ai.github.io/langgraph/ --out-dir docs_cache_langgraph(from the skill folder)
- Use seeds in
Reference map (use these, don’t guess)
references/langchain_create_agent_middleware.md:create_agent+ middleware hooks + migration mappings.references/langchain_multiagent_handoffs.md: choosing between subagents vs handoffs vs multi-node subgraphs.references/langgraph_graph_api_primitives.md: reducers, Send API, subgraphs, and common error codes.references/memory_and_context_engineering.md: state vs store vs runtime context; memory design.references/mcp_integration_patterns.md: MCP client/interceptors and ToolRuntime-driven auth/DI.references/testing_evaluation.md: tests + eval strategy for safe migrations.references/deployment_agent_server.md:langgraph.jsonand Agent Server deployment basics.references/security_threat_model.md: threat model + mitigations for tool calling systems.references/upgrades_and_versioning.md: repeatable upgrade process.references/audit_and_migration_methodology.md: end-to-end audit→plan→execute methodology.references/api_map_python.md: import-path cheat sheet (verify for your version).references/ui_nextjs_rsc.md: Next.js App Router (RSC) + React UI integration (Agent Server +useStream).references/ui_nextjs_ai_sdk.md: Next.js App Router using AI SDK v6 (useChat) + Streamdown (alternative UI stack).references/ui_fastapi_backend.md: FastAPI backend patterns for in-process agents or Agent Server BFF/proxy.references/ui_streaming_protocol.md: simple SSE event protocol for custom UI backends.references/ui_streamlit.md: Streamlit UI integration patterns (streaming + HITL).
Build workflow (LangChain v1 + LangGraph v1+)
- Select the right multi-agent topology (start simple):
- Supervisor + subagents (tool-calling): a main “supervisor” calls subagents as tools for context isolation.
- Orchestrator-worker (fan-out): use LangGraph’s worker primitives (e.g., Send-style fanout) for parallelizable tasks; use reducers to avoid concurrent state-update errors.
- Deterministic workflow + agent nodes: keep control-flow explicit; use agents only where needed.
- Define strict tool schemas and failure behavior:
- Make tools idempotent where possible; add timeouts and retries; “fail open” on non-critical model/rerank steps.
- Do context engineering explicitly:
- Use typed state for dynamic runtime context; keep “store” for long-term memory; use runtime context injection for user/org-scoped dependencies.
- Add safety controls early:
- Use middleware guardrails and human-in-the-loop for high-stakes tools (payments, outbound emails, destructive actions).
- Add observability before scaling:
- Enable tracing (LangSmith/OpenTelemetry), record tool latency/cost, and add regression evaluations.
Use references/patterns.md for design templates and “gotchas” per topology.
Audit & migrate workflow (deprecated patterns → modern stack)
- Inventory current architecture (dependencies + imports + runtime behavior):
- Run
scripts/audit_repo_agents.pyand expand it with repo-specific patterns if needed.
- Run
- Identify deprecated patterns and their replacements using official docs:
- Prioritize the LangGraph v1 and LangChain v1 migration guides.
- Verify whether
langgraph.prebuilt.create_react_agentusage should move tolangchain.agents.create_agent.
- Plan the migration in slices (keep it shippable):
- Preserve external behavior; migrate internals; then upgrade prompts/tools; finally tighten types and add tests.
- Execute migration and harden:
- Replace legacy agent frameworks (LlamaIndex agents, CrewAI, Agno, OpenAI Agents) with LangGraph/LangChain equivalents.
- Add offline tests with fixtures/mocks, and add trace-based regression checks when possible.
Use references/migration_supervisor.md and references/migration_other_frameworks.md for playbooks.
Performance & scale workflow
- Add measurement first:
- Capture per-node latency, token usage, tool call counts, cache hit rates, and error taxonomy.
- Apply the “cheapest win” stack in order:
- Reduce context size (summaries, retrieval, subagent isolation).
- Cache tool results (deterministic and keyed); add request coalescing.
- Parallelize only where state updates are safe (reducers, fanout collection keys).
- Use smaller/faster models for routing/validation; reserve larger models for synthesis.
- Constrain worst-case behavior:
- Recursion limits, max steps, max tool calls, and circuit breakers for flaky dependencies.
Use references/performance_cost.md for checklists and tactics.
Bundled resources in this skill
scripts/: deterministic helpers (audits,llms.txtextraction,opensrcsnapshot runner)references/: playbooks + templates for docs research, patterns, migrations, safety, observability, and performanceassets/: copy/paste templates for Python + UI integrations (Next.js, FastAPI, Streamlit)
Start with references/docs_index.md and references/research_playbook.md.
Source
git clone https://github.com/BjornMelin/dev-skills/blob/main/skills/langgraph-multiagent/SKILL.mdView on GitHub Overview
Architect-level toolkit to build, audit, and migrate multi-agent systems with LangGraph v1+ and LangChain v1+. It covers supervisor/subagent architectures, orchestrator-worker workflows, memory and context engineering, guardrails, human-in-the-loop, observability, and deployment—plus migration from deprecated patterns and libs.
How This Skill Works
The skill enforces API version discipline by always resolving current LangGraph/LangChain APIs from official docs and installed versions. It establishes ground truth via pinned versions, repo lockfiles, importlib.metadata, and canonical references (Context7), and, for edge cases, inspects internals using opensrc to cite exact paths and versions in evidence-backed writeups.
When to Use It
- Building or refactoring supervisor/subagent architectures
- Implementing orchestrator-worker workflows with routing and handoffs
- Developing agentic RAG, short/long-term memory, and state/context engineering
- Adding guardrails, human-in-the-loop, MCP tool integration, observability
- Migrating away from deprecated patterns and libraries (e.g., langgraph.prebuilt.create_react_agent, langgraph-supervisor(-py), LlamaIndex agents)
Quick Start
- Step 1: Run audit to generate a deprecations + framework audit report: python scripts/audit_repo_agents.py --root .
- Step 2: Optionally export audit JSON: python scripts/audit_repo_agents.py --root . --json agent_audit.json
- Step 3: Generate a migration plan from the audit: python scripts/generate_migration_plan.py --audit-json agent_audit.json --out migration_plan.md
Best Practices
- Treat LangGraph/LangChain APIs as versioned; do not code surface areas from memory
- Always establish ground truth: check installed/pinned versions via lockfiles and importlib.metadata
- Query official docs with langchain-docs.SearchDocsByLangChain and use Context7 for canonical references
- When docs are ambiguous, inspect internals with opensrc and cite exact opensrc paths/versions
- Use a rigorous doc-sweep and migration plan workflow for audits and migrations
Example Use Cases
- Refactor a supervisor/subagent system to LangGraph v1+ with LangChain v1+ for improved alignment with docs
- Implement agentic RAG with short and long-term memory, ensuring consistent state management
- Introduce guardrails and human-in-the-loop controls in high-stakes workflows
- Migrate from deprecated patterns like langgraph.prebuilt.create_react_agent and related libs
- Enhance observability with LangSmith/OpenTelemetry integration during deployment and runtime