scrape-do-automation
Scannednpx machina-cli add skill ComposioHQ/awesome-claude-skills/scrape-do-automation --openclawScrape Do Automation via Rube MCP
Automate Scrape Do operations through Composio's Scrape Do toolkit via Rube MCP.
Toolkit docs: composio.dev/toolkits/scrape_do
Prerequisites
- Rube MCP must be connected (RUBE_SEARCH_TOOLS available)
- Active Scrape Do connection via
RUBE_MANAGE_CONNECTIONSwith toolkitscrape_do - Always call
RUBE_SEARCH_TOOLSfirst to get current tool schemas
Setup
Get Rube MCP: Add https://rube.app/mcp as an MCP server in your client configuration. No API keys needed — just add the endpoint and it works.
- Verify Rube MCP is available by confirming
RUBE_SEARCH_TOOLSresponds - Call
RUBE_MANAGE_CONNECTIONSwith toolkitscrape_do - If connection is not ACTIVE, follow the returned auth link to complete setup
- Confirm connection status shows ACTIVE before running any workflows
Tool Discovery
Always discover available tools before executing workflows:
RUBE_SEARCH_TOOLS
queries: [{use_case: "Scrape Do operations", known_fields: ""}]
session: {generate_id: true}
This returns available tool slugs, input schemas, recommended execution plans, and known pitfalls.
Core Workflow Pattern
Step 1: Discover Available Tools
RUBE_SEARCH_TOOLS
queries: [{use_case: "your specific Scrape Do task"}]
session: {id: "existing_session_id"}
Step 2: Check Connection
RUBE_MANAGE_CONNECTIONS
toolkits: ["scrape_do"]
session_id: "your_session_id"
Step 3: Execute Tools
RUBE_MULTI_EXECUTE_TOOL
tools: [{
tool_slug: "TOOL_SLUG_FROM_SEARCH",
arguments: {/* schema-compliant args from search results */}
}]
memory: {}
session_id: "your_session_id"
Known Pitfalls
- Always search first: Tool schemas change. Never hardcode tool slugs or arguments without calling
RUBE_SEARCH_TOOLS - Check connection: Verify
RUBE_MANAGE_CONNECTIONSshows ACTIVE status before executing tools - Schema compliance: Use exact field names and types from the search results
- Memory parameter: Always include
memoryinRUBE_MULTI_EXECUTE_TOOLcalls, even if empty ({}) - Session reuse: Reuse session IDs within a workflow. Generate new ones for new workflows
- Pagination: Check responses for pagination tokens and continue fetching until complete
Quick Reference
| Operation | Approach |
|---|---|
| Find tools | RUBE_SEARCH_TOOLS with Scrape Do-specific use case |
| Connect | RUBE_MANAGE_CONNECTIONS with toolkit scrape_do |
| Execute | RUBE_MULTI_EXECUTE_TOOL with discovered tool slugs |
| Bulk ops | RUBE_REMOTE_WORKBENCH with run_composio_tool() |
| Full schema | RUBE_GET_TOOL_SCHEMAS for tools with schemaRef |
Powered by Composio
Source
git clone https://github.com/ComposioHQ/awesome-claude-skills/blob/master/composio-skills/scrape-do-automation/SKILL.mdView on GitHub Overview
This skill automates Scrape Do operations through Composio's Scrape Do toolkit using Rube MCP. It emphasizes fetching current tool schemas with RUBE_SEARCH_TOOLS before execution, then guiding you through discovery, connection validation, and tool execution to ensure reliable workflows.
How This Skill Works
Technically, you discover available tools with RUBE_SEARCH_TOOLS to obtain slugs and input schemas, verify the Scrape Do connection with RUBE_MANAGE_CONNECTIONS and ensure ACTIVE, then execute the chosen tool using RUBE_MULTI_EXECUTE_TOOL with the discovered slug and schema-compliant arguments. Always include a memory object and a valid session_id, and rely on fresh tool schemas for accurate execution.
When to Use It
- When you need up-to-date Scrape Do tool schemas before running a task.
- When setting up or validating the Rube MCP connection for the Scrape Do toolkit.
- When orchestrating a Scrape Do workflow: discover, connect, then execute.
- When running multiple tools in a single workflow and reusing a session ID.
- When tool schemas change or you must handle pagination by refreshing tool data.
Quick Start
- Step 1: Discover available Scrape Do tools with RUBE_SEARCH_TOOLS for current schemas.
- Step 2: Check the Scrape Do connection with RUBE_MANAGE_CONNECTIONS and ensure ACTIVE.
- Step 3: Execute a tool via RUBE_MULTI_EXECUTE_TOOL using the discovered slug, proper arguments, memory: {}, and a session_id.
Best Practices
- Always call RUBE_SEARCH_TOOLS before execution to get current tool slugs.
- Verify RUBE_MANAGE_CONNECTIONS shows ACTIVE before running tools.
- Use exact field names and types from the discovered tool schemas.
- Include memory in RUBE_MULTI_EXECUTE_TOOL calls, even if empty ({}).
- Reuse session IDs within a workflow; generate new IDs for new workflows; check for pagination tokens.
Example Use Cases
- Discover Scrape Do tools with RUBE_SEARCH_TOOLS and run a chosen task using the discovered slug.
- Add the MCP server https://rube.app/mcp, then validate the connection is ACTIVE with RUBE_MANAGE_CONNECTIONS.
- Execute a Scrape Do tool by passing the schema-compliant arguments obtained from the discovery results.
- Run several Scrape Do tools in one session by reusing the same session_id and including a memory object.
- If tool schemas change, re-run RUBE_SEARCH_TOOLS to refresh slugs and schemas before execution.