Get the FREE Ultimate OpenClaw Setup Guide →

scrape-do-automation

Scanned
npx machina-cli add skill ComposioHQ/awesome-claude-skills/scrape-do-automation --openclaw
Files (1)
SKILL.md
2.9 KB

Scrape Do Automation via Rube MCP

Automate Scrape Do operations through Composio's Scrape Do toolkit via Rube MCP.

Toolkit docs: composio.dev/toolkits/scrape_do

Prerequisites

  • Rube MCP must be connected (RUBE_SEARCH_TOOLS available)
  • Active Scrape Do connection via RUBE_MANAGE_CONNECTIONS with toolkit scrape_do
  • Always call RUBE_SEARCH_TOOLS first to get current tool schemas

Setup

Get Rube MCP: Add https://rube.app/mcp as an MCP server in your client configuration. No API keys needed — just add the endpoint and it works.

  1. Verify Rube MCP is available by confirming RUBE_SEARCH_TOOLS responds
  2. Call RUBE_MANAGE_CONNECTIONS with toolkit scrape_do
  3. If connection is not ACTIVE, follow the returned auth link to complete setup
  4. Confirm connection status shows ACTIVE before running any workflows

Tool Discovery

Always discover available tools before executing workflows:

RUBE_SEARCH_TOOLS
queries: [{use_case: "Scrape Do operations", known_fields: ""}]
session: {generate_id: true}

This returns available tool slugs, input schemas, recommended execution plans, and known pitfalls.

Core Workflow Pattern

Step 1: Discover Available Tools

RUBE_SEARCH_TOOLS
queries: [{use_case: "your specific Scrape Do task"}]
session: {id: "existing_session_id"}

Step 2: Check Connection

RUBE_MANAGE_CONNECTIONS
toolkits: ["scrape_do"]
session_id: "your_session_id"

Step 3: Execute Tools

RUBE_MULTI_EXECUTE_TOOL
tools: [{
  tool_slug: "TOOL_SLUG_FROM_SEARCH",
  arguments: {/* schema-compliant args from search results */}
}]
memory: {}
session_id: "your_session_id"

Known Pitfalls

  • Always search first: Tool schemas change. Never hardcode tool slugs or arguments without calling RUBE_SEARCH_TOOLS
  • Check connection: Verify RUBE_MANAGE_CONNECTIONS shows ACTIVE status before executing tools
  • Schema compliance: Use exact field names and types from the search results
  • Memory parameter: Always include memory in RUBE_MULTI_EXECUTE_TOOL calls, even if empty ({})
  • Session reuse: Reuse session IDs within a workflow. Generate new ones for new workflows
  • Pagination: Check responses for pagination tokens and continue fetching until complete

Quick Reference

OperationApproach
Find toolsRUBE_SEARCH_TOOLS with Scrape Do-specific use case
ConnectRUBE_MANAGE_CONNECTIONS with toolkit scrape_do
ExecuteRUBE_MULTI_EXECUTE_TOOL with discovered tool slugs
Bulk opsRUBE_REMOTE_WORKBENCH with run_composio_tool()
Full schemaRUBE_GET_TOOL_SCHEMAS for tools with schemaRef

Powered by Composio

Source

git clone https://github.com/ComposioHQ/awesome-claude-skills/blob/master/composio-skills/scrape-do-automation/SKILL.mdView on GitHub

Overview

This skill automates Scrape Do operations through Composio's Scrape Do toolkit using Rube MCP. It emphasizes fetching current tool schemas with RUBE_SEARCH_TOOLS before execution, then guiding you through discovery, connection validation, and tool execution to ensure reliable workflows.

How This Skill Works

Technically, you discover available tools with RUBE_SEARCH_TOOLS to obtain slugs and input schemas, verify the Scrape Do connection with RUBE_MANAGE_CONNECTIONS and ensure ACTIVE, then execute the chosen tool using RUBE_MULTI_EXECUTE_TOOL with the discovered slug and schema-compliant arguments. Always include a memory object and a valid session_id, and rely on fresh tool schemas for accurate execution.

When to Use It

  • When you need up-to-date Scrape Do tool schemas before running a task.
  • When setting up or validating the Rube MCP connection for the Scrape Do toolkit.
  • When orchestrating a Scrape Do workflow: discover, connect, then execute.
  • When running multiple tools in a single workflow and reusing a session ID.
  • When tool schemas change or you must handle pagination by refreshing tool data.

Quick Start

  1. Step 1: Discover available Scrape Do tools with RUBE_SEARCH_TOOLS for current schemas.
  2. Step 2: Check the Scrape Do connection with RUBE_MANAGE_CONNECTIONS and ensure ACTIVE.
  3. Step 3: Execute a tool via RUBE_MULTI_EXECUTE_TOOL using the discovered slug, proper arguments, memory: {}, and a session_id.

Best Practices

  • Always call RUBE_SEARCH_TOOLS before execution to get current tool slugs.
  • Verify RUBE_MANAGE_CONNECTIONS shows ACTIVE before running tools.
  • Use exact field names and types from the discovered tool schemas.
  • Include memory in RUBE_MULTI_EXECUTE_TOOL calls, even if empty ({}).
  • Reuse session IDs within a workflow; generate new IDs for new workflows; check for pagination tokens.

Example Use Cases

  • Discover Scrape Do tools with RUBE_SEARCH_TOOLS and run a chosen task using the discovered slug.
  • Add the MCP server https://rube.app/mcp, then validate the connection is ACTIVE with RUBE_MANAGE_CONNECTIONS.
  • Execute a Scrape Do tool by passing the schema-compliant arguments obtained from the discovery results.
  • Run several Scrape Do tools in one session by reusing the same session_id and including a memory object.
  • If tool schemas change, re-run RUBE_SEARCH_TOOLS to refresh slugs and schemas before execution.

Frequently Asked Questions

Add this skill to your agents
Sponsor this space

Reach thousands of developers