Get the FREE Ultimate OpenClaw Setup Guide →

gemini-batch

npx machina-cli add skill akrindev/google-studio-skills/gemini-batch --openclaw
Files (1)
SKILL.md
11.5 KB

Gemini Batch Processing

Process large volumes of requests efficiently using Gemini Batch API through executable scripts for cost savings and high throughput.

When to Use This Skill

Use this skill when you need to:

  • Process hundreds/thousands of requests
  • Generate content in bulk (blogs, emails, descriptions)
  • Reduce costs for high-volume tasks
  • Run async jobs without blocking
  • Process large datasets with AI
  • Generate multiple documents at once
  • Create scalable content pipelines
  • Process requests that don't need real-time responses

Available Scripts

scripts/create_batch.py

Purpose: Create a batch job from a JSONL file

When to use:

  • Starting any batch processing task
  • Uploading multiple requests for processing
  • Creating async jobs for large workloads

Key parameters:

ParameterDescriptionExample
input_fileJSONL file path (required)requests.jsonl
--model, -mModel to usegemini-3-flash-preview
--name, -nDisplay name for job"my-batch-job"

Output: Job name/ID to track with check_status.py

scripts/check_status.py

Purpose: Monitor batch job progress and completion

When to use:

  • Checking if a batch job is complete
  • Polling job status until finished
  • Monitoring async job execution

Key parameters:

ParameterDescriptionExample
job_nameBatch job name/ID (required)batches/abc123
--wait, -wPoll until completionFlag

Output: Job status and final state

scripts/get_results.py

Purpose: Retrieve completed batch job results

When to use:

  • Downloading completed batch results
  • Parsing batch job output
  • Extracting generated content

Key parameters:

ParameterDescriptionExample
job_nameBatch job name/ID (required)batches/abc123
--output, -oOutput file pathresults.jsonl

Output: Results content or file

Workflows

Workflow 1: Basic Batch Processing

# 1. Create JSONL file
echo '{"key": "req1", "request": {"contents": [{"parts": [{"text": "Explain photosynthesis"}]}]}}' > requests.jsonl
echo '{"key": "req2", "request": {"contents": [{"parts": [{"text": "What is gravity?"}]}}]}' >> requests.jsonl

# 2. Create batch job
python scripts/create_batch.py requests.jsonl --name "science-questions"

# 3. Check status
python scripts/check_status.py <job-name> --wait

# 4. Get results
python scripts/get_results.py <job-name> --output results.jsonl
  • Best for: Basic bulk processing, cost efficiency
  • Typical time: Minutes to hours depending on job size

Workflow 2: Bulk Content Generation

# 1. Generate JSONL with content requests
python3 << 'EOF'
import json

topics = ["sustainable energy", "AI in healthcare", "space exploration"]
with open("content-requests.jsonl", "w") as f:
    for i, topic in enumerate(topics):
        req = {
            "key": f"blog-{i}",
            "request": {
                "contents": [{
                    "parts": [{
                        "text": f"Write a 500-word blog post about {topic}"
                    }]
                }]
            }
        }
        f.write(json.dumps(req) + "\n")
EOF

# 2. Process batch
python scripts/create_batch.py content-requests.jsonl --name "blog-posts" --model gemini-3-flash-preview
python scripts/check_status.py <job-name> --wait
python scripts/get_results.py <job-name> --output blog-posts.jsonl
  • Best for: Blog generation, article creation, bulk writing
  • Combines with: gemini-text for content needs

Workflow 3: Dataset Processing

# 1. Load dataset and create batch requests
python3 << 'EOF'
import json

# Your dataset
data = [
    {"product": "laptop", "features": ["fast", "lightweight"]},
    {"product": "headphones", "features": ["wireless", "noise-cancelling"]},
]

with open("product-descriptions.jsonl", "w") as f:
    for item in data:
        features = ", ".join(item["features"])
        prompt = f"Write a product description for {item['product']} with these features: {features}"
        req = {
            "key": item["product"],
            "request": {
                "contents": [{"parts": [{"text": prompt}]}]
            }
        }
        f.write(json.dumps(req) + "\n")
EOF

# 2. Process
python scripts/create_batch.py product-descriptions.jsonl
python scripts/check_status.py <job-name> --wait
python scripts/get_results.py <job-name> --output results.jsonl
  • Best for: Product descriptions, dataset enrichment, bulk analysis

Workflow 4: Email Campaign Generation

# 1. Create personalized email requests
python3 << 'EOF'
import json

customers = [
    {"name": "Alice", "product": "premium plan"},
    {"name": "Bob", "product": "basic plan"},
]

with open("emails.jsonl", "w") as f:
    for cust in customers:
        prompt = f"Write a personalized email to {cust['name']} about upgrading to our {cust['product']}"
        req = {
            "key": f"email-{cust['name'].lower()}",
            "request": {
                "contents": [{"parts": [{"text": prompt}]}]
            }
        }
        f.write(json.dumps(req) + "\n")
EOF

# 2. Process batch
python scripts/create_batch.py emails.jsonl --name "email-campaign"
python scripts/check_status.py <job-name> --wait
python scripts/get_results.py <job-name> --output email-results.jsonl
  • Best for: Marketing campaigns, personalized outreach
  • Combines with: gemini-text for email content

Workflow 5: Async Job Monitoring

# 1. Create job
python scripts/create_batch.py large-batch.jsonl --name "big-job"

# 2. Check status periodically (non-blocking)
while true; do
    python scripts/check_status.py <job-name>
    sleep 60  # Check every minute
done

# 3. Get results when done
python scripts/get_results.py <job-name> --output final-results.jsonl
  • Best for: Long-running jobs, background processing
  • Use when: You don't need immediate results

Workflow 6: Cost-Optimized Bulk Processing

# 1. Use flash model for cost efficiency
python scripts/create_batch.py requests.jsonl --model gemini-3-flash-preview --name "cost-optimized"

# 2. Monitor and retrieve
python scripts/check_status.py <job-name> --wait
python scripts/get_results.py <job-name>
  • Best for: High-volume, cost-sensitive applications
  • Savings: Batch API typically 50%+ cheaper than real-time

Workflow 7: Multi-Stage Pipeline

# Stage 1: Generate content
python scripts/create_batch.py content-requests.jsonl --name "stage1-content"
python scripts/check_status.py <job1> --wait

# Stage 2: Summarize content
python scripts/create_batch.py summaries.jsonl --name "stage2-summaries"
python scripts/check_status.py <job2> --wait

# Stage 3: Convert to audio (gemini-tts)
# Process results from stage 2
  • Best for: Complex workflows, multi-step processing
  • Combines with: Other Gemini skills for complete pipelines

Parameters Reference

JSONL Format

Each line is a separate JSON object:

{
  "key": "unique-identifier",
  "request": {
    "contents": [
      {
        "parts": [
          {
            "text": "Your prompt here"
          }
        ]
      }
    ]
  }
}

Model Selection

ModelBest ForCostSpeed
gemini-3-flash-previewGeneral bulk processingLowestFast
gemini-3-pro-previewComplex reasoning tasksMediumMedium
gemini-2.5-flashStable, reliableLowFast
gemini-2.5-proCode/math/STEMMediumSlow

Job States

StateDescription
JOB_STATE_PENDINGJob queued, waiting to start
JOB_STATE_RUNNINGJob actively processing
JOB_STATE_SUCCEEDEDJob completed successfully
JOB_STATE_FAILEDJob failed (check error message)
JOB_STATE_CANCELLEDJob was cancelled
JOB_STATE_EXPIREDJob timed out

Size Limits

MethodMax SizeBest For
File uploadUnlimitedLarge batches (recommended)
Inline requests<20MBSmall batches

Output Interpretation

Results JSONL

Each line contains:

{
  "key": "your-identifier",
  "response": {
    "text": "Generated content here..."
  }
}

Error Handling

  • Failed requests appear in results with error information
  • Check for error field in response
  • Partial failures don't stop entire job

Result Access

  • Use --output to save to file
  • Script prints preview of results
  • Parse JSONL line by line for processing

Common Issues

"google-genai not installed"

pip install google-genai

"JSONL file not found"

  • Verify file path is correct
  • Check file extension is .jsonl (not .json)
  • Use absolute paths if relative paths fail

"Invalid JSONL format"

  • Each line must be valid JSON
  • No trailing commas between objects
  • Check for syntax errors in JSON
  • Use JSON validator if unsure

"Job failed"

  • Check error message in status
  • Verify request format is correct
  • Check model availability
  • Review API quota limits

"No results found"

  • Ensure job state is JOB_STATE_SUCCEEDED
  • Wait for job completion before retrieving
  • Check job status first with check_status.py

"Processing stuck in RUNNING state"

  • Large jobs can take hours
  • Use --wait flag for automated polling
  • Check job size and model choice
  • Contact support if stuck >24 hours

Best Practices

JSONL Creation

  • Use unique keys for each request
  • Validate JSON before uploading
  • Test with small batch first (5-10 requests)
  • Include error handling in your scripts

Job Management

  • Use descriptive display names (--name)
  • Save job names for tracking
  • Monitor status before retrieving results
  • Keep backup of original JSONL file

Performance Optimization

  • Use flash models for cost efficiency
  • Batch as many requests as possible
  • File upload preferred over inline
  • Process during off-peak hours if timing sensitive

Error Handling

  • Check for failed requests in results
  • Retry failed requests individually
  • Log job names for audit trails
  • Validate output format before use

Cost Management

  • Batch API is 50%+ cheaper than real-time
  • Use flash models when possible
  • Monitor token usage
  • Process in chunks if quota limited

Related Skills

  • gemini-text: Generate individual text requests
  • gemini-image: Batch image generation
  • gemini-tts: Batch audio generation
  • gemini-embeddings: Batch embedding creation

Quick Reference

# Basic workflow
python scripts/create_batch.py requests.jsonl
python scripts/check_status.py <job-name> --wait
python scripts/get_results.py <job-name> --output results.jsonl

# With model and name
python scripts/create_batch.py requests.jsonl --model gemini-3-flash-preview --name "my-job"

# Create JSONL programmatically
echo '{"key":"1","request":{"contents":[{"parts":[{"text":"Prompt"}]}]}}' > batch.jsonl

Reference

Source

git clone https://github.com/akrindev/google-studio-skills/blob/main/skills/gemini-batch/SKILL.mdView on GitHub

Overview

This skill lets you run large volumes of requests through Gemini Batch API using executable scripts. It covers batch creation, status monitoring, and result retrieval to enable cost-efficient, high-throughput AI tasks. Ideal for bulk content generation, JSONL processing, and async workflows.

How This Skill Works

Prepare a JSONL input file and start a batch with scripts/create_batch.py, selecting a model and a job name. Monitor progress with scripts/check_status.py and retrieve finished results using scripts/get_results.py. The workflow runs asynchronously, enabling parallel processing and scalable throughput.

When to Use It

  • Process hundreds or thousands of requests
  • Generate content in bulk (blogs, emails, descriptions)
  • Reduce costs for high-volume AI tasks
  • Run async jobs without blocking
  • Create scalable content pipelines and multi-document outputs

Quick Start

  1. Step 1: Prepare a JSONL input file requests.jsonl with one request per line and the needed contents
  2. Step 2: Create and start a batch job: python scripts/create_batch.py requests.jsonl --name "my-batch-job" --model gemini-3-flash-preview
  3. Step 3: Monitor and fetch: python scripts/check_status.py <job-name> --wait; python scripts/get_results.py <job-name> --output results.jsonl

Best Practices

  • Design each line in the JSONL input to match the examples (key and request.contents with text prompts)
  • Always set a descriptive job name (--name) to simplify tracking with check_status
  • Validate a small JSONL sample first before scaling to full workloads
  • Choose an appropriate model (e.g., gemini-3-flash-preview) to balance cost and performance
  • Use get_results.py to fetch outputs and parse results.jsonl, handling large outputs in chunks if needed

Example Use Cases

  • Bulk blog post generation for a content studio by feeding topics in a JSONL file and collecting results via get_results
  • Generating hundreds of product descriptions for an e-commerce catalog using batch prompts
  • Async batch data summarization or literature reviews from large JSONL datasets
  • Mass email or newsletter content creation with per-item prompts and centralized result retrieval
  • Building a scalable reporting pipeline that assembles multi-document summaries from JSONL prompts

Frequently Asked Questions

Add this skill to your agents
Sponsor this space

Reach thousands of developers