Get the FREE Ultimate OpenClaw Setup Guide →

nano-banana

Scanned
npx machina-cli add skill NikiforovAll/claude-code-rules/nano-banana --openclaw
Files (1)
SKILL.md
5.5 KB

Nano Banana Skill

Python scripting with Gemini image generation using uv. Write small, focused scripts using heredocs for quick tasks—no files needed for one-off operations.

Choosing Your Approach

Quick image generation: Use heredoc with inline Python for one-off image requests.

Complex workflows: When multiple steps are needed (generate -> refine -> save), break into separate scripts and iterate.

Scripting tasks: For non-image Python tasks, use the same heredoc pattern with uv run.

Writing Scripts

Execute Python inline using heredocs with inline script metadata for dependencies:

uv run - << 'EOF'
# /// script
# dependencies = ["google-genai", "pillow"]
# ///
from google import genai
from google.genai import types

client = genai.Client()

response = client.models.generate_content(
    model="gemini-3-pro-image-preview",
    contents=["A cute banana character with sunglasses"],
    config=types.GenerateContentConfig(
        response_modalities=['IMAGE']
    )
)

for part in response.parts:
    if part.inline_data is not None:
        image = part.as_image()
        image.save("tmp/generated.png")
        print("Image saved to tmp/generated.png")
EOF

The # /// script block declares dependencies inline using TOML syntax. This makes scripts self-contained and reproducible.

Why these dependencies:

  • google-genai - Gemini API client
  • pillow - Required for .as_image() method (converts base64 to PIL Image) and saving images

Only write to files when:

  • The script needs to be reused multiple times
  • The script is complex and requires iteration
  • The user explicitly asks for a saved script

Basic Template

uv run - << 'EOF'
# /// script
# dependencies = ["google-genai", "pillow"]
# ///
from google import genai
from google.genai import types

client = genai.Client()

# Generate image
response = client.models.generate_content(
    model="gemini-3-pro-image-preview",
    contents=["YOUR PROMPT HERE"],
    config=types.GenerateContentConfig(
        response_modalities=['IMAGE']
    )
)

# Save result
for part in response.parts:
    if part.text is not None:
        print(part.text)
    elif part.inline_data is not None:
        image = part.as_image()
        image.save("tmp/output.png")
        print("Saved: tmp/output.png")
EOF

Key Principles

  1. Small scripts: Each script should do ONE thing (generate, refine, save)
  2. Evaluate output: Always save images and print status to decide next steps
  3. Use tmp/: Save generated images to tmp/ directory by default
  4. Stateless execution: Each script runs independently, no cleanup needed

Workflow Loop

Follow this pattern for complex tasks:

  1. Write a script to generate/process one image
  2. Run it and observe the output
  3. Evaluate - did it work? Check the saved image
  4. Decide - refine prompt or task complete?
  5. Repeat until satisfied

Image Configuration

Configure aspect ratio and resolution:

config=types.GenerateContentConfig(
    response_modalities=['IMAGE'],
    image_config=types.ImageConfig(
        aspect_ratio="16:9",  # "1:1", "16:9", "9:16", "4:3", "3:4"
        image_size="2K"       # "1K", "2K", "4K" (uppercase required)
    )
)

Models

  • gemini-3-pro-image-preview - Fast, general purpose image generation
  • gemini-3-pro-image-preview - Advanced, professional asset production (Nano Banana Pro)

Default to gemini-3-pro-image-preview (Nano Banana Pro) for all image generation unless:

  • The user explicitly requests a different model
  • The user wants to save budget/costs
  • The user specifies a simpler or quick generation task

Nano Banana Pro provides higher quality results and should be the recommended choice.

Text + Image Output

To receive both text explanation and image:

config=types.GenerateContentConfig(
    response_modalities=['TEXT', 'IMAGE']
)

Image Editing

Edit existing images by including them in the request:

uv run - << 'EOF'
# /// script
# dependencies = ["google-genai", "pillow"]
# ///
from google import genai
from google.genai import types
from PIL import Image

client = genai.Client()

# Load existing image
img = Image.open("input.png")

response = client.models.generate_content(
    model="gemini-3-pro-image-preview",
    contents=[
        "Add a party hat to this character",
        img
    ],
    config=types.GenerateContentConfig(
        response_modalities=['IMAGE']
    )
)

for part in response.parts:
    if part.inline_data is not None:
        part.as_image().save("tmp/edited.png")
        print("Saved: tmp/edited.png")
EOF

Debugging Tips

  1. Print response.parts to see what was returned
  2. Check for text parts - model may include explanations
  3. Save images immediately to verify output visually
  4. Use Read tool to view saved images after generation

Error Recovery

If a script fails:

  1. Check error message for API issues
  2. Verify GOOGLE_API_KEY is set
  3. Try simpler prompt to isolate the issue
  4. Check image format compatibility for edits

Advanced Scenarios

For complex workflows including thinking process, Google Search grounding, multi-turn conversations, and professional asset production, load references/guide.md.

Source

git clone https://github.com/NikiforovAll/claude-code-rules/blob/main/plugins/handbook-nano-banana/skills/nano-banana/SKILL.mdView on GitHub

Overview

Nano Banana enables inline Python scripting for Gemini image generation using uv run. Create quick AI art, edit images with AI, or run Python tasks using heredocs without needing to manage files for one-off operations. It emphasizes small, focused scripts with inline dependencies for reproducibility.

How This Skill Works

Write a one-off Python script inside a heredoc and execute it with uv run. Declare dependencies inline using the # /// script block (TOML style) such as google-genai and pillow, then call Gemini models (default gemini-3-pro-image-preview) to generate images and save them to tmp/ using the provided image handling methods.

When to Use It

  • Generate a quick image from a prompt using a one-off heredoc script
  • Create AI art or edit an image with AI via Gemini in a compact script
  • Run small Python tasks inline with uv run to avoid creating files
  • Handle complex workflows by splitting into multiple scripts (generate -> refine -> save)
  • Prototype image prompts quickly with inline dependencies and save results to tmp/

Quick Start

  1. Step 1: Create a one-off inline Python script using a heredoc and run it with uv run
  2. Step 2: Inside the script, declare dependencies with the # /// script block and call Gemini to generate an image, saving to tmp/output.png
  3. Step 3: Review tmp/output.png and repeat or refine as needed

Best Practices

  • Keep each script focused on a single task (generate, refine, save)
  • Declare dependencies inline using the # /// script block (TOML style)
  • Save generated images to tmp/ and print status updates
  • Default to the gemini-3-pro-image-preview model unless the user requests otherwise
  • Use a simple, iterative workflow loop to refine prompts and outputs

Example Use Cases

  • Generate an image of a cute banana character with sunglasses and save to tmp/generated.png
  • Create AI art of a futuristic cityscape using a one-off inline script
  • Edit an image by generating a refined version and saving to tmp/refined.png
  • Prototype a 16:9, 2K image via image_config in the script and save to tmp/output.png
  • Iterate prompts: generate, evaluate the result, then re-run with refinements

Frequently Asked Questions

Add this skill to your agents
Sponsor this space

Reach thousands of developers