Get the FREE Ultimate OpenClaw Setup Guide →

gemini-embeddings

Scanned
npx machina-cli add skill akrindev/google-studio-skills/gemini-embeddings --openclaw
Files (1)
SKILL.md
9.6 KB

Gemini Embeddings

Generate high-quality text embeddings for semantic search, similarity analysis, clustering, and RAG (Retrieval Augmented Generation) applications through executable scripts.

When to Use This Skill

Use this skill when you need to:

  • Find semantically similar documents or texts
  • Build semantic search engines
  • Implement RAG (Retrieval Augmented Generation)
  • Cluster or group similar documents
  • Calculate text similarity scores
  • Power recommendation systems
  • Enable semantic document retrieval
  • Create vector databases for AI applications

Available Scripts

scripts/embed.py

Purpose: Generate embeddings and calculate similarity

When to use:

  • Creating vector representations of text
  • Comparing text similarity
  • Building semantic search systems
  • Implementing RAG pipelines
  • Clustering documents

Key parameters:

ParameterDescriptionExample
textsText(s) to embed (required)"Your text here"
--model, -mEmbedding modelgemini-embedding-001
--task, -tTask typeSEMANTIC_SIMILARITY
--dim, -dOutput dimensionality768, 1536, 3072
--similarity, -sCalculate pairwise similarityFlag
--json, -jOutput as JSONFlag

Output: Embedding vectors or similarity scores

Workflows

Workflow 1: Single Text Embedding

python scripts/embed.py "What is the meaning of life?"
  • Best for: Basic embedding generation
  • Output: Vector with 3072 dimensions (default)
  • Use when: Storing single document vectors

Workflow 2: Semantic Search

# 1. Generate embedding for query
python scripts/embed.py "best practices for coding" --task RETRIEVAL_QUERY > query.json

# 2. Generate embeddings for documents (batch)
python scripts/embed.py "Coding best practices include version control" "Clean code is essential" --task RETRIEVAL_DOCUMENT > docs.json

# 3. Compare and find most similar (calculate similarity separately)
  • Best for: Building search functionality
  • Task types: RETRIEVAL_QUERY, RETRIEVAL_DOCUMENT
  • Combines with: Similarity calculation for ranking

Workflow 3: Text Similarity Comparison

python scripts/embed.py "What is the meaning of life?" "What is the purpose of existence?" "How do I bake a cake?" --similarity
  • Best for: Comparing multiple texts, finding duplicates
  • Output: Pairwise similarity scores (0-1)
  • Use when: Need to rank text similarity

Workflow 4: Dimensionality Reduction for Efficiency

python scripts/embed.py "Text to embed" --dim 768
  • Best for: Faster storage and comparison
  • Options: 768, 1536, or 3072 (default)
  • Trade-off: Lower dimensions = less accuracy but faster

Workflow 5: Document Clustering

# 1. Generate embeddings for multiple documents
python scripts/embed.py "Machine learning is AI" "Deep learning is a subset" "Neural networks power AI" --json > embeddings.jsonl

# 2. Process embeddings with clustering algorithm (your code)
# Use scikit-learn, KMeans, etc.
  • Best for: Grouping similar documents, topic discovery
  • Task type: CLUSTERING
  • Combines with: Clustering libraries (scikit-learn)

Workflow 6: RAG Implementation

# 1. Create document embeddings (one-time setup)
python scripts/embed.py "Document 1 content" "Document 2 content" --task RETRIEVAL_DOCUMENT --dim 1536

# 2. For each query, find similar documents
python scripts/embed.py "User query here" --task RETRIEVAL_QUERY

# 3. Use retrieved documents in prompt to LLM (gemini-text)
python skills/gemini-text/scripts/generate.py "Context: [retrieved docs]. Answer: [user query]"
  • Best for: Building knowledge-based AI systems
  • Combines with: gemini-text for generation with context

Workflow 7: JSON Output for API Integration

python scripts/embed.py "Text to process" --json
  • Best for: API responses, database storage
  • Output: JSON array of embedding vectors
  • Use when: Programmatic processing required

Workflow 8: Batch Document Processing

# 1. Create JSONL with documents
echo '{"text": "Document 1"}' > docs.jsonl
echo '{"text": "Document 2"}' >> docs.jsonl

# 2. Process with script or custom code
python3 << 'EOF'
import json
from google import genai

client = genai.Client()

texts = []
with open("docs.jsonl") as f:
    for line in f:
        texts.append(json.loads(line)["text"])

response = client.models.embed_content(
    model="gemini-embedding-001",
    contents=texts,
    task_type="RETRIEVAL_DOCUMENT"
)

embeddings = [e.values for e in response.embeddings]
print(f"Generated {len(embeddings)} embeddings")
EOF
  • Best for: Large document collections
  • Combines with: Vector databases (Pinecone, Weaviate)

Parameters Reference

Task Types

Task TypeBest ForWhen to Use
SEMANTIC_SIMILARITYComparing text similarityGeneral comparison tasks
RETRIEVAL_DOCUMENTEmbedding documentsStoring documents for retrieval
RETRIEVAL_QUERYEmbedding search queriesFinding similar documents
CLASSIFICATIONText classificationCategorizing text
CLUSTERINGGrouping similar textsDocument clustering

Dimensionality Options

DimensionsUse CaseTrade-off
768High-volume, real-timeLower accuracy, faster
1536Balanced performanceGood accuracy/speed balance
3072Highest accuracySlower, more storage

Similarity Scores

ScoreInterpretation
0.8 - 1.0Very similar (likely duplicates)
0.6 - 0.8Highly related (same topic)
0.4 - 0.6Moderately related
0.2 - 0.4Weakly related
0.0 - 0.2Unrelated

Output Interpretation

Embedding Vector

  • Format: List of float values (768, 1536, or 3072)
  • Range: Typically -1.0 to 1.0
  • Normalized for cosine similarity
  • Can be stored in vector databases

Similarity Output

Pairwise Similarity:
  'What is the meaning of life?...' <-> 'What is the purpose of existence?...': 0.8742
  'What is the meaning of life?...' <-> 'How do I bake a cake?...': 0.1234
  • Higher scores = more similar
  • Use threshold (e.g., 0.7) for matching

JSON Output

[[0.123, -0.456, 0.789, ...], [0.234, -0.567, 0.890, ...]]
  • Array of embedding vectors
  • One per input text
  • Ready for database storage

Common Issues

"google-genai not installed"

pip install google-genai numpy

"numpy not installed" (for similarity)

pip install numpy

"Invalid task type"

  • Use available tasks: SEMANTIC_SIMILARITY, RETRIEVAL_DOCUMENT, RETRIEVAL_QUERY, CLASSIFICATION, CLUSTERING
  • Check spelling (case-sensitive)
  • Use correct task for your use case

"Invalid dimension"

  • Options: 768, 1536, or 3072 only
  • Check model supports requested dimension
  • Default to 3072 if unsure

"No similarity calculated"

  • Need multiple texts for similarity comparison
  • Use --similarity flag
  • Check that at least 2 texts provided

"Embedding size mismatch"

  • All embeddings must have same dimensionality
  • Use consistent --dim parameter
  • Recompute if dimensions differ

Best Practices

Task Selection

  • SEMANTIC_SIMILARITY: General text comparison
  • RETRIEVAL_DOCUMENT: Storing documents for search
  • RETRIEVAL_QUERY: Querying for similar documents
  • CLASSIFICATION: Categorization tasks
  • CLUSTERING: Grouping similar content

Dimensionality Choice

  • 768: Real-time applications, high volume
  • 1536: Balanced choice for most use cases
  • 3072: Maximum accuracy, offline processing

Performance Optimization

  • Use lower dimensions for speed
  • Batch multiple texts in one request
  • Cache embeddings for repeated queries
  • Precompute document embeddings for search

Storage Tips

  • Use vector databases (Pinecone, Weaviate, Chroma)
  • Normalize vectors for consistent comparison
  • Store metadata with embeddings
  • Index for fast retrieval

RAG Implementation

  • Precompute document embeddings
  • Use RETRIEVAL_DOCUMENT for docs
  • Use RETRIEVAL_QUERY for user questions
  • Combine top results with gemini-text

Similarity Thresholds

  • 0.9+: Exact duplicates or near-duplicates
  • 0.7-0.9: Same topic/subject
  • 0.5-0.7: Related concepts
  • <0.5: Different topics

Related Skills

  • gemini-text: Generate text with retrieved context (RAG)
  • gemini-batch: Process embeddings in bulk
  • gemini-files: Upload documents for embedding
  • gemini-search: Implement semantic search (if available)

Quick Reference

# Basic embedding
python scripts/embed.py "Your text here"

# Semantic search
python scripts/embed.py "Query" --task RETRIEVAL_QUERY

# Document embedding
python scripts/embed.py "Document text" --task RETRIEVAL_DOCUMENT

# Similarity comparison
python scripts/embed.py "Text 1" "Text 2" "Text 3" --similarity

# Dimensionality reduction
python scripts/embed.py "Text" --dim 768

# JSON output
python scripts/embed.py "Text" --json

Reference

Source

git clone https://github.com/akrindev/google-studio-skills/blob/main/skills/gemini-embeddings/SKILL.mdView on GitHub

Overview

Generate high-quality text embeddings for semantic search, similarity analysis, clustering, and RAG using the Gemini Embedding API via executable scripts. These vector representations power semantic retrieval, similarity scoring, topic discovery, and scalable AI workflows.

How This Skill Works

Run the embedded script (scripts/embed.py) to produce embeddings or compute similarity. Specify the Gemini model (e.g., gemini-embedding-001) and a task type such as RETRIEVAL_QUERY, RETRIEVAL_DOCUMENT, or CLUSTERING; you can also request pairwise similarity with --similarity. Outputs include embedding vectors and, when requested, similarity scores for ranking.

When to Use It

  • Build semantic search engines to retrieve documents by meaning, not just keywords
  • Find semantically similar documents or texts for deduplication or recommendations
  • Implement Retrieval Augmented Generation (RAG) pipelines by aligning queries with relevant docs
  • Cluster or group documents to discover topics and structure large text corpora
  • Create vector databases for fast similarity lookups in AI applications

Quick Start

  1. Step 1: Generate a single embedding for a text sample python scripts/embed.py "What is the meaning of life?"
  2. Step 2: For semantic search, create query and document embeddings python scripts/embed.py "best practices for coding" --task RETRIEVAL_QUERY > query.json python scripts/embed.py "Coding best practices include version control" "Clean code is essential" --task RETRIEVAL_DOCUMENT > docs.json
  3. Step 3: Compare embeddings to obtain similarity scores and rank results Python code or --similarity can be used to compute and sort results

Best Practices

  • Choose the right model and dimensionality (gemini-embedding-001, with 768, 1536, or 3072 dimensions) based on speed and accuracy needs
  • Use the appropriate task types: RETRIEVAL_QUERY for queries, RETRIEVAL_DOCUMENT for documents, CLUSTERING for grouping
  • Start with single-text embeddings to validate output before batch processing
  • Export embeddings in JSON or JSONL format (using --json) for downstream processing and analysis
  • Leverage workflows for real use cases: semantic search (RETRIEVAL_QUERY/RETRIEVAL_DOCUMENT) and clustering (CLUSTERING) to structure results

Example Use Cases

  • Power a semantic search engine over a corpus of articles or manuals
  • Enable a RAG system by retrieving relevant documents to answer user queries
  • Cluster customer feedback into topics for trend analysis
  • Compute text similarity scores to detect duplicates or near-duplicates across documents
  • Build a vector database of product descriptions for fast similarity-based recommendations

Frequently Asked Questions

Add this skill to your agents
Sponsor this space

Reach thousands of developers