pinecone-integration
npx machina-cli add skill a5c-ai/babysitter/pinecone-integration --openclawFiles (1)
SKILL.md
1.2 KB
Pinecone Integration Skill
Capabilities
- Set up Pinecone index and environment
- Configure index parameters and pods
- Implement upsert and query operations
- Design namespace strategies for multi-tenancy
- Configure metadata filtering
- Implement batch operations and optimization
Target Processes
- vector-database-setup
- rag-pipeline-implementation
Implementation Details
Core Operations
- Index Management: Create, configure, delete indices
- Upsert: Single and batch vector uploads
- Query: Similarity search with metadata filters
- Fetch/Delete: Direct vector operations
- Index Stats: Monitor index usage
Configuration Options
- Index dimension and metric
- Pod type and replicas
- Serverless vs pod-based deployment
- Namespace configuration
- Metadata schema design
Best Practices
- Use appropriate metric for embeddings
- Design namespaces for isolation
- Batch upserts for efficiency
- Implement proper error handling
- Monitor index performance
Dependencies
- pinecone-client
- langchain-pinecone
Source
git clone https://github.com/a5c-ai/babysitter/blob/main/plugins/babysitter/skills/babysit/process/specializations/ai-agents-conversational/skills/pinecone-integration/SKILL.mdView on GitHub Overview
This skill covers Pinecone index creation, configuration, and core vector operations (upsert, query, fetch, delete) tailored for RAG pipelines. It also implements namespace-based multi-tenancy, metadata filtering, and batch optimization to improve throughput and relevance.
How This Skill Works
It orchestrates core Pinecone operations via the pinecone-client, including index management, upsert (single and batch), query with metadata filters, and fetch/delete, along with index stats monitoring. Configuration options govern index dimension, metric, pod type, replicas, deployment mode (serverless vs pod-based), namespace, and metadata schema to support scalable, multi-tenant RAG workflows.
When to Use It
- Initializing a new Pinecone index for a RAG pipeline
- Ingesting large document sets via batch upserts
- Running similarity queries with metadata-based filters
- Isolating user or tenant data with namespaces
- Monitoring index usage and tuning deployment (pods/replicas)
Quick Start
- Step 1: Install dependencies and initialize the Pinecone client (pinecone-client, langchain-pinecone)
- Step 2: Create and configure an index with the desired dimension, metric, pod type, and namespace
- Step 3: Upsert vectors (single or batch), then perform a query with optional metadata filters and fetch/delete as needed
Best Practices
- Choose the right embedding metric for your embeddings
- Design namespaces to ensure isolation between tenants or datasets
- Batch upserts to improve throughput and reduce API calls
- Implement robust error handling and retries
- Monitor index performance and adjust resources (pods/replicas) accordingly
Example Use Cases
- A multi-tenant support bot needing isolated document spaces
- Internal docs search across departments with metadata filters
- Product catalog Q&A with per-document metadata (source, author)
- Legal document retrieval with batch ingestion
- Research assistant indexing scientific papers and querying with filters
Frequently Asked Questions
Add this skill to your agents