qdrant-integration
Scannednpx machina-cli add skill a5c-ai/babysitter/qdrant-integration --openclawFiles (1)
SKILL.md
1.3 KB
Qdrant Integration Skill
Capabilities
- Set up Qdrant (local, cloud, self-hosted)
- Create collections with configuration
- Implement advanced filtering with payloads
- Configure quantization for efficiency
- Set up sparse vectors for hybrid search
- Implement batch operations and optimization
Target Processes
- vector-database-setup
- rag-pipeline-implementation
Implementation Details
Deployment Modes
- Local Memory: For testing
- Local Disk: Persistent local storage
- Qdrant Cloud: Managed service
- Self-Hosted: Docker/Kubernetes deployment
Core Operations
- Collection management with parameters
- Point upsert with vectors and payloads
- Search with filters (must, should, must_not)
- Scroll for pagination
- Batch operations
Configuration Options
- Vector parameters (size, distance)
- Quantization (scalar, product)
- Sparse vector configuration
- Payload indexes
- Replication and sharding
Best Practices
- Use quantization for large collections
- Design payload indexes for filters
- Implement proper batch sizes
- Configure appropriate distance metrics
Dependencies
- qdrant-client
- langchain-qdrant
Source
git clone https://github.com/a5c-ai/babysitter/blob/main/plugins/babysitter/skills/babysit/process/specializations/ai-agents-conversational/skills/qdrant-integration/SKILL.mdView on GitHub Overview
This skill integrates Qdrant as a vector database, enabling collection configuration, advanced filtering with payloads, and quantization for efficiency. It supports local memory, local disk storage, Qdrant Cloud, and self-hosted deployments, with sparse vectors for hybrid search. It also supports batch operations to scale RAG pipelines.
How This Skill Works
The skill provisions Qdrant deployments (local, cloud, or self-hosted) and creates collections with configurable vector size and distance. It supports point upserts with vectors and payloads, and enables search with must/should/must_not filters, plus pagination via scroll. Batch operations optimize throughput during large-scale updates.
When to Use It
- When you need scalable semantic search for large document sets.
- When you must filter results using payload-based metadata.
- When deploying in local, cloud, or self-hosted environments.
- When optimizing retrieval with quantization and sparse vectors for hybrid search.
- When performing batched upserts and searches to improve throughput.
Quick Start
- Step 1: Install qdrant-client and langchain-qdrant and prepare your environment.
- Step 2: Choose deployment mode (Local Memory/Disk, Cloud, or Self-Hosted) and start/connect to Qdrant; create a collection with size and distance.
- Step 3: Upsert vectors with payloads, then perform filtered searches (must/should/must_not) and use scroll for pagination; enable batch operations as needed.
Best Practices
- Use quantization for large collections.
- Design payload indexes for filters.
- Implement proper batch sizes for upserts and queries.
- Configure appropriate distance metrics (cosine, dot, euclidean).
- Leverage sparse vectors for hybrid search when needed.
Example Use Cases
- Set up a local Qdrant collection for a RAG pipeline with batch upserts and filter-based retrieval.
- Deploy Qdrant Cloud for cloud-based document search and apply payload-based filtering.
- Configure sparse vectors to enable hybrid search across text and structured metadata.
- Paginate results using scroll to fetch large result sets.
- Tune quantization and shard settings to optimize throughput on a self-hosted cluster.
Frequently Asked Questions
Add this skill to your agents