Back to Skills

qdrant-vector-search

davila7
Updated 4 days ago
304 views
18,478
1,685
18,478
View on GitHub
MetaRAGVector SearchQdrantSemantic SearchEmbeddingsSimilarity SearchHNSWProductionDistributed

About

The qdrant-vector-search skill provides a high-performance vector similarity search engine for building production RAG systems. It enables fast nearest neighbor search, hybrid search with filtering, and scalable vector storage powered by Rust. Use it when you need low-latency semantic search with horizontal scaling capabilities and full data control.

Quick Install

Claude Code

Recommended
Primary
npx skills add davila7/claude-code-templates -a claude-code
Plugin CommandAlternative
/plugin add https://github.com/davila7/claude-code-templates
Git CloneAlternative
git clone https://github.com/davila7/claude-code-templates.git ~/.claude/skills/qdrant-vector-search

Copy and paste this command in Claude Code to install this skill

GitHub Repository

davila7/claude-code-templates
Path: cli-tool/components/skills/ai-research/rag-qdrant
0
anthropicanthropic-claudeclaudeclaude-code

Related Skills

huggingface-tokenizers

Documents

This skill provides high-performance tokenization using HuggingFace's Rust-based library, processing 1GB of text in under 20 seconds. It supports BPE, WordPiece, and Unigram algorithms while enabling custom tokenizer training and alignment tracking. Use it when you need production-fast tokenization or to build custom tokenizers integrated with the transformers ecosystem.

View skill

crewai-multi-agent

Meta

CrewAI is a lightweight multi-agent orchestration framework for building teams of specialized AI agents that collaborate autonomously on complex tasks. It enables role-based agent collaboration with memory and supports sequential or hierarchical workflows for production use. The framework is built without LangChain dependencies for lean, fast execution.

View skill

chroma

Documentation

Chroma is an open-source embedding database for AI applications that provides vector search, metadata filtering, and a simple API. It's ideal for building RAG applications and semantic search, scaling from local development to production. Use it when you need a self-hosted vector database for document retrieval and embedding storage.

View skill

training-llms-megatron

Design

This skill trains massive LLMs (2B-462B parameters) using NVIDIA's Megatron-Core framework for maximum GPU efficiency. Use it when training models over 1B parameters and needing advanced parallelism like tensor, pipeline, or expert parallelism. It's a production-ready framework proven on models like Nemotron and LLaMA.

View skill