🛠️ エージェントDBベクター検索
AgentDBを活用し、文書の内容を理解して関連性の
📺 まず動画で見る(YouTube)
▶ 【衝撃】最強のAIエージェント「Claude Code」の最新機能・使い方・プログラミングをAIで効率化する超実践術を解説! ↗
※ jpskill.com 編集部が参考用に選んだ動画です。動画の内容と Skill の挙動は厳密には一致しないことがあります。
📜 元の英語説明(参考)
Implement semantic vector search with AgentDB for intelligent document retrieval, similarity matching, and context-aware querying. Use when building RAG systems, semantic search engines, or intelligent knowledge bases.
🇯🇵 日本人クリエイター向け解説
AgentDBを活用し、文書の内容を理解して関連性の
※ jpskill.com 編集部が日本のビジネス現場向けに補足した解説です。Skill本体の挙動とは独立した参考情報です。
⚠️ ダウンロード・利用は自己責任でお願いします。当サイトは内容・動作・安全性について責任を負いません。
🎯 このSkillでできること
下記の説明文を読むと、このSkillがあなたに何をしてくれるかが分かります。Claudeにこの分野の依頼をすると、自動で発動します。
📦 インストール方法 (3ステップ)
- 1. 上の「ダウンロード」ボタンを押して .skill ファイルを取得
- 2. ファイル名の拡張子を .skill から .zip に変えて展開(macは自動展開可)
- 3. 展開してできたフォルダを、ホームフォルダの
.claude/skills/に置く- · macOS / Linux:
~/.claude/skills/ - · Windows:
%USERPROFILE%\.claude\skills\
- · macOS / Linux:
Claude Code を再起動すれば完了。「このSkillを使って…」と話しかけなくても、関連する依頼で自動的に呼び出されます。
詳しい使い方ガイドを見る →- 最終更新
- 2026-05-17
- 取得日時
- 2026-05-17
- 同梱ファイル
- 1
💬 こう話しかけるだけ — サンプルプロンプト
- › AgentDB Vector Search を使って、最小構成のサンプルコードを示して
- › AgentDB Vector Search の主な使い方と注意点を教えて
- › AgentDB Vector Search を既存プロジェクトに組み込む方法を教えて
これをClaude Code に貼るだけで、このSkillが自動発動します。
📖 Claude が読む原文 SKILL.md(中身を展開)
この本文は AI(Claude)が読むための原文(英語または中国語)です。日本語訳は順次追加中。
AgentDB Vector Search
What This Skill Does
Implements vector-based semantic search using AgentDB's high-performance vector database with 150x-12,500x faster operations than traditional solutions. Features HNSW indexing, quantization, and sub-millisecond search (<100µs).
Prerequisites
- Node.js 18+
- AgentDB v1.0.7+ (via agentic-flow or standalone)
- OpenAI API key (for embeddings) or custom embedding model
Quick Start with CLI
Initialize Vector Database
# Initialize with default dimensions (1536 for OpenAI ada-002)
npx agentdb@latest init .$vectors.db
# Custom dimensions for different embedding models
npx agentdb@latest init .$vectors.db --dimension 768 # sentence-transformers
npx agentdb@latest init .$vectors.db --dimension 384 # all-MiniLM-L6-v2
# Use preset configurations
npx agentdb@latest init .$vectors.db --preset small # <10K vectors
npx agentdb@latest init .$vectors.db --preset medium # 10K-100K vectors
npx agentdb@latest init .$vectors.db --preset large # >100K vectors
# In-memory database for testing
npx agentdb@latest init .$vectors.db --in-memory
Query Vector Database
# Basic similarity search
npx agentdb@latest query .$vectors.db "[0.1,0.2,0.3,...]"
# Top-k results
npx agentdb@latest query .$vectors.db "[0.1,0.2,0.3]" -k 10
# With similarity threshold (cosine similarity)
npx agentdb@latest query .$vectors.db "0.1 0.2 0.3" -t 0.75 -m cosine
# Different distance metrics
npx agentdb@latest query .$vectors.db "[...]" -m euclidean # L2 distance
npx agentdb@latest query .$vectors.db "[...]" -m dot # Dot product
# JSON output for automation
npx agentdb@latest query .$vectors.db "[...]" -f json -k 5
# Verbose output with distances
npx agentdb@latest query .$vectors.db "[...]" -v
Import/Export Vectors
# Export vectors to JSON
npx agentdb@latest export .$vectors.db .$backup.json
# Import vectors from JSON
npx agentdb@latest import .$backup.json
# Get database statistics
npx agentdb@latest stats .$vectors.db
Quick Start with API
import { createAgentDBAdapter, computeEmbedding } from 'agentic-flow$reasoningbank';
// Initialize with vector search optimizations
const adapter = await createAgentDBAdapter({
dbPath: '.agentdb$vectors.db',
enableLearning: false, // Vector search only
enableReasoning: true, // Enable semantic matching
quantizationType: 'binary', // 32x memory reduction
cacheSize: 1000, // Fast retrieval
});
// Store document with embedding
const text = "The quantum computer achieved 100 qubits";
const embedding = await computeEmbedding(text);
await adapter.insertPattern({
id: '',
type: 'document',
domain: 'technology',
pattern_data: JSON.stringify({
embedding,
text,
metadata: { category: "quantum", date: "2025-01-15" }
}),
confidence: 1.0,
usage_count: 0,
success_count: 0,
created_at: Date.now(),
last_used: Date.now(),
});
// Semantic search with MMR (Maximal Marginal Relevance)
const queryEmbedding = await computeEmbedding("quantum computing advances");
const results = await adapter.retrieveWithReasoning(queryEmbedding, {
domain: 'technology',
k: 10,
useMMR: true, // Diverse results
synthesizeContext: true, // Rich context
});
Core Features
1. Vector Storage
// Store with automatic embedding
await db.storeWithEmbedding({
content: "Your document text",
metadata: { source: "docs", page: 42 }
});
2. Similarity Search
// Find similar documents
const similar = await db.findSimilar("quantum computing", {
limit: 5,
minScore: 0.75
});
3. Hybrid Search (Vector + Metadata)
// Combine vector similarity with metadata filtering
const results = await db.hybridSearch({
query: "machine learning models",
filters: {
category: "research",
date: { $gte: "2024-01-01" }
},
limit: 20
});
Advanced Usage
RAG (Retrieval Augmented Generation)
// Build RAG pipeline
async function ragQuery(question: string) {
// 1. Get relevant context
const context = await db.searchSimilar(
await embed(question),
{ limit: 5, threshold: 0.7 }
);
// 2. Generate answer with context
const prompt = `Context: ${context.map(c => c.text).join('\n')}
Question: ${question}`;
return await llm.generate(prompt);
}
Batch Operations
// Efficient batch storage
await db.batchStore(documents.map(doc => ({
text: doc.content,
embedding: doc.vector,
metadata: doc.meta
})));
MCP Server Integration
# Start AgentDB MCP server for Claude Code
npx agentdb@latest mcp
# Add to Claude Code (one-time setup)
claude mcp add agentdb npx agentdb@latest mcp
# Now use MCP tools in Claude Code:
# - agentdb_query: Semantic vector search
# - agentdb_store: Store documents with embeddings
# - agentdb_stats: Database statistics
Performance Benchmarks
# Run comprehensive benchmarks
npx agentdb@latest benchmark
# Results:
# ✅ Pattern Search: 150x faster (100µs vs 15ms)
# ✅ Batch Insert: 500x faster (2ms vs 1s for 100 vectors)
# ✅ Large-scale Query: 12,500x faster (8ms vs 100s at 1M vectors)
# ✅ Memory Efficiency: 4-32x reduction with quantization
Quantization Options
AgentDB provides multiple quantization strategies for memory efficiency:
Binary Quantization (32x reduction)
const adapter = await createAgentDBAdapter({
quantizationType: 'binary', // 768-dim → 96 bytes
});
Scalar Quantization (4x reduction)
const adapter = await createAgentDBAdapter({
quantizationType: 'scalar', // 768-dim → 768 bytes
});
Product Quantization (8-16x reduction)
const adapter = await createAgentDBAdapter({
quantizationType: 'product', // 768-dim → 48-96 bytes
});
Distance Metrics
# Cosine similarity (default, best for most use cases)
npx agentdb@latest query .$db.sqlite "[...]" -m cosine
# Euclidean distance (L2 norm)
npx agentdb@latest query .$db.sqlite "[...]" -m euclidean
# Dot product (for normalized vectors)
npx agentdb@latest query .$db.sqlite "[...]" -m dot
Advanced Features
HNSW Indexing
- O(log n) search complexity
- Sub-millisecond retrieval (<100µs)
- Automatic index building
Caching
- 1000 pattern in-memory cache
- <1ms pattern retrieval
- Automatic cache invalidation
MMR (Maximal Marginal Relevance)
- Diverse result sets
- Avoid redundancy
- Balance relevance and diversity
Performance Tips
- Enable HNSW indexing: Automatic with AgentDB, 10-100x faster
- Use quantization: Binary (32x), Scalar (4x), Product (8-16x) memory reduction
- Batch operations: 500x faster for bulk inserts
- Match dimensions: 1536 (OpenAI), 768 (sentence-transformers), 384 (MiniLM)
- Similarity threshold: Start at 0.7 for quality, adjust based on use case
- Enable caching: 1000 pattern cache for frequent queries
Troubleshooting
Issue: Slow search performance
# Check if HNSW indexing is enabled (automatic)
npx agentdb@latest stats .$vectors.db
# Expected: <100µs search time
Issue: High memory usage
# Enable binary quantization (32x reduction)
# Use in adapter: quantizationType: 'binary'
Issue: Poor relevance
# Adjust similarity threshold
npx agentdb@latest query .$db.sqlite "[...]" -t 0.8 # Higher threshold
# Or use MMR for diverse results
# Use in adapter: useMMR: true
Issue: Wrong dimensions
# Check embedding model dimensions:
# - OpenAI ada-002: 1536
# - sentence-transformers: 768
# - all-MiniLM-L6-v2: 384
npx agentdb@latest init .$db.sqlite --dimension 768
Database Statistics
# Get comprehensive stats
npx agentdb@latest stats .$vectors.db
# Shows:
# - Total patterns$vectors
# - Database size
# - Average confidence
# - Domains distribution
# - Index status
Performance Characteristics
- Vector Search: <100µs (HNSW indexing)
- Pattern Retrieval: <1ms (with cache)
- Batch Insert: 2ms for 100 vectors
- Memory Efficiency: 4-32x reduction with quantization
- Scalability: Handles 1M+ vectors efficiently
- Latency: Sub-millisecond for most operations
Learn More
- GitHub: https:/$github.com$ruvnet$agentic-flow$tree$main$packages$agentdb
- Documentation: node_modules$agentic-flow/docs/AGENTDB_INTEGRATION.md
- MCP Integration:
npx agentdb@latest mcpfor Claude Code - Website: https:/$agentdb.ruv.io
- CLI Help:
npx agentdb@latest --help - Command Help:
npx agentdb@latest help <command>