AI/ML · Global library
RAG Evaluation Framework Builder
Build comprehensive evaluation frameworks for RAG systems with retrieval metrics, generation metrics, and end-to-end assessment
CodexClaude CodeKimi Codeorchestrator-mcp
Best use case
Use RAG Evaluation Framework Builder when you need to build comprehensive evaluation frameworks for RAG systems with retrieval metrics, generation metrics, and end-to-end assessment, especially when the work is driven by RAG evaluation and retrieval metrics.
Trigger signals
RAG evaluationretrieval metricsgeneration metricsfaithfulnessanswer relevancecontext precision
Validation hooks
metric-coveragebenchmark-quality
Install surface
Copy the exact command path you need.
Inspect
pip install "orchestrator-mcp[dashboard]"
orchestrator-mcp skills show rag-evaluation-framework-builder
Use
orchestrator-mcp skills export rag-evaluation-framework-builder --to ./skillforge-packs
# copy the exported pack into your preferred agent environment
Export
cp -R skills/rag-evaluation-framework-builder ./your-agent-skills/rag-evaluation-framework-builder
# or open skills/rag-evaluation-framework-builder/SKILL.md in a markdown-first client
File patterns
*.pyeval*.pymetrics*.pyrag/*.py
Model preferences
claude-sonnet-4gpt-4oclaude-haiku-3
Related skills
Adjacent packs to compose next.
Design robust communication protocols for agent systems with message schemas, serialization, and delivery guarantees
CodexClaude Code
Manage complete agent lifecycles from initialization through graceful shutdown with health monitoring, scaling, and resource optimization
CodexClaude Code
Design short-term, long-term, and episodic memory layers for agents without turning retrieval into an unbounded context leak.
CodexClaude Code