AI/ML · Global library
AI Safety Evaluator
Design and execute comprehensive safety evaluations for AI systems with red-teaming, adversarial testing, and safety metric frameworks
CodexClaude CodeKimi Codeorchestrator-mcp
Best use case
Use AI Safety Evaluator when you need to design and execute comprehensive safety evaluations for AI systems with red-teaming, adversarial testing, and safety metric frameworks, especially when the work is driven by safety evaluation and red team.
Trigger signals
safety evaluationred teamadversarial testsafety metricsharmful contentjailbreak
Validation hooks
coverage-checkthreshold-validation
Install surface
Copy the exact command path you need.
Inspect
pip install "orchestrator-mcp[dashboard]"
orchestrator-mcp skills show ai-safety-evaluator
Use
orchestrator-mcp skills export ai-safety-evaluator --to ./skillforge-packs
# copy the exported pack into your preferred agent environment
Export
cp -R skills/ai-safety-evaluator ./your-agent-skills/ai-safety-evaluator
# or open skills/ai-safety-evaluator/SKILL.md in a markdown-first client
File patterns
*.pyeval*.pysafety/*.pytest*.py
Model preferences
claude-opus-4gpt-4oclaude-haiku-3
Related skills
Adjacent packs to compose next.
Design robust communication protocols for agent systems with message schemas, serialization, and delivery guarantees
CodexClaude Code
Manage complete agent lifecycles from initialization through graceful shutdown with health monitoring, scaling, and resource optimization
CodexClaude Code
Design short-term, long-term, and episodic memory layers for agents without turning retrieval into an unbounded context leak.
CodexClaude Code