Claude Skill

eval-view

by hidai25

Proof your AI agent still works. Regression testing with golden baselines, tool-call diffing, and output drift detection. MCP server + Claude Code skills. LangGraph, CrewAI, Anthropic, OpenAI.

47 stars6 forksActivePython
79
Good

📊 Score Breakdown

🛡️Security30%
5.0/5
Utility30%
2.0/5
🔄Maintenance25%
5.0/5
💎Uniqueness15%
4.0/5

Overall = Security (30%) + Utility (30%) + Maintenance (25%) + Uniqueness (15%). Full methodology →

ℹ️ Details

Category

💻 Code Execution & Dev Tools

Ecosystem

Claude Skill

Language

Python

Pricing

Free

License

Apache-2.0

Status

Active

Platforms

claude

📈 GitHub Signals

47

Stars

6

Forks

0

Commits (30d)

12

Open Issues

Last commit: 2 weeks ago

agentagent-benchmarkagent-evaluationagentic-aiai-agentsanthropiccrewaicrewai-toolsevaluationlangchainlanggraphlanggraph-pythonllmllmopsmlopsopenai-assistantspytesttestingtools

The Weekly Index 📬

New MCP servers, Claude skills, stale alerts, and picks — every Thursday.

Data last verified: 2 weeks ago. See something wrong? Report it →