Claude Skill

MindTrial

by petmal

MindTrial: Evaluate and compare AI language models (LLMs) on text-based tasks with optional file/image attachments and tool use. Supports multiple providers (OpenAI, Google, Anthropic, DeepSeek, Mistral AI, xAI, Alibaba, Moonshot AI, OpenRouter), custom tasks in YAML, and HTML/CSV reports.

5 stars1 forksActiveGo
73
Good

📊 Score Breakdown

🛡️Security30%
5.0/5
Utility30%
1.0/5
🔄Maintenance25%
5.0/5
💎Uniqueness15%
4.0/5

Overall = Security (30%) + Utility (30%) + Maintenance (25%) + Uniqueness (15%). Full methodology →

ℹ️ Details

Category

💻 Code Execution & Dev Tools

Ecosystem

Claude Skill

Language

Go

Pricing

Free

License

MPL-2.0

Status

Active

Platforms

claude

📈 GitHub Signals

5

Stars

1

Forks

0

Commits (30d)

0

Open Issues

Last commit: 1 months ago

ai-benchmarkai-evaluation-toolsai-model-comparisonai-toolanthropicartificial-intelligence-projectsdeepseekgoogle-gemini-aigrok-ailanguage-models-aillm-benchmarkingllm-comparisonllm-evaluation-frameworkmistral-aimoonshot-aiopenaiopenrouteropensourceqwenxai

The Weekly Index 📬

New MCP servers, Claude skills, stale alerts, and picks — every Thursday.

Data last verified: 1 months ago. See something wrong? Report it →