Get the FREE Ultimate OpenClaw Setup Guide →

Evaluation

(5 skills)

AI agent skills tagged “Evaluation” for Claude Code, Cursor, Windsurf, and more.

phoenix-observability

Orchestra-Research/AI-Research-SKILLs

4.3k

Open-source AI observability platform for LLM tracing, evaluation, and monitoring. Use when debugging LLM applications with detailed traces, running evaluations on datasets, or monitoring production AI systems with real-time insights.

evaluating-code-models

Orchestra-Research/AI-Research-SKILLs

4.3k

Evaluates code generation models across HumanEval, MBPP, MultiPL-E, and 15+ benchmarks with pass@k metrics. Use when benchmarking code models, comparing coding abilities, testing multi-language support, or measuring code generation quality. Industry standard from BigCode Project used by HuggingFace leaderboards.

langsmith-observability

Orchestra-Research/AI-Research-SKILLs

4.3k

LLM observability platform for tracing, evaluation, and monitoring. Use when debugging LLM applications, evaluating model outputs against datasets, monitoring production systems, or building systematic testing pipelines for AI applications.

nemo-evaluator-sdk

Orchestra-Research/AI-Research-SKILLs

4.3k

Evaluates LLMs across 100+ benchmarks from 18+ harnesses (MMLU, HumanEval, GSM8K, safety, VLM) with multi-backend execution. Use when needing scalable evaluation on local Docker, Slurm HPC, or cloud platforms. NVIDIA's enterprise-grade platform with container-first architecture for reproducible benchmarking.

evaluating-llms-harness

Orchestra-Research/AI-Research-SKILLs

4.3k

Evaluates LLMs across 60+ academic benchmarks (MMLU, HumanEval, GSM8K, TruthfulQA, HellaSwag). Use when benchmarking model quality, comparing models, reporting academic results, or tracking training progress. Industry standard used by EleutherAI, HuggingFace, and major labs. Supports HuggingFace, vLLM, APIs.

Sponsor this space

Reach thousands of developers