llm-evaluation

Master comprehensive evaluation strategies for LLM applications, from automated metrics to human evaluation and A/B testing.

Content Preview
---
name: llm-evaluation
description: "Master comprehensive evaluation strategies for LLM applications, from automated metrics to human evaluation and A/B testing."
risk: unknown
source: community
date_added: "2026-02-27"
---

# LLM Evaluation

Master comprehensive evaluation strategies for LLM applications, from automated metrics to human evaluation and A/B testing.

## Do not use this skill when

- The task is unrelated to llm evaluation
- You need a different domain or tool outside this scope
How to Use

Recommended: Install to project (local)

mkdir -p .claude/skills
curl -o .claude/skills/llm-evaluation.md \
  https://raw.githubusercontent.com/sickn33/antigravity-awesome-skills/main/skills/llm-evaluation/SKILL.md

Skill is scoped to this project only. Add .claude/skills/ to your .gitignoreif you don't want to commit it.

Alternative: Clone full repo

git clone https://github.com/sickn33/antigravity-awesome-skills

Then reference at skills/llm-evaluation/SKILL.md

Related Skills