llm-agentsFeatured

llm-evaluation

5.2k starsUpdated 2025-12-28
Compatible with:claudecodex

Description

Implement comprehensive evaluation strategies for LLM applications using automated metrics, human feedback, and benchmarking.

How to Use

  1. Visit the GitHub repository to get the SKILL.md file
  2. Copy the file to your project root or .cursor/rules directory
  3. Restart your AI assistant or editor to apply the new skill

Full Skill Documentation

name

llm-evaluation

description

Implement comprehensive evaluation strategies for LLM applications using automated metrics, human feedback, and benchmarking. Use when testing LLM performance, measuring AI application quality, or establishing evaluation frameworks.

Tags

#llm#evaluation#testing