Output Quality Evaluator (LLM-as-a-Judge)

Output Quality Evaluator (LLM-as-a-Judge)

Model: GPT-4
Difficulty: Intermediate
Upvotes: 0
evaluationLLM-as-a-judgequality

โ€” Prompt โ€”

Evaluate the following output [output text] for [criteria such as accuracy, clarity, relevance], and provide a score and reasoning.

Why It Works

Provides structured evaluation metrics, improving feedback quality.

Example Output

Score: 8/10 โ€” Accurate but lacks detail in explaining cause-effect relationships.