Red-Team a Claim
Take a claim about AI in evaluation — a statement about how AI should or shouldn't be used in evaluation practice — and systematically interrogate its assumptions, evidence base, and potential harms. The goal is to stress-test assertions about AI's role, value, or risks in the evaluation field.
Steps:
- Identify a claim about AI in evaluation — something asserting how AI should or shouldn't be used in evaluation practice. Use one you've encountered in the field, generate one with an AI tool, or start from a provided example.
- Ask the AI: "What assumptions underlie this claim about AI in evaluation?" — then push back on the ones it misses or glosses over.
- Ask: "Who might this recommendation harm, and how?" Probe for blind spots specific to evaluation contexts (e.g. marginalized communities, low-resource settings, data privacy).
- Draft a short critique: one paragraph summarizing what the claim gets right and where it falls short as guidance for evaluation practitioners.
Include in your submission:
- The claim about AI in evaluation (copied or paraphrased)
- Your one-line framing of the core issue with the claim
- Evidence of your critique (pasted exchange, notes, or written summary)
- A 140-character reflection on what surprised you
Can't think of a claim? Check out our 100% not fabricated claim library →