Back to Academy
Relevance 7/10Safety and PolicyAdvanced6 min read
Adversarial Example Awareness
Adversarial example awareness identifies inputs crafted to trigger model errors.
Why it matters for annotators
Awareness improves robustness testing and safety evaluation quality.
Visual mental model
Adversarial input -> model failure mode -> defensive label.
Examples (bad vs good)
Scenario: Real annotation scenario involving Adversarial Example Awareness
Bad: Labeling quickly without applying project rubric.
Good: Applying rubric criteria, documenting rationale, and escalating uncertainty.
Common mistakes
- Skipping guideline details for edge cases.
- Applying inconsistent criteria across similar samples.
- Avoiding escalation even when uncertain.
Submission checklist
- Read the latest guideline update before each batch.
- Apply rubric dimensions explicitly in each decision.
- Escalate ambiguous items with concise rationale.