Description
Why This Role Exists
At Mercor, we believe the safest AI is the one that’s already been attacked — by us. That’s why we’re building a pod of safety specialists to assess the capabilities and limitations of frontier models.
What You’ll Do
-
Generate high-quality human data: annotate failures, classify vulnerabilities, and flag systemic risks
-
Apply structure: follow taxonomies, benchmarks, and playbooks to keep testing consistent Document reproducibly: produce reports, datasets, and attack cases customers can act on
-
Flex across projects: support different customers, from LLM jailbreaks to socio-technical abuse testing
Who You Are
-
You’re curious and adversarial: you instinctively push systems to breaking points
-
You’re structured: you use frameworks or benchmarks, not just random hacks
-
You’re communicative: you explain risks clearly to technical and non-technical stakeholders
-
You’re adaptable: thrive on moving across projects and customers
Why Join Mercor
-
Build experience in human data AI work at the frontier of safety
-
Play a direct role in making AI systems more robust, safe, and trustworthy
We consider all qualified applicants without regard to legally protected characteristics and provide reasonable accommodations upon request.
About Mercor
Mercor partners with leading AI labs and enterprises to train frontier models using human expertise. You will work on projects that focus on training and enhancing AI systems.
Interested in this position?
Apply directly on the company's website