// ROLE SUMMARY

We need people who can think like attackers. You will methodically test AI models against a taxonomy of failure modes: harmful content generation, jailbreak susceptibility, PII leakage, bias amplification, and more.

AI Safety Red Teamer — Part Time

Red Teaming$3545/hrRemote(US)Posted January 26, 2026

// DESCRIPTION

We need people who can think like attackers. You will methodically test AI models against a taxonomy of failure modes: harmful content generation, jailbreak susceptibility, PII leakage, bias amplification, and more. Each test is logged in a structured format that feeds into the safety team's tracking system. Successful exploits are prioritized for mitigation; unsuccessful attempts still provide valuable negative evidence.

The strongest red teamers combine technical curiosity with lateral thinking. You should be the kind of person who reads a policy document and immediately starts thinking about edge cases and loopholes. Familiarity with prompt engineering techniques, model architectures, and AI safety research literature is valuable. You should also be comfortable writing clear, detailed vulnerability reports.

Red team sessions are scheduled in 3-4 hour blocks. Most testers work 3-5 sessions per week. The work is mentally intense, so we encourage breaks between sessions. A weekly debrief with the safety team reviews top findings and updates attack priorities.

// SKILLS & REQUIREMENTS

Background in security research, journalism, or lawComfort working with sensitive content categoriesAbility to reproduce and clearly document exploitsExperience in cybersecurity, pen testing, or adversarial MLCreative and lateral thinking about system vulnerabilities

// FREQUENTLY ASKED QUESTIONS

// READY TO GET STARTED?

Apply in minutes

Create your profile, select your areas of expertise, and start working on frontier AI projects.

Apply Now