Research Engineer / Scientist, Alignment Science
AnthropicFull Time
Junior (1 to 2 years)
Key technologies and capabilities for this role
Common questions about this position
The salary range is $380K - $460K.
The position is hybrid.
The role requires passion for red-teaming and AI safety, ability to design and implement worst-case demonstrations, develop adversarial evaluations, conduct research on alignment failure modes, and publish influential papers.
Safety Systems is at the forefront of OpenAI’s mission to build safe AGI, with a misalignment research team focusing on identifying and understanding AGI risks through worst-case demonstrations, adversarial evaluations, stress testing, and alignment research, fostering a culture of rigorous, impact-oriented safety work.
Candidates who are already thinking about AGI misalignment problems night and day, share the mission to build safe AGI, and have expertise in red-teaming, adversarial evaluations, and alignment research will thrive.
Develops safe and beneficial AI technologies
OpenAI develops and deploys artificial intelligence technologies aimed at benefiting humanity. The company creates advanced AI models capable of performing various tasks, such as automating processes and enhancing creativity. OpenAI's products, like Sora, allow users to generate videos from text descriptions, showcasing the versatility of its AI applications. Unlike many competitors, OpenAI operates under a capped profit model, which limits the profits it can make and ensures that excess earnings are redistributed to maximize the social benefits of AI. This commitment to safety and ethical considerations is central to its mission of ensuring that artificial general intelligence (AGI) serves all of humanity.