OpenAI

Model Policy Manager

San Francisco, California, United States

Not SpecifiedCompensation
Senior (5 to 8 years), Expert & Leadership (9+ years)Experience Level
Full TimeJob Type
UnknownVisa
Artificial Intelligence, BiotechnologyIndustries

About the Team

The Safety Systems team is at the forefront of OpenAI's mission to build and deploy safe AGI, driving our commitment to AI safety and fostering a culture of trust and transparency.

The Model Policy team aligns model behavior with desired human values and norms. We co-design policy with models and for models. Key focus areas include:

  • Addressing critical societal challenges like info-hazard risks and how the model should respond in mental health contexts.
  • Defining evaluation criteria for foundational models’ ability to reason about safety, values, and questions of cultural norms.
  • Driving rapid policy taxonomy iteration based on data.

About the Role

Providing access to powerful AI models introduces a host of challenging questions when it comes to model safety: How do we define safe behavior for how a model should behave? To what end? How do we do this in such a way that is actionable, objective and sustains replicability?

This is a senior role in which you’ll help shape policy creation and development at OpenAI and make an impact by helping ensure that our groundbreaking technologies do not create harm. The ideal candidate can identify and develop cohesive and thoughtful taxonomies of harm on high risk topics with a sense of urgency. They can balance internal and external input in making complex decisions, carefully think through trade-offs, and write principled, enforceable policies based on our values. Importantly, this role is embedded in our research teams and directly informs model training.

This role is based in San Francisco, CA. We use a hybrid work model of 3 days in the office per week and offer relocation assistance to new employees.

In this role, you’ll:

  • Design model policies that govern safe model behavior in an objective and defensible way - e.g. how should the model respond in risky/unsafe scenarios? What does unsafe mean?
  • Develop taxonomies that inform data collection campaigns, model behaviour and monitoring strategies and also toe the line between maximizing utility and preventing catastrophic risk.
  • Lead prioritization for safety efforts across the company for new model launches, understanding and addressing technical and business trade-offs.
  • Develop a broad range of subject matter expertise while maintaining agility across topics.
  • Work across many internal teams which will require high organizational acumen and confident decision making.

You might thrive in this role if you:

  • Have extensive experience researching LLMs, ML, AI, tech policy, moral reasoning, and/or enjoy classification problems.
  • Have extensive experience defining, refining and enforcing policies for ML models.
  • Deeply understand the operational challenges of enforcing policies with RLHF and can incorporate this into policy design.
  • Can analyze the benefits and risks of open-ended problem spaces; can generate ideas required to solve ambiguous problems and take full ownership of the solution.

Most Relevant Publications

  • Introducing HealthBench
  • Preparing for future AI capabilities in biology
  • Safety evaluations hub
  • OpenAI o1 System Card (Section 3)
  • GPT-4o System Card (Section 3)
  • GPT-4 System Card (Sections 2,3,4)
  • Evaluating Fairness in ChatGPT
  • Improving Model Safety Behavior with Rule-Based Rewards
  • OpenAI Model Spec

About OpenAI

OpenAI is an AI research and deployment company dedicated to ensuring that general-purpose artificial intelligence benefits all of humanity. We push the boundaries of the capabilities of AI systems and seek to safely deploy them to the world through our products. AI is an extremely powerful tool that must be created with safety and human needs at its core, and to achieve our mission, we must encompass and value the many different perspectives, voices, and experiences that form the full spectrum of humanity.

We are an equal opportunity employer, and we do not discriminate on the basis of race, religion, color, national origin, sex, sexual orientation, age, v

Employment Details

  • Salary: $325K
  • Location Type: Hybrid
  • Employment Type: FullTime

Skills

AI safety
Policy development
Taxonomy development
Risk assessment
Data analysis
Research
Cross-functional collaboration

OpenAI

Develops safe and beneficial AI technologies

About OpenAI

OpenAI develops and deploys artificial intelligence technologies aimed at benefiting humanity. The company creates advanced AI models capable of performing various tasks, such as automating processes and enhancing creativity. OpenAI's products, like Sora, allow users to generate videos from text descriptions, showcasing the versatility of its AI applications. Unlike many competitors, OpenAI operates under a capped profit model, which limits the profits it can make and ensures that excess earnings are redistributed to maximize the social benefits of AI. This commitment to safety and ethical considerations is central to its mission of ensuring that artificial general intelligence (AGI) serves all of humanity.

San Francisco, CaliforniaHeadquarters
2015Year Founded
$18,433.2MTotal Funding
LATE_VCCompany Stage
AI & Machine LearningIndustries
1,001-5,000Employees

Benefits

Health insurance
Dental and vision insurance
Flexible spending account for healthcare and dependent care
Mental healthcare service
Fertility treatment coverage
401(k) with generous matching
20-week paid parental leave
Life insurance (complimentary)
AD&D insurance (complimentary)
Short-term/long-term disability insurance (complimentary)
Optional buy-up life insurance
Flexible work hours and unlimited paid time off (we encourage 4+ weeks per year)
Annual learning & development stipend
Regular team happy hours and outings
Daily catered lunch and dinner
Travel to domestic conferences

Risks

Elon Musk's legal battle may pose financial and reputational challenges for OpenAI.
Customizable ChatGPT personas could lead to privacy and ethical concerns.
Competitors like Anthropic raising capital may intensify market competition.

Differentiation

OpenAI's capped profit model prioritizes ethical AI development over unlimited profit.
OpenAI's AI models, like Sora, offer unique video creation from text descriptions.
OpenAI's focus on AGI aims to create AI systems smarter than humans.

Upsides

OpenAI's $6.6 billion funding boosts its AI research and computational capacity.
Customizable ChatGPT personas enhance user engagement and satisfaction.
OpenAI's 'Operator' AI agent could revolutionize workforce automation by 2025.

Land your dream remote job 3x faster with AI