Research Engineer / Scientist, Alignment Science, London at Anthropic

London, England, United Kingdom

Anthropic Logo
Not SpecifiedCompensation
Senior (5 to 8 years)Experience Level
Full TimeJob Type
UnknownVisa
Artificial IntelligenceIndustries

Requirements

  • Significant software, ML, or research engineering experience
  • Based at least 25% in London with occasional travel to San Francisco

Responsibilities

  • Testing the robustness of our safety techniques by training language models to subvert our safety techniques, and seeing how effective they are at subverting our interventions
  • Run multi-agent reinforcement learning experiments to test out techniques like AI Debate
  • Build tooling to efficiently evaluate the effectiveness of novel LLM-generated jailbreaks
  • Write scripts and prompts to efficiently produce evaluation questions to test models’ reasoning abilities in safety-relevant contexts
  • Contribute ideas, figures, and writing to research papers, blog posts, and talks
  • Run experiments that feed into key AI safety efforts at Anthropic, like the design and implementation of our Responsible Scaling Policy

Skills

Key technologies and capabilities for this role

Machine LearningSoftware EngineeringResearch EngineeringAI SafetyAI AlignmentInterpretabilityFine-TuningAI Control

Questions & Answers

Common questions about this position

What is the work location and arrangement for this role?

The position is hybrid, requiring candidates to be based at least 25% or more in London with occasional travel to San Francisco. The team's hub is currently in San Francisco.

What is the salary for this Research Engineer position?

This information is not specified in the job description.

What skills and experience are required for this role?

Significant software, ML, or research engineering experience is required. The role involves building and running machine learning experiments, testing safety techniques, and contributing to research papers.

What is the company culture like at Anthropic?

Anthropic has a quickly growing team of committed researchers, engineers, policy experts, and business leaders working together to build safe and beneficial AI systems.

What makes a strong candidate for this Alignment Science role?

Strong candidates are those who can describe themselves as both scientists and engineers, with significant experience in software, ML, or research engineering, and a passion for building elegant experiments to ensure AI is helpful, honest, and harmless.

Anthropic

Develops reliable and interpretable AI systems

About Anthropic

Anthropic focuses on creating reliable and interpretable AI systems. Its main product, Claude, serves as an AI assistant that can manage tasks for clients across various industries. Claude utilizes advanced techniques in natural language processing, reinforcement learning, and code generation to perform its functions effectively. What sets Anthropic apart from its competitors is its emphasis on making AI systems that are not only powerful but also understandable and controllable by users. The company's goal is to enhance operational efficiency and improve decision-making for its clients through the deployment and licensing of its AI technologies.

San Francisco, CaliforniaHeadquarters
2021Year Founded
$11,482.1MTotal Funding
GROWTH_EQUITY_VCCompany Stage
Enterprise Software, AI & Machine LearningIndustries
1,001-5,000Employees

Benefits

Flexible Work Hours
Paid Vacation
Parental Leave
Hybrid Work Options
Company Equity

Risks

Ongoing lawsuit with Concord Music Group could lead to financial liabilities.
Technological lag behind competitors like OpenAI may impact market position.
Reliance on substantial funding rounds may indicate financial instability.

Differentiation

Anthropic focuses on AI safety, contrasting with competitors' commercial priorities.
Claude, Anthropic's AI assistant, is designed for tasks of any scale.
Partnerships with tech giants like Panasonic and Amazon enhance Anthropic's strategic positioning.

Upsides

Anthropic's $60 billion valuation reflects strong investor confidence and growth potential.
Collaborations like the Umi app with Panasonic tap into the growing wellness AI market.
Focus on AI safety aligns with increasing industry emphasis on ethical AI development.

Land your dream remote job 3x faster with AI