Enable job alerts via email!
A leading company in AI safety seeks a Research Engineer to advance large language models through innovative reinforcement learning techniques. Collaborate with a diverse team to enhance reasoning abilities in code generation and mathematics. Ideal candidates have strong software engineering skills and experience with deep learning frameworks. Embrace a culture of collaboration and impact in AI research.
London, UK
Anthropic’s mission is to create reliable, interpretable, and steerable AI systems. We aim for AI to be safe and beneficial for users and society. Our team is a growing group of researchers, engineers, policy experts, and business leaders working together to build beneficial AI systems.
As a Research Engineer on the Reinforcement Learning Fundamentals team, you will collaborate with researchers and engineers to advance the capabilities and safety of large language models through fundamental research in reinforcement learning, enhancing reasoning abilities in areas like code generation and mathematics, and exploring reinforcement learning for agentic / open-ended tasks.
Deadline to apply: Rolling review; no fixed deadline.
The expected salary range is provided separately.
Education: Bachelor's degree or equivalent experience in a related field.
Location-based hybrid policy: Expectation of being in the office at least 25% of the time; some roles may require more.
Visa sponsorship: We sponsor visas when possible; we will make reasonable efforts to assist if you are offered the role.
We encourage applicants from diverse backgrounds and those who may not meet every qualification to apply.
We focus on big science and impactful research, valuing collaboration, communication, and impact in AI safety and steerability. Our research directions include GPT-3, interpretability, multimodal neurons, scaling laws, and more.
Based in San Francisco, we offer competitive compensation, benefits, flexible hours, and a collaborative environment.
Fields marked with * are required.