
Enable job alerts via email!
Generate a tailored resume in minutes
Land an interview and earn more. Learn more
A leading research organization in London is hiring Evaluation Engineers to oversee evaluation campaigns for cutting-edge AI models. The role involves automating pipelines, improving evaluation processes, and working closely with frontier labs. Ideal candidates will have a strong background in Python and data analysis, and be passionate about AI model testing. This full-time, in-person position offers a competitive salary, flexible hours, and numerous benefits.
Applications deadline: We're accepting applications until 03 January 2026. We encourage early submissions and will start interviews in December 2025.
We’re looking for Evaluation Engineers who will run and own “evaluation campaigns” (pre-deployment testing for unreleased frontier models), build out our evaluation infrastructure, and automate the evals pipeline.
You will get to work with frontier labs like OpenAI, Anthropic, and Google DeepMind and be amongst the first to interact with new models before anyone else.
The ideal candidate loves rigorously testing frontier AI models, and enjoys building efficient pipelines and automating them.
The rapid rise in AI capabilities offer tremendous opportunities, but also present significant risks. At Apollo Research, we’re primarily concerned with risks from Loss of Control, i.e. risks coming from the model itself rather than e.g. humans misusing the AI. We’re particularly concerned with deceptive alignment / scheming, a phenomenon where a model appears to be aligned but is, in fact, misaligned and capable of evading human oversight. We work on the detection of scheming (e.g., building evaluations), the science of scheming (e.g., model organisms), and scheming mitigations (e.g., anti-scheming and control). We closely work with multiple frontier AI companies, e.g. to test their models before deployment or collaborate on scheming mitigations. At Apollo, we aim for a culture that emphasizes truth-seeking, being goal-oriented, giving and receiving constructive feedback, and being friendly and helpful. If you’re interested in more details about what it’s like working at Apollo, you can find more information here.
The current evals team consists of Jérémy Scheurer, Alex Meinke, Rusheb Shah, Bronson Schoen, Andrei Matveiakin, Felix Hofstätter, Axel Højmark, Teun van der Weij, Alex Lloyd, Alex Kedryk and Glen Rodgers. Alex Meinke leads and Marius Hobbhahn advises the evals team, though team members lead individual projects. You will mostly work with the evals team, but you will likely sometimes interact with the governance team to translate technical knowledge into concrete recommendations. You can find our full team here.
Equality Statement: Apollo Research is an Equal Opportunity Employer. We value diversity and are committed to providing equal opportunities to all, regardless of age, disability, gender reassignment, marriage and civil partnership, pregnancy and maternity, race, religion or belief, sex, or sexual orientation.
Please complete the application form with your CV. The provision of a cover letter is not necessary. Please also feel free to share links to relevant work samples. About the interview process: Our multi-stage process includes a screening interview, a take-home test (approx. 2.5 hours), 3 technical interviews, and a final interview with Marius (CEO). The technical interviews will be closely related to tasks the candidate would do on the job. There are no LeetCode-style general coding interviews. If you want to prepare for the interviews, we suggest running existing evaluations in Inspect evals and using those results to compare different models.
Applications deadline: We\'re accepting applications until 03 January 2026. We encourage early submissions and will start interviews in December 2025.
Your Privacy and Fairness in Our Recruitment Process: We are committed to protecting your data, ensuring fairness, and adhering to workplace fairness principles in our recruitment process. To enhance hiring efficiency, we use AI-powered tools to assist with tasks such as resume screening. These tools are designed and deployed in compliance with internationally recognized AI governance frameworks. Your personal data is handled securely and transparently. If you have questions about how your data is processed or wish to report concerns about fairness, please contact us at info@apolloresearch.ai.