
Enable job alerts via email!
A future-focused AI consultancy is looking for QA contributors to validate and improve AI agents' evaluation frameworks. Ideal candidates will possess strong analytical skills and attention to detail while enjoying a flexible, remote work environment. Responsibilities include reviewing evaluation tasks, defining expected behaviors for agents, and collaborating with cross-functional teams to enhance AI systems. This role provides an opportunity to influence future AI technologies and build a unique portfolio.
We’re looking for curious and intellectually proactive contributors who double‑check assumptions and play devil’s advocate. Are you comfortable with ambiguity and complexity? Does an async, remote, flexible opportunity sound exciting? Would you like to learn how modern AI systems are tested and evaluated?
We’re on the hunt for QAs for autonomous AI agents for a new project focused on validating and improving complex task structures, policy logic, and agent evaluation frameworks. Throughout the project, you’ll have to balance quality assurance, research, and logical problem‑solving. This project opportunity is ideal for people who enjoy looking at systems holistically and thinking through scenarios, implications, and edge cases.
Apply to this post, qualify, and get the chance to contribute to a project aligned with your skills, on your own schedule. Shape the future of AI while building tools that benefit everyone.