Contract Type: 6 month contract outsourced via agency on an hourly rate
Location: Egham
Hybrid: 3 days onsite (minimum) and 2 days working from home
Rate: Very much dependant on level of experience.
Key responsibilities include:
- Performance Optimization: Profile and debug performance bottlenecks at the OS, runtime, and model levels.
- Model Deployment: Work across the stack—from model conversion, quantization, and optimization to runtime integration of AI models on-device.
- Toolchain Evaluation: Compare deployment toolchains and runtimes for latency, memory, and accuracy trade-offs.
- Open-Source Contribution: Enhance open-source libraries by adding new features and improving capabilities.
- Experimentation & Analysis: Conduct rigorous experiments and statistical analysis to evaluate algorithms and systems.
- Prototyping: Lead the development of software prototypes and experimental systems with high code quality.
- Collaboration: Work closely with a multidisciplinary team of researchers and engineers to integrate research findings into products.
We not require a PhD holder this time which is unusual for the AI Team.
We're looking for someone with:
- Technical Expertise: Strong OS fundamentals (memory management, multithreading, user/kernel mode interaction) and expertise in ARM CPU architectures.
- Programming Skills: Expert proficiency in Python and Rust, with desirable knowledge in C and C++.
- AI Knowledge: Solid understanding of machine learning and deep learning fundamentals, including architectures and evaluation metrics.
- Problem-Solving: Strong analytical skills and the ability to design and conduct rigorous experiments.
- Team Player: Excellent communication and collaboration skills, with a results-oriented attitude
Desirable Skills:
- Experience with ARM 64-bit architecture and CPU hardware architectures.
- Knowledge of trusted execution environments (confidential computing).
- Hands-on experience with deep learning model optimization (quantization, pruning, distillation).
- Familiarity with lightweight inference runtimes (ExecuTorch, llama.cpp, Candle).