Enable job alerts via email!

AI Research Engineer (Model Evaluation)

Tether Operations Limited

London

Remote

GBP 50,000 - 80,000

Full time

2 days ago
Be an early applicant

Boost your interview chances

Create a job specific, tailored resume for higher success rate.

Job summary

Join Tether as an AI evaluation framework developer and play a key role in pioneering the future of digital finance. You will work with advanced model architectures, develop metrics for performance indicators, and collaborate with teams to enhance fintech solutions. Your expertise in AI methodologies and strong programming skills will contribute to shaping innovative projects in a fully remote environment.

Qualifications

  • Proficient in developing assessment frameworks for AI models.
  • Strong programming skills with experience in scalable evaluation pipelines.
  • Ability to translate technical insights into actionable recommendations.

Responsibilities

  • Develop and deploy evaluation frameworks tracking KPIs like accuracy, latency, and memory usage.
  • Collaborate with various teams to align evaluation metrics with business goals.
  • Conduct research to refine evaluation methodologies.

Skills

AI model evaluation
Programming
Collaboration
Data analysis
Iterative experiments

Education

Degree in Computer Science
PhD in NLP or Machine Learning

Job description

Join Tether and Shape the Future of Digital Finance

At Tether, we're pioneering a global financial revolution with innovative solutions that enable seamless integration of reserve-backed tokens across blockchains. Our offerings include the trusted stablecoin USDT, energy solutions for Bitcoin mining, data sharing apps like KEET, and educational initiatives to democratize digital learning.

We are a global, remote team looking for passionate individuals to contribute to our cutting-edge projects. If you have excellent English communication skills and a drive to innovate in fintech, Tether is the place for you.

About the job:

As part of our AI model team, you will develop evaluation frameworks for pre-training, post-training, and inference stages. Your focus will be on designing metrics and assessment strategies to ensure models are responsive, efficient, and reliable across various applications, from resource-limited devices to multi-modal architectures.

You should have expertise in advanced model architectures, evaluation practices, and benchmarking. Your work will involve developing novel evaluation strategies, testing, and implementing them to track performance indicators such as accuracy, latency, throughput, and memory footprint. Collaboration with cross-functional teams to share findings and improve deployment strategies will be key.

Responsibilities:

  • Develop and deploy evaluation frameworks assessing models at all stages, tracking KPIs like accuracy, latency, and memory usage.
  • Create high-quality datasets and benchmarks to reliably measure model robustness and improvements.
  • Collaborate with product, engineering, and operations teams to align evaluation metrics with business goals, presenting insights via dashboards and reports.
  • Analyze evaluation data to identify bottlenecks and propose optimizations for model performance and resource utilization.
  • Conduct research to refine evaluation methodologies, staying updated on emerging techniques to enhance benchmarking and model reliability.

Minimum requirements:

  • A degree in Computer Science or related field; PhD in NLP, Machine Learning, or similar is preferred, with a strong publication record in top conferences.
  • Experience in designing and evaluating AI models across different stages, with proficiency in developing assessment frameworks.
  • Strong programming skills and experience building scalable evaluation pipelines; familiarity with performance metrics like latency, throughput, and memory footprint.
  • Ability to conduct iterative experiments and research to improve evaluation practices.
  • Experience collaborating with diverse teams and translating technical insights into actionable recommendations.
Get your free, confidential resume review.
or drag and drop a PDF, DOC, DOCX, ODT, or PAGES file up to 5MB.

Similar jobs

AI Research Engineer (Model Evaluation)

Tether Operations Limited

London

Remote

GBP 70,000 - 110,000

2 days ago
Be an early applicant

AI Research Engineer (Model Evaluation)

Tether Operations Limited

London

Remote

GBP 70,000 - 100,000

3 days ago
Be an early applicant

AI Research Engineer (Model Serving & Inference)

Tether Operations Limited

London

Remote

GBP 60,000 - 100,000

Yesterday
Be an early applicant

AI Research Engineer (Model Serving & Inference)

Tether Operations Limited

London

Remote

GBP 60,000 - 85,000

3 days ago
Be an early applicant

AI Research Engineer (Model Evaluation - 100%)

Tether Operations Limited

London

On-site

GBP 70,000 - 100,000

2 days ago
Be an early applicant

AI Research Engineer (Fine-tuning)

Tether Operations Limited

London

Remote

GBP 60,000 - 100,000

3 days ago
Be an early applicant

AI Research Engineer (Fine-tuning - 100% Remote UK)

Tether Operations Limited

London

Remote

GBP 60,000 - 100,000

5 days ago
Be an early applicant

Research Engineer - Brain-Inspired AI Systems (Contractor)

Huawei R&D UK

London

On-site

GBP 50,000 - 80,000

5 days ago
Be an early applicant

Research Engineer, Data (Foundational Research, Machine Learning)

Refinitiv

London

Hybrid

GBP 50,000 - 80,000

2 days ago
Be an early applicant