Aktiviere Job-Benachrichtigungen per E-Mail!

Research Engineer (Pre-training & Post-training)

WaveForms AI

San Francisco (CA)

Vor Ort

USD 120’000 - 180’000

Vollzeit

Vor 30+ Tagen

Erhöhe deine Chancen auf ein Interview

Erstelle einen auf die Position zugeschnittenen Lebenslauf, um deine Erfolgsquote zu erhöhen.

Zusammenfassung

An innovative firm is seeking a Research Engineer specializing in pre-training and post-training of large-scale language models. This role is pivotal in enhancing human-AI interactions through advanced audio intelligence. The ideal candidate will spearhead the optimization of model performance using cutting-edge techniques like reinforcement learning and generative modeling. You'll collaborate with cross-functional teams to develop robust data pipelines and ensure seamless integration of workflows. Join a forward-thinking company where your expertise will shape the future of multimodal AI systems and push the boundaries of technology.

Qualifikationen

  • 4-5 years of experience in training large language models and optimizing performance.
  • Strong background in distributed systems and scaling model training infrastructure.

Aufgaben

  • Lead pre-training and fine-tuning of large-scale language models for optimal performance.
  • Build scalable data pipelines for multimodal AI training involving audio and text.

Kenntnisse

Training large language models (LLMs)
Distributed systems
Compute efficiency
Data pipeline management
RLHF (Reinforcement Learning from Human Feedback)
Python
PyTorch
GANs (Generative Adversarial Networks)
Diffusion models
Cloud platforms (AWS, GCP, Azure)

Tools

Fully Sharded Data Parallel

Jobbeschreibung

Job title: Research Engineer (Pre-training & Post-training) / Member of Technical Staff

Who We Are
WaveForms AI is an Audio Large Language Models (LLMs) company building the future of audio intelligence through advanced research and products. Our models will transform human-AI interactions making them more natural, engaging and immersive.

Role overview: The Research Engineer – Pre-training & Post-training role integrates responsibilities across all phases of the AI model lifecycle, including pre-training, post-training, and data preparation. This position involves building and optimizing large-scale data pipelines, handling multimodal datasets (audio and text), conducting pre-training with a focus on compute efficiency and scalability, and refining models with cutting-edge techniques like supervised fine-tuning (SFT), reinforcement learning from human feedback (RLHF) and generative modeling. The ideal candidate will leverage advanced methods, including GANs and diffusion models, to push the boundaries of multimodal AI systems focused on audio and text.

Key Responsibilities

  • Lead the pre-training and fine-tuning of large-scale language models (LLMs), maximizing compute efficiency and scaling infrastructure.

  • Optimize model performance using advanced techniques, including RLHF, reward modeling (RM), instruction-tuning, distillation, GANs, and diffusion models.

  • Develop robust evaluation pipelines to monitor, refine, and improve model performance throughout training phases.

  • Build and optimize scalable, distributed data pipelines to support multimodal (audio + text) AI training.

  • Handle and process massive datasets (PiB scale) for pre-training and post-training, ensuring efficient preparation, annotation, and data flow.

  • Collaborate with research and engineering teams to ensure seamless integration of data preparation and training workflows for multimodal systems.

Required Skills & Qualifications

  • Proven experience in training large language models (LLMs), including pre-training, fine-tuning, and post-training optimization.

  • Strong background in distributed systems, compute efficiency, and scaling model training infrastructure.

  • Expertise in designing and managing large-scale, distributed data pipelines for multimodal datasets, particularly audio + text.

  • Proficiency in advanced techniques such as RLHF, instruction-tuning, reward modeling, distillation, GANs, and diffusion models.

  • Proficiency in Python, PyTorch, and distributed frameworks (e.g., Fully Sharded Data Parallel)

  • Familiarity with cloud platforms like AWS, GCP, or Azure for managing distributed environments.

  • Knowledge of multimodal AI systems combining audio and text for training and evaluation.

Minimum Experience

  • 4-5 years of relevant professional experience is required

Hol dir deinen kostenlosen, vertraulichen Lebenslauf-Check.
eine PDF-, DOC-, DOCX-, ODT- oder PAGES-Datei bis zu 5 MB per Drag & Drop ablegen.