Overview
Join Tether and Shape the Future of Digital Finance. At Tether, we’re building solutions that enable reserve-backed tokens across blockchains. Our technology supports storing, sending, and receiving digital tokens globally with transparency and trust.
Why Join Us
Our team is a global, remote workforce. If you’re passionate about fintech, collaborate with leading minds and contribute to a fast-growing leader in the industry. Excellent English communication skills are a plus.
About the job
We’re seeking experienced AI infrastructure engineers to design and implement robust, scalable pipelines for massive data workloads. Join Tether’s applied research team to contribute to high-impact projects across thousands of GPUs and video generation foundation development.
Responsibilities
- Build and scale high-throughput data infrastructure optimized for video and multimodal content processing across large GPU clusters (e.g., H100/H200).
- Design core preprocessing algorithms for video, audio, text, and image modalities, enabling efficient extraction, synchronization, and normalization of temporal data.
- Build automated acquisition pipelines for sourcing large-scale video datasets, handling diverse formats, frame rates, annotations, and embedded audio.
- Architect robust systems for scalable evaluation and annotation, including prompt-based scoring, perceptual metrics, caption generation, and retrieval-based diagnostics.
- Collaborate with model researchers to co-design video model architectures (e.g. DiTs, VAEs, spatio-temporal transformers) and training schedules across pretraining and fine-tuning stages.
- Optimize distributed data loading and pipeline throughput for training at scale, ensuring robustness across model variants and modality combinations.
- Manage infrastructure to support experiment tracking, model versioning, and cross-team deployment workflows, integrating with production and research platforms.
- Support backend engineering across research, product, and creative teams to ensure seamless integration of data and model workflows from prototyping to inference.
- Proficient in Python with strong programming skills across backend, infrastructure, and data tooling domains.
- Strong software engineering experience, including 2+ years working with petabyte-scale data pipelines and systems across thousands of GPUs.
- Proven ability to architect and maintain large-scale distributed systems for data processing and delivery.
- Deep expertise in orchestration frameworks such as Kubernetes and SLURM with hands-on experience deploying and managing high-throughput workloads.
Preferred Qualifications
- Practical experience on building pipelines and infrastructure with visual and multimodal datasets, including image/video pipelines.
- Experience in building video foundation infrastructure pipelines and workflows with collaboration of LLM and/or video foundation research and engineering teams is a strong advantage.
Important information for candidates
- Apply only through our official channels. We do not use third-party platforms or agencies for recruitment unless clearly stated. All open roles are listed on our official careers page: tether.recruitee.com
- Verify the recruiter’s identity. All our recruiters have verified LinkedIn profiles. If you’re unsure, confirm their identity by checking their profile or contacting us through our website.
- Be cautious of unusual communication methods. We do not conduct interviews over WhatsApp, Telegram, or SMS. All communication is done through official company emails and platforms.
- Double-check email addresses. All communication from us will come from emails ending in tether.to or tether.io.
- We will never request payment or financial details. If someone asks for personal financial information or payment during the hiring process, it is a scam. Please report it immediately.
When in doubt, feel free to reach out through our official website.