Enable job alerts via email!
A pioneering fintech company in the digital finance sector is seeking an AI model team member. You will drive innovations in AI architecture development while enhancing the performance capabilities of various models. Ideal candidates will have a PhD in a related field and hands-on experience with large-scale model training. Join a global team dedicated to pushing the boundaries of technology and innovation in fintech.
Join Tether and shape the future of digital finance. At Tether, we're not just building products; we're pioneering a global financial revolution. Our solutions empower businesses—from exchanges and wallets to payment processors and ATMs—to seamlessly integrate reserve-backed tokens across blockchains. By harnessing blockchain technology, Tether enables you to store, send, and receive digital tokens instantly, securely, and globally, at a fraction of the cost. Transparency underpins our operations, ensuring trust in every transaction.
Our global team works remotely from around the world. If you're passionate about fintech and have excellent English communication skills, this is your chance to collaborate with top minds, push boundaries, and set new standards. We've grown fast, stayed lean, and secured our leadership position. Join us to contribute to the most innovative platform on the planet.
As a member of the AI model team, you'll drive innovation in architecture development for models of various scales, including small, large, and multi-modal systems. Your work will enhance intelligence, efficiency, and capabilities in AI. You should have deep expertise in LLM architectures, pre-training optimization, and a research-driven approach. Your mission includes exploring and implementing novel techniques to advance AI performance, including data curation, improving baselines, and resolving pre-training bottlenecks.
Requirements include a degree in Computer Science or related fields, ideally a PhD in NLP, Machine Learning, or similar, with a strong record in AI R&D. Hands-on experience with large-scale LLM training on distributed GPU servers, familiarity with training frameworks and libraries like PyTorch and Hugging Face, and deep knowledge of transformer architectures are essential.