Enable job alerts via email!

Model Optimization Engineer (PyTorch Infrastructure Development)

Apple Inc.

Cupertino (CA)

On-site

USD 143,000 - 265,000

Full time

7 days ago
Be an early applicant

Boost your interview chances

Create a job specific, tailored resume for higher success rate.

Job summary

Apple is seeking a Model Optimization Engineer to develop and optimize algorithms within their Core ML stack. You will design APIs, manage training jobs, and collaborate with teams to implement cutting-edge compression techniques, contributing to advancements in machine learning applications.

Benefits

Comprehensive medical and dental coverage
Retirement benefits
Discounted products and free services
Educational reimbursement
Discretionary bonuses available
Employee stock purchase plan

Qualifications

  • 3+ years of industry and/or research experience required.
  • Proficient in ML authoring frameworks.
  • Experience with optimization libraries for ML frameworks.

Responsibilities

  • Design and develop core infrastructure for compression algorithms.
  • Collaborate with engineering teams for new compression operations.
  • Run experiments and profile algorithms on models.

Skills

Python programming
Machine Learning
Model Compression

Education

Bachelors in Computer Sciences, Engineering, or related discipline

Tools

PyTorch
TensorFlow
JAX

Job description

Cupertino, California, United States Software and Services

Description

We work on a python library that implements a variety of training time and post training quantization algorithms and provides them to developers as simple to use, turnkey APIs, and ensures that these optimizations work seamlessly with the Core ML inference stack and Apple hardware. Our algorithms are implemented using PyTorch. We optimize models across domains, including NLP, vision, text, generative models etc.In this role, the Model Optimization Engineer will be an expert in understanding the internal workings of PyTorch, graph capturing and graph editing mechanisms, methods to observe and modify intermediate activations and weights, tensor subclasses, custom ops, different types of parallelism for training models, and use this knowledge to implement and update the core infrastructure of the optimization library which enables an efficient and scalable implementation of various classes of compression algorithms. You'll also set up and debug training jobs, datasets, evaluation, performance benchmarking pipelines.Additionally, you will...- Design and develop the core infrastructure which powers the implementations of various compression algorithms (training time, post training, data free, calibration data based etc)- Implement the latest algorithms from research papers for model compression in the optimization library.- Collaborate with software and hardware engineers, from the ML compiler inference stack, to co-develop new compression operations, and model export flows for on device deployment.- Design clean, intuitive, maintainable APIs- Run detailed experiments and ablation studies to profile algorithms on various models and tasks, across different model sizes.

Minimum Qualifications
  • Bachelors in Computer Sciences, Engineering, or related discipline.
  • 3+ years of industry and/or research experience
  • Highly proficient in Python programming
  • Proficiency in at least one ML authoring framework, such as PyTorch, TensorFlow, JAX, MLX
  • Experience in the area of model compression and quantization techniques, specially in one of the optimization libraries for an ML framework (e.g. torch.ao).
  • Ability to ramp up quickly on new training code bases and run experiments.
Preferred Qualifications
  • Demonstrated ability to design user friendly and maintainable APIs
  • Experience in training, fine tuning, and optimizing neural network models
  • Primary contributor to a model optimization/compression library.
  • Self prioritize and adjust to changing priorities and asks
  • Improving model optimization documentation, writing tutorials and guides
  • Good communication skills, including ability to communicate with cross-functional audiences
At Apple, base pay is one part of our total compensation package and is determined within a range. This provides the opportunity to progress as you grow and develop within a role. The base pay range for this role is between $143,100 and $264,200, and your base pay will depend on your skills, qualifications, experience, and location.

Apple employees also have the opportunity to become an Apple shareholder through participation in Apple’s discretionary employee stock programs. Apple employees are eligible for discretionary restricted stock unit awards, and can purchase Apple stock at a discount if voluntarily participating in Apple’s Employee Stock Purchase Plan. You’ll also receive benefits including: Comprehensive medical and dental coverage, retirement benefits, a range of discounted products and free services, and for formal education related to advancing your career at Apple, reimbursement for certain educational expenses — including tuition. Additionally, this role might be eligible for discretionary bonuses or commission payments as well as relocation. Learn more about Apple Benefits.

Note: Apple benefit, compensation and employee stock programs are subject to eligibility requirements and other terms of the applicable plan or program.

Apple is an equal opportunity employer that is committed to inclusion and diversity. We seek to promote equal opportunity for all applicants without regard to race, color, religion, sex, sexual orientation, gender identity, national origin, disability, Veteran status, or other legally protected characteristics. Learn more about your EEO rights as an applicant .

Get your free, confidential resume review.
or drag and drop a PDF, DOC, DOCX, ODT, or PAGES file up to 5MB.

Similar jobs

Model Optimization Engineer (PyTorch Infrastructure Development)

Apple

Cupertino

On-site

USD 143,000 - 265,000

5 days ago
Be an early applicant