Enable job alerts via email!

Member of Technical Staff - Edge AI Inference Engineer

Liquid AI

San Francisco (CA)

On-site

USD 60,000 - 240,000

Full time

2 days ago
Be an early applicant

Boost your interview chances

Create a job specific, tailored resume for higher success rate.

Job summary

Liquid AI is seeking a Member of Technical Staff for Edge AI Inference Engineering. This role involves optimizing AI model inference stacks across diverse hardware platforms including CPUs and GPUs. Ideal candidates will possess deep technical expertise and have a passion for enhancing AI system performance.

Qualifications

  • Extensive experience in inference on embedded hardware.
  • Deep understanding of CPU, NPU, and GPU architectures.
  • Proficiency in building and enhancing edge inference stacks.

Responsibilities

  • Optimize inference stacks tailored to various edge device types.
  • Work with frameworks like llama.cpp, Executorch, and TensorRT.
  • Independently deliver optimized edge inference stack.

Skills

ML Experience
Hardware Awareness
Proficient in Coding
Optimization of Low-Level Primitives
Self-Guided and Ownership

Job description

Member of Technical Staff - Edge AI Inference Engineer

Join to apply for the Member of Technical Staff - Edge AI Inference Engineer role at Liquid AI

Member of Technical Staff - Edge AI Inference Engineer

1 month ago Be among the first 25 applicants

Join to apply for the Member of Technical Staff - Edge AI Inference Engineer role at Liquid AI

Liquid AI, an MIT spin-off, is a foundation model company headquartered in Boston, Massachusetts. Our mission is to build capable and efficient general-purpose AI systems at every scale.

Our goal at Liquid is to build the most capable AI systems to solve problems at every scale, such that users can build, access, and control their AI solutions. This is to ensure that AI will get meaningfully, reliably and efficiently integrated at all enterprises. Long term, Liquid will create and deploy frontier-AI-powered solutions that are available to everyone.

What this role actually is:

As we prepare to deploy our models across various edge device types, including CPUs, embedded GPUs, and NPUs, we seek an expert to optimize inference stacks tailored to each platform. We're looking for someone who can take our models, dive deep into the task, and return with a highly optimized inference stack, leveraging existing frameworks like llama.cpp, Executorch, and TensorRT to deliver exceptional throughput and low latency.

The ideal candidate is a highly skilled engineer with extensive experience in inference on embedded hardware and a deep understanding of CPU, NPU, and GPU architectures. They should be self-motivated, capable of working independently, and driven by a passion for optimizing performance across diverse edge hardware platforms.

Proficiency in building and enhancing edge inference stacks is essential. Additionally, experience with mobile development and expertise in cache-aware algorithms will be highly valued.

Responsibilities

  • Strong ML Experience: Proficiency in Python and PyTorch to effectively interface with the ML team at a deeply technical level
  • Hardware Awareness: Must understand modern hardware architecture, including cache hierarchies and memory access patterns, and their impact on performance
  • Proficient in Coding: Expertise in Python, C++, or Rust for AI-driven real-time embedded systems
  • Optimization of Low-Level Primitives: Responsible for optimizing core primitives to ensure efficient model execution
  • Self-Guided and Ownership: Ability to independently take a PyTorch model and inference requirements and deliver a fully optimized edge inference stack with minimal guidance

Seniority level
  • Seniority level
    Not Applicable
Employment type
  • Employment type
    Full-time
Job function
  • Job function
    Engineering and Information Technology
  • Industries
    Software Development

Referrals increase your chances of interviewing at Liquid AI by 2x

Get notified about new Member of Technical Staff jobs in San Francisco, CA.

Berkeley, CA $60,000.00-$240,000.00 1 year ago

Consulting Member of Technical Staff (IC5)
Member of Technical Staff - Computational Biologist
Member of Technical Staff - Head of Engineering

San Francisco, CA $101,500.00-$156,750.00 2 months ago

Alameda, CA $96,000.00-$137,500.00 8 hours ago

San Francisco, CA $136,947.00-$239,699.00 5 months ago

San Francisco, CA $90,000.00-$130,000.00 6 months ago

Associate Director of Counseling & Psychological Services - (Administrator II) - Counseling and Psychological Services

San Francisco, CA $10,000.00-$120,000.00 7 months ago

Project Archaeologist/ Cultural Resources Specialist

San Francisco, CA $141,800.00-$221,600.00 6 hours ago

San Mateo, CA $141,800.00-$221,600.00 6 hours ago

San Francisco, CA $116,000.00-$200,100.00 1 day ago

Member of Technical Staff (Senior/Staff)

San Francisco, CA $145,000.00-$220,000.00 3 months ago

Member of Technical Staff - General Interest
Member of Technical Staff - Compute Platform
Quantum Engineer - Member of Technical Staff

San Francisco, CA $120,000.00-$180,000.00 1 month ago

Member of Technical Staff, Founding Design Engineer

San Francisco, CA $130,000.00-$200,000.00 5 months ago

Member of Technical Staff, Founding Frontend Engineer

San Francisco, CA $130,000.00-$200,000.00 4 months ago

Member of Technical Staff, Founding Backend Engineer

San Francisco, CA $150,000.00-$200,000.00 5 months ago

San Francisco, CA $160,000.00-$175,000.00 4 days ago

San Francisco, CA $200,000.00-$260,000.00 2 weeks ago

San Francisco, CA $200,000.00-$260,000.00 4 days ago

We’re unlocking community knowledge in a new way. Experts add insights directly into each article, started with the help of AI.

Get your free, confidential resume review.
or drag and drop a PDF, DOC, DOCX, ODT, or PAGES file up to 5MB.

Similar jobs

Member of Technical Staff - Edge AI Inference Engineer

Liquid AI

San Francisco

On-site

USD 60,000 - 240,000

17 days ago

Software Engineer, ML Inference Compiler & Deployment, AI Frameworks

Tesla

Palo Alto

On-site

USD 132,000 - 390,000

Yesterday
Be an early applicant

Member of Technical Staff, Post-Training

Microsoft

Mountain View

On-site

USD 117,000 - 230,000

23 days ago

AI Software Engineer, Inference

Nexus

San Francisco

On-site

USD 135,000 - 200,000

21 days ago

Member of Technical Staff - Copilot AI Web Engineering Manager

Microsoft

Mountain View

On-site

USD 188,000 - 305,000

17 days ago

Member of Technical Staff, MLE (North)

Cohere

New York

Remote

USD 120,000 - 180,000

30+ days ago

Member of Technical Staff - Pretraining / Inference Optimization Freiburg (Germany), San Franci[...]

Global Trade Plaza

Mississippi

Remote

USD 60,000 - 80,000

30+ days ago

Member of Technical Staff

dynamism vc

San Francisco

On-site

USD 200,000 - 250,000

30+ days ago

Member of Technical Staff - Agentic Systems

Context

Palo Alto

On-site

USD 120,000 - 160,000

26 days ago