Let’s get started
By clicking ‘Next’, I agree to the Terms of Service
and Privacy Policy
Jobs / Job page
Member of Technical Staff - Edge AI Inference Engineer image - Rise Careers
Job details

Member of Technical Staff - Edge AI Inference Engineer

Liquid AI, an MIT spin-off, is a foundation model company headquartered in Boston, Massachusetts. Our mission is to build capable and efficient general-purpose AI systems at every scale.


Our goal at Liquid is to build the most capable AI systems to solve problems at every scale, such that users can build, access, and control their AI solutions. This is to ensure that AI will get meaningfully, reliably and efficiently integrated at all enterprises. Long term, Liquid will create and deploy frontier-AI-powered solutions that are available to everyone.


What this role actually is:


As we prepare to deploy our models across various edge device types, including CPUs, embedded GPUs, and NPUs, we seek an expert to optimize inference stacks tailored to each platform. We're looking for someone who can take our models, dive deep into the task, and return with a highly optimized inference stack, leveraging existing frameworks like llama.cpp, Executorch, and TensorRT to deliver exceptional throughput and low latency.


The ideal candidate is a highly skilled engineer with extensive experience in inference on embedded hardware and a deep understanding of CPU, NPU, and GPU architectures. They should be self-motivated, capable of working independently, and driven by a passion for optimizing performance across diverse edge hardware platforms.


Proficiency in building and enhancing edge inference stacks is essential. Additionally, experience with mobile development and expertise in cache-aware algorithms will be highly valued.


Responsibilities
  • Strong ML Experience: Proficiency in Python and PyTorch to effectively interface with the ML team at a deeply technical level.
  • Hardware Awareness: Must understand modern hardware architecture, including cache hierarchies and memory access patterns, and their impact on performance.
  • Proficient in Coding: Expertise in Python, C++, or Rust for AI-driven real-time embedded systems
  • Optimization of Low-Level Primitives: Responsible for optimizing core primitives to ensure efficient model execution.
  • Self-Guided and Ownership: Ability to independently take a PyTorch model and inference requirements and deliver a fully optimized edge inference stack with minimal guidance


Average salary estimate

$140000 / YEARLY (est.)
min
max
$120000K
$160000K

If an employer mentions a salary or salary range on their job, we display it as an "Employer Estimate". If a job has no salary data, Rise displays an estimate if available.

What You Should Know About Member of Technical Staff - Edge AI Inference Engineer, Liquid AI

At Liquid AI, we're on a mission to revolutionize the way artificial intelligence is integrated into everyday enterprises, and we're excited to announce that we're looking for a talented Member of Technical Staff - Edge AI Inference Engineer to join our dynamic team. Based in Boston, but with flexibility in location, this role is all about optimizing inference stacks for a variety of edge devices like CPUs, embedded GPUs, and NPUs. If you're a passionate engineer who is deeply familiar with hardware architectures and has hands-on experience with frameworks such as llama.cpp, Executorch, and TensorRT, this is the job for you! You'll dive deep into technical challenges, crafting high-performance inference solutions that push the limits of what's possible with edge hardware. We're seeking self-motivated individuals who thrive on ownership and aim to enhance model throughput and reduce latency. With a strong foundation in Python and PyTorch, you will be collaborating closely with the machine learning team to ensure seamless integration and execution of AI models. Fluidity in C++, Rust, or similar languages will also be beneficial! If you enjoy the thrill of optimizing low-level machine primitives and you have experience with mobile development, we would love to hear from you. At Liquid AI, we believe in empowering our team members to take initiative and deliver killer solutions that help shape the future of AI. Join us on this incredible journey and let's create the next generation of AI together!

Frequently Asked Questions (FAQs) for Member of Technical Staff - Edge AI Inference Engineer Role at Liquid AI
What are the key responsibilities of a Member of Technical Staff - Edge AI Inference Engineer at Liquid AI?

The Member of Technical Staff - Edge AI Inference Engineer at Liquid AI is responsible for optimizing inference stacks for various edge hardware platforms. This involves understanding modern hardware architectures and performance optimization, effectively working with frameworks such as llama.cpp, Executorch, and TensorRT, and delivering optimized solutions for AI models. You'll work closely with the machine learning team to enhance model execution and ensure low latency in real-time AI applications.

Join Rise to see the full answer
What qualifications are required for the Edge AI Inference Engineer role at Liquid AI?

Applicants for the Edge AI Inference Engineer position at Liquid AI should have extensive experience with inference on embedded hardware, proficiency in Python and PyTorch, and a strong understanding of CPU, NPU, and GPU architectures. Additionally, expertise in C++, Rust, or similar languages is beneficial, as is familiarity with cache-aware algorithms and mobile development.

Join Rise to see the full answer
How does an Edge AI Inference Engineer contribute to product development at Liquid AI?

An Edge AI Inference Engineer plays a critical role in product development at Liquid AI by optimizing machine learning model performance on embedded devices. By ensuring efficient execution of AI models tailored to various hardware architectures, they enable our products to achieve exceptional throughput and low latency, thereby enhancing the overall user experience with AI-powered solutions.

Join Rise to see the full answer
What skills are essential for success as a Member of Technical Staff - Edge AI Inference Engineer?

To succeed as a Member of Technical Staff - Edge AI Inference Engineer at Liquid AI, you should have a strong background in machine learning, proficiency in relevant programming languages, and the ability to work independently. Understanding hardware implications on performance is crucial, as is experience with optimizing low-level primitives. Additionally, you should have a passion for exploring new challenges in edge AI systems.

Join Rise to see the full answer
What kind of work environment can I expect as an Edge AI Inference Engineer at Liquid AI?

At Liquid AI, you can expect a collaborative and innovative work environment where creativity and technical expertise are highly valued. We emphasize team ownership and encourage self-motivated individuals to take initiative. Our culture fosters continuous learning and exploration, with opportunities to contribute to meaningful projects that advance the future of AI technology.

Join Rise to see the full answer
Common Interview Questions for Member of Technical Staff - Edge AI Inference Engineer
How would you optimize an AI model for edge devices?

To optimize an AI model for edge devices, I'd start by profiling the model to identify bottlenecks, followed by streamlining the computation by leveraging optimized inference frameworks like TensorRT. I would ensure that the model memory usage is efficient and reduce its complexity without sacrificing performance, potentially exploring quantization techniques.

Join Rise to see the full answer
What experience do you have with optimizing low-level primitives in AI?

In my previous role, I worked extensively on optimizing low-level primitives for real-time AI applications, focusing on enhancing performance through memory-efficient algorithms and cache optimization. I'm proficient in tools like PyTorch for implementing these optimizations and have a good grasp of hardware architectures.

Join Rise to see the full answer
Can you explain your understanding of cache hierarchies and their impact on model execution?

Cache hierarchies are critical for improving data access speeds during model execution. I understand how different levels of cache (L1, L2, L3) can be leveraged to keep frequently accessed data readily available, minimizing latency. My approach involves optimizing data structures to fit within these caches effectively.

Join Rise to see the full answer
What frameworks are you familiar with for optimizing inference, and how have you applied them?

I have experience with several optimization frameworks including llama.cpp, Executorch, and TensorRT. I've applied these frameworks to reduce model inference time considerably by implementing layer fusion and weight pruning, ultimately leading to smooth execution on edge devices.

Join Rise to see the full answer
Describe your coding experience in Python and how it relates to AI.

I have extensive coding experience in Python, particularly in relation to machine learning libraries such as PyTorch. This expertise has empowered me to interface effectively with ML teams, develop prototypes quickly, and analyze model performance, directly translating to impactful AI solutions.

Join Rise to see the full answer
How do you ensure code quality in your projects?

I ensure code quality by adhering to best practices, utilizing code reviews, and implementing unit tests during the development process. This attention to detail not only helps in maintaining efficiency but also ensures that optimal performance is consistently achieved during model execution.

Join Rise to see the full answer
Describe a challenging problem you faced with hardware optimization and how you resolved it.

One challenging problem I faced was processing speed bottlenecks on an embedded GPU. I had to profile the model and identified inefficient memory access patterns. By reorganizing the data flow and optimizing the model structure, I dramatically enhanced the processing speed and performance.

Join Rise to see the full answer
What experience do you have with mobile development in the context of AI?

I have foundational experience in mobile development, particularly in deploying AI models on Android platforms. This involved optimizing models for lower computational power, ensuring they run efficiently without compromising user experience or model accuracy.

Join Rise to see the full answer
How would you approach integrating a new AI model into an existing edge inference stack?

I would begin by analyzing the architecture of the existing inference stack, followed by assessing the requirements and constraints of the new AI model. After ensuring compatibility, I would proceed with performance tuning and integration testing to ensure seamless operation and optimal performance.

Join Rise to see the full answer
How do you stay updated with the latest developments in AI and hardware optimization?

I stay updated by actively engaging with the AI community through forums, attending relevant workshops, and following leading researchers. I also regularly read scientific papers and blogs on emerging trends in AI and hardware optimization to ensure my skills are aligned with industry advancements.

Join Rise to see the full answer
Similar Jobs

Join Liquid AI to contribute to groundbreaking AI systems as a Member of Technical Staff specializing in foundational model data.

Join Liquid AI as a Member of Technical Staff to advance Vision-Language models and innovate in AI solutions.

Photo of the Rise User
Posted 8 days ago

Join Jobgether as a Senior Software Engineer and help build secure and scalable software solutions.

Photo of the Rise User
Posted 14 days ago
Photo of the Rise User
Posted 8 days ago

Join Fortanix, a leader in data-centric cybersecurity, as a Scrum Master to drive our agile practices within hybrid multicloud solutions.

Photo of the Rise User
Posted yesterday
Medical Insurance
Dental Insurance
Vision Insurance
Mental Health Resources
Learning & Development
Equity
Paid Holidays
Paid Time-Off
WFH Reimbursements
Child Care stipend
Maternity Leave
Paternity Leave

Be part of Flock Safety's mission to eliminate crime as a Software Engineer II by building advanced search and sharing systems.

Photo of the Rise User
Posted 14 days ago
Photo of the Rise User
Posted 14 days ago
Photo of the Rise User

HighLevel is looking for a Software Development Engineer II to enhance their Calendar Integrations and optimize systems while promoting a culture of continuous learning.

Photo of the Rise User
Posted 2 days ago

Lead innovative technology strategies as a Chief Software Engineer at Visa, focusing on AI-driven risk and fraud management solutions.

MATCH
Calculating your matching score...
FUNDING
SENIORITY LEVEL REQUIREMENT
TEAM SIZE
No info
HQ LOCATION
No info
EMPLOYMENT TYPE
Full-time, on-site
DATE POSTED
April 4, 2025

Subscribe to Rise newsletter

Risa star 🔮 Hi, I'm Risa! Your AI
Career Copilot
Want to see a list of jobs tailored to
you, just ask me below!
LATEST ACTIVITY
Photo of the Rise User
Someone from OH, Dayton just viewed Inventory Control Analyst II at Aretum
Photo of the Rise User
Someone from OH, Dayton just viewed Business Analyst (Supply Chain project) at Nagarro
Photo of the Rise User
9 people applied to Front end developer at Viseven
Photo of the Rise User
Someone from OH, Dayton just viewed Sr. Logistics Analyst at Innio
Photo of the Rise User
Someone from OH, Cincinnati just viewed Forensic Nurse Examiner-Prn Shift Varies at TriHealth
Photo of the Rise User
Someone from OH, New Albany just viewed Junior Buyer at CSC Generation
Photo of the Rise User
Someone from OH, Columbus just viewed Financial Administrator Intern at Finalsite
F
Someone from OH, Columbus just viewed Part Time Support Lead at Five Below
Photo of the Rise User
Someone from OH, North Olmsted just viewed Art Director - Creative- KY at Photon
Photo of the Rise User
Someone from OH, Cleveland just viewed Account Executive, Army SOF/COCOMs at Pure Storage
Photo of the Rise User
Someone from OH, Kent just viewed IT Compliance Analyst I at Fidelity National Financial
Photo of the Rise User
Someone from OH, Dayton just viewed Music Production / Creative Intern at Landor
Photo of the Rise User
Someone from OH, Cleveland just viewed Double Remote Assistant (Central US) at Zirtual
S
Someone from OH, Cincinnati just viewed Product Manager - Remote at Substance
Photo of the Rise User
162 people applied to Scrum Master-Remote at DICE
Photo of the Rise User
Someone from OH, Mason just viewed IT General Controls Tester at ING
Photo of the Rise User
Someone from OH, Columbus just viewed Contact Center Representative - 4882 at Advantmed