Job title: Software Engineer, LLM Inference Engine and Product / Member of Technical Staff
Who We Are
WaveForms AI is an Audio Large Language Models (LLMs) company building the future of audio intelligence through advanced research and products. Our models will transform human-AI interactions making them more natural, engaging and immersive.
Role overview: The Software Engineer, LLM Inference Engine and Product will focus on developing and optimizing a real-time inference engine for multimodal large language models (LLMs) that handle audio and text inputs seamlessly. This role involves leveraging technologies such as LiveKit, RTC engines, WebRTC, and FastAPI to create an efficient, real-time API layer. You will contribute to cutting-edge AI systems that enable smooth user experiences across platforms, including iOS, Android, and desktop.
Key Responsibilities
Real-time Inference Development: Build and optimize a robust inference engine that supports multimodal LLMs, handling real-time audio and text inputs.
Technology Integration: Leverage tools like LiveKit, RTC engines, WebRTC, and FastAPI to enable low-latency, real-time communication and inference.
End-to-End Pipeline Design: Create and maintain the complete inference pipeline, from data ingestion to model serving, ensuring real-time performance.
Cross-platform Compatibility: Ensure the inference engine operates efficiently across various platforms, including mobile (iOS/Android) and desktop.
Optimization & Performance Tuning: Optimize the inference system to reduce latency, improve throughput, and enhance user experience.
API Development: Design and maintain scalable APIs to support real-time LLM interaction for diverse applications.
Required Skills & Qualifications
Inference Engine Expertise: Proven experience in building and optimizing inference engines for multimodal AI systems, particularly combining audio and text inputs.
Technical Proficiency: Strong experience with LiveKit, RTC engines, WebRTC, and FastAPI for real-time communication and model inference.
Real-time System Design: Expertise in creating real-time pipelines and maintaining low-latency performance in production systems.
Cross-platform Development: Familiarity with iOS, Android, and desktop app development, ensuring seamless integration with inference systems.
Performance Optimization: Proficiency in optimizing inference engines to reduce latency and improve computational efficiency.
API Development: Experience in designing and maintaining APIs for real-time AI applications.
Join the innovative team at WaveForms AI as a Software Engineer, LLM Inference Engine and Product! In this exciting role, you’ll be at the forefront of audio intelligence, helping to develop and optimize a real-time inference engine for multimodal large language models (LLMs). Imagine being part of a company that is set to revolutionize human-AI interactions by creating more natural and engaging experiences. Your responsibilities will include leveraging cutting-edge technologies like LiveKit, RTC engines, WebRTC, and FastAPI to build a robust inference engine that seamlessly integrates audio and text inputs. This position isn't just about coding; it's about crafting an infrastructure that supports real-time communication and ensures smooth performance across platforms such as iOS, Android, and desktop. You'll be in charge of maintaining the entire inference pipeline, design scalable APIs, and fine-tune system performance to provide the best possible user experience. If you have a passion for AI technologies and a knack for cross-platform development, we invite you to contribute to groundbreaking projects that shape the future of AI in audio. Your expertise will play a crucial role in enabling our advanced systems to perform flawlessly and efficiently. Come join us at WaveForms AI and be part of the transformation!
Subscribe to Rise newsletter