Letta is a company founded around the MemGPT project (13k+ GitHub stars). The founding team comes from the same research lab and PhD advisors at Berkeley that produced Spark (→ Databricks) and Ray (→ Anyscale). We have deep expertise in both AI and systems, are currently hiring a founding team of exceptional engineers to join us in building the next generation of LLM agent technology.
👾 You can read more about Letta on TechCrunch and our blog.
LLM-driven intelligence is bounded by the inherent limits of the Transformer architecture and its training methods. We believe that LLMs are just one piece of a complete agentic system - to build human-like AI that can reason, plan, learn, and remember, we need to engineer the new computer. At Letta, our research is focused on understanding the fundamental limitations of LLM-driven intelligence, and discovering new ways to advance it.
Unlike research groups at foundation model companies, we focus on research that is model agnostic. This allows us to mix-and-match the best models depending on their capabilities and move fast by avoiding training. We believe that the same level of research advancements that happened at the model layer to enable today’s LLMs will also need to happen at the layer above the models to fully leverage model capabilities in compound systems.
As a research engineer, you will work closely with a world-class research team (PhDs from UC Berkeley’s BAIR and Sky research labs, behind the MemGPT and “Lost in the Middle” work) on agentic systems, memory, reasoning, and scaling test-time compute. You will help productionize research developments in Letta’s OSS framework and cloud platform.
Responsibilities
Developing Letta’s core agentic reasoning loops, which encompasses tool execution, stream parsing, reasoning, and more.
Evaluating and improving performance of Letta’s agents framework with open models and new model types (e.g. reasoning models)
Integrating LLM API providers into Letta’s framework
Skills
Extremely strong programming skills in Python
Understanding of AI/ML fundamentals (Large Language Models)
Ability to read and understand research papers in the agents/memory space
Experience in software engineering
Examples of what you might work on
Adding memory management capabilities to voice agents
Implementing constrained decoding to improve tool calling in open models
Systems for multi-agent coordination
Benchmarking model performance
We are hiring a small, tight-knit team of exceptionally talented founding engineers. Every hire matters, so we take the hiring process very seriously.
Initial phone interview (30m video call): We want to learn more about your background, your skills, your opinions on open source AI, and why you want to work at an early stage AI startup.
Technical take-home (<1hr assessment): To get a better sense of your skillset, we’ll give you an example problem to work that’s as targeted to your potential day-to-day work as possible.
Paid workday (in-person recommended): As the final step in the interview process, we’ll simulate working together as closely as possible by giving you a real (or as close to real as possible) task to work on for a day - and paying for your time of course. If you live in the Bay Area, we highly recommend visiting our offices in-person! We’re an in-person company, so working at our office will give you a great idea of what it will be like to join as a full-time member of the team.
If an employer mentions a salary or salary range on their job, we display it as an "Employer Estimate". If a job has no salary data, Rise displays an estimate if available.
Are you ready to be part of something groundbreaking? Letta, a pioneering company born from the MemGPT project, is searching for an exceptional Founding Research Engineer to join our dynamic team in San Francisco. With our roots in innovative research from Berkeley, we specialize in advancing the capabilities of large language models (LLMs) and creating an agentic system that can learn, reason, and adapt like a human. As a Founding Research Engineer at Letta, you'll collaborate with a world-class team of PhDs from renowned labs, working on core agentic reasoning loops, optimizing performance of our agents framework, and integrating various LLM API providers into our cloud platform. Your responsibilities will stretch from developing advanced memory management techniques for voice agents to designing systems for multi-agent coordination. You'll have a strong programming background in Python and a grasp of AI/ML fundamentals, especially LLMs, enabling you to turn research insights into tangible contributions. If you thrive in a fast-paced environment and are eager to solve complex challenges, we'd love to hear from you! We're building a close-knit, talented, and passionate team where every hire counts, so be prepared for a thoughtful and thorough interview process.
Subscribe to Rise newsletter