Eventual is a data platform that helps data scientists and engineers build data applications across ETL, analytics and ML/AI.
Our distributed data engine Daft is open-sourced and runs on 800k CPU cores daily. This is more compute than Frontier, the world's largest supercomputer!
Daft is used at leading AI/ML companies such as Amazon, TogetherAI, EssentialAI, CloudKitchens and more. It makes ML/AI workloads easy and performant to run alongside traditional relational tabular workloads.
Today's “Big Data” data tooling (Spark, Trino, Snowflake) was built for a world of tabular data analytics. They do not generalize well to the needs of modern ML/AI data workloads. We built Daft to be the successor to these Big Data technologies along these core principles:
Python-native: Python is the native language of ML/AI and most of data engineering today
First-Class Local Development UX: Interactive development in a local Python notebook or script is where the magic happens
Multimodal Data Support: Modern workloads require support for operations on complex types such as long-form text, images, tensors and more
Heterogenous Compute (GPUs): GPUs are a requirement for workloads that perform model batch inference as part of the overall query
As a Software Engineer on the Core Engine team, you will build key capabilities for the Daft distributed data engine.
You will be working on core architectural design and implementation of various components in Daft including:
Planning/Query Optimizer: intelligently optimize users’ workloads with modern database techniques
Execution Engine: improve memory stability through the use of streaming computation and more efficient data structures
Distributed Scheduler: improve Daft’s resource utilization, task scheduling and fault tolerance
Storage: improve Daft integrations with modern data lake technologies such as Apache Parquet, Apache Iceberg and Delta Lake
Our goal is to build the world’s best open-source distributed query engine, becoming the leading framework for data engineering and analytics.
We are a young startup - so be prepared to wear many hats such as tinkering with infrastructure, talking to customers and participating heavily in the core design process of our product!
We are looking for a candidate with a strong foundation in systems programming and ideally experience with building distributed data systems or databases (e.g. Hadoop, Spark, Dask, Ray, BigQuery, PostgreSQL etc)
3+ years of experience working with distributed data systems (query planning, optimizations, workload pipelining, scheduling, networking, fault tolerance etc)
Strong fundamentals in systems programming (e.g. C++, Rust, C) and Linux
Familiarity and experience with cloud technologies (e.g. AWS S3 etc)
Most importantly, we are looking for someone who works well in small, focused teams with fast iterations and lots of autonomy. If you are passionate, intellectually curious and excited to build the next generation of distributed data technologies, we want you on the team!
We are believers in both having the flexibility of remote work but also the importance of in-person work, especially at the earliest stages of a startup. We have a flexible hybrid approach to in-person work with at least 3 days of in-person work typically from Monday - Wednesday at our office in San Francisco.
We believe in providing employees with best-in-class compensation and benefits including meal allowances, comprehensive health coverage including medical, dental, vision and more.
A short phone screen over video call with one of our co-founders for us to get acquainted, understand your aspirations and evaluate if there is a good fit in terms of the type of role you are looking for.
A technical phone screen question over video call to understand your technical abilities.
Technical interviews with the rest of the Eventual team with questions to further understand your technical strengths, weaknesses and experiences.
As many chats as necessary to get to know us - come have a coffee with our co-founders and existing team members to understand who we are and our goals, motivations and ambitions.
We look forward to meeting you!
We are well funded by investors such as YCombinator, Caffeinated Capital, Array.vc and top angels in the valley from Databricks, Meta and Lyft.
Our team has deep expertise in high performance computing, big data technologies, cloud infrastructure and machine learning. Our team members have previously worked in top technology companies such as Amazon, Databricks, Tesla and Lyft.
We are looking for exceptional individuals with a passion for technology and a strong sense of intellectual curiosity.
If that sounds like you, please reach out even if you don't see a specific role listed that matches your skillsets - we'd love to chat!
If an employer mentions a salary or salary range on their job, we display it as an "Employer Estimate". If a job has no salary data, Rise displays an estimate if available.
Join the innovative team at Eventual as a Software Engineer focused on our Core Engine! Located in the vibrant city of San Francisco, Eventual is all about enabling data scientists and engineers to create cutting-edge data applications seamlessly. You’ll be diving into the heart of our open-source distributed data engine, Daft, which is already making waves by running on a staggering 800k CPU cores daily. Your role will be pivotal in designing and implementing core components of Daft, including the Planning/Query Optimizer and Execution Engine. We're not just another tech company; we aim to revolutionize how ML/AI workloads are handled alongside traditional data processes. If you're passionate about systems programming and have a knack for building distributed data systems, you're the kind of engineer we're looking for. With your expertise, you'll help improve Daft's memory stability and resource utilization. This is a dynamic role where you’ll collaborate closely with our focused team, participate in design discussions, and even interact with customers. We value autonomy and quick iterations, and ideally, you have experience in C++, Rust, or similar languages as well as familiarity with cloud technologies. Plus, you’ll enjoy the flexibility of a hybrid work model, ensuring you have the best of both in-person collaboration and remote flexibility. If you’re excited to be part of a passionate team aiming to create the future of data technology, we’d love to hear from you!
Subscribe to Rise newsletter