At Tensorwave, we’re leading the charge in AI compute, building a versatile cloud platform that’s driving the next generation of AI innovation. We’re focused on creating a foundation that empowers cutting-edge advancements in intelligent computing, pushing the boundaries of what’s possible in the AI landscape.
Job Description:
TensorWave is seeking an ML Framework Engineer to lead the integration, optimization, and maintenance of PyTorch (and select AI libraries) on AMD ROCm GPUs. This role is critical in ensuring our AI cloud platform remains at the cutting edge of performance, stability, and compatibility by tracking upstream framework changes, debugging compatibility issues, and automating builds, testing, and benchmarking. You will be responsible for maintaining a registry of validated AI libraries, debugging low-level performance issues, and working with external maintainers to upstream fixes. You will collaborate with DevOps, MLOps, and AI researchers to ensure a seamless deployment and development experience across TensorWave’s infrastructure. This role is ideal for an engineer with deep PyTorch internals knowledge, strong GPU debugging experience, and a passion for optimizing AI workloads at the framework level.
We’re looking for resilient, adaptable people to join our team—folks who enjoy collaborating and tackling tough challenges. We’re all about offering real opportunities for growth, letting you dive into complex problems and make a meaningful impact through creative solutions. If you're a driven contributor, we encourage you to explore opportunities to make an impact at Tensorwave. Join us as we redefine the possibilities of intelligent computing.
What We Bring:
In addition to a competitive salary, we offer a variety of benefits to support your needs, including:
Stock Options
100% paid Medical, Dental, and Vision insurance
Life and Voluntary Supplemental Insurance
Short Term Disability Insurance
Flexible Spending Account
401(k)
Flexible PTO
Paid Holidays
Parental Leave
Mental Health Benefits through Spring Health
If an employer mentions a salary or salary range on their job, we display it as an "Employer Estimate". If a job has no salary data, Rise displays an estimate if available.
At Tensorwave in vibrant Las Vegas, NV, we are taking the world of AI computing by storm, and we want you to be part of that journey. We’re looking for a talented ML Framework Engineer to join our innovative team. If you are passionate about optimizing AI workloads and have a solid grasp of PyTorch internals, this is the opportunity for you! In this role, you will spearhead the integration and maintenance of PyTorch on AMD ROCm GPUs, which is crucial for keeping our AI cloud platform at the forefront of performance and stability. Your day-to-day will involve debugging compatibility issues, managing a registry of validated AI libraries, and automating builds and tests to ensure our infrastructure runs smoothly. You’ll collaborate closely with DevOps, MLOps, and AI researchers, creating seamless development and deployment experiences. We believe in pushing boundaries, so you’ll also be triaging ROCm-related issues and optimizing performance by analyzing AI workloads. If you love working with third-party contributors to enhance framework compatibility and you're excited to dive into complex challenges, we want to hear from you. At Tensorwave, we’re committed to your growth and offer a supportive environment where you can truly make an impact in the world of intelligent computing.
Supercharge your large-scale PyTorch LLM workloads with our cloud powered by AMD MI300X
3 jobsSubscribe to Rise newsletter