Job Description
About Cartesia
Our mission is to architect AI that learns from and interacts with the world like humans do.
We're pioneering the model architectures that will make this possible. Our founding team met as PhDs at the Stanford AI Lab, where we invented State Space Models or SSMs, a new primitive for training efficient, large-scale foundation models. Our team combines deep expertise in model innovation and systems engineering paired with a design-minded product engineering team to build and ship cutting edge models and experiences.
We're funded by leading investors at Index Ventures and Lightspeed Venture Partners, along with Factory, Conviction, A Star, General Catalyst, SV Angel, Databricks and others. We're fortunate to have the support of many amazing advisors, and 90+ angels across many industries, including the world's foremost experts in AI.
About the Role
We're hiring an Inference Engineer to advance our mission of building real-time multimodal intelligence.
Your Impact
Design and build low latency, scalable, and reliable model inference and serving stack for our cutting edge foundation models using Transformers, SSMs and hybrid models.
Work closely with our research team and product engineers to serve our suite of products in a fast, cost-effective, and reliable manner.
Design and build robust inference infrastructure and monitoring for our products.
Have significant autonomy to shape our products and directly impact how cutting-edge AI is applied across various devices and applications.
What You Bring
Given the scale and difficulty of problems we work on, we value strong engineering skills at Cartesia.
Strong engineering skills, comfortable navigating complex codebases and an eye for writing clean and maintainable code.
Experience building large-scale distributed systems with high demands on performance, reliability, and observability.
Technical leadership with the ability to execute and deliver zero-to-one results amidst ambiguity.
Background in or experience working on inference pipelines with machine learning and generative models.
Experience implementing state of the art Machine Learning models and research to applied problems.
Preferable: experience with vLLM, SGLang, Continuous Batching or other inference frameworks.
Preferable: experience working in CUDA, Triton or similar
More Details
๐ข In-office policy: Weโre an in-person team based out of offices in ๐บ๐ธ San Francisco, ๐ฌ๐ง London and ๐ฎ๐ณ Bangalore We love being in the office, hanging out together, and learning from each other every day.
๐ Visa sponsorship: We provide visa sponsorship support and assess each circumstance on a case-by-case basis. However, visa sponsorship is dependent on many factors, including the role you are applying for, and the location you are going to be based, and so we can't always guarantee success. Your Recruiter will work with you to understand your visa sponsorship needs from the first call.
๐ข We ship fast. All of our work is novel and cutting edge, and execution speed is paramount. We have a high bar, and we donโt sacrifice quality or design along the way.
๐ค We support each other. We have an open & inclusive culture thatโs focused on giving everyone the resources they need to succeed.
Our Benefits
๐ฐ Compensation. Competitive base salary alongside attractive equity package.
๐ Commuter Allowance. A monthly stipend to help you get to and from the office.
๐๏ธ Flexible PTO. Take as much time as you need to recharge your batteries.
๐ฒ Meals & Snacks. Lunch, dinner and plenty of snacks, provided daily.
๐ฆ Your own personal Yoshi.