Hey we’re Joe, Jake, and Matthew, cofounders of Luminal.
Luminal is an open-source ML compiler that generates blazingly fast CUDA kernels and makes deploying AI to production one line of code. We’re already powering research at Yale, production workloads at VC-backed startups and several research labs.
TLDR:
Ask:
—
Most people running their own AI models are lighting money on fire and don’t even know it. AI teams are frustrated when moving their model from dev to production. They either fall into dependency hell or kill their model’s speed. Today, companies waste millions of dollars on GPU engineering teams to optimize new models before they can be served to users.
—
Luminal replaces a process that companies pay GPU engineers $300k+ a year to do.
Our key insight is that by treating optimization as a search problem, we’re able to automatically discover extremely complex optimizations in minutes that would take a GPU expert weeks. We use a series of ‘rewrite rules’ to create millions of graphs to describe your model, generate kernel code for each and then search for the fastest one based on runtime.
Example of a complicated kernel on the left that Luminal automatically sped up, on the right.
Luminal is:
—
If you want your team spending more time making the world’s best models and less time optimizing hardware and cloud infrastructure, we’re for you!
—
Joe - ex Intel, every Intel chip sold has the AI accelerator Joe worked on. He has extensive experience optimizing performance at the silicon level.
Matt - ex Amazon where his software handled finding and automatically fixing issues within the global inventory network 24/7
Jake - ex Apple Jake worked on imaging at Apple for your iPhone. He’s been a founder (with an exit) and head of growth at another startup that he grew to ~$5M ARR.
Send us an email: contact@luminalai.com
Most people running their own AI models are lighting money on fire and don’t even know it. The rush to get to market first means nearly every AI company is leaving thousands of dollars on the table every MONTH by not optimizing their models. That’s why we made Luminal.
If you want your team spending more time making the world’s best models and less time optimizing hardware and cloud infrastructure, we’re for you!
At Luminal, we’re taking a bold swing at redefining how machine learning engineers interact with the core compute layer powering AI. When we succeed, the world’s intelligence stacks will be faster, simpler, and more robust. The hardware is ready for AI performance, the low-level code is not.
Solving the software layer destroys NVIDIA’s moat and democratizes compute for AI