HomeLaunchesLuminal
124

Luminal - PyTorch for Production

Luminal automatically speeds up and deploys your PyTorch models

Hey we’re Joe, Jake, and Matthew, cofounders of Luminal.

Luminal is an open-source ML compiler that generates blazingly fast CUDA kernels and makes deploying AI to production one line of code. We’re already powering research at Yale, production workloads at VC-backed startups and several research labs.

https://youtu.be/Rgo15NY9K-8

TLDR:

  • Our drop-in upgrade to PyTorch automatically generates complex GPU kernel code, like Flash Attention, with zero hand-engineering
  • We provide truly serverless deployments; no idle costs or cold starts.
  • By not relying on pretrained LLMs or hand-written optimizations, we can reliably speed up any model

Ask:

The problem:

Most people running their own AI models are lighting money on fire and don’t even know it. AI teams are frustrated when moving their model from dev to production. They either fall into dependency hell or kill their model’s speed. Today, companies waste millions of dollars on GPU engineering teams to optimize new models before they can be served to users.

How Luminal solves it:

Luminal replaces a process that companies pay GPU engineers $300k+ a year to do.

Our key insight is that by treating optimization as a search problem, we’re able to automatically discover extremely complex optimizations in minutes that would take a GPU expert weeks. We use a series of ‘rewrite rules’ to create millions of graphs to describe your model, generate kernel code for each and then search for the fastest one based on runtime.

Example of a complicated kernel on the left that Luminal automatically sped up, on the right.

Luminal is:

  1. Closing the research → engineering gap. Because we auto-generate optimized GPU code directly from your PyTorch, a brand-new model can go from research notebook to production in hours instead of weeks
  2. Writing blazing fast models. Luminal generates kernels that would take a seasoned GPU engineer weeks of painstaking profiling and tuning to write.
  3. Future-proofing AI for new chips. We can rerun our compiler on fresh hardware to get new optimal kernels so you're never out of date.

You should reach out if:

  • You’re an AI researcher or ML engineer
  • You’d like to lower your compute bill by tens of thousands of dollars a month
  • You work at a company that runs their own custom models

If you want your team spending more time making the world’s best models and less time optimizing hardware and cloud infrastructure, we’re for you!

The Team:

Joe - ex Intel, every Intel chip sold has the AI accelerator Joe worked on. He has extensive experience optimizing performance at the silicon level.

Matt - ex Amazon where his software handled finding and automatically fixing issues within the global inventory network 24/7

Jake - ex Apple Jake worked on imaging at Apple for your iPhone. He’s been a founder (with an exit) and head of growth at another startup that he grew to ~$5M ARR.

Send us an email: contact@luminalai.com

Book a call with us