HomeCompaniesWafer

Wafer

Optimized AI Inference. Custom Built for Your Stack.

Host AI models on custom infrastructure optimized for time-sensitive applications. Custom Trained Models, Dedicated Infrastructure, and Protected Data. We provide scalable compute, optimized inference engines, intelligent routing and autoscaling, real-time observability, and kernel-level performance optimizations, with strict SLAs and white-glove support. Herdora's long-term bet is maximizing intelligence per watt. AI is clearly going to be everywhere. The constraint won't be what models can do, but how much intelligence we can afford to run. Most companies waste enormous compute running inference inefficiently. They use the wrong models, bad serving stacks, and infrastructure that wasn't built for their actual workload. We focus on serving enterprises that need the best performance and reliability in the world while keeping control and visibility.
Active Founders
Emilio Andere
Emilio Andere
Founder
prev @ argonne, uchicago sand lab, elicit. math at uchicago.
Steven Arellano
Steven Arellano
Founder
prev @ two sigma, google, sei labs, and axlab. cs + econ at uchicago.
Company Launches
👑 Herdora - Cursor for CUDA
See original launch post

TL;DR

The best GPU engineers work at Nvidia and OpenAI for $2M+ packages. The rest of us are stuck with PyTorch code that uses <50% of our hardware. We built Herdora for every team that can't afford a GPU optimization expert. Herdora takes your PyTorch and optimizes it with custom kernels. It monitors your inference workloads, catches inefficiencies, and fixes them automatically. You keep shipping fast. Herdora makes it run fast. No CUDA knowledge required, no million-dollar hires needed.

Ask: Are you trying to make your inference faster, writing CUDA by hand, or do you know someone who is? Fill our our interest form or send us an email contact@herdora.com to be added to our pilot program. See below for details.

https://youtu.be/K6nEevvlOro










🧙 The Problem

Getting AI models to run fast on GPUs requires a rare breed of engineer.

These GPU wizards:

  • Command $300k+ base salaries (if you can even find them)
  • Spend months hand-writing custom kernels for each model
  • Need to rewrite every kernel for new hardware

Even worse: If you want to run on alternative GPUs or other accelerators, you need different experts to rewrite all your kernels. Most teams are stuck choosing between:

  • Paying $$$ for GPU talent
  • Accepting <50% hardware utilization
  • Locked to one chip longer than they want

🐄 How Herdora solves it

Herdora does what those GPU engineering wizards do, automatically. Feed it your PyTorch code, and it generates the optimized kernels that would normally take months of manual work.

What you get:

  • Auto optimization - Your PyTorch goes in, optimized GPU code comes out. Same model, 1.5 to 5x faster. The kernels Herdora writes are what a senior CUDA engineer would write after weeks of profiling and tuning.
  • Hardware portability - NVIDIA can’t give you enough capacity? Want to try out AMD’s new MI350 series? Herdora makes your code fast on both in seconds. No rewrites, no performance penalties.
  • Production monitoring - Herdora watches your models in production, spots inefficiencies, and fixes them. Think of it as having a GPU expert on-call 24/7.

You write PyTorch. Herdora makes it run like it was hand-tuned by the best in the business. Herdora then monitors prod to track performance. No CUDA expertise needed on your team.

📢 Who This is For

You should reach out if any of this sounds familiar:

  • You're an ML engineer but you don't have time to learn CUDA
  • Your cloud bill is killing you because your models only use 30-50% of each GPU
  • You want to try cheaper GPUs but can't risk your models running slower

👥 The Team

  • Steven Arellano - Researched LLMs at UChicago and worked as an engineer at Two Sigma, Google, and Sei Labs.
  • Emilio Andere - Researched transformers for weather prediction at Argonne National Lab, adversarial ML at SANDLab, and an engineer at Elicit.

🤝 Friends since freshman year at UChicago, where we were also roommates

🙏 Our Ask

If you're running ML workloads on GPUs (or know someone who is), we'd love to add you to our pilot program. Email us at contact@herdora.com or fill out our interest form with your model type and current GPU setup.

You'll get the same optimizations that companies pay GPU consultants $50k+ to deliver except automated and in days instead of months.

Currently Onboarding:

  • ML teams running inference at scale
  • Teams hitting latency requirements in production
  • Companies with growing GPU compute bills
  • Companies evaluating AMD or Intel GPUs

Send us an email: contact@herdora.com

Book a call: https://cal.com/herdora/20min

Interest: https://www.herdora.com/interest



Jobs at Wafer
San Francisco, CA, US
$6K - $10K / monthly
Any
San Francisco, CA, US
$6K - $10K / monthly
Any
San Francisco, CA, US
$180K - $230K
2.00% - 4.00%
Any (new grads ok)
Wafer
Founded:2025
Batch:Summer 2025
Team Size:2
Status:
Active
Location:San Francisco
Primary Partner:Jared Friedman