Mark Lubin

← Back

I Spent Weeks Mapping the World's Most Effective Startup Incubator. It's Not What You Think.

I went looking for what makes successful AI founders. I found something I wasn't expecting: a hidden system that's been minting billion-dollar companies while disguised as academic research.


This started as a research project. I wanted to understand the AI infrastructure landscape — who's building what, where the opportunities are, what patterns predict success. I deployed research agents, scraped GitHub, read papers, tracked VC announcements. The usual.

What I found instead was a reckoning with my own assumptions about how the world works.


The World I Didn't Know Existed

About 15 years ago, when I was making career decisions, my mental model of “academia” was clear: pyramid scheme. Ten thousand people competing for two tenure-track jobs at universities nobody wants to live near. Mediocre pay. Work that doesn't matter to anyone outside a tiny peer group. You publish papers that three people read, you stress about grant applications, and if you're lucky, you end up at a state school teaching undergrads who don't want to be there.

This model was accurate. For most of academia, it still is. Physics, humanities, most of biology — that's exactly what it looks like.

So I did what seemed rational. I went into industry. Big Tech. Good salary, interesting problems, impressive resume line.

Here's what I didn't know: somewhere in the last decade, a handful of computer science departments at maybe ten schools worldwide transformed into something completely different. And nobody told me.


What I Actually Found

I started this research trying to answer a simple question: where do successful AI infrastructure companies come from?

So I mapped it. Seventeen major labs. Fifty-plus spin-out companies. Valuations, timelines, founders, advisors, funding rounds.

The concentration was the first thing that hit me.

UC Berkeley's Sky Computing Lab — one research group, one building — has produced over $150 billion in company value. Databricks alone is worth $134 billion. Then there's Anyscale, Letta, Covariant, Conviva, and as of the last few weeks, two new companies: Inferact (vLLM commercialization, $800M valuation out of the gate) and RadixArk (SGLang commercialization, $400M valuation).

One building. $150 billion.

Stanford's Hazy Research group: another $20 billion. Snorkel, SambaNova, Together AI.

CMU's robotics program: Skild AI just raised at a $14 billion valuation.

I kept pulling the thread, and the same names kept appearing. Ion Stoica. Chris Re. Pieter Abbeel. A small network of professors who function less like academics and more like... I don't know, general partners at a venture fund who happen to have tenure.


The Incubator Hiding in Plain Sight

Here's the insight that reframed everything for me.

The top CS/AI programs aren't really academic research programs anymore. They're startup incubators disguised as graduate schools. And they're better than traditional incubators by almost every measure.

Think about what a normal accelerator offers: space, mentorship, network access, maybe some funding. You pay for it with equity — 7-10% at YC. And you get three months.

Now think about what a Berkeley PhD gets:

The university provides the space and resources. For free. For five years.

The advisor provides mentorship. But more importantly — and this is the part I didn't understand — the advisor does product discovery.

Ion Stoica knows what industry needs. He's on boards. He has corporate sponsors. He consults. He sees the pain points. Then he takes those pain points, frames them as “research problems,” and hands them to capable students to run with.

This is crucial: the professor isn't teaching students to find problems. The professor is handing them pre-validated, industry-shaped problems and saying “solve this.”

The advisor is the network. When you're in Stoica's lab, you don't need to figure out how to meet VCs. They're waiting for you. They're monitoring your GitHub stars. They already know who you are.

And when your paper gets traction — when your open source project hits 10,000 stars — you don't have to convince anyone there's a company here. The company is obvious. The funding is obvious. Everyone's been waiting.

Compare that to a traditional incubator:

Traditional Incubator Academic Lab Pipeline
Founders pay for space, mentorship University provides for free
3-month intensive program 5-year deep expertise building
Founders must find their own problems Professor hands them the problem
Network building is hard Professor is the network
VC access is uncertain VCs wait outside the lab
Takes 7-10% equity Professor gets advisor shares + reputation

The professor does the hardest part of founding a company — identifying a problem worth solving — and the student executes. It's an assembly line for startups, and nobody outside the system realizes it's happening.


The Great Compression

The timeline data is what really got me.

I went back through the history. Spark, the paper, came out around 2010. Databricks was founded in 2013. Three to four years from research to company.

Ray came out of Berkeley in 2017. Anyscale was founded in 2019. Still two to three years.

Then something shifted.

MemGPT — the paper that showed you could give LLMs persistent memory — was published in October 2023. Letta, the company, raised a $70 million seed round in late 2024. Eleven months.

Eleven months from research paper to heavily-funded startup.

The compression is accelerating. Academic research used to be a lagging indicator — by the time it was published, the idea was already old news. Now it's a leading indicator. If you're watching the right GitHub repos, you can see companies forming in real time.


The Inversion That Nobody Talks About

This is where it gets philosophically interesting.

Universities exist, in theory, to produce fundamental knowledge. The whole point of academia — the reason we fund it, the reason tenure exists — is to let people pursue truth without commercial pressure. Industry is supposed to take that foundational knowledge and build products on top of it.

That's the mental model. And in AI right now, it's completely backwards.

The fundamental breakthroughs are coming from industry labs.

The transformer architecture? Google. Scaling laws? OpenAI and Anthropic. RLHF? DeepMind and OpenAI. The actual science — the discoveries that change our understanding of what's possible — is happening at corporations.

What academic labs do, at least in ML infrastructure, is the applied engineering work. Making things faster. Making things cheaper. Fixing what's broken in production.

vLLM is brilliant systems engineering — it figured out how to manage GPU memory efficiently for LLM inference. FlashAttention is brilliant systems engineering — it restructured attention computation to be hardware-aware. DSPy is brilliant systems engineering — it gives developers programming abstractions instead of prompt engineering.

But it's engineering. It's product development dressed as research.

This is a genuine inversion of what universities and industry are “supposed” to do. DeepMind publishes papers that reshape our understanding of intelligence. Berkeley publishes papers that make inference 6x faster.

I'm not saying one is better than the other. But the inversion is real, and it matters. If you're looking for foundational breakthroughs, you should be reading industry lab publications. If you're looking for production infrastructure, you should be watching academic repos.

That's weird. And I don't think most people have updated their mental models to account for it.


The Big Tech Dead End

Here's something I wish I'd understood earlier.

Working at Big Tech is fine. Good salary. Interesting problems. Smart colleagues. But it's not a generative system. It's not a flywheel that builds on itself.

What I mean is: the academic lab path has compounding effects. You get handed a problem, you build expertise, you publish, you build a network, your open source gets traction, VCs notice, you spin out a company, your advisor takes a board seat and introduces you to everyone, and the whole thing feeds back into itself. Each step enables the next step.

Big Tech doesn't work like that. You're an employee. You have a manager, not an advisor. Nobody's doing product discovery for you. There's no spin-out infrastructure. When you leave, you leave with your salary and your resume line, and that's it.

It's not that Big Tech is bad. It's that it's static. You get paid for your time, but you're not building positional leverage. You're not accumulating the kind of network effects that let some people turn a research paper into a billion-dollar company in eleven months.

The academic lab path is a flywheel. Big Tech is a treadmill.


What's Actually Getting Built

Let me share what I actually learned about the landscape.

The inference moment is now. Training was the 2018-2022 story. Inference is 2023-2026. Every major lab is working on making LLMs cheaper and faster to run. Costs have dropped 1000x in three years.

The problem is: this space is crowded. vLLM, SGLang, TensorRT-LLM, FlashAttention derivatives — everyone's working on it. The easy wins are taken.

Where's the gap? Agent reliability.

57% of enterprises have AI agents in production now. The number one blocker, cited by almost half of them? Reliability. The agents don't work consistently.

And here's the thing: there's almost zero academic work on agent debugging, observability, or reliability engineering. The labs are focused on making agents more capable. Nobody's focused on making them more reliable.

This is a massive unfilled gap. And the academic pipeline isn't competing for it.


The Self-Referential Problem

One more thing I noticed, and I'm not sure what to do with it.

The LLM infrastructure space is somewhat self-referential. The benchmarks are created by insiders — Berkeley created BFCL for function calling, Stanford created HELM for general evaluation. The success criteria are defined by insiders. Then the insiders build things that score well on their own benchmarks.

The outsiders are judged by insider-created standards.

I don't know if this is a problem exactly. The benchmarks seem reasonable. But there's something circular about it. The arbiter of truth is the insider community, not external reality. That feels like a vulnerability somewhere.


So What Now

I started this research trying to understand where opportunities are. I ended up understanding something about how knowledge production and company formation have merged in ways I hadn't seen.

The paths forward, if you're not already in the system:

Get inside a generative startup — not Big Tech, but the spin-outs themselves. Together AI, Modal, Anyscale, Fireworks. These companies are developing their own flywheel dynamics.

Build in the gaps — Agent debugging. Compliance tooling. Vertical applications. The stuff the labs aren't working on because it's not publishable.

Take the YC path — It exists specifically to create network effects for people who don't have them. They don't care about your credentials. They care about traction.

Build open source credibility — Contribute, build reputation, become known. Takes years, but works.

Build something undeniably useful — If it works and people use it, the network eventually finds you. This is the hardest path, but it's a path.


What I'm Watching

For what it's worth, here's where I think the near-term action is:

DSPy (Stanford) — 160,000 monthly downloads. “Programming, not prompting.” Every developer hates prompt engineering; this makes it feel like regular code. Omar Khattab is the PhD to watch. Company feels imminent.

The agent reliability gap — 57% in production, reliability is the top blocker, near-zero tooling. This is the biggest unfilled space I found.

Compliance infrastructure — EU AI Act is active. US states are following. The tooling is primitive. This is regulatory-driven, not research-driven, which means the academic pipeline doesn't own it.


I wrote this to share what I learned, but also to process it.

The insight that sticks with me: universities are doing engineering while industry does science. Professors are doing product discovery while students do execution. The whole system is inverted from what you'd expect, and it works incredibly well — if you're inside it.

Understanding this is an asset. But only if converted to action.

So: what are you going to do with it?


I'm still figuring out my own answer to that question. If you're thinking about similar things, I'd like to hear about it.


Sources: