7 Lessons - Building and Investing in Vertical AI
We spent the last 12+ months elbows-deep in model prompts, user data, and real-world edge cases
If you haven’t come across this piece of news (screenshot below), pay attention. I felt its a good place to start.
Now, take a second to absorb that.
Everyone talks about Chips/GPUs, compute and model APIs, but data infrastructure - the human-in-the-loop layer that teaches AI what “good” looks like has quietly become the next big revenue engine in the stack.
Companies like Scale AI and Surge AI are reportedly generating close to a billion dollars annually (each) by combining data labeling, model evaluation, and feedback loops for training and fine-tuning large models.
A friend of mine in fintech casually mentioned he’s doing a side gig for one of the relatively smaller AI data annotation companies helping them annotate data in the field of finance. And that’s the real story here. Domain expertise and data not model size is the chasm AI has to cross to actually become useful deep in the industries. It’s not enough for AI to “know finance”; it has to think like a banker, reason like an analyst, and speak in the language of Excel. who will make that happen?
That’s where the next trillion dollars of enterprise AI value will be unlocked.
Last year when we started gAI Ventures (alongwith Kushal - my co-founder/CTO, ex-AI researcher and AI architect) we had a hypothesis that Vertical AI is different and will require domain expertise and propriety data and more than just AI. We believed that Industry-specific AI startups that embed deeply into existing workflows, automate high-value processes, and drive measurable economic impact will win. So we assembled a team of AI engineers and partnered with Vijay (ex-500 Global and BBVA Ventures). Together we started building gAI Ventures brick by brick, putting together the systems and processes to identify ideas/opportunities and the right founders. Its been a lot of hard work and you have to stay focused and avoid distractions. Since then we have done a lot of R&D, gulped dozens of research papers (arXiv) and built a lot of AI native stuff including a product that went to #1 product on product hunt globally and another one that serves 13+ customers including a $1.5 B Aum RIA in 6-7 months from just an idea.
Here are some lessons we have learned on the way building and investing in Vertical AI:
Generative AI tech can do wonders. This is a generational opportunity to blend human intuition with AI efficiency
Thomas Dohmke (CEO of GitHub)
“Sooner than later, 80% of the code is going to be written by Copilot. And that doesn’t mean the developer is going to be replaced.”
Andy Jassy (CEO of Amazon)
“Today, in virtually every corner of the company, we’re using Generative AI… We have over 1,000 Generative AI services and applications in progress or built.”
Goldman Sachs CEO David Solomon. He stated that AI can generate “95% of an IPO prospectus (S-1 document)” in minutes, leaving only the final 5% as the human-differentiator.
We have seen it first-hand building AI native products and hearing that “wow” from customers. Imagine from gathering data via meeting bots and document processor AI, extracting information and updating CRM and financial planning software while allowing wealth advisors to spend more time with customers (investors) and letting AI doing the back-end heavy lifting.
Another Example: Harvey AI - AI drafts contracts, analyzes clauses, or summarizes regulations. Lawyers then apply domain expertise, negotiation skill, and risk judgment. Impact:
Associates save 30–40% of drafting/review time.
Partners focus on strategy, not boilerplate.
And oh… “AGI is still a decade away”
Andrej Karpathy in the excellent podcast (Dwarkesh) said AI can beat us at Go, schoolwork, analyzing medical images, and many well defined tasks. At the same time, it lacks the reward systems that humans use to improve including curiosity, empowerment, play, intrinsic motivation, and culture.
We now know that we’re years away from a “magical” superintelligence that will do anything and everything for us. Many insiders who are building in AI knew but the latest Karpathy podcast cleared the air. What we have right now, increasingly, are AI models that exhibit strong general reasoning capabilities with improving skills in specific domains. This will all get better and better over time, unquestionably. But they still need a ton to be useful in most domains.
Karpathy said in a tweet:
Podcast link if you want to go deeper.
AI is not good at many things today and autonomous agents fail all the time
We’ve learned that AI progress isn’t linear. It’s loops of sweat, code, and small wins. Turns out, building AI products is 10% models and 90% messy human workflows. I read a post recently …
Two identical tax AI startups launched on the same day. Six months later, one has 10,000 users, the other has 12. They used the exact same model. Here’s what made the difference:
Founder A dumped everything into the context window:
→ Full transcripts
→ Entire IRS documents
→ Generic instructions
Results were inconsistent. Sometimes wrong.
Founder B built a tight pipeline:
→ User intent selector (”file,” “explain,” “plan”)
→ IRS content retrieved at section level only
→ Compact user profile summaries
→ Output constrained by formatting rules
→ Injected reasoning steps before answers
Same underlying model. Very different product.
Founder B’s assistant felt smarter, more reliable, more professional. That’s why his product exploded over the past few months.
and then I saw this tweet as well:
and the other day I got this email and look at it carefully (after gazillion emails being sent by AI, email campaigns are at their lowest performance levels):
writing is on the wall.
No, vibe-coding cannot help you build enterprise and/or production ready products. I will start with a deeply concerning chart…
Cursor and Lovable are awesome products. we use them all the time to build smaller modules. but you still can’t build a production ready enterprise ready product with them.
» “Coding Agents are Slopware App-crappers” said Chamath Palihapitiya
Basically Chamath is saying: these “coding agents” (AI code bots, generative tools) are churning out low-quality software (“slopware”), producing many apps or features that are shallow, half-baked, maybe hype-driven rather than value-driven.
In other words: Using coding agents often leads to a volume play (lots of app/features) but with weak discipline, resulting in many “app-crappers” rather than high-quality, well-engineered products.
The implication: Just because you have coding agents doesn’t mean you build something good. you risk building more junk/mediocre apps.
don’t confuse tool → output quantity with output quality. Without rigorous engineering + product design + domain embedding you risk building “slopware app-crappers”.
And on top of that - In the vertical AI domain, what matters is domain depth, data, embedding into workflows, strong engineering, proper product-market fit and… not just “generate lots of features with an agent”.
AI’s capabilities are not superhuman (yet) and have limitations - If you watch how the best work gets done with AI, it happens in chunks where a human supervises the output and gives iterative feedback to the AI. Large scale autonomous agents are brittle and fail quickly. Watch anyone vibe code an application with any level of novel complexity.
For vertical AI you need to have the prop data, understand the domain, the workflows, data privacy, security and did I say “data”.
AI Models are great but what they need, to be most effective, in real world situations is context and usually an applied set of features to make the AI useful. They need the data that you’re working with, they need to show up in user interfaces that are relevant to the task, they need to have a sense of the domain that they’re being applied to.
Vertical AI needs more in the beginning
AI also faces integration barriers into existing organizations. AI is not capable of pulling a lot of the levers you need to be effective like coordinating with multiple stakeholders, building trust, authenticity, and interacting with different modalities across time and space. You could argue that the average human doesn’t either, but people know when they’re interacting with an AI and don’t allow it the same agency as they do to people. The most successful AI B2B companies in many cases actually need more humans to integrate what they’ve built (forward deployed engineers) than traditional B2B SaaS. something builders and insiders know
You can’t just rename a legacy software or saas company and make it AI
Some companies are sleeping. Some legacy software cos are complacent (one guy told me we have been doing AI for 20 years). While 90% of incumbents haven’t adapted, it is possible. Figma, Notion, Vercel, Box, and Intercom have done a great job of tearing down what they have and rebuilding AI native products. They have teams who are close to the current capabilities of the models. They also understand their problem domain and as new capabilities come out know what capabilities will map well to what problems in what way. They are able to deliver on AI’s promise to their customers. Whereas the majority of existing companies will die.
Okay….to wrap this up, I’ll leave you with a tweet that captures the spirit of what’s really happening beneath all the AI hype. That’s the quiet revolution. The real story is more that just about GPUs, model sizes, or even benchmarks. It’s about the knowledge transfer. The messy, human, domain-specific wisdom that turns general intelligence into something useful inside a spreadsheet, hospital, or logistics dashboard.












