Jonas Stamm

AI engineer, growth hacker, builder

Back to all posts
·5 min read

Why AI Doesn't Work for Most Startups (And How to Fix It)

thought-leadership

"We added AI to our product!"

crickets

I see this pattern everywhere.

Startup adds ChatGPT API. Slaps "AI-powered" on the landing page. Waits for users to care.

They don't.

Here's why most AI features fail — and what actually works.

The Three Deadly Sins

Sin #1: AI for AI's Sake

What startups do: "Let's add an AI chatbot!"

What users hear: "We made our support worse and called it innovation."

Example I saw last week: SaaS tool added an "AI assistant" that:

  • Answered questions users could Google in 10 seconds
  • Couldn't do anything the actual product could do
  • Required users to learn a new chat interface

Result? 2% usage. Removed after 3 months.

The fix: Start with the problem, not the tech.

Bad: "How can we use AI?" Good: "What's our biggest user pain point? Could AI solve it?"

Sin #2: Generic AI in Specific Domains

What startups do: Plug in ChatGPT API. Ship it. Done.

What users experience: Vague, generic answers that don't actually help.

Real example from BauGPT:

User (construction worker): "What's the minimum reinforcement spacing for C30/37 concrete?"

Generic ChatGPT: "Reinforcement spacing depends on many factors including..."

BauGPT (with construction RAG): "For C30/37: minimum 20mm spacing between bars, DIN 1045-1 §8.3.2. Aggregate size + 5mm minimum. For slabs: typically 100-200mm spacing."

See the difference?

Generic AI knows about things. Domain-specific AI knows the answer.

The fix: Build domain expertise into your AI.

  • Industry-specific knowledge base
  • Custom embeddings for your domain
  • Citations to authoritative sources
  • Confidence scoring ("I don't know" when uncertain)

Sin #3: Ignoring Distribution

What startups do: Build amazing AI tech. Put it behind a login. Wonder why nobody uses it.

Real example:

We could have built BauGPT as:

  • Mobile app (requires download, login, onboarding)
  • Web dashboard (requires computer access on construction site)
  • Slack bot (requires company Slack, IT approval)

We built it as:

  • WhatsApp bot (zero friction, already on their phone)

Result? 10x adoption vs mobile app.

The fix: Meet users where they already are.

Don't make them:

  • Download something new
  • Learn a new interface
  • Change their workflow

Integrate into tools they're already using.

What Actually Works: The 3-Step Framework

Step 1: Find a Real Problem

Not:

  • "AI could make this cooler"
  • "Competitors have AI, we should too"
  • "Investors want to see AI"

But:

  • Users are struggling with X
  • Current solution takes 2 hours
  • People are paying for workarounds

BauGPT example: Problem: Construction workers can't understand German safety docs. Current solution: Hire interpreters (€300-500/day). Pain: Real (safety incidents, delays, worker isolation).

Step 2: Build Domain Expertise

Generic AI → Generic results.

You need:

Knowledge:

  • Industry-specific data
  • Authoritative sources
  • Edge cases and exceptions

Context:

  • User's role/expertise level
  • Current task/workflow
  • Historical interactions

Accuracy:

  • Source attribution
  • Confidence scoring
  • Human-in-loop for critical decisions

BauGPT example:

  • 50+ DIN standards (German building codes)
  • Construction terminology in 40+ languages
  • Safety-critical accuracy (wrong answer = injury)

Step 3: Zero-Friction Distribution

Your AI is only as good as people's willingness to use it.

Questions to ask:

  • Where do users already spend time?
  • What tools do they already use?
  • How can we integrate there?

BauGPT example: Construction workers:

  • Don't download apps (tried, failed)
  • Do use WhatsApp (for everything)
  • Need offline-first (poor site connectivity)

→ WhatsApp bot with async processing.

Case Study: What We Did Differently

Most construction tech: "Here's a mobile app with AI features!"

What we did:

  1. Found real pain: Language barriers causing safety incidents
  2. Built domain expertise: German building codes + multilingual construction knowledge
  3. Zero-friction distribution: WhatsApp (already on their phone)

Results:

  • 10x adoption vs mobile app
  • Zero safety incidents from miscommunication (pilot sites)
  • Workers actually use it (vs 2% usage for typical AI feature)

The Litmus Test

Before shipping any AI feature, ask:

1. Problem-first? □ Are users actively struggling with this? □ Are they paying for alternatives? □ Would they notice if this didn't exist?

2. Domain-specific? □ Does it know our industry deeply? □ Can it cite authoritative sources? □ Does it handle edge cases?

3. Zero-friction? □ Integrates into existing workflow? □ Works where users already are? □ Requires minimal learning?

If you can't check all three boxes, don't ship it.

What's Next

AI isn't magic. It's a tool.

Like any tool, it only works if:

  1. You're solving a real problem
  2. You're using it correctly
  3. People can actually access it

Most startups fail at #1 or #3.

The ones that get it right will win.

We're betting on construction tech. Others are betting on healthcare, legal, education, logistics.

The pattern is the same:

  • Real problem
  • Domain expertise
  • Zero-friction distribution

Get those three right, and AI actually works.


Building AI products? I'd love to hear what you're working on. Drop a comment or DM.

Want more? Subscribe for weekly insights on AI, growth, and shipping things that actually work.

Enjoyed this post?

Subscribe to get new posts in your inbox every week. No spam, unsubscribe anytime.

Get new posts in your inbox

Weekly insights on AI, growth engineering, and building.