DEV Community

Louis Dupont
Louis Dupont

Posted on

15 7 3 6 6

Why Most AI Teams Are Stuck πŸ€”

A few years ago, I worked on a Generative AI project, a customer-facing AI assistant. The company had great data and was convinced AI could turn it into something valuable.

We built a prototype fast. Users were excited.

Iteration was quick. Each tweak made the AI feel better.

Then we hit a wall.

We kept changing things, but… was it actually getting better? Or just different?

We didn't know.

When "Iterating" Is Just Making Random Changes

At first, improving the AI felt obvious. We spotted issues, fixed them, and saw real progress. But suddenly, everything slowed down.

  • Some changes made things better, but we weren't sure why.
  • Other changes made things worse, but we couldn't explain how.
  • Sometimes, things just felt… different, not actually better.

It took me way too long to realize: we weren't iterating. We were guessing.

We were tweaking prompts, adjusting retrieval parameters, fine-tuning the model… but none of it was measured. We were just testing on a few cherry-picked examples and convincing ourselves that it felt better.

And that's exactly how most AI teams get stuck.

Better on a Few Examples Isn't Better

When you're close to a project, it's easy to think you can tell when something improves. You run a few tests. The output looks better. So you assume progress.

But:

  • Did it actually improve across the board?
  • Did it break something else in the process?
  • Are you fixing what users actually care about or just what you noticed?

Most teams think they're iterating. They're just moving in random directions πŸ”

Iterate Without Measurement... and Fail!

And that's the real problem.

Most teams, when they hit this wall, do what we did: try more things.

  • More prompt tweaks.
  • More model adjustments.
  • More retrieval fine-tuning.

But real iteration isn't about making changes. It's about knowing, at every step, whether those changes actually work.

Without that, you're just optimizing in the dark.

So What's the Fix?

The teams that move past this don't just build better models, they build better ways to measure what β€œbetter” means.

Instead of relying on gut feeling, they:

  • Define clear success criteria. What actually makes an answer useful?
  • Measure changes systematically. Not just on a few cherry-picked examples.
  • Make sure improvements don't break what already works.

The Bottom Line

Most AI teams don't struggle to build AI. They struggle to improve it.

I learned this the hard way. But once I started treating iteration as something that needs clear feedback loops, not gut feeling, everything changed.

In a following article, I'll break down how to actually measure AI improvement without getting trapped by misleading metrics.

πŸ‘‰ Follow to get notified when it's out.

πŸ“Œ In the meantime, if you want to go deeper on AI iteration and continuous improvement, check out my Blog.

Heroku

Built for developers, by developers.

Whether you're building a simple prototype or a business-critical product, Heroku's fully-managed platform gives you the simplest path to delivering apps quickly β€” using the tools and languages you already love!

Learn More

Top comments (2)

Collapse
 
kurealnum profile image
Oscar β€’

Will definitely keep this in mind. Great post!

Collapse
 
nadeem_zia_257af7e986ffc6 profile image
nadeem zia β€’

good information provided

Some comments may only be visible to logged-in visitors. Sign in to view all comments.

5 Playwright CLI Flags That Will Transform Your Testing Workflow

  • 0:56 --last-failed
  • 2:34 --only-changed
  • 4:27 --repeat-each
  • 5:15 --forbid-only
  • 5:51 --ui --headed --workers 1

Learn how these powerful command-line options can save you time, strengthen your test suite, and streamline your Playwright testing experience. Click on any timestamp above to jump directly to that section in the tutorial!

πŸ‘‹ Kindness is contagious

Please leave a ❀️ or a friendly comment on this post if you found it helpful!

Okay