newsence
來源篩選

There Is an AI Code Review Bubble

Hacker News

The article argues that the AI code review market is experiencing a 'bubble' with numerous companies entering the space, making differentiation difficult. The author questions how to stand out when many tools claim to be excellent at bug detection, a subjective and hard-to-prove metric.

newsence

AI程式碼審查存在泡沫

Hacker News
大約 1 個月前

AI 生成摘要

本文認為AI程式碼審查市場正經歷一個「泡沫期」,眾多公司紛紛投入此領域,導致差異化困難。作者質疑在許多工具都聲稱擅長抓蟲的情況下,該如何脫穎而出,因為抓蟲能力是主觀且難以驗證的指標。

There is an AI Code Review Bubble | Greptile Blog

Image Image Image Image Image Image

There is an AI Code Review Bubble

Daksh Gupta • Jan 24, 2026

Table of Contents

Table of Contents

Today, we're in the hard seltzer era of AI code review: everybody's doing them. OpenAI, Anthropic, Cursor, Augment, now Cognition, and even Linear. Of course, there's also the "White Claws" of code review: pure-play code review agents like Greptile (that's us!), CodeRabbit, Macroscope, and a litter of fledgling YC startups. Then there are the adjacent Budweisers of this world:

Image Image

Amazingly, these two were announced practically within 24 hours of each other.

Image

As the proprietors of an, er, AI code review tool suddenly beset by an avalanche of competition, we're asking ourselves: what makes us different?

How does one differentiate?

Based on our benchmarks, we are uniquely good at catching bugs. However, if all company blogs are to be trusted, this is something we have in common with every other AI code review product. One just has to try a few, and pick the one that feels the best.

Unfortunately, code review performance is ephemeral and subjective, and is ultimately not an interesting way to discern the agents before trying them. It's useless for me to try to convince you that we're the best. You should just try us and make up your own mind.

Instead of telling you how our product is differentiated, I am going to tell you how our viewpoint is differentiated - how we think code review will look in the long-term, and what we're doing today to prepare our customers for that future.

Our thesis can be distilled into three pillars: independence, autonomy, and feedback loops.

Independence

We strongly believe that the review agent should be different from the coding agent. We are opinionated on the importance of independent code validation agents. In spite of multiple requests, we have never shipped codegen features. We don't write code; an auditor doesn't prepare the books, a fox doesn't guard the henhouse, and a student doesn't grade their own essays.

Today's agents are better than the median human code reviewer at catching issues and enforcing standards, and they're only getting better. It's clear that in the future a large percentage of code at companies will be auto-approved by the code review agent. In other words, there will be some instances where a human writes a ticket, an agent writes the PR, and another agent validates, approves, and merges it.

This might seem far-fetched but the counterfactual is Kafkaesque. A human rubber-stamping code being validated by a super intelligent machine is the equivalent of a human sitting silently in the driver's seat of a self-driving car, "supervising".

If agents are approving code, it would be quite absurd and perhaps non-compliant to have the agent that wrote the code also approve the code. Only once would you have X write a PR, then have X approve and merge it to realize the absurdity of what you just did.

Autonomy

Something that Greptiles generally agree on is that everything that can be automated, will be automated.

Code validation - which to us is the combination of review, test, and QA, is an excellent candidate for full automation. It's work that humans don't want to do, and aren't particularly good at. It also requires little in the way of creative expression, unlike programming. In addition, success is generally pretty well-defined. Everyone wants correct, performant, bug-free, secure code.

While some other products have built out great UIs for humans to review code in an AI-assisted paradigm, we have chosen to build for what we consider to be an inevitable future - one where code validation requires vanishingly little human participation. We have no code review UI, and view ourselves as more of a background automation or "pipes" product. Human engineers should be focused only on two things - coming up with brilliant ideas for what should exist, and expressing their vision and taste to agents that do the cruft of turning it all into clean, performant code.

Loops

Not long ago, we released our Claude Code plugin. It can do many things - but most notably, you can ask Claude Code to pull down and address Greptile's comments from the PR. You can ask it to keep going until there are no new comments, waiting a few minutes for a review after each push.

This is a step towards the future we're excited about: Human expresses intent, coding agent executes, validation/review agent finds issues and hands them back - kicking off a loop until it approves and merges. If there is ambiguity at any point, the agents Slack the human to clarify.

Image

The question of how these things are different is important. Unlike picking IDEs and coding agents that ostensibly have low switching costs, code review products are harder to rip out, so your decision will very likely turn out to be a long-term one, especially if you're a large company.

We've been around for about as long as AI code review has been around. It has gone from a fringe interest of the world's most adventurous vibecoders to a mainstream product that our enterprise users (including two of the Mag7) often describe as a "no-brainer" purchase.

Yet, our guess on where this goes is about as good as anyone else's. Meanwhile, we'll keep doing what we've always done - trying to make things our users love.

Keep Reading

Sandboxing agents at the kernel level

Tracing the open syscall to understand how containers conceal files from agents.

AI Code Review: Should the Author Be The Reviewer?

AI reviewing AI code? Exploring the paradox of using AI to review AI-generated code & whether this creates conflict or genius.

The Content-ification of Software

Explore how AI will transform software development by enabling highly specialized software companies to profitably service extremely niche markets.

Subscribe to our blog

Get the latest posts and product updates delivered to your inbox

See Greptile in action

Book a personalized demo to see how Greptile can help your team ship faster

Image Image

© 2026 Tabnam, Inc.