Back to Blog

Fraud Detection Is a Product Problem, Not a Data Science Problem

Most AI fraud detection initiatives underperform not because the model is wrong, but because fraud detection is treated as a data science problem instead of a product problem.

Views expressed are personal and do not represent any employer, partner, or client.

A model with 95% accuracy that generates 10,000 daily alerts is less valuable than a model with 90% accuracy that generates 500 alerts an analyst can realistically act on. Most organizations learn this the hard way.

There is a version of this story that plays out at nearly every financial institution investing in AI-powered fraud detection. A data science or ML team builds a model. The model performs well in testing. It gets deployed. And then, slowly or suddenly, things go sideways.

False positive rates climb. Operations teams drown in alerts they cannot action. Customers get locked out of their accounts. The fraud team loses trust in the system, starts overriding it manually, and the model's value quietly erodes. Six months later, someone asks why the AI investment has not delivered the returns that were promised.

The instinct is to blame the model. Retrain it. Add more features. Bring in a new vendor. But in most cases, the model is not the problem. The problem is that fraud detection was treated as a data science and ML challenge when it is fundamentally a product problem.

The Model Is Not the Product

A fraud detection model is a component. It is an important one, but it is still just a component. The product is the entire system that surrounds it: how alerts get surfaced, how analysts prioritize and investigate them, how decisions get made under time pressure, how feedback from those decisions flows back into the model, and how the experience feels to the end customer who may never know any of this is happening.

When organizations treat the model as the product, they optimize for the wrong thing. They chase AUC scores and precision-recall curves in isolation, without asking whether the humans downstream can actually use what the model produces.

This is not a technical distinction. It is a product design decision about where to place the burden of judgment, and it changes everything about how the system should be built.

Why Product Thinking Changes the Outcome

When you approach fraud detection as a product problem, the questions change. Instead of "how do we improve the model," you ask "how do we improve the decision." That shift opens up a different set of levers.

First, it forces you to define what a good outcome actually looks like for every stakeholder. For the fraud analyst, a good outcome might be fewer alerts with higher confidence and enough context to make a decision in under two minutes. For the customer, it is never knowing the system exists unless something is genuinely wrong. These are product requirements, not model requirements.

Second, it puts the feedback loop at the center of the architecture instead of treating it as an afterthought. In most fraud systems I have seen, the loop from alert to investigation to outcome to model retraining is either broken or so slow that the model is always fighting the last war. Product-led fraud systems treat that feedback loop as a first-class feature, not a quarterly data science or ML exercise.

Third, it forces honest measurement. Data science and ML teams naturally measure model performance. Product teams measure business outcomes. The difference matters. A model might be improving its detection rate while the overall system is getting worse because alert fatigue is causing analysts to miss the ones that matter. Only product-level measurement catches that.

The False Positive Problem Is a Design Problem

If there is one metric that exposes the gap between data science thinking and product thinking in fraud detection, it is the false positive rate.

False positives are not just a model accuracy issue. They are a customer experience issue, an operational cost issue, and a trust issue. Every false positive is a real person whose transaction was declined, whose account was frozen, or whose onboarding was interrupted. At scale, a false positive rate that looks acceptable in a confusion matrix can translate to thousands of legitimate customers having a terrible experience every day.

The product-minded approach to false positives is not simply "make the model more precise." It is "design the system so that false positives cause the least possible damage." That might mean tiered alert routing where low-confidence flags get a soft challenge instead of a hard block. It might mean giving analysts richer context so they can clear false positives faster. It might mean building customer-facing recovery flows that feel respectful instead of adversarial.

These are design decisions, not data science or ML decisions. And they often have a bigger impact on the customer and the business than any model improvement could.

What This Means for How Teams Are Structured

If fraud detection is a product problem, it follows that the team building it needs product leadership, not just data science and ML leadership.

The most effective fraud teams I have worked with or built have a product manager who owns the end-to-end experience, not just the model roadmap. That PM understands threat intelligence well enough to prioritize which fraud vectors to address. They understand the analyst workflow well enough to design useful tooling. They understand the customer journey well enough to know where friction is acceptable and where it is not. And they have the organizational authority to make trade-off decisions that cut across data science, ML engineering, operations, and compliance.

This does not diminish the role of data science or ML. It elevates it by giving those teams a clearer problem definition and a tighter feedback loop. The best data scientists and ML engineers I have worked with prefer this structure because it means their work actually ships and actually gets measured against real outcomes, not just offline benchmarks.

The Shift That Needs to Happen

Billions are flowing into AI fraud prevention. The investment is not the problem. The discipline around deployment is.

The shift is easy to describe. It is harder to execute, because it requires organizational change, not just technical change. Treat the model as a component, not the product. Measure business outcomes, not just model metrics. Design for the humans in the system, both the analysts and the customers. Build feedback loops that actually close. And put someone in charge who can see the whole picture, not just the algorithm.

Fraud detection is hard. The adversaries are adaptive, the stakes are high, and the tolerance for error is low. That is exactly why it deserves product thinking, not just better models.


Shyam Menon is a product leader specializing in AI-powered fraud prevention and identity verification for financial services. He writes at shyammenon.com.