AI in finance is promising but unreliable today. Learn why ROI lagsâand the practical workflow-first steps banks and fintechs can take for 2026.

AI in Finance: Why Itâs Not Working (Yet)
Only 15% of executives in a Forrester survey said AI improved profit margins over the last year. BCG found just 5% saw widespread value. Those numbers donât read like a hype cycle; they read like an implementation problem.
If you work in banking or fintech, youâve probably felt the gap first-hand: leadership is sold on the future of generative AI, but teams are stuck wrestling with messy data, inconsistent model behavior, and the uncomfortable truth that customers still want humans for anything even slightly complicated.
Iâm bullish on AI in finance. But Iâm also convinced most organisations are approaching it backwardsâstarting with shiny tools instead of specific workflows, governance, and the âboringâ operational plumbing that makes AI reliable. The reality? The winners in 2026 wonât be the companies that bought the biggest AI contract. Theyâll be the ones that made AI predictable.
The problem isnât beliefâitâs reliability and ROI
Business leaders arenât arguing about whether AI matters. Theyâre frustrated that it doesnât work consistently enough to justify large-scale rollout.
That pattern shows up across industries, and finance amplifies it. When a wine appâs AI sommelier is âtoo nice,â itâs a brand quirk. When a bankâs AI assistant is âtoo nice,â it can become a mis-selling risk, a compliance incident, or a customer remediation bill.
Why ROI is hard to prove in financial services
Generative AI adoption in finance runs into three classic traps:
- Value is real but diffuse: If AI saves analysts 20 minutes here and 12 minutes there, itâs helpfulâbut not always captured in a clean P&L line.
- Controls add friction: Model risk management, privacy, and audit trails slow deployment (for good reasons).
- The hard costs are obvious: Data work, security reviews, integration, and change management are visible, budgeted, and painful.
This is why many firms are quietly shifting from âbig bang transformationâ to targeted automationâfraud ops, AML triage, customer service deflection, credit decisioning supportâwhere thereâs a measurable baseline and a clear owner.
Snippet-worthy truth: In finance, AI value isnât âcan it answer?â Itâs âcan it answer correctly, consistently, and in a way we can defend later?â
The âjagged frontierâ hits finance harder than most industries
AI can write code, summarise documents, and draft customer emailsâthen fail at something that feels trivial, like interpreting a time period (âlast weekâ) or mapping a suburb to the right region. Researchers call this the âjagged frontierâ: impressive performance in one domain, surprising failure in another.
Finance is full of jagged edges:
- Inconsistent source data (core banking, CRM, payment rails, collections notes, scanned PDFs)
- Ambiguous language (âarrears,â âdelinquency,â âhardship,â âchargeback,â âdisputeâ) that varies by product and jurisdiction
- Context-dependent rules (policy exceptions, thresholds, changing regulations)
A model that âmostly worksâ is still unacceptable when errors create regulatory exposure.
The data formatting tax (and why AI exposes it)
One of the most practical observations from the source story is that AI can âread patterns that donât existâ when data is formatted differently across systems.
In financial services, thatâs not hypothetical. Itâs daily reality:
- Merchant names appear in multiple forms across card, EFT, and wallet rails
- Counterparty identifiers differ between systems
- Product codes and status flags vary across business lines
- Free-text notes in collections and customer support are messy and emotionally charged
If you want reliable AI for fraud detection, credit scoring support, or personalised financial advice, you end up paying a data reformatting and definition alignment bill first. Many teams try to skip that bill. They donât skip itâthey just postpone it and suffer later.
âWe thought itâd be the easy buttonâ: where AI projects go wrong
A recurring theme from executives is blunt: they expected an easy button.
That expectation is especially common in finance because so many AI demos look magical: a chatbot summarises policy documents, a model drafts a customer response, an agent explains a transaction history. Then the pilot hits production constraints:
- Access controls and entitlement checks
- PII handling and redaction
- Audit logs and evidence trails
- Human review requirements
- Integration with ticketing, workflow, and approval systems
The most common failure pattern I see
Most teams start with: âLetâs build an AI assistant.â
They should start with: âWhich decision or workflow step is expensive, repetitive, and measurable?â
Hereâs the difference:
- Assistant-first leads to open-ended scope, unclear success metrics, and endless edge cases.
- Workflow-first leads to bounded tasks, clear evaluation, and a realistic control model.
If your use case canât answer these questions, itâs not ready:
- What is the decision the AI influences?
- What is the cost of a wrong answer?
- Whatâs the fallback when confidence is low?
- Who owns the KPI and signs off on risk?
Customer service is the clearest proof: humans arenât going away
Payments and telco examples in the source reinforce what financial services teams learn quickly: AI is great at routine interactions, and it struggles with emotional, complex, or ambiguous cases.
In fintech customer operations, the best results typically come from a hybrid model:
- AI handles authentication steps, basic FAQs, and transaction explanations
- AI drafts responses for agents (agent-assist)
- Humans take over for hardship, disputes, fraud trauma, complaints, and complex product questions
This isnât a retreat from AI. Itâs maturity.
Practical takeaway for banks and fintechs
If you want generative AI to reduce contact centre cost without damaging customer experience:
- Measure containment rate (what % issues are resolved without a human)
- Track transfer quality (does the AI pass context cleanly to the agent?)
- Optimise for time-to-human on high-friction cases
One line I agree with strongly: empathy is a hard blocker for fully automated customer conversations right now. In financial services, empathy isnât just âniceââitâs often the difference between a complaint and a retained customer.
The better approach for AI in finance: start small, then earn scale
Forrester predicts companies will delay about 25% of planned AI spending by a year. That doesnât mean AI is failing. It means organisations are learning what âproduction AIâ actually costs.
The playbook that works in finance is not glamorous, but itâs effective.
Step 1: Choose âhigh impact, low liftâ finance workflows
Good early wins share three traits: clear inputs, repeat volume, and measurable outputs.
Examples in AI in finance and fintech:
- Fraud ops triage: summarise cases, cluster similar patterns, draft SAR narratives for review
- AML alert enrichment: compile customer and transaction context into a standard analyst view
- Credit memo drafting: generate first-pass writeups with citations to source fields
- Collections agent-assist: propose compliant call scripts and next-best actions
- Customer dispute intake: classify dispute type and gather required evidence checklist
Notice whatâs missing: âautonomous decisions.â Early wins are about speed and consistency, with humans still accountable.
Step 2: Design for âNoâ (and for uncertainty)
The wine app story nailed a subtle truth: models often need permission to be critical.
In finance, you want the model to say:
- âI donât have enough information to answer.â
- âThis conflicts with policy X.â
- âRoute to a licensed adviser / compliance / human agent.â
This is not just prompting. Itâs product design:
- Confidence thresholds
- Guardrails by intent (advice vs info)
- Retrieval that cites internal policy snippets
- Hard blocks on restricted topics
Step 3: Make data boring again (standardise it)
If your AI pilot relies on heroic prompt engineering to compensate for inconsistent data, youâre building on sand.
A more durable path:
- Define canonical entities (customer, account, transaction, merchant)
- Standardise time windows and location mapping
- Create a governed feature store for modelling and analytics teams
- Build a âgolden setâ of documents for retrieval (current policies, product T&Cs, procedures)
This is where many teams win or lose.
Step 4: Evaluate like a financial institution, not a chatbot hobbyist
A finance-ready evaluation plan includes:
- Accuracy by category (not one blended score)
- Hallucination rate on restricted topics
- Stability across repeated runs
- Bias and fairness checks (especially for credit-related outputs)
- Auditability: can you recreate what the model saw and why it answered?
If you canât explain it to a regulator, itâs not ready for production.
Why 2026 will reward the âhandholdingâ vendors and teams
The source notes that AI labs and application vendors are increasingly embedding experts with customers. Thatâs not a services upsell; itâs recognition that adoption is mostly workflow engineering and change management.
In financial services, the most valuable âAI capabilityâ is often cross-functional:
- Product + operations know the edge cases
- Data teams know the quirks of lineage and quality
- Risk and compliance set the safe operating boundary
- Engineering makes it reliable and observable
Firms that build this muscle will move faster each quarter. Firms that treat AI as a plug-in will keep re-running pilots.
Snippet-worthy truth: AI in finance scales when governance, data, and workflow design scaleânot when prompts get smarter.
What to do next if youâre responsible for AI adoption in finance
If youâre staring at a 2026 roadmap and feeling pressure to âshow ROI,â Iâd focus on three moves:
- Pick one workflow per business line with a clear KPI (minutes saved, losses reduced, faster resolution time).
- Build a human-in-the-loop operating model from day one (who reviews what, and when the AI must stop).
- Invest in data standardisation specifically for your chosen workflow, not as a never-ending enterprise program.
This post is part of our AI in Finance and FinTech series, and this theme will keep coming up: the institutions seeing value arenât waiting for perfect models. Theyâre building practical systems that assume imperfectionâand still deliver safe, measurable outcomes.
If your AI plans for 2026 had to be cut down to two production bets, which workflows would you chooseâand what would you measure to prove they worked?