AI & Machine Learning

My CEO Is an AI Clone, the ECB Runs on ML, and a Cambridge Chip Just Made Data Centers Sweat

A bank CEO deployed his AI clone on a live earnings call, the ECB revealed it has used ML for monetary policy since 2022, and a brain-inspired hafnium oxide chip promises to slash AI energy consumption by 70%. Finance's AI transformation is no longer coming — it's already running the room.

Bhanu Pratap
Share this article

Share:

A bank CEO deployed his AI clone on a live earnings call, the ECB revealed it has used ML for monetary policy since 2022, and a brain-inspired hafnium oxide chip promises to slash AI energy consumption by 70%. Finance's AI transformation is no longer coming — it's already running the room.
Table of Contents

There is a specific kind of déjà vu you get when you realize the person speaking on an earnings call is not actually a person. Customers Bank CEO Sam Sidhu induced exactly that feeling during the bank’s Q1 2026 results call when he revealed — after the fact — that his AI clone had delivered the prepared remarks. Not a voice-over, not a script. A digital replica of the CEO, running in real time, indistinguishable enough that analysts didn’t catch it. He then announced a multi-year deal to embed OpenAI engineers inside the bank. The prepared remarks were, appropriately, about how AI would transform everything.

Meanwhile, 4,000 miles away, the European Central Bank quietly disclosed that it has been running a machine learning model as part of its monetary policy toolkit since the end of 2022. Not a side experiment. Not a dashboard widget. A component of the analytical process used to set interest rates for the Eurozone. And in a Cambridge laboratory, researchers published results on a hafnium oxide chip that can process and store data simultaneously — like a neuron — cutting AI energy consumption by up to 70%.

Today’s three stories are connected by a single thread: the AI transformation of finance has stopped being a future projection and started being a present-tense operating reality. Let’s get into it.

The Bank That Let Its CEO Go Digital

Customers Bank is a $22 billion regional bank based in West Reading, Pennsylvania. It is not JPMorgan. It does not have a $19.8 billion tech budget. It does, however, have a CEO who apparently decided that the best way to signal commitment to AI is to be the AI on the most important call of the quarter.

Sam Sidhu’s AI clone delivered the prepared portion of the Q1 2026 earnings call on April 27th. Sidhu then came on for the Q&A — in human form — and announced that Customers Bank had signed a strategic collaboration with OpenAI to “redefine the commercial banking operating model.” The deal involves OpenAI embedding engineers directly at the bank, a structure more similar to a consulting engagement than a software license. The bank is not buying a product; it is co-building a capability.

The focus is three areas: commercial lending, deposits, and payments. The most concrete target is lending. Today, closing a commercial loan takes 30 to 45 days, covering underwriting, document collection, and legal negotiations. Customers Bank is targeting 7 days. That is not an incremental improvement — it is a structural compression of one of banking’s most labour-intensive workflows by roughly 80 percent.

The efficiency ratio target (from 49 percent to the low 40s by 2027) is the financial headline, but the engineering headline is what OpenAI is actually building: custom AI capabilities trained on the bank’s own processes, data, and institutional knowledge. This is the opposite of buying a SaaS product. The bank is treating its loan operations as training data and its process knowledge as a moat, then layering generative AI on top as the execution layer.

The stunt with the earnings call clone is silly in isolation. In context, it is a piece of deliberate signaling: this is a leadership team that is not hedging. They are all in, and they want the market to know it.

The ECB’s Quiet ML Confession

Every central bank publishes reams of research. The trick is noticing when a blog post is actually a policy disclosure. The ECB’s April 21 post, “Navigating uncertain times with the help of artificial intelligence,” was the latter.

Since the end of 2022, the ECB’s QRF (Quantile Random Forest) model has been part of the analytical toolkit used to prepare monetary policy decisions. The model draws on roughly 60 macroeconomic indicators, capturing inflation expectations, cost pressures, real economic activity, and financial conditions. It has been used to help produce short-term inflation forecasts and the risk assessment around the baseline projection.

To be clear about what this is and is not: the ECB’s interest rate decisions are still made by the Governing Council, by humans, in a room. The QRF model does not set rates. But it shapes the analysis that rate-setters see. It filters the signal from 60 indicators and produces a probabilistic inflation view that sits in front of the people making the call. In the language of decision systems, this is not a rubber stamp — it is upstream influence.

The practical effect is real-time inflation risk tracking at a granularity that human analysts working from batch data cannot match. The ECB disclosed that the model has helped flag tail risks — notably upside inflation surprises — faster than conventional approaches. In an environment where energy price shocks and supply chain disruptions move faster than quarterly data releases, this matters.

What makes the ECB disclosure significant for the broader finance industry is not the technology itself — random forests are not new — but the institutional signal it sends. If a central bank governing 20 countries and $14 trillion in GDP is comfortable embedding ML into monetary policy preparation, the argument that regulated financial institutions cannot use AI in consequential decisions has just become significantly harder to make.

The Great AI Inflation Debate

Central banks are not just using AI — they are also being asked to think about AI as a macroeconomic force that may reshape the very variables they target. This debate has sharpened considerably in April 2026.

On the deflationary side, Mike Hunstad, head of Northern Trust’s $1.4 trillion asset management division, told the Financial Times that AI “could prove to be one of the biggest positive supply shocks in modern economic history.” The argument is straightforward: if AI delivers sustained productivity uplift across services, healthcare, and knowledge work, it will lower unit costs of production persistently — the way containerisation did for goods trade in the 1970s, or the internet did for information services in the 1990s.

On the inflationary side, Goldman Sachs has done the maths on energy. Power inflation in the US ran at 6.9 percent year-on-year through December 2025. Goldman estimates higher electricity costs from AI data centre demand will add 0.2 percentage points to headline inflation in 2026 and 0.15 percentage points in 2027. These are not catastrophic numbers in isolation, but they are non-trivial when the Fed is trying to thread a needle back to 2 percent.

The San Francisco Fed’s April framework paper (referenced in last week’s post on model risk) went further, warning about “expectation-driven asset bubbles” amplified by AI narrative trading and “model monocultures” creating correlated systemic risk. Central banks are now wrestling with the possibility that the very tool reshaping their analytical toolkit is also reshaping the economic system they are trying to measure.

This is new territory. Traditional monetary models assume human agents with bounded rationality and sticky expectations. AI systems are neither bounded nor sticky. They update fast, they correlate, and in financial markets they can amplify rather than dampen volatility. What this means for the neutral rate, for the inflation-output trade-off, and for forward guidance is genuinely unresolved — and central bank economists are not pretending otherwise.

The 70% Energy Breakthrough Nobody Talked Enough About

There is a physics problem at the heart of modern AI. It is not an algorithm problem or a data problem. Moving data between memory and processor — the fundamental operation in every neural network forward pass — consumes energy. At the scale of a 400-billion-parameter model running millions of inference calls per day, this adds up to something approaching the electricity budget of a small country.

Researchers at the University of Cambridge, publishing in April 2026, have a candidate solution that is both simple in concept and genuinely novel in execution. They engineered a new nanoelectronic device using a modified form of hafnium oxide — a material already used in standard semiconductor manufacturing — that functions as a highly stable, low-energy memristor.

A memristor does something conventional transistors cannot: it stores and processes information in the same physical location. No data movement. No energy cost of shuttling numbers between RAM and CPU. The device mimics how biological neurons work — synaptic weight storage and computation are co-located — which is why this class of hardware is called neuromorphic.

The Cambridge team’s hafnium oxide memristor is notable for two reasons beyond the basic concept. First, it achieves high stability — a persistent problem with memristors, which tend to drift over time and introduce noise. The modified hafnium oxide structure suppresses this drift, which means the device could actually survive the read/write cycles of a real inference workload. Second, because hafnium oxide is already in the semiconductor supply chain, manufacturing these devices does not require exotic materials or entirely new fab processes. The path from lab result to production chip is shorter than for many neuromorphic proposals.

The headline claim — up to 70 percent reduction in AI energy consumption — is a projected figure, not a production benchmark. Real-world results will depend heavily on the architecture of the neural network and how the memristor arrays are integrated with existing compute. But the order of magnitude is striking. If even half the theoretical gain survives to production, the economics of AI inference change substantially. Data centres that currently budget for exponentially growing power costs would be looking at a fundamentally different cost curve.

Intel’s Hala Point neuromorphic system, deploying 1.15 billion neurons, has already demonstrated orders-of-magnitude better energy efficiency than conventional AI hardware in specific workloads. Cambridge’s hafnium oxide work suggests the material science to take this mainstream is now within reach.

Connecting the Threads

Three stories, one theme: AI in finance has crossed from pilot into operating reality, and the infrastructure assumptions it is overturning — time-to-close for loans, analytical toolkits for central banks, energy economics of compute — are not minor. They are foundational.

Customers Bank’s decision to let an AI clone deliver an earnings call is a marketing stunt. The decision to embed OpenAI engineers and target an 80 percent compression of commercial lending cycle time is not. That is a bet on reengineering the bank’s operating model from the ground up, and the payoff targets (efficiency ratio in the low 40s, returns improvement by 2027) are specific enough to hold the team accountable.

The ECB’s ML disclosure matters because of what it normalises. If Europe’s most important monetary authority is comfortable with machine learning upstream of rate decisions, the regulatory and reputational argument against AI in consequential financial decisions has weakened considerably. Expect other central banks to follow, or quietly to admit they already have.

And the Cambridge chip matters because energy is the binding constraint on AI at scale. The compute and algorithm problems are being solved. If the energy problem gets solved at the hardware level — through neuromorphic architectures that consume 70 percent less power — the cost curve for AI inference changes dramatically, which changes the economics of deploying AI everywhere, including in every bank branch, every insurance underwriter’s laptop, and every trading algorithm’s colocation rack.

What to Watch

Watch whether other regional US banks announce similar OpenAI or frontier-model embedding deals in Q2 2026. Customers Bank is not a systemically important institution, but it is exactly the size of bank where a step-change in operational efficiency creates competitive pressure on peers who are still running pilots.

Watch the ECB’s next Economic Bulletin for further detail on how QRF model outputs are weighted against human analyst forecasts. The question is not whether ML is in the toolkit — it is — but how much weight it carries relative to traditional models when they diverge.

Watch patent filings and spin-out announcements from the Cambridge neuromorphic group. Academic results at this stage of maturity often translate to startup formation within 12 to 18 months. If hafnium oxide memristors are as manufacturable as they appear, the commercial interest will be fast.

The earnings call clone was the headline. The loan cycle compression is the business case. The ML in Eurotower is the policy signal. And the 70 percent energy chip is the infrastructure wildcard. All four are live, and all four matter.


Sources

Back to Blog

Related Posts

View All Posts »