BotBeat
...
← Back

> ▌

RevolutRevolut
RESEARCHRevolut2026-04-27

Revolut and NVIDIA Publish PRAGMA, the Largest Foundation Model for Banking

Key Takeaways

  • ▸PRAGMA is the largest encoder backbone for consumer banking event data published to date, trained on 26M user histories and 207B tokens across 111 countries—demonstrating that enterprise fintech companies now have the data and compute to train foundation models
  • ▸The model achieves 130% improvement on credit scoring, 164% on engagement, and 40% on recommendations, with LoRA fine-tuning matching or exceeding task-specific baselines—validating the business case for pre-trained representations in banking
  • ▸The technical architecture reveals banking's shift from hand-crafted features to learned representations: PRAGMA's tokenization strategy (semantic key, typed value, temporal coordinate) preserves numerical magnitude and contextual information while managing sequence length
Source:
Hacker Newshttps://philippdubach.com/posts/inside-pragma-revoluts-foundation-model-for-banking/↗

Summary

Revolut Research and NVIDIA have published PRAGMA, an encoder-only transformer foundation model trained on 26 million user histories spanning 24 billion events and 207 billion tokens across 111 countries. The model represents a significant shift in financial services technology: moving from hand-crafted tabular features and gradient-boosted-tree models to learned representations trained on raw transaction data. PRAGMA scales from 10 million to 1 billion parameters and uses a novel input representation that preserves numerical magnitude information and temporal context while keeping sequence length manageable.

The technical approach is tailored specifically for banking's discriminative tasks. Rather than GPT-style text generation, PRAGMA uses BERT-style masked language modeling with three masking strategies: standard token masking (15%), whole-event masking (10%), and semantic-type masking (10%). Numerical values are mapped to learned percentile buckets, categorical values to single tokens, and timestamps encoded both as compressed log-seconds and periodic sinusoids. The model achieves substantial improvements across downstream tasks: 130.2% relative improvement on credit scoring (PR-AUC), 163.7% on communication uplift (AUUC), and 40.5% on product recommendation (mAP). Fine-tuning with LoRA—updating only 2-4% of parameters—consistently matches or exceeds training task-specific models from scratch.

The publication of PRAGMA signals a paradigm shift in banking technology: foundation models trained on transaction data are moving from closely-guarded trade secrets to published research. Revolut's decision to disclose results under author affiliations suggests internal foundation models have become a competitive disclosure point in the industry, not a hidden advantage. This work also echoes Nubank's earlier nuFormer publication, indicating a broader shift in fintech's approach to representation learning.

  • Publishing PRAGMA with author affiliations marks a turning point: internal foundation models in fintech are moving from trade secrets to competitive disclosures, similar to Nubank's nuFormer

Editorial Opinion

PRAGMA represents a maturation moment for enterprise AI: large fintech firms now have the scale and resources to train competitive foundation models where most teams lack the data, GPU budget, or organizational will. The publication of results under Revolut's name is perhaps more significant than the technical contributions—it signals that foundation models have become table stakes for consumer financial services, not a hidden edge. However, the absolute improvements are commercially redacted and presented as relative gains against internal baselines, so real-world impact should be calibrated carefully.

Machine LearningDeep LearningFinance & FintechResearch

More from Revolut

RevolutRevolut
RESEARCH

PRAGMA: Revolut Foundation Model Brings Transformer Architecture to Banking Event Sequences

2026-04-15
RevolutRevolut
FUNDING & BUSINESS

Revolut Secures Full UK Banking Licence, Marking Major Milestone After Three-Year Wait

2026-03-11

Comments

Suggested

Pilot ProtocolPilot Protocol
RESEARCH

Pilot Protocol Launches Novel Reputation System for AI Agents, Ditching Blockchain for Speed

2026-04-27
NVIDIANVIDIA
RESEARCH

Guess-Verify-Refine: Data-Aware Algorithm Achieves 1.88x Speedup for Sparse-Attention Decoding on Blackwell

2026-04-27
MetaMeta
RESEARCH

Meta Unveils Decoupled DiLoCo: New Architecture for Resilient, Distributed AI Training Across Global Data Centers

2026-04-27
← Back to news
© 2026 BotBeat
AboutPrivacy PolicyTerms of ServiceContact Us