HRT Trains AI Models on Trading Data
The quant firm has developed transformer-based models using decades of market microstructure data.
Hey, it’s Matt. Welcome back to AI Street. This week:
HRT on Building “Foundation Models for Automated Trading”
Top Papers on AI in Finance Q4 2025: SSRN
JPMorgan’s Dimon: banks must invest in AI or get left behind. + More News
Hudson River Trading is building foundation-style models trained on decades of global market data, applying techniques similar to those used in frontier language models for automated trading.
The firm is training these models on more than two decades of data spanning equities, futures, and cryptocurrencies, totaling over 100 terabytes. That translates into “something like trillions of tokens, in the same realm as what you train frontier language models on,” said Marc Khoury, a researcher on HRT’s AI team, speaking at an academic conference.
At a high level, HRT’s goal is to model markets as sequences of interactions. Electronic markets generate detailed streams of activity, including full limit order books, executed trades, and order-level events such as placements, cancellations, and fills. According to Khoury, much of the predictive signal lies in how these sequences evolve over time, especially during fast-moving conditions.
(We’ve previously discussed how training transformers on specific domains — the weather, payments, grocery store sales — have shown promising results.)
Khoury showed charts indicating that more data boosted the model’s predictive performance. “As I increase the model size, the model continues to improve,” Khoury said, adding that the pattern mirrors what researchers see when scaling large language models.
To train a transformer model on markets, trade data has to be converted into tokens. In language models, tokenization typically breaks text into subword pieces. See the below example from OpenAI’s tokenizer:
In market data, tokenization means mapping events into discrete units a model can learn from. One approach groups events into fixed time intervals, such as one-minute windows. Another bundles a fixed number of events together. HRT described both as active design choices, each with tradeoffs for model performance and cost.
“Language modeling people would call this tokenization,” Khoury said. In this context, each token represents a unit of market activity rather than a word. The goal is to preserve market structure while keeping training and inference computationally feasible.
For more background on tokenization of market data, check out my chat with Juho Kanniainen, a professor at Tampere University’s Data Science Research Centre, on treating limit order books as “language” here.
Transformer models offer one way to address this problem. Because they are designed to process long sequences of diverse events, transformers can adapt to changing conditions without relying entirely on fixed, hand-engineered features.
Training models at this scale requires substantial infrastructure. Khoury did not disclose how many GPUs the firm operates, but he said HRT runs its own state-of-the-art data center and that at one point, the firm’s hardware purchases were large enough that he joked it “bottlenecked” GPU deliveries on the U.S. East Coast.
The research effort comes as HRT posts record results. Bloomberg reported that the firm generated an estimated $12.3 billion in net trading revenue in 2025. The company has also been expanding beyond high-speed trading into longer-horizon strategies, while continuing to invest heavily in artificial intelligence and data-driven trading models. The firm accounts for roughly 10 percent of U.S. equity trading volume.
An HRT spokesperson declined to comment on the firm’s AI training efforts.
Takeaway
After tokenizing 100 terabytes of market events into a trillion-token-scale training set, HRT showed internal results where predictive accuracy kept improving as they scaled both data and model size, mirroring the scaling behavior seen in LLMs.
Outlook
There’s a lot to unpack with the implications of HRT building a foundation model for finance. Khoury’s presentation shows they’re getting promising results, which to me leads into all sorts of other questions:
How many other funds are pursuing similar market-sequence models?
Are computing resources becoming a meaningful edge for market makers?
If transformers keep outperforming traditional AI models, how big of an impact is that on GPU demand and data-center buildouts?
Further Reporting
If you’re building in this space, please reach out by replying to this email. I’m still mapping the landscape and have a lot more questions.
Related Reading
RESEARCH
Top Papers on AI in Finance
SSRN just published its Q4 2025 leaderboard for AI-in-finance research: the 10 most-downloaded papers posted that quarter, plus which organizations did the downloading. The list is a quick snapshot of what’s getting traction right now, with “agentic” workflows, LLMs in systematic investing, and GenAI model risk showing up alongside core quant portfolio work.
The most downloaded paper was “Agentic Artificial Intelligence in Finance: A Comprehensive Survey” by lead author Irene Aldridge. I interviewed her last month about how AI agents are the next evolution in trading.
Big themes: agentic AI (survey + real estate), ChatGPT for systematic investing, GenAI model risk management, ML-meets-Markowitz portfolio construction
Also notable: learning firm characteristics for asset management, investor automation and effort allocation, LLMs for central bank communication
COMPUTE
Tracking “Compute” Constraints
We hear more about AI spending and the building out of data centers than we do about the overall demand for compute, or the hardware that trains and runs AI. Or at least, I’ve found it hard to track down this sort of information.
So I “built” a compute capacity constraint index using Claude Code. The name is a bit misleading. You do not spend much time coding with it. It does the coding for you.
My goal is to track whether the Amazons, Microsofts, Googles, and other infrastructure have more customer demand than they do compute capacity. My thinking is that this gives some sense of AI demand and whether or not all this supply being built will be met with more demand.
This is very much a work in progress, so if you have ideas on how to structure this, please reply to this email.
SPONSORSHIPS
Reach Wall Street’s AI Decision-Makers
Advertise on AI Street to reach a highly engaged audience of decision-makers at firms including JPMorgan, Citadel, BlackRock, Skadden, McKinsey, and more. Sponsorships are reserved for companies in AI, markets, and finance. Email me (Matt@ai-street.co) for more details.
NEWS
AI Adoption
Dimon says JPMorgan has to invest in AI or risk getting ‘left behind’
JPMorgan’s Jamie Dimon used the bank’s Tuesday morning earnings call to defend JPM’s heavy spending on tech and AI, framing it as a competitive necessity, not a nice-to-have.
He said JPM isn’t just fighting the usual Wall Street rivals. It is also up against fintechs like Stripe, SoFi, and Revolut, which “are good players.”
Dimon’s punchiest line came in response to a question on spending from Wells Fargo analyst Mike Mayo.
“We are going to stay out front, so help us God,” Dimon said.
And he made clear JPM is not managing the business to hit a near-term efficiency line item if it risks long-term positioning.
“We’re not going to try to meet some expense target, and then 10 years from now, you’d be asking us a question, how did JPMorgan get left behind?” Business Insider
Investors Can Use AI for Proxy Votes: SEC Official
An SEC investment-management official said advisers can use AI to help make proxy-voting decisions, with one clear boundary: do not let models replace human judgment, per Bloomberg.
JPMorgan said last week it’s ditching proxy advisors to use an in-house, AI-powered platform to cast shareholder votes, according to the WSJ.
The bank will rely on its internal AI platform, it’s calling Proxy IQ, to assist with company votes by analyzing data from more than 3,000 annual company meetings and provide recommendations to portfolio managers.
SocGen Turns to Microsoft’s Copilot After Scrapping Own AI Tool
Societe Generale is dropping its internal assistant, SoGPT, and rolling out Microsoft Copilot after realizing that the gap with other tools was widening.
It’s a clear example of how even big banks are finding that building and maintaining in-house AI tools can get expensive fast. A year ago, I heard a lot about firms building these tools internally but many have since abandoned those efforts for the same reasons SocGen is. Bloomberg
Advisors warm to AI, but still have trust issues
Financial advisors are leaning into artificial intelligence to streamline their workdays, but they are drawing a firm line when it comes to who makes the final call on client advice, according to new research from Advisor360. Investment News
← There are no standards for AI or any industry-backed standards to follow, so it’s a real hurdle to adoption. At some point there will be but it will take time to get regulators, trade associations and market participants in the same room.
ROUNDUP
What Else I’m Reading
Viking Global’s Head of Trading is Stepping Down BI
LLMs contain a LOT of parameters. But what’s a parameter? MIT
Emerging AI patterns in finance (what to watch in 2026) Gradient Flow
NY Federal Judge Questions if NOT Using AI Could Be Malpractice BBG
AI in Investment Management: 2026 Outlook Two Sigma
Bank of America CEO Says AI Paying Off as Bank Cuts Costs PYMNTS
CALENDAR
Upcoming AI + Finance Conferences
CDAO Financial Services – Feb. 18–19 • NYC
Data strategy and AI implementation in the financial sector.
AI in Finance Summit NY – Apr. 15–16 • NYC
The latest developments and applications of AI in the financial industry.
AI & RegTech for Financial Services & Insurance – May 20–21 • NYC
Covers AI, regulatory technology, and compliance in finance and insurance.




