Discussion about this post

User's avatar
Guido's avatar

Interesting that the core bet here is treating market data as a 'language' to learn from scratch. What I'm curious about is the trade off: pre-training from zero vs. continued pre-training on an open-source base, then fine-tuning for trading objectives. The compute costs are orders of magnitude apart, but perhaps I'm failing to consider that HFs can go through expensive training runs so...

Great article.

Neural Foundry's avatar

Fascinating dive into the compute arms race. The parallel between quant shops and AI labs is esp apt, both essentially running prediction optimization pipelines at scale. That XTX €1B data center investment signals this isn't experimental anymore, its infrastructural. The point about smaller firms lacking GPU access being at a real disadvanatge makes sense when training becomes teh edge itself.

2 more comments...

No posts

Ready for more?