AI Street

AI Street

Quantifying Intangible Value with AI

Matt Robinson's avatar
Matt Robinson
Aug 24, 2025
∙ Paid

Hey, it’s Matt. In this AI Street Markets:

🎙️ An interview with Kai Wu, Sparkline Capital founder & CIO, on how he’s using LLMs to better quantify intangible assets.


INTERVIEW

Quantitative investors have historically relied on accounting data and price metrics.

Kai Wu thinks they're missing the soft factors that drive stock performance today.

As Sparkline Capital founder and CIO, he uses AI to analyze patents, corporate communications, and other unstructured data to identify what he calls "intangible value"—the intellectual property, brand strength, and human capital that he believes traditional financial statements understate.

He started his career at GMO working on Jeremy Grantham's $40 billion asset allocation team, helping manage a $2.5 billion global macro hedge fund. In 2014, he co-founded Kaleidoscope Capital, a quantitative hedge fund in Boston that grew to $350 million in assets before selling his stake in 2018.

He founded Sparkline Capital that same year, spending time exploring where the investment industry was headed and discovering large language models—well before they became mainstream. He launched his first ETF in 2021 and has since built a suite of active ETFs centered on his intangible value framework.

In our conversation, Wu explains how he applies AI on centuries of patent data and culture indicators, why he thinks the line between quantitative and fundamental investing is blurring, and why transfer learning made text-based factor investing viable. He also shares his view on what investors can learn from Renaissance Technologies’ use of unstructured data.

This interview has been edited for clarity and length. 

❝  

"The four largest companies today by market value do not need any net tangible assets. They are not like AT&T, GM, or Exxon Mobil, requiring lots of capital to produce earnings. We have become an asset-light economy."

 Warren Buffett in 2018

Tell me about Sparkline Capital

The main business at Sparkline is asset management through ETFs. We’re still trying to create alpha using quantitative techniques but in terms of structure we are trying to skate to where the puck is going. A lot of assets and investor interest are moving into ETFs, specifically active ETFs.

Historically, ETFs were synonymous with index funds. But due to a variety of changes, we’re now seeing more active strategies put into ETF wrappers. That provides efficiency, operational benefits, and tax advantages compared with traditional hedge funds. There’s a lot of interest in that category.

I launched my first fund four years ago, a second one about a year ago, and now I’m building out a suite of products centered on the concept of intangible value. I believe that if value investing, in the Ben Graham and Warren Buffett sense, is going to thrive in the digital economy, then we need to adapt the definition of intrinsic value to include intangible assets.

The techniques we use—LLMs and unstructured data—are what make this possible. If you just look at accounting data, you’re missing out on the most valuable information on intangible assets. There’s simply not enough information. Why wouldn’t you also look at the 80-plus percent of data that’s unstructured? And why wouldn’t you use the latest tools to analyze it?

Nobody I know is really trying to solve this problem.

How did you end up focusing on intangible value?

Historically quants have excelled in some dimensions, right? We have the ability to process larger amounts of data faster and in a more disciplined way. We're less emotional, so we're not gonna just sell all our stocks in ‘08.

The downside of being a quant is that historically only a small percentage of the potential universe of information on companies is accessible. Until more recently, quants have been restricted to accounting-based information, price, volume, PE ratios, asset turnover ratios, all that kind of stuff.

But a lot of information isn't even digital. And even that which is digital has historically been very difficult for quants to ingest because you can't take these textual documents and put them through linear regression.

And that’s where LLMs are a huge breakthrough for us, because now we can start saying, let’s base things on text. I wrote a paper called Text-Based Factor Investing, and you can probably guess what that means. The idea was: can we create factors—like Value, Carry, Momentum—but derived from textual data instead? Using NLP, we can generate culture scores or innovation scores and turn those into factors that can be incorporated alongside traditional ones in an investment process.

I think we’re seeing a convergence. Quants are starting to encroach on the discretionary investor’s world, and are now able to incorporate information that historically wouldn’t have been accessible. At the same time, it’s moving in the other direction too. Discretionary investors are being given tools they can use without needing to be master coders. A lot of what we’ve mentioned is increasingly available off the shelf—though of course, there’s still the challenge of sorting through all the different vendors.

They theoretically enable an analyst with no programming experience to benefit from many of the insights AI can provide. Over time, I think these things are going to meet in the middle, where the distinctions between quant and fundamental will matter less.

What tools do you use?

One of the challenges today is that there's been a proliferation in the number of vendors. If you're a fund manager, you're being pitched a million things from different startups.

We can count the number of foundational model companies on one hand, but on top of that there's a whole layer claiming to offer specialized services to investors. It's just really difficult to diligence.

What do you actually use?

I generally try to go homegrown, although I'm probably unique because I've been working with large language models since about 2019.

What were you doing back then? Not too many people knew about LLMs at that time.

I had a career transition. I sold my last hedge fund and was starting my business. It gave me some time to reset and say: Where are the big industry trends? And that's where I discovered large language models and natural language processing techniques. My goal was to quantify intangible assets from the perspective of a value investor. I used to work for GMO, a quant value investment manager, and the problem I recognized was that a lot of the intangible assets were not accurately measured by accounting statements.

The question became, how can we go about quantifying the value hidden in patents or trademarks, these unstructured data sets? It became clear to me that LLMs and AI provided the key to unlocking these data.

When did you first hear about LLMs?

Obviously the [Attention is All You Need] paper and BERT were the big breakthroughs in 2017-8. But I think that the bigger breakthrough was actually less about the models and more about the data.

Deep neural networks were invented decades ago. It was just that computers weren't fast enough and there wasn't enough data to train them in an effective way.

So the architectural breakthrough of the transformer was better than the alternatives at the time. But I don't think that was the actual game changer. The game changer was transfer learning. At the time, you could develop a specialized model trained on 10-Ks, but the problem was there just weren’t many 10-Ks. You’re talking about an extremely small sample to train a large model on so the results wouldn’t be very accurate.

User's avatar

Continue reading this post for free, courtesy of Matt Robinson.

Or purchase a paid subscription.
© 2026 Matt Robinson · Privacy ∙ Terms ∙ Collection notice
Start your SubstackGet the app
Substack is the home for great culture