<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[AI Street : Research ]]></title><description><![CDATA[Analysis of the latest AI research with practical implications for trading, risk, and operations. ]]></description><link>https://www.ai-street.co/s/research</link><image><url>https://substackcdn.com/image/fetch/$s_!ezC3!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8a4fc97a-4b2d-4478-92be-ea095be05d61_800x800.png</url><title>AI Street : Research </title><link>https://www.ai-street.co/s/research</link></image><generator>Substack</generator><lastBuildDate>Mon, 20 Apr 2026 01:26:01 GMT</lastBuildDate><atom:link href="https://www.ai-street.co/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Matt Robinson]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[matt@ai-street.co]]></webMaster><itunes:owner><itunes:email><![CDATA[matt@ai-street.co]]></itunes:email><itunes:name><![CDATA[Matt Robinson]]></itunes:name></itunes:owner><itunes:author><![CDATA[Matt Robinson]]></itunes:author><googleplay:owner><![CDATA[matt@ai-street.co]]></googleplay:owner><googleplay:email><![CDATA[matt@ai-street.co]]></googleplay:email><googleplay:author><![CDATA[Matt Robinson]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[What 16,000 SEC Filings Say About AI Adoption on Wall Street]]></title><description><![CDATA[Analysis of ADV filings reveals how financial firms report AI adoption, governance policies, and related costs.]]></description><link>https://www.ai-street.co/p/what-16000-sec-filings-say-about</link><guid isPermaLink="false">https://www.ai-street.co/p/what-16000-sec-filings-say-about</guid><dc:creator><![CDATA[Matt Robinson]]></dc:creator><pubDate>Wed, 15 Apr 2026 15:31:33 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/d2ed5c0f-85e2-4e74-b41d-eea806a5387e_2190x1369.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>Matt&#8217;s note: This post was updated on Friday April 17 with more granular data from the ADV analysis.</em> </p><div><hr></div><p>As I&#8217;ve said before, AI excels at organizing tedious, repetitive material that no one would review by hand. Form ADV brochures fit that description. They&#8217;re long, dense, and there are thousands of them.</p><p>So I ran them through a local model. </p><p>I downloaded roughly 16,000 Form ADV filings filed in March, the annual disclosures registered investment advisers file with the SEC describing how they run their businesses. Of those, roughly 15,000 were standalone Part 2A brochures, covering about 12,600 advisory firms across hedge funds, private equity, venture, real estate, and traditional RIAs.</p><h3><strong>What the Filings Show</strong></h3><p>I scanned every filing for AI-related keywords. About 5,800, just under 40%, mentioned AI at all. In most cases, that meant boilerplate risk disclosures warning about potential impacts on portfolio companies or financial markets.</p><p>More than 1,200 firms described using AI in their operations. Fewer than 320 named a specific product. Of those, more than 450 firms disclosed a formal internal AI policy, and 88 said AI-related costs were being charged directly to clients or fund investors.</p><div class="callout-block" data-callout="true"><p>&#8226; <strong>Point72 Asset Management</strong> &#8212; Uses generative AI and large language models &#8220;in the <a href="https://files.adviserinfo.sec.gov/IAPD/Content/Common/crd_iapd_Brochure.aspx?BRCHR_VRSN_ID=1037764">operation of its business</a>, including in connection with investment and non-investment processes.&#8221; </p><p>&#8226; <strong>Rexford Capital </strong>&#8212; &#8220;Rexford Capital <a href="https://files.adviserinfo.sec.gov/IAPD/Content/Common/crd_iapd_Brochure.aspx?BRCHR_VRSN_ID=1038825">subscribes exclusively</a> to enterprise-grade, paid versions of major AI platforms, including offerings from OpenAI, Anthropic, and Microsoft. This matters for clients: unlike free consumer-facing versions of these tools, our enterprise subscriptions include enhanced data privacy protections, end-to-end encryption, and contractual assurances.&#8221;</p><p>&#8226; <strong>Jupiter Asset Management</strong> &#8212; Names Aladdin, FactSet, Northfield, ICE, Bloomberg, and Style Analytics. &#8220;Artificial Intelligence<a href="https://files.adviserinfo.sec.gov/IAPD/Content/Common/crd_iapd_Brochure.aspx?BRCHR_VRSN_ID=1035287"> is not used to generate </a>investment decisions but may be used as a tool in the broader analytical approach deployed.&#8221;</p></div><p>The detailed disclosures are concentrated at the top. The largest managers &#8212; those with more than $100 billion under management &#8212; were meaningfully more likely to describe specific AI use cases and governance frameworks than smaller firms. Building an AI governance framework requires lawyers, compliance staff, and engineers working in concert. Most smaller firms don&#8217;t have that infrastructure. Explicit references to AI driving investment decisions are less common, and usually more carefully qualified.</p><p>Where firms do describe AI in concrete terms, it&#8217;s mostly operational. Concentric Capital Strategies, with $3.1 billion in regulatory assets under management, <a href="https://files.adviserinfo.sec.gov/IAPD/Content/Common/crd_iapd_Brochure.aspx?BRCHR_VRSN_ID=1024799">discloses</a> the use of LLMs, such as ChatGPT, within its investment research and business processes.</p><p>Caveat: Form ADV filings capture what firms consider material enough to disclose. Routine or limited AI use may not appear at all. </p><div class="callout-block" data-callout="true"><h2><strong>Data &amp; Methodology</strong></h2><p>Each ADV was converted from PDF to plain text and scanned for AI-related keywords (artificial intelligence, machine learning, large language model, generative AI, and close variants). Filings with at least one hit were passed to Gemma 4 for structured extraction. Gemma read the most AI-relevant passage from each filing &#8212; typically drawn from the section with the highest density of AI-related language. It returned a set of flags covering own-use vs. portfolio theme, investment vs. operational use, tool names, governance policy, human oversight language, and cost disclosures.</p></div><div><hr></div><h3><strong>A Closer Look at the Largest Managers</strong></h3><p>I selected 100 of the world&#8217;s largest and most recognizable money managers &#8212; firms spanning traditional asset management, hedge funds, private equity, and venture capital &#8212; and scanned their 2026 ADV filings with Gemma. </p><p>Seventy-five of the 100 disclosed some form of AI use. Twenty-four named a formal governance policy. Thirteen disclosed that AI-related costs may be passed to investors. </p><div class="callout-block" data-callout="true"><h3><strong>AI Street Data</strong> </h3><p>I&#8217;m still working out what to do with the underlying dataset &#8212; governance flags, named tools, and cost-charging disclosures across firms. If that would be useful to you, reply and let me know how you&#8217;d use it.</p></div><p>The following is for paid subscribers and examines the firms that named governance policies and what those policies actually say, the private equity pattern of charging AI infrastructure costs to fund investors, and three large managers that went from no AI disclosure in 2025 to named frameworks in 2026.</p><div><hr></div><h3><strong>The Governance Tier</strong></h3>
      <p>
          <a href="https://www.ai-street.co/p/what-16000-sec-filings-say-about">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[Ex-BlackRock Exec Ang Details 50-Agent Investment Process]]></title><description><![CDATA[Ang and Altbridge researchers lay out an architecture for autonomous portfolio management]]></description><link>https://www.ai-street.co/p/ex-blackrock-exec-ang-details-50</link><guid isPermaLink="false">https://www.ai-street.co/p/ex-blackrock-exec-ang-details-50</guid><dc:creator><![CDATA[Matt Robinson]]></dc:creator><pubDate>Wed, 08 Apr 2026 15:31:11 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/1ea40c45-c2cf-48ab-bbf8-065e1318cff8_1884x1094.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Financial firms have been moving away from the one-size-fits-all approach when it comes to AI. </p><p>Companies are using smaller, specialized systems because they&#8217;re easier to test, control, and generally, are more consistent. </p><p>Some examples: </p><ul><li><p>Capital One <a href="https://static.rainfocus.com/nvidia/gtc26/sess/1769189647938001MN4L/FinalPresPDF/EX82362_1773262728080001E7iG.pdf">replaced</a> a single LLM with a multi-agent system for call summaries, where agents interpret, reason, cross-check, and document the interaction before finalizing it.</p></li><li><p><a href="https://daloopa.com/?utm_source=www.ai-street.co&amp;utm_medium=newsletter&amp;utm_campaign=the-rise-of-ai-market-models&amp;_bhlid=058e7cc3b93a6457a395b9cb5f03dcf3e1f2d537">Daloopa</a> built dozens of narrow models, each trained on structured financial data and focused on one task. &#8220;Our models have an IQ of 250 on one task and 2 on something else,&#8221; <a href="https://www.linkedin.com/in/thomas-li-a6189245/?utm_source=www.ai-street.co&amp;utm_medium=newsletter&amp;utm_campaign=the-rise-of-ai-market-models&amp;_bhlid=c02b2dd5c6da4e3c42c28a74ec8b9746466bf949">Thomas Li</a>, CEO and co-founder of Daloopa, <a href="https://www.ai-street.co/p/the-rise-of-ai-market-models">told me</a> in September. </p></li><li><p>At BlackRock, researchers <a href="https://www.ai-street.co/i/183582065/blackrock-researchers-develop-ai-agent-system-for-stock-picks-study">broke stock screening</a> into specialized agents&#8212;fundamentals, sentiment, and valuation&#8212;that debate and cross-check each other before reaching a final decision.</p></li></ul><p>A new paper from <a href="https://www.linkedin.com/in/andrew-ang-a9a65a89/">Andrew Ang</a>, a former BlackRock executive, <a href="https://www.linkedin.com/in/nazym/">Nazym Azimbayev</a>, a sovereign wealth fund CIO, and <a href="https://www.linkedin.com/in/kimandrik/">Andrey Kim</a> PhD, a Deutsche Bank quant, takes the BlackRock debate architecture further.</p><p>The paper, <a href="https://arxiv.org/pdf/2604.02279">the Self-Driving Portfolio: Agentic Architecture for Institutional Asset Management</a>, asks if autonomous driving is here, why not autonomous investing? </p><p>Their answer is a 50-agent pipeline that runs the process and documents each step of its reasoning.</p><div><hr></div><h3><strong>Manage Email Preferences</strong></h3><p>If you prefer to receive one weekly email with all AI Street content, deselect Research and Interviews.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.ai-street.co/account&quot;,&quot;text&quot;:&quot;Manage How Often You Receive AI Street&quot;,&quot;action&quot;:null,&quot;class&quot;:&quot;button-wrapper&quot;}" data-component-name="ButtonCreateButton"><a class="button primary button-wrapper" href="https://www.ai-street.co/account"><span>Manage How Often You Receive AI Street</span></a></p><div><hr></div><p>They view running a strategic asset allocation process at an institutional level as a bandwidth problem as much as an analytical one. A CIO can supervise maybe 10 to 15 investment departments. A research team can realistically cover 20 to 30 asset classes before the process bottlenecks and an investment committee meets quarterly.</p><p>To speed this process up, they built a 50 agent pipeline that produced a documented strategic asset allocation with capital market assumptions, portfolio construction, peer review, and a board memo.</p><h3><strong>Here&#8217;s what they built </strong></h3><p>The pipeline is organized around the Investment Policy Statement (IPS), which governs the whole system the way it would govern human portfolio managers. Every agent reads it; the chief risk officer agent checks compliance for every portfolio candidate; the final output must satisfy it. </p><p>To illustrate how the pipeline works in practice, the authors ran it in March 2026 against the following mandate: 18 liquid asset classes (6 equity, 8 fixed income, 4 alternatives), a target real return of CPI +3&#8211;4%, a volatility band of 8&#8211;12%, a maximum drawdown of &#8722;25%, and a tracking error ceiling of 6% relative to a 60/40 benchmark.</p><ul><li><p><strong>Macro agent</strong>: Classifies the current economic regime &#8212; expansion, late-cycle, recession, or recovery &#8212; using macro data, market indicators, and web searches for real-time readings. Output flows downstream to every other agent.</p></li><li><p><strong>Asset class agents: </strong>Agents run in parallel, one per asset class. For equity classes, each estimates expected returns using six different methods, then blends them into a seventh composite. An LLM-as-judge step reads all seven alongside the current macro regime and valuations, and selects a final estimate with explicit weights and a written rationale.</p></li><li><p><strong>Portfolio construction agents</strong>: 20 agents each build a portfolio using a different method, ranging from simple rules of thumb to more sophisticated approaches. A 21st researcher agent scans the academic literature and proposes methods not yet in the pipeline. A separate adversarial diversifier, one of the original 20, deliberately constructs the portfolio most different from the consensus of all the others.</p></li><li><p><strong>Strategy review:</strong> Each agent reviews two others &#8212; one using a similar approach, one using a different one &#8212; and all reviews are released simultaneously. Agents then vote, ranking their top five and flagging a bottom pick. Votes are combined with a performance score, and the final shortlist must include methods from at least three of the four broad categories.</p></li><li><p><strong>CIO agent: </strong>Combines the top candidates using seven different aggregation methods and selects the one best suited to the current environment. Produces a board memo written for non-technical stakeholders.</p></li><li><p><strong>Meta-agent: </strong>After each rebalancing cycle, compares past forecasts against realized returns, identifies systematic weaknesses, and updates both the code and instructions governing the other agents. All changes are logged.</p><div><hr></div></li></ul><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.ai-street.co/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">AI Street is reader supported. Access 30+ expert interviews, 18+ months of reporting, and Subscriber Chat with a paid subscription.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><div><hr></div><h3><strong>Here&#8217;s what they found </strong></h3><ul><li><p>The macro agent classified the current environment as late-cycle with stagflationary risk.</p></li><li><p>When each asset class agent settled on its return forecast, the pattern was consistent: the more expensive the market, the more the agent discounted historical estimates. US Growth stocks had their forecast cut 2.0 percentage points below the composite; US Large Cap was cut 1.1 points; Emerging Markets were barely adjusted. The agents weren&#8217;t pessimistic across the board &#8212; they were specifically skeptical of backward-looking estimates for the assets where current prices already implied low future returns.</p></li><li><p>The same reasoning surfaced in the portfolio construction vote. In a late-cycle environment where return forecasts are uncertain, the agents collectively favored methods that lean on historical volatility and correlation data rather than return predictions. Maximum Diversification &#8212; a method that spreads risk across assets without relying heavily on return forecasts &#8212; ranked first. The portfolio that was deliberately constructed to be as different as possible from all the others came last, which was expected: its value is in the final blending step, not as a standalone recommendation.</p></li><li><p>The final portfolio came out modestly underweight stocks (44.9% vs. 60% in a standard balanced portfolio), roughly in line on bonds (41.7%), with an 8.1% cash position. Over a backtest from 1996 to 2026, it produced nearly the same return profile as a 60/40 portfolio &#8212; but with a peak-to-trough loss of 25.6% versus 34.3%.</p></li></ul><p>To be sure, this is a proof of concept, not an investment strategy. One run producing a sensible-looking portfolio doesn't tell you much given the short-time horizon.</p><p>I asked the paper&#8217;s authors about the results and received email responses from Azimbayev, who is also CEO of <a href="https://www.altbridge.ai/">Altbridge</a>, which describes itself as an AI-native hedge fund. </p>
      <p>
          <a href="https://www.ai-street.co/p/ex-blackrock-exec-ang-details-50">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[JPMorgan Taught AI the Language of Markets]]></title><description><![CDATA[Researchers apply the architecture behind ChatGPT to create a model that simulates market behavior.]]></description><link>https://www.ai-street.co/p/jpmorgan-taught-ai-the-language-of</link><guid isPermaLink="false">https://www.ai-street.co/p/jpmorgan-taught-ai-the-language-of</guid><dc:creator><![CDATA[Matt Robinson]]></dc:creator><pubDate>Tue, 31 Mar 2026 15:31:45 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/6ef821e9-186b-4139-a1d8-7b9fafa98b34_2816x1536.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Much of the AI conversation is focused on the latest capabilities of Anthropic&#8217;s Claude or ChatGPT, which deserve our attention, but this is a narrow view of the power of the transformer breakthrough. </p><div><hr></div><h3><strong>Manage Email Preferences</strong></h3><p>If you prefer to receive one weekly email with all AI Street content, turn off Research and Interviews here: </p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.ai-street.co/account&quot;,&quot;text&quot;:&quot;Manage How Often You Receive AI Street&quot;,&quot;action&quot;:null,&quot;class&quot;:&quot;button-wrapper&quot;}" data-component-name="ButtonCreateButton"><a class="button primary button-wrapper" href="https://www.ai-street.co/account"><span>Manage How Often You Receive AI Street</span></a></p><div><hr></div><p>The transformer breakthrough began in text, but researchers are adapting the architecture to other kinds of sequential data. With enough data, transformer models can learn the patterns of &#8220;language&#8221; in that dataset in ways that traditional models missed. </p><p>For example, AlphaFold is a transformer-based system trained on protein data to predict how proteins fold into their 3D shapes from amino acid sequences, which determine how they function. It effectively solved the protein folding problem. (The work later contributed to a Nobel Prize in Chemistry awarded to its creators, including Demis Hassabis, who is <a href="https://www.nobelprize.org/prizes/chemistry/2024/hassabis/facts/">not a chemist</a>.)</p><p>As I&#8217;ve written before, no one knows <em>exactly</em> how these models work. They&#8217;re grown rather than built, as the CEO of Anthropic likes to <a href="https://www.darioamodei.com/essay/the-adolescence-of-technology">say</a>. We didn&#8217;t know how aspirin worked for like 70 years, but we knew it was effective. </p><p>This brings us to a new paper from JPMorgan researchers, who trained a transformer model on market data.</p><h2>The market as a language</h2><p>Every buy, sell, order submission, or cancellation leaves a trace: what happened, how much size was involved, how far from the market mid-price it was placed, and when it occurred. Multiply that across thousands of stocks and millions of events per day, and you get a massive stream of sequential data.</p><p>TradeFM is a 524-million-parameter model trained on 10.7 billion training tokens drawn from more than 9,000 U.S. equities, using data spanning 368 trading days from February 2024 to September 2025.</p><p>Instead of predicting the next word in a sentence, their model &#8212; called TradeFM &#8212; predicts the next event in a sequence: its timing, size, price depth, and direction.</p><p>Trading data is messy. Stocks trade at different prices. A $5 move on a $2 stock is massive. A $5 move on one that&#8217;s $500 isn&#8217;t news.</p><p>If you feed those raw numbers into a model, it can&#8217;t really compare one stock to another, so it struggles to learn general patterns.</p><p>So the researchers adjusted the data before training. They expressed price-related features in relative terms, compressed volumes so large and small trades are easier to compare, and measured time as the gap between events.</p><p>That puts different stocks on a common scale, so moves are comparable whether it&#8217;s a $2 stock or a $200 stock.</p><p>They then discretized each event&#8217;s features and combined timing, price depth, volume, side, and action type into a single composite token. The result was a vocabulary of 16,384 trade event tokens.</p><div><hr></div><h2><strong>Related Research</strong></h2><div class="digest-post-embed" data-attrs="{&quot;nodeId&quot;:&quot;77034d90-899a-4487-9e57-caede79c7bda&quot;,&quot;caption&quot;:&quot;Hey, it&#8217;s Matt. Welcome back to AI Street. This week:&quot;,&quot;cta&quot;:&quot;Read full story&quot;,&quot;showBylines&quot;:true,&quot;size&quot;:&quot;sm&quot;,&quot;isEditorNode&quot;:true,&quot;title&quot;:&quot;HRT Trains AI Models on Trading Data&quot;,&quot;publishedBylines&quot;:[{&quot;id&quot;:227819155,&quot;name&quot;:&quot;Matt Robinson&quot;,&quot;bio&quot;:&quot;I write AI Street &#8212; how Wall Street uses AI from trading floors to the C-suite. Former Bloomberg News reporter &quot;,&quot;photo_url&quot;:&quot;https://substackcdn.com/image/fetch/$s_!JhAn!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2b2b35a3-1ee4-4f02-8d99-c6019ea474eb_1181x1181.png&quot;,&quot;is_guest&quot;:false,&quot;bestseller_tier&quot;:null}],&quot;post_date&quot;:&quot;2026-01-15T16:30:37.344Z&quot;,&quot;cover_image&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/830b51cb-b61b-4a72-8e80-e9c20b92157f_2456x1378.png&quot;,&quot;cover_image_alt&quot;:null,&quot;canonical_url&quot;:&quot;https://www.ai-street.co/p/hrt-trains-ai-models-on-trading-data&quot;,&quot;section_name&quot;:null,&quot;video_upload_id&quot;:null,&quot;id&quot;:184024628,&quot;type&quot;:&quot;newsletter&quot;,&quot;reaction_count&quot;:12,&quot;comment_count&quot;:3,&quot;publication_id&quot;:4098119,&quot;publication_name&quot;:&quot;AI Street &quot;,&quot;publication_logo_url&quot;:&quot;https://substackcdn.com/image/fetch/$s_!ezC3!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8a4fc97a-4b2d-4478-92be-ea095be05d61_800x800.png&quot;,&quot;belowTheFold&quot;:true,&quot;youtube_url&quot;:null,&quot;show_links&quot;:null,&quot;feed_url&quot;:null}"></div><div class="digest-post-embed" data-attrs="{&quot;nodeId&quot;:&quot;212077be-05a8-400c-b3be-7e59b4dfbd78&quot;,&quot;caption&quot;:&quot;RESEARCH&quot;,&quot;cta&quot;:&quot;Read full story&quot;,&quot;showBylines&quot;:true,&quot;size&quot;:&quot;sm&quot;,&quot;isEditorNode&quot;:true,&quot;title&quot;:&quot;Treating Trading Data As \&quot;Language\&quot; &quot;,&quot;publishedBylines&quot;:[{&quot;id&quot;:227819155,&quot;name&quot;:&quot;Matt Robinson&quot;,&quot;bio&quot;:&quot;I write AI Street &#8212; how Wall Street uses AI from trading floors to the C-suite. Former Bloomberg News reporter &quot;,&quot;photo_url&quot;:&quot;https://substackcdn.com/image/fetch/$s_!JhAn!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2b2b35a3-1ee4-4f02-8d99-c6019ea474eb_1181x1181.png&quot;,&quot;is_guest&quot;:false,&quot;bestseller_tier&quot;:null}],&quot;post_date&quot;:&quot;2026-01-02T14:06:00.000Z&quot;,&quot;cover_image&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f2de16a0-c5fd-4e8b-aa1a-55900366048c_1280x720.png&quot;,&quot;cover_image_alt&quot;:null,&quot;canonical_url&quot;:&quot;https://www.ai-street.co/p/treating-trading-data-as-language&quot;,&quot;section_name&quot;:null,&quot;video_upload_id&quot;:null,&quot;id&quot;:183581943,&quot;type&quot;:&quot;newsletter&quot;,&quot;reaction_count&quot;:5,&quot;comment_count&quot;:0,&quot;publication_id&quot;:4098119,&quot;publication_name&quot;:&quot;AI Street &quot;,&quot;publication_logo_url&quot;:&quot;https://substackcdn.com/image/fetch/$s_!ezC3!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8a4fc97a-4b2d-4478-92be-ea095be05d61_800x800.png&quot;,&quot;belowTheFold&quot;:true,&quot;youtube_url&quot;:null,&quot;show_links&quot;:null,&quot;feed_url&quot;:null}"></div><div><hr></div><h2><strong>What they found</strong></h2><p>The researchers tested the model inside a simulated exchange, where it predicts trades in a continuous loop. The resulting data reproduces core patterns seen in real markets, including clustered volatility and large price swings. Across 9 stocks, 3 liquidity tiers, and 9 months of held-out data, it matched those patterns 2 to 3 times more closely than a standard baseline known as a Compound Hawkes process.</p><p>What&#8217;s most interesting is that <a href="https://arxiv.org/html/2602.23784v1">TradeFM</a>&#8217;s behavior extends beyond the U.S. data it was trained on. JPMorgan tested the model, without any adjustments, on trading data from China and Japan, where market structure differs meaningfully. Japan uses batch auctions at the open. China imposes 10% daily price limits. Spreads in both markets are several times wider than in the U.S. Despite those differences, the model&#8217;s performance degraded only moderately. It had never seen these markets, yet it still captured their core dynamics.</p><p>The model appears to be learning structure that carries across markets.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.ai-street.co/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.ai-street.co/subscribe?"><span>Subscribe now</span></a></p><p><a href="https://www.linkedin.com/in/armankhaledian/?utm_source=www.ai-street.co&amp;utm_medium=newsletter&amp;utm_campaign=ai-startup-filters-out-the-noise-in-financial-news&amp;_bhlid=821bb636d94ce5dd3099b83433064009ba97b0ab">Arman Khaledian</a>, PhD, a former quant at Millennium and now CEO of <a href="https://zanista.ai/">Zanista AI</a>, said: &#8220;That&#8217;s not a toy result. It means the model is picking up something real about how markets work at a structural level.&#8221;</p>
      <p>
          <a href="https://www.ai-street.co/p/jpmorgan-taught-ai-the-language-of">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[Tracking Shifts in Earnings Call Narratives]]></title><description><![CDATA[In earnings calls, LLMs are better at spotting changes in the metrics companies highlight.]]></description><link>https://www.ai-street.co/p/tracking-shifts-in-earnings-call</link><guid isPermaLink="false">https://www.ai-street.co/p/tracking-shifts-in-earnings-call</guid><dc:creator><![CDATA[Matt Robinson]]></dc:creator><pubDate>Tue, 24 Mar 2026 15:30:52 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/93a8630f-e724-42ce-90e2-234b4865e62e_2584x1476.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Regulatory rules dictate how companies report performance, but there are virtually no rules governing what management chooses to talk about on an earnings call. </p><p>The C-suite can choose what numbers to highlight &#8212; revenue per user, lifetime value, total addressable market, etc. </p><p>These selected metrics support the corporate narrative, but they&#8217;re not static. When a number is strong, it gets airtime. When it softens, it tends to quietly disappear from the script, replaced by whatever metric tells a better story that quarter.</p><p>Shifting corporate narratives happens so often it has a name: &#8220;moving targets.&#8221; This tracks the fraction of previously highlighted metrics that go missing from the next comparable earnings call. Research showed that firms with high metric turnover tend to underperform in subsequent months. The more a company reshuffles the numbers it talks about, the worse its stock tends to do.</p><p>The challenge is detection. Most approaches rely on keyword matching across transcripts, comparing terms quarter over quarter. You need to be able to link &#8220;revenue growth&#8221; to &#8220;top-line expansion.&#8221; Keyword search can&#8217;t distinguish &#8220;North America cloud revenue&#8221; from &#8220;revenue.&#8221;</p><p>At scale, this becomes difficult to track. Following the metrics that appear and disappear across thousands of earnings calls is not feasible to do consistently by hand. This is where LLMs fit: extracting and standardizing how companies describe performance over time. This is tedious work for humans and easy to scale with AI.</p><p>A group of researchers at MIT, BlackRock, and J.P. Morgan asked whether LLMs could close this detection gap.</p><p><strong>Here&#8217;s what they did:</strong></p><ul><li><p>Instead of scanning transcripts for predefined terms, they use an LLM to extract full phrases with context. Where keyword methods pull &#8220;revenue,&#8221; the model pulls &#8220;North America cloud revenue.&#8221; Where it grabs &#8220;dividends,&#8221; the model also captures &#8220;cash flow,&#8221; &#8220;share repurchases,&#8221; and &#8220;cash flow from operations.&#8221;</p></li><li><p>They then compare metrics across quarters using semantic similarity rather than exact matches. Instead of forcing a binary match, they allow for an &#8220;ambiguous&#8221; range where similarity is scaled.</p></li><li><p>They apply this across firms listed in the S&amp;P 100 index from January 2010 to December 2024, yielding 5,615 firm-quarter observations across 64 quarters.</p></li><li><p>To test it, they sort stocks by how much their metrics shift and compare returns, then run cross-sectional regressions with standard controls for size, valuation, and prior returns.</p></li></ul><p><strong>Here&#8217;s what they found:</strong></p>
      <p>
          <a href="https://www.ai-street.co/p/tracking-shifts-in-earnings-call">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[AI Still Falls Short in Excel: Study]]></title><description><![CDATA[Even the best model wouldn't make it on Wall Street]]></description><link>https://www.ai-street.co/p/ai-still-falls-short-in-excel-study</link><guid isPermaLink="false">https://www.ai-street.co/p/ai-still-falls-short-in-excel-study</guid><dc:creator><![CDATA[Matt Robinson]]></dc:creator><pubDate>Wed, 18 Mar 2026 15:30:10 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/1904cc40-4d75-4d8c-b38f-8c7c2880e786_2816x1536.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>AI for Excel has improved dramatically over the last six months, but it still wouldn&#8217;t make it as a junior analyst on Wall Street. </p><p>The three leading AI providers, OpenAI, Anthropic and Google&#8217;s Gemini, have all released updated capabilities for Excel and are reporting higher accuracy on benchmarks.  </p><p>For example, OpenAI <a href="https://openai.com/index/chatgpt-for-excel/">said</a> this month that performance on its internal investment banking benchmark jumped to 87.3% (GPT-5.4 Thinking) from 43.7% (GPT-5).  </p><p>While these are <em>material</em> improvements, AI still can&#8217;t be relied on without supervision. &#8220;Mostly right&#8221; is not good enough. </p><p>A recent benchmark on AI in Excel points to the same issue: performance drops sharply as tasks get more complex. </p><p><a href="https://arxiv.org/pdf/2603.07316">FinSheet-Bench</a> tests how models handle real-world private equity workbooks with messy layouts, multiple funds, and non-standard formatting. Across 10 models from OpenAI, Google, and Anthropic, the best result came from Gemini 3.1 Pro at 82.4% accuracy, followed closely by GPT-5.2 with reasoning and Claude Opus 4.6 with thinking, both around 80%.</p><p>On simple lookups, top models exceed 90% accuracy.</p><p>The gap widens further on large, realistic files. On the most complex workbook tested, with 152 companies across eight funds, average accuracy was worse than a coin flip. </p><p>One reason: models don&#8217;t actually &#8220;see&#8221; the spreadsheet. They operate on a text-serialized version that strips out layout, formatting, and visual structure.</p><p>So this: </p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!ZLyT!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70430b79-2093-4a06-98ec-f82a359a66fe_1588x282.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!ZLyT!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70430b79-2093-4a06-98ec-f82a359a66fe_1588x282.png 424w, https://substackcdn.com/image/fetch/$s_!ZLyT!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70430b79-2093-4a06-98ec-f82a359a66fe_1588x282.png 848w, https://substackcdn.com/image/fetch/$s_!ZLyT!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70430b79-2093-4a06-98ec-f82a359a66fe_1588x282.png 1272w, https://substackcdn.com/image/fetch/$s_!ZLyT!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70430b79-2093-4a06-98ec-f82a359a66fe_1588x282.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!ZLyT!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70430b79-2093-4a06-98ec-f82a359a66fe_1588x282.png" width="1456" height="259" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/70430b79-2093-4a06-98ec-f82a359a66fe_1588x282.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:259,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:28265,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.ai-street.co/i/191054535?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70430b79-2093-4a06-98ec-f82a359a66fe_1588x282.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!ZLyT!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70430b79-2093-4a06-98ec-f82a359a66fe_1588x282.png 424w, https://substackcdn.com/image/fetch/$s_!ZLyT!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70430b79-2093-4a06-98ec-f82a359a66fe_1588x282.png 848w, https://substackcdn.com/image/fetch/$s_!ZLyT!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70430b79-2093-4a06-98ec-f82a359a66fe_1588x282.png 1272w, https://substackcdn.com/image/fetch/$s_!ZLyT!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70430b79-2093-4a06-98ec-f82a359a66fe_1588x282.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p>Becomes this </p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!xdTj!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1147bf08-47b5-4356-83d7-cc9a8269d66e_1444x174.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!xdTj!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1147bf08-47b5-4356-83d7-cc9a8269d66e_1444x174.png 424w, https://substackcdn.com/image/fetch/$s_!xdTj!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1147bf08-47b5-4356-83d7-cc9a8269d66e_1444x174.png 848w, https://substackcdn.com/image/fetch/$s_!xdTj!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1147bf08-47b5-4356-83d7-cc9a8269d66e_1444x174.png 1272w, https://substackcdn.com/image/fetch/$s_!xdTj!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1147bf08-47b5-4356-83d7-cc9a8269d66e_1444x174.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!xdTj!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1147bf08-47b5-4356-83d7-cc9a8269d66e_1444x174.png" width="1444" height="174" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/1147bf08-47b5-4356-83d7-cc9a8269d66e_1444x174.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:174,&quot;width&quot;:1444,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:29273,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.ai-street.co/i/191054535?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1147bf08-47b5-4356-83d7-cc9a8269d66e_1444x174.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!xdTj!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1147bf08-47b5-4356-83d7-cc9a8269d66e_1444x174.png 424w, https://substackcdn.com/image/fetch/$s_!xdTj!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1147bf08-47b5-4356-83d7-cc9a8269d66e_1444x174.png 848w, https://substackcdn.com/image/fetch/$s_!xdTj!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1147bf08-47b5-4356-83d7-cc9a8269d66e_1444x174.png 1272w, https://substackcdn.com/image/fetch/$s_!xdTj!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1147bf08-47b5-4356-83d7-cc9a8269d66e_1444x174.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p>Spokespeople for OpenAI, Anthropic and Google didn&#8217;t respond to requests for comment on FinSheet-Bench&#8217;s results. </p><p>What I&#8217;d like to see from model providers, rather than the latest numbers from internal benchmarks, is basic operating data: real error rates, how often outputs need to be corrected, and what level of reliability users should expect.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.ai-street.co/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.ai-street.co/subscribe?"><span>Subscribe now</span></a></p><div><hr></div><h2><strong>ICYMI Research</strong></h2><div class="digest-post-embed" data-attrs="{&quot;nodeId&quot;:&quot;ba5156c8-82c9-4f90-a93b-52cd3a45665a&quot;,&quot;caption&quot;:&quot;AI looks impressive when you ask a narrow question about a single company filing, such as revenue last quarter.&quot;,&quot;cta&quot;:&quot;Read full story&quot;,&quot;showBylines&quot;:true,&quot;size&quot;:&quot;sm&quot;,&quot;isEditorNode&quot;:true,&quot;title&quot;:&quot;Why AI Struggles With Real Analyst Work &quot;,&quot;publishedBylines&quot;:[{&quot;id&quot;:227819155,&quot;name&quot;:&quot;Matt Robinson&quot;,&quot;bio&quot;:&quot;I write AI Street &#8212; how Wall Street uses AI from trading floors to the C-suite. Former Bloomberg News reporter &quot;,&quot;photo_url&quot;:&quot;https://substackcdn.com/image/fetch/$s_!JhAn!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2b2b35a3-1ee4-4f02-8d99-c6019ea474eb_1181x1181.png&quot;,&quot;is_guest&quot;:false,&quot;bestseller_tier&quot;:null}],&quot;post_date&quot;:&quot;2026-02-25T12:03:54.382Z&quot;,&quot;cover_image&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/959d5a45-b276-47ad-aeef-accd61e7d236_1786x1076.png&quot;,&quot;cover_image_alt&quot;:null,&quot;canonical_url&quot;:&quot;https://www.ai-street.co/p/why-ai-struggles-with-real-analyst&quot;,&quot;section_name&quot;:&quot;Research &quot;,&quot;video_upload_id&quot;:null,&quot;id&quot;:188912344,&quot;type&quot;:&quot;newsletter&quot;,&quot;reaction_count&quot;:9,&quot;comment_count&quot;:0,&quot;publication_id&quot;:4098119,&quot;publication_name&quot;:&quot;AI Street &quot;,&quot;publication_logo_url&quot;:&quot;https://substackcdn.com/image/fetch/$s_!ezC3!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8a4fc97a-4b2d-4478-92be-ea095be05d61_800x800.png&quot;,&quot;belowTheFold&quot;:true,&quot;youtube_url&quot;:null,&quot;show_links&quot;:null,&quot;feed_url&quot;:null}"></div><div class="digest-post-embed" data-attrs="{&quot;nodeId&quot;:&quot;f8f08081-b32f-4e3f-b1b1-124526ed7d44&quot;,&quot;caption&quot;:&quot;&#8220;The price of intelligence is going to zero.&#8221;&quot;,&quot;cta&quot;:&quot;Read full story&quot;,&quot;showBylines&quot;:true,&quot;size&quot;:&quot;sm&quot;,&quot;isEditorNode&quot;:true,&quot;title&quot;:&quot;AI Boosts Retail Trading Volume: Research&quot;,&quot;publishedBylines&quot;:[{&quot;id&quot;:227819155,&quot;name&quot;:&quot;Matt Robinson&quot;,&quot;bio&quot;:&quot;I write AI Street &#8212; how Wall Street uses AI from trading floors to the C-suite. Former Bloomberg News reporter &quot;,&quot;photo_url&quot;:&quot;https://substackcdn.com/image/fetch/$s_!JhAn!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2b2b35a3-1ee4-4f02-8d99-c6019ea474eb_1181x1181.png&quot;,&quot;is_guest&quot;:false,&quot;bestseller_tier&quot;:null}],&quot;post_date&quot;:&quot;2026-02-12T12:02:23.938Z&quot;,&quot;cover_image&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/9f75fbf1-7f11-4f4e-9aa5-69c4d6ba896c_1794x1686.png&quot;,&quot;cover_image_alt&quot;:null,&quot;canonical_url&quot;:&quot;https://www.ai-street.co/p/ai-boosts-retail-trading-volume-research&quot;,&quot;section_name&quot;:&quot;Research &quot;,&quot;video_upload_id&quot;:null,&quot;id&quot;:187608461,&quot;type&quot;:&quot;newsletter&quot;,&quot;reaction_count&quot;:3,&quot;comment_count&quot;:0,&quot;publication_id&quot;:4098119,&quot;publication_name&quot;:&quot;AI Street &quot;,&quot;publication_logo_url&quot;:&quot;https://substackcdn.com/image/fetch/$s_!ezC3!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8a4fc97a-4b2d-4478-92be-ea095be05d61_800x800.png&quot;,&quot;belowTheFold&quot;:true,&quot;youtube_url&quot;:null,&quot;show_links&quot;:null,&quot;feed_url&quot;:null}"></div><div class="digest-post-embed" data-attrs="{&quot;nodeId&quot;:&quot;dd01ac51-eb74-41fd-a866-03e6746d3c70&quot;,&quot;caption&quot;:&quot;Hey, it&#8217;s Matt. This week on AI Street:&quot;,&quot;cta&quot;:&quot;Read full story&quot;,&quot;showBylines&quot;:true,&quot;size&quot;:&quot;sm&quot;,&quot;isEditorNode&quot;:true,&quot;title&quot;:&quot;BlackRock Study Tests AI Agents for Stock Picks &quot;,&quot;publishedBylines&quot;:[{&quot;id&quot;:227819155,&quot;name&quot;:&quot;Matt Robinson&quot;,&quot;bio&quot;:&quot;I write AI Street &#8212; how Wall Street uses AI from trading floors to the C-suite. Former Bloomberg News reporter &quot;,&quot;photo_url&quot;:&quot;https://substackcdn.com/image/fetch/$s_!JhAn!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2b2b35a3-1ee4-4f02-8d99-c6019ea474eb_1181x1181.png&quot;,&quot;is_guest&quot;:false,&quot;bestseller_tier&quot;:null}],&quot;post_date&quot;:&quot;2025-08-21T15:30:00.000Z&quot;,&quot;cover_image&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/646fa0b1-ec1c-485d-aec9-8b5085615c69_1200x630.png&quot;,&quot;cover_image_alt&quot;:null,&quot;canonical_url&quot;:&quot;https://www.ai-street.co/p/blackrock-tests-multi-agent-ai-for-stock-picks&quot;,&quot;section_name&quot;:null,&quot;video_upload_id&quot;:null,&quot;id&quot;:183582065,&quot;type&quot;:&quot;newsletter&quot;,&quot;reaction_count&quot;:6,&quot;comment_count&quot;:0,&quot;publication_id&quot;:4098119,&quot;publication_name&quot;:&quot;AI Street &quot;,&quot;publication_logo_url&quot;:&quot;https://substackcdn.com/image/fetch/$s_!ezC3!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8a4fc97a-4b2d-4478-92be-ea095be05d61_800x800.png&quot;,&quot;belowTheFold&quot;:true,&quot;youtube_url&quot;:null,&quot;show_links&quot;:null,&quot;feed_url&quot;:null}"></div><div><hr></div><p>The problems of AI in Excel echo the broader challenges of AI, namely that if the model has to traverse thousands of pages or dozens of tabs in Excel, accuracy plummets. </p><p>Over the last few weeks here, we&#8217;ve <a href="https://www.ai-street.co/p/what-actually-makes-ai-reliable">covered</a> how there&#8217;s more of a premium on reliable models rather than the most powerful. </p>
      <p>
          <a href="https://www.ai-street.co/p/ai-still-falls-short-in-excel-study">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[AI Replicates Human Investor Biases]]></title><description><![CDATA[A study of 48 models found framing and sunk-cost effects distort AI investment decisions.]]></description><link>https://www.ai-street.co/p/ai-replicates-human-investor-biases</link><guid isPermaLink="false">https://www.ai-street.co/p/ai-replicates-human-investor-biases</guid><dc:creator><![CDATA[Matt Robinson]]></dc:creator><pubDate>Wed, 11 Mar 2026 13:02:45 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/abacd065-4c26-4f56-b2f5-3ef472a43783_1408x768.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>AI hallucinations get a lot of attention. But another risk is bias. </p><p>The way you write a prompt can steer a model toward different answers, even when the underlying question is exactly the same. I <a href="https://www.ai-street.co/i/183581991/how-simple-word-choices-lead-ai-astray">wrote</a> in November: </p><blockquote><p>If you asked a (human) financial analyst whether Microsoft or Apple is the better investment, the answer wouldn&#8217;t depend on whether you said <em>Microsoft or Apple</em> or <em>Apple or Microsoft.</em> For LLMs, that word order matters, according to new research. </p></blockquote><p>This risk is harder to detect since an answer isn&#8217;t necessarily <em>wrong</em>, the model just chooses to highlight a different point. </p><p>Individual investors and, I suspect, some institutional ones as well, are likely falling for this risk by asking AI for research ideas and stock-picking guidance. </p><p><a href="https://www.etoro.com/news-and-analysis/etoro-updates/retail-investors-flock-to-ai-tools-with-usage-up-46-in-one-year/">More and more retail investors</a> are relying on AI tools, and almost three quarters of millennials do so, according to an October eToro survey. </p><p>To investigate how widespread this issue is, researchers at Auburn University and the University of Tulsa evaluated 48 large language models across investment-style decision tasks.</p><p>They presented identical financial scenarios twice, changing only how the information was framed, such as wording risk as a gain versus a loss, adding a prestigious source, or mentioning prior spending. Many of the same biases have long been documented in human investors. The difference is that AI systems can reproduce them consistently and at scale.</p><h3>What they did</h3><ul><li><p>Showed each model the same scenario twice: once neutral, once with a subtle wording or context change</p></li><li><p>Tested 11 well-known investor errors, including framing, anchoring, herding, narrative appeal, and sunk costs</p></li><li><p>Ran 25 scenario pairs per error across all 48 models</p></li><li><p>Evaluated mitigation methods such as debiasing instructions and prompt rewriting</p></li></ul><h3>Results</h3><ul><li><p><strong>Framing alone moved ratings by 1.62 points on a 10-point scale</strong>, enough to flip decisions around common thresholds</p></li><li><p><strong>Narrative cues dominated fundamentals:</strong> describing founders as fitting a familiar archetype raised ratings by 65% or when attributing the analysis to a Nobel Laureate.  </p></li></ul>
      <p>
          <a href="https://www.ai-street.co/p/ai-replicates-human-investor-biases">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[Why AI Struggles With Real Analyst Work ]]></title><description><![CDATA[Due diligence requires evidence across firms and time, where current systems fail, according to new research.]]></description><link>https://www.ai-street.co/p/why-ai-struggles-with-real-analyst</link><guid isPermaLink="false">https://www.ai-street.co/p/why-ai-struggles-with-real-analyst</guid><dc:creator><![CDATA[Matt Robinson]]></dc:creator><pubDate>Wed, 25 Feb 2026 12:03:54 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/959d5a45-b276-47ad-aeef-accd61e7d236_1786x1076.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>AI looks impressive when you ask a narrow question about a single company filing, such as revenue last quarter.</p><p>Ask it to compare two firms&#8217; risk disclosures or track strategy over several years, and performance drops fast. Fin-RATE, a <a href="https://arxiv.org/abs/2602.07294">new benchmark</a> from researchers at Yale and Goldman Sachs, measures that gap and identifies where the model breaks.</p><p>Most financial benchmarks reduce SEC filings to lookup tasks: find a number in a 10-K and repeat it back accurately. That design misses how analysts actually work. Real due diligence requires synthesizing disclosures across companies, time periods, and filing types simultaneously. A pass/fail system doesn&#8217;t tell you whether errors came from retrieval, hallucination, or broken reasoning chains.</p><p><strong>Here&#8217;s what they did:</strong></p><ul><li><p>Built a body of 15,311 document segments from 2,472 SEC filings (10-K, 10-Q, 8-K, DEF 14A, and others) covering 43 companies across 36 industries, 2020&#8211;2025. Sourced from EDGAR, segmented at official SEC item boundaries, converted to structured Markdown.</p></li><li><p>Designed three task types</p><ul><li><p>Single-document questions </p></li><li><p>Cross-company comparisons </p></li><li><p>Multi-year analysis within one firm</p></li></ul></li><li><p>Created 7,500 question-answer pairs with numbers manually verified against source filings. </p></li><li><p>Evaluated 17 models, including closed-source systems, major open-source models, and finance-tuned variants</p></li><li><p>Tested performance with passages provided directly versus retrieved using four RAG methods. </p></li></ul><p><strong>The findings:</strong></p>
      <p>
          <a href="https://www.ai-street.co/p/why-ai-struggles-with-real-analyst">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[Can We Break Open AI’s Black Box?]]></title><description><![CDATA[Researchers are developing new frameworks to break open AI&#8217;s "black box," building interpretability and causal reasoning into models from the start.]]></description><link>https://www.ai-street.co/p/can-we-break-open-ais-black-box</link><guid isPermaLink="false">https://www.ai-street.co/p/can-we-break-open-ais-black-box</guid><dc:creator><![CDATA[Matt Robinson]]></dc:creator><pubDate>Tue, 24 Feb 2026 11:10:23 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/a377d549-74e8-464f-ae78-380508a80338_1744x1672.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>This is an excerpt from an article I wrote that was originally <a href="https://www.chicagobooth.edu/review/can-we-break-open-ais-black-box">published</a> in the Chicago Booth Review, a publication of the University of Chicago Booth School of Business.</em></p><div><hr></div><p>Demis Hassabis is not a chemist, yet he was one of three recipients of the 2024 Nobel Prize in Chemistry. The prize recognized major contributions to the study of protein structures. Hassabis, a computer scientist who runs Google&#8217;s AI research lab DeepMind, and his fellow honoree John Jumper, who also works at DeepMind, developed an AI prediction model that the chair of the Nobel committee said fulfilled &#8220;a 50-year-old dream: predicting protein structures from their amino acid sequences.&#8221; Another committee member called it &#8220;one of the really first big scientific breakthroughs of AI.&#8221;</p><p>For decades, uncovering the shape of a single protein meant spending months, even years, of painstaking lab work and hundreds of thousands of dollars toward research and development with no guarantee of success.</p><p>With DeepMind&#8217;s deep-learning model AlphaFold2, revealing these structures takes minutes, not months. The DeepMind team trained AlphaFold2 with data from lab-determined protein shapes, along with extra examples it created on its own from patterns found in huge protein-sequence databases. The model examined protein shapes and amino acid sequences to determine the physical and evolutionary constraints dictating protein structure.</p><p>The team has since predicted more than 200 million protein structures and made them freely available, creating a global resource for scientific research.</p><p>AlphaFold2 is one in a growing list of scientific breakthroughs driven by AI. It also represents a new paradigm in scientific discovery: AI models that achieve breakthroughs in ways their creators can&#8217;t fully explain. While traditional science builds understanding through hypotheses we can test and verify, these AI systems are discovering solutions by finding patterns in data that remain opaque to human analysis.</p><p>There is currently no easy way to examine what AlphaFold2 learned about protein evolution. Its inner workings, and those of other AI systems making important contributions to science and society, remain hidden.</p><p>As these models get better, the gap between their performance and our understanding of them is only widening.</p><p>Nonetheless, AI adoption is racing ahead. Modern AI works incredibly well. The latest models can perform tasks that, 10 years ago, sounded like science fiction: generating movie-quality videos from a few lines of text, writing entire codebases for working apps, even driving cars without human input.</p><p>These advances have quickly entered our personal and professional lives. But this rapid deployment of black-box systems creates a fundamental tension in our relationship with AI: We&#8217;re becoming dependent on tools that have reasoning we can&#8217;t verify or build upon.</p><div class="pullquote"><p>&#8220;You can go as crazy as you want and build the biggest, deepest neural network and still have interpretability baked in from the beginning.&#8221;<br>&#8212; Bryon Aragam</p></div><p>Even the architects of modern AI admit to being troubled by their lack of insight.</p><p>Dario Amodei, a cofounder of the AI lab Anthropic and the company&#8217;s CEO, wrote in April 2025: &#8220;People outside the field are often surprised and alarmed to learn that we do not understand how our own AI creations work. They are right to be concerned: this lack of understanding is essentially unprecedented in the history of technology.&#8221;</p><p>This has made interpretability, the science of cracking open AI&#8217;s &#8220;mind,&#8221; a pressing priority, and a new wave of research is taking a novel approach. AI-interpretability research has long been a form of detective work done <em>after </em>an AI system has already been trained and deployed. By then, the AI has already &#8220;decided&#8221; which data matter most when making its predictions.</p><p>Instead of trying to work backward to understand AI models after they&#8217;re built, scientists are now using new research frameworks to build interpretability into the training process from the start&#8212;a notion considered impossible just a few years ago.</p><h2><strong>Reverse engineering AI</strong></h2><p>When researchers try to parse the reasoning of an AI model after it has already been fully developed, they are essentially trying to reverse engineer a system that, in many ways, built itself, and attempting to uncover the internal patterns and definitions it formed along the way.</p><p>&#8220;People think these things are built systems, but they&#8217;re really not built per se,&#8221; says Ted Sumers, a researcher at Anthropic. &#8220;It&#8217;s much more like growing a plant than building a building.&#8221;</p><p>Understanding how a model &#8220;grows&#8221; has become a central focus for researchers.</p><p>One branch of this work, called mechanistic interpretability, maps which neurons activate when a user asks AI a question, and traces how information flows through the network&#8217;s intricate layers.</p><p>Anthropic, a rival to OpenAI, has been at the vanguard of this type of approach, dissecting neural networks by studying the roles of individual neurons and circuits.</p><p>This has yielded practical results. Teams can, without damaging overall performance, identify and remove specific circuits that lead to biased or unwanted outputs. They can also locate the exact parts of a model that enforce safety rules&#8212;like refusal to answer harmful queries&#8212;and adjust those directly. Since the techniques go down to the neuron level, they offer a way to audit whether a model is memorizing sensitive data. Together, these advances make models easier to edit, test, and trust as they continue to grow more capable.</p><p>Still, it&#8217;s like peering into a house through a keyhole.</p><h3><strong>The struggle to understand AI</strong></h3><p>Unlike traditional software, which relies on top-down, hard-coded rules, a neural network&#8212;a type of artificial-intelligence model that&#8217;s often described as resembling the structure of a human brain&#8212;learns from the bottom up, ingesting training data and making internal adjustments based on what it observes. Such models learn patterns from massive datasets, some with trillions of data points.</p><p>For example, to learn to identify pictures of dogs, a neural network reviews millions of labeled images of the animals rather than relying on a fixed set of definitions.</p><p>During training, the model guesses what each image shows and compares its answer to the correct label. If it guesses &#8220;not dog&#8221; for an image labeled &#8220;dog,&#8221; it recognizes the mistake and adjusts its internal settings to reduce the error. This process repeats again and again.</p><p>After enough examples, it becomes very good at identifying dogs. But it does so in a way that&#8217;s fundamentally different from how humans process and recall information. AI relies on statistical analysis to identify patterns, rather than mental imagery.</p><p>AI doesn&#8217;t &#8220;see&#8221; the way we do.</p><p>It sees the world through numerical representations of data. All types of data that AI works with&#8212;whether text, images, or audio&#8212;are converted into numbers that the system can mathematically manipulate. For example, the sentence, &#8220;AI sees the world through numerical representations of data&#8221; is converted into: [17527, 27432, 290, 2375, 1819, 57979, 63700, 328, 1238] according to OpenAI&#8217;s <a href="https://platform.openai.com/tokenizer">tool</a>, which displays how a piece of text might be tokenized by a language model. (Different models tokenize the same words differently.)</p><p>Turning data into strings of numbers makes them usable by AI models. Computers may not be able to see or read in the traditional sense, but they can run mathematical operations on numbers. That&#8217;s how AI detects patterns, compares inputs, and ultimately learns from data instead of relying on fixed rules.</p><p>These numbers aren&#8217;t stored in a database or an Excel spreadsheet. They exist in what&#8217;s called a high-dimensional space.</p><p>We can visualize and understand the difference between two and three dimensions. Schoolchildren are taught that a rectangle has two dimensions&#8212;length and width. A cube adds another dimension: depth. It&#8217;s much harder for us to grasp a fourth dimension.</p><p>But AI can understand hundreds, even thousands, of dimensions.</p><p>To navigate these vast high-dimensional spaces, a model learns during training which direction matters most by adjusting its internal weights&#8212;think of them as groups of dials that turn up or down to chart a course through this mathematical terrain. As training progresses, turning up weights for &#8220;furry&#8221; and &#8220;four legs&#8221; steers it deeper into dog country, while dialing down irrelevant features such as &#8220;fire hydrants&#8221; prevents the model from wandering into dead ends.</p><p>Through training, the model groups together the features that typically appear in pictures of dogs without being told what exactly a dog is. However, there is no index, as you might find in the back of a book, that you can consult to find the exact &#8220;dial&#8221; or weight corresponding to doglike features; those features are intertwined across the model&#8217;s complex architecture. Researchers have to go find them.</p><p>This gets at the core challenge of AI interpretability. Researchers know how to build and train these models. But they often can&#8217;t see what, exactly, in an image causes the model to adjust one specific dial out of billions.</p><p>Understanding how a model makes its predictions can help illuminate how much we can trust it&#8212;or, if necessary, how to fix it when its behavior deviates from what we want or expect.</p><p></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.chicagobooth.edu/review/can-we-break-open-ais-black-box&quot;,&quot;text&quot;:&quot;Continue Reading at CBR&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.chicagobooth.edu/review/can-we-break-open-ais-black-box"><span>Continue Reading at CBR</span></a></p>]]></content:encoded></item><item><title><![CDATA[AI Stock Picks Beat Benchmark in Live Market Test: Study ]]></title><description><![CDATA[AI autonomously searched the web, scored all Russell 1000 stocks, and constructed a daily portfolio.]]></description><link>https://www.ai-street.co/p/ai-stock-picks-beat-benchmark-in</link><guid isPermaLink="false">https://www.ai-street.co/p/ai-stock-picks-beat-benchmark-in</guid><dc:creator><![CDATA[Matt Robinson]]></dc:creator><pubDate>Thu, 19 Feb 2026 13:03:26 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/a4f6311a-176e-42a7-a92f-e514f1828268_1762x1566.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h6><strong>RESEARCH </strong></h6><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!9eZZ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9d86ddd9-3f63-4ceb-9244-b9603a5205f7_1162x606.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!9eZZ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9d86ddd9-3f63-4ceb-9244-b9603a5205f7_1162x606.png 424w, https://substackcdn.com/image/fetch/$s_!9eZZ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9d86ddd9-3f63-4ceb-9244-b9603a5205f7_1162x606.png 848w, https://substackcdn.com/image/fetch/$s_!9eZZ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9d86ddd9-3f63-4ceb-9244-b9603a5205f7_1162x606.png 1272w, https://substackcdn.com/image/fetch/$s_!9eZZ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9d86ddd9-3f63-4ceb-9244-b9603a5205f7_1162x606.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!9eZZ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9d86ddd9-3f63-4ceb-9244-b9603a5205f7_1162x606.png" width="1162" height="606" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/9d86ddd9-3f63-4ceb-9244-b9603a5205f7_1162x606.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:606,&quot;width&quot;:1162,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:116087,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.ai-street.co/i/188476725?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9d86ddd9-3f63-4ceb-9244-b9603a5205f7_1162x606.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!9eZZ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9d86ddd9-3f63-4ceb-9244-b9603a5205f7_1162x606.png 424w, https://substackcdn.com/image/fetch/$s_!9eZZ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9d86ddd9-3f63-4ceb-9244-b9603a5205f7_1162x606.png 848w, https://substackcdn.com/image/fetch/$s_!9eZZ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9d86ddd9-3f63-4ceb-9244-b9603a5205f7_1162x606.png 1272w, https://substackcdn.com/image/fetch/$s_!9eZZ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9d86ddd9-3f63-4ceb-9244-b9603a5205f7_1162x606.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Many AI + investing research papers suffer from the same problem: the models were trained on historical internet data that often contains the outcomes they are asked to predict. Ask a model today what happened to a stock in 2022, and it may already know. </p><p>This look-ahead bias makes me skeptical of many papers with &#8220;big&#8221; conclusions. But since the models have become ubiquitous, researchers can test their theories in real time. </p><p>Two Peking University researchers, Zefeng Chen and Darcy Pu, did just that. They ran a live, nine-month experiment asking a frontier AI model to pick stocks every night across the Russell 1000. </p><p><strong>Here&#8217;s what they did:</strong></p><ul><li><p>Every night from April 2025 through January 2026, they queried a leading U.S. frontier AI model via its web interface with live search enabled, with no pre-selected news or filings fed to the model. The model autonomously searched the web, synthesized what it found, and returned a score (&#8722;5 to +5) for each Russell 1000 stock.</p></li><li><p>Signals were generated after the 4pm close and before the next open. Portfolios were entered at the opening auction and exited the following open.</p></li><li><p>They ranked about 1,000 stocks by the model&#8217;s daily score, built a portfolio of the top 20 weighted by market value, and tested its performance using standard factor models.</p></li></ul><p><strong>Here&#8217;s what they found:</strong></p>
      <p>
          <a href="https://www.ai-street.co/p/ai-stock-picks-beat-benchmark-in">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[AI Boosts Retail Trading Volume: Research]]></title><description><![CDATA[Evidence from Italy&#8217;s ChatGPT ban shows AI expands the set of assets retail investors trade.]]></description><link>https://www.ai-street.co/p/ai-boosts-retail-trading-volume-research</link><guid isPermaLink="false">https://www.ai-street.co/p/ai-boosts-retail-trading-volume-research</guid><dc:creator><![CDATA[Matt Robinson]]></dc:creator><pubDate>Thu, 12 Feb 2026 12:02:23 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/9f75fbf1-7f11-4f4e-9aa5-69c4d6ba896c_1794x1686.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>&#8220;The price of intelligence is going to zero.&#8221;</p><p>I&#8217;ve thought of this quote often since interviewing <a href="https://www.youtube.com/watch?v=CGACi1Elh4c&amp;t=5s">Tharsis Souza</a>, now at Citadel, in a podcast episode in January last year. </p><p>Souza presciently highlighted the diminishing costs of expert analysis. (Check out the recording <a href="https://www.youtube.com/watch?v=CGACi1Elh4c&amp;t=5s">here</a>.) </p><p>I&#8217;ve seen more and more examples of this, in my own life, like trying to sort out a hairy Italian-U.S. tax issue with the help of ChatGPT. </p><p>More empirically, studies are coming out showing how much investors are using AI to broaden their investing universe. </p><h2><strong>When Italy Banned ChatGPT, Retail Traders Narrowed Their Bets</strong></h2><p>In March 2023, Italian regulators forced OpenAI to suspend service over data privacy concerns. </p><p>This 28-day blackout created a natural experiment, revealing how the sudden loss of a leading AI tool&#8212;available to neighboring countries but denied to Italians&#8212;altered trading behavior.</p><h3>The Study</h3><p>A research team led by Omri Even-Tov (UC Berkeley) analyzed granular, account-level data from the brokerage platform <strong>eToro</strong>.</p><ul><li><p><strong>The Scope:</strong> 3 million accounts across 100+ countries, covering stocks, crypto, ETFs, and commodities. The paper&#8217;s final analysis sample consists of 24,185 investors and 169,295 investor-month observations from Italy and neighboring control countries between January and July 2023.</p></li><li><p><strong>The Method:</strong> They compared <strong>Italian investors</strong> (the treatment group) against peers in <strong>France, Switzerland, Austria, and Slovenia</strong> (the control group).</p></li></ul><h3>The Key Finding: Narrowed Horizons</h3><p>The researchers used the <strong>Herfindahl-Hirschman Index (HHI)</strong> to measure trade concentration. While the average retail investor typically sticks to 2&#8211;3 assets, the ban caused Italian portfolios to shrink even further.</p><ul><li><p><strong>The Spike:</strong> During the ban, Italian trade concentration <strong>rose by ~3.1%</strong> relative to the control group.</p></li><li><p><strong>The Proof:</strong> The data showed a &#8220;dynamic&#8221; shift&#8212;the concentration spike appeared <em><strong>only</strong></em> during the month of the ban.</p></li><li><p><strong>The Rebound:</strong> Once the ban was lifted, Italian investors did not immediately return to their pre-ban behavior, even showing a brief &#8220;overcorrection&#8221; as they explored new assets to make up for lost time.</p></li></ul><p>The ban also reduced the likelihood that investors would initiate trades in assets they hadn&#8217;t previously touched. The odds of adding a new asset declined by roughly 10.3% during the ban. And trading shifted toward popular assets (the top 50, top 100, top 200 by aggregate retail volume) while pulling away from the long tail.</p><h3><strong>Not Just Behavior &#8212; Prices Moved Too</strong></h3><p>Even-Tov&#8217;s study captures what happened at the account level. But a separate team had already used the same Italy ban to measure something different: what happened to stock prices.</p><p>A separate study used the same Italy ChatGPT ban to examine market-level effects. The authors find that Italian firms with higher exposure to generative AI underperformed by roughly 9 percent during the ban. Bid-ask spreads widened, and analysts based in Italy issued fewer forecasts.</p><p>That&#8217;s two studies exploiting the same 28-day shock, measuring different outcomes, and arriving at the same directional conclusion: removing access to GenAI compressed both participation and market quality.</p><h3><strong>Where the Effect Was Strongest</strong></h3><p>Researchers sliced the data by asset class, investor characteristics, and firm characteristics.</p><p>By asset class, the results were driven by stocks and cryptocurrencies. ETFs, commodities, and currencies showed no significant effect. That&#8217;s consistent with the information processing story: stocks and crypto involve large volumes of unstructured, or hard-to-verify information. ETFs and commodities track standardized benchmarks. There&#8217;s less for an AI tool to help with.</p><p>By investor type, the effect was strongest among investors with lower income or less trading experience (consistent with those investors facing higher information processing costs) and among investors in technology-related occupations or students (consistent with higher likelihood of actually using ChatGPT in the first place).</p><p>By firm characteristics, the decline in trading activity was steeper for hard-to-value stocks. Firms with lower profitability or higher volatility saw the largest drop in Italian investor participation during the ban. The effect also intensified around earnings announcements, when investors face concentrated releases of financial data.</p><h3><strong>The Portfolio Consequences</strong></h3><p>Higher trade concentration has downstream effects. During the ban, Italian investors&#8217; portfolios exhibited more return comovement across holdings and higher portfolio volatility. Fewer distinct positions, more overlap in the stocks they held, less diversification.</p><p>One thing to note: the ban had no detectable effect on risk-adjusted trading performance. Investors weren&#8217;t making worse picks during the ban. They were just making fewer of them, and those picks looked more alike. GenAI appears to broaden the set of assets retail investors consider, but there&#8217;s no evidence (at least in this 28-day window) that it helps them choose better ones.</p><div class="pullquote"><p> Investors weren&#8217;t making worse picks during the ban. They were just making fewer of them, and those picks looked more alike.</p></div><h3><strong>A Pattern, Not a One-Off</strong></h3>
      <p>
          <a href="https://www.ai-street.co/p/ai-boosts-retail-trading-volume-research">
              Read more
          </a>
      </p>
   ]]></content:encoded></item></channel></rss>