Contrary to the hype around alternative data, the most crucial input for intraday trading AI is standard market data feeds from exchanges. This raw, high-volume data on quotes and trades is the truest expression of market intent, far outweighing the predictive value of news or social media feeds.
Hedge funds have a constant, daily need to make informed buy, sell, or hold decisions, creating a clear business problem that data solves. Corporations often lack this frequent, high-stakes decision-making cycle, making the value proposition of external data less immediate and harder to justify.
High-frequency trading firms are expanding into medium-frequency horizons (days to weeks). They use their sophisticated short-term AI models, which can predict optimal prices within the next hour, to inform the execution strategy for their longer-term positions, creating a cascading effect where intraday precision enhances multi-day trading performance.
For years, access to compute was the primary bottleneck in AI development. Now, as public web data is largely exhausted, the limiting factor is access to high-quality, proprietary data from enterprises and human experts. This shifts the focus from building massive infrastructure to forming data partnerships and expertise.
High-frequency trading (HFT) firms use proprietary exchange data feeds to legally front-run retail and institutional orders. This systemic disadvantage erodes investor confidence, pushing them toward high-risk YOLO call options and sports betting to seek returns.
Hudson River Trading shifted from handcrafted features based on human intuition to training models on raw, internet-scale market data. This emergent approach, similar to how ChatGPT is trained, has entirely overtaken traditional quant methods that relied on simpler techniques like linear regression.
Cliff Asnes explains that integrating machine learning into investment processes involves a crucial trade-off. While AI models can identify complex, non-linear patterns that outperform traditional methods, their inner workings are often uninterpretable, forcing a departure from intuitively understood strategies.
AI models can predict short-term stock prices, defying the efficient market hypothesis. However, the predictions are only marginally better than random, with an accuracy akin to "50.1%". The profitability comes not from magic, but from executing this tiny statistical edge millions of times across the market.
The primary reason multi-million dollar AI initiatives stall or fail is not the sophistication of the models, but the underlying data layer. Traditional data infrastructure creates delays in moving and duplicating information, preventing the real-time, comprehensive data access required for AI to deliver business value. The focus on algorithms misses this foundational roadblock.
While the "quad" economic outlook is crucial, the ultimate authority is the market's "signal"—a multi-factor model of price, volume, and volatility. Keith McCullough states if he had to choose only one, he would rely on the signal, as it reflects what the market *is* doing, not what it *should* be doing.
MDT deliberately avoids competing on acquiring novel, expensive datasets (informational edge). Instead, they focus on their analytical edge: applying sophisticated machine learning tools to long-history, high-quality standard datasets like financials and prices to find differentiated insights.