<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom">
  <channel>
    <title>Ooak Data Research</title>
    <link>https://ooakdata.com/research</link>
    <description>Research from Ooak Data on evaluation methodology, dataset design, and the gap between benchmark performance and real-world capability.</description>
    <language>en</language>
    <atom:link href="https://ooakdata.com/rss.xml" rel="self" type="application/rss+xml" />
    <item>
      <title>Why synthetic benchmarks are holding AI back</title>
      <link>https://ooakdata.com/research/why-synthetic-benchmarks-are-holding-ai-back</link>
      <guid isPermaLink="true">https://ooakdata.com/research/why-synthetic-benchmarks-are-holding-ai-back</guid>
      <description>The AI industry over-indexes on synthetic benchmarks that don&apos;t reflect real-world complexity. Models that ace leaderboards routinely fail in production.</description>
      <content:encoded><![CDATA[
<p>AI models have never looked better on paper. GPT-4, Claude, Gemini — they top leaderboard after leaderboard. MMLU scores climb past 90%. HumanEval gets saturated. New benchmarks appear, and within months, they too are nearly solved.</p>

<p>And yet, when these same models are deployed into real company workflows — reading messy documents, navigating CRM systems, coordinating across Slack and email — they break in ways that no benchmark predicted.</p>

<p>This is not a coincidence. It is a structural problem with how we evaluate AI.</p>

<h2>The saturation problem</h2>

<p>Benchmarks are being solved faster than we can create them. The <a href="https://hai.stanford.edu/ai-index/2025-ai-index-report" target="_blank" rel="noopener">2025 Stanford HAI AI Index Report</a> documents that traditional benchmarks like MMLU, GSM8K, and HumanEval have effectively reached saturation — AI appears "nearly perfect" on them, yet still produces incorrect answers and unwanted outputs in practice. New benchmarks introduced in 2023, including MMMU, GPQA, and SWE-bench, saw performance jump by 18.8, 48.9, and 67.3 percentage points within a single year, suggesting they too will soon become ceiling tests rather than discriminative measures.</p>

<p>A recent large-scale study by <a href="https://arxiv.org/abs/2602.16763" target="_blank" rel="noopener">Akhtar et al. (2025)</a>, published at ICML, analyzed 60 AI benchmarks systematically and found that nearly half show high or very high saturation. Among older benchmarks (those over 60 months old), the saturation rate reaches 54.5%. Crucially, neither private test sets nor open-ended question formats were sufficient to prevent saturation — the mechanisms we assumed would keep benchmarks useful are failing.</p>

<p>When a benchmark is saturated, it stops telling us anything meaningful about the difference between models, or about whether a model is ready for production. It becomes a checkbox, not a test.</p>

<h2>What gets measured gets gamed</h2>

<p>There is a deeper problem than saturation: optimization pressure. Charles Goodhart's observation that "when a measure becomes a target, it ceases to be a good measure" applies with particular force to AI evaluation.</p>

<p><a href="https://www.cell.com/patterns/fulltext/S2666-3899(22)00056-3" target="_blank" rel="noopener">Rachel Thomas and David Uminsky (2022)</a> argued this point in <em>Patterns</em>, writing that "current AI approaches have weaponized Goodhart's law by centering on optimizing a particular measure as a target." They document how over-optimizing metrics leads to manipulation, short-termism, and outcomes that look good on the metric while missing the underlying objective.</p>

<p>This is not theoretical. <a href="https://arxiv.org/abs/2210.10760" target="_blank" rel="noopener">Gao, Schulman, and Hilton at OpenAI (2022)</a> empirically measured Goodhart's Law in the context of RLHF, showing that optimizing against imperfect proxy reward models causes the true objective to degrade after a point — and that this relationship follows predictable scaling laws. The more aggressively you optimize for a proxy metric, the more reliably you degrade real performance.</p>

<p><a href="https://arxiv.org/abs/2308.14272" target="_blank" rel="noopener">Hsia et al. (2023)</a> demonstrated this in NLP explanation benchmarks specifically, showing that standard metrics like ERASER comprehensiveness and EVAL-X scores can be inflated dramatically without altering model predictions or explanations on real test inputs. The scores go up. The model's actual behavior does not change.</p>

<h2>The contamination problem</h2>

<p>Even setting aside optimization pressure, there is a more direct problem: the models have seen the test data.</p>

<p><a href="https://arxiv.org/abs/2311.09783" target="_blank" rel="noopener">Deng et al. (2023)</a>, in a paper presented at NAACL 2024, developed a protocol to detect benchmark contamination in proprietary LLMs by testing whether models could guess missing answer options. The results: ChatGPT and GPT-4 achieved exact match rates of 52% and 57%, respectively, in guessing missing options on MMLU items — strong evidence that these models have memorized substantial portions of one of the most widely used benchmarks in the field.</p>

<p>A comprehensive survey by <a href="https://arxiv.org/abs/2406.04244" target="_blank" rel="noopener">Xu et al. (2024)</a> found contamination rates ranging from 1% to 45% across tested LLMs, categorizing contamination into four severity levels. Their conclusion is blunt: it is "impracticable to fully remove the risks associated with contamination" due to the scale of pre-training data and the proliferation of AI-generated content that may itself contain benchmark questions.</p>

<p>When a model scores 90% on a benchmark it has partially memorized, that score tells us about memorization, not capability. And since we often cannot determine the extent of contamination in proprietary models, we cannot know how much of any given score is genuine.</p>

<h2>The real-world gap</h2>

<p>If benchmark scores were just noisy but directionally correct, the problem would be manageable. But growing evidence suggests the gap between benchmark performance and real-world deployment is not just noise — it is systematic and large.</p>

<p>A striking example comes from clinical AI. <a href="https://www.jmir.org/2025/1/e84120" target="_blank" rel="noopener">Gong et al. (2025)</a>, in a systematic review published in the <em>Journal of Medical Internet Research</em>, found that <strong>diagnostic accuracy drops from 82% on traditional case vignettes to 62.7% on multi-turn patient dialogues</strong> — a 19.3 percentage-point decrease. And only 5% of the 761 LLM evaluation studies they reviewed assessed performance on real patient care data.</p>

<p>As <a href="https://techcrunch.com/2024/03/07/heres-why-most-ai-benchmarks-tell-us-so-little/" target="_blank" rel="noopener">Kyle Wiggers noted in TechCrunch (2024)</a>, the most commonly used benchmarks "haven't been adapted to reflect how models are used today." Researchers at the Allen Institute for AI have described an "evaluation crisis" — benchmarks like GPQA test PhD-level science while most real users write emails and cover letters. The mismatch between what we test and what we deploy is not a gap; it is a chasm.</p>

<p><a href="https://arxiv.org/abs/2402.09880" target="_blank" rel="noopener">McIntosh et al. (2024)</a>, in a critical assessment of 23 state-of-the-art LLM benchmarks published in <em>IEEE Transactions on Artificial Intelligence</em>, found significant limitations across the board: biases in question construction, inability to measure genuine reasoning, implementation inconsistencies across environments, and systematic failure to account for cultural norms.</p>

<h2>What would better evaluation look like?</h2>

<p>If static benchmarks are failing us, what should replace them?</p>

<p>One promising direction is human evaluation at scale. The <a href="https://arxiv.org/abs/2403.04132" target="_blank" rel="noopener">Chatbot Arena platform (Chiang et al., 2024)</a>, presented at ICML, demonstrates that crowdsourced pairwise comparison achieves 89.1% agreement with expert raters and provides far better separability between models than standard benchmarks. With over 240,000 votes from 90,000 users across 100+ languages, it suggests that human preference is both scalable and more informative than static tests.</p>

<p>But human evaluation alone is not sufficient for agents — systems that take actions, use tools, and complete multi-step workflows. For these, we need evaluation environments that replicate the conditions of deployment: real tools, real data patterns, real organizational complexity.</p>

<p>This means moving from static question-answer pairs to dynamic environments where agents interact with realistic systems — reading documents with real formatting quirks, navigating CRM interfaces, coordinating across communication channels. It means building evaluation infrastructure from real company data, anonymized but structurally authentic, so that the test reflects the mess an agent will actually face.</p>

<p>The goal is not to make evaluation harder. It is to make it honest.</p>

<h2>The cost of comfortable metrics</h2>

<p>The current benchmark paradigm is not just inadequate — it is actively harmful. It creates false confidence. It directs research effort toward optimizing scores rather than solving real problems. It allows models to be marketed as "state-of-the-art" on metrics that bear little relation to the tasks organizations actually need performed.</p>

<p>The 2025 Stanford HAI report puts it plainly: AI appears "nearly perfect" on existing benchmarks yet "still gives incorrect answers, produces unwanted outputs, and is difficult to interact with." This gap between measured performance and experienced performance is not closing. If anything, as models get better at gaming benchmarks, it is widening.</p>

<p>The AI industry needs evaluation infrastructure built on real-world data, grounded in actual deployment conditions, and designed to expose failures rather than confirm successes.</p>

<hr />

<h2>References</h2>

<ul>
<li>Akhtar, M., Reuel, A., Soni, P., et al. (2025). When AI Benchmarks Plateau: A Systematic Study of Benchmark Saturation. <em>ICML 2025</em>.</li>
<li>Chiang, W.-L., Zheng, L., Sheng, Y., et al. (2024). Chatbot Arena: An Open Platform for Evaluating LLMs by Human Preference. <em>ICML 2024</em>.</li>
<li>Deng, C., Zhao, Y., Tang, X., Gerstein, M., &amp; Cohan, A. (2023). Investigating Data Contamination in Modern Benchmarks for Large Language Models. <em>NAACL 2024</em>.</li>
<li>Gao, L., Schulman, J., &amp; Hilton, J. (2022). Scaling Laws for Reward Model Overoptimization. <em>ICML 2023</em>.</li>
<li>Gong, E. J., Bang, C. S., Lee, J. J., &amp; Baik, G. H. (2025). Knowledge-Practice Performance Gap in Clinical Large Language Models. <em>JMIR</em>.</li>
<li>Hsia, J., Pruthi, D., Singh, A., &amp; Lipton, Z. C. (2023). Goodhart's Law Applies to NLP's Explanation Benchmarks. <em>EACL 2024</em>.</li>
<li>McIntosh, T. R., et al. (2024). Inadequacies of Large Language Model Benchmarks in the Era of Generative AI. <em>IEEE Trans. on AI</em>.</li>
<li>Stanford HAI. (2025). The 2025 AI Index Report.</li>
<li>Thomas, R. &amp; Uminsky, D. (2022). Reliance on metrics is a fundamental challenge for AI. <em>Patterns</em>, 3(5).</li>
<li>Xu, C., Guan, S., Greene, D., &amp; Kechadi, M-T. (2024). Benchmark Data Contamination of Large Language Models: A Survey.</li>
</ul>
]]></content:encoded>
      <pubDate>Wed, 18 Mar 2026 00:00:00 GMT</pubDate>
    </item>
    <item>
      <title>The data problem in agentic AI</title>
      <link>https://ooakdata.com/research/the-data-problem-in-agentic-ai</link>
      <guid isPermaLink="true">https://ooakdata.com/research/the-data-problem-in-agentic-ai</guid>
      <description>AI agents need fundamentally different data than chatbots — multi-step interaction trajectories, tool-use demonstrations, and multi-modal workflow data. This data barely exists.</description>
      <content:encoded><![CDATA[
<p>The AI industry is in the middle of a paradigm shift. The focus is moving from models that answer questions to agents that complete tasks — systems that read documents, send messages, update databases, navigate tools, and execute multi-step workflows autonomously.</p>

<p>But there is a problem that the excitement around agents tends to obscure: we do not have the data to train them.</p>

<p>The data infrastructure that powered the LLM revolution — web scrapes, books, code repositories — is fundamentally inadequate for agents. Agents need something different: multi-step interaction trajectories, tool-use demonstrations, multi-modal workflow data, and environments where they can learn through trial and error. This data barely exists, and building it is harder than most people realize.</p>

<h2>Agents are not chatbots</h2>

<p>The distinction matters because it determines what training data looks like.</p>

<p>A chatbot takes a prompt and returns a response. Its training data is input-output pairs. The internet is full of this data. It is comparatively easy to collect, clean, and scale.</p>

<p>An agent takes a goal and executes a sequence of actions across multiple tools and systems to achieve it. Its training data needs to capture trajectories — chains of reasoning, tool invocations, observations, corrections, and completions that unfold over many steps. This kind of data is vanishingly rare in the wild.</p>

<p><a href="https://arxiv.org/abs/2210.03629" target="_blank" rel="noopener">Yao et al. (2023)</a> formalized this distinction with ReAct, an architecture that interleaves reasoning traces with action steps. The key insight: agents require fundamentally different data than standard LLMs. They need trajectories of interleaved thought and action, not just input-output pairs.</p>

<p><a href="https://arxiv.org/abs/2302.04761" target="_blank" rel="noopener">Schick et al. (2023)</a> demonstrated a related point with Toolformer: LLMs can learn to use tools, but the training data must be augmented with API call annotations at positions where tool use helps predict future tokens. Standard text corpora contain no tool-use demonstrations. The data must be synthetically generated or manually curated — and neither approach scales cleanly.</p>

<h2>The benchmark evidence</h2>

<p>Recent agent benchmarks have made the data gap quantitatively visible.</p>

<p><strong><a href="https://arxiv.org/abs/2307.13854" target="_blank" rel="noopener">WebArena</a></strong> (Zhou et al., 2023), published at ICLR 2024, evaluates agents on realistic web tasks using fully functional self-hosted websites. The best GPT-4-based agent achieved only <strong>14.41% end-to-end task success, compared to 78.24% for humans</strong>. This is a 5x performance differential on tasks that most knowledge workers perform routinely.</p>

<p><strong><a href="https://arxiv.org/abs/2404.07972" target="_blank" rel="noopener">OSWorld</a></strong> (Xie et al., 2024), presented at NeurIPS, extends this to real computer environments across Ubuntu, Windows, and macOS. The best model achieves <strong>12.24% success versus 72.36% for humans</strong>. The primary failure modes: GUI grounding and operational knowledge — capabilities that require visual, multi-modal training data that barely exists.</p>

<p><strong><a href="https://arxiv.org/abs/2401.13649" target="_blank" rel="noopener">VisualWebArena</a></strong> (Koh et al., 2024), published at ACL, finds that the best vision-language model agents achieve only <strong>16.4% success versus 88.7% for humans</strong> on visually grounded web tasks. Text-only agent training data is fundamentally insufficient.</p>

<p><strong><a href="https://arxiv.org/abs/2308.03688" target="_blank" rel="noopener">AgentBench</a></strong> (Liu et al., 2023), accepted at ICLR 2024, evaluated 29 LLMs across 8 diverse environments and found a significant performance gap between top commercial models and open-source alternatives. The authors identify poor long-term reasoning, decision-making, and instruction-following as the main obstacles — all capabilities that require multi-step interaction data to develop.</p>

<p>The pattern is consistent: models that perform well on static benchmarks struggle dramatically when placed in realistic, multi-step, multi-tool environments. The models are not incapable. They are undertrained — specifically, they lack the right kind of training data.</p>

<h2>The reliability problem</h2>

<p>Even when agents succeed on a task, they do not succeed reliably.</p>

<p><a href="https://arxiv.org/abs/2406.12045" target="_blank" rel="noopener">Yao et al. (2024)</a> introduced <strong>tau-bench</strong>, a benchmark for tool-agent-user interaction in real-world customer service domains. Their findings: even state-of-the-art agents succeed on fewer than 50% of customer service tasks, and <strong>consistency drops to approximately 25% when the same task is repeated eight times</strong>.</p>

<p>For enterprise deployment, this is the critical metric. A system that works 50% of the time is not half as useful as one that works 100% of the time — it is essentially unusable for any workflow where errors have consequences.</p>

<h2>Why synthetic data is not enough</h2>

<p>The obvious response to the data scarcity problem is to generate synthetic training data — use a large model to produce agent trajectories, then fine-tune smaller models on those trajectories.</p>

<p>This approach works, up to a point. <a href="https://arxiv.org/abs/2310.05915" target="_blank" rel="noopener">Chen et al. (2023)</a> showed with FireAct that fine-tuning Llama2-7B with just 500 agent trajectories generated by GPT-4 leads to a <strong>77% performance increase</strong> on HotpotQA. The data is clearly valuable. But it creates two problems.</p>

<p>First, there is a <strong>quality ceiling</strong>. Synthetic trajectories inherit the limitations of the generating model. They do not contain patterns the generating model has never seen — and the messy, ambiguous, context-dependent workflows of real companies are precisely what no model has been trained on.</p>

<p>Second, there is an <strong>entanglement problem</strong>. <a href="https://arxiv.org/abs/2403.12881" target="_blank" rel="noopener">Chen et al. (2024)</a> found with Agent-FLAN that current agent training corpora entangle format-following with agent reasoning, causing distribution shift. Worse, improving agent abilities through naive fine-tuning introduces hallucinations as a side effect.</p>

<p><a href="https://arxiv.org/abs/2310.12823" target="_blank" rel="noopener">Zeng et al. (2023)</a>, with AgentTuning, demonstrated that the quality and diversity of agent trajectory data is the bottleneck, not model architecture. Their AgentLM-70B became comparable to GPT-3.5-turbo on unseen agent tasks, confirming that the right training data can close capability gaps that no amount of architectural innovation addresses.</p>

<h2>What agent data actually needs to look like</h2>

<p><a href="https://www.anthropic.com/research/building-effective-agents" target="_blank" rel="noopener">Anthropic's research team (2024)</a>, in their guide to building effective agents, makes the point that the practical bottleneck is not model architecture but "the quality of tool definitions, context engineering, and evaluation data." The data and evaluation layer — not the model layer — is the binding constraint.</p>

<p>A <a href="https://www.wing.vc/content/rl-environments-for-agentic-ai-who-will-win-the-training-verification-layer-by-2030" target="_blank" rel="noopener">2025 analysis by Wing Venture Capital</a> frames this in market terms: RL environments are playing for AI agents the same role that EDA played for silicon design. Anthropic alone is estimated to be spending tens of millions annually on RL environments, with 3-5x growth expected into 2026.</p>

<p>This points to what agent training data actually needs:</p>

<ul>
<li><strong>Real-world grounding.</strong> Training data sourced from actual company workflows — documents with real formatting inconsistencies, conversations with organizational context, project management tools with actual task dependencies.</li>
<li><strong>Multi-modal coherence.</strong> Documents, conversations, and tool data that are interconnected — the Slack message references the Jira ticket which references the Google Doc.</li>
<li><strong>Expert-level task calibration.</strong> Tasks designed to challenge the current frontier of model capability, continuously updated as models improve.</li>
<li><strong>RL environments, not static datasets.</strong> Agents learn through interaction — taking actions, observing outcomes, adjusting strategies. The shift from datasets to environments is as fundamental as the shift from chatbots to agents.</li>
</ul>

<h2>The infrastructure gap</h2>

<p>The data problem in agentic AI is not a temporary inconvenience that will be solved by scaling existing approaches. It is a structural gap in the AI infrastructure stack. We have world-class models, increasingly capable architectures, and billions of dollars in deployment demand. What we lack is the data infrastructure to connect model capability to real-world performance.</p>

<p>Closing this gap requires building something new: pipelines that source real-world data through established partnerships, anonymize it while preserving structural fidelity, and transform it into RL environments where agents can train against the complexity they will actually face.</p>

<p>The model layer has had its revolution. The data layer is next.</p>

<hr />

<h2>References</h2>

<ul>
<li>Anthropic. (2024). Building Effective Agents.</li>
<li>Chen, B., et al. (2023). FireAct: Toward Language Agent Fine-tuning.</li>
<li>Chen, Z., et al. (2024). Agent-FLAN: Designing Data and Methods of Effective Agent Tuning. <em>ACL 2024</em>.</li>
<li>Koh, J. Y., et al. (2024). VisualWebArena: Evaluating Multimodal Agents on Realistic Visual Web Tasks. <em>ACL 2024</em>.</li>
<li>Liu, X., et al. (2023). AgentBench: Evaluating LLMs as Agents. <em>ICLR 2024</em>.</li>
<li>Schick, T., et al. (2023). Toolformer: Language Models Can Teach Themselves to Use Tools.</li>
<li>Wing Venture Capital. (2025). RL Environments for Agentic AI.</li>
<li>Xie, T., et al. (2024). OSWorld: Benchmarking Multimodal Agents. <em>NeurIPS 2024</em>.</li>
<li>Yao, S., et al. (2023). ReAct: Synergizing Reasoning and Acting. <em>ICLR 2023</em>.</li>
<li>Yao, S., et al. (2024). tau-bench: A Benchmark for Tool-Agent-User Interaction.</li>
<li>Zeng, A., et al. (2023). AgentTuning: Enabling Generalized Agent Abilities for LLMs.</li>
<li>Zhou, S., et al. (2023). WebArena: A Realistic Web Environment. <em>ICLR 2024</em>.</li>
</ul>
]]></content:encoded>
      <pubDate>Wed, 11 Feb 2026 00:00:00 GMT</pubDate>
    </item>
    <item>
      <title>From sandbox to production: the gap nobody talks about</title>
      <link>https://ooakdata.com/research/from-sandbox-to-production</link>
      <guid isPermaLink="true">https://ooakdata.com/research/from-sandbox-to-production</guid>
      <description>80% of AI projects fail in production. The problem is not the models — it&apos;s the gap between evaluation conditions and deployment conditions.</description>
      <content:encoded><![CDATA[
<p>Here is a pattern that plays out across the AI industry, over and over:</p>

<p>A team builds an AI system. It works impressively in the development environment. They demo it to stakeholders. Everyone is excited. They push it toward production. It fails — not catastrophically, not in a way that makes headlines, but persistently, in small ways that erode trust and eventually lead the project to be quietly shelved.</p>

<p>The industry has a name for this now: the sandbox-to-production gap. But we have been remarkably slow to treat it as the structural problem it is.</p>

<h2>The numbers are worse than you think</h2>

<p>The failure rate of AI projects in production is not a secret, but the scale is still surprising.</p>

<p>A <a href="https://www.rand.org/pubs/research_reports/RRA2680-1.html" target="_blank" rel="noopener">2024 RAND Corporation study</a>, based on interviews with 65 data scientists and engineers, estimates that <strong>80% of AI projects fail</strong> — double the failure rate of non-AI IT projects.</p>

<p>In 2024, <a href="https://www.gartner.com/en/newsroom/press-releases/2024-07-29-gartner-predicts-30-percent-of-generative-ai-projects-will-be-abandoned-after-proof-of-concept-by-end-of-2025" target="_blank" rel="noopener">Gartner predicted</a> that at least <strong>30% of generative AI projects would be abandoned after proof of concept</strong> by end of 2025, citing poor data quality, inadequate risk controls, escalating costs, and unclear business value.</p>

<p>By 2025, <a href="https://www.spglobal.com/market-intelligence/en/news-insights/research/ai-experiences-rapid-adoption-but-with-mixed-outcomes-highlights-from-vote-ai-machine-learning" target="_blank" rel="noopener">S&amp;P Global's Voice of the Enterprise survey</a> made that prediction look optimistic: <strong>42% of companies had abandoned the majority of their AI initiatives before reaching production</strong> — up from 17% the year before.</p>

<p><a href="https://www.nttdata.com/global/en/insights/focus/2024/between-70-85p-of-genai-deployment-efforts-are-failing" target="_blank" rel="noopener">NTT DATA's 2024 global survey</a> of 2,300+ leaders across 34 countries found that <strong>70-85% of GenAI deployments fail to meet their desired ROI</strong>, despite 83% of organizations claiming a well-defined GenAI strategy.</p>

<p><a href="https://www.mckinsey.com/capabilities/quantumblack/our-insights/the-state-of-ai" target="_blank" rel="noopener">McKinsey's 2025 State of AI survey</a> confirms the pattern from the other direction: while 78% of organizations now use AI in at least one business function, <strong>only about 5.6% report that more than 5% of their organization's EBIT is attributable to AI</strong>. Adoption is nearly universal. Value creation is not.</p>

<h2>Why demos work and production doesn't</h2>

<p>The sandbox-to-production gap is not random. It has specific, identifiable causes — and most of them are about data, not models.</p>

<h3>Underspecification</h3>

<p><a href="https://arxiv.org/abs/2011.03395" target="_blank" rel="noopener">D'Amour et al. (2022)</a>, in a landmark paper from Google published in the <em>Journal of Machine Learning Research</em>, introduced the concept of "underspecification": ML pipelines return many predictors with equivalently strong held-out performance in the training domain, but these predictors can behave very differently in deployment.</p>

<p>This means that good test-set performance does not guarantee good deployment performance — not because the model is bad, but because the evaluation does not constrain the model's behavior enough to ensure it generalizes correctly. The authors demonstrated this across computer vision, NLP, medical imaging, clinical risk prediction, and genomics.</p>

<h3>Data cascades</h3>

<p><a href="https://dl.acm.org/doi/10.1145/3411764.3445518" target="_blank" rel="noopener">Sambasivan et al. (2021)</a>, in a study published at CHI based on interviews with 53 AI practitioners, identified "data cascades" — compounding negative downstream effects from data quality issues that are invisible during development but devastating in production.</p>

<p><strong>92% of practitioners had experienced data cascades.</strong> The root cause: conventional AI/ML practices systematically undervalue data quality work in favor of model work. As one practitioner put it (giving the paper its title): "Everyone wants to do the model work, not the data work."</p>

<h3>Hidden technical debt</h3>

<p><a href="https://papers.neurips.cc/paper/5656-hidden-technical-debt-in-machine-learning-systems.pdf" target="_blank" rel="noopener">Sculley et al. (2015)</a>, in one of the most cited ML systems papers, observed that only a tiny fraction of real-world ML systems consists of the ML model code itself. The surrounding infrastructure — data pipelines, feature stores, serving systems, monitoring, configuration management — is vast and prone to what they call "hidden technical debt."</p>

<p>This debt includes boundary erosion, entanglement, hidden feedback loops, and undeclared consumers. None of these failure modes appear in the sandbox. All of them appear in production.</p>

<h2>The organizational gap</h2>

<p>The sandbox-to-production gap is not purely technical. It is also organizational.</p>

<p><a href="https://dl.acm.org/doi/abs/10.1145/3510003.3510209" target="_blank" rel="noopener">Nahar et al. (2022)</a>, in a study published at ICSE based on interviews with 45 practitioners from 28 organizations, found that the transition breaks down at the boundary between data science and software engineering — miscommunication about model assumptions, inadequate documentation, and misaligned practices.</p>

<p>A <a href="https://www.brookings.edu/articles/the-last-mile-problem-in-ai/" target="_blank" rel="noopener">2024 analysis from the Brookings Institution</a> by Fleming, Li, and Thompson quantifies a specific aspect of this gap: while 80% of computer vision tasks are technically automatable, <strong>only 23% are economically viable</strong> when customization costs are included.</p>

<p><a href="https://hbr.org/2026/03/the-last-mile-problem-slowing-ai-transformation" target="_blank" rel="noopener">Lakhani, Spataro, and Stave (2026)</a>, writing in the <em>Harvard Business Review</em>, identify seven organizational frictions that prevent AI from scaling beyond pilots. Among them: organizations become "pilot-rich but transformation-poor"; legacy workflows ("process debt") prevent AI from operating reliably; and tribal knowledge hoarding creates invisible dependencies.</p>

<h2>The evaluation problem at the root</h2>

<p>These causes — underspecification, data cascades, technical debt, organizational friction — share a common root: <strong>the evaluation environment does not match the deployment environment.</strong></p>

<p>When we test models on clean benchmarks with well-formed inputs, we learn whether the model has a capability. We do not learn whether it can exercise that capability in the conditions where it will actually operate.</p>

<p>When we evaluate agents in sandboxes with synthetic data, we learn whether the agent can complete a task in controlled conditions. We do not learn whether it can handle the messy Slack threads, the ambiguous CRM entries, the documents that don't follow templates, the organizational structures that affect who can access what information.</p>

<h2>What closing the gap requires</h2>

<p>The solution is not better models. The current generation of models is already capable of far more than production deployment rates suggest. The solution is better data infrastructure:</p>

<ul>
<li><strong>Sourcing data from real companies</strong>, not generating it synthetically. Real organizational data has properties — inconsistency, ambiguity, implicit context, access control structures — that synthetic data cannot replicate.</li>
<li><strong>Anonymizing without simplifying.</strong> The anonymization process must preserve the structural complexity of the original data. An anonymized digital twin should be just as messy and interconnected as the original.</li>
<li><strong>Building dynamic environments, not static test sets.</strong> Agents need to be evaluated in environments where they take actions and observe consequences, not just answer questions.</li>
<li><strong>Calibrating for the frontier.</strong> Evaluation environments that can be solved by current models are not useful for long. Tasks need continuous calibration against the latest models.</li>
</ul>

<h2>The cost of the gap</h2>

<p>Every AI project that succeeds in the sandbox but fails in production represents wasted engineering time, wasted compute, and — most importantly — wasted organizational trust. When a team demos an impressive AI system that then fails to deliver, the organization does not just lose a project. It loses confidence in the next project.</p>

<p>The S&amp;P Global finding — that the proportion of organizations citing positive impact from GenAI fell across every objective assessed between 2024 and 2025 — is not a sign that AI does not work. It is a sign that the gap between what AI can do in controlled conditions and what it delivers in production is eroding the confidence needed to invest in deployment.</p>

<p>Closing this gap is not a research problem in the traditional sense. The models are good enough. The architectures are capable enough. What is missing is the data infrastructure that connects model capability to deployment reality — evaluation environments built from the real world, not from our idealized version of it.</p>

<hr />

<h2>References</h2>

<ul>
<li>D'Amour, A., et al. (2022). Underspecification Presents Challenges for Credibility in Modern ML. <em>JMLR</em>, 23, 1-61.</li>
<li>Fleming, M., Li, W., &amp; Thompson, N. C. (2024). The Last Mile Problem in AI. <em>Brookings Institution</em>.</li>
<li>Gartner. (2024). Gartner Predicts 30% of Generative AI Projects Will Be Abandoned After PoC.</li>
<li>Lakhani, K. R., Spataro, J., &amp; Stave, J. (2026). The "Last Mile" Problem Slowing AI Transformation. <em>HBR</em>.</li>
<li>McKinsey &amp; Company. (2025). The State of AI: Global Survey.</li>
<li>Nahar, N., et al. (2022). Collaboration Challenges in Building ML-Enabled Systems. <em>ICSE '22</em>.</li>
<li>NTT DATA. (2024). Between 70-85% of GenAI Deployments Are Failing to Meet Their Desired ROI.</li>
<li>Paleyes, A., Urma, R.-G., &amp; Lawrence, N. D. (2022). Challenges in Deploying Machine Learning. <em>ACM Computing Surveys</em>.</li>
<li>Ryseff, J., De Bruhl, B. F., &amp; Newberry, S. J. (2024). Root Causes of Failure for AI Projects. <em>RAND Corporation</em>.</li>
<li>S&amp;P Global. (2025). AI Experiences Rapid Adoption, But with Mixed Outcomes.</li>
<li>Sambasivan, N., et al. (2021). Data Cascades in High-Stakes AI. <em>CHI '21</em>.</li>
<li>Sculley, D., et al. (2015). Hidden Technical Debt in ML Systems. <em>NeurIPS 2015</em>.</li>
</ul>
]]></content:encoded>
      <pubDate>Thu, 22 Jan 2026 00:00:00 GMT</pubDate>
    </item>
  </channel>
</rss>