AI Readiness in Portfolio Companies: Signal, Noise, and What Actually Matters

AI Readiness in Portfolio Companies: Signal, Noise, and What Actually Matters

Navigating the hype to uncover operational truth: How deal teams can assess true AI maturity, avoid 'AI washing', and underwrite realistic value creation.

Global enterprises invested an estimated $684 billion in artificial intelligence initiatives during 2025, yet a staggering $547 billion of that capital failed to deliver its intended business value [3]. For private equity deal teams and operating partners evaluating target companies, this massive gap between capital expenditure and realized value represents the single greatest underwriting risk of the current technological cycle. Acquirers are routinely presented with aggressive growth models built on the assumption of immediate artificial intelligence integration. However, beneath the polished management presentations and carefully curated software demonstrations lies a complex operational reality. Assessing a portfolio company is no longer about determining whether they use modern algorithms. It is about rigorously quantifying whether their foundational digital architecture, their data pipelines, and their human workflows are mature enough to support non-deterministic technology without destroying gross margins.

The AI Paradox in Private Markets: High Adoption, Low Impact

Dashboard showing high AI adoption versus low productivity impact

The current enterprise technology environment is characterized by a profound paradox. On one side of the equation, baseline adoption of artificial intelligence appears exceptionally high, creating an illusion that the entire market has successfully modernized. On the other side, the measurable financial returns on these deployments are virtually non-existent for the vast majority of organizations. Investors must navigate conflicting data regarding this baseline adoption rate. For example, a 2025 McKinsey & Company study reports that up to 88% of surveyed organizations regularly use artificial intelligence in at least one business function [1]. Conversely, the National Bureau of Economic Research notes a 69% active use rate [2], while broader U.S. Census Bureau estimates put general enterprise use significantly lower at roughly 18% [9]. This variance highlights a critical diligence issue. Executive surveys tend to capture superficial experimentation rather than structural integration.

The Illusion of Enterprise Ubiquity

When a target company claims to be an artificial intelligence innovator, deal teams must immediately investigate the depth of that adoption. The 88% adoption figure widely cited by McKinsey masks a severe bottleneck in enterprise operationalization [1]. In reality, providing employees with corporate licenses to large language models does not constitute enterprise adoption. It is merely decentralized software provisioning. True ubiquity requires algorithmic processes to be embedded directly into the core revenue generating or cost saving workflows of the business. Unfortunately, most organizations treat these new tools as isolated experiments. They deploy basic conversational interfaces on top of fragmented data silos and declare the initiative a success for internal marketing purposes. This illusion of ubiquity frequently inflates valuations during early stage deal discussions, leading acquirers to overpay for a capability that is fundamentally disjointed from the core product architecture.

Source: McKinsey & Company (2025) / NBER (2026)

Trapped in Pilot Purgatory

The failure to move beyond isolated experimentation has left nearly two thirds of organizations trapped in what industry analysts call pilot purgatory. A proof of concept is relatively inexpensive and easy to build. An engineering team can quickly connect an external application programming interface to a limited, sanitized dataset and produce a compelling demonstration. However, graduating that pilot into a secure, scalable production environment requires structural capabilities that most mid-market companies simply lack. They lack the data governance to ensure sensitive customer information is not leaked into the training data. They lack the cloud elasticity to handle unpredictable compute costs. As a result, the pilot remains isolated. It functions perfectly in a controlled sandbox but completely fails when exposed to the messy, unstructured realities of live enterprise data. For private equity operating partners, identifying a company stuck in pilot purgatory is essential. These companies have already incurred the initial software development costs, but they will never realize the corresponding operational efficiencies without a massive, unbudgeted overhaul of their core digital infrastructure.

The Labor Productivity Disconnect

The most concerning metric for financial sponsors involves the complete disconnect between technology deployment and actual labor productivity. The fundamental investment thesis for many software assets is that algorithmic automation will reduce headcount requirements and expand gross margins. The empirical data thoroughly contradicts this assumption. The National Bureau of Economic Research conducted a massive survey of approximately 6,000 executives across four major economies, finding that 89% of firms reported absolutely zero impact on labor productivity over the last three years [2]. Furthermore, the average measured productivity boost across all 6,000 surveyed firms was a mere 0.29% [2]. This creates a massive underwriting vulnerability. If an investment thesis assumes a 15% reduction in operating expenses due to automated data entry or customer service, and the statistical reality is a 0.29% improvement, the post-close earnings before interest and taxes will severely miss projections. The failure to generate measurable labor productivity stems from organizations attempting to layer advanced algorithms over inefficient, legacy human workflows. Instead of eliminating the work, the technology simply creates a parallel process, requiring human workers to constantly monitor, correct, and verify the algorithmic outputs, effectively doubling the operational friction.

Decoding the Trajectory of AI Project Failure

Funnel chart illustrating AI project abandonment and failure rates

To accurately model post-close value creation, investment committees must account for the statistical probability that a target company's ongoing technological initiatives will fail. Traditional enterprise software deployments carry known risks, but the failure rates associated with non-deterministic algorithms are entirely unprecedented in modern enterprise technology. Relying on standard software due diligence frameworks to evaluate these projects will inevitably lead to massive capital misallocation.

The Rising Rate of Abandoned Initiatives

The trajectory of project failure is not stabilizing; it is actively worsening as companies attempt more complex integrations. According to comprehensive data from the RAND Corporation, enterprise artificial intelligence initiatives face an 80.3% failure rate [3]. This is exactly twice the failure rate of traditional non-AI information technology projects [3]. More alarmingly, a 2025 report by S&P Global Market Intelligence indicates a dramatic year-over-year spike in project abandonment. Fully 42% of companies abandoned at least one major initiative in 2025, up drastically from just 17% in 2024 [4]. This surge in abandonment occurs because companies are finally hitting the architectural limits of their legacy systems. A project may appear viable during the first three months of development, but once the engineering team attempts to scale the solution to handle millions of daily unstructured data queries, latency and accuracy issues render the system unusable. For deal teams, this means that any ongoing technology project evaluated during due diligence has a statistically probable outcome of zero value realization.

Source: S&P Global Market Intelligence (2025)

Sector-Specific Vulnerabilities

The failure rates are not distributed evenly across the market. Highly regulated sectors characterized by intense data privacy requirements and fragmented legacy databases are experiencing catastrophic implementation failures. The RAND Corporation data shows that the Financial Services sector suffers an 82.1% failure rate, while Healthcare projects fail at a rate of 78.9% [3]. In financial services, the inability to clearly explain how an algorithm arrived at a specific credit or trading decision violates strict regulatory frameworks. If the model operates as a black box, compliance officers will refuse to authorize its deployment into production, forcing the abandonment of the entire initiative. In healthcare, patient data is typically distributed across dozens of incompatible electronic health record systems. Attempting to unify this highly sensitive, unstructured data to feed a predictive model frequently violates compliance standards or results in dangerous data hallucinations. Acquirers looking at health tech or fintech assets must apply a massive risk premium to any unproven technological roadmap.

Sunk Costs and EBITDA Drag

The financial implications of these abandoned projects represent a direct threat to post-close value creation. When a traditional software project fails, the losses are generally contained to developer hours and basic server hosting. When an advanced algorithmic project fails, the sunk costs are exponential. Companies spend heavily on specialized data scientists, premium external consulting fees, massive cloud compute bills for model training, and expensive third-party data licensing. The average sunk cost per abandoned enterprise project currently ranges from $4.2 million to $8.4 million. If a private equity sponsor acquires a platform company with three distinct initiatives currently in the development pipeline, they could easily be inheriting over $20 million in impending sunk costs. This represents a severe drag on post-close earnings before interest, taxes, depreciation, and amortization. Thorough technology due diligence must identify these impending failures before the transaction closes, allowing the deal team to either reduce the enterprise valuation or immediately defund the doomed projects on day one of the 100-day operational plan.

The True Prerequisites of AI Readiness: Architecture Over Algorithms

Network graph visualization of AI-ready enterprise data architecture

The consensus among leading technology due diligence practitioners is that true platform readiness is determined entirely by underlying infrastructure maturity, not by raw algorithmic capability. A target company boasting about its proprietary machine learning models is frequently hiding a fragile, unscalable digital foundation. High performing organizations understand that algorithmic capabilities change on a weekly basis, but a flexible, elastic data architecture serves as a permanent competitive advantage.

Securing the API Supply Chain

Modern software relies heavily on Application Programming Interfaces to connect disparate services. However, in the context of advanced algorithmic systems, these connections represent a complex and highly vulnerable API supply chain. An organization is no longer just connecting to a payment gateway; it is streaming massive volumes of proprietary corporate data through third-party endpoints managed by external vendors like OpenAI or Anthropic. If a target company does not have absolute visibility and control over this supply chain, it is completely exposed to data leakage, vendor lock in, and security breaches. True readiness requires transitioning to standardized, network-based agentic interactions. Implementing frameworks like the Model Context Protocol allows a company to manage how automated agents interact with secure corporate data. The Model Context Protocol ensures that external systems cannot arbitrarily pull unauthorized information, standardizing the security layer. During due diligence, if a company cannot map its entire API supply chain or lacks strict access control protocols for external language models, its architecture is fundamentally insecure.

Moving from Schema-First to Zero-ETL Data Lakehouses

The single greatest technical bottleneck preventing successful implementation is the rigid nature of traditional data architecture. Historically, companies built schema-first relational databases. If the business wanted to analyze a new type of data, data engineers had to spend weeks restructuring the entire database schema to accommodate it. Research indicates that modifying traditional schema-first architectures to support new algorithmic requirements takes an average of 4 to 8 weeks [8]. In a fast moving market, an 8-week delay for a single data pipeline is fatal to product velocity. True readiness requires abandoning these rigid structures in favor of Data Lakehouses and zero-ETL federation. A Data Lakehouse allows a company to store massive amounts of raw, unstructured data without forcing it into a predefined format. Zero-ETL refers to the ability to query and analyze this data instantly, without the laborious process of extracting, transforming, and loading it into separate analytical databases. This modern approach drastically reduces latency, allowing vector databases to feed Retrieval-Augmented Generation pipelines in real time. If a target company is still relying on batch processed, schema-first databases, any claims of real time predictive capabilities are mathematically impossible.

AI Observability and Structural Governance

Traditional software monitoring tools were designed to track deterministic outcomes. They tell an engineer if a server is online, if a database is overloaded, or if a specific page failed to load. These tools are completely useless for monitoring non-deterministic algorithms, which can generate entirely different responses to the exact same input based on subtle shifts in temperature parameters or background context. To operate safely, a mature technology stack must implement AI Observability. This is a specialized layer of monitoring designed to track semantic drift, token consumption, inference latency, and groundedness. Semantic drift occurs when a model slowly begins to change the nature of its outputs over time, potentially providing incorrect or biased information to customers. Groundedness monitoring detects hallucinations, ensuring that every claim generated by the system can be directly traced back to a verified corporate document. A platform lacking robust AI observability is flying blind. They will only discover that their system is hallucinating false information when a customer formally complains or threatens legal action. During an acquisition assessment, the absence of observability tooling is an immediate operational red flag.

Rewiring Workflows: The Missing Link in AI Operationalization

Process workflow diagram illustrating integrated AI and human-in-the-loop systems

The most sophisticated data architecture in the world will fail to generate financial returns if the human beings using the system refuse to adapt. Technology is not a substitute for process engineering. The vast majority of failed implementations occur because management teams treat complex algorithmic tools as standard software upgrades, entirely ignoring the profound human behavioral changes required to unlock their value.

The Fallacy of the AI 'Bolt-On'

A persistent trend in the software market is the attempt to leapfrog a poor digital foundation by simply bolting on a conversational interface to a legacy product. A vendor will take a 15-year-old, cumbersome software application, add a chatbot window to the user interface, and rebrand the product as an intelligent platform. This is a guaranteed implementation failure. Attempting to force modern, probabilistic workflows onto rigid, legacy software architecture creates immense user frustration. The underlying system cannot process the complex requests generated by the interface, resulting in massive latency and persistent error messages. According to industry data, only 21% of organizations have proactively redesigned their core workflows to accommodate these new tools [8]. The remaining 79% are merely bolting new software onto outdated human processes. When deal teams encounter a target company claiming high artificial intelligence adoption, they must ask the management team exactly how operational workflows have changed. If the underlying process remains identical, the technology is merely decorative.

Source: MIT Sloan / Project NANDA (2025)

Human-in-the-Loop Process Redesign

Achieving measurable value creation requires rigorous, human-in-the-loop process redesign. The technology is rarely capable of completing a complex business process entirely on its own from start to finish. It requires human oversight at critical junctures. Workflow redesign is the single most correlated factor with successful value creation in portfolio companies. Instead of expecting an algorithm to fully automate a financial audit, a successful operating partner restructures the audit team. They assign the algorithm the task of extracting unstructured data from thousands of contracts and flagging anomalies. The human auditor's job shifts from manual data entry to reviewing and resolving the anomalies flagged by the system. This requires rewriting job descriptions, changing performance metrics, and retraining staff. If a target company has not demonstrated a willingness to engage in this deep organizational change management, their technology investments will simply become expensive overhead, utilized by a fraction of the workforce while the rest revert to their legacy spreadsheets.

Mitigating Generative AI Pilot Failure

The failure to rewire these operational workflows is the primary driver behind the staggering pilot failure rates. Research from MIT Sloan and Project NANDA demonstrates that 95% of generative AI pilots fail to reach production or deliver any measurable profit and loss impact [7]. This is an incredibly precise indicator of market dysfunction. The friction is rarely technical; it is almost entirely operational. Pilots are typically built in a vacuum by small, isolated innovation teams. They do not consult the actual frontline workers who will be forced to use the tool. When the pilot is finally deployed, it inevitably violates undocumented business rules or disrupts established team dynamics. Mitigating this 95% failure rate requires deal teams to shift their diligence focus away from the engineering team and toward the operations team. Before underwriting any future growth predicated on operational automation, acquirers must verify that the target company possesses a proven methodology for integrating new tools into the daily habits of its workforce.

Valuation Premiums, 'AI Washing', and Diligence Pitfalls

Bar charts depicting valuation multiples and AI premiums in private markets

The intersection of artificial intelligence and private market deal flow has created a highly volatile valuation environment. Authentic technical capability commands a massive premium, while the desperate pursuit of those multiples has spawned widespread misrepresentation. Private equity deal teams must adjust their deal frameworks to accurately price genuine infrastructure assets while aggressively discounting heavily marketed vaporware.

M&A Multiples and Legitimate Premium Assets

When a target company possesses a proprietary data moat, a mature zero-ETL data lakehouse, and proven workflow redesign, the market rewards them substantially. Legitimate infrastructure assets saw valuation multiples surge by 40% in 2025, frequently reaching 12 to 15 times earnings before interest, taxes, depreciation, and amortization. In highly specialized sectors, the premiums are even more pronounced. Early stage FinTech companies with deeply embedded algorithmic underwriting capabilities commanded a staggering 242% valuation premium over their traditional peers. The Digital Health sector similarly captured a massive influx of capital, with AI startups commanding a 19% premium on average deal size, peaking at a 61% premium for late stage Series C funding rounds. These premiums are entirely justified if the asset fundamentally alters the gross margin profile of its industry. A healthcare platform capable of accurately parsing unstructured medical records without human intervention possesses a structural cost advantage that traditional competitors cannot replicate.

The Anatomy of 'AI Washing'

However, the immense financial incentive to secure these valuation premiums has led to an epidemic of 'AI washing', the deliberate practice of mischaracterizing or falsifying algorithmic capabilities to deceive investors. Software Equity Group notes that while 61% of buyers expect future targets to be meaningfully driven by advanced technology, only 26% of evaluated targets actually met that standard during detailed diligence in 2025 [5]. The most catastrophic example of this phenomenon is the collapse of Builder.ai. The company successfully secured a massive $1.5 billion valuation by claiming its proprietary software could automate the entire application development process. Rigorous post-investment audits eventually revealed the truth. The sophisticated software was largely an illusion. The company was secretly routing complex development tasks to human offshore labor disguised behind a digital interface [6]. This manual process severely constrained their margins and scalability, leading to a complete collapse of investor confidence. Spotting AI washing requires deal teams to meticulously correlate claimed automation capabilities with actual offshore headcount and operational expenditure.

Adjusting Deal Frameworks for the GenAI Era

Headline mergers and acquisitions data frequently distorts the true health of the market, requiring careful adjustment of standard deal frameworks. For example, aggregate data suggested that e-commerce startups leveraging advanced algorithms generated massive exit values in 2025. However, this data requires deep scrutiny. Excluding the massive, anomalous $14.9 billion initial public offering of Klarna, the rest of the e-commerce AI cohort actually underperformed their traditional peers in exit value, generating only $1.5 billion compared to $2 billion for non-AI platforms.

Deal teams cannot rely on generic technology consultants to uncover these discrepancies. Traditional quality assurance checks, uptime monitoring, and basic code reviews are entirely insufficient for evaluating non-deterministic systems, algorithmic bias, or the massive data latency limits hidden inside legacy architectures. True diligence must verify foundational digital maturity first. They must prove the existence of cloud elasticity, modular microservices, clean data governance, and secure API supply chains before assigning a single dollar of enterprise value to a predictive model.

Understanding these technical realities is what separates a successful acquisition from a massive write down. The gap between a visionary pitch deck and operational truth can only be bridged by specialized, architecture focused assessment. Altimi's Rapid Tech DD provides a definitive, risk-adjusted investment recommendation in just two to three weeks. By combining targeted code sampling, deep architectural assessment, and precise risk scoring, our senior advisors separate true structural maturity from superficial market noise, starting from €8,500. Book a call today to secure the operational truth behind your next software transaction.

Frequently Asked Questions

What is the primary difference between a traditional tech diligence and an AI readiness assessment in PE?

Traditional diligence focuses on code quality, technical debt, and system uptime. An AI readiness assessment PE framework prioritizes data architecture (like zero-ETL), the security of the API supply chain, AI observability mechanisms to track model hallucinations, and the foundational maturity of the digital infrastructure required to host non-deterministic algorithms.

How can deal teams quickly spot 'AI washing' in early-stage discussions?

Look for a disconnect between operational costs and claimed automation. If a company claims high AI-driven efficiency but has bloated offshore manual data-entry teams (as seen in the Builder.ai collapse), or if they lack robust AI observability tooling to monitor semantic drift, their 'AI' is likely superficial or vaporware.

Why are generative AI pilot failure rates so high (95%)?

Most companies treat generative AI as a standalone software 'bolt-on' rather than an architectural shift. Failure occurs because target companies do not redesign workflows to integrate human-in-the-loop processes, and their legacy, schema-first databases cannot feed real-time data to RAG (Retrieval-Augmented Generation) pipelines without immense latency.

Is it worth paying a 20-40% valuation premium for an AI-enabled SaaS asset?

It is only justified if the target possesses a proprietary data moat, a mature data lakehouse architecture, and proven workflow redesign that generates verifiable labor productivity or increased NRR. Premiums paid for companies merely wrapping third-party APIs (like OpenAI) without a proprietary data advantage will likely result in a poor ROI as those features become commoditized.

What is 'Zero-ETL' and why does it matter for our operating partners?

Zero-ETL (Extract, Transform, Load) refers to an architecture where data doesn't need to be manually moved and transformed between databases to be analyzed. For operating teams implementing AI, it means large language models can query real-time enterprise data instantly without waiting weeks for data engineering teams to prep the data pipelines, drastically reducing time-to-value.

How should we adjust our post-close 100-day plan for AI initiatives?

Before deploying algorithms, the 100-day plan must focus on foundational digital maturity: establishing cloud elasticity, migrating to data lakehouses, implementing strict AI governance/observability, and auditing third-party APIs. Only after this 'AI readiness' baseline is achieved should you begin applying AI to optimize workflows.

Ready to De-Risk Your Next Deal?

Altimi's Rapid Tech DD provides a clear investment recommendation in 2–3 weeks — combining code sampling, AI assessment, and risk scoring.

Book a 20-minute call