IntuitionLabs
Back to ArticlesBy Adrien Laurent

Enterprise AI Infrastructure: The Eli Lilly Case Study

Executive Summary

Eli Lilly’s recent announcements illustrate a fundamental shift in enterprise AI strategy: major companies are building their own AI infrastructure rather than relying solely on packaged “AI apps” or cloud chat services. In October 2025 Lilly revealed a partnership with NVIDIA to construct the world’s most powerful AI supercomputer dedicated to pharmaceuticals, along with an “AI factory” that manages every stage of model development ([1]) ([2]). Simultaneously, Lilly launched Lilly TuneLab, a federated AI/ML platform exposing Lilly’s proprietary drug‐discovery models (trained on over $1 billion of R&D data) to biotech partners in exchange for their data contributions ([3]) ([4]). These initiatives underscore the enterprise need for compute capacity, data access, and specialized platforms that generic tools (such as public chatbots) cannot provide.

In contrast to consumer AI (e.g. ChatGPT) – which offers convenient interfaces but limited domain knowledge, data security, and scale – Lilly’s approach builds custom infrastructure: thousands of GPUs networked with high-speed fabric, federated learning to safeguard data, and integrated software (e.g. NVIDIA’s BioNeMo, MONAI, Mission Control). This allows Lilly to train models on millions of experiments, apply AI at every step of drug discovery, and embed “digital assistants” into lab workflows ([5]) ([6]). Industry analysts note this is part of a broader enterprise trend: organizations see “cloud-based AI” (pay-per-call APIs) becoming cost-prohibitive at scale, commanding millions of dollars per month ([7]) ([8]). Surveys indicate many companies (often nearly half) are now developing AI in-house or considering on-premises solutions to gain control and reduce cost ([9]) ([10]).

This report explores why enterprises need full AI stacks to truly capitalize on AI, with Lilly as a case study. We examine the compute hardware, networking, data platforms, and AI software required – far beyond the scope of simple chat tools – and review other leading examples (e.g. Samsung’s “AI Megafactory,” DOE’s 110k‐GPU supercomputer, Microsoft’s multi‐datacenter AI cloud). Through data, expert analysis, and specific examples, we show that custom infrastructure is essential for mission-critical domains. While chat interfaces can boost productivity in narrow tasks, only a dedicated AI infrastructure (compute, data pipelines, internal ML platforms) can deliver the scale, customization and integration required in fields like drug discovery. The Lilly‐NVIDIA supercomputer exemplifies this principle: it is not an “app” but a foundation for next-generation R&D and operations ([1]) ([11]).

Introduction & Background

Artificial intelligence has become ubiquitous in consumer applications – from virtual assistants to language chatbots – but its role in mission-critical enterprise and scientific settings remains fundamentally different. Enterprises cannot rely solely on public chat tools or one-size-fits-all AI services, because their challenges demand specialized data, compliance, and scale. In regulated fields like pharmaceuticals, proprietary data (clinical trials, lab experiments, patents) must remain confidential, and AI models must be trained or fine-tuned on domain-specific knowledge that generic systems lack. As one Lilly executive notes, a 150-year legacy means the company’s “most powerful asset is decades of data.” Leveraging that data for AI requires custom models and hardware, not just off-the-shelf apps ([12]) ([13]).

Over recent years, many companies experimented with “Demo AI” pilots using cloud-based APIs (e.g. calling ChatGPT or Azure OpenAI). These pilots often yield quick wins, but analysts warn they seldom scale. A Forbes Council participant notes that simply licensing ChatGPT may accelerate individual tasks (e.g. drafting text), but “complex customer conversations” or B2B problems require training models on internal data and orchestrating AI with backend systems, not just a chat interface ([14]) ([8]). A 2026 analysis underscores this “enterprise AI trap”: pilot projects cheer initial results but then stall when there is no infrastructure to handle compliance, data integration, or volume ([15]). In short, enterprise AI success depends on robust infrastructure and platform readiness, not just flashy front-end demos.

Why “Chat Tools” Aren’t Enough

Chat-based AI tools (like public chatbots and generative AI APIs) have driven enthusiasm for AI in the enterprise. They provide intuitive interfaces and can boost productivity on discrete tasks (e.g. summarizing reports, drafting emails). However, experts emphasize their limitations for strategic, large-scale applications. Chatbots are typically built on general-purpose models with broad training data, and they do not inherently incorporate a company’s proprietary knowledge. They can hallucinate or give non-deterministic outputs without proper grounding. Crucially, security and compliance can be problematic: enterprise data sent to external services may violate privacy rules or IP protection.

In contrast, enterprise AI applications often require heavy customization and integration with internal systems. For example, a drug discovery AI needs to ingest lab instruments’ data, molecular simulation results, and patient outcomes from Lilly’s own research. A customer-service AI needs access to private CRM and historical support logs. These require pipelines that secure and preprocess data, and models that understand domain-specific terminology. Chat tools alone cannot handle this.

Industry commentators point out that smaller, domain-specific AI models often make more sense for companies than one massive generic LLM. Tech analyst Mike Vizard notes that most enterprises don’t need an all-purpose 175-billion-parameter model – “they need a model that ‘knows the ropes’ of [their field] and doesn’t blow up the cloud bill every time someone hits Enter” ([16]). Specialized models (or “agent” systems) can run faster, be fine-tuned more easily, and keep data on-site.In practice, companies are moving toward modular agent architectures, where multiple lightweight models handle subtasks (e.g. one agent for data extraction, another for reasoning, another for execution), orchestrated by a controller ([17]). This modular approach demands internal platforms and infrastructure: model training environments, APIs, data storage, and orchestration tools.

The need is acute because, while computation costs have plummeted (some studies cite a 280-fold drop in inference cost), usage is exploding ([7]). An analyst warns that enterprises now face “monthly AI bills approaching tens of millions of dollars,” making naïve cloud-API usage unsustainable ([7]). TechTarget reported that large enterprises using generative AI can incur > $1 million per month in cloud costs, driving them to in-house deployments ([8]). Indeed, surveys by Menlo Ventures and ESG find that around 45–47% of companies are developing generative AI capabilities in-house or consider on-premises options for AI ([9]) ([10]). This reflects a broader realization: at scale, enterprises must build and own their AI infrastructure – from data pipelines and compute hardware to MLOps platforms – to achieve ROI, rather than relying on siloed chat interfaces.

In summary, while chat tools can illustrate AI’s potential, they cannot replace the comprehensive stack needed for enterprise applications. As Lilly’s CIO Diogo Rau puts it, achieving a new standard of innovation demands “excellence not just in science but also in technology,” leveraging Lilly’s deep R&D data with purpose-built AI models ([12]). Lilly’s strategy exemplifies this principle by combining proprietary data, custom models, and vast compute resources into an “AI factory” – a far cry from a standalone chatbot.

Historical Context: Computing in Pharma

Enterprise computing has a long history of enabling complex science. Even decades ago, pharmaceutical companies invested in supercomputers for simulation and data analysis. For instance, Lilly’s leadership proudly remembers that the company was “the first adopter of the Cray-2 supercomputer among any commercial institution” in the early 1980s ([18]). That pioneering spirit is echoed today: Lilly calls its new system LillyPod, indicating it’s the first of its kind in pharma.

In recent years, other industries have similarly begun building in-house AI infrastructure. For example, NVIDIA and HPE recently unveiled turnkey “AI factory” solutions for enterprise, combining on-premises GPU clusters with managed software stacks ([19]) ([20]). In manufacturing, Samsung announced an “AI Megafactory” with more than 50,000 NVIDIA GPUs and digital-twin simulation to automate chip production ([21]) ([22]). Government labs like the U.S. Department of Energy are commissioning exascale AI supercomputers (110,000+ GPUs) to accelerate science in fields ranging from materials to healthcare ([23]). Even AI startups are building physical computing presence: Anthropic and OpenAI have leased record amounts of U.S. data-center capacity (over 7.4 gigawatts in one quarter) to meet AI demand ([24]).

These developments underscore a new reality: AI at industrial scale requires industrial-scale infrastructure. It is no longer sufficient to “rent” a bit of GPU-time via a cloud API when tasks involve petabytes of data, proprietary models, and national-scale research. Lilly’s initiative to construct its own supercomputer and collaborative platform is part of this broader trend of enterprise-grade AI engineering across sectors.

Lilly’s AI Infrastructure Strategy

Eli Lilly’s announcements in late 2025 and early 2026 provide a concrete case study of what enterprise-scale AI infrastructure looks like. The centerpiece is a partnership with NVIDIA to build “the most powerful AI supercomputer owned and operated by a pharmaceutical company” ([1]). This on-premises GPU cluster – dubbed LillyPod – is a first-of-its-kind NVIDIA DGX SuperPOD B300 system with 1,016 NVIDIA Blackwell GPUs ([25]) ([26]). All GPUs and storage are connected via NVIDIA’s high-speed Spectrum-X Ethernet fabric, meaning the entire cluster (compute, storage, networking) effectively shares one unified, ultra-fast network ([26]). This allows data to move instantly between thousands of cores during training or inference. NVIDIA’s Mission Control management software will orchestrate workloads across the full pod, monitoring performance and automating operations securely across all 1,000+ GPUs ([27]). The system is housed in Lilly’s own data center and – to align with Lilly’s sustainability goals – will run on 100% renewable electricity using the company’s existing chilled-water cooling systems ([28]).

Importantly, Lilly does not view this as “just another computer,” but as an AI factory – a term Lilly uses to emphasize that the system will manage the entire AI lifecycle. As the official press release states, the supercomputer “will power an AI factory, a specialized computing infrastructure that manages the entire AI lifecycle from data ingestion and training to fine-tuning and high-volume inference.” ([1]). In other words, rather than using external AI services or small proof-of-concept models, Lilly is building an integrated pipeline: raw R&D data goes in, models are trained or retrained on in-house and partner data, then those models are tuned and deployed at scale on Lilly’s own inference infrastructure. This closed-loop cycle allows rapid iteration: scientists can take results from experiments, quickly update models, and refine hypotheses in days instead of years.

Lilly’s executives stress that this infrastructure is needed to transform their science. CTO Diogo Rau emphasizes that Lilly combines “scientific expertise with NVIDIA’s leadership in AI computing” to “turn data into actionable insights” ([29]). He explains that by having this capacity, the company can “unlock deeper biological knowledge down to the individual cell and molecule levels.” For instance, one goal is to go beyond fine-tuning public AI models with widely available data: instead, Lilly can train “powerful models using our own internal data”, including the “lessons learned from millions of molecules that didn’t succeed” ([13]). This deep reservoir of internal pharmaceutical data – decades’ worth – becomes fuel for AI, enabling Lilly to search chemical space far more intelligently than any generic model could.

Chief AI Officer Thomas Fuchs puts it this way: “Lilly is shifting from using AI as a tool to embracing it as a scientific collaborator.” ([30]) In practice, the shift means that every step of drug R&D is augmented by AI agents and models. Lilly plans to develop “scientific AI agents” – essentially autonomous assistants that help researchers plan experiments, interpret results, and connect digital simulations to lab work ([31]). For example, generative models can propose novel molecular structures grounded in biochemical rules, or predict patient responses to therapies. AI will also be embedded in advanced medical imaging workflows: by using deep learning frameworks (e.g. NVIDIA MONAI), Lilly expects to analyze massive 3D imaging datasets “in days instead of months”, enabling more precise biomarker discovery ([32]).

Beyond discovery, Lilly foresees AI in manufacturing and supply chains. Digital twins of production lines (using platforms like NVIDIA Omniverse and RTX-powered servers) will allow Lilly to simulate and stress-test manufacturing processes virtually ([33]). This can improve yield, reduce downtime, and scale capacity to meet demand for blockbuster drugs. Robotic automation – already used for quality control in Lilly factories – can be further enhanced by AI for tasks like visual inspection or material transport. In short, Lilly expects its AI infrastructure to touch discovery, development, and delivery of medicines, effectively accelerating the entire pipeline.

A key element of Lilly’s strategy is collaboration via a controlled ecosystem. Many of the AI models trained on LillyPod will be made available through Lilly TuneLab, an AI platform unveiled in September 2025 ([34]). TuneLab lets external biotech firms access Lilly’s proprietary drug-discovery models (built on Lilly’s datasets) in a privacy-preserving manner. In return, partner biotechs provide their own data back to Lilly, improving the models for everyone. The initial TuneLab release includes models based on “proprietary data obtained at a cost of over $1 billion,” making it one of the largest private datasets (hundreds of thousands of molecules) ever used in an AI system available to smaller companies ([3]) ([35]). This federated approach (built on NVIDIA FLARE technology) means Lilly can share intelligence without sharing raw data ([4]) ([35]). As Nisha Nanda of Lilly’s Catalyze360 program notes, TuneLab ‘compresses decades of learning into instantly accessible intelligence’ for biotechs ([36]).

In summary, Lilly’s approach goes far beyond purchasing a chatbot or using third-party AI apps. It involves building intensive compute resources, constructing data pipelines, and creating internal/external platforms. The goal is to make Lilly itself an “AI-native” company where intelligence is woven into every layer of R&D and operations ([37]) ([6]). This enterprise strategy – massive on-prem compute plus rich data and software – is what Lilly believes will supercharge medicine discovery and delivery faster than conventional methods ([1]) ([30]).

Key Components of Lilly’s AI Infrastructure

Lilly’s announcements highlight several critical components of enterprise AI infrastructure. These are summarized below, with references to Lilly’s statements and industry context:

  • High-Performance Compute (HPC) Hardware: LillyPod is built on NVIDIA DGX SuperPOD architecture. A SuperPOD is a turnkey cluster combining multiple DGX systems with high-speed interconnect. Lilly’s will be the first to use the DGX B300 systems, specifically the new NVIDIA Blackwell GPUs. In total, over 1,000 Blackwell Ultra GPUs will be connected on a unified spectrum fabric ([2]) ([26]). Such scale means Lilly can train models on extremely large datasets (e.g. millions of chemical simulations) in parallel. For context, NVIDIA’s announcement of DOE systems (Solstice/Equinox) shows that 110,000 GPUs yield ~2,200 exaflops of performance ([23]) – Lilly’s 1,000-GPU cluster, while smaller, is similarly unprecedented in pharma and will deliver multi-exaflop AI performance. These GPUs will be located on Lilly’s campus, enabling low-latency access to lab data and easy integration with Lilly’s IT environment. The cluster will be managed for energy efficiency and sustainability: it uses chilled water cooling and 100% renewable power per Lilly’s carbon goals ([28]).

  • Networking and Orchestration: To harness so many GPUs, Lilly employs advanced networking and management software. The entire SuperPOD uses NVIDIA’s Spectrum-X Ethernet, ensuring a high-speed, unified interconnect among all compute and storage nodes. This means data flows seamlessly between GPUs during distributed training or inference ([26]). Orchestration is handled by NVIDIA’s Mission Control software, which Lilly will use to monitor and manage the HPC cluster. Mission Control provides a central dashboard for scheduling jobs, automating workflows, and ensuring security across the multi-node system ([27]). In effect, Mission Control treats the thousands of GPUs as a single compute continuum optimized for AI workloads, a critical capability for enterprise-scale usage. (Analogously, Microsoft’s Azure Air system interconnects datacenters to act as one giant supercomputer ([38]).)

  • Software Stack and AI Platforms: Lilly’s AI factory comes with a full software ecosystem. The NVIDIA “full-stack” includes both low-level libraries (like cuDNN) and high-level toolkits. Notably, Lilly plans to use platforms like NVIDIA BioNeMo to train generative models for molecules. BioNeMo allows combining Lilly’s experi­mental datasets with public scientific knowledge to propose novel antibodies, proteins, and other biotherapeutics ([39]). For medical imaging, Lilly will leverage the open-source MONAI framework, which is optimized for healthcare AI; this can reduce image analysis from months to days ([32]). These are examples of how an enterprise AI infrastructure integrates specialized ML frameworks tailored to its domain. Additionally, Lilly’s TuneLab itself is an AI/ML platform built on federated learning (using NVIDIA FLARE) so that multiple organizations can use and improve shared models without exposing data ([4]). This kind of internally-managed AI platform is a key part of infrastructure, beyond mere hardware: it governs data access, model deployment, and cross-company collaboration.

  • Data Pipelines and Proprietary Data: A cornerstone of Lilly’s plan is its vast, proprietary datasets from decades of drug R&D. The HPC will ingest pre-processed data from discovery, preclinical, and clinical experiments (e.g. assay results, chemical libraries, safety studies). Lilly explicitly notes it has “decades of data” at its disposal ([12]). For example, TuneLab’s models were trained on Lilly’s full “drug disposition, safety, and preclinical” datasets covering hundreds of thousands of unique molecules ([35]). Building an AI infrastructure means creating automated pipelines to curate, label, and feed these data into model training. Lilly’s system must also integrate new data continuously: as scientists run new assays or adopt the digital twins, the AI pipeline will capture those results to retrain models. We see a sign of this continuous learning vision in the LillyPod announcement: it will enable “rapid learning and iteration”, letting scientists train models on “millions of experiments” to dramatically widen the scope of discovery ([40]). In short, enterprise AI infrastructure must ingest both existing and emerging data across the company, far beyond what a standalone chatbot could access.

  • Federated and Collaborative Learning: To extend infrastructure beyond Lilly’s walls while respecting privacy, Lilly’s TuneLab platform employs federated learning. This means biotechs can query Lilly’s models without giving up their data, and Lilly can incorporate partner insights by training on encrypted updates. The press release explains that curated biotech data enhances the models “for the benefit of others in the ecosystem and ultimately patients” ([41]). Such federated platforms are becoming an important part of enterprise AI infrastructure, allowing cross-organization learning. As outlined in [5], TuneLab’s design (federated AI/ML platform) enables Lilly to “expand access to advanced discovery tools” without compromising any party’s data ([42]). This collaborative model is itself an infrastructure component – akin to how some large tech consortiums share models, but with secure data governance built-in.

  • Governance, Security, and Compliance: Enterprise AI cannot ignore regulation, especially in healthcare. Lilly, as a pharmaceutical company, must meet strict data privacy and quality standards, which are central to its infrastructure design. For instance, TuneLab’s federated approach ensures no raw data leaves a company’s control, addressing IP concerns ([4]) ([35]). Internally, Lilly will use encryption, SOC 2 controls, and identity management on its platforms (similar to enterprise-grade ChatGPT offerings ([43])). The infrastructure must also comply with FDA and privacy regulations. In practice, this means auditing, drift detection, and version control become part of the pipeline. Indeed, NVIDIA’s appeal to Lilly emphasized “secure, scalable, purpose-built” systems for healthcare ([2]). The setup effectively gives Lilly the same enterprise control (data lineage, access policies, audit logs) that CIOs demand, something a consumer chat tool cannot provide.

In summary, Lilly’s AI infrastructure spans hardware (GPUs, network), software (tools, libraries, orchestration), data (pipelines, governance), and collaboration platforms. Each component matches a critical enterprise need. For example, unlike a simple AI service, Lilly’s HPC can handle high-volume training of large foundation models, then immediately switch to high-throughput inference for internal apps, all within its own secure network ([1]) ([2]). This end-to-end integration is precisely what enterprise experts say is lacking in “lighter” AI solutions ([15]) ([16]). Lilly’s program is an archetype of moving from reading science papers to doing AI-enabled science at industrial scale.

Case Studies and Industry Comparisons

Lilly’s approach is one of several recent examples showing how major enterprises are building AI infrastructures. Below we compare Lilly with other high-profile AI initiatives:

OrganizationAI Infrastructure InitiativeComponents & ScalePrimary Goal
Eli Lilly (Pharma)Formation of the Lilly AI Factory (announced Oct 2025) and Lilly TuneLab (Sept 2025) ([1]) ([34]).On-prem NVIDIA DGX SuperPOD B300: >1,000 Blackwell GPUs with unified Spectrum-X network ([26]). Managed by Nvidia Mission Control. Dedicated HPC in Lilly data center (100% renewable power). TuneLab platform with federated learning.Accelerate drug discovery and development with AI: train models on millions of internal experiments, share models with biotech partners via TuneLab ([5]) ([44]).
U.S. DOE / Argonne (Science)Solstice & Equinox AI Supercomputers (announced Oct 2025) ([23]).Solstice: 100,000 NVIDIA Blackwell GPUs; Equinox: 10,000 GPUs; high-speed interconnect. Combined ~2,200 exaflops ([23]). Built by NVIDIA and Oracle.Enable “agentic AI” for scientific discovery across research areas (healthcare, energy, materials) ([23]) ([45]).
Samsung (Tech/Manufacturing)AI Megafactory (announced late 2025) ([21]) ([22]).>50,000 NVIDIA GPUs integrated across semiconductor fab operations. Uses NVIDIA Omniverse for digital twins and robotics.Embed AI across chipmaking, robotics, and mobile manufacturing. AI-managed processes (e.g. lithography, quality control) to speed up production and optimize yield ([21]).
Microsoft (Cloud)Fairwater AI Superfactory (announced 2025-2026) ([38]).Networked datacenters in Wisconsin and Atlanta (and others), linked by ultra-fast fiber. “Hundreds of thousands” of GPUs, exabytes of storage, millions of CPUs across sites ([38]).Pool distributed resources as a unified platform to train large AI models (e.g. OpenAI’s models, Copilot). Reduce training time from months to weeks through parallelism ([38]).
Anthropic (AI Startup)Claude Data Centers (announced Nov 2025) ([46]).New AI datacenters planned in Texas and New York; $50 billion investment announced; working with Fluidstack. Large capacity for Claude chatbot (Cloud TPU/GPUs not specified). Industry-wide datacenter leasing surged by 7.4 GW in Q3 2025 for AI ([24]).Scale up computing to meet enterprise demand for Anthropic’s AI services. The investment reflects need for owned infrastructure as API usage grows ([46]).
HPE & NVIDIA (Enterprise)NVIDIA AI Computing by HPE turnkey solution (announced Nov 2025) ([19]) ([20]).Private cloud for AI: integrated NVIDIA AI stack on HPE hardware. Supports inference, fine-tuning, retrieval-augmented generation on private data ([20]). Features HPE GreenLake service, sustainability focus.Give enterprises a plug-and-play AI platform to run generative AI on-premises. Emphasizes data privacy/control while accelerating use-case deployment ([19]) ([20]).
(Emerging) Other SectorsFinancial, Retail, Manufacturing (2023–26 trends): Many firms are shifting AI in-house.Surveys (Menlo, ESG): ~47% of large enterprises develop GenAI in-house ([9]). Spending on AI servers up. Tech companies (Oracle, Dell, HPE) shipping NVIDIA-powered servers and appliances for AI.Cost-control and customization: reduce cloud spend (~$1M/month savings) and meet data regulations. Enable AI for internal workflows (finance, supply chain, etc.) as strategic asset ([8]) ([7]).

This table illustrates that Lilly’s initiative is part of a wider movement: leading organizations are investing in bespoke AI infrastructure tailored to their domain. For instance, while Lilly focuses on molecular data and lab automation, Samsung focuses on manufacturing processes, and Microsoft on general AI model training. In each case, the solution is not an off-the-shelf AI app, but rather a computational platform combining hardware, software, and data pipelines.

Key Takeaway: Chatbots and generic AI services have democratized AI access but cannot solve the unique challenges of industries like pharma. Both industry leaders (Samsung, Microsoft, DOE) and specialist players (Lilly, Anthropic) are proving that real AI value comes from owning the stack – from silicon up to specialized algorithms – not from surface-level tools. Enterprises that understand this are building “AI factories” (a turnkey concept also popularized by NVIDIA and HPE ([19])) to maintain competitive advantage and control over their data.

Data and Collaboration in Enterprise AI

A recurring theme in Lilly’s planning (and in enterprise AI at large) is that data is as important as compute. Lilly emphasizes the value of its proprietary datasets, and plans to harness decades of R&D knowledge. As Diogo Rau stated, Lilly has “decades of data” from its 150-year history, and the goal is to use “purpose-built models” on that data to accelerate discovery ([12]). The initial set of TuneLab models alone was built on Lilly data whose acquisition cost exceeded $1 billion ([3]).

Building AI infrastructure means establishing data pipelines that can feed this precious data into AI models continuously. In practice, Lilly will need to aggregate data from chemistry, biology, preclinical assays, learned synthetic routes, clinical trials, etc., into a coherent “AI-ready” format. This requires ETL processes, metadata tagging, quality control, and a fast storage backend. It also includes data from digital twins and sensors in manufacturing. For example, Lilly’s supercomputer will support digital-twin simulations of supply chains and production lines, generating new data on process dynamics. Integrating this with R&D data (e.g. correlating synthesis failures in the lab with manufacturing bottlenecks) could reveal new insights. Rich, cross-domain data is a hallmark of “AI infrastructure” as opposed to siloed apps.

Lilly’s strategy for data is twofold: internal use and ecosystem sharing. Internally, the company will use all available data to train its models. For example, the TuneLab is backed by Lilly’s complete preclinical dataset covering “hundreds of thousands of unique molecules” ([35]). In effect, Lilly compresses its historical research into AI models: these models implicitly encode lessons from failed compounds, side-effect profiles, and large screening campaigns. By continually retraining on new experiments, the models become ever more attuned to Lilly’s specific drug-discovery space.

Externally, Lilly’s approach is collaborative. Recognizing that innovation thrives in an ecosystem, TuneLab invites biotech partners to contribute data. The PR notes that partner contributions “fuel continuous improvement” of the shared AI models ([41]). In this federated learning setup, Lilly infrastructure enables smaller companies to benefit from Lilly’s AI while giving back their learnings. This symbiotic data collaboration is possible only because Lilly built a dedicated platform (with privacy-first design); a generic AI product would not so easily allow such controlled data combinatorics.

Industry-wide, this pattern is emerging: organizations are finding new ways to pool and reuse data for AI. For example, in healthcare some consortia use federated networks to train models on patient data across hospitals without exchanging records. Lilly’s TuneLab is an analogous model for pharma R&D. This shows how data infrastructure in enterprise AI includes not just in-house databases but also protocols and platforms for data collaboration. Without its own infrastructure, Lilly could not offer federated training at scale.

Finally, enterprise data infrastructure must deal with privacy, IP, and compliance. Lilly’s solution – federated learning – addresses this head-on: partner companies run training locally and only model updates move to Lilly’s servers. According to NVIDIA’s description, “data stays private and separate” for each user ([4]). Thus, the AI infrastructure is built with data sovereignty in mind. In industries like pharmaceuticals, this is crucial: sensitive data can never be carelessly placed on a public cloud. Lilly explicitly highlights that TuneLab operates on a third-party host “employ [ing] federated learning, a privacy-preserving approach” ([41]), underscoring that data governance was a core design goal of their AI platform.

In summary, Lilly’s example shows that enterprise AI infrastructure must provide comprehensive data management: massive storage and processing capacity on-premises, pipelines to collect lab/manufacturing data, and secure collaboration mechanisms for sharing models. In contrast, a tool like ChatGPT offers none of this: it runs on OpenAI’s data, cannot ingest private research unless separately engineered, and provides no way to integrate new experiment data. The Lilly case demonstrates that the true infrastructure includes every link in the chain from raw data to model to end-user application.

Enterprise AI Platforms and Tools

Beyond hardware and raw data, Lilly’s plan includes specialized software and platforms – effectively treating AI as an integrated system. Several concrete examples emerge from Lilly’s announcements:

  • Lilly TuneLab Platform: This AI/ML platform (launched Sept 2025 ([34])) is perhaps the clearest example of an in-house tool. TuneLab is hosted off-site (to handle scale) but tightly controlled by Lilly. It offers a catalog of Lilly-trained models for various discovery tasks (chemistry predictions, ADME/toxicity, etc.) and, crucially, it implements federated AI/ML. Lilly describes TuneLab as “the first drug discovery platform to offer Lilly models and NVIDIA Clara [open models] in a federated environment” ([47]). This means biotech firms can run Lilly’s models against their own data without sending that data to Lilly. Internally, Lilly must maintain the orchestration (e.g. distributing models via Nvidia’s FLARE framework and aggregating model updates securely). TuneLab also integrates contributions from partners: as each biotech uses the models, Lilly aggregates anonymized learnings to refine them. Thus, TuneLab is a custom platform built on Lilly infrastructure – a far cry from a generic analytics tool.

  • AI Lifecycle Orchestration: The concept of an “AI factory” implies handling end-to-end workflows. Lilly explicitly says the new system will manage fine-tuning and high-volume inference in addition to training ([1]). In practice, this means Lilly is investing in MLOps tools: software for continuous integration/continuous deployment of models, experiment tracking, monitoring performance, and versioning. NVIDIA’s Mission Control provides part of this, but Lilly likely also needs platforms akin to Databricks or Kubernetes for ML. They will have pipelines that take new lab data, retrain models, validate them, and push updates to researcher interfaces. For example, if a model suggests a novel molecule, Lilly researchers can synthesize it, test it, and the results would flow back into the model retraining cycle. This closed-loop demands a robust software pipeline – something that out-of-the-box chat systems simply cannot replicate. It is akin to building a mini “open-source Lab-to-Model pipeline” across the enterprise network.

  • Integration with Scientific Tools: Lilly’s AI infrastructure is not an island; it links to existing scientific software. For instance, Lilly will integrate AI with its digital twin initiatives for manufacturing, using NVIDIA Omniverse and RTX servers ([33]). Omniverse allows detailed simulation of Lilly’s production lines and supply chains, which can then feed into AI-driven optimization. Likewise, Lilly can use AI for real-time monitoring: sensor data from bioreactors or lab instruments could be analyzed by models trained on historical process data. Integrations with lab information management systems (LIMS), electronic lab notebooks, and clinical data systems are all part of making this infrastructure work. This all-encompassing approach is necessary for enterprise – it’s not just building a new app, but weaving AI into the digital fabric of the company.

  • Productivity and Collaboration Tools: Internally, Lilly will likely deploy AI-enhanced applications for employees across functions. The press mentions using large language models to accelerate clinical writing and documentation, and to power internal GPT-based agents for workflows ([48]). For example, one use-case is speedier medical writing: automatically drafting sections of regulatory documents or summarizing research findings. Lilly scientists may get custom LLM-based assistants trained on Lilly’s own medical and scientific corpus, again requiring fine-tuning on Lilly data. Traditional productivity suites may incorporate these AI copilot features on-premises. None of this would be feasible with a simple "ChatGPT subscription" – it requires a secure, bespoke platform and middleware to handle legal and IP concerns.

  • Governance and Monitoring Tools: Finally, enterprise AI infrastructure includes oversight software. Lilly’s CIO stresses security and governance: data is encrypted in transit and at rest, and no prompts train public OpenAI models by default (akin to ChatGPT Enterprise features) ([49]). They will apply user roles, audit logs, and bias/hallucination checks to their models. NVIDIA’s blog emphasizes that Lilly’s platform is “secure, scalable, purpose-built” for compliance-heavy industries ([2]). Building built-in safeguards – error monitoring, drift detection, ethical filters – is part of the platform story. In pharma, regulators would expect documentation and validation of any AI used in drug trials or patient care, so Lilly is likely investing in explainability tools and compliance workflows. All this governance is an “infrastructure” component, again not found in a standalone chatbot.

In essence, Lilly’s undertaking involves building multiple new layers of software, from bespoke model repositories (TuneLab) to data integration pipelines to AI-enhanced user applications. Each layer is tightly integrated with Lilly’s data, objectives, and compliance needs. Analyst commentary on enterprise AI often echoes this: simple plug-and-play solutions give way to heavy engineering. For example, one analysis of IT trends predicts that enterprises will “rebuild everything” – from network architecture to talent structure – around AI ([50]). Lilly’s investments mirror that thesis: they are effectively architecting the organization around this AI platform, not just running a pilot project.

Data Analysis and Expert Perspectives

Multiple industry experts and studies corroborate Lilly’s strategy. Several salient points arise from the research:

  • ROI and Scale of Investments: Building such infrastructure is expensive, but the payoff can be huge. Lilly itself has signaled a willingness to invest (the Let’s build a $1B dataset is one clue). IBM, Microsoft, and Dell note that “enterprise control” over AI – including running on-prem – is a major driver in purchasing decisions ([8]) ([19]). A LinkedIn publication by technology analysts emphasizes that while chat tools tutdere adoption, “the cracks show” at scale if underlying infrastructure is weak, potentially leading to wasted investment ([15]). When companies moved AI from pilots to real products, unexpected costs appeared: e.g. a consumer brand might find their “pay-per-use” cloud bills skyrocketing to millions per month – making a capital investment in dedicated servers economically sensible ([7]) ([8]). In other words, the math often favors owning hardware for steady, heavy workloads.

  • Talent and Cultural Challenges: Analysts stress that technology is only half the battle – the other half is people and process. The Foreign Affairs Forum piece on “infrastructure reckoning” predicts that enterprises will have to retrain roughly 30% of their roles towards AI skills ([50]). Techstrong’s analysis similarly warns of a talent gap: “skilled practitioners are scarce”, so companies must commit to upskilling rather than just buying tools ([51]). Lilly’s move implicitly acknowledges this: it is hiring AI scientists and engineers (the presentation at NVIDIA GTC, others) and involving researchers in iterative AI work. To truly leverage an AI factory, Lilly will need biologists and chemists who can use AI outputs, data scientists to maintain models, and IT staff to run the infrastructure. This holistic development – building not just an app but an ecosystem of skilled users and engineers – underpins success. A survey cited by Techtarget shows that although almost half of companies develop AI in-house ([9]), only a subset have built the necessary pipelines and teams. Lilly’s approach tries to do both.

  • Empirical Gains: While Lilly’s announcements focus on future capability, some evidence from other projects illustrates potential gains. Microsoft’s Materials Science collaboration, for instance, demonstrates that AI+HPC can compress years of lab work into weeks. In that DOE partnership, researchers used Azure Quantum and AI to sift through 32 million candidate materials, isolating 18 promising compounds for batteries in just 80 hours ([52]). This kind of speedup – finding useful outputs in days instead of decades – typifies what Lilly hopes to achieve in drug discovery. If LillyPod enables even a fraction of that acceleration for pharmaceutical R&D, the investment pays dividends in faster drug candidates and reduced costs per discovery. Moreover, industry data suggest AI adopters see outsized improvements: Bhardwaj’s analysis claims AI-driven companies release products 400% faster than peers and achieve ~15% higher margins when blending human expertise with AI ([53]). While these numbers are from broader AI studies, they support the idea that enterprise AI infrastructure can transform performance, not just automate small tasks.

  • Multi-stakeholder Benefits: Lilly explicitly frames its platform as benefiting a broad ecosystem. By sharing models via TuneLab, Lilly creates a collective learning environment: each participant’s data makes the shared AI models better. Analysts note that such federated platforms could become standard in pharma and biotech, ensuring that innovation isn’t locked in silos. The implication for enterprise AI is that the infrastructure not only powers patient company’s labs, but also generates intellectual property and partnerships. From a business perspective, Lilly can recoup its AI R&D costs partly by offering AI-as-a-service to partners, a model that bigger tech companies (e.g. AWS, Oracle, NVIDIA) have pursued. Lilly’s CTO mentions this in the context of Lilly Catalyze360’s initiatives ([54]). This extension beyond an App toward a platform with external stakeholders further shows why a robust infrastructure – data, compute, trust frameworks – is necessary.

  • Long-term Industry Shift: Finally, thought leaders see Lilly’s strategy as indicative of a broader era where AI is the new lab instrument. Huang (NVIDIA’s CEO) famously predicted that drug research would move away from traditional wet labs towards AI-driven platforms ([55]). He specifically cited Lilly’s upcoming supercomputer as evidence: three years ago “all of their R&D budget was probably wet labs,” but now they’re pivoting to AI ([11]). This endorsement from an outside leader reinforces that Lilly’s approach is not a one-off but part of a structural change. In effect, companies in science-intensive industries will increasingly allocate R&D dollars to computing estates and AI engineers, just as manufacturing companies invest in robotics and automation.

Implications and Future Directions

Lilly’s investments foreshadow several implications for enterprise AI:

  • Competitive Edge in Pharma: By building an AI infrastructure, Lilly aims to shorten drug cycles and bring therapies to market faster. If successful, this could translate to blockbuster drugs reaching patients years earlier than competitors. Other pharma companies are likely to follow suit or partner with AI-specialists to avoid falling behind. The Carnegie Mellon Regression Model warns that “60% of new economic value by 2030 will accrue to enterprises investing heavily in AI now” ([53]). For Lilly, the gamble is that the upfront cost of GPUs and data platforms will pay off in innovation leadership. Early results (e.g. new targets found, faster trial designs) will determine whether this becomes a standard model across the industry.

  • Organizational Transformation: Lilly is not just buying hardware; it is restructuring how it does research. Expect Lilly to develop new roles (AI scientists, data engineers, ML ops managers) and possibly even change team structures (e.g. “AI centers of excellence”). The Foreign Affairs analysis suggests companies will evolve AI-centric governance and federated teams ([56]). Lilly’s case encourages enterprises to rethink job designs – for example, pairing biologists with AI specialists, or funding internal tools innovation rather than only external vendors.

  • Vendor Ecosystems: Lilly’s partnership with NVIDIA and the involvement of multiple tech providers indicates the rise of AI infrastructure vendors. Beyond NVIDIA, companies like HPE, Cisco, Dell, et al. will offer “AI stack” solutions (often co-developed with partners). Lilly’s strategic choice of partners suggests it wants cutting-edge tech, but also robust enterprise support. We may see more pharma/biotech co-innovation labs (as Lilly has one with NV) and specialized services for life sciences AI. Vendors will compete to provide the secure, compliant AI platforms that Lilly needs.

  • Data as Strategic Asset: Lilly’s emphasis on decades of research data highlights a broader point: companies with rich data histories have a huge advantage in this new AI race. This could place greater value on data governance and curation roles. The fact that Lilly shares some data only through AI models (not raw data) may spark new norms of how to monetize or share proprietary knowledge. Other industries might similarly build AI platforms to leverage their transactional or sensor data.

  • Ethical and Regulatory Considerations: The move to enterprise AI infrastructure also raises regulatory questions. If AI influences drug design or patient treatment, oversight bodies (FDA, EMA, etc.) will scrutinize the models. Lilly will need to align its infrastructure with guidelines for AI in healthcare. For other enterprises, Lilly’s example will underline the need to embed compliance checks and ethical review into the AI pipeline from day one.

  • Beyond Chat – Toward Autonomy: Looking forward, companies like Lilly may extend this infrastructure to agentic AI – systems that autonomously perform research loops. NVIDIA has already hinted that such supercomputers could enable “agentic scientists” ([57]) that propose hypotheses and experiments. While dangerous to anthropomorphize, the vision is clear: Lilly is laying the groundwork for AI that can actively suggest new science. This is a far cry from passive chat tools, and it requires guardrails (as even Techstrong warns, “clear guardrails” are needed to avoid catastrophic errors) ([58]). Enterprise infrastructure will need to include safety layers as AI agents become more independent.

In conclusion, Lilly’s AI infrastructure initiative exemplifies the shift from viewing AI as a “nice help” to treating it as core infrastructure – as fundamental to the company as the labs and factories. This supports the adage that “software is eating the world”; here, AI-powered software is reshaping the world of drug discovery. As Lilly’s case shows, success demands heavy investment in tech and data. Chat tools and cloud APIs have their place for quick wins, but for a true enterprise transformation AI must be built from the ground up, not simply bought off the shelf ([1]) ([7]).

Conclusion

The case of Eli Lilly’s AI investments highlights a crucial lesson: Enterprise AI requires infrastructure, not just standalone apps. Lilly’s building of a 1,000+ GPU supercomputer and creation of a federated AI platform demonstrate how deeply computing, data, and software platforms must be integrated to solve real-world problems at scale. Unlike ad hoc chatbots, Lilly’s “AI factory” is designed as a continuous, secure environment spanning R&D, manufacturing, and partnership networks. This infrastructure will allow Lilly to train models on its unique domain data, iterate rapidly on hypotheses, and apply AI in areas from molecular design to supply-chain optimization.

Enterprise decision-makers should heed Lilly’s example. While tools like ChatGPT offer impressive immediate benefits, they cannot replace the tailored pipelines and compute needed for mission-critical applications. As industry insights confirm, companies rapidly find that AI at scale means building or co-owning hardware, software and data capabilities – from GPU clusters to federated learning platforms ([1]) ([16]). Otherwise, they will hit limits of cost, privacy, or relevance. Lilly’s bold step into hybrid AI infrastructure (on-prem plus partner networks) underscores why simply subscribing to a chatbot is not enough for tasks like drug discovery.

Looking ahead, we anticipate more enterprises following suit: developing private AI clouds, linking centers with fiber (as Microsoft is doing ([38])), and co-investing with tech partners. This shift has profound implications: IT budgets will increasingly include AI hardware and services; talent development will focus on AI engineering; and cross-industry collaborations (like Lilly with biotechs) will become common as data-sharing platforms. The future of enterprise AI is not a chatbot on someone’s shelf, but a factory – a robust, comprehensive system that learns and evolves. Eli Lilly’s initiative is a vivid demonstration of that future, and a roadmap for why infrastructure matters as much as algorithms in unlocking AI’s potential.

References:

  • Lilly press releases and web articles ([1]) ([59]) ([3]) ([35])
  • Nvidia blog posts and press releases ([2]) ([23]) ([19])
  • FierceBiotech analysis of Lilly/NVIDIA partnership ([60]) ([30])
  • Axios and AP News coverage of Nvidia and Anthropic statements ([55]) ([46])
  • TechTarget and TechRadar industry reports ([38]) ([8])
  • Microsoft and NVIDIA publications on AI infrastructure use cases ([52]) ([39])
  • Expert commentary on enterprise AI trends ([16]) ([7]).
External Sources (60)
Adrien Laurent

Need Expert Guidance on This Topic?

Let's discuss how IntuitionLabs can help you navigate the challenges covered in this article.

I'm Adrien Laurent, Founder & CEO of IntuitionLabs. With 25+ years of experience in enterprise software development, I specialize in creating custom AI solutions for the pharmaceutical and life science industries.

DISCLAIMER

The information contained in this document is provided for educational and informational purposes only. We make no representations or warranties of any kind, express or implied, about the completeness, accuracy, reliability, suitability, or availability of the information contained herein. Any reliance you place on such information is strictly at your own risk. In no event will IntuitionLabs.ai or its representatives be liable for any loss or damage including without limitation, indirect or consequential loss or damage, or any loss or damage whatsoever arising from the use of information presented in this document. This document may contain content generated with the assistance of artificial intelligence technologies. AI-generated content may contain errors, omissions, or inaccuracies. Readers are advised to independently verify any critical information before acting upon it. All product names, logos, brands, trademarks, and registered trademarks mentioned in this document are the property of their respective owners. All company, product, and service names used in this document are for identification purposes only. Use of these names, logos, trademarks, and brands does not imply endorsement by the respective trademark holders. IntuitionLabs.ai is an AI software development company specializing in helping life-science companies implement and leverage artificial intelligence solutions. Founded in 2023 by Adrien Laurent and based in San Jose, California. This document does not constitute professional or legal advice. For specific guidance related to your business needs, please consult with appropriate qualified professionals.

Related Articles

Need help with AI?

© 2026 IntuitionLabs. All rights reserved.