A foundation model is a large AI model pre-trained on massive amounts of data that can be adapted for a wide range of tasks. GPT-4, Claude, Gemini, Llama — these are all foundation models. They’re called “foundation” because they serve as the base layer that businesses build on top of, rather than training a model from scratch.

This matters for your business because foundation models fundamentally changed the economics of AI. Five years ago, building an AI application meant collecting training data, hiring ML engineers, and training a custom model — a six-figure project minimum. Today, you can build on top of a foundation model through an API call. Understanding this shift helps you make better technology and budget decisions.

How They Work

Foundation models are trained on enormous datasets — billions of web pages, books, code repositories, and other text — using vast computational resources. This training gives them broad capabilities: they can understand language, generate text, reason through problems, analyze documents, write code, and more.

The key insight is generality. A foundation model isn’t trained for one specific task. It’s trained to understand and generate language broadly, which means it can be adapted for almost any language-based task: customer support, document analysis, code generation, content creation, data extraction, and hundreds of other applications.

You adapt a foundation model to your needs through three mechanisms:

Prompting. You give the model instructions and context in natural language. “You are a customer support agent for a medical device company. Answer questions using the following product documentation…” This is the simplest and cheapest approach.

RAG (Retrieval-Augmented Generation). You connect the model to your data at query time, so it can reference your documents when generating responses. This gives the model access to current, proprietary information without modifying the model itself.

Fine-tuning. You train the model further on your specific data to change its default behavior — its style, format, or domain expertise. This is the most expensive adaptation method and only necessary when prompting and RAG aren’t sufficient.

When a Business Would Use One

You’re probably already using foundation models — through AI coding tools, chatbots, document processing services, or search features. The more strategic question is when to build directly on a foundation model versus buying a product built on one.

Build on a foundation model directly when you need a custom AI application tailored to your specific workflow, data, and requirements. This gives you maximum flexibility and control, but requires engineering investment.

Buy a product built on a foundation model when the use case is well-served by existing SaaS products — customer support chatbots, AI writing tools, code assistants. Don’t build what you can buy off the shelf.

Use multiple foundation models for different tasks. A smaller, cheaper model for simple classification tasks. A more capable model for complex reasoning. This is increasingly common and reduces costs significantly.

What to Watch Out For

Vendor lock-in is real. If you build your entire AI strategy around one model provider’s API, you’re dependent on their pricing, availability, and technical decisions. Design your architecture so you can swap models without rewriting your application. This isn’t theoretical — I’ve seen clients forced to migrate on short notice when model versions were deprecated.

Bigger isn’t always better. The most capable model is also the most expensive and slowest. For many business tasks — classification, extraction, simple Q&A — a smaller, cheaper model performs just as well. Match the model to the task complexity.

Foundation models hallucinate. They generate plausible-sounding but incorrect information. This is not a bug that will be fixed in the next version — it’s inherent to how these models work. Any production application needs guardrails: fact-checking mechanisms, source attribution, confidence scoring, or human-in-the-loop validation.

Open source vs. proprietary is a real choice. Models like Llama and Mistral are open source — you can run them on your own infrastructure. This eliminates API costs and data privacy concerns but requires more engineering investment. For regulated industries or sensitive data, self-hosted open-source models may be the right choice.

The Verdict

Foundation models are the most significant shift in how businesses build software since cloud computing. They make AI accessible to companies that couldn’t afford custom model development. But accessibility doesn’t mean simplicity — choosing the right model, the right adaptation strategy, and the right architecture still requires informed technical leadership.

The companies getting the most value from foundation models are the ones that understand both the capabilities and the constraints, and make pragmatic decisions about which problems are actually worth solving with AI.


Related: What Is an LLM and What Can It Do for Business | AI Strategy for Non-Technical CEOs