The AI guide to the galaxy
Don't panic
Someone on your board said "we need an AI strategy." Or maybe it was your CEO, or a competitor's press release, or the third consulting deck this quarter with "AI-powered" on every slide. However the mandate arrived, it arrived. And now it's yours.
Here's what you're probably feeling: the urgency is real, but the path from "we need to do AI" to actually doing something useful is unclear. Every vendor has a pitch. Every conference has a keynote. The coverage oscillates between "AI will replace your entire workforce" and "AI is just autocomplete." Neither helps when you're trying to figure out what to do on Monday morning.
You're not wrong to feel stuck. Seventy-four percent of companies now rank AI as a top-three strategic priority. Forty-two percent abandoned most of their AI initiatives last year. Those numbers can both be true because the problem isn't whether to do this. It's how. And most companies are getting the "how" wrong.
This guide won't make you a machine learning engineer. It will give you enough understanding to evaluate pitches without getting snowed, enough data to separate signal from noise, and a practical framework for building a strategy that actually reaches production. That last part matters more than it sounds. Ninety-five percent of AI pilots never get there.
We've written detailed surveys of both the commercial AI landscape and the open source AI landscape as companion pieces. They cover who the players are, what the market looks like, and where the money is going. This piece is different. This is the "what do I actually do?" piece.
How to think about the technology
You don't need a computer science degree to make good AI decisions. But you need one core concept, because it changes how you evaluate everything you hear.
The prediction machine
Large language models (the technology behind ChatGPT, Claude, Gemini, and most of the tools making news) work by predicting what comes next. You give them text. They predict the most likely continuation, one word at a time. That's it.
This sounds simple because it is. The complexity is in the scale: these models have processed trillions of words during training and learned patterns of language so deeply that their predictions can feel like understanding. They can write fluently, summarize accurately, translate between languages, analyze documents, and generate working code. All by predicting what plausible text looks like given the input you provided.
The reason this matters for your decisions: an LLM doesn't know things the way a database knows things. It doesn't look up answers. It generates plausible text. Sometimes that text is factually correct. Sometimes it isn't. The model can't tell the difference, and it will state a fabrication with the same confidence as a fact. The industry calls this "hallucination." It's not a bug being worked on. It's a property of how the technology functions.
What this means in practice
AI is excellent at tasks where the output is a starting point, not a final answer. Drafting documents, summarizing meetings, analyzing unstructured text, generating code for a developer to review, classifying information, extracting structured data from messy sources. These are the use cases delivering real value today, because "good enough" is good enough and a human checks the result.
AI is unreliable for tasks where precision is non-negotiable. Medical diagnoses, legal citations, financial calculations, anything where a confident-sounding wrong answer is worse than no answer at all. AI can still participate in these domains. But the workflow must be designed so that AI assists and humans verify. The companies getting burned are the ones that skipped the verification step.
The practical framework: think of AI as the best junior analyst you've ever hired. Fast, tireless, surprisingly capable, and absolutely not someone you'd let sign a contract without review.
What the data actually says
The AI conversation is drowning in hype. Here's what's happening, measured in dollars and outcomes rather than keynote slides.
The adoption numbers
Eighty-eight percent of organizations now use AI in at least one business function, up from 78% a year ago (McKinsey). Two-thirds use it in multiple functions. Enterprise AI spending hit $180 billion in 2025 and is on track to pass $250 billion this year.
That sounds like a success story. It isn't.
The failure numbers
Ninety-five percent of generative AI pilots fail to deliver measurable financial returns. That comes from MIT research based on 150 leader interviews, a 350-person survey, and analysis of 300 public AI deployments. Forty-two percent of companies abandoned most of their AI initiatives in 2025, more than double the rate from the year before (S&P Global). The average organization scrapped nearly half its proofs of concept before they reached production.
Only 39% of organizations attribute any EBIT impact to AI (McKinsey). Among those, most say AI accounts for less than 5% of earnings. Deloitte's 2026 survey found that only 20% report actual revenue growth from AI. Most see payback in two to four years, compared to the seven-to-twelve-month payback typical for other technology investments.
The gap that matters
BCG studied 1,250 executives in September 2025 and found that only 5% of companies globally are capturing substantial value from AI. Those companies see 1.7x revenue growth, 3.6x three-year total shareholder return, and 1.6x EBIT margins compared to laggards. Top performers report $10.30 in value for every dollar invested.
The gap is not between companies that adopted AI and companies that didn't. Almost everyone has adopted something. The gap is between companies that deployed with strategic clarity and companies that bought tools and ran pilots and called it a strategy. Sixty percent of organizations report minimal gains (BCG). Not because the technology failed. Because the approach did.
Where companies go wrong
The patterns of failure are consistent and well-documented. Five mistakes account for the vast majority of AI initiatives that stall or die.
Starting with the solution
"We need to use AI" is not a strategy. It's a solution looking for a problem. The most common and most expensive mistake is choosing the technology before understanding the workflow. Which processes in your business are bottlenecked by human attention? Where do your people spend hours on repetitive, structured tasks that don't require judgment? Those questions come first. The AI discussion comes after.
MIT Sloan found that 70% of AI's value comes from investments in people and process, not from the sophistication of the technology. Yet most mid-market companies spend 60-70% of their AI budget on technology. They have the ratio backwards.
Pilot purgatory
Over 80% of companies have experimented with tools like ChatGPT or Copilot. Fewer than 5% have moved custom AI solutions into production (McKinsey). Two-thirds of organizations remain in experiment or pilot mode. The industry has a name for this: pilot purgatory. It looks like progress because things are happening. It isn't progress because nothing is shipping.
The trap works like this: a technical team runs a pilot. It shows promising results in a sandbox. Getting to production requires a business owner, integration with existing systems, governance, and change management. Nobody owns those things. So the team runs another pilot. The organization accumulates proofs of concept and zero production deployments.
The companies that escape share one trait: someone on the business side owns the outcome and ties it to a P&L line item. When AI success is measured in model accuracy, it stays in the lab. When it's measured in ticket resolution time or processing cost per document, it ships.
Building when they should be buying
MIT's "GenAI Divide" report found that purchasing AI tools from specialized vendors succeeds about 67% of the time. Internal builds succeed only 33% of the time. For mid-market companies without dedicated ML teams, building custom AI is almost always the wrong first move.
The exception is specific and you'll know if you're it: strict data residency requirements, a highly specialized workflow that no vendor addresses, or an existing ML team with spare capacity. If those don't describe your situation, start with commercial tools and build only when you've exhausted what buying can do.
Bolting AI onto broken processes
Adding AI to a workflow that was never designed for it is like putting a turbocharger on a car with flat tires. Only about 5% of organizations have redesigned workflows end-to-end around AI capabilities (McKinsey). The rest are automating the existing process, including all its inefficiencies.
The companies seeing real returns asked a harder question: if we were designing this workflow from scratch today, knowing what AI can do, what would it look like? That question produces different answers than "where can we plug AI into what we already have?"
Ignoring the shadow AI problem
While only 40% of companies provide official AI access, 90% of workers report using personal AI tools daily for job tasks (MIT). A separate 2025 study found 68% of employees used personal accounts to access free AI tools, with 57% inputting sensitive company data. Cisco found that 46% of organizations experienced internal data leaks through generative AI.
Your employees are already using AI. The question is whether they're using it through channels you control, with policies you've set, on data you've decided is appropriate to share. If you haven't provided sanctioned tools and clear guidelines, the answer is no. And you're accepting risk you haven't evaluated.
What the 5% are doing differently
BCG's "future-built" companies share patterns worth studying. Not because you need to replicate all of them, but because the patterns reveal what actually matters versus what looks like it matters.
They spend on people, not models
The research-supported "70-20-10 rule" says the right AI budget allocation is 70% on people and processes, 20% on infrastructure and integration, 10% on algorithms and models. One documented case showed a company spending $4 million with this framework outperforming a competitor spending $8 million with a technology-first approach across adoption rates, productivity gains, and ROI.
The skills most in demand aren't what companies expected. They thought they needed AI engineers. What they actually need are people who understand both the technology and the business well enough to identify the right problems, scope implementations, and manage change. Three-quarters of AI skill demand is concentrated in management and business operations, not pure engineering.
They buy more than they build
This bears repeating because it contradicts the instinct of most technical leaders. The 67% success rate for vendor solutions versus 33% for internal builds isn't a marginal difference. The companies creating value aren't spending months on custom implementations. They're deploying commercial tools fast, learning from production data, and building custom only where they have a clear, defensible reason.
They redesign workflows
Half of AI high performers are actively redesigning business processes around AI capabilities. The laggards are automating existing steps. The difference: high performers report 3x higher revenue impact and 30% higher EBIT (McKinsey).
Workflow redesign doesn't mean tearing everything apart. It means asking: given what AI can do now, which steps in this process no longer need to exist? What handoffs can be eliminated? What decisions can be made faster because the information is already summarized and ready?
They govern from the top
PwC's 2026 Global CEO Survey found that companies with strong AI governance are three times more likely to report meaningful financial returns. The governance isn't slowing them down. It's what makes scaling possible.
The minimum framework is simpler than most companies assume: an acceptable use policy (what tools, what data, what review requirements), a risk classification for use cases, quality monitoring for AI output, and vendor evaluation criteria. Four documents. Most companies could write them in a week. Most haven't started.
Your first 90 days
Mid-market companies have one structural advantage over large enterprises: speed. Large enterprises take nine months on average to scale an AI initiative. Mid-market firms do it in 90 days. That advantage only works if you use it.
Weeks one and two: find the real problems
Don't start with technology. Start with a simple audit. Where is your team spending time on repetitive, information-heavy tasks that don't require human judgment? Where are bottlenecks? Where is work sitting in queues waiting for someone to review, classify, summarize, or respond?
Also: find out where AI is already happening. Survey your teams. The shadow AI data suggests your people are already using personal ChatGPT or Claude accounts for work. Find out what they're using it for. Their actual usage patterns are better market research than any vendor pitch.
Weeks three and four: pick one problem
Not the biggest. The clearest. Choose a use case where you can define what "good enough" looks like in specific, measurable terms. A 50% reduction in document processing time. Ninety percent accuracy in classification. Response drafts that require editing instead of writing from scratch.
The best first use case involves unstructured text or data that humans currently process manually. That's where AI delivers the fastest, most measurable returns.
Month two: ship something
Pick a commercial platform. Not three platforms to evaluate against each other. One. If you're already in Microsoft, start with Copilot. If you're on AWS, look at Bedrock. If neither fits, read our commercial landscape survey and make a call.
Deploy the solution for your chosen use case. Get it into production, not into another pilot. Production means real users, real data, real feedback. A pilot with synthetic data in a sandbox tells you almost nothing about whether AI will work in your actual workflow.
Month three: measure, learn, expand
One use case in production teaches more than ten in pilot. Measure outcomes against the success criteria you defined. Talk to the people using it. Find out what's working, what's failing, and what they wish it did differently.
Then pick the next problem. The second deployment will be faster because you've already built the muscle: the governance framework, the vendor relationship, the change management playbook, the internal expertise. This is the learning loop. It compounds.
Write the policy now
Don't wait for the data leak or the compliance question. Write your AI acceptable use policy before you need it. Cover four things: which tools are approved, what data can be shared with external AI providers, which outputs require human review before reaching customers or regulators, and what happens when someone finds a problem with an AI output. One page. If it's longer, it won't get read.
What's changing
The market moves fast. Four trends should inform your decisions this year.
The pricing collapse. AI costs are falling roughly 10x every 18 months. Tasks that were too expensive to automate twelve months ago are now trivially cheap. If you evaluated a use case last year and rejected it on cost, run the numbers again.
Reasoning models. Models that think before they answer (OpenAI's o-series, DeepSeek R1, Claude's extended thinking) break complex problems into steps instead of generating a response in one pass. They're slower and more expensive per query, but measurably more accurate on multi-step tasks like financial analysis, code generation, and document review. If your use cases involve anything beyond basic search and summarization, evaluate these specifically.
Agentic AI. Systems that don't just generate text but pursue goals: breaking tasks into steps, using tools, making decisions, adapting based on results. An estimated 40% of enterprise applications will include some form of AI agent by end of year. The technology works. The challenge is organizational: defining clear boundaries, building oversight, and integrating with existing workflows.
Vendor consolidation. This is the year companies stop experimenting with seven AI tools and consolidate around one or two providers. The cost of maintaining multiple vendors (different APIs, security reviews, billing, training) adds up. If you haven't picked your primary provider, this is the year.
For the full picture of who the players are and how the market is shaping up, see our surveys of the commercial and open source landscapes.
The bottom line
The AI mandate sitting on your desk is real. The urgency is real. But the most expensive mistake you can make is treating this as a technology problem.
The 5% of companies capturing real value didn't start by picking the best model or building the most sophisticated system. They started by identifying a specific problem, deploying a capable tool, measuring what happened, and doing it again. They spent on people and process, not just software. They governed early. They shipped to production instead of running another pilot.
The good news for mid-market companies: you can move faster than the enterprises burning nine months on each initiative. Ninety days from decision to production is real. That's the advantage of being smaller, if you use it.
The technology is ready. The data on what works is clear. The companies pulling ahead aren't the ones with the biggest AI budgets. They're the ones that started with the right question.
Not "how do we do AI?" but "what are we actually trying to solve?"
Start there.
More Playbooks
See all playbooksKeeping your brand voice alive in the age of AI content
AI can write faster than you. It cannot write like you. A practical guide to using AI for content creation without sounding like everyone else.
February 4, 2026
The AI innovation pipeline
Most companies go too big or too small with AI. A systematic approach to finding, scoring, and sequencing the initiatives that actually reach production.
January 21, 2026
The AI readiness scorecard
Most companies pick AI projects based on excitement. A framework for figuring out which workflows are actually ready and which ones need work first.
January 7, 2026
Get The Pepper Report
The short list of what we're writing and what we're reading.