JM

Justin McKelvey

Fractional CTO · 15 years, 50+ products shipped

AI for Business 10 min read Apr 28, 2026

AI Operations: How I Scope AI Projects That Actually Ship

TL;DR: AI Operations in Three Phases

Most AI projects fail before they start because they were scoped to impress, not to ship. AI operations — the discipline of selecting, scoping, integrating, and running AI inside a business — is where almost all the value (or wasted budget) lives. The model isn't the hard part anymore. The hard part is getting the AI output to actually be used by your team in a workflow that produces a measurable result. The 3-phase AI operations scoping framework: Discovery → Build & Integrate → Operate. Most successful AI projects ship in 4-12 weeks across these three phases. Most failed AI projects skipped Phase 1 and tried to start at Phase 2.

What "AI Operations" Actually Means (And Why It's Different from AIOps)

The term gets confused regularly, so let's separate the two cleanly.

AIOps (one word) is the established enterprise IT category. Splunk, Dynatrace, Datadog, and similar players use "AIOps" to mean using AI to monitor and manage IT infrastructure — alerting, anomaly detection, log analysis. If you search for AIOps and you're not in IT operations, you're probably in the wrong category.

AI operations (two words) is the broader business discipline. It's how any company — accounting firms, agencies, ecommerce shops, professional services — selects AI use cases, scopes projects, integrates the output into real workflows, and runs the resulting systems over time. It's the management layer around AI, not the AI itself.

This post is about the second one. If you're an owner or operator at a profitable business trying to figure out how to actually use AI in your operations — not how to monitor your servers — keep reading.

Why Most AI Projects Don't Ship

I run an AI ops shop called SuperDupr where this is the work. Across the engagements I've scoped, the same failure patterns repeat. They're not technical failures. They're scoping failures.

The demo trap. An owner sees an AI demo. The demo summarizes a contract in 30 seconds with one clean input and one perfect output. They back into a project that recreates the demo. Six months later, the project hasn't shipped because the real workflow involves 12 contract types, 4 systems of record, 3 stakeholders who need to review, and edge cases the demo never showed. The scope didn't account for any of that. It accounted for the demo.

The pilot-as-stall. Stakeholders aren't convinced. Someone proposes "let's run a pilot." The pilot is scoped to prove AI works. It does prove that. Then everyone discusses the pilot for four months while no real implementation lands. The pilot becomes the project. AI operations gets confused with AI testing.

The scope-from-tool, not from-problem. Someone reads about a new AI tool and decides "we should use this." A project is scoped around the tool. Halfway through, the team realizes the tool doesn't fit the problem. The project either fails or pivots to a different tool. Scope-from-tool is the most common mistake at companies under 100 people.

The fix to all three is the same: a real scoping phase before anything gets built. That's what Phase 1 is for.

The 3-Phase AI Project Plan I Use (Scoping AI Project Engagements That Ship)

Scoping an AI project well is the single highest-leverage activity in this whole arc. Get the scoping right and the rest of the engagement runs cleanly. Get it wrong and no amount of model quality or engineering rescues it. The 3-phase AI project plan below is the structure I use on every engagement.

Phase 1: Discovery (Weeks 1-2). Phase 2: Build & Integrate (Weeks 3-8). Phase 3: Operate (ongoing, with a 30-day stabilization). The phases aren't optional and they aren't in a different order. Discovery before building. Building before operating. The shape doesn't change; what changes is how big each phase is for the specific project.

Phase 1: Discovery (Weeks 1-2)

The job in Phase 1 is to define the workflow we're trying to change, the people whose work changes, the data that already exists, and what "success" looks like in measurable terms. No AI gets built. No tools get selected. The output is a one-page scoping doc that anyone in the business can read.

Specifically, Phase 1 produces:

The workflow being changed. Drawn out as a sequence — what happens today, who does it, where the data comes from, what the output looks like. If we can't draw the current state in under a page, the project isn't scoped tight enough yet. Keep narrowing.

The success metric. One number. Time per task, error rate, throughput, customer turnaround, revenue per hour — whichever applies. Without one specific number, the project has no way to declare success and no way to declare failure. Vague projects don't ship.

The data we have. Where it lives, what shape it's in, how clean it is, who owns access to it. Most AI projects break in Phase 2 because Phase 1 didn't honestly assess data. "We have customer data" isn't an assessment — "we have 12,000 customer records spread across HubSpot and a Google Sheet, with 30% missing email and inconsistent name formatting" is.

The integration surface. Which tools the AI output has to flow into and out of. CRM, accounting system, ticketing tool, email — whatever's relevant. Each integration is a project unto itself; counting them honestly in Phase 1 prevents Phase 2 surprises.

The risk of being wrong. If the AI output is wrong 10% of the time, what happens? For some workflows (drafting first-pass email replies for human review), 10% wrong is fine. For others (auto-categorizing transactions for tax purposes), 10% wrong is catastrophic. The tolerance for error is the most-skipped scoping question.

Phase 2: Build & Integrate (Weeks 3-8)

This phase varies most across projects. A focused use case with clean data and one integration is 3-4 weeks. A multi-stakeholder rollout with messy data and 3-5 integrations is 6-8 weeks. The shape, though, is consistent.

Pick the model and the orchestration layer. Most projects don't need a custom model — a foundation model (Claude, GPT, Gemini) plus a thin orchestration layer (a small Rails or Node app, or a workflow tool) is sufficient. Reserve fine-tuning and custom models for cases where you've already proven the workflow works without them.

Build the integration plumbing. Webhooks in, webhooks out. Auth. Error handling. Logging. The unglamorous engineering work that makes the AI output flow from where it's generated to where it gets used. This is where 60-70% of the actual build time goes — not on the AI.

Run real data through it. Not synthetic test data. Real records from the business, with all their messiness. The first end-to-end run is almost always ugly. Phase 2 is partially about making the ugly parts less ugly through iteration on prompts, integration code, and edge-case handling.

Build the human review/feedback loop. AI outputs need human eyes for the first 30-90 days at minimum, and often forever. Build the review interface as part of the project, not as an afterthought. The team using the AI needs to be able to flag bad outputs and have those flags actually train the next iteration.

Phase 3: Operate (Ongoing, with 30-Day Stabilization)

This is the phase most projects skip — and it's where AI operations actually lives. The first 30 days after launch are stabilization. Things will break. Edge cases will surface. The team will use the system in ways nobody anticipated. The job in Phase 3 is to absorb that signal and improve the system.

What Phase 3 looks like in practice:

Daily monitoring for the first 30 days. Output quality, error rates, usage rates. If the team isn't using the AI output, it doesn't matter how good the model is. Adoption is a metric.

Weekly review of flagged outputs. Whatever the team flagged as wrong — review them, find the patterns, update prompts or rules accordingly. This is the loop that turns a 70%-accurate AI workflow into a 90%-accurate one over 8-12 weeks of operation.

Quarterly scope reviews. What the AI does well, what it doesn't, what new use cases the team has surfaced. Most successful AI deployments find adjacent use cases within 3-6 months that weren't visible at scoping. Phase 3 is when those get evaluated honestly.

What AI Implementation Actually Costs (AI Implementation Cost Ranges 2026)

If you searched for "ai implementation cost," here are honest numbers — not vendor brochures. AI implementation cost ranges widely depending on scope, integration depth, and how clean your data is going in. The model itself is nearly free; the implementation around it is where every dollar goes.

Here are real cost ranges I see across engagement sizes. Honest numbers, not vendor brochures.

Focused single-use-case AI project (one workflow, one integration, one team): $15,000-$30,000 fixed price for a 4-6 week engagement. This is the right shape for most first-time AI projects at small/mid businesses. Get one use case shipping; learn from it; expand from there.

Multi-stakeholder AI rollout (one workflow, 3-5 integrations, multiple teams): $30,000-$75,000 for an 8-12 week engagement. The cost driver isn't the AI; it's the integration count and the change management. Each stakeholder group adds friction.

Enterprise AI platform build (custom orchestration, multiple use cases, ongoing operation): $75,000-$250,000+ over 6-12 months. Reserve this for businesses that have already shipped 2-3 focused use cases successfully and have a real platform need.

If a vendor quotes you under $5K for an AI project: they're selling you an API call wrapper, not an implementation. The "AI part" of an AI project is genuinely cheap now. The work that produces business value is the scoping, integration, and operation around it — and that's never $5K.

The Pilot-Project Trap (And How to Avoid It)

Pilots get a bad rap because most of them are stalling tactics. But there's a real version of the AI pilot project that's worth running.

A real pilot has a kill criterion. "If after 30 days the model's accuracy is below 80% on real data, we kill the project" is a real criterion. "Let's see how it goes" is not. Without a written kill criterion, the pilot will outlive its usefulness because nobody wants to be the person who "killed" it.

A real pilot has a time box. 30 days, 60 days, 90 days. The clock starts when the pilot launches and stops on a specific calendar date. After that date, you decide: ship it, kill it, or extend with explicit rationale. No silent extensions.

A real pilot has a single use case. Not "AI for our business." A specific workflow with a specific success metric. If the pilot scope drifts during the pilot, you're not running a pilot anymore — you're running a project with extra steps.

If your proposed pilot doesn't have all three (kill criterion, time box, single use case), it's a stalling tactic. Either tighten it into a real pilot or skip the pilot phase and ship a small focused implementation directly.

Who Should Own AI Operations Inside a Company

Three rough thresholds based on the engagements I see.

Under 50 people: the owner or COO owns AI operations as part of their broader role. They scope projects, prioritize use cases, and oversee outside support. A dedicated AI hire at this scale is premature; the work isn't 40 hours a week and the right shape is fractional or productized AI ops support.

50 to 250 people: a designated lead — usually a head of operations or chief of staff — owns AI operations part-time, often 5-10 hours a week. Outside or fractional support handles the implementation work; the internal lead handles prioritization, stakeholder coordination, and adoption.

250+ people: a dedicated role makes sense. Titles vary — Chief AI Officer, Head of AI Operations, AI Program Manager. The role manages a portfolio of AI projects across the company and increasingly owns the relationships with outside vendors and consultants.

The wrong move at any scale is hiring a "Head of AI" too early, before the work is defined. Generic Head of AI hires often turn into expensive position-builders who write strategy decks for 6 months before any AI ships.

How This Connects to My Work

The 3-phase scoping framework is what we run at SuperDupr — that's the productized AI ops business I founded. SuperDupr handles the implementation: Phase 2 build + integration, Phase 3 operate + stabilize. The Phase 1 discovery work is included in every engagement because it's the part that makes the rest work.

If you're trying to figure out whether AI operations are the right fit for your business right now, that's a different conversation — strategic advisory rather than productized implementation. Book a strategy call and I'll help you decide whether you're ready to scope a project, whether a pilot is appropriate, or whether you should hold off until your data and workflows are in better shape. If you've already scoped a project and want a productized implementation, that's what SuperDupr is built for.

For the broader question of how to choose between an AI consultant, an AI agency, and a productized AI ops service, see AI Consultant vs Agency vs Productized AI Ops. For the related decision of whether to build AI yourself or buy it, see Build vs Buy AI: A Decision Framework.

Frequently Asked Questions

What is AI operations?
AI operations means the discipline of running, scoping, and maintaining AI systems inside a business — selecting the right use cases, integrating them with existing tools, monitoring their output, and improving them over time. It's the management layer around AI implementation, not the AI itself. It's distinct from AIOps (one word), which is an enterprise IT category for using AI to monitor infrastructure. AI operations applies to any business deploying AI; AIOps is a narrower IT discipline.
How is AI operations different from AIOps?
AIOps (one word) is the established enterprise IT category — companies like Splunk, Dynatrace, and Datadog use 'AIOps' to mean using AI to monitor and manage IT infrastructure. AI operations (two words) is the broader business discipline: how a non-IT business runs AI projects from scoping through deployment to ongoing improvement. If you're hiring someone for AIOps, you want infrastructure monitoring expertise. If you're hiring for AI operations, you want someone who scopes business problems and ships AI solutions to them.
How long does an AI project take to ship?
A well-scoped AI project ships in 4 to 12 weeks. The variable is scope and integration depth. A focused use case with clean data and one integration ships in 4-6 weeks. A multi-stakeholder rollout with messy data and 3-5 integrations takes 8-12 weeks. Anything scoped to take more than 12 weeks at this stage of AI tooling maturity is almost always over-scoped — break it into phases that ship value every 4-6 weeks.
How much does an AI implementation cost?
AI implementation costs range from $5,000 for a focused single-use-case engagement to $150,000+ for enterprise rollouts. The honest mid-range for a profitable small or mid-sized business is $15,000 to $60,000 per scoped project. The number is less about model costs (which keep dropping) and more about discovery, integration, change management, and the human work of getting AI output to actually be used by the team. If a vendor is quoting you under $5K for an AI project, they're not scoping it; they're selling you a model API call.
Should I run an AI pilot first?
Sometimes. The right reason to run a pilot: you have a specific question about the use case (will the model output be accurate enough? will users actually adopt it?) that can only be answered by testing. The wrong reason: stakeholders aren't convinced and need to be convinced before approving a real project. Pilots aimed at convincing skeptics usually become permanent — they ship 'proof' that gets discussed for months while no real implementation ever lands. If you can't define what would make the pilot succeed (or fail) in advance, you're not running a pilot, you're stalling.
Who should own AI operations inside a company?
It depends on company size. Under 50 people: the owner or COO owns AI operations as part of their role; outside support handles scoping and implementation. 50 to 250 people: a designated lead — often a head of operations or a chief of staff — owns it part-time, with outside or fractional AI operations support. 250+ people: a dedicated role makes sense, often titled Chief AI Officer or Head of AI Operations. Below that scale, hiring a full-time AI ops lead is premature; the work isn't yet 40 hours a week.
What's the most common AI scoping mistake?
Scoping for the demo, not for the workflow. Founders and ops leaders see an AI demo and back into a project that recreates the demo. The demo runs in 30 seconds with clean inputs and one happy-path test case. Real workflows have messy data, multiple stakeholders, edge cases, and integration with tools that don't have clean APIs. A project scoped from a demo systematically underestimates integration and change-management work — usually by 3-5x. Scope from the workflow you want to change, not from the demo that excited you.

If this was useful, here are two ways I can help: