Michigan GPT

What an AI pilot looks like in a mid-sized manufacturer

How to scope an AI pilot inside a fifty to two hundred person manufacturer so it produces a real result in a quarter, instead of a slide deck the team forgets about.

Most AI pilots inside mid-sized manufacturers produce one of two outcomes. They produce a measurable improvement in the operation, the team adopts the result, and the project becomes a foundation for the next one. Or they produce a polished demonstration, the team nods, and the project quietly disappears six months later when nobody can quite remember whether the system is still in use.

The difference between the two outcomes is mostly determined at the pilot scoping stage, before any technology is purchased or any vendor is engaged. The pilots that produce results are scoped narrowly, measured honestly, and run by people inside the operation rather than by people outside it. The pilots that produce demonstrations are scoped broadly, measured loosely, and run by external consultants who do not stay long enough to see the result.

This is the working version of how to scope an AI pilot in a mid-sized manufacturer so it produces a real result in a quarter.

Pick a problem with a number attached

The first scoping decision is the most important. The pilot has to address a specific problem whose current state can be measured in a number, and whose target state can be expressed in a different number.

A problem like "improve quoting" is not yet a pilot. A problem like "reduce average quoting time per RFQ from forty-five minutes to fifteen, with the same accuracy on jobs over fifty thousand dollars" is a pilot. The first version is aspirational. The second version is operational.

The team that puts a number on the current state and the target state has done the most important work of the pilot before the pilot has started. The team that skips this step tends to produce a project that ends up "improving things in some way" without a clear answer to the question of whether the improvement is real.

The number does not have to be perfect. It has to be defensible enough that the team can compare results against it honestly. A current state estimate that is approximately right is much better than no current state estimate. The most common scoping mistake is to wait for perfect measurement before starting. There is rarely going to be perfect measurement. Start with the rough number and improve it as the pilot runs.

Pick a scope that fits a quarter

The next scoping decision is the timeline. A pilot that requires more than a quarter to produce a result is not actually a pilot. It is a project, with all the risk and overhead a project carries. Most mid-sized manufacturers cannot absorb that risk, and the projects that span multiple quarters tend to lose internal sponsorship before they finish.

The right scope for a real pilot is something the operation can implement, run, and measure in roughly twelve weeks. That is short. It forces real choices about what is in scope and what is not. The forcing function is healthy.

The kinds of projects that fit in a quarter at this scale are narrow versions of larger ambitions. Not "AI quoting across the operation" but "AI-assisted quoting for the part family that produces forty percent of our revenue." Not "production scheduling automation" but "AI-assisted morning schedule recommendations for the highest-throughput cell." Not "computer vision quality inspection" but "computer vision pre-inspection for the specific defect that costs us the most rework."

Each of these is meaningful. Each of them is achievable. The pilot that proves out one of them earns the right to scope a larger project the next quarter. The pilot that tries to do the larger project in the first attempt usually does neither.

Run the pilot inside the operation

The third scoping decision is who runs the pilot. The pilots that produce results are run by people inside the operation, with external help where needed. The pilots that produce demonstrations are run by external consultants with operation people kept in the loop.

The reason for this is operational rather than ideological. The person inside the operation knows what the part actually looks like, how the schedule actually breaks, and what the team will and will not adopt. The external consultant has process expertise but is rarely going to be there when the pilot result has to be operated next quarter.

The right shape is for an internal owner to lead the pilot, with external technical help on the parts that require expertise the operation does not have. The internal owner is named, has explicit authority for the pilot's decisions, and has the time to run it. The external help is bounded, brought in for specific work, and exits when the work is done.

The mistake to avoid is an internal owner who has the title but not the time. A pilot that is the third or fourth priority of someone running a department does not get the attention it needs to succeed. The owner needs roughly a quarter of their time, dedicated, for the pilot quarter. If that is not available, the pilot is not actually scheduled, and the operation should either find someone whose time is available or wait until next quarter.

Define the success and failure criteria explicitly

The pilot needs explicit criteria for what counts as success and what counts as failure. The criteria are documented at the start, signed off by the executive sponsor, and not revised mid-pilot.

The success criteria are usually some version of "we hit the target number, the team adopted the change, and the result has held for at least a few weeks." All three matter. A target number hit in a one-week burst that is not sustainable is not a success. A target number hit by a system the team did not adopt is not a success. A team adopting a system that did not move the number is not a success either.

The failure criteria are equally important. A pilot that does not hit its target should be honestly recognized as not having hit it, with a clear understanding of what did not work. The team that gets this right learns from each pilot, even when the pilot does not succeed. The team that gets it wrong tends to claim partial success on every pilot and never quite figures out which approaches actually work.

The most common mistake is to define success too softly. "Demonstrated capability" is not success. "Generated insights" is not success. "Increased awareness" is not success. The pilot has to produce a measurable change in the operation, or the pilot has not produced what a pilot is supposed to produce.

Plan for the second quarter at the start

A pilot that succeeds is not the end of the work. It is the start. The team has to operate the result, support it, retrain when conditions change, and integrate it into the rest of the operation. The plan for the second quarter, when the pilot becomes the new normal, is part of the pilot scope.

This planning step is consistently undervalued. Pilots that are technically successful but operationally orphaned tend to fade quickly. A scheduling tool that is not being maintained, a quoting model that is not being retrained, an inspection system that is not being recalibrated all degrade over months until the team stops using them.

The right planning includes a named operator for the post-pilot system, a budget for ongoing maintenance, a defined retraining or recalibration cadence, and integration with the operation's existing tooling. The plan does not need to be detailed. It does need to exist, and the executive sponsor needs to have signed off on the resources before the pilot starts.

What this looks like across the quarter

A pilot that follows this scoping has a recognizable shape across the twelve weeks. The first two weeks are spent installing the measurement, locking the scope, and aligning the team. Weeks three through eight are the build, with iterative checkpoints against the success criteria. Weeks nine and ten are real operation against live conditions, with the team adjusting as needed. Weeks eleven and twelve are the honest assessment against the criteria and the planning for the post-pilot operation.

A pilot run this way costs the operation real money and real time. The cost is bounded. The result is honest. The team learns whether the technology works for their specific situation, and they learn it within a quarter rather than within a year.

The mid-sized manufacturers who handle their first pilot well tend to compound. The second pilot scopes faster, draws on the lessons of the first, and produces a result with less help. The third pilot starts to feel like a normal operating capability rather than a special initiative. Within two years, the operation has a working AI competence that competitors have not built, and the cumulative margin from a series of small successful pilots is meaningful.

The manufacturers who skip this discipline tend to produce one or two demonstrations, lose sponsor confidence, and end up further behind than where they started. The cost of doing it right is small relative to the cost of doing it wrong. The choice is mostly about whether the operation is willing to scope narrowly and measure honestly. The technology is not the hard part.