Most enterprises make the operating model decision for their AI Center of Excellence the same way they make furniture purchases: based on what looks right in the showroom, not what fits the house. The result is predictable. Hub models that cannot scale beyond four business units. Hub-and-Spoke structures where the spokes spend their days waiting for hub approvals. Platform models deployed before the organization has enough AI maturity to use them.
The AI CoE operating model is the single biggest determinant of whether your AI program delivers 12 models in production in year one or 2 models in year three. It is not a technical architecture question. It is an organizational design question, and it needs to be answered using data about your organization specifically, not best practices from a vendor case study.
This article gives you the framework to make that choice correctly, implement it deliberately, and migrate to a more sophisticated model as your maturity grows.
The Three Operating Models: What They Actually Are
Enterprise AI CoE literature tends to describe these three models in aspirational terms. This is not helpful. Here is what each model actually looks like when deployed in a real organization with real politics, real budget constraints, and real talent shortages.
- Clear accountability for all AI output
- Fastest to build from zero talent base
- Consistent governance and standards
- Efficient use of scarce ML talent
- Becomes bottleneck beyond 3 to 4 units
- Business units feel ownership is absent
- High bus-factor risk on key staff
- Domain knowledge gap widens over time
- Combines central governance with domain proximity
- Scales across multiple business units
- Builds distributed AI capability faster
- Business units have real ownership
- Requires mature governance to avoid fragmentation
- Spoke practitioners often become isolated
- Dual reporting creates ambiguity
- Coordination overhead is significant
- Maximum scalability once mature
- Business units fully own their AI programs
- CoE becomes strategic lever not delivery team
- Enables AI-native product thinking
- Requires 3 to 5 years of maturity to work
- Self-service tools fail without ready users
- Risk of governance collapse at scale
- Very high upfront infrastructure investment
The critical insight here is that these models are not equally valid options for every organization. They represent a maturity progression. Deploying a Platform model before your organization has AI literacy, trained practitioners, and functioning governance is not ambitious leadership. It is expensive failure.
How to Choose: The Selection Criteria Matrix
The right operating model is determined by seven organizational factors. Assess each one honestly before making the decision. If you are receiving advice to skip this assessment, the advisor is optimizing for something other than your success.
If you are honest about these seven factors for your organization, the right model is usually obvious. Where we consistently see enterprises go wrong is overstating their AI literacy, their existing talent count, and their governance maturity. The organizations that skip this assessment and jump to Hub-and-Spoke or Platform models based on peer benchmarking spend the next 18 months building the foundations they should have built first.
The Hub Model: Making It Work Without Creating a Bottleneck
The centralized Hub is the right starting point for most large enterprises. The problem is not the model itself. The problem is how most organizations implement it. A Hub that functions as a shared services team, processing use case requests from a queue, is not an AI Center of Excellence. It is an expensive consulting department that happens to sit inside your organization.
A functioning Hub model has three structural properties that distinguish it from an internal consulting shop. First, it operates as a demand filter, not a demand processor. Not every business unit request becomes a project. The Hub applies a consistent triage framework to prioritize use cases by strategic value and feasibility, declining low-value requests rather than building everything submitted. Second, it owns a reusable asset base. Each project produces components: data pipelines, feature stores, model patterns, and deployment templates that accelerate subsequent work. A Hub that starts from scratch on every project is rebuilding instead of compounding. Third, it measures velocity, not busyness. The metric is models in production per quarter, not projects in progress. Teams that track activity rather than output consistently deliver less.
The bottleneck problem emerges specifically when the Hub takes on more than it can deliver at quality. The structural fix is not adding headcount. It is demand management: making the triage criteria public, communicating backlogs transparently to business units, and defining clear thresholds at which Hub capacity triggers a model transition rather than a hiring spree.
Hub Capacity Thresholds
The Hub model breaks at predictable points. Knowing these thresholds in advance lets you plan the transition rather than react to the breakdown. Based on patterns across our client portfolio, there are three warning signals that indicate a Hub is approaching its structural limit.
Backlog exceeds 8 use cases: When more than 8 use cases are waiting in the Hub queue, average time-to-production exceeds 9 months and business unit frustration becomes a political problem. This is the earliest signal that Hub capacity is insufficient for organizational demand.
More than 4 business units actively sponsoring work: Beyond 4 concurrent business unit relationships, the Hub's coordination overhead exceeds 30% of team capacity. The team spends more time in alignment meetings than building models.
Hub team size exceeds 15 practitioners: At this size, the Hub's internal coordination cost becomes significant. Team cohesion decreases, knowledge sharing requires formal processes, and the cost per delivered model begins rising. This is typically the point at which Hub-and-Spoke becomes structurally superior.
The Hub-and-Spoke Model: Avoiding the Coordination Trap
Hub-and-Spoke is the most commonly adopted model in large enterprises with 5,000 or more employees, and it has the highest variance in outcomes of the three models. Done well, it combines the governance and infrastructure advantages of a centralized Hub with the domain proximity and ownership advantages of embedded practitioners. Done poorly, it creates a coordination nightmare where the hub practitioners and spoke practitioners have overlapping responsibilities, unclear decision rights, and competing incentives.
The architecture that works consistently has a clean separation of responsibility between hub and spoke functions. The hub owns the shared infrastructure layer: data platforms, model registry, monitoring tooling, security standards, and the governance framework. It also owns the talent pipeline, meaning it recruits, develops, and, in some cases, rotates spoke practitioners. The spokes own domain problem definition, feature engineering specific to their business unit, and deployment and adoption within their unit. They are responsible for making AI work in their context, not for rebuilding common infrastructure.
The coordination trap occurs when this separation is not enforced. If spoke practitioners can build their own tooling because the hub's tooling is too slow or inflexible, you now have fragmentation. If hub practitioners routinely intervene in spoke delivery because spoke skills are insufficient, you have rebuilt the bottleneck in a distributed form. Both failure modes are common and both stem from the same root cause: the organization adopted Hub-and-Spoke without the prerequisite maturity for spoke practitioners to operate independently.
The Dual Reporting Question
Spoke practitioners typically report to a business unit leader for day-to-day delivery and to the CoE hub for technical standards, career development, and platform alignment. This dual reporting structure is the most common source of interpersonal friction in Hub-and-Spoke implementations. The practical resolution is not to eliminate dual reporting, which removes the accountability tension that makes the model work, but to define which decisions belong to each reporting line.
Business unit reporting covers: project prioritization, stakeholder relationships, sprint goals, and performance evaluation inputs. CoE reporting covers: technical architecture decisions, standards compliance, tooling choices, and career development. When these boundaries are explicit and enforced from day one, dual reporting functions. When they are ambiguous, the spoke practitioner resolves the ambiguity by deferring to whichever relationship is more comfortable, which is almost always the business unit, leading to technical fragmentation.
The Platform Model: Why Most Organizations Are Not Ready for It
The Platform model is genuinely the right end state for large enterprises with significant AI ambition. It is also the operating model most frequently adopted prematurely, because it is what AI thought leadership describes when it discusses organizational AI maturity, and because internal CoE teams prefer it as it elevates their organizational status.
A functioning Platform model requires four structural prerequisites that most enterprises take three to five years to build. First, a self-service ML platform that business unit practitioners can use without CoE support. This requires significant investment in tooling, documentation, and training, and it requires business unit practitioners who have enough ML literacy to work with self-service tooling responsibly. Second, a comprehensive governance framework that business units can apply independently, because in a Platform model the CoE no longer reviews every model before it goes to production. Instead, business units self-certify against governance standards. This requires governance frameworks sophisticated enough to catch problems and business units mature enough to apply them honestly. Third, established talent pipelines at the business unit level. You cannot run a Platform model if business units need to import talent from the hub for every non-trivial project. Fourth, a cultural shift in how the CoE sees its role, from delivery team to platform team. CoE practitioners who define their professional identity by building models will resist this shift actively.
Migration Paths: Moving Between Models Without Losing Momentum
The AI CoE operating model is not a one-time decision. Organizations mature, talent bases grow, and the right model at year one is rarely the right model at year four. The problem is that model transitions are almost always reactive, triggered by a bottleneck crisis or a delivery failure, rather than planned as a deliberate maturation step.
Planned transitions take four to six months and maintain delivery continuity throughout. Reactive transitions take twelve to eighteen months, lose key personnel in the reorganization, and typically set the AI program back six to twelve months. The difference is whether the transition was anticipated and structured before the trigger event.
Governance Integration by Operating Model
The AI governance framework that works for a Hub model is structurally different from what works for a Hub-and-Spoke or Platform model. This is one of the most commonly overlooked aspects of operating model selection. Organizations that adopt Hub-and-Spoke without redesigning their governance find that governance becomes the bottleneck that the model transition was supposed to eliminate.
| Governance Function | Hub Model | Hub-and-Spoke | Platform |
|---|---|---|---|
| Model approval | Hub reviews all | Spoke self-certifies, Hub spot-checks | Automated gate with audit trail |
| Risk classification | Centralized by CoE | Spoke classifies, Hub validates High/Critical | Self-service with mandatory review for Critical |
| Standards enforcement | Build-time review | Quarterly audit plus incident-triggered review | Continuous automated policy enforcement |
| Incident response | Hub owns entirely | Spoke responds, Hub escalation path | Distributed with CoE oversight for Sev1 |
| Vendor governance | Hub owns all contracts | Hub owns platform vendors, Spoke owns tools | Catalog-based with Hub approval for new vendors |
| Data access control | Centralized approval | Hub sets policy, Spoke enforces | Policy-as-code with Hub-defined guardrails |
The governance transition between Hub and Hub-and-Spoke is the highest-risk period in any operating model migration. During this period, the hub is delegating approval authority to spokes that may not yet have the judgment to exercise it correctly, and the hub is reducing its review cadence before automated checks are in place to substitute for human review. The practical mitigation is to run a parallel review period of at least 90 days where both the spoke and hub review models independently, then compare outcomes. If their risk assessments agree more than 85% of the time, the spoke is ready for delegated authority. If not, continue the parallel period and address the gaps in spoke governance training.
Readiness Thresholds: When to Transition and When to Wait
Transitioning to a more sophisticated operating model before you are ready is significantly more damaging than staying in a simpler model longer than necessary. A Hub model that delivers 12 models in production per year is more valuable than a Hub-and-Spoke model that delivers 4 models while the organization reorganizes around the coordination overhead.
Five Structural Mistakes That Derail Operating Model Implementation
Across 200+ enterprise AI programs, the same structural mistakes appear repeatedly. None of these are strategic failures. They are execution failures, which means they are preventable.
Mistake 1: Treating the operating model as an org chart decision. The operating model is not primarily about reporting lines. It is about decision rights, workflow, and accountability. Organizations that redesign the org chart without redesigning the decision-making process discover that the new boxes on the chart behave identically to the old ones.
Mistake 2: Hiring for the target model instead of the current model. If you are a Hub model today, your highest-value hire is a senior practitioner who builds production models. If you hire a platform architect to build the infrastructure for a Platform model you will not need for three years, you have spent significant budget on capability you cannot use. Hire for where you are, not where you aspire to be.
Mistake 3: Underinvesting in spoke enablement. In Hub-and-Spoke implementations, most organizations invest heavily in hub capability and treat spoke enablement as a secondary concern. The result is spokes that cannot function independently and a hub that continues bearing the delivery burden that the transition was supposed to distribute. Budget a minimum of 20% of the first-year Hub-and-Spoke operating budget for spoke training and enablement.
Mistake 4: Launching without a governance framework for the operating model. The governance design for your AI program must match the operating model. See the AI governance framework design guide for model-specific governance architecture. Organizations that launch Hub-and-Spoke with Hub-level governance, where everything requires central approval, have not changed their delivery model at all.
Mistake 5: Measuring the wrong things. The default measurement for a new CoE is activity: number of projects started, number of use cases evaluated, number of models trained. None of these metrics tell you if the CoE is delivering business value. The only metrics that matter are models in production, business impact per model, and time from use case approval to production deployment. If you are not measuring these three things from day one, you are optimizing for the wrong outcomes.
The Decision That Compounds
The operating model decision is unusual because its consequences compound over time. A Hub model that is well-executed in year one builds the asset base that makes Hub-and-Spoke viable in year two. A Hub-and-Spoke model with strong spoke enablement builds the distributed capability that makes Platform viable in year three. Each model, done correctly, creates the conditions for the next.
The reverse is equally true. A Hub model that accepts every use case request and measures success by projects in progress rather than models in production produces a backlog and a frustrated organization. A Hub-and-Spoke model that deploys spokes without governance training produces fragmentation. A Platform model adopted before spoke practitioners have self-service capability produces shadow AI programs at scale. The wrong model, or the right model implemented incorrectly, actively destroys the conditions needed for future maturity.
Use the selection criteria honestly. Build the prerequisites before you transition. Measure the outputs that matter. The organizations running 30 or more models in production three years after starting their CoE did not get there by selecting the most sophisticated operating model on day one. They got there by selecting the right model for where they were and executing it with discipline.
For the structural design and 12-month launch roadmap, see our AI CoE setup guide. For the governance framework that supports each operating model, see the AI CoE complete operational guide. If you are ready to assess your current state and determine the right operating model for your organization, start with the free AI readiness assessment.