iCentric Insights Insight

AI and Consulting: Strategy, Build & Adoption | iCentric

AI and consulting explained: how to scope, build and scale AI across your organisation. Frameworks, use cases and a delivery model that actually ships.

May 15, 2026
ai and consulting
AI and Consulting: Strategy, Build & Adoption | iCentric

What we mean by "AI and consulting"

The phrase AI and consulting is doing a lot of work. For some buyers it means a strategy deck and a maturity assessment. For others it means an engineering team that turns up, plugs into the data warehouse and ships a working agent in eight weeks. For still others it means an outsourced AI capability that runs alongside the in-house team for years. Used loosely, the term hides three different services that need very different skills.

At iCentric we split the category into three layers that we believe every buyer should be able to name before signing a contract.

Advisory. This is the work of framing the right problem, sizing the prize and choosing where to invest first. It looks like maturity assessments, opportunity maps, target operating models, business cases and roadmaps. Good advisory is short, sharp and ends with a decision rather than a deck.

Build. This is the engineering, design and product work that turns a chosen use case into a live system. It looks like data pipelines, retrieval-augmented generation, fine-tuned models, evaluation harnesses, agent orchestration, front-end interfaces, integration with systems of record and the unglamorous plumbing that makes any of it survive contact with real users.

Operate. This is the run capability that keeps models healthy in production. It looks like monitoring drift, refreshing knowledge bases, re-running evaluations as foundation models change underneath you, handling incidents, retraining and slowly expanding scope as confidence grows.

The big strategy houses historically owned advisory, the system integrators owned build and almost no one owned operate. The arrival of generative AI has scrambled that picture. Strategy firms are hiring engineers. Engineering shops are hiring strategists. Hyperscalers are pushing reference architectures down into the boardroom. Specialist AI automation agencies are taking work that used to belong to the big four. The buyer needs a clearer map than ever.

Most organisations that come to us for AI consulting are mid-market or enterprise, with revenue large enough to feel the pressure of competitors using AI but not so large that they can spin up a thousand-person internal AI function. They have ambition, a backlog of ideas, anxious boards and a handful of failed pilots. They want a partner who can advise, build and operate without three handovers in between. That is the gap our practice sits in, and it shapes everything that follows in this guide.

Why AI is rewiring the consulting operating model

It is worth being honest about how the supply side of consulting is changing, because it shows up in every proposal you read. The traditional model was a pyramid: a partner at the top, principals and managers in the middle, and a wide base of analysts and associates doing the research, modelling and slide production. Margins relied on leverage, billing many junior days for every senior day.

Large language models hollow out the bottom of that pyramid. Desk research, document synthesis, market sizing, first-draft analysis, slide formatting and even initial code can now be produced by a competent generalist with the right tooling in a fraction of the time. The economic consequence is that firms cannot sustain the same ratios. The pyramid is becoming what the Harvard Business Review recently called an obelisk: fewer layers, smaller teams, and three sharply defined roles.

The first role is the AI facilitator, a practitioner fluent in the current generation of tools, evaluation techniques and data pipelines. They are not academic researchers; they are operators who know which model to reach for, how to wire it into the client's stack and how to spot when it is hallucinating. The second is the engagement architect, who frames the problem, decides what is worth solving, interprets model outputs and translates them into decisions a board can defend. The third is the client leader, who holds the long relationship with senior executives and helps them make sense of constant change.

For buyers, this matters in three ways.

First, the shape of the team you are paying for should look different. If a vendor proposes a classic pyramid of eight juniors to two seniors, you are paying for work the models should be doing for free. Ask what the AI facilitators are using and what the leverage looks like with tools in the loop.

Second, the commercial model is shifting away from time and materials towards outcomes, capacity-based subscriptions and value share. Hours are a worse and worse proxy for value when a senior with the right tooling can do in a morning what a team used to do in a week.

Third, the speed of engagements has compressed. Discovery that took twelve weeks now takes four. Prototypes that took three months now take three sprints. If your prospective partner is quoting timelines that sound like the old world, they are either inflating scope or they have not adopted their own advice.

The firms that have rebuilt their delivery model around AI are not just selling AI; they are using it to deliver, and the difference shows up in price-per-outcome long before it shows up in the brand.

The AI maturity ladder: where most organisations actually sit

Before framing any engagement we run clients through a five-step maturity ladder. It is deliberately simple because most of the elaborate maturity models on the market collapse on contact with a real organisation.

Stage 0 — Ad hoc experimentation. A few enthusiasts have personal subscriptions, marketing has played with a content tool, an engineer has written some scripts that call an API. There is no policy, no inventory, no shared infrastructure. Most organisations are here for longer than their leadership believes.

Stage 1 — Assistant-led productivity. The organisation has deployed a sanctioned assistant (Microsoft Copilot, Google Gemini for Workspace, ChatGPT Enterprise or similar) with single sign-on, data protection terms and a basic usage policy. Productivity gains are real but diffuse, hard to measure and concentrated in heavy users.

Stage 2 — Workflow automation with humans in the loop. Specific workflows have been redesigned around AI: contract review, customer service triage, marketing brief generation, financial reporting commentary. There are evaluation suites, prompts under version control, and feedback loops. Value is now measurable per workflow.

Stage 3 — Agentic, multi-step execution. Agents are doing chains of work: pulling data, drafting outputs, calling tools, asking for approval, recording decisions. The org has invested in orchestration, observability and guardrails. Human roles have shifted from doing the work to supervising it.

Stage 4 — Enterprise AI platform. A shared platform serves dozens of use cases. There is a central evaluation lab, a model garden, a governance committee, a product mindset and a clear separation between platform engineering and application teams. AI is now part of how the company builds software, not a parallel programme.

Most mid-market clients we meet are somewhere between Stage 0 and Stage 2. Most large enterprises are between Stage 1 and Stage 3, with islands at Stage 4 and pockets still at Stage 0. The honest diagnostic is to look at three things: how many use cases are in production with monitored outcomes, how much of the spend goes to platform vs point solutions, and whether non-technical staff can describe the rules under which they are allowed to use AI. If any of those answers are fuzzy, you are lower on the ladder than your slides claim.

Where you are on the ladder determines the consulting engagement you should buy. A Stage 0 organisation needs policy, literacy and quick wins, not an agentic platform. A Stage 3 organisation needs platform thinking and governance maturity, not another pilot. Misdiagnosing the stage is the most common reason engagements fail.

The end-to-end AI consulting engagement

A well-run AI engagement has six phases. They are not waterfall; they overlap, repeat and feed back into each other. But naming them helps you and your partner agree on what done looks like at each stage.

Phase 1 — Discovery and opportunity mapping. Two to four weeks of structured listening: stakeholder interviews, process shadowing, data inventory, technology audit, competitive scan. Output: an opportunity map scored on value and feasibility, an honest maturity diagnosis and a shortlist of candidate use cases.

Phase 2 — Use case prioritisation and business case. One to two weeks. For each shortlisted use case, build a thin business case: lever, expected magnitude, data needs, integration footprint, change burden, risk class. Sequence the portfolio. Decide which two or three move to build.

Phase 3 — Data and architecture readiness. Two to six weeks, often in parallel with phase 4. Confirm source systems, set up retrieval infrastructure, define data contracts, stand up the evaluation harness, choose models, design guardrails and human-in-the-loop checkpoints.

Phase 4 — Solution design and build. Four to twelve weeks per use case. Iterative sprints: prototype, evaluate, refine, integrate. The evaluation harness is the gating mechanism, not opinion. Each sprint ends with a working system measured against the metrics defined in phase 2.

Phase 5 — Deployment, adoption and change. Two to eight weeks, but really continuous. Training, communications, role redesign, incentive alignment, support model, escalation paths, success metrics dashboard. Adoption is where most pilots silently die; treat it as a workstream, not a slide.

Phase 6 — Operate, monitor and iterate. Ongoing. Drift monitoring, prompt and model refreshes, knowledge base updates, evaluation re-runs, incident handling, expansion of scope. This is where compounding value lives and where most consultancies disappear.

The single biggest delivery risk across all six phases is treating any of them as optional. Skipping discovery means building the wrong thing. Skipping evaluation means shipping a system that gets worse without anyone noticing. Skipping adoption means a beautiful piece of software no one uses. Skipping operate means a depreciating asset that quietly becomes a liability.

Discovery: framing the right problem before touching a model

A good discovery does not start with AI. It starts with outcomes. We use an outcome tree: a top-level commercial outcome decomposed into the operational outcomes that drive it, then the friction points that block each one. Only at the bottom of the tree do we ask which of those friction points are addressable with AI versus other interventions.

This matters because the natural drift of an AI programme is to start with the capability and look for problems to apply it to. That is how organisations end up with an internal chatbot no one uses and a contract analysis tool that solves a problem the legal team did not have. Working top-down from outcomes keeps the portfolio honest.

In practice, our discovery weeks contain four parallel streams. Stakeholder interviews surface stated needs and political reality. Process shadowing surfaces the unspoken friction that no interview will reveal — the rekeying, the screen-switching, the small humiliations of bad software. A data inventory establishes what is actually available, where it lives and what state it is in. A technology audit maps the existing stack, current AI investments and contractual constraints. The four streams converge into a single artefact: a prioritised opportunity map with each opportunity scored on value at stake, feasibility, time-to-value and strategic fit.

The trap to avoid in discovery is the long list with no opinions. A consulting deliverable that lists forty use cases and asks the client to pick is a failure of the consultant's job. A good opportunity map has a clear top three, a defensible reason for each and a clear no-pile of opportunities that look attractive but are not worth doing yet. The willingness to put a finger on the scale is what the engagement architect is for.

The other trap is treating discovery as a one-off. Markets, models and your own organisation move fast enough that the opportunity map should be revisited every two quarters. Build that cadence into the relationship from day one or accept that you will commission an expensive refresh exercise in eighteen months.

A practical use case prioritisation framework

With an opportunity map in hand, the next job is to choose what to build first. We use a two-by-two whose axes are value at stake and feasibility, and we score feasibility on five sub-dimensions: data availability, integration complexity, regulatory load, change burden and technical risk.

The top-right quadrant — high value, high feasibility — is the obvious start. The trap is that everyone overestimates feasibility. We force a defensive scoring pass: for each high-feasibility opportunity, ask what would have to be true for it to fail, then evidence each assumption. Anything that fails the evidence test gets demoted.

The bottom-right quadrant — high value, low feasibility — is the flagship territory. These are the opportunities worth a multi-quarter build, often involving platform investments, data remediation and meaningful change management. Treat them as programmes, not projects, and resource them accordingly.

The top-left — low value, high feasibility — is the seductive trap. It is full of quick demos that look great in showcases and produce no measurable outcome. A disciplined portfolio includes a small allocation here for learning and storytelling but never lets it dominate.

The bottom-left — low value, low feasibility — is the no-pile. Naming what you are not doing is as important as naming what you are.

Alongside the two-by-two, we layer a horizon mix: roughly 60% near-term assistants and workflow automations, 30% medium-term agentic systems and 10% longer-term platform bets. The exact mix varies by maturity, but the principle is that a portfolio of only quick wins starves the future, and a portfolio of only moonshots starves the present.

A worked example from a mid-market professional services client illustrates the point. Their initial wish list had eighteen ideas. Scored honestly, three survived the first cut: an AI-assisted proposal drafting workflow (assistant, near-term, high value), an agentic research and synthesis tool for client teams (workflow, medium-term, high value) and a knowledge platform feeding both (platform, medium-term, foundational). The other fifteen ideas were either too small to matter, dependent on a data set that did not yet exist, or were better solved by fixing a process rather than adding AI. Eighteen months later, the three survivors are in production and have funded the next wave.

Data foundations: the unsexy work that decides everything

The single largest delta between AI projects that succeed and AI projects that quietly die is the state of the underlying data. Foundation models are extraordinary general reasoners, but they will happily produce confident answers from incomplete, contradictory or out-of-date sources. The work to make that not happen is data work, not model work.

We treat data readiness in five layers.

Source-of-truth mapping. For each use case, which systems hold the canonical data? Where are the shadow copies? What is the refresh cadence? Who owns each source? It sounds elementary; it almost never is.

Retrieval design. Most enterprise generative AI use cases are some form of retrieval-augmented generation. Retrieval design — how documents are chunked, embedded, indexed and ranked — has more impact on quality than which foundation model you choose. Get this layer right and a mid-tier model outperforms a frontier model with bad retrieval.

Data contracts and lineage. When AI systems consume data from upstream sources, those sources gain new obligations. A schema change that was previously an annoyance can now silently degrade a customer-facing assistant. Data contracts make the obligations explicit; lineage makes them traceable.

Quality SLOs. Define what "good enough" means for each source: completeness, freshness, accuracy thresholds. Monitor against the SLOs. Page someone when they break. This is the discipline that converts AI from a science experiment into infrastructure.

Privacy, residency and access control. Who is allowed to ask the system what? Where does the data live? What gets logged and for how long? These questions get hand-waved in pilots and become showstoppers at production. Solve them once, centrally, and reuse the answers.

A realistic data-readiness assessment usually surfaces uncomfortable findings: that the "single customer view" is actually three views stitched together by a nightly batch job; that the knowledge base the support team relies on is a SharePoint folder last reorganised by someone who left two years ago; that the product catalogue used for personalisation is not the catalogue shown on the website. The advisory job is to surface these findings and price them honestly into the roadmap rather than pretend the model will paper over them.

Architecture choices: build, buy, compose

The architecture conversation in AI is unusually loud because the technology is moving fast and every vendor has a strong incentive to convince you their layer is the one that matters. We try to anchor architecture decisions in a small number of durable principles rather than current-quarter fashions.

Compose, don't monolith. The stack has at least five distinct layers — foundation models, orchestration, retrieval, evaluation, application. Treat them as replaceable. A monolithic vendor that owns all five sells convenience now and lock-in later.

Foundation model selection. No single model wins every task. We typically design for two or three models behind a thin abstraction: a frontier model for hard reasoning tasks, a mid-tier model for the majority of workflow tasks and an open-weights or small model for high-volume, latency-sensitive or sensitive-data tasks. Routing logic decides which is used per call.

Orchestration and agents. Frameworks like LangGraph, the OpenAI Agents SDK, Microsoft's Semantic Kernel and a handful of others compete here. Choose for engineering ergonomics and observability rather than feature surface. The framework you can debug at 2am is the right one.

RAG vs fine-tuning vs prompt engineering. The decision tree is simpler than it looks. Prompt first. Add retrieval when the answer needs facts the model does not have. Fine-tune only when you have a stable, narrow task with enough labelled data and have exhausted the cheaper options. Most projects never need fine-tuning; the ones that do tend to know it from the start.

Avoid vendor lock-in without paralysis. Pure best-of-breed is exhausting; pure single-vendor is dangerous. The middle path is to pick a primary platform for the boring parts (identity, data plane, basic assistants) and remain composable at the application layer where differentiation lives.

Our reference architecture for mid-market clients runs along these lines: identity and data governance on the existing Microsoft or Google estate; a thin orchestration layer in TypeScript or Python deployed on the client's existing cloud; retrieval against a managed vector store; multi-model routing through a gateway with logging, tracing and cost monitoring; evaluation harness as code; application interfaces inside the tools people already use (Teams, Slack, Salesforce, the CRM) rather than yet another tab to open. It is not glamorous but it ships, it survives a vendor change and it is debuggable.

Generative, predictive and agentic AI: when each earns its place

The public conversation conflates three different things under the AI umbrella. Each has its own strengths, failure modes and consulting implications.

Generative AI produces new artefacts — text, code, images, audio, structured data — conditioned on inputs. It excels at synthesis, drafting, translation and pattern reproduction. It fails when accuracy on specific facts is required without retrieval, when outputs must be verifiable against ground truth and when the cost of a confident-but-wrong answer is high. Use it for first drafts, summarisation, ideation, classification and any task where a human will review the output.

Predictive AI estimates a value or class from features — churn likelihood, demand forecast, fraud probability, equipment failure window. It has been in production for decades under the names statistics, machine learning and data science. It excels at narrow, well-defined prediction tasks with enough labelled data. It fails when the world drifts away from the training distribution and when stakeholders cannot interpret the outputs. Use it for forecasting, scoring, optimisation and recommendation.

Agentic AI chains multiple model calls and tool uses to accomplish multi-step tasks: read an email, look up a record, draft a reply, file a ticket, escalate if needed. It excels at workflows that previously required a human to coordinate several systems. It fails spectacularly when steps cascade without checkpoints, when tools return ambiguous results and when the cost of an error compounds across steps. Use it carefully, with human approval gates on consequential actions, narrow tool surfaces and strong observability.

The most valuable systems we build for clients are usually hybrids. A claims triage system uses predictive AI to score severity, generative AI to draft the customer communication and agentic patterns to coordinate the human handoff. A sales workflow uses predictive AI to score lead intent, generative AI to draft the outreach and an agent to schedule, log and follow up. Treating the three as separate disciplines competing for the same budget is a mistake; treating them as complementary tools in one toolbox is how mature programmes are built.

AI in the front office: marketing, sales and service

Front-office AI is where revenue lives and where the customer's perception of your brand is forged. It is also where the reputational risk of getting AI wrong is highest. The principles below have survived contact with enough programmes to be worth stating.

Personalisation at scale, without the creepiness. Generative AI makes it cheap to produce variants of content, emails, offers and pages. The constraint is no longer production capacity; it is judgement about what the customer actually wants to receive. The best programmes invest in segmentation logic and editorial guardrails first, then in generation. Volume without judgement is just spam at lower marginal cost.

AI-assisted content production. A well-run content operation now treats writers and designers as editors and curators of AI output, not as producers from scratch. The role shift is real; the editorial standards must not slip. Style guides, fact-checking processes and approval workflows become the centre of gravity. The teams that ship best have explicit policies on which content types are AI-assisted, which are human-only and how AI involvement is disclosed.

Sales copilots and call intelligence. Reps spend a remarkable amount of their week on research, note-taking and CRM hygiene. AI copilots that prepare call briefs, transcribe and summarise calls, draft follow-ups and update CRM fields buy back hours per rep per week. The trap is rolling these out as standalone tools rather than weaving them into the existing sales cadence. The ones that stick are invisible inside the tools reps already use.

Tier-zero customer service. Self-service powered by retrieval and generation now handles a meaningful share of customer queries for well-implemented programmes. The discipline that separates good from bad is the escalation path: AI handles what it can confidently answer, hands off to a human at clear thresholds and never traps a customer in a loop. The metric to watch is not pure deflection; it is first-contact resolution including AI, which incentivises the system to escalate well rather than deflect aggressively.

Measuring outcomes. Front-office AI should be measured against the metrics the business already cares about: conversion, retention, CSAT, NPS, average handle time, revenue per customer. If a programme is only measured against tool-specific KPIs (queries answered, content pieces produced) it is not measuring value.

AI in the back office: finance, HR, legal and operations

Back-office AI is less glamorous and often more valuable. The work is repetitive, document-heavy, rule-bound and ripe for augmentation. The buyers are usually risk-aware and willing to invest in evaluation and governance — which is exactly the discipline AI needs.

Legal and contracts. Contract abstraction, clause libraries, redline suggestions and obligation tracking are now table stakes. The best implementations have lawyers working alongside AI rather than supervising it at arm's length: AI surfaces, humans decide. The ROI shows up as faster turnaround on routine agreements and capacity for legal teams to focus on bespoke work.

Finance and procurement. Spend classification, anomaly detection, forecast commentary, scenario modelling and procurement copilots all have well-trodden patterns. Finance functions are usually conservative in adoption but precise in their requirements, which leads to robust implementations once the political work is done.

HR and talent. Skills inference from CVs and project histories, internal mobility recommendations, training pathway personalisation and policy assistants for managers are the use cases that consistently deliver. Bias and fairness considerations are non-negotiable here; we always design with vendor-independent evaluation and clear human decision points.

Operations and supply chain. Demand forecasting, inventory optimisation, route planning and predictive maintenance have been ML use cases for years. Generative AI adds value at the edges: turning forecasts into commentary executives can read, drafting supplier communications, summarising incident reports. The wins come from connecting predictive and generative layers, not replacing one with the other.

Risk and compliance. Policy assistants, control testing automation, regulatory horizon scanning and audit support are all maturing fast. The trick is to use AI to augment second-line functions, not to replace the human judgement that regulators expect.

Industry deep dives

The shape of an AI programme varies meaningfully by industry. The same architecture will not survive the move from a retail bank to a hospital trust to a manufacturing plant. Below is a condensed view of how the priorities shift.

Financial services. Heavy regulation, mature data estates and high transaction volumes make financial services a natural home for AI at scale. The highest-leverage use cases sit in KYC and onboarding (document handling, identity verification, ongoing monitoring), advisory and wealth (copilots for relationship managers, personalised communications), fraud and AML (pattern detection augmented by generative reasoning over flagged cases) and operations (reconciliation, exceptions handling, regulatory reporting). The constraints are model risk management frameworks, explainability requirements and audit trails — none of which are blockers if designed in from the start.

Retail and consumer. The opportunity surface is wide: assortment planning, demand forecasting, dynamic pricing, content production for thousands of SKUs, search relevance, recommendation, customer service, store operations. The winning programmes pick one core lever (often search and discovery or content production) and execute it deeply rather than spreading thinly. Brand voice and editorial control are the recurring stress point.

Manufacturing and supply chain. Predictive maintenance, quality inspection, planning optimisation and field service support are mature ML domains gaining generative augmentation — for example, turning sensor anomalies into natural-language work orders for technicians. The constraints are connectivity at the edge, data quality from older equipment and the change management of shop-floor adoption.

Professional services. Law firms, accounting practices, consultancies, agencies and architecture practices live or die by knowledge work productivity, which is precisely what generative AI accelerates. Proposal drafting, research synthesis, document review, knowledge management and timesheet support are all proven use cases. The honest tension is that AI compresses the billable pyramid, which is a strategic question, not a tooling one.

Healthcare and life sciences. Clinical applications require regulatory pathways and careful evidence; administrative and research applications are open ground. Documentation assistants for clinicians, prior-authorisation support, research literature synthesis, patient communication and operational scheduling all have well-trodden patterns. Patient safety, data protection and equity considerations are central, not optional.

Public sector. Casework support, citizen service assistants, fraud detection, policy analysis and internal knowledge management are the dominant patterns. Transparency, accessibility, procurement constraints and democratic accountability shape every decision. The pace is slower; the impact when delivered well is enormous.

The consulting implication across all six is that industry context is not optional knowledge. A team that has shipped in your sector knows which use cases stall, which regulators care about what and which vendors actually deliver. Pay for that.

Responsible AI, governance and assurance

Responsible AI is no longer a values statement bolted onto a programme. It is an operating discipline with concrete artefacts, owners and cadences. Done well, it accelerates adoption by making the safe path the default path. Done poorly, it becomes a slow committee that everyone routes around.

We build governance against three reference points: ISO/IEC 42001 for AI management system structure, the EU AI Act risk classification for regulatory alignment and the NIST AI Risk Management Framework for risk language and controls. None of these is a checklist; together they form a defensible baseline.

The practical artefacts that matter are these.

A use case inventory with risk classification, owner, status and review date. Every AI use case in the organisation lives on it. If it is not on the inventory, it is shadow AI.

Model cards and system cards documenting purpose, data sources, intended and prohibited uses, evaluation results, known limitations and human oversight design. Updated when material changes happen.

An evaluation harness with task-specific test sets, automated metrics and periodic human review. Evaluations run on every release and on a schedule against production traffic samples to detect drift. The harness is the single most important investment most clients underweight.

Red teaming and adversarial testing for high-risk use cases, with findings tracked to closure. External red teams for the most consequential systems.

Human oversight design. Each use case has explicit answers to: who can override the system, who reviews aggregate behaviour, who is accountable for outcomes, what is logged and for how long, and what triggers a roll-back.

A governance forum with representation from legal, risk, security, data protection, the relevant business functions and engineering. It meets often enough to keep up — usually fortnightly in the build phase, monthly thereafter — and has the authority to stop a launch.

The organisations that get this right treat governance the same way mature engineering organisations treat security: a partner to product, not a blocker. The ones that get it wrong either have no governance until a regulator turns up, or have so much governance that nothing ships. Aim for the middle.

Change management and adoption

Most AI pilots do not die because the model is wrong. They die because no one uses the thing. Adoption is where strategy meets reality, and it deserves the same rigour as any other workstream.

The pattern we see in failed adoption is consistent: a beautifully engineered tool launched into an organisation that has not changed its processes, incentives, training or communications. People default to the way they have always worked because that is what the rest of the system rewards. The fix is not louder announcements; it is to design the surrounding operating model deliberately.

The components of a serious adoption programme are these.

Operating model redesign. Which steps in the workflow are now done by AI, which by humans and which jointly? What do the new role descriptions look like? What is the new definition of done? Without explicit answers, individuals improvise and reversion is inevitable.

Reskilling and capability. Generic AI training is a waste of time. Role-specific training in the tools that role uses, with practice problems from that role's actual workflow, is where adoption is built. The most effective formats are short, repeated and embedded in the work itself rather than separate courses.

Communications. A consistent story from leadership about why the change is happening, what is in it for the individual and how the organisation will support people through the transition. Silence is read as threat.

Incentives. The metrics on which people are measured must reward the new way of working. If the sales rep is still measured on dials made, they will not adopt a tool that reduces dials in favour of better-targeted conversations. Align incentives or accept slow adoption.

Champion networks. Practitioners across the business who use the tools heavily, support their peers and feed insights back to the central team. The champion network is the bridge between the AI programme and the rest of the organisation. Resource it explicitly.

Support and feedback loops. Clear channels for users to report issues, request features and share workarounds. A visible queue with visible response times. The fastest way to kill adoption is to ignore feedback for a quarter.

A useful sanity check: when adoption is going well, you hear stories. Stories about a deal won, a customer saved, a week given back, a manager who was sceptical now evangelising. When adoption is failing, you hear metrics about logins and queries with no narrative behind them. Listen for the stories.

Measuring value: a payback model that survives the board

The value conversation in AI gets distorted by two opposite errors. The first is vanity measurement: counting queries, users and pieces of content produced as if they were value. The second is paralysis by attribution: refusing to claim any value until causality is proven beyond reasonable doubt. Both are wrong. The board needs a measurement discipline that is honest, defensible and decision-useful.

We split AI value into three lever families.

Productivity levers. Time saved per task, throughput per person, capacity unlocked. The honest measurement requires before-and-after task instrumentation, not survey self-reports. Survey self-reports systematically overstate gains by a factor of two to three.

Quality levers. Error rates, customer satisfaction, decision quality, compliance findings. Quality gains often matter more than productivity gains because they unlock revenue or avoid losses that productivity alone cannot.

Revenue levers. Conversion, retention, cross-sell, new product enabled. Revenue attribution to AI is the hardest case and requires controlled rollouts, holdouts or rigorous before-and-after analysis.

For each use case, we define leading and lagging indicators. Leading indicators (usage depth, evaluation scores, user feedback) tell you within weeks whether the programme is on track. Lagging indicators (the business metric) tell you within quarters whether it actually moved the number. Reporting both keeps short-term confidence and long-term honesty aligned.

Time-to-value varies by use case class. Assistant rollouts typically show measurable productivity gains within one to two quarters. Workflow automations show ROI within two to four quarters once adoption stabilises. Agentic systems and platform investments often take three to six quarters before the compounding effect is visible. Setting these expectations upfront protects programmes from being killed in their second quarter for failing to deliver fourth-quarter results.

The most important measurement discipline is the reinvestment loop. Productivity gained must be deliberately reinvested — into higher-value work, into capacity for growth, into the next wave of use cases — or it leaks back out as slack. The programmes that compound do so because leadership treats unlocked capacity as a strategic asset to deploy, not as a cost line to harvest.

Common AI consulting pitfalls and how to avoid them

A shortlist of the failure patterns we see most often, with the antidotes.

Buying a platform before defining a problem. A platform purchased without a portfolio of named use cases becomes a sunk cost in search of justification. Define the portfolio, then choose the platform that fits.

Underestimating data and integration effort. The model is 20% of the work. Data, integration and operations are the other 80%. Build the plan around the 80%.

Overweighting demos in vendor selection. Demos are scripted. Reference customers are curated. The only honest test is a paid bake-off on your own data with your own evaluation criteria. Insist on it.

Ignoring evaluation infrastructure. Without evaluations, you do not know whether the system is getting better or worse over time. Build the harness in week one, not week twelve.

Treating governance as an afterthought. Governance designed at the end of a programme is always either too restrictive (and slows adoption) or too permissive (and embarrasses leadership). Design it in from the start.

Pyramid pricing on AI work. Be wary of partners who quote teams of eight juniors to one senior. The economics no longer require it and you are paying for hours rather than outcomes.

Skipping operate. A pilot is not a product. Without an operate model, AI systems silently degrade as the world changes around them. Budget for ongoing operations from day one.

Single-vendor lock-in. The model market moves fast enough that whichever vendor is best today will not be best in two years. Compose your stack so you can swap layers.

Ignoring the human cost. Roles will change. Some jobs will shrink. Pretending otherwise damages trust. Honest communication, supported transitions and credible reskilling are non-negotiable.

Treating AI as an IT programme. AI is a business transformation enabled by technology. Run it with business leadership in the room, not just technology leadership.

How to choose an AI consulting partner

The market is crowded and the brochures all look similar. A short list of questions cuts through most of the noise.

**Can they advise, build and operate?** Or are they only advisors who will hand off to someone else for the hard parts? Each handoff is a tax on velocity and a place where strategy and execution can drift apart.

Show me a system in production. Not a demo, not a case study slide, an actual system with traffic. Walk through the architecture, the evaluation harness, the incident history. The conversation reveals depth quickly.

Who actually does the work? Sometimes the names on the proposal are not the names on the project. Insist on meeting the engagement architect and AI facilitators who will deliver, and on continuity through the programme.

What is their model portfolio? A partner with strong opinions about one foundation model and weak fluency with the rest will design around their bias. You want partners who routinely deploy across the major model families.

How do they evaluate? Ask to see their evaluation framework. If they cannot describe it crisply, they do not have one. Without evaluation, you have no idea what you are buying.

Industry context. Have they delivered in your sector? Not adjacent sectors — yours. Industry-specific risk, language and stakeholder dynamics matter.

Commercial alignment. Are they incentivised by outcomes or by hours? Where does the risk sit if the programme underdelivers? The cleanest engagements are the ones where the partner has skin in the game.

Cultural fit and knowledge transfer. Will your team be more capable at the end of the engagement than at the start? A good partner deliberately transfers capability rather than building dependency.

The firms that pass all of these tests are a small subset of the market. They are worth finding.

How iCentric delivers AI and consulting

Our practice is built around the three layers we described at the start — advisory, build and operate — delivered by integrated pods rather than handed off between functions.

A typical pod blends an engagement architect, two AI engineers, a data engineer, a product designer and a delivery lead. Pods are small on purpose. Tools do the leverage; people do the judgement. The pod stays with the client across phases, which removes the handoff cost that traditional firms charge for and accelerates learning compounding.

Our advisory work is time-boxed. Discovery engagements run four to six weeks and end with a decision-ready opportunity map, a portfolio sequenced for build and a maturity diagnosis the leadership team can act on. We do not sell six-month strategy projects when the answer can be reached in six weeks.

Our build work runs in sprints against an evaluation-first cadence. Every sprint produces a working system measured against agreed metrics. We use the reference architecture described earlier — composable, multi-model, observability-rich — because it ships and survives.

Our operate offering picks up where build ends. Drift monitoring, knowledge refreshes, evaluation reruns, model upgrades, incident handling and scope expansion all sit inside a managed service that keeps systems healthy and improving. Clients who care about compounding value run AI as a continuous capability, not a project.

Across all three layers, we are deliberate about knowledge transfer. By the end of a programme your internal team should be able to extend, debug and operate the systems we have built together. If you would prefer us to continue operating them at scale, that option is on the table — but it should be a choice rather than a dependency.

If any of this maps to a problem you are sitting with, the next step is a short conversation. We will tell you honestly whether we are the right partner for what you need, and if we are not, we will tell you who is. Get in touch via the contact page or read more about our AI automation services and digital transformation consulting for context on how the practice fits together.

Frequently asked questions

What does an AI consultancy actually do? A good AI consultancy helps you frame the right problems, build working systems and operate them at scale. The mix between advisory, build and operate varies by partner and by engagement; the better partners are credible across all three.

How long before we see results? Assistant rollouts and well-scoped workflow automations show measurable results within one to two quarters. Agentic systems and platform investments tend to show value within three to six quarters. Anyone promising transformational results in weeks is selling demos rather than outcomes.

Do we need a data lake before we start? No. You need enough data hygiene around the specific use cases you are starting with. Boil-the-ocean data programmes are a classic way to delay value indefinitely. Start narrow, prove value, expand the data foundation as the portfolio grows.

Should we use ChatGPT, Copilot or build our own? Almost always: use sanctioned assistant subscriptions for general productivity and build workflow-specific systems on top of foundation model APIs for the use cases where you need control. The answer is rarely either-or.

How do we keep humans in the loop responsibly? Define for each use case which actions require human approval, which are reviewed periodically and which are fully automated. Log decisions. Monitor aggregate behaviour. Empower users to override and feed those overrides back into improvement cycles. Responsible AI is mostly about good system design, not about restraint.

How is AI consulting different from digital transformation? Digital transformation is the broader programme of modernising how an organisation works using technology. AI consulting is a focused capability within that programme, with its own technical depth, governance requirements and operating disciplines. The best programmes treat AI as one stream within a coherent transformation, not as a separate initiative competing for the same change capacity.

What does an AI consultancy actually do?

A good AI consultancy helps you frame the right problems, build working systems and operate them at scale. The work spans three layers: advisory (strategy, opportunity mapping, business cases), build (data, models, integration, evaluation, applications) and operate (monitoring, refreshes, incident handling, scope expansion). The better partners are credible across all three rather than handing off between them.

How long before we see results from AI consulting?

Assistant rollouts and well-scoped workflow automations typically show measurable results within one to two quarters. Agentic systems and platform investments usually show value within three to six quarters as adoption stabilises and reinvestment loops kick in. Anyone promising transformational results in a few weeks is selling demos rather than outcomes.

Do we need a data lake before we start with AI?

No. You need enough data hygiene around the specific use cases you are starting with, not a perfect enterprise data platform. Boil-the-ocean data programmes are a classic way to delay value indefinitely. Start narrow, prove value on a small portfolio of use cases and expand the data foundation deliberately as the portfolio grows.

Should we use ChatGPT, Copilot or build our own AI?

Almost always the answer is both. Use sanctioned assistant subscriptions such as Microsoft Copilot, Google Gemini or ChatGPT Enterprise for general productivity, and build workflow-specific systems on top of foundation model APIs for the use cases where you need data integration, evaluation and control. The decision is rarely either-or and a sensible architecture composes layers from multiple vendors.

How do we keep humans in the loop responsibly?

Define for each use case which actions require human approval, which are reviewed periodically and which are fully automated. Log decisions, monitor aggregate behaviour and give users a clear way to override the system, then feed those overrides back into improvement cycles. Responsible AI is mostly about deliberate system design and clear accountability rather than blanket restraint.

How is AI consulting different from digital transformation?

Digital transformation is the broader programme of modernising how an organisation works using technology, covering process, data, platforms, ways of working and culture. AI consulting is a focused capability within that programme with its own technical depth, governance requirements and operating disciplines. The best programmes treat AI as one stream within a coherent transformation, not as a separate initiative competing for the same change capacity.

ai and consulting

Get in touch today

Book a call at a time to suit you, or fill out our enquiry form or get in touch using the contact details below

iCentric
May 2026
MONTUEWEDTHUFRISATSUN

How long do you need?

What time works best?

Showing times for 18 May 2026

No slots available for this date