The L&D leader's playbook for rolling out AI training to a global workforce

The head of L&D was given twelve months to get 2,000 people "AI-ready." She had a budget, an executive sponsor…

Ijan Kruizinga·

Start with the result, not the curriculum

The first mistake most L&D leaders make is shopping for content before defining the outcome. They get a pitch from a vendor, see a slick demo, and start mapping modules to job families. Six months later they have a beautifully structured catalogue and no measurable change in how anyone works.

A good AI training rollout starts the other way around. What does the business actually want? Higher Copilot license utilisation? Faster cycle times in claims processing? Fewer manual handoffs in finance? An AI-fluent engineering org that can ship agentic features? Each of those needs a different curriculum, a different cohort design, and a different set of metrics. None of them are solved by "general AI awareness for all staff."

Before you talk to a single vendor, get your executive sponsor in a room and force the question: what does success look like in twelve months, in numbers? If the answer is fuzzy, the rollout will be fuzzy. We've written more about how to think about this in measuring ROI on enterprise AI training, and the short version is that the metric you choose at the start determines the program you end up with.

Segment your workforce honestly

The second mistake is treating the workforce as one audience. "AI training for all employees" sounds democratic and is operationally useless. A risk analyst in Sydney, a data engineer in Bangalore, a customer service lead in Manila, and a board director in London need radically different things from your program.

A workable segmentation for most enterprises looks something like this:

  • Executives and board. Strategic literacy. What AI can and can't do, how to govern it, how to ask the right questions. Half a day, in person where possible.

  • People leaders. How to redesign work, spot AI use cases in their team, and manage the change. One to two days, with follow-up.

  • Knowledge workers. Practical fluency in the tools you've licensed (Copilot, Gemini, Claude, internal agents). Workflow redesign, not feature tours. Two to four hours, role-specific.

  • Technical builders. Engineers, data scientists, ML platform teams. Deep technical training on the platforms you actually use. Multi-week, hands-on.

  • High-risk roles. Legal, compliance, audit, fraud, anyone making decisions where AI errors have material consequences. Specialised training on AI risk, model limitations, and verification workflows.

That's five distinct programs, not one. The mistake is trying to make a single curriculum stretch across all of them. The other mistake is buying five off-the-shelf courses and calling it strategy. We've written separately on when off-the-shelf training is the right call and when it isn't.

Sequence by impact, not by org chart

Once you've segmented, the next decision is sequencing. Most L&D leaders default to either alphabetical (start with the loudest division) or political (start with the executive who sponsored the budget). Neither produces results.

Sequence by impact. Which cohort, if trained well in the next 90 days, will produce a visible business outcome that justifies the rest of the program? Usually it's one of three:

  1. A team with high license spend and low utilisation, where training pays for itself in tool ROI.

  2. A team with a clear, measurable workflow that AI can compress (claims, contract review, ticket triage, code review).

  3. A team that's already trying to do something with AI and is stuck, where training unblocks an in-flight initiative.

Train that cohort first. Measure the outcome. Use it to fund and de-risk the rest of the rollout. The Microsoft research on Copilot adoption patterns is consistent on this point: the teams that get value early are the ones with structured enablement tied to specific workflows, not the ones with the broadest training catalogue.

Build a governance layer before you scale

This is the boring part of the playbook and the part that separates real rollouts from training theatre.

A global AI training rollout touches data privacy, IP, model risk, vendor management, and in regulated industries, prudential or sectoral compliance. If your training tells engineers to paste production data into a public model, you've created a problem your CISO will not thank you for. If your training tells legal to "use AI to draft contracts" without addressing verification workflows, you've created a different problem your general counsel will not thank you for.

Before you scale, get four things written down:

  • Acceptable use. Which tools, for which tasks, with which data classifications. Make this concrete. "Use approved tools responsibly" is not a policy.

  • Verification workflows. When AI output needs human review, who reviews it, and what does the audit trail look like.

  • Vendor and model inventory. What's approved, what's pending, what's prohibited. Update it monthly.

  • Incident pathway. What happens when AI gets something materially wrong. Who's notified, what's logged, how is it learned from.

The training program then teaches into this governance, not around it. We cover this in more depth in enterprise AI risk management. The point for L&D is that you cannot outsource this to legal or risk and expect them to do it for you. You have to drive it because the training won't work without it.

Choose between train-the-trainer and external delivery deliberately

For a global rollout, the delivery model matters as much as the curriculum. There are three real options:

  • Pure external. A vendor delivers every cohort. Highest quality control, highest cost, slowest to scale.

  • Train-the-trainer. A vendor trains your internal facilitators, who then deliver to the workforce. Lower cost at scale, faster to roll out, quality depends entirely on facilitator selection and ongoing support.

  • Hybrid. External delivery for high-stakes cohorts (executives, technical builders, high-risk roles), train-the-trainer for broad knowledge worker rollout.

For most enterprises above 5,000 staff, hybrid is the right answer. The trick is being honest about which cohorts can be served by an internal facilitator with two days of vendor prep, and which can't. A workshop on Copilot fundamentals for a marketing team can be run by a trained internal facilitator. A program on building production AI agents for a Databricks engineering team cannot.

If you're going train-the-trainer, the failure mode to watch for is facilitator drift. Six months in, your trainers have made the content their own, simplified the bits they didn't fully understand, and quietly dropped the parts they found uncomfortable. Build in a refresh cycle and a quality audit. Otherwise the program degrades silently.

Localise without diluting

For a global workforce, localisation is real but often overdone. The genuine localisation needs are:

  • Language, where business English isn't the working language

  • Regulatory context, especially around privacy (GDPR in Europe, Privacy Act in Australia, sector rules in financial services)

  • Tool stack, where regional teams use different licensed products

  • Time zones for live delivery

What you don't need to localise is the core methodology. The principles of prompting, verification, workflow redesign, and governance are the same in Sydney, Singapore, and São Paulo. Resist the pressure from regional leaders to build "their" version of the program. You'll end up with twelve curricula, none of them maintained.

Measure what you said you'd measure

Twelve months in, your CFO will ask what they got for the spend. The answer "we trained 40,000 people and average satisfaction was 4.6" is the wrong answer.

The right answer connects back to the outcome you defined at the start. If the goal was license utilisation, show the utilisation curve before and after, by trained cohort versus untrained control. If the goal was cycle time in a specific workflow, show the cycle time. If the goal was AI-driven product features shipped, show the feature count and the engineering hours saved.

Three traps to avoid:

  • Smile sheets as the primary metric. Satisfaction scores tell you whether people enjoyed the session. They tell you nothing about whether anyone changed how they work.

  • Completion rates as the primary metric. "92% completed the module" is a measure of how compliant your LMS reminders are, not how capable your workforce is.

  • Self-reported confidence. "85% feel more confident using AI" is the most flattering and least useful number in L&D.

Measure behaviour change and business outcomes. They're harder to capture, which is why most rollouts don't bother, which is why most rollouts can't justify their next budget cycle.

Plan for the second wave before the first one finishes

The last piece of the playbook: AI capability is not a one-and-done program. The tools, the models, and the use cases will look meaningfully different in twelve months. Whatever you teach this year will be partially obsolete next year.

Build the rollout assuming a refresh cadence. Quarterly updates for technical builders, half-yearly for knowledge workers, annual for executives. Bake it into the budget now, because trying to find new money for "AI training, again" in year two is a fight you don't want to have.

The L&D leaders who get this right in the next eighteen months will define what AI fluency means in their organisation for the rest of the decade. The ones who run a generic awareness campaign and tick the box will be back in front of the board in 2026 explaining why nothing changed. If you're starting now, the work is choosing which one you'll be. If you want to talk through what your specific rollout should look like, get in touch.

Ijan Kruizinga

Co-founder of Better People. 20+ years across technology and marketing leadership. Previously CEO of Crucial, CEO/COO of OMG and Jaywing.

Ready to talk?

30-minute discovery call.