Back to Insights
AI Change Management & TrainingFramework

Designing an AI Training Program: A Framework for L&D Leaders

November 16, 202510 min readMichael Lansdowne Hauge
Updated March 15, 2026
For:CHROConsultantCEO/FounderCFOCTO/CIOHead of OperationsIT Manager

Comprehensive framework for AI training program design covering audience segmentation, curriculum development, delivery methods, and effectiveness measurement.

Summarize and fact-check this article with:
Australian Malaysian Collab - ai change management & training insights

Key Takeaways

  • 1.Design comprehensive AI training programs for the workforce
  • 2.Structure training across foundational and role-specific levels
  • 3.Select appropriate training modalities and delivery methods
  • 4.Build internal AI training capability and resources
  • 5.Scale AI training effectively across the organization

AI adoption succeeds or fails on people, not technology. Organizations that invest in training see adoption rates 3-significantly higher than those that don't. This guide provides a framework for designing effective AI training programs.

Executive Summary

  • AI training isn't IT training—it requires understanding of capabilities, limitations, and judgment
  • Different audiences need different training: executives, managers, practitioners, specialists
  • Balance conceptual understanding with hands-on practice—neither alone is sufficient
  • Training should address fear and resistance, not just skills
  • Measure effectiveness beyond completion rates: adoption, quality of use, business impact
  • Continuous learning matters more than one-time training as AI evolves rapidly
  • Vendor training is rarely sufficient—supplement with internal and third-party resources
  • Budget 10-20% of AI implementation cost for training and change management

Why This Matters Now

Most AI implementations underperform expectations. The common thread? Insufficient investment in helping people actually use the AI effectively.

Training addresses multiple barriers:

  • Skill gaps: People don't know how to use AI tools
  • Fear: People worry about job security or making mistakes
  • Skepticism: People don't trust AI outputs
  • Habit: People default to old ways of working

Without training, you've bought software that sits unused or gets misused.

Definitions and Scope

AI Training: Education and skill development for using AI tools effectively and responsibly.

AI Literacy: Baseline understanding of AI capabilities, limitations, and implications.

Upskilling: Developing new capabilities in existing employees.

Scope of this guide: Designing training programs for commercial AI tool adoption—not training data scientists or AI developers.


Training Needs Assessment

Step 1: Define Audiences

Audience segmentation:

AudienceCharacteristicsTraining Focus
ExecutivesStrategic decision-makersAI strategy, risk, governance, investment decisions
ManagersOperational leadersTeam management, use case identification, change leadership
PractitionersDaily AI usersTool proficiency, effective prompting, quality judgment
SpecialistsPower users, adminsAdvanced features, troubleshooting, optimization
EveryoneAll employeesAI literacy, acceptable use, policy awareness

Step 2: Assess Current State

Skills assessment methods:

  • Self-assessment surveys
  • Skills testing
  • Manager input
  • Usage data analysis

Assessment dimensions:

  • AI awareness and literacy
  • Tool-specific proficiency
  • Critical evaluation skills
  • Responsible use understanding

Example skills matrix:

Skill AreaBeginnerIntermediateAdvanced
Understanding AI capabilitiesAwarenessCan evaluateCan strategize
Using AI toolsBasic promptsEffective promptingAdvanced techniques
Evaluating outputsAccepts without reviewBasic verificationCritical evaluation
Responsible useAware of policyFollows guidelinesChampions practices

Step 3: Define Learning Objectives

For each audience, define:

  • What they need to know (knowledge)
  • What they need to do (skills)
  • How they should think (mindset)

Example objectives for practitioners:

After completing training, participants will be able to:

  1. Explain what AI can and cannot do effectively (knowledge)
  2. Construct prompts that generate useful outputs (skill)
  3. Evaluate AI outputs for accuracy and bias (skill)
  4. Apply organizational AI policies in daily work (skill)
  5. Make informed decisions about when to use AI vs. other approaches (mindset)

Curriculum Design

AI Literacy (Everyone)

Duration: 1-2 hours Format: Online, self-paced

Topics:

  • What AI is (and isn't)
  • Capabilities and limitations
  • How AI makes decisions
  • AI risks and safeguards
  • Company AI policy and expectations
  • When to use and when not to use AI

Executive AI Training

Duration: Half-day workshop Format: Facilitated session with case studies

Topics:

  • AI strategic landscape and trends
  • Business applications and ROI
  • Risk and governance considerations
  • Board and regulatory expectations
  • Decision-making frameworks for AI investment
  • Leading AI adoption

Manager Training

Duration: Full-day or two half-days Format: Workshop with practice exercises

Topics:

  • AI literacy foundations
  • Identifying AI opportunities in your area
  • Managing AI-augmented teams
  • Change management for AI adoption
  • Measuring AI effectiveness
  • Coaching team members on AI use

Practitioner Training

Duration: 1-2 days depending on tool complexity Format: Hands-on workshop with exercises

Topics:

  • Tool-specific training (navigation, features)
  • Effective prompting and interaction
  • Understanding and evaluating outputs
  • Quality control and verification
  • Exception handling
  • Workflow integration
  • Responsible use in practice

Specialist Training

Duration: 2-3 days plus ongoing Format: Technical workshop with lab exercises

Topics:

  • Advanced tool features
  • Configuration and customization
  • Integration and administration
  • Performance monitoring
  • Troubleshooting
  • Training and supporting other users

Delivery Methods

Method Comparison

MethodBest ForAdvantagesLimitations
In-person workshopComplex skills, interactionEngagement, hands-on practiceCost, scheduling
Live virtualGeographically distributed teamsConvenience, interactionEngagement challenges
Self-paced onlineFoundational knowledge, scaleFlexibility, consistencyLimited practice
MicrolearningReinforcement, just-in-timeFits into workflowNot for complex skills
Coaching/mentoringAdvanced skills, behavior changePersonalized, effectiveResource-intensive
Practice labsHands-on skillsSafe environment to experimentSetup complexity

For most AI rollouts:

AudiencePrimary MethodSupplement
All employeesSelf-paced onlineTeam discussions
ExecutivesFacilitated workshop1:1 coaching
ManagersFacilitated workshopPeer learning
PractitionersHands-on workshopPractice labs, microlearning
SpecialistsTechnical workshopVendor training, certification

RACI Example: AI Training Program

ActivityL&DITHRBusiness UnitVendor
Training needs assessmentRCCAI
Curriculum designRCICC
Content developmentRCICC
Platform setupCRIIC
SchedulingRIACI
FacilitationRCIIC
EvaluationRICAI
OptimizationRCCCI

R = Responsible, A = Accountable, C = Consulted, I = Informed


Addressing Fear and Resistance

Common Concerns

ConcernRoot CauseTraining Response
"AI will take my job"Job security fearShow AI as augmentation, not replacement
"I'll make mistakes"Competency fearSafe practice environment, permission to fail
"I don't trust AI"SkepticismTeaching critical evaluation, showing limitations
"It's too complicated"ConfidenceGradual skill building, quick wins
"Why change?"Habit, comfortDemonstrate clear value, peer examples

Tactics for Resistance

  1. Acknowledge concerns directly
  • Don't dismiss fears
  • Create space to discuss
  1. Start with low-stakes wins
  • Build confidence gradually
  • Celebrate early successes
  1. Use peer champions
  • Early adopters who share positive experiences
  • Peer learning is more credible than corporate messaging
  1. Demonstrate personal value
  • Show time savings
  • Highlight reduced tedious work
  • Frame as capability expansion
  1. Provide safety net
  • Permission to make mistakes during learning
  • Support resources readily available
  • No penalties for learning-phase errors

Measuring Training Effectiveness

Kirkpatrick Model Applied to AI Training

LevelWhat to MeasureHow to Measure
ReactionParticipant satisfactionPost-training surveys
LearningKnowledge/skill acquisitionAssessments, skill tests
BehaviorOn-the-job applicationUsage data, manager observation
ResultsBusiness impactProductivity metrics, quality metrics

Specific Metrics

MetricTargetMeasurement Method
Training completion rate>90%LMS data
Knowledge assessment pass rate>80%Post-training quiz
Tool adoption rate>70%Usage analytics
Effective usage (quality)Defined per toolOutput review, manager assessment
Time savingsPer business caseTime tracking, surveys
Employee confidenceIncreaseBefore/after survey

Implementation Checklist

Planning:

  • Defined training audiences
  • Completed needs assessment
  • Established learning objectives
  • Designed curriculum by audience
  • Selected delivery methods
  • Allocated budget and resources

Development:

  • Created or sourced content
  • Set up learning platform
  • Developed assessments
  • Prepared facilitators
  • Created practice environments

Delivery:

  • Scheduled training sessions
  • Communicated to participants
  • Delivered training by audience
  • Collected feedback
  • Provided support resources

Evaluation:

  • Analyzed completion and satisfaction
  • Measured knowledge acquisition
  • Tracked adoption and usage
  • Assessed business impact
  • Identified improvements

Tooling Suggestions

Learning Management Systems: For delivery, tracking, and reporting Video platforms: For recording and async delivery Practice environments: Sandbox AI environments for safe practice Assessment tools: For knowledge and skill testing Survey tools: For feedback collection Analytics: For usage tracking and behavior analysis


FAQ

Q: How much should we budget for AI training? A: Plan for 10-20% of AI implementation budget. More for complex tools or significant change.

Q: Should we use vendor training or develop our own? A: Usually both. Vendor training covers tool mechanics; internal training addresses your specific use cases, policies, and culture.

Q: How long should training take? A: AI literacy: 1-2 hours. Practitioner training: 1-2 days. Plan for refreshers as tools evolve.

Q: What if people don't attend training? A: Make training part of rollout—access to tools contingent on training completion. Get manager support for attendance.

Q: How do we keep training current as AI evolves? A: Build modular curriculum. Establish quarterly review process. Use microlearning for updates.

Q: Should training be mandatory? A: AI literacy and policy training: yes. Tool-specific training: yes for users. Optional for those who won't use tools.

Q: How do we train executives who "don't have time"? A: Short, high-impact sessions (90-120 minutes). Executive-specific content. Peer discussions. 1:1 coaching.


Next Steps

Training is not a one-time event but an ongoing investment. Design programs that build foundational literacy, develop practical skills, and evolve with the technology.

Ready to develop your AI training program?

Book an AI Readiness Audit to get expert guidance on training design and change management for AI adoption.


Sustaining AI Training Impact Beyond the Initial Program

L&D leaders face a common challenge: AI training generates enthusiasm during delivery but fails to translate into sustained behavioral change in the workplace. Three program design elements significantly improve long-term training impact.

First, implement spaced learning reinforcement by scheduling brief follow-up sessions at 2-week, 6-week, and 12-week intervals after the initial training. These reinforcement sessions are not full retraining but targeted exercises where participants apply AI tools to current work challenges and share results with peers. spaced reinforcement improves knowledge retention by 40 to 60 percent compared to single-session training. Second, create AI practice communities where trained employees meet monthly to share successful AI applications, troubleshoot challenges, and discover new use cases relevant to their roles. Peer learning accelerates skill development and creates social accountability for continued practice. Third, integrate AI competency into performance development conversations so that AI tool proficiency becomes a recognized professional skill rather than an optional add-on. When AI competency appears in performance frameworks, employees allocate ongoing attention to maintaining and expanding their capabilities.

Common Questions

Segment audiences by role and skill level, combine foundational literacy with role-specific applications, use multiple modalities, and build in practice opportunities and feedback loops.

Everyone needs AI literacy basics. Managers need governance and decision-making. Technical staff need tool-specific skills. Executives need strategic understanding.

Use a train-the-trainer model, develop self-paced modules for basics, conduct live sessions for complex topics, and build internal communities of practice.

References

  1. AI Risk Management Framework (AI RMF 1.0). National Institute of Standards and Technology (NIST) (2023). View source
  2. ISO/IEC 42001:2023 — Artificial Intelligence Management System. International Organization for Standardization (2023). View source
  3. Model AI Governance Framework (Second Edition). PDPC and IMDA Singapore (2020). View source
  4. Training Subsidies for Employers — SkillsFuture for Business. SkillsFuture Singapore (2024). View source
  5. ASEAN Guide on AI Governance and Ethics. ASEAN Secretariat (2024). View source
  6. OECD Principles on Artificial Intelligence. OECD (2019). View source
  7. Model AI Governance Framework for Generative AI. Infocomm Media Development Authority (IMDA) (2024). View source
Michael Lansdowne Hauge

Managing Director · HRDF-Certified Trainer (Malaysia), Delivered Training for Big Four, MBB, and Fortune 500 Clients, 100+ Angel Investments (Seed–Series C), Dartmouth College, Economics & Asian Studies

Managing Director of Pertama Partners, an AI advisory and training firm helping organizations across Southeast Asia adopt and implement artificial intelligence. HRDF-certified trainer with engagements for a Big Four accounting firm, a leading global management consulting firm, and the world's largest ERP software company.

AI StrategyAI GovernanceExecutive AI TrainingDigital TransformationASEAN MarketsAI ImplementationAI Readiness AssessmentsResponsible AIPrompt EngineeringAI Literacy Programs

EXPLORE MORE

Other AI Change Management & Training Solutions

INSIGHTS

Related reading

Talk to Us About AI Change Management & Training

We work with organizations across Southeast Asia on ai change management & training programs. Let us know what you are working on.