Defining clear rules, responsibilities, and boundaries for AI use in your organization.

AI is already entering your workplace — formally or informally.

This course helps your team define, together, how AI should be used in your specific organizational context.

Rather than importing generic guidelines, you will:

  • identify where AI creates value,

  • surface ethical, social, legal, and cognitive risks,

  • define what is permitted, sensitive, or off-limits,

  • and establish shared principles that can evolve into an internal AI policy.

This is governance built from within — not imposed from outside.

What you walk away with

  • A shared understanding of AI in your context

  • Clearly articulated AI principles

  • Defined boundaries and responsibilities

  • Your organisation’s specific AI Policy Foundation Map: a clear, shared set of AI principles, structured so they can directly inform an internal AI policy.

  • Alignment across leadership and operational roles

How it works

This course is delivered in a structured, facilitated team setting.

Through guided discussion, real-case exploration, and participatory exercises, your team will:

  • examine current and emerging AI use in your organization,

  • clarify risks and opportunities specific to your context,

  • articulate shared principles and boundaries,

  • and align on responsibilities and oversight.

In Practice:

  • Work together to set the conditions for good AI governance in your organisation:
    1.
    You will define, together, how AI can and cannot be used in your organization.
    2. You will agree on clear principles that reflect your values, your responsibilities, and your context—creating a solid foundation for an internal AI policy, including boundaries, expectations and opportunities.

The process is collaborative and confidential.
At the end, your collective thinking is synthesized into a clear, usable policy foundation document.

Duration

Online: 2 × 2-hour sessions
On-site: 1 full day

Why this course is essential

AI adoption without clear principles leads to:

  • shadow use,

  • inconsistency,

  • legal and reputational risk,

  • loss of trust internally.

Setting governance early ensures:

  • clarity before scale,

  • innovation within boundaries,

  • and shared ownership of AI decisions.

Without this step, AI integration remains fragmented and fragile.

For whom

  • Leadership teams

  • HR / People & Culture

  • IT / Digital / Innovation leads

  • Department heads

  • Public sector administrators

  • School leadership teams

Best delivered with a cross-functional internal group.

THIS WORKSHOP IS OFFERED AS A STAND ALONE OR AVAILABLE AS PART OF A LARGER AI GOVERNANCE PROGRAMME, “Working Responsibly with AI“, DEDICATED TO TEAMS WITHIN BUSINESSES, PUBLIC, EDUCATIONAL OR CIVIL SOCIETY ORGANISATIONS

Previous
Previous

Future-Proofing Roles & Skills in the age of AI.