How AI actually works — the core concept
All AI is doing is working with
word relationships. Large Language Models (LLMs) tie every word in a language to every other word — the word "dog" is close to "bark", which is close to "tree", which is close to "leaf". When you send a prompt, the AI looks at all those word relationships and predicts the next most likely word, then the next, then the next.
Critical exam point: AI does not fact-check. It is not finding the most correct answer — it is predicting the next token. This is why human oversight is essential and why "human in the loop" is a recurring theme throughout the entire exam.
Traditional AI
Makes predictions based on patterns in previous data. Output is deterministic — ask the same question, get the same answer. Used for trend analysis, classification, pattern recognition.
Example: Analysing emails to determine if someone gives positive or negative reviews — always produces the same result for the same input.
Generative AI
Uses LLMs and word relationships to create something new — content that didn't exist before. Output is non-deterministic (indeterministic) — ask the same question three times, get three different answers.
Example: Writing an email, creating a slide deck, summarising a meeting — generates new content each time.
Exam trap: If the task is creating something new (email, document, response) → Generative AI. If the task is finding a pattern in existing data → Traditional AI. The key differentiator: are we creating something that didn't exist before?
Machine learning & deep learning
Machine learning
The computer's ability to get better over time by picking up on patterns. For the exam you just need the definition and the lifecycle: Define the task → Collect and prepare data → Train and validate the model → Deploy → Monitor and manage. You don't need technical depth.
Deep learning
Machine learning that uses many layers of connected neural networks to discover complex patterns in unstructured data automatically. Exam tip: tie the words "deep learning" to "neural network" — that's all you need to know.
AI challenges and weaknesses
Fabrication (hallucination)
Microsoft is moving from the term "hallucination" to "fabrication". AI doesn't lie — it just predicts the next token. If that prediction is wrong, it confidently states something incorrect. Expect questions on this.
Reliability of output
Ask Copilot to write the same email three times — you'll get three different results. The output is non-deterministic. This is different from fabrication — it's about consistency, not correctness.
Lack of explainability
Understanding exactly HOW the AI produced its output is difficult. Users generally don't understand what data it referenced or how it reasoned. This is a real challenge for building trust.
Data quality
All AI does is predict the next token based on the data it has access to. Flawed data → flawed responses. If an agent is grounded in bad data, its answers will be bad. Data must be accurate, up-to-date, deduplicated, and representative.
Bias and representativeness
If your training or grounding data is biased, your AI output will be biased. Ensure data is representative of ALL the people your AI serves — not just the majority. This is how you address bias in agent output.
Privacy concerns
Be careful about what data the AI has access to, especially PII. Have we cited sources? Have we considered who can see the output? These are questions the exam will ask about.
Sustainability
Running powerful AI models uses significant compute, power, and water. As a transformation leader, are we running unnecessarily powerful models for simple tasks? Cost and environmental impact are real considerations.
Grounding, RAG, and prompting
Grounding
What data is the AI using to generate its responses? By default, most agents are grounded in the general web. The power of Microsoft Copilot is grounding it in your business data — SharePoint, Teams chats, emails — so responses are specific and relevant to your organisation.
RAG (Retrieval Augmented Generation)
Before sending your prompt to the LLM, the system searches available data sources, finds relevant information, and adds it to your prompt automatically behind the scenes. You don't need to memorise the acronym — understand the concept: it's how agents pull in relevant business data to improve responses and reduce fabrication.
Semantic indexing
Ensuring that the word "customer" means the same thing across Sales, Marketing, and Shipping in your organisation. Consistent data definitions across your business are essential for AI to give coherent answers.
Writing a good prompt — four parts: (1) Context — who/what is this for? (2) Goal — what do you want? (3) Source — what data should it use? (4) Expectation — how should it respond (length, tone, format)?
Few-shot prompting: Including examples of the desired output inside your prompt. Gives the AI more context about what "good" looks like, enabling more consistent results. This is a Microsoft exam term — be aware of it.
Microsoft's six Responsible AI principles
Commit these to memory. They are thematically woven into every question on the exam. Microsoft's view: AI must be governed responsibly. We are not moving fast and breaking things — the potential for harm is too significant.
1. Fairness
AI must not discriminate against any group of people. Every user interacting with the AI has a fair chance of getting a helpful response. Example of a fairness violation: a loan approval model that automatically declines anyone who wore a red hoodie in their photo — the model is discriminating based on an irrelevant attribute.
2. Reliability & Safety
Have we tested the AI rigorously to ensure we are not getting harmful, fabricated, or inconsistent outputs? Have we protected against exploitation and jailbreaking (getting the AI to say things it shouldn't)? Is the AI grounded in the right data and regularly tested?
3. Privacy & Security
Ensure the AI is not connected to systems containing PII unnecessarily. Apply the principle of least privilege — the AI should only access the data it needs for its specific task. Data handling must comply with privacy regulations.
4. Inclusiveness
All users must have an equal chance of interacting with the AI regardless of their abilities. A chatbot that works perfectly for sighted users but is unusable for visually impaired users fails inclusiveness — not fairness. Think: accessibility for people with disabilities, hearing impairments, neurodivergence.
5. Transparency
Users should understand what the AI is doing and where its answers come from. Copilot cites its sources — this is transparency in action. Users should know they're interacting with AI and understand (at a reasonable level) how it generates responses. It's not magic — it's math.
6. Accountability
The human is responsible for AI output. If Copilot drafts a document and you send it to a client without reviewing it and it contains errors, you are accountable — not Copilot. This is why humans must be in the loop, reviewing AI output before it reaches consequential decisions or external parties.
Fairness vs Inclusiveness — the exam trap: Fairness = the AI discriminates against a group (red hoodie example). Inclusiveness = some users can't effectively use the AI due to their abilities (visually impaired user can't interact with the chatbot). These feel similar but are distinctly different principles.
Sensitive use cases to avoid
Microsoft outlines three categories where AI agents should generally be avoided or handled with extreme caution and human oversight:
Denial of consequential services
An agent that could automatically deny someone's ability to buy a home (mortgage approval), access healthcare, or lose employment. The consequences of a wrong answer are too significant to leave to AI alone.
Risk of harm
Using AI to diagnose illness or prescribe medication without human medical oversight. Any scenario where an incorrect AI response could physically harm a person.
Infringement on human rights
If the output of the agent could, in any way, restrict a person's fundamental human rights, the agent should not be built or must have robust human oversight at every decision point.
Note: These are not absolute prohibitions — organisations do use AI in sensitive scenarios. But they require humans in the loop for final decisions, governance systems, and clear accountability frameworks.
Governance systems & Microsoft's governing bodies
Chief AI Ethics Officer
A designated individual who owns AI governance decisions — understands both the technology and the business process. Makes decisions about what is acceptable in sensitive AI scenarios.
Cross-functional committee
AI governance must NOT sit with IT alone. It requires executive leadership + technical staff + business process owners + third-party experts. Diverse, cross-department, cross-level representation from the very beginning of AI strategy development.
Microsoft Senior Leadership Team
Fully integrated into how Microsoft manages AI and develops responsible AI practices and tools.
Office of Responsible AI
Microsoft's cross-organisational responsible AI governance function. Shapes norms and standards for AI inside and outside Microsoft.
AETHER Committee
AI and Ethics in Engineering and Research — a body of members that develops tools, best practices, and standards for responsible AI at Microsoft.
Key terms — flash reference
Fabrication
Microsoft's preferred term for hallucination. AI producing incorrect output because it predicts the next token — not because it intends to deceive.
Non-deterministic
Generative AI produces different output each time for the same input. Ask for the same email three times → three different emails.
Deterministic
Traditional AI produces the same output for the same input every time — because it's pattern-based.
Grounding
What data the AI uses to generate responses. Copilot's power = grounded in your business data (SharePoint, Teams, emails), not just the web.
RAG
Retrieval Augmented Generation. The system pulls relevant business data and adds it to your prompt before sending to the LLM. Reduces fabrication, improves relevance.
Few-shot prompting
Including examples of the desired output inside your prompt to guide the AI toward more consistent results.
Agent orchestration
Multiple specialised agents each doing one task, connected together — rather than one agent doing everything.
Human in the loop
A human reviews and is accountable for AI output before it reaches consequential decisions or external parties. Fundamental to responsible AI.
Tokens
The unit of AI consumption and cost. More tokens used = more money spent. Good prompting = fewer tokens = lower cost.
Buy / Extend / Build
Buy = M365 Copilot (productivity). Extend = Copilot Studio (business processes). Build = Foundry (custom models). Match the approach to what you actually need.
Fairness vs Inclusiveness
Fairness = AI doesn't discriminate against groups. Inclusiveness = all users can interact with the AI regardless of their abilities.
Semantic indexing
Ensuring consistent data definitions across your organisation ("customer" means the same thing in every department).
Data readiness
Continuous — not a one-time exercise. Ongoing data hygiene is required for trustworthy AI.
Scenario Q&A — click to reveal ↓
Your organisation wants to deploy an AI agent that automatically approves or denies mortgage applications. From a Responsible AI perspective, what should you advise?
Advise against full automation — require human review of every decision. This is a denial of consequential services scenario. An agent that automatically denies a mortgage denies someone the ability to buy a home — a consequential outcome too significant to leave to AI alone. A human must be in the loop for final decisions. Additionally, the agent must be checked for fairness (not discriminating on irrelevant factors) and the data used to train/ground it must be representative.
A company has three Copilot Studio agents running in production, widely used across the organisation. They want to optimise their Azure costs. Which Foundry licensing model should they use?
Prepaid (reserved capacity). With three production agents running at scale, consumption is predictable. Prepaid gives better pricing by committing capacity upfront, supports long-term planning, and provides more cost control than pay-as-you-go. Pay-as-you-go is for pilots and testing when consumption is unknown.
A sales team wants to use AI to help prepare for customer meetings. They need it to search through recent emails and Teams conversations and generate a briefing. Which platform and which M365 Copilot feature is most appropriate?
M365 Copilot — Researcher agent. This is a productivity task (meeting prep) fully within the M365 ecosystem using M365 data (emails, Teams). No custom line-of-business systems needed → M365 Copilot, not Copilot Studio. Researcher is most appropriate because the task involves finding and synthesising information from documents and communications (not a structured data/spreadsheet task, which would be Analyst).
A manufacturing company wants to use AI to detect defects on a production line by analysing live camera feeds. Which Azure AI service should they use?
Azure Vision. This is a scene-detection / object-detection use case using live camera feeds — not document processing. Azure Vision enables AI to see and identify objects in images and video streams. Azure Document Intelligence would be wrong — it extracts structured fields from uploaded documents like invoices, not live camera scenes.
Your organisation is considering an AI initiative. Leadership asks: "What questions should we be asking to evaluate the business value?" What are the three questions?
1. What problem does it solve? 2. What measurable outcomes will it deliver? 3. Does it align with our organisation's strategic goals? These are the Microsoft-defined questions for evaluating AI business value. On the exam, if one answer references these questions and another just says "build an agent in Copilot Studio," the business-value answer is the most correct choice.
What is the difference between Traditional AI and Generative AI? Give the key distinguishing features.▼
Traditional AI identifies patterns in previous data to make predictions about what will happen next. It is deterministic — the same input always produces the same output. Use cases: fraud detection, demand forecasting, sentiment classification. Generative AI uses LLMs and word/semantic relationships to create entirely new content that didn't exist before. It is non-deterministic — the same prompt produces different outputs each time. Use cases: drafting emails, generating documents, answering questions in natural language. The key differentiator: are we creating something new (generative) or finding a pattern in existing data (traditional)?
What are the six Microsoft Responsible AI principles and what makes Fairness different from Inclusiveness?▼
The six principles are: Fairness, Reliability & Safety, Privacy & Security, Inclusiveness, Transparency, and Accountability. Fairness means the AI does not discriminate against groups of people — it gives everyone an equally fair chance of a helpful response. Example violation: a loan model that automatically declines people based on an irrelevant attribute (red hoodie). Inclusiveness means all users can effectively interact with the AI regardless of their abilities — sighted vs visually impaired, hearing vs hearing impaired. Example violation: a chatbot that works perfectly for sighted users but is unusable for visually impaired users. The distinction: Fairness is about whether the AI treats different groups equitably in its decisions. Inclusiveness is about whether all users can access and use the AI effectively regardless of their physical or cognitive abilities.
What are the three elements of an agent, and what is agent orchestration?▼
Every agent has three elements: (1) Model — the LLM or AI model that powers the agent's reasoning (the brain). Choice affects performance, cost, and reliability. (2) Instructions — the text defining who the agent is, who it serves, what it does, and how it behaves (the job description). (3) Tools — the connections to external systems and APIs that allow the agent to actually execute tasks — not just talk, but act (the hands). Agent orchestration is the practice of building multiple specialised agents that each perform one specific task, then connecting them so they work together as a system. Instead of one agent doing six tasks, build six agents doing one task each. This mirrors the parent/child flow pattern in Power Automate.
What are the five pillars of the AI readiness framework, and which one is explicitly described as continuous?▼
The five pillars are: (1) Business Strategy — identifying the outcomes and KPIs your organisation cares about, before any AI discussion. (2) Technology & Data — preparing your data estate: breaking down silos, improving quality, ensuring clean/deduplicated/representative data, creating data dictionaries and access controls. This is the pillar explicitly described as continuous — data readiness is not a one-time exercise but an ongoing cycle of hygiene. (3) AI Strategy & Experience — starting small, tightly scoped pilots with 6–12 week timelines, clear success criteria, and fast iteration loops. (4) Organisational & Culture — leading from the top, building diverse cross-functional teams, embedding AI goals in performance reviews, building executive sponsorship. (5) AI Governance — covering data governance, AI governance (managing model risks), and regulatory governance (legal compliance and ethics).