HELP

AI-900 Generative AI Domain: Copilot & Azure OpenAI Explained

AI Certification Exam Prep — Beginner

AI-900 Generative AI Domain: Copilot & Azure OpenAI Explained

AI-900 Generative AI Domain: Copilot & Azure OpenAI Explained

Learn AI-900 fundamentals plus Copilot and Azure OpenAI—then pass confidently.

Beginner ai-900 · microsoft · azure-ai-fundamentals · azure-openai

Prepare for Microsoft AI-900 with a clear, exam-aligned blueprint

This course is a beginner-friendly, exam-prep roadmap for the Microsoft AI-900: Azure AI Fundamentals certification. It is designed for learners with basic IT literacy who want a structured way to understand what Microsoft tests, how to recognize the right Azure AI approach for a scenario, and how to answer common question patterns under time pressure.

The blueprint maps directly to the official AI-900 exam domains:

  • Describe AI workloads
  • Fundamental principles of ML on Azure
  • Computer vision workloads on Azure
  • NLP workloads on Azure
  • Generative AI workloads on Azure

Why this course focuses on Copilot and Azure OpenAI (without losing exam coverage)

AI-900 increasingly expects you to recognize where generative AI fits, what it can (and cannot) do, and how Azure services support safe deployment. This course keeps the fundamentals intact while making generative AI practical by explaining core concepts such as tokens, embeddings, grounding, and responsible AI, plus how these ideas appear in Microsoft’s ecosystem through Copilot patterns and Azure OpenAI capabilities.

How the 6-chapter structure helps you pass

Chapter 1 orients you to the exam itself—registration, scoring expectations, and a study strategy that aligns your time to the official domains. Chapters 2 through 5 each go deep into one or two domains at a time, using plain-language explanations and exam-style practice prompts that mirror typical AI-900 scenario questions (service selection, workload identification, and basic metric interpretation). Chapter 6 then consolidates everything with a full mock exam experience, analysis of weak areas, and an exam-day checklist.

  • Chapter 1: Exam setup, scheduling, scoring expectations, and a practical plan for beginners
  • Chapter 2: Describe AI workloads and match real scenarios to the right AI approach
  • Chapter 3: Fundamental ML principles on Azure, including training vs inference and evaluation basics
  • Chapter 4: Computer vision workloads on Azure, with service selection patterns and common task types
  • Chapter 5: NLP plus generative AI workloads on Azure, including Azure OpenAI and Copilot concepts
  • Chapter 6: Full mock exam + final review to build confidence and speed

Built for beginners, optimized for exam performance

You do not need prior certification experience. The course focuses on the “why” behind each answer choice so you can reliably eliminate distractors. You’ll also learn the vocabulary Microsoft uses (for example, how prompts describe a workload vs a service, and how responsible AI principles show up in scenario wording).

When you’re ready, start by setting your baseline and planning your schedule. If you’re new to Edu AI, begin here: Register free. Or explore other certification tracks anytime: browse all courses.

Outcome

By the end, you’ll be able to map a question to an AI-900 domain quickly, select the best-fit Azure AI approach, and complete a timed mock exam with a repeatable review method—helping you walk into the Microsoft AI-900 exam prepared and calm.

What You Will Learn

  • Describe AI workloads and identify common AI solution scenarios
  • Explain fundamental principles of machine learning on Azure, including training vs inference and model evaluation
  • Describe computer vision workloads on Azure and choose appropriate Azure AI services
  • Describe NLP workloads on Azure, including text analytics and conversational AI concepts
  • Describe generative AI workloads on Azure, including Azure OpenAI, responsible AI, and Copilot concepts

Requirements

  • Basic IT literacy (files, browsers, cloud concepts)
  • No prior Microsoft certification experience required
  • Access to a computer with internet for reading objectives and taking practice quizzes
  • Optional: An Azure account for exploring services (not required)

Chapter 1: AI-900 Exam Orientation and Study Strategy

  • Understand AI-900 format, skills measured, and question styles
  • Register for the exam and set up a realistic study plan
  • Learn scoring, passing expectations, and time management
  • Build your exam-day toolkit: notes, labs, and revision loops
  • Baseline diagnostic quiz and goal setting

Chapter 2: Describe AI Workloads (Domain Deep Dive)

  • Differentiate AI, ML, deep learning, and generative AI in exam terms
  • Map business problems to AI workloads and Azure solution types
  • Apply responsible AI principles to real scenarios
  • Practice set: AI workloads scenario questions
  • Mini-review: domain summary and memory cues

Chapter 3: Fundamental Principles of ML on Azure

  • Understand core ML concepts: features, labels, training, inference
  • Choose ML approaches and metrics for typical AI-900 prompts
  • Explain Azure Machine Learning capabilities at a fundamentals level
  • Practice set: ML lifecycle and metrics questions
  • Mini-review: key definitions and exam word cues

Chapter 4: Computer Vision Workloads on Azure

  • Identify vision tasks and pick the right Azure AI service
  • Explain detection, OCR, and image classification in plain language
  • Understand vision customization and model deployment at a high level
  • Practice set: computer vision service selection questions
  • Mini-review: vision cheat sheet and last-minute recalls

Chapter 5: NLP + Generative AI Workloads on Azure (Copilot & Azure OpenAI)

  • Understand NLP tasks and Azure service choices for text and conversation
  • Explain generative AI concepts: tokens, embeddings, grounding, retrieval
  • Use Azure OpenAI and Copilot concepts in exam language (no coding required)
  • Practice set: NLP + GenAI + responsible AI questions
  • Mini-review: final domain map and cross-domain comparisons

Chapter 6: Full Mock Exam and Final Review

  • Mock Exam Part 1
  • Mock Exam Part 2
  • Weak Spot Analysis
  • Exam Day Checklist
  • Final rapid review and confidence pass plan

Nadia Karim

Microsoft Certified Trainer (MCT)

Nadia Karim is a Microsoft Certified Trainer who designs exam-first learning paths for Azure and AI fundamentals. She has coached beginners through Microsoft certification prep by translating official objectives into hands-on mental models and exam-style practice.

Chapter 1: AI-900 Exam Orientation and Study Strategy

This chapter sets your operating system for AI-900: how the exam is built, what it rewards, and how to study efficiently—especially if you’re learning Generative AI concepts like Copilot and Azure OpenAI alongside classic AI fundamentals. AI-900 is not a coding exam; it evaluates whether you can recognize AI workloads, select appropriate Azure AI services, and describe core concepts with correct terminology. Most missed questions come from confusing similarly named services, mixing up training vs. inference, or overthinking “best” when the prompt really asks for “most appropriate.”

Your goal is to build a repeatable revision loop: (1) learn the objective, (2) see it in Azure and Microsoft Learn, (3) test recognition with practice items, (4) fix weak spots with targeted notes and short labs. By the end of this chapter you should have a realistic schedule, know how the exam is scored and delivered, and have a practical exam-day toolkit (notes, labs, and checkpoints) that you’ll refine as you progress.

Exam Tip: Treat AI-900 as a vocabulary-and-scenarios exam. If you can map a scenario to a workload (vision/NLP/generative AI) and then to the correct Azure service family, you’ll answer most questions quickly—even when distractors sound plausible.

Practice note for Understand AI-900 format, skills measured, and question styles: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Register for the exam and set up a realistic study plan: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Learn scoring, passing expectations, and time management: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Build your exam-day toolkit: notes, labs, and revision loops: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Baseline diagnostic quiz and goal setting: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Understand AI-900 format, skills measured, and question styles: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Register for the exam and set up a realistic study plan: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Learn scoring, passing expectations, and time management: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Build your exam-day toolkit: notes, labs, and revision loops: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Baseline diagnostic quiz and goal setting: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Sections in this chapter
Section 1.1: What AI-900 Azure AI Fundamentals covers (official domains)

AI-900 (Azure AI Fundamentals) measures your ability to describe AI workloads and identify common solution scenarios on Azure. Think of it as “recognize and explain,” not “build and tune.” The exam is organized around official skills domains that typically include: describing AI workloads and considerations, describing fundamental principles of machine learning (ML) on Azure, describing features of computer vision workloads, describing features of natural language processing (NLP) workloads, and describing features of generative AI workloads (including Azure OpenAI and responsible AI concepts). Your course outcomes map directly to these domains: workloads and scenarios; ML fundamentals such as training vs. inference and evaluation; vision and NLP service selection; and generative AI with Azure OpenAI, Copilot concepts, and responsible AI.

What the test is really checking is whether you can: (a) identify the workload type from a short business prompt (for example, “classify images” vs “extract key phrases” vs “generate an email draft”), (b) choose the right Azure AI service category, and (c) describe core ideas accurately (training vs inference, precision/recall basics, prompt vs completion, embeddings vs chat, etc.).

Common traps at this stage are domain-mixing and naming confusion. For example, candidates may treat “Copilot” as a single product rather than a pattern (assistant experience layered over models + grounding + safety + orchestration). Another trap is assuming Azure OpenAI is the answer to every language scenario; the exam often expects you to choose Azure AI Language (Text Analytics capabilities) for extraction/classification tasks, reserving generative AI for content generation, summarization, or conversational experiences that benefit from open-ended output.

Exam Tip: When a question asks you to “describe,” pick the option that uses precise definitions (training builds the model; inference uses it). When it asks you to “choose,” map scenario → workload → service family before reading the answer choices.

Section 1.2: Exam registration (Microsoft Learn/Pearson VUE) and scheduling

Registering correctly is part of exam success because scheduling decisions influence your study plan and stress level. AI-900 is delivered through Pearson VUE. Start at Microsoft Learn’s certification pages to locate the AI-900 exam, confirm the measured skills outline, and follow the “Schedule exam” flow into Pearson VUE. You’ll choose online proctoring or a test center. Online is convenient but has stricter environment rules; test centers reduce the risk of technical interruptions.

Build a realistic timeline by counting backwards from your desired test date. For most learners, a two-to-four-week plan works well if you study 30–60 minutes most days; longer if you’re new to cloud concepts. Align your schedule with the domains: dedicate specific days to ML fundamentals, vision, NLP, and generative AI. Keep one weekly slot for cumulative review, because AI-900 is broad and memory fades quickly without spaced repetition.

Be intentional about time-of-day selection. If you do your best focused work in the morning, schedule the exam then. Also plan a buffer week if possible: it’s not “extra time,” it’s risk management for life events and weak-domain remediation.

Exam Tip: Before scheduling, verify your ID name matches your registration exactly. Mismatches are a surprisingly common reason for day-of delays or forfeits. Also, if you choose online proctoring, run the system test early and again the day before.

Section 1.3: Scoring, question formats, and common traps

AI-900 uses scaled scoring. The key practical point: you need to clear the passing threshold, not achieve perfection. Because it’s scaled, don’t waste energy trying to infer “points per question.” Focus on consistency across domains. If you’re strong in ML and weak in generative AI or vice versa, the exam can expose that imbalance through scenario distribution.

Expect multiple-choice and multiple-response items, plus scenario-based questions. Some items ask you to match capabilities to services (for example, “OCR,” “object detection,” “key phrase extraction,” “sentiment analysis,” “chat completion,” “embeddings”). Others test conceptual boundaries: training vs inference, supervised vs unsupervised learning, classification vs regression, and when to use prebuilt models vs custom models. In the generative AI domain, expect conceptual questions about responsible AI, content filters, grounding, and what Azure OpenAI provides (models, deployments, prompts) versus what Copilot experiences provide (user-facing assistant + orchestration and data access patterns).

Common traps include absolute words (“always,” “never”), confusing similar service names, and selecting an overpowered service. For instance, if a prompt asks to “extract entities and key phrases from text,” the safest choice is Azure AI Language, not Azure OpenAI—unless the question explicitly emphasizes generative output. Another trap is mixing “model” with “deployment.” In Azure OpenAI, you deploy a model and then call that deployment; the exam may test whether you understand that distinction at a high level.

Exam Tip: When two answers both seem correct, pick the one that most directly satisfies the requirement with the simplest service match. AI-900 favors appropriate fit over “coolest” technology.

Section 1.4: Study strategy mapped to domains and objectives

Your study strategy should mirror the exam’s structure. Organize your notes and revision by the official domains, not by random resources. Create a simple tracker with each outcome and a confidence rating (for example, 1–5). Your goal is to move every item to “4+” before exam day. This is where you integrate a baseline diagnostic (without writing questions in your notes): use a short assessment or practice set to identify which domains are weakest, then allocate time accordingly.

For ML fundamentals, prioritize: what ML is used for, the difference between training and inference, and how evaluation works conceptually (accuracy, precision/recall, confusion matrix basics). Don’t overlearn formulas; learn what each metric indicates and when it matters. For computer vision, know which tasks are classification, object detection, OCR, and image analysis—and which Azure service category supports them. For NLP, distinguish extraction (entities, key phrases), classification (sentiment, language detection), and conversational AI concepts. For generative AI, focus on what makes it different (open-ended output), the Azure OpenAI service role, prompt engineering basics, embeddings/grounding patterns, and responsible AI principles that show up in enterprise deployments.

Plan your weeks in layers: first pass (recognize terms), second pass (scenario mapping), third pass (timed practice and trap avoidance). Insert mini-labs at the end of each domain: even a lightweight “click-through” in the Azure portal or Microsoft Learn sandbox improves recall. Your exam-day toolkit begins here: a one-page “service-to-workload” map, a glossary of confusing pairs (classification vs regression; OCR vs image tagging; Azure OpenAI vs Azure AI Language), and a checklist of responsible AI concepts (fairness, reliability/safety, privacy/security, inclusiveness, transparency, accountability).

Exam Tip: If you can’t explain a concept in one or two sentences (e.g., “inference”), you don’t own it yet. Tight definitions are what the exam rewards.

Section 1.5: How to use Microsoft Learn and objective-driven revision

Microsoft Learn is your primary source of truth because it aligns terminology and service descriptions with what the exam expects. Use it in an objective-driven way: start from the AI-900 skills outline, open the matching Learn modules, and take notes that answer “What is it?”, “When do I use it?”, and “How is it different from nearby options?” Avoid passive reading. After each module, summarize the decision rule in your own words—this is what you’ll recall under time pressure.

Build revision loops. A practical loop looks like: Learn module → 10–15 minutes of recap notes → quick scenario mapping practice (no deep build needed) → correct misunderstandings → update your one-page map. Repeat weekly. The key is to keep notes minimal and usable: tables that map workload to service are better than long paragraphs. For generative AI, add a small “concept map” that distinguishes: model vs deployment, prompt vs completion, chat vs embeddings, and Copilot as an application pattern that can ground model responses in enterprise data with safety controls.

When you hit confusion, go back to the objective and re-anchor. For example, if you’re unsure whether a task is “NLP” or “generative AI,” ask: is the output deterministic extraction/classification, or is it free-form generation? The exam usually separates these intentionally.

Exam Tip: Don’t chase every new product name. AI-900 tests stable fundamentals: workloads, service families, and responsible AI. Use Microsoft Learn to keep your terminology current, but keep your study centered on the measured skills.

Section 1.6: Practice approach: eliminate distractors and manage time

Your practice approach should train two skills: accuracy under ambiguity and speed under time limits. Start untimed to learn patterns, then switch to timed sets to build endurance. AI-900 questions often include distractors that are “related but not best.” Your job is to eliminate quickly using workload keywords. For example: “detect objects in an image” points to a vision detection capability; “extract text from images” points to OCR; “summarize a long document” leans generative; “identify key phrases” is classic NLP extraction.

Use a three-pass method during timed practice: (1) answer easy recognition questions immediately; (2) return to medium questions where you can narrow to two; (3) only then spend time on the hardest ones. Manage cognitive load by writing down (on the provided whiteboard tool or scratch paper, where allowed) a short “service shortlist” for the scenario before reading all options. This prevents being pulled toward shiny distractors.

Time management is less about rushing and more about avoiding unproductive loops. If you’re stuck, mark it and move on. When reviewing, don’t just note the right answer—note the rule that made the other option wrong. This is how you build trap resistance.

Exam Tip: For multiple-response items, treat each option as true/false against the requirement. Many candidates fail by selecting one extra plausible statement. Only select choices that are explicitly supported by the scenario and the service’s core purpose.

Chapter milestones
  • Understand AI-900 format, skills measured, and question styles
  • Register for the exam and set up a realistic study plan
  • Learn scoring, passing expectations, and time management
  • Build your exam-day toolkit: notes, labs, and revision loops
  • Baseline diagnostic quiz and goal setting
Chapter quiz

1. You are creating a study plan for AI-900. A colleague asks whether they need to write code during the exam. Which statement best describes what AI-900 evaluates?

Show answer
Correct answer: Your ability to recognize AI workloads, choose appropriate Azure AI services, and use correct AI terminology
AI-900 is a fundamentals exam focused on identifying AI workloads (vision, NLP, generative AI, etc.), mapping scenarios to the right Azure AI service family, and describing concepts accurately—not writing code. Option B is more aligned with practitioner/implementation exams where coding and model-building are assessed. Option C aligns with architecture/security-focused certifications rather than AI fundamentals.

2. During practice questions, you frequently miss items where Azure services have similar names. Which action most directly targets the likely cause of these misses for AI-900?

Show answer
Correct answer: Create a quick-reference mapping from common scenarios (vision/NLP/generative AI) to the correct Azure AI service families and key differentiators
AI-900 commonly tests recognition: scenario → workload → service family, and many wrong answers come from confusing similarly named services. A scenario-to-service mapping with differentiators directly reduces those errors. Pricing/regions (B) are not a primary objective for AI-900. Deep implementation details like custom training pipelines (C) go beyond the non-coding, conceptual nature of the exam and can distract from what’s tested.

3. A team member tends to overthink questions that ask for the "best" choice and often changes correct answers to incorrect ones. What exam strategy is most appropriate for AI-900-style questions?

Show answer
Correct answer: Identify the workload and the minimum requirement in the prompt, then select the most appropriate service/concept rather than an over-engineered option
AI-900 rewards selecting the most appropriate option for the scenario, which often means mapping to the simplest correct workload/service that meets stated requirements. Choosing the most feature-rich option (B) often leads to wrong answers when the question is about appropriateness rather than maximum capability. Managed services (C) are frequently the correct choice on Azure fundamentals exams; dismissing them is an invalid test-taking heuristic.

4. You are mentoring a candidate who confuses training with inference in multiple practice questions. Which statement correctly distinguishes them in an exam scenario context?

Show answer
Correct answer: Training is creating or fitting a model using data; inference is using a trained model to generate predictions or outputs for new input
AI-900 expects you to use correct terminology: training produces a trained model from data, while inference uses that model to produce outputs on new data (including generative outputs). Option B reverses the meanings and would cause incorrect service/concept selection in scenarios. Option C is wrong because AI-900 explicitly assesses core AI concepts and vocabulary in addition to recognizing services.

5. You want a repeatable revision loop for AI-900 that reduces weak areas quickly while keeping effort realistic. Which sequence best matches an effective approach for this exam?

Show answer
Correct answer: Learn the objective → see it in Azure/Microsoft Learn → test recognition with practice items → fix weak spots with targeted notes and short labs
AI-900 preparation is most effective when you iterate: objective learning, hands-on familiarity (Azure/Microsoft Learn), practice to test recognition, then targeted remediation with notes/labs. Option B delays feedback and typically preserves misconceptions until late. Option C emphasizes rote memorization and poor scheduling discipline, which increases confusion between similar services and leaves gaps unaddressed.

Chapter 2: Describe AI Workloads (Domain Deep Dive)

This chapter is where many AI-900 candidates either gain easy points—or lose them due to vague wording. The exam does not expect you to build models, but it does expect you to recognize workloads, match them to common solution scenarios, and pick the Azure service category that fits. In Copilot and Azure OpenAI questions, you’ll also be tested on whether you can tell “generative AI” apart from traditional ML, and whether you can apply Responsible AI principles to realistic business situations.

Throughout this chapter, practice a simple exam habit: when you read a scenario, underline the input type (text, image, audio, tabular data), the desired output (label, number, summary, generated content), and whether there is an explicit “right answer” label available (supervised vs unsupervised). Those three clues usually identify the workload in under 10 seconds.

Exam Tip: Many wrong answers are “adjacent” services or adjacent workloads. If a scenario asks for extracting insights from documents, don’t jump straight to “chatbot.” If it asks for generating new text, don’t pick “text analytics” (which analyzes existing text). Train yourself to spot the verb: classify, predict, cluster, detect, extract, search, converse, generate.

  • Differentiate AI, ML, deep learning, and generative AI in exam terms
  • Map business problems to AI workloads and Azure solution types
  • Apply responsible AI principles to real scenarios
  • Practice set: scenario matching habits (without needing to build anything)
  • Mini-review: memory cues that survive exam-day pressure

Now, let’s do the domain deep dive section by section.

Practice note for Differentiate AI, ML, deep learning, and generative AI in exam terms: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Map business problems to AI workloads and Azure solution types: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Apply responsible AI principles to real scenarios: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Practice set: AI workloads scenario questions: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Mini-review: domain summary and memory cues: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Differentiate AI, ML, deep learning, and generative AI in exam terms: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Map business problems to AI workloads and Azure solution types: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Apply responsible AI principles to real scenarios: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Practice set: AI workloads scenario questions: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Sections in this chapter
Section 2.1: Describe AI workloads and common solution scenarios

Section 2.1: Describe AI workloads and common solution scenarios

On AI-900, “AI workload” is shorthand for the type of problem you are solving with AI. The exam commonly frames workloads as: machine learning (prediction from data), computer vision (images/video), natural language processing (text/language), knowledge mining (search + extraction from unstructured content), conversational AI (bots), and generative AI (creating new content). You’re expected to connect a business goal to one of these workload categories and then to an Azure solution type.

Use exam definitions (not research-paper definitions). AI is the broad umbrella of systems that exhibit “intelligent” behavior. ML is a subset where a model learns patterns from data. Deep learning is a subset of ML that uses neural networks with multiple layers and is often used for vision, language, and generative models. Generative AI is a subset focused on producing new outputs (text, images, code) based on prompts. In exam terms: if the scenario says “generate,” “draft,” “summarize,” “rewrite,” “create an image,” or “produce code,” you are likely in generative AI territory.

Exam Tip: Don’t overthink the math. The exam is checking whether you can map: (1) business problem, (2) workload, (3) Azure capability. If a scenario mentions “train a model using historical labeled data,” that’s classic ML. If it mentions “use a prebuilt API to extract key phrases,” that’s an Azure AI service. If it mentions “ground the model in company documents,” think retrieval/knowledge mining patterns plus generative AI.

  • Prediction of a number: forecast demand, estimate house price → ML (regression)
  • Choose a category/label: approve/deny loan, spam/not spam → ML (classification)
  • Group similar items: customer segmentation → ML (clustering)
  • Find unusual behavior: fraud, sensor anomalies → anomaly detection
  • Analyze images: detect objects, read text in images → computer vision
  • Understand text: sentiment, language detection → NLP/text analytics
  • Search across documents: index PDFs, extract entities → knowledge mining
  • Chat interface: helpdesk bot, FAQs → conversational AI
  • Create new content: draft emails, summarize meetings → generative AI/Azure OpenAI

Common trap: candidates answer with a specific product name when the question is asking for the workload type. Read the prompt carefully: if it says “which AI workload,” answer with the workload (e.g., NLP), not the service (e.g., Azure AI Language). If it says “which service,” then choose the service category that matches.

Section 2.2: Classification vs regression vs clustering vs anomaly detection

Section 2.2: Classification vs regression vs clustering vs anomaly detection

This is a high-frequency AI-900 objective because it tests whether you can interpret outputs. The fastest way to decide is to look at the shape of the output and whether there is labeled training data. Classification outputs a category (often yes/no or a set of classes). Regression outputs a continuous numeric value. Clustering outputs groups based on similarity without predefined labels. Anomaly detection outputs a flag/score indicating “unusualness,” often in time series or transactional data.

Training vs inference is also tested here. Training is when you fit the model using historical data (often compute-heavy, iterative). Inference is when you use the trained model to make predictions on new data (often latency-sensitive). The exam likes to embed this distinction in operational scenarios: “batch scoring overnight” vs “real-time scoring at checkout.”

Exam Tip: If the scenario includes “known outcomes” like churned/not churned or fraud/not fraud, that strongly suggests supervised learning (classification). If it includes “we don’t know the groups ahead of time,” that suggests clustering. If it includes “detect spikes,” “outliers,” “unusual patterns,” that points to anomaly detection—even if the output is just a score.

  • Classification: spam detection, disease diagnosis categories, product defect yes/no
  • Regression: price prediction, energy usage forecasting, delivery time estimation
  • Clustering: market segmentation, grouping similar news articles, organizing customers by behavior
  • Anomaly detection: fraud discovery, equipment failure early warning, network intrusion signals

Model evaluation appears often as “which metric is appropriate?” At AI-900 level, remember: classification commonly uses accuracy/precision/recall; regression commonly uses error metrics (like MAE/RMSE); anomaly detection often uses precision/recall style thinking because anomalies are rare. The trap is picking “accuracy” for imbalanced fraud scenarios: accuracy can be misleading if 99.9% transactions are normal.

Exam Tip: When you see “rare event” language (fraud, intrusion, defect), think class imbalance and avoid simplistic “accuracy” reasoning. The exam isn’t asking for metric formulas, but it is checking whether you can spot what “good” means in context.

Section 2.3: Computer vision vs NLP vs knowledge mining vs conversational AI

Section 2.3: Computer vision vs NLP vs knowledge mining vs conversational AI

These workloads are easy points if you lock onto the input modality. Computer vision uses images or video frames: object detection, image classification, face detection (not identification unless allowed), OCR (extracting printed/handwritten text), and spatial understanding. NLP uses language as input: detecting sentiment, extracting entities, identifying key phrases, summarizing existing text (note: summarization can be NLP or generative depending on phrasing and tool choice). Knowledge mining is about search and extraction from large collections of unstructured content—documents, PDFs, scans—typically involving indexing plus enrichment. Conversational AI is about dialogue: managing multi-turn interactions, intents, and responses.

The exam often tests whether you can separate “analyze” from “generate.” For example, extracting entities from customer emails is NLP/text analytics. Generating a reply email is generative AI. Indexing 100,000 policy PDFs so employees can search them is knowledge mining (often paired with Azure AI Search). Creating a chat experience that answers HR policy questions could be conversational AI, but if it also “reads” policy PDFs to answer accurately, the underlying pattern is knowledge mining + conversational interface (and possibly generative AI for natural responses).

Exam Tip: “Search across documents” is a keyword for knowledge mining. “Chatbot” is not automatically the right answer if the real need is document indexing and retrieval. Many scenarios describe a bot interface, but the core workload is still search + extraction.

  • Computer vision clues: images, camera feed, X-rays, invoices as scanned images, detect/locate objects
  • NLP clues: emails, reviews, call transcripts (as text), sentiment, key phrases, translation
  • Knowledge mining clues: corpus, repository, index, search, enrich, extract metadata from documents
  • Conversational AI clues: intents, utterances, multi-turn, escalation to human, dialogue flow

Common trap: OCR belongs to computer vision even though the output is text. If the scenario is “extract text from images,” the input is still an image workload. Another trap: “speech” scenarios can blur boundaries. If the exam mentions converting speech to text, that’s speech AI; once text is produced, NLP can analyze it. Read for the primary task.

Section 2.4: Generative AI workload recognition (prompt-in, content-out patterns)

Section 2.4: Generative AI workload recognition (prompt-in, content-out patterns)

Generative AI questions typically follow a “prompt-in, content-out” pattern: a user provides instructions (and possibly context), and the system outputs newly composed text, code, images, or structured content. For AI-900 in the Copilot/Azure OpenAI domain, you must recognize when the requirement is creation rather than analysis. Drafting marketing copy, writing a job description, summarizing a meeting into action items, generating SQL from natural language, and rephrasing content in a brand voice are all generative patterns.

Azure OpenAI is the canonical “service family” the exam expects you to associate with generative text/code/image scenarios. Copilot scenarios often describe the same idea but embedded into productivity tools: “help users write,” “help users summarize,” “help users generate a plan.” Treat Copilot as an application experience built on generative models plus grounding, security, and orchestration. The exam generally won’t require deep architecture, but it will check that you understand the role: Copilot helps users produce outcomes inside workflows.

Exam Tip: If the scenario mentions “use company data to answer accurately,” look for retrieval/grounding language (documents, policies, product catalog). This is a common pattern: retrieval-augmented generation (RAG). The trap is assuming the model “already knows” internal documents. On the exam, when proprietary data is required, the correct approach usually involves retrieving relevant content and then generating an answer grounded in that content.

  • Prompt engineering clues: instructions, tone/style constraints, examples, few-shot prompts
  • Content-out clues: draft, rewrite, summarize, generate, compose, create, brainstorm
  • Grounding clues: “based on our documents,” “use the product manual,” “company policy says…”

Common trap: “Summarize this text” can be answered by either extractive summarization (traditional NLP) or generative summarization (Azure OpenAI). If the question emphasizes “generate a well-written summary,” “rewrite,” or “create,” prefer generative. If it emphasizes “analyze text to extract key phrases,” prefer NLP/text analytics.

Section 2.5: Responsible AI: fairness, reliability, privacy, inclusiveness, transparency, accountability

Section 2.5: Responsible AI: fairness, reliability, privacy, inclusiveness, transparency, accountability

Responsible AI is not a side topic on AI-900; it is a scoring area that appears in both classic ML and generative AI questions. You must be able to match a scenario risk to a principle. Memorize the six principles and tie each to a “what could go wrong” example.

  • Fairness: avoid bias that disadvantages groups (e.g., loan approvals skewed by demographic proxies).
  • Reliability & Safety: system performs consistently and avoids harmful failures (e.g., hallucinated medical guidance, unsafe instructions).
  • Privacy & Security: protect sensitive data, limit exposure and retention (e.g., PII in prompts/logs, data leakage).
  • Inclusiveness: usable by people with different abilities, languages, and contexts (e.g., accessibility features, multilingual support).
  • Transparency: users understand system limitations and that AI is being used (e.g., disclose AI-generated content, explain confidence/limits).
  • Accountability: humans remain responsible; governance and auditability exist (e.g., approval workflows, monitoring, incident response).

Exam Tip: The exam often disguises these principles as operational requirements. “We need to explain why a decision was made” maps to transparency. “We need a human to review before sending” maps to accountability and safety. “We must not store customer SSNs in logs” maps to privacy/security. Choose the principle that is most directly threatened in the scenario.

For generative AI, reliability and transparency show up as “hallucinations” and “over-trust.” A safe design might include grounded answers, citations to retrieved sources, user warnings, and fallback behavior when confidence is low. For ML, fairness and accountability show up in decision systems (hiring, credit, insurance). The common trap is selecting “privacy” for every sensitive scenario; instead, identify whether the issue is data handling (privacy) or outcome bias (fairness) or inability to explain/trace decisions (transparency/accountability).

Section 2.6: Exam-style practice: scenario matching and workload selection

Section 2.6: Exam-style practice: scenario matching and workload selection

This section is your practical “scenario matching” toolkit. You are not answering practice questions here; you are building a repeatable method that works under time pressure. AI-900 scenarios are short, so your goal is to extract the signal quickly and avoid distractors.

Step 1: Identify the artifact. Are you dealing with tabular data (rows/columns), text, images, documents, or a conversation? Tabular usually points to ML (classification/regression/clustering/anomaly). Images point to computer vision. Large document sets point to knowledge mining and search. Multi-turn interaction points to conversational AI.

Step 2: Identify the verb. Predict/forecast/estimate → regression. Decide/label/categorize → classification. Group/similar segments → clustering. Unusual/outlier/spike → anomaly detection. Extract entities/key phrases/sentiment → NLP. Detect objects/recognize text in image → vision. Search/index/enrich documents → knowledge mining. Draft/rewrite/summarize/generate → generative AI.

Step 3: Decide if it’s “analyze” or “create.” This is where Azure OpenAI and Copilot become obvious. If the system is producing original prose/code/images from a prompt, it’s generative. If it’s identifying properties of existing content, it’s analytics.

Exam Tip: Watch for mixed scenarios. Many real solutions combine workloads (e.g., OCR + NLP; search + generative answers; speech-to-text + sentiment). The exam usually asks for the primary workload or the best first step. Don’t pick a later-stage component if the scenario is explicitly about the earlier transformation (e.g., “convert scanned forms to text” is vision/OCR, even if the next step is NLP).

  • Memory cue: “Number → regression; Name → classification; Neighborhoods → clustering; Needle-in-haystack → anomaly.”
  • Memory cue: “Pixels → vision; Words → NLP; Documents at scale → knowledge mining; Dialogue → conversational; Prompt-to-content → generative.”

Common traps to avoid: (1) picking “deep learning” as a workload when the question wants the business task (vision/NLP/etc.); (2) confusing chatbot UX with the underlying knowledge mining requirement; (3) calling all language tasks “generative” when many are simple text analytics; and (4) forgetting Responsible AI in scenarios with hiring, lending, healthcare, or any user-facing generated content. If you can label the workload, you can usually eliminate two options immediately—and that’s how you turn domain knowledge into exam points.

Chapter milestones
  • Differentiate AI, ML, deep learning, and generative AI in exam terms
  • Map business problems to AI workloads and Azure solution types
  • Apply responsible AI principles to real scenarios
  • Practice set: AI workloads scenario questions
  • Mini-review: domain summary and memory cues
Chapter quiz

1. A retail company wants to automatically route incoming customer emails into categories such as "billing", "returns", and "technical support". They have thousands of previously labeled emails for training. Which AI workload best fits this requirement?

Show answer
Correct answer: Classification (supervised machine learning)
This is a supervised learning scenario because the company has labeled examples and wants to assign one of several predefined labels to new emails (classification). Clustering is incorrect because it groups data without known labels and would not reliably map to specific business categories. Generative AI is incorrect because the goal is to label/rout existing text, not generate new content.

2. A manufacturer wants to predict next month’s energy consumption for each facility using historical meter readings and weather data. The output should be a numeric value (kWh). Which AI workload should you choose?

Show answer
Correct answer: Regression
Predicting a continuous numeric value is a regression workload. Object detection is incorrect because it applies to images and locating objects in them. Sentiment analysis is incorrect because it evaluates opinions in text, not numeric forecasting from time-series/tabular data.

3. A legal team has thousands of contracts (PDFs) and wants to extract key fields such as party names, effective date, and termination clause text to populate a database. Which Azure solution type best aligns with this workload?

Show answer
Correct answer: Document intelligence / information extraction
The scenario is about extracting structured information from documents, which maps to document processing/document intelligence (OCR + field extraction). A chatbot is incorrect because conversation is not required; the goal is structured extraction rather than interactive Q&A. Image generation is incorrect because the output is extracted text/fields from existing documents, not newly generated images.

4. A marketing team wants an assistant that can draft new product descriptions in the company’s style based on a short bullet list of features. Which approach is the best match?

Show answer
Correct answer: Use a generative AI model to generate text from a prompt
Drafting new descriptions is content creation, which is a generative AI workload (text generation). Key phrase extraction is incorrect because it analyzes existing text rather than producing new copy. Clustering is incorrect because grouping items does not generate marketing descriptions.

5. A bank deploys an AI system to assist with loan pre-screening. After a pilot, the bank discovers that applicants from a specific neighborhood are disproportionately rejected due to historical data patterns. Which Responsible AI principle should be prioritized to address this issue?

Show answer
Correct answer: Fairness
Disproportionate negative outcomes for a group indicate a fairness/bias problem, so the fairness principle should be prioritized (e.g., evaluate bias, rebalance data, adjust thresholds, and monitor outcomes). Reliability and safety is incorrect because the issue is not primarily about system stability or avoiding harmful operation failures. Privacy and security is incorrect because the scenario describes biased decisions, not unauthorized data access or data leakage.

Chapter 3: Fundamental Principles of ML on Azure

This chapter targets the AI-900 expectation that you can describe machine learning (ML) at a fundamentals level and connect those fundamentals to Azure services and common exam scenarios. Even though this course focuses on Copilot and Azure OpenAI, AI-900 still tests classic ML vocabulary and lifecycle concepts because they underpin how data-driven systems (including generative AI solutions) are built, evaluated, and deployed.

As you read, keep an exam mindset: AI-900 questions often provide a short scenario and then ask you to identify (1) the ML approach (supervised vs unsupervised), (2) the correct evaluation metric (classification vs regression), or (3) the correct Azure capability (training vs inference, workspace vs compute). The most common mistakes come from missing a single word cue in the prompt—terms like “predict a category,” “forecast a value,” “group similar items,” “deploy,” or “real-time scoring.”

We’ll first lock in core ML definitions (features, labels, training, inference), then connect them to approach selection and evaluation metrics, then map to Azure Machine Learning fundamentals (workspace, compute, pipelines, Automated ML, and Designer). You’ll finish with a practice-oriented mini-review of word cues and decision patterns you can apply under exam time pressure.

Practice note for Understand core ML concepts: features, labels, training, inference: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Choose ML approaches and metrics for typical AI-900 prompts: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Explain Azure Machine Learning capabilities at a fundamentals level: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Practice set: ML lifecycle and metrics questions: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Mini-review: key definitions and exam word cues: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Understand core ML concepts: features, labels, training, inference: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Choose ML approaches and metrics for typical AI-900 prompts: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Explain Azure Machine Learning capabilities at a fundamentals level: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Practice set: ML lifecycle and metrics questions: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Mini-review: key definitions and exam word cues: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Sections in this chapter
Section 3.1: ML basics: datasets, features/labels, training vs inference

Section 3.1: ML basics: datasets, features/labels, training vs inference

AI-900 expects you to use ML terminology precisely. A dataset is a collection of examples (rows) with attributes (columns). In supervised learning, each example has features (inputs) and a label (the output you want to predict). Features might be “age,” “income,” and “tenure”; a label might be “will churn” or “monthly sales.” If the prompt says “known outcomes,” “historical results,” or “labeled data,” that’s your cue that labels exist and supervised learning is likely.

Training is the process of fitting a model to data—learning patterns that map features to labels. Inference (also called scoring or prediction) is using the trained model to make predictions on new data. On the exam, these are commonly confused with deployment. Deployment is about packaging and hosting a model so inference can happen in an application; training is not deployment.

Exam Tip: When a scenario includes words like “build,” “train,” “fit,” “learn,” or “create a model,” think training. When you see “predict,” “classify,” “score,” “real-time,” “batch scoring,” or “during app usage,” think inference.

  • Common trap: Treating “test data” as the same as “new production data.” Test/validation datasets are used during model development to estimate performance; production inference happens after deployment.
  • Common trap: Assuming more features always improves accuracy. Extra features can add noise, leakage, or overfitting—AI-900 won’t ask feature engineering details, but it will test the concept that model quality depends on appropriate data, not just more data.

Finally, remember the broad ML lifecycle: define the problem, prepare data, train, evaluate, and deploy for inference. Questions may ask you to order steps; choose the option where evaluation occurs after training but before deployment.

Section 3.2: Supervised vs unsupervised learning and when to use each

Section 3.2: Supervised vs unsupervised learning and when to use each

Approach selection is a frequent AI-900 prompt pattern. Supervised learning uses labeled data to learn a mapping from features to a known target. It covers two big families: classification (predict a category) and regression (predict a number). If the question says “approve/deny,” “spam/not spam,” “diagnosis type,” or “which product category,” you are in classification territory. If it says “forecast demand,” “predict price,” “estimate time,” or “temperature,” that points to regression.

Unsupervised learning uses unlabeled data to find structure. The most tested unsupervised technique in fundamentals exams is clustering—grouping similar items. Word cues include “segment customers,” “group documents,” “find similar behaviors,” or “identify natural groupings.” Another unsupervised idea you might see is anomaly detection (finding unusual patterns), which can be framed as “find outliers” or “detect unusual transactions.”

Exam Tip: Look for the “what is the output?” clue. If the output is a known label from history, it’s supervised. If the output is a grouping or “discover patterns,” it’s unsupervised.

  • Common trap: Confusing clustering with classification. Clustering does not require pre-defined categories; classification does.
  • Common trap: Misreading “recommendations” as clustering. Recommendations can use many approaches; on AI-900, if the prompt emphasizes “similar users/items,” clustering is plausible, but if it emphasizes “predict rating/likelihood,” it can be supervised.

In Azure-centric wording, you may see “train a model to predict” (supervised) versus “analyze data to find segments” (unsupervised). Your job is to match the approach to the scenario, not to name a specific algorithm.

Section 3.3: Model evaluation basics: accuracy, precision/recall, RMSE, overfitting

Section 3.3: Model evaluation basics: accuracy, precision/recall, RMSE, overfitting

Evaluation metrics are chosen based on the task type and the cost of mistakes. For classification, accuracy (correct predictions / total predictions) is a basic metric, but it can be misleading when classes are imbalanced (for example, fraud detection where “fraud” is rare). In those cases, the exam often wants you to prefer precision and recall. Precision answers: “When the model predicts positive, how often is it correct?” Recall answers: “Of all actual positives, how many did we catch?”

Exam Tip: If false positives are expensive (flagging many legitimate transactions as fraud), precision matters. If false negatives are expensive (missing a fraudulent transaction or missing a disease), recall matters. Prompts sometimes describe the business impact; map that impact to the error type.

For regression, a common metric is RMSE (root mean squared error), which penalizes larger errors more than smaller ones. If the prompt says “predict a numeric value” and asks which metric to use, RMSE is a safe exam choice. You may also see MAE, but RMSE is the frequent fundamentals cue.

AI-900 also expects you to recognize overfitting: a model performs very well on training data but poorly on unseen data. The scenario clue is typically “high training accuracy but low test accuracy.” Overfitting is not solved by “train longer” by default; it’s often addressed by more data, simplifying the model, regularization, or better validation practices.

  • Common trap: Choosing accuracy for every classification scenario. If the prompt mentions rare events, imbalance, or critical misses, think precision/recall.
  • Common trap: Confusing evaluation with inference. Evaluation uses labeled validation/test data; inference is predicting on new, unlabeled production data.

When the exam asks how to improve generalization, the best fundamentals answer usually references addressing overfitting and validating on separate data, not tweaking a deployment setting.

Section 3.4: Azure Machine Learning concepts: workspaces, compute, pipelines (fundamentals)

Section 3.4: Azure Machine Learning concepts: workspaces, compute, pipelines (fundamentals)

At the fundamentals level, Azure Machine Learning (Azure ML) is the Azure service used to build, train, and manage ML solutions end-to-end. AI-900 won’t test deep configuration, but it will test what major components are for and how they relate to training and deployment.

A workspace is the top-level container for Azure ML resources: experiments, models, datasets, registered assets, and related configuration. If a prompt says “central place to manage models, runs, and artifacts,” the workspace is the answer. Compute refers to the resources used to run training or inference workloads—examples include compute instances (development) and compute clusters (scalable training). If a question asks what you scale to reduce training time, choose compute (often a cluster), not the workspace.

Pipelines are used to define and automate ML workflows as repeatable steps (data preparation, training, evaluation, and sometimes deployment). Pipelines help with reproducibility and operationalizing the ML lifecycle. On the exam, the cue is “repeatable workflow,” “orchestrate steps,” or “automate training process.”

Exam Tip: Map nouns to responsibilities: workspace = management boundary, compute = where it runs, pipeline = how steps are chained. This quick mapping helps you answer many “which component” questions without needing product depth.

  • Common trap: Thinking a workspace is the same as compute. A workspace is not where training runs; it organizes and tracks what happened.
  • Common trap: Assuming pipelines are only for deployment. Pipelines can include deployment steps, but their core value is orchestrating the whole ML workflow.

In a Copilot/Azure OpenAI world, Azure ML still matters for classic predictive models and for managing ML operations. The exam wants you to recognize where Azure ML fits when the solution is not purely generative—especially for forecasting, classification, and tabular ML scenarios.

Section 3.5: Automated ML and designer concepts (when and why)

Section 3.5: Automated ML and designer concepts (when and why)

Two Azure ML capabilities frequently tested at a high level are Automated ML (AutoML) and Designer. AutoML helps you automatically try multiple algorithms and hyperparameters to find a strong model for your dataset and task type (classification, regression, forecasting). The key exam idea: AutoML is for accelerating model selection and tuning when you know the target (label) and want a good baseline quickly.

Designer is a drag-and-drop visual interface for building ML workflows without writing much code. It’s about assembling data prep, training, and evaluation steps visually. If the prompt says “no-code,” “visual pipeline,” or “drag-and-drop,” Designer is the cue. If it says “automatically choose the best algorithm” or “optimize model,” AutoML is the cue.

Exam Tip: AutoML answers “Which model configuration is best?” Designer answers “How do I build the workflow visually?” Don’t swap them.

  • Common trap: Using AutoML for unlabeled data. AutoML assumes you have a defined target for supervised tasks; clustering is not the typical AutoML fundamentals scenario.
  • Common trap: Treating Designer as a separate service from Azure ML. Designer is a capability within Azure ML that produces pipelines and runs within the workspace.

Also remember that AutoML and Designer relate primarily to the training phase. The exam may include “deploy the best model” as a follow-on step, but AutoML/Designer are not synonymous with deployment; they help create and evaluate models that can later be deployed for inference.

Section 3.6: Exam-style practice: metrics selection, lifecycle ordering, scenario questions

Section 3.6: Exam-style practice: metrics selection, lifecycle ordering, scenario questions

This section consolidates the chapter into the patterns AI-900 repeatedly uses: choose the metric, choose the lifecycle step, and choose the ML approach. When you see a prompt, first classify it as classification/regression/clustering, then select the metric and Azure capability that match that type. The exam is less about math and more about correct matching.

Metrics selection cues: If the output is a category, think classification metrics. Accuracy is acceptable for balanced classes; precision/recall are better when the scenario describes imbalance or costly mistakes. If the output is a number, think regression metrics like RMSE. If the prompt is about grouping, metrics might not be central; instead, the right approach is clustering and the evaluation discussion is typically more qualitative at fundamentals level.

Lifecycle ordering cues: Most correct sequences include data preparation before training, training before evaluation, and evaluation before deployment. If an option evaluates before training, it’s wrong. If an option deploys before any evaluation, it’s usually wrong in an exam context unless the question explicitly says “quick proof of concept” (even then, evaluation is typically implied).

Azure component cues: “Track experiments, register models, manage artifacts” points to the workspace. “Run training at scale” points to compute clusters. “Automate a repeatable workflow” points to pipelines. “Automatically try algorithms/hyperparameters” points to AutoML. “No-code visual workflow” points to Designer.

Exam Tip: Under time pressure, underline the output noun in the scenario (category/number/group) and the operational verb (train/evaluate/deploy/predict). Those two highlights typically determine the correct answer set.

  • Common trap: Overthinking and searching for advanced algorithm names. AI-900 rarely requires naming specific algorithms; it rewards correct categorization and metric choice.
  • Common trap: Mixing “validation” with “production.” Validation/test sets are for evaluation; production data is for inference.

In your mini-review, memorize the word cues: “segment” → clustering; “forecast” → regression (RMSE); “classify”/“approve” → classification (accuracy/precision/recall); “high train, low test” → overfitting; “central management” → workspace; “scale training” → compute; “orchestrate steps” → pipeline; “best algorithm automatically” → AutoML; “drag-and-drop” → Designer.

Chapter milestones
  • Understand core ML concepts: features, labels, training, inference
  • Choose ML approaches and metrics for typical AI-900 prompts
  • Explain Azure Machine Learning capabilities at a fundamentals level
  • Practice set: ML lifecycle and metrics questions
  • Mini-review: key definitions and exam word cues
Chapter quiz

1. A retail company wants to predict whether an online transaction is fraudulent (Yes/No) based on features such as purchase amount, country, device type, and account age. Which machine learning approach should you choose?

Show answer
Correct answer: Supervised classification
This is a supervised learning scenario because historical transactions can be labeled as fraudulent or not fraudulent. The target is a category (Yes/No), so it is classification. Regression is incorrect because it predicts a continuous numeric value (for example, a dollar amount). Clustering is incorrect because it groups similar records without known labels and does not directly output a fraud/not-fraud decision.

2. A utility provider wants to forecast next month’s electricity consumption (kWh) for each household using past monthly usage, home size, and number of occupants. Which metric is most appropriate to evaluate the model?

Show answer
Correct answer: Mean Absolute Error (MAE)
The goal is to predict a continuous numeric value (kWh), which is a regression problem. Regression models are commonly evaluated with error-based metrics such as MAE or RMSE. Accuracy and F1 score are classification metrics, so they are not appropriate for evaluating a numeric forecast.

3. A marketing team has a dataset of customer purchase behavior but no labels. They want to group customers into segments with similar behavior patterns to target promotions. Which approach best fits this requirement?

Show answer
Correct answer: Unsupervised clustering
Customer segmentation with no predefined labels is an unsupervised learning scenario, and clustering is the standard approach to group similar records. Classification is incorrect because it requires labeled categories to learn from (for example, known segment labels). Regression is incorrect because it predicts a continuous value rather than forming groups.

4. You are preparing an ML solution in Azure. You want a central place to manage datasets, experiments, model training runs, and registered models for a project. Which Azure Machine Learning component should you use?

Show answer
Correct answer: Azure Machine Learning workspace
An Azure Machine Learning workspace is the top-level resource used to organize and manage assets such as data, experiments, runs, models, and deployments. A compute instance is primarily a development compute resource for authoring and running code, not the central management boundary. An inference endpoint is used for deployment/real-time scoring, not for managing the overall ML project lifecycle.

5. A company has trained a model and now needs to use it in production to return predictions for new data through a REST call. In ML terminology, which phase is this?

Show answer
Correct answer: Inference
Using a trained model to generate predictions on new data is inference (often called scoring), and in Azure this is commonly delivered via a deployed endpoint. Training is the phase where the algorithm learns model parameters from labeled (or structured) data. Feature engineering is the process of selecting/transforming input variables, which can occur before training and inference but is not the act of serving predictions.

Chapter 4: Computer Vision Workloads on Azure

In AI-900, “computer vision” means turning images and video into structured information that applications can act on. The exam does not test you on deep math or model architectures; it tests whether you can recognize a vision task, choose the right Azure AI service, and explain outputs at a high level. This chapter ties directly to the objective “Describe computer vision workloads on Azure and choose appropriate Azure AI services,” and it also supports the broader course outcomes around training vs. inference and service selection.

As you work through the lessons in this chapter, keep an exam mindset: the wording of the scenario almost always signals the workload. Words like “identify the main subject” or “categorize images” typically point to classification; “find and locate items” points to detection; “extract text” points to OCR; and “find per-pixel boundaries” points to segmentation. Exam Tip: On AI-900, you’re often graded on choosing the most appropriate managed service (Azure AI Vision, Azure AI Face, Azure AI Custom Vision, Azure AI Document Intelligence) rather than describing how to build a model from scratch.

You’ll also see a recurring theme: out-of-the-box (prebuilt) vs customized (trained) solutions. Many scenarios can be handled with prebuilt models using Azure AI services; customization becomes relevant when labels are domain-specific (e.g., defect types in manufacturing) or when you need to train on your own images. The lessons in this chapter guide you through identifying tasks, explaining detection/OCR/classification plainly, understanding customization and deployment at a high level, practicing service selection, and finishing with a quick vision cheat sheet.

Practice note for Identify vision tasks and pick the right Azure AI service: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Explain detection, OCR, and image classification in plain language: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Understand vision customization and model deployment at a high level: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Practice set: computer vision service selection questions: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Mini-review: vision cheat sheet and last-minute recalls: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Identify vision tasks and pick the right Azure AI service: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Explain detection, OCR, and image classification in plain language: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Understand vision customization and model deployment at a high level: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Sections in this chapter
Section 4.1: Describe computer vision workloads on Azure (official objective mapping)

Section 4.1: Describe computer vision workloads on Azure (official objective mapping)

The AI-900 objective expects you to recognize common vision workloads and map them to Azure services. At a foundational level, computer vision workloads include: describing what’s in an image (tags/captions), detecting objects and their locations (bounding boxes), reading text (OCR), identifying people-related attributes (face detection/verification), and analyzing documents/receipts/forms (document understanding).

For the exam, focus on “workload recognition” more than implementation details. You should be able to explain what the system outputs during inference: labels (classification), coordinates + labels (detection), pixel masks (segmentation), extracted text (OCR), or structured fields (document extraction). Exam Tip: When a prompt mentions “real-time camera feed,” “safety monitoring,” or “counting items,” it’s typically an inference scenario where you call an API endpoint—no training required unless the classes are unique to your business.

Also know the rough boundary between “image understanding” and “document understanding.” Images like photos of products, streets, or animals are usually handled by Azure AI Vision (or Custom Vision if you need custom labels). Documents like invoices, receipts, IDs, and forms often fit Azure AI Document Intelligence because the goal is structured field extraction rather than just text lines.

  • Classification: assign a category to an image (e.g., “damaged” vs “not damaged”).
  • Object detection: find instances and locations (e.g., “3 helmets detected,” with boxes).
  • Segmentation: identify the exact shape/area of objects (e.g., outline defects on a surface).
  • OCR: extract printed/handwritten text (e.g., serial numbers, signs).
  • Document extraction: extract key-value pairs and tables (e.g., invoice total, vendor name).

A common trap is overcomplicating the choice: AI-900 typically rewards selecting the managed AI service aligned to the workload rather than proposing Azure Machine Learning training pipelines unless the scenario explicitly asks to train and manage models end-to-end.

Section 4.2: Image classification vs object detection vs segmentation (exam framing)

Section 4.2: Image classification vs object detection vs segmentation (exam framing)

Classification, detection, and segmentation are easy to confuse under exam pressure because all three “recognize things in images.” The difference is the granularity of the answer. Classification returns “what it is” for the whole image (or a region you provide). Detection returns “what and where” using bounding boxes. Segmentation returns “what and exactly which pixels,” producing a mask for each object/region.

Use scenario keywords to decide quickly. If the question wants “categorize images in a folder” or “determine if an image contains a forklift,” classification is the match. If it wants “locate every forklift and count them,” detection is the match. If it wants “measure the area of corrosion” or “separate background from foreground precisely,” segmentation is the match.

Exam Tip: Bounding boxes imply detection. “Outline,” “mask,” “pixel-level,” or “exact shape” implies segmentation. If neither location nor shape is required, default to classification.

At a service level, Azure AI Vision provides general-purpose analysis and can perform tasks like detecting objects and generating tags/captions. Azure AI Custom Vision is commonly associated with training your own classification or object detection model when the built-in labels are not sufficient (for example, custom defect categories). A frequent exam trap is selecting Custom Vision when the scenario does not require custom labels; if a prebuilt capability can solve it, AI-900 often expects the prebuilt service selection.

  • Classification output: class label(s) with confidence scores.
  • Detection output: class label(s) + confidence + bounding box coordinates.
  • Segmentation output: class label(s) + confidence + mask/region information.

Finally, remember the high-level training vs inference distinction: classification/detection/segmentation can be prebuilt (no training by you) or custom (you label images, train, then deploy). On AI-900, you mainly need to articulate that training happens once (or periodically), while inference is the repeated act of scoring new images.

Section 4.3: OCR and document/image text extraction scenarios

Section 4.3: OCR and document/image text extraction scenarios

OCR (optical character recognition) turns text in an image into machine-readable text. AI-900 often frames OCR as “read text from photos,” “extract serial numbers,” “read street signs,” or “digitize scanned forms.” The key exam skill is distinguishing plain OCR from document understanding. OCR returns text lines/words and their positions; document understanding returns structured fields like totals, dates, and line items.

Azure AI Vision can perform OCR on images (including printed text in photos). When the scenario involves business documents—receipts, invoices, purchase orders, ID documents—Azure AI Document Intelligence is typically the better fit because it goes beyond raw text to extract structure, key-value pairs, and tables. Exam Tip: If the prompt mentions “fields,” “table extraction,” “key-value pairs,” “invoice total,” or “receipt merchant name,” that’s a Document Intelligence signal, not just OCR.

Another common trap is assuming OCR is the same as search. OCR is a vision workload (extract text), while search solutions require indexing and querying; the exam typically wants you to pick the vision service first, then optionally you could store results elsewhere for retrieval. Don’t jump straight to a search service if the core requirement is “read text.”

  • OCR best fits: signage translation prep, extracting a product code, reading a shipping label, capturing meter readings from a photo.
  • Document extraction best fits: automating invoice processing, parsing receipts into totals/taxes, extracting form fields into a database.

From an “interpret outputs” standpoint, OCR results commonly include the recognized text and positional data (bounding boxes for words/lines). The exam may describe output as “coordinates” or “regions” for text; that is still OCR. Document extraction outputs are more semantic: “InvoiceTotal: 123.45,” “VendorName: …,” often with confidence per field.

Section 4.4: Face-related capabilities and responsible use considerations (fundamentals)

Section 4.4: Face-related capabilities and responsible use considerations (fundamentals)

Face-related scenarios appear in AI-900 as “detect faces in an image,” “compare two faces to verify the same person,” or “group similar faces.” On Azure, these are typically associated with Azure AI Face (a dedicated service for face detection and analysis). At a fundamental level, face detection finds face regions; face verification compares whether two faces likely belong to the same person; identification matches a face against a known set (when allowed and configured).

The exam also expects responsible AI awareness. Face capabilities can be sensitive and may have usage constraints, approval requirements, or limitations depending on region and policy. You should know that responsible use includes consent, privacy, bias considerations, and using the technology only for appropriate scenarios. Exam Tip: If a question hints at surveillance, tracking individuals without consent, or sensitive attribute inference, the “best answer” often involves emphasizing responsible use, governance, and selecting compliant capabilities rather than blindly enabling face features.

A common trap is confusing “face detection” with “object detection.” Face detection is a specialized capability: it focuses on human faces and can return face rectangles and face-related metadata. Object detection is broader (cars, helmets, animals). If the scenario explicitly mentions “faces” or “verify identity,” choose Azure AI Face rather than generic vision object detection.

  • Face detection: find faces and return locations.
  • Face verification: compare two images (1:1) for similarity/identity likelihood.
  • Face identification: match against a gallery (1:many), subject to policy/feature availability.

Keep your explanation at the AI-900 level: you are not expected to describe embeddings in detail, but you should be able to say the service produces a numeric representation internally to compare similarity scores, and outputs a match/confidence-type result.

Section 4.5: Azure AI Vision and related service selection patterns

Section 4.5: Azure AI Vision and related service selection patterns

This section connects the “pick the right service” lesson with a repeatable decision pattern. On AI-900, the most common vision-related services you’ll see are Azure AI Vision (general image analysis, tagging, captions, OCR, object detection), Azure AI Custom Vision (train custom classifiers or detectors using your own labeled images), Azure AI Face (face-specific), and Azure AI Document Intelligence (structured document extraction). Your job is to map the scenario to the simplest service that meets requirements.

Start with the question: is it a photo-like image or a business document? If it’s a document with fields/tables, lean Document Intelligence. If it’s a photo and the task is “describe what’s in it,” lean Azure AI Vision. If it requires custom labels (like proprietary parts or niche defects), Custom Vision becomes the likely answer because you can upload images, label them, train, and publish an endpoint.

Exam Tip: The phrase “custom model” or “train with our own images” is a direct pointer to Custom Vision. If the requirement is generic (animals, common objects, captions), prebuilt Vision is usually correct.

At a high level, customization and deployment follow a predictable loop: collect representative images, label them, train the model, evaluate accuracy, then publish for inference. AI-900 expects you to recognize that training and evaluation occur before deployment, and that deployment exposes an endpoint used by apps. A trap is mixing up training location: Custom Vision is a managed training experience; Azure Machine Learning is the broader platform for building and managing custom ML, but it’s not the default answer for a straightforward “custom image classifier” scenario.

  • Azure AI Vision: prebuilt image analysis, captions/tags, OCR, common detection.
  • Azure AI Custom Vision: custom classification/detection with your labeled images.
  • Azure AI Document Intelligence: invoices/receipts/forms to structured data.
  • Azure AI Face: face detection/verification with responsible use.

When two services seem plausible, choose the one that best matches the output format requested. “Extract fields” beats “extract text.” “Train on our data” beats “use prebuilt.” “Verify person” beats “detect object.” This is how you avoid 50/50 errors on the exam.

Section 4.6: Exam-style practice: pick the service, interpret outputs, scenario matching

Section 4.6: Exam-style practice: pick the service, interpret outputs, scenario matching

To prepare for the practice set in this chapter, train yourself to do three actions quickly: (1) identify the vision task, (2) select the best-fit Azure AI service, and (3) interpret what the output would look like. AI-900 questions often hide the answer in nouns: “caption,” “tags,” “bounding box,” “mask,” “key-value,” “table,” “verify,” “gallery,” and “OCR.” Your goal is to match those nouns to the right workload and service.

When interpreting outputs, remember the exam’s level of detail: you should recognize that classification returns labels + confidence; detection returns labels + confidence + coordinates; OCR returns text + layout coordinates; document extraction returns fields/tables; and face verification returns a similarity/match result. Exam Tip: If a scenario says “we need coordinates,” ask: coordinates of what? Objects implies detection; words/lines implies OCR; invoice fields implies Document Intelligence.

Common scenario-matching traps include: choosing Custom Vision when the scenario does not mention custom labels; choosing OCR when the scenario really needs structured invoice fields; and choosing generic Vision for identity verification instead of Face. Another trap is ignoring the deployment hint. If the question mentions “publish an endpoint” after training, that’s a Custom Vision workflow; if it mentions “call a prebuilt API,” that’s typically Azure AI Vision or Document Intelligence prebuilt models.

  • Service selection heuristic: Photo understanding → Azure AI Vision; Custom categories → Custom Vision; Documents to fields → Document Intelligence; Face verification → Azure AI Face.
  • Output heuristic: Label-only → classification; Box → detection; Mask → segmentation; Text → OCR; Fields/tables → document extraction.

Finish your mini-review with a “vision cheat sheet” mindset: identify the keywords, map to task, map to service, and predict the output. If you can do that consistently, you’ll be ready for AI-900 computer vision questions even when the scenario is wrapped in business context (retail, manufacturing, healthcare, or security).

Chapter milestones
  • Identify vision tasks and pick the right Azure AI service
  • Explain detection, OCR, and image classification in plain language
  • Understand vision customization and model deployment at a high level
  • Practice set: computer vision service selection questions
  • Mini-review: vision cheat sheet and last-minute recalls
Chapter quiz

1. A retail company wants to process in-store camera images to locate and count products on shelves. The output must include bounding boxes for each detected product. Which computer vision workload best matches this requirement?

Show answer
Correct answer: Object detection
Object detection is used to find and locate items in an image and returns coordinates (for example, bounding boxes) for each instance. Image classification typically assigns a label to the whole image (or a single main subject) but doesn’t provide locations. OCR extracts printed or handwritten text and is not intended for locating physical objects like products.

2. A logistics company scans shipping labels and needs to extract the recipient name, street address, and tracking number from the label image. Which Azure AI service is the most appropriate choice for this scenario?

Show answer
Correct answer: Azure AI Document Intelligence
Azure AI Document Intelligence is designed for document-centric OCR and structured data extraction (forms, receipts, invoices, labels), which aligns with extracting fields like names and tracking numbers. Azure AI Vision can perform OCR, but it’s generally used for broader image analysis and basic text extraction rather than form-like field extraction. Azure AI Custom Vision is used to train custom classification or detection models and is not the primary service for reading and extracting structured label fields.

3. A wildlife organization has thousands of images from trail cameras and wants to automatically categorize each image into one of several animal species (for example, deer, bear, fox). The organization does not need object locations, only the category per image. Which vision task is being described?

Show answer
Correct answer: Image classification
Image classification assigns a category label to an image (or its main subject) and fits a scenario where you only need the species per photo. Object detection is used when you must find and locate animals with bounding boxes. Segmentation is used when you need per-pixel boundaries (for example, outlining the exact shape of an animal), which is beyond what this scenario asks for.

4. A manufacturer wants to identify specific types of defects on circuit boards that are unique to their production process. Off-the-shelf labels like "damaged" are not sufficient, and they want to train using their own labeled images and deploy the resulting model for predictions. Which Azure service best fits this requirement?

Show answer
Correct answer: Azure AI Custom Vision
Azure AI Custom Vision is used when you need to train a custom image classification or object detection model using your own labeled images and then deploy it for inference. Azure AI Vision focuses on prebuilt capabilities (image analysis, OCR, and related tasks) and is not primarily about training custom defect categories. Azure AI Face is specialized for face detection/analysis and is not appropriate for circuit board defect detection.

5. You are designing a solution and must choose the most appropriate managed Azure AI service. A customer wants to take photos of store signs and extract the visible text so it can be translated and displayed to users. Which service should you recommend?

Show answer
Correct answer: Azure AI Vision
Azure AI Vision is the best fit for extracting text from general images (OCR on photos of signs) as part of a broader computer vision workload. Azure AI Document Intelligence is optimized for document/form processing where you typically want structured field extraction from documents (invoices, receipts, forms), which is not the main requirement here. Azure AI Custom Vision is for training custom classifiers/detectors and does not inherently perform OCR for sign text extraction.

Chapter 5: NLP + Generative AI Workloads on Azure (Copilot & Azure OpenAI)

This chapter maps directly to the AI-900 outcomes around NLP workloads and generative AI workloads, with special attention to the exam’s “choose the right service” style questions. Expect the test to probe whether you can distinguish classic NLP (extract, classify, summarize) from GenAI (generate, transform, chat), and whether you understand the deployment and safety vocabulary used with Azure OpenAI and Copilots. A common exam pattern is a short scenario (customer support emails, internal policy search, meeting notes) followed by “which Azure service fits best?” Your job is to recognize the workload type, the required output (labels vs free-form text), and the risk constraints (privacy, hallucinations, safety).

As you read, keep a mental decision tree: (1) Is the task deterministic extraction/analysis? Think Azure AI Language features (Text Analytics). (2) Is it conversation/knowledge base? Think bot concepts and Q&A patterns. (3) Is it generation, rewriting, or code/content creation? Think Azure OpenAI. (4) Does the business want GenAI embedded into everyday tools (M365, Power Platform, Dynamics)? Think Copilot and governance.

Practice note for Understand NLP tasks and Azure service choices for text and conversation: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Explain generative AI concepts: tokens, embeddings, grounding, retrieval: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Use Azure OpenAI and Copilot concepts in exam language (no coding required): document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Practice set: NLP + GenAI + responsible AI questions: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Mini-review: final domain map and cross-domain comparisons: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Understand NLP tasks and Azure service choices for text and conversation: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Explain generative AI concepts: tokens, embeddings, grounding, retrieval: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Use Azure OpenAI and Copilot concepts in exam language (no coding required): document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Practice set: NLP + GenAI + responsible AI questions: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Mini-review: final domain map and cross-domain comparisons: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Sections in this chapter
Section 5.1: Describe NLP workloads on Azure (text analytics, extraction, sentiment)

Section 5.1: Describe NLP workloads on Azure (text analytics, extraction, sentiment)

AI-900 frequently tests “NLP workloads” as analysis of text rather than generation. In Azure, these are commonly delivered through Azure AI Language capabilities (often referred to in exam prep as “Text Analytics” features). The scenario clues are words like detect, classify, extract, score, or identify—outputs that are structured (labels, entities, key phrases) rather than open-ended prose.

Core workload types you should recognize: sentiment analysis (positive/neutral/negative scoring), key phrase extraction (main topics from text), entity recognition (people, organizations, locations, dates, product names), and language detection. You may also see “PII detection” as a compliance-oriented extraction task (find sensitive data in documents). These fit classic NLP because they turn unstructured text into structured insights that you can store, filter, and trend over time.

Exam Tip: If the question asks for “insights from text” and expects a list of entities, key phrases, or sentiment scores, choose Azure AI Language/Text Analytics-type options—not Azure OpenAI. GenAI can do these tasks, but the exam expects you to pick the purpose-built analytics service when the output is structured and repeatable.

Common trap: confusing OCR and text analytics. OCR (extracting text from images) is a computer vision/document workload, while sentiment/entity extraction is NLP. Another trap is assuming “summarization” always means GenAI. Azure AI Language can provide summarization features; the exam will often reward you for selecting an NLP summarization capability when the requirement is straightforward summarizing rather than creative generation.

How to identify the correct service choice: look for the input type (plain text), the output type (scores/fields), and the operational need (batch processing, dashboards, compliance). If the scenario mentions reporting, trends, or KPI monitoring from customer feedback, sentiment + key phrases + entities is usually the right mental bundle.

Section 5.2: Conversational AI fundamentals: bots, intents, entities, Q&A patterns

Section 5.2: Conversational AI fundamentals: bots, intents, entities, Q&A patterns

Conversational AI on AI-900 is less about building bots and more about recognizing the building blocks. A “bot” is the interaction layer (channels like web chat, Teams, etc.). Underneath, the system needs to understand what the user wants (intent) and pull out important details (entities). For example, in “Book a flight to Seattle tomorrow,” the intent might be BookFlight and the entities include destination=Seattle and date=tomorrow.

Many exam questions focus on Q&A patterns (sometimes described as knowledge base or FAQ style). Here, the user asks a question and the system retrieves the best-matching answer from curated content. The key distinction: Q&A is primarily retrieval (find an existing answer), while generative chat is primarily generation (compose new text). In enterprise scenarios, Q&A is popular because it is more controllable and easier to validate than free-form generation.

Exam Tip: If the requirement says “respond with answers from our approved documentation” or “use an FAQ,” think Q&A retrieval patterns. If it says “draft a new response” or “converse naturally across many topics,” that leans toward GenAI chat (often via Azure OpenAI) or a hybrid approach.

Common trap: assuming a conversational interface always implies GenAI. The exam can present a chatbot scenario where the best answer is still classic intent/entity extraction and a deterministic workflow (e.g., route a ticket, look up an order status). Another trap: ignoring escalation and fallback. When intent confidence is low or no answer is found, the correct design is often to hand off to a human or request clarification—this is both a user experience and responsible AI expectation.

To select correct answers, watch for keywords: “intents/entities” suggests NLU-style conversational AI fundamentals; “knowledge base/FAQ” suggests Q&A retrieval; “grounded responses” suggests retrieval plus generation (RAG), which you’ll revisit in Section 5.6.

Section 5.3: Generative AI workloads on Azure: use cases, benefits, limitations

Section 5.3: Generative AI workloads on Azure: use cases, benefits, limitations

Generative AI workloads are defined by producing new content: text, summaries, explanations, translations, and sometimes code or structured outputs. On the exam, the scenario cues are verbs like draft, rewrite, brainstorm, compose, summarize with a specific tone, or generate variants. These are ideal fits for large language models (LLMs) accessed through Azure OpenAI or productized through Copilot experiences.

Benefits the exam expects you to articulate: improved productivity (first drafts, faster analysis), natural language interfaces (users ask questions instead of navigating complex apps), and adaptability (the same model can handle multiple language tasks via prompting). Generative AI also supports “transformations,” like converting bullet points into a formal email or turning a meeting transcript into action items.

Limitations are heavily tested, especially under the “responsible AI” lens. LLMs can hallucinate (produce plausible but incorrect statements), amplify bias present in training data, and leak sensitive information if prompts or outputs are mishandled. They also require careful evaluation: you don’t “prove correctness” the way you might for deterministic extraction; you validate quality, safety, and grounding.

Exam Tip: When a scenario requires high factual accuracy tied to enterprise data (policies, inventory, customer records), assume you need grounding (connecting the model to trusted sources) rather than asking the model to “just know” the answer. This points toward retrieval-augmented generation concepts discussed later.

Common trap: treating GenAI as a replacement for all NLP. If the question needs consistent labels, compliance scoring, or entity extraction at scale, classic NLP services are often the best match. GenAI excels when the output is flexible language, but it increases governance and validation needs.

Section 5.4: Azure OpenAI fundamentals: models, prompts, temperature/top-p (conceptual)

Section 5.4: Azure OpenAI fundamentals: models, prompts, temperature/top-p (conceptual)

AI-900 uses Azure OpenAI terminology in a conceptual way: model families, prompting, and key generation controls. You are not expected to code, but you are expected to understand what terms mean and why they matter in solution selection and responsible deployment.

Tokens are chunks of text used by the model for input and output. Token limits affect how much context you can send (prompt + conversation history + retrieved documents) and how long the response can be. Embeddings are numeric vector representations of text used to measure similarity; they power semantic search and retrieval patterns that “ground” answers. The exam often frames embeddings as the foundation for “find related documents” and “search by meaning, not keywords.”

Prompts are the instructions and context you provide. Good prompts specify role, task, constraints, and desired format. On the exam, prompt engineering is usually assessed through outcomes: does the prompt reduce ambiguity, require citations, or constrain the model to provided content?

Generation controls: temperature and top-p both influence randomness/creativity. Higher temperature (or broader top-p) generally yields more varied outputs; lower values tend to be more deterministic and repeatable. In regulated or customer-facing scenarios (policy answers, support guidance), the safer default is lower creativity to reduce unexpected phrasing and errors.

Exam Tip: If you see “consistent answers” or “reduce variability,” choose lower temperature/top-p. If you see “brainstorm ideas” or “creative copy,” choose higher. The trap is thinking higher creativity improves factuality—it usually does the opposite.

Finally, remember “grounding” vocabulary: the model can be instructed to rely on provided context, but without retrieval it may still hallucinate. This is why Azure OpenAI solutions commonly pair with search/retrieval to anchor responses to enterprise content.

Section 5.5: Copilot concepts: where it fits, productivity patterns, governance basics

Section 5.5: Copilot concepts: where it fits, productivity patterns, governance basics

Copilot is the exam’s way of describing productized generative AI assistants embedded into tools users already work in. Rather than building an app from scratch, organizations adopt Copilot experiences to accelerate common tasks: summarizing meetings, drafting emails, generating presentations, answering questions over business content, and automating workflows. On AI-900, you’re tested on recognizing when “use a Copilot” is the most appropriate answer versus “build a custom Azure OpenAI solution.”

Productivity patterns to remember: (1) Draft and refine (first draft → edit with constraints), (2) Summarize and extract actions (notes → decisions, tasks), (3) Ask-and-answer over organizational knowledge (policies, projects), and (4) Automate routine work (connect to business processes through low-code tooling). Copilot is often described as a natural-language front end that orchestrates actions across apps while respecting identity and permissions.

Governance basics appear as exam-friendly terms: data protection, access control, auditability, and responsible AI safeguards. You should assume enterprises require controls around what data is used, who can access it, and how outputs are monitored. Even without deep implementation detail, the exam expects you to prefer solutions that align with organizational compliance, privacy, and safety requirements.

Exam Tip: When the scenario is “employees want AI inside Microsoft productivity tools” (email, documents, meetings) and emphasizes quick adoption, Copilot is often the intended answer. When the scenario is a unique customer-facing app with custom workflows and domain data integration, Azure OpenAI is often the intended answer.

Common trap: treating Copilot as “just a chatbot.” Copilot is typically positioned as an assistant integrated into workflows with governance and enterprise identity; the exam will reward you for linking Copilot to productivity and organizational controls rather than generic chat.

Section 5.6: Exam-style practice: prompt engineering basics, RAG/grounding, safe deployment

Section 5.6: Exam-style practice: prompt engineering basics, RAG/grounding, safe deployment

This section ties together what AI-900 tests most: can you describe how to get useful, safer outcomes from GenAI without overstating its capabilities. Three recurring concepts appear in exam language: prompt engineering, retrieval-augmented generation (RAG), and responsible deployment.

Prompt engineering basics: specify the task, the audience, the constraints, and the output format. Strong prompts reduce ambiguity (which reduces hallucinations) and make evaluation easier (because outputs are consistent). Look for “must use bullet points,” “must cite provided policy,” or “if you don’t know, say you don’t know” as exam indicators of good prompting. A common trap is prompts that ask for “the best answer” without giving context—these invite the model to fill gaps with guesses.

RAG/grounding: grounding means anchoring responses to trusted content. RAG is the common pattern: (1) create embeddings for documents, (2) retrieve the most relevant passages by similarity, (3) provide those passages to the model as context, and (4) generate an answer that references that context. On the exam, the right choice often appears when the scenario says “use our internal documentation,” “avoid hallucinations,” “keep answers up to date,” or “don’t retrain the model.” RAG is preferred when data changes frequently because you update the source documents and index rather than retrain the LLM.

Safe deployment: the exam expects you to acknowledge content safety, privacy, and monitoring. You should be ready to describe safeguards such as filtering harmful content, limiting sensitive data in prompts, applying least privilege access to data sources, and adding human review for high-impact use cases. Also remember that evaluation is ongoing—prompt changes, new documents, or new user behaviors can change risk.

Exam Tip: If a question contrasts “fine-tune the model” vs “use grounding/retrieval,” the exam commonly favors grounding/RAG for enterprise knowledge scenarios because it’s faster, cheaper, and keeps content current. Fine-tuning is more appropriate for style/format consistency or specialized behavior—not as a primary method for injecting private facts.

Final mini-review: Classic NLP (sentiment, entities) produces structured insights; conversational patterns (intents, Q&A) focus on routing or retrieving known answers; GenAI (Azure OpenAI) generates and transforms language; Copilot applies GenAI in productivity contexts with governance. Many real solutions combine them—your exam task is to identify which component is doing the “main job” in the scenario and select the service accordingly.

Chapter milestones
  • Understand NLP tasks and Azure service choices for text and conversation
  • Explain generative AI concepts: tokens, embeddings, grounding, retrieval
  • Use Azure OpenAI and Copilot concepts in exam language (no coding required)
  • Practice set: NLP + GenAI + responsible AI questions
  • Mini-review: final domain map and cross-domain comparisons
Chapter quiz

1. A support team receives thousands of customer emails per day. They need to automatically detect the customer’s sentiment and extract key phrases from each email. The output must be consistent and explainable (not free-form generation). Which Azure service is the best fit?

Show answer
Correct answer: Azure AI Language (Text Analytics features such as sentiment analysis and key phrase extraction)
Azure AI Language provides classic NLP analysis capabilities like sentiment analysis and key phrase extraction with deterministic, structured outputs suited for classification/extraction workloads. Azure OpenAI is optimized for generative tasks (creating or rewriting text) and is more prone to variability/hallucination when you only need extracted fields. Microsoft Copilot for Microsoft 365 is an end-user productivity experience embedded in M365 apps, not the primary service for building an email-processing NLP pipeline.

2. A company wants an internal chat experience that answers employee questions using the company’s HR policy documents. They want to reduce hallucinations by ensuring responses are grounded in retrieved content from their documents. Which concept best describes this approach?

Show answer
Correct answer: Grounding with retrieval (RAG) to provide source content to the model
Retrieval-augmented generation (RAG) and grounding describe the pattern of retrieving relevant enterprise content and providing it to the model so responses are based on that data, reducing hallucinations. Increasing temperature generally increases randomness/creativity and can worsen factual consistency. Tokens relate to how text is chunked/processed and to context/response limits, but token settings alone do not ensure the model uses authoritative HR documents.

3. You are designing a solution to find semantically similar documents (for example, policies related to “parental leave” even when the phrase isn’t used). Which generative AI concept is most directly used to enable semantic similarity search?

Show answer
Correct answer: Embeddings
Embeddings represent text as vectors in a way that captures semantic meaning, enabling similarity search and retrieval even when wording differs. Tokens are the units the model processes and are mainly relevant to context length and cost/limits, not semantic search by themselves. Content filtering is a responsible AI/safety control to detect and block certain categories of harmful content, not a mechanism for semantic similarity.

4. A business user asks for generative AI capabilities inside Outlook and Teams to summarize long email threads and meetings, and to draft responses using organizational context and governance controls. They want a product experience rather than building custom apps. Which option best fits?

Show answer
Correct answer: Microsoft Copilot for Microsoft 365
Microsoft Copilot for Microsoft 365 is designed to bring generative AI into M365 apps like Outlook and Teams with enterprise controls and user-facing workflows. Azure AI Language focuses on classic NLP tasks (extract/classify/analyze) rather than drafting and summarizing across M365 experiences. Azure OpenAI is the platform to build custom generative solutions, but the scenario explicitly asks for an embedded product experience rather than developing a bespoke application.

5. A company is deploying an Azure OpenAI chat solution for employees. They must reduce the risk of harmful or inappropriate outputs and enforce safety policies. Which capability best addresses this requirement?

Show answer
Correct answer: Azure OpenAI content filtering / safety system to detect and block unsafe content
Azure OpenAI includes safety features (often referred to as content filtering) to help detect and restrict certain categories of harmful content in prompts and completions—this aligns to responsible AI exam objectives. Embeddings are used for semantic search/retrieval and do not enforce output safety policies. Key phrase extraction is an Azure AI Language NLP feature for extracting terms from text; it does not control or constrain generative model outputs.

Chapter 6: Full Mock Exam and Final Review

This chapter is your consolidation and execution plan. AI-900 rewards clear thinking more than memorization: identify the workload, map it to the correct Azure AI capability, and avoid distractors that sound “AI-ish” but don’t solve the scenario. You’ll complete two mixed-domain mock blocks, then run a structured answer review that trains you to eliminate distractors quickly. Finally, you’ll build a weak-spot remediation plan aligned to official AI-900 outcomes—AI workloads, machine learning fundamentals, computer vision, NLP, and generative AI (Azure OpenAI + Copilot + responsible AI)—and lock in an exam-day checklist that reduces avoidable mistakes.

As you go, keep a single goal: convert knowledge into repeatable decisions under time pressure. The mock exam is not just a score; it’s a diagnostic tool. Your final review is not re-reading; it’s targeted remediation and confidence building through deliberate practice.

Practice note for Mock Exam Part 1: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Mock Exam Part 2: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Weak Spot Analysis: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Exam Day Checklist: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Final rapid review and confidence pass plan: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Mock Exam Part 1: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Mock Exam Part 2: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Weak Spot Analysis: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Exam Day Checklist: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Final rapid review and confidence pass plan: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Mock Exam Part 1: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Practice note for Mock Exam Part 2: document your objective, define a measurable success check, and run a small experiment before scaling. Capture what changed, why it changed, and what you would test next. This discipline improves reliability and makes your learning transferable to future projects.

Sections in this chapter
Section 6.1: Mock exam instructions and timing strategy (AI-900 pacing)

Section 6.1: Mock exam instructions and timing strategy (AI-900 pacing)

Treat the mock exam like the real one: quiet environment, single sitting, no notes, and a hard time cap. Your objective is to practice decision-making with enough speed to leave room for review. AI-900 questions are typically short, but the traps are subtle: similar service names, overlapping capabilities (for example, “search” vs “language understanding”), and scenario clues that imply constraints like real-time inference, data residency, or responsible AI requirements.

Use a three-pass pacing method. Pass 1 is “confident picks”: answer everything you can in under a minute and flag anything that requires rereading. Pass 2 is “scenario decoding”: return to flagged items and slow down enough to extract the workload type (vision/NLP/generative/ML) and the key verbs (classify, extract, generate, chat, detect objects). Pass 3 is “risk management”: only change answers when you can name the exact clue that proves the new choice is correct.

Exam Tip: In AI-900, the fastest way to get unstuck is to restate the problem as a workload label. If you can say “this is OCR,” “this is sentiment analysis,” or “this is text generation with safety requirements,” the correct service family usually becomes obvious.

  • Pass 1 target: complete the whole set quickly while flagging uncertainty.
  • Pass 2 target: resolve flags using keywords and service capability boundaries.
  • Pass 3 target: verify no misreads (negations like “NOT,” “best,” “most cost-effective”).

Common pacing trap: overinvesting early in one ambiguous question. Flag it and move on. Many later questions jog memory and reduce uncertainty when you return.

Section 6.2: Mock Exam Part 1: mixed-domain questions and scenarios

Section 6.2: Mock Exam Part 1: mixed-domain questions and scenarios

Part 1 is designed to mix AI fundamentals with practical scenario identification. Expect rapid context switches: a question may start with “predict” (machine learning) and then shift to “extract key phrases” (NLP), or mention “images from a factory line” (computer vision) alongside “generate a report” (generative AI). Your job is to separate what is being asked from what is merely background.

AI-900 frequently tests the difference between training and inference. In training, you are building a model (choosing algorithms, evaluating metrics). In inference, you are using a trained model to score new data. If a scenario describes deploying an endpoint, scoring new transactions, or running predictions in production, it’s inference. If it describes splitting data, evaluating accuracy, or reducing overfitting, it’s training/evaluation.

Exam Tip: When you see “evaluate,” “validate,” “test set,” “precision/recall,” or “confusion matrix,” you are in model evaluation territory—not just “using AI.” Questions often reward knowing which metric fits classification vs regression.

Also watch for foundational workload identification: classification vs regression vs clustering. Classification predicts a category (fraud/not fraud). Regression predicts a number (demand forecast). Clustering groups unlabeled items (customer segmentation). A common distractor is presenting regression-like language (“predict”) but the output is clearly categorical.

In mixed scenarios, prioritize the primary objective. If the goal is to extract text from receipts, OCR is the core workload; any downstream summarization is secondary. If the goal is to create marketing copy from a product brief, that is generative AI; sentiment analysis may be an add-on but not the main service selection.

Section 6.3: Mock Exam Part 2: mixed-domain questions and service selection

Section 6.3: Mock Exam Part 2: mixed-domain questions and service selection

Part 2 leans harder into Azure service selection, including generative AI. The exam expects you to recognize which Azure AI service matches a workload without overengineering. For vision tasks, distinguish between image analysis (labels/tags), object detection, face-related capabilities, and OCR/document extraction. For NLP, separate text analytics (sentiment, entities, language detection) from conversational AI patterns. For generative AI, focus on Azure OpenAI use cases (text generation, summarization, chat, embeddings) and how Copilot fits as an application layer experience.

Service selection traps often use near-synonyms. “Translate text” indicates translation capabilities (language services), not generative AI. “Extract key phrases” is text analytics, not Q&A chat. “Search a knowledge base” may hint at Azure AI Search plus embeddings; however, if the question is only about understanding intent and entities from user utterances, it’s closer to conversational language understanding. Use the action verb and input/output types to decide.

Exam Tip: For generative AI questions, look for words like “generate,” “draft,” “summarize,” “rewrite,” “chat,” or “create embeddings.” If the requirement includes grounding on enterprise data, think retrieval augmentation patterns (often Azure AI Search + Azure OpenAI embeddings) rather than “fine-tune” as a default.

Copilot concepts are tested as “what it is” and “what it is not.” Copilot is typically an integrated assistant experience built on LLMs and orchestration (prompts, plugins/connectors, grounding, and safety). The trap is assuming Copilot equals “a model you train.” In most exam framing, you configure and govern; you do not train the underlying foundation model. Likewise, responsible AI shows up as requirements: content filtering, privacy, human oversight, transparency, and avoiding harmful outputs. If a scenario emphasizes safety constraints, choose options that mention governance controls rather than “increase creativity” or “remove filters.”

Finally, keep an eye on “best choice” questions that implicitly prefer managed services over custom builds. AI-900 is fundamentals: select the simplest managed Azure AI capability that meets the need.

Section 6.4: Answer review method: why the distractors are wrong

Section 6.4: Answer review method: why the distractors are wrong

Your score improves fastest when you review answers like an examiner. Don’t just note the correct option—write a one-sentence justification and a one-sentence rejection for each distractor. This builds “discrimination skill,” which is what the test actually measures: can you tell similar concepts apart under time pressure?

Use a consistent four-step review method. Step 1: identify the workload (ML, vision, NLP, generative). Step 2: identify the task subtype (classification vs regression; OCR vs object detection; sentiment vs entity extraction; generation vs embeddings). Step 3: identify constraints (real-time inference, structured vs unstructured input, safety/governance). Step 4: verify the service mapping aligns with those constraints.

Exam Tip: If you can’t explain why three options are wrong, you don’t truly “own” the concept yet. In the real exam, uncertainty usually comes from not being able to eliminate distractors confidently.

Common distractor patterns include:

  • Overbroad tool: A general “AI” answer that doesn’t match the specific task (e.g., choosing a chatbot solution for simple sentiment scoring).
  • Wrong modality: Picking a text service for an image problem, or vice versa, because the scenario mentions both in passing.
  • Wrong lifecycle phase: Confusing training tasks (data splitting, feature engineering) with inference/deployment tasks (endpoint scoring).
  • Unnecessary complexity: Selecting custom ML when a prebuilt cognitive capability is implied by the requirement.

When you miss a question, label the miss type (misread, concept gap, service confusion, or overthinking). Your remediation plan in the next section should target the miss type, not just the topic.

Section 6.5: Weak-spot remediation plan mapped to official domains

Section 6.5: Weak-spot remediation plan mapped to official domains

After both mock parts, categorize every miss and every “guessed but correct” item. Guesses are hidden weaknesses: they tend to fail under exam stress. Map them to the course outcomes (which align to exam domains) and create a short remediation loop for each.

Domain 1: AI workloads and solution scenarios. Remediate by drilling workload keywords and outputs. Practice converting a scenario into: input type → task verb → output type. Trap: “AI” vs “automation.” Not every automation needs AI; the exam may bait you with simple rule-based cases.

Domain 2: ML fundamentals on Azure (training vs inference, evaluation). Remediate by reviewing metric-to-task mapping: accuracy/precision/recall for classification; MAE/MSE/RMSE for regression; overfitting signals and train/validation/test roles. Trap: thinking higher accuracy always wins—imbalanced datasets often require precision/recall focus.

Domain 3: Computer vision workloads. Remediate by separating OCR/document extraction from image classification and object detection. Trap: “detect text” vs “analyze image.” If the output is extracted strings and fields, it’s OCR/document intelligence style thinking.

Domain 4: NLP workloads (text analytics and conversational concepts). Remediate by practicing the boundaries: sentiment/key phrases/entities/language detection vs intent-based conversations. Trap: equating “chat” with “NLP service selection”—sometimes the scenario only needs analytics, not a bot.

Domain 5: Generative AI on Azure (Azure OpenAI, responsible AI, Copilot). Remediate by outlining: generation vs embeddings vs retrieval augmentation; and responsible AI controls (content filtering, monitoring, human oversight, privacy). Trap: defaulting to fine-tuning when grounding via retrieval is safer and faster for most enterprise Q&A use cases.

Keep remediation tight: 20–30 minutes per weak domain, then re-attempt similar items or review notes until you can reject distractors confidently.

Section 6.6: Exam day checklist: environment, time, review passes, and last-minute tips

Section 6.6: Exam day checklist: environment, time, review passes, and last-minute tips

On exam day, your goal is to protect focus and reduce preventable errors. Start with environment control: reliable network, quiet space, and all required identification if proctored. Close unnecessary apps, silence notifications, and ensure your display settings are comfortable for reading—fatigue leads to misreading “best” vs “least” questions.

Use the same three-pass strategy you practiced. Pass 1: answer fast and flag. Pass 2: resolve flags by mapping to workload and service boundaries. Pass 3: audit for misreads, especially negatives (“NOT”), superlatives (“most appropriate”), and scope (“which service,” “which feature,” “which principle”).

Exam Tip: Do not change an answer in the final minutes unless you can point to a specific requirement word in the prompt that contradicts your original choice. “A feeling” is not evidence.

  • Time management: Keep a steady rhythm; avoid deep dives early.
  • Confidence pass plan: If you finish early, re-check only flagged questions first, then scan for misreads.
  • Service sanity check: Ask: does this option match the input modality and the requested output?
  • Responsible AI reminder: If the scenario mentions safety, governance, privacy, or harmful content, prioritize controls and managed safeguards.

Last-minute review should be lightweight: revisit your personal “trap list” (common confusions you identified during weak-spot analysis) and the high-frequency mappings: classification/regression/clustering; training vs inference; OCR vs image analysis; sentiment/entities vs conversational intent; Azure OpenAI generation/embeddings and Copilot as an experience layer with responsible AI guardrails. Walk in with a plan, execute your passes, and let disciplined elimination do the work.

Chapter milestones
  • Mock Exam Part 1
  • Mock Exam Part 2
  • Weak Spot Analysis
  • Exam Day Checklist
  • Final rapid review and confidence pass plan
Chapter quiz

1. You are reviewing results from a full AI-900 mock exam. A user story reads: “We need to extract key phrases and detect the language from customer support tickets.” Which Azure AI capability is the best fit for this workload?

Show answer
Correct answer: Azure AI Language (Text Analytics features such as language detection and key phrase extraction)
Language detection and key phrase extraction are NLP tasks covered under Azure AI Language. Azure AI Vision targets image/video scenarios (OCR, tagging) and does not perform text key phrase extraction from raw text. Azure AI Search helps with indexing and querying content, but it does not natively perform language detection/key phrase extraction as the core workload; it typically complements NLP services rather than replacing them.

2. A company wants a chatbot that can draft answers using the company’s policy documents. They want to reduce hallucinations by grounding responses in their own content. Which approach best matches Azure OpenAI recommended patterns for this scenario?

Show answer
Correct answer: Use Retrieval-Augmented Generation (RAG): retrieve relevant passages (e.g., via Azure AI Search) and provide them as context to an Azure OpenAI model
RAG is the standard approach: retrieve relevant enterprise content and pass it to the model as context to improve factuality and relevance. Training a new foundation model from scratch is typically unnecessary, expensive, and not aligned with common Azure OpenAI usage for enterprise Q&A. Azure AI Vision can help extract text from scanned images (OCR), but it is not the mechanism for generating grounded conversational answers.

3. During weak-spot analysis, you notice you often confuse classification and regression questions. Which scenario describes a classification workload?

Show answer
Correct answer: Predict whether a credit card transaction is fraudulent (yes/no)
Fraud detection framed as yes/no is a classification task (discrete categories). Predicting electricity consumption is regression (continuous numeric output). Summarization is a generative AI/NLP task using a language model, not a supervised learning classification/regression outcome.

4. You’re applying an exam-day checklist strategy. A question asks: “Select the Azure service that detects objects in images and returns bounding boxes.” Which service should you select?

Show answer
Correct answer: Azure AI Vision
Object detection with bounding boxes is a computer vision workload supported by Azure AI Vision. Azure AI Language focuses on text-based NLP (entities, sentiment, key phrases). Azure OpenAI is for generative tasks (chat, summarization, code, embeddings) and is not the primary service for object detection in images.

5. Your final rapid review focuses on responsible AI. A team is deploying a generative AI assistant using Azure OpenAI and wants to reduce the risk of generating harmful content. What is the most appropriate built-in capability to apply?

Show answer
Correct answer: Use Azure OpenAI content filtering/safety system to detect and block disallowed content categories
Azure OpenAI includes safety/content filtering designed to help detect and mitigate harmful content, aligning with responsible AI expectations on the exam. Disabling encryption is unrelated to harmful content risk and is a security regression. Increasing temperature typically increases randomness/creativity, which can make outputs less predictable and may increase risk rather than mitigate it.
More Courses
Edu AI Last
AI Course Assistant
Hi! I'm your AI tutor for this course. Ask me anything — from concept explanations to hands-on examples.